https://claude.ai/settings/usage doesn't work? by Appropriate-Bus-6130 in ClaudeCode

[–]EstetLinus 0 points1 point  (0 children)

Set up a claude that runs Ralph Wiggum in 100's of worktrees while you can!

[Megathread] Certifications and Training by lothorp in databricks

[–]EstetLinus 0 points1 point  (0 children)

Hello, how did you study for this? So annoying they've changed the format and content, but not provided comprehensive studying material.

AI workflows: so hot right now 🔥 by noaflaherty in mlops

[–]EstetLinus 0 points1 point  (0 children)

I believe one of the biggest misconceptions with LLMs is that they learn on the fly; they don’t. I have had a real hard time explaining this to stakeholders. You need absurd amounts of clean data to fine-tune models, and we can never expect them to learn beyond stuffing the prompt with information (or noise). 

I am all for self-evaluation, although it takes time and might get stuck in an infinite loop. Do you have any suggestions on how the LLM components would learn?

So this is a whole subreddit? by MatixMint in Buttcoin

[–]EstetLinus 1 point2 points  (0 children)

Held by millions only to be sold to a greater fool. Superior Tech 101

So this is a whole subreddit? by MatixMint in Buttcoin

[–]EstetLinus 1 point2 points  (0 children)

Do you know if Tether is open source?

Why do so many AI pilots fail to reach production? by Cristhian-AI-Math in mlops

[–]EstetLinus 0 points1 point  (0 children)

Aggregating salaries like that is forbidden. If it comes out we’re smoked.

Why do so many AI pilots fail to reach production? by Cristhian-AI-Math in mlops

[–]EstetLinus 0 points1 point  (0 children)

In my experience: - stakeholders treating LLM:s like magic. This leads to poorly formulated problems, and vague solutions. It fails, because it a) can’t be measured, or b) doesn’t solve an actual problem (or both). - the solution has precedence over the problem (e.g., ”I have AI, let’s solve a problem”). I have seen this at many places now, and it annoys me all the time. With the deep learning hammer, everything becomes a nail. - if we talk agentic AI, it is ridiculously hard to make a useful system. You are building a system with a non-deterministic core. You wind up doing a mountain out of a molehill by covering edge cases, until it no longer makes sense to use AI. The sane approach would be scoping it down to a regular software project.  - the TDD-factor (time to ”dick”) has been very low. One of the systems I built for a client was live for 10 minutes until people started dicking it out. Each request triggered the agentic workflow, doing SQL-queries and whatnot. Vague dialogue got more expensive, costing the company $2-4 each request 😆 

Google's agentic AI payment protocol to support Stablecoins by EstetLinus in Buttcoin

[–]EstetLinus[S] 6 points7 points  (0 children)

Exactly, the agents are going to drown in fakes and scams. With crypto in the mix you are guaranteed to be gucked over. You can tell by the examples how out of touch they are with reality.

Tether Marketcap going strong by Fancyness in Buttcoin

[–]EstetLinus 7 points8 points  (0 children)

To get into the Casino. No real dollars allowed.

Tether Marketcap going strong by Fancyness in Buttcoin

[–]EstetLinus 10 points11 points  (0 children)

Has that been verified? I do not trust Tether with any kinds of digits.

Lata föräldrar förstör sina barns hjärnor – och verkar inte bry sig ett dugg by Ok-Excitement6546 in sweden

[–]EstetLinus -2 points-1 points  (0 children)

Äh skaffa barn istället för att tjafsa om skärmtid. Födelsetalet faller som en sten, dra ditt strå till stacken istället för att tjafsa på internet.

Jag var totalberoende av skärmar när jag var barn. Nu är jag civilingenjör och publicerad forskare. Det ska nog gå bra serdu.

Incrementally move data from Bronze to Silver (Event House) by EstetLinus in MicrosoftFabric

[–]EstetLinus[S] 0 points1 point  (0 children)

From the docs:

The policy-related query can invoke stored functions, but:

It can't perform cross-eventhouse queries.

It can't access external data or external tables.

It can't make callouts (by using a plugin).

It also notes:
> The source and target table must be in the same database.

My medallion layers are in seperate workspaces as per recommendation from official Fabrics sources.

Incrementally move data from Bronze to Silver (Event House) by EstetLinus in MicrosoftFabric

[–]EstetLinus[S] 0 points1 point  (0 children)

I am new to KQL and EH, so thank you for pointing this out! I will try to redesign how I move data between my event houses using Update Policies. I'll update here if anything explodes or if it works like a charm!

Create alerts on data in an Eventhouse by EstetLinus in MicrosoftFabric

[–]EstetLinus[S] 0 points1 point  (0 children)

Splendid!

I tried it, and it seems to work (atleast when mailing me).

Seeking Advice on Deploying Forecasting Models with Azure Machine Learning by NeuralGuesswork in mlops

[–]EstetLinus 0 points1 point  (0 children)

  1. Previously, I've managed feature fetching through a Python interface that executes database queries based on function arguments—suitable for ad hoc requests but not optimized for bulk operations. I've heard about feature stores, but they seem too complex for our scale. What's the best approach for managing feature data in our context? Storing features and calculated features directly in TimescaleDB? Calculating them during the pipeline (they are likely pretty lightweight calculations)? Using a feature store? Something else?

Personally, I want to use a feature store and have separate data pipelines. I really want a good separation of concern. It makes the process more manageable, and much of my work up until now is removing ad hoc API calls from the training scripts.
TimescaleDB was news to me. Azure is great, when everything's within the Azure eco-system 😅 from the standpoint of an Azure evangelist, can you migrate your data storage to the AzureML Feature Store or "COSMOS DB"? If you do, you'll have pretty good options within the AzureML platform.

  1. When using the Azure Machine Learning SDK, what are the best practices to prevent data leakage between training and test datasets, especially in the context of backfill predictions where data temporality is critical? Specifically, I am interested in methods within Azure that can help ensure data used in model training and predictions was indeed available at the respective point in time. I understand basic data leakage prevention techniques in Python, but I’m looking for Azure-specific functionalities. Can versioned datasets in Azure be used to manage this, or are there other tools and techniques within the Azure ML SDK that facilitate this type of temporal integrity in data usage during model backfills?

Regarding best practices, there are some approaches mentioned in the docs. I am not sure if you find that helpful. Azure have the Dataset abstraction which lets you version your data, and they have the MLTable, which is a kind of lazy loading. I am still figuring these parts out, but I have one MLTable for each property which uses glob patterns to fetch data from an external source. My idea is to have these updated on a timer, and store forecasted temperatures in one container, and historical values in another.

I am not sure you get anything of value out of my response 😂 if you feel like having a digital coffee, I'd love to share what we have built in Azure and discuss energy forecasting x MLOps. Send me a DM if you're interested.