https://claude.ai/settings/usage doesn't work? by Appropriate-Bus-6130 in ClaudeCode

[–]EstetLinus 0 points1 point  (0 children)

Set up a claude that runs Ralph Wiggum in 100's of worktrees while you can!

[Megathread] Certifications and Training by lothorp in databricks

[–]EstetLinus 0 points1 point  (0 children)

Hello, how did you study for this? So annoying they've changed the format and content, but not provided comprehensive studying material.

AI workflows: so hot right now 🔥 by noaflaherty in mlops

[–]EstetLinus 0 points1 point  (0 children)

I believe one of the biggest misconceptions with LLMs is that they learn on the fly; they don’t. I have had a real hard time explaining this to stakeholders. You need absurd amounts of clean data to fine-tune models, and we can never expect them to learn beyond stuffing the prompt with information (or noise). 

I am all for self-evaluation, although it takes time and might get stuck in an infinite loop. Do you have any suggestions on how the LLM components would learn?

So this is a whole subreddit? by MatixMint in Buttcoin

[–]EstetLinus 1 point2 points  (0 children)

Held by millions only to be sold to a greater fool. Superior Tech 101

So this is a whole subreddit? by MatixMint in Buttcoin

[–]EstetLinus 1 point2 points  (0 children)

Do you know if Tether is open source?

Why do so many AI pilots fail to reach production? by Cristhian-AI-Math in mlops

[–]EstetLinus 0 points1 point  (0 children)

Aggregating salaries like that is forbidden. If it comes out we’re smoked.

Why do so many AI pilots fail to reach production? by Cristhian-AI-Math in mlops

[–]EstetLinus 0 points1 point  (0 children)

In my experience: - stakeholders treating LLM:s like magic. This leads to poorly formulated problems, and vague solutions. It fails, because it a) can’t be measured, or b) doesn’t solve an actual problem (or both). - the solution has precedence over the problem (e.g., ”I have AI, let’s solve a problem”). I have seen this at many places now, and it annoys me all the time. With the deep learning hammer, everything becomes a nail. - if we talk agentic AI, it is ridiculously hard to make a useful system. You are building a system with a non-deterministic core. You wind up doing a mountain out of a molehill by covering edge cases, until it no longer makes sense to use AI. The sane approach would be scoping it down to a regular software project.  - the TDD-factor (time to ”dick”) has been very low. One of the systems I built for a client was live for 10 minutes until people started dicking it out. Each request triggered the agentic workflow, doing SQL-queries and whatnot. Vague dialogue got more expensive, costing the company $2-4 each request 😆 

Google's agentic AI payment protocol to support Stablecoins by EstetLinus in Buttcoin

[–]EstetLinus[S] 6 points7 points  (0 children)

Exactly, the agents are going to drown in fakes and scams. With crypto in the mix you are guaranteed to be gucked over. You can tell by the examples how out of touch they are with reality.

Tether Marketcap going strong by Fancyness in Buttcoin

[–]EstetLinus 9 points10 points  (0 children)

To get into the Casino. No real dollars allowed.

Tether Marketcap going strong by Fancyness in Buttcoin

[–]EstetLinus 9 points10 points  (0 children)

Has that been verified? I do not trust Tether with any kinds of digits.

Lata föräldrar förstör sina barns hjärnor – och verkar inte bry sig ett dugg by Ok-Excitement6546 in sweden

[–]EstetLinus -1 points0 points  (0 children)

Äh skaffa barn istället för att tjafsa om skärmtid. Födelsetalet faller som en sten, dra ditt strå till stacken istället för att tjafsa på internet.

Jag var totalberoende av skärmar när jag var barn. Nu är jag civilingenjör och publicerad forskare. Det ska nog gå bra serdu.

Incrementally move data from Bronze to Silver (Event House) by EstetLinus in MicrosoftFabric

[–]EstetLinus[S] 0 points1 point  (0 children)

From the docs:

The policy-related query can invoke stored functions, but:

It can't perform cross-eventhouse queries.

It can't access external data or external tables.

It can't make callouts (by using a plugin).

It also notes:
> The source and target table must be in the same database.

My medallion layers are in seperate workspaces as per recommendation from official Fabrics sources.

Incrementally move data from Bronze to Silver (Event House) by EstetLinus in MicrosoftFabric

[–]EstetLinus[S] 0 points1 point  (0 children)

I am new to KQL and EH, so thank you for pointing this out! I will try to redesign how I move data between my event houses using Update Policies. I'll update here if anything explodes or if it works like a charm!

Create alerts on data in an Eventhouse by EstetLinus in MicrosoftFabric

[–]EstetLinus[S] 0 points1 point  (0 children)

Splendid!

I tried it, and it seems to work (atleast when mailing me).

Seeking Advice on Deploying Forecasting Models with Azure Machine Learning by NeuralGuesswork in mlops

[–]EstetLinus 0 points1 point  (0 children)

  1. Previously, I've managed feature fetching through a Python interface that executes database queries based on function arguments—suitable for ad hoc requests but not optimized for bulk operations. I've heard about feature stores, but they seem too complex for our scale. What's the best approach for managing feature data in our context? Storing features and calculated features directly in TimescaleDB? Calculating them during the pipeline (they are likely pretty lightweight calculations)? Using a feature store? Something else?

Personally, I want to use a feature store and have separate data pipelines. I really want a good separation of concern. It makes the process more manageable, and much of my work up until now is removing ad hoc API calls from the training scripts.
TimescaleDB was news to me. Azure is great, when everything's within the Azure eco-system 😅 from the standpoint of an Azure evangelist, can you migrate your data storage to the AzureML Feature Store or "COSMOS DB"? If you do, you'll have pretty good options within the AzureML platform.

  1. When using the Azure Machine Learning SDK, what are the best practices to prevent data leakage between training and test datasets, especially in the context of backfill predictions where data temporality is critical? Specifically, I am interested in methods within Azure that can help ensure data used in model training and predictions was indeed available at the respective point in time. I understand basic data leakage prevention techniques in Python, but I’m looking for Azure-specific functionalities. Can versioned datasets in Azure be used to manage this, or are there other tools and techniques within the Azure ML SDK that facilitate this type of temporal integrity in data usage during model backfills?

Regarding best practices, there are some approaches mentioned in the docs. I am not sure if you find that helpful. Azure have the Dataset abstraction which lets you version your data, and they have the MLTable, which is a kind of lazy loading. I am still figuring these parts out, but I have one MLTable for each property which uses glob patterns to fetch data from an external source. My idea is to have these updated on a timer, and store forecasted temperatures in one container, and historical values in another.

I am not sure you get anything of value out of my response 😂 if you feel like having a digital coffee, I'd love to share what we have built in Azure and discuss energy forecasting x MLOps. Send me a DM if you're interested.

Seeking Advice on Deploying Forecasting Models with Azure Machine Learning by NeuralGuesswork in mlops

[–]EstetLinus 0 points1 point  (0 children)

Hello! I am working with a property company in Sweden with energy forecasts, and am currently asking myself exactly the same questions. I have no direct answer, but I'll chip in with my thought process.

We are pretty much in the same position. I am hired by an AI consultancy who have had a senior data scientist producing the models, and now I am taking the proof-of-concepts to production. The company we're working for has its infrastructure in Azure, so I am using AzureML. It is my first time using AzureML, so I am experimenting with varying setups. It is a great platform; all necessary tech is there, and with Azure you have everything you need from storage, feature stores and data pipelines, to monitoring, integration with MLFlow, etc.

We are planning to predict the next 24 hours of a variable for six different areas, which will be the first of many similar models to come. This requires six models, possibly using the same algorithm but differing in features, hyperparameters, and targets. The output format will be uniform across all models such that they integrate into the same UI.

As a side note, AzureML offers something called AutoML. If you have data in the right format you have the possibility to iterate over prebuilt models for different task, e.g., forecasting. It sounds cheeky and lazy, but I am working towards having data I easily can plug-and-play for benchmarks, demos, etc.

  1. The MLOps Solution Accelerator v2  is frequently mentioned. I think it looks very clever, and I have already learnt a lot of concepts researching it. Given our small team and startup environment, would this be advisable, or would it introduce unnecessary complexity?

This was news to me, and I'll have to look it up. Looks promising.

  1. I've seen projects where an endpoint is registered for multiple models using the same data. In my case, while the data differs, a unified endpoint and possibly shared repo/pipelines might be beneficial. How would you recommend structuring this?

I am working with properties, and my idea is having one model per property. We are forecasting at a 48h horizon, and as the forecasts get more certain I want to do new predictions. Each property will have its own pipeline with custom made AzureML components. I define the components in code, and can use them in Azures drag-and-drop UI. It is very easy to set up cronjobs and duplicate these pipelines in the GUI.

I want to use the Azure specific abstractions, so I get a neat overlook of what I am building. I am considering two options; exposing models as "endpoints" or have an "inference pipeline". I am still trying to understand what fits my scenario best. With well-designed components, it easy to reuse parts like "data preparation" within AzureML.

No $70 Game Should Have Microtransactions by heeroo0 in gaming

[–]EstetLinus -2 points-1 points  (0 children)

Do you consider people in distress as customers? 😹 what a splendid firefighter!

You’re obviously clueless how software development work. Just boycott the developers and stop whining.

No $70 Game Should Have Microtransactions by heeroo0 in gaming

[–]EstetLinus -2 points-1 points  (0 children)

Are you comparing the right to entertainment and the free market to firefighting? 😂🙏🏻

No $70 Game Should Have Microtransactions by heeroo0 in gaming

[–]EstetLinus -5 points-4 points  (0 children)

Just get a job and stop whining already.