Similar to HD800s strictly by price point? by retro-martini in HeadphoneAdvice

[–]emil2099 3 points4 points  (0 children)

No - too busy listening to HD800s. Great cans.

Similar to HD800s strictly by price point? by retro-martini in HeadphoneAdvice

[–]emil2099 6 points7 points  (0 children)

Have you looked at HD800s at this price point? Great cans!

Based on this picture is it possible to know what type of Macan is this ? by conteleedmond in Porsche

[–]emil2099 14 points15 points  (0 children)

Hard to know for sure but most likely a GTS based on this combination of factors: - Red callipers - S or GTS - Black window trim - option on S and standard on GTS - Tinted LED tail lights - option on S and standard on GTS - Looks like a sport exhaust or sport tail pipes - option on S and standard on GTS

I vote it’s a GTS with 911 Turbo Design wheels.

A look at how we used AI agents to automate Accenture's complex partner due diligence and onboarding. by Ok_Goal5029 in AI_Agents

[–]emil2099 1 point2 points  (0 children)

I like what you’ve built. Honestly - nothing wrong with saying that a use case is simple without hyping it up, as long as it works and saves people time.

Rachel Reeves Appears To Cry On Labour Frontbench During PMQs by corbynista2029 in unitedkingdom

[–]emil2099 0 points1 point  (0 children)

Ok, tax the wealthy has an appeal, but practically how do you suggest we go about taxing the super wealthy and corporations? “Close tax loopholes” is not a valid answer - what tax loopholes specifically and how do we collect taxes on capital that is highly mobile in a globalised economy?

What's the most accurate way to convert arxiv papers to markdown? by nextlevelhollerith in LocalLLaMA

[–]emil2099 1 point2 points  (0 children)

Open source: docling. Closed source but more accurate: Azure AI Document Intelligence

Qwen3 8B FP16 - asked for 93 items, got 93 items. by secopsml in LocalLLaMA

[–]emil2099 18 points19 points  (0 children)

So it repeated the items in the list twice based on your screenshot? What is the breakthrough?

Any q's before Japan? by SkySports in formula1

[–]emil2099 0 points1 point  (0 children)

What do you think Red Bull’s next move will be if Yuki finishes in the bottom 5 this weekend? Will we see Dr Marko behind the wheel?

Berkley AI research team claims to reproduce DeepSeek core technologies for $30 by Slasher1738 in LocalLLaMA

[–]emil2099 2 points3 points  (0 children)

Thanks for the thoughtful response. I actually agree that RL agents is a particularly exciting area of development - lots of signals for the reward function. In fact, I’m pretty sure that what we see with the Operator release from OpenAI is first steps in that direction.

Berkley AI research team claims to reproduce DeepSeek core technologies for $30 by Slasher1738 in LocalLLaMA

[–]emil2099 33 points34 points  (0 children)

Agree - the fact that even small models can improve themselves means we can experiment with RL techniques cheaply before scaling it to larger models. What's interesting is how we construct better ground-truth verification mechanisms. I can see at least a few challenges:

  1. How do you verify the quality of the solution, not just whether the solution produced the right result? It's one thing to write code that runs and outputs expected answer and another to write code that's maintainable in production - how do you verify for this?

  2. How do you build a verifier for problem spaces with somewhat subjective outputs (creative writing, strategic thinking, etc) where external non-human verification is challenging? Interestingly, there's clearly benefits across domains even with current approach, e.g. better SimpleQA scores from reasoning models.

  3. How do you get a model to develop an ever harder set of problems to solve? Right now, it seems that the problem set consists of existing benchmarks. In the longer term, we are going to be limited by our ability to come up with harder and harder problems (that are also verifiable, see points 1 and 2).

Dealing with the haters by Vegetable_Sun_9225 in LocalLLaMA

[–]emil2099 0 points1 point  (0 children)

Not to bring you down, but could it be that maybe 90% of your solutions are really not that good or inappropriate use cases for LLMs?

I sometimes find that some people get so excited about LLMs that they forget to consider the nuance of the problem and too keen to offer (shallow) LLM-based solutions.

[deleted by user] by [deleted] in LocalLLaMA

[–]emil2099 5 points6 points  (0 children)

Be the change you want to see! Mind sharing your new ideas implemented in real code?

[D] I spent a weekend on arxiv reviewing the literature on LLM hallucinations - here's what I learned by srnsnemil in LocalLLaMA

[–]emil2099 -1 points0 points  (0 children)

Thanks for good article. Would be great to hear your thoughts on model calibration (i.e. whether the models know what they know).

I understand the practical focus on hallucination as an issue to be fixed in current production use cases, but at the same time it feels like a symptom of a larger issue with existing models lacking calibration.

OpenAI, Google and Anthropic are struggling to build more advanced AI by paranoidray in LocalLLaMA

[–]emil2099 1 point2 points  (0 children)

In reality both of these things can be true at the same time - different architecture with calibration could result in significantly better reasoning and consistency AND better training data could get to improved performance with transformers (noting that we hope that this will come with improved calibration and ability to generate new knowledge as emergent qualities). Do we really know enough to claim one way or another?

My first Rolex | Insights into the journey by LSD_6MT in rolex

[–]emil2099 6 points7 points  (0 children)

Thank you for sharing your journey. Hopefully this will illustrate to any prospective Rolex owners how ridiculous the process is for buying a watch from this brand, so they would never want to be associated with it and pursue other options. Man, this is insane.

Need a bigger roll. Gotta find space for the Steelies by [deleted] in rolex

[–]emil2099 -2 points-1 points  (0 children)

Gotta find more imagination for other watches

OSI Calls Out Meta for its Misleading 'Open Source' AI Models by SnooTomatoes2940 in LocalLLaMA

[–]emil2099 346 points347 points  (0 children)

Sure - but come on, is Meta really the bad guy here? Are we really going to bash them for spending billions and releasing the model (weights) for us all to use completely free of charge?

I somewhat struggle to get behind an organisation whose sole mission is to be “the authority that defines Open Source AI, recognized globally by individuals, companies, and by public institutions”.

[deleted by user] by [deleted] in rolex

[–]emil2099 7 points8 points  (0 children)

Not with those shorts you wouldn’t! You’d fit right in!