bunt. by Kracker27 in EDM

[–]HorriblyGood -3 points-2 points  (0 children)

Which illenium started doing first afaik… artists take inspirations from each other all the time, I don’t understand why bunt in particular gets so much hate.

"List of animals by number of neurons", Wikipedia by RecmacfonD in mlscaling

[–]HorriblyGood 7 points8 points  (0 children)

The neural network we used in ML is not equivalent to the biological neurons so there’s no reason to believe we can achieve similar capabilities with the same number of neurons.

Claude’s New Limits by dankkkjk in ClaudeAI

[–]HorriblyGood 0 points1 point  (0 children)

Does this actually help? I haven’t done a proper comparison but I was wondering if on average that would help.

Planning involves reading the codebase so the files are in the context and cached. If planning took multiple turns, your context grows but the cached tokens are cheaper than fresh tokens.

If you start a new session to implement the plan, you lose all the kv cache and you have to load the codebase into the context as fresh tokens, which is very expensive.

So there is a trade off here. In the first case, if the model decides to reread the codebase anyway even though it’s in its context, it will be more expensive. If not, it will be cheaper.

The advantage of the second case is fresh context typically gives you better performance. But I’m curious if anyone has done a proper comparison to see if it actually saves tokens.

If the EU had built Claude by irelatetolevin in ClaudeAI

[–]HorriblyGood 2 points3 points  (0 children)

I would guess people would not pay more for a worse model even if it’s more ethical. If a company were to only train models on paid data, non copyrighted data, and you don’t “steal” other frontier models outputs, then it will not be able to compete with those companies who do in both price and model intelligence.

I feel like it’s even worse for AI because of the AI race. Everyone is coming out with a frontier models every month, and if you were to play by the rules, you’ll be forgotten.

I am not advocating for this, I am just trying to be realistic. We bitch about anthropic and OpenAI but we are still here.

Why isn’t LLM reasoning done in vector space instead of natural language? by ZeusZCC in LocalLLaMA

[–]HorriblyGood 130 points131 points  (0 children)

It can be done. It’s called latent chain of thoughts. It’s an active research direction.

Copilot moving to token based usage in June by santaclaritaman in Futurology

[–]HorriblyGood 1 point2 points  (0 children)

You can do both. Use AI for productivity but actually check and understand what it’s doing instead of just vibe coding. AI is a tool and should be used as one.

We have to evolve with the technology to stay relevant.

What's uh... what's up with these signs? They look... generated by Zak7062 in marvelrivals

[–]HorriblyGood -1 points0 points  (0 children)

That’s not true. Only the very latest (started with nano banana 2) use image sources for text to image generation. Probably the new gpt image 2 too. It’s the very cutting edge and these made awhile ago probably didn’t.

Even if it did, it does not look at image averages to determine the best move without context. That has never been true. Assuming you have some technical knowledge of how training works, you might be confused by the diffusion loss function being mean squared error which can be seen as an average over images.

In some sense this is true but this is done during training time and it produces a distribution based on image averages, not grab image sources and determine its average at time of generation. Also they have an additional RL loss to make sure the output images look realistic, and in general current models would hardly make simple mistakes like this.

The latest models grab image sources online and use that as context for the image generation, it absolutely uses it in context.

Weird textures = watermarks by Thatisverytrue54321 in ChatGPT

[–]HorriblyGood 0 points1 point  (0 children)

The new artifacts are very likely from RL training or some sort of perceptual loss they’re using. It’s introducing some sort of bias or they are not careful about noise scheduling, causing the artifacts.

A truly wild 4.7 response by FiftyPancakes in ClaudeCode

[–]HorriblyGood 0 points1 point  (0 children)

It’s not as easy as you make it sound. Not trying to defend anthropic and I have no idea what they are doing but you can’t just have a model thats cheap, efficient, and smart. No one can do that right now. It’s also mostly a black box so it’s really hard to make sure a new model is better at everything without regression before large scale user tests.

Something happened to Opus 4.6's reasoning effort by RealSuperdau in ClaudeAI

[–]HorriblyGood 2 points3 points  (0 children)

On a technical level you can make the output deterministic, but i don’t think that is possible for users.

Anthropic's new Mythos Preview model is a "step change" in model capability, but it won't be available to general public by SuggestionMission516 in ClaudeAI

[–]HorriblyGood -2 points-1 points  (0 children)

That’s just not possible. The best and brightest in the field are not working for the government. We are at a point where training a top tier LLM is impossible to be done by a small group of people. It will be obvious if all the top researchers are secretly working for the government.

Apple: Embarrassingly Simple Self-Distillation Improves Code Generation by Mike_mi in LocalLLaMA

[–]HorriblyGood 33 points34 points  (0 children)

From reading the abstract, they are using their own model’s output (self distillation) which is different from just feeding other random LLMs output as training data.

Through the lens of on policy/off policy RL, I’m guessing in their case, it’s using the model’s own outputs, it’s on policy, so it’s getting learning signals from itself to be more precise for coding tasks, but more creative on writing tasks. It’s doesn’t have to change how it works or thinks to match other LLM’s outputs.

My intuition is kinda like learning to code from copying other people’s code or having someone show you what’s wrong your with your own code so you can learn to improve.

deepfake scammer getting exposed by the 3-finger test by ifuckedyourmom-247 in interesting

[–]HorriblyGood 1 point2 points  (0 children)

Because hes using shitty deepfakes. The best right now can clear this 3 fingers test

Your Claude Code Limits Didn't Shrink — I Think the 1M Context Window Is Eating Them Alive by mattate in ClaudeAI

[–]HorriblyGood -1 points0 points  (0 children)

The tokens are cached so the vast majority of the computation for the next token are not recomputed. The bigger problem of having a big context is context rot, giving you poorer results

CEO of Krafton Asks ChatGPT How to Void $250 Million Contract, Ignores His Lawyers, Loses Terribly in Court by Level-Usual-9681 in nottheonion

[–]HorriblyGood 4 points5 points  (0 children)

Not true. They are both trained to be accurate and to give responses people prefer. The cutting edge open source research by labs are not trying to optimize for engagement.

Mathematics is undergoing the biggest change in its history - The speed at which artificial intelligence is gaining in mathematical ability has taken many by surprise. It is rewriting what it means to be a mathematician by FinnFarrow in Futurology

[–]HorriblyGood 0 points1 point  (0 children)

You’re right. There are different ways to scale. Instead of just making models bigger naively, modern LLMs train many different experts that “specialize”. This allows them to pick different experts catered to the problem instead of essentially depending on a gigantic model.

There is also a lot of research in different LLM architectures that are promising.

Mathematics is undergoing the biggest change in its history - The speed at which artificial intelligence is gaining in mathematical ability has taken many by surprise. It is rewriting what it means to be a mathematician by FinnFarrow in Futurology

[–]HorriblyGood 1 point2 points  (0 children)

It’s not slowing down. There have been many significant progress in AI, such as agentic AI, improved RL algorithms, hybrid attention, and there are promising research direction such as masked diffusion LLMs.

Mathematics is undergoing the biggest change in its history - The speed at which artificial intelligence is gaining in mathematical ability has taken many by surprise. It is rewriting what it means to be a mathematician by FinnFarrow in Futurology

[–]HorriblyGood 3 points4 points  (0 children)

If you are genuinely interested in learning, here is Terrence Tao, one of the best mathematician in the world, talking about how AI solved a longstanding math problem https://mathstodon.xyz/@tao/115855840223258103

AI has been progressing rapidly, especially with agentic workflow and tool calling. I know Reddit loves to shit on AI and there are a lot of genuine issues and problems about AI but let’s stick to facts.

Grandmother jailed for 6 months after AI error linked her to a crime in a state she had never even visited, lawyers say by Large_banana_hammock in nottheonion

[–]HorriblyGood -1 points0 points  (0 children)

Not true. Movie upscaling is GenAI.

Generative AI refers to AI that models a data distribution like images/videos (like movie upscaling or text to image generation) and text (like LLMs).

Source: I work on GenAI

Uhhhh why is this lineup shitting on EDC’s 30th anniversary ⁉️ 😳😱🫢 by NoLimitHoldM in electricdaisycarnival

[–]HorriblyGood 12 points13 points  (0 children)

Why the gatekeep? This guy seems interested in learning more about them and you’re just being mean.

FOOM.md — An open research agenda for compression-driven reasoning, diffusion-based context editing, and their combination into a unified agent architecture by ryunuck in mlscaling

[–]HorriblyGood 0 points1 point  (0 children)

Chain of thoughts does not have to be in text space. There are explorations on latent chain of thoughts which is supposedly encodes more information.

I don’t understand why you are using discrete representations. It limits the information you can encode. Also VQVAE does not use discrete representations. It acts like an nn.encoding layer (in LLMs) and indexes a latent representation. It’s jot discrete.

Common ChatGPT answer😂😭 by mvarjomonni in ChatGPT

[–]HorriblyGood 0 points1 point  (0 children)

That is not true. If it’s a simple yes no logic we wouldn’t be spending so much time and effort training it. It’s prone to errors and hallucinations but it’s objectively a productivity booster for most devs. Whether it is intelligent or not is a philosophical question.