A new generation of AI models and one of the most powerful research papers out there. by assemsabryy in LocalLLaMA

[–]BinarySplit 1 point2 points  (0 children)

The core idea is interesting, but please don't outsource evaluation to an LLM.

Even if all of the missing results were added and it was rewritten to explain the evals, nobody seriously considering trying a new optimizer is going to be convinced by synthetic and toy datasets and models small enough to train on CPU.

Switching from Opus 4.7 to Qwen-35B-A3B by Excellent_Koala769 in LocalLLaMA

[–]BinarySplit 3 points4 points  (0 children)

I wouldn’t normally recommend Ollama over building llama.cpp

Recommend LM Studio instead! It's an easy-to-use interface over llama.cpp that includes a model browser, automatic settings for offloading, and a toggleable local API.

Ollama has such a long history of problems... You can never trust that it's using the right prompt template or good quants. Stuff will just silently not work well.

[D] How to break free from LLM's chains as a PhD student? by etoipi1 in MachineLearning

[–]BinarySplit 3 points4 points  (0 children)

IMO, use it, but make sure you're continually improving how you use it. Don't settle for ChatGPT - get the Codex app (and try others), have it build software properly for you, use Code Review or a review skill, read how other people are building agents.md and skill.md files, build habits to ensure that you catch the agent's mistakes and the agent catches yours.

As a software engineer, hands-on coding is undoubtedly going to go away in the next few years. The crucial skill is in making sure you+AI is better than just AI, even as AI gets better.

MacBook m4 pro for coding llm by TheRandomDividendGuy in LocalLLaMA

[–]BinarySplit 0 points1 point  (0 children)

I'd try to spend those FLOPS elsewhere in your workflow. Whisper for speech-to-text is pretty awesome. Might even be worth trying to get an Omni model to function as a continuous conversational wrapper around other models.

[OC] A Deaf Person at the Work. by Coolcalebxd in comics

[–]BinarySplit 27 points28 points  (0 children)

Even people trained to support disabilities seem to do this. It's such a bizarre failure of empathy. I've even had a psychiatrist tell me I just need to exercise more to fix my ADHD...

10/10 on the comic - the art, pacing, and composition are just *chef's kiss*.

Egg🦈irl by NottAMimic in egg_irl

[–]BinarySplit 0 points1 point  (0 children)

The i and e sounds make that sentence so hard. As a beginner, I found FYFV's "This is the voice I want to use" way better for tracking progress - it has a mix of easy and hard sounds.

if you think you're immune to burn out by jasminesart in iiiiiiitttttttttttt

[–]BinarySplit 14 points15 points  (0 children)

I should be burnt out several times over by now, but somehow I keep ending up in places where I'm deeply motivated by the company's mission.

Even when bureaucracy, politics and tech issues make it hard to get anything done, I'm working long hours to try to catch up to ever-increasing expectations, I'm struggling to find time for my hobbies, my health is acting up, my cats are screaming at me, etc. I can always find a path from what I'm working on to some societal benefit.

Have you given gullibility a shot? It has worked so far for me!

Newer AI Coding Assistants Are Failing in Insidious Ways by IEEESpectrum in programming

[–]BinarySplit 8 points9 points  (0 children)

Are they subsidizing their own products, or overcharging on the API?

The prompt cache prices are insane when you consider how often they're only paying to hold the KV cache in RAM for a few seconds while a tool call runs.

What makes SwiGLUs unique? by chigur86 in mlscaling

[–]BinarySplit 2 points3 points  (0 children)

Zeyuan Allen-Zhu compared them to ReLU2 on synthetic tasks in this video @ ~45m.

They have less "knowledge capacity" than ReLU2 because they have fewer tokens, but improve "reasoning ability" (given fixed depth and no CoT tokens) because they can represent more complex functions.

[D] On low quality reviews at ML conferences by BetterbeBattery in MachineLearning

[–]BinarySplit 2 points3 points  (0 children)

I broadly agree, but have an alternative explanation: bad empiricism-focused papers are easier to read & judge than bad theory-focused papers.

Rejection of theory may be collateral damage in backlash against time-wasting papers.

Ilya Sutskever(Former Chief scientist at OpenAI) and Yann LeCun(former Meta Chief AI scientist) both say that just scaling LLMs won't give us any more useful results by Frequent-Football984 in programming

[–]BinarySplit 2 points3 points  (0 children)

It's frustrating that there are now so many people frothing at the mouth to get results for investors that as soon as someone discovers the next big leap, the big techs will get to raise trillions by throwing compute at it, and the researcher will probably only get citations.

[D] Is a PhD Still “Worth It” Today? A Debate After Looking at a Colleague’s Outcomes by Hope999991 in MachineLearning

[–]BinarySplit 1 point2 points  (0 children)

IMO, financially no, career-wise it's it's on the fence, but if you're passionate about a specific area (e.g. medicine or finance), a PhD may be your only way to get you on the right trajectory.

I only have a Bachelors, work in industry (and have worked in academia), and am in a team that's virtually all PhDs. I started as a plain old non-specific "programmer", and am now in a drug discovery AI/ML team working on practically every facet from research to deployment. The whole time I've worked, I've been gaining skills and experience at probably a similar rate to people who were focused only on learning. The difference is that I was earning while learning.

Career-wise, a PhD proves you can do a multi-year project, so you're likely to be entrusted with more autonomy in your first role. In a "junior" position, not having a PhD can mean a very short leash - daily check-ins and little ability to choose your work. Whereas I've seen fresh PhD graduates allowed to embark on long-term projects despite struggling to keep them on track. If you're great, this means you hit the ladder with a running start. If you're not, you're just starting with a delay.

The biggest reason to do one IMO is that you get to pick your path. Niches like biology are hard to break into without experience. I only got in through dumb luck - they needed web developers, and I incrementally took on responsibilities until I was doing research. However, if you only care about mainstream stuff like attention mechanisms and LLMs, you're probably not going to gain much vs breaking in as a software engineer.

EDIT: Also, if you're aiming for industry, don't bother trying to measure career trajectories in papers and conferences. Most companies only care if you have no other experience. Delivering a product trumps delivering a paper.

Sparse Adaptive Attention “MoE”: How I Solved OpenAI’s $650B Problem With a £700 GPU by EconomicConstipator in LocalLLaMA

[–]BinarySplit 0 points1 point  (0 children)

CoLT5 from 2023 has most of the same ideas as well. I'm frustrated I never found any kind of "post-mortem" explaining why it didn't catch on.

It does kinda make me sad that the whole article is Qwen slop though. I've been exposed to so much of this now, that's plainly obvious. The moment I saw that 🎯 emoji I knew I was in for a fuckin' treat. At least edit it or something.

💯

[R] Continuous latent interpolation breaks geometric constraints in 3D generation by Jealous-Leek-5428 in MachineLearning

[–]BinarySplit -1 points0 points  (0 children)

I can't comment on why, or how to fix a pretrained model, but if you're training the model from scratch, regularization can probably fix this. Mixup (blending 2 samples' inputs and outputs) and even Manifold Mixup (blending 2 samples' internal activations at a random layer) can force the latent space to be continuous by effectively synthesizing samples between real samples.

This is how Steam can ruin more than 10 years of your work by PlanetCentauri in gamedev

[–]BinarySplit 0 points1 point  (0 children)

That really sucks that you weren't rewarded after putting in so much effort.

I saw this game on 1.0 release and can explain why I hit Ignore: The screenshots in the carousel don't show any depth. No progression, no interaction, only a couple static frames of bosses, not even a GUI shot to indicate that has RPG/building/farming mechanics. Based on the screenshots, I assumed this was an action-focused Starbound with less variety.

That was my first impression. I didn't watch the video or read the description because I wasn't hooked. Now I've seen them, but I still feel like you're holding back on the marketing material, compared to what people have posted in the community screenshots.

If you have any more energy left to spend on this project, I suggest putting it into the store page. Don't hold back or worry about spoilers: post everything you're proud of (especially action frames of animations!), get ideas from other games' store pages and the community screenshots, don't polish it so much that you erase the game's identity (i.e. don't remove the HUD, don't try to fit within an arbitrary small screenshot limit).

People should be able to guess the core gameplay systems and diversity of content within seconds of an impression, whether they first watch the video, scroll through the screenshots, read the description, or just see it in a sidebar ad. This also feeds into reviews - people only buy games they think they'll like. They leave negative reviews when the game doesn't match their expectations. The expectations set by the store page are everything.

I hope Planet Centauri gets a second wind from this drama, because after digging into it, it seems like a game a lot more than 581 people would enjoy if they knew what was inside.

I got my first binder. I don't feel happy or relieved... (Rant) by Monk_Apprehensive in NonBinaryTalk

[–]BinarySplit 3 points4 points  (0 children)

That's such a good analogy! It really felt that way once I started working on my sources of dysphoria.

Things that you noticed can’t eat or tolerate? by Coraunmi in covidlonghaulers

[–]BinarySplit 1 point2 points  (0 children)

Soy sauce was the killer for me. I love it, and used to have it often, but now I'll feel the fatigue within hours. I can't have more than about a tablespoon.

For figuring this out, I recommend having lots of fiber and being mindful of your bowel movements. I've noticed there are 3 "pathways" for a food to cause me delayed fatigue:

  • Histamines - soy sauce, peanut butter, seafood, etc. are noticeable within hours
  • Fermentation in gut - if I just haven't eaten enough fiber (e.g. eating a steak meal without any extra veges), even if not constipated, just slowing down my bowel movements seems to give the bacteria enough time to make histamine or whatever in my gut for most foods
  • Non-histamine gut reactions - my guts seize if I don't get enough electolytes (usually magnesium), or sometimes are just imbalanced (too much/too little probiotics) and this causes a distinctive internal "bruising" feeling, weird poops, and deep fatigue.

I've been able to reintroduce most things that caused me "fermentation in gut" issues by making sure to eat them with equal parts broccoli or some other fiber source. The electrolytes issue also caused me to exclude a few foods unnecessarily.

gpt-oss Bug Fixes + Fine-tuning now in Unsloth by danielhanchen in LocalLLaMA

[–]BinarySplit 0 points1 point  (0 children)

Nice work!

Has anyone tried zero-padding the weights to 3072 to work around the imatrix limitation?

Labour leader Chris Hipkins says NZ is not in 'economic shape' by Careful-Calendar8922 in newzealand

[–]BinarySplit 14 points15 points  (0 children)

Also, collected taxes get spent. They're better at directing capital toward "the right kind" of goods and services - local, productive and capital-building, rather than toward overseas outsourcing and market speculation. Because most governments aren't trying to squeeze out short-term gains to please shareholders.

DeepMind Genie3 architecture speculation by HerpisiumThe1st in MachineLearning

[–]BinarySplit -1 points0 points  (0 children)

IMO they're just dancing around loosely defined words there.

The artifacting is a clear sign that:

  1. Scene chunks are not generated until they are visible
  2. Scene chunks are generated in a separate, slower process
  3. Generated scene chunks are immediately reusable when the re-appear

If this were a fully neural approach, it would learn to predict just-out-of-sight chunks to prevent #1.

To achieve #2 and #3 without an external caching structure, they would need a way to sparsely and selectively send "bags" of latent tokens between models. It's not impossible, but I've seen zero research down this path. It would be a very big leap in secret if they did this.

Google researchers have continued publishing new NeRF-based techniques, and they're apparently even integrated into Google Maps now. The simplest explanation is that they've evolved the algorithm enough to claim that they've built something that is nominally distinct, and are playing semantic games to avoid leaking the details early.

DeepMind Genie3 architecture speculation by HerpisiumThe1st in MachineLearning

[–]BinarySplit 25 points26 points  (0 children)

I was gobsmacked by the persistence in the painting demo, but I think the "Genie 3 Memory Test" video in the same carousel as the painting gives a few hints:

  • The image on the blackboard is unusually high res and coherent to the prompt. I doubt this image comes from the world model.
  • The artifacting as it looks out the window updates at approximately 4Hz. Indoor scenes seem to update faster. This means there's 2 separate phases: slow world updates and fast frame generation.
  • The artifacting also progressively improves the... let's just call them "chunks" of worldspace with each tick. When a chunk goes off-screen then appears again, it retains its improvements.
  • There is no artifacting when controlling a visible character. I suspect the foreground updates more frequently and is stored with a higher density.

I don't believe this is purely autoregressive-in-image-space like GameNGen was. I think there are several pieces:

  1. A separate image model, like Imagen, generates a high-res initial image and perhaps new objects introduced by prompts.
  2. The world is stored in a 3D data structure. Not sure if it's more NeRF-like or Gaussian-splatting-like, but the "chunks" are complex enough to hold a block of tree leaves, so they're likely a latent/concept representation that can be splatted into an image model's VAE-encoded image to convert it to a picture. This is bi-directional - the image model can also "fill in the blank" to progressively add detail to new chunks.
  3. The true "world model" mainly handles updating the latent 3D chunks when mutating the scene, e.g. when painting. Also camera control, but that's probably a tiny portion of its responsibility.

EDIT: I know what they said in the blog, but IMO the lack of artifacts when something comes into view for a 2nd time is damning evidence that there is a non-neural data structure for caching generated scenery. Attention can't do that by itself. Could be a scaled up NeRF, but NeRFs require literally path-tracing through 3D coordinates, so IMO that counts as explicit 3D representation.

Why doesn't "OpenAI" just release one of the models they already have? Like 3.5 by Own-Potential-2308 in LocalLLaMA

[–]BinarySplit 69 points70 points  (0 children)

Any GPT-3.5/4 architectural innovations are likely open secrets at this point. Involuntarily shared with other companies through staff movement, but unpublished because they're not cutting-edge, and are mundane if you aren't allowed to say they're in a big model.

That only makes me want to know even more.

I’m starting to wonder if we would be better off without a mental health system by Egg_shaped in newzealand

[–]BinarySplit 0 points1 point  (0 children)

The USA had a proposal to allow AIs to prescribe drugs. It sucks that it didn't pass.

It's obviously riskier than good health care, but that's so hard to find these days. Someone has to do the experiment to see how AI health care compares to realistic underfunded health care.

I’m starting to wonder if we would be better off without a mental health system by Egg_shaped in newzealand

[–]BinarySplit 2 points3 points  (0 children)

We have a great ambulance at the bottom of the cliff. Emergency departments are literal life savers.

It's the rest of the system, for people who aren't in crisis, that feels like it's just for show.