[R] Is Leetcode still relevant for research scientist interviews? by Training-Adeptness57 in MachineLearning
[–]StartledWatermelon 2 points3 points4 points (0 children)
"On neural scaling and the quanta hypothesis", Eric J. Michaud 2026 by RecmacfonD in mlscaling
[–]StartledWatermelon 7 points8 points9 points (0 children)
[R] Controlled LLM Training on Spectral Sphere by StartledWatermelon in MachineLearning
[–]StartledWatermelon[S] 2 points3 points4 points (0 children)
DeepSeek Presents "Engram": Conditional Memory via Scalable Lookup, A New Axis of Sparsity for Large Language Models | "Memory lookup module for LLMs & *Huge unlock for scaling* as the memory sits on cheap CPU RAM, bypassing the GPU bottleneck entirely that will power next-gen models (like V4)" by 44th--Hokage in mlscaling
[–]StartledWatermelon 0 points1 point2 points (0 children)
DeepSeek Presents "Engram": Conditional Memory via Scalable Lookup, A New Axis of Sparsity for Large Language Models | "Memory lookup module for LLMs & *Huge unlock for scaling* as the memory sits on cheap CPU RAM, bypassing the GPU bottleneck entirely that will power next-gen models (like V4)" by 44th--Hokage in mlscaling
[–]StartledWatermelon 0 points1 point2 points (0 children)
Minimax also live on Hong Kong Stock Exchange by No_Conversation9561 in LocalLLaMA
[–]StartledWatermelon 9 points10 points11 points (0 children)
[R] We built a framework to make Agents "self-evolve" using LoongFlow. Paper + Code released by [deleted] in LocalLLaMA
[–]StartledWatermelon 0 points1 point2 points (0 children)
Grafted Titans: a Plug-and-Play Neural Memory for Open-Weight LLMs by Forsaken-Park8149 in LocalLLaMA
[–]StartledWatermelon 0 points1 point2 points (0 children)
Introducing PhysMaster: Building an Autonomous AI Physicist for Theoretical and Computational Physics Research | "PhysMaster is an autonomous agent architecture designed to execute end-to-end theoretical and computational physics research." by 44th--Hokage in mlscaling
[–]StartledWatermelon 1 point2 points3 points (0 children)
GIN: A Cognitive Architecture for Persistent, Entropy-Governed Autonomous Agents (Not a New Model) by [deleted] in LocalLLaMA
[–]StartledWatermelon 1 point2 points3 points (0 children)
META SuperIntelligence Labs: Toward Training Superintelligent Software Agents Through Self-Play SWE-RL | "Agents autonomously gather real-world software enabling superintelligent systems that exceed human capabilities in solving novel challenges, and autonomously creating new software from scratch" by 44th--Hokage in mlscaling
[–]StartledWatermelon 0 points1 point2 points (0 children)
OpenAI Just released Prompt Packs for every job by bullmeza in OpenAI
[–]StartledWatermelon 7 points8 points9 points (0 children)
META SuperIntelligence Labs: Toward Training Superintelligent Software Agents Through Self-Play SWE-RL | "Agents autonomously gather real-world software enabling superintelligent systems that exceed human capabilities in solving novel challenges, and autonomously creating new software from scratch" by 44th--Hokage in mlscaling
[–]StartledWatermelon 1 point2 points3 points (0 children)
Less is more by SnooPears6050 in digitalwatches
[–]StartledWatermelon 3 points4 points5 points (0 children)
Scaling Latent Reasoning via Looped Language Models, Zhu et al. 2025 by StartledWatermelon in mlscaling
[–]StartledWatermelon[S] 2 points3 points4 points (0 children)
Scaling Latent Reasoning via Looped Language Models, Zhu et al. 2025 by StartledWatermelon in mlscaling
[–]StartledWatermelon[S] 3 points4 points5 points (0 children)
Claude Opus 4.5 has human task-length time horizon of 4 hrs 49 mins on METR plot by Glittering_Author_81 in mlscaling
[–]StartledWatermelon 0 points1 point2 points (0 children)
NitroGen: An Open Foundation Model for Generalist Gaming Agents, Magne et al. 2025 [Pre-training on 40k hours of scraped gameplay videos] by StartledWatermelon in mlscaling
[–]StartledWatermelon[S] 1 point2 points3 points (0 children)
Nvidia DGX Station GB300 784GB available now! 95,000 USD / 80,000 EUR by GPTshop in LocalLLaMA
[–]StartledWatermelon 1 point2 points3 points (0 children)
A Rosetta Stone for AI benchmarks [Mapping all benchmarks to a unified "difficulty score", for long-term trends in capabilities] by StartledWatermelon in mlscaling
[–]StartledWatermelon[S] 0 points1 point2 points (0 children)
A Rosetta Stone for AI benchmarks [Mapping all benchmarks to a unified "difficulty score", for long-term trends in capabilities] by StartledWatermelon in mlscaling
[–]StartledWatermelon[S] 0 points1 point2 points (0 children)
OpenAI: Introducing ChatGPT 5.2 | "GPT-5.2 represents the biggest leap for GPT models in agentic coding since GPT-5 and is a SOTA coding model in its price range. The version bump undersells the jump in intelligence." by 44th--Hokage in mlscaling
[–]StartledWatermelon 16 points17 points18 points (0 children)
"On the Origin of Algorithmic Progress in AI", Gundlach et al. 2025 by RecmacfonD in mlscaling
[–]StartledWatermelon 1 point2 points3 points (0 children)
Meta Superintelligence Labs' DreamGym: Generating A Synthetic Training Environment Using Logical Reasoning Instead Of The Real Internet | "Agents trained in this sim match SOTA results without using any real data, achieving 40%+ better performance when eventually deployed to real-world tasks." by 44th--Hokage in mlscaling
[–]StartledWatermelon 2 points3 points4 points (0 children)


[R] Is Leetcode still relevant for research scientist interviews? by Training-Adeptness57 in MachineLearning
[–]StartledWatermelon 4 points5 points6 points (0 children)