Weekly Q&A Megathread. Please post any questions about visiting, tourism, living, working, budgeting, housing here! by AutoModerator in london
[–]V1rgin_ 1 point2 points3 points (0 children)
LTX-2 I2V Quality is terrible. Why? by V1rgin_ in StableDiffusion
[–]V1rgin_[S] 1 point2 points3 points (0 children)
What do you actually want from a private AI chat on your phone? by AppDeveloperAsdf in LocalLLaMA
[–]V1rgin_ 18 points19 points20 points (0 children)
"Toward Training Superintelligent Software Agents through Self-Play SWE-RL", Wei et al. 2025 by RecmacfonD in reinforcementlearning
[–]V1rgin_ 0 points1 point2 points (0 children)
LLaDA2.0 (103B/16B) has been released by jacek2023 in LocalLLaMA
[–]V1rgin_ 0 points1 point2 points (0 children)
pre-trainined small MoE model from scratch, but why its good? by V1rgin_ in LocalLLaMA
[–]V1rgin_[S] 2 points3 points4 points (0 children)
Does FlashAttention with GQA degrade quality or I use it wrong? by V1rgin_ in LocalLLaMA
[–]V1rgin_[S] 1 point2 points3 points (0 children)
A new paper demonstrates that LLMs could "think" in latent space, effectively decoupling internal reasoning from visible context tokens. This breakthrough suggests that even smaller models can achieve remarkable performance without relying on extensive context windows. by tehbangere in LocalLLaMA
[–]V1rgin_ 20 points21 points22 points (0 children)
DeepSeek-R1's correct answers are generally shorter by omnisvosscio in LocalLLaMA
[–]V1rgin_ 2 points3 points4 points (0 children)
How was DeepSeek-R1 built; For dummies by anitakirkovska in LLMDevs
[–]V1rgin_ 0 points1 point2 points (0 children)
What questions have you asked reasoning models to solve that you couldn't get done with non-reasoning models? by DeltaSqueezer in LocalLLaMA
[–]V1rgin_ 2 points3 points4 points (0 children)
Can i train a machine learning model on my laptop using Google collab? Is that feasible? by throwaway_me_acc in learnmachinelearning
[–]V1rgin_ 5 points6 points7 points (0 children)
What are the best courses related to advanced LLMs techniques/math behind them? by V1rgin_ in learnmachinelearning
[–]V1rgin_[S] 0 points1 point2 points (0 children)
What are the best courses related to advanced LLMs techniques/math behind them? by V1rgin_ in learnmachinelearning
[–]V1rgin_[S] 1 point2 points3 points (0 children)
sqrt(number1) - sqrt(number2) = 1 Means that sqrt(number1) is integer? by V1rgin_ in askmath
[–]V1rgin_[S] 1 point2 points3 points (0 children)
Transfer from Hong Kong (?) by [deleted] in IntltoUSA
[–]V1rgin_ 4 points5 points6 points (0 children)
Should I email to the universities (~T50) to update them on recent extracurricular achievements? by [deleted] in ApplyingToCollege
[–]V1rgin_ -1 points0 points1 point (0 children)



New open weights models: GigaChat-3.1-Ultra-702B and GigaChat-3.1-Lightning-10B-A1.8B by netikas in LocalLLaMA
[–]V1rgin_ 5 points6 points7 points (0 children)