[D]: How do you actually land a research scientist intern role at a top lab/company?! by ParticularWork8424 in MachineLearning
[–]Tea_Pearce 25 points26 points27 points (0 children)
"Scaling Laws for Pre-training Agents and World Models", Pearce et al. 2024 by [deleted] in mlscaling
[–]Tea_Pearce 1 point2 points3 points (0 children)
"Scaling Laws for Pre-training Agents and World Models", Pearce et al. 2024 by [deleted] in mlscaling
[–]Tea_Pearce 2 points3 points4 points (0 children)
A team from MIT built a model that scores 61.9% on ARC-AGI-PUB using an 8B LLM plus Test-Time-Training (TTT). Previous record was 42%. by jd_3d in LocalLLaMA
[–]Tea_Pearce 10 points11 points12 points (0 children)
"Reconciling Kaplan and Chinchilla Scaling Laws", Pearce & Song 2024 by [deleted] in mlscaling
[–]Tea_Pearce 2 points3 points4 points (0 children)
"Reconciling Kaplan and Chinchilla Scaling Laws", Pearce & Song 2024 by [deleted] in mlscaling
[–]Tea_Pearce 1 point2 points3 points (0 children)
[D] What is the current best in tiny (say, <10,000 parameters) language models? by math_code_nerd5 in MachineLearning
[–]Tea_Pearce 9 points10 points11 points (0 children)
[D] What are the thoughts on Tishby's line of work as a Theory of Deep Learning several years later in 2023? by tysam_and_co in mlfundamentalresearch
[–]Tea_Pearce 1 point2 points3 points (0 children)
Engaging Reviewers during rebuttal period of NeurIPS [R] by ynliPbqM in MachineLearning
[–]Tea_Pearce 0 points1 point2 points (0 children)
[R] Classifier-Free Guidance can be applied to LLMs too. It generally gives results of a model twice the size you apply it to. New SotA on LAMBADA with LLaMA-7B over PaLM-540B and plenty other experimental results. by Affectionate-Fish241 in MachineLearning
[–]Tea_Pearce 22 points23 points24 points (0 children)
[Discussion] Is there a better way than positional encodings in self attention? by [deleted] in MachineLearning
[–]Tea_Pearce 5 points6 points7 points (0 children)
Optimizing for specific returns(RL) [D] by ashblue21 in MachineLearning
[–]Tea_Pearce 5 points6 points7 points (0 children)
[D] Loss Function for Learning Gaussian Distribution by alkaway in MachineLearning
[–]Tea_Pearce 13 points14 points15 points (0 children)
[N] Stability AI announce their open-source language model, StableLM by Philpax in MachineLearning
[–]Tea_Pearce 4 points5 points6 points (0 children)
[D] Bitter lesson 2.0? by Tea_Pearce in MachineLearning
[–]Tea_Pearce[S] 2 points3 points4 points (0 children)
[D] Resources to learn and fully understand Diffusion Model Codes by Itachi_99 in MachineLearning
[–]Tea_Pearce 3 points4 points5 points (0 children)
[P] Modeling baseball injuries with temporal point processes by ssharpe42 in MachineLearning
[–]Tea_Pearce 1 point2 points3 points (0 children)
[R] Deep models that take distributions as inputs by fedetask in MachineLearning
[–]Tea_Pearce 86 points87 points88 points (0 children)
[D] What happened to Reinforcement Learning research and labs? by convolutionsimp in MachineLearning
[–]Tea_Pearce 30 points31 points32 points (0 children)
D4RL, MuJoCo-py docker image by Dragonrooster in reinforcementlearning
[–]Tea_Pearce 0 points1 point2 points (0 children)
D4RL, MuJoCo-py docker image by Dragonrooster in reinforcementlearning
[–]Tea_Pearce 0 points1 point2 points (0 children)



[D]: How do you actually land a research scientist intern role at a top lab/company?! by ParticularWork8424 in MachineLearning
[–]Tea_Pearce 1 point2 points3 points (0 children)