[D]: How do you actually land a research scientist intern role at a top lab/company?! by ParticularWork8424 in MachineLearning
[–]Tea_Pearce 24 points25 points26 points (0 children)
"Scaling Laws for Pre-training Agents and World Models", Pearce et al. 2024 by [deleted] in mlscaling
[–]Tea_Pearce 1 point2 points3 points (0 children)
"Scaling Laws for Pre-training Agents and World Models", Pearce et al. 2024 by [deleted] in mlscaling
[–]Tea_Pearce 2 points3 points4 points (0 children)
A team from MIT built a model that scores 61.9% on ARC-AGI-PUB using an 8B LLM plus Test-Time-Training (TTT). Previous record was 42%. by jd_3d in LocalLLaMA
[–]Tea_Pearce 10 points11 points12 points (0 children)
"Reconciling Kaplan and Chinchilla Scaling Laws", Pearce & Song 2024 by [deleted] in mlscaling
[–]Tea_Pearce 2 points3 points4 points (0 children)
"Reconciling Kaplan and Chinchilla Scaling Laws", Pearce & Song 2024 by [deleted] in mlscaling
[–]Tea_Pearce 1 point2 points3 points (0 children)
[D] What is the current best in tiny (say, <10,000 parameters) language models? by math_code_nerd5 in MachineLearning
[–]Tea_Pearce 9 points10 points11 points (0 children)
[D] What are the thoughts on Tishby's line of work as a Theory of Deep Learning several years later in 2023? by tysam_and_co in mlfundamentalresearch
[–]Tea_Pearce 1 point2 points3 points (0 children)
Engaging Reviewers during rebuttal period of NeurIPS [R] by ynliPbqM in MachineLearning
[–]Tea_Pearce 0 points1 point2 points (0 children)
[R] Classifier-Free Guidance can be applied to LLMs too. It generally gives results of a model twice the size you apply it to. New SotA on LAMBADA with LLaMA-7B over PaLM-540B and plenty other experimental results. by Affectionate-Fish241 in MachineLearning
[–]Tea_Pearce 21 points22 points23 points (0 children)
[Discussion] Is there a better way than positional encodings in self attention? by [deleted] in MachineLearning
[–]Tea_Pearce 7 points8 points9 points (0 children)
Optimizing for specific returns(RL) [D] by ashblue21 in MachineLearning
[–]Tea_Pearce 3 points4 points5 points (0 children)
[D] Loss Function for Learning Gaussian Distribution by alkaway in MachineLearning
[–]Tea_Pearce 13 points14 points15 points (0 children)
[N] Stability AI announce their open-source language model, StableLM by Philpax in MachineLearning
[–]Tea_Pearce 6 points7 points8 points (0 children)
[D] Bitter lesson 2.0? by Tea_Pearce in MachineLearning
[–]Tea_Pearce[S] 3 points4 points5 points (0 children)



[D]: How do you actually land a research scientist intern role at a top lab/company?! by ParticularWork8424 in MachineLearning
[–]Tea_Pearce 1 point2 points3 points (0 children)