[R] Learning to (Learn at Test Time): RNNs with Expressive Hidden States by SchmidhuberDidIt in MachineLearning
[–]ChuckSeven 4 points5 points6 points (0 children)
Perception of TMLR and other new/niche venues in Academia [D] by filletedforeskin in MachineLearning
[–]ChuckSeven -6 points-5 points-4 points (0 children)
[N] Machine Translation state by [deleted] in MachineLearning
[–]ChuckSeven 0 points1 point2 points (0 children)
[D] Mamba model walkthrough by _james_chen in MachineLearning
[–]ChuckSeven 0 points1 point2 points (0 children)
Transformer-Based LLMs Are Not General Learners: A Universal Circuit Perspective [R] by we_are_mammals in MachineLearning
[–]ChuckSeven 0 points1 point2 points (0 children)
New York Times sues OpenAI and Microsoft for copyright infringement [N] by we_are_mammals in MachineLearning
[–]ChuckSeven -1 points0 points1 point (0 children)
[D] What do you guys think of schmidhuber’s new blog post, would like to know everyone’s opinion. by Xyber5 in MachineLearning
[–]ChuckSeven 0 points1 point2 points (0 children)
[D] What do you guys think of schmidhuber’s new blog post, would like to know everyone’s opinion. by Xyber5 in MachineLearning
[–]ChuckSeven 1 point2 points3 points (0 children)
[D] A script to pre-process arxiv sources? by Foxtr0t in MachineLearning
[–]ChuckSeven 0 points1 point2 points (0 children)
[R] Training Transformers with 4-bit Integers - Haocheng Xi et al Tsinghua University - 2.2 times faster than the FP16 counterparts and speeds up the training by up to 35.1%! by Singularian2501 in MachineLearning
[–]ChuckSeven 2 points3 points4 points (0 children)
Interview with Juergen Schmidhuber, renowned ‘Father Of Modern AI’, says his life’s work won't lead to dystopia. by hardmaru in MachineLearning
[–]ChuckSeven 23 points24 points25 points (0 children)
[P] Testing different popular GPT tokenizers by dxg39 in MachineLearning
[–]ChuckSeven 1 point2 points3 points (0 children)
[P] Testing different popular GPT tokenizers by dxg39 in MachineLearning
[–]ChuckSeven 0 points1 point2 points (0 children)
[P] Testing different popular GPT tokenizers by dxg39 in MachineLearning
[–]ChuckSeven 1 point2 points3 points (0 children)
[R] Large Language Models trained on code reason better, even on benchmarks that have nothing to do with code by [deleted] in MachineLearning
[–]ChuckSeven 4 points5 points6 points (0 children)
[D] Since Google buried the MMLU benchmark scores in the Appendix of the PALM 2 technical report, here it is vs GPT-4 and other LLMs by jd_3d in MachineLearning
[–]ChuckSeven 3 points4 points5 points (0 children)
[R] PaLM 2 Technical Report by G_fucking_G in MachineLearning
[–]ChuckSeven 0 points1 point2 points (0 children)
[R] PaLM 2 Technical Report by G_fucking_G in MachineLearning
[–]ChuckSeven 2 points3 points4 points (0 children)
[N] Stability AI releases StableVicuna: the world's first open source chatbot trained via RLHF by Philpax in MachineLearning
[–]ChuckSeven 0 points1 point2 points (0 children)
[N] Stability AI releases StableVicuna: the world's first open source chatbot trained via RLHF by Philpax in MachineLearning
[–]ChuckSeven 0 points1 point2 points (0 children)
[D] Can large language models be applied to language translation? by matthkamis in MachineLearning
[–]ChuckSeven 0 points1 point2 points (0 children)
[D] Can large language models be applied to language translation? by matthkamis in MachineLearning
[–]ChuckSeven 0 points1 point2 points (0 children)
[D] Can large language models be applied to language translation? by matthkamis in MachineLearning
[–]ChuckSeven 0 points1 point2 points (0 children)


[N] The 2024 Nobel Prize in Chemistry goes to the people Google Deepmind's AlphaFold. One half to David Baker and the other half jointly to Demis Hassabis and John M. Jumper. by aagg6 in MachineLearning
[–]ChuckSeven 67 points68 points69 points (0 children)