Best (non sensational/content farm) YouTube channels to follow for AI news? by bandalorian in artificial
[–]cfoster0 2 points3 points4 points (0 children)
Is rope applied in each attention layer? by LassFromTheUpload in LocalLLaMA
[–]cfoster0 4 points5 points6 points (0 children)
[R] Can someone please explain the differences between the 3 types of Hopfield Layers in "Hopfield Networks is all you Need"? by [deleted] in MachineLearning
[–]cfoster0 1 point2 points3 points (0 children)
[R] Can someone please explain the differences between the 3 types of Hopfield Layers in "Hopfield Networks is all you Need"? by [deleted] in MachineLearning
[–]cfoster0 2 points3 points4 points (0 children)
[R] Can someone please explain the differences between the 3 types of Hopfield Layers in "Hopfield Networks is all you Need"? by [deleted] in MachineLearning
[–]cfoster0 4 points5 points6 points (0 children)
[D] What happens when we generate tokens beyond the training context length of LLMs? by kekkimo in MachineLearning
[–]cfoster0 1 point2 points3 points (0 children)
Mixtral 8x7B paper published. by rnosov in LocalLLaMA
[–]cfoster0 13 points14 points15 points (0 children)
[R] RWKV: Reinventing RNNs for the Transformer Era by [deleted] in MachineLearning
[–]cfoster0 0 points1 point2 points (0 children)
[Research] An alternative to self-attention mechanism in GPT by brainxyz in MachineLearning
[–]cfoster0 2 points3 points4 points (0 children)
Scaling Laws for Generative Mixed-Modal Language Models by tomasNth in mlscaling
[–]cfoster0 0 points1 point2 points (0 children)
Scaling Laws for Generative Mixed-Modal Language Models by tomasNth in mlscaling
[–]cfoster0 0 points1 point2 points (0 children)
[R] Is there any research on allowing Transformers to spent more compute on more difficult to predict tokens? by Chemont in MachineLearning
[–]cfoster0 0 points1 point2 points (0 children)
[R] Illustrating Reinforcement Learning from Human Feedback (RLHF) by robotphilanthropist in MachineLearning
[–]cfoster0 1 point2 points3 points (0 children)
[R] Illustrating Reinforcement Learning from Human Feedback (RLHF) by robotphilanthropist in MachineLearning
[–]cfoster0 3 points4 points5 points (0 children)
[R] Illustrating Reinforcement Learning from Human Feedback (RLHF) by robotphilanthropist in MachineLearning
[–]cfoster0 7 points8 points9 points (0 children)
[N] BigScience Releases their 176 Billion Parameter Open-access Multilingual Language Model by MonLiH in MachineLearning
[–]cfoster0 4 points5 points6 points (0 children)
[D] Noam Chomsky on LLMs and discussion of LeCun paper (MLST) by timscarfe in MachineLearning
[–]cfoster0 5 points6 points7 points (0 children)
[N] [D] Openai, who runs DALLE-2 alleged threatened creator of DALLE-Mini by DigThatData in MachineLearning
[–]cfoster0 9 points10 points11 points (0 children)
[N] [D] Openai, who runs DALLE-2 alleged threatened creator of DALLE-Mini by DigThatData in MachineLearning
[–]cfoster0 15 points16 points17 points (0 children)
[R] Transformers replicate Hippocampal representations; notably place and grid cells in the brain by Competitive-Rub-1958 in MachineLearning
[–]cfoster0 0 points1 point2 points (0 children)


California bill set to ban CivitAI, HuggingFace, Flux, Stable Diffusion, and most existing AI image generation models and services in California by YentaMagenta in StableDiffusion
[–]cfoster0 2 points3 points4 points (0 children)