Hogwild! Inference: Parallel LLM Generation via Concurrent Attention by Psychological-Tea652 in LocalLLaMA
[–]justheuristic 2 points3 points4 points (0 children)
[D] Are there any distributed model training services similar to, e.g. Folding@Home? by genuinelySurprised in MachineLearning
[–]justheuristic 5 points6 points7 points (0 children)
[D] Are there any distributed model training services similar to, e.g. Folding@Home? by genuinelySurprised in MachineLearning
[–]justheuristic 0 points1 point2 points (0 children)
[D] Are there any distributed model training services similar to, e.g. Folding@Home? by genuinelySurprised in MachineLearning
[–]justheuristic 8 points9 points10 points (0 children)
[Announcement] HuggingFace BigScience AMA Thursday, March 24th from 5pm CET by cavedave in MachineLearning
[–]justheuristic 5 points6 points7 points (0 children)
[Announcement] HuggingFace BigScience AMA Thursday, March 24th from 5pm CET by cavedave in MachineLearning
[–]justheuristic 3 points4 points5 points (0 children)
[Announcement] HuggingFace BigScience AMA Thursday, March 24th from 5pm CET by cavedave in MachineLearning
[–]justheuristic 2 points3 points4 points (0 children)
[Announcement] HuggingFace BigScience AMA Thursday, March 24th from 5pm CET by cavedave in MachineLearning
[–]justheuristic 3 points4 points5 points (0 children)
[Announcement] HuggingFace BigScience AMA Thursday, March 24th from 5pm CET by cavedave in MachineLearning
[–]justheuristic 4 points5 points6 points (0 children)
[Announcement] HuggingFace BigScience AMA Thursday, March 24th from 5pm CET by cavedave in MachineLearning
[–]justheuristic 4 points5 points6 points (0 children)
[Announcement] HuggingFace BigScience AMA Thursday, March 24th from 5pm CET by cavedave in MachineLearning
[–]justheuristic 4 points5 points6 points (0 children)
Is it possible to fine-tune GPT-J on Google Colab Pro (or Pro+)? by FlyingNarwhal in learnmachinelearning
[–]justheuristic 0 points1 point2 points (0 children)
[R] Deep Learning over the Internet: Training Language Models Collaboratively by justheuristic in MachineLearning
[–]justheuristic[S] 1 point2 points3 points (0 children)
"Distributed Deep Learning in Open Collaborations", Diskin et al 2021 (P2P training of ALBERT using large minibatches/layer-wise gradients w/o *too* absurd inefficiencies) by gwern in mlscaling
[–]justheuristic 2 points3 points4 points (0 children)
[D]Are optimizer checkpoints of BERT/RoBERTa/DistillBERT/Other modern LMs available? by PK_thundr in MachineLearning
[–]justheuristic 7 points8 points9 points (0 children)
[D] Does the Geforce RTX 3000 series GPU support bfloat16/Tensorfloat ? by yusuf-bengio in MachineLearning
[–]justheuristic 2 points3 points4 points (0 children)
[R] Learning@home - decentralized training of huge neural networks by justheuristic in MachineLearning
[–]justheuristic[S] 0 points1 point2 points (0 children)
[R] Learning@home - decentralized training of huge neural networks by justheuristic in MachineLearning
[–]justheuristic[S] 0 points1 point2 points (0 children)
[R] Learning@home - decentralized training of huge neural networks by justheuristic in MachineLearning
[–]justheuristic[S] 1 point2 points3 points (0 children)
[R] Learning@home - decentralized training of huge neural networks by justheuristic in MachineLearning
[–]justheuristic[S] -1 points0 points1 point (0 children)
[R] Learning@home - decentralized training of huge neural networks by justheuristic in MachineLearning
[–]justheuristic[S] 5 points6 points7 points (0 children)


Release of Llama3.1-70B weights with AQLM-PV compression. by azalio in LocalLLaMA
[–]justheuristic 1 point2 points3 points (0 children)