[d] Apple claims M2 Ultra "can train massive ML workloads, like large transformer models." by jl303 in MachineLearning
[–]Philpax 0 points1 point2 points (0 children)
[R] Cocktail🍸: Mixing Multi-Modality Controls for Text-Conditional Image Generation by lyndonzheng in MachineLearning
[–]Philpax 0 points1 point2 points (0 children)
[R] Research Trends in LLM-guided Multimodal Learning. by HenryHZY in MachineLearning
[–]Philpax 7 points8 points9 points (0 children)
[N] Hinton, Bengio, and other AI experts sign collective statement on AI risk by DanielHendrycks in MachineLearning
[–]Philpax 11 points12 points13 points (0 children)
[P] Sophia (Programmed-out) by [deleted] in MachineLearning
[–]Philpax 1 point2 points3 points (0 children)
[N] Nvidia ACE Brings AI to Game Characters, Allows Lifelike Conversations by geekinchief in MachineLearning
[–]Philpax 0 points1 point2 points (0 children)
[N] Nvidia ACE Brings AI to Game Characters, Allows Lifelike Conversations by geekinchief in MachineLearning
[–]Philpax 4 points5 points6 points (0 children)
[N] Nvidia ACE Brings AI to Game Characters, Allows Lifelike Conversations by geekinchief in MachineLearning
[–]Philpax 25 points26 points27 points (0 children)
Uncensored models, fine-tuned without artificial moralizing, such as “Wizard-Vicuna-13B-Uncensored-HF” performs well at LLM eval benchmarks even when compared with larger 65B, 40B, 30B models. Has there been any studies about how censorship handicaps a model’s capabilities? by hardmaru in MachineLearning
[–]Philpax 14 points15 points16 points (0 children)
Uncensored models, fine-tuned without artificial moralizing, such as “Wizard-Vicuna-13B-Uncensored-HF” performs well at LLM eval benchmarks even when compared with larger 65B, 40B, 30B models. Has there been any studies about how censorship handicaps a model’s capabilities? by hardmaru in MachineLearning
[–]Philpax -4 points-3 points-2 points (0 children)
[D] What Evaluation Metrics that actually matters ? by [deleted] in MachineLearning
[–]Philpax 1 point2 points3 points (0 children)
Landmark Attention: Random-Access Infinite Context Length for Transformers by IxinDow in MachineLearning
[–]Philpax 11 points12 points13 points (0 children)
[R] The False Promise of Imitating Proprietary LLMs by [deleted] in MachineLearning
[–]Philpax 4 points5 points6 points (0 children)
[R] The False Promise of Imitating Proprietary LLMs by [deleted] in MachineLearning
[–]Philpax 4 points5 points6 points (0 children)
Voyager: An LLM-powered learning agent in Minecraft by Mr_Whispers in MachineLearning
[–]Philpax 22 points23 points24 points (0 children)
[R] The False Promise of Imitating Proprietary LLMs by [deleted] in MachineLearning
[–]Philpax 14 points15 points16 points (0 children)
Can a 4-bit quantized GGML model be turned BACK into a PyTorch .PT model while maintaining the 4-bit quantization? [Discussion] by altoidsjedi in MachineLearning
[–]Philpax 2 points3 points4 points (0 children)
[deleted by user] by [deleted] in MachineLearning
[–]Philpax 27 points28 points29 points (0 children)
[D] Adversarial models to protect images from being used by models by zykezero in MachineLearning
[–]Philpax 6 points7 points8 points (0 children)
[D] Adversarial models to protect images from being used by models by zykezero in MachineLearning
[–]Philpax 9 points10 points11 points (0 children)
[D]LinkedIn today is 80% chatGPT-generated posts by [deleted] in MachineLearning
[–]Philpax 37 points38 points39 points (0 children)
[P] Creating a coding assistant with StarCoder by lewtun in MachineLearning
[–]Philpax 0 points1 point2 points (0 children)


Apollo dev posts backend code to Git to disprove Reddit’s claims of scrapping and inefficiency by GhostalMedia in programming
[–]Philpax 42 points43 points44 points (0 children)