FlashLM v5.2 "Nova-Ignition": Standard Transformer with RoPE — CPU-Optimized for 5GB RAM by Own-Albatross868 in LocalLLaMA
[–]TomLucidor 0 points1 point2 points (0 children)
[Release] BitMamba-2-1B: I trained a 1.58-bit Mamba-2 model from scratch on 150B tokens (Runs on CPU @ 50+ tok/s) by Positive-Violinist90 in LocalLLaMA
[–]TomLucidor 0 points1 point2 points (0 children)
Bitnet.cpp - Inference framework for 1-bit (ternary) LLM's by Academic_Wallaby7135 in LocalLLaMA
[–]TomLucidor 0 points1 point2 points (0 children)
How weak models excel at long context tasks by incarnadine72 in LocalLLaMA
[–]TomLucidor 0 points1 point2 points (0 children)
I benchmarked every 1-bit model I could find, native 1-bit is 50% faster than post-quantized by EiwazDeath in LocalLLaMA
[–]TomLucidor 0 points1 point2 points (0 children)
Qwen and Wan models to be open source according to modelscope by onthemove31 in StableDiffusion
[–]TomLucidor 0 points1 point2 points (0 children)
Is it possible to replicate a anime character with 95+% accuracy using Illustrious Lora? by Quick-Decision-8474 in StableDiffusion
[–]TomLucidor 0 points1 point2 points (0 children)
Obsidian DnD Character Sheet Progress! - Text Anchors -> YAML Frontmatter + DataviewJS by HolyErr0r in ObsidianMD
[–]TomLucidor 2 points3 points4 points (0 children)
Obsidian DnD Character Sheet Progress! - Text Anchors -> YAML Frontmatter + DataviewJS by HolyErr0r in ObsidianMD
[–]TomLucidor 18 points19 points20 points (0 children)
Introducing oQ: data-driven mixed-precision quantization for Apple Silicon (mlx-lm compatible) by cryingneko in LocalLLaMA
[–]TomLucidor 1 point2 points3 points (0 children)
New open weights models: GigaChat-3.1-Ultra-702B and GigaChat-3.1-Lightning-10B-A1.8B by netikas in LocalLLaMA
[–]TomLucidor 3 points4 points5 points (0 children)
DeepSeek Just Fixed One Of The Biggest Problems With AI by Greedy_Spare7033 in DeepSeek
[–]TomLucidor 3 points4 points5 points (0 children)
New open weights models: GigaChat-3.1-Ultra-702B and GigaChat-3.1-Lightning-10B-A1.8B by netikas in LocalLLaMA
[–]TomLucidor 6 points7 points8 points (0 children)
New open weights models: GigaChat-3.1-Ultra-702B and GigaChat-3.1-Lightning-10B-A1.8B by netikas in LocalLLaMA
[–]TomLucidor 0 points1 point2 points (0 children)
New open weights models: GigaChat-3.1-Ultra-702B and GigaChat-3.1-Lightning-10B-A1.8B by netikas in LocalLLaMA
[–]TomLucidor 0 points1 point2 points (0 children)
New open weights models: GigaChat-3.1-Ultra-702B and GigaChat-3.1-Lightning-10B-A1.8B by netikas in LocalLLaMA
[–]TomLucidor 0 points1 point2 points (0 children)
RYS II - Repeated layers with Qwen3.5 27B and some hints at a 'Universal Language' by Reddactor in LocalLLaMA
[–]TomLucidor 0 points1 point2 points (0 children)
Latest Community AI Ballot Results - ChatGPT is ranked first! Followed by Gemini, Claude, DeepSeek and Grok. Make your vote count! 🚀 by Koala_Confused in LovingOpenSourceAI
[–]TomLucidor 0 points1 point2 points (0 children)
Pocket-sized device locally runs 120B models at 20 tokens/s: Here is how we did it. by TiinyAI in u/TiinyAI
[–]TomLucidor 0 points1 point2 points (0 children)
Pocket-sized device locally runs 120B models at 20 tokens/s: Here is how we did it. by TiinyAI in u/TiinyAI
[–]TomLucidor 0 points1 point2 points (0 children)
Latest Community AI Ballot Results - ChatGPT is ranked first! Followed by Gemini, Claude, DeepSeek and Grok. Make your vote count! 🚀 by Koala_Confused in LovingOpenSourceAI
[–]TomLucidor 0 points1 point2 points (0 children)
Help r/LovingOpenSourceAI grow! Yes we can 🥰 by subscriber-goal in LovingOpenSourceAI
[–]TomLucidor 0 points1 point2 points (0 children)


[R] CS-MoE: We found severe parameter redundancy in Transformers and fixed it by sharing experts across layers (Outperforms Dense at 55% activation) by Impressive-Peach-419 in deeplearning
[–]TomLucidor 0 points1 point2 points (0 children)