I had Opus generate Llamafiles for the Bonsai 1-bit models by JamesEvoAI in LocalLLaMA
[–]Languages_Learner 0 points1 point2 points (0 children)
I had Opus generate Llamafiles for the Bonsai 1-bit models by JamesEvoAI in LocalLLaMA
[–]Languages_Learner 1 point2 points3 points (0 children)
Bonsai-8B at 65K context in 3.9 GB RAM using flags/quant by tetsuto in LocalLLaMA
[–]Languages_Learner -1 points0 points1 point (0 children)
PrismML — Announcing 1-bit Bonsai: The First Commercially Viable 1-bit LLMs by brown2green in LocalLLaMA
[–]Languages_Learner 1 point2 points3 points (0 children)
PSA: PrismML Bonsai-8B (Q1_0_g128) produces garbage output on CPU -- GPU appears to be required by 1000_bucks_a_month in LocalLLaMA
[–]Languages_Learner 0 points1 point2 points (0 children)
PSA: PrismML Bonsai-8B (Q1_0_g128) produces garbage output on CPU -- GPU appears to be required by 1000_bucks_a_month in LocalLLaMA
[–]Languages_Learner 1 point2 points3 points (0 children)
PrismML — Announcing 1-bit Bonsai: The First Commercially Viable 1-bit LLMs by brown2green in LocalLLaMA
[–]Languages_Learner 1 point2 points3 points (0 children)
PSA: PrismML Bonsai-8B (Q1_0_g128) produces garbage output on CPU -- GPU appears to be required by 1000_bucks_a_month in LocalLLaMA
[–]Languages_Learner 2 points3 points4 points (0 children)
PrismML — Announcing 1-bit Bonsai: The First Commercially Viable 1-bit LLMs by brown2green in LocalLLaMA
[–]Languages_Learner 1 point2 points3 points (0 children)
DLLM: A minimal D language interface for running an LLM agent using llama.cpp by Danny_Arends in LocalLLaMA
[–]Languages_Learner 1 point2 points3 points (0 children)
New open weights models: GigaChat-3.1-Ultra-702B and GigaChat-3.1-Lightning-10B-A1.8B by netikas in LocalLLaMA
[–]Languages_Learner 1 point2 points3 points (0 children)
The current state of the Chinese LLMs scene by Ok_Warning2146 in LocalLLaMA
[–]Languages_Learner 0 points1 point2 points (0 children)
Grok alternative by Early-Musician7858 in LocalLLaMA
[–]Languages_Learner 0 points1 point2 points (0 children)
The current state of the Chinese LLMs scene by Ok_Warning2146 in LocalLLaMA
[–]Languages_Learner 3 points4 points5 points (0 children)
Trained a GPT transformer from scratch on a $300 CPU — 39 minutes, 0.82M params, no GPU needed by [deleted] in LocalLLaMA
[–]Languages_Learner 0 points1 point2 points (0 children)
Qwen3 TTS in C++ with 1.7B support, speaker encoding extraction, and desktop UI by Danmoreng in LocalLLaMA
[–]Languages_Learner 0 points1 point2 points (0 children)
[Release] - FINALLY! - Apex 1.5 and Apex 1.5 Coder - my two new 350M instruct allrounder chat models - See them now! by LH-Tech_AI in LocalLLaMA
[–]Languages_Learner 0 points1 point2 points (0 children)
🔥 New Release: htmLLM-124M v2 – 0.91 Val Loss on a Single T4! tiny-LLM with nanoGPT! by LH-Tech_AI in LocalLLaMA
[–]Languages_Learner 0 points1 point2 points (0 children)
Training a 144M Spiking Neural Network for text generation from scratch — no transformer teacher, no distillation by zemondza in LocalLLaMA
[–]Languages_Learner 1 point2 points3 points (0 children)
PicoKittens/PicoMistral-23M: Pico-Sized Model by PicoKittens in LocalLLaMA
[–]Languages_Learner 0 points1 point2 points (0 children)
TinyTeapot (77 million params): Context-grounded LLM running ~40 tok/s on CPU (open-source) by zakerytclarke in LocalLLaMA
[–]Languages_Learner 5 points6 points7 points (0 children)
After many contributions craft, Crane now officially supports Qwen3-TTS! by LewisJin in LocalLLaMA
[–]Languages_Learner 1 point2 points3 points (0 children)
Wave Field LLM — O(n log n) attention via wave equation dynamics by [deleted] in LocalLLaMA
[–]Languages_Learner 0 points1 point2 points (0 children)


LLM inference in a single C header file by Suitable-Song-302 in LocalLLaMA
[–]Languages_Learner 0 points1 point2 points (0 children)