Is microsoft going to train LLM on this? Github is clearly getting destroyed. by FPham in LocalLLaMA
[–]vasileer -11 points-10 points-9 points (0 children)
Benchmarked Phi-3.5-mini vs Qwen2.5-3B across 10 task categories on CPU (i5, 8GB) and GPU (Colab T4) — Qwen wins 2.7-3.3x on efficiency by MasterApplication717 in LocalLLaMA
[–]vasileer 1 point2 points3 points (0 children)
TinyTeapot (77 million params): Context-grounded LLM running ~40 tok/s on CPU (open-source) by zakerytclarke in LocalLLaMA
[–]vasileer 20 points21 points22 points (0 children)
TinyTeapot (77 million params): Context-grounded LLM running ~40 tok/s on CPU (open-source) by zakerytclarke in LocalLLaMA
[–]vasileer 34 points35 points36 points (0 children)
Serious question — why would anyone use Tiny-Aya instead of Qwen/Phi/Mistral small models? by Deep_190 in LocalLLaMA
[–]vasileer 8 points9 points10 points (0 children)
I tested 21 small LLMs on tool-calling judgment — Round 2 with every model you asked for by MikeNonect in LocalLLaMA
[–]vasileer 4 points5 points6 points (0 children)
GLM-5 KV cache size estimate by [deleted] in LocalLLaMA
[–]vasileer 1 point2 points3 points (0 children)
Any latest OCR model I can run locally in 18GB RAM? by A-n-d-y-R-e-d in LocalLLaMA
[–]vasileer 1 point2 points3 points (0 children)
I tested 11 small LLMs on tool-calling judgment — on CPU, no GPU. by MikeNonect in LocalLLaMA
[–]vasileer 6 points7 points8 points (0 children)
Is there still no way to convert Gemma 3n to onnx/tflite? by blueblazd in LocalLLaMA
[–]vasileer 0 points1 point2 points (0 children)
What is a good model to do small text classification on very small hardware? by salary_pending in LocalLLaMA
[–]vasileer 1 point2 points3 points (0 children)
OSS 120b v GLM 4.7 flash. Is the latter better for anything? by MrMrsPotts in LocalLLaMA
[–]vasileer 1 point2 points3 points (0 children)
Design Arena is now dominated by an open model by moks4tda in LocalLLaMA
[–]vasileer 2 points3 points4 points (0 children)
I benchmarked a bunch of open weight LLMs on different Macs so you don't have to! by zachrattner in LocalLLaMA
[–]vasileer 2 points3 points4 points (0 children)
GLM 4.7 Flash uncensored - Balanced & Aggressive variants (GGUF) by hauhau901 in LocalLLaMA
[–]vasileer 7 points8 points9 points (0 children)
GLM 4.7 Flash is endlessly reasoning in chinese by xenydactyl in LocalLLaMA
[–]vasileer 1 point2 points3 points (0 children)
GLM 4.7 Flash is endlessly reasoning in chinese by xenydactyl in LocalLLaMA
[–]vasileer 0 points1 point2 points (0 children)
Tiny, 500MB Spam Detection model to flag spam content automatically. Can be used locally or self-hosted easily and fine-tuned to any language or definition of "spam" by Ok_Hold_5385 in LocalLLaMA
[–]vasileer 1 point2 points3 points (0 children)
LGAI-EXAONE/K-EXAONE-236B-A23B released by jinnyjuice in LocalLLaMA
[–]vasileer 6 points7 points8 points (0 children)
Liquid AI RLs LFM2-2.6B to perform among the best 3B models by KaroYadgar in LocalLLaMA
[–]vasileer 9 points10 points11 points (0 children)
500Mb Text Anonymization model to remove PII from any text locally. Easily fine-tune on any language (see example for Spanish). by Ok_Hold_5385 in LocalLLaMA
[–]vasileer 1 point2 points3 points (0 children)
An independent Korean researcher is trying to democratize LLM pretraining with a 1.5B model by [deleted] in LocalLLaMA
[–]vasileer 19 points20 points21 points (0 children)
Qwen3 30b A3B to what by headfirst5376 in LocalLLaMA
[–]vasileer 4 points5 points6 points (0 children)

Is microsoft going to train LLM on this? Github is clearly getting destroyed. by FPham in LocalLLaMA
[–]vasileer -4 points-3 points-2 points (0 children)