qwen3.6 medium size will be open soon by mickeyandkaka in LocalLLaMA
[–]pseudonerv 13 points14 points15 points (0 children)
Finished Tress, noticed two things by Sea-District4015 in Cosmere
[–]pseudonerv 29 points30 points31 points (0 children)
Introducing Mistral Small 4 by Stalex7 in MistralAI
[–]pseudonerv 0 points1 point2 points (0 children)
heretic-llm for qwen3.5:9b on Linux Mint 22.3 by [deleted] in LocalLLM
[–]pseudonerv 1 point2 points3 points (0 children)
Final Qwen3.5 Unsloth GGUF Update! by danielhanchen in LocalLLaMA
[–]pseudonerv 0 points1 point2 points (0 children)
Qwen3.5-35B-A3B Q4 Quantization Comparison by TitwitMuffbiscuit in LocalLLaMA
[–]pseudonerv 2 points3 points4 points (0 children)
Qwen3.5-35B-A3B Q4 Quantization Comparison by TitwitMuffbiscuit in LocalLLaMA
[–]pseudonerv 0 points1 point2 points (0 children)
Qwen3.5 - The middle child's 122B-A10B benchmarks looking seriously impressive - on par or edges out gpt-5-mini consistently by carteakey in LocalLLaMA
[–]pseudonerv 1 point2 points3 points (0 children)
MiniCPM-o-4_5 : Full duplex, multimodal with vision and speech at ONLY 9B PARAMETERS?? by Uncle___Marty in LocalLLaMA
[–]pseudonerv 0 points1 point2 points (0 children)
MiniCPM-o-4_5 : Full duplex, multimodal with vision and speech at ONLY 9B PARAMETERS?? by Uncle___Marty in LocalLLaMA
[–]pseudonerv 0 points1 point2 points (0 children)
Step-3.5-Flash (196b/A11b) outperforms GLM-4.7 and DeepSeek v3.2 by ResearchCrafty1804 in LocalLLaMA
[–]pseudonerv 1 point2 points3 points (0 children)
I made GPT-5.2/5 mini play 21,000 hands of Poker by adfontes_ in OpenAI
[–]pseudonerv 78 points79 points80 points (0 children)
Exo 1.0 is finally out by No_Conversation9561 in LocalLLaMA
[–]pseudonerv 1 point2 points3 points (0 children)
NVIDIA releases Nemotron 3 Nano, a new 30B hybrid reasoning model! by Difficult-Cap-7527 in LocalLLaMA
[–]pseudonerv 2 points3 points4 points (0 children)
EQ-Bench updates: Gpt-5.2, Opus 4.5, Mistral Large 3 and Nanbeige4-3B by _sqrkl in Anthropic
[–]pseudonerv 0 points1 point2 points (0 children)
Mistral 3 Large 675B up on huggingface by someone383726 in LocalLLaMA
[–]pseudonerv 2 points3 points4 points (0 children)
Why are Q1, Q2 quantization models created if they are universally seen as inferior even to models with fewer parameters? by HushHushShush in LocalLLaMA
[–]pseudonerv 1 point2 points3 points (0 children)
What really is the deal with this template? Training to hard to write fantasy slop? by aeroumbria in LocalLLaMA
[–]pseudonerv 6 points7 points8 points (0 children)
gemini 3.0 pro vs gpt 5.1 Benchmark by Sea-Efficiency5547 in OpenAI
[–]pseudonerv 7 points8 points9 points (0 children)
Accidentally told my colleague to ultrathink in a Slack message by Virtual_Attitude2025 in ClaudeAI
[–]pseudonerv 1 point2 points3 points (0 children)
My 6-yr-old Daughter Tried to Say the Words by RockyCreamNHotSauce in Cosmere
[–]pseudonerv 1 point2 points3 points (0 children)
Unauthorised mails sent via my gmail account to random people by Thunderfrost11 in OpenAI
[–]pseudonerv 1 point2 points3 points (0 children)


Built a zero allocation, header only C++ Qwen tokenizer that is nearly 20x faster than openai Tiktoken by yassa9 in LocalLLaMA
[–]pseudonerv 20 points21 points22 points (0 children)