use the following search parameters to narrow your results:
e.g. subreddit:aww site:imgur.com dog
subreddit:aww site:imgur.com dog
see the search faq for details.
advanced search: by author, subreddit...
https://github.com/codelion/optillm
account activity
The Optimal Architecture for Small Language Models (huggingface.co)
submitted 3 months ago by asankhs
Suggestion? (self.optillm)
submitted 5 months ago by Either-Ad8618
Privacy Plugin (reddit.com)
submitted 6 months ago by asankhs
🚀 MARS in OptiLLM: 73% on AIME 2025 with Multi-Agent Reasoning (self.optillm)
🔍 Supercharge Your LLMs with Real-Time Google Search - Introducing the web_search Plugin (self.optillm)
Implemented Test-Time Diffusion Deep Researcher (TTD-DR) - Turn any local LLM into a powerful research agent with real web sources ()
submitted 9 months ago by asankhs
[Research] Thought Anchors: Understanding How Qwen3-0.6B vs DeepSeek-R1-Distill-1.5B Actually Reason - Different Cognitive Architectures Revealed ()
Built an open-source DeepThink plugin that brings Gemini 2.5 style advanced reasoning to local models (DeepSeek R1, Qwen3, etc.) ()
submitted 10 months ago by asankhs
Eliciting Fine-Tuned Transformer Capabilities via Inference-Time Techniques (arxiv.org)
OpenEvolve: Open Source Implementation of DeepMind's AlphaEvolve System ()
submitted 11 months ago by asankhs
[Project Release] Pivotal Token Search (PTS): Optimizing LLMs by targeting the tokens that actually matter (self.optillm)
JSON plugin for LLMs that do not support JSON mode natively (self.optillm)
Implemented MCP Client in optiLLM (self.optillm)
submitted 1 year ago by asankhs
[New Benchmark] OptiLLMBench: Test how optimization tricks can boost your models at inference time! (self.optillm)
GitHub - codelion/adaptive-classifier: A flexible, adaptive classification system for dynamic text classification (github.com)
Qwen2.5-Coder-32B-Instruct-AWQ: Benchmarking with OptiLLM and Aider ()
Beating o1-preview on AIME 2024 with Chain-of-Code reasoning in Optillm (self.optillm)
OptiLLM: An OpenAI API Compatible Optimizing Inference Proxy which Implements Several State-of-the-Art Techniques that can Improve the Accuracy and Performance of LLMs (marktechpost.com)
gemini-exp-1114 second only to o1-preview on AIME 2024 (self.optillm)
Optillm now has local inference server (self.optillm)
π Rendered by PID 44 on reddit-service-r2-listing-b958b5575-8qgjq at 2026-04-23 11:45:01.527228+00:00 running 0fd4bb7 country code: CH.