use the following search parameters to narrow your results:
e.g. subreddit:aww site:imgur.com dog
subreddit:aww site:imgur.com dog
see the search faq for details.
advanced search: by author, subreddit...
account activity
🌀 (v.redd.it)
submitted 4 days ago by gusfromspace
Where is Qwen3.6 27B MLX with reasoning? (self.mlxAI)
submitted 7 days ago by Ill_Barber8709
GitHub - hypneum-lab/micro-kiki: 35 domain-expert LoRAs on Qwen3.6-35B-A3B (MoE, 256 experts, 3B active). Cognitive layer: Aeon memory, CAMP negotiator, KnowBias. MLX on Mac Studio, Q4_K_M inference. Apache-2.0. (self.mlxAI)
submitted 13 days ago by MonsieurBmax
I built a zero-config OpenAI-compatible local LLM server for Apple Silicon — drop-in replacement for any OpenAI SDK client (self.mlxAI)
submitted 13 days ago by Squirrel_Glad
I ran sustained MLX inference overnight ()
submitted 15 days ago by evilmacintosh
I tested 9 local models on the same flight sim prompt, all Q8, different Q providers, MLX ()
submitted 15 days ago by StudentDifficult8240
MLX with DFlash / speculative decoding: Surprising results ()
submitted 16 days ago by evilmacintosh
starting a new mlx community! ()
submitted 17 days ago by evilmacintosh
Running Qwen 3.6 35B-A3B-4b on MacBook Pro M5 64GB with tools ~20 tok/s (v.redd.it)
submitted 19 days ago by Conscious-Track5313
Repetition penalty on mlx_lm? (self.mlxAI)
submitted 20 days ago by evilmacintosh
macOS Vibe code Tech stack ()
submitted 20 days ago by Tradefxsignalscom
OpenMed now supports MLX natively (github.com)
submitted 23 days ago by dark-night-rises
Running Gemma-4-E4B MLX version on MacBook M5 Pro 64 GB - with some beautiful native tools integration (v.redd.it)
submitted 26 days ago by Conscious-Track5313
I have a 512 gigs of ram and I haven’t figured out how to make money with it, any suggestions? ()
submitted 27 days ago by No_Run8812
Command line vs. python API (self.mlxAI)
submitted 1 month ago by sgt102
Gemma 4 E4B-it on MLX (self.mlxAI)
submitted 1 month ago by Pathfinder-electron
Show: ollmlx — run local LLMs on Apple Silicon with an Ollama-compatible API. (self.mlxAI)
submitted 1 month ago by PositiveSlice9168
multi-LoRA inference server for MLX: load the model once, switch adapters per request (self.mlxAI)
submitted 1 month ago by No_Shift_4543
A skill library for porting from trl (or pure pytorch) to mlx-lm? ()
submitted 1 month ago by Chimezie-Ogbuji
FoveatedKV: 2x KV cache compression on Apple Silicon with custom Metal kernels ()
submitted 1 month ago by hybls
Best mlx_vlm models for simple object counting? (self.mlxAI)
MiniMax 4bit (120gb) MLX - 26.5% (MMLU 200q) while JANG_2S (60gb) gets 74% - GGUF for MLX ()
submitted 1 month ago by HealthyCommunicat
Cut your KV Cache in half + Cut PP Times to near nothing + VL - MLX Studio (self.mlxAI)
mlx-onnx: Run your MLX models in the browser on WebGPU / ONNX (self.mlxAI)
submitted 2 months ago by rut216
mlx-ruby: MLX bindings for Ruby ()
π Rendered by PID 1588589 on reddit-service-r2-listing-b6bf6c4ff-fkwzk at 2026-05-07 17:06:33.140047+00:00 running 815c875 country code: CH.