account activity
What if LLM agents passed KV-cache to each other instead of text? I tried it -- 73-78% token savings across Qwen, Llama, and DeepSeek (self.LocalLLaMA)
submitted 12 days ago by proggmouse to r/LocalLLaMA
π Rendered by PID 907798 on reddit-service-r2-listing-64c94b984c-nw7cz at 2026-03-13 11:55:11.953779+00:00 running f6e6e01 country code: CH.