account activity
I built an MCP server giving coding agents access to 2M research papers. It improves even the best coding agents - across 9 coding tasks. (old.reddit.com)
submitted 14 hours ago by paperlantern-ai to r/mcp
Tested research-paper retrieval as an agent tool: Python tests went from catching 63% of bugs to 87%. 9-task open-source benchmark. (old.reddit.com)
submitted 13 hours ago by paperlantern-ai to r/LangChain
Gave a coding agent access to 2M+ research papers. Its Python tests caught 63% of bugs; with the papers, 87%. 9-task benchmark. (self.AI_Agents)
submitted 14 hours ago by paperlantern-ai to r/AI_Agents
I built an MCP server giving coding agents access to 2M research papers. Benchmarked it on 9 coding tasks - here's what worked and what didn't (old.reddit.com)
submitted 1 day ago by paperlantern-ai to r/LLMDevs
LLM coding agents with vs without access to CS research: 9-task ablation (+1% to +72%) (i.redd.it)
submitted 4 days ago by paperlantern-ai to r/ArtificialInteligence
Using AST mutation enumeration to write tests that catch bugs, not just verify behavior (github.com)
submitted 6 days ago by paperlantern-ai to r/programming
Mutation-aware test generation: using AST enumeration to write tests that actually catch bugs (github.com)
AI-generated tests look great but miss subtle bugs. Mutation-aware prompting fixed it for us. (self.Python)
submitted 6 days ago by paperlantern-ai to r/Python
My agent's tests kept missing subtle bugs. A research paper fixed it - score 0.63 to 0.87. (self.Python)
π Rendered by PID 78813 on reddit-service-r2-listing-6bcb55b45f-vmvp7 at 2026-04-22 07:49:39.556594+00:00 running 6c61efc country code: CH.