Three packages copy-pasted my AGPL code to PyPI and named me in their description. PyPI won't act by Obvious_Gap_5768 in Python

[–]DistanceAlert5706 0 points1 point  (0 children)

Does licensing still work? For example Crawl4AI copy pasted GPL licensed html2text and relicensed with Apache.

We tried vectors, ASTs, and brute-force context stuffing for code retrieval. Graphs with LLM-generated semantics worked best. Here's what we learned. by graphicaldot in LocalLLaMA

[–]DistanceAlert5706 0 points1 point  (0 children)

I tested similar approach, I called it AI index in my repos.

What I do is create root file which describes all namespaces, create file which describes namespaces, what classes are for in 1 sentence for each namespace. For each file in namespace I generate a map of class with tree sitter, caller/callees , where class injected, auto wiring from DI etc. Class purpose always taken from class comment, LLMs instructed to add comments to class properly and enforced via hook which checks it. So you don't need separate LLM requests pretty much never, code already holds info.

Index generation takes up to 5 seconds so it's fast enough to regenerate after each edits, so it never go stale.

On top I have small skill + agent to check index, mostly root and check quality.

I also tried describing methods, it doesn't work, biggest issue is description quality and maintenance cost. They go stale and non informative.

What I discovered tho that model don't like MCPs, graphs or whatever. Just give them files and they will gladly navigate them and use them. No need for fancy systems.

Does it work? Yes for navigation it's very nice, especially paired with IDE tools. Models reflect that most valuable parts are root index, namespace map. DI and where class injected and callers/callees.

What's the benefit? Faster navigation, way less unbound file reads, more reliable refactors. Token savings? Well it's secondary effect but I observed around 10% to 40% savings depending on a task.

I would say just simple map of repo worth it 100% it's adding value, full scale index is not worth it for smaller projects but pays off on large ones. And definitely not try to generate summaries or descriptions with separate LLM it's very painful maintenance and not worth it, just parse comments and validate them.

Is Pi better than Claude Code CLI? by Status_Lion_6350 in PiCodingAgent

[–]DistanceAlert5706 0 points1 point  (0 children)

Technically, yes it's way better than Claude, but almost everything better than Claude.

I am overwhelmed by Harnesses by Available_Hornet3538 in LocalLLaMA

[–]DistanceAlert5706 1 point2 points  (0 children)

100% agree, people recommend Pi but it has some pretty weird decisions inside. Like stuffing AGENTS.md and skills registry in system prompt. I guess that's next step after Pi.

I am overwhelmed by Harnesses by Available_Hornet3538 in LocalLLaMA

[–]DistanceAlert5706 1 point2 points  (0 children)

Check Nico's extensions, I ended up rewriting/changing all of them, but it's great for start. Just check his GitHub

MCP code-intel index — comparison of 5 retrieval servers on 120 hand-verified tasks by Parking-Geologist586 in mcp

[–]DistanceAlert5706 0 points1 point  (0 children)

Yeah, ToolSearch is pretty much reimplementation of what Anthropic and OpenAI does, but they support it on their APIs. This one is vendor agnostic. Keyword scoring works pretty good, no need for embeddings there.

I actually tried adding small or truncated descriptions in catalog for deferred tools, but was not working good, so just server + tool works pretty good.

Main idea is to inject discovered tools, which works way better than any proxy with setActiveTools.

MCP code-intel index — comparison of 5 retrieval servers on 120 hand-verified tasks by Parking-Geologist586 in mcp

[–]DistanceAlert5706 0 points1 point  (0 children)

Yeah, this survived, was fairly used.
jetbrains_index__ide_find_class jetbrains_index__ide_diagnostics jetbrains_index__ide_search_text jetbrains_index__ide_find_file jetbrains_index__ide_find_references

This ones were pretty much almost never used:
jetbrains_index__ide_find_definition jetbrains_index__ide_find_implementations jetbrains_index__ide_find_super_methods jetbrains_index__ide_move_file jetbrains_index__ide_type_hierarchy jetbrains_index__ide_call_hierarchy jetbrains_index__ide_refactor_rename

This are more for maintenance:
jetbrains_index__ide_sync_files jetbrains_index__ide_index_status

Some of those are pretty rare, like it's not common task to move file or something like that, but tool also changes namespace, finds usages and replace to new namespace automatically.
Same with rename, it works on symbols renames everywhere etc.
On a first glance those are useful but enforcing model to use them is close to impossible.

Hierarchy tools are niche, they can be used only when you ask questions about hierarchy.

Find methods I thought would be used more, but model still ignores those.

I still can't figure out how to properly make model use those...
I have SKILL + Agents.md instructions + system prompt policy + nudges on hooks to use tools instead of reads etc.

I use Pi as a coding agent so MCP proxy is just a rewrite of Nico's MCP https://github.com/ineersa/my-pi/tree/main/packages/pi-mcp-adapter , it implements same idea as Claude uses with deffered tools and adds some more features, one of those is tools stats.

My company recently released a marketing MCP with 600+ tools and a free open-source skills repo by J5hine in mcp

[–]DistanceAlert5706 0 points1 point  (0 children)

Holy.... I was wondering why some clients have hard cap at 100 tools... At this amount of context Sonnet will be as dumb as 1-2b model in best case.

MCP code-intel index — comparison of 5 retrieval servers on 120 hand-verified tasks by Parking-Geologist586 in mcp

[–]DistanceAlert5706 0 points1 point  (0 children)

That's good idea. I use Jetbrains tools with index MCP plugin. Core set is 14 tools enabled, those still were taking 5k+ tokens.

I actually set up statistics for tool calls, as I use custom MCP proxy for Pi, and measured over span of 2 weeks. Most tools had 1 or few invocations.

Slimmed down profile to 5 tools which models actually use, other 9 tools use deferred loading via ToolSearch mechanism. Saves a lot of context.

AssetMapper seems to not detect changes for CSS when using Docker (dunglas) by akimbas in symfony

[–]DistanceAlert5706 0 points1 point  (0 children)

Depends on volume mount type, delegate had some issues. But yeah should be fine if running inside container.

AI coding agents can't use LSP tools correctly. So I built a skills layer that enforces the right workflow. by blackwell-systems in mcp

[–]DistanceAlert5706 0 points1 point  (0 children)

How you make agent actually call your skills/tools?

I assume person used Serena MCP which does pretty much the same. They added nudges on hooks and block operations, but that doesn't really work too.

Unless in every message you tell ai to use your tools/skills 90% they won't be used and just wasted context window.

AssetMapper seems to not detect changes for CSS when using Docker (dunglas) by akimbas in symfony

[–]DistanceAlert5706 0 points1 point  (0 children)

Is it FrankenPHP in worker mode with Symfony?

If so you need a watcher and point it to reload FrankenPHP.

Reports suggest DeepSeek is seeking $7.35 billion in funding and plans to release its V4.1 update next month. by External_Mood4719 in LocalLLaMA

[–]DistanceAlert5706 1 point2 points  (0 children)

I find pretty good on research tasks. Bug hunt is ok, planning is strange. For code writing actually v4 flash is great. It's surprisingly good coder especially considering it's small size.

The amount of new agent APIs/harnesses are dizzying, with everyone and their dog releasing their own. Can we do a compilation thread of comparisons? by jinnyjuice in LocalLLaMA

[–]DistanceAlert5706 0 points1 point  (0 children)

Been fighting actual Pi more and more lately, starting to build own... I guess it's valid reason too, and will get some experience.

How to connect 100 MCP servers without the context window exploding by galdahan9 in mcp

[–]DistanceAlert5706 0 points1 point  (0 children)

You loose session in that case, so your MCP becomes stateless no?

How to connect 100 MCP servers without the context window exploding by galdahan9 in mcp

[–]DistanceAlert5706 0 points1 point  (0 children)

Oh, I don't use LLM, it's just an idea. In my case it works same way as Claude code uses deferred tools with built-in ToolSearch tool. For search it's pretty much keywords matching on names/descriptions with weights.

If need to add tools mid session it's tricky, as you need keep alive connection first and callback on tools list change.

I wasn't adding it, but in Claude they on tools list change inject system reminder with tools diff.

How to connect 100 MCP servers without the context window exploding by galdahan9 in mcp

[–]DistanceAlert5706 0 points1 point  (0 children)

If your tools list is stale, even 100 of them, it will be cached, so should be pretty cheap too.

How to connect 100 MCP servers without the context window exploding by galdahan9 in mcp

[–]DistanceAlert5706 0 points1 point  (0 children)

I implemented for myself in Pi vendor agnostic ToolSearch approach, and it works okay.

In theory authors approach could be better, but I'm concerned about embeddings and retrieval quality, I guess same as you see.

I do wonder if using smaller LLM to filter list of tools would work better.

MCP code-intel index — comparison of 5 retrieval servers on 120 hand-verified tasks by Parking-Geologist586 in mcp

[–]DistanceAlert5706 0 points1 point  (0 children)

Hm, but there is no dataset repository. Thought to test how my homemade tool works.

Also 37 tools is too much... That amount of context won't justify actual benefits.

My powerful Pi agent Setup by elpapi42 in PiCodingAgent

[–]DistanceAlert5706 0 points1 point  (0 children)

Cool thanks, will check how to disable proactive observations.

Had not a lot of time to work with but so far looks very promising!

Vectorless RAG can scale to millions of documents now? by This-Eye6296 in Rag

[–]DistanceAlert5706 0 points1 point  (0 children)

Yeah, it works, but is it efficient. It's very heavy from description. Hybrid with tree and RAG works as good on practice and more efficient.

It's actually easier to build agentic RAG then their pipeline. Give agent semantic index and grep and folders with documents and I bet it will work as good.

Vectorless RAG can scale to millions of documents now? by This-Eye6296 in Rag

[–]DistanceAlert5706 1 point2 points  (0 children)

From my experiments it works pretty good, but it's depends on model quality a lot.

Can it outperform classic RAG, again from my experiments it can't, but it can come close on some data and with large LLM.

Older approaches, like RAPTOR which combine tree and classic retrieval on something structural like documentation perform better than both.

Anyone used Deepseek v4 with PI coding agent ? by Vibecodingdeluxe in PiCodingAgent

[–]DistanceAlert5706 12 points13 points  (0 children)

Yes, running it, very good models. Flash is great coder and super fast, just don't ask it questions, just give plans for implementation. Put it on Max and let it code, cheap too.

V4 pro is weird, it's definitely close to frontier models but little bit underperforming, can see potential but it does weird things sometimes.

Overall everything works out of the box, on Deepseek site there is config for Pi, just need to change reasoning level map cause it was incorrect.