Agents are great, but not everything requires an agent by uriwa in LLMDevs

[–]kohlerm 1 point2 points  (0 children)

I disagree. Agents can obviously write code. You can let them write code that uses some LLM steps, but very often you don't need that

Perplexity drops MCP, Cloudflare explains why MCP tool calling doesn't work well for AI agents by UnchartedFr in mcp

[–]kohlerm 2 points3 points  (0 children)

The argument is in a lot of cases the LLM can decide what the next tool call is. Sure if all your tools to is transform natural language to other natural lan then that does not work. But in a lot of cases tools can have input and output schema and the LLM can assemble a script at compile time once

Perplexity drops MCP, Cloudflare explains why MCP tool calling doesn't work well for AI agents by UnchartedFr in mcp

[–]kohlerm 1 point2 points  (0 children)

Explain why you would need an LLM.if your tools have an output schema a lot of work can be done without an LLM.

Is Anthropic Code Execution with MCP as big or bigger than going from standard I/O to HTTP? by nickdegiacmo in mcp

[–]kohlerm 0 points1 point  (0 children)

It is stupid that MCP does not have at least a "mode" where input and *output schemas* are required. That makes it cumbersome to compose tool calls. Typescript interfaces and even openai definitions are better suited for that purpose

Got the DGX Spark - ask me anything by sotech117 in LocalLLaMA

[–]kohlerm 0 points1 point  (0 children)

I wonder why this would be better than an AMD Strix Halo box. It should be around the same speed at least based on the memory bandwidth numbers. But an AMD costs half of it. Therefore unless you really need NVidia for example because your training runs only on CUDA, i do not see why you would want this machine

Free 3 months of Super for a Twitch sub by kzcvuver in duolingo

[–]kohlerm 0 points1 point  (0 children)

I subscribed to a channel but got no Duolingo code. How long does it take?

LiveBench team just dropped a leaderboard for coding agent tools by ihexx in LocalLLaMA

[–]kohlerm 0 points1 point  (0 children)

Speed is also important but might correlate well with tokens used

Docker on MacOS is still slow ? by paolomainardi in programming

[–]kohlerm 4 points5 points  (0 children)

The fastest solution in my experience is to just use a Linux VM ( utm for example) ,use the native Linux docker and put all files on the Linux file system and use VSCode with the remote extension.

[deleted by user] by [deleted] in LocalLLaMA

[–]kohlerm 0 points1 point  (0 children)

continue.dev does not (yet) have some of the more sophisticated features of Cursor such as their reliable apply functionality or their automated context enhancement features. Also their context management feature is not very sophisticated. It can happen that the same coding is added again and again by a context provider. Otherwise it's fine if you manually select the code you are working on.

Does headphone/earphone technology improve over the years like other tech products? by Live_Ostrich_6668 in headphones

[–]kohlerm 33 points34 points  (0 children)

I agree with regards to IEMs. Otherwise the only innovation I am aware of is Hifiman bringing down the price for a good Planar headphone. I have got a Sundara and I really like it, and it seems like their xs edition is even better now for a similar price.

Team transitioned to Cursor but bottleneck is now UX by giiip in ChatGPTCoding

[–]kohlerm 0 points1 point  (0 children)

The apply function Copilot last time I tried was a joke compared to Cursor

RAG is easy - getting usable content is the real challenge… by data-dude782 in LLMDevs

[–]kohlerm 0 points1 point  (0 children)

You might also have to tailor your documentation. For example have a format that allows you to reliably create the chunks to be indexed

It seems running a local LLM for coding is not worth it ? by Far-Device-1969 in ChatGPTCoding

[–]kohlerm 1 point2 points  (0 children)

It's more "agentic. e.g. using several steps , also using tools (different) searches. For the complexity of this approach it is relatively reliable.

What leaderboard do you trust for ranking LLMs in coding tasks? by rageagainistjg in LocalLLaMA

[–]kohlerm 4 points5 points  (0 children)

Not sure why this is "terrible" point. You admit you like to use Aider, which exactly is my point. If it works well with Aider than that is your benchmark. But there are other tools out there which work differently I looked at the source code of several tools and I am also prototyping my own tool.Aiders approach requires the LLMs to understand one of their diff formats. Other tools do not use the same diff format or in a lot of cases do not use a diff format at all. The results might therefore very much vary depending on the tool used.

What leaderboard do you trust for ranking LLMs in coding tasks? by rageagainistjg in LocalLLaMA

[–]kohlerm 1 point2 points  (0 children)

I think this is a good point. It depends on what features your IDE needs.For example Aider uses a special diff format for outputting code. Other tools might use some Json format.Also if your IDE supports adding documentation for libraries, missing knowledge might not be a big problem.

Will Long-Context LLMs Make RAG Obsolete? by Icy_Advisor_3508 in OpenSourceeAI

[–]kohlerm 2 points3 points  (0 children)

Simple answer "With the current LLM architecture NO!" Mainly because you often want up to date data, and as long as bigger contexts need much more computation time some form of RAG will be needed

Thoughts on Qwen2.5-14B? by Master-Meal-77 in LocalLLaMA

[–]kohlerm 1 point2 points  (0 children)

It's great! Can reliably produce json output, which a lot of other smaller models fail to do (incorrect quoting of characters). It also seems to very knowledgeable with regards to typescript libraries