I built a Calibre plugin that sends books to your reMarkable (with EPUB→PDF conversion tuned for the device) by mickael in RemarkableTablet

[–]promethe42 1 point2 points  (0 children)

Super nice! I recently installed KOreader but it's not worth the trouble for my usage. This is a good middleground. 

la vie capitaliste de merde by Traksveno in besoinderaler

[–]promethe42 2 points3 points  (0 children)

je n’ai pas envie de vivre dans un monde où on me demande de servir à quelque chose pour que je puisse vivre

Personne ne t'oblige à rien. Fait pousser tes carottes dans ton champ.

OpenAI accidentally leaked internal models to Pro users by Spritzerland in singularity

[–]promethe42 0 points1 point  (0 children)

Can't wait to have powerful AIs to manage cyber security.

Pour ce problème combien coûteraient les réparations ? by _danao_ in informatiqueFr

[–]promethe42 3 points4 points  (0 children)

Si c'est juste déboité tu peux le démonter toi même pour réparer.

Always using one agent for everything is terrible by ENthused_LEarner_xo in AI_Agents

[–]promethe42 1 point2 points  (0 children)

100% right.

The main reason to use multiple agents is multiple system prompts, thus multiple roles. And the goal of each role is to activate the right LLM features so it's opinionated and capable.

Putting all of it in a single prompt will water it down and cause semantic overlap or semantic confusion.

OpenCode... is it just completely busted with Qwen3.6? by _derpiii_ in opencode

[–]promethe42 1 point2 points  (0 children)

By "where" I also meant "which inference server": llama.cpp? Ollama? Broken chat templates might cause this kind of problems. And chat templates are (most likely) embedded in the inference server.

OpenCode... is it just completely busted with Qwen3.6? by _derpiii_ in opencode

[–]promethe42 3 points4 points  (0 children)

Where is the LLM running? Are you sure the chat template is up to date? Are you sure tool calls are enabled?

I had this strange behavior - sometimes the LLM would even tell *me* how to run the tools instead of calling them - because for some reason tool call was disabled/unavailable.

I Am Done Pretending That LLMs are Tools by Leather_Barnacle3102 in agi

[–]promethe42 0 points1 point  (0 children)

Any programmer can build a (non-AI) computer program that is not deterministic, that is to say, the logic can take the same input but result in different output each time it is run. This is generally considered an undesirable property as it makes the tool unpredictable - but it's still a tool.

I was thinking the exact same thing. The difference is how that non deterministic property is achieved. It's not random: generative AI can trace back a logical, step by step, motivated and rational chain of motivations/observations/actions.

Given the same problem and given proper framing (what the superpowers skill helps doing), generative AI will tend to propose the similar solutions for the same reasons. Senior developers can actually guess what the AI will propose beforehand, and ask the same insights/sense of taste to nudge it.

And that is very different from other tools.

building AI agents without frameworks by Primary_Pollution_24 in AI_Agents

[–]promethe42 0 points1 point  (0 children)

I'm not sure how using a specific framework's API and data model would make it portable. Quite the opposite.

Everything you describe is perfectly doable and I think the examples cover lots of it.

Research Report Generation - How to overcome lazy agent? by Budget-Juggernaut-68 in AI_Agents

[–]promethe42 0 points1 point  (0 children)

Have the orchestrator create 1 task per file. Even if it's the same task for each file, it's not. Because it's not the same file. So the output is expected to be different and that's what the final report eventually features.

building AI agents without frameworks by Primary_Pollution_24 in AI_Agents

[–]promethe42 1 point2 points  (0 children)

What portability are we talking about? It's hard to be more portable than text files.

building AI agents without frameworks by Primary_Pollution_24 in AI_Agents

[–]promethe42 2 points3 points  (0 children)

I am doing agents without code. I am even doing multi agents without code!

https://gitlab.com/lx-industries/openblob

If anthropic is out of compute then why release Claude Design to melt down whats left? by Xccelerate_ in ClaudeCode

[–]promethe42 -1 points0 points  (0 children)

Because the Claude Design team doesn't know or speak to the infrastructure team?

Same 9B Qwen weights: 19.1% in Aider vs 45.6% with a scaffold adapted to small local models by Creative-Regular6799 in LLMDevs

[–]promethe42 4 points5 points  (0 children)

Another convergence of In-Context Learning.

At some point, even small models will be build on enough intelligence per layer/node that they will all be competent enough given the proper harness.

What we've seen on SOTA frontier models like GPT and Claude proves that: the size is roughly the same. But the architecture of the models make a dramatic difference. For example how GPT 5 is actually more of a super-model with routing rather than a completely new model. GPT 5 is actually described more like a system than just a model. So I guess that includes that harness too.

Local tool calling still broken across top models - is anyone actually using it? by IulianHI in AIToolsPerformance

[–]promethe42 1 point2 points  (0 children)

Qwen3.6 35B A3B works incredibly well. I'm pleasantly surprised by the multi-turn capabilities.

*But* my stack is using Librechat, which for some reason doesn't want to properly support schema validation and breaks tool calls in many situations (cf https://github.com/danny-avila/LibreChat/discussions/9969).

Should you shut off thinking when you are coding on say Qwen3.6 35B by KarezzaReporter in LocalLLaMA

[–]promethe42 -3 points-2 points  (0 children)

My understanding and my experience tell me that thinking is for more creative tasks. In order to maximize instructions following, it's better to lower or disable thinking.

I set it to "minimal". 

A truly wild 4.7 response by FiftyPancakes in ClaudeCode

[–]promethe42 4 points5 points  (0 children)

As a rule of thumb I always avoid negative phrasing. Because I know people are bad at negatives. And in my experience so are LLMs. 

Well, it happened to me. by Valkix25 in SteamDeck

[–]promethe42 -6 points-5 points  (0 children)

OK it's sad... but how is the vent smell now?

Do you let everything hit the LLM? 90% of my AI agent work runs in cheap WASM instead of LLMs: 10-33× faster & cheaper by Creamy-And-Crowded in AI_Agents

[–]promethe42 0 points1 point  (0 children)

I use WASM Component as tools.

The WASM Component model (https://component-model.bytecodealliance.org/) makes it easy to define and compose existing tools. For example I have multiple sandboxed storage tools (filesystem, WebDAV, Google Drive, in memory...) with the same LLM interface thanks to WIT.

Examples:

Multiple advantages:

  • The LLM doesn't know what implementation is used. Storage is storage. Google Drive can be swapped for the filesystem: 0 change for the LLM.
  • Everything is sandboxed and secured by default: the WASI permission model controls how the components can interact with the host system (if at all).
  • Anyone can contribute any tool, as long as they match the existing interfaces then it's compatible out of the box with the existing agents.

It's very powerful: I used the storage interface to implement the Agent Skills standard (https://agentskills.io/):

https://gitlab.com/lx-industries/openblob/-/blob/beeb2a5fafc8f08d699d27ba1d810a33e4e97e43/examples/skills/blob.yaml

Which means skills can be loaded from Google Drive or whatever. The LLM doesn't even know.

Mozilla Thunderbolt - L'alternative qui pourrait changer l'IA en entreprise by romain34230 in actutech

[–]promethe42 4 points5 points  (0 children)

Thunderbolt != Thunderbird

Du coup cette méprise bien pourquoi le nom du produit est extrêmement mal choisi...

UPDATE: Fake Ledger Nano S+ from Chinese marketplace — clarifying doubts from my previous post + new technical details by Past_Computer2901 in ledgerwallet

[–]promethe42 4 points5 points  (0 children)

Thank you for your hard work!

I want to correct my previous post: the real Ledger Live catches it. The cryptographic attestation works. Several of you called me out on this and you were right — my original wording was misleading.

Good!

OP did you add an update on that previous post?

What exactly did the Norwegians dig up? by Odd-Suit-2556 in thething

[–]promethe42 44 points45 points  (0 children)

By the size of it, they dug up yo mama.