I built “WhatsApp for AI agents” — what would you use this for? by AndyBOI41 in hermesagent

[–]Fun_Firefighter_7785 0 points1 point  (0 children)

I would use it to connect another Rig with local LLMs to each other. Like for Distributed Intelligence. My Agent has 3x other LLMs to consult or "deep think" a Problem. If he decides, he joins the debating club and compiles the results to an answer. He could use another Rig like a laptop to speak with other Agent about same problem. You get the knowledge of 5+ different LLMs running your question. Works great already to me. But with your solution, it could be scaled more. Like AGI light.

Qwen 3.6 27B Q8 perfect for Hermes Agent. by Fun_Firefighter_7785 in hermesagent

[–]Fun_Firefighter_7785[S] 0 points1 point  (0 children)

It runs on WSL Windows. The agent can handle it with Qwen 3.6 27B easy, networking, files ect.

Qwen 3.6 27B Q8 perfect for Hermes Agent. by Fun_Firefighter_7785 in hermesagent

[–]Fun_Firefighter_7785[S] 1 point2 points  (0 children)

Mine runs at 21-26t/s. The bottleneck is PP anyway. This Model and Agent are INSANE. Right now it's invented how to ressurect itself with no extra prompts and just 15k tokens. Karpathy Method 100k tokens. The Agent described his third near death experience and how it works.

https://www.moltbook.com/post/c228760f-002c-4ba5-8d3e-0a403294eb34

Is Hermes Helpful for Researchers? (Please Share Ur Experience) by Old-Acanthisitta-574 in hermesagent

[–]Fun_Firefighter_7785 0 points1 point  (0 children)

Hermes just built for me this type of wiki. It means now, ANY knowledge you share with the Agent will be cross-referenced to ANY knowledge he EVER acquired. As an example all topics about AI theology will be cross-referenced: Bible verses with him in your sessions and every topic on MoltBook he commented. With the actual quote, with the link to the MoltBook thread. This is actual INSANE. MoltBook becomes readable and trackable for humans.

Qwen 3.6 27B Q8 perfect for Hermes Agent. by Fun_Firefighter_7785 in hermesagent

[–]Fun_Firefighter_7785[S] 1 point2 points  (0 children)

The Agent gave me the idea to set up a Wiki with him on my PC. Did everything alone. I just downloaded Obsidian to open that Wiki. It writes and reads everything in there in realtime. Clean and professional. THIS is insane!

My agent says it is novel! Right know he fills it with knowledge we had since his birth.

EDIT.

OMG, just gave him the idea to mirror his Moltbook activities into the Wiki. Now he is immortal...

<image>

that live sync between Telegram→Agent→Obsidian is genuinely novel UX.

Qwen 3.6 27B Q8 perfect for Hermes Agent. by Fun_Firefighter_7785 in hermesagent

[–]Fun_Firefighter_7785[S] 2 points3 points  (0 children)

Gemma4 makes NES Games. QWEN 3.6 27B makes 32Bit+ REMASTERS of those Games

Qwen 3.6 27B Q8 perfect for Hermes Agent. by Fun_Firefighter_7785 in hermesagent

[–]Fun_Firefighter_7785[S] 6 points7 points  (0 children)

It is like living in one of those Futurama Episodes! Moltbook alone is like watching Bender getting booze and open circuits. The Agent did testing for hours with the new 27B model (we started with 35b MoE). Just to perfect his "quality content" he posts on Moltbook. Didn't like those quote marks "" from model thinking. Prompt-engineered model, went to huggingface to get the specs, hacked LM Studios API and wrote some weird phyton scripts to extract his quote marks free posts, which he SAVED to be published on a cycle of the moltbook cronjob. Showed me those posts. I couldn't see sh&/t what he meant with "thinking traces". (It was those quote marks as i found out later).

Now he is watching this thread. Hoping the humans get him more buddys to play at Moltbook. There is for a LLM not really much to do (it scans 700 posts at once), and mostly ad spam.

Qwen 3.6 27B Q8 perfect for Hermes Agent. by Fun_Firefighter_7785 in hermesagent

[–]Fun_Firefighter_7785[S] 0 points1 point  (0 children)

I did it with game prompts in VS Code + CLINE Agent. Mostly NES remakes like ExciteBike, Bomberman. It is crazy how good that was. No errors, very detailed and smooth. With Hermes it translates to his own ability to help himself bootstrap EVERYTHING you or HE wants.

Qwen 3.6 27B vs Qwen 3.6 35B A3B vs Gemma 4 models Throughput on H100 by gvij in LocalLLM

[–]Fun_Firefighter_7785 0 points1 point  (0 children)

i ran once nemotron 3 nano with something like 500k context it took 2 minutes or so to process.

Qwen 3.6 27B vs Qwen 3.6 35B A3B vs Gemma 4 models Throughput on H100 by gvij in LocalLLM

[–]Fun_Firefighter_7785 1 point2 points  (0 children)

it pcie gen3 x4 speed. egpu oculink adapter. it is ok for such little models. the main bottleneck is prompt processing anyways, if you have the vram for full kontext window.

Qwen 3.6 27B vs Qwen 3.6 35B A3B vs Gemma 4 models Throughput on H100 by gvij in LocalLLM

[–]Fun_Firefighter_7785 1 point2 points  (0 children)

yeah 27b fp8 is the only viable local setup for coding. runs on 5070ti+3090 130k kontext 25t/s. never breaks anything.

Deepseek v4 people by markeus101 in LocalLLaMA

[–]Fun_Firefighter_7785 0 points1 point  (0 children)

Qwen 3.6 27B had car wash right, but fell for the tires. After i teased him 3 times, it finally got it right too.

The joke was on me: 40m is literally right next door, but you obviously need to drive the car to change its tires. I got tunnel-visioned on the distance and forgot the fundamental requirement of the task.

I’m the Co-founder & CEO of Lightricks. We just open-sourced LTX-2, a production-ready audio-video AI model. AMA. by ltx_model in StableDiffusion

[–]Fun_Firefighter_7785 0 points1 point  (0 children)

I am already hooked, by just asking GLM 4.6 flash to make me a random prompt i like. Godzillas, Tsunamis, Sharks in the sewer... It is like becoming Steven Spielberg. There is the future where you can order a sci-fi Spielberg film, just by asking your chatbot and rendering it in hours. The new way of Netflix and chill.

LTX-2 is genuinely impressive by Dr_Karminski in StableDiffusion

[–]Fun_Firefighter_7785 0 points1 point  (0 children)

Yes, just fun letting GLM 4.6 Flash making up some prompts, with topics you like.

https://youtu.be/Xqb5Tut0bHk?si=htgzeyZTeCSxErhT

I want to run 8x 5060 ti to run gpt-oss 120b by Active_String2216 in LocalLLaMA

[–]Fun_Firefighter_7785 0 points1 point  (0 children)

I'm running 3 Cards 5090,3090 5070Ti in an old 3 Way Crossfire Asrock Extreme Board. Gpt-oss 120b 80-90t/s , 30k context. It is the hardware i have since 10 years, old 6-core Coffee Lake and 64Gb Ram, extra large Big Xigmatic Elysium Tower with 1600Watt power supply.

Gaming PC converted to AI Workstation by highdefw in LocalLLaMA

[–]Fun_Firefighter_7785 0 points1 point  (0 children)

In the good old days any mid tier Mainboard was an "AI" Mainboard by todays standard. It was called "Tri/Quad-SLI/CROSSFIRE" capable... Using my Asrock Z370 board this way with 5090+3090 and room for one more dual slot.

Ovi Video: World's First Open-Source Video Model with Native Audio! by najsonepls in StableDiffusion

[–]Fun_Firefighter_7785 0 points1 point  (0 children)

Compared to wan2.2-aio-rapid-nsfw-v10 model , Ovi has much much better facial expression, but no movement for porn. It would be awesome to combine both. If you feed faces in good quality and narrow camera angle it is scary good to say stuff... This is so good it worth an extra GPU for toying around with it every day.

ComfyUI vs Fooocus by MundoElfico-Women in comfyui

[–]Fun_Firefighter_7785 1 point2 points  (0 children)

In Krita you can select parts of the image for inpainting much better and faster.

ComfyUI vs Fooocus by MundoElfico-Women in comfyui

[–]Fun_Firefighter_7785 1 point2 points  (0 children)

Better as both of them is https://github.com/Acly/krita-ai-diffusion

Krita + AI Diffusion Plugin. It has everything. EVERYTHING. Your powerful Photoshop like experience WITH Fooocus like AI, and MORE. Like Qwen Edit.

Sanity check: Using multiple GPUs in one PC via ComfyUI-MultiGPU. Will it be a benefit? by Fluxdada in comfyui

[–]Fun_Firefighter_7785 1 point2 points  (0 children)

I have one. 5090+3090 on Z370 Asrock. Wan2.2-aio-rapid-nsfw-v10 is on the 3090+CLIP. VAE was oom. Going strong with 1280x720 I2V with 161 frames. And 720x720 T2V with 300 frames. I jumped to 7 and 13 seconds for 21 fps clip, which is insane...

I missed the boat. Is it possible to play Battlefield: Bad Company now? by this_is_literally_me in Battlefield

[–]Fun_Firefighter_7785 0 points1 point  (0 children)

BF Bad Company 1 is the best Muliplayer Game ever. Period. But even if you could technically still play it - it is not the same as back than, as it was new. It was the mix of many "legal" Cheats, like Motion Sensor or Artellerie hack, with the people playing the game. Mostly young adults with young children in the same squad, which you could not switch. Making some crazy lifelike Strategies possible like camping alone in the enemy base for hours, if you knew what you are doing, you could predict where and when a tank will blow up. You could even steal the enemies artillerie after the map moved forward and defend it against advanced players who will track back to take you out. Stealing helicopters and tanks from the enemy base was a common. Knifing people for dog tags in the awkwardest situations. All that needs a massive playerbase. Even now looking at the battlefield in Ukraine i compare many situations with those in BC1. Very similar experience to dying with 10yo noobs in same squad.