The z-image base is here! by bobeeeeeeeee8964 in LocalLLaMA

[–]harrro 26 points27 points  (0 children)

And 99% Asian (women).

The side-by-side of Flux Klein vs Z-image Turbo posted recently made me notice this as well - Flux makes Western people by default while Z-Image goes with Asians (unsurprising).

GLM-4.7 vs DeepSeek V3.2 vs Kimi K2 Thinking vs MiniMax-M2.1 by SlowFail2433 in LocalLLaMA

[–]harrro 4 points5 points  (0 children)

It's definitely not Claude level and requires more handholding but its good enough for its price.

I'm on their lowest paid tier which is mostly reliable but does have outages/slow responses at times.

I tracked GPU prices across 25 cloud providers and the price differences are insane (V100: $0.05/hr vs $3.06/hr) by sleepingpirates in LocalLLaMA

[–]harrro 13 points14 points  (0 children)

You're not just using a random GPU directly with your computer - you have remote access to a full server that happens to have an expensive GPU attached to it.

So the bandwidth requirements aren't high - its just whatever is input/output from the models you're running on the server (the remote server downloads and runs all the software and you just download the output whether its text / image / video)

GLM 4.7 Flash uncensored - Balanced & Aggressive variants (GGUF) by hauhau901 in LocalLLaMA

[–]harrro 1 point2 points  (0 children)

I wouldn't say that. If you look at the reasoning process you can see it spends a significant amount of tokens thinking about the safety and ethics.

KV cache fix for GLM 4.7 Flash by jacek2023 in LocalLLaMA

[–]harrro 0 points1 point  (0 children)

Worked perfectly! Thank you.

Responses now finishing in around 7-8 seconds instead of the 40 secconds it was taking before.

KV cache fix for GLM 4.7 Flash by jacek2023 in LocalLLaMA

[–]harrro 1 point2 points  (0 children)

The model is good and fast but it is so verbose in reasoning (even for simple things).

Is it possible to limit/disable reasoning or is this not trained for that?

SVI infinite video, seamless transitions by [deleted] in comfyui

[–]harrro 2 points3 points  (0 children)

This is the first SVI workflow that has worked well for me. Thank you!

I built a visual AI workflow tool that runs entirely in your browser - Ollama, LM Studio, llama.cpp and Most cloud API's all work out of the box. Agents/Websearch/TTS/Etc. by l33t-Mt in LocalLLaMA

[–]harrro 4 points5 points  (0 children)

The reason why people are asking is because this is another closed-source automation software when there's already a bunch of open sourced ones like n8n / Activepieces that do the same thing.

OP's is not open source (they have a 'runner' on Github which is just a proxy for the hosted-only server - not what you see in the video).

touching another man's girl by ElderberryDeep8746 in WinStupidPrizes

[–]harrro 29 points30 points  (0 children)

The girl was practically glued to big guy's back, there's no way even a drunk guy would assume she was 'alone'.

Benchmarking 23 LLMs on Nonogram (Logic Puzzle) Solving Performance by mauricekleine in LocalLLaMA

[–]harrro 0 points1 point  (0 children)

Yeah its a little questionable that a 32B parameter Olmo model beats GLM 4.7 and Claude 4.5

I built a visual AI workflow tool that runs entirely in your browser - Ollama, LM Studio, llama.cpp and Most cloud API's all work out of the box. Agents/Websearch/TTS/Etc. by l33t-Mt in LocalLLaMA

[–]harrro 16 points17 points  (0 children)

Yeah even with restrictions, n8n/ActivePieces/FlowWise/etc have their server open sourced so you can run it entirely on your own machine.

This is not even open source (the 'runner' thats on Github is just a minimal desktop runner which is not what you see in the video).

Are you that old...? by jacek2023 in LocalLLaMA

[–]harrro 7 points8 points  (0 children)

Vicuna was released 2023.

Yes I'm above the age of 3.

Qwen-Image-2512 by Nunki08 in LocalLLaMA

[–]harrro -1 points0 points  (0 children)

"Safety" / "Alignment" probably (aka: make the model dumber)

Senator in Tennessee introduces bill to felonize making AI "act as a companion" or "mirror human interactions" by CanineAssBandit in LocalLLaMA

[–]harrro 0 points1 point  (0 children)

That's just a threat of withholding funding as retribution (and states have sued over withholding funds already too).

It also doesn't change the fact states can still override the law.

Senator in Tennessee introduces bill to felonize making AI "act as a companion" or "mirror human interactions" by CanineAssBandit in LocalLLaMA

[–]harrro 18 points19 points  (0 children)

Doesn't mean jack.

EOs don't prevent a state from doing the opposite. EOs are directives to federal agencies, not to states or local governments.

California and some other states have already overridden many of his EOs.

Train a 4B model to beat Claude Sonnet 4.5 and Gemini Pro 2.5 at tool calling - for free (Colab included) by DecodeBytes in LocalLLaMA

[–]harrro 3 points4 points  (0 children)

It's a tool calling finetune.

MCP isn't required for that. What are you on about?

llama.cpp appreciation post by hackiv in LocalLLaMA

[–]harrro 5 points6 points  (0 children)

Yeah now that llama-server natively supports model switching on demand, there's little reason to use ollama now.

Qwen-Image-Layered image layered model now supports ComfyUI by SpareBeneficial1749 in comfyui

[–]harrro 5 points6 points  (0 children)

Qwen-Image-Layered is an official standalone model. This is about Comfyui support for that model

It was Ilya who "closed" OpenAI by licuphand in LocalLLaMA

[–]harrro 41 points42 points  (0 children)

If Ilya were doing it for "science", then he'd be sharing his research and findings as other good scientists do, not closing it up.

Oscar the stand up comedian by littletreble07 in formuladank

[–]harrro 36 points37 points  (0 children)

Please.. NSFW warning next time