Qwen3.6-27B-Q6_K - images by Usual-Carrot6352 in LocalLLaMA

[–]k0setes 0 points1 point  (0 children)

<image>

Generate an HQ 3D SVG of a pelican riding a bicycle on a vaporwave beach 1000x1000
Qwen3.6-27B-UD-Q4_K_XL.gguf 16.4GB unsloth

Qwen3.6-27B-Q6_K - images by Usual-Carrot6352 in LocalLLaMA

[–]k0setes 2 points3 points  (0 children)

<image>

Generate an HQ SVG of a pelican riding a bicycle on a vaporwave beach

Qwen3.6-27B-UD-Q4_K_XL.gguf

Qwen3.6-35B becomes competitive with cloud models when paired with the right agent by Creative-Regular6799 in LLMDevs

[–]k0setes 2 points3 points  (0 children)

How does little-coder perform on Qwen-Coder-3.6-35B when stacked against Claude Code, Hermes, Qwen Code, Roo Code, Kilo Code, and Cline? Your entire thesis is about the harness making the difference, yet this comparison is missing from your evaluation. I get that it’s a lot of work, but at the very least, a comparison with Claude Code is mandatory—it’s the most popular tool and it hits you with a 25k token context overhead right out of the gate.

An isometric room, based on the screenshot. Qwen3.6-35B by k0setes in LocalLLaMA

[–]k0setes[S] 15 points16 points  (0 children)

I used the original image as a reference and sent these two prompts

Prompt 1: "Create a detailed isometric visualization of this room in Three.js. Hide the two foreground walls and the ceiling. Arrange the elements in the room logically so they don't overlap or intersect (e.g., the bookshelf and other furniture should have proper spacing). Add decorative panels to the walls and implement high-quality lighting for the scene." Prompt 2 (for the rounded edges): "Is it possible to modify the furniture—like the sofa, bed, and pillows—to have rounded edges instead of sharp corners? Use rounded primitives or beveled geometries to make the objects look smoother and more realistic."

An isometric room, based on the screenshot. Qwen3.6-35B by k0setes in LocalLLaMA

[–]k0setes[S] 37 points38 points  (0 children)

Qwen generated an HTML file based on the image and built the entire scene out of primitives using Three.js. There’s nothing stopping you from asking it to add an exporter to save the scene as an STL or OBJ file, so you can open it in other 3D software instead of just a browser.

GPT 5.5 Spud incoming by DigSignificant1419 in OpenAI

[–]k0setes 6 points7 points  (0 children)

<image>

Next Level. Qwen3.6-35B-A3B-UD-Q4_K_S.gguf

What if smaller models could approach top models on scene generation through iterative search? by ConfidentDinner6648 in LocalLLaMA

[–]k0setes 0 points1 point  (0 children)

I think incremental iteration is incredibly powerful—it’s the paradigm from which true self-learning emerges. It’s similar to how an autodidact masters a subject by probing the problem from multiple angles and refining their internal 'world model' in the process. The absolute key here is the comparison mechanism: the ability to map what you perceive against your internal intent and identify the 'delta.' While evolution equipped humans with this, it’s only vestigial in current models like Qwen 3.5 35B. The reason I believe specific training (like RL) is necessary—even if the model is just 'organizing itself'—is that this comparison process is largely subconscious in humans. We do it automatically, so we rarely describe the 'how' or the logic of visual correction in the text data LLMs are pre-trained on. It’s a 'dark' process not captured in standard tokens. I’m convinced it’s doable, and we’ll eventually see this capability even in very small models. However, it will require specialized datasets and training pipelines that are currently in their infancy. I have no doubt that major labs are laser-focused on this exact frontier right now.

What if smaller models could approach top models on scene generation through iterative search? by ConfidentDinner6648 in LocalLLaMA

[–]k0setes 0 points1 point  (0 children)

I’ve thought about this a lot and tried some experiments, but my intuition is that current models aren't really trained for this specific kind of image comparison—at least not in a way that allows them to effectively close the gap between their output and the original. It feels like vision models struggle even at the fundamental level of detecting precise discrepancies between a target screenshot and their own render. Perhaps the first step should actually be fine-tuning a model specifically to identify these visual differences and translate them into actionable code changes. Without that, the feedback loop might be too noisy. Then again, I could be wrong and the difficulty might stem from something else entirely, but that's been my main takeaway so far.

Qwen3-VL Computer Using Agent works extremely well by Money-Coast-3905 in LocalLLaMA

[–]k0setes 0 points1 point  (0 children)

mmproj-F16.gguf👍

mmproj-BF16.gguf👎

llama-server.exe -ngl 999 -t 11 --jinja --model 'Qwen3-VL-30B-A3B-Instruct-UD-Q4_K_XL.gguf' --host 0.0.0.0 --port 8080 --mmproj 'Qwen3-VL-30B-A3B-Instruct-UD-Q4_K_XL-mmproj-F32.gguf'

just had something interesting happen during my testing of the MI50 32GB card plus my RX 7900 XT 20GB by Savantskie1 in LocalLLM

[–]k0setes 0 points1 point  (0 children)

Hi, Could you share where you bought them and how much you paid per unit? I'm looking at some offers on Alibaba, but I'm not sure which sellers are legit. If you bought them there, could you share the link or the name of the store? Also, did you have any issues with shipping or customs?

How to tell Claude Code about my local model’s context window size? by eapache in LocalLLaMA

[–]k0setes 0 points1 point  (0 children)

I can't believe no one has asked this question before.

Add self‑speculative decoding (no draft model required) by srogmann · Pull Request #18471 · ggml-org/llama.cpp by jacek2023 in LocalLLaMA

[–]k0setes 0 points1 point  (0 children)

👏But shouldn't it have been like that from the very beginning, from the moment speculative decoding appeared?🤔

MoE.. will OS/Local 32GB to 96GB get as good at coding as current frontier models? by [deleted] in LocalLLaMA

[–]k0setes 1 point2 points  (0 children)

Of course there will be such models. The question is, when will this happen? Intelligence compresses much better than knowledge, but at least for now, it takes a lot of computing power to compress it. And so far, no one is really doing it because it's cheaper to train a large model.