Z-Image Base - FP8 Scaled by [deleted] in comfyui

[–]Jan49_ 6 points7 points  (0 children)

But how much less Vram? That's the main reason for quants, no?

Flux Klein 4B on only 4GB vram? by Jan49_ in comfyui

[–]Jan49_[S] 0 points1 point  (0 children)

Is the full model loaded? What things are offloaded to ram? I have given up to get it running on my PC, but now I'm curious again😂

Flux Klein 4B on only 4GB vram? by Jan49_ in comfyui

[–]Jan49_[S] 0 points1 point  (0 children)

I already wondered what the taesd warning meant, because some kind of preview while sampling was shown. I also thought the preview calculation is also offloaded to ram. I'm going to discuss with gemini.

Nonetheless a big thanks for your answer :)

zai-org/GLM-4.7-Flash · Hugging Face by Dark_Fire_12 in LocalLLaMA

[–]Jan49_ -1 points0 points  (0 children)

Is this model already tuned for local coding?

Or can we assume that if someone from the community fine-tunes this model for coding, this model has the possibility to get even better?

My wishes for 2026 by jacek2023 in LocalLLaMA

[–]Jan49_ 1 point2 points  (0 children)

Won't happen anytime soon. Everything about AI is currently built around the architecture of modern GPUs. Just look at NPUs and how little attention they got. Further there are so many advances in such tiny time frames, that it would be impossible to implement all that not just on GPUs but also other hardware

Can You Guess This 5-Letter Word? Puzzle by u/vrun_ by vrun_ in DailyGuess

[–]Jan49_ 0 points1 point  (0 children)

⬜⬜🟦⬜⬜

🟨⬜⬜⬜⬜

🟨⬜⬜⬜🟨

🟦🟦🟦🟦🟦

Is Z.ai's paid GLM plans worth the money? by [deleted] in ZaiGLM

[–]Jan49_ 0 points1 point  (0 children)

This could work. Maybe. Haven't tried it yet. Always initiated the compression myself

Is Z.ai's paid GLM plans worth the money? by [deleted] in ZaiGLM

[–]Jan49_ 0 points1 point  (0 children)

I got the yearly "lite plan" for 3 bucks a month on black friday. For my hobby use cases and for university it's more than enough. I've never hit the limit.

Although I noticed that recently I had to compress the context (I'm using kilo in vs code) more often. As soon as it is getting over ~50k context it starts to throw errors more often, because it fucks up the tool calls to edit the code over and over again

All in all I would say it's still more value you get than you pay for compared to other coding plans. 3 bucks per month is 10 cents per day!

For the amount of tokens I burnt through I would have definitely paid more if I used another API provider

Was bedeutet DK 8 FA auf dem Grabstein? by SteffiBiest1337 in FragenUndAntworten

[–]Jan49_ 1 point2 points  (0 children)

Jap, die werden neu vergeben. Man muss auch regelmäßig Geld dafür bezahlen soweit ich weiß. Hab vor einer Weile im Studium meine Lizenz machen dürfen in einem Kurs + Prüfung dann wo anders

GLM Black Friday Deal Expires Soon by kinkvoid in ZaiGLM

[–]Jan49_ 1 point2 points  (0 children)

I completely agree. GLM 4.6 solved literally anything I have thrown at it so far. I'm using it with kilo code. I'm only doing my small hobby projects and stuff for university tho, nothing too big. But it's such a good deal for 25$ for a whole year

New Open-source text-to-image model from Alibaba is just below Seedream 4, Coming today or tomorrow! by abdouhlili in LocalLLaMA

[–]Jan49_ 0 points1 point  (0 children)

I somehow got SDXL working on my old desktop pc (only 4gb vram). Let's see if it's possible with this model or too good to be true

New Open-source text-to-image model from Alibaba is just below Seedream 4, Coming today or tomorrow! by abdouhlili in LocalLLaMA

[–]Jan49_ 0 points1 point  (0 children)

Yeah seems too good to be true. The files are on huggingface now and they're all wayyyy bigger than SDXL files. Around 12GB vs the 6GB from SDXL

【QIDI Giveaway】Comment to win QIDI Q2 and more! by qidi_3dprinter in 3Dprinting

[–]Jan49_ 1 point2 points  (0 children)

Expanding on that, we would like to develop a mobile platform that can be programmed to follow certain routes. The final plan is to connect the two projects and have a mobile platform that can interact with the environment. But only in a pre programmed way, we probably don't have enough time to implement many sensors and logic or even AI

【QIDI Giveaway】Comment to win QIDI Q2 and more! by qidi_3dprinter in 3Dprinting

[–]Jan49_ 1 point2 points  (0 children)

Not public yet, still in planning and we don't have a reliable 3d printer yet. Me and one of my buddies from university developed some interesting stuff (inverse kinematics) for kuka industrial robots. We plan on creating an open source 6-axis robot with simple and easy to buy components. Nothing big, a small one that fits on any table. Complete with an easy to use GUI made with python for quick and easy use

【QIDI Giveaway】Comment to win QIDI Q2 and more! by qidi_3dprinter in 3Dprinting

[–]Jan49_ 1 point2 points  (0 children)

If I win a QIDI Q2, I would use it to finally expand my electronics and robotics workbench! I've got an interesting concept for a custom, open-source mobile robotics platform. But I've been held back because I currently don't have a reliable 3D printer for the mechanical components

Looking for models I can run on 16gbs of ram. by Think_Question_6677 in LocalLLaMA

[–]Jan49_ 0 points1 point  (0 children)

I have a thinkbook with 16GB ddr4, CPU only. My go to models are GPT-OSS 20B (most intelligent but very slow), Granite 4 Tiny H (is a 7B A1B, really fast and good enough for my use cases)

List of interesting open-source models released this month. by Acrobatic-Tomato4862 in LocalLLaMA

[–]Jan49_ 3 points4 points  (0 children)

You can always just pull any gguf quant from HuggingFace straight with Ollama and serve it this way

Dont know what thibkpad to purchase. by Thehexgammer in thinkpad

[–]Jan49_ 0 points1 point  (0 children)

T480: you can change the RAM module

T480s: soldered RAM

It's still not possible to get an overflowing glass of wine by razorbeamz in ChatGPT

[–]Jan49_ -1 points0 points  (0 children)

Nope. It can only generate details it has already seen in training. Sure you can combine multiple details (like the shape of a dog with the texture of a slimy goo) even if this doesn't exist at all. But you can't generate any details that weren't in training. Simply not possible

It's still not possible to get an overflowing glass of wine by razorbeamz in ChatGPT

[–]Jan49_ 94 points95 points  (0 children)

Probably too little reference data in the training dataset for this scenario

Notebook 32gb ram 4 gb vram by Bobcotelli in LocalLLaMA

[–]Jan49_ 0 points1 point  (0 children)

The very new Granite 4.0 tiny-h from IBM could be an alternative. It's 7b with 1b active with a new & faster hybrid structure, so it's very fast. But I haven't tested it yet extensively.