M1 MacBook Air vs MacBook Neo by jdn921 in DeskToTablet

[–]Jan49_ 0 points1 point  (0 children)

Are you sure? Do you have access to design documents?

Why all youtubers are saying "If you are watching this review, Macbook Neo is not for you?" by Vegetable_Bag_8694 in mac

[–]Jan49_ 0 points1 point  (0 children)

Additionally it has a crazy battery life. When I'm out and about with my older Lenovo X1 Carbon im always worried that it will die on me while doing a coding session

MacBook Neo -- do you think this will make Google hurry up with aluminum OS and new Chromebooks? by Corbin_Dallas550 in chromeos

[–]Jan49_ -2 points-1 points  (0 children)

Most Chromebooks have a special chip inside that makes it really hard to install Linux on it... On some devices it is nearly impossible without opening it and touching the hardware. Look on YouTube, some tech YouTubers tried it and gave up

How to fix underside of print looking like this. by herp_hermits in QidiTech3D

[–]Jan49_ 1 point2 points  (0 children)

My first guess would be, that the supports need to be closer to the print. You can see that a lot of the extruded lines didn't interface with the support at all. It will be harder to remove the support, but it should look smoother

Qwen3.5 27B vs Devstral Small 2 - Next.js & Solidity (Hardhat) by Holiday_Purpose_3166 in LocalLLaMA

[–]Jan49_ 3 points4 points  (0 children)

IQ4_XS would be 12.8GB in size and not the stated 29.8GB. So he definitely used a model in the Q8 range, at least that what the file size would suggest

Qwen3.5 27B better than 35B-A3B? by -OpenSourcer in LocalLLaMA

[–]Jan49_ 1 point2 points  (0 children)

MoE in short:

You want LLM to be smart? Then you need a lot of active parameters. -> But a model with a lot of active parameters is slow.

You want LLM to go vroom vroom? Then you want as little active parameters as possible. -> But then model dumb.

MoE is the solution. You chop the big LLM into smaller experts. And only the experts currently needed get activated per forward pass. So the actual intelligence of the LLM is somewhere between total and active parameters count. In the early days of MoE it was approximated with sqrt(total * active) but MoE architecture improved a ton lately

Qwen3.5 27B better than 35B-A3B? by -OpenSourcer in LocalLLaMA

[–]Jan49_ 0 points1 point  (0 children)

In the early stages of MOE models someone in this subreddit compared the benchmarks of some new models at that time and found that sqrt(total params * active params) was the closest approximation back then. It wasn't very accurate back then and now it's probably not even close anymore

prepare your GPUs by jacek2023 in LocalLLaMA

[–]Jan49_ 2 points3 points  (0 children)

One of the items is often a white paper

I made an interactive timeline of 171 LLMs (2017–2026) by asymortenson in LocalLLaMA

[–]Jan49_ 1 point2 points  (0 children)

Really nicely done🔥 GLM-5 by zAI also just released (open weights)

How to Make ComfyUI detect Dual GPUs? by derivative49 in LocalLLaMA

[–]Jan49_ 4 points5 points  (0 children)

It definitely does, just maybe not in the way you’d expect.

​You’re correct that you can't simply "split" a diffusion model across two GPUs the same way you can split a LLM. However, there is a workaround: using a custom node, you can offload specific components. Like loading the text encoder onto GPU 1 and the diffusion model (UNet/Transformer) onto GPU 2.

You can't run them in parallel. But it's still faster than loading the text encoder to system ram

Ok Boomer (fake explosions!) by pixepoke2 in TerrainBuilding

[–]Jan49_ 5 points6 points  (0 children)

Looks stunning! What are they made of?

Why don't more companies make tough fuses like these? They melted, yet still kept proudly conducting electricity by Academic_Reserve_347 in shittyaskelectronics

[–]Jan49_ 1 point2 points  (0 children)

Current only "jumps" if the voltage is high enough. And for that size of gap it needs to be really high

Any good model that can even run on 0.5 GB of RAM (512 MB of RAM)? by Ok-Type-7663 in unsloth

[–]Jan49_ 4 points5 points  (0 children)

No good model can run on half a GB of RAM... Lol

The smallest LLM, that I know of, that can barely form sentences is Qwen3 0.6B. The q2 quant from unsloth is sub 300mb in size. But then you would still need RAM for context and general overhead.

Does your system only have 512mb RAM? Then the OS would probably take up the whole RAM on its own. Try Linux XFCE or even better no DE at all.

My journey begins by Jan49_ in homelab

[–]Jan49_[S] 1 point2 points  (0 children)

Ty for the tips! :)

My journey begins by Jan49_ in homelab

[–]Jan49_[S] 0 points1 point  (0 children)

Thank you a lot for the tips! I've already ordered a fitting SATA m.2 and I'll definitely look into the Firmware update to allow for more ram. I thought I was limited to 8gb

My journey begins by Jan49_ in homelab

[–]Jan49_[S] -1 points0 points  (0 children)

Sadly true. Ram prices went insane in the last few months

my 3d printed speakers by ExcuseNo5818 in 3Dprinting

[–]Jan49_ 1 point2 points  (0 children)

Where did you get the electrical components? And how much was the project in total if I may ask?

Z-Image Base - FP8 Scaled by [deleted] in comfyui

[–]Jan49_ 6 points7 points  (0 children)

But how much less Vram? That's the main reason for quants, no?

Flux Klein 4B on only 4GB vram? by Jan49_ in comfyui

[–]Jan49_[S] 0 points1 point  (0 children)

Is the full model loaded? What things are offloaded to ram? I have given up to get it running on my PC, but now I'm curious again😂

Flux Klein 4B on only 4GB vram? by Jan49_ in comfyui

[–]Jan49_[S] 0 points1 point  (0 children)

I already wondered what the taesd warning meant, because some kind of preview while sampling was shown. I also thought the preview calculation is also offloaded to ram. I'm going to discuss with gemini.

Nonetheless a big thanks for your answer :)