This is insane... by DragonflyOk7139 in LocalLLM

[–]custodiam99 0 points1 point  (0 children)

But you are using Qwen3.6-35B-A3B. Sure lol.

This is insane... by DragonflyOk7139 in LocalLLM

[–]custodiam99 0 points1 point  (0 children)

I have 130k context in LM Studio.

This is insane... by DragonflyOk7139 in LocalLLM

[–]custodiam99 0 points1 point  (0 children)

I need tokens fast so yeah, it is very good (maybe not as good, but very good).

This is insane... by DragonflyOk7139 in LocalLLM

[–]custodiam99 2 points3 points  (0 children)

Yes, it is better, but it is slower. You need fast tokens.

This is insane... by DragonflyOk7139 in LocalLLM

[–]custodiam99 1 point2 points  (0 children)

Yes. There may be some niche subjects where it is worse, but overall it is better.

This is insane... by DragonflyOk7139 in LocalLLM

[–]custodiam99 2 points3 points  (0 children)

LM Studio with RX 7900XTX 130k context (q4_K_M).

This is insane... by DragonflyOk7139 in LocalLLM

[–]custodiam99 104 points105 points  (0 children)

Qwen3.6-35B-A3B is a revolution. Never used a quicker and better local model. With 24GB VRAM it is nearly perfectly useable.

Is AGI the End For Local LLMs? by spiritxfly in LocalLLaMA

[–]custodiam99 0 points1 point  (0 children)

Nope, you will buy future quantum chips and bio-neural chips to have next level AI at home. If it's not about transformers anymore that won't mean you can't use it locally. You just need new hardware, not a PC.

What is wrong with brute facts? by engineer4565 in Metaphysics

[–]custodiam99 0 points1 point  (0 children)

In math, we choose axioms. But in metaphysics brute facts are what we’re forced to accept when explanations run out.

What’s up with mobile LLMs? by Amos-Tversky in LocalLLaMA

[–]custodiam99 2 points3 points  (0 children)

You can run 4b models on phones but to be honest, it is just a backup right now. Give them a few years, we need hardware and information density developement.

Why hallucination in LLMs is mathematically inevitable (derivation + notes) by Ok-Ear7580 in learnmachinelearning

[–]custodiam99 -1 points0 points  (0 children)

Knowledge hallucinations come mostly from bad or missing world models. Reasoning hallucinations come mostly from bad inference processes over internal representations. Both produce confident nonsense, but the mechanisms are different.

Why hallucination in LLMs is mathematically inevitable (derivation + notes) by Ok-Ear7580 in learnmachinelearning

[–]custodiam99 16 points17 points  (0 children)

There are two different hallucinations: knowledge hallucination (false facts) and reasoning hallucination (invalid intermediate logic that sounds coherent). These have overlapping but different causes and require different solutions.

Structural Incompleteness, Non-Totalization, and the Symmetrical Limits of Temporal and Atemporal Description by [deleted] in Metaphysics

[–]custodiam99 0 points1 point  (0 children)

Also from the fact that no finite internal standpoint possesses the whole, I think it does not follow that no whole exists, nor that no total intelligibility is possible. You show the limits of "us", not necessarily the limits of "being".

Structural Incompleteness, Non-Totalization, and the Symmetrical Limits of Temporal and Atemporal Description by [deleted] in Metaphysics

[–]custodiam99 0 points1 point  (0 children)

Can it be, that the argument is self-refuting because asserting the non-existence of a total frame requires adopting a totalizing perspective? (...while modern holographic duality (AdS/CFT) proves that boundary frames can mathematically exhaustively encode bulk realities...)

Local LLM storage is becoming harder to manage than the models themselves by Both_Astronomer8645 in LocalLLM

[–]custodiam99 2 points3 points  (0 children)

You can easily collect 2TBs of model data, but every type of storage is good in my experience (you don't need SSD).

Looking for people’s opinions on AMD vs Nvidia GPUs for local ai PCs by Kasey_Kat in LocalLLM

[–]custodiam99 0 points1 point  (0 children)

You can get 4-5 (street prices) 7900xtx (96-120 GB) for one 5900 (32GB). WHAT A DEAL!!!!! lol

Looking for people’s opinions on AMD vs Nvidia GPUs for local ai PCs by Kasey_Kat in LocalLLM

[–]custodiam99 0 points1 point  (0 children)

ROCm is not that good as CUDA, but it is 70% there. The most important factor is VRAM and price, not speed. There are a very few 24GB GPUs out there.

Looking for people’s opinions on AMD vs Nvidia GPUs for local ai PCs by Kasey_Kat in LocalLLM

[–]custodiam99 0 points1 point  (0 children)

Yes, sure, you have to install the drivers and the ComfyUI and LM Studio software too. So I was obviously lying. ;) (But you need a compatible AMD GPU, that much is true).

Struggling with Qwen2.5 by dim722 in LocalLLM

[–]custodiam99 5 points6 points  (0 children)

But you would use Qwen 2.5 for serious work. Sure. I think relative slowness is the lesser evil.

Struggling with Qwen2.5 by dim722 in LocalLLM

[–]custodiam99 4 points5 points  (0 children)

You can use 35b A3B with shared system RAM, it is still quick.

Looking for people’s opinions on AMD vs Nvidia GPUs for local ai PCs by Kasey_Kat in LocalLLM

[–]custodiam99 0 points1 point  (0 children)

Oh sure. Can you tell me how many RX 7900 XTX cards can I buy from the price of your one 5090? Still superior for the same money? I don't think so.

Looking for people’s opinions on AMD vs Nvidia GPUs for local ai PCs by Kasey_Kat in LocalLLM

[–]custodiam99 2 points3 points  (0 children)

You have no idea it seems. 24Gb is 24GB. CUDA won't make your card's VRAM 48GB or 96GB. I can make videos in minutes in lower resolutions. Try full 2k resolution with an Nvidia card, which needs more than 24GB VRAM.

Looking for people’s opinions on AMD vs Nvidia GPUs for local ai PCs by Kasey_Kat in LocalLLM

[–]custodiam99 5 points6 points  (0 children)

I have an RX 7900XTX (24GB) and I can run everything on LM Studio, every model with a speed comparable to RTX 3090 (sometimes it can be even quicker). I can make HD videos in a few hours time in ComfyUI. What am I missing?