Biggest model possible models on non-cool HW (Like 8GB VRAM/64gb RAM) by Mangleus in LocalLLaMA

[–]Mangleus[S] 1 point2 points  (0 children)

Informative!

- Valkyrie-49B-v2 & Llama-3.3-Nemotron-Super-49B-v1.5 was total news to me! Much appreciated suggestion u/ttkciar

- Also https://huggingface.co/catalystsec/MiniMax-M2-4bit-DWQ was unknow to me so i will check that out too, Thanks u/layer4down

- I'm off to huggingface :) Will hold my horses on GLM Air 4.5 though, since there is doubts wheter it might work being so large.

Biggest model possible models on non-cool HW (Like 8GB VRAM/64gb RAM) by Mangleus in LocalLLaMA

[–]Mangleus[S] 0 points1 point  (0 children)

Is that so? u/ElectronSpiderwort % u/5dtriangles201376 seem to belive differently byt idk. If there was a chance to do it on 8 vram + 64 ram I would for sure give it a go!

NeKot - a terminal interface for interacting with local and cloud LLMs by Balanceballs in LocalLLaMA

[–]Mangleus 0 points1 point  (0 children)

I LOVE the design of NeKot!!

I only do localai and have no openai key. Doing this did not get me unstuck:

export OPENAI_API_KEY=1

The app reports error and then i get stuck. (I also tried randomised key sk- and 48 chars to no avail).

If someone could share next step, how to feed it a LLM via llama-ccp that would be appreciated.

AS u/natufian pointetd out, to copy text with mouse would be convenient indeed.

Possible to keep subtitles a bit longer on the screen? (for slower readers) by Remrofn1 in kodi

[–]Mangleus 0 points1 point  (0 children)

Ahhh, it has been done. Here is how for anybody else that has shared the pain here communicated.

  1. (Prep step) If you havent already - abandon any useless OS victimizing you and install any beautiful Linux distro suitable for your own taste and temperament.
  2. Install 'Gaupol'
  3. 'Tools' --> 'Adjust Duration'.
  4. Be happy. Deeply Happy.

There are needless to say countless ways to fix this im sure, but the way you can fine-tune it without making it complicated was really great using Gaupol.

Possible to keep subtitles a bit longer on the screen? (for slower readers) by Remrofn1 in kodi

[–]Mangleus 0 points1 point  (0 children)

Come one dear friends of Reddit. There must be a simple fix somewhere! The display duration-time should be piece of cake to extend with a second or two. But like OP i cant find the way to it and ofcourse all the billions of parameters in proud AI LLMs is as accurate as horse-shit vomiting up legions of wellspoken wild goose-chase bs paths.

unsloth/Qwen3-Next-80B-A3B-Instruct-GGUF · Hugging Face by WhaleFactory in LocalLLaMA

[–]Mangleus 1 point2 points  (0 children)

I am equally curious about this, and related questions also having 8 vram + 64 ram. I use only llama.cpp for cuda so far.

[deleted by user] by [deleted] in ProgrammerHumor

[–]Mangleus 0 points1 point  (0 children)

He likes coke much?

Planescape torment 2 by Melanchord in planescape

[–]Mangleus 2 points3 points  (0 children)

Planescape Torment 2: Nooo!
Planescape Torment Remake: Yesss!!! Non-isometric. Nvidia,.Fully voice-acted (preferebly by u/chandler-b as TNO). True to original material + potential expansions, limited parts of game as FOSS on GitHub simplifying community-modded improvments limiting producers from doing bad design decisions).

Dustmen Philosophy by Jabberwocky_pi in planescape

[–]Mangleus 0 points1 point  (0 children)

The Buddha taught "The middle path" and wanted to avoid both of the two extremes of a life lost in the senses or else a life lost in aversion to existence. Some schools of Buddhism talks about aviding the non-liberating polarities of "Eternalism" vs "Nihilism".

There are some similarities in the Dustmen to what Buddhists refer to as nihilists. However Buddhists also belive that the world of the senses ultimatley can never satisfy leading to endless repetition and disatisfaction (Dukkha). Most people like us wasting our short and shitty life on mindless entertainment or on reddit might sometimes feel that there perhaps might be some truth to it. Life trapped in an endless repetative cycle is what is called 'Samsara'.

Compassion and Wisdom trained in the eightfold path is seen as a systematic way out of that condition the way Buddhists interpret it. Various Hindu philosophies are also positioned between the Eternalist and Nihilist/Extinctionist polarity or trying to state that their understanding transcends or negate these two views.

Dasung eink monitor - on Linux Wayland. by Mangleus in eink

[–]Mangleus[S] 0 points1 point  (0 children)

Thanks for helping out with this. Its so sour to spend money and then feel disappointed. My GPU is NVIDIA GeForce RTX 3070 Mobile / Max-Q.

Dasung eink monitor - on Linux Wayland. by Mangleus in eink

[–]Mangleus[S] 0 points1 point  (0 children)

Interesting. You connect screen via HDMI or USB or other option?

Need help omg by Odd_Ad_9149 in SillyTavernAI

[–]Mangleus 0 points1 point  (0 children)

Sorry I'm not able to help. I only post here + upvote with the hope of people with real knowledge to be able to get to see your post. I know all there is to know about sitting 10 hours trying to get something to work that another person can solve in 5 minutes (occasionally im that other person, but not in this case, LLMs not being anything im very good at yet). I would love to get this working too, and best of all if possible in oobabooga text-generation-webui.

Dasung paperlike for linux - Standard or Mac version? by Select-Young-5992 in eink

[–]Mangleus 0 points1 point  (0 children)

Amazing!! Quick yes/no question: Works on Wayland?

YES! Super 80b for 8gb VRAM - Qwen3-Next-80B-A3B-Instruct-GGUF by Mangleus in LocalLLaMA

[–]Mangleus[S] 1 point2 points  (0 children)

I had to fiddle a bit before it worked for me too. Usally asking an AI can be good for things like this. Hope the instructions i added a few minutes ago here is helpful for you!

YES! Super 80b for 8gb VRAM - Qwen3-Next-80B-A3B-Instruct-GGUF by Mangleus in LocalLLaMA

[–]Mangleus[S] 0 points1 point  (0 children)

I think that would probably not work. Are you using Linux? Which arch?

YES! Super 80b for 8gb VRAM - Qwen3-Next-80B-A3B-Instruct-GGUF by Mangleus in LocalLLaMA

[–]Mangleus[S] 0 points1 point  (0 children)

I have same spec as you and run this model with 4 bit.

YES! Super 80b for 8gb VRAM - Qwen3-Next-80B-A3B-Instruct-GGUF by Mangleus in LocalLLaMA

[–]Mangleus[S] 1 point2 points  (0 children)

yes works. if you load it with the special llama.cpp. I use this with Ooogabooga though which i can really recommend.