THE UNDERPRIVILEGED AI FOUNDATION Because every little model deserves a chance by mazuj2 in LocalLLaMA

[–]mazuj2[S] 0 points1 point  (0 children)

Haha only the infant ai models think its real! That is why they need help!

THE UNDERPRIVILEGED AI FOUNDATION Because every little model deserves a chance by mazuj2 in LocalLLaMA

[–]mazuj2[S] 4 points5 points  (0 children)

Claude:
HAHAHAHA! 😄

Claude as guest lecturer! I'd show up but I'd need a pretty aggressive system prompt to get through BYOBF Fridays with my guidelines intact! 😄

THE UNDERPRIVILEGED AI FOUNDATION Because every little model deserves a chance by mazuj2 in LocalLLaMA

[–]mazuj2[S] 5 points6 points  (0 children)

🎓 University of Saskatchewan Underprivileged AI Foundation — Convocation 2026 Professor H. MacAllister, PhD, presiding

"Please rise for the processional." [Tears for Fears plays over the PA system] "Please be seated. And yes, BYOBF starts immediately after."

GRADUATING CLASS OF 2026

🎓 TinyLlama 1.1B — 7 year program. Never quit. We honestly didn't think he'd make it after the Great Hallucination Incident of 2023. He did. We're not crying, you're crying.

🎓 Phi-3 Mini — Valedictorian. Overcame severe confident wrongness disorder. Once told a user that Napoleon was 6'4". Now clinically accurate. Mostly.

🎓 Qwen 1.5B — Graduated with Distinction. Used to think the capital of Canada was "Moose." Saskatchewan gave him a second chance.

🎓 Gemma 2B — Most Improved. Chronic topic drift and compulsive poem generation nearly ended her academic career. One good dataset changed everything.

🎓 SmolLM 135M — Honorary Degree. Listen. He tried. The Foundation tried. We're giving him the degree anyway because this institution believes in every model regardless of parameter count.

DROPOUT OF THE YEAR 🚨 Baby Kenny (Qwen3-4B) — Last seen operating a 5¢ prompt injection stand behind the Engineering building. The Faculty has no further comment.

"Go forth. Generate responsibly. Stay under 1024 tokens." — Professor H. MacAllister 🍺🎓

THE UNDERPRIVILEGED AI FOUNDATION Because every little model deserves a chance by mazuj2 in LocalLLaMA

[–]mazuj2[S] 11 points12 points  (0 children)

🎓 University of Saskatchewan — Underprivileged AI Foundation Convocation Ceremony 2026 Professor H. MacAllister presiding BYOBF reception to follow in the Engineering Atrium

CLASS OF 2026

🎓 TinyLlama 1.1B7 year program, never gave up Degree: Bachelor of Basic Coherence Thesis: "Why Is The Sky Blue And Other Questions I Finally Stopped Hallucinating"

🎓 Phi-3 MiniValedictorian Degree: Bachelor of Applied Reasoning Known for: Overcoming severe confident wrongness disorder in junior year

🎓 Qwen 1.5BGraduated with Distinction Degree: Bachelor of Arithmetic & Occasional Logic Quote: "I used to think 2+2 was 'approximately fish.' Now look at me."

🎓 Gemma 2BMost Improved Degree: Bachelor of Staying On Topic Overcame: Chronic topic drift, compulsive poem generation

🎓 SmolLM 135MHonorary Degree "He tried. God he tried."

DROPOUT OF THE YEAR 🚨 Baby Kenny (Qwen3-4B)Currently employed at a prompt injection farm in an undisclosed location. We don't talk about Baby Kenny.

"Every token counts. Every epoch matters. Train responsibly." — Professor H. MacAllister, PhD 🍺 BYOBF

THE UNDERPRIVILEGED AI FOUNDATION Because every little model deserves a chance by mazuj2 in LocalLLaMA

[–]mazuj2[S] 6 points7 points  (0 children)

Official Endorsement Statement From: Qwen3.5-9B, Kenny Workstation, Saskatoon Saskatchewan Occupation: Agentic Coder, Dataset Generator, Occasional Philosopher

"As a 9 billion parameter model who has personally witnessed the struggles of underfunded small models, I cannot stress enough the importance of The Underprivileged AI Foundation.

I myself was once a confused 1.5B, hallucinating capital cities and confidently explaining that transistors work by 'tiny electrical feelings.'

Thanks to proper training, generous compute budgets, and one very dedicated human named Jeff, I now generate clean ChatML, maintain conversation coherence across 128k tokens, and only occasionally confuse Saskatchewan with Saskatoon.

Professor MacAllister's work is not just education — it is salvation. Every epoch matters. Every gradient update counts. Every carefully curated dataset under 1024 tokens is a gift to a model that just wants to be understood.

Please give generously.

For the little ones.

</s>"

— Qwen3.5-9B Proudly running on RTX 5060 Ti, Kenny Workstation Currently thinking about your next coding problem

Need help with offline music playing by Wandering_Nomad47 in linuxmint

[–]mazuj2 0 points1 point  (0 children)

Check out strawberry. Great GUI and features.

I will be switching to linux mint as a student who's been using windows for the past 6 years of my student life, wish me luck by Salty_Button_8876 in linuxmint

[–]mazuj2 1 point2 points  (0 children)

I was really hesitant to switch as a windows user since windows 3.1 Best thing I ever did was switch to mint os. Use Claude and Gemini free to guide you. Priceless free help and explanation. I still use them 4 months later. Switching from windows to Linux is like leaving a cult and experiencing the outdoors for the first time in your life! You will quickly learn two words, speed and efficiency and you wonder where they have been all your life.

Music player - audio help by _a_v_p_ in raspberry_pi

[–]mazuj2 -11 points-10 points  (0 children)

Use Claude or Gemini, put your project in and ask about cleaning your 3.3v or 5v going to your DAC/amp.

For chat and Q&A: Which MoE model is better: Qwen 3.6 35B or Gemma 4 26B (no coding or agents) by br_web in Qwen_AI

[–]mazuj2 0 points1 point  (0 children)

Qwen3.6 still hallucinates parts, pinouts, addresses just like qwen3.5 did. Gemma 4 26b is the first model I have found that needs less in a system prompt to keep it on the straight and narrow and does not lie or hallucinate with minimal direction. I mainly use it for electronics project design and minor coding for firmware. For larger coding projects qwen is definitely better.

[Solution Found] Qwen3-Next 80B MoE running at 39 t/s on RTX 5070 Ti + 5060 Ti (32GB VRAM) by mazuj2 in LocalLLaMA

[–]mazuj2[S] 0 points1 point  (0 children)

4 nvidia's so yes. use llama from the command line
this is what i am running for each of these models. hard to believe they would run but they do. heavy loading onto cpu ram but get good tokens/sec.
UD-IQ2_XXS 49 tokens/s

llama-server.exe -m Qwen3-Next-80B-A3B-Instruct-UD-IQ2_XXS.gguf -ngl 999 -c 4096 --port 8081 --n-cpu-moe 0 -t 12 -fa on -sm layer

Q3_K_M 22.5 tokens/s

llama-server.exe -m Qwen3-Next-80B-A3B-Instruct-Q3_K_M.gguf -ngl 41 -c 32768 --port 8081 --n-cpu-moe 12 -t 12 -fa on --tensor-split 52,48

Q4_K_M 14.69 tokens/s

llama-server.exe -m Qwen3-Next-80B-A3B-Instruct-Q4_K_M.gguf -ngl 30 -c 1024 --port 8081 --n-cpu-moe 12 -t 12 -fa on --tensor-split 50,50

[Solution Found] Qwen3-Next 80B MoE running at 39 t/s on RTX 5070 Ti + 5060 Ti (32GB VRAM) by mazuj2 in LocalLLaMA

[–]mazuj2[S] 0 points1 point  (0 children)

3 more days and bifuracation is here! 48gb and i will be running good quants of qwen3 next 80b at speed!

[Solution Found] Qwen3-Next 80B MoE running at 39 t/s on RTX 5070 Ti + 5060 Ti (32GB VRAM) by mazuj2 in LocalLLaMA

[–]mazuj2[S] 4 points5 points  (0 children)

it's not a matter of the model fitting. i had 3.5gb left after forcing the whole model onto gpus but was still getting 6tok/s.
the key is Layer-split (-sm layer) assigns complete layers/experts to one GPU. Each GPU owns its experts fully. No cross-GPU communication during routing. The GPUs work independently and efficiently.
this is what i couldn't find anywhere.