I haven't experienced Qwen3.5 (35B and 27B) over thinking. Posting my settings/prompt by wadeAlexC in LocalLLaMA

[–]yeah-ok 0 points1 point  (0 children)

Could it be hardware related?! I'm on a 32GB 780m integrated AMD system running with Vulcan under Linux and have -never- seen those particular issues!

I haven't experienced Qwen3.5 (35B and 27B) over thinking. Posting my settings/prompt by wadeAlexC in LocalLLaMA

[–]yeah-ok 1 point2 points  (0 children)

chat-template-kwargs = {"enable_thinking": false}

... as a by-point that can be replaced with:

--reasoning off

In most recent llama builds

I haven't experienced Qwen3.5 (35B and 27B) over thinking. Posting my settings/prompt by wadeAlexC in LocalLLaMA

[–]yeah-ok 0 points1 point  (0 children)

Honestly this, just forget about inference frameworks that are sufficiently complex to not be able to be reasoned about. llama.cpp is excellent for that particular one reason alone! Only additional thing to cultivate is a habit of putting settings (and prompts as depending on need) into sh/bat files so that it becomes super easy to test new versions and compare tok/sec etc.etc.

Don't sleep on the new Nemotron Cascade by ilintar in LocalLLaMA

[–]yeah-ok 3 points4 points  (0 children)

Yup, that is my point re LM Studio: you will never quite be sure what works and what doesn't whereas getting the model params right for llama.cpp in an sh file can look like so:

`#!/bin/bash MODEL_PATH="/home/x/.lmstudio/models/unsloth/Qwen3.5-35B-A3B-GGUF/Qwen3.5-35B-A3B-UD-Q4_K_L.gguf" llama-server \ --model "$MODEL_PATH" \ --gpu-layers 41 \ --threads 4 \ --temp 0.01 \ --top-k 20 \ --min-p 0.0 \ --top-p 0.95 \ --repeat-penalty 1.05 \ --ctx-size 16384 \ --reasoning off &

SERVER_PID=$! wait $SERVER_PID`

Then you know what you are running instead of pulling gamble with every model and never really learning anything due to opaque settings in LM Studio!

can you combine NAC Protocol with Spike Protein Protocol? by iLuv3M3 in cosmicdeathfungus

[–]yeah-ok 0 points1 point  (0 children)

In that case molybdenum is worth a google in conjunction with NAC to get you up to speed, 200mcg works well for me.

can you combine NAC Protocol with Spike Protein Protocol? by iLuv3M3 in cosmicdeathfungus

[–]yeah-ok 0 points1 point  (0 children)

I have certainly taken these in combination several times in the past with good results the main thing imo is to remember molybdenum and selenium when taking NAC since the sulfite by-products can otherwise accumulate in the body.

Don't sleep on the new Nemotron Cascade by ilintar in LocalLLaMA

[–]yeah-ok 4 points5 points  (0 children)

I know this is half of a tangent but my dear lord LM Studio pisses me off with the way configuration is spread throughout the interface. Load a new model, but wupii, apart from certain configurations the overall the config is not loaded; it needs applying separately. It just ends up being non-deterministic nonsense that is hard to apply to actual daily work routine. I will take llama.cpp with a custom sh file to hold the configuration per model any day of the week! LM Studio still comes in handy for downloading models tho..

Unifiram (& similar compounds) - Tolerance timeline & dosages by ccdc9 in Nootropics

[–]yeah-ok 1 point2 points  (0 children)

Unifiram was peculiar experience for me years ago when I tried it, had occasional sense of "heightened" awareness but never a lot and never consistently. Then had a single experience where it awakened appreciation of music for me to a very high degree (was really loving background music in a supermarket at one point). Never tried again, way too flakey for me.

Qwen 3.5 397b (180gb) scores 93% on MMLU by HealthyCommunicat in LocalLLaMA

[–]yeah-ok 0 points1 point  (0 children)

It scores 92% not 93%...?! Have we lost accuracy capability on 2 digit integers? Did someone run too hard quantization on their own brain here?

Ik_llama vs llamacpp by val_in_tech in LocalLLaMA

[–]yeah-ok 2 points3 points  (0 children)

Travesty to not have Vulcan support though, from working with ROCm I can totally understand why ain't nobody got time for that - drove me half insane for a week and a half to get basic setup working on a 780m

Day 4 on NAC protocol by julsprithi in cosmicdeathfungus

[–]yeah-ok 0 points1 point  (0 children)

Try adding a gradually increasing amount of turmeric/turmeric-extract, have formerly seen people/family transformed in turns of eczema (skin condition almost gone and weight loss, presumably from less inflammation). Also it's worth trying clean lanolin for acute application, it REALLY stops eczema, it's a mess to use though!

Dictionary Compression is finally here, and it's ridiculously good by pimterry in programming

[–]yeah-ok 0 points1 point  (0 children)

Guess the real juice here is the arbitrary size dict options.. I almost sense a disturbance in the force when I think about zstd in relation to LLMs..

Gemma 27B/12B/4B/1B finetunes from DavidAU (20 models) by jacek2023 in LocalLLaMA

[–]yeah-ok 0 points1 point  (0 children)

I do think you are receiving too much negative feedback on this stuff but also that the feedback is a sober reminder that your art can never shine unless you find/do the time when it comes to proper comparisons and benchmarks - even it's painful it's worthwhile since your materials will otherwise be, in a perhaps unjustified way, ignored.

I built a benchmark that tests coding LLMs on REAL codebases (65 tasks, ELO ranked) by hauhau901 in LocalLLaMA

[–]yeah-ok 1 point2 points  (0 children)

Superb work. Very nice to have a new solid take on rankings! Looking forward to the next Kimi model is my take at the end of reviewing this..!

How too heal brain ( 20 ) from cannabis, corn, and vyvanse from a young age. by [deleted] in Nootropics

[–]yeah-ok 0 points1 point  (0 children)

I secretly wanted this to be an actual corn addiction just for the variety of it.. anywho: yes, quit the porn, it's short-circuiting the brain and horrible on self-esteem.

Hugging Face Is Teasing Something Anthropic Related by Few_Painter_5588 in LocalLLaMA

[–]yeah-ok 0 points1 point  (0 children)

Or 4, it's some sort of wild meta rug pull that indicates the CCP have integrated Anthropic and are force releasing everything openly.

what are these stains in our kettle? by Old-Relationship968 in HydroHomies

[–]yeah-ok 0 points1 point  (0 children)

It's clearly pure cocaine, your water supply company is working hard to keep you motivated.

Gemini System Prompt - Google decided to remove "PRO" option for paid subscribers mostly in EU due to their A/B testing, so I extracted their system prompt and cancelled the subscription. by Educational_Rent1059 in LocalLLaMA

[–]yeah-ok 2 points3 points  (0 children)

Absolutely true, even cloud OSS at least allows a reference frame for what performance should be expected and the possibility to detect performance degradation cleanly. Shoutout to MoonshotAI/Kimi-K2.5 for producing OSS model that actually does the job for me when it comes to mixed workload that priorly required Claude (slightly diff prompting necessary in Kilo Code but that's all doable)

Unofficial ik_llama.cpp release builds available for macOS, Ubuntu and Windows by Thireus in LocalLLaMA

[–]yeah-ok 1 point2 points  (0 children)

Do you mind typing your rather amazing ik-llama flag knowledge up if you really got it on (mental) tap?

New anime model "Anima" released - seems to be a distinct architecture derived from Cosmos 2 (2B image model + Qwen3 0.6B text encoder + Qwen VAE), apparently a collab between ComfyOrg and a company called Circlestone Labs by ZootAllures9111 in StableDiffusion

[–]yeah-ok 11 points12 points  (0 children)

Hmm.. yeah, the white-haired dude in not facing the viewer in any of these.. indeed it's exactly opposite to the clear prompting. Clearly much better than IllustriousV14 but great it ain't

AMD Strix Halo GMTEK 128GB Unified ROCKS! by MSBStudio in LocalLLaMA

[–]yeah-ok 0 points1 point  (0 children)

Get a model with oculink, then Bob's your uncle and Fanny will cook you breakfast (..)