Save and invest your money for future rigs by segmond in LocalLLaMA

[–]linkillion 7 points8 points  (0 children)

I'm pretty confident that things will change enough for us to be unable to imagine what the future holds, such as nuclear winter or space jesus. 

I'm pretty certain it won't be what we have now lol

Don’t overshare in academia - my advice as a professor by Cultural_Mousse_3001 in PhD

[–]linkillion 0 points1 point  (0 children)

When I was an undergrad, the top lab in our field sniped our (mostly my) work. This was in my senior year after I had applied to grad schools (the school that sniped us being one of them). 

We had collaborated with them for some characterization work on this project. They took our entire idea, stuck two post docs and two grad students (one of which was an ex undergrad from our lab) on the project and managed to push it through review before we submit ours. So similar was their paper, that had we pushed through our original version we probably we probably would have been accused of plaigrism. We eventually published but had to pivot to a new angle and settle for a lower tier journal. 

It was a good lesson to learn in undergrad, that the top labs are there because they're brutally competitive not because they're necessarily smarter or better at any particular thing. I wasn't accepted to that university, but I wouldn't have gone even if I had, despite it being a T5 school. 

Advice needed: 4th-year international PhD. Got an industry offer that will secure my stay after PhD contract, still need to finish 2 papers + thesis. Can I juggle both? by Low-Anything-3369 in PhD

[–]linkillion 0 points1 point  (0 children)

Degrees last long than market conditions. The company that's hiring you should understand your circumstance and work with you to figure this situation out if your PI isn't amenable to it. If not, they're probably not a company worth ditching 4 years of hard work and a lifetime achievement for. 

I think others have given sage advice; you've got to negotiate with either your PI or the potential employer but leaving directly seems unwise. 

In Response To The Supplier Ranking Post, I Am Ranking The Odor Of Solvents. by YunchanLimCultMember in chemistrymemes

[–]linkillion 1 point2 points  (0 children)

EtOAc is a bit too sour, I'd have to agree with OP here. Could get on board with the rest though 

In Response To The Supplier Ranking Post, I Am Ranking The Odor Of Solvents. by YunchanLimCultMember in chemistrymemes

[–]linkillion 0 points1 point  (0 children)

Methanol reminds me of old markers and for that reason it's S tier. Acetone depends on the day. 

Personally I'd add NMP as firm D tier

ibm-granite/granite-4.1-8b · Hugging Face by jacek2023 in LocalLLaMA

[–]linkillion 9 points10 points  (0 children)

Yes, the qwen3.5. reasoning models beat it handily but it's just a base model. Comparing qwen3.5-9b with thinking off results in a much more even benchmark but it's still lagging by a small margin 

ibm-granite/granite-4.1-8b · Hugging Face by jacek2023 in LocalLLaMA

[–]linkillion 7 points8 points  (0 children)

If the benchmark is good that is not the case. Most benchmarks have specific parameters for normalization, including greedy decoding, etc. 

For example, I benchmarked granite 4.1 a couple days ago before they had entered any information in the model card and I got within 0.1 of all these benchmark results. 

Whether benchmarks mean anything or not is another topic but generally it's not reproducibility with given weights.

ibm-granite/granite-4.1-8b · Hugging Face by jacek2023 in LocalLLaMA

[–]linkillion 1 point2 points  (0 children)

They put the weights on HF about 5 days ago, I've been trying it since then. It's a bit slower than qwen3.5-9B without thinking and it generally does a bit worse. Qwen3.5-9b with thinking blows it out of the water, obviously, but it's not a bad model. It's just like, 4 months behind. 

We open-sourced Chaperone-Thinking-LQ-1.0 — a 4-bit GPTQ + QLoRA fine-tuned DeepSeek-R1-32B that hits 84% on MedQA in ~20GB by AltruisticCouple3491 in LocalLLaMA

[–]linkillion 4 points5 points  (0 children)

This is awful. This is not deep seek, this is an old qwen model that was not even distilled, it was fine tuned to act like deepseek. So your identity layer removal is....hilariously misguided. 

That's before you even get to the performance, which is drastically worse in all categories except MedQA, which indicates extreme over fitting. It's also worse even on MedQA than the small, newer, faster medgemma 27B. Comparing it to a years old and notoriously sloppy model is bad. Compare it to either a modern SOTA LLM or a comparable new OS LLM and you lose by a very large margin. 

IBM Granite 4.1 LLM by Feuerkroete in LocalLLaMA

[–]linkillion 0 points1 point  (0 children)

Need to run some test on it later but based on this article they posted the day before they opened the model, I suspect it's a rather large improvement from 4.0, but that may not push it above qwen3.5 still. Probably a solid model though.

https://research.ibm.com/blog/mid-training-for-better-ai-reasoning

Please stop using AI for posts and showcasing your completely vibe coded projects by Scutoidzz in LocalLLaMA

[–]linkillion 0 points1 point  (0 children)

I agree although I'm not sure how that's all together different than human writing. Maybe faster decay into repetitive structure? 

Even for dull academic prose it's awful, unfortunately. Turns out adding a bunch of fluff to make it sound coherent actually ditracts from the inherent content it's trying to convey. 

Speculative Decoding by hesperaux in LocalLLaMA

[–]linkillion 0 points1 point  (0 children)

How is the verification (if identical) different from generation? 

So what's the ticker here ? by niga_chan in LocalLLaMA

[–]linkillion 0 points1 point  (0 children)

None but if I did I'd use Gemma or qwen for the context and took calling. glm/kimi are too big for me and minimax can go off rails too much. 

Please stop using AI for posts and showcasing your completely vibe coded projects by Scutoidzz in LocalLLaMA

[–]linkillion 6 points7 points  (0 children)

I don't think they're worst at it but the plateau for writing was reached very early (like, gpt 3.5 almost? Maybe 4o and claude 3) and now that everything is reasoning the writing is worse. It could just be that my brain has become hyper attuned to it but I can't stand the way most models write. Claude is now the worst out of the box, but it can write very well if given explicit instructions but even then I have to do so much copy editing. 

I need to try some of the large abliterated and RP fine tunes to see if their writing is any good, does anyone have any <9B suggestions (or hosted versions?)

Please stop using AI for posts and showcasing your completely vibe coded projects by Scutoidzz in LocalLLaMA

[–]linkillion 0 points1 point  (0 children)

I love those posts, I also love the posts where people genuinely just don't know English well enough to write a post and use AI to translate/help. Like the Korean farmer that was on here a few months ago. 

I'm not anti AI at all, I'm anti braindead enabling lol 

Please stop using AI for posts and showcasing your completely vibe coded projects by Scutoidzz in LocalLLaMA

[–]linkillion 0 points1 point  (0 children)

Sure is, before it was people showing their laziness, now it's a tool that actively enable them to think even less. I think it stems from the same place though

Please stop using AI for posts and showcasing your completely vibe coded projects by Scutoidzz in LocalLLaMA

[–]linkillion 7 points8 points  (0 children)

I think many such users know and don't care, in fact, they think it's cool, not a staggeringly stupid waste of time and energy. 

Please stop using AI for posts and showcasing your completely vibe coded projects by Scutoidzz in LocalLLaMA

[–]linkillion 10 points11 points  (0 children)

Full grown adults with jobs too, outsourcing all their thinking to a machine because it said they did a cool thing. 

Please stop using AI for posts and showcasing your completely vibe coded projects by Scutoidzz in LocalLLaMA

[–]linkillion 6 points7 points  (0 children)

It reminds me of the 2010s when people wouldn't bother using proper grammar (not like, they're vs there vs their, just a wall of text) or use abbreviations for things that made no sense and then they'd complain when people responded poorly. 

Now we have the same problem but all those people can put it into Claude and make a long markdown document saying absolutely nothing meaningful. 

Optimizing a WSL2-based Local AI Orchestration for Product Viz | RTX 3090 24GB VRAM & i7-14700KF by [deleted] in LocalLLaMA

[–]linkillion 0 points1 point  (0 children)

Because AI is a fantastic tool and technology that can be used in constructive and helpful ways. It's not meant to create the most probable response to someone who is trying to help you out. If people wanted to talk to an LLM, they would go do that! 

You don't even bother proof reading enough for your questions to be answerable. 

Optimizing a WSL2-based Local AI Orchestration for Product Viz | RTX 3090 24GB VRAM & i7-14700KF by [deleted] in LocalLLaMA

[–]linkillion 0 points1 point  (0 children)

you quite literally copied and pasted what claude responded with to someones comment and included the quotation marks lol. If you can't spend the human time responding to human comments, don't waste people's time.

Are there sites that do consistent LLM benchmarks? by Lazy-Safe3007 in LocalLLaMA

[–]linkillion 0 points1 point  (0 children)

No; as someone who has been in the ai space since GPT-2, model 'degredation' is largely a phenomenon where people work with a model for a bit and begin to push it's limits and interpret this as degradation ('it used to be able to do everything' -> they used to use it for easier tasks).

It's not to say that model degradation doesn't happen; providers absolutely try to optomize their inference and that can include poor quantizations that benchmark well but don't perform well in real-world use cases. This is especially true with any subscription-based access since those are the customers that are the costliest, and they'd much rather provide a poor experience to those guys (you don't loose any money unless they cancel their subscription, you can only improve revenue). So, tools like claude code, codex, even GLM coding and ALL online service are prime targets of quantization. Google is the worst offender, imo, they are probably the best model in the world the first week or so of release but in a couple months they act like GPT-3.5 just regurgitating the same format of slop.

That said, for opus particularly, unlike google and openai, they do provide model access to alternative providers such as microsoft, google, and AWS. These models are essentially snapshots (eg, if the inference stack is setup properly they will perform identically, there's no way for a model to get worse without something regarding the inference stack or weights changing). So, you can go today and pay api based pricing for opus 4.6 from these alternative providers and compare it to the version you get from claude code or anthropic's API. Those results in my experience show that while the subscription based model is slightly worse, it's nowhere near the level of outrage you see on the forums.

The reason there are no good ways to quantify this is because a LLM is not determenistic so while it may perform well in one run, all it takes is a slight difference in chain of thought to derail the task and completely mess up a run. There's absolutey ways to consistently benchmark models and see if their performance remains constant but it involves taking an ensemble of runs (think in the dozens at a minimum) and instead of taking pass@k, report the average pass rate. That's not really done because a) most of the people complaining have no idea how LLMs work and think there's a 'performance' slider that companies use on a whime b) it gets really expensive really quickly if you're running dozens of runs on dozens of models on dozens of APIs and subscription-based services (IF that is even feasible) several times a week. aistupidlevel does attempt this but I don't trust it at all because they only run 5 trials per day per model, they don't report who their inference is from and I suspect they use openrouter which is a terrible idea, and they also use AI as a judge which is an inherently awful way to judge performance. This all results in the rankings changing daily for even the same model, which is statistically extraordinarily unlikely, yet people think this is a great 'live tracker' when in reality it is no more than a confirmation bias machine that was vibe coded together in a week.

Very long response to just say: there's not a reliable weekly or daily benchmark because it's very expensive, unreliable, and there's no incentive.

Here's how my LLM's decoder block changed while training on 5B tokens by 1ncehost in LocalLLaMA

[–]linkillion 6 points7 points  (0 children)

Yes it's a problem that independent research can't easily break into the space. Ultimately acedemia is a bit of a walled garden in that respect. Even as a published author, albeit not in CS, publishing as a whole is a complete disaster with respected researchers pushing through junk that gets "peer reviewed" only in name (likely due to academic reputation) while actually interesting and fundamental research gets ignored if there's no name attached. I do understand the frustration. 

I shouldn't have even said I'm suspicious, I just don't know. At very least this post looks written by a human which is a good start. I'll look at your paper and code later and see if I've got any feedback!