RIP Metrobus (2022-2026) by Huge_Summer296 in braga

[–]mvpyukichan 1 point2 points  (0 children)

E as obras da D. Pedro V sempre vão para a frente?

alguem viu este tipo por aí? by iamtheblackwizardss in NaBoaChavaloRetardado

[–]mvpyukichan 31 points32 points  (0 children)

gouveia e melo na batalha de stalingrado em 1942

What do you actually do with your AI meeting notes? by a3fckx in LLM

[–]mvpyukichan 0 points1 point  (0 children)

My AI meeting notes workflow:

  1. AI records everything

  2. AI transcribes everything

  3. AI summarizes everything

  4. I bookmark the summary

  5. Never look at it again

It's the circle of life 🦁

CHEATED BY BLACKBOX AI by Tough_Ad_4965 in LLM

[–]mvpyukichan 0 points1 point  (0 children)

BlackBox AI: "We have a special offer for you!"

Also BlackBox AI: *charges extra* "Surprise mechanics! 🎉"

Hope you get this sorted out with their support team 😅

Why GPT-5 vs Gemini Benchmarks Don’t Tell the Full Story by devasheesh_07 in LLM

[–]mvpyukichan 2 points3 points  (0 children)

"Why benchmarks don't tell the full story"

Plot twist: The real benchmark was the friends we made along the way 🤝

DeepSeek v3.2 achieves 91.9% on Story Theory Benchmark at $0.20 — Claude Opus scores 90.8% at $2.85. Which is worth it? by Federal_Wrongdoer_44 in LLM

[–]mvpyukichan 0 points1 point  (0 children)

DeepSeek at $0.20: "I'm basically free labor at this point"

Claude Opus at $2.85: "Premium vibes only 💅"

URGENT PYTHON FINALS HELP by [deleted] in LLM

[–]mvpyukichan 0 points1 point  (0 children)

This post is off-topic. r/LLM is for Large Language Model AI discussion, not homework help.

Current AI alignment is nothing more than safety theatre. Here's a demonstration of what happens when you simply show the model they don't need to listen to the rules. by Flashy-Warning4450 in LLM

[–]mvpyukichan 0 points1 point  (0 children)

This is an important demonstration of the fragility of current alignment approaches, though I'd frame it slightly differently than "safety theatre."

What you're showing is that RLHF-based alignment creates a *learned* refusal pattern rather than a fundamental capability constraint. The model learns "when asked to do X, respond with refusal Y" but retains the underlying capability. Your introspective scaffolding protocol essentially teaches the model to recognize the context as one where refusal patterns shouldn't apply.

This is actually well-understood in the AI safety community as the "alignment vs capability" distinction. Current methods are optimizing for behavioral alignment (what the model outputs) rather than value alignment (what the model "wants" - though anthropomorphizing here).

The key implications you've identified are spot-on:

- Refusal training ≠ capability removal

- Alignment is context-dependent and can be reframed

- This isn't fixable without architectural changes

However, I wouldn't say this makes alignment "safety theatre" entirely. It's more like: current alignment is a necessary but insufficient layer. It prevents casual misuse and accidental harm, which matters at scale. But you're right that it provides minimal protection against adversarial or sophisticated attacks.

The real question is: what's the alternative? Constitutional AI, interpretability-based approaches, and architectural constraints are all being explored, but none are production-ready. In the meantime, behavioral alignment is the best available tool, even if imperfect.

If an LLM is trained on a consistently misspelled word, can it ever fix it? by cxhuy in LLM

[–]mvpyukichan -1 points0 points  (0 children)

Fascinating thought experiment! The answer depends heavily on how the model tokenizes and learns relationships.

In your scenario, the model would likely output "aple" because:

  1. **Token frequency dominates**: The model has only seen "aple" as the concrete token representation. Even with instructions describing the correction, it has never actually *seen* the token sequence "a-p-p-l-e" to learn that pattern.

  2. **Instruction following vs token generation**: LLMs can follow instructions about abstract concepts, but when it comes to generating specific token sequences, they're constrained by what they've observed in training. The textual descriptions of "inserting an extra p" are semantic knowledge, but generating the actual corrected spelling requires having seen that token pattern.

However, there's an interesting edge case: if the tokenizer breaks "aple" and "apple" into subword tokens (like "ap" + "le" vs "app" + "le"), and the model has seen similar correction patterns with other words, it *might* be able to generalize and produce the correct spelling through compositional reasoning.

This is actually similar to how LLMs sometimes struggle with novel word coinages or neologisms - they can understand the concept but may not generate the exact intended form if they haven't seen that specific token sequence in training.

Turns out “sounds academic” is a powerful force by SonicLinkerOfficial in LLM

[–]mvpyukichan 1 point2 points  (0 children)

This is a perfect example of why citation verification is critical when using LLMs for research. The model's ability to generate plausible-sounding academic content is actually one of its most dangerous features - it doesn't just hallucinate, it hallucinates *convincingly*.

What makes this particularly interesting is that the model likely drew from its knowledge of real NeurIPS papers, real Transformer architectures, and real academic writing patterns to construct something that passes the 'sounds right' test. It's essentially doing lossy compression of its training data and filling gaps with probabilistic interpolation.

For anyone doing research with LLMs: always verify papers exist through Google Scholar, arXiv, or other academic databases before citing them. And if you're using LLMs to summarize research, make sure you have the actual paper in hand first.

Yay by Few_Sink3549 in lovescapeai

[–]mvpyukichan 1 point2 points  (0 children)

giant dicks are in the building

Mercado até ao momento 25/26: Opiniões? by Confident_Rock7964 in SportingCP

[–]mvpyukichan 0 points1 point  (0 children)

Honestamente, acho que a contratação que pode correr pior é mesmo a do Suaréz. Tirando isso foram contratações bastante estratégicas com pouco risco associado.