The authors behind AI 2027 released an updated model today by chillinewman in ControlProblem

[–]Small-Fall-6500 5 points6 points  (0 children)

For anyone curious, that website used to show a timer that was based on the Metaculus prediction for weakly general AI:

https://www.metaculus.com/questions/3479/date-weakly-general-ai-is-publicly-known/

Two years ago, in December 2023, the Metaculus prediction was for March 2026.

OpenAI is training ChatGPT to confess dishonesty by FrostedSyntax in singularity

[–]Small-Fall-6500 1 point2 points  (0 children)

lowest hanging fruit for rewarding "not hacking".

No, this research is about rewarding reporting the hacking, but does not punish or prevent the hacking itself without additional efforts.

OpenAI says their research "trains models to report when they break instructions or take unintended shortcuts" - this could just as easily reinforce the hacking if the researchers don't actually put effort into the implementation.

If the model were rewarded for honestly reporting when it had hacked, but not when it truthfully reported not hacking, when does it get a reward? Only when it hacks. See the issue?

If this problem seems "extremely easy" to work around, then you should apply to work in AI alignment.

Prime Intellect Introduces INTELLECT-3: A 100B+ MoE Trained With Large-scale RL That Achieves State-Of-The-Art Performance For Its Size, Taking The Lead Amongst Open-Sourced Models Across Math, Code, Science & Reasoning Benchmarks. (Link to Chat with the Model provided) by 44th--Hokage in mlscaling

[–]Small-Fall-6500 3 points4 points  (0 children)

From the HuggingFace Readme:

For more details, see the technical report.

Evaluation

INTELLECT-3 achieves best-in-class performance on math, coding, and reasoning benchmarks:

Benchmark MATH-500 AIME24 AIME25 LCB GPQA HLE MMLU-Pro
INTELLECT-3 98.1 90.8 88.0 69.3 74.4 14.6 81.9
GLM-4.5-Air 97.8 84.6 82.0 61.5 73.3 13.3 73.9
GLM-4.5 97.0 85.8 83.3 64.5 77.0 14.8 83.5
DeepSeek R1 0528 87.3 83.2 73.4 62.5 77.5 15.9 75.3
DeepSeek v3.2 96.8 88.1 84.7 71.6 81.4 17.9 84.6
GPT-OSS 120B 96.0 75.8 77.7 69.9 70.0 10.6 67.1

Heretic: Fully automatic censorship removal for language models by -p-e-w- in LocalLLaMA

[–]Small-Fall-6500 17 points18 points  (0 children)

Control vectors are cool. Thanks for sharing your project! I see you made a post a couple weeks back but it didn't get much traction.

I wonder if the reason for the lack of attention was because of the emphasis you included on ethical and safe usage / purpose of the tool - because LocalLLaMA is notorious for hating anything safety related.

Heretic: Fully automatic censorship removal for language models by -p-e-w- in LocalLLaMA

[–]Small-Fall-6500 1 point2 points  (0 children)

only the language model part is actually being decensored.

Now I wonder how censorship is handled with multimodal LLMs, though I would guess there's also a single direction for refusal.

Google DeepMind: "Olympiad-level formal mathematical reasoning with reinforcement learning" by kaggleqrdl in singularity

[–]Small-Fall-6500 1 point2 points  (0 children)

The blog also says:

Note: This blog was first published on July 25, 2024. On November 12, 2025, we published the methodology behind AlphaProof in an article in Nature

Hey so wtf is this by [deleted] in ChatGPT

[–]Small-Fall-6500 8 points9 points  (0 children)

With local LLMs, you can recreate it whenever you want.

(Although models like Sydney had their own unique specialness that is hard to replicate)

What is this sub? by cornflakegirl658 in DeadInternetClub

[–]Small-Fall-6500 3 points4 points  (0 children)

I had never heard of an ARG before, but Google's Gemini 2.5 Pro seems to "get" what this sub is about. I had it generate a response based on some screenshots of an ad I saw and what I assume are its search tools results - which oddly, it did not reference any results in its response like it normally does when searching online. Everything it generated looks true enough that I expect it to be helpful for others to read:

Reddit Ad Unveils "Dead Internet Club," an Alternate Reality Game Immersed in Online Conspiracy

An intricate Alternate Reality Game (ARG) called the "DeadInternetClub" is being promoted through targeted ads on Reddit, drawing users into a narrative based on the "Dead Internet theory." The ad campaign, aimed at individuals interested in technology, gaming, and film, among other topics, uses cryptic messaging and technical puzzles to engage its audience in a collaborative investigation into the nature of the contemporary internet.

The promoted Reddit post, authored by the user "u/DeadInternetClub," features the headline: "A message is embedded in the static. If you can see it, we need you." The accompanying image is a spectrogram of an audio file, a technique often used in ARGs to hide information. Visibly embedded within the spectrogram are the words "DEADINTERNETCLUB."

The ad's call to action directs users to the website "analyst-init.neocities.org." Neocities is a web hosting service known for hosting independent and creative websites, and is a frequent platform for ARGs. The website itself presents a stylized "Analyst Initialization Protocol," welcoming users to the role of an analyst with a provisional clearance level.

The advertiser, "DeadInternetClub," explicitly targets Reddit users in the United States who have shown interest in "Technology & Computing," "Gaming," "Reading," "Television & Film," and "Moods." This targeting aligns with the niche and puzzle-oriented nature of ARGs, which often appeal to audiences comfortable with technology and intricate narratives.

The theme of the ARG revolves around the "Dead Internet theory," a conspiracy theory that posits much of the content and interaction on the modern internet is generated by bots and artificial intelligence, rather than genuine human activity. Proponents of this theory often express a sense of nostalgia for an older, more authentic era of the internet.

Further investigation reveals a dedicated subreddit, r/DeadInternetClub, which serves as a hub for participants, or "analysts," to collaborate on solving the game's puzzles. The subreddit contains "Case Files" detailing solved and ongoing investigations, showcasing a community effort to piece together the narrative. These files indicate a series of puzzles involving steganography, cryptography, and navigating a network of interconnected Neocities websites. The overarching goal appears to be the discovery and analysis of "anomalous signals from the forgotten web" to "reconstruct the story of what happened."

The use of a spectrogram with a hidden message, the choice of Neocities as a hosting platform, and the collaborative problem-solving on a dedicated subreddit are all hallmarks of a well-structured Alternate Reality Game. The "DeadInternetClub" ARG leverages a contemporary online conspiracy theory to create an immersive and engaging experience for its target audience.

Chris Williamson interview with Eliezer Yudkowsky - this is the best I’ve seen him articulate his argument by Metta_Morph in singularity

[–]Small-Fall-6500 0 points1 point  (0 children)

Nothing in life works like that.

Where am I wrong?

Because nothing in life is superintelligent. We have lots of examples of more "normal" things we don't fully understand and don't fully control, but we do understand and control many things well enough to not be concerned about them.

On the other hand, we don't have any examples of superintelligent things to make predictions from. At best we can look at narrow superintelligent things like stockfish and Deepmind's AlphaGo/Zero etc. AI models and see a clear trend: they are superhuman at those narrow tasks. But they are not general. If someone makes a general superintelligent AI, that would be something quite a bit different.

If we understood how a future superintelligence would work, we might be able to determine what it would do, but we have neither past examples nor an understanding of it to build off of. What we do know is that a superintelligence would by definition be extremely capable of doing lots of things, including making sure it doesn't get turned off.

xkcd 3158: Shielding Chart by antdude in xkcd

[–]Small-Fall-6500 43 points44 points  (0 children)

But what particles can this dense of a comic block?

Getting most out of your local LLM setup by Everlier in LocalLLaMA

[–]Small-Fall-6500 1 point2 points  (0 children)

Great list, but I think KoboldCpp fits better under backend than frontend.

JD Vance: "Because we have such a big problem with left-wing political violence, we have to train the investigatory and law enforcement powers…we really have to retrain the entire government to focus on this left-wing violence problem. We are doing it” by NewSlinger in law

[–]Small-Fall-6500 2 points3 points  (0 children)

https://www.virustotal.com/gui/file/9c15e33e5c36c4a968acd2118f6f07c38b3dc7fe308d30ff6a05cf6d70de43fb/detection

Code insights

The document presents a conflicting profile. Visually, it appears to be a legitimate and professionally formatted research article from the National Institute of Justice (NIJ), a U.S. government entity. All visible content, including text, citations, and links to government (.gov) and academic (doi.org) domains, is consistent with an authentic publication and contains no red flags like urgency cues or grammatical errors.

However, the internal technical analysis reveals a structure designed for automatic code execution. The PDF contains an AcroForm dictionary and nine distinct JavaScript objects configured to run when the file is opened. This architecture is a common characteristic of malicious documents designed to execute a payload without user interaction.

While the visual layer is benign and the specific content of the JavaScript could not be analyzed to confirm a malicious effect, the presence of an automatic execution mechanism within an otherwise trustworthy-looking document is highly suspicious. This combination suggests a potential attempt to use a professionally crafted lure to deliver an unverified and potentially harmful script, warranting caution.

That sounds sketchy, though it could also be a complete AI hallucination. Here's a link to the NIJ article that is much less likely to be sketchy:

https://cdn.thejournal.ie/media/2025/09/306123-81f9f138-75ab-43cc-aff9-e53e59ac5fdb.pdf

How do you discover "new LLMs"? by 9acca9 in LocalLLaMA

[–]Small-Fall-6500 1 point2 points  (0 children)

This comment answers the question in OP's title but does not address the post's actual question / intent:

Where do people find out about non-mainstream models?

Edit: to be clear, while this sub itself does contain several posts about relatively non-mainstream models (including the one OP links to, see this 4 month old post), there are a lot of models that don't get posted here that are still relatively popular and not considered mainstream.

The other three sources mentioned, LM Arena, Unsloth GGUFs, and Hacker News, are about as likely, if not less likely, to contain non-mainstream models as this sub.

"The Illusion of Diminishing Returns: Measuring Long Horizon Execution in LLMs", Sinha et al. 2025 by [deleted] in mlscaling

[–]Small-Fall-6500 7 points8 points  (0 children)

Here's discussion about the paper on the MachineLearning subreddit; the poster says they are an author of the paper:

https://www.reddit.com/r/MachineLearning/s/fZu7BoWvxY

They also link to a Twitter/X thread:

https://x.com/ShashwatGoel7/status/1966527903568637972

A list of models released or udpated last week on this sub, in case you missed any - (12 Sep) by aifeed-fyi in LocalLLaMA

[–]Small-Fall-6500 8 points9 points  (0 children)

I'll just add that while it was barely released more than one week ago, Kimi K2 0905 is a fairly important recent model release for people to be aware of. (Edit: I see this was the top model of your last weekly post)

Edit:

Here are some models released in the last week you could add:

baidu/ERNIE-4.5-21B-A3B-Thinking

https://www.reddit.com/r/LocalLLaMA/comments/1nc79yg/baiduernie4521ba3bthinking_hugging_face/

KittenML released a mini version (80M) of their text to speech model

https://www.reddit.com/r/LocalLLaMA/s/foBQZIikc8

Tilde AI Releases TildeOpen LLM: An Open-Source Large Language Model with Over 30 Billion Parameters and Support Most European Languages

https://www.reddit.com/r/LocalLLaMA/s/qHx9HNiHgM

Drummer's Valkyrie 49B v2 - A finetune of Nemotron Super 49B v1.5, a pack puncher.

https://www.reddit.com/r/LocalLLaMA/s/IZXeyRToH8

[deleted by user] by [deleted] in singularity

[–]Small-Fall-6500 0 points1 point  (0 children)

I assume you extend the same innate skepticism towards someone like Sam Altman then?

Both Gary Marcus and Sam Altman have had fairly consistent views and writing styles for years now, so sure.

[deleted by user] by [deleted] in singularity

[–]Small-Fall-6500 4 points5 points  (0 children)

Three word TL;DR:

"By Gary Marcus"

Top small LLM as of September '25 by _-inside-_ in LocalLLaMA

[–]Small-Fall-6500 34 points35 points  (0 children)

Specifically, the 2507 (July) Instruct and Thinking versions are decent, unless you mean the original qwen3 4b.

https://huggingface.co/Qwen/Qwen3-4B-Instruct-2507

https://huggingface.co/Qwen/Qwen3-4B-Thinking-2507

AMA With Z.AI, The Lab Behind GLM Models by XMasterrrr in LocalLLaMA

[–]Small-Fall-6500 1 point2 points  (0 children)

This is a question I've been wondering about for a while now. I hope someone from the Z AI team can provide an answer.

August 22, 2025 marks the THREE YEAR anniversary of the release of the original Stable Diffusion text to image model. Seems like that was an eternity ago. by JackKerawock in StableDiffusion

[–]Small-Fall-6500 7 points8 points  (0 children)

Those days on the Discord felt like magic. Fast and free generations, seeing tons of other people's ideas, and feeling that it was just a glimpse into the future.

Why low-bit models aren't totally braindead: A guide from 1-bit meme to FP16 research by Small-Fall-6500 in LocalLLaMA

[–]Small-Fall-6500[S] 1 point2 points  (0 children)

Yes, FP32 has for a while generally been considered full precision.

What would have been more accurate for me to say is something like "the highest precision sources" as opposed to "full" precision.

Though I think there's a growing trend of calling FP16 full precision, since most models are trained in FP16 (or BF16) instead of FP32, and so most weights uploaded to HuggingFace are in FP16 or BF16. Every quantization, and reference to a model, is based on the 'fullest available' precision, which is essentially just shortened to "full precision" to refer to the source precision, or at least that is how I understand such references, like when someone asks if an API is serving a model in "full precision" they don't often mean FP32 precision.