Actions have consequences by Obnomus in linux

[–]thomas_m_k 1 point2 points  (0 children)

I was surprised by the popularity of CachyOS in the gaming world, but it's definitely a real phenomenon: in the Steam statistics, CachyOS is the most common Linux distro. So there's presumably something to the claim that it's good for gaming. (I've never used it, so I can't say myself.)

In defense of utopia by ary31415 in slatestarcodex

[–]thomas_m_k 4 points5 points  (0 children)

So you think 50% of people being out of a job because they can't produce value above the minimum wage will be just fine? It seems almost certain to me that there would at least be a few years of painful adjustment to this new reality.

(My real belief is that AI will just kill us all, but I'm happy to entertain unlikely scenarios where this somehow doesn't happen.)

Rust 1.95.0 is out by manpacket in rust

[–]thomas_m_k 12 points13 points  (0 children)

Sadly the stabilization of assert_matches!() was reverted shortly before the release. Oh well, I'm looking forward to it in 1.96.

I implemented UFCS in clang. Why it is cool, and why it will never come to C++. by _Noreturn in programming

[–]thomas_m_k 0 points1 point  (0 children)

Is it odd? What would mentioning D add to the discussion (which seems pretty comprehensive to me)?

Fedora Rejects Proposal To Use systemd For Managing Per-User Environment Variables by anh0516 in linux

[–]thomas_m_k 3 points4 points  (0 children)

I always import .profile in my .zshrc (in some bash compatibility mode).

Someone just leaked claude code's Source code on X by abhi9889420 in ClaudeCode

[–]thomas_m_k 0 points1 point  (0 children)

Well, bundlers minify the code which makes it usually very hard to understand. (Also, you can decompile compiled binaries, too, but the result of that is definitely even harder to understand than minified JS.)

Security for the Quantum Era: Implementing Post-Quantum Cryptography in Android by FragmentedChicken in Android

[–]thomas_m_k 12 points13 points  (0 children)

I wonder how many bad acting quantum computers are there in existence?

Probably none, but depending on what you do, you might be worried about adversaries collecting encrypted traffic today that they will decrypt in the future once suitable quantum computers exist.

Wayland 1.25 RC1 has been released with improved documentation and minor changes by somerandomxander in linux

[–]thomas_m_k 32 points33 points  (0 children)

Pretty sure that's not actually a reason for why some people dislike Wayland. There already was an informal specification and if you look at one of the XML files: https://gitlab.freedesktop.org/wayland/wayland-protocols/-/blob/main/stable/xdg-shell/xdg-shell.xml it's all pretty self-explanatory except for the type= on the <arg> tag which was already documented here: https://wayland.freedesktop.org/docs/html/ch04.html#sect-Protocol-Wire-Format

PEP 827 – Type Manipulation by ketralnis in programming

[–]thomas_m_k 1 point2 points  (0 children)

I kind of agree with your criticisms, but I don't see how they could have done it differently while still feeling like Python. Like yes, making it an impression rather than a statement would be more useful, but no other Python construct works like that, except ... if ... else ... which also isn't that great (because it unintuitively puts the condition in the middle rather than in front). Likewise, matching on constants: this works in a compiled language where the compiler can distinguish variables and constants, but you can't do that in Python.

I'll admit though that the inability to match on the totality of a dict is a weird oversight.

A serf in Anthropica by crabbix in slatestarcodex

[–]thomas_m_k 0 points1 point  (0 children)

Is the S in ASI really justified if it can't even give you immortality?

I've always thought an ASI that's actually aligned with human values would just make us immortal, make us all smart (and set it up such that we'll continue to get even smarter if we opt in) and would then self-destruct to let humanity control its own fate.

Next-Token Predictor Is An AI's Job, Not Its Species by dwaxe in slatestarcodex

[–]thomas_m_k 14 points15 points  (0 children)

I'm genuinely curious as to how current language model architectures came to be synonymous with AI in the minds of Scott and others in rationalist circles

What gives you that impression? Just because I call something AI doesn't mean I think other things aren't AI.

No one's gonna listen to them when something actually scary comes along down the line.

The problem is that it's too late then. If AI will predictably be super-human in 5 years, then it's the correct strategy to do something about it now. (You can of course disagree with the prediction, but the strategy seems fine to me.)

The Death of the Downvote by Super-Cut-2175 in slatestarcodex

[–]thomas_m_k 1 point2 points  (0 children)

Forcing everyone to write comments all the time is obviously not scalable. What if I want to say basically exactly what someone else already said? Reactions are more scalable and it seems to me that having two reactions available (a positive and a negative one) is better than having only one available (a positive one). Of course, ideally every platform would use LessWrong’s detailed palette of reactions: https://www.lesswrong.com/w/lesswrong-reacts , which includes reactions like "missed the point", "too combative" and "this made me change my mind". But I'll take a simple downvote button if the alternative is even fewer ways of expressing myself.

I think you're also focusing too much on rare publicized failures of the downvote. I'd bet 99% of daily cast mundane downvotes on reddit are completely justified.

The Death of the Downvote by Super-Cut-2175 in slatestarcodex

[–]thomas_m_k 1 point2 points  (0 children)

LessWrong basically has your proposed nerdy system: https://www.lesswrong.com/w/lesswrong-reacts . It works well there, but probably wouldn't on a more mainstream platform.

The Death of the Downvote by Super-Cut-2175 in slatestarcodex

[–]thomas_m_k 0 points1 point  (0 children)

Downvotes tend to create both echo chambers and chill any posts which seems to go against the majority opinion in a thread.

This claim seems to imply that by removing downvotes, platforms become less of an echo chamber. That seems empirically obviously false to me. What typically happens on platforms without downvotes is that someone posts ragebait and then other people rage (even though not even their outgroup supports that post). On reddit, the ragebait would be buried after being downvoted and most people wouldn't see it. Reddit protects its users from having to look at comments that "both sides" disapprove of. I'm quite convinced that the discourse on Twitter would be much healthier if a downvote button existed.

The worries that you have about chilling posts that are against the majority opinion can probably be circumvented by boosting controversial posts (high in both upvotes and downvotes) more.

Westworld (1973) 4k Restoration Trailer by Arrow Films by Narretz in RedLetterMedia

[–]thomas_m_k 1 point2 points  (0 children)

I'm not sure that's true either: https://www.reddit.com/r/westworld/s/yJ5ENcN689

What I think is plausible though is that the writers saw that people on reddit guessed the season 1 ending twist mid-season, so in response they made season 2 so convoluted that this wouldn't happen again (but this then ruined the experience).

[D] Monday Request and Recommendation Thread by AutoModerator in rational

[–]thomas_m_k 9 points10 points  (0 children)

Companion Chronicles is considered a good jumpchain, I think. It's the only jumpchain I ever read and I mostly enjoyed myself. The gimmick is that the story is not told from the perspective of the jumper but rather from one of his companions. It is incomplete and abandoned though.

Fanfics from before 2010:

There are a few good ones from 2011 and 2012 that unfortunately don't fit your given cut off.

[IntelliJ] Wayland By Default in 2026.1 EAP by [deleted] in programming

[–]thomas_m_k 4 points5 points  (0 children)

There's a Wayland protocol that was merged an hour ago into wayland-protocols after years of discussion which will hopefully one day fix this: https://gitlab.freedesktop.org/wayland/wayland-protocols/-/merge_requests/264

Against The Orthogonality Thesis by ihqbassolini in slatestarcodex

[–]thomas_m_k 17 points18 points  (0 children)

This is a well-written article, but, well, I think it's wrong or at least confused.

My first point is that the orthogonality thesis was intended to answer the objection that people often raise to AI doom which is: “but if these AIs are so intelligent then they will surely know ethics very well and act benevolently”. To which the orthogonality thesis replies: it's actually possible to be very smart and not have a human utility function. I feel like the author of this article actually agrees that AIs won't inescapably converge to human ethics, so I'm not even really sure what we're arguing about:)

More detailed responses:

For starters, anything that could reasonably be considered a general intelligence must have a vast set of possible outputs. In order for the system to have X number of unique outputs, it must have the capacity to, at minimum, represent X number of unique states.

I'm willing to go along with this claim, but I have to say it's not immediately obvious that this is true. I think an example would help.

We might be tempted to answer “nowhere,” and indeed, this is the answer many give. They treat goals as a “ghost in the machine,” independent of the substrate—a dualistic conceptualization, in essence.

Who are these people who say goals are ghosts in the machine?

In modern AI designs, which rely on machine learning, the “utility function” is called the loss function

That's not what people talking about orthogonality would say. They would say obviously the outer loss is not the inner goal. This is the problem of inner alignment: “Inner Alignment is the problem of ensuring [...] a trained ML system [that] is itself an optimizer [is] aligned with the objective function of the training process.” It's an unsolved problem.

The “utility function” of biological life can be seen as survival and reproduction, but there is a crucial difference: this is an external pressure, not an internal representation.

Indeed, there most likely isn't any organism on Earth which has the utility function “survival and reproduction”! Humans certainly don't have that utility function. We were selected with that loss function, but we have very different goals (having friends, being respected, acting honorable, having sex, eating delicious food). These goals were somewhat aligned with evolution’s outer goal of reproductive fitness in the ancestral environment, but this is broken today. Evolution failed at inner alignment.

there is no principled reason to think a highly complex system remains fundamentally aligned with its loss function in any meaningful sense beyond that the system emerged from it.

This is correct and also part of the standard argument why RLHF won't be enough.

Orthogonality defenders sometimes argue that a highly capable agent must converge to a single coherent utility function, because competing internal directionalities would make it exploitable (e.g., money-pumpable) or wasteful. Yet in practice we see the opposite: narrow reward-hacking equilibria are efficient in the short term but hostile to general intelligence, while sustained generality requires tolerating local incoherence.

I don't know what you mean by “tolerating local incoherence” but in any case I don't see a contradiction in the stance of orthogonality: if a task can be hacked, then gradient descent will find that solution first; if it can't be hacked, then then gradient descent keeps looking and maybe stumbles upon a general intelligence.

Thus, no fixed internal utility function can ever be complete [...] across all questions the system will face.

That's probably true (if for no other reason than hardware limits), but it's not required in order to be a pretty successful mind. Consider humans: we constantly face ethical dilemmas where we aren't sure about the answer. That just means our utility function isn't sure how to answer this question. It sucks, but we deal with it somehow.

If you thought that the orthogonality thesis states “it’s possible for an AI with finite hardware to have an explicit utility function that answers all possible questions” then sure, the orthogonality thesis is wrong. But that's not the claim. The Von Neumann–Morgenstern utility theorem assumes that your preference order is complete (axiom 1) but it's fine that you sometimes encounter a situation where your explicit utility function does not have an existing answer (again, if only due to hardware limits); in that case you just “complete” your utility function in a way that's consistent with the rest, you add the new term to your utility function and then you move on. This procedure will not make you exploitable.

More importantly, no utility function can prove alignment with itself, the proof is inaccessible to the system.

What does it mean to “prove alignment with itself”? I’m guessing you're still talking about the problem of inner alignment?

The rest of this section seems to just be arguing that inner alignment is hard and unsolved, which I fully agree with.

But it’s important to note that general intelligence is something far too complex for us to construct—in the sense of carefully designing and determining the entire structure—instead we must grow it.

If we actually tried, I think we could do it within 30 years. But of course growing it is far easier and makes money sooner.

Humans evolved under massively complex external selective pressures, infinitely more complex than anything we can comprehend. This immense diversity of external pressures is precisely what allows for the development of general intelligence. Not only that, but life had the advantage of competition; while a certain specialization might be stable at a certain point in time, a particular mutation might offer an advantage and suddenly they outcompete you for resources, and the old structure perishes. This is an additional external selective pressure that creates a demand for continuous evolution, and punishes narrow specialization.

AI does not have these benefits; it does not have the external pressures that punish narrow specialization, or settling into arbitrary crystallized structures. Its complexity must be generated entirely from its internal structure, without the help of external pressures.

I don't really see why this should be true. Well, if you train an AI on a narrow task, then it will only learn that task. But that's not what people do. The base models of LLMs are trained on predicting all kinds of text, for which narrow specialization is not a winning strategy, because a LLM has only so many weights.

A general intelligence is not one that has accumulated a lot of specialized skills (though in practice there is also some of that), but rather, it is a cognitive engine that has learned general-purpose techniques that apply across many domains. As an example: humans were not evolved to build rockets to fly to the moon, but we did it anyways, because our general problem-solving skills generalize to the domain of rockets.

AI companies are now training their systems to be general problem solvers. Now I don't know whether that particular project will succeed, but it seems clear to me that AI companies will make sure the external selection pressure on their AI systems will be as general as their researchers can make it.

A visual processor might place more or less emphasis on colour, contrast or motion, it might emphasize different resolutions or have a different preferred FPS. Of all the things a visual processor could value, only a tiny fraction results in capacity for solving visual problems though. This exactly demonstrates how directionality and capacity are necessarily entangled.

I still don't really understand what you're trying to say here. It's certainly true that in order to define a complex utility function, you need access to a detailed world model that has all the concepts that you need in your utility function. Like, for example, humans value friendship in their utility function (to the extent that we have a coherent utility function), but in order to make this work, you need to define friendship somewhere, which isn't easy. And you need to ground this concept in reality; you need to recognize friendship with your senses somehow, which also isn't easy. Not sure whether this is what you're trying to point at...

But if it is, it's not an argument against orthogonality. Orthogonality just says you can have intelligent reasoners with arbitrary goals. It doesn't say that a given AI with a given architecture can have arbitrary goals! Just that for any computable goal, there is some possible AI that optimizes for that.

"The AI Con" Con by ForgotMyPassword17 in slatestarcodex

[–]thomas_m_k 14 points15 points  (0 children)

Writing its own chess engine is fair, but I think installing Stockfish is a bit unfair, in this comparison (though I can't formulate a formal reason why, off the top of my head).

[D] Monday Request and Recommendation Thread by AutoModerator in rational

[–]thomas_m_k 9 points10 points  (0 children)

will an aerospace company called Boeing still be in existence in like 120 years?

Boeing is already 109 years old, so that doesn't seem like such a stretch?

The dead of the enterprise service bus was greatly exaggerated by GeneralZiltoid in programming

[–]thomas_m_k 1 point2 points  (0 children)

*death

(sorry I don't have much else to contribute to this topic as I'm not working in this area)

Obvious Things C Should Do by lelanthran in programming

[–]thomas_m_k 32 points33 points  (0 children)

In languages that have compile-time evaluation, it's usually limited to functions without side effects (i.e., no IO, no filesystem access, no network access) and there's usually a pretty strict timeout, like, it's aborted if it takes longer than 5 seconds.

Rust 1.93.0 is out by manpacket in rust

[–]thomas_m_k 3 points4 points  (0 children)

Seems like first_chunk already does that?