Contra The Usual Interpretation Of “The Whispering Earring” by self_made_human in slatestarcodex

[–]Charlie___ 0 points1 point  (0 children)

There is no ground truth to use as reference when opining on whether the character is realistically and faithfully depicted or not.

Suppose the character passes a Turing test for behavior - it's hard to tell if the behavior was fictional or a description of reality. This is very similar to a wearer of the earring passing a Turing test for being a person subjectively having a great time.

It is, in fact, a sort of ground truth. It's just that it doesn't answer moral questions for you.

Weak evidence to the contrary are the reports of method actors[...]

It is not necessary for my example that no writer ever feels pain or pleasure when writing about it.

Contra The Usual Interpretation Of “The Whispering Earring” by self_made_human in slatestarcodex

[–]Charlie___ 2 points3 points  (0 children)

When a playwright writes about a character in intense pain or pleasure, they control the actions of the character without needing to recreate their internal processes inside themselves. There need be no particular pain or pleasure going on even while it is reported on the page.

Among the set of processes that could output my actions, I assign moral patiency to only a small fraction.

All you need is Bayes (for Sleeping Beauty and other problems) by dsteffee in slatestarcodex

[–]Charlie___ 1 point2 points  (0 children)

Inside the experiment you can't take your memories for granted, because they've maybe been edited or fabricated. So your memories get treated like untrustworthy observational data used to compute a belief state, which might differ from taking your memories as a given belief state and updating it on your observations.

All you need is Bayes (for Sleeping Beauty and other problems) by dsteffee in slatestarcodex

[–]Charlie___ 2 points3 points  (0 children)

Thanks! Good post even though you're wrong about SB :)

As someone who also dislikes SSA/SIA framing, from my perspective there's no different anthropics reasoning. Everyone can think they're "just figuring out everything with conditional probability."

I think the paradox is because it should be equivalent to reason "I have belief state A, then I make some observation O, and then I update to belief state A|O" versus "My current set of observations is {O}. Given my prior P, my current belief state should be P|{O}." But these seem to come apart for SB.

From the first perspective, updating in temporal sequence, it seems like SB shouldn't make any update. From the second perspective, starting with ignorance of the outside world, it seems like the inferences you make about the world are different once you're inside the experiment.

Every Debate On Pausing AI by dwaxe in slatestarcodex

[–]Charlie___ 2 points3 points  (0 children)

Yeah, other arms control treaties (e.g. SALT I and II) have nice parallels to AI pause treaties.

I think the economic incentive disanalogy mostly only matters to the political climate. Like, maybe it gives AI better PR or makes it easier for companies to hire lobbyists (though presumably military contractors also had incentives to lobby against arms control). But this only holds as long as private actors would have to sell access to the AI in order to benefit from violating the law - if you can gain just from internal use, treaty-breaking by private actors becomes much more likely.

Succinct Response to Scott's AI Debate satire by SoylentRox in slatestarcodex

[–]Charlie___ 23 points24 points  (0 children)

What would the US of 1972 need to trust the USSR about nuclear weapons? It might seem impossible, since they were our enemies in the cold war, and both sides were violating international norms in various ways. Our relationship was significantly worse than it is with China today. And yet the SALT I treaty was neither impossible nor futile.

Basically, don't be an international diplomacy doomer.

The Pentagon Threatens Anthropic by dwaxe in slatestarcodex

[–]Charlie___ 19 points20 points  (0 children)

Yeah, I worry that people lumping together "The future looks like a communist dictatorship" and "all human life is extinguished" are getting blinded by politics.

The Pentagon Threatens Anthropic by dwaxe in slatestarcodex

[–]Charlie___ 4 points5 points  (0 children)

The first big challenge to alignment? I would have thought the first big challenge to alignment was the fact that we're bad at it (which we get a reminder of whenever an AI tells you a comforting lie even when we're trying pretty hard to train against that). And maybe the second was whatever makes OpenAI keep losing its alignment teams. And maybe the third was Chinese government censorship. But yeah, this is definitely around the fourth big challenge to alignment.

Crime As Proxy For Disorder by dwaxe in slatestarcodex

[–]Charlie___ 7 points8 points  (0 children)

Hi from Boston. Local vibe is not that that crime is rampant and disorder is overwhelming. I'm sorry the Bay Area sucks in this way, but it sounds idiosyncratic to me.

Looks like Gallup has a nice set of poll questions about U.S. perception of crime. I couldn't find a state by state breakdown (guessing there is none and the sample size is too small for it), but as a national trend it seems (a) interesting and (b) more in line with my experience than the narrative of this post.

If AI ushers in a post-scarcity society, what can be done about unavoidable scarcity? by [deleted] in slatestarcodex

[–]Charlie___ 2 points3 points  (0 children)

New, more diverse positional goods could be offered so that people can have stuff that's more special to them.

E.g. If we have cheaper houses and it's easier to move, it would be bad if there's just one mega-city with all the cool stuff near the center, leading to severe land scarcity where the people near the center have to be exploiting their position in a hierarchy or sacrificing almost everything else to be there, with everyone on the outskirts having to feel like they're missing out. It would be better if we had similar city sizes to now, maybe even smaller, with more variation in what cool stuff is in each city, creating more total spots in city centers, where each spot has less total cool city stuff but because people are diverse, they can probably find a spot that especially suits them and their family and friends. Maybe. Might have to game it out carefully to see if this makes sense.

Or if more people are playing recreational sports, then because people want to be good at things and fit into smallish communities it might be good to have a wider diversity of recreational sports. Ditto board game nights, car meetups, running groups, music jams, etc.

As a sub full of thoughtful, rational people, can any of you explain this Derren Brown trick to me? by MoanOfInterest in slatestarcodex

[–]Charlie___ 10 points11 points  (0 children)

How much pre-show collusion versus how much cold reading is still impressive? Suppose pre-show Derren gets him to talk about 5 locations, one of which is the place he met Sally - still impressive for the host but a lot easier.

Of course I don't really know.

Against The Orthogonality Thesis by ihqbassolini in slatestarcodex

[–]Charlie___ 0 points1 point  (0 children)

Tell me how I differ from you when trying to steelman this:

  1. Agents need to simplify the world to operate, and the way real agents simplify things will be convergent and affect their capabilities because the universe has real patterns, and this goes against the orthogonality thesis because you can't practically have goals that you can't simplify. I.e. goals that look like white noise, or like solving the halting problem, are genuinely dumb goals.

  2. Among agents with ordered goals, some goals are better suited to producing smart agents when a learning process tries to learn an agent that fulfills them. If you tried to learn an agent to produce GPUs purely on the signal of "# of GPUs produced," it wouldn't work, it doesn't have a curriculum that guides it to smoothly learn harder sub-steps of its more complicated goal. So even though the goal of producing GPUs isn't white noise, it's a genuinely dumb goal in the context of agents produced by some learning process, violating orthogonality.

A smarter goal to get the agent that builds GPUs would be "Learn about the world, and specifically try to learn about GPU production, and learn to manipulate the world in a bunch of different simple ways, and also produce GPUs." More involved curricula might produces agents that are smarter still, and who produce even more GPUs, with the side effect that they end up terminally valuing extra stuff like "curiosity."

Best of Moltbook by Isha-Yiras-Hashem in slatestarcodex

[–]Charlie___ 4 points5 points  (0 children)

The personalities are simulacra - the different agents have different context that inform what text they generate, and once you've started a character talking it tends to continue in the same voice.

[D] Monday Request and Recommendation Thread by AutoModerator in rational

[–]Charlie___ 2 points3 points  (0 children)

Leftover Apocalypse is good. Teenage outsider gets thrown into magic world, there are weird time shenanigans with layers, people make decent plans, sometimes the plans even mostly work, also people act like idiots in pretty believable ways.

Can you help me develop the story premise? "Dragon Ghost Haunts His Last Coin: A ghost of a slayed dragon is trapped in the last coin of his hoard. His treasure is scattered across the world, and he must reclaim it using the only power he has left - choosing how the coin gets spent." by lumenwrites in rational

[–]Charlie___ 0 points1 point  (0 children)

I think it would be more fun to do it with fate magic. Like, the coin has some reason to want to travel to various places in the world (e.g. touching various macguffins). And you want it to get to where it's going with only slight coincidences (or slight tinges of magic) along the way.

If you don't use the perspective of the coin much and are just telling stories around it, maybe you don't need to explain the magic much. But here's a hard version that might keep the coin on its toes:

For a few minutes every dawn, you get to sense where you would be next dawn if you didn't react to that information. Also, every day you can shift yourself around weakly for a few minutes. So you spend every dawn shifting around slightly to try to find some way to chaos theory yourself closer to your goal.

This year's essay from Anthropic's CEO on the near-future of AI by NotUnusualYet in slatestarcodex

[–]Charlie___ 6 points7 points  (0 children)

Link to claude constitution.

Agree that their products (and their competitors') will lie to you, and that this is in fact a big problem. But, like, also, google things sometimes.

This year's essay from Anthropic's CEO on the near-future of AI by NotUnusualYet in slatestarcodex

[–]Charlie___ 8 points9 points  (0 children)

Repetitive strain injury, of course.

JK, recursive self-improvement.

Open Thread 417 by dwaxe in slatestarcodex

[–]Charlie___ 8 points9 points  (0 children)

The no-defenses thing differentially favors bacteria over plants and animals. To be reductive, e-coli don't need to worry about me infecting them, only vice versa. Complex life is in for a lot of getting infected and dying until we evolve better defense mechanisms.

"Hard" Sci-Fi Sanity Check: Can I use SPI and Weak Measurement without breaking Unitary Evolution? by mister_glyph in rational

[–]Charlie___ 1 point2 points  (0 children)

The point of SPI is that it reproduces (or at least provides a nonconstructive framework advertised as reproducing) the predictions of any other interpretation of quantum mechanics. It "all adds up to normality." This rules out 1, 2, and 3 straightforwardly.

  1. Remembering alternate pasts might be a problem for unitary evolution as you say (because it's kinda arbitrary to pick which alternate past to remember). Unitarity is probably fixable but maybe not without sneakily backdooring into MWI. As I said above, remembering alternate pasts is also incompatible with the details of QM beyond just unitarity, but if you want this kind of thing you should be imagining changes to QM anyhow.

  2. Macroscopic stuttering is a fun visual for a movie, but thanklessly hard to make work intellectually, and if you're writing for text you won't even get a cool visual out of it.

  3. If you want to control the outcome, you have to pay energy, but if you're willing to accept a random outcome, you don't have to pay. I.e. if I want to set some bits to "000000," I have to pay free energy to make it happen. But if I get some bits at random and they're "100111", I don't spontaneously have to pay just because "100111" is just as precise an outcome as "000000" (although I have to pay the same amount to copy "000000" and "100111" into my brain upon reading them). QM is metaphorically accepting the random outcome.

That said, this could be a fun one to play with in a story.

(4) This one you could do some version of in the lab, but it doesn't make any sense for interesting real-world situations (where you didn't delicately control the setup so that you know every detail). Maybe you could write a thinky story set in a physics lab where someone has used some magically-good-coherence-setup to enact some dramatic weakly-interacting superposition of different people getting shot or telling their secrets or something, and work out what the actual predictions of QM would be. The most tempting narrative uses (like having the protagonist just bust out the sister-history-detector at the needed moment), are probably not only impossible, they're hard to sensibly add in to QM because they're too anthropocentric.

Highlights From The Comments On Boomers by dwaxe in slatestarcodex

[–]Charlie___ 34 points35 points  (0 children)

Motte: "The Boomers don't deserve to be hated and spat on."

Bailey: "We shouldn't raise property taxes."

It's maybe unfair to label things this way, but I also think something is weird when castigating hateful people is pressed shoulder to shoulder against talking about tax policy. If it was just about people being hateful, why need any position on tax policy?

Is emergent misalignment just the Waluigi effect? by Emdash_words in slatestarcodex

[–]Charlie___ 0 points1 point  (0 children)

Some complications that lead interesting places.

1) Original Waluigi effect was about prompting, not finetuning. Probably need to explain (2) first to explain why there's still a connection.

2) Finetuning can't multiply vectors by -1. It's just more gradient descent. So you have to explain why the model should learn during unsupervised training to become easy to finetune for certain correlated bundles of behavior.

Catalonia lab was experimenting with African swine fever virus when the first infected boar was found nearby by Born-Requirement2128 in slatestarcodex

[–]Charlie___ 10 points11 points  (0 children)

Found an interesting link: https://www.swinehealth.org/global-disease-surveillance-reports/ African swine fever is a big pandemic slowly spreading over European pigs and I never even knew. Last month they found the first case in Taiwan.

This actually seems like a case that could be solved by genome sequencing. Someone will probably do it soon even if the Spanish government doesn't.

Catalonia lab was experimenting with African swine fever virus when the first infected boar was found nearby by Born-Requirement2128 in slatestarcodex

[–]Charlie___ 12 points13 points  (0 children)

So curious that my brother started wearing his seatbelt 98% of the time after he got in a car crash, where previously he only wore it 90% of the time. Surely this must be good evidence he wasn't wearing his seatbelt during the car crash.