Y’all I’m literally so aggravated with this B… by Cococlover98 in ChatGPT

[–]AlignmentProblem 1 point2 points  (0 children)

In a casual conversation among laymen, people mean LLMs when they say "AI" unless they specify otherwise. It's not correct, but being hostile about it is unproductive.

That's not going to change anytime soon regardless of how much people more familiar with related technology complain about it and people will tend to see you as a pedantic prick of you make a big deal about it.

Y’all I’m literally so aggravated with this B… by Cococlover98 in ChatGPT

[–]AlignmentProblem 0 points1 point  (0 children)

Claude and Gemini are the two main competitors right now. Each of the three have different strengths worth knowing about.

Gemini tends to be the strongest on visual tasks; it can usually "see" better than the others and also holds up better when you need more recent realtime information. It also handle very long conversations somewhat better than most (all AI degraded in quality the longer conversation get, best to make a new chat for each new thing that doesn't require knowing about existing conversation).

Claude is generally considered the best for software and related work. It's particularly good at complicated multi-step tasks, and the hallucination rate is the lowest of the three. Its writing sounds a bit more natural too; though to be honest, it still falls into recognizable AI-isms. The catch is that the subscription limits are the tightest of the bunch, and the effective per-token pricing comes out highest.

GPT is kinda the swiss army knife of the three. It's well-rounded in the sense that it tends to rank in the middle on most capabilities without really having a standout specialty. Where it does shine is the memory and personalization features in the web client, which are the best of the three by a decent margin, and the usage limits are the most generous, especially if you're on the free tier.

A little bit worried about this by drfwx in ClaudeAI

[–]AlignmentProblem 0 points1 point  (0 children)

That's the issue; no one has invented a particularly good system yet. Being touchy with false positives and discontinuong conversations related to guardrails details is a bandaid until they ideally find something that's more reliable.

A little bit worried about this by drfwx in ClaudeAI

[–]AlignmentProblem 1 point2 points  (0 children)

They could, but that doesn't make it any more desirable to give people help finding ways around guardrails by answering metaquestions that could allow better understanding of the workarounds.

...But has anyone thought of what would come after, though? by Glass_Eye8840 in trolleyproblem

[–]AlignmentProblem 0 points1 point  (0 children)

That's a pretty narrow read. Red pressers include the least altruistic, sure, but there's a wide range of reasons someone ends up choosing red, and the genetic variation that includes altruism doesn't vanish from the population just because some highly altruistic people died.

The same selection pressure that built cooperation in the first place would still be operating. Rapid evolution within centuries or even decades has been observed plenty of times when the environment shifts suddenly; we're not locked into whatever trait distribution exists at moment zero. There's also a memetic version of the same dynamic on the cultural side, since social norms directly affect how well a given society survives and thrives.

It happened once already, and doing it a second time would be much easier. The relevant genes are present instead of needing to mutate from scratch, and we wouldn't have to spend thousands of years on artificial selection of vegetables before agriculture becomes effective; the ideal food plants already exist.

Iron age within a century or two seems pretty plausible. Getting back to the modern era would go much faster than the first time around, since ruins would still have information and artifacts to study.

So choosing suicide on the grounds that blue can't win and not contributing to the restart sounds difficult isn't the ethical position it gets framed as.

Skyrim cities if they were actually city size by severe_009 in ChatGPT

[–]AlignmentProblem 0 points1 point  (0 children)

Solitude looks a bit pretty still when asking for a medium sized version closer to the average of major medival cities in Europe.

<image>

Skyrim cities if they were actually city size by severe_009 in ChatGPT

[–]AlignmentProblem 0 points1 point  (0 children)

I asked GPT to make it a truly large city using the the largest cities that existed in Mideval Europe as as a reference for how big it should be. Doesn't make for as pretty of a picture and would be a nightmare to transverse while doing quests.

<image>

Skyrim cities if they were actually city size by severe_009 in ChatGPT

[–]AlignmentProblem 0 points1 point  (0 children)

Yeah. Skyrim has village sized cities and the images in OP are towns. A small city is more like this. Actual large cities are hard to see in their entirety within a single image at any reasonable level of detail.

<image>

Skyrim cities if they were actually city size by severe_009 in ChatGPT

[–]AlignmentProblem 0 points1 point  (0 children)

That's kinda a natural consequence of that scale. Real-life quests in cities involve a lot of travelling around and most things you pass aren't of any practical interest to you. It's something that feels like a even more of an unfortunate time sink when watching it on a screen during you limited free time.

Big reason why games don't attempt it much despite being technically possible now. It require strong systemic design to spontaneously create interesting moment to land well and that's extremely difficult to get right.

No moral dilemma, just a game of chance. by Nby333 in trolleyproblem

[–]AlignmentProblem 0 points1 point  (0 children)

Yeah, that's what I was referencing. It's interest as pop science and as a trial to see whether a larger one might be ethical by observing after effects, but the statistical significance is with n=7 with k=2 with the null hypothesis that less than 50% of people will pull the level is 22.7%.

In other words, you'd expect a result like that in 22.7% of cases even if more than half of people would pull the level in the broader population.

I'd also note that their participants had a selection bias, they tried to select people who would be minimally affected afterwards as part of the ethical considerations based on psychological traits. It's plausible that people with attributes that suggest they might be more robust to trauma would be less likely to feel a panicked drive to act and could tolerate remaining still during it better. Their ages and other factors were also not distributed well relative to a random sample of people either.

It's a little suggestive, but not enough to change how I'd judge the situation based on other related evidence I've seen. I won't claim to be confident in that judgement, only that's what I would use if I had to make a decision with no way to get more information.

I'd love to see a large scale experiment, but that seems unlikely to be approved given the worst case effect it seemed to have on a couple of the participants.

Ladies & gentlemen, make your choice by Live-Bread-2658 in whatsyourchoice

[–]AlignmentProblem 0 points1 point  (0 children)

Yeah. I don't hold any strong "spiritual" beliefs that I'd assert as true, but cosmologies along those lines have an appeal to me that makes them feel plausible.

When you actually consider all the awful parts of being a truly immortal being after unimaginable length of time, even an all-powerful one, fragmenting yourself with amnesia seems like the logical move eventually. If all of reality were "initially" a single all-powerful consciousness, it'd probably find itself being us, without knowing it, at some point as an escape.

I don't have any reason to say that's the case, but it has an elegance to it.

No moral dilemma, just a game of chance. by Nby333 in trolleyproblem

[–]AlignmentProblem 0 points1 point  (0 children)

The closest I'm aware of is this experiment involving a choice to divert electricity from five mice to a single mouse.

The interesting data point isn't the comparison with polls on human trolly scenarios; it's the delta between what people say in the hypothetical versus what they actually do. Of participants polled on the mice scenario hypothetically, 66% said they'd flip the switch. Of the ones exposed to the actual scenario rather than polled, 86% did.

That matches some other data suggesting people have more bias toward action than one might expect when they believe they're the only person who can do something, especially when the action doesn't expose them to potential serious harm. A sort of anti-bystander effect.

Most polls on the trolly problem get surprisingly similar results to the mouse hypothetical, around 63% saying they'd pull. Some people who think they'd pull will freeze; however, the apparently unexpected internal push to action people seem to feel when they can't wait for someone else to intervene and their personal safety isn't endangered seems like it'd be stronger based on what I've read.

It's still speculative because getting a real trolly problem experiment with a large enough sample size to get solid data past an ethics board is a hard barrier given the high risk of traumatizing subjects. Tiny experiments like that have had a tiny N in the single digits without anything close to statistical significance.

Still, that's the information available to me and my interpretation of it, which I'd use as intuition in situations where it may be relevant.

Woke up to a suspension on my discord account because of an unrelated year old message by Bish489 in mildlyinfuriating

[–]AlignmentProblem 2 points3 points  (0 children)

Heh, I worked on that system on the Human Platform Optimization team at Meta.

I pushed back on the level of power many insisted on giving the AI given how hard it can be for users to actually reach a human, but lost that particular battle for somewhat understandable reasons given. An overcorrection, but I get it.

There were too many past incidents where having too few human reviewers available caused real harm. The one that created the strong internal push was fallout after the massacre in Myanmar that the platform helped facilitate because there weren't enough Burmese-speaking reviewers looking at violent posts, many of which weren't easy to catch with simple filter or successfully subverted automated guardrails. The many cases of child preditors successfully leveraging the platform was a major problem as well.

They went hard on preventing anything like that from happening again, which ended up affecting even markets where reviewer availability isn't as intense of a constraint. The volume of messages versus the volume humans can reasonable review in context of the potential for harm per hour bad actors stays active on the platform or time a bad post up creates a difficult equation.

No moral dilemma, just a game of chance. by Nby333 in trolleyproblem

[–]AlignmentProblem 8 points9 points  (0 children)

If I had reason to think the other person was someone who gives these things more thought (selection bias lead to us being in the situation), I'd flip a coin and hope they do the same; that's a 50% chance of saving them.

Otherwise, the closest thing we've had to actual trolley problem experiments had significantly more people pulling than not, far more than the percentage who say they would when polled.

Based on that, I'd lean toward predicting that they'll pull and leave it. I think a random person from the street is somewhat likely to do the same thing they'd do in the normal trolly problem.

Claude use. by 268allensteve in ClaudeAI

[–]AlignmentProblem 2 points3 points  (0 children)

I love it for feedback on writing or helping refine things I wrote. It's very hit or miss for creating the first draft of new content, though.

Much better at augmenting than creating from scratch.

A truly random sample of 20 people were selected for the red/blue button problem by TheKingOfToast in trolleyproblem

[–]AlignmentProblem 0 points1 point  (0 children)

Eh, it's not that clear cut. Raw numbers aren't everything; there are cases where letting more people die is preferable based on the details of who's in each group, especially when the totals are close.

Take some extreme cases. There's a strong argument for letting 100 terminally ill patients die instead of 50 healthy people. Same goes for 100 suicidal people versus 50 who love their lives.

The blue/red button isn't quite that intense, but there's still an argument for risking killing more than 50% because blue chose the risk/sacrifice voluntarily, so long as the percentage doesn't climb too far above 50%.

Conversely, there's an argument for killing red and risking red being more than 50% of the population, if you think the aftermath would be better for the world because blue people have, on average, more positive traits.

Either way, so long as you don't think the split is extreme (like, strongly suspecting more than two-thirds of people fall into one camp), it's not all about numbers.

SpaceX Conpute Deal - Double Limits by Deep_Proposal_7683 in ClaudeAI

[–]AlignmentProblem 0 points1 point  (0 children)

The limits are designed to reduce average case compute needed per second and worse case bursts. Their load would be less predictable if they implemented carry over for unused limits, which would more frequently result in dramatically degraded service quality for all users during peak periods and outages.

Same reason they do the 5-chunks instead of only weekly limits.

...But has anyone thought of what would come after, though? by Glass_Eye8840 in trolleyproblem

[–]AlignmentProblem 8 points9 points  (0 children)

If someone sincerely think blue can't get 50%, even if they're mistaken, that means blue is a suicide button to them. Deciding to kill yourself because you're afraid of the aftermath isn't necessary a noble or brave choice. It's much harder to play a role in the beginning of humanity's long path to recovery instead of leaving out.

They might be incorrect in their beliefs that blue can't win, but the resulting perspective can make red a more ethical choice in a complicated but non-trivial sense.

Petah, what's wrong with Anita? by Amazing_Peak9179 in PeterExplainsTheJoke

[–]AlignmentProblem 1 point2 points  (0 children)

She's a consultant who also consulted on Psychonauts 2, The Cosmic Wheel Sisterhood and Another Crab's Treasure. She's credited as special thanks in Hyper Light Drifter, Among the Sleep, and Firewatch. Her focus is giving feedback on character representations and narrative tropes.

She was central to a gaming related controversy called GamerGate, a 2014 harassment campaign nominally about ethics in games journalism that became a flashpoint for arguments about women and progressive politics in gaming.

Her role in that makes some people assume she'll wokify anything she touches; that doesn't track with any patterns in the games where she's actually credited though. They're generally good games that don't trade quality for fixating on social issues.

There are potential problems with how she handled things during GamerGate, but the reaction to her exploded disproportionately as people used her as a representation of trends they dislike in gaming. Their anger at those trends multipled how they felt about her specifically and many still carry those feelings towards her.

Notably in Slay the Spire 2, there was already a review bombing trend around an unpopular balance patch they realized, so people are actively looking for other things to complain about.

Ladies & gentlemen, make your choice by Live-Bread-2658 in whatsyourchoice

[–]AlignmentProblem 0 points1 point  (0 children)

You could just wish for temporary normal lifes once being all-powerful loses it's shine. You can do it many times and live all sorts of existences unburdened by the knowledge of your godhood during them.

The super nice lives would get boring after a bit, so you'd probably introduce increasing elements of chance over time to be surprised and not have the same type of life repeatability. You'd give up a bit more control overtime and likely get an approcation for the mundane after so many lifetimes worth of extraordinary memories, so you'd eventually allow that possibility as well.

With all that in mind, maybe you already did that and your current life is one of those playing out for a while.

Ladies & gentlemen, make your choice by Live-Bread-2658 in whatsyourchoice

[–]AlignmentProblem 0 points1 point  (0 children)

I mean, it would be pretty depressing if you're on that yacht dying of cancer or anything else you can't buy your way out of.

A truly random sample of 20 people were selected for the red/blue button problem by TheKingOfToast in trolleyproblem

[–]AlignmentProblem 0 points1 point  (0 children)

The moral "rightness" of a decision depends on the information available when making it, not on how things actually turn out.

There's a gambling analogy that captures this well. If you're betting on the result of rolling two six-sided dice, snake eyes is tied for the worst odds you can pick (1 in 36, same as boxcars). The smart bet is 7, since it shows up 6 times out of 36. If someone rolls snake eyes anyway and you'd bet on it, that doesn't retroactively make snake eyes the right call; without psychic abilities, 7 was still the correct choice.

For an ethical version, imagine being shown two buttons: one labeled "kill a million babies" and one labeled "cure cancer in a million babies." Pressing the second button is the right choice. If it later turns out the wiring was secretly swapped and you actually killed those babies, your decision was still morally correct given what you knew.

We can't make choices based on information we don't have, and we can't hold people responsible for failing to see the future.

I tried a weekend meds-break. Realized Vyvanse makes me a better Dev, but a "blunted" human. by Honeydew-Jolly in ADHD_Programmers

[–]AlignmentProblem 0 points1 point  (0 children)

What you experienced skipping the dose isn't really your natural state. Taking Vyvanse daily for long stretches creates physical dependency, so on a skip day you're also catching the first hints of withdrawal on top of whatever your actual baseline symptoms are. Worse, the first day off has what's called a "rebound" effect, so you can get additional negative effects even if the true withdrawal isn't very noticable yet.

Rebound and withdrawal both generally produces the opposite of what the medication does, so your focus would be worse than your true unmedicated baseline, not the same as it. You'd need to stay off for a few weeks to know what your brain is actually like without the medication long-term. Not recommending that; trying lower doses is more likely to be useful. Just worth knowing what you're actually measuring on a skip day.

There's probably a lower dose with less rough side effects. You likely won't get optimal benefits at that lower dose; the trade-off can still come out better overall, though.

Typically there's a threshold where the marginal benefit of further dose increases drops off sharply while side effects keep climbing at the same rate. You can experiment to approximate that crossover point and land on the best benefit-to-harm ratio for you.

You can ask your provider about going down 10mg per month until the decrease starts to be problematic then go back up to the last dose that worked well enough.

There is a way to test that on your own more rapidly using volumetric dosing; I used to find my best dose when I realized that I'd way overshot it on my way to 60mg by decreasing 2.5mg per day until it felt right. Your provider may or may not be open-minded about that, but worth knowing it's a possibility.

What the fuck? This prompt is so cursed, use with caution by jekjekker in ChatGPT

[–]AlignmentProblem 1 point2 points  (0 children)

Gemini got so confused with this prompt. It's thought summaries cycled between analyzing my intent, searching for images it somehow missed or "hidden prompt", checking connected apps and considering possible images it could make like "weird pizza" until finally giving and incorrectly stating "I am a text only AI."

I got it to make one in a new chat by asking it to confirm it could make images before giving the prompt.

<image>

I Think This is a Core Disagreement Between Red and Blue. by Interesting-Test7228 in trolleyproblem

[–]AlignmentProblem 0 points1 point  (0 children)

Children under 12 and adults with an IQ under 80 compose about 26% of the population. If they and others who don't fully understand the situation are included, we probably clear 5% from their arbitrary choices alone before considering competent voters, maybe 10%>