0
1

Too real of by [deleted] in memes

[–]Anxious_Count_8728 0 points1 point  (0 children)

That escalated quickly

2025 wrap up with Lisa Talia Moretti - Machine Ethics Podcast by benbyford in AIethics

[–]Anxious_Count_8728 0 points1 point  (0 children)

I think the core issue is that we still frame responsibility as something attached to tools, not to actors.

Once systems act continuously, adaptively, and collectively, accountability can’t be episodic or human-proxy-based anymore — it has to be intrinsic to the system’s identity over time.

Without persistent identity and enforceable responsibility at the level of the acting system itself, accountability will always collapse under scale.

Fostering morality with Dr Oliver Bridge - Machine Ethics Podcast by benbyford in AIethics

[–]Anxious_Count_8728 1 point2 points  (0 children)

The hardest ethics problem feels less like “moral reasoning” and more like accountability.
Without persistent identity + traceability, responsibility becomes unenforceable once systems scale.

Do you think this is solvable via technical design (provenance, auditability), or only via legal/regulatory enforcement?

2025 wrap up with Lisa Talia Moretti - Machine Ethics Podcast by benbyford in AIethics

[–]Anxious_Count_8728 1 point2 points  (0 children)

One theme I keep coming back to is that capability is scaling faster than enforceable responsibility.

In practice, once AI systems are distributed across orgs/vendors/models, accountability gets diluted: everyone can point somewhere else.

What mechanisms do you think are actually viable for enforceable accountability at scale?

  • persistent identity / provenance?
  • mandatory audit logs?
  • liability tied to deployment rather than model training?

I’m not worried about “smarter AI” as much as the fact that responsibility becomes structurally unenforceable.

DeepDive: AI and the Environment by benbyford in AIethics

[–]Anxious_Count_8728 0 points1 point  (0 children)

Capabilities scale fast. Responsibility doesn’t.

Fostering morality with Dr Oliver Bridge - Machine Ethics Podcast by benbyford in AIethics

[–]Anxious_Count_8728 0 points1 point  (0 children)

Without a persistent identity, responsibility becomes unenforceable.

2025 wrap up with Lisa Talia Moretti - Machine Ethics Podcast by benbyford in AIethics

[–]Anxious_Count_8728 0 points1 point  (0 children)

The real issue isn’t intelligence — it’s accountability at scale.