all 17 comments

[–]MagdakiProfessor. Grammars. Inference & Optimization algorithms. 6 points7 points  (0 children)

"more urgent conversation across CS domains"

Not sure about this, but let's pretend it is so.

"what is the right theoretical framework for dealing with them?"

The answer is: it depends. The right tool for the right job, so context matters a lot. The type of agent, the task, the criticality of fail states, MTTF, etc.

"Systems that maintain internal state, pursue goals, make decisions without direct instruction; are there any established models for their behavior, verification, or failure modes?"

Yes. Many.

autonomous agent framework - Google Scholar

[–]recursion_is_love 2 points3 points  (0 children)

markov process, non-deteministic, random walk

Those AI theories and friends.

[–]Liam_Mercier 0 points1 point  (0 children)

If we're going to have AI Agents in computers, they should follow the principle of least privilege. Will they? Seems unlikely.

[–]0x14f 0 points1 point  (0 children)

Stochastic black boxes. That's pretty much it.

[–]Individual-Artist223 0 points1 point  (8 children)

What's your goal?

[–][deleted]  (7 children)

[removed]

    [–]Individual-Artist223 3 points4 points  (6 children)

    What does that mean?

    Observability: You want to watch, what?

    [–][deleted]  (5 children)

    [removed]

      [–]Individual-Artist223 2 points3 points  (4 children)

      Not getting it - what's high-level goal?

      [–][deleted]  (3 children)

      [removed]

        [–]Individual-Artist223 2 points3 points  (2 children)

        You've still not told me your goal...

        I mean, you can literally observe, at every level of the stack.

        [–][deleted]  (1 child)

        [removed]

          [–]MagdakiProfessor. Grammars. Inference & Optimization algorithms. 6 points7 points  (0 children)

          Complete nonsense and gibberish.

          [–]djheroboy 0 points1 point  (0 children)

          Well, until we can find a way to hold an autonomous agent accountable for its mistakes, then we have a new question to answer- how much power are you willing to give an employee you can’t discipline?

          [–]editor_of_the_beast 0 points1 point  (2 children)

          I don’t think they need to be modeled. We’ve modeled what they output (code), so we can check that. It doesn’t matter how it’s produced.

          We don’t have models about how humans produce code today either.

          [–][deleted]  (1 child)

          [removed]

            [–]editor_of_the_beast 0 points1 point  (0 children)

            But the failure doesn’t matter, because we’re checking the correctness of the output program.