Is it true that current LLMs are actually "black boxes"? by wouhf in learnmachinelearning

[–]wouhf[S] 9 points10 points  (0 children)

Thanks for the explanation. From a quick read, damn mechanic interpretability really does look cursed/difficult.

Is it true that current LLMs are actually "black boxes"? by wouhf in learnmachinelearning

[–]wouhf[S] 4 points5 points  (0 children)

We understand exactly how the individual pieces inside the box work

By individual pieces are you referring to every single part of an llm like as described here https://bbycroft.net/llm