It’s a goddamn stochastic parrot, starting from zero on each invocation and spitting out something passing for coherence according to its training set.
“Not understanding what is happening” in regards to AI is NOT “we don’t jniw how it works mechanically” it’s “yeah there are so many parameters, it’s just not possible to make sense of / keep track of them all”.
A analogy might be a static biological “brain” custom grown to predict a list of possible next words in a block of text. It’s thinking, sorta. Maybe it could acknowledge itself in a mirror. That doesn’t mean it’s self aware, though: It’s an unchanging organ.
And if one wants to go down the rabbit hole of “well there are different types of sentience, lines blur,” yada yada, with the end point of that being to treat things like they are…
It’s a goddamn stochastic parrot, starting from zero on each invocation and spitting out something passing for coherence according to its training set.
“Not understanding what is happening” in regards to AI is NOT “we don’t jniw how it works mechanically” it’s “yeah there are so many parameters, it’s just not possible to make sense of / keep track of them all”.
There’s no awareness or thought.
There may be thought in a sense.
A analogy might be a static biological “brain” custom grown to predict a list of possible next words in a block of text. It’s thinking, sorta. Maybe it could acknowledge itself in a mirror. That doesn’t mean it’s self aware, though: It’s an unchanging organ.
And if one wants to go down the rabbit hole of “well there are different types of sentience, lines blur,” yada yada, with the end point of that being to treat things like they are…
All ML models are static tools.
For now.