You could say the same thing about rewiring a human’s neurons randomly. It’s not the powerful argument you think it is.
We don’t really know exactly how brains work. But when, say, Wernicke’s area is damaged (but not Broca’s area), then you can get people spouting meaningless but syntactically valid sentences that look a lot like autocorrect. So it could be that there’s some part of our language process which is essentially no more or less powerful than an LLM.
Anyway, it turns out that you can do a lot with LLMs, and they can reason (insofar as they can produce logically valid chains of text, which is good enough). The takeaway for me is not that LLMs are really smart – rather it’s that the MVP of intelligence is a lot lower a bar than anyone was expecting.
You could say the same thing about rewiring a human’s neurons randomly. It’s not the powerful argument you think it is.
We don’t really know exactly how brains work. But when, say, Wernicke’s area is damaged (but not Broca’s area), then you can get people spouting meaningless but syntactically valid sentences that look a lot like autocorrect. So it could be that there’s some part of our language process which is essentially no more or less powerful than an LLM.
Anyway, it turns out that you can do a lot with LLMs, and they can reason (insofar as they can produce logically valid chains of text, which is good enough). The takeaway for me is not that LLMs are really smart – rather it’s that the MVP of intelligence is a lot lower a bar than anyone was expecting.