Preferences

JonChesterfield parent
The state machine with a random number generator is soundly beating some people in cognition already. That is, if the test for intelligence is set high enough that chatgpt doesn't pass it, nor do quite a lot of the human population.

If you can prove this can't happen, your axioms are wrong or your deduction in error.


moomin
I’m beginning to feel like the tests are part of the problem. Our intelligence tests are all tests of specialisation. We’ve established LLMs are part of the problem. Plenty of people who would fail a bar exam yet still know how many Rs there are in strawberry, could learn a new game just by reading the rules, know how to put up a set of shelves.
roenxi
I think the problem is that, as far as we can tell, AIs are just more generally intelligent than humans and people are trying to figure out how to assert that they are not. A specialist human in their area of competence can still outperform an AI, but there don't seem to be any fields now where a human novice can reliably out-think a computer.

We're seeing a lot more papers like this one where we have to define humans as non-general-intelligences.

moomin
I don't really buy this. It's apparently not possible to build an Estonian LLM with a satisfactory level of performance. Does that mean Estonians are general intelligences and English-speakers aren't? Or just that our ways of assessing intelligence aren't valid?
If you rarely got to see letters and just saw fragments of words as something like Chinese characters (tokens), could you count the R's in arbitrary words well?

The bigger issue is LLMs still need way way more data than humans get tons what they do. But they also have many less parameters than the human brain.

ben_w
> If you rarely got to see letters and just saw fragments of words as something like Chinese characters (tokens), could you count the R's in arbitrary words well?

While this seems correct, I'm sure I tried this when it was novel and observed that it could split the word into separate letters and then still count them wrong, which suggested something weird is happening internally.

I just now tried to repeat this, and it now counts the "r"'s in "strawberry" correctly (presumably enough examples of this specifically on the internet now?), but I did find it making the equivalent mistake with a German word (https://chatgpt.com/share/6859289d-f56c-8011-b253-eccd3cecee...):

  How many "n"'s are in "Brennnessel"?
But even then, having it spell the word out first, fixed it: https://chatgpt.com/share/685928bc-be58-8011-9a15-44886bb522...
kbelder
Counting letters is such a dull test. LLMs generally have a hard time with this question because letters are tokenized before they receive them, and they have to go through an involved reasoning process to figure it out. It's like asking a color blind person what color the street light is, and declaring him unintelligent because he sometimes gets the answer wrong.
moomin
I mean, if you don’t want to include tests that LLMs are, by definition, bad at, why don’t we do the same thing for humans?
"tons what they" autocorrected from "to do what they do."

"Paucity of the stimulus" is the term for what I'm talking about with the brain needing much less data, but beyond just more parameters we may have innate language processing that isn't there in other animals; Chomsky has been kind of relegated away now after LLMs but he may still have been right if it isn't just parameter count and or the innate thing different from animals isn't something like transformers. If you look at the modern language program in Chomsky's later years, it does have some remarkably similar things to transformers: permutation independent internal representation, and the merge operation being very similar to transformer's soft max. It's kind of describing something very like single head attention.

We know animals have rich innate neural abilities beyond just beating the heart and breathing etc.: a baby horse can be blind folded from birth, several days later blind fold taken off and it can immediately walk and navigate. Further development goes on, but other animals like cats have a visual system that doesn't seem to develop at all if it doesn't get natural stimulus in a critical early period. Something like that may apply to human language, it may be multiple systems missing from other apes and early hominids, but whatever it is we don't think it had many generations to evolve. Researchers have identified circuits in songbird brains that are also in humans but not apes, and something like that may be a piece of it for tracking sequences.

croes
Would you consider those who fail intelligent?

This item has no comments currently.