A new test of AI capabilities consists of puzzles that humans are able to solve without too much trouble, but which all leading AI models struggle with. To improve and pass the test, AI companies will need to balance problem-solving abilities with cost.
Yes, you’re anthropomorphizing far too much. An LLM can’t understand, or recall (in the common sense of the word, i.e. have a memory), and is not aware.
Those are all things that intelligent, thinking things do. LLMs are none of that. They are a giant black box of math that predicts text. It doesn’t even understand what a word is, orthe meaning of anything it vomits out. All it knows is what is the statistically most likely text to come next, with a little randomization to add “creativity”.