Borg
Science Advisor
Gold Member
- 2,265
- 4,860
There are a lot of humans in this world that the same could be said of.nsaspook said:Maybe it's because these things don't really reason or understand anything.
Seriously though, we are still early in the evolution of these systems. Models and architectures that perform the best will be selected over the ones that don't. What we perceive as their 'reasoning' capabilities will continue to improve - just as people will continue to move the goalposts on what constitutes the ability to reason.
There is a lot of work that goes into designing these things. Building them is like trying to mimic the human brain which we don't fully understand. Now think about how the human brain works when presented with a question. How would you design a system that can recieve a question and then respond in a way that is similar to how a human brain would respond?
I am in the middle of building a relatively complex multi-agent system. In its simplest form of mimicking a human response, the system needs to accept a user question and answer it using previously learned or researched information. The process involves many smaller, specialized agents that are good at specific tasks like understanding dates, calculating numbers, web searches, etc. In many ways, the human brain operates in a similar manner with some areas that are good at recognizing faces, some that are good at math, some that are good at spatial problems, etc.
Once the information is gathered, there is typically a validation process with more agents. As noted in the article, when the system has the capability to search the internet, its accuracy can improve.
OpenAI’s GPT-4o with web search achieves 90% accuracy on SimpleQA, another one of OpenAI’s accuracy benchmarks.
Then, after gathering all of the information, models and humans alike have to decide which pieces of information are most relevant in answering the question - more agents in the AI case and specialized areas of the brain for humans.
Finally, if any of these processes generates bad information, there can be downstream failures in accuracy - this applies to models and humans alike. I personally see AI systems as evolving from a Savant Syndrome stage when they first arrived to now having far fewer social or intellectual impairments. Yes, at their core, they are still statistical engines but I don't see the human brain as being much different in its various components. Even with the best available information, people still make really bad judgements.