Count Lucanor wrote: ↑March 25th, 2023, 6:43 pmAn article in New York Times cites several researchers with a similar claim as the Microsoft researchers.value wrote: ↑March 25th, 2023, 5:46 amMicrosoft engineers are claiming in a recent paper that it's GPT-4 is already showing signs of AGI (Artificial General Intelligence).They are wrong. AI does not and cannot show signs of intelligence, it only simulates behavior that looks like intelligence, but it doesn't understand a thing it does, even though the technology is impressive. Look at the Chinese Room Experiment, which completely refuted such claims.
(2023) Microsoft Research Paper Claims Sparks of Artificial Intelligence in GPT-4
https://www.lesswrong.com/posts/FinfRNL ... artificial
(2023) Researchers Claim AI Chatbots Have Developed Theory of Mind
Some researchers claim that chatbots have developed theory of mind.
Michal Kosinski, a psychologist at the Stanford Graduate School of Business, made just that argument: that large language models like OpenAI’s ChatGPT and GPT-4 — next-word prediction machines trained on vast amounts of text from the internet — have developed theory of mind.
His studies have not been peer reviewed, but they prompted scrutiny and conversation among cognitive scientists, who have been trying to take the often asked question these days — Can ChatGPT do this? — and move AI into the realm of more robust scientific inquiry. What capacities do these models have, and how might they change our understanding of our own minds?
The Sally-Anne test, in which a girl, Anne, moves a marble from a basket to a box when another girl, Sally, isn’t looking. To know where Sally will look for the marble, researchers claimed, a viewer would have to exercise theory of mind, reasoning about Sally’s perceptual evidence and belief formation: Sally didn’t see Anne move the marble to the box, so she still believes it is where she last left it, in the basket.
Dr. Kosinski presented 10 large language models with 40 unique variations of these theory of mind tests — descriptions of situations like the Sally-Anne test, in which a person (Sally) forms a false belief. Then he asked the models questions about those situations, prodding them to see whether they would attribute false beliefs to the characters involved and accurately predict their behavior. He found that GPT-3.5, released in November 2022, did so 90 percent of the time, and GPT-4, released in March 2023, did so 95 percent of the time.
The conclusion? Machines have theory of mind.
Maarten Sap, a computer scientist at Carnegie Mellon University, fed more than 1,000 theory of mind tests into large language models and found that the most advanced transformers, like ChatGPT and GPT-4, passed about 70 percent of the time.
In general, Dr. Kosinski’s work and the responses to it fit into the debate about whether the capacities of these machines can be compared to the capacities of humans — a debate that divides researchers who work on natural language processing. Are these machines stochastic parrots, or alien intelligences, or fraudulent tricksters? A 2022 survey of the field found that, of the 480 researchers who responded, 51 percent believed that large language models could eventually “understand natural language in some nontrivial sense,” and 49 percent believed that they could not.
https://www.nytimes.com/2023/03/27/scie ... tbots.html