Announcements
No announcements yet
New blog posts
Completion of Workshop on Water Recling Simulation and Modelling: Unlocking the Future of Water Management
19 March, 2024 by Charlotte Lee
We are thrilled to announce the successful...
IJITIS Journal Meeting and SWOT Analysis at TULTECH
15 January, 2024 by Charlotte Lee
Greetings, TULTECH community! In our...
A Milestone Meeting for EIL: Shaping the Future of Environmental Industry Letters
15 December, 2023 by Charlotte Lee
Dear TULTECH Community, We are delighted to...
Weather
-2°C
Calendar of Events
Psychologists have shown that GPT-3 has the same level of reasoning ability as a college student
Posted on 4 August, 2023 by Charlotte Lee
Summary:
Standardised test-like logic problems were no match for the artificial intelligence language model GPT-3, which performed as well as college students. The experiment's authors argue that their findings raise the question of whether the technology is emulating human thinking or employing a novel cognitive mechanism. To get an answer, you'd need to go inside the code that powers GPT-3 and other AI programmes.
UCLA psychologists have shown that the AI language model GPT-3 does as well as college freshmen when presented with the types of reasoning difficulties normally seen in IQ testing and standardised exams like the SAT. Nature Human Behaviour has published the study.
However, the authors of the publication state that the research prompts the following question: Is GPT-3 employing a fundamentally different form of cognitive process, or is it just a result of its large language training dataset that makes it behave like a human brain?
Since OpenAI, the business that developed GPT-3, is protecting its secretive inner workings, the scientists at UCLA cannot definitively comment on the nature of GPT-3's reasoning skills. They also note that despite GPT-3's impressive performance in some areas of reasoning, the widely used AI tool still falls short in others.
"It's important to emphasise that this system has major limitations," said Taylor Webb, the study's first author and a postdoctoral researcher in psychology at UCLA.
Forty first-year students at UCLA were given the identical issues to tackle by the researchers.
According to the study's principal author and UCLA psychology professor Hongjing Lu: "Surprisingly, not only did GPT-3 do about as well as humans but it made similar mistakes as well."
GPT-3 was successful at solving 80% of the questions, which is above the average score of slightly around 60% for human participants and within the range of the top human scores.
The only way to find out is to gain access to the programme and the data used to train the software, and then to give the software tests that it hasn't already been given, which is a daunting task. They claimed it would be the next stage in determining the proper direction for AI.
Webb said that "having the backend to GPT models would be very useful for AI and cognitive researchers." To paraphrase, "We're just doing inputs and getting outputs, and it's not as decisive as we'd like it to be."
source: sciencedaily.com/releases/2023/07/230731110750.htm
Event Categories
Past Events
Workshop on Artificial Intelligence Applications in Smart Cities
20 August, 2024Workshop on Advanced Water Treatment Processes
10 July, 2024Workshop on Water Recycling Simulation and Modelling
15 March, 2024Today In History
Here are some interesting facts ih history happened on 23 November.
- just past midnight a sharp jolt causes Lake Merced to drop 30 feet
- Patent granted for a process of making color photographs
- Post Hospital at Presidio SF renamed Letterman General Hospital
- Lens to provide zoom effects patented - FG Back
- Merle Oberon dies at 68
- Mae West dies at 88