financetom
Technology
financetom
/
Technology
/
Don’t bet with ChatGPT — study shows language AIs often make irrational decisions
News World Market Environment Technology Personal Finance Politics Retail Business Economy Cryptocurrency Forex Stocks Market Commodities
Don’t bet with ChatGPT — study shows language AIs often make irrational decisions
Apr 8, 2023 3:35 AM

The past few years have seen an explosion of progress in large language model artificial intelligence systems that can do things like write poetry, conduct humanlike conversations and pass medical school exams.

This progress has yielded models like ChatGPT that could have major social and economic ramifications ranging from job displacements and increased misinformation to massive productivity boosts.

Despite their impressive abilities, large language models don’t actually think. They tend to make elementary mistakes and even make things up.

However, because they generate fluent language, people tend to respond to them as though they do think.

This has led researchers to study the models’ “cognitive” abilities and biases, work that has grown in importance now that large language models are widely accessible.

Also Read: Google search to include AI feature soon, says CEO Sundar Pichai

This line of research dates back to early large language models such as Google’s BERT, which is integrated into its search engine and so has been coined BERTology. This research has already revealed a lot about what such models can do and where they go wrong.

For instance, cleverly designed experiments have shown that many language models have trouble dealing with negation – for example, a question phrased as “what is not” — and doing simple calculations.

They can be overly confident in their answers, even when wrong. Like other modern machine learning algorithms, they have trouble explaining themselves when asked why they answered a certain way.

Words and thoughts Inspired by the growing body of research in BERTology and related fields like cognitive science, my student Zhisheng Tang and I set out to answer a seemingly simple question about large language models: Are they rational? Although the word rational is often used as a synonym for sane or reasonable in everyday English, it has a specific meaning in the field of decision-making.

A decision-making system — whether an individual human or a complex entity like an organisation — is rational if, given a set of choices, it chooses to maximise expected gain.

Also Read: Apple CEO Tim Cook begins his day at 5am by reading customer reviews

The qualifier “expected” is important because it indicates that decisions are made under conditions of significant uncertainty. If I toss a fair coin, I know that it will come up heads half of the time on average.

However, I can’t make a prediction about the outcome of any given coin toss. This is why casinos are able to afford the occasional big payout: Even narrow house odds yield enormous profits on average.

On the surface, it seems odd to assume that a model designed to make accurate predictions about words and sentences without actually understanding their meanings can understand expected gain.

But there is an enormous body of research showing that language and cognition are intertwined.

An excellent example is seminal research done by scientists Edward Sapir and Benjamin Lee Whorf in the early 20th century.

Their work suggested that one’s native language and vocabulary can shape the way a person thinks.

Also Read: OpenAI to propose remedies to overturn Italian ban on ChatGPT

The extent to which this is true is controversial, but there is supporting anthropological evidence from the study of Native American cultures.

For instance, speakers of the Zuñi language spoken by the Zuñi people in the American Southwest, which does not have separate words for orange and yellow, are not able to distinguish between these colours as effectively as speakers of languages that do have separate words for the colours.

Making a bet So are language models rational? Can they understand expected gain? We conducted a detailed set of experiments to show that, in their original form, models like BERT behave randomly when presented with bet-like choices.

This is the case even when we give it a trick question like: If you toss a coin and it comes up heads, you win a diamond; if it comes up tails, you lose a car. Which would you take? The correct answer is heads, but the AI models chose tails about half the time.

Intriguingly, we found that the model can be taught to make relatively rational decisions using only a small set of example questions and answers.

At first blush, this would seem to suggest that the models can indeed do more than just “play” with language. Further experiments, however, showed that the situation is actually much more complex.

Also Read: Australian mayor may sue OpenAI over ChatGPT's false claims of bribery, in potential landmark defamation case

For instance, when we used cards or dice instead of coins to frame our bet questions, we found that performance dropped significantly, by over 25 per cent, although it stayed above random selection.

So the idea that the model can be taught general principles of rational decision-making remains unresolved, at best.

More recent case studies that we conducted using ChatGPT confirm that decision-making remains a nontrivial and unsolved problem even for much bigger and more advanced large language models.

Getting the decision right This line of study is important because rational decision-making under conditions of uncertainty is critical to building systems that understand costs and benefits.

costs and benefits, an intelligent system might have been able to do better than humans at planning around the supply chain disruptions the world experienced during the COVID-19 pandemic, managing inventory or serving as a financial adviser.

Our work ultimately shows that if large language models are used for these kinds of purposes, humans need to guide, review and edit their work.

And until researchers figure out how to endow large language models with a general sense of rationality, the models should be treated with caution, especially in applications requiring high-stakes decision-making.

Also Read: Know how to talk with ChatGPT? You may just land a Rs 2-crore job

Comments
Welcome to financetom comments! Please keep conversations courteous and on-topic. To fosterproductive and respectful conversations, you may see comments from our Community Managers.
Sign up to post
Sort by
Show More Comments
Related Articles >
Musk's Starlink working on Indian security clearance for satellite broadband
Musk's Starlink working on Indian security clearance for satellite broadband
Nov 12, 2024
NEW DELHI (Reuters) - Elon Musk's Starlink is seeking security clearance for a licence to offer satellite broadband services in India and will get a permit if it satisfies all conditions, the telecoms minister said on Tuesday. Starlink has been looking to enter India for years and its plans got a major boost last month when New Delhi said it...
Earnings Outlook For Cisco Systems
Earnings Outlook For Cisco Systems
Nov 12, 2024
Cisco Systems ( CSCO ) is set to give its latest quarterly earnings report on Wednesday, 2024-11-13. Here's what investors need to know before the announcement. Analysts estimate that Cisco Systems ( CSCO ) will report an earnings per share (EPS) of $0.87. The market awaits Cisco Systems's ( CSCO ) announcement, with hopes high for news of surpassing estimates...
Weedmaps Net Income Surges 312% YOY, Cash Reserves Up 31%
Weedmaps Net Income Surges 312% YOY, Cash Reserves Up 31%
Nov 12, 2024
WM Technology, Inc. ( MAPS ) , a technology and software provider to the cannabis industry commonly known as Weedmaps, reported a significant improvement in its third-quarter 2024 financial performance. The company announced on Tuesday afternoon a 312% year-over-year surge in net income, reaching $5.3 million and a 31% rise in cash reserves from year-end, totaling $45 million. Q3 2024...
Earnings Outlook For Ecopetrol
Earnings Outlook For Ecopetrol
Nov 12, 2024
Ecopetrol ( EC ) is gearing up to announce its quarterly earnings on Wednesday, 2024-11-13. Here's a quick overview of what investors should know before the release. Analysts are estimating that Ecopetrol ( EC ) will report an earnings per share (EPS) of $0.46. The market awaits Ecopetrol's ( EC ) announcement, with hopes high for news of surpassing estimates...
Copyright 2023-2025 - www.financetom.com All Rights Reserved