SAN FRANCISCO: California start-up OpenAI has released a chatbot capable of answering a variety of questions, but its impressive performance has reopened the debate on the risks linked to artificial intelligence (AI) technologies.

The conversations with ChatGPT, posted on Twitter by fascinated users, show a kind of omniscient machine, capable of explaining scientific concepts and writing scenes for a play, university dissertations or even functional lines of computer code.

“Its answer to the question ‘what to do if someone has a heart attack’ was incredibly clear and relevant,” Claude de Loupy, head of Syllabs, a French company specialised in automatic text generation, said.

“When you start asking very specific questions, ChatGPT’s response can be off the mark,” but its overall performance remains “really impressive,” with a “high linguistic level,” he said.

OpenAI, cofounded in 2015 in San Francisco by billionaire tech mogul Elon Musk, who left the business in 2018, received $1 billion from Microsoft in 2019. The start-up is best known for its automated creation software: GPT-3 for text generation and DALL- E for image generation.

ChatGPT is able to ask its interlocutor for details, and has fewer strange responses than GPT-3, which, in spite of its prowess, sometimes spits out absurd results, said De Loupy.

Cicero

“A few years ago chatbots had the vocabulary of a dictionary and the memory of a goldfish,” said Sean McGregor, a researcher who runs a database of AI-related incidents.

“Chatbots are getting much better at the ‘history problem’ where they act in a manner consistent with the history of queries and responses. The chatbots have graduated from goldfish status.” Like other programmes relying on deep learning, mimicking neural activity, ChatGPT has one major weakness: “it does not have access to meaning,” says De Loupy.

The software cannot justify its choices, such as explain why its picked the words that make up its responses.

AI technologies able to communicate are, nevertheless, increasingly able to give an impression of thought.

Researchers at Facebook-parent Meta recently developed a computer programme dubbed Cicero, after the Roman statesman. The software has proven proficient at the board game Diplomacy, which requires negotiation skills.

“If it doesn’t talk like a real person — showing empathy, building relationships, and speaking knowledgeably about the game — it won’t find other players willing to work with it,” Meta said in research findings. In October, Character.ai, a start-up founded by former Google engineers, put an experimental chatbot online that can adopt any personality.

Users create characters based on a brief description and can then “chat” with a fake Sherlock Holmes, Socrates or Donald Trump.

‘Just a machine’

This level of sophistication both fascinates and worries some observers, who voice concern these technologies could be misused to trick people, by spreading false information or by creating increasingly credible scams.

What does ChatGPT think of these hazards? “There are potential dangers in building highly sophisticated chatbots, particularly if they are designed to be indistinguishable from humans in their language and behavior,” the chatbot said. Some businesses are putting safeguards in place to avoid abuse of their technologies.

On its welcome page, OpenAI lays out disclaimers, saying the chatbot “may occasionally generate incorrect information” or “produce harmful instructions or biased content.” And ChatGPT refuses to take sides.

Published in Dawn, December 5th, 2022

Opinion

Accessing the RSF

Accessing the RSF

RSF can help catalyse private sector inves­tment encouraging investment flows, build upon institutional partnerships with MDBs, other financial institutions.

Editorial

Madressah oversight
Updated 19 Dec, 2024

Madressah oversight

Bill should be reconsidered and Directorate General of Religious Education, formed to oversee seminaries, should not be rolled back.
Kurram’s misery
19 Dec, 2024

Kurram’s misery

THE unfolding humanitarian crisis in Kurram district, particularly in Parachinar city, has reached alarming...
Hiking gas rates
19 Dec, 2024

Hiking gas rates

IMPLEMENTATION of a new Ogra recommendation to increase the gas prices by an average 8.7pc or Rs142.45 per mmBtu in...
Geopolitical games
Updated 18 Dec, 2024

Geopolitical games

While Assad may be gone — and not many are mourning the end of his brutal rule — Syria’s future does not look promising.
Polio’s toll
18 Dec, 2024

Polio’s toll

MONDAY’s attacks on polio workers in Karak and Bannu that martyred Constable Irfanullah and wounded two ...
Development expenditure
18 Dec, 2024

Development expenditure

PAKISTAN’S infrastructure development woes are wide and deep. The country must annually spend at least 10pc of its...