Skip to content
Hindi News, हिंदी समाचार, Samachar, Breaking News, Latest Khabar – Pratirodh

Hindi News, हिंदी समाचार, Samachar, Breaking News, Latest Khabar – Pratirodh

Primary Menu Hindi News, हिंदी समाचार, Samachar, Breaking News, Latest Khabar – Pratirodh

Hindi News, हिंदी समाचार, Samachar, Breaking News, Latest Khabar – Pratirodh

  • Home
  • Newswires
  • Politics & Society
  • The New Feudals
  • World View
  • Arts And Aesthetics
  • For The Record
  • About Us
  • Featured

Why We Need To Kick The Habit Of Treating AI Like A Human

May 20, 2023 | Pratirodh Bureau

FILE PHOTO: A robot delivers a birthday cake at a restaurant as the coronavirus disease outbreak continues, in Milton Keynes, Britain in October 2020

The artificial intelligence (AI) pioneer Geoffrey Hinton recently resigned from Google, warning of the dangers of the technology “becoming more intelligent than us”. His fear is that AI will one day succeed in “manipulating people to do what it wants”.

There are reasons we should be concerned about AI. But we frequently treat or talk about AIs as if they are human. Stopping this, and realising what they actually are, could help us maintain a fruitful relationship with the technology.

In a recent essay, the US psychologist Gary Marcus advised us to stop treating AI models like people. By AI models, he means large language models (LLMs) like ChatGPT and Bard, which are now being used by millions of people on a daily basis.

He cites egregious examples of people “over-attributing” human-like cognitive capabilities to AI that have had a range of consequences. The most amusing was the US senator who claimed that ChatGPT “taught itself chemistry”. The most harrowing was the report of a young Belgian man who was said to have taken his own life after prolonged conversations with an AI chatbot.

Marcus is correct to say we should stop treating AI like people – conscious moral agents with interests, hopes and desires. However, many will find this difficult to near-impossible. This is because LLMs are designed – by people – to interact with us as though they are human, and we’re designed – by biological evolution – to interact with them likewise.

Good Mimics

The reason LLMs can mimic human conversation so convincingly stems from a profound insight by computing pioneer Alan Turing, who realised that it is not necessary for a computer to understand an algorithm in order to run it. This means that while ChatGPT can produce paragraphs filled with emotive language, it doesn’t understand any word in any sentence it generates.

The LLM designers successfully turned the problem of semantics – the arrangement of words to create meaning – into statistics, matching words based on their frequency of prior use. Turing’s insight echos Darwin’s theory of evolution, which explains how species adapt to their surroundings, becoming ever-more complex, without needing to understand a thing about their environment or themselves.

The cognitive scientist and philosopher Daniel Dennett coined the phrase “competence without comprehension”, which perfectly captures the insights of Darwin and Turing.

Another important contribution of Dennett’s is his “intentional stance”. This essentially states that in order to fully explain the behaviour of an object (human or non-human), we must treat it like a rational agent. This most often manifests in our tendency to anthropomorphise non-human species and other non-living entities.

But it is useful. For example, if we want to beat a computer at chess, the best strategy is to treat it as a rational agent that “wants” to beat us. We can explain that the reason why the computer castled, for instance, was because “it wanted to protect its king from our attack”, without any contradiction in terms.

We may speak of a tree in a forest as “wanting to grow” towards the light. But neither the tree, nor the chess computer represents those “wants” or reasons to themselves; only that the best way to explain their behaviour is by treating them as though they did.

Intentions And Agency

Our evolutionary history has furnished us with mechanisms that predispose us to find intentions and agency everywhere. In prehistory, these mechanisms helped our ancestors avoid predators and develop altruism towards their nearest kin. These mechanisms are the same ones that cause us to see faces in clouds and anthropomorphise inanimate objects. No harm comes to us when we mistake a tree for a bear, but plenty does the other way around.

Evolutionary psychology shows us how we are always trying to interpret any object that might be human as a human. We unconsciously adopt the intentional stance and attribute all our cognitive capacities and emotions to this object.

With the potential disruption that LLMs can cause, we must realise they are simply probabilistic machines with no intentions, or concerns for humans. We must be extra-vigilant around our use of language when describing the human-like feats of LLMs and AI more generally. Here are two examples.

The first was a recent study that found ChatGPT is more empathetic and gave “higher quality” responses to questions from patients compared with those of doctors. Using emotive words like “empathy” for an AI predisposes us to grant it the capabilities of thinking, reflecting and of genuine concern for others – which it doesn’t have.

The second was when GPT-4 (the latest version of ChatGPT technology) was launched last month, capabilities of greater skills in creativity and reasoning were ascribed to it. However, we are simply seeing a scaling up of “competence”, but still no “comprehension” (in the sense of Dennett) and definitely no intentions – just pattern matching.

Safe And Secure

In his recent comments, Hinton raised a near-term threat of “bad actors” using AI for subversion. We could easily envisage an unscrupulous regime or multinational deploying an AI, trained on fake news and falsehoods, to flood public discourse with misinformation and deep fakes. Fraudsters could also use an AI to prey on vulnerable people in financial scams.

Last month, Gary Marcus and others, including Elon Musk, signed an open letter calling for an immediate pause on the further development of LLMs. Marcus has also called for an international agency to promote safe, secure and peaceful AI technologies” – dubbing it a “Cern for AI”.

Furthermore, many have suggested that anything generated by an AI should carry a watermark so that there can be no doubt about whether we are interacting with a human or a chatbot.

Regulation in AI trails innovation, as it so often does in other fields of life. There are more problems than solutions, and the gap is likely to widen before it narrows. But in the meantime, repeating Dennett’s phrase “competence without comprehension” might be the best antidote to our innate compulsion to treat AI like humans.

(This article is republished from The Conversation under a Creative Commons license. Read the original article)

Tags: AI, Artificial Intelligence, human intelligence, Pratirodh, semantics

Continue Reading

Previous Is ‘2nd Demonetisation’ A Cover-Up Of Wrong Decision Made Earlier?
Next 4 Major River Basins In HP See Decline In Snow Cover: Study

More Stories

  • Featured

Women’s Resistance & Rebellion: What Greek Mythology Teaches Us

19 hours ago Pratirodh Bureau
  • Featured

Wrestlers Manhandled: ‘How Can You Sleep At Night?’ Lalan Asks PM Modi

19 hours ago Pratirodh Bureau
  • Featured

India Should Fence Cheetah Habitats, Worst Yet To Come: Expert

23 hours ago Pratirodh Bureau

Recent Posts

  • Women’s Resistance & Rebellion: What Greek Mythology Teaches Us
  • Wrestlers Manhandled: ‘How Can You Sleep At Night?’ Lalan Asks PM Modi
  • India Should Fence Cheetah Habitats, Worst Yet To Come: Expert
  • How Climate Change Worsens Avalanches In The Himalayas
  • “PM Treating Inauguration Of New Parliament Building As Coronation”
  • Book Review: A Deep Dive Into The Imbalances Of Climate Justice In India
  • DU Replaces Paper On Mahatma Gandhi With One On Savarkar
  • India Faces ‘Very Complicated Challenge’ From China: EAM
  • If We Are Smart About Water, We Can Stop Our Cities Sinking
  • Nine Years Of Modi Govt: Congress Poses Nine Questions To PM
  • Explainer: Why Has India Been Soft On Russia?
  • Bihar Heatwave Response Reveals Flaws In Our Heat Strategy
  • Death Of Six Cheetahs At Kuno: NTCA Sets Up ‘High-Power’ Committee
  • Wrestling With Untruths, Abuse & Scorn: When Will It End?
  • Heritage Pumps Used To Green Gardens At Taj Mahal Lost And Found
  • Two Billion People Will Struggle To Survive In A Warming World: Study
  • The Hidden Side Of Human-Elephant Conflicts: Orphaned Calves
  • Cong Says Modi Govt’s ‘Arrogance’ Has All But ‘Destroyed’ Parliamentary System
  • Debt-For-Nature Swaps Can Help Create A More Resilient South Asia
  • ‘Sengol’ Set To Be Installed In Parl Linked To TN

Search

Main Links

  • Home
  • Newswires
  • Politics & Society
  • The New Feudals
  • World View
  • Arts And Aesthetics
  • For The Record
  • About Us

Related Stroy

  • Featured

Women’s Resistance & Rebellion: What Greek Mythology Teaches Us

19 hours ago Pratirodh Bureau
  • Featured

Wrestlers Manhandled: ‘How Can You Sleep At Night?’ Lalan Asks PM Modi

19 hours ago Pratirodh Bureau
  • Featured

India Should Fence Cheetah Habitats, Worst Yet To Come: Expert

23 hours ago Pratirodh Bureau
  • Featured

How Climate Change Worsens Avalanches In The Himalayas

23 hours ago Pratirodh Bureau
  • Featured

“PM Treating Inauguration Of New Parliament Building As Coronation”

2 days ago Pratirodh Bureau

Recent Posts

  • Women’s Resistance & Rebellion: What Greek Mythology Teaches Us
  • Wrestlers Manhandled: ‘How Can You Sleep At Night?’ Lalan Asks PM Modi
  • India Should Fence Cheetah Habitats, Worst Yet To Come: Expert
  • How Climate Change Worsens Avalanches In The Himalayas
  • “PM Treating Inauguration Of New Parliament Building As Coronation”
Copyright © All rights reserved. | CoverNews by AF themes.