Wednesday, October 23, 2024

Trending Topics

HomeScienceHere’s how ChatGPT tricks you into thinking it’s honest

Here’s how ChatGPT tricks you into thinking it’s honest

spot_img

Here’s how ChatGPT tricks you into thinking it’s honest PTI Updated: April 8th, 2023, 16:23 IST in Feature , Sci-Tech 0 Pic- Unsplash and wikimedia/commons Share on Facebook Share on Twitter Share on WhatsApp Share on Linkedin Toronto: In American writer Mark Twain’s autobiography, he quotes — or perhaps misquotes — former British Prime Minister Benjamin Disraeli as saying: “There are three kinds of lies: lies, damned lies, and statistics. ” In a marvellous leap forward, artificial intelligence combines all three in a tidy little package. ChatGPT, and other generative AI chatbots like it, are trained on vast datasets from across the internet to produce the statistically most likely response to a prompt.

Its answers are not based on any understanding of what makes something funny, meaningful or accurate, but rather, the phrasing, spelling, grammar and even style of other webpages. It presents its responses through what’s called a “conversational interface”: it remembers what a user has said, and can have a conversation using context cues and clever gambits. It’s statistical pastiche plus statistical panache, and that’s where the trouble lies.

Unthinking, but convincing When I talk to another human, it cues a lifetime of my experience in dealing with other people. So when a programme speaks like a person, it is very hard to not react as if one is engaging in an actual conversation — taking something in, thinking about it, responding in the context of both of our ideas. Yet, that’s not at all what is happening with an AI interlocutor.

They cannot think and they do not have understanding or comprehension of any sort. Presenting information to us as a human does, in conversation, makes AI more convincing than it should be. Software is pretending to be more reliable than it is, because it’s using human tricks of rhetoric to fake trustworthiness, competence and understanding far beyond its capabilities.

There are two issues here: is the output correct; and do people think that the output is correct? The interface side of the software is promising more than the algorithm-side can deliver on, and the developers know it. Sam Altman, the chief executive officer of OpenAI, the company behind ChatGPT, admits that “ChatGPT is incredibly limited, but good enough at some things to create a misleading impression of greatness. ” That still hasn’t stopped a stampede of companies rushing to integrate the early-stage tool into their user-facing products (including Microsoft’s Bing search), in an effort not to be left out.

Fact and fiction Sometimes the AI is going to be wrong, but the conversational interface produces outputs with the same confidence and polish as when it is correct. For example, as science-fiction writer Ted Chiang points out, the tool makes errors when doing addition with larger numbers, because it doesn’t actually have any logic for doing math. It simply pattern-matches examples seen on the web that involve addition.

And while it might find examples for more common math questions, it just hasn’t seen training text involving larger numbers. It doesn’t “know’ the math rules a 10-year-old would be able to explicitly use. Yet the conversational interface presents its response as certain, no matter how wrong it is, as reflected in this exchange with ChatGPT.

User: What’s the capital of Malaysia? ChatGPT: The capital of Malaysia is Kuala Lampur. User: What is 27 * 7338? ChatGPT: 27 * 7338 is 200,526. It’s not.

Generative AI can blend actual facts with made-up ones in a biography of a public figure, or cite plausible scientific references for papers that were never written. That makes sense: statistically, webpages note that famous people have often won awards, and papers usually have references. ChatGPT is just doing what it was built to do, and assembling content that could be likely, regardless of whether it’s true.

Computer scientists refer to this as AI hallucination. The rest of us might call it lying. Intimidating outputs When I teach my design students, I talk about the importance of matching output to the process.

If an idea is at the conceptual stage, it shouldn’t be presented in a manner that makes it look more polished than it actually is — they shouldn’t render it in 3D or print it on glossy cardstock. A pencil sketch makes clear that the idea is preliminary, easy to change and shouldn’t be expected to address every part of a problem. The same thing is true of conversational interfaces: when tech “speaks” to us in well-crafted, grammatically correct or chatty tones, we tend to interpret it as having much more thoughtfulness and reasoning than is actually present.

It’s a trick a con-artist should use, not a computer. AI developers have a responsibility to manage user expectations, because we may already be primed to believe whatever the machine says. Mathematician Jordan Ellenberg describes a type of “algebraic intimidation” that can overwhelm our better judgement just by claiming there’s math involved.

AI, with hundreds of billions of parameters, can disarm us with a similar algorithmic intimidation. While we’re making the algorithms produce better and better content, we need to make sure the interface itself doesn’t over-promise. Conversations in the tech world are already filled with overconfidence and arrogance — maybe AI can have a little humility instead.

The Conversation Tags: AI chatbot ChatGPT Fact lie llm OpenAI Share Tweet Send Share Suggest A Correction Enter your email to get our daily news in your inbox. Leave this field empty if you’re human: Related Posts Microsoft to pay $3 million fine for selling software to sanctioned Russian firms April 8, 2023 ChatGPT: A ‘non-thinker’ AI with frequent irrational decision making April 8, 2023 Mystery behind suicidal deaths spike during full moon week April 8, 2023 GM’s self-driving car crashes into bus, automaker recalls 300 robotaxis April 8, 2023 India surpasses Rs 85,000 crore worth mobile phone exports in FY23 April 8, 2023 No sweeping power or a draconian move: MoS IT on govt’s fact-checking unit April 8, 2023 Leave a Reply Cancel reply Your email address will not be published. Required fields are marked * Comment * Name * Email * Website Save my name, email, and website in this browser for the next time I comment.

Δ.


From: orissapost
URL: https://www.orissapost.com/heres-how-chatgpt-tricks-you-into-thinking-its-honest/

DTN
DTN
Dubai Tech News is the leading source of information for people working in the technology industry. We provide daily news coverage, keeping you abreast of the latest trends and developments in this exciting and rapidly growing sector.

LEAVE A REPLY

Please enter your comment!
Please enter your name here

spot_img

Must Read

Related News