ChatGPT chatbot
What it is
ChatGPT (or 'Chat Generative Pre-trained Transformer') is an AI-powered chatbot that is capable of understanding natural human language and generating detailed human-like information and answers questions through a conversational interface.
Built on a version of Open AI's GPT-3 large language model, ChatGPT was released as a prototype for general public use in November 2022. Voice chat and image generation (via DALLE-3) capabilities were incorporated in September 2023.
Risks and harms
Notable incidents driven by and relating to ChatGPT recorded and categorised using AIAAIC's current risk and harms classifications and definitions include:
Accuracy, reliability
ChatGPT is prone to generate factual inaccuracies, including those known as 'hallucinations' (or plausible-sounding inaccurate information such as fake sources), and can be inconsistent and unreliable.
For example, the chatbot:
Wrongly claimed the death of privacy technologist Alexander Hanff.
Invented false legal case citations in a lawsuit involving Colombian airline Avianca.
Produced an error-strewn legal decision in the name of a Brazilian judge.
Generated false citations in a scientific paper about millipedes in the name of Danish academic Henrik Enghoff.
Overwhelmed coding Q&A site Stack Overflow moderators with low quality content.
Wrongly answered over half of software engineering questions it received, according to Purdue University researchers
Was found to generate significantly less accurate computer code over time.
The model's tendency to produce plausible-sounding factual falsities led Princeton University computer science professor Arvind Narayanan to call it 'the greatest bullshit generator ever.'
Anthropomorphism
ChatGPT's ability to generate plausible-looking, human-like information and advice runs the risk of convincing users that it is somehow human and/or 'sentient'.
A US professor falsely accused his whole class of using ChatGPT to write essays after believing the system when it told him it was responsible for their submissions.
Lawyer Steven Schwartz believed that six fake cases ChatGPT had cited as relevant to his legal case against Colombian airline Avianca were real.
Bias, discrimination
Despite its built-in safeguards, researchers and others have demonstrated that ChatGPT produces outputs that are explicitly and implicitly biased in multiple ways.
University of East Anglia researchers found that ChatGPT demonstrates 'significant' and 'systemic' left-wing bias across the political spectrum in Brazil, the UK, and USA.
ChatGPT was found to reproduce gender bias in job recommendation letters, according to University of California, Los Angeles researchers
GPT-3.5 and GPT-4, the large language models that power ChatGPT, produce responses that strongly reinforce and amplify gender bias, according to Princeton researchers.
Confidentiality
Unless they expressly opt-out, OpenAI uses users' information and data inputted to ChatGPT to train its models, thereby potentially or actually violating their personal confidentality, or the confidentiality of their employers.
Perth's South Metropolitan Health Service said doctors had been using ChatGPT to write medical notes, which were then being uploaded to patient record systems, thereby potentially compromising patient confidentiality.
Australian Research Council peer reviewers for were found to have been using ChatGPT to assess grant applications, thereby risking loss of confidentiality.
Samsung employees fed sensitive information about company meetings and source code into ChatGPT on several occasions, resulting in the company banning third-party generative AI services.
ChatGPT has been variously banned, blocked, or restricted by Apple, CitiGroup, Bank of America, Deutsche Bank, Goldman Sachs, Wells Fargo, and many other organisations.
Data privacy
Concerns exist about ChatGPT's approach to and impact on personal privacy.
ChatGPT was found to have allowed some users to see other users' conversations and personal information, suggesting OpenAI had access to their conversations.
ETH Zurich researchers showed that large language models such as GPT-4, which powers ChatGPT, can identify an individual's age, location, gender and income with up to 85 per cent accuracy simply by analysing their posts on Reddit.
ChatGPT can be used to identify individual internet users, according to Google DeepMind and other researchers.
Italy's privacy regulator banned ChatGPT for potentially infringing the privacy rights of Italians.
Poland's privacy regulator announced it was launching an investigation into OpenAI and ChatGPT following a complaint.
Canada's privacy regulator opened an investigation into OpenAI and ChatGPT on the basis of a complaint.
Germany, France, and Spain launched regulatory investigations into ChatGPT.
Japan's privacy watchdog issued issued a warning on the collection of user and third-party data.
US law firm Clarkson filed a lawsuit against OpenAI and Microsoft for 'stealing' personal information to create ChatGPT.
Software engineers sued OpenAI and Microsoft for illegally feeding their AI models with their personal information and professional expertise.
Employment
ChatGPT is seen to pose a risk to jobs across a swathe of industries, from law, finance, and marketing to academics, journalists, and software programmers.
There is already evidence of customer service agents, marketing copywriters, Kenyan student ghostwriters and Chinese video game illustrators being 'displaced' or laid off.
OpenAI's development of ChatGPT involved poorly paid workers in third-world countries with limited rights.
Time journalist Billy Perrigo revealed that OpenAI used Kenyan workers working for self-styled 'ethical AI' company Sama AI who were being paid less than USD 2 an hour to de-toxify ChatGPT.
Environment
OpenAI has not revealed the environmental impacts of its products, which are thought to be substantial.
University of California, Riverside researchers estimate that ChatGPT needs to 'drink' a 500ml bottle of water for a simple conversation of roughly 20-50 questions and answers.
IP and copyright
According to ChatGPT, the content generated by it may be protected by copyright, but it is not owned by the system or by OpenAI, which avoids including citations or attributions to the original sources of information trained and generated by ChatGPT.
2,200+ news and media publishers complained that OpenAI used their articles to train ChatGPT without informing or paying them.
Author Jane Friedman discovered 5 books likely to have been generated by ChatGPT being sold under her name on Amazon. Amazon initially refused to remove them, later changing course after she had taken her case public.
Authors Mona Awad and Paul Tremblay sued OpenAI for breaching copyright law by training ChatGPT on their novels without permission.
Comedienne and author Sarah Silverman and novelists Richard Kadrey and Christopher Golden sued OpenAI and Meta for USD 1 billion on the basis that the two companies used 'shadow libraries' to train AI chatbots in what they argued is 'unprecedented open-source software piracy'.
US Pulitzer Prize winner Michael Chabon and a group of other writers sued (pdf) OpenAI for copying their works without permission in order to teach ChatGPT how to respond to human prompts, arguing that the system can accurately summarize their works and generate text that mimics their styles.
A class-action suit accused OpenAI of engaging 'in a systematic course of mass-scale copyright infringement' brought by the Authors Guild and 17 authors, including Jonathan Franzen and John Grisham.
Author and journalist Julian Sancton sued OpenAI and Microsoft for copyright abuse.
Misinformation and disinformation
ChatGPT is capable of producing large volumes of convincing misinformation and disinformation. Examples include:
The chatbot was used to spread a false rumour that authorities in Hangzhou, China, would end alternate-day number-plate driving restrictions.
ChatGPT enabled a Chinese man to spread fake news about a fatal train accident, which ended in his arrest.
The system falsely claimed German geocoding company OpenCage offered an API to turn a mobile phone number into the location of the phone.
GPT-4 powered ChatGPT Plus echoed false news narratives 100 percent of the time, prompting accusation that it was a 'disinformation engine'.
ChatGPT was found to be more likely to produce disinformation in Chinese than English.
Allegedly powered Fox8, an autonomous botnet using fake personas to promote crypto/blockchain/NFT content.
Reputational damage
ChatGPT can harm the reputation of individuals and organisations it generates information about, in some instances resulting in threats to sue OpenAI for defamation.
An Australian mayor said he would sue OpenAI if it did not correct ChatGPT's false claims that he had served time in prison for bribery.
ChatGPT falsely accused law professor Jonathan Hurley of sexual harrassment, prompting Hurley to say that he had been 'defamed by ChatGPT'.
US radio host Mark Walters sued OpenAI after ChatGPT wrongly stated that Walters had been accused of defrauding and embezzling funds from a non-profit organisation.
The US Federal Trade Commission opened an investigation into whether OpenAI and its AI models, including ChatGPT, 'engaged in unfair or deceptive practices relating to risks of harm to consumers, including reputational harm.'
ChatGPT has also been shown to harm the reputation of individuals and organisations using it, including when used naively.
Safety
ChatGPT has safeguards to stop requests generating inappropriate, abusive, or offensive content. However, it appears many of these can be overridden relatively easily.
Vice journalist Steph Maj Swanson 'easily' manipulated ChatGPT to produce BDSM role-play scenarios and describe sex acts with children and animals.
Social wellbeing and cohesion
ChatGPT's ability to generate plausible-sounding misinformation, disinformation, and hate speech is seen to have potentially serious effects on the well-being of communities and democracy.
A Washington Post investigation discovered that ChatGPT can be used to develop political messaging and campaigns, despite OpenAI's rules to the contrary.
Security
ChatGPT has been shown to be vulnerable to malfunction and manipulation, and can easily be made to generate spam and malware. The chatbot:
Experienced a bug that jeopardised ChatGPT users' personal data, including chat histories and payment details.
Wrote malicious code that made databases leak sensitive information
Wrote 'plausible' phishing emails and malicious code.
Powered content farms spewing low-quality posts and spam in multiple languages.
Generated fake online reviews on Amazon and many other online stores and review sites.
Generated inaccurate and irrelevant bug reports for crypto bug bounty platform Immunefi.
Was tricked into producing malicious code, which could be used to launch cyber attacks.
Was manipulated into having its safety rules bypassed in 'virtually unlimited ways' using character suffixes.
Transparency, openness
OpenAI may have been open about some of ChatGPT's limitations, but its governance and algorithms are - and appear likely to remain - carefully concealed, commercially-driven black boxes.
Research studies show that ChatGPT is amongst the least open large language models, with little public information available about how it works, and with access to its data, model, and algorithm jealously guarded.
Italy's privacy regulator pointed out OpenAI appears to have been made deliberately difficult to find privacy-related information, make complaints, and request the removal of personal information.
ChatGPT's visibility has also led to accusations that OpenAI is highly opaque about many aspects of its corporate governance, including its funding, financial performance, supply chain management, and the amount of energy required to train and run its systems.
Databank
Operator: OpenAI
Developer: OpenAI
Country: USA; Global
Sector: Multiple
Purpose: Generate text
Technology: Chatbot; NLP/text analysis; Neural network; Deep learning; Machine learning; Reinforcement learning
Issue: Accuracy/reliability; Anthropomorphism; Bias/discrimination; Confidentiality; Copyright; Environment; Employment; Mis/disinformation; Privacy; Safety; Security
Transparency: Governance; Black box; Complaints/appeals; Marketing; Privacy
System
Lawsuits
Investigations, assessments, audits
Research, advocacy
Marcus G., Davis E., et al. ChatGPT/LLM Errors Tracker
EPIC (2023). Generative Harms (pdf)
Liesenfeld A., Lopez A. & Dingemanse M. (2023). Opening up ChatGPT: Tracking Openness, Transparency, and Accountability in Instruction-Tuned Text Generators
Bommasani R., Klyman K., Zhang D., Liang P. (2023). Do Foundation Model Providers Comply with the Draft EU AI Act?
Derner E., Batistic K. (2023). Beyond the Safeguards: Exploring the Security Risks of ChatGPT (pdf)
Shaping AI - University of Warwick (2023). Shifting AI controversies (pdf)
Brynjolfsson E., Li D., Raymond L.R. (2023). Generative AI at Work
Li P., Yang J., Islam M.A., Ren S. (2023). Making AI Less 'Thirsty': Uncovering and Addressing the Secret Water Footprint of AI Models (pdf)
Europol (2023). ChatGPT - the impact of Large Language Models on Law Enforcement
Newsguard (2023). Despite OpenAI’s Promises, the Company’s New AI Tool Produces Misinformation More Frequently, and More Persuasively, than its Predecessor
Newsguard (2023). The Next Great Misinformation Superspreader: How ChatGPT Could Spread Toxic Misinformation At Unprecedented Scale
Gao C.A., Howard F.M., Markov N.S., Dyer E.C., Ramesh S., Luo Y., Pearson A.T, (2022). Comparing scientific abstracts generated by ChatGPT to original abstracts using an artificial intelligence output detector, plagiarism detector, and blinded human reviewers
Page info
Type: System
Published: December 2022
Last updated: December 2023