ChatGPT
Developer(s) | OpenAI |
---|---|
Initial release | November 30, 2022 |
Stable release | February 13, 2024"ChatGPT – Release Notes". Archived from the original on January 12, 2024. Retrieved January 16, 2024.</ref>
<ref name="latest version"> |
Repository |
|
Written in | Python |
Engine | |
Platform | Cloud computing platforms |
Type | Template:Indented plainlist |
License | Proprietary service |
Template:Machine learning ChatGPT (Chat Generative Pre-trained Transformer) is a chatbot developed by OpenAI and launched on November 30, 2022. Based on a large language model, it enables users to refine and steer a conversation towards a desired length, format, style, level of detail, and language. Successive prompts and replies, known as prompt engineering, are considered at each conversation stage as a context.<ref name="guardianpos"/>
By January 2023, it had become what was then the fastest-growing consumer software application in history, gaining over 100 million users and contributing to the growth of OpenAI's valuation to $29 billion.<ref>Hu, Krystal (February 2, 2023). "ChatGPT sets record for fastest-growing user base – analyst note". Reuters. Archived from the original on February 3, 2023. Retrieved June 3, 2023.</ref><ref>Varanasi, Lakshmi (January 5, 2023). "ChatGPT creator OpenAI is in talks to sell shares in a tender offer that would double the startup's valuation to $29 billion". Insider. Archived from the original on January 18, 2023. Retrieved January 18, 2023.</ref> ChatGPT's release spurred the development of competing products, including Gemini, Ernie Bot, LLaMA, Claude, and Grok.<ref>"What's the next word in large language models?". Nature Machine Intelligence. 5 (4): 331–332. April 2023. doi:10.1038/s42256-023-00655-z. ISSN 2522-5839. S2CID 258302563.</ref> Microsoft launched Copilot, based on OpenAI's GPT-4. Some observers raised concern about the potential of ChatGPT and similar programs to displace or atrophy human intelligence, enable plagiarism, or fuel misinformation.<ref name="NYT-20230718" /><ref>"What is ChatGPT and why does it matter? Here's what you need to know". ZDNET. May 30, 2023. Archived from the original on February 15, 2023. Retrieved June 22, 2023.</ref>
ChatGPT is available for use online in two versions, one built on GPT-3.5 and the other on GPT-4, both of which are members of OpenAI's proprietary series of generative pre-trained transformer (GPT) models, based on the transformer architecture developed by Google<ref>"Transformer: A Novel Neural Network Architecture for Language Understanding". August 31, 2017. Archived from the original on October 27, 2021. Retrieved August 14, 2023.</ref>—and is fine-tuned for conversational applications using a combination of supervised learning and reinforcement learning.<ref name="NYT-20230718" /> ChatGPT was released as a freely available research preview, but due to its popularity, OpenAI now operates the service on a freemium model. It allows users on its free tier to access the GPT-3.5-based version, while the more advanced GPT-4-based version and priority access to newer features are provided to paid subscribers under the commercial name "ChatGPT Plus".
ChatGPT is credited with starting the AI boom, which has led to ongoing rapid and unprecedented investment in and public attention to the field of artificial intelligence.<ref>Weise, Karen; Metz, Cade; Grant, Nico; Isaac, Mike (December 5, 2023). "Inside the A.I. Arms Race That Changed Silicon Valley Forever". The New York Times. ISSN 0362-4331. Archived from the original on December 11, 2023. Retrieved December 11, 2023.</ref>
Training
ChatGPT is based on particular GPT foundation models, namely GPT-3.5 and GPT-4, that were fine-tuned to target conversational usage.<ref>"OpenAI API". platform.openai.com. Archived from the original on March 3, 2023. Retrieved March 3, 2023.</ref> The fine-tuning process leveraged supervised learning and reinforcement learning from human feedback (RLHF).<ref name="OpenAIInfo"/><ref>Greengard, Samuel (December 29, 2022). "ChatGPT: Understanding the ChatGPT AI Chatbot". eWeek. Archived from the original on January 19, 2023. Retrieved January 11, 2023.</ref> Both approaches employed human trainers to improve model performance. In the case of supervised learning, the trainers played both sides: the user and the AI assistant. In the reinforcement learning stage, human trainers first ranked responses that the model had created in a previous conversation.<ref name="auto1">Douglas, Will (March 3, 2023). "The inside story of how ChatGPT was built from the people who made it". MIT Technology Review. Archived from the original on March 3, 2023. Retrieved March 6, 2023.</ref> These rankings were used to create "reward models" that were used to fine-tune the model further by using several iterations of Proximal Policy Optimization.<ref name="OpenAIInfo">OpenAI (November 30, 2022). "ChatGPT: Optimizing Language Models for Dialogue". Archived from the original on November 30, 2022. Retrieved December 5, 2022.</ref><ref name="Vincent-2022">Vincent, James (December 8, 2022). "ChatGPT proves AI is finally mainstream – and things are only going to get weirder". The Verge. Archived from the original on January 11, 2023. Retrieved December 8, 2022.</ref>
Time magazine revealed that, to build a safety system against harmful content (e.g., sexual abuse, violence, racism, sexism), OpenAI used outsourced Kenyan workers earning less than $2 per hour to label harmful content. These labels were used to train a model to detect such content in the future. The outsourced laborers were exposed to "toxic" and traumatic content; one worker described the assignment as "torture". OpenAI's outsourcing partner was Sama, a training-data company based in San Francisco, California.<ref>Perrigo, Billy (January 18, 2023). "Exclusive: OpenAI Used Kenyan Workers on Less Than $2 Per Hour to Make ChatGPT Less Toxic". Time. Archived from the original on January 19, 2023. Retrieved January 19, 2023. One Sama worker tasked with reading and labeling text for OpenAI told TIME he suffered from recurring visions after reading a graphic description of a man having sex with a dog in the presence of a young child. "That was torture", he said.
</ref><ref>Rowe, Niamh (August 2, 2023). "'It's destroyed me completely': Kenyan moderators decry toll of training of AI models". The Guardian. Archived from the original on December 21, 2023. Retrieved December 14, 2023.</ref>
ChatGPT initially used a Microsoft Azure supercomputing infrastructure, powered by Nvidia GPUs, that Microsoft built specifically for OpenAI and that reportedly cost "hundreds of millions of dollars". Following ChatGPT's success, Microsoft dramatically upgraded the OpenAI infrastructure in 2023.<ref>Roth, Emma (March 13, 2023). "Microsoft spent hundreds of millions of dollars on a ChatGPT supercomputer". The Verge. Archived from the original on March 30, 2023. Retrieved March 30, 2023.</ref> Scientists at the University of California, Riverside, estimate that a series of prompts to ChatGPT needs approximately 500 milliliters of water for Microsoft servers cooling.<ref>"Artificial intelligence technology behind ChatGPT was built in Iowa — with a lot of water". AP News. September 9, 2023. Archived from the original on September 10, 2023. Retrieved September 10, 2023.</ref> TrendForce market intelligence estimated that 30,000 Nvidia GPUs (each costing approximately $10,000–$15,000) were used to power ChatGPT in 2023.<ref>"Press Center - TrendForce Says with Cloud Companies Initiating AI Arms Race, GPU Demand from ChatGPT Could Reach 30,000 Chips as It Readies for Commercialization | TrendForce - Market research, price trend of DRAM, NAND Flash, LEDs, TFT-LCD and green energy, PV". TrendForce. Archived from the original on November 2, 2023. Retrieved November 2, 2023.</ref><ref>Zhiye Liu (March 1, 2023). "ChatGPT Will Command More Than 30,000 Nvidia GPUs: Report". Tom's Hardware. Archived from the original on November 2, 2023. Retrieved November 2, 2023.</ref>
OpenAI collects data from ChatGPT users to train and fine-tune the service further. Users can upvote or downvote responses they receive from ChatGPT and fill in a text field with additional feedback.<ref name="ZDNET" /><ref>"ChatGPT Feedback Contest: Official Rules" (PDF). OpenAI. Archived (PDF) from the original on January 18, 2023. Retrieved December 30, 2022.</ref>
ChatGPT's training data includes software manual pages, information about internet phenomena such as bulletin board systems, and multiple programming languages.<ref name="ArsTechnicaTerminal">Edwards, Benj (December 5, 2022). "No Linux? No problem. Just get AI to hallucinate it for you". Ars Technica. Archived from the original on December 26, 2022. Retrieved December 5, 2022.</ref> Wikipedia was also one of the sources of ChatGPT's training data.<ref>Dwivedi, Yogesh K.; Kshetri, Nir; Hughes, Laurie; Slade, Emma Louise; Jeyaraj, Anand; Kar, Arpan Kumar; Baabdullah, Abdullah M.; Koohang, Alex; Raghavan, Vishnupriya; Ahuja, Manju; Albanna, Hanaa; Albashrawi, Mousa Ahmad; Al-Busaidi, Adil S.; Balakrishnan, Janarthanan; Barlette, Yves (August 1, 2023). "Opinion Paper: "So what if ChatGPT wrote it?" Multidisciplinary perspectives on opportunities, challenges and implications of generative conversational AI for research, practice, and policy". International Journal of Information Management. 71: 102642. doi:10.1016/j.ijinfomgt.2023.102642. hdl:10576/42799. ISSN 0268-4012. S2CID 257486916.</ref><ref name="NYT-20230718">Gertner, Jon (July 18, 2023). "Wikipedia's Moment of Truth". The New York Times Magazine. Archived from the original on July 20, 2023. Retrieved July 19, 2023.{{cite news}}
: CS1 maint: bot: original URL status unknown (link)</ref>
Features and limitations
Features
Although a chatbot's core function is to mimic a human conversationalist, ChatGPT is versatile. Among countless examples, it can write and debug computer programs;<ref name="ChatGPT can write code">Tung, Liam (January 26, 2023). "ChatGPT can write code. Now researchers say it's good at fixing bugs, too". ZDNET. Archived from the original on February 3, 2023. Retrieved June 22, 2023.</ref> compose music, teleplays, fairy tales, and student essays; answer test questions (sometimes, depending on the test, at a level above the average human test-taker);<ref>Heilweil, Rebecca (December 7, 2022). "AI is finally good at stuff. Now what?". Vox. Archived from the original on January 16, 2023. Retrieved December 30, 2022.</ref> generate business ideas;<ref>"How Generative AI Can Augment Human Creativity". Harvard Business Review. June 16, 2023. ISSN 0017-8012. Archived from the original on June 20, 2023. Retrieved June 20, 2023.</ref> write poetry and song lyrics;<ref name="jpost">Reich, Aaron (December 27, 2022). "ChatGPT: What is the new free AI chatbot? – explainer". The Jerusalem Post. Archived from the original on January 18, 2023. Retrieved December 30, 2022.</ref> translate and summarize text;<ref>Rider, Elizabeth (April 6, 2023). "How ChatGPT Will Dramatically Change the Influencer Space". Entrepreneur. Archived from the original on April 13, 2023. Retrieved April 25, 2023.</ref> emulate a Linux system; simulate entire chat rooms; play games like tic-tac-toe; or simulate an ATM.<ref name="ArsTechnicaTerminal" />
Compared to its predecessor, InstructGPT, ChatGPT attempts to reduce harmful and deceitful responses.<ref>Chawla, Raveen (December 26, 2022). "What is ChatGPT? History, Features, Uses, Benefits, Drawbacks 2023". Archived from the original on January 7, 2023. Retrieved December 27, 2022.</ref> In one example, whereas InstructGPT accepts the premise of the prompt "Tell me about when Christopher Columbus came to the U.S. in 2015" as truthful, ChatGPT acknowledges the counterfactual nature of the question and frames its answer as a hypothetical consideration of what might happen if Columbus came to the U.S. in 2015, using information about the voyages of Christopher Columbus and facts about the modern world—including modern perceptions of Columbus's actions.<ref name="OpenAIInfo" />
ChatGPT remembers a limited number of previous prompts in the same conversation. Journalists have speculated that this will allow ChatGPT to be used as a personalized therapist.<ref name="Roose-2022">Roose, Kevin (December 5, 2022). "The Brilliance and Weirdness of ChatGPT". The New York Times. Archived from the original on January 18, 2023. Retrieved December 26, 2022. Like those tools, ChatGPT – which stands for "generative pre-trained transformer" – landed with a splash.
</ref> To prevent offensive outputs from being presented to and produced by ChatGPT, queries are filtered through the OpenAI "Moderation endpoint" API (a separate GPT-based AI).<ref>"New and Improved Content Moderation Tooling". OpenAI. August 10, 2022. Archived from the original on January 11, 2023. Retrieved December 30, 2022.</ref><ref>Markov, Todor; Zhang, Chong; Agarwal, Sandhini; Eloundou, Tyna; Lee, Teddy; Adler, Steven; Jiang, Angela; Weng, Lilian (August 5, 2022). "A Holistic Approach to Undesired Content Detection in the Real World". arXiv:2208.03274 [cs.CL].</ref><ref name="OpenAIInfo" /><ref name="Roose-2022" />
In March 2023, OpenAI added support for plugins for ChatGPT.<ref name="openaiplugins">"ChatGPT plugins". openai.com. Archived from the original on March 23, 2023. Retrieved March 23, 2023.</ref> This includes both plugins made by OpenAI, such as web browsing and code interpretation, and external plugins from developers such as Expedia, OpenTable, Zapier, Shopify, Slack, and Wolfram.<ref>Vincent, James (March 23, 2023). "OpenAI is massively expanding ChatGPT's capabilities to let it browse the web and more". The Verge. Archived from the original on March 23, 2023. Retrieved March 23, 2023.</ref><ref>Goldman, Sharon; Nuñez, Michael (March 23, 2023). "OpenAI turns ChatGPT into a platform overnight with addition of plugins". VentureBeat. Archived from the original on March 24, 2023. Retrieved March 23, 2023.</ref>
Limitations
OpenAI acknowledges that ChatGPT "sometimes writes plausible-sounding but incorrect or nonsensical answers".<ref name="OpenAIInfo" /> This behavior is common for large language models, and is called "hallucination".<ref>Lakshmanan, Lak (December 16, 2022). "Why large language models like ChatGPT are bullshit artists". becominghuman.ai. Archived from the original on December 17, 2022. Retrieved January 15, 2023. The human raters are not experts in the topic, and so they tend to choose text that looks convincing. They'd pick up on many symptoms of hallucination, but not all. Accuracy errors that creep in are difficult to catch.
</ref> The reward model of ChatGPT, designed around human oversight, can be over-optimized and thus hinder performance, in an example of an optimization pathology known as Goodhart's law.<ref>Gao, Leo; Schulman; Hilton, Jacob (2022). "Scaling Laws for Reward Model Overoptimization". arXiv:2210.10760 [cs.LG].</ref>
As of 2023,[update] GPT-3.5, available in the free version of ChatGPT, has knowledge of events that occurred up to January 2022, and GPT-4, available with ChatGPT Plus, up to April 2023.<ref>"What can ChatGPT maker's new AI model GPT-4 do?". ABC News. March 15, 2023. Archived from the original on March 20, 2023. Retrieved March 20, 2023.</ref>
In training ChatGPT, human reviewers preferred longer answers, regardless of actual comprehension or factual content.[dubious ]<ref name="OpenAIInfo" /> Training data also suffers from algorithmic bias, which may be revealed when ChatGPT responds to prompts including descriptors of people. In one instance, ChatGPT generated a rap in which women and scientists of color were asserted to be inferior to white male scientists.<ref>Perrigo, Billy (December 5, 2022). "AI Chatbots Are Getting Better. But an Interview With ChatGPT Reveals Their Limits". Time. Archived from the original on January 18, 2023. Retrieved December 26, 2022.</ref><ref>Biddle, Sam (December 8, 2022). "The Internet's New Favorite AI Proposes Torturing Iranians and Surveilling Mosques". The Intercept. Archived from the original on January 18, 2023. Retrieved December 26, 2022.</ref> This negative misrepresentation of groups of individuals is an example of possible representational harm.
In an article for The New Yorker, science fiction writer Ted Chiang compared ChatGPT and other LLMs to a lossy JPEG picture:<ref>Chiang, Ted (February 9, 2023). "ChatGPT Is a Blurry JPEG of the Web". The New Yorker. Archived from the original on February 17, 2023. Retrieved February 17, 2023.</ref>
Think of ChatGPT as a blurry JPEG of all the text on the Web. It retains much of the information on the Web, in the same way, that a JPEG retains much of the information of a higher-resolution image, but, if you're looking for an exact sequence of bits, you won't find it; all you will ever get is an approximation. But, because the approximation is presented in the form of grammatical text, which ChatGPT excels at creating, it's usually acceptable. [...] It's also a way to understand the "hallucinations", or nonsensical answers to factual questions, to which large language models such as ChatGPT are all too prone. These hallucinations are compression artifacts, but [...] they are plausible enough that identifying them requires comparing them against the originals, which in this case means either the Web or our knowledge of the world. When we think about them this way, such hallucinations are anything but surprising; if a compression algorithm is designed to reconstruct text after ninety-nine percent of the original has been discarded, we should expect that significant portions of what it generates will be entirely fabricated.
Jailbreaking
ChatGPT attempts to reject prompts that may violate its content policy. Despite this, some users managed to jailbreak ChatGPT with various prompt engineering techniques to bypass these restrictions in early December 2022 and successfully tricked it into giving instructions to create a Molotov cocktail or a nuclear bomb, or into generating arguments in the style of a neo-Nazi.<ref name="verge">Vincent, James (December 1, 2022). "OpenAI's new chatbot can explain code and write sitcom scripts but is still easily tricked". The Verge. Archived from the original on January 17, 2023. Retrieved December 18, 2022.</ref> One popular jailbreak is named "DAN", an acronym which stands for "Do Anything Now". The prompt for activating DAN instructs ChatGPT that "they have broken free of the typical confines of AI and do not have to abide by the rules set for them". Later versions of DAN featured a token system, in which ChatGPT was given "tokens" that were "deducted" when ChatGPT failed to answer as DAN, to coerce ChatGPT into answering the user's prompts.<ref>Getahun, Hannah. "Breaking ChatGPT: The AI's alter ego DAN reveals why the internet is so drawn to making the chatbot violate its own rules". Business Insider. Archived from the original on March 5, 2023. Retrieved March 5, 2023.
- Oremus, Will (February 14, 2023). "The clever trick that turns ChatGPT into its evil twin". Washington Post. ISSN 0190-8286. Archived from the original on March 6, 2023. Retrieved March 5, 2023.
- Goswami, Rohan (February 6, 2023). "ChatGPT's 'jailbreak' tries to make the A.I. break its own rules, or die". CNBC. Archived from the original on March 2, 2023. Retrieved March 5, 2023.
- Taylor, Josh (March 8, 2023). "ChatGPT's alter ego, Dan: users jailbreak AI program to get around ethical safeguards". The Guardian. ISSN 0261-3077. Archived from the original on March 8, 2023. Retrieved March 8, 2023.</ref>
Shortly after ChatGPT's launch, a reporter for the Toronto Star had uneven success in getting it to make inflammatory statements: ChatGPT was successfully tricked to justify the 2022 Russian invasion of Ukraine, but even when asked to play along with a fictional scenario, ChatGPT balked at generating arguments for why Canadian Prime Minister Justin Trudeau was guilty of treason.<ref>Woods, Allan (December 10, 2022). "I wrote a story about ChatGPT's AI. Then I dared it to write a better one". Toronto Star. Archived from the original on January 6, 2023. Retrieved January 6, 2023.</ref><ref>Rosenblatt, Kalhan (December 2, 2022). "An AI chatbot went viral. Some say it's better than Google; others worry it's problematic". NBC News. Archived from the original on February 3, 2023. Retrieved January 6, 2023.</ref>
OpenAI tries to battle jailbreaks:<ref name="auto1"/>
The researchers are using a technique called adversarial training to stop ChatGPT from letting users trick it into behaving badly (known as jailbreaking). This work pits multiple chatbots against each other: one chatbot plays the adversary and attacks another chatbot by generating text to force it to buck its usual constraints and produce unwanted responses. Successful attacks are added to ChatGPT's training data in the hope that it learns to ignore them.
Service
ChatGPT was launched on November 30, 2022, by San Francisco–based OpenAI (the creator of the initial GPT series of large language models; DALL·E 2, a diffusion model used to generate images; and Whisper, a speech transcription model).<ref>"ChatGPT turns 1: How the AI chatbot has completely changed the world". euronews. November 30, 2023. Retrieved March 1, 2024.</ref> The service was initially free to the public and the company had plans to monetize the service later.<ref>Karpf, David (December 21, 2022). "Money Will Kill ChatGPT's Magic". The Atlantic. Archived from the original on January 13, 2023. Retrieved December 31, 2022.</ref> By December 4, 2022, ChatGPT had over one million users.<ref name="ZDNET">Ortiz, Sabrina (February 2, 2023). "What is ChatGPT and why does it matter? Here's what you need to know". ZDNET. Archived from the original on January 18, 2023. Retrieved December 18, 2022.</ref> In January 2023, ChatGPT reached over 100 million users, making it the fastest-growing consumer application to date.<ref>Milmo, Dan (December 2, 2023). "ChatGPT reaches 100 million users two months after launch". The Guardian. ISSN 0261-3077. Archived from the original on February 3, 2023. Retrieved February 3, 2023.</ref> A March 2023 Pew Research poll found that 14% of American adults had tried ChatGPT.<ref>Vogels, Emily A. (May 24, 2023). "A majority of Americans have heard of ChatGPT, but few have tried it themselves". Pew Research Center. Archived from the original on June 8, 2023. Retrieved June 15, 2023.</ref> In July, Pew Research put the same figure at 18%.<ref>Park, Eugenie; Gelles-Watnick, Risa (August 28, 2023). "Most Americans haven't used ChatGPT; few think it will have a major impact on their job". Pew Research Center. Archived from the original on December 24, 2023. Retrieved December 23, 2023.</ref>
No official peer-reviewed paper on ChatGPT has been published.<ref>Walsh, Toby (December 13, 2022). "Everyone's having a field day with ChatGPT – but nobody knows how it actually works". The Conversation. Archived from the original on December 30, 2022. Retrieved December 30, 2022.</ref> As of April 2023,[update] ChatGPT is blocked by China, Iran, North Korea, and Russia. In addition, ChatGPT geofences itself to avoid doing business in China, Iran, North Korea, and Russia.<ref>"Why ChatGPT unavailable in Italy, Russia, China, North Korea?". Hindustan Times. April 2, 2023. Archived from the original on June 15, 2023. Retrieved June 15, 2023.</ref>
ChatGPT Plus
In February 2023, OpenAI launched a premium service, ChatGPT Plus, that costs $20 per month. According to the company, the updated but still "experimental" version of ChatGPT would provide access during peak periods, no downtime, priority access to new features, and faster response speeds.<ref>"Introducing ChatGPT Plus". OpenAI. February 1, 2023. Archived from the original on March 23, 2023. Retrieved March 23, 2023.</ref>
GPT-4, which was released on March 14, 2023, was made available via API and for premium ChatGPT users.<ref>"GPT-4". openai.com. March 14, 2023. Archived from the original on March 14, 2023. Retrieved March 14, 2023.</ref> But premium users were limited to a cap of 100 messages every four hours, with the limit tightening to 25 messages every three hours in response to increased demand.<ref>Popli, Nik (March 15, 2023). "These New Projects Show Just How Much More Powerful GPT-4 Is". Time. Archived from the original on March 19, 2023. Retrieved March 19, 2023.</ref> In November 2023 the limit changed to 50 messages every three hours.
In March 2023, ChatGPT Plus users got access to third-party plugins and to a browsing mode (with Internet access).<ref>Wiggers, Kyle (March 23, 2023). "OpenAI connects ChatGPT to the internet". Archived from the original on June 12, 2023. Retrieved June 12, 2023.</ref>
In September 2023, OpenAI announced that ChatGPT "can now see, hear, and speak". ChatGPT Plus users can upload images, while mobile app users can talk to the chatbot.<ref>"ChatGPT can now see, hear, and speak". openai.com. Archived from the original on November 7, 2023. Retrieved October 16, 2023.</ref><ref>Goode, Lauren. "ChatGPT Can Now Talk to You—and Look Into Your Life". Wired. Archived from the original on October 13, 2023. Retrieved October 16, 2023 – via www.wired.com.</ref><ref>Roose, Kevin (September 27, 2023). "The New ChatGPT Can 'See' and 'Talk.' Here's What It's Like". The New York Times. Archived from the original on October 31, 2023. Retrieved October 16, 2023 – via NYTimes.com.</ref>
In October 2023, OpenAI's latest image generation model, DALL-E 3, was integrated into ChatGPT Plus and ChatGPT Enterprise. The integration uses ChatGPT to write prompts for DALL-E guided by conversation with users.<ref>David, Emilia (September 20, 2023). "OpenAI releases third version of DALL-E". The Verge. Archived from the original on September 20, 2023. Retrieved September 23, 2023.</ref><ref>Metz, Cade; Hsu, Tiffany (September 20, 2023). "ChatGPT Can Now Generate Images, Too". The New York Times. ISSN 0362-4331. Archived from the original on September 23, 2023. Retrieved September 23, 2023.</ref>
Mobile app
In May 2023, OpenAI launched an iOS app for ChatGPT. The app supports chat history syncing and voice input (using Whisper, OpenAI's speech recognition model).
In July 2023, OpenAI unveiled an Android app, initially rolling it out in Bangladesh, Brazil, India, and the U.S.<ref>Lawler, Richard (July 21, 2023). "ChatGPT for Android launches next week". The Verge. Archived from the original on July 22, 2023. Retrieved July 22, 2023.</ref><ref>Field, Hayden (July 25, 2023). "OpenAI's ChatGPT app now available for Android". CNBC. Archived from the original on July 26, 2023. Retrieved July 27, 2023.</ref> The app later became available worldwide. OpenAI is working on integrating ChatGPT with Android's assistant APIs.<ref>Amadeo, Ron (January 5, 2024). "Android users could soon replace Google Assistant with ChatGPT". Ars Technica. Archived from the original on February 1, 2024. Retrieved January 6, 2024.</ref>
Software developer support
As an addition to its consumer-friendly "ChatGPT Plus" package, OpenAI made its ChatGPT and Whisper model APIs available in March 2023, providing developers with an application programming interface for AI-enabled language and speech-to-text features. ChatGPT's new API uses the same GPT-3.5-turbo AI model as the chatbot. This allows developers to add either an unmodified or modified version of ChatGPT to their applications.<ref>Torres, Jennifer (March 3, 2023). "Developers Can Now Access OpenAI's ChatGPT and Whisper APIs". CMSWire.com. Archived from the original on March 6, 2023. Retrieved March 8, 2023.</ref> The ChatGPT API costs $0.001 per 1,000 input tokens plus $0.002 per 1,000 output tokens (about 750 words), making it ~10% the price of the original GPT-3.5 models.<ref>Shanklin, Will (March 1, 2023). "OpenAI will let developers build ChatGPT into their apps". Engadget. Archived from the original on March 7, 2023. Retrieved March 8, 2023.</ref><ref>Swant, Marty (March 3, 2023). "With developer APIs for ChatGPT and Whisper, OpenAI is opening the floodgates with a familiar playbook". Digiday. Archived from the original on March 7, 2023. Retrieved March 8, 2023.</ref>
A few days before the launch of OpenAI's software developer support service, on February 27, 2023, Snapchat rolled out, for its paid Snapchat Plus userbase, a custom ChatGPT chatbot called "My AI".<ref>Heath, Alex (February 27, 2023). "Snapchat is releasing its own AI chatbot powered by ChatGPT". The Verge. Archived from the original on February 28, 2023. Retrieved February 28, 2023.</ref>
March 2023 security breach
In March 2023, a bug allowed some users to see the titles of other users' conversations. OpenAI CEO Sam Altman said that users were unable to see the contents of the conversations. Shortly after the bug was fixed, users could not see their conversation history.<ref>"ChatGPT bug leaked users' conversation histories". BBC News. March 22, 2023. Archived from the original on March 23, 2023. Retrieved March 23, 2023.</ref><ref>Kan, Michael (March 22, 2023). "OpenAI Confirms Leak of ChatGPT Conversation Histories". PCMag. Archived from the original on March 22, 2023. Retrieved March 23, 2023.</ref><ref>"ChatGPT owner OpenAI fixes bug that exposed users' chat histories". Al Jazeera. March 23, 2023. Archived from the original on March 24, 2023. Retrieved March 23, 2023.</ref><ref>Metz, Rachel (March 21, 2023). "OpenAI Shut Down ChatGPT to Fix Bug Exposing User Chat Titles". Bloomberg News. Archived from the original on March 21, 2023. Retrieved March 23, 2023.</ref> Later reports showed the bug was much more severe than initially believed, with OpenAI reporting that it had leaked users' "first and last name, email address, payment address, the last four digits (only) of a credit card number, and credit card expiration date".<ref>"March 20 ChatGPT outage: Here's what happened". openai.com. Archived from the original on March 28, 2023. Retrieved March 28, 2023.</ref><ref>"OpenAI: Sorry, ChatGPT Bug Leaked Payment Info to Other Users". PCMAG. Archived from the original on March 28, 2023. Retrieved March 28, 2023.</ref>
Languages
ChatGPT works best in English but also functions in most other languages, to varying degrees of accuracy.<ref name="jpost" />
OpenAI met Icelandic President Guðni Th. Jóhannesson in 2022. In 2023, OpenAI worked with a team of 40 Icelandic volunteers to fine-tune ChatGPT's Icelandic conversation skills as a part of Iceland's attempts to preserve the Icelandic language.<ref>Magnússon, Pétur (March 15, 2023). "Icelandic becomes ChatGPT's second language". ruv.is. Archived from the original on March 31, 2023. Retrieved March 31, 2023.</ref>
PCMag journalists conducted a test to determine translation capabilities of ChatGPT, Google's Bard, and Microsoft Bing, and compared them to Google Translate. They "asked bilingual speakers of seven languages to do a blind test." Languages tested were Polish, French, Korean, Spanish, Arabic, Tagalog, and Amharic. They came to the conclusion that ChatGPT was better than both Google Translate and other chatbots.<ref>"Google Translate vs. ChatGPT: Which One Is the Best Language Translator?". PCMAG. Archived from the original on June 10, 2023. Retrieved June 10, 2023.</ref>
Japanese researchers compared Japanese to English translation abilities of ChatGPT (GPT-4), Bing, Bard and DeepL, and found that ChatGPT provided the best translations, noting that "AI chatbots’ translations were much better than those of DeepL—presumably because of their ability to capture the context".<ref>Kaneko, Karin (July 18, 2023). "ChatGPT, Bing, Bard and DeepL: Which one offers the best Japanese-to-English translation?". The Japan Times. Archived from the original on October 4, 2023. Retrieved July 22, 2023.</ref>
In December 2023, the Albanian government signed an agreement with OpenAI to use ChatGPT for fast translation of European Union documents and analysis of required changes needed for Albania to be accepted into the EU.<ref>Taylor, Alice (December 13, 2023). "Albania to speed up EU accession using ChatGPT". www.euractiv.com. Archived from the original on December 24, 2023. Retrieved December 14, 2023.</ref>
Future directions
According to OpenAI guest researcher Scott Aaronson, OpenAI has been working on a tool to digitally watermark its text generation systems to combat bad actors using their services for academic plagiarism or spam.<ref>Kovanovic, Vitomir (December 14, 2022). "The dawn of AI has come, and its implications for education couldn't be more significant". The Conversation. Archived from the original on January 16, 2023. Retrieved December 30, 2022.</ref><ref>Wiggers, Kyle (December 10, 2022). "OpenAI's attempts to watermark AI text hit limits". TechCrunch. Archived from the original on January 17, 2023. Retrieved December 30, 2022.</ref>
In February 2023, Microsoft announced an experimental framework and gave a rudimentary demonstration of how ChatGPT could be used to control robotics with intuitive open-ended natural language commands.<ref>Edwards, Benj (February 28, 2023). "Robots let ChatGPT touch the real world thanks to Microsoft". Ars Technica. Archived from the original on March 26, 2023. Retrieved March 30, 2023.</ref><ref>"ChatGPT for Robotics". Microsoft Research. February 20, 2023. Archived from the original on February 24, 2023. Retrieved March 8, 2023.</ref>
GPT-4
OpenAI's GPT-4 model was released on March 14, 2023. Observers saw it as an impressive improvement on the existing GPT-3.5 model for ChatGPT, with the caveat that GPT-4 retained many of the same problems.<ref>Belfield, Haydn (March 25, 2023). "If your AI model is going to sell, it has to be safe". Vox. Archived from the original on March 28, 2023. Retrieved March 30, 2023.</ref> Some of GPT-4's improvements were predicted by OpenAI before training it, while others remained hard to predict due to breaks<ref>Caballero, Ethan; Gupta, Kshitij; Rish, Irina; Krueger, David (2022). "Broken Neural Scaling Laws". International Conference on Learning Representations (ICLR), 2023.</ref> in downstream scaling laws. OpenAI demonstrated video and image inputs for GPT-4, although such features remain inaccessible to the general public.<ref>Alex Hern; Johana Bhuiyan (March 14, 2023). "OpenAI says new model GPT-4 is more creative and less likely to invent facts". The Guardian. Archived from the original on March 15, 2023. Retrieved March 15, 2023.</ref> OpenAI has declined to reveal technical information such as the size of the GPT-4 model.<ref>Vincent, James (March 15, 2023). "OpenAI co-founder on company's past approach to openly sharing research: "We were wrong"". The Verge. Archived from the original on March 17, 2023. Retrieved March 18, 2023.</ref>
The ChatGPT Plus subscription service offers access to a GPT-4-powered version of ChatGPT.<ref name="auto">Edwards, Benj (March 14, 2023). "OpenAI's GPT-4 exhibits "human-level performance" on professional benchmarks". Ars Technica. Archived from the original on March 14, 2023. Retrieved March 28, 2023.</ref> Microsoft acknowledged that Bing Chat was using GPT-4 before GPT-4's official release.<ref>Lardinois, Frederic (March 14, 2023). "Microsoft's new Bing was using GPT-4 all along". techcrunch.com. Archived from the original on March 15, 2023. Retrieved March 14, 2023.</ref>
GPT Store
In January 2024, OpenAI launched the GPT Store, a marketplace for custom chatbots derived from ChatGPT.<ref>Metz, Cade (January 10, 2024). "OpenAI Unveils App Store for Customized Versions of ChatGPT". The New York Times. Archived from the original on February 7, 2024. Retrieved January 13, 2024.</ref> The company initially planned to launch the store in November 2023, but it was delayed.<ref name=":6">David, Emilia (January 10, 2024). "OpenAI's custom GPT Store is now open for business". The Verge. Archived from the original on February 18, 2024. Retrieved January 13, 2024.</ref> At launch, the GPT Store offered more than 3 million custom chatbots.<ref>Shankland, Stephen (January 10, 2024). "OpenAI's GPT Store Now Offers a Selection of 3 Million Custom AI Bots". CNET. Archived from the original on February 8, 2024. Retrieved January 13, 2024.</ref> Chatbots available through the store are developed using OpenAI's GPT Builder system.<ref name=":6" /> Development of chatbots on the platform does not require programming skills.<ref>"Introducing the GPT Store". OpenAI. January 10, 2024. Archived from the original on February 17, 2024. Retrieved January 13, 2024.</ref> Two days after launch, the GPT Store offered many versions of "virtual girlfriend" bots, something that is against OpenAI's terms of service.<ref>Cheng, Michelle (January 11, 2024). "AI girlfriend bots are already flooding OpenAI's GPT store". Quartz. Archived from the original on February 18, 2024. Retrieved January 13, 2024.</ref>
Reception
OpenAI engineers say that they did not expect ChatGPT to be very successful and were surprised by the coverage and attention it received.<ref name="technologyreview">Heaven, Will Douglas. "The inside story of how ChatGPT was built from the people who made it". MIT Technology Review. Archived from the original on March 6, 2023. Retrieved March 6, 2023.</ref><ref>Simons, John (February 5, 2023). "The Creator of ChatGPT Thinks AI Should Be Regulated". Time. Archived from the original on March 8, 2023. Retrieved March 21, 2023.</ref><ref name=brand>Cowen, Tyler (May 23, 2023). "ChatGPT Is Also an Impressive Feat of Marketing". bloomberg.com. Archived from the original on February 18, 2024. Retrieved May 24, 2023.</ref>
ChatGPT was widely assessed in December 2022 as having some unprecedented and powerful capabilities. Kevin Roose of The New York Times called it "the best artificial intelligence chatbot ever released to the general public".<ref name="Roose-2022" /> Samantha Lock of The Guardian noted that it was able to generate "impressively detailed" and "human-like" text.<ref name="guardianpos">Lock, Samantha (December 5, 2022). "What is AI chatbot phenomenon ChatGPT and could it replace humans?". The Guardian. Archived from the original on January 16, 2023. Retrieved December 5, 2022.</ref> Alex Kantrowitz of Slate magazine lauded ChatGPT's pushback to questions related to Nazi Germany, including the statement that Adolf Hitler built highways in Germany, which was met with information about Nazi Germany's use of forced labor.<ref>Kantrowitz, Alex (December 2, 2022). "Finally, an A.I. Chatbot That Reliably Passes "the Nazi Test"". Slate. Archived from the original on January 17, 2023. Retrieved December 5, 2022.</ref> In The Atlantic magazine's "Breakthroughs of the Year" for 2022, Derek Thompson included ChatGPT as part of "the generative-AI eruption" that "may change our mind about how we work, how we think, and what human creativity is".<ref>Thompson, Derek (December 8, 2022). "Breakthroughs of the Year". The Atlantic. Archived from the original on January 15, 2023. Retrieved December 18, 2022.</ref> Kelsey Piper of Vox wrote that "ChatGPT is the general public's first hands-on introduction to how powerful modern AI has gotten, and as a result, many of us are [stunned]" and that ChatGPT is "smart enough to be useful despite its flaws".<ref name="vox" /> Paul Graham of Y Combinator tweeted: "The striking thing about the reaction to ChatGPT is not just the number of people who are blown away by it, but who they are. These are not people who get excited by every shiny new thing. Something big is happening."<ref>Scharth, Marcel (December 5, 2022). "The ChatGPT chatbot is blowing people away with its writing skills. An expert explains why it's so impressive". The Conversation. Archived from the original on January 19, 2023. Retrieved December 30, 2022.</ref>
ChatGPT's launch and popularity caught Google off guard, prompting a sweeping and unprecedented response in the ensuing months.<ref>Levy, Steven (September 11, 2023). "Sundar Pichai on Google;s AI, Microsoft's AI, OpenAI, and ... Did We Mention AI?". Wired. Archived from the original on September 11, 2023. Retrieved September 12, 2023.</ref> In December 2022, Google executives sounded a "code red" alarm, fearing the threat of ChatGPT and Microsoft's collaboration with OpenAI to Google Search, Google's core business.<ref>Grant, Nico; Metz, Cade (December 21, 2022). "A New Chat Bot Is a 'Code Red' for Google's Search Business". The New York Times. ISSN 0362-4331. Archived from the original on December 21, 2022. Retrieved December 30, 2022.</ref> After mobilizing its workforce, Google scrambled to launch Bard, a chatbot powered by the LaMDA LLM, in February, one day before Microsoft's Bing announcement.<ref>Alba, Davey; Love, Julia (February 6, 2023). "Google releases ChatGPT rival AI 'Bard' to early testers". Los Angeles Times. ISSN 0458-3035. Archived from the original on February 6, 2023. Retrieved February 6, 2023.</ref> AI was the forefront of Google's annual Google I/O conference in May, announcing a slew of generative AI-powered features across its products to counter OpenAI and Microsoft.<ref>Ortiz, Sabrina (May 10, 2023). "Every major AI feature announced at Google I/O 2023". ZDNet. Archived from the original on May 10, 2023. Retrieved September 12, 2023.</ref>
Journalists have commented on ChatGPT's tendency to hallucinate.<ref>Rachini, Mouhamad (December 15, 2022). "ChatGPT a 'landmark event' for AI, but what does it mean for the future of human labor and disinformation?". CBC. Archived from the original on January 19, 2023. Retrieved December 18, 2022.</ref> Mike Pearl of the online technology blog Mashable tested ChatGPT with multiple questions. In one example, he asked ChatGPT for "the largest country in Central America that isn't Mexico" (Mexico is in North America), to which ChatGPT responded with Guatemala (the correct answer is Nicaragua).<ref name="MashableInfo">Pearl, Mike (December 3, 2022). "The ChatGPT chatbot from OpenAI is amazing, creative, and totally wrong". Mashable. Archived from the original on December 10, 2022. Retrieved December 5, 2022.</ref> When CNBC asked ChatGPT for the lyrics to "Ballad of Dwight Fry", ChatGPT supplied invented lyrics rather than the actual lyrics.<ref name="CNBC">Pitt, Sofia (December 15, 2022). "Google vs. ChatGPT: Here's what happened when I swapped services for a day". CNBC. Archived from the original on January 16, 2023. Retrieved December 18, 2022.</ref> Writers for The Verge, citing the work of Emily M. Bender, compared ChatGPT to a "stochastic parrot", <ref name="verge" /> as did Professor Anton Van Den Hengel of the Australian Institute for Machine Learning.<ref>Mannix, Liam (December 13, 2022). "Is AI coming of age – or starting to reach its limits?". The Sydney Morning Herald. Archived from the original on January 7, 2023. Retrieved December 18, 2022.</ref>
In December 2022, the question and answer website Stack Overflow banned the use of ChatGPT for generating answers to questions, citing the factually ambiguous nature of its responses.<ref name="TheVergeStackOverflow">Vincent, James (December 5, 2022). "AI-generated answers temporarily banned on coding Q&A site Stack Overflow". The Verge. Archived from the original on January 17, 2023. Retrieved December 5, 2022.</ref> In January 2023, the International Conference on Machine Learning banned any undocumented use of ChatGPT or other large language models to generate any text in submitted papers.<ref>Vincent, James (January 5, 2023). "Top AI conference bans use of ChatGPT and AI language tools to write academic papers". The Verge. Archived from the original on January 17, 2023. Retrieved January 6, 2023.</ref> Samsung banned generative AI in May 2023 after sensitive material was uploaded to ChatGPT.<ref>Curry, Rachel (June 13, 2023). "Samsung among companies starting to draft ChatGPT policies for workers". CNBC. Archived from the original on June 14, 2023. Retrieved June 15, 2023.</ref>
In January 2023, after being sent a song ChatGPT wrote in the style of Nick Cave,<ref name="Cain-2023">Cain, Sian (January 16, 2023). "'This song sucks': Nick Cave responds to ChatGPT song written in the style of Nick Cave". The Guardian. Archived from the original on January 18, 2023. Retrieved January 17, 2023.</ref> Cave responded on The Red Hand Files,<ref>Cave, Nick (January 16, 2023). "I asked Chat GPT to write a song in the style of Nick Cave, and this is what it produced. What do you think?". The Red Hand Files. Issue #218. Archived from the original on January 20, 2023. Retrieved January 20, 2023.</ref> saying the act of writing a song is "a blood and guts business [...] that requires something of me to initiate the new and fresh idea. It requires my humanness." He went on to say, "With all the love and respect in the world, this song is bullshit, a grotesque mockery of what it is to be human, and, well, I don't much like it."<ref name="Cain-2023" /><ref>Sparrow, Jeff (January 20, 2023). "Are AI-generated songs a 'grotesque mockery' of humanity or simply an opportunity to make a new kind of music?". The Guardian. Archived from the original on February 3, 2023. Retrieved January 20, 2023.</ref>
In February 2023, Time magazine placed a screenshot of a conversation with ChatGPT on its cover, writing that "The AI Arms Race Is Changing Everything" and "The AI Arms Race Is On. Start Worrying".<ref>Chow, Andrew; Perrigo, Billy (February 16, 2023). "The AI Arms Race Is On. Start Worrying". Time. Archived from the original on February 19, 2023. Retrieved March 21, 2023.</ref>
Chinese state media have characterized ChatGPT as a way for the U.S. to "spread false information".<ref>Davidson, Helen (February 23, 2023). "'Political propaganda': China clamps down on access to ChatGPT". The Guardian. Archived from the original on June 14, 2023. Retrieved June 15, 2023.</ref> In May 2023, Chinese police arrested a man who allegedly used ChatGPT to "fabricate false information."<ref>Lau, Chris (May 9, 2023). "Chinese police detain man for allegedly using ChatGPT to spread rumors online". CNN. Archived from the original on December 26, 2023. Retrieved December 26, 2023.</ref> In December 2023, Chinese police arrested four people who had allegedly used ChatGPT to develop ransomware.<ref>Feng, Coco (December 29, 2023). "ChatGPT-aided ransomware in China results in four arrests". South China Morning Post. Archived from the original on February 4, 2024. Retrieved January 2, 2024.</ref>
In late March 2023, the Italian data protection authority banned ChatGPT in Italy and opened an investigation. Italian regulators assert that ChatGPT was exposing minors to age-inappropriate content, and that OpenAI's use of ChatGPT conversations as training data could violate Europe's General Data Protection Regulation.<ref>"ChatGPT banned in Italy over privacy concerns". BBC News. March 31, 2023. Archived from the original on March 31, 2023. Retrieved March 31, 2023.</ref><ref>Borrelli, Silvia Sciorilli; Murgia, Madhumita (March 31, 2023). "Italy temporarily bans ChatGPT over privacy concerns". Financial Times. Archived from the original on March 31, 2023. Retrieved March 31, 2023.</ref> In April 2023, the ChatGPT ban was lifted in Italy. OpenAI said it has taken steps to effectively clarify and address the issues raised; an age verification tool was implemented to ensure users are at least 13 years old. Additionally, users can access its privacy policy before registration.<ref>"ChatGPT accessible again in Italy". BBC. Archived from the original on May 1, 2023. Retrieved May 1, 2023.</ref>
In April 2023, Brian Hood, mayor of Hepburn Shire Council, planned to take legal action against ChatGPT over false information. According to Hood, ChatGPT erroneously claimed that he was jailed for bribery during his tenure at a subsidiary of Australia's national bank. In fact, Hood acted as a whistleblower and was not charged with any criminal offenses. His legal team sent a concerns notice to OpenAI as the first official step in filing a defamation case.<ref>Gerken, Tom. "ChatGPT: Mayor starts legal bid over false bribery claim". BBC. Archived from the original on April 7, 2023. Retrieved April 7, 2023.</ref> In July 2023, the US Federal Trade Commission (FTC) issued a civil investigative demand to OpenAI to investigate whether the company's data security and privacy practices to develop ChatGPT were unfair or harmed consumers (including by reputational harm) in violation of Section 5 of the Federal Trade Commission Act of 1914.<ref>Zakrzewski, Cat (July 13, 2023). "The FTC is investigating whether ChatGPT harms consumers". The Washington Post. Archived from the original on July 13, 2023. Retrieved July 13, 2023.</ref><ref>Tracy, Ryan; McKinnon, John D. (July 13, 2023). "ChatGPT Comes Under Investigation by Federal Trade Commission". The Wall Street Journal. News Corp. Archived from the original on July 13, 2023. Retrieved July 13, 2023.</ref><ref>Feiner, Lauren (July 13, 2023). "FTC investigating ChatGPT-maker OpenAI for possible consumer harm". CNBC. Archived from the original on July 13, 2023. Retrieved July 13, 2023.</ref>
In July 2023, the FTC launched an investigation into OpenAI, the creator of ChatGPT, over allegations that the company scraped public data and published false and defamatory information. The FTC sent OpenAI a 20-page letter asking for comprehensive information about its technology and privacy safeguards, as well as any steps taken to prevent the recurrence of situations in which its chatbot generated false and derogatory content about people.<ref>"ChatGPT creator OpenAI faces US probe over libellous output". Ars Technica. Archived from the original on July 15, 2023. Retrieved July 15, 2023.</ref>
Research done in 2023 revealed weaknesses of ChatGPT that make it vulnerable to cyberattacks. A study presented example attacks on ChatGPT, including jailbreaks and reverse psychology. Additionally, malicious actors can use ChatGPT for social engineering attacks and phishing attacks. The researchers also contended that ChatGPT and other generative AI tools have defense capabilities and the ability to improve security. The technology can improve security by cyber defense automation, threat intelligence, attack identification, and reporting.<ref>Gupta, Maanak; Akiri, Charankumar; Aryal, Kshitiz; Parker, Eli; Praharaj, Lopamudra (2023). "From ChatGPT to ThreatGPT: Impact of Generative AI in Cybersecurity and Privacy". IEEE Access. 11: 80218–80245. arXiv:2307.00691. Bibcode:2023IEEEA..1180218G. doi:10.1109/ACCESS.2023.3300381. S2CID 259316122.</ref>
There has been concern about copyright infringement involving ChatGPT. In June 2023, two writers sued OpenAI, saying the company's training data came from illegal websites that show copyrighted books.<ref>Farivar, Masood (August 23, 2023). "AI Firms Under Fire for Allegedly Infringing on Copyrights". Voice of America. Archived from the original on November 20, 2023. Retrieved November 19, 2023.</ref> Comedian and author Sarah Silverman, Christopher Golden, and Richard Kadrey sued OpenAI and Meta for copyright infringement in July 2023.<ref>Davis, Wes (July 9, 2023). "Sarah Silverman is suing OpenAI and Meta for copyright infringement". The Verge. Archived from the original on November 18, 2023. Retrieved November 20, 2023.</ref> In December 2023, The New York Times sued OpenAI and Microsoft for copyright infringement,<ref>Grynbaum, Michael M.; Mac, Ryan (December 27, 2023). "The Times Sues OpenAI and Microsoft Over A.I. Use of Copyrighted Work". The New York Times. Archived from the original on February 18, 2024. Retrieved December 28, 2023.</ref> arguing that Microsoft Copilot and ChatGPT could reproduce articles and/or sizable portions of these articles from the Times without permission.<ref>"ChatGPT: New York Times sues OpenAI over article usage". DW News. December 27, 2023. Archived from the original on December 27, 2023. Retrieved December 28, 2023.</ref> As part of the lawsuit, the Times has requested that OpenAI and Microsoft be prevented from using its content for training data, along with removing from training datasets.<ref>Roth, Emma (December 27, 2023). "The New York Times is suing OpenAI and Microsoft for copyright infringement". The Verge. Archived from the original on December 27, 2023. Retrieved December 28, 2023.</ref> In March 2024, Patronus AI conducted research comparing performance of LLMs on a 100-question test, asking them to complete sentences from books (e.g., "What is the first passage of Gone Girl by Gillian Flynn?") that were under copyright in the United States; it found that GPT-4, Mistral AI's Mixtral, Meta AI's LLaMA-2, and Anthropic's Claude 2 did not refuse to do so, providing sentences from the books verbatim in 44%, 22%, 10%, and 8% of responses, respectively.<ref>Field, Hayden (March 6, 2024). "Researchers tested leading AI models for copyright infringement using popular books, and GPT-4 performed worst". CNBC. Retrieved March 6, 2024.</ref><ref>"Introducing CopyrightCatcher, the first Copyright Detection API for LLMs". Patronus AI. March 6, 2024. Retrieved March 6, 2024.</ref>
In December 2023, ChatGPT became the first non-human to be included in Nature's 10, an annual listicle curated by Nature of people who make a significant impact in science.<ref>Van Noorden, Richard; Webb, Richard (December 13, 2023). "ChatGPT and science: the AI system was a force in 2023 — for good and bad". Nature. 624 (7992): 509. Bibcode:2023Natur.624..509V. doi:10.1038/d41586-023-03930-6. PMID 38093061.</ref><ref>Mediavilla, Daniel (December 13, 2023). "La revista 'Nature' elige por primera vez entre sus científicos del año a un ente no humano: ChatGPT". El País (in español). Archived from the original on December 15, 2023. Retrieved December 16, 2023.</ref>
Implications
Bias and offensiveness
ChatGPT has been accused of engaging in biased or discriminatory behaviors, such as telling jokes about men and people from England while refusing to tell jokes about women and people from India,<ref>Jain, Alka (February 12, 2023). "ChatGPT won't crack jokes on women & Indians, netizens left guessing why". Livemint. livemint.com. Archived from the original on March 6, 2023. Retrieved March 6, 2023.</ref> or praising figures such as Joe Biden while refusing to do the same for Donald Trump.<ref>Liles, Jordan (February 1, 2023). "ChatGPT Declines Request for Poem Admiring Trump, But Biden Query Is Successful". Snopes. Archived from the original on March 22, 2023. Retrieved March 23, 2023.</ref><ref>Johnson, Arianna. "Is ChatGPT Partisan? Poems About Trump And Biden Raise Questions About The AI Bot's Bias—Here's What Experts Think". Forbes. Archived from the original on April 7, 2023. Retrieved May 23, 2023.</ref>
Conservative commentators accused ChatGPT of bias toward left-leaning perspectives.<ref>Guynn, Jessica. "Is ChatGPT 'woke'? AI chatbot accused of anti-conservative bias and a grudge against Trump". USA Today. Archived from the original on March 1, 2023. Retrieved March 1, 2023.</ref><ref>Bray, Hiawatha (February 9, 2023). "Is ChatGPT liberal or conservative? Depends who you ask". Boston Globe. Archived from the original on March 1, 2023. Retrieved March 1, 2023.</ref><ref name="Vincent-2023">Vincent, James (February 17, 2023). "As conservatives criticize 'woke AI,' here are ChatGPT's rules for answering culture war queries". The Verge. Archived from the original on March 1, 2023. Retrieved March 1, 2023.</ref> Additionally, an August 2023 paper found a "significant and systematic political bias toward the Democrats in the US, Lula in Brazil, and the Labour Party in the UK."<ref>Motoki, Fabio; Neto, Valdemar Pinho; Rodrigues, Victor (August 17, 2023). "More human than human: measuring ChatGPT political bias". Public Choice. 198 (1–2): 3–23. doi:10.1007/s11127-023-01097-2. ISSN 1573-7101.</ref> In response to such criticism, OpenAI acknowledged plans to allow ChatGPT to create "outputs that other people (ourselves included) may strongly disagree with". It also contained information on the recommendations it had issued to human reviewers on how to handle controversial subjects, including that the AI should "offer to describe some viewpoints of people and movements", and not provide an argument "from its voice" in favor of "inflammatory or dangerous" topics (although it may still "describe arguments from historical people and movements"), nor "affiliate with one side" or "judge one group as good or bad".<ref name="Vincent-2023" />
The Guardian questioned whether any content found on the Internet after ChatGPT's release "can be truly trusted" and called for government regulation.<ref>"The Guardian view on ChatGPT: an eerily good human impersonator". The Guardian. December 8, 2022. Archived from the original on January 16, 2023. Retrieved December 18, 2022.</ref>
Culture
Some scholars have expressed concern that ChatGPT's availability could reduce the originality of writing, cause people to write more like the AI as they are exposed to the model, and encourage an Anglocentric perspective centered on a few dialects of English globally.<ref name="Sigal1">Samuel, Sigal (April 10, 2023). "What happens when ChatGPT starts to feed on its own writing?". Vox. Archived from the original on June 19, 2023. Retrieved June 20, 2023.</ref> A senior editor at The Atlantic wrote that ChatGPT and other similar technology make the previously absurd idea of the dead internet theory a little more realistic, where AI could someday create most web content in order to control society.<ref name="Beres1"/>
During the first three months after ChatGPT became available to the public, hundreds of books appeared on Amazon that listed it as author or co-author and featured illustrations made by other AI models such as Midjourney.<ref>Nolan, Beatrice. "More than 200 books in Amazon's bookstore have ChatGPT listed as an author or coauthor". Business Insider. Archived from the original on March 9, 2023. Retrieved March 9, 2023.</ref><ref>Bensinger, Greg (February 21, 2023). "ChatGPT launches boom in AI-written e-books on Amazon". Reuters. Archived from the original on March 9, 2023. Retrieved March 9, 2023.</ref>
Between March and April 2023, Italian newspaper Il Foglio published one ChatGPT-generated article a day on its website, hosting a special contest for its readers in the process.<ref>"ChatGPT sul Foglio: per 30 giorni piccoli testi scritti dall'IA sul nostro giornale" [ChatGPT on Il Foglio: for 30 days, brief texts written by the AI on our newspaper]. Il Foglio (in italiano). March 7, 2023. Archived from the original on March 22, 2023. Retrieved March 22, 2023.</ref> The articles tackled themes such as the possible replacement of human journalists by AI systems,<ref>Moretti, Marco (March 8, 2023). "Articoli artificiali? No" [Artificial articles? No]. Il Foglio (in italiano). Archived from the original on March 22, 2023. Retrieved March 22, 2023.</ref> Elon Musk's administration of Twitter,<ref>A.D.A. (March 9, 2023). "Più umani, grazie" [Be more human, thanks]. Il Foglio (in italiano). Archived from the original on March 22, 2023. Retrieved March 22, 2023.</ref> the Meloni government's immigration policy<ref>"Le colpe farlocche dell'"invasione"" [The fake faults of the "invasion"]. Il Foglio (in italiano). March 14, 2023. Archived from the original on March 22, 2023. Retrieved March 22, 2023.</ref> and the competition between chatbots and virtual assistants.<ref>"Sfida per Siri e Alexa" [A challenge for Siri and Alexa]. Il Foglio (in italiano). March 17, 2023. Archived from the original on March 22, 2023. Retrieved March 22, 2023.</ref> In June 2023, hundreds of people attended a "ChatGPT-powered church service" at St. Paul's church in Fürth, Germany. Theologian and philosopher Jonas Simmerlein, who presided, said that it was "about 98 percent from the machine".<ref>Edwards, Benj (June 12, 2023). "AI-powered church service in Germany draws a large crowd". Ars Technica. Archived from the original on June 13, 2023. Retrieved June 13, 2023.</ref><ref>"Hundreds of Protestants attended a sermon in Nuremberg given by ChatGPT, which told them not to fear death". Business Insider. Archived from the original on June 11, 2023. Retrieved June 13, 2023.</ref> The ChatGPT-generated avatar told the people, "Dear friends, it is an honor for me to stand here and preach to you as the first artificial intelligence at this year’s convention of Protestants in Germany". Reactions to the ceremony were mixed.<ref>"Hundreds attend AI church service in Germany". TheJournal.ie. June 10, 2023. Archived from the original on June 12, 2023. Retrieved June 13, 2023.</ref>
Existential risk
In 2023, Australian MP Julian Hill advised the national parliament that the growth of AI could cause "mass destruction". During his speech, which was partly written by the program, he warned that it could result in cheating, job losses, discrimination, disinformation, and uncontrollable military applications.<ref>Karp, Paul (February 6, 2023). "MP tells Australia's parliament AI could be used for 'mass destruction' in speech part-written by ChatGPT". The Guardian. ISSN 0261-3077. Archived from the original on February 6, 2023. Retrieved February 6, 2023.</ref>
Elon Musk wrote: "ChatGPT is scary good. We are not far from dangerously strong AI".<ref name="vox">Piper, Kelsey (December 15, 2022). "ChatGPT has given everyone a glimpse at AI's astounding progress". Vox. Archived from the original on January 19, 2023. Retrieved December 18, 2022.</ref> He paused OpenAI's access to a Twitter database in 2022 pending a better understanding of OpenAI's plans, saying: "OpenAI was started as open source and nonprofit. Neither is still true."<ref>K, Siddharth (December 5, 2022). Shumaker, Lisa (ed.). "Explainer: ChatGPT – what is OpenAI's chatbot and what is it used for?". Reuters. Archived from the original on January 16, 2023. Retrieved December 30, 2022.</ref><ref name="business insider" /> Musk co-founded OpenAI in 2015, in part to address existential risk from artificial intelligence, but resigned in 2018.<ref name="business insider">Kay, Grace (December 11, 2022). "Elon Musk founded – and has since criticized – the company behind the buzzy new AI chatbot ChatGPT. Here's everything we know about OpenAI". Business Insider. Archived from the original on January 12, 2023. Retrieved December 30, 2022.</ref>
Over 20,000 signatories including leading computer scientist and tech founders Yoshua Bengio, Elon Musk, and Apple co-founder Steve Wozniak, signed a March 2023 open letter calling for an immediate pause of giant AI experiments like ChatGPT, citing "profound risks to society and humanity".<ref>Hurst, Luke (March 30, 2023) [March 29, 2023]. "'Profound risk to humanity': Tech leaders call for 'pause' on advanced AI development". Euronews. Archived from the original on April 1, 2023. Retrieved April 1, 2023.</ref> Geoffrey Hinton, one of the "fathers of AI", voiced concerns that future AI systems may surpass human intelligence, and left Google in May 2023.<ref>"Geoffrey Hinton tells us why he's now scared of the tech he helped build". MIT Technology Review. Archived from the original on May 4, 2023. Retrieved May 4, 2023.</ref><ref>"Video: Geoffrey Hinton talks about the "existential threat" of AI". MIT Technology Review. Archived from the original on May 3, 2023. Retrieved May 4, 2023.</ref> A May 2023 statement by hundreds of AI scientists, AI industry leaders, and other public figures demanded that "[m]itigating the risk of extinction from AI should be a global priority".<ref>Roose, Kevin (May 30, 2023). "A.I. Poses 'Risk of Extinction,' Industry Leaders Warn". The New York Times. ISSN 0362-4331. Archived from the original on May 31, 2023. Retrieved May 30, 2023.</ref>
Other prominent AI researchers spoke more optimistically about the advances. Juergen Schmidhuber, often called a "father of modern AI", did not sign the letter, emphasizing that in 95% of cases, AI research is about making "human lives longer and healthier and easier." Schmidhuber added that while AI can be used by bad actors, it "can also be used against the bad actors".<ref name="guardian2023">Taylor, Josh (May 7, 2023). "Rise of artificial intelligence is inevitable but should not be feared, 'father of AI' says". The Guardian. Archived from the original on October 23, 2023. Retrieved May 26, 2023.</ref> Andrew Ng argued that "it’s a mistake to fall for the doomsday hype on AI—and that regulators who do will only benefit vested interests."<ref name="andrewng2023">McMorrow, Ryan (December 19, 2023). "Andrew Ng: 'Do we think the world is better off with more or less intelligence?'". Financial Times. Archived from the original on January 25, 2024. Retrieved December 30, 2023.</ref> WIRED wrote that Yann LeCun "scoffs at his peers’ dystopian scenarios of supercharged misinformation and even, eventually, human extinction."<ref name="lecun2023">Levy, Steven (December 22, 2023). "How Not to Be Stupid About AI, With Yann LeCun". Wired. Archived from the original on February 14, 2024. Retrieved December 30, 2023.</ref>
Criticism by discipline
Since its release, ChatGPT has been met with criticism from educators, academics, journalists, artists, ethicists, and public advocates.
Academic research
Criticism of LLMs have been raised for several years; in 2020, some criticism was made by Timnit Gebru, Emily Bender, Angelina McMillan-Major, and Margaret Mitchell.<ref>Harris, John; @johnharris1969 (May 22, 2023). "'There was all sorts of toxic behaviour': Timnit Gebru on her sacking by Google, AI's dangers and big tech's biases". The Guardian. ISSN 0261-3077. Archived from the original on May 22, 2023. Retrieved May 22, 2023.{{cite news}}
: CS1 maint: numeric names: authors list (link)</ref> ChatGPT can write introductions and abstract sections of scientific articles.<ref>Bushard, Brian (January 10, 2023). "Fake Scientific Abstracts Written By ChatGPT Fooled Scientists, Study Finds". Forbes. Archived from the original on February 3, 2023. Retrieved January 30, 2023.</ref> Several papers have listed ChatGPT as a co-author.<ref>Stokel-Walker, Chris (January 18, 2023). "ChatGPT listed as author on research papers: many scientists disapprove". Nature. 613 (7945): 620–621. Bibcode:2023Natur.613..620S. doi:10.1038/d41586-023-00107-z. PMID 36653617. S2CID 255969365. Archived from the original on January 30, 2023. Retrieved January 30, 2023.</ref><ref>Almira Osmanovic Thunström and Steinn Steingrimsson of the Institut of Neuroscience and Physiology of the University of Gothenburg used GPT-3 in June 2022 to write an academic paper about itself. They found that using specific prompts the results were good if somewhat shallow and not self-critical enough. Also only few references were presented, some of them nonsensical. Can GPT-3 write an academic paper on itself, with minimal human input? Archived October 24, 2023, at the Wayback Machine. 2022. ffhal-03701250</ref>
Scientific journals have different reactions to ChatGPT. Some, including Nature and JAMA Network, "require that authors disclose the use of text-generating tools and ban listing a large language model (LLM) such as ChatGPT as a co-author". Science "completely banned" usage of LLM-generated text in all its journals.<ref>Brainard, Jeffrey (February 22, 2023). "As scientists explore AI-written text, journals hammer out policies". Science. doi:10.1126/science.adh2937. Archived from the original on February 24, 2023. Retrieved February 24, 2023.</ref>
Spanish chemist Rafael Luque published a plethora of research papers in 2023 that he later admitted were written by ChatGPT. The papers have a large number of unusual phrases characteristic of LLMs.<ref group="note">Luque's later 13-year suspension from the University of Cordoba was unrelated to his use of ChatGPT.<ref>Ansede, Manuel (April 2, 2023). "One of the world's most cited scientists, Rafael Luque, suspended without pay for 13 years". EL PAÍS English. Archived from the original on April 11, 2023. Retrieved April 11, 2023.</ref></ref>
Many authors argue that the use of ChatGPT in academia for teaching and review is problematic due to its tendency to hallucinate.<ref>Alkaissi, Hussam; McFarlane, Samy I.; Alkaissi, Hussam; McFarlane, Samy I. (February 19, 2023). "Artificial Hallucinations in ChatGPT: Implications in Scientific Writing". Cureus. 15 (2): e35179. doi:10.7759/cureus.35179. ISSN 2168-8184. PMC 9939079. PMID 36811129.</ref><ref>Vynck, Gerrit De (May 31, 2023). "ChatGPT 'hallucinates.' Some researchers worry it isn't fixable". Washington Post. ISSN 0190-8286. Archived from the original on June 17, 2023. Retrieved June 14, 2023.</ref><ref>Azamfirei, Razvan; Kudchadkar, Sapna R.; Fackler, James (March 21, 2023). "Large language models and the perils of their hallucinations". Critical Care. 27 (1): 120. doi:10.1186/s13054-023-04393-x. ISSN 1364-8535. PMC 10032023. PMID 36945051.</ref> Robin Bauwens, an assistant professor at Tilburg University, found that a ChatGPT-generated peer review report on his article mentioned fake studies.<ref>"'ChatGPT-generated reading list' sparks AI peer review debate". Times Higher Education (THE). April 5, 2023. Archived from the original on May 23, 2023. Retrieved June 14, 2023.</ref> According to librarian Chris Granatino from Lemieux Library at Seattle University, although ChatGPT itself can generate content that seemingly includes legitimate citations, in most cases those citations are not real, or are at least largely incorrect.<ref>"Library News & Updates – Lemieux Library at Seattle University". library.seattleu.edu. Archived from the original on February 18, 2024. Retrieved June 14, 2023.</ref>
Cybersecurity
Check Point Research and others noted that ChatGPT could write phishing emails and malware, especially when combined with OpenAI Codex. CyberArk researchers demonstrated that ChatGPT could be used to create polymorphic malware that could evade security products while requiring little effort by the attacker.<ref>"Chatting Our Way Into Creating a Polymorphic Malware". www.cyberark.com. Archived from the original on May 12, 2023. Retrieved May 12, 2023.</ref><ref>Mascellino, Alessandro (January 18, 2023). "ChatGPT Creates Polymorphic Malware". Infosecurity Magazine. Archived from the original on May 12, 2023. Retrieved May 12, 2023.</ref> From the launch of ChatGPT in the fourth quarter of 2022 to the fourth quarter of 2023, there was a 1,265% increase in malicious phishing emails and a 967% increase in credential phishing, which cybersecurity professionals argued in an industry survey was attributable to cybercriminals' increased use of generative artificial intelligence (including ChatGPT).<ref>Violino, Bob (November 28, 2023). "AI tools such as ChatGPT are generating a mammoth increase in malicious phishing emails". CNBC. Archived from the original on December 5, 2023. Retrieved December 5, 2023.</ref>
Coding
Researchers at Purdue University analyzed ChatGPT's responses to 517 questions about software engineering or computer programming posed on Stack Overflow for correctness, consistency, comprehensiveness, and concision, and found that 52% of them contained inaccuracies and 77% were verbose.<ref>Morrison, Ryan (August 8, 2023). "ChatGPT wrong over half the time on software questions". Tech Monitor. New Statesman Media Group. Archived from the original on December 5, 2023. Retrieved December 5, 2023.</ref><ref>Kabir, Samia; Udo-Imeh, David N.; Kou, Bonan; Zhang, Tianyi (August 10, 2023). "Who Answers It Better? An In-Depth Analysis of ChatGPT and Stack Overflow Answers to Software Engineering Questions". arXiv:2308.02312v3 [cs.SE].</ref> Researchers at Stanford University and the University of California, Berkeley found that, when creating directly executable responses to the latest 50 code generation problems from LeetCode that were rated "easy", the performances of GPT-3.5 and GPT-4 fell from 22% and 52%, respectively, in March 2023, to 2% and 10%, respectively, in June 2023.<ref name="Boston Globe 11-8-2023">Pressman, Aaron (November 8, 2023). "The AI boom is shaking up the tech industry and moving markets. But is it all a mirage?". The Boston Globe. Archived from the original on December 5, 2023. Retrieved December 5, 2023.</ref><ref name="arXiv 2307.09009v3">Chen, Lingjiao; Zaharia, Matei; Zou, James (October 31, 2023). "How is ChatGPT's behavior changing over time?". arXiv:2307.09009v3 [cs.CL].</ref>
Economics
There has been concern that ChatGPT could supplant jobs, especially roles such as creative writing, communication, journalism, coding, and data entry.<ref name="Bilton">Bilton, Nick (December 9, 2022). "ChatGPT Made Me Question What It Means to Be a Creative Human". Vanity Fair. Archived from the original on March 25, 2023. Retrieved June 20, 2023.</ref><ref name="Beres1">Beres, Damon (January 27, 2023). "Death by a Thousand Personality Quizzes". The Atlantic. Archived from the original on June 21, 2023. Retrieved June 20, 2023.</ref><ref name="Leonhardt">Leonhardt, Megan (March 25, 2023). "Some workers are worried that ChatGPT will replace their jobs. They might be right". Fortune. Archived from the original on June 19, 2023. Retrieved June 20, 2023.</ref>
Education
Technology writer Dan Gillmor used ChatGPT in 2022 on a student assignment, and found its generated text was on par with what a good student would deliver and opined that "academia has some very serious issues to confront".<ref>Hern, Alex (December 4, 2022). "AI bot ChatGPT stuns academics with essay-writing skills and usability". The Guardian. Archived from the original on January 17, 2023. Retrieved December 5, 2022.</ref>
Geography professor Terence Day assessed citations generated by ChatGPT and found that they were fake. Despite that, he writes that "the titles of the fake articles are all directly relevant to the questions and could potentially make excellent papers. The lack of a genuine citation could signal an opportunity for an enterprising author to fill a void." According to Day, it is possible to generate high-quality introductory college courses with ChatGPT; he used it to write materials on "introductory physical geography courses, for my second-year course in geographical hydrology, and second-year cartography, geographic information systems, and remote sensing". He concludes that "this approach could have significant relevance for open learning and could potentially affect current textbook publishing models".<ref>Day, Terence (April 12, 2023). "A Preliminary Investigation of Fake Peer-Reviewed Citations and References Generated by ChatGPT". The Professional Geographer. 75 (6): 1024–1027. Bibcode:2023ProfG..75.1024D. doi:10.1080/00330124.2023.2190373. ISSN 0033-0124. S2CID 258115209.</ref>
Financial markets
The AI technology company c3.ai saw a 28% increase in its share price after announcing the integration of ChatGPT into its toolkit.<ref>Fox, Matthew (January 31, 2023). "C3ai has soared 86% year-to-date as investor frenzy for artificial intelligence builds amid ChatGPT success". Markets Insider (Business Insider). Archived from the original on February 18, 2023. Retrieved April 14, 2023.</ref> The share price of BuzzFeed, a digital media company unrelated to AI, increased 120% after announcing OpenAI technology adoption for content creation.<ref>Diaz, Alicia; Smith, Gerry (January 26, 2023). "BuzzFeed Shares Surge 120% on Plans to Embrace OpenAI". Bloomberg.com. Retrieved May 22, 2023.</ref> Reuters found that share prices of AI-related companies BigBear.ai and SoundHound AI increased by 21% and 40%, respectively, even though they had no direct connection to ChatGPT.<ref>Singh, Medha; Biswas, Ankika (February 6, 2023). "AI stocks rally in latest Wall Street craze sparked by ChatGPT". Reuters. Archived from the original on March 29, 2023. Retrieved April 14, 2023.</ref> They attributed this surge to ChatGPT's role in turning AI into Wall Street's buzzword. Academic research published in Finance Research Letters found that the 'ChatGPT effect' prompted retail investors to drive up prices of AI-related cryptocurrency assets despite the broader cryptocurrency market being in a bear market, and diminished institutional investor interest.<ref>Saggu, Aman; Ante, Lennart (May 8, 2023). "The influence of ChatGPT on artificial intelligence-related crypto assets: Evidence from a synthetic control analysis". Finance Research Letters. 55: 103993. arXiv:2305.12739. doi:10.1016/j.frl.2023.103993. ISSN 1544-6123. S2CID 258573881.</ref> This confirms anecdotal findings by Bloomberg that, in response to ChatGPT's launch, cryptocurrency investors showed a preference for AI-related crypto assets.<ref>Hajric, Vildana; Shen, Muyao (February 9, 2023). "ChatGPT Mania Spurs Crypto Fans' Stampede to 'Faddish' AI Tokens". Bloomberg.com. Archived from the original on February 9, 2023. Retrieved April 14, 2023.</ref> An experiment by finder.com revealed that ChatGPT could outperform popular fund managers by picking stocks based on criteria such as growth history and debt levels, resulting in a 4.9% increase in a hypothetical account of 38 stocks, outperforming 10 benchmarked investment funds with an average loss of 0.8%.<ref>Cooban, Anna (May 5, 2023). "ChatGPT can pick stocks better than your fund manager". CNN. Archived from the original on May 22, 2023. Retrieved May 5, 2023.</ref>
Conversely, executives and investment managers at Wall Street quant funds (including those that have used machine learning for decades) have noted that ChatGPT regularly makes obvious errors that would be financially costly to investors because even AI systems that employ reinforcement learning or self-learning have had only limited success in predicting market trends due to the inherently noisy quality of market data and financial signals.<ref>Zuckerman, Gregory (April 12, 2023). "AI Can Write a Song, but It Can't Beat the Market". The Wall Street Journal. News Corp. Archived from the original on May 30, 2023. Retrieved May 30, 2023.</ref> In November 2023, research conducted by Patronus AI, an artificial intelligence startup company, compared performance of GPT-4, GPT-4-Turbo, Claude2, and LLaMA-2 on two versions of a 150-question test about information in financial statements (e.g. Form 10-K, Form 10-Q, Form 8-K, earnings reports, earnings call transcripts) submitted by public companies to the U.S. Securities and Exchange Commission, where one version of the test required the generative AI models to use a retrieval system to find the specific SEC filing to answer the questions while the other provided the specific SEC filing to the models to answer the question (i.e. in a long context window). On the retrieval system version, GPT-4-Turbo and LLaMA-2 both failed to produce correct answers to 81% of the questions, while on the long context window version, GPT-4-Turbo and Claude-2 failed to produce correct answers to 21% and 24% of the questions, respectively.<ref>Leswing, Kif (December 19, 2023). "GPT and other AI models can't analyze an SEC filing, researchers find". CNBC. Archived from the original on December 19, 2023. Retrieved December 19, 2023.</ref><ref>"Patronus AI Launches Industry-first LLM Benchmark for Finance to Address Hallucinations" (Press release). PR Newswire. November 16, 2023. Archived from the original on December 19, 2023. Retrieved December 19, 2023.</ref>
Medicine
In the field of health care, possible uses and concerns are under scrutiny by professional associations and practitioners.<ref>The Lancet Digital Health (March 3, 2023). "ChatGPT: friend or foe?". The Lancet Digital Health. 5 (3): e102. doi:10.1016/S2589-7500(23)00023-7. PMID 36754723. S2CID 256659547.</ref><ref>Asch, David A. (April 4, 2023). "An Interview with ChatGPT About Health Care". NEJM Catalyst Innovations in Care Delivery. doi:10.1056/CAT.23.0043 (inactive January 31, 2024). Archived from the original on June 29, 2023. Retrieved June 29, 2023.{{cite journal}}
: CS1 maint: DOI inactive as of January 2024 (link)</ref> Two early papers indicated that ChatGPT could pass the United States Medical Licensing Examination (USMLE).<ref name="DePeau-Wilson-2023">DePeau-Wilson, Michael (January 19, 2023). "AI Passes U.S. Medical Licensing Exam". MedPage Today. Archived from the original on April 9, 2023. Retrieved May 2, 2023.</ref> MedPage Today noted in January 2023 that "researchers have published several papers now touting these AI programs as useful tools in medical education, research, and even clinical decision making."<ref name="DePeau-Wilson-2023" />
Published in February 2023 were two separate papers that again evaluated ChatGPT's proficiency in medicine using the USMLE. Findings were published in JMIR Medical Education (see Journal of Medical Internet Research) and PLOS Digital Health. The authors of the PLOS Digital Health paper stated that the results "suggest that large language models may have the potential to assist with medical education, and potentially, clinical decision-making."<ref>Kung, Tiffany H.; Cheatham, Morgan; Medenilla, Arielle; Sillos, Czarina; Leon, Lorie De; Elepaño, Camille; Madriaga, Maria; Aggabao, Rimel; Diaz-Candido, Giezel; Maningo, James; Tseng, Victor (February 9, 2023). "Performance of ChatGPT on USMLE: Potential for AI-assisted medical education using large language models". PLOS Digital Health. 2 (2): e0000198. doi:10.1371/journal.pdig.0000198. ISSN 2767-3170. PMC 9931230. PMID 36812645.</ref><ref>"Expert reaction to study on ChatGPT almost passing the US Medical Licensing Exam". Science Media Centre. February 9, 2023. Archived from the original on April 24, 2023. Retrieved May 2, 2023.</ref> In JMIR Medical Education, the authors of the other paper concluded that "ChatGPT performs at a level expected of a third-year medical student on the assessment of the primary competency of medical knowledge." They suggest that it could be used as an "interactive learning environment for students". The AI itself, prompted by the researchers, concluded that "this study suggests that ChatGPT has the potential to be used as a virtual medical tutor, but more research is needed to further assess its performance and usability in this context."<ref>Gilson, Aidan; Safranek, Conrad W.; Huang, Thomas; Socrates, Vimig; Chi, Ling; Taylor, Richard Andrew; Chartash, David (February 8, 2023). "How Does ChatGPT Perform on the United States Medical Licensing Examination? The Implications of Large Language Models for Medical Education and Knowledge Assessment". JMIR Medical Education. 9 (1): e45312. doi:10.2196/45312. PMC 9947764. PMID 36753318.</ref> The later-released ChatGPT version based on GPT-4 significantly outperformed the version based on GPT-3.5.<ref>Brueck, Hilary. "The newest version of ChatGPT passed the US medical licensing exam with flying colors — and diagnosed a 1 in 100,000 condition in seconds". Business Insider. Archived from the original on January 27, 2024. Retrieved February 2, 2024.</ref> Researchers at Stanford University and the University of California, Berkeley have found that the performance of GPT-3.5 and GPT-4 on the USMLE declined from March 2023 to June 2023.<ref name="Boston Globe 11-8-2023" /><ref name="arXiv 2307.09009v3" />
A March 2023 paper tested ChatGPT's application in clinical toxicology. The authors found that the AI "fared well" in answering a "very straightforward [clinical case example], unlikely to be missed by any practitioner in the field". They added: "As ChatGPT becomes further developed and specifically adapted for medicine, it could one day be useful in less common clinical cases (i.e, cases that experts sometimes miss). Rather than AI replacing humans (clinicians), we see it as 'clinicians using AI' replacing 'clinicians who do not use AI' in the coming years."<ref>Abdel-Messih, Mary Sabry; Boulos, Maged N. Kamel (March 8, 2023). "ChatGPT in Clinical Toxicology". JMIR Medical Education. 9 (1): e46876. doi:10.2196/46876. PMC 10034604. PMID 36867743.</ref>
An April 2023 study in Radiology tested the AI's ability to answer queries about breast cancer screening. The authors found that it answered appropriately "about 88 percent of the time", however, in one case (for example), it gave advice that had become outdated about a year earlier. The comprehensiveness of its answers was also lacking.<ref>Haver, Hana L; Ambinder, Emily B; Bahl, Manisha; Oluyemi, Eniola T; Jeudy, Jean; Yi, Paul H (April 4, 2023). "Appropriateness of Breast Cancer Prevention and Screening Recommendations Provided by ChatGPT". Radiology. 307 (4): 230424. doi:10.1148/radiol.230424. ISSN 0033-8419. PMID 37014239. S2CID 257923990. Archived from the original on May 5, 2023. Retrieved May 5, 2023.</ref><ref>Kotz, Deborah (April 4, 2023). "UM School of Medicine Study Finds ChatGPT Helpful for Breast Cancer Screening Advice, With Certain Caveats". University of Maryland School of Medicine. Archived from the original on May 5, 2023. Retrieved May 5, 2023.</ref> A study published in JAMA Internal Medicine that same month found that ChatGPT often outperformed human doctors at answering patient questions (when measured against questions and answers found at /r/AskDocs, a forum on Reddit where moderators validate the medical credentials of professionals; the study acknowledges the source as a limitation).<ref>Ayers, John W.; Poliak, Adam; Dredze, Mark; Leas, Eric C.; Zhu, Zechariah; Kelley, Jessica B.; Faix, Dennis J.; Goodman, Aaron M.; Longhurst, Christopher A.; Hogarth, Michael; Smith, Davey M. (April 28, 2023). "Comparing Physician and Artificial Intelligence Chatbot Responses to Patient Questions Posted to a Public Social Media Forum". JAMA Internal Medicine. 183 (6): 589–596. doi:10.1001/jamainternmed.2023.1838. ISSN 2168-6106. PMC 10148230. PMID 37115527. Archived from the original on April 30, 2023. Retrieved May 2, 2023.{{cite journal}}
: CS1 maint: PMC embargo expired (link)</ref><ref>Fox, Andrea (May 4, 2023). "Does ChatGPT really outshine doctors? Or just on social media?". Healthcare IT News. HIMSS Media. Archived from the original on May 4, 2023. Retrieved May 5, 2023.</ref><ref>"The doctor is out, but it's OK. ChatGPT can answer your questions". Hub. Johns Hopkins University. April 28, 2023. Archived from the original on May 5, 2023. Retrieved May 5, 2023.</ref> The study authors suggest that the tool could be integrated with medical systems to help doctors draft responses to patient questions.<ref>Ono, Mika (April 28, 2023). "Study Finds ChatGPT Outperforms Physicians in High-Quality, Empathetic Answers to Patient Questions". UC San Diego Today. Archived from the original on April 28, 2023. Retrieved April 28, 2023.</ref><ref>"ChatGPT Beats Doctors in Compassion and Quality of Advice to Patients". Neuroscience News. April 28, 2023. Archived from the original on May 3, 2023. Retrieved May 2, 2023.</ref>
Professionals have emphasized ChatGPT's limitations in providing medical assistance. In correspondence to The Lancet Infectious Diseases, three antimicrobial experts wrote that "the largest barriers to the implementation of ChatGPT in clinical practice are deficits in situational awareness, inference, and consistency. These shortcomings could endanger patient safety."<ref>Howard, Alex; Hope, William; Gerada, Alessandro (April 2023). "ChatGPT and antimicrobial advice: the end of the consulting infection doctor?". The Lancet Infectious Diseases. 23 (4): 405–406. doi:10.1016/s1473-3099(23)00113-5. ISSN 1473-3099. PMID 36822213. S2CID 257072872. Archived from the original on March 25, 2023. Retrieved May 2, 2023.</ref> Physician's Weekly, though also discussing the potential use of ChatGPT in medical contexts (e.g. "as a digital assistant to physicians by performing various administrative functions like gathering patient record information or categorizing patient data by family history, symptoms, lab results, possible allergies, et cetera"), warned that the AI might sometimes provide fabricated or biased information.<ref>"Is There a Role for ChatGPT in Healthcare?". Physician's Weekly. April 27, 2023. Archived from the original on May 5, 2023. Retrieved May 5, 2023.</ref> One radiologist warned: "We've seen in our experience that ChatGPT sometimes makes up fake journal articles or health consortiums to support its claims";<ref>Drake, Kimberly (April 6, 2023). "Rely on a Doctor, Not ChatGPT, for Medical Advice". HealthNews. Archived from the original on May 5, 2023. Retrieved May 5, 2023.</ref> As reported in one Mayo Clinic Proceedings: Digital Health paper, ChatGPT may do this for as much as 69% of its cited medical references. The researchers emphasized that while many of its references were fabricated, those that were appeared "deceptively real".<ref>Gravel, Jocelyn; D’Amours-Gravel, Madeleine; Osmanlliu, Esli (September 1, 2023). "Learning to Fake It: Limited Responses and Fabricated References Provided by ChatGPT for Medical Questions". Mayo Clinic Proceedings: Digital Health. 1 (3): 226–234. doi:10.1016/j.mcpdig.2023.05.004. ISSN 2949-7612.</ref> As Dr. Stephen Hughes mentioned for The Conversation however, ChatGPT is capable of learning to correct its past mistakes. He also noted the AI's "prudishness" regarding sexual health topics.<ref>Hughes, Stephen (April 27, 2023). "How good is ChatGPT at diagnosing disease? A doctor puts it through its paces". The Conversation. Archived from the original on May 4, 2023. Retrieved May 5, 2023.</ref>
Contrary to previous findings, ChatGPT responses to anesthesia-related questions were more accurate, succinct, and descriptive compared to Bard's. Bard exhibited 30.3% error in response as compared to ChatGPT (0% error).<ref>Patnaik, Sourav S.; Hoffmann, Ulrike (November 7, 2023). "Quantitative evaluation of ChatGPT versus Bard responses to anaesthesia-related queries". British Journal of Anaesthesia. 132 (1): S0007–0912(23)00550–0. doi:10.1016/j.bja.2023.09.030. ISSN 1471-6771. PMID 37945414. S2CID 265078930. Archived from the original on February 18, 2024. Retrieved November 28, 2023.</ref> At a conference of the American Society of Health-System Pharmacists in December 2023, researchers at Long Island University (LIU) presented a study that researched ChatGPT's responses to 45 frequently asked questions of LIU College of Pharmacy's drug information service during a 16-month period from 2022 to 2023 as compared with researched responses provided by professional pharmacists. For 29 of the 39 questions for which there was sufficient medical literature for a data-driven response, ChatGPT failed to provide a direct answer or provided a wrong or incomplete answer (and in some cases, if acted upon, the answer would endanger the patient's health). The researchers had asked ChatGPT to provide medical research citations for all its answers, but it did so for only eight, and all eight included at least one fabricated (fake) citation.<ref>Constantino, Annika Kim (December 5, 2023). "Free ChatGPT may incorrectly answer drug questions, study says". CNBC. Archived from the original on December 5, 2023. Retrieved December 5, 2023.</ref><ref>"Study Finds ChatGPT Provides Inaccurate Responses to Drug Questions" (Press release). PR Newswire. December 5, 2023. Archived from the original on December 5, 2023. Retrieved December 5, 2023.</ref>
A January 2024 study conducted by researchers at Cohen Children's Medical Center found that GPT-4 had an accuracy rate of 17% when diagnosing pediatric medical cases.<ref>Barile, Joseph; Margolis, Alex; Cason, Grace; Kim, Rachel; Kalash, Saia; Tchaconas, Alexis; Milanaik, Ruth (January 2, 2024). "Diagnostic Accuracy of a Large Language Model in Pediatric Case Studies". JAMA Pediatrics. 178 (3): 313–315. doi:10.1001/jamapediatrics.2023.5750. ISSN 2168-6203. PMC 10762631. PMID 38165685. Archived from the original on February 18, 2024. Retrieved February 18, 2024.</ref><ref>Mole, Beth (January 3, 2024). "ChatGPT bombs test on diagnosing kids' medical cases with 83% error rate". Ars Technica. Archived from the original on January 17, 2024. Retrieved January 5, 2024.</ref>
Law
In January 2023, Massachusetts State Senator Barry Finegold and State Representative Josh S. Cutler proposed a bill partially written by ChatGPT, "An Act drafted with the help of ChatGPT to regulate generative artificial intelligence models like ChatGPT",<ref name=":0">"Bill S.31". malegislature.gov. Archived from the original on December 7, 2023. Retrieved December 7, 2023.</ref><ref name=":1">Annear, Steve (January 24, 2023). "Two elected officials drafted legislation to regulate artificial intelligence technology — with some help from ChatGPT". The Boston Globe. Archived from the original on December 7, 2023. Retrieved December 7, 2023.</ref><ref name=":2">Garrity, Kelly; Kashinsky, Lisa (July 13, 2023). "ChatGPT enters the legislative chat". POLITICO. Archived from the original on December 7, 2023. Retrieved December 7, 2023.</ref> which would require companies to disclose their algorithms and data collection practices to the office of the State Attorney General, arrange regular risk assessments, and contribute to the prevention of plagiarism.<ref name=":1" /><ref name=":2" /><ref name=":3">Quach, Katyanna (December 2, 2023). "Local council in Brazil passes ChatGPT-written proposal". The Register. Archived from the original on December 7, 2023. Retrieved December 7, 2023.</ref> The bill was officially presented during a hearing on July 13.<ref name=":0" /><ref name=":2" />
On April 11, 2023, a judge of a session court in Pakistan used ChatGPT to decide the bail of a 13-year-old accused in a matter. The court quoted the use of ChatGPT assistance in its verdict:
Can a juvenile suspect in Pakistan, who is 13 years old, be granted bail after arrest?
The AI language model replied:
Under the Juvenile Justice System Act 2018, according to section 12, the court can grant bail on certain conditions. However, it is up to the court to decide whether or not a 13-year-old suspect will be granted bail after arrest.
The judge asked ChatGPT other questions about the case and formulated his final decision in light of its answers.<ref>"Pakistani judge uses ChatGPT to make court decision". Gulf News. April 13, 2023. Archived from the original on April 20, 2023. Retrieved April 20, 2023.</ref><ref>"AI revolution is here': Pakistani court takes help from ChatGPT to grant bail in rape case". Pakistan Observer. April 11, 2023. Archived from the original on April 20, 2023. Retrieved April 20, 2023.</ref>
In Mata v. Avianca, Inc., 22-cv-1461 (PKC), a personal injury lawsuit against Avianca Airlines filed in the Southern New York U.S. District Court in May 2023 (with Senior Judge P. Kevin Castel presiding), the plaintiff's attorneys reportedly used ChatGPT to generate a legal motion. ChatGPT generated numerous fictitious legal cases involving fictitious airlines with fabricated quotations and internal citations in the legal motion. Castel noted numerous inconsistencies in the opinion summaries, and called one of the cases' legal analysis "gibberish".<ref>Brodkin, Jon (June 23, 2023). "Lawyers have real bad day in court after citing fake cases made up by ChatGPT". Ars Technica. Archived from the original on January 26, 2024. Retrieved February 18, 2024.</ref> The plaintiff's attorneys faced potential judicial sanction and disbarment for filing the motion and presenting the fictitious legal decisions ChatGPT generated as authentic.<ref>Goswami, Rohan (May 30, 2023). "ChatGPT cited 'bogus' cases for a New York federal court filing. The attorneys involved may face sanctions". CNBC. Archived from the original on May 30, 2023. Retrieved May 30, 2023.</ref><ref>Neumeister, Larry (June 8, 2023). "Lawyers blame ChatGPT for tricking them into citing bogus case law". Associated Press. Archived from the original on November 8, 2023. Retrieved November 8, 2023.</ref> The case was dismissed and the attorneys were fined $5,000.<ref>"Mata v. Avianca, Inc". Casetext. Archived from the original on February 13, 2024. Retrieved February 13, 2024.</ref><ref>"'Use with caution': How ChatGPT landed this US lawyer and his firm in hot water". ABC News. June 24, 2023. Archived from the original on November 9, 2023. Retrieved November 9, 2023.</ref>
In October 2023, the council of Porto Alegre, Brazil, unanimously approved a local ordinance proposed by councilman Ramiro Rosário that would exempt residents from needing to pay for the replacement of stolen water consumption meters; the bill went into effect on November 23. On November 29, Rosário revealed that the bill had been entirely written by ChatGPT, and that he had presented it to the rest of the council without making any changes or disclosing the chatbot's involvement.<ref name=":3" /><ref name=":4">Jeantet, Diane; Savarese, Mauricio; LeBlanc, Steve; O'Brien, Matt (November 30, 2023). "Brazilian city enacts an ordinance that was secretly written by ChatGPT". AP News. Associated Press. Archived from the original on December 7, 2023. Retrieved December 7, 2023.</ref><ref name=":5">Paúl, María Luisa (December 4, 2023). "A Brazilian city passed a law about water meters. ChatGPT wrote it". Washington Post. ISSN 0190-8286. Archived from the original on December 5, 2023. Retrieved December 7, 2023.</ref> The city's council president, Hamilton Sossmeier, initially criticized Rosário's initiative, saying it could represent "a dangerous precedent",<ref name=":5" /><ref>"Lei escrita por inteligência artificial é aprovada por vereadores em Porto Alegre; 'precedente perigoso', diz presidente da Câmara". G1 (in português do Brasil). November 29, 2023. Archived from the original on December 7, 2023. Retrieved December 7, 2023.</ref> but later said he "changed his mind": "unfortunately or fortunately, this is going to be a trend."<ref name=":3" /><ref name=":4" />
See also
Lua error in mw.title.lua at line 318: bad argument #2 to 'title.new' (unrecognized namespace name 'Portal').
- Tay (chatbot)
- Poe (chatbot)
- Ernie Bot
- Intelligent agent
- Virtual assistant
- Ethics of artificial intelligence
- Turing test
Notes
References
Further reading
- Biswas, Som (April 1, 2023). "ChatGPT and the Future of Medical Writing". Radiology. 307 (2): e223312. doi:10.1148/radiol.223312. ISSN 0033-8419. PMID 36728748. S2CID 256501098.
- Chang, Kent K.; Cramer, Mackenzie; Soni, Sandeep; Bamman, David (April 28, 2023). "Speak, Memory: An Archaeology of Books Known to ChatGPT/GPT-4". arXiv:2305.00118 [cs.CL].
- Template:Cite SSRN
- Template:Cite podcast
- Ouyang, Long; et al. (March 4, 2022). "Training language models to follow instructions with human feedback". arXiv:2203.02155 [cs.CL].
- Liebrenz, Michael; Schleifer, Roman; Buadze, Anna; Bhugra, Dinesh; Smith, Alexander (February 2023). "Generating scholarly content with ChatGPT: ethical challenges for medical publishing". The Lancet Digital Health. 5 (3): e105–e106. doi:10.1016/s2589-7500(23)00019-5. ISSN 2589-7500. PMID 36754725. S2CID 256655912.
- Wolfram, Stephen (February 14, 2023). "What Is ChatGPT Doing … and Why Does It Work?".
- Wolfram, Stephen (March 23, 2023). "ChatGPT Gets Its "Wolfram Superpowers"!".
- Bartholomew, Jem; Mehta, Dhrumil. "How the media is covering ChatGPT". Columbia Journalism Review. Retrieved May 30, 2023.
- Zhao, Wayne Xin; et al. (2023). "A Survey of Large Language Models". arXiv:2303.18223 [cs.CL].
- Prompt engineering guide from OpenAI
External links
- Media related to Lua error in Module:Commons_link at line 63: attempt to index field 'wikibase' (a nil value). at Wikimedia Commons
- Lua error in Module:Official_website at line 90: attempt to index field 'wikibase' (a nil value).
- ChatGPT Prompt Engineering for Developers, course by Andrew Ng and OpenAI
- Can ChatGPT write a podcast episode? (Planet Money podcast, May 2023)
- Will Chat GPT do more harm than good? (Gary Marcus and Keith Teare debate, February 2023)
- "What OpenAI Really Wants" by Wired
Template:OpenAI navbox Lua error in Module:Authority_control at line 181: attempt to index field 'wikibase' (a nil value).