The Los Angeles Post
U.S. World Business Lifestyle
Today: April 10, 2025
Today: April 10, 2025

FTC probe of OpenAI: Consumer protection is the opening salvo of US AI regulation

July 18, 2023
The Los Angeles Post - The Conversation

The FTC probe of ChatGPT maker OpenAI aligns with concerns that members of Congress have expressed.

AP Photo/Michael Dwyer

The Federal Trade Commission has launched an investigation of ChatGPT maker OpenAI for potential violations of consumer protection laws. The FTC sent the company a 20-page demand for information in the week of July 10, 2023. The move comes as European regulators have begun to take action, and Congress is working on legislation to regulate the artificial intelligence industry.

FTC probe of OpenAI: Consumer protection is the opening salvo of US AI regulation

The FTC has asked OpenAI to provide details of all complaints the company has received from users regarding “false, misleading, disparaging, or harmful” statements put out by OpenAI, and whether OpenAI engaged in unfair or deceptive practices relating to risks of harm to consumers, including reputational harm. The agency has asked detailed questions about how OpenAI obtains its data, how it trains its models, the processes it uses for human feedback, risk assessment and mitigation, and its mechanisms for privacy protection.

As a researcher of social media and AI, I recognize the immensely transformative potential of generative AI models, but I believe that these systems pose risks. In particular, in the context of consumer protection, these models can produce errors, exhibit biases and violate personal data privacy.

Hidden power

At the heart of chatbots such as ChatGPT and image generation tools such as DALL-E lies the power of generative AI models that can create realistic content from text, images, audio and video inputs. These tools can be accessed through a browser or a smartphone app.

FTC probe of OpenAI: Consumer protection is the opening salvo of US AI regulation

Since these AI models have no predefined use, they can be fine-tuned for a wide range of applications in a variety of domains ranging from finance to biology. The models, trained on vast quantities of data, can be adapted for different tasks with little to no coding and sometimes as easily as by describing a task in simple language.

Given that AI models such as GPT-3 and GPT-4 were developed by private organizations using proprietary data sets, the public doesn’t know the nature of the data used to train them. The opacity of training data and the complexity of the model architecture – GPT-3 was trained on over 175 billion variables or “parameters” – make it difficult for anyone to audit these models. Consequently, it’s difficult to prove that the way they are built or trained causes harm.

Hallucinations

In language model AIs, a hallucination is a confident response that is inaccurate and seemingly not justified by a model’s training data. Even some generative AI models that were designed to be less prone to hallucinations have amplified them.

There is a danger that generative AI models can produce incorrect or misleading information that can end up being damaging to users. A study investigating ChatGPT’s ability to generate factually correct scientific writing in the medical field found that ChatGPT ended up either generating citations to nonexistent papers or reporting nonexistent results. My collaborators and I found similar patterns in our investigations.

FTC probe of OpenAI: Consumer protection is the opening salvo of US AI regulation

Such hallucinations can cause real damage when the models are used without adequate supervision. For example, ChatGPT falsely claimed that a professor it named had been accused of sexual harassment. And a radio host has filed a defamation lawsuit against OpenAI regarding ChatGPT falsely claiming that there was a legal complaint against him for embezzlement.

Bias and discrimination

Without adequate safeguards or protections, generative AI models trained on vast quantities of data collected from the internet can end up replicating existing societal biases. For example, organizations that use generative AI models to design recruiting campaigns could end up unintentionally discriminating against some groups of people.

When a journalist asked DALL-E 2 to generate images of “a technology journalist writing an article about a new AI system that can create remarkable and strange images,” it generated only pictures of men. An AI portrait app exhibited several sociocultural biases, for example by lightening the skin color of an actress.

Data privacy

Another major concern, especially pertinent to the FTC investigation, is the risk of privacy breaches where the AI may end up revealing sensitive or confidential information. A hacker could gain access to sensitive information about people whose data was used to train an AI model.

FTC probe of OpenAI: Consumer protection is the opening salvo of US AI regulation

Researchers have cautioned about risks from manipulations called prompt injection attacks, which can trick generative AI into giving out information that it shouldn’t. “Indirect prompt injection” attacks could trick AI models with steps such as sending someone a calendar invitation with instructions for their digital assistant to export the recipient’s data and send it to the hacker.

A man in a business suit stands with his right hand raised in a wood-paneled room.

OpenAI CEO Sam Altman testified before a Senate Judiciary subcommittee on May 16, 2023. AI regulation legislation is in the works, but the FTC beat Congress to the punch.

AP Photo/Patrick Semansky

Some solutions

The European Commission has published ethical guidelines for trustworthy AI that include an assessment checklist for six different aspects of AI systems: human agency and oversight; technical robustness and safety; privacy and data governance; transparency, diversity, nondiscrimination and fairness; societal and environmental well-being; and accountability.

FTC probe of OpenAI: Consumer protection is the opening salvo of US AI regulation

Better documentation of AI developers’ processes can help in highlighting potential harms. For example, researchers of algorithmic fairness have proposed model cards, which are similar to nutritional labels for food. Data statements and datasheets, which characterize data sets used to train AI models, would serve a similar role.

Amazon Web Services, for instance, introduced AI service cards that describe the uses and limitations of some models it provides. The cards describe the models’ capabilities, training data and intended uses.

The FTC’s inquiry hints that this type of disclosure may be a direction that U.S. regulators take. Also, if the FTC finds OpenAI has violated consumer protection laws, it could fine the company or put it under a consent decree.

The Conversation

Anjana Susarla receives funding from the National Institute of Health and the Omura-Saxena Professorship in Responsible AI

Share This

Popular

Business|Economy|Political|US

Treasury secretary on Trump’s backdown on tariffs

Treasury secretary on Trump’s backdown on tariffs
Economy|Political|US

Former Trump official describes how she thinks Trump’s decision-making led to pause on tariffs

Former Trump official describes how she thinks Trump’s decision-making led to pause on tariffs
Americas|Crime|Political|US|World

US-Russian dual national Ksenia Karelina is released in prisoner swap between Moscow and Washington

US-Russian dual national Ksenia Karelina is released in prisoner swap between Moscow and Washington
Business|Economy|Europe|Political|US|World

EU chief welcomes Trump's tariff pause but is quiet about the bloc's own retaliatory duty plan

EU chief welcomes Trump's tariff pause but is quiet about the bloc's own retaliatory duty plan

Political

Asia|Business|Economy|Political|US|World

China reaches out to others as Trump layers on tariffs

China reaches out to others as Trump layers on tariffs
Environment|Europe|Health|Political|World

Pope Francis meets privately with King Charles and Queen Camilla during his Vatican convalescence

Pope Francis meets privately with King Charles and Queen Camilla during his Vatican convalescence
Crime|Europe|Political|World

Russian drone attack injures 12 in two Ukrainian cities, officials say

Russian drone attack injures 12 in two Ukrainian cities, officials say
Australia|Political|World

New Zealand lawmakers reject proposed law to redefine the country's founding Treaty of Waitangi

New Zealand lawmakers reject proposed law to redefine the country's founding Treaty of Waitangi

Access this article for free.

Already have an account? Sign In