Connect with us

Hi, what are you looking for?

The Independent TradersThe Independent Traders

World News

The FTC is investigating whether ChatGPT harms consumers

The Federal Trade Commission has opened an expansive investigation into OpenAI, probing whether the maker of the popular ChatGPT bot has run afoul of consumer protection laws by putting personal reputations and data at risk.

The agency this week sent the San Francisco company a 20-page demand for records about how it addresses risks related to its AI models, according to a document reviewed by The Washington Post. The salvo represents the most potent regulatory threat to date to OpenAI’s business in the United States, as the company goes on a global charm offensive to shape the future of artificial intelligence policy.

Analysts have called OpenAI’s ChatGPT the fastest-growing consumer app in history, and its early success set off an arms race among Silicon Valley companies to roll out competing chatbots. The company’s chief executive, Sam Altman, has emerged as an influential figure in the debate over AI regulation, testifying on Capitol Hill, dining with lawmakers and meeting with President Biden and Vice President Harris.

But now the company faces a new test in Washington, where the FTC has issued multiple warnings that existing consumer protection laws apply to AI, even as the administration and Congress struggle to outline new regulations. Senate Majority Leader Charles E. Schumer (D-N.Y.) has predicted that new AI legislation is months away.

The FTC’s demands of OpenAI are the first indication of how it intends to enforce those warnings. If the FTC finds that a company violates consumer protection laws, it can levy fines or put a business under a consent decree, which can dictate how the company handles data. The FTC has emerged as the federal government’s top Silicon Valley cop, bringing large fines against Meta, Amazon and Twitter for alleged violations of consumer protection laws.

The FTC called on OpenAI to provide detailed descriptions of all complaints it had received of its products making “false, misleading, disparaging or harmful” statements about people. The FTC is investigating whether the company engaged in unfair or deceptive practices that resulted in “reputational harm” to consumers, according to the document.

The FTC also asked the company to provide records related to a security incident that the company disclosed in March when a bug in its systems allowed some users to see payment-related information, as well as some data from other users’ chat history. The FTC is probing whether the company’s data security practices violate consumer protection laws. OpenAI said in a blog post that the number of users whose data was revealed to someone else was “extremely low.”

OpenAI and the FTC did not immediately respond to requests for comment sent on Thursday morning.

News of the probe comes as FTC Chair Lina Khan is likely to face a combative hearing Thursday before the House Judiciary Committee, where Republican lawmakers are expected to analyze her enforcement record and accuse her of mismanaging the agency. Khan’s ambitious plans to rein in Silicon Valley have suffered key losses in court. On Tuesday, a federal judge rejected the FTC’s attempt to block Microsoft’s $69 billion deal to buy the video game company Activision.

The agency has repeatedly warned that action is coming on AI, in speeches, blog posts, op-eds and news conferences. In a speech at Harvard Law School in April, Samuel Levine, the director of the agency’s Bureau of Consumer Protection, said the agency was prepared to be “nimble” in getting ahead of emerging threats.

“The FTC welcomes innovation, but being innovative is not a license to be reckless,” Levine said. “We are prepared to use all our tools, including enforcement, to challenge harmful practices in this area.”

The FTC also has issued several colorful blog posts about its approach to regulating AI, at times invoking popular science fiction movies to warn the industry against running afoul of the law. The agency has warned against AI scams, using generative AI to manipulate potential customers and falsely exaggerating the capabilities of AI products. Khan also participated in a news conference with Biden administration officials in April about the risk of AI discrimination.

“There is no AI exemption to the laws on the books,” Khan said at that event.

Among the information the FTC is seeking from Open AI is any research, testing or surveys that assess how well consumers understand “the accuracy or reliability of outputs” generated by its AI tools. The agency made extensive demands about records related to ways OpenAI’s products could generate disparaging statements, asking the company to provide records of the complaints people send about its chatbot making false statements.

The agency’s focus on such fabrications comes after numerous high-profile reports of the chatbot producing incorrect information that could damage people’s reputations. Mark Walters, a radio talk show host in Georgia sued OpenAI for defamation, alleging the chabot made up legal claims against him. The lawsuit alleges that ChatGPT falsely claimed that Walters, the host of “Armed American Radio,” was accused of defrauding and embezzling funds from the Second Amendment Foundation. The response was provided in response to a question about a lawsuit about the foundation that Walters is not a party to, according to the complaint.

ChatGPT also said that a lawyer had made sexually suggestive comments and attempted to touch a student on a class trip to Alaska, citing an article that it said had appeared in The Washington Post article. But no such article existed, the class trip never happened and the lawyer said he was never accused of harassing a student, The Post reported previously.

The FTC in its request also asked the company to provide extensive details about its products and the way it advertises them. It also demanded details about the policies and procedures that OpenAI takes before it releases any new product to the public, including a list of times that OpenAI held back a large language model because of safety risks.

The agency also demanded a detailed description of the data that OpenAI uses to train its products, which mimic humanlike speech by ingesting text, mostly scraped from Wikipedia, Scribd and other sites across the open web. The agency also asked OpenAI to describe how it refines its models to address their tendency to “hallucinate,” making up answers when the models don’t know the answer to a question.

OpenAI also has to turn over details about how many people were affected by the March security incident and information about all the steps it took to respond.

The FTC’s records request, which is called a Civil Investigative Demand, primarily focuses on potential consumer protection abuses, but it also asks OpenAI to provide some details about how it licenses its models to other companies.

The United States has trailed other governments in drafting AI legislation and regulating the privacy risks associated with the technology. Countries within the European Union have taken steps to limit U.S. companies’ chatbots under the bloc’s privacy law, the General Data Protection Regulation. Italy temporarily blocked ChatGPT from operating there due to data privacy concerns, and Google had to postpone the launch of its chatbot Bard after receiving requests for privacy assessments from the Irish Data Protection Commission. The European Union is also expected to pass AI legislation by the end of the year.

There is a flurry of activity in Washington to catch up. On Tuesday, Schumer hosted an all-senator briefing with officials from the Pentagon and intelligence community to discuss the national security risks of artificial intelligence, as he works with a bipartisan group of senators to craft new AI legislation. Schumer told reporters after the session that it’s going to be “very hard” to regulate AI, as lawmakers try to balance the need for innovation with ensuring there are proper safeguards on the technology.

On Wednesday, Vice President Harris hosted a group of consumer protection advocates and civil liberties leaders at the White House for a discussion on the safety and security risks of AI.

“It is a false choice to suggest that we either can advance innovation or we protect consumers,” Harris said. “We can do both.”

Will Oremus contributed to this report.

This post appeared first on The Washington Post

The Federal Trade Commission has opened an expansive investigation into OpenAI, probing whether the maker of the popular ChatGPT bot has run afoul of consumer protection laws by putting personal reputations and data at risk.

The agency this week sent the San Francisco company a 20-page demand for records about how it addresses risks related to its AI models, according to a document reviewed by The Washington Post. The salvo represents the most potent regulatory threat to date to OpenAI’s business in the United States, as the company goes on a global charm offensive to shape the future of artificial intelligence policy.

Analysts have called OpenAI’s ChatGPT the fastest-growing consumer app in history, and its early success set off an arms race among Silicon Valley companies to roll out competing chatbots. The company’s chief executive, Sam Altman, has emerged as an influential figure in the debate over AI regulation, testifying on Capitol Hill, dining with lawmakers and meeting with President Biden and Vice President Harris.

But now the company faces a new test in Washington, where the FTC has issued multiple warnings that existing consumer protection laws apply to AI, even as the administration and Congress struggle to outline new regulations. Senate Majority Leader Charles E. Schumer (D-N.Y.) has predicted that new AI legislation is months away.

The FTC’s demands of OpenAI are the first indication of how it intends to enforce those warnings. If the FTC finds that a company violates consumer protection laws, it can levy fines or put a business under a consent decree, which can dictate how the company handles data. The FTC has emerged as the federal government’s top Silicon Valley cop, bringing large fines against Meta, Amazon and Twitter for alleged violations of consumer protection laws.

The FTC called on OpenAI to provide detailed descriptions of all complaints it had received of its products making “false, misleading, disparaging or harmful” statements about people. The FTC is investigating whether the company engaged in unfair or deceptive practices that resulted in “reputational harm” to consumers, according to the document.

The FTC also asked the company to provide records related to a security incident that the company disclosed in March when a bug in its systems allowed some users to see payment-related information, as well as some data from other users’ chat history. The FTC is probing whether the company’s data security practices violate consumer protection laws. OpenAI said in a blog post that the number of users whose data was revealed to someone else was “extremely low.”

OpenAI and the FTC did not immediately respond to requests for comment sent on Thursday morning.

News of the probe comes as FTC Chair Lina Khan is likely to face a combative hearing Thursday before the House Judiciary Committee, where Republican lawmakers are expected to analyze her enforcement record and accuse her of mismanaging the agency. Khan’s ambitious plans to rein in Silicon Valley have suffered key losses in court. On Tuesday, a federal judge rejected the FTC’s attempt to block Microsoft’s $69 billion deal to buy the video game company Activision.

The agency has repeatedly warned that action is coming on AI, in speeches, blog posts, op-eds and news conferences. In a speech at Harvard Law School in April, Samuel Levine, the director of the agency’s Bureau of Consumer Protection, said the agency was prepared to be “nimble” in getting ahead of emerging threats.

“The FTC welcomes innovation, but being innovative is not a license to be reckless,” Levine said. “We are prepared to use all our tools, including enforcement, to challenge harmful practices in this area.”

The FTC also has issued several colorful blog posts about its approach to regulating AI, at times invoking popular science fiction movies to warn the industry against running afoul of the law. The agency has warned against AI scams, using generative AI to manipulate potential customers and falsely exaggerating the capabilities of AI products. Khan also participated in a news conference with Biden administration officials in April about the risk of AI discrimination.

“There is no AI exemption to the laws on the books,” Khan said at that event.

Among the information the FTC is seeking from Open AI is any research, testing or surveys that assess how well consumers understand “the accuracy or reliability of outputs” generated by its AI tools. The agency made extensive demands about records related to ways OpenAI’s products could generate disparaging statements, asking the company to provide records of the complaints people send about its chatbot making false statements.

The agency’s focus on such fabrications comes after numerous high-profile reports of the chatbot producing incorrect information that could damage people’s reputations. Mark Walters, a radio talk show host in Georgia sued OpenAI for defamation, alleging the chabot made up legal claims against him. The lawsuit alleges that ChatGPT falsely claimed that Walters, the host of “Armed American Radio,” was accused of defrauding and embezzling funds from the Second Amendment Foundation. The response was provided in response to a question about a lawsuit about the foundation that Walters is not a party to, according to the complaint.

ChatGPT also said that a lawyer had made sexually suggestive comments and attempted to touch a student on a class trip to Alaska, citing an article that it said had appeared in The Washington Post article. But no such article existed, the class trip never happened and the lawyer said he was never accused of harassing a student, The Post reported previously.

The FTC in its request also asked the company to provide extensive details about its products and the way it advertises them. It also demanded details about the policies and procedures that OpenAI takes before it releases any new product to the public, including a list of times that OpenAI held back a large language model because of safety risks.

The agency also demanded a detailed description of the data that OpenAI uses to train its products, which mimic humanlike speech by ingesting text, mostly scraped from Wikipedia, Scribd and other sites across the open web. The agency also asked OpenAI to describe how it refines its models to address their tendency to “hallucinate,” making up answers when the models don’t know the answer to a question.

OpenAI also has to turn over details about how many people were affected by the March security incident and information about all the steps it took to respond.

The FTC’s records request, which is called a Civil Investigative Demand, primarily focuses on potential consumer protection abuses, but it also asks OpenAI to provide some details about how it licenses its models to other companies.

The United States has trailed other governments in drafting AI legislation and regulating the privacy risks associated with the technology. Countries within the European Union have taken steps to limit U.S. companies’ chatbots under the bloc’s privacy law, the General Data Protection Regulation. Italy temporarily blocked ChatGPT from operating there due to data privacy concerns, and Google had to postpone the launch of its chatbot Bard after receiving requests for privacy assessments from the Irish Data Protection Commission. The European Union is also expected to pass AI legislation by the end of the year.

There is a flurry of activity in Washington to catch up. On Tuesday, Schumer hosted an all-senator briefing with officials from the Pentagon and intelligence community to discuss the national security risks of artificial intelligence, as he works with a bipartisan group of senators to craft new AI legislation. Schumer told reporters after the session that it’s going to be “very hard” to regulate AI, as lawmakers try to balance the need for innovation with ensuring there are proper safeguards on the technology.

On Wednesday, Vice President Harris hosted a group of consumer protection advocates and civil liberties leaders at the White House for a discussion on the safety and security risks of AI.

“It is a false choice to suggest that we either can advance innovation or we protect consumers,” Harris said. “We can do both.”

Will Oremus contributed to this report.

This post appeared first on The Washington Post

 

You May Also Like

Tech News

Unity Earlier this week, Unity, the company that makes the Unity video game engine popular with indie developers, announced that it was changing its...

Tech News

Illustration: The Verge X CEO Linda Yaccarino announced a series of changes to her executive team, including a shakeup to the company’s sales organization...

Tech News

Image: Brazil Climate Summit At the moment I arrived at the Brazil Climate Summit event, it felt like home to me. As I opened...

Tech News

The Logitech G Pro X Superlight 2 mouse. | Photo by Sean Hollister / The Verge I called it the real magic mouse, but...