FTC to enforce law to curb risks of artificial intelligence
Technology
The sudden popularity of ChatGPT this year has prompted global calls for regulation
WASHINGTON (Reuters) - The Federal Trade Commission's chief said the agency was committed to using existing laws to rein in some of the dangers of artificial intelligence, such as enhancing the power of dominant firms and "turbocharging" fraud.
"Although these (AI) tools are novel, they are not exempt from existing rules, and the FTC will vigorously enforce the laws we are charged with administering, even in this new market," FTC Chair Lina Khan wrote in an opinion piece in the New York Times on Wednesday.
The sudden popularity of Microsoft-backed (MSFT.O) OpenAI's ChatGPT this year has prompted global calls for regulation amid concerns about its possible use for wrongdoing even as companies seek to use it to enhance efficiency. read more
She described the agency as "well equipped" to handle the job.
One risk she noted was that firms that dominate cloud services and computing would become even more powerful as they help startups and other firms launch their own AI. AI tools could also be used to facilitate collusion to raise prices.
Khan expressed concern that generative AI, which writes in conversational English, could be used to help scammers write more specific and effective phishing emails.
"When enforcing the law's prohibition on deceptive practices, we will look not just at the fly-by-night scammers deploying these tools but also at the upstream firms that are enabling them," she wrote.