Attorney General Phil Weiser warns voters on AI-created “deepfakes,” issues public advisory
Sept. 9, 2024 (DENVER) – Attorney General Phil Weiser today warned voters to be on the lookout for election misinformation and disinformation in the form of realistic-looking images, videos, and audio created using artificial intelligence, known as “deepfakes.” Weiser accompanied his warning with a public advisory educating Coloradans about a new law designed to prevent political campaigns from using AI deepfakes without properly disclosing content origins to voters.
This year, lawmakers passed and Gov. Polis signed into law HB24-1147. The new law requires anyone using AI to create communications to voters featuring images, videos, or audio of candidates for office to include a disclaimer explaining that the content is not real. Failure to provide such a disclaimer can result in fines and other penalties.
“Because images, videos, and audio created with artificial intelligence are becoming difficult to distinguish from the real thing, you should be cautious when forming opinions based on what you see and hear online, on TV, and receive in the mail,” said Weiser. “The sad reality is that even AI-powered tools designed to detect these deepfakes have difficulty catching them. I encourage voters to do your research, get your news and information from trusted sources, and be mindful that the sophistication of AI means you can’t always believe what you see and hear anymore.”
The new law applies to communications to voters within 60 days of a primary election and 90 days of a general election. Since it is less than 90 days from the November 5 general election, the protections created by the law are currently in effect.
In the public advisory Weiser issued, he lays out what voters, candidates, and campaigns need to know about the new law:
- Any visual or audio communication regarding candidates for office using deepfake images, audio, video, or multimedia are prohibited unless properly disclosed.
- The required disclosures must be clear and conspicuous. A disclaimer notifying voters that the content “has been edited and depicts speech or conduct that falsely appears to be authentic or truthful” must be displayed or otherwise appear in the communication, and the law provides for exact font sizes and other requirements.
- Exceptions to the law include protections for outlets that discuss deepfake material in news stories, so long as the broadcast makes clear the content includes a deepfake. Additionally, radio and television broadcast stations are exempt if they run political advertisements that contain deepfakes that lack proper disclaimers. The law also exempts satires and parodies.
- Violations can result in legal action to prevent dissemination of the deepfake in question, and violators could be subject to financial liabilities or even criminal penalties.
###