AI regulations

Microsoft Urges Congress to Regulate AI-Generated Deepfakes

Microsoft calls for new laws against AI-generated deepfake fraud.

Microsoft Calls for Regulations on AI-Generated Deepfakes

In a recent statement, Microsoft has urged Congress to take significant steps to regulate the use of AI-generated deepfakes, highlighting the necessity of safeguarding against fraud, abuse, and manipulation. Brad Smith, the vice chair and president of Microsoft, emphasized the urgent need for policymakers to act to protect vulnerable populations, including seniors from fraud and children from abuse.

The Need for Legislative Action

"While the tech sector and non-profit groups have taken recent steps to address this problem, it has become apparent that our laws will also need to evolve to combat deepfake fraud," Smith stated in a blog post. He noted that one of the most critical actions the US could undertake is to establish a comprehensive deepfake fraud statute, which would empower law enforcement to prosecute cybercriminal activities utilizing this technology to exploit innocent individuals.

Proposed Deepfake Fraud Statute

Microsoft is advocating for a specific legal framework aimed at deepfake fraud, which would give authorities the ability to target and combat AI-generated scams effectively. Further, Smith urged legislators to modernize existing federal and state laws surrounding child sexual exploitation and abuse, ensuring they account for AI-generated content.

Legislative Developments

Recently, the Senate took a noteworthy step by passing a bill aimed at addressing sexually explicit deepfakes. This legislation allows victims of nonconsensual AI-generated explicit content to initiate lawsuits against the individuals responsible for creating these damaging materials. This bill was prompted by alarming incidents involving middle and high school students who fabricated explicit images of their female peers, as well as instances of graphic deepfake content circulating on social media.

Microsoft’s Commitment to Safety

Following incidents where Microsoft’s own AI tools were misused, the company has strengthened its safety controls. A notable example was a loophole in its Designer AI image creator that enabled users to produce explicit images of celebrities, including Taylor Swift. Smith asserted, "The private sector has a responsibility to innovate and implement safeguards that prevent the misuse of AI." This responsibility aligns with the broader call for establishing robust regulations and standards.

The Challenge of Generative AI

The Federal Communications Commission (FCC) has already implemented measures against robocalls utilizing AI-generated voices. However, generative AI technologies pose a broader challenge, facilitating the creation of convincing fake audio, images, and videos. As the 2024 presidential election approaches, the risk of misinformation through deepfake content rises. Recently, Elon Musk shared a deepfake video parodying Vice President Kamala Harris on social media, which could potentially breach X’s policies regarding synthetic and manipulated media.

A Call for Transparent Labelling

Smith emphasized the importance of transparency in combating misinformation: "Congress should require AI system providers to use state-of-the-art provenance tooling to label synthetic content." This initiative is pivotal in fostering trust within the information ecosystem and aiding the public in discerning AI-generated or manipulated content from genuine material.

Conclusion

As deepfake technology continues to evolve and pose significant challenges, the call for comprehensive regulations becomes increasingly critical. By implementing laws that address the misuse of AI-generated content, protecting vulnerable populations, and ensuring transparency, Congress can help secure a safer digital environment for all.

Czytaj dalej

OnePlus Pad 2 tablet showcasing its sleek design and features.
Delta Airlines logo representing compensation claim against CrowdStrike and Microsoft

Zostaw komentarz

Wszystkie komentarze są moderowane przed opublikowaniem.

Ta strona jest chroniona przez hCaptcha i obowiązują na niej Polityka prywatności i Warunki korzystania z usługi serwisu hCaptcha.