Microsoft Calls for Regulations on AI-Generated Deepfakes
In a recent statement, Microsoft has urged Congress to take significant steps to regulate the use of AI-generated deepfakes, highlighting the necessity of safeguarding against fraud, abuse, and manipulation. Brad Smith, the vice chair and president of Microsoft, emphasized the urgent need for policymakers to act to protect vulnerable populations, including seniors from fraud and children from abuse.
The Need for Legislative Action
"While the tech sector and non-profit groups have taken recent steps to address this problem, it has become apparent that our laws will also need to evolve to combat deepfake fraud," Smith stated in a blog post. He noted that one of the most critical actions the US could undertake is to establish a comprehensive deepfake fraud statute, which would empower law enforcement to prosecute cybercriminal activities utilizing this technology to exploit innocent individuals.
Proposed Deepfake Fraud Statute
Microsoft is advocating for a specific legal framework aimed at deepfake fraud, which would give authorities the ability to target and combat AI-generated scams effectively. Further, Smith urged legislators to modernize existing federal and state laws surrounding child sexual exploitation and abuse, ensuring they account for AI-generated content.
Legislative Developments
Recently, the Senate took a noteworthy step by passing a bill aimed at addressing sexually explicit deepfakes. This legislation allows victims of nonconsensual AI-generated explicit content to initiate lawsuits against the individuals responsible for creating these damaging materials. This bill was prompted by alarming incidents involving middle and high school students who fabricated explicit images of their female peers, as well as instances of graphic deepfake content circulating on social media.
Microsoft’s Commitment to Safety
Following incidents where Microsoft’s own AI tools were misused, the company has strengthened its safety controls. A notable example was a loophole in its Designer AI image creator that enabled users to produce explicit images of celebrities, including Taylor Swift. Smith asserted, "The private sector has a responsibility to innovate and implement safeguards that prevent the misuse of AI." This responsibility aligns with the broader call for establishing robust regulations and standards.
The Challenge of Generative AI
The Federal Communications Commission (FCC) has already implemented measures against robocalls utilizing AI-generated voices. However, generative AI technologies pose a broader challenge, facilitating the creation of convincing fake audio, images, and videos. As the 2024 presidential election approaches, the risk of misinformation through deepfake content rises. Recently, Elon Musk shared a deepfake video parodying Vice President Kamala Harris on social media, which could potentially breach X’s policies regarding synthetic and manipulated media.
A Call for Transparent Labelling
Smith emphasized the importance of transparency in combating misinformation: "Congress should require AI system providers to use state-of-the-art provenance tooling to label synthetic content." This initiative is pivotal in fostering trust within the information ecosystem and aiding the public in discerning AI-generated or manipulated content from genuine material.
Conclusion
As deepfake technology continues to evolve and pose significant challenges, the call for comprehensive regulations becomes increasingly critical. By implementing laws that address the misuse of AI-generated content, protecting vulnerable populations, and ensuring transparency, Congress can help secure a safer digital environment for all.
댓글 남기기
모든 댓글은 게시 전 검토됩니다.
이 사이트는 hCaptcha에 의해 보호되며, hCaptcha의 개인 정보 보호 정책 과 서비스 약관 이 적용됩니다.