Controversy Surrounds Minnesota’s Deepfake Technology Law
A federal lawsuit has emerged in Minnesota, focusing on the state law concerning the use of deepfake technology for electoral influence. Advocates and challengers of the law are now clashing, particularly over the implications of artificial intelligence (AI) in legal documentation.
Allegations of AI-Generated Text in Legal Filing
Reports from the Minnesota Reformer suggest that the affidavit supporting the law may contain elements of AI-generated content, raising questions about its authenticity. Attorneys contesting the law pointed out that the submission, authored by Jeff Hancock, the founding director of the Stanford Social Media Lab, includes citations that do not exist, hinting at AI hallucinations—misinformation typically produced by language models like ChatGPT.
Questionable Citations in Hancock’s Affidavit
Among the claims made in Hancock's affidavit is a reference to a 2023 study purportedly published in the Journal of Information Technology & Politics titled "The Influence of Deepfake Videos on Political Attitudes and Behavior." However, investigations reveal that there are no records of this study existing in any academic publication.
Another mentioned source, “Deepfakes and the Illusion of Authenticity: Cognitive Processes Behind Misinformation Acceptance,” has similarly been found to be non-existent. The absence of these sources raises the integrity of the affidavit into question.
Immediate Repercussions for Legal Proceedings
In light of these developments, the plaintiffs, which include Minnesota State Representative Mary Franson and conservative YouTuber Christopher Khols (known as Mr. Reagan), are arguing that Hancock’s declaration may be compromising the legal proceedings. They emphasized that the document lacks methodological rigor and analytical logic.
Lawyers representing Franson and Khols stated, “The citation bears the hallmarks of being an artificial intelligence (AI) ‘hallucination,’ suggesting that at least the citation was generated by a large language model like ChatGPT.” This line of reasoning calls into question not only Hancock's affidavit but also the broader ethical implications surrounding the use of AI-generated content in legal contexts.
Contact and Comments
The Verge has attempted to reach Hancock for comments on this matter, but no response has been received.
The Broader Implications of AI in Legal Frameworks
This evolving scenario underscores the growing intersection between AI technology and legal frameworks, prompting urgent discussions on how such technologies should be utilized and regulated, particularly as they relate to electoral integrity.
As deepfake technology becomes more advanced and widespread, ensuring transparency and truthfulness in legal submissions is paramount. This case in Minnesota could set a precedent for how AI-related content is handled in the legal system, highlighting the need for careful scrutiny of AI-generated text.
댓글 남기기
모든 댓글은 게시 전 검토됩니다.
이 사이트는 hCaptcha에 의해 보호되며, hCaptcha의 개인 정보 보호 정책 과 서비스 약관 이 적용됩니다.