Preview

Russian Journal of Economics and Law

Advanced search

(Synthetic) stump speech: crafting generative AI disclosure regulations for political advertisements

https://doi.org/10.21202/2782-2923.2025.3.603-634

Abstract

Objective: to analyze the benefits and drawbacks of legal regulation of relations in the sphere of the US elections in the context of using artificial intelligence.

Methods: the article employs the dialectical method of cognition, as well as general scientific (analysis, synthesis, induction, deduction) and specific scientific (formal-legal) research methods.

Results: The article assesses the current regulatory landscape for synthetic media in political advertising to analyze the benefits and drawbacks of greater regulation. Based on current and emerging regulatory approaches, the author examines how governments and private actors have limited synthetic media usage within existing First Amendment jurisprudence. Although initial prohibitions served a necessary role, the author proposes that transparency enforcement is the best approach and should be built upon by creating a repository that contains information on an advertisement's synthetic content.

Scientific novelty: Synthetic media, or content generated using artificial intelligence, has begun to infect political advertising. Federal legislation has spent most of its time stalled in committees, but states and online platforms have rapidly implemented regulations. Although synthetic media may pose harms through voter manipulation and democratic distortion, it also can lower campaign costs and more vividly illustrate conceptions of a political choice's consequences. Some governments and commentators have sought to prohibit the most harmful forms, while others have focused more on transparent approaches to regulation. In the face of yet another contentious election cycle, the question of how to ensure choices are made based on belief and not manipulation looms large.

Practical significance: the main provisions and conclusions of the article can be used in scientific, pedagogical and law enforcement activities when considering the issues related to regulation of the US elections.

About the Author

A. Paget
Fordham University School of Law
United States

Alex Paget - J.D. Candidate, 2025, Fordham University School of Law

Fordham


Competing Interests:

No conflict of interest is declared by the author



References

1. Ardia, D. S. (2022). Beyond the Marketplace of Ideas: Bridging Theory and Doctrine to Promote Self-Governance. Harv. L. & Pol'y Rev., 16, 275, 285–293.

2. Ardia, D. S., & Ringel, E. (2022). First Amendment Limits on State Laws Targeting Election Misinformation. First Amend. L. Rev., 20, 291.

3. Bambauer, J. R. (2017). The Empirical First Amendment. Ohio St. L.J., 78, 947.

4. Blitz, M. J. (2018). Lies, Line Drawing, and (Deep) Fake News. Okla. L. Rev., 71, 59.

5. Blitz, M. J. (2020). Deepfakes and Other Non-testimonial Falsehoods: When is Belief Manipulation (Not) First Amendment Speech? Yale J. L. & Tech., 23, 160.

6. Brown, N. I. (2021). Regulatory Goldilocks: Finding the Just and Right Fit for Content Moderation on Social Platforms. Tex. A & M L. Rev., 8, 451. https://doi.org/10.37419/lr.v8.i3.1

7. Canen, N., & Martin, G. J. (2023). How Campaign Ads Stimulate Political Interest. Rev. Econ. & Stat., 105, 292. https://doi.org/10.1162/rest_a_01062

8. Charles, G.-U. E. (2020). Motivated Reasoning, Post-truth, and Election Law. St. Louis U. L. J., 64, 595.

9. Chesney, B., & Citron, D. (2019). Deep Fakes: A Looming Challenge for Privacy, Democracy, and National Security. Calif. L. Rev., 107, 1753. https://doi.org/10.2139/ssrn.3213954

10. Garvey, J. B. (2022). Let's Get Real: Weak Artificial Intelligence Has Free Speech Rights: Note. Fordham L. Rev., 91, 953.

11. Goodman, E. P. (2006). Stealth Marketing and Editorial Integrity. Tex. L. Rev., 85, 83.

12. Goodman, E. P. (2021). Digital Fidelity and Friction. Nev. L. J., 21, 623.

13. Green, R. (2019). Counterfeit Campaign Speech. Hastings L. J., 70, 1445.

14. Haenschen, K. (2023). The Conditional Effects of Microtargeted Facebook** Advertisements on Voter Turnout. Pol. Behav., 45, 1661. https://doi.org/10.1007/s11109-022-09781-7

15. Hasen, R. L. (2020). Deep Fakes, Bots, and Siloed Justices: American Election Law in a "Post-truth" World. St. Louis U. L. J., 64, 535.

16. Hodge, S. D. Jr. (2021). Don't Always Believe What You See: Shallowfake and Deepfake Media Has Altered the Perception of Reality. Hofstra L. Rev., 50, 51.

17. Huh, J., Nelson, M. R., & Russell, C. A. (2023). ChatGPT, AI Advertising, and Advertising Research and Education. J. Advertising, 52, 477, 477. https://doi.org/10.1080/00913367.2023.2227013

18. Matwyshyn, A. M., & Mowbray, M. (2021). Fake. Cardozo L. Rev., 43, 643.

19. Mokadem, S. Sh. El. (2023). The Effect of Media Literacy on Misinformation and Deep Fake Video Detection. Arab Media & Soc'y, 35, 115.

20. Mukhamediev, R. I., Popova, Y., Kuchin, Ya., Zaitseva, E., Kalimoldayev, A., Symagulov, A., Levashenko, V., Abdoldina, F., Gopejenko, V., Yakunin, K., Muhamedijeva, E., & Yelis, M. (2022). Review of Artificial Intelligence and machine learning technologies: classification, restrictions, opportunities and challenges. Mathematics, MDPI, 10(15), 1–25. https://doi.org/10.3390/math10152552

21. Navarrine, P. G. (2023). Political advertising on free streaming sites: conflicts with first amendment and exploring viability of regulation: note. Cornell L. Rev., 108, 1821.

22. Ostrowski, J. R. (2023). Shallowfakes. The New Atlantis, 72, 96.

23. Parsons, M. (2020). Fighting for attention: democracy, free speech, and the marketplace of ideas. Minn. L. Rev., 104, 2157.

24. Russell, L. (2020). Weaver, fake news (& deep fakes) and democratic discourse. J. Tech. L. & Pol'y, 24, 35.

25. Schroeder, J. (2022). The marketplace of ideas and the problem of networked truths. U. Tol. L. Rev., 54, 27.

26. Volokh, E. (2018). The law of compelled speech. Tex. L. Rev., 97, 355.

27. Wood, A. K., & Ravel, A. M. (2018). Fool me once: regulating “fake news” and other online advertising. S. Cal. L. Rev., 91, 1223.

28. Yamaoka-Enkerlin, A. (2020). Disrupting disinformation: deepfakes and the law. Comment. N.Y.U. J. Legis. & Pub. Pol'y, 22, 725.

29. Paget, A. (2024). (Synthetic) stump speech: crafting generative AI disclosure regulations for political advertisements. Fordham Law Review, 93(1), 321–358.


Review

For citations:


Paget A. (Synthetic) stump speech: crafting generative AI disclosure regulations for political advertisements. Russian Journal of Economics and Law. 2025;19(3):603-634. (In Russ.) https://doi.org/10.21202/2782-2923.2025.3.603-634

Views: 25


ISSN 2782-2923 (Print)