The rise of artificial intelligence (AI) is undoubtedly transforming various industries, and politics is no exception. AI voice cloning startups are venturing into the realm of political advertising, despite growing concerns surrounding disinformation. Instreamatic, an AI audio/video ad platform based in Boca Raton, Florida, has recently announced its foray into political advertising. With its innovative solution, candidate campaigns can generate highly-targeted AI-driven contextual video and audio ads featuring voiceovers that adapt to changing events or locations. This expansion of AI in political advertising raises important questions about the potential for disinformation in the upcoming 2024 US elections.
AI Voice Cloning: A Powerful Tool or Disinformation Minefield?
Instreamatic’s offering allows candidates to alter any audio or video political ad by replicating a voice without the need for time-consuming studio re-recordings. This feature was demonstrated through a video campaign involving the replication of Barack Obama’s voice. The AI-powered voiceovers can automatically create unlimited ad versions, incorporating specific details such as the audience’s location, time of day, app or platform where they receive the ad, or even the nearest store. While this technology presents exciting possibilities for personalization and targeting, the use of AI in political campaigns is expected to become a disinformation minefield.
As AI-generated content raises concerns about its potential misuse, Instreamatic emphasizes that it has implemented guardrails to prevent its product from being used for election disinformation. Stas Tushinskiy, the CEO and co-founder of Instreamatic, explains that clients must confirm they have permission to use a voice for any campaign. Furthermore, the political advertising offering will not be available to everyone, as Instreamatic will be actively engaged in campaign creation. Tushinskiy affirms that the platform is intended for legitimate use and expresses a commitment to address any issues promptly. If necessary, Instreamatic will take immediate action, including deleting problematic political ads and making public safety statements.
Tushinskiy clarifies that Instreamatic’s offering does not seek to reinvent political advertising but rather aims to automate the existing manual process. Traditionally, the process involves candidates or voice talents spending hours in a studio, followed by the extensive involvement of multiple individuals to upload and check for errors. Instreamatic streamlines this process, compressing it from weeks to mere minutes. Additionally, the back-and-forth communication between agencies and clients, often requiring new takes due to word changes, becomes more efficient with voice cloning technology. This allows for more contextual ads that mention specific destinations or locations, thereby increasing the effectiveness of ad spend.
Despite the potential benefits of AI voice cloning in political advertising, concerns persist due to the absence of federal regulations for AI-generated content in political campaigns. As highlighted by experts, the current landscape of political ads is fraught with the risk of disinformation. The use of AI, whether in the form of chatbots or deepfakes, can significantly complicate the 2024 US elections. The lack of clear guidelines and rules raises red flags regarding the potential manipulation of voters through AI-generated political content.
As AI voice cloning startups like Instreamatic continue to enter the political arena, the ethical and regulatory challenges associated with this technology must be addressed. Striking a balance between innovation and responsible use is crucial to avoid disinformation and uphold the integrity of democratic processes. It is imperative for policymakers and industry stakeholders to collaborate and establish comprehensive regulations for AI-generated content in political campaigns. Looking ahead, the future of AI in politics hinges on finding ethical solutions that harness the power of technology while safeguarding against its potential misuse.
Leave a Reply