The European Commission has initiated an inquiry into the strategies employed by big tech companies to address the risks associated with generative Artificial Intelligence (AI) that could potentially mislead voters during elections. The commission made this formal request on March 14 and has set an April 3 deadline for the companies to submit relevant documents and information.

The request was issued to prominent technology companies such as Facebook, Instagram, X (formerly Twitter), TikTok, Google Search, Bing, Snapchat, and YouTube. The European Commission is seeking to understand the measures and approaches these companies have implemented to mitigate the potential risks posed by generative AI in the context of electoral processes.

By scrutinizing the strategies employed by these tech giants, the European Commission aims to ensure the integrity and transparency of elections within its jurisdiction. This inquiry reflects the growing concerns surrounding the influence of AI-generated content on electoral outcomes and the need to address potential misinformation or manipulation.

It will be important to monitor the response of these companies to the commission’s request and observe any subsequent actions or policy changes that may emerge as a result of this inquiry.

European Commission Questions Big Techs over Generative AI’s Hazards

he European Commission’s request for information (RFI) is aimed at obtaining further details about the potential hazards associated with generative AI. The concern arises from the fact that big tech companies enable users to create and disseminate content using this technology.

The European Union (EU) is particularly focused on understanding the precautions and measures taken by these companies to mitigate the risks of generative AI, as it relates to election voters. The EU is worried about the potential impact of viral deepfakes and manipulation of automated services on voters’ perceptions during elections.

By seeking information from these tech companies, the European Commission aims to gain insights into the strategies and safeguards in place to address these concerns. This inquiry reflects the EU’s commitment to ensuring the integrity and security of electoral processes, as well as its efforts to address the potential risks posed by emerging technologies like generative AI.

Monitoring the companies’ responses and any subsequent actions taken by the European Commission will provide further clarity on how these issues are being addressed and regulated within the EU.


Following the information requests, the European Commission has the authority to impose penalties for errors, inconsistencies, or misrepresentation.

Indeed, the European Commission’s request for information regarding the impact of generative AI on election processes aligns with the EU’s broader efforts to regulate the digital sphere. The Digital Services Act (DSA) is one of the recent regulations introduced by the EU to govern e-commerce and online platforms.

Under the DSA, certain platforms have been classified as Very Large Online Platforms (VLOPs) and are subject to specific obligations. These VLOPs are required to assess and manage systemic risks, along with complying with other guidelines outlined in the act.

In addition to election security, the European Commission is also concerned about various other issues related to online platforms. This includes addressing gender-based violence, combating the distribution of illicit content, safeguarding fundamental rights, protecting minors, and promoting mental health.

By focusing on generative AI’s impact on election processes, the European Commission is considering one aspect of the broader challenges associated with online platforms. The aim is to ensure the responsible and secure use of emerging technologies while safeguarding the rights and well-being of individuals within the EU.

As the EU continues to develop and implement regulations, it will be important to monitor how these measures evolve and their impact on online governance and digital services.

EU Moves Towards Election Security Rules

The European Commission’s active examination of the impact of AI on voters reflects its recognition of the increasing use of AI across various domains. The potential for AI to amplify misinformation among voters has been highlighted in recent studies, underscoring the need for robust measures to address this issue.

The EU has been diligently working on election security regulations and is on track to finalize them by March 27. The European Commission has sought feedback on election security regulations through public consultations, indicating its commitment to engaging stakeholders and considering a broad range of perspectives.

The recent request for information (RFI) directed at big tech companies is considered a crucial step in shaping the EU’s election security policy. The insights and information provided by these companies will aid in the formulation of effective measures to address election security challenges, including the spread of misinformation among voters.

By developing a comprehensive election security policy, the EU aims to tackle the persistent challenges associated with misinformation during elections. This policy will likely encompass a range of strategies and guidelines to ensure the integrity, transparency, and fairness of electoral processes in the digital age.

As the EU progresses in creating its election security policy, it will be important to assess its effectiveness and adaptability to evolving technological landscapes.

By ailf

Leave a Reply

Your email address will not be published. Required fields are marked *