EU Investigates Facebook’s and Instagram’s Handling of Disinformation Ahead of Elections
The European Commission has initiated an investigation into Meta Platforms, the parent company of Facebook and Instagram, over alleged failures to curb disinformation and deceptive advertising in the lead-up to the European Parliament elections. The move follows concerns regarding potential sources of disinformation both within and outside the EU.
According to a report by Reuters, EU tech regulators have raised alarms over the proliferation of misleading information — not only from external actors like Russia, China and Iran, but also from political parties and organizations within the EU. These concerns have prompted the European Commission to take action amid preparations for the upcoming elections scheduled for June 6-9.
The investigation is rooted in suspicions that Meta Platforms may be in breach of EU online content rules, particularly the Digital Services Act (DSA), which came into effect last year. Under the DSA, major tech companies are obligated to take more robust measures to combat illegal and harmful content on their platforms, with potential fines reaching up to 6% of their global annual turnover.
Related: Brussels to Investigate Meta Platforms’ Handling of Disinformation on Facebook and Instagram
One focal point of the probe will be the activities of a Russia-based influence operation network known as Doppelganger, which was previously exposed by Meta in 2022. People familiar with the matter, as cited by Reuters, state that the EU investigation aims to assess Meta’s compliance with DSA obligations, particularly regarding the dissemination of deceptive advertisements, disinformation campaigns and coordinated inauthentic behavior within the EU.
In response to the investigation, Margrethe Vestager, the EU’s digital chief, expressed concerns about Meta’s moderation practices and transparency regarding advertisements and content moderation procedures. She stated, “We suspect that Meta’s moderation is insufficient, that it lacks transparency of advertisements and content moderation procedures.”
Meta Platforms, with over 250 million monthly active users in the European Union, defended its approach to risk mitigation, asserting that it has a well-established process for identifying and addressing risks on its platforms. A spokesperson for Meta emphasized the company’s commitment to cooperating with the European Commission and providing further details of its efforts to mitigate risks.
The Commission’s investigation signals a concerted effort to ensure that tech giants like Meta comply with EU regulations aimed at safeguarding the integrity of elections and combating the spread of misinformation and deceptive advertising online.
Source: Reuters
Featured News
Big Tech Braces for Potential Changes Under a Second Trump Presidency
Nov 6, 2024 by
CPI
Trump’s Potential Shift in US Antitrust Policy Raises Questions for Big Tech and Mergers
Nov 6, 2024 by
CPI
EU Set to Fine Apple in First Major Enforcement of Digital Markets Act
Nov 5, 2024 by
CPI
Six Indicted in Federal Bid-Rigging Schemes Involving Government IT Contracts
Nov 5, 2024 by
CPI
Ireland Secures First €3 Billion Apple Tax Payment, Boosting Exchequer Funds
Nov 5, 2024 by
CPI
Antitrust Mix by CPI
Antitrust Chronicle® – Remedies Revisited
Oct 30, 2024 by
CPI
Fixing the Fix: Updating Policy on Merger Remedies
Oct 30, 2024 by
CPI
Methodology Matters: The 2017 FTC Remedies Study
Oct 30, 2024 by
CPI
U.S. v. AT&T: Five Lessons for Vertical Merger Enforcement
Oct 30, 2024 by
CPI
The Search for Antitrust Remedies in Tech Leads Beyond Antitrust
Oct 30, 2024 by
CPI