Social Media Platforms Grapple with Disinformation and Harmful Content During Pakistan’s 2024 General Elections

The 2024 general elections in Pakistan witnessed a surge in disinformation and harmful content disseminated across major social media platforms, raising serious concerns about the efficacy of content moderation mechanisms and their impact on democratic processes. A new report by the Digital Rights Foundation (DRF), "Platforms At The Polls: Disinformation, Political Ads & Accountability During The 2024 Pakistan General Elections," highlights the inconsistent enforcement of content policies and the disproportionate targeting of women in the public sphere, particularly journalists and politicians. The report scrutinizes the performance of platforms like TikTok, Facebook/Instagram, YouTube, and X (formerly Twitter), revealing a stark contrast between their content moderation practices in Pakistan and those implemented in the Global North.

The pervasiveness of harmful content during the election period was alarming. The DRF’s analysis of 225 posts between November 2023 and March 2024 exposed a disturbing trend: 163 posts constituted gendered disinformation, while the remaining content encompassed broader disinformation, misinformation, fake news, AI-generated content, and hate speech. This surge in online abuse was particularly egregious against women. Technology-facilitated gender-based violence (TFGBV) and the use of deepfakes became potent weapons deployed against female journalists and politicians, threatening not only their reputations but also their safety. The vulnerability of marginalized communities, including ethnic, indigenous, and transgender groups, was further underscored by the DRF’s findings.

Political parties also became targets of online disinformation campaigns, with the Pakistan Muslim League-N (PML-N) bearing the brunt of the attacks. Maryam Nawaz, a prominent figure within the PML-N, was the subject of a significant proportion of the gendered disinformation posts. The Pakistan Tehreek-e-Insaf (PTI) and the Pakistan People’s Party (PPP) also faced online attacks, albeit to a lesser extent. The prevalence of these targeted campaigns underscores the need for robust content moderation strategies that can effectively identify and remove harmful content before it can inflict significant damage.

While social media platforms boast comprehensive policies on electoral information integrity, their implementation in Pakistan and other South Asian countries has been demonstrably insufficient. The DRF report reveals a concerning gap between stated policies and actual practices. Despite having clear community guidelines and content moderation policies, platforms struggled to control the spread of disinformation and harmful content related to the elections. This discrepancy raises questions about the platforms’ commitment to upholding these policies in regions outside the Global North. The lack of transparency surrounding political advertising expenditure further complicates the issue.

The financial aspect of online political campaigning also came under scrutiny in the DRF report. An analysis of political ad spending on Facebook and Instagram revealed a collective expenditure of Rs54 million (approximately $190,000) by major political parties. The Pakistan Tehreek-e-Insaf (PTI) topped the list, spending Rs25 million, followed by the Istehkam-e-Pakistan Party (IPP) with Rs13 million, and other parties with lesser expenditures. Significantly, only Meta (the parent company of Facebook and Instagram) provided transparency regarding political ad spending. Other platforms failed to disclose similar data, hindering efforts to understand the full extent of online political campaigning and its potential influence on the election outcome. This lack of transparency raises questions about accountability and the need for regulatory frameworks to ensure greater transparency in online political advertising.

The DRF report serves as a wake-up call, highlighting the urgent need for social media platforms to strengthen their content moderation practices, particularly in regions susceptible to disinformation campaigns and online abuse. The inconsistent enforcement of content policies, coupled with the lack of transparency surrounding political advertising, undermines democratic processes and poses a threat to free and fair elections. As Nighat Dad, Executive Director of DRF, emphasizes, maintaining electoral integrity on social media platforms is crucial for upholding democratic norms. To achieve this, tech companies must prioritize information integrity and implement robust content moderation strategies that are consistently applied across the globe. This includes greater transparency regarding political advertising and a commitment to addressing election-related information manipulation promptly and effectively. The findings of the DRF report emphasize the need for a collaborative approach involving tech companies, policymakers, and civil society organizations to combat disinformation and ensure that social media platforms contribute positively to democratic discourse.

Share.
Exit mobile version