Media and Internal Security — The Dual Role
Media and social networking platforms play a dual role in internal security — they can both strengthen and threaten it.
| Role | Positive Impact | Negative Impact |
|---|---|---|
| Information | Early warning during disasters; civic awareness | Fake news triggers panic, mob violence, communal riots |
| Accountability | Exposing corruption, human rights violations | Trial by media; undermining investigations |
| Mobilisation | Democratic movements (anti-corruption, protests) | Radicalisation; recruitment by extremist groups |
| Communication | Government outreach to citizens | Propaganda by hostile state and non-state actors |
| Surveillance | Intelligence gathering through open-source monitoring | Privacy violations; mass surveillance risks |
Social Media — Scale and Security Implications
India's Digital Landscape
| Platform | Users in India (est. 2025-26) |
|---|---|
| 550+ million | |
| YouTube | 500+ million |
| 360+ million | |
| 340+ million | |
| Telegram | 150+ million |
| X (Twitter) | 30+ million |
India has 90+ crore internet users (2025) — the largest online population after China. This massive digital presence creates both opportunities and vulnerabilities for internal security.
Threats from Social Media
1. Fake News and Misinformation
| Aspect | Detail |
|---|---|
| Scale | India is the world's largest market for misinformation due to high internet penetration + low digital literacy |
| Triggers | Communal tensions (doctored videos of religious incidents), mob lynching (WhatsApp rumours of child kidnappers), election manipulation |
| Deepfakes | AI-generated realistic fake videos/audio; used in 2024 elections for synthetic speeches and videos of deceased leaders |
| Impact | At least 30+ deaths linked to WhatsApp rumour-driven mob violence (2017-2019); communal riots fuelled by fabricated content |
WhatsApp Lynchings (2018): A spate of mob killings swept India in 2018 after fake messages about child kidnappers and organ harvesters went viral on WhatsApp. At least 24 people were killed in mob attacks in 2018 alone. The incidents began in Jharkhand in May 2017 but escalated nationally by mid-2018. In response, WhatsApp launched a newspaper advertising campaign warning against fake news in July 2018, labelled forwarded messages, and disabled the "quick forward" button in India.
For Mains: Fake news is not just a law-and-order problem — it is a national security threat. During India-Pakistan tensions (Pulwama/Balakot, 2019), fake images and videos circulated widely, inflaming public sentiment and complicating government communication. Hostile state actors can weaponise fake news to destabilise society without a single soldier crossing the border. This is hybrid warfare.
2. Radicalisation and Recruitment
| Threat Actor | How They Use Social Media |
|---|---|
| Jihadi groups (IS, AQ) | Propaganda videos, encrypted recruitment via Telegram, radicalisation of lone wolves |
| Left-Wing Extremism | Maoist/Naxal propaganda; mobilisation of cadre; disinformation about security forces |
| Separatist movements | Khalistan, Kashmir separatism — diaspora-driven social media campaigns; foreign funding coordination |
| Right-wing extremism | Communal hate speech; targeted harassment campaigns; mob mobilisation |
Case: NIA investigations have found that over 100 Indians who attempted to join ISIS were radicalised primarily through social media content. Encrypted messaging apps (Telegram, Signal) make it difficult for intelligence agencies to monitor communications.
Global scale of online radicalisation: ISIS mobilised an estimated 40,000 foreign nationals from 110 countries, largely through social media. Over 40,000 Twitter accounts actively supported ISIS, with approximately 2,000 tweeting in English. The dark web and encrypted platforms (Telegram channels, Signal groups) serve as secondary recruitment layers where operatives share operational manuals and coordinate logistics beyond the reach of law enforcement.
3. Information Warfare and Hybrid Threats
Hybrid warfare blends conventional military operations, cyberattacks, information operations, and economic pressure to destabilise a target state without a formal declaration of war. The information warfare component uses troll farms — organised groups of paid operatives and automated bots — to create fake social media accounts, amplify divisive content, and manipulate public opinion. State-sponsored troll farms have been documented in Russia, China, and several other countries. During the 2016 US presidential election, Russian troll farms flooded social media with polarising content to deepen societal divisions — a textbook hybrid warfare operation.
Relevance for India: Pakistan-based accounts routinely amplify separatist narratives during Kashmir incidents. China has been linked to influence operations through media investments and data harvesting via apps. The challenge of attribution — proving state sponsorship — makes countering information warfare particularly difficult.
4. Cyber Propaganda by Hostile States
| Actor | Method |
|---|---|
| Pakistan | Coordinated info-ops during Kashmir incidents; fake accounts amplifying separatist narratives |
| China | Influence operations via apps, media investments; data harvesting through banned apps (TikTok) |
| Non-state proxies | Bot networks amplifying divisive content during elections, communal incidents |
5. Impact on Law Enforcement
| Challenge | Detail |
|---|---|
| Mob violence | Social media accelerates crowd formation — police have minutes, not hours, to respond |
| Investigation compromise | Trial by media can prejudice public opinion; leaked investigation details compromise operations |
| Encrypted communications | End-to-end encryption on WhatsApp, Signal prevents lawful interception |
| Jurisdiction | Platforms hosted overseas (Meta, Google, X) — enforcement of Indian laws is complex |
Legal Framework
Information Technology Act, 2000
| Provision | Relevance |
|---|---|
| Section 66A | Punished "offensive" online content — struck down by Supreme Court in Shreya Singhal v. Union of India (2015) as violating Article 19(1)(a) |
| Section 69A | Government can direct blocking of content/websites in the interest of sovereignty, security, public order |
| Section 79 | Safe harbour — intermediaries not liable for user-generated content IF they comply with due diligence (IT Rules) |
| Section 87 | Power to make rules (basis for IT Rules 2021) |
Landmark: Shreya Singhal v. Union of India (2015): The Supreme Court struck down Section 66A as unconstitutional — it was vague ("grossly offensive", "menacing character") and chilled free speech. However, Section 69A (government blocking orders) was upheld as having sufficient procedural safeguards. This is a must-know case for both GS2 and GS3.
IT (Intermediary Guidelines and Digital Media Ethics Code) Rules, 2021
| Provision | Requirement |
|---|---|
| Due diligence | Platforms must have a grievance officer, compliance officer, and nodal contact for law enforcement |
| Content takedown | Remove content within 24 hours of government/court order; remove intimate images within 24 hours of complaint |
| First originator tracing | Significant Social Media Intermediaries (SSMIs) with 5 million+ users must identify the first originator of a message when ordered by court/government |
| Significant intermediary | Platforms with 5 million+ registered users must comply with additional obligations (monthly compliance report, etc.) |
| Digital media ethics | OTT platforms and digital news media regulated under a three-tier self-regulatory structure |
First originator controversy: WhatsApp challenged the tracing requirement, arguing it would require breaking end-to-end encryption. The case is sub judice. The tension: government needs tracing for crime investigation (fake news, terrorism); privacy advocates argue mass tracing undermines encrypted communication for all users.
Other Relevant Laws
| Law | Relevance |
|---|---|
| Bharatiya Nyaya Sanhita (BNS), 2023 | Replaced IPC; Section 197 (imputations, assertions prejudicial to national integration); Section 353 (statements creating enmity); Section 356 (defamation) |
| Unlawful Activities (Prevention) Act, 1967 | Online content supporting terrorism or unlawful activities |
| National Security Act, 1980 | Preventive detention for social media posts threatening public order (controversial use) |
| Digital Personal Data Protection Act, 2023 | Data privacy framework; regulates how platforms collect, process, and store user data |
| Press Council of India Act, 1978 | Statutory body to preserve press freedom and maintain standards; can warn, admonish, or censure — but has no power to impose penalties |
| Cable Television Networks (Regulation) Act, 1995 | Mandatory registration for cable operators; content standards prohibiting obscene, defamatory, or violence-inciting material; compulsory carriage of Doordarshan channels |
| Telecommunications Act, 2023 | Empowers government to order temporary internet shutdowns in the interest of public safety |
Digital Personal Data Protection Act, 2023 — Key Features
| Feature | Detail |
|---|---|
| Core principles | Consent, transparency, purpose limitation, data minimisation, accuracy, storage limitation, security safeguards, accountability |
| Consent standard | Must be free, specific, informed, unconditional, and unambiguous; can be withdrawn at any time |
| Data Protection Board | Central government establishes the Data Protection Board of India — monitors compliance, imposes penalties, directs breach response, hears grievances |
| Significant Data Fiduciary (SDF) | Government can designate certain fiduciaries as SDFs based on volume/sensitivity of data and national security risk; SDFs must appoint a Data Protection Officer, conduct audits and impact assessments |
| Penalties | Up to Rs 250 crore for non-compliance; up to Rs 200 crore for failure to report data breaches |
| DPDP Rules, 2025 | Notified on 13 November 2025; established the Data Protection Board and set phased compliance timelines (full compliance by May 2027) |
Traditional Media vs Social Media Regulation — A Comparison
| Parameter | Traditional Media (Print/TV) | Social Media |
|---|---|---|
| Regulator | Press Council of India (print); Cable TV Act, 1995 / NBDA (broadcast) | IT Act 2000 + IT Rules 2021 (MeitY) |
| Entry barrier | Registration/licensing required | No barrier; anyone can publish |
| Content accountability | Editor legally responsible | Platform claims safe harbour (Section 79); individual user responsible |
| Speed of regulation | Established frameworks; slower content but easier to regulate | Content goes viral before regulation can act |
| Cross-border challenge | Largely domestic operations | Global platforms; servers abroad; jurisdiction issues |
| Anonymity | Known publishers/editors | Pseudonymous or anonymous accounts; bot networks |
| Self-regulation | Press Council, NBDA, Editors Guild | Three-tier IT Rules structure; largely untested |
Internet Shutdowns
India leads the world in internet shutdowns.
| Feature | Detail |
|---|---|
| Total shutdowns (2012-2025) | 800+ (highest globally) |
| Justification | Prevention of violence, communal harmony, examination malpractice |
| Longest | Kashmir (August 2019 — 18+ months of varying restrictions after Article 370 abrogation) |
| SC ruling | Anuradha Bhasin v. Union of India (2020) — internet access is a fundamental right under Article 19(1)(a); shutdowns must be proportionate, necessary, and time-bound; must be published as a written order subject to judicial review |
For Mains: Internet shutdowns are a blunt tool with enormous economic cost (~$1.9 billion per year). They prevent violence in the short term but also disable emergency communication, healthcare, banking, and livelihood for millions. The Anuradha Bhasin ruling established proportionality requirements, but implementation remains inconsistent. Discuss alternatives: targeted content blocking (Section 69A), platform cooperation, digital literacy programmes.
Case Studies — Social Media and Security
| Case | What Happened | Security Lesson |
|---|---|---|
| Arab Spring (2011) | Protesters in Tunisia and Egypt used Facebook and Twitter to organise mass protests; tweets from Egypt surged from 2,300/day to 230,000/day in the week before Mubarak's resignation; Facebook users in the Arab world grew 30% in Q1 2011 | Social media can topple authoritarian regimes; also creates power vacuums and instability |
| Cambridge Analytica (2018) | Data of up to 87 million Facebook profiles harvested through the app "This Is Your Digital Life" by researcher Aleksandr Kogan; used for political micro-targeting; Facebook fined $5 billion by FTC | Data harvesting by platforms is a sovereignty and electoral security risk |
| WhatsApp Lynchings, India (2018) | Fake child-kidnapping rumours on WhatsApp led to at least 24 mob killings; WhatsApp responded by labelling forwards and disabling quick-forward | End-to-end encryption complicates law enforcement; digital literacy is critical |
| Pulwama/Balakot (2019) | Fake images and videos circulated during India-Pakistan tensions, inflaming public sentiment and complicating government communication | Hostile actors weaponise crises through information warfare |
Positive Role of Media in Security
| Area | How Media Helps |
|---|---|
| Disaster response | Social media as real-time alert system (Kerala floods 2018, Chennai floods 2015) — crowd-sourced rescue coordination |
| Community policing | Police social media accounts build public trust; tip-offs from citizens; missing persons alerts |
| Counter-narrative | Government and civil society counter extremist propaganda online |
| Transparency | CCTV footage, body cameras, citizen journalism hold security forces accountable |
| Intelligence | Open Source Intelligence (OSINT) from social media aids threat assessment |
| Democratisation of information | Citizens access government data, RTI information, and policy debates directly — reduces information asymmetry |
| Election monitoring | Social media enables real-time reporting of booth-level irregularities; Election Commission uses platforms for voter awareness campaigns |
Balancing Security and Freedom
| Approach | Argument |
|---|---|
| More regulation | National security requires content control; fake news kills; radicalisation is a clear danger |
| Less regulation | Free speech (Article 19(1)(a)) is fundamental; government overreach leads to censorship; democratic dissent is not a security threat |
| Balanced approach | Targeted action against specific content (not blanket shutdowns); transparent processes with judicial oversight; platform accountability without state control of speech |
For Mains framework: The ideal answer acknowledges that social media is BOTH a threat and a tool for security. Recommend: (1) strengthen digital literacy to build citizen resilience against fake news, (2) platform accountability under IT Rules with independent oversight, (3) proportionate content regulation with judicial safeguards, (4) international cooperation on cross-border cyber threats, (5) invest in OSINT capabilities for intelligence agencies. Avoid binary positions — neither "ban social media" nor "leave it unregulated" is a good answer.
Key IT and Cyber Laws at a Glance
| Law/Rule | Year | Key Provision for Internal Security |
|---|---|---|
| IT Act | 2000 | Foundational cyber law; Sections 69A (blocking), 79 (safe harbour) |
| IT Act Amendment | 2008 | Added cyber terrorism (Section 66F), data protection (Section 43A) |
| Shreya Singhal judgment | 2015 | Section 66A struck down; Section 69A upheld |
| IT Rules (Intermediary Guidelines) | 2021 | Due diligence for platforms; SSMI obligations; first originator tracing |
| IT Rules Amendment | 2023 | Fact-check unit provision (challenged in courts); three-tier digital media regulation |
| DPDP Act | 2023 | Consent-based data processing; Data Protection Board; penalties up to Rs 250 crore |
| Telecommunications Act | 2023 | Internet shutdown powers; telecom licensing framework |
| DPDP Rules | 2025 | Implementation framework; phased compliance; Data Protection Board constituted |
UPSC Relevance
Prelims Focus Areas
- IT Act 2000 — Section 66A (struck down), Section 69A (blocking), Section 79 (safe harbour)
- IT Rules 2021 — first originator tracing, significant intermediary threshold (5 million users)
- Shreya Singhal v. Union of India (2015) — Section 66A struck down as unconstitutionally vague; Section 69A upheld
- Anuradha Bhasin v. Union of India (2020) — internet access is a fundamental right; shutdowns must be proportionate
- Digital Personal Data Protection Act, 2023 — Data Protection Board, consent framework, Significant Data Fiduciaries
- Telecommunications Act, 2023
- Press Council of India — established under Press Council Act, 1978; advisory powers only
Mains Focus Areas
- Social media as a tool for radicalisation and recruitment (ISIS, Maoist, separatist groups)
- Fake news and its impact on communal harmony and public order (WhatsApp lynchings 2018)
- Internet shutdowns — necessity vs proportionality (Anuradha Bhasin framework)
- Balancing free speech with national security
- Role of social media in hybrid warfare and information operations
- Platform regulation — IT Rules, Section 69A, and the safe harbour debate
- Deepfakes and AI-generated misinformation — emerging challenges
- Positive use of media in disaster management and community policing
- Cambridge Analytica and data sovereignty concerns
- Traditional media vs social media regulation — gaps and challenges
Vocabulary
Deepfake
- Pronunciation: /ˈdiːpˌfeɪk/
- Definition: A convincingly realistic but fabricated image, video, or audio recording created using artificial intelligence — particularly deep learning techniques — that superimposes one person's likeness onto another or generates entirely synthetic media.
- Origin: A blend of deep learning + fake; coined in 2017 by a Reddit user who demonstrated face-swapping technology using deep neural networks.
Disinformation
- Pronunciation: /dɪsˌɪnfərˈmeɪʃən/
- Definition: False or misleading information that is deliberately created and disseminated with the intent to deceive, manipulate public opinion, or cause harm — distinct from misinformation, which is spread without deliberate intent.
- Origin: From dis- ("negation, reversal") + information; attested in English from 1939, modelled on Russian dezinformatsiya, a term used by Soviet intelligence for coordinated propaganda campaigns.
Censorship
- Pronunciation: /ˈsɛnsərʃɪp/
- Definition: The suppression or restriction of speech, publication, or other forms of expression by a government, regulatory body, or institution, typically on grounds of national security, public order, morality, or political control.
- Origin: From censor (Latin censor, a Roman magistrate responsible for public morals and the census) + -ship; the Roman censorship dates to 443 BCE.
Key Terms
Fake News
- Pronunciation: /feɪk njuːz/
- Definition: Fabricated or misleading content deliberately designed and disseminated under the guise of legitimate news reporting in order to deceive readers, manipulate public opinion, inflame communal or political tensions, generate advertising revenue through clickbait, or advance specific political agendas — often amplified virally through social media platforms and encrypted messaging services like WhatsApp, where content verification before forwarding is minimal.
- Context: The term gained global prominence during and after the 2016 US presidential election, with Collins Dictionary naming it Word of the Year in 2017 after its usage surged 365% year-on-year. In India, fake news became a lethal internal security concern during the 2017-2018 WhatsApp lynchings — a spate of mob killings triggered by fabricated rumours about child abduction and organ harvesting spread via WhatsApp, commencing with seven men killed in Jharkhand in May 2017 and resulting in over a dozen deaths across multiple states. These incidents prompted WhatsApp to limit message forwarding to five chats at a time (July 2018) and India to strengthen intermediary obligations under the IT (Intermediary Guidelines) Rules, 2021. The British government has criticised "fake news" as a "poorly defined" term that conflates genuine error with deliberate fabrication, preferring "disinformation" (deliberate) and "misinformation" (unintentional).
- UPSC Relevance: GS3 Internal Security — Mains asks about the role of social media in spreading fake news, its impact on communal harmony (WhatsApp lynchings 2018), and government measures (IT Rules 2021 — SSMI obligations, first originator tracing; Section 69A content blocking). Also tested as a hybrid warfare dimension — hostile state actors weaponising misinformation and deepfakes for information operations. Links to GS2 (free speech under Article 19(1)(a) vs regulation under 19(2)), GS4 (media ethics, paid news, trial by media), and the emerging challenge of AI-generated deepfakes that make fake news harder to detect.
Social Media Regulation
- Pronunciation: /ˈsoʊʃəl ˈmiːdiə ˌrɛɡjuˈleɪʃən/
- Definition: The body of laws, rules, and institutional mechanisms that govern the conduct of social media platforms and their users, addressing content moderation, data privacy, platform accountability, intermediary liability, and the balance between free expression and public safety. In India, the regulatory framework primarily comprises the IT Act, 2000, the IT (Intermediary Guidelines and Digital Media Ethics Code) Rules, 2021, the Digital Personal Data Protection Act, 2023, and the Telecommunications Act, 2023 — supplemented by landmark judicial decisions that define the boundaries of online free speech.
- Context: Emerged as a distinct policy domain in the 2010s as the scale and influence of social media platforms grew to billions of users globally; between 2011 and 2022, 78 countries enacted laws targeting the spread of false or harmful content on social media. India's IT (Intermediary Guidelines) Rules, 2021 introduced the category of Significant Social Media Intermediaries (SSMIs) — platforms with 50 lakh (5 million) or more registered users in India — imposing enhanced obligations including appointment of a Chief Compliance Officer, Nodal Contact Person, and Resident Grievance Officer (all based in India), content takedown within 36 hours of government/court order, and enabling identification of the first originator of information on messaging platforms (a requirement that clashes with end-to-end encryption on WhatsApp and Signal). Key judicial landmarks include Shreya Singhal v. Union of India (2015, struck down Section 66A as unconstitutionally vague while upholding Section 69A) and Anuradha Bhasin v. Union of India (2020, held that internet access is a fundamental right and shutdowns must be proportionate and time-bound).
- UPSC Relevance: GS3 Internal Security — Mains asks "How should India balance free speech with social media regulation?" and tests knowledge of IT Rules 2021 (SSMI obligations, first originator tracing, 36-hour content takedown), Section 79 safe harbour (conditional on due diligence compliance), and the Shreya Singhal and Anuradha Bhasin judgments. Prelims tests SSMI threshold (5 million users), content takedown timelines, and the fact-check unit provision (2023 amendment, challenged in courts). A cross-cutting topic spanning GS3 (internal security, cyber security), GS2 (governance, fundamental rights), and GS4 (media ethics, information ethics).
BharatNotes