Published On: February 4th 2026
Authored By: Aayesha Gupta
Kirit P. Mehta School of Law, Narsee Monjee Institute of Management Deemed to be University (NMIMS), Mumbai
Introduction
There was a time when truth had a face.
You could look at a photograph and trust that the moment existed. You could hear a voice and believe it belonged to the body it came from. Law itself was built on this assumption: that evidence had a physical anchor, that testimony was tethered to reality, that falsity required effort.
Deepfakes shattered that assumption quietly. No explosions. No warning. Just a slow erosion of certainty.
What makes this rupture unprecedented is scale. Since 2019, the number of detected deepfake videos online has increased multiple times over, with monitoring reports showing exponential rather than linear growth. What once required technical expertise can now be done using consumer-facing AI systems such as OpenAI, Stability AI, and Runway, often with no coding knowledge at all. The barrier to impersonation has effectively collapsed.
96 percent of all deepfake videos online are non-consensual porn.
Nearly all of them feature women.
This is not misuse. This is a pattern.
Deepfakes are already the largest non-consensual sexual content ecosystem ever created.
Independent deepfake landscape analyses consistently show that more than 90 percent of identified deepfake videos are pornographic in nature, and within that category, over 95 percent involve the non-consensual use of women’s faces. Most victims are not public figures but private individuals. This makes deepfake abuse one of the fastest-growing forms of gendered digital violence, operating largely outside effective legal remedy.
Today, a woman can watch herself commit a crime she never committed. A politician can declare war without opening their mouth. A teenager can be digitally undressed without consent and told by the world that it is “just pixels.” The harm is real. The law is confused. And truth, for the first time in history, is on the defensive.
This is not a technology problem. It is a moral one that law has not yet learned how to name. When Seeing Is No Longer Believing
Empirical studies now show that between 60 and 75 percent of viewers are unable to reliably distinguish a high-quality deepfake video from a real one. Detection accuracy drops further when videos are short, emotionally charged, or viewed on mobile devices. Even trained participants perform close to chance when faced with sophisticated manipulations, eroding the legal assumption that visual evidence carries inherent reliability.
Deepfakes are not impressive because they are clever. They are terrifying because they are convincing.
Deepfake-related fraud losses crossed $1.2 billion globally in a single year.
Voice cloning scams have been used to impersonate CEOs, parents, and even emergency calls.
One deepfake voice can empty bank accounts in minutes.
The law still treats this as “cyber fraud,” as if it were an email typo.
Voice-based deception has expanded rapidly due to AI systems such as ElevenLabs, Resemble AI, and Descript, which can replicate a person’s voice using seconds of audio. Reported cases show individual financial losses ranging from ₹50 lakh to over ₹40 crore, with law enforcement agencies acknowledging sharp growth since 2022.
They operate in the most intimate spaces of trust: the human face, the human voice, the emotional cues we evolved to rely on. The danger is not that people will believe lies. People have always believed lies. The danger is that people will stop believing anything at all.
This is called the “liar’s dividend.” When everything can be fake, the guilty gain a new defence: That wasn’t me. The authentic becomes suspect. Accountability dissolves.
In courtrooms, this uncertainty is lethal. Evidence is no longer evidence simply because it exists. Judges must ask questions the law was never designed to answer: Was this video generated? By whom? With what model? Trained on whose data? Altered at what point?
The burden of proof grows heavier just as the ground beneath it collapses.
From a legal point of view, less than 10 percent of reported deepfake abuse cases actually reach court, and convictions are rare. Most cases fail because the creator cannot be identified, the content crosses borders, or platforms do not cooperate quickly. The problem is not lack of harm, but the system’s inability to respond in time.
Over 70 percent of people cannot reliably distinguish a deepfake video from a real one.
When shown side-by-side comparisons, even trained professionals fail. Eyewitness testimony is already fragile.
Visual evidence is next.
Consent, Identity, and the Theft of the Self
Surveys of women journalists, politicians, and public-facing professionals indicate that over half fear deepfake-based sexual or reputational abuse. Many report self-censorship, withdrawal from online spaces, or avoidance of public participation altogether. The chilling effect is measurable and disproportionate, creating indirect exclusion rather than overt censorship.
The most brutal use of deepfakes is not political propaganda. It is sexual humiliation.
Women disproportionately bear this violence. Their faces are grafted onto pornographic bodies, circulated anonymously, consumed endlessly. The law often responds with indifference: No physical contact occurred. No explicit law was broken.
But something fundamental was stolen.
Identity is not merely biological. It is reputational. Psychological. Social. When a deepfake erases the line between the real and the fabricated, it hijacks a person’s presence in the world. The harm is not hypothetical. Victims lose jobs, families, safety, and sometimes their will to exist.
Existing legal frameworks struggle because they were built around tangible harm. Assault required touch. Defamation required false statements made knowingly. Privacy violations required intrusion.
Deepfakes exploit the gaps between these definitions.
The Law’s Old Language in a New World
Most legal systems are attempting to retrofit outdated statutes rather than confront the philosophical rupture deepfakes represent.
In India, the Information Technology Act, 2000 was never written with synthetic media in mind. Its provisions on impersonation and identity theft gesture toward protection, but enforcement is reactive, fragmented, and technologically illiterate.
Globally, the General Data Protection Regulation recognizes biometric data as sensitive, yet struggles when the data is generated, not collected. Who owns a face that never existed but looks exactly like yours?
Courts are trapped between two fears: overregulation that stifles innovation, and under regulation that normalizes abuse. In this paralysis, harm accelerates.
Law, by nature, moves slower than technology. But here, slowness is not neutrality. It is permission.
Misinformation as a Weapon, Not a Bug
Deepfakes are not just lies. They are strategic lies.
During elections, a single manipulated clip released hours before voting can shift outcomes irrevocably. By the time it is debunked, the damage is complete. Retractions do not travel as far as outrage. Truth does not trend as well as spectacle.
Democracy assumes an informed electorate. Deepfakes poison that assumption.
Institutions like the Election Commission of India can issue guidelines, but enforcement across millions of devices, platforms, and anonymous accounts borders on fantasy. The battlefield is asymmetric. Bad actors need one viral moment. The law needs certainty, procedure, jurisdiction, and time.
Time is exactly what deepfakes weaponize.
Election monitoring bodies and researchers observe sharp spikes in manipulated political audio and video in the final days before voting. These materials often circulate faster than any legal or factual correction can reach the public. Temporary belief, even if later disproven, is sufficient to distort democratic choice, making post-facto remedies largely symbolic.
Platforms, Profits, and Plausible Deniability
Social media companies insist they are neutral intermediaries. They are not.
Algorithms amplify what provokes emotion. Deepfakes provoke rage, fear, desire, and tribal loyalty. They are algorithmic gold.
Companies like Meta Platforms and others deploy detection tools, but these systems are often opaque, underpowered, and reactive. Content is removed after virality, not before harm.
Legal immunity frameworks were designed when platforms hosted content, not when they shaped reality. The distinction between publisher and intermediary has collapsed in practice, even if law pretends otherwise.
Accountability remains outsourced to victims who must prove damage in systems stacked against them.
Evidence, Courts, and the Crisis of Proof
There is currently no uniform judicial standard for authenticating AI-generated or AI-altered digital evidence. Courts differ widely in admissibility thresholds, forensic expectations, and technical understanding. This results in the same piece of evidence being accepted in one jurisdiction and rejected in another, turning justice into a matter of geography and resources rather than principle.
The courtroom has always relied on the visual. Surveillance footage. Recordings. Photographs. Deepfakes corrode this foundation.
Soon, every piece of digital evidence will require forensic validation. Smaller courts, underfunded systems, and developing jurisdictions simply cannot keep up. Justice becomes a privilege of resources.
Worse, the fear of falsification may lead courts to distrust legitimate evidence, allowing the powerful to escape accountability. When truth becomes expensive to prove, it becomes optional.
The law must now decide: Do we treat digital media as presumptively unreliable, or do we invest in systems that can authenticate reality itself?
Both choices are costly. Neither can be avoided.
Beyond Criminal Law: The Need for Ethical Architecture
Criminalization alone will not solve this crisis. By the time a deepfake qualifies as a crime, the harm is already irreversible.
What is needed is preventive legal architecture:
- Mandatory watermarking of AI-generated media
- Clear consent regimes for biometric likeness
- Strict liability for platforms that fail to act after notice
- Rapid response judicial mechanisms for takedowns
Most importantly, we need a shift in legal philosophy. The law must recognize synthetic harm as real harm. Emotional devastation. Reputational death. Democratic corrosion.
These are not abstract injuries. They are the currency of the digital age. The Question We Are Afraid to Ask
At its core, the deepfake crisis forces an uncomfortable question:
What happens to law when reality itself becomes disputable?
Legal systems depend on shared facts. Without them, power fills the vacuum. Whoever controls narrative, controls outcome. Truth becomes a matter of influence, not evidence.
This is not dystopian fiction. It is unfolding now.
The danger is not that machines will replace humans. The danger is that humans will stop believing each other.
Conclusion: Choosing What We Protect
Every legal era is defined by what it chooses to defend.
The industrial age protected labour.
The post-war era protected dignity.
The digital age must decide whether it protects truth.
Deepfakes are not merely a technological evolution. They are a test of whether law can adapt without losing its soul. Whether accountability can survive in a world where anyone can be everywhere, saying anything, with no body attached.
If the law fails here, it will not fail loudly. It will fail quietly, one fake video at a time, until reality itself feels negotiable.
And when that happens, justice will not collapse.
It will simply stop being recognizable.




