Nigeria, Dec. 14 -- When AI-generated nude images of singer Ayra Starr spread across X last month, the platform responded with remarkable indifference: no policy violated, nothing to see. Actress Kehinde Bankole endured the same violation earlier this year. Across Nigeria, women's faces and bodies are being digitally stripped, weaponised, and circulated in acts of violence that would trigger national outrage if they occurred offline. Instead, they vanish into a legal void. Our institutions pretend they cannot see what is happening. But the truth is starker: they have chosen not to look.

One victim told fact-checking site DUBAWA: "This post is causing more insults and humiliation to me, and as it spreads, it ruins my reputation." When she...