• Powered by
ContextKnow better. Do better.

Auction of Muslim women on Indian app shows tech weaponised for abuse

The logos of mobile apps, Google, Amazon, Facebook, Apple and Netflix, are displayed on a screen in this illustration picture taken December 3, 2019. REUTERS/Regis Duvignau/File Photo

The logos of mobile apps, Google, Amazon, Facebook, Apple and Netflix, are displayed on a screen in this illustration picture taken December 3, 2019. REUTERS/Regis Duvignau/File Photo

What’s the context?

Women increasingly targeted online by new technologies that make it easy to track and abuse anonymously

  • Scores of Muslim women featured in app that was taken down
  • Technologies such as deepfake, tracking used to harass women
  • Victims struggle to be taken seriously, get justice

By Rina Chandran

Six months ago, pilot Hana Khan saw her picture on an app that appeared to be auctioning scores of Muslim women in India. The app was quickly taken down, no one was charged, and the issue shelved - until a similar app popped up on New Year's Day.

Khan was not on the new app called Bulli Bai - a slur for Muslim women - that was hawking activists, journalists, an actor, politicians and Nobel Laureate Malala Yousafzai as maids.

Amid growing outrage, the app was taken down, and four suspects arrested this week.

The fake auctions that were shared widely on social media are just the latest examples of how technology is being used - often with ease, speed and little expense - to put women at risk through online abuse, theft of privacy or sexual exploitation.

For Muslim women in India who are often abused online, it is an everyday risk, even as they use social media to call out hatred and discrimination against their minority community.

Sneha Chakraborty posing with her artwork. Artists like Sneha are fast gaining a following in the crypto space./Photos courtesy of Sneha Chakraborty
Go DeeperCrypto queens: Women stake out space in blockchain world
The social audio app Clubhouse is seen on a mobile phone in this illustration picture taken February 8, 2021. REUTERS/Florence Lo/Illustration
Go DeeperArabs talk sex and taboos on Clubhouse app, but for how long?
A photo illustration shows a chain and a padlock in front of a displayed Whatsapp logo January 13, 2017. REUTERS/Dado Ruvic/Illustration
Go DeeperIraq Yazidis want Big Tech held to account for Islamic State crimes

"When I saw my picture on the app, my world shook. I was upset and angry that someone could do this to me, and I became angrier as I realised this nameless person was getting away with it," said Khan, who filed a police complaint against the first app, Sulli Deals, another pejorative term for Muslim women.

"This time, I felt so much dread and despair that it was happening again to my friends, to Muslim women like me. I don't know how to make it stop," Khan, a commercial pilot in her 30s, told the Thomson Reuters Foundation.

Mumbai police said they were investigating whether the Bulli Bai app was "part of a larger conspiracy".

A spokesperson for GitHub, which hosted both apps, said it had "longstanding policies against content and conduct involving harassment, discrimination, and inciting violence.

"We suspended a user account following the investigation of reports of such activity, all of which violate our policies."

Misconception

Advances in technology have heightened risks for women across the world, be it trolling or doxxing with their personal details revealed, surveillance cameras, location tracking, or deepfake pornographic videos featuring doctored images.

Deepfakes - or artificial, intelligence-generated, synthetic media - are used to create porn, with apps that let users strip clothes off women or swap their faces into explicit videos.

Digital abuse of women is pervasive because "everybody has a device and a digital presence," said Adam Dodge, chief executive of EndTAB, a U.S.-based nonprofit tackling tech-enabled abuse.

"The violence has become easier to perpetrate, as you can get at somebody anywhere in the world. The order of magnitude of harm is also greater because you can upload something and show it to the world in a matter of seconds," he said.

"And there is a permanency to it because that photo or video exists forever online," he added.

The emotional and psychological impact of such abuse is "just as excruciating" as physical abuse, with the effects compounded by the virality, public nature, and permanence of the content online, said Noelle Martin, an Australian activist.

At 17, Martin discovered her image had been photoshopped into pornographic images and distributed. Her campaign against image-based abuse helped change the law in Australia.

But victims struggle to be heard, she said.

"There is a dangerous misconception that the harms of technology-facilitated abuse are not as real, serious, or potentially lethal as abuse with a physical element," she said.

"For victims, this misconception makes speaking out, seeking support, and accessing justice much more difficult."

Persecution

Tracking lone creators and rogue coders is hard, and technology platforms tend to shield anonymous users who can easily create a fake email or social media profile.

Even lawmakers are not spared: in November, the U.S. House of Representatives censured Republican Paul Gosar over a photoshopped anime video that showed him killing Democrat Alexandra Ocasio-Cortez. He then retweeted the video.

"With any new technology we should immediately be thinking about how and when it will be misused and weaponised to harm girls and women online," said Dodge.

"Technology platforms have created a very imbalanced atmosphere for victims of online abuse, and the traditional ways of seeking help when we are harmed in the physical world are not as available when the abuse occurs online," he said .

Some technology firms are taking action.

Following reports that its AirTags - locator devices that can be attached to keys and wallets - were being used to track women, Apple launched an app to help users shield their privacy.

In India, the women on the auction apps are still shaken.

Ismat Ara, a journalist showcased on Bulli Bai, called it "nothing short of online harassment."

It was "violent, threatening and intending to create a feeling of fear and shame in my mind, as well as in the minds of women in general and the Muslim community," Ara said in a police complaint that she posted on social media.

Arfa Khanum Sherwani, also featured for sale, wrote on Twitter: "The auction may be fake but the persecution is real."


Context is powered by the Thomson Reuters Foundation Newsroom.

Our Standards: Thomson Reuters Trust Principles


Tags

Gender equity
Polarisation
Race and inequality
Tech and inequality
Tech regulation
Underground economies
Data rights
Corporate responsibility

FEATURED PODCAST

An illustration photo shows the globe with a tree standing on top. On the left hand side, a red backed illustration shows barren trees and oil refinery towers. On the right hand side, a green backed illustration shows wind turbines and solar panels. A sound equaliser image crosses the screen to indicates audio.
6 EPISODES
Podcast

Just Transition

The human stories behind the shift to a green economy

Just Transition promo image
Podcast




Get our data & surveillance newsletter. Free. Every week.

By providing your email, you agree to our Privacy Policy.


Today On Context