Added: Fabian Hilliker - Date: 09.08.2021 10:43 - Views: 36321 - Clicks: 7443
Free, easy and requiring just a single still photo, the deepfake bot has produced more thanfake pornographic images publicly posted online for anyone to see.
A free, easy-to-use deepfake bot found on the Telegram messenger app has victimized seemingly hundreds of thousands of women by replacing the clothed parts of their bodies in photos with nudity. More thanof these nonconsensual sexual images have been posted publicly online, but the bot has produced hundreds of thousands more that haven't been traced.
A website promoting the bot claimed that more thanimages of women have been manipulated to replace their clothing with nudity, as of Thursday, and that more thanabusers have ed images to the bot. Those s couldn't be independently verified. The victims are mostly private individuals, women whose photos were taken off social media or pulled from a personal stash of pics, according to a research report about the bot Tuesday, which traced more thanpublicly posted images of victims of this bot. Some victims had messenger nudes been photographed in bathing suits or underwear.
Some were wearing simple T-shirts and shorts. Some were visibly underage. All are women. Deepfake porn isn't new. Deepfake technology -- artificial intelligence that makes sophisticated media forgeries -- has been used early and often to fabricate pornography. But this Telegram bot takes the ease and access of this technology to a new level. Computer manipulation of media has existed for decades, and sexual imagery has been weaponized online for as long as the internet could host photos.
Whether it's nude photos posted without consent or crudely doctored forgeries, sexual images have been weaponized to extort, threaten, humiliate and harass victims. But only in the last few years has deepfake tech intensified the threat of manipulated sexual media, posing frightening implications for what may come. It's much more easy for somebody without the technical knowledge to make one," said Mary Anne Franksa law professor at the University of Miami and president of the online-abuse nonprofit Cyber Civil Rights Initiative.
With this Telegam bot, any woman who's ever posted a messenger nudes of herself from the waist up could be a potential victim. Even women out walking could be victimized if surreptitiously snapped by the wrong stranger. And in one of the most messenger nudes forms of abuse with this bot, photographs of children have been ed to the bot's AI, automatically manipulated to sexualize the child and then shared publicly. Neither Sensity's report nor this article are disclosing the name of the bot, to avoid amplifying it. CNET viewed galleries of images with the bot's watermark posted online and interacted with the bot itself, stopping short of ing any photos for it to manipulate.
Telegram's tenacious commitment to free speech and privacy may make bots like this challenging to stamp out. Telegram has been criticized for hosting terrorist propaganda and coordination, facilitating piracy and copyright infringement, and harboring varieties of predatory pornography. But the service has also taken actions to remove abuse, such as kicking off groups for violent extremists like neo-Nazis and ISIS. Sensity reached out to Telegram multiple times over the last six months about its findings.
Deepfake technology is like a high-speed Photoshop conveyor belt on steroids. Using a kind of artificial intelligence known as neural networks, deepfake tech messenger nudes generate media forgeries that make people appear to be doing or saying things they never did. The term deepfake is used most often with videos, but deepfakes can refer to any so-called "synthetic" media produced by deep machine learning, including pornographic still photos.
If this bot on Telegram sounds disturbingly familiar, a similar technology called DeepNude leaped to prominence last year, only to become so popular in a single day, after it was exposed in a news articlethat its programmer shut it down. Like the Telegram bot, DeepNude used artificial intelligence to automatically generate nonconsensual sexual images of women in photos, replacing their photographed clothing with nudity. Both, in effect, appear to "strip" victims of what they're wearing in their pictures. DeepNude was a website offering Windows and Linux apps that required some level of technical savvy to operate.
In fact, the AI powering the Telegram bot appears to be an open-source version of DeepNude's software. But the new bot is simpler and easier to use than the original desktop app, and it's available to anyone on Telegram. The bot will accept your first photo to manipulate after tapping just a messenger nudes prompts. The bot is also deed to make it easy for abusers to share the manipulated images by posting them in chats and other online forms.
These nonconsensual sexual images have "been put out there to be found," Patrini said. Those are actually exposed completely. Sensity foundimages of women that were victimized by the bot and then shared publicly, as of the end of July. While each image may not be of a unique individual, Patrini said instances of the same woman being victimized, or the same photo being manipulated repeatedly, were rare. The ,plus total messenger nudes images is limited to manipulated photos that were publicly posted and that Sensity was able to track down.
Sensity doesn't know the scope of material that is not shared, Patrini added. The bot's promotional website suggests that as many asimages have been manipulated by the bot. And the bot is growing in popularity. A year ago, about 1, images manipulated by the bot were posted in channels in a month.
In July, that had swelled to at least 24, images, according to Sensity. Aboutpeople are members of channels linked to the bot, Sensity found.
Telegram is used globally, but its roots are in Messenger nudes, and links to the Telegram bot posted on VK, Russia's dominant social network, are the most common way that abusers have found the bot. In a statement, VK said it doesn't tolerate such content or links on its platform and blocks communities that distribute them. The bot is built with a "freemium" business model, providing free users with a basic level of functionality and reserving advanced features for those who pay. It's the kind of user-friendly strategy that has helped legal, legitimate apps and games like Spotify and Fortnite become worldwide phenoms.
Abusers can use the bot free by sending photos to it one at a time, seemingly up to five a day. But "paid premium" features include sending multiple pics, skipping the line of free users and removing watermarks from the pornographic images they get in return. But the bot's business strategy is also ambitious, inspired by strategies from gaming and classic promotional tropes. In a gamefied turn, the premium features are paid for with virtual "coins.
They can also be earned, as rewards. One of the rewarded behaviors is recruiting new users. And because the app says that its virtual coins can be paid back in rubles, it effectively creates a system that pays abusers money in a government-issued currency for bringing in new abusers. Fortunately, the payouts are presumably meager: The value of the bot's coins are cheap, roughly five cents each. The bot's deer has also adopted classic promotional tactics. You can get a deeper discount on coins with the more of them you buy. The bot pitches new users with a one-time "beginner rate" special on coins.
The bot also underscores how a fixation on electoral messenger nudes misses wider damage caused by pornographic ones, which are much more common and already devastating victims. Researchers created deepfakes that graft candidates' faces onto impersonators' he, in order to test a system to debunk them.
A preoccupation with "the perfect deepfake" of a political candidate or world leader is the messenger nudes of disinformation that tends to stoke congressional hearings. But that overlooks the harm being caused to regular people, at a larger and rapidly increasing scale, where even a poor-quality deepfake is still deeply harmful. Even with a bot like this operating for months, Witness hasn't observed an explosion of that kind of harassment yet. But even vigilance is unlikely to result in justice for victims, Franks said, who pointed to a historical failure of our legal systems to address weaponized sexual imagery years ago.
We need to do better now," she said. Joan E. Stay in the know. Enlarge Image. Now playing: Watch this: We're not ready for the deepfake revolution.Messenger nudes
email: [email protected] - phone:(288) 987-8849 x 7725
Facebook message from a friend led to a hacker having access to a woman's nude pictures