AI’s recent focus has been its potential use in helping doctors diagnose and treat cancer, but AI technology is being misused to create nude images of minors that are then distributed online in TikTok and WhatsApp groups – including targeting teachers and students themselves. Two Miami middle school students were arrested last December for doctoring pictures of female classmates; these charges resulted in third-degree felonies for promoting sexually explicit material, sparking national discussions over this problem of deepfake images.
Images created and shared are made using “generative AI,” an app which uses an algorithm to alter images to make someone appear nude. Some of these apps are free while others require subscription, with free ones usually having limited numbers of nude images per month or watermarks or ads; paid apps provide higher-quality images as well as greater customization options.
One of the easiest and most widely-used generative AI apps, Nudeitnow, transforms clothed photos into nude images by simulating what someone would look like without clothing – all available on both iOS and Android devices with just a few clicks needed to upload one and transform into naked image; user can select their body type, adjust amount of skin showing and even add backgrounds or images as needed.
Undress-app.ai is another generative AI app, which allows users to digitally unclothe a subject in photos taken using ClothOff software and generate a nude version that can then be saved to an SD card or shared through social media. Though originally intended for security and safety reasons, many are using it create lewd images of their peers; similar laws have already been proposed and introduced in some states against its use.
As use of generative AI increases in schools, educators and parents need to be aware of its associated risks and take measures to stop students using these apps in the classroom by informing them about them and monitoring student devices or school networks for nude AI tools.
Misogyny and child sexual abuse have long been issues, yet new laws and technologies do little to solve this issue. Generative AI will continue to create and share lewd images of children and teens; therefore it is crucial that we target any developers knowingly creating software which facilitates such practices instead of simply trying to stop teenage users themselves.
Keep in mind that using nude AI may have legal ramifications, including copyright infringement and invasion of privacy violations. Users can minimize these risks by complying with local regulations and seeking consent prior to engaging with such technologies; should any violations arise they should notify appropriate authorities immediately.https://undress-app.ai/