NudeNet: An ensemble of Neural Nets for Nudity Detection and Censoring

Praneeth Bedapudi
Mar 30, 2019 · 6 min read

Note: This post can also be read from here

Part 1: Nudity detection with image classification

With the advent of excellent DL libraries and plethora of open-source implementations and papers, Image Classification is very easy to implement. That is, if you have the dataset. There are plenty of open datasets available to test and refine classification models. But obtaining a task specific dataset is tough. Nudity/ NSFW detection is one such use-case where there are no practically useful open datasets available.

In the first part of this two part project, I collect data for and implement nudity detection using Image Classification. My goal is to build a good open source dataset for this use-case and provide a pre-trained model for the same. In the second part I aim to implement and open-source private/ exposed part detection using Object Detection.

The Why: I sincerely believe in not re-inventing the wheel. At the time of starting this project, Yahoo’s open_nsfw is the only decently working nudity detection module available. It is quiet outdated and the data isn’t available to public. There is no (AFAIK) open-source project available for censoring of exposed parts.

Collecting Nude images: For getting these images, I started with Reddit. I made a list of NSFW sub-reddits mostly collected from the website scrolller (thanks to the hard work of /u/faroix). After using the excellent RipMe application to download ~1000 images from each of these sub-reddits and going through them (not recommended for the lighthearted), I found a major problem with this data.
Almost all the images collected from these sub-reddits are of very high quality. In fact, combined these images came out to ~260 GB. This isn’t ideal, since a lot of porn is of potato quality. To balance this out, I crawled thumbnails of videos from PornHub.
While I was doing this, another open-source enthusiast GantMan open-sourced a model trained on the data collected by alexkimxyz. So, I contacted GantMan and obtained the dataset he used.

After using the above two commands to reduce normalize the images and remove duplicates, I ended up with 1,78,601 from PornHub, 1,21,644 from Reddit and 1,30,266 from GantMan’s dataset.

Collecting Safe images: After I was done with collection of nude images, I though that the tough part was over and I couldn’t be more wrong. Collecting some random images would be very easy, but doing so will result in the classifier learning to classify all humans into nude category. Since, neural networks try to minimize loss, not having a lot of human images in the “safe” category will result in our model being a “Human detector” (which is an easy task) rather than being a “nudity detector”.

For building a good classifier, I needed lot of non-nude images that had people in them. After going through GantMan’s sfw data, I observed that though the total number of images is decent, the number of images with people in them was very less.
After some thinking, I realized that Facebook is a great place to collect images with people in them. So, I crawled Facebook profile pictures using their Graph API. I also made a list of safe sub-reddits and crawled ~1000 images from each one using ripme. After resizing and cleaning, I ended up with 68,948 from Facebook, 98,359 from GantMan’s dataset and 55,137 from Reddit.

Processing Pipeline: I used the excellent image augmentation library Augmentor with some added fail-safes for on the fly image augmentation to use with Keras’s fit_generator.

The following snippet is the augmentation used for training data.

The Training: For the training I rented a machine with GTX 1080Ti, with 12GB of vram and 64GB system memory from for 0.11$ per hour. I was able to to use a batch size of 32 with Xception and 256x256 input image size.

Note: The implementations of image classification models provided in Keras’s applications, do not use any type of regularization. For adding regularization (dropout or l2) loop over each layer and add regularization.

Using SGD with momentum, the model converges at 0.9347 accuracy on GantMan’s data.

Evaluation: Finding the right data for evaluating a nudtiy detection module is very tough. Nudity/ Porn has thousands of variations (See Rule 34) and building a comprehensive test data is next to impossible. But, just to get a simple idea of how this is working, I use the data collected by Aditya at Note that, even this is not a good test data. The best way to test and improve something like this is by community help. Since, his test data is in different categories, I map the categories “nsfw_porn”, “nsfw_explicit_nudity”, “nsfw_simulated_porn” to “nude” category and the category “sfw” to “safe”.

Image for post
Image for post
Precision and Recall of NudeNet’s classifier

For testing with GantMan’s test data, I map the the classes “hentai”, “porn” to “nude” category and “drawings”, “neutral”, “sexy” to safe category. This is because, according to his definition of classes nudes of people will be labelled as “porn” and cartoon nudes will be labelled as “hentai”.

Image for post
Image for post
Precision and Recall of GantMan’s nsfw_model
Image for post
Image for post
Precision and Recall of Yahoo’s Open NSFW

Funnily enough, All the three projects get similar scores, although they fail at different images. For example, GantMan’s model funnily fails with Jeff Goldblum’s images where as NudeNet doesn’t. Similarly, NudeNet fails with some images, for which GantMan’s model doesn’t. In the couple of months of working on this project, I came across a lot of images for which Google’s cloud vision api fails miserably. Because most of these images are NSFW (duh!), I am unable to add them here. If you want to take a look at some of these examples, please leave a comment.

Update: I was able to find another manually curated dataset for testing. This dataset is available at .

Image for post
Image for post
NudeNet precision and recall on the dataturks dataset
Image for post
Image for post
GantMan’s nsfw_model precision and recall on the dataturks dataset

The project can be found at

The pre-trained models at

To install and use NudeNet, take a look at the following snippet.

# Using the classifier
from NudeNet import NudeClassifier
classifier = NudeClassifier('classifier_checkpoint_path')

In the second part of this post, I implement Exposed Part Detection and Censoring using Object Detection. Please find the second part of this post at

Edit: The data is available

Welcome to a place where words matter. On Medium, smart voices and original ideas take center stage - with no ads in sight. Watch

Follow all the topics you care about, and we’ll deliver the best stories for you to your homepage and inbox. Explore

Get unlimited access to the best stories on Medium — and support writers while you’re at it. Just $5/month. Upgrade

Get the Medium app

A button that says 'Download on the App Store', and if clicked it will lead you to the iOS App store
A button that says 'Get it on, Google Play', and if clicked it will lead you to the Google Play store