Nudity Detection

Image Moderation / Introduction

Introduction to Image Moderation

Sightengine's Image Moderation API is used to moderate Images and detect whether they contain unwanted content (such as adult content, offensive content, commercial text, children, weapons...).

The API works both with standard images (such as JPG, PNG, WebP...) and animated multi-frame images such as GIF images.

How the API works

Sightengine Image Moderation is very straightforward to use:

  1. You submit an image to the Sightengine API, along with the list of models and/or image lists you want to check the image against
  2. You immediately get a detailed response from the API giving you a description of what was found (if anything), along with moderation scores.

All needed data is given in the API response. There are no callbacks, no moderation queues, no need to wait for updates or track state.

Head to our Quickstart guide to start using the API with a few lines of code, or head to the API documentation for a deeper look into the structure of requests and responses.


A Model is a filter that will look for a specific type of content in your Image. For instance, nudity is the name of a model that has been trained to look for any adult content, racy content, suggestive content and specifically flags scenes ranging from explicit to mild nudity.

By specifying the list of Models you wish to apply to an image, you tell the API what you would like to detect and filter.

Head to our Model reference to see all the available Models along with their detection capabilities.

Was this page helpful?