Smartclick

NSFW

Our adult content detection technology scans images and filters out any offensive content that can be considered not suitable or safe for work. The AI-powered NSFW detector recognizes offensive content in real time and helps keep the applications safe and free from improper content.

Features

  • Scanning image for Nudity(Hentai, Sexy, Pornographic)
  • Scanning image for Drawing and Neutral

Usage

let detector = Detector()
detector.scan(image: image) { result in    
    switch result {    
    case .success(let successResult):
        print(successResult.neutral)
        print(successResult.drawing)
        print(successResult.hentai)
        print(successResult.sexy)
        print(successResult.pornagraphic)                
    case .error(let error):
        print("Processing failed: \(error.localizedDescription)")
    }
}

Requirements

  • iOS 13.0+ (if you use only UIKit)
  • Swift 4.2+

Installation

CocoaPods

pod 'NSFW'

Carthage

github "smartclick/NSFW"

GitHub

View Github