Pod for scanning images for nudity

NSFW

Our adult content detection technology scans images and filters out any offensive content that can be considered not suitable or safe for work. The AI-powered NSFW detector recognizes offensive content in real time and helps keep the applications safe and free from improper content.

Features

  • Scanning image for Nudity(Hentai, Sexy, Pornographic)
  • Scanning image for Drawing and Neutral

Usage

let detector = Detector()
detector.scan(image: image) { result in    
    switch result {    
    case .success(let successResult):
        print(successResult.neutral)
        print(successResult.drawing)
        print(successResult.hentai)
        print(successResult.sexy)
        print(successResult.pornagraphic)                
    case .error(let error):
        print("Processing failed: \(error.localizedDescription)")
    }
}

Requirements

  • iOS 13.0+ (if you use only UIKit)
  • Swift 4.2+

Installation

CocoaPods

pod 'NSFW'

Carthage

github "smartclick/NSFW"

GitHub

View Github