Pod for scanning images for nudity
NSFW
Our adult content detection technology scans images and filters out any offensive content that can be considered not suitable or safe for work. The AI-powered NSFW detector recognizes offensive content in real time and helps keep the applications safe and free from improper content.
Features
- Scanning image for Nudity(Hentai, Sexy, Pornographic)
- Scanning image for Drawing and Neutral
Usage
let detector = Detector()
detector.scan(image: image) { result in
switch result {
case .success(let successResult):
print(successResult.neutral)
print(successResult.drawing)
print(successResult.hentai)
print(successResult.sexy)
print(successResult.pornagraphic)
case .error(let error):
print("Processing failed: \(error.localizedDescription)")
}
}
Requirements
- iOS 13.0+ (if you use only UIKit)
- Swift 4.2+
Installation
CocoaPods
pod 'NSFW'
Carthage
github "smartclick/NSFW"