
lucataco/nsfw_video_detection
Detect NSFW content in videos. Accepts a video and returns a single label: "normal" or "nsfw", with adjustable strictnes...
Found 15 models (showing 1-15)
Detect NSFW content in videos. Accepts a video and returns a single label: "normal" or "nsfw", with adjustable strictnes...
Detect NSFW content in images. Accepts an image and returns a binary label of "nsfw" or "normal" for content moderation...
Detect NSFW content in images for moderation. Accepts an image and returns category flags and confidence scores for nudi...
Moderate images and prompts for NSFW, public figures, and copyright risk. Accepts an image and/or text prompt and return...
Moderate text-to-image prompts by scoring NSFW/toxicity on a 0–10 safety scale. Takes a text prompt and returns an integ...
Classify text prompts, model responses, and multiple images for safety policy compliance. Accepts text and a list of ima...
Moderate LLM conversations by classifying user prompts and assistant responses as SAFE or UNSAFE and listing violated po...
Classify text for safety policy compliance. Takes a user prompt and/or an assistant response and returns a safe/unsafe l...
Detect hate speech and toxic language in text. Accepts a text string and returns probability scores for toxicity, severe...
Classify images for safety violations across sexually_explicit, dangerous_content, and violence_gore policies. Takes an...
Moderate images and accompanying user messages by classifying safety risks. Takes an image and optional text input; outp...
Detect NSFW content in images for content moderation. Takes a single image as input and returns a binary label (safe or...
Detect NSFW content in images. Analyze an input image with Stable Diffusion’s content filter and return a structured mod...
Moderate text prompts and assistant responses for safety policy compliance. Accepts a user prompt and/or an assistant me...
Moderate images for safety and policy compliance. Accepts an image (optional custom prompt) and returns structured JSON...