meta/llama-guard-4-12b
Moderate text prompts, model responses, and images for safety compliance. Accepts text with optional multiple images and...
Found 7 models (showing 1-7)
Moderate text prompts, model responses, and images for safety compliance. Accepts text with optional multiple images and...
Moderate LLM prompts and responses for safety policy compliance. Accepts a user prompt and/or an assistant reply as text...
Moderate text prompts and assistant responses for safety and policy compliance. Accepts a user message (prompt) and/or a...
Classify text prompts and assistant responses for safety. Accepts a user message and/or assistant reply and outputs a sa...
Classify the safety of multimodal inputs (image and user message) for content moderation. Accepts an image (required) an...
Classify text against custom safety policies with rationale. Accepts a plain‑English policy and a text input, and return...
Moderate text by classifying user prompts and optional assistant responses as Safe, Unsafe, or Controversial. Accepts a...