🤖 Model
lucataco/prompt-guard-86m
Detect prompt injection and jailbreak attempts in text inputs. Classify a string as benign, injection, and/or jailbreak...
Found 2 models (showing 1-2)
Detect prompt injection and jailbreak attempts in text inputs. Classify a string as benign, injection, and/or jailbreak...
Moderate text prompts and assistant responses for safety and policy compliance. Accepts a user message (prompt) and/or a...