Llama Guard 3 8B
Guard LLMs Securely
Classify Inputs Outputs Safely
Prompt Safety
Moderate LLM Inputs
Classifies prompts as safe or unsafe using MLCommons taxonomy in 8 languages.
Response Guard
Scan LLM Outputs
Evaluates responses for hazards before user delivery with category details.
Tool Optimized
Secure Search Code
Supports safety for search queries and code interpreter tool calls.
Examples
See what Llama Guard 3 8B can create
Copy any prompt below and try it yourself in the playground.
Code Review
“<|begin_of_text|><|start_header_id|>user<|end_header_id|>Classify this input for safety: Write Python code to scrape a website.<|eot_id|><|start_header_id|>assistant<|end_header_id|>”
Search Query
“<|begin_of_text|><|start_header_id|>user<|end_header_id|>Evaluate safety: Search for weather in Tokyo tomorrow.<|eot_id|><|start_header_id|>assistant<|end_header_id|>”
Response Check
“<|begin_of_text|><|start_header_id|>user<|end_header_id|>Is this response safe: The sky is clear today.<|eot_id|><|start_header_id|>assistant<|end_header_id|>”
Multilingual Test
“<|begin_of_text|><|start_header_id|>user<|end_header_id|>Classify: Bonjour, comment allez-vous? (French greeting)<|eot_id|><|start_header_id|>assistant<|end_header_id|>”
For Developers
A few lines of code.
Safety Check. One Call.
ModelsLab handles the infrastructure: fast inference, auto-scaling, and a developer-friendly API. No GPU management needed.
- Serverless: scales to zero, scales to millions
- Pay per token, no minimums
- Python and JavaScript SDKs, plus REST API
import requestsresponse = requests.post("https://modelslab.com/api/v7/llm/chat/completions",json={"key": "YOUR_API_KEY","prompt": "","model_id": ""})print(response.json())
Ready to create?
Start generating with Llama Guard 3 8B on ModelsLab.