Content Moderation/Safety Models
Overview
With our API, you can use content moderation / safety models to classify input content as safe or unsafe instantly.
Model list
meta-llama/Meta-Llama-Guard-3-8B
open-source
Meta-Llama/Llama-Guard-7b
open-source
meta-llama/Llama-Guard-3-11B-Vision-Turbo
open-source
meta-llama/LlamaGuard-2-8b
open-source
Key Features
Text Analysis: Check text for security.
Image Analysis: Check image for security.
Flexible Input Methods: Supports both image URLs and base64 encoded images.
Multiple Image Inputs: Analyze multiple images in a single request.
Quick Start
Ensure you replace "my_key"
with your actual API key before running the code.
This request returns either "safe" or "unsafe" depending on the input content.
Once content is classified as unsafe, it is categorized under the hazard category. This process is unique to each model.
For example:
unsafe \n 04
Content moderation models are perfect for scenarios where content safety is crucial
moderate user-generated content on websites
filter harmful inputs in chatbots
safeguard sensitive systems from unsafe data
ensure compliance with safety standards in applications
Example
Last updated
Was this helpful?