Content Moderation
Identify content that may be inappropriate or offensive.
What is Content Moderation?
Content moderation reviews content to ensure it meets certain standards or guidelines and flags violations. This can include identifying content that may be inappropriate, offensive, or harmful.
The path to use content moderation:
Content Moderation Topics
By default, content moderation flags for the following topics:
- Accidents: Any man-made incident that happens unexpectedly and results in damage, injury, or death.
- Alcohol: Content that discusses any alcoholic beverage or its consumption.
- Crime Violence: Content that discusses any type of criminal activity or extreme violence that is criminal in nature.
- Drugs: Content that discusses illegal drugs or their usage.
- Gambling: Gambling on casino-based games such as poker, slots, etc. as well as sports betting.
- Hate Speech: Content that’s a direct attack against people or groups based on their sexual orientation, gender identity, race, religion, ethnicity, national origin, disability, etc.
- Health Issues: Content that discusses any medical or health-related problems.
- Marijuana: Content that discusses marijuana or its usage.
- Natural Disasters: Phenomena that happens infrequently and results in damage, injury, or death. Such as hurricanes, tornadoes, earthquakes, volcano eruptions, and firestorms.
- Negative News: News content with a negative sentiment which typically occur in the third person as an unbiased recapping of events.
- NSFW (Adult Content): Content considered “Not Safe for Work” and consists of content that a viewer would not want to be heard/seen in a public environment.
- Pornography: Content that discusses any sexual content or material.
- Profanity: Any profanity or cursing.
- Sensitive Social Issues: Content that may be considered insensitive, irresponsible, or harmful to certain groups based on their beliefs, political affiliation, sexual orientation, or gender identity.
- Terrorism: Terrorist acts as well as terrorist groups. Examples include bombings, mass shootings, and ISIS. Note that many texts corresponding to this topic may also be classified into the crime violence topic.
- Tobacco: Content that discusses tobacco and tobacco usage, including e-cigarettes, nicotine, vaping, and general discussions about smoking.
- Weapons: Content that discusses any type of weapon including guns, ammunition, shooting, knives, missiles, torpedoes, etc.
Running Content Moderation
To run content moderation:
Content moderation only runs on one video at a time.
Limiting Moderation Topics
To limit the topics that are moderated, you can use the topics
parameter. For example, to only moderate content related to drugs, you can use the following code:
Using Custom Prompts
To use a custom moderation prompt:
Content Moderation Response Format
The response format for content moderation:
Top-level container holding the list of policy-violation objects detected in the video
A violation item contains the following:
Category of the policy violation (e.g., “Gambling”)
Category of the policy violation (e.g., “Gambling”)
Start timestamp of the violating segment (HH:MM:SS)
End timestamp of the violating segment (HH:MM:SS)
Confidence (0 – 1) that the detection is correct
Relevance (0 – 1) of the violation to the policy category
Modalities contributing to the detection (e.g., “visual”, “text on screen”)
Severity level of the violation (e.g., “low”, “medium”, “high”)
Human-readable summary of the violation
For example:
Content Moderation Complete Example
A complete content moderation example:
Content Moderation JSON
TODO