Identifying hate speech in audio or video files
Our Content Moderation model can help you ensure that your content is safe and appropriate for all audiences.
The model pinpoints sensitive discussions in spoken data and provides information on the severity to which they occurred.
In this guide, we’ll learn how to use the Content Moderation model, and look at an example response to understand its structure.
Get started
Before we begin, make sure you have an AssemblyAI account and an API key. You can sign up for a free account and get your API key from your dashboard.
The complete source code for this guide can be viewed here.
Here is an audio example for this guide:
Step-by-step instructions
Set up the API endpoint and headers. The headers should include your API key.
Python SDK
Python (requests)
TypeScript
PHP
Ruby
C#
Python SDK
Create a TranscriptionConfig
with content_safety
set to True
.
Python SDK
Python (requests)
TypeScript
PHP
Ruby
C#
Python SDK
Create a Transcriber
object and pass in the configuration.
Python SDK
Python (requests)
TypeScript
PHP
Ruby
C#
Understanding the response
In the JSON response, there’ll be an additional key called content_safety_labels
that contains information about any sensitive content detected. The full text is contained in the text
key, and each problematic utterance has its own labels
and timestamp
. The entire audio is assigned a summary
and a severity_score_summary
for each category of unsafe content. Each label is returned with a confidence score and a severity score.
For more information, see Content Moderation model documentation and API reference.
Conclusion
The AssemblyAI API supports many different content safety labels. Identifying hate speech is only a single, important use case for automated content moderation, and you can learn about others on the AssemblyAI blog.