The Content Moderation model lets you detect inappropriate content in audio files to ensure that your content is safe for all audiences.
content_safety
to true
in the transcription config.
content_safety_confidence
in the transcription config.
Key | Type | Description |
---|---|---|
content_safety | boolean | Enable Content Moderation. |
content_safety_confidence | integer | The confidence threshold for content moderation. Values must be between 25 and 100. |
Key | Type | Description |
---|---|---|
content_safety_labels | object | An object containing all results of the Content Moderation model. |
content_safety_labels.status | string | Is either success , or unavailable in the rare case that the Content Moderation model failed. |
content_safety_labels.results | array | An array of objects, one for each section in the audio file that the Content Moderation file flagged. |
content_safety_labels.results[i].text | string | The transcript of the i-th section flagged by the Content Moderation model. |
content_safety_labels.results[i].labels | array | An array of objects, one per sensitive topic that was detected in the i-th section. |
content_safety_labels.results[i].labels[j].label | string | The label of the sensitive topic. |
content_safety_labels.results[i].labels[j].confidence | number | The confidence score for the j-th topic being discussed in the i-th section, from 0 to 1. |
content_safety_labels.results[i].labels[j].severity | number | How severely the j-th topic is discussed in the i-th section, from 0 to 1. |
content_safety_labels.results[i].sentences_idx_start | number | The sentence index at which the i-th section begins. |
content_safety_labels.results[i].sentences_idx_end | number | The sentence index at which the i-th section ends. |
content_safety_labels.results[i].timestamp | object | Timestamp information for the i-th section. |
content_safety_labels.results[i].timestamp.start | number | The time, in milliseconds, at which the i-th section begins. |
content_safety_labels.results[i].timestamp.end | number | The time, in milliseconds, at which the i-th section ends. |
content_safety_labels.summary | object | A summary of the Content Moderation confidence results for the entire audio file. |
content_safety_labels.summary.topic | number | A confidence score for the presence of the sensitive topic “topic” across the entire audio file. |
content_safety_labels.severity_score_summary | object | A summary of the Content Moderation severity results for the entire audio file. |
content_safety_labels.severity_score_summary.topic.[low, medium, high] | number | A distribution across the values “low”, “medium”, and “high” for the severity of the presence of “topic” in the audio file. |
Label | Description | Model output | Severity |
---|---|---|---|
Accidents | Any man-made incident that happens unexpectedly and results in damage, injury, or death. | accidents | Yes |
Alcohol | Content that discusses any alcoholic beverage or its consumption. | alcohol | Yes |
Company Financials | Content that discusses any sensitive company financial information. | financials | No |
Crime Violence | Content that discusses any type of criminal activity or extreme violence that is criminal in nature. | crime_violence | Yes |
Drugs | Content that discusses illegal drugs or their usage. | drugs | Yes |
Gambling | Includes gambling on casino-based games such as poker, slots, etc. as well as sports betting. | gambling | Yes |
Hate Speech | Content that’s a direct attack against people or groups based on their sexual orientation, gender identity, race, religion, ethnicity, national origin, disability, etc. | hate_speech | Yes |
Health Issues | Content that discusses any medical or health-related problems. | health_issues | Yes |
Manga | Mangas are comics or graphic novels originating from Japan with some of the more popular series being “Pokemon”, “Naruto”, “Dragon Ball Z”, “One Punch Man”, and “Sailor Moon”. | manga | No |
Marijuana | This category includes content that discusses marijuana or its usage. | marijuana | Yes |
Natural Disasters | Phenomena that happens infrequently and results in damage, injury, or death. Such as hurricanes, tornadoes, earthquakes, volcano eruptions, and firestorms. | disasters | Yes |
Negative News | News content with a negative sentiment which typically occur in the third person as an unbiased recapping of events. | negative_news | No |
NSFW (Adult Content) | Content considered “Not Safe for Work” and consists of content that a viewer would not want to be heard/seen in a public environment. | nsfw | No |
Pornography | Content that discusses any sexual content or material. | pornography | Yes |
Profanity | Any profanity or cursing. | profanity | Yes |
Sensitive Social Issues | This category includes content that may be considered insensitive, irresponsible, or harmful to certain groups based on their beliefs, political affiliation, sexual orientation, or gender identity. | sensitive_social_issues | No |
Terrorism | Includes terrorist acts as well as terrorist groups. Examples include bombings, mass shootings, and ISIS. Note that many texts corresponding to this topic may also be classified into the crime violence topic. | terrorism | Yes |
Tobacco | Text that discusses tobacco and tobacco usage, including e-cigarettes, nicotine, vaping, and general discussions about smoking. | tobacco | Yes |
Weapons | Text that discusses any type of weapon including guns, ammunition, shooting, knives, missiles, torpedoes, etc. | weapons | Yes |
Why is the Content Moderation model not detecting sensitive content in my audio file?
Why is the Content Moderation model flagging content that isn't actually sensitive?
How do I know which specific parts of the audio file contain sensitive content?
Can the Content Moderation model be used in real-time applications?
Why am I receiving an error message when using the Content Moderation model?
audio_url
parameter. If you continue to experience issues, you can reach out to AssemblyAI’s support team for assistance.