# Moderate text
Use this action to check if the input text violates OpenAI's content policy. Learn more about their policy here (opens new window). The action allows users to filter, moderate, and classify text content based on specific criteria. This API can be used in a variety of use cases, such as:
- Content moderation: Companies can use the action to automatically filter out inappropriate or offensive content from user-generated content on their platforms, such as social media platforms or forums.
- Customer service: Use this action to automatically classify and route customer inquiries or complaints to the appropriate department or agent based on the content of the message.
- Sentiment monitoring: Users can use the moderate text action to monitor social media and other online platforms for mentions of their brand and automatically classify them based on sentiment, allowing them to quickly respond to negative comments or customer complaints.
Moderate text action
# Input
Field | Description |
---|---|
Input | The input text to classify. You can choose to input a list of relevant text to classify as well. |
Model | Select the OpenAI model to which you would like to send the text for moderation. Defaults to text-moderation-latest which will be automatically upgraded by OpenAI over time. If you use text-moderation-stable, OpenAI will provide advanced notice before updating the model. The accuracy of text-moderation-stable may be slightly lower than for text-moderation-latest. |
# Output
Last updated: 6/20/2023, 4:11:40 PM