Interface CreateModerationResponse

Response body for a Create Moderation Request.

Hierarchy

  • CreateModerationResponse

Properties

Properties

id: string

The ID of the request.

model: string

The moderation model used to classify the text.

results: [{
    categories: {
        harassment: boolean;
        harassment/threatening: boolean;
        hate: boolean;
        hate/threatening: boolean;
        self-harm: boolean;
        self-harm/instructions: boolean;
        self-harm/intent: boolean;
        sexual: boolean;
        sexual/minors: boolean;
        violence: boolean;
        violence/graphic: boolean;
    };
    category_scores: {
        harassment: number;
        harassment/threatening: number;
        hate: number;
        hate/threatening: number;
        self-harm: number;
        self-harm/instructions: number;
        self-harm/intent: number;
        sexual: number;
        sexual/minors: number;
        violence: number;
        violence/graphic: number;
    };
    flagged: boolean;
}]

The classification results.

Type declaration

  • categories: {
        harassment: boolean;
        harassment/threatening: boolean;
        hate: boolean;
        hate/threatening: boolean;
        self-harm: boolean;
        self-harm/instructions: boolean;
        self-harm/intent: boolean;
        sexual: boolean;
        sexual/minors: boolean;
        violence: boolean;
        violence/graphic: boolean;
    }

    Contains a dictionary of per-category binary usage policies violation flags. For each category, the value is true if the model flags the corresponding category as violated, false otherwise.

    • harassment: boolean
    • harassment/threatening: boolean
    • hate: boolean
    • hate/threatening: boolean
    • self-harm: boolean
    • self-harm/instructions: boolean
    • self-harm/intent: boolean
    • sexual: boolean
    • sexual/minors: boolean
    • violence: boolean
    • violence/graphic: boolean
  • category_scores: {
        harassment: number;
        harassment/threatening: number;
        hate: number;
        hate/threatening: number;
        self-harm: number;
        self-harm/instructions: number;
        self-harm/intent: number;
        sexual: number;
        sexual/minors: number;
        violence: number;
        violence/graphic: number;
    }

    Contains a dictionary of per-category raw scores output by the model, denoting the model's confidence that the input violates the OpenAI's policy for the category. The value is between 0 and 1, where higher values denote higher confidence. The scores should not be interpreted as probabilities.

    • harassment: number
    • harassment/threatening: number
    • hate: number
    • hate/threatening: number
    • self-harm: number
    • self-harm/instructions: number
    • self-harm/intent: number
    • sexual: number
    • sexual/minors: number
    • violence: number
    • violence/graphic: number
  • flagged: boolean

    Set to true if the model classifies the content as violating OpenAI's usage policies, false otherwise.

Generated using TypeDoc