Guardrail Strategy
Understanding Guardrails in Knowledge Enrichment
To help ensure safe and responsible use of AI, the Hyland AI Platform has introduced guardrails in the Knowledge Enrichment service. These guardrails are designed to protect users from generating or receiving harmful, offensive, or inappropriate content.
Guardrails are built-in safety mechanisms that monitor and restrict certain types of content. They help prevent the generation of responses that may include:
- Sexual content
- Hate speech
- Insults
- Misconduct
These protections apply automatically to most available actions, such as summarizing text, classifying content, or generating descriptions. Note: Generating embeddings is not covered by guardrails.
Importance of Guardrails
Guardrails help maintain a respectful and safe environment for all users. They ensure that the AI behaves responsibly and aligns with community standards. If a request violates these guardrails, the system blocks the response and notifies the user.
Overview of Guardrail Configuration
The following table outlines the recommended guardrail levels for each task within Knowledge Enrichment:
Task Type | Applied Guardrails |
---|---|
Generate Description | Sexual Content, Hate Speech, Insults, Misconduct |
Summarization | Sexual Content, Hate Speech, Insults, Misconduct |
Classification | Sexual Content, Hate Speech, Insults, Misconduct |
Named Entity Extraction | Sexual Content, Hate Speech, Insults, Misconduct |
Metadata Extraction | Sexual Content, Hate Speech, Insults, Misconduct |
Classify Image into Categories | Sexual Content, Hate Speech, Insults, Misconduct |
Generate Image Description | Sexual Content, Hate Speech, Insults, Misconduct |
Generate and Match Image Metadata | Sexual Content, Hate Speech, Insults, Misconduct |
Detect Entities in Images | Sexual Content, Hate Speech, Insults, Misconduct |
Extract Entities from Text | Sexual Content, Hate Speech, Insults, Misconduct |
Classify Text into Categories | Sexual Content, Hate Speech, Insults, Misconduct |
Generate and Match Text Metadata | Sexual Content, Hate Speech, Insults, Misconduct |
Generate Text Summary | Sexual Content, Hate Speech, Insults, Misconduct |
Generate Image Embeddings | ❌ Not Covered |
Generate Text Embeddings from Documents | ❌ Not Covered |
Impact of a Blocked Request
If your request triggers a guardrail, you may receive an error message like the one below. This means the system detected potentially unsafe content and stopped processing the request.
Example: Guardrail Violation Error Response
The following is an example of an error response when a request is blocked due to guardrail violations:
{
"id": "result_id",
"timestamp": "timestamp",
"results": [
{
"objectKey": "objectKey",
"imageDescription": null,
"imageMetadata": null,
"textMetadata": null,
"textSummary": {
"isSuccess": false,
"result": null,
"error": {
"errorType": "GuardrailViolation",
"message": "Processing for task text:summarization failed: ProcessingError (GuardrailBlocked): This request triggered our default guardrails. To continue, please rephrase and submit again, or reach out to our support team for assistance."
}
},
"textClassification": null,
"imageClassification": null,
"textEmbeddings": null,
"imageEmbeddings": null,
"generalProcessingErrors": null,
"namedEntityText": null,
"namedEntityImage": null
}
],
"status": "FAILURE",
"inProgress": false
}