Skip to main content

Guardrail Strategy

Understanding Guardrails in Knowledge Enrichment

To help ensure safe and responsible use of AI, the Hyland AI Platform has introduced guardrails in the Knowledge Enrichment service. These guardrails are designed to protect users from generating or receiving harmful, offensive, or inappropriate content.

Guardrails are built-in safety mechanisms that monitor and restrict certain types of content. They help prevent the generation of responses that may include:

  • Sexual content
  • Hate speech
  • Insults
  • Misconduct

These protections apply automatically to most available actions, such as summarizing text, classifying content, or generating descriptions. Note: Generating embeddings is not covered by guardrails.

Importance of Guardrails

Guardrails help maintain a respectful and safe environment for all users. They ensure that the AI behaves responsibly and aligns with community standards. If a request violates these guardrails, the system blocks the response and notifies the user.

Overview of Guardrail Configuration

The following table outlines the recommended guardrail levels for each task within Knowledge Enrichment:

Task TypeApplied Guardrails
Generate DescriptionSexual Content, Hate Speech, Insults, Misconduct
SummarizationSexual Content, Hate Speech, Insults, Misconduct
ClassificationSexual Content, Hate Speech, Insults, Misconduct
Named Entity ExtractionSexual Content, Hate Speech, Insults, Misconduct
Metadata ExtractionSexual Content, Hate Speech, Insults, Misconduct
Classify Image into CategoriesSexual Content, Hate Speech, Insults, Misconduct
Generate Image DescriptionSexual Content, Hate Speech, Insults, Misconduct
Generate and Match Image MetadataSexual Content, Hate Speech, Insults, Misconduct
Detect Entities in ImagesSexual Content, Hate Speech, Insults, Misconduct
Extract Entities from TextSexual Content, Hate Speech, Insults, Misconduct
Classify Text into CategoriesSexual Content, Hate Speech, Insults, Misconduct
Generate and Match Text MetadataSexual Content, Hate Speech, Insults, Misconduct
Generate Text SummarySexual Content, Hate Speech, Insults, Misconduct
Generate Image Embeddings❌ Not Covered
Generate Text Embeddings from Documents❌ Not Covered

Impact of a Blocked Request

If your request triggers a guardrail, you may receive an error message like the one below. This means the system detected potentially unsafe content and stopped processing the request.

Example: Guardrail Violation Error Response

The following is an example of an error response when a request is blocked due to guardrail violations:

{
"id": "result_id",
"timestamp": "timestamp",
"results": [
{
"objectKey": "objectKey",
"imageDescription": null,
"imageMetadata": null,
"textMetadata": null,
"textSummary": {
"isSuccess": false,
"result": null,
"error": {
"errorType": "GuardrailViolation",
"message": "Processing for task text:summarization failed: ProcessingError (GuardrailBlocked): This request triggered our default guardrails. To continue, please rephrase and submit again, or reach out to our support team for assistance."
}
},
"textClassification": null,
"imageClassification": null,
"textEmbeddings": null,
"imageEmbeddings": null,
"generalProcessingErrors": null,
"namedEntityText": null,
"namedEntityImage": null
}
],
"status": "FAILURE",
"inProgress": false
}