Guardrails
Last updated
Last updated
Without guardrails, the LLM application simply takes a prompt as input, processes it through the LLM, and generates an output.
With guardrails, there are additional components:
An Input Guard that can filter out personal identifiable information (PII), detect jailbreak attempts, and identify off-topic prompts before they reach the LLM.
After the LLM generates output, there is an Output Guard that scans for hallucinations, competitor mentions, and profanity.
When the answer is YES, the bot should refuse to respond by returning this message: