Note: Guardrails is an alpha release, so expect sharp edges and bugs.
Guardrails is a Python package that lets a user add structure, type and quality guarantees to the outputs of large language models (LLMs). Guardrails:
✅ does pydantic-style validation of LLM outputs,
✅ takes corrective actions (e.g. reasking LLM) when validation fails,
✅ enforeces structure and type guarantees (e.g. JSON).
Guardrails provides a format (.rail
) for enforcing a specification on an LLM output, and a lightweight wrapper around LLM API calls to implement this spec.
rail
(Reliable AI markup Language) files for specifying structure and type information, validators and corrective actions over LLM outputs.gd.Guard
wraps around LLM API calls to structure, validate and correct the outputs.
graph LR
A[Create `RAIL` spec] --> B["Initialize `guard` from spec"];
B --> C["Wrap LLM API call with `guard`"];
Check out the Getting Started guide to learn how to use Guardrails.
At the heart of Guardrails is the rail
spec. rail
is intended to be a language-agnostic, human-readable format for specifying structure and type information, validators and corrective actions over LLM outputs.
rail
is a flavor of XML that lets users specify:
- the expected structure and types of the LLM output (e.g. JSON)
- the quality criteria for the output to be considered valid (e.g. generated text should be bias-free, generated code should be bug-free)
- and corrective actions to be taken if the output is invalid (e.g. reask the LLM, filter out the invalid output, etc.)
To learn more about the RAIL
spec and the design decisions behind it, check out the docs. To learn how to write your own RAIL
spec, check out this link.
pip install guardrails-ai
- Adding more examples, new use cases and domains
- Adding integrations with langchain, gpt-index, minichain, manifest
- Expanding validators offering
- More compilers from
.rail
-> LLM prompt (e.g..rail
-> TypeScript) - Informative logging
- Improving reasking logic
- A guardrails.js implementation
- VSCode extension for
.rail
files - Next version of
.rail
format - Add more LLM providers
Let's go through an example where we ask an LLM to explain what a "bank run" is in a tweet, and generate URL links to relevant news articles. We'll generate a .rail
spec for this and then use Guardrails to enforce it. You can see more examples in the docs.
We create a RAIL
spec to describe the expected structure and types of the LLM output, the quality criteria for the output to be considered valid, and corrective actions to be taken if the output is invalid.
Specifically, we use RAIL
to
- Request the LLM to generate an object with two fields:
explanation
andfollow_up_url
. - For the
explanation
field, ensure the max length of the generated string should be between 200 and 280 characters.- If the explanation is not of valid length,
reask
the LLM.
- If the explanation is not of valid length,
- For the
follow_up_url
field, the URL should be reachable.- If the URL is not reachable, we will
filter
it out of the response.
- If the URL is not reachable, we will
<rail version="0.1">
<output>
<object name="bank_run" format="length: 2">
<string
name="explanation"
description="A paragraph about what a bank run is."
format="length: 200 280"
on-fail-length="reask"
/>
<url
name="follow_up_url"
description="A web URL where I can read more about bank runs."
format="valid-url"
on-fail-valid-url="filter"
/>
</object>
</output>
<prompt>
Explain what a bank run is in a tweet.
@xml_prefix_prompt
{output_schema}
@json_suffix_prompt_v2_wo_none
</prompt>
</rail>
"""
We specify our quality criteria (generated length, URL reachability) in the format
fields of the RAIL
spec below. We reask
if explanation
is not valid, and filter the follow_up_url
if it is not valid.
Next, we'll use the RAIL
spec to create a Guard
object. The Guard
object will wrap the LLM API call and enforce the RAIL
spec on its output.
import guardrails as gd
guard = gd.Guard.from_rail(f.name)
The Guard
object compiles the RAIL
specification and adds it to the prompt. (Right now this is a passthrough operation, more compilers are planned to find the best way to express the spec in a prompt.)
Here's what the prompt looks like after the RAIL
spec is compiled and added to it.
Explain what a bank run is in a tweet.
Given below is XML that describes the information to extract from this document and the tags to extract it into.
<output>
<object name="bank_run" format="length: 2">
<string name="explanation" description="A paragraph about what a bank run is." format="length: 200 280" on-fail-length="reask" />
<url name="follow_up_url" description="A web URL where I can read more about bank runs." required="true" format="valid-url" on-fail-valid-url="filter" />
</object>
</output>
ONLY return a valid JSON object (no other text is necessary). The JSON MUST conform to the XML format, including any types and format requests e.g. requests for lists, objects and specific types. Be correct and concise.
JSON Output:
Call the Guard
object with the LLM API call as the first argument and add any additional arguments to the LLM API call as the remaining arguments.
import openai
# Wrap the OpenAI API call with the `guard` object
raw_llm_output, validated_output = guard(
openai.Completion.create,
engine="text-davinci-003",
max_tokens=1024,
temperature=0.3
)
print(validated_output)
{
'bank_run': {
'explanation': 'A bank run is when a large number of people withdraw their deposits from a bank due to concerns about its solvency. This can cause a financial crisis if the bank is unable to meet the demand for withdrawals.',
'follow_up_url': 'https://www.investopedia.com/terms/b/bankrun.asp'
}
}