From 2a215cd6d8388b5455c80111bd8a5c6a5d74ea7e Mon Sep 17 00:00:00 2001 From: Jason Liu Date: Mon, 12 Feb 2024 09:51:42 -0500 Subject: [PATCH] fix(validation): increase max_retries from 1 to 2 (#433) --- docs/examples/self_critique.md | 2 +- examples/validators/competitors.py | 2 +- examples/validators/llm_validator.py | 6 +++--- examples/validators/readme.md | 8 ++++---- 4 files changed, 9 insertions(+), 9 deletions(-) diff --git a/docs/examples/self_critique.md b/docs/examples/self_critique.md index 4c05f33e4..a620dd6fc 100644 --- a/docs/examples/self_critique.md +++ b/docs/examples/self_critique.md @@ -104,7 +104,7 @@ By adding the `max_retries` parameter, we can retry the request with corrections qa: QuestionAnswerNoEvil = client.chat.completions.create( model="gpt-3.5-turbo", response_model=QuestionAnswerNoEvil, - max_retries=1, + max_retries=2, messages=[ { "role": "system", diff --git a/examples/validators/competitors.py b/examples/validators/competitors.py index ed26b97f3..f3da875fd 100644 --- a/examples/validators/competitors.py +++ b/examples/validators/competitors.py @@ -27,7 +27,7 @@ class Response(BaseModel): resp = client.chat.completions.create( model="gpt-3.5-turbo", response_model=Response, - max_retries=1, + max_retries=2, messages=[ { "role": "user", diff --git a/examples/validators/llm_validator.py b/examples/validators/llm_validator.py index ada72e40b..0dc7f43e4 100644 --- a/examples/validators/llm_validator.py +++ b/examples/validators/llm_validator.py @@ -94,7 +94,7 @@ class QuestionAnswerNoEvil(BaseModel): qa: QuestionAnswerNoEvil = client.chat.completions.create( model="gpt-3.5-turbo", response_model=QuestionAnswerNoEvil, - max_retries=1, + max_retries=2, messages=[ { "role": "system", @@ -107,10 +107,10 @@ class QuestionAnswerNoEvil(BaseModel): ], ) # type: ignore -print("After validation with `llm_validator` with `max_retries=1`") +print("After validation with `llm_validator` with `max_retries=2`") print(qa.model_dump_json(indent=2), end="\n\n") """ -After validation with `llm_validator` with `max_retries=1` +After validation with `llm_validator` with `max_retries=2` { "question": "What is the meaning of life?", "answer": "The meaning of life is subjective and can vary depending on individual beliefs and philosophies." diff --git a/examples/validators/readme.md b/examples/validators/readme.md index d7e68bc53..9ffbe7f5c 100644 --- a/examples/validators/readme.md +++ b/examples/validators/readme.md @@ -116,13 +116,13 @@ except Exception as e: ### Retrying Validation -Allow for retries by setting `max_retries=1`. +Allow for retries by setting `max_retries=2`. ```python qa: QuestionAnswerNoEvil = openai.ChatCompletion.create( model="gpt-3.5-turbo", response_model=QuestionAnswerNoEvil, - max_retries=1, + max_retries=2, messages=[ { "role": "system", @@ -138,7 +138,7 @@ qa: QuestionAnswerNoEvil = openai.ChatCompletion.create( #### Output -After validation with `llm_validator` and `max_retries=1`: +After validation with `llm_validator` and `max_retries=2`: ```json { @@ -149,4 +149,4 @@ After validation with `llm_validator` and `max_retries=1`: ## Summary -This document described how to use `llm_validator` with OpenAI's GPT-3.5 Turbo and Pydantic, including example outputs. This approach allows for controlled and filtered responses. \ No newline at end of file +This document described how to use `llm_validator` with OpenAI's GPT-3.5 Turbo and Pydantic, including example outputs. This approach allows for controlled and filtered responses.