From 22aba6dd8a010bf4d9d9d4492657c238359bb365 Mon Sep 17 00:00:00 2001 From: Reinier van der Leer Date: Wed, 14 Feb 2024 11:17:34 +0100 Subject: [PATCH] fix(agent/llm): Include bad response in parse-fix prompt in `OpenAIProvider.create_chat_completion` Apparently I forgot to also append the response that caused the parse error before throwing it back to the LLM and letting it fix its mistake(s). --- autogpts/autogpt/autogpt/core/resource/model_providers/openai.py | 1 + 1 file changed, 1 insertion(+) diff --git a/autogpts/autogpt/autogpt/core/resource/model_providers/openai.py b/autogpts/autogpt/autogpt/core/resource/model_providers/openai.py index 494035c73..3f7457e8a 100644 --- a/autogpts/autogpt/autogpt/core/resource/model_providers/openai.py +++ b/autogpts/autogpt/autogpt/core/resource/model_providers/openai.py @@ -466,6 +466,7 @@ class OpenAIProvider( extras={"assistant_msg": assistant_msg, "i_attempt": attempts}, ) if attempts < self._configuration.fix_failed_parse_tries: + model_prompt.append(assistant_msg) model_prompt.append( ChatMessage.system(f"ERROR PARSING YOUR RESPONSE:\n\n{e}") )