Merge remote-tracking branch 'upstream/master' into browse_scrape_links_test_and_validate

pull/780/head
Itamar Friedman 2023-04-12 12:18:16 +03:00
commit 1a7159095a
4 changed files with 9 additions and 4 deletions

3
.gitignore vendored
View File

@ -7,9 +7,10 @@ package-lock.json
auto_gpt_workspace/* auto_gpt_workspace/*
*.mpeg *.mpeg
.env .env
venv/* *venv/*
outputs/* outputs/*
ai_settings.yaml ai_settings.yaml
.vscode .vscode
.idea/*
auto-gpt.json auto-gpt.json
log.txt log.txt

View File

@ -6,7 +6,7 @@ To contribute to this GitHub project, you can follow these steps:
2. Clone the repository to your local machine using the following command: 2. Clone the repository to your local machine using the following command:
``` ```
git clone https://github.com/Torantulino/Auto-GPT git clone https://github.com/<YOUR-GITHUB-USERNAME>/Auto-GPT
``` ```
3. Create a new branch for your changes using the following command: 3. Create a new branch for your changes using the following command:

View File

@ -58,7 +58,7 @@ Your support is greatly appreciated
## 📋 Requirements ## 📋 Requirements
- [Python 3.8 or later](https://www.tutorialspoint.com/how-to-install-python-in-windows) - [Python 3.8 or later](https://www.tutorialspoint.com/how-to-install-python-in-windows)
- OpenAI API key - [OpenAI API key](https://platform.openai.com/account/api-keys)
- [PINECONE API key](https://www.pinecone.io/) - [PINECONE API key](https://www.pinecone.io/)
Optional: Optional:

View File

@ -266,6 +266,7 @@ def prompt_user():
def parse_arguments(): def parse_arguments():
"""Parses the arguments passed to the script""" """Parses the arguments passed to the script"""
global cfg global cfg
cfg.set_debug_mode(False)
cfg.set_continuous_mode(False) cfg.set_continuous_mode(False)
cfg.set_speak_mode(False) cfg.set_speak_mode(False)
@ -297,6 +298,9 @@ def parse_arguments():
print_to_console("GPT4 Only Mode: ", Fore.GREEN, "ENABLED") print_to_console("GPT4 Only Mode: ", Fore.GREEN, "ENABLED")
cfg.set_fast_llm_model(cfg.smart_llm_model) cfg.set_fast_llm_model(cfg.smart_llm_model)
if args.debug:
print_to_console("Debug Mode: ", Fore.GREEN, "ENABLED")
cfg.set_debug_mode(True)
# TODO: fill in llm values here # TODO: fill in llm values here
@ -388,7 +392,7 @@ while True:
f"COMMAND = {Fore.CYAN}{command_name}{Style.RESET_ALL} ARGUMENTS = {Fore.CYAN}{arguments}{Style.RESET_ALL}") f"COMMAND = {Fore.CYAN}{command_name}{Style.RESET_ALL} ARGUMENTS = {Fore.CYAN}{arguments}{Style.RESET_ALL}")
# Execute command # Execute command
if command_name.lower().startswith( "error" ): if command_name is not None and command_name.lower().startswith( "error" ):
result = f"Command {command_name} threw the following error: " + arguments result = f"Command {command_name} threw the following error: " + arguments
elif command_name == "human_feedback": elif command_name == "human_feedback":
result = f"Human feedback: {user_input}" result = f"Human feedback: {user_input}"