diff --git a/.env.template b/.env.template index 525cd61c..01735615 100644 --- a/.env.template +++ b/.env.template @@ -2,13 +2,14 @@ PINECONE_API_KEY=your-pinecone-api-key PINECONE_ENV=your-pinecone-region OPENAI_API_KEY=your-openai-api-key ELEVENLABS_API_KEY=your-elevenlabs-api-key -SMART_LLM_MODEL="gpt-4" -FAST_LLM_MODEL="gpt-3.5-turbo" +SMART_LLM_MODEL=gpt-4 +FAST_LLM_MODEL=gpt-3.5-turbo GOOGLE_API_KEY= CUSTOM_SEARCH_ENGINE_ID= USE_AZURE=False -OPENAI_API_BASE=your-base-url-for-azure -OPENAI_API_VERSION=api-version-for-azure -OPENAI_DEPLOYMENT_ID=deployment-id-for-azure +OPENAI_AZURE_API_BASE=your-base-url-for-azure +OPENAI_AZURE_API_VERSION=api-version-for-azure +OPENAI_AZURE_DEPLOYMENT_ID=deployment-id-for-azure IMAGE_PROVIDER=dalle -HUGGINGFACE_API_TOKEN= \ No newline at end of file +HUGGINGFACE_API_TOKEN= +USE_MAC_OS_TTS=False diff --git a/README.md b/README.md index 118131e4..749c8791 100644 --- a/README.md +++ b/README.md @@ -59,7 +59,7 @@ Your support is greatly appreciated ## 📋 Requirements - [Python 3.8 or later](https://www.tutorialspoint.com/how-to-install-python-in-windows) - OpenAI API key -- PINECONE API key +- [PINECONE API key](https://www.pinecone.io/) Optional: - ElevenLabs Key (If you want the AI to speak) @@ -92,8 +92,8 @@ pip install -r requirements.txt 4. Rename `.env.template` to `.env` and fill in your `OPENAI_API_KEY`. If you plan to use Speech Mode, fill in your `ELEVEN_LABS_API_KEY` as well. - Obtain your OpenAI API key from: https://platform.openai.com/account/api-keys. - - Obtain your ElevenLabs API key from: https://beta.elevenlabs.io. You can view your xi-api-key using the "Profile" tab on the website. - - If you want to use GPT on an Azure instance, set `USE_AZURE` to `True` and provide the `OPENAI_API_BASE`, `OPENAI_API_VERSION` and `OPENAI_DEPLOYMENT_ID` values as explained here: https://pypi.org/project/openai/ in the `Microsoft Azure Endpoints` section + - Obtain your ElevenLabs API key from: https://elevenlabs.io. You can view your xi-api-key using the "Profile" tab on the website. + - If you want to use GPT on an Azure instance, set `USE_AZURE` to `True` and provide the `OPENAI_AZURE_API_BASE`, `OPENAI_AZURE_API_VERSION` and `OPENAI_AZURE_DEPLOYMENT_ID` values as explained here: https://pypi.org/project/openai/ in the `Microsoft Azure Endpoints` section ## 🔧 Usage @@ -179,8 +179,7 @@ MEMORY_INDEX=whatever ## 🌲 Pinecone API Key Setup -Pinecone enable a vector based memory so a vast memory can be stored and only relevant memories -are loaded for the agent at any given time. +Pinecone enables the storage of vast amounts of vector-based memory, allowing for only relevant memories to be loaded for the agent at any given time. 1. Go to app.pinecone.io and make an account if you don't already have one. 2. Choose the `Starter` plan to avoid being charged. diff --git a/scripts/ai_config.py b/scripts/ai_config.py index 8fabf3dc..1d5832c1 100644 --- a/scripts/ai_config.py +++ b/scripts/ai_config.py @@ -34,7 +34,7 @@ class AIConfig: @classmethod def load(cls: object, config_file: str=SAVE_FILE) -> object: """ - Returns class object with parameters (ai_name, ai_role, ai_goals) loaded from yaml file if yaml file exists, + Returns class object with parameters (ai_name, ai_role, ai_goals) loaded from yaml file if yaml file exists, else returns class with no parameters. Parameters: @@ -42,7 +42,7 @@ class AIConfig: config_file (int): The path to the config yaml file. DEFAULT: "../ai_settings.yaml" Returns: - cls (object): A instance of given cls object + cls (object): A instance of given cls object """ try: @@ -61,11 +61,11 @@ class AIConfig: """ Saves the class parameters to the specified file yaml file path as a yaml file. - Parameters: + Parameters: config_file(str): The path to the config yaml file. DEFAULT: "../ai_settings.yaml" Returns: - None + None """ config = {"ai_name": self.ai_name, "ai_role": self.ai_role, "ai_goals": self.ai_goals} @@ -76,7 +76,7 @@ class AIConfig: """ Returns a prompt to the user with the class information in an organized fashion. - Parameters: + Parameters: None Returns: diff --git a/scripts/ai_functions.py b/scripts/ai_functions.py index dc774b37..8ad77441 100644 --- a/scripts/ai_functions.py +++ b/scripts/ai_functions.py @@ -27,7 +27,7 @@ def evaluate_code(code: str) -> List[str]: def improve_code(suggestions: List[str], code: str) -> str: """ - A function that takes in code and suggestions and returns a response from create chat completion api call. + A function that takes in code and suggestions and returns a response from create chat completion api call. Parameters: suggestions (List): A list of suggestions around what needs to be improved. diff --git a/scripts/browse.py b/scripts/browse.py index c15214e7..09f376a7 100644 --- a/scripts/browse.py +++ b/scripts/browse.py @@ -5,12 +5,21 @@ from llm_utils import create_chat_completion cfg = Config() +# Define and check for local file address prefixes +def check_local_file_access(url): + local_prefixes = ['file:///', 'file://localhost', 'http://localhost', 'https://localhost'] + return any(url.startswith(prefix) for prefix in local_prefixes) + def scrape_text(url): """Scrape text from a webpage""" # Most basic check if the URL is valid: if not url.startswith('http'): return "Error: Invalid URL" + # Restrict access to local files + if check_local_file_access(url): + return "Error: Access to local files is restricted" + try: response = requests.get(url, headers=cfg.user_agent_header) except requests.exceptions.RequestException as e: @@ -126,4 +135,4 @@ def summarize_text(text, question): max_tokens=300, ) - return final_summary \ No newline at end of file + return final_summary diff --git a/scripts/commands.py b/scripts/commands.py index 073ecc56..ce5d04ff 100644 --- a/scripts/commands.py +++ b/scripts/commands.py @@ -42,9 +42,6 @@ def get_command(response): # Use an empty dictionary if 'args' field is not present in 'command' object arguments = command.get("args", {}) - if not arguments: - arguments = {} - return command_name, arguments except json.decoder.JSONDecodeError: return "Error:", "Invalid JSON" diff --git a/scripts/config.py b/scripts/config.py index 50432c42..24911bce 100644 --- a/scripts/config.py +++ b/scripts/config.py @@ -37,7 +37,7 @@ class Config(metaclass=Singleton): self.continuous_mode = False self.speak_mode = False - self.fast_llm_model = os.getenv("FAST_LLM_MODEL", "gpt-3.5-turbo") + self.fast_llm_model = os.getenv("FAST_LLM_MODEL", "gpt-3.5-turbo") self.smart_llm_model = os.getenv("SMART_LLM_MODEL", "gpt-4") self.fast_token_limit = int(os.getenv("FAST_TOKEN_LIMIT", 4000)) self.smart_token_limit = int(os.getenv("SMART_TOKEN_LIMIT", 8000)) @@ -46,15 +46,18 @@ class Config(metaclass=Singleton): self.use_azure = False self.use_azure = os.getenv("USE_AZURE") == 'True' if self.use_azure: - self.openai_api_base = os.getenv("OPENAI_API_BASE") - self.openai_api_version = os.getenv("OPENAI_API_VERSION") - self.openai_deployment_id = os.getenv("OPENAI_DEPLOYMENT_ID") + self.openai_api_base = os.getenv("OPENAI_AZURE_API_BASE") + self.openai_api_version = os.getenv("OPENAI_AZURE_API_VERSION") + self.openai_deployment_id = os.getenv("OPENAI_AZURE_DEPLOYMENT_ID") openai.api_type = "azure" openai.api_base = self.openai_api_base openai.api_version = self.openai_api_version self.elevenlabs_api_key = os.getenv("ELEVENLABS_API_KEY") + self.use_mac_os_tts = False + self.use_mac_os_tts = os.getenv("USE_MAC_OS_TTS") + self.google_api_key = os.getenv("GOOGLE_API_KEY") self.custom_search_engine_id = os.getenv("CUSTOM_SEARCH_ENGINE_ID") diff --git a/scripts/main.py b/scripts/main.py index 0bca5b37..a64a7241 100644 --- a/scripts/main.py +++ b/scripts/main.py @@ -1,6 +1,7 @@ import json import random import commands as cmd +import utils from memory import get_memory import data import chat @@ -16,9 +17,18 @@ from ai_config import AIConfig import traceback import yaml import argparse +import logging cfg = Config() +def configure_logging(): + logging.basicConfig(filename='log.txt', + filemode='a', + format='%(asctime)s,%(msecs)d %(name)s %(levelname)s %(message)s', + datefmt='%H:%M:%S', + level=logging.DEBUG) + return logging.getLogger('AutoGPT') + def check_openai_api_key(): """Check if the OpenAI API key is set in config.py or as an environment variable.""" if not cfg.openai_api_key: @@ -29,7 +39,6 @@ def check_openai_api_key(): print("You can get your key from https://beta.openai.com/account/api-keys") exit(1) - def print_to_console( title, title_color, @@ -39,10 +48,12 @@ def print_to_console( max_typing_speed=0.01): """Prints text to the console with a typing effect""" global cfg + global logger if speak_text and cfg.speak_mode: speak.say_text(f"{title}. {content}") print(title_color + title + " " + Style.RESET_ALL, end="") if content: + logger.info(title + ': ' + content) if isinstance(content, list): content = " ".join(content) words = content.split() @@ -133,12 +144,12 @@ def load_variables(config_file="config.yaml"): # Prompt the user for input if config file is missing or empty values if not ai_name: - ai_name = input("Name your AI: ") + ai_name = utils.clean_input("Name your AI: ") if ai_name == "": ai_name = "Entrepreneur-GPT" if not ai_role: - ai_role = input(f"{ai_name} is: ") + ai_role = utils.clean_input(f"{ai_name} is: ") if ai_role == "": ai_role = "an AI designed to autonomously develop and run businesses with the sole goal of increasing your net worth." @@ -148,7 +159,7 @@ def load_variables(config_file="config.yaml"): print("Enter nothing to load defaults, enter nothing when finished.") ai_goals = [] for i in range(5): - ai_goal = input(f"Goal {i+1}: ") + ai_goal = utils.clean_input(f"Goal {i+1}: ") if ai_goal == "": break ai_goals.append(ai_goal) @@ -181,7 +192,7 @@ def construct_prompt(): Fore.GREEN, f"Would you like me to return to being {config.ai_name}?", speak_text=True) - should_continue = input(f"""Continue with the last settings? + should_continue = utils.clean_input(f"""Continue with the last settings? Name: {config.ai_name} Role: {config.ai_role} Goals: {config.ai_goals} @@ -216,7 +227,7 @@ def prompt_user(): "Name your AI: ", Fore.GREEN, "For example, 'Entrepreneur-GPT'") - ai_name = input("AI Name: ") + ai_name = utils.clean_input("AI Name: ") if ai_name == "": ai_name = "Entrepreneur-GPT" @@ -231,7 +242,7 @@ def prompt_user(): "Describe your AI's role: ", Fore.GREEN, "For example, 'an AI designed to autonomously develop and run businesses with the sole goal of increasing your net worth.'") - ai_role = input(f"{ai_name} is: ") + ai_role = utils.clean_input(f"{ai_name} is: ") if ai_role == "": ai_role = "an AI designed to autonomously develop and run businesses with the sole goal of increasing your net worth." @@ -243,7 +254,7 @@ def prompt_user(): print("Enter nothing to load defaults, enter nothing when finished.", flush=True) ai_goals = [] for i in range(5): - ai_goal = input(f"{Fore.LIGHTBLUE_EX}Goal{Style.RESET_ALL} {i+1}: ") + ai_goal = utils.clean_input(f"{Fore.LIGHTBLUE_EX}Goal{Style.RESET_ALL} {i+1}: ") if ai_goal == "": break ai_goals.append(ai_goal) @@ -281,7 +292,7 @@ def parse_arguments(): if args.debug: print_to_console("Debug Mode: ", Fore.GREEN, "ENABLED") - cfg.set_debug_mode(True) + cfg.set_debug_mode(True) if args.gpt3only: print_to_console("GPT3.5 Only Mode: ", Fore.GREEN, "ENABLED") @@ -292,6 +303,7 @@ def parse_arguments(): # TODO: fill in llm values here check_openai_api_key() cfg = Config() +logger = configure_logging() parse_arguments() ai_name = "" prompt = construct_prompt() @@ -341,7 +353,7 @@ while True: f"Enter 'y' to authorise command, 'y -N' to run N continuous commands, 'n' to exit program, or enter feedback for {ai_name}...", flush=True) while True: - console_input = input(Fore.MAGENTA + "Input:" + Style.RESET_ALL) + console_input = utils.clean_input(Fore.MAGENTA + "Input:" + Style.RESET_ALL) if console_input.lower() == "y": user_input = "GENERATE NEXT COMMAND JSON" break diff --git a/scripts/speak.py b/scripts/speak.py index 10dd7c07..5d1e153c 100644 --- a/scripts/speak.py +++ b/scripts/speak.py @@ -39,9 +39,15 @@ def gtts_speech(text): playsound("speech.mp3") os.remove("speech.mp3") +def macos_tts_speech(text): + os.system(f'say "{text}"') + def say_text(text, voice_index=0): if not cfg.elevenlabs_api_key: - gtts_speech(text) + if cfg.use_mac_os_tts == 'True': + macos_tts_speech(text) + else: + gtts_speech(text) else: success = eleven_labs_speech(text, voice_index) if not success: diff --git a/scripts/utils.py b/scripts/utils.py new file mode 100644 index 00000000..5039796f --- /dev/null +++ b/scripts/utils.py @@ -0,0 +1,8 @@ +def clean_input(prompt: str=''): + try: + return input(prompt) + except KeyboardInterrupt: + print("You interrupted Auto-GPT") + print("Quitting...") + exit(0) + diff --git a/tests/test_browse_scrape_text.py b/tests/test_browse_scrape_text.py index 27ebc0f6..bfc3b425 100644 --- a/tests/test_browse_scrape_text.py +++ b/tests/test_browse_scrape_text.py @@ -37,7 +37,7 @@ Additional aspects: class TestScrapeText: - # Tests that scrape_text() returns the expected text when given a valid URL. + # Tests that scrape_text() returns the expected text when given a valid URL. def test_scrape_text_with_valid_url(self, mocker): # Mock the requests.get() method to return a response with expected text expected_text = "This is some sample text" @@ -50,7 +50,7 @@ class TestScrapeText: url = "http://www.example.com" assert scrape_text(url) == expected_text - # Tests that the function returns an error message when an invalid or unreachable url is provided. + # Tests that the function returns an error message when an invalid or unreachable url is provided. def test_invalid_url(self, mocker): # Mock the requests.get() method to raise an exception mocker.patch("requests.get", side_effect=requests.exceptions.RequestException) @@ -60,7 +60,7 @@ class TestScrapeText: error_message = scrape_text(url) assert "Error:" in error_message - # Tests that the function returns an empty string when the html page contains no text to be scraped. + # Tests that the function returns an empty string when the html page contains no text to be scraped. def test_no_text(self, mocker): # Mock the requests.get() method to return a response with no text mock_response = mocker.Mock() @@ -72,7 +72,7 @@ class TestScrapeText: url = "http://www.example.com" assert scrape_text(url) == "" - # Tests that the function returns an error message when the response status code is an http error (>=400). + # Tests that the function returns an error message when the response status code is an http error (>=400). def test_http_error(self, mocker): # Mock the requests.get() method to return a response with a 404 status code mocker.patch('requests.get', return_value=mocker.Mock(status_code=404)) @@ -83,7 +83,7 @@ class TestScrapeText: # Check that the function returns an error message assert result == "Error: HTTP 404 error" - # Tests that scrape_text() properly handles HTML tags. + # Tests that scrape_text() properly handles HTML tags. def test_scrape_text_with_html_tags(self, mocker): # Create a mock response object with HTML containing tags html = "
This is bold text.
" @@ -96,4 +96,4 @@ class TestScrapeText: result = scrape_text("https://www.example.com") # Check that the function properly handles HTML tags - assert result == "This is bold text." \ No newline at end of file + assert result == "This is bold text."