name: Bug report 🐛 description: Create a bug report for Auto-GPT. labels: ['status: needs triage'] body: - type: markdown attributes: value: | ### ⚠️ Before you continue * Check out our [backlog], [roadmap] and join our [discord] to discuss what's going on * If you need help, you can ask in the [discussions] section or in [#tech-support] * **Throughly search the [existing issues] before creating a new one** [backlog]: https://github.com/orgs/Significant-Gravitas/projects/1 [roadmap]: https://github.com/orgs/Significant-Gravitas/projects/2 [discord]: https://discord.gg/autogpt [discussions]: https://github.com/Significant-Gravitas/Auto-GPT/discussions [#tech-support]: https://discord.com/channels/1092243196446249134/1092275629602394184 [existing issues]: https://github.com/Significant-Gravitas/Auto-GPT/issues?q=is%3Aissue - type: checkboxes attributes: label: ⚠️ Search for existing issues first ⚠️ description: > Please [search the history](https://github.com/Torantulino/Auto-GPT/issues) to see if an issue already exists for the same problem. options: - label: I have searched the existing issues, and there is no existing issue for my problem required: true - type: markdown attributes: value: | Please provide a searchable summary of the issue in the title above ⬆️. ⚠️ SUPER-busy repo, please help the volunteer maintainers. The less time we spend here, the more time we spend building AutoGPT. Please help us help you: - Does it work on `stable` branch (https://github.com/Torantulino/Auto-GPT/tree/stable)? - Does it work on current `master` (https://github.com/Torantulino/Auto-GPT/tree/master)? - Search for existing issues, "add comment" is tidier than "new issue" - Ask on our Discord (https://discord.gg/autogpt) - Provide relevant info: - Provide commit-hash (`git rev-parse HEAD` gets it) - If it's a pip/packages issue, provide pip version, python version - If it's a crash, provide traceback. - type: dropdown attributes: label: Which Operating System are you using? description: > Please select the operating system you were using to run Auto-GPT when this problem occurred. options: - Windows - Linux - MacOS - Docker - Devcontainer / Codespace - Windows Subsystem for Linux (WSL) - Other (Please specify in your problem) validations: required: true - type: dropdown attributes: label: GPT-3 or GPT-4? description: > If you are using Auto-GPT with `--gpt3only`, your problems may be caused by the [limitations](https://github.com/Significant-Gravitas/Auto-GPT/issues?q=is%3Aissue+label%3A%22AI+model+limitation%22) of GPT-3.5. options: - GPT-3.5 - GPT-4 validations: required: true - type: textarea attributes: label: Steps to reproduce 🕹 description: | **⚠️ Issues that we can't reproduce will be closed.** - type: textarea attributes: label: Current behavior 😯 description: Describe what happens instead of the expected behavior. - type: textarea attributes: label: Expected behavior 🤔 description: Describe what should happen. - type: textarea attributes: label: Your prompt 📝 description: > If applicable please provide the prompt you are using. Your prompt is stored in your `ai_settings.yaml` file. value: | ```yaml # Paste your prompt here ``` - type: textarea attributes: label: Your Logs 📒 description: | Please include the log showing your error and the command that caused it, if applicable. You can copy it from your terminal or from `logs/activity.log`. This will help us understand your issue better!
Example ```log INFO NEXT ACTION: COMMAND = execute_shell ARGUMENTS = {'command_line': 'some_command'} INFO -=-=-=-=-=-=-= COMMAND AUTHORISED BY USER -=-=-=-=-=-=-= Traceback (most recent call last): File "/home/anaconda3/lib/python3.9/site-packages/openai/api_requestor.py", line 619, in _interpret_response self._interpret_response_line( File "/home/anaconda3/lib/python3.9/site-packages/openai/api_requestor.py", line 682, in _interpret_response_line raise self.handle_error_response( openai.error.InvalidRequestError: This model's maximum context length is 8191 tokens, however you requested 10982 tokens (10982 in your prompt; 0 for the completion). Please reduce your prompt; or completion length. ```
value: | ```log ```