diff --git a/.github/ISSUE_TEMPLATE/bug_report.md b/.github/ISSUE_TEMPLATE/bug_report.md index b36ec484e6..2ebd50d9f4 100644 --- a/.github/ISSUE_TEMPLATE/bug_report.md +++ b/.github/ISSUE_TEMPLATE/bug_report.md @@ -10,15 +10,15 @@ assignees: '' **Describe the bug** A clear and concise description of what the bug is. -**To Reproduce** -Steps to reproduce the behavior: -1. Go to '...' -2. Click on '....' -3. Scroll down to '....' -4. See error +**Please describe your setup** -**Expected behavior** -A clear and concise description of what you expected to happen. +-[ ] MemGPT version + - What is the output of `memgpt version`? (eg "0.2.4") +-[ ] How did you install memgpt? + - `pip install pymemgpt`? `pip install pymemgpt-nightly`? `git clone`? +-[ ] Describe your setup + - What's your OS (Windows/MacOS/Linux)? + - How are you running `memgpt`? (`cmd.exe`/Powershell/Anaconda Shell/Terminal) **Screenshots** If applicable, add screenshots to help explain your problem. @@ -28,32 +28,12 @@ Add any other context about the problem here. --- -**How did you install MemGPT?** - -- From the official package? `pip install pymemgpt` -- From source? - - With [poetry](https://github.com/cpacker/MemGPT#development)? - - With `git clone git@github.com:cpacker/MemGPT.git` and `pip install -r requirements.txt`? - -**Your setup (please complete the following information)** - -- Your OS (Linux, MacOS, Windows) -- Where you're trying to run MemGPT from - - Terminal? PowerShell? cmd.exe? -- Your python version (run `python --version`) -- If you installed with `pip install pymemgpt`: - - `pymemgpt` version (run `pip list | grep pymemgpt`) -- If you installed from source: - - Your current git commit (run `git describe --tags`) +If you're not using OpenAI, please provide additional information on your local LLM setup: **Local LLM details** If you are trying to run MemGPT with local LLMs, please provide the following information: -- The exact model you're trying to use (link to the HuggingFace page you downloaded it from) - - We need the EXACT model name, including quantization scheme, e.g.: `dolphin-2.1-mistral-7b.Q6_K.gguf` (not just `dolphin-2.1-mistral`) -- The local LLM backend you are using (web UI? LM Studio?) -- Your hardware for the local LLM backend (local computer? operating system? remote RunPod?) -- Your hardware for the MemGPT command (same computer as the local LLM backend?) -- The full output (or as much as possible) of where the LLM is failing - - If you can include screenshots, even better! +-[ ] The exact model you're trying to use (e.g. `dolphin-2.1-mistral-7b.Q6_K.gguf`) +-[ ] The local LLM backend you are using (web UI? LM Studio?) +-[ ] Your hardware for the local LLM backend (local computer? operating system? remote RunPod?)