diff --git a/docs/source/assets/deployment/anything-llm-chat-with-doc.png b/docs/source/assets/deployment/anything-llm-chat-with-doc.png new file mode 100644 index 000000000000..f9b57f5c3cec Binary files /dev/null and b/docs/source/assets/deployment/anything-llm-chat-with-doc.png differ diff --git a/docs/source/assets/deployment/anything-llm-chat-without-doc.png b/docs/source/assets/deployment/anything-llm-chat-without-doc.png new file mode 100644 index 000000000000..952a43bcd677 Binary files /dev/null and b/docs/source/assets/deployment/anything-llm-chat-without-doc.png differ diff --git a/docs/source/assets/deployment/anything-llm-provider.png b/docs/source/assets/deployment/anything-llm-provider.png new file mode 100644 index 000000000000..bb699f7571f4 Binary files /dev/null and b/docs/source/assets/deployment/anything-llm-provider.png differ diff --git a/docs/source/assets/deployment/anything-llm-upload-doc.png b/docs/source/assets/deployment/anything-llm-upload-doc.png new file mode 100644 index 000000000000..00c70e9c01f6 Binary files /dev/null and b/docs/source/assets/deployment/anything-llm-upload-doc.png differ diff --git a/docs/source/deployment/frameworks/anything-llm.md b/docs/source/deployment/frameworks/anything-llm.md new file mode 100644 index 000000000000..d430c170ef54 --- /dev/null +++ b/docs/source/deployment/frameworks/anything-llm.md @@ -0,0 +1,47 @@ +(deployment-anything-llm)= + +# Anything LLM + +[Anything LLM](https://github.com/Mintplex-Labs/anything-llm) is a full-stack application that enables you to turn any document, resource, or piece of content into context that any LLM can use as references during chatting. + +It allows you to deploy a large language model (LLM) server with vLLM as the backend, which exposes OpenAI-compatible endpoints. + +## Prerequisites + +- Setup vLLM environment + +## Deploy + +- Start the vLLM server with the supported chat completion model, e.g. + +```console +vllm serve Qwen/Qwen1.5-32B-Chat-AWQ --max-model-len 4096 +``` + +- Download and install [Anything LLM desktop](https://anythingllm.com/desktop). + +- On the bottom left of open settings, AI Prooviders --> LLM: + - LLM Provider: Generic OpenAI + - Base URL: http://{vllm server host}:{vllm server port}/v1 + - Chat Model Name: `Qwen/Qwen1.5-32B-Chat-AWQ` + +:::{image} /assets/deployment/anything-llm-provider.png +::: + +- Back to home page, New Workspace --> create `vllm` workspace, and start to chat: + +:::{image} /assets/deployment/anything-llm-chat-without-doc.png +::: + +- Click the upload button: + - upload the doc + - select the doc and move to the workspace + - save and embed + +:::{image} /assets/deployment/anything-llm-upload-doc.png +::: + +- Chat again: + +:::{image} /assets/deployment/anything-llm-chat-with-doc.png +::: diff --git a/docs/source/deployment/frameworks/index.md b/docs/source/deployment/frameworks/index.md index 8be38d56cd28..a1b405386b77 100644 --- a/docs/source/deployment/frameworks/index.md +++ b/docs/source/deployment/frameworks/index.md @@ -3,6 +3,7 @@ :::{toctree} :maxdepth: 1 +anything-llm bentoml cerebrium dstack