Your fully proficient, AI-powered and local chatbot assistant🤖
git clone https://github.com/AstraBert/everything-ai.git
cd everything-ai
.env
fileModify:
VOLUME
variable in the .env file so that you can mount your local file system into Docker container.MODELS_PATH
variable in the .env file so that you can tell llama.cpp where you stored the GGUF models you downloaded.MODEL
variable in the .env file so that you can tell llama.cpp what model to use (use the actual name of the gguf file, and do not forget the .gguf extension!)MAX_TOKENS
variable in the .env file so that you can tell llama.cpp how many new tokens it can generate as output.An example of a .env
file could be:
VOLUME="c:/Users/User/:/User/"
MODELS_PATH="c:/Users/User/.cache/llama.cpp/"
MODEL="stories260K.gguf"
MAX_TOKENS="512"
This means that now everything that is under “c:/Users/User/” on your local machine is under “/User/” in your Docker container, that llama.cpp knows where to look for models and what model to look for, along with the maximum new tokens for its output.
docker pull astrabert/everything-ai:latest
docker pull qdrant/qdrant:latest
docker pull ghcr.io/ggerganov/llama.cpp:server
docker compose up
localhost:8670
and choose your assistantYou will see something like this:
Choose the task among:
qdrant
backend to build a retrieval-friendly knowledge base, which you can query and tune the response of your model on. You have to pass either a pdf/a bunch of pdfs specified as comma-separated paths or a directory where all the pdfs of interest are stored (DO NOT provide both); you can also specify the language in which the PDF is written, using ISO nomenclature - MULTILINGUAL./
├── test/
| ├── label1/
| └── label2/
└── train/
├── label1/
└── label2/
You can query the database starting from your own pictures.
localhost:7860
and start using your assistantOnce everything is ready, you can head over to localhost:7860
and start using your assistant: