Chat with your documents offline using AI. No data leaves your system. Internet connection is only required to install the tool and download the AI models. It is based on PrivateGPT but has more features.
Contents
- Supports GGML/GGUF models via CTransformers
- Supports 🤗 Transformers models
- Supports GPTQ models
- Web UI
- GPU support
- Highly configurable via
chatdocs.yml
Show supported document types
| Extension | Format |
|---|---|
.csv |
CSV |
.docx, .doc |
Word Document |
.enex |
EverNote |
.eml |
|
.epub |
EPub |
.html |
HTML |
.md |
Markdown |
.msg |
Outlook Message |
.odt |
Open Document Text |
.pdf |
Portable Document Format (PDF) |
.pptx, .ppt |
PowerPoint Document |
.txt |
Text file (UTF-8) |
Install the tool using:
pip install chatdocsDownload the AI models using:
chatdocs downloadNow it can be run offline without internet connection.
Add a directory containing documents to chat with using:
chatdocs add /path/to/documentsThe processed documents will be stored in
dbdirectory by default.
Chat with your documents using:
chatdocs uiOpen http://localhost:5000 in your browser to access the web UI.
It also has a nice command-line interface:
chatdocs chatAll the configuration options can be changed using the chatdocs.yml config file. Create a chatdocs.yml file in some directory and run all commands from that directory. For reference, see the default chatdocs.yml file.
You don't have to copy the entire file, just add the config options you want to change as it will be merged with the default config. For example, see tests/fixtures/chatdocs.yml which changes only some of the config options.
To change the embeddings model, add and change the following in your chatdocs.yml:
embeddings:
model: hkunlp/instructor-largeNote: When you change the embeddings model, delete the
dbdirectory and add documents again.
To change the CTransformers (GGML/GGUF) model, add and change the following in your chatdocs.yml:
ctransformers:
model: TheBloke/Wizard-Vicuna-7B-Uncensored-GGML
model_file: Wizard-Vicuna-7B-Uncensored.ggmlv3.q4_0.bin
model_type: llamaNote: When you add a new model for the first time, run
chatdocs downloadto download the model before using it.
You can also use an existing local model file:
ctransformers:
model: /path/to/ggml-model.bin
model_type: llamaTo use 🤗 Transformers models, add the following to your chatdocs.yml:
llm: huggingfaceTo change the 🤗 Transformers model, add and change the following in your chatdocs.yml:
huggingface:
model: TheBloke/Wizard-Vicuna-7B-Uncensored-HFNote: When you add a new model for the first time, run
chatdocs downloadto download the model before using it.
To use GPTQ models with 🤗 Transformers, install the necessary packages using:
pip install chatdocs[gptq]To enable GPU (CUDA) support for the embeddings model, add the following to your chatdocs.yml:
embeddings:
model_kwargs:
device: cudaYou may have to reinstall PyTorch with CUDA enabled by following the instructions here.
To enable GPU (CUDA) support for the CTransformers (GGML/GGUF) model, add the following to your chatdocs.yml:
ctransformers:
config:
gpu_layers: 50You may have to install the CUDA libraries using:
pip install ctransformers[cuda]To enable GPU (CUDA) support for the 🤗 Transformers model, add the following to your chatdocs.yml:
huggingface:
device: 0You may have to reinstall PyTorch with CUDA enabled by following the instructions here.

