1.2 KiB
ChromaDB
If you want chromaDB to use the more accurate (but much slower) instructor embeddings add the following to config.yaml
:
chromadb:
embeddings: instructor
instructor_device: cpu
instructor_model: hkunlp/instructor-xl
You will need to restart the backend for this change to take effect.
Note that the first time you do this it will need to download the instructor model you selected. This may take a while, and the talemate backend will be un-responsive during that time.
Once the download is finished, if talemate is still un-responsive, try reloading the front-end to reconnect. When all fails just restart the backend as well.
GPU support
If you want to use the instructor embeddings with GPU support, you will need to install pytorch with CUDA support.
To do this on windows, run install-pytorch-cuda.bat
from the project root. Then change your device in the config to cuda
:
chromadb:
embeddings: instructor
instructor_device: cuda
instructor_model: hkunlp/instructor-xl
Instructor embedding models:
hkunlp/instructor-base
(smallest / fastest)hkunlp/instructor-large
hkunlp/instructor-xl
(largest / slowest) - requires about 5GB of GPU memory