ritual/projects/tgi-llm/tgi/README.md

491 B

TGI Service

The Makefile for this service simply invokes huggingface's huggingface/text-generation-inference:1.4 docker image. Ensure that you are running this on a machine with a GPU.

For example, to run the TGI container with model mistralai/Mistral-7B-v0.1, you can use the following command:

make run model=mistralai/Mistral-7B-v0.1 volume=/path/to/your/data
  • model: is defaulted to mistralai/Mistral-7B-v0.1
  • volume: is defaulted to ./data