16 lines
491 B
Markdown
16 lines
491 B
Markdown
|
# TGI Service
|
||
|
|
||
|
The [Makefile](./Makefile) for this service simply invokes
|
||
|
huggingface's `huggingface/text-generation-inference:1.4`
|
||
|
docker image. Ensure that you are running this on a machine with a GPU.
|
||
|
|
||
|
For example, to run the TGI container with model `mistralai/Mistral-7B-v0.1`, you can
|
||
|
use the following command:
|
||
|
|
||
|
```bash
|
||
|
make run model=mistralai/Mistral-7B-v0.1 volume=/path/to/your/data
|
||
|
```
|
||
|
|
||
|
* `model`: is defaulted to `mistralai/Mistral-7B-v0.1`
|
||
|
* `volume`: is defaulted to `./data`
|