feat: publishing infernet-container-starter v0.2.0

This commit is contained in:
ritual-all
2024-03-29 10:50:13 -04:00
parent 41aaa152e6
commit 4545223364
155 changed files with 6086 additions and 257 deletions

View File

@ -0,0 +1,8 @@
.phony: run
volume ?= $(PWD)/data
model ?= mistralai/Mistral-7B-v0.1
run:
docker run --gpus all --shm-size 1g -p 8080:80 -v $(volume):/data \
ghcr.io/huggingface/text-generation-inference:1.4 --model-id $(model)

View File

@ -0,0 +1,15 @@
# TGI Service
The [Makefile](./Makefile) for this service simply invokes
huggingface's `huggingface/text-generation-inference:1.4`
docker image. Ensure that you are running this on a machine with a GPU.
For example, to run the TGI container with model `mistralai/Mistral-7B-v0.1`, you can
use the following command:
```bash
make run model=mistralai/Mistral-7B-v0.1 volume=/path/to/your/data
```
* `model`: is defaulted to `mistralai/Mistral-7B-v0.1`
* `volume`: is defaulted to `./data`