feat: publishing infernet-container-starter v0.2.0
This commit is contained in:
8
projects/tgi-llm/tgi/Makefile
Normal file
8
projects/tgi-llm/tgi/Makefile
Normal file
@ -0,0 +1,8 @@
|
||||
.phony: run
|
||||
|
||||
volume ?= $(PWD)/data
|
||||
model ?= mistralai/Mistral-7B-v0.1
|
||||
|
||||
run:
|
||||
docker run --gpus all --shm-size 1g -p 8080:80 -v $(volume):/data \
|
||||
ghcr.io/huggingface/text-generation-inference:1.4 --model-id $(model)
|
15
projects/tgi-llm/tgi/README.md
Normal file
15
projects/tgi-llm/tgi/README.md
Normal file
@ -0,0 +1,15 @@
|
||||
# TGI Service
|
||||
|
||||
The [Makefile](./Makefile) for this service simply invokes
|
||||
huggingface's `huggingface/text-generation-inference:1.4`
|
||||
docker image. Ensure that you are running this on a machine with a GPU.
|
||||
|
||||
For example, to run the TGI container with model `mistralai/Mistral-7B-v0.1`, you can
|
||||
use the following command:
|
||||
|
||||
```bash
|
||||
make run model=mistralai/Mistral-7B-v0.1 volume=/path/to/your/data
|
||||
```
|
||||
|
||||
* `model`: is defaulted to `mistralai/Mistral-7B-v0.1`
|
||||
* `volume`: is defaulted to `./data`
|
Reference in New Issue
Block a user