api-for-open-llm docker启动补充

因为大模型数据是在启动后下载在容器内部,所以先

mkdir ~/huggingface

然后再启动,这样就会下载到~/huggingface 文件夹中了

docker run -it -d --gpus all --ipc=host -p 7891:8000 --name=llm-api \
    --ulimit memlock=-1 --ulimit stack=67108864 \
    -v `pwd`:/workspace \
    -v ~/huggingface:/root/.cache/huggingface \
    llm-api:pytorch \
    python api/server.py
Cache management
We’re on a journey to advance and democratize artificial intelligence through open source and open science.