访问HuggingFace网站获取所需模型:https://huggingface.co/BAAI/bge-large-zh-v1.5/tree/main
以BAAI/bge-large-zh-v1.5为例,下载“Files and versions”下所有文件。
访问链接https://www.hiascend.com/developer/ascendhub/detail/07a016975cc341f3a5ae131f2b52399d,根据NPU类型下载相应镜像。
docker run -u root -e ASCEND_VISIBLE_DEVICES=0 -itd --name=tei --net=host \ -e HOME=/home/HwHiAiUser \ -e TEI_NPU_DEVICE=0 \[可选] --privileged=true \ -v /root/bge-large-zh-v1.5:/home/HwHiAiUser/model/bge-large-zh-v1.5 \ -v /usr/local/bin/npu-smi:/usr/local/bin/npu-smi \ -v /usr/local/Ascend/driver:/usr/local/Ascend/driver \ --entrypoint /home/HwHiAiUser/start.sh \[可选,一般不加] swr.cn-south-1.myhuaweicloud.com/ascendhub/mis-tei:6.0.0-300I-Duo-aarch64 BAAI/bge-large-zh-v1.5 ip port
以下参数可根据实际情况进行配置:
curl ip:port/embed \ -X POST \ -d '{"inputs":"What is Deep Learning?"}' \ -H 'Content-Type: application/json'