diff --git a/ACL_PyTorch/contrib/cv/image_retrieval/BLIP/readme.md b/ACL_PyTorch/contrib/cv/image_retrieval/BLIP/readme.md index b002bfde7f2cf44016baec1a814dd5f952909221..3bf14836af0b59343c17386e02a30e2868f5361a 100644 --- a/ACL_PyTorch/contrib/cv/image_retrieval/BLIP/readme.md +++ b/ACL_PyTorch/contrib/cv/image_retrieval/BLIP/readme.md @@ -202,7 +202,7 @@ BLIP模型为一种新的Vision-Language Pre-training框架,它可以灵活地 1. 配置环境变量。 ``` - source /usr/local/Ascend/...... + source /usr/local/Ascend/ascend-toolkit/set_env.sh ``` 2. 执行命令查看芯片名称($\{chip\_name\})。 diff --git a/ACL_PyTorch/contrib/cv/image_retrieval/BLIP/requirement.txt b/ACL_PyTorch/contrib/cv/image_retrieval/BLIP/requirement.txt index f9acda5fe1c3ffb6ceb247fe17b50d60644d1b79..53ed61d2ffda71314e002e7ccf71375cd52ec91c 100644 --- a/ACL_PyTorch/contrib/cv/image_retrieval/BLIP/requirement.txt +++ b/ACL_PyTorch/contrib/cv/image_retrieval/BLIP/requirement.txt @@ -4,4 +4,4 @@ torchvision==0.8.1 transformers==4.18.0 fairscale==0.4.0 timm==0.6.11 -onnx==0.12.0 \ No newline at end of file +onnx==1.12.0 \ No newline at end of file