Are you sure you want to delete this task? Once this task is deleted, it cannot be recovered.
huolongshe 4d6cf7184d | 2 months ago | |
---|---|---|
app | 2 months ago | |
demo_data | 4 months ago | |
docs | 4 months ago | |
.gitignore | 4 months ago | |
Dockerfile | 4 months ago | |
LICENSE | 4 months ago | |
README.md | 2 months ago | |
application.yml | 4 months ago | |
build-docker.sh | 4 months ago | |
pack_model.py | 4 months ago | |
pip-install-reqs.sh | 4 months ago | |
requirements.txt | 2 months ago | |
run_model_server.py | 4 months ago |
Vision Transformer (ViT) model pre-trained on ImageNet-21k (14 million images, 21,843 classes) at resolution 224x224, and fine-tuned on ImageNet 2012 (1 million images, 1,000 classes) at resolution 224x224. It was introduced in the paper An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale by Dosovitskiy et al. and first released in this repository. However, the weights were converted from the timm repository by Ross Wightman, who already converted the weights from JAX to PyTorch. Credits go to him.
模型来源: https://hf-mirror.com/google/vit-base-patch16-224
本模型基于 ServiceBoot微服务引擎 进行服务化封装,参见: 《CubeAI模型开发指南》
$ sh pip-install-reqs.sh
$ serviceboot start
或
$ python3 run_model_server.py
一键式本地容器化部署和运行,参见: 《CubeAI模型独立部署指南》 或 CubeAI Docker Builder
本模型服务可一键发布至 CubeAI智立方平台 进行共享和部署,参见: 《CubeAI模型发布指南》
No Description
Text Python Shell Dockerfile other
Dear OpenI User
Thank you for your continuous support to the Openl Qizhi Community AI Collaboration Platform. In order to protect your usage rights and ensure network security, we updated the Openl Qizhi Community AI Collaboration Platform Usage Agreement in January 2024. The updated agreement specifies that users are prohibited from using intranet penetration tools. After you click "Agree and continue", you can continue to use our services. Thank you for your cooperation and understanding.
For more agreement content, please refer to the《Openl Qizhi Community AI Collaboration Platform Usage Agreement》