You can not select more than 25 topics Topics must start with a chinese character,a letter or number, can include dashes ('-') and can be up to 35 characters long.
 
 
 
 
Lin-Bert 8470b5f601 fix docs. 5 months ago
..
bert 删除optimizer_shard开关 6 months ago
blip2 !1469 【dev】BLIP2一阶段训练,评估适配910B,支持多卡并行评估,适配CANN7.0 6 months ago
bloom fix docs. 5 months ago
clip 删除optimizer_shard开关 6 months ago
codegeex2 修改codealpaca训练参数 6 months ago
general 删除optimizer_shard开关 6 months ago
glm [BugFix]Fix glm infer configuration error. 6 months ago
glm2 glm2添加边训边推功能 6 months ago
gpt2 gpt2-13B 910B 6 months ago
llama fix docs. 5 months ago
llama2 新增分布式评测,lite推理配置导致bug修复 5 months ago
mae 删除optimizer_shard开关 6 months ago
pangualpha 删除optimizer_shard开关 6 months ago
qa 删除optimizer_shard开关 6 months ago
sam 删除optimizer_shard开关 6 months ago
swin 删除optimizer_shard开关 6 months ago
t5 删除optimizer_shard开关 6 months ago
tokcls 删除optimizer_shard开关 6 months ago
txtcls 删除optimizer_shard开关 6 months ago
vit 删除optimizer_shard开关 6 months ago
README.md 增加对config中不同tokenizer的说明 6 months ago

MindSpore Transformers套件的目标是构建一个大模型训练、推理、部署的全流程套件: 提供业内主流的Transformer类预训练模型, 涵盖丰富的并行特性。 期望帮助用户轻松的实现大模型训练。 文档:https://mindformers.readthedocs.io/zh-cn/latest/

Jupyter Notebook Python Markdown Shell