ModelLink2/examples/mcore/mixtral
shenjiarun f79c5dcabe !1817 整改两处公网地址
Merge pull request !1817 from shenjiarun/master
2024-10-29 11:05:59 +00:00
..
ckpt_convert_mixtral_hf2mcore.sh !1597 新增权重转换ut模板和mixtral用例,支持legacy和mcore互转 2024-09-10 07:40:15 +00:00
data_convert_mixtral_instruction.sh !1817 整改两处公网地址 2024-10-29 11:05:59 +00:00
data_convert_mixtral_pretrain.sh !1555 新增 Mixtral 8x22B 预训练、推理和评估 2024-09-02 03:43:11 +00:00
evaluate_mixtral_8x7b_ptd.sh !1741 更新llama2-7b,13b,70b性能 2024-10-09 02:35:47 +00:00
evaluate_mixtral_8x22b_ptd.sh !1555 新增 Mixtral 8x22B 预训练、推理和评估 2024-09-02 03:43:11 +00:00
generate_mixtral_8x7b_ptd.sh !1741 更新llama2-7b,13b,70b性能 2024-10-09 02:35:47 +00:00
generate_mixtral_8x22b_ptd.sh !1555 新增 Mixtral 8x22B 预训练、推理和评估 2024-09-02 03:43:11 +00:00
pretrain_mixtral_8x7b_ptd.sh !1604 增加mixtral_8*7b推理评估 2024-09-04 01:05:21 +00:00
pretrain_mixtral_8x22b_32k_ptd_C.sh !1613 修复Mixtral预训练加载权重 2024-09-05 02:09:37 +00:00
pretrain_mixtral_8x22b_32k_ptd.sh !1613 修复Mixtral预训练加载权重 2024-09-05 02:09:37 +00:00
pretrain_mixtral_8x22b_64k_ptd_C.sh !1613 修复Mixtral预训练加载权重 2024-09-05 02:09:37 +00:00
tune_mixtral_8x7B_full_ptd.sh !1696 新增Mixtral-8x7B-mcore MOE全参微调Loss对齐脚本 2024-09-21 06:22:22 +00:00
tune_mixtral_8x7b_lora_ptd.sh !1774 lora适配moe模型 2024-10-25 06:28:18 +00:00