# 模型支持列表
## NLP
### masked_language_modeling
### [text_generation](task_cards/text_generation.md)
| 模型
model | 模型规格
type | 数据集
dataset | 评估指标
metric | 评估得分
score | 配置
config |
|:-----------------------------------------------------:|:-------------------------------------------------------------------------------------:|:-------------------:|:-------------------------------------:|:---------------------------------------------------------------------:|:--------------------------------------------------------------------------------:|
| [llama2](model_cards/llama2.md) | llama2_7b
llama2_13b
llama2_7b_lora
llama2_13b_lora
llama2_70b | alpaca | PPL / EM / F1 | 6.58 / 39.6 / 60.5
6.14 / 27.91 / 44.23
-
-
- | [configs](https://gitee.com/mindspore/mindformers/tree/dev/configs/llama) |
| [glm2](model_cards/glm2.md) | glm2_6b
glm2_6b_lora | ADGEN | BLEU-4 / Rouge-1 / Rouge-2 / Rouge-l | 7.47 / 30.78 / 7.07 / 24.77
7.23 / 31.06 / 7.18 / 24.23 | [configs](https://gitee.com/mindspore/mindformers/tree/dev/configs/glm2) |
| [glm3](model_cards/glm3.md) | glm3_6b | ADGEN | - | - | [configs](https://gitee.com/mindspore/mindformers/tree/dev/configs/glm3) |
| [gpt2](model_cards/gpt2.md) | gpt2_small
gpt2_13b | wikitext-2 | - | - | [configs](https://gitee.com/mindspore/mindformers/tree/dev/configs/gpt2) |
| [baichuan2](../research/baichuan2/baichuan2.md) | baichuan2_7b
baichuan2_13b
baichuan2_7b_lora
baichuan2_13b_lora | belle | - | - | [configs](https://gitee.com/mindspore/mindformers/tree/dev/research/baichuan2) |
| [codegeex2](model_cards/codegeex2.md) | codegeex2_6b | CodeAlpaca | - | - | [configs](https://gitee.com/mindspore/mindformers/tree/dev/configs/codegeex2) |
| [codellama](model_cards/codellama.md) | codellama_34b | CodeAlpaca | - | - | [configs](https://gitee.com/mindspore/mindformers/tree/dev/configs/codellama) |
| [deepseek](../research/deepseek/deepseek.md) | deepseek_33b | - | - | - | [configs](https://gitee.com/mindspore/mindformers/tree/dev/research/deepseek) |
| [internlm](../research/internlm/internlm.md) | internlm_7b
internlm_20b | alpaca | - | - | [configs](https://gitee.com/mindspore/mindformers/tree/dev/research/internlm) |
| [mixtral](../research/mixtral/mixtral.md) | mixtral-8x7b | wikitext-2 | - | - | [configs](https://gitee.com/mindspore/mindformers/tree/dev/research/mixtral) |
| [qwen](../research/qwen/qwen.md) | qwen_7b
qwen_14b
qwen_7b_lora
qwen_14b_lora | alpaca | C-Eval | 63.3
72.13
-
- | [configs](https://gitee.com/mindspore/mindformers/tree/dev/research/qwen) |
| [qwen1.5](../research/qwen1_5/qwen1_5.md) | qwen1.5-14b
qwen1.5-72b | - | - | - | [configs](https://gitee.com/mindspore/mindformers/tree/dev/research/qwen1_5) |
| [wizardcoder](../research/wizardcoder/wizardcoder.md) | wizardcoder_15b | CodeAlpaca | MBPP Pass@1 | 50.8 | [configs](https://gitee.com/mindspore/mindformers/tree/dev/research/wizardcoder) |
| [yi](../research/yi/yi.md) | yi_6b
yi_34b | alpaca_gpt4_data_zh | - | - | [configs](https://gitee.com/mindspore/mindformers/tree/dev/research/yi) |
## LLM大模型能力支持一览
| 模型 \ 特性 | 低参微调 | 边训边评 | Flash Attention | 并行推理 | 流式推理 | Chat | 多轮对话 |
|:-----------------:|:---------------:|:--------------:|:---------------:|:-----:|:----:|:----:|:----:|
| Llama2-7B/13B/70B | Lora | PPL | √ | dp/mp | √ | √ | √ |
| GLM2-6B | Lora/P-TuningV2 | PPL/Bleu/Rouge | √ | dp/mp | √ | √ | √ |
| GLM3-6B | Lora | × | √ | dp/mp | √ | √ | √ |
| CodeGeex2-6B | × | PPL/Bleu/Rouge | √ | dp/mp | √ | √ | √ |
| CodeLlama-34B | Lora | pass@1 | √ | dp/mp | √ | √ | × |
| GPT2-128m/13B | Lora | PPL | √ | dp/mp | √ | × | × |
| BaiChuan2-7B/13B | Lora | PPL | √ | dp/mp | √ | √ | √ |
| Qwen-7B/14B | √ | × | √ | dp/mp | √ | √ | √ |
| Qwen1.5-14B | × | × | × | dp/mp | √ | √ | √ |
| Qwen1.5-72B | Lora | × | √ | dp/mp | √ | √ | √ |
| InternLM-7B/20B | Lora | PPL | √ | dp/mp | √ | √ | √ |
| Wizardcoder-15B | × | PPL | × | dp/mp | √ | √ | √ |
| Deepseek-33B | × | × | × | × | × | √ | × |
| Mixtral-8×7B | × | × | √ | √ | × | × | × |
| Yi-6B | Lora | × | √ | dp/mp | √ | √ | √ |
| Yi-34B | × | × | × | dp/mp | √ | × | × |