Key Takeaways
- Qwen-72B achieves 73.5% on MMLU benchmark
- Qwen1.5-72B-Instruct scores 80.5% on MMLU
- Qwen2-72B-Instruct reaches 84.2% on MMLU 5-shot
- Qwen-72B has 72 billion parameters
- Qwen1.5-110B contains 110 billion parameters
- Qwen2-72B features 72 billion parameters
- Qwen trained on over 2 trillion tokens
- Qwen1.5 pre-trained on 7 trillion tokens including multilingual data
- Qwen2-72B trained on 7+ trillion high-quality tokens
- Qwen excels in 29 languages with C-Eval score of 85.2% for Qwen-72B
- Qwen1.5-72B achieves 81.7% on MultiICL benchmark
- Qwen2-72B scores 74.5% on MGSM multilingual math
- Qwen model repository has over 50 million downloads on Hugging Face
- Qwen2 series garnered 10 million downloads in first month
- Qwen1.5-7B has 15 million total downloads
Qwen models have high MMLU scores, high params, and big downloads.
Benchmark Performance
Benchmark Performance Interpretation
Community and Adoption
Community and Adoption Interpretation
Model Architecture
Model Architecture Interpretation
Multilingual Support
Multilingual Support Interpretation
Training Details
Training Details Interpretation
Sources & References
- Reference 1QWENLMqwenlm.github.ioVisit source
- Reference 2ARXIVarxiv.orgVisit source
- Reference 3HUGGINGFACEhuggingface.coVisit source
- Reference 4PLATFORMplatform.lmsys.orgVisit source
- Reference 5GITHUBgithub.comVisit source
- Reference 6ALIBABACLOUDalibabacloud.comVisit source
- Reference 7MODELSCOPEmodelscope.cnVisit source






