Key Takeaways
- Qwen2.5-72B-Instruct achieved 85.4% on MMLU benchmark
- Qwen2-72B-Instruct scored 84.2% on MMLU 5-shot
- Qwen1.5-72B-Chat reached 78.1% on MMLU
- Qwen2.5-72B has 7.37 billion parameters
- Qwen2-72B model supports 128K context length
- Qwen1.5-32B uses Grouped-Query Attention (GQA)
- Qwen trained on over 7 trillion tokens for Qwen2.5 series
- Qwen2 pre-trained on 7T tokens including code data
- Qwen1.5 used 2.5T multilingual tokens
- Qwen first released on September 1, 2023
- Qwen1.5 series launched February 1, 2024
- Qwen2 released June 6, 2024
- Qwen repo 1B downloads on Hugging Face as of Nov 2024
- Qwen2.5-72B-Instruct 50M downloads HF
- Qwen GitHub repo 35K stars
Alibaba Qwen models show strong benchmark performance across various metrics.
Adoption Metrics
Adoption Metrics Interpretation
Performance Benchmarks
Performance Benchmarks Interpretation
Release Timeline
Release Timeline Interpretation
Technical Specifications
Technical Specifications Interpretation
Training Resources
Training Resources Interpretation
Sources & References
- Reference 1QWENLMqwenlm.github.ioVisit source
- Reference 2HUGGINGFACEhuggingface.coVisit source
- Reference 3ARENAarena.lmsys.orgVisit source
- Reference 4LEADERBOARDleaderboard.lmsys.orgVisit source
- Reference 5PAPERSWITHCODEPaperswithcode.comVisit source
- Reference 6OPENLEADERBOARDopenleaderboard.vercel.appVisit source
- Reference 7BLOGblog.qwen.aiVisit source
- Reference 8GITHUBgithub.comVisit source
- Reference 9VLLMvllm.aiVisit source
- Reference 10DISCORDdiscord.ggVisit source
- Reference 11ALIBABACLOUDalibabacloud.comVisit source
- Reference 12DASHSCOPEdashscope.aliyun.comVisit source
- Reference 13ARXIVarxiv.orgVisit source






