Alibaba Cloud has unveiled Qwen2.5, the latest AI model in its Tongyi Qianwen series. This model offers a range of parameter sizes from 50 million to 72 billion to accommodate diverse application needs. During its training, Qwen2.5 utilized a large dataset containing approximately 18 trillion tokens, resulting in a performance improvement of at least 18% over the previous version.
Qwen2.5 is capable of handling longer text inputs and outputs, supporting context lengths of up to 128,000 tokens and generating content of up to 8,000 tokens. Additionally, the model supports over 29 languages, enhancing its multilingual capabilities.
For specific domain requirements, Qwen2.5 provides optimized versions: Qwen2.5-Coder for programming tasks and Qwen2.5-Math for mathematical problem-solving. The former is trained on a dataset of 5.5 trillion tokens covering 92 programming languages, while the latter incorporates various reasoning methods, demonstrating strong problem-solving abilities in both Chinese and English math problems.
According to public information, Qwen2.5 has shown impressive performance in multiple benchmark tests. For example, it achieved scores of 86.8 in the MMLU-rudex general knowledge assessment, 88.2 in the MBPP code writing capability test, and 83.1 in the MATH mathematical proficiency evaluation. These improvements not only enhance Qwen2.5's natural language processing capabilities but also demonstrate its increased effectiveness in specialized applications.