New Business

Alibaba launches new AI model targeting rival DeepSeek, China’s hottest start-up

In a statement posted on WeChat, the e-commerce giant’s cloud computing and AI arm Alibaba Cloud said its new Qwen 2.5-Max model also outperformed OpenAI’s GPT-4o and Meta Platforms’ Llama-3.1-405B in LLM performance benchmark platforms Arena-Hard and LiveBench. Alibaba owns the South China Morning Post.
The benchmark performance of Qwen 2.5-Max, part of Alibaba’s Tongyi Qianwen LLM family, was on par with Anthropic’s Claude-3.5-Sonnet model, according to Alibaba Cloud. LLMs are the technology underpinning generative AI services like ChatGPT.
Alibaba’s multimodal model is offered in various sizes, from 3 billion to 72 billion parameters, and includes both base and instruction-tuned versions. The flagship model, Qwen2.5-VL-72B-Instruct, is now accessible through the Qwen Chat platform, while the entire Qwen2.5-VL series is available on open-source platform Hugging Face and Alibaba’s own open-source community Model Scope.
Alibaba Cloud’s new Qwen 2.5-Max artificial intelligence model is touted to have outperformed rival large language models from DeepSeek and OpenAI. Photo: AFP
Alibaba Cloud’s new Qwen 2.5-Max artificial intelligence model is touted to have outperformed rival large language models from DeepSeek and OpenAI. Photo: AFP

Parameter is a machine-learning term for variables present in an AI system during training, which helps establish how data prompts yield the desired output. Open source gives public access to a software program’s source code, allowing third-party developers to modify or share its design, fix broken links or scale up its capabilities.


Source link

Related Articles

Leave a Reply

Your email address will not be published. Required fields are marked *

Back to top button