All Updates

All Updates

icon
Filter
Product updates
Alibaba releases AI model ‘Qwen2-VL’ for video and image analysis
Foundation Models
Aug 29, 2024
This week:
Partnerships
T-Mobile partners with OpenAI to develop AI-powered customer service platform
Generative AI Applications
Yesterday
Partnerships
Runway partners with Lionsgate to develop AI video tools using studio's movie catalog
Generative AI Applications
Sep 18, 2024
Funding
QMill raises EUR 4 million in seed funding to provide quantum computing industrial applications
Quantum Computing
Sep 18, 2024
Product updates
QuiX Quantum launches 'Bia' quantum cloud computing service for quantum solutions
Quantum Computing
Sep 18, 2024
Partnerships
Oxford Ionics and Infineon Technologies partner to build portable quantum computer for Cyberagentur
Quantum Computing
Sep 18, 2024
Product updates
Partnerships
Tencent Ai Lab launches EzAudio AI for text-to-audio generation with Johns Hopkins University
Foundation Models
Sep 18, 2024
Funding
TON secures USD 30 million in investment from Bitget and Foresight Ventures
Web3 Ecosystem
Sep 18, 2024
Funding
Hemi Labs raises USD 15 million in funding to launch blockchain network
Web3 Ecosystem
Sep 18, 2024
Product updates
Fivetran launches Hybrid Deployment for data pipeline management
Machine Learning Infrastructure
Sep 18, 2024
Product updates
Fivetran launches Hybrid Deployment for data pipeline management
Data Infrastructure & Analytics
Sep 18, 2024
Foundation Models

Foundation Models

Aug 29, 2024

Alibaba releases AI model ‘Qwen2-VL’ for video and image analysis

Product updates

  • Alibaba Cloud has launched “Qwen2-VL,” an advanced vision-language AI model to enhance visual understanding, video comprehension, and multilingual text-image processing.

  • The model is available in 72 billion, 7 billion, and 2 billion parameters. Additionally, it can analyze and describe handwriting in multiple languages, identify objects in images, analyze live video in near real-time, and analyze videos longer than 20 minutes while answering questions about content. Furthermore, it supports English, Chinese, most European languages, and Japanese, Korean, Arabic, and Vietnamese. 

  • The company claims the model can be integrated into mobile phones and robots, allowing automated operations based on visual environments and text instructions. It also supports function calling, enabling integration with third-party software and visual extraction of information from external sources.

Contact us

Gain access to all industry hubs, market maps, research tools, and more
Get a demo
arrow
menuarrow

By using this site, you agree to allow SPEEDA Edge and our partners to use cookies for analytics and personalization. Visit our privacy policy for more information about our data collection practices.