On August 25,Watch Boarding House Hungry Wolves 2 Online Alibaba Cloud launched an open-source Large Vision Language Model (LVLM) named Qwen-VL. The LVLM is based on Alibaba Cloud’s 7 billion parameter foundational language model Qwen-7B. In addition to capabilities such as image-text recognition, description, and question answering, Qwen-VL introduces new features including visual location recognition and image-text comprehension, the company said in a statement. These functions enable the model to identify locations in pictures and to provide users with guidance based on the information extracted from images, the firm added. The model can be applied in various scenarios including image and document-based question answering, image caption generation, and fine-grained visual recognition. Currently, both Qwen-VL and its visual AI assistant Qwen-VL-Chat are available for free and commercial use on Alibaba’s “Model as a Service” platform ModelScope. [Alibaba Cloud statement, in Chinese]
Related Articles
Google is adding icons to let you know where image search results come from
2025-06-25 23:54
1645 views
Read More
Kendall Jenner is proving just how meaningless everything she does is
2025-06-25 23:44
2666 views
Read More
Stocks app Robinhood goes down again, and people are not happy
2025-06-25 23:42
2488 views
Read More
First impressions of Devolver Digital's 2020 games: A mixed bag
2025-06-25 23:31
2573 views
Read More
U.S. wireless carriers face $200 million in FCC fines for mishandling customer data
2025-06-25 23:09
1295 views
Read More