On August 25,Film noir Archives Alibaba Cloud launched an open-source Large Vision Language Model (LVLM) named Qwen-VL. The LVLM is based on Alibaba Cloud’s 7 billion parameter foundational language model Qwen-7B. In addition to capabilities such as image-text recognition, description, and question answering, Qwen-VL introduces new features including visual location recognition and image-text comprehension, the company said in a statement. These functions enable the model to identify locations in pictures and to provide users with guidance based on the information extracted from images, the firm added. The model can be applied in various scenarios including image and document-based question answering, image caption generation, and fine-grained visual recognition. Currently, both Qwen-VL and its visual AI assistant Qwen-VL-Chat are available for free and commercial use on Alibaba’s “Model as a Service” platform ModelScope. [Alibaba Cloud statement, in Chinese]
(Editor: {typename type="name"/})
Dyson V8 Plus cordless vacuum: $120 off at Amazon
Best vacuum deal: Save $100 on Dyson V15 at Amazon
Google Pixel 9a leak shows a drastically different look
Alabama vs. Georgia football livestreams: kickoff time, streaming deals, and more
The Kindle Scribe just dropped to its lowest price ever, but is it worth it?
Best air purifier deal: Save up to 46% off during Amazon's early Prime Day sale
Colorado vs. UCF football without cable: kickoff time, streaming deals, and more
Google Pixel Buds Pro 2: $40 off at Amazon
The 26 best horror movies now streaming on Peacock
Best Amazon Fire TV Cube deal: Save $30 at Amazon
Ohio State vs. MSU football livestreams without cable: kickoff time, streaming deals, and more
接受PR>=1、BR>=1,流量相当,内容相关类链接。