The company mainly trained Phi-4-reasoning-vision-15B on open-source data. The data included images and text-based descriptions of the objects depicted in those images. Before it started training the ...
B, an open-weight multimodal vision AI model designed to deliver strong math, science, document and UI reasoning with far less training data and compute than much larger systems.
Start working toward program admission and requirements right away. Work you complete in the non-credit experience will transfer to the for-credit experience when you ...
Modern AI Models for Vision and Multimodal Understanding is a course that will enable you to understand and build systems that interpret images, text, and more—just like today’s leading AI models.
Twelve Labs Inc., the developer of generative artificial intelligence foundation models that can understand videos like humans, said today that the company has raised $50 million in early-stage ...
Explore how vision-language-action models like Helix, GR00T N1, and RT-1 are enabling robots to understand instructions and act autonomously.
“GPT-4o is especially better at vision and audio understanding compared to existing models,” OpenAI said in its announcement. During an on-stage event, Murati said GPT-4o will also have new memory ...
Sarvam has gained attention at the AI Impact Summit 2026 by unveiling its advanced AI model, Sarvam Vision, which claims ...
Canadian AI startup Cohere launched in 2019 specifically targeting the enterprise, but independent research has shown it has so far struggled to gain much of a market share among third-party ...
Chicago, Feb. 11, 2026 (GLOBE NEWSWIRE) -- The global vision-Language Models (VLM) market size was valued at USD 3.84 billion in 2025 and is projected to hit the market valuation of USD 41.75 billion ...
Some results have been hidden because they may be inaccessible to you
Show inaccessible results