B, an open-weight multimodal vision AI model designed to deliver strong math, science, document and UI reasoning with far ...
Explore how vision-language-action models like Helix, GR00T N1, and RT-1 are enabling robots to understand instructions and ...
Microsoft's Phi-4-reasoning-vision-15B uses careful data curation and selective reasoning to compete with models trained on ...
As I highlighted in my last article, two decades after the DARPA Grand Challenge, the autonomous vehicle (AV) industry is still waiting for breakthroughs—particularly in addressing the “long tail ...
February brought new coding models, and vision-language models impress with OCR. Open Responses aims to establish itself as a ...
Figure AI has unveiled HELIX, a pioneering Vision-Language-Action (VLA) model that integrates vision, language comprehension, and action execution into a single neural network. This innovation allows ...
Canadian AI startup Cohere launched in 2019 specifically targeting the enterprise, but independent research has shown it has so far struggled to gain much of a market share among third-party ...
OpenAI Group PBC today launched a new large language model that it says is more adept at automating work tasks than its ...
Unlock the full InfoQ experience by logging in! Stay updated with your favorite authors and topics, engage with content, and download exclusive resources. Cory Benfield discusses the evolution of ...
Phi-3-vision, a 4.2 billion parameter model, can answer questions about images or charts. Phi-3-vision, a 4.2 billion parameter model, can answer questions about images or charts. is a reporter who ...