News

Chance AI, the multi-agent visual AI for explorers, artists, and creatives, today announces its most substantial model ...
A team of researchers has successfully developed a more efficient method to control the outputs of large language models ...
On Tuesday, NVIDIA announced its new Llama Nemotron Nano VL, a new multimodal vision-language model (VLM) that now leads the ...
Deepseek VL-2 is a sophisticated vision-language model designed to address complex ... or even understand humor in visual contexts—all while optimizing performance. In this overview AICodeKing ...
Hugging Face Inc. today open-sourced SmolVLM-256M, a new vision language model with the lowest parameter ... a range of tasks that involve processing visual data. It can answer questions about ...
The most capable open source AI model with visual abilities yet could see more ... the Allen Institute for AI (Ai2), the Multimodal Open Language Model, or Molmo, can interpret images as well ...
H Company launches its next-generation European agent suite, comprising Runner H, Tester H, and Surfer H, and the open-sourcing of Holo-1 H, its ...
According to Hugging Face, SmolVLM-256M has 256 million parameters, making it the world's smallest visual language model (VLM). SmolVLM-500M also has 500 million parameters, making it an ultra ...
Anthropic PBC today launched Claude 3.5 Sonnet, the company’s first release in a forthcoming artificial intelligence large language model family that outperforms both competing models and its ...
UroBot is based on OpenAI's most powerful language model, GPT-4o. It uses a customized method of retrieval-augmented generation (RAG) that is able to retrieve relevant information from hundreds of ...
With this release, Chance AI becomes the first true visual reasoning model, offering an unprecedented ... the joy of exploration in their own native language, by text or voice.” ...