News

Jakub Pachocki, who leads the firm’s development of advanced models, is excited to release an open version to researchers.
The new ImageBind model combines text, audio, visual, movement, thermal, and depth data. It’s only a research project but shows how future AI models could be able to generate multisensory content.
Engineers at RMIT University have invented a small "neuromorphic" device that detects hand movement, stores memories and ...
On Wednesday, Microsoft Research introduced Magma, an integrated AI foundation model that combines visual and language processing to ... only processes multimodal data (like text, images, and ...
On Monday, a group of AI researchers from Google and the Technical University of Berlin unveiled PaLM-E, a multimodal embodied visual-language model ... to pre-process or annotate the data ...
“On the Dangers of Stochastic Parrots: Can Language Models Be Too Big?” lays out the risks of large language models—AIs trained on staggering amounts of text data. These have grown ...
Pierre Ferragu of New Street Research ... with more data and more computer power creating more intelligence. But once a model was trained, it was hard to use extra processing power well.