Chinese researchers unveil LLaVA-o1 to challenge OpenAI’s o1 model

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More

OpenAI‘s o1 model has shown that inference-time scaling—using more compute during inference—can significantly boost a language model’s reasoning abilities. LLaVA-o1, a new model developed by researchers from multiple universities in China, brings this paradigm to open-source vision language models (VLMs).

Early open-source VLMs typically use a direct predict...

To read the content, please register or login