Tag: Vision-Language Models
-

New Method Lets AI Locate Personalised Objects Across Scenes
New Training Approach Improves Personalised Object Localization in AI A collaboration between MIT and the MIT-IBM Watson AI Lab has yielded a practical breakthrough in generative vision-language models (VLMs). The researchers developed a training regime that enables these models to locate personalised objects across different scenes—a capability that has traditionally been challenging. The work aims…
-

New Training Method Enables AI to Localise Personalised Objects Across Scenes
Overview: A step forward in personalised object localisation Researchers at MIT and the MIT-IBM Watson AI Lab have devised a training regime that lets generative vision-language models (VLMs) pinpoint personalised objects in novel scenes. This advances the field beyond generic category recognition to reliably locating a specific item—like a beloved pet or a unique chair—across…
-

New method helps AI models locate personalised objects in scenes
Overview Researchers from MIT and the MIT-IBM Watson AI Lab have unveiled a training technique that enables generative vision-language models (VLMs) to pinpoint personalised objects across different scenes. Previously, these models excelled at recognizing general object categories like dogs or chairs but struggled to identify a specific instance—such as a particular cat—in varied environments. The…
-

MIT researchers teach AI to locate personalized objects in scenes
Overview: Beyond general recognition Vision-language models (VLMs) blend visual understanding with language processing, enabling them to recognize broad categories like “dog” or “car.” But users increasingly want these systems to locate a specific, personalized object—think your French bulldog Bowser or a child’s backpack—across different moments in time. A team from MIT and the MIT-IBM Watson…
-

How MIT’s New Method Teaches AI to Find Personalized Objects in Video Context
Overview: Teaching AI to Localize Personalized Objects Vision-language models (VLMs) like GPT-5 have made strides in recognizing general objects in complex scenes. However, researchers at MIT and the MIT-IBM Watson AI Lab identified a critical gap: these models struggle to locate personalized objects across time and varied contexts, such as a specific pet or a…
-

Method Teaches Generative AI Models To Locate Personalized Objects
Overview: Teaching AI to Find What Makes Objects Personal Researchers from MIT and the MIT-IBM Watson AI Lab are tackling a persistent challenge in vision-language models (VLMs): locating a specific, personalized object within a scene. While such models can recognize general objects like dogs, they often struggle to identify Bowser the French Bulldog among many…
