🚀New from Meta FAIR: today we’re introducing Seamless Interaction, a research project dedicated to modeling interpersonal dynamics. The project features a family of audiovisual behavioral models, developed in collaboration with Meta’s Codec Avatars lab + Core AI lab, that render speech between two individuals into diverse, expressive full-body gestures and active listening behaviors, allowing the creation of fully embodied avatars in 2D and 3D. These models have potential to create more natural, interactive virtual agents that can engage in human-like social interactions across a variety of settings. Learn more:
We’re also publicly releasing the Seamless Interaction Dataset, with 4,000+ participants and 4,000+ hours of interactions, making it the largest known video dataset of its kind. This dataset enables our audiovisual behavioral models to understand and generate human-like social behaviors, and represents a crucial stepping stone to understanding and modeling how people communicate and behave when they’re together. We’re sharing it here to help the research community advance their work:
One more thing… you can also check out this technical report detailing our methodology to build motion models on the dataset, along with an evaluation framework for this type of model:
Head to our blog to go deeper on the full story:
32,85K