Microsoft teases lifelike avatar AI tech but gives no release date

Microsoft teases lifelike avatar AI tech but gives no release date

Microsoft researchers say an AI model they have developed lets avatars engage in realistic seeming conversations complete with nuanced facial expressions
Microsoft researchers say an AI model they have developed lets avatars engage in realistic seeming conversations complete with nuanced facial expressions. Photo: Pau BARRENA / AFP/File
Source: AFP

Researchers at Microsoft have revealed a new artificial tool that can create deeply realistic human avatars -- but offered no timetable to make it available to the public, citing concerns about facilitating deep fake content.

The AI model known as VASA-1, for "visual affective skills," can create an animated video of a person talking, with synchronized lip movements, using just a single image and a speech audio clip.

Disinformation researchers fear rampant misuse of AI-powered applications to create "deep fake" pictures, video, and audio clips in a pivotal election year.

"We are opposed to any behavior to create misleading or harmful contents of real persons," wrote the authors of the VASA-1 report, released this week by Microsoft Research Asia.

"We are dedicated to developing AI responsibly, with the goal of advancing human well-being," they said.

Read also

Kyrgyzstan's TikTok block builds censorship fears

"We have no plans to release an online demo, API, product, additional implementation details, or any related offerings until we are certain that the technology will be used responsibly and in accordance with proper regulations."

Microsoft researchers said the technology can capture a wide spectrum of facial nuances and natural head motions.

"It paves the way for real-time engagements with lifelike avatars that emulate human conversational behaviors," researchers said in the post.

VASA can work with artistic photos, songs, and non-English speech, according to Microsoft.

Researchers touted potential benefits of the technology such as providing virtual teachers to students or therapeutic support to people in need.

"It is not intended to create content that is used to mislead or deceive," they said.

VASA videos still have "artifacts" that reveal they are AI-generated, according to the post.

Read also

Quarter of UK 5 to 7-year-olds have smart phone: study

ProPublica technology lead Ben Werdmuller said he'd be "excited to hear about someone using it to represent them in a Zoom meeting for the first time."

"Like, how did it go? Did anyone notice?" he said on social network Threads.

ChatGPT-maker OpenAI in March revealed a voice-cloning tool called "Voice Engine" that can essentially duplicate someone's speech based on a 15-second audio sample.

But it said it was "taking a cautious and informed approach to a broader release due to the potential for synthetic voice misuse."

Earlier this year, a consultant working for a long-shot Democratic presidential candidates admitted he was behind a robocall impersonation of Joe Biden sent to voters in New Hampshire, saying he was trying to highlight the dangers of AI.

The call featured what sounded like Biden's voice urging people not to cast ballots in the state's January's primary, sparking alarm among experts who fear a deluge of AI-powered deep fake disinformation in the 2024 White House race.

Source: AFP

Authors:
AFP avatar

AFP AFP text, photo, graphic, audio or video material shall not be published, broadcast, rewritten for broadcast or publication or redistributed directly or indirectly in any medium. AFP news material may not be stored in whole or in part in a computer or otherwise except for personal and non-commercial use. AFP will not be held liable for any delays, inaccuracies, errors or omissions in any AFP news material or in transmission or delivery of all or any part thereof or for any damages whatsoever. As a newswire service, AFP does not obtain releases from subjects, individuals, groups or entities contained in its photographs, videos, graphics or quoted in its texts. Further, no clearance is obtained from the owners of any trademarks or copyrighted materials whose marks and materials are included in AFP material. Therefore you will be solely responsible for obtaining any and all necessary releases from whatever individuals and/or entities necessary for any uses of AFP material.