DeepMind’s lip-reading AI is already more accurate than humans, but still has a way to go

Even professional lip-readers can figure out only 20% to 60% of what a person is saying. Slight movements of a person’s lips at the speed of natural speech are immensely difficult to reliably understand, especially from a distance or if the lips are obscured. And lip-reading isn’t just a plot point in NCIS: It’s an essential tool to understand the world for the hearing-impaired, and if automated reliably, could help millions. A new paper (pdf) from Alphabet’s DeepMind and the University of Oxford details an artificial intelligence system, called LipNet, that watches video of a person speaking and matches text to the movement of their mouth with 93.4% accuracy. The previous state of the art system operated word-by-word, and had an accuracy of 79.6%. The DeepMind and Oxford researchers say…


Link to Full Article: DeepMind’s lip-reading AI is already more accurate than humans, but still has a way to go

Pin It on Pinterest

Share This

Join Our Newsletter

Sign up to our mailing list to receive the latest news and updates about homeAI.info and the Informed.AI Network of AI related websites which includes Events.AI, Neurons.AI, Awards.AI, and Vocation.AI

You have Successfully Subscribed!