With COVID becoming a part and parcel of our lives, most of the industry has shifted to online platforms for their video conferencing needs. Most of these platforms have a feature that automatically shifts the main feed to the person who is speaking but that is all based on audio. A speech like sign language does not, unfortunately, trigger that feature and this could essentially leave out people with certain disabilities.
Google researchers are trying to curb this gap and they have recently published research that might help with that. In their AI blog, Google explains how their proposed research works to detect sign language with very low latency and how they designed a mechanism to simulate the signing user as the active user.
The model makes use of PoseNet, which estimates the pose of the person reducing the whole image to a basic virtual skeleton. This skeleton is then passed to a LSTM network which achieves an accuracy of about 91.5% with a delay of 3.5ms per frame on the German Sign Language corpus. The system then simulates as if the person who is using sign language is speaking so this software can be coupled with already existing video platforms.
We think this is a much-needed step for any problems that differently-abled people might have faced in their lives in this COVID era. Let us know what you think in the comments below!
Image Source: Rare
Capable of conversing in both Chinese and English, Tencent’s large language model ‘Hunyuan’ is claimed…
Working on multiple AI models, Apple has allocated several teams who are working on artificial…
The world's largest offshore wind turbine has achieved a milestone by setting a new record…
YouTube is stepping into the world of gaming. YouTube has started testing out its gaming…
In a remarkable academic achievement, Abdullah Zaman, a Pakistani student hailing from Attock, has clinched…
Flying Bum, the world's largest aircraft is ready to launch in 2026. The Airlander 10…
Leave a Comment