性视界传媒

AI-assisted computer vision research aims to improve accessibility for deaf, hard of hearing

In This Story

People Mentioned in This Story
Body

Digital assistants like Amazon鈥檚 Alexa aren鈥檛 currently useful for, say, the hard of hearing and deaf community. 性视界传媒 researchers led by Jana Ko拧eck谩 are making the Internet of Things more听inclusive and accessible to those for whom it has not been designed. For the next year, her work to improve "seeing" computer systems to translate continuous American Sign Language into English will be funded by Amazon鈥檚 Fairness in AI Research Program.听

portrait of Jana Kosecka
Jana Ko拧eck谩. Photo by Ron Aira/Creative Services

Having worked at Mason for more than 20 years, Ko拧eck谩 began studying computer vision as it applies to American Sign Language in 2019 with colleagues Huzefa Rangwala and Parth Pathak in collaboration with Gallaudet University. Their work resulted in three academic publications on the topic in 2020. The team鈥檚 initial work focused on computer vision recognizing American Sign Language at the word level. 听

Ko拧eck谩 describes her current work as a continuation of earlier work, but now, especially with the help of AI, she鈥檚 tackling more complex ASL content, such as sentence-level communication, facial expressions, and very specific hand gesticulation.

鈥淭he challenge of extending some of these ideas [of computer translation] to American Sign Language translation is the input is video as opposed to text; it's continuous, and you have a lot of challenges, because you have a lot of variations about how people sign,鈥 says Ko拧eck谩.

The project is accordingly multifaceted.听鈥淲e are focusing on better hand modeling, focusing on incorporating the facial features and extending to continuous sign language, so you can have short phrases the model can translate to English,鈥 Ko拧eck谩 explains. 鈥淲e are basically trying to capture continuous sign language and not just individual words."听

To accomplish this goal,Ko拧eck谩 is using weakly supervised learning machine learning methods that provide mechanisms to teach the system without excessive human labelling effort.

Weakly supervised learning techniquesdon't needperfect alignment of video sequences that contain multiple words,鈥 she says. In the word-level recognition, the听model is presented听with examples of a听video听snippet of a signed word听and the听word text,听so it听has听perfect supervision. Given many examples of听the sign apple听as a video snippet,听the system will learn to recognize the word 'apple.' 鈥

鈥淭here are some techniques which can discover patterns without this need of direct supervision. If you just give the model a lot of examples, the model will figure out repeating patterns of certain words occurring in certain contexts,鈥 she says. 鈥淪o we are听applying these machine-learning techniques to the setting of American Sign Language.鈥 听

Relating her work to AI-powered chatbots like chatGPT, Ko拧eck谩 says, 鈥淭here has been a lot of headway made in this space for written and spoken languages, and we would like to make a little bit of听headway for American Sign Language, using some of these insights and ideas.鈥澨

Ko拧eck谩 envisions her research helping improve the interface between hard of hearing people and their environment, whether that be when they鈥檙e communicating with Amazon鈥檚 Alexa or ordering at a restaurant counter. No doubt her work will help improve inclusivity and accessibility for the deaf and hard of hearing听both at Mason and beyond. 听