U.S. flag

An official website of the United States government

Here's how you know

Official websites use .gov A .gov website belongs to an official government organization in the United States.

Secure .gov websites use HTTPS A lock ( Lock A locked padlock ) or https:// means you've safely connected to the .gov website. Share sensitive information only on official, secure websites.

Home

Assistive Devices for People with Hearing, Voice, Speech, or Language Disorders

On this page:

What are assistive devices?

What types of assistive devices are available, what types of assistive listening devices are available, what types of augmentative and alternative communication devices are available for communicating face-to-face, what augmentative and alternative communication devices are available for communicating by telephone, what types of alerting devices are available, what research is being conducted on assistive technology, where can i get more information.

The terms assistive device or assistive technology can refer to any device that helps a person with hearing loss or a voice, speech, or language disorder to communicate. These terms often refer to devices that help a person to hear and understand what is being said more clearly or to express thoughts more easily. With the development of digital and wireless technologies, more and more devices are becoming available to help people with hearing, voice, speech, and language disorders communicate more meaningfully and participate more fully in their daily lives.

Health professionals use a variety of names to describe assistive devices:

  • Assistive listening devices (ALDs) help amplify the sounds you want to hear, especially where there’s a lot of background noise. ALDs can be used with a hearing aid or cochlear implant to help a wearer hear certain sounds better.
  • Augmentative and alternative communication (AAC) devices help people with communication disorders to express themselves. These devices can range from a simple picture board to a computer program that synthesizes speech from text.
  • Alerting devices connect to a doorbell, telephone, or alarm that emits a loud sound or blinking light to let someone with hearing loss know that an event is taking place.

Several types of ALDs are available to improve sound transmission for people with hearing loss. Some are designed for large facilities such as classrooms, theaters, places of worship, and airports. Other types are intended for personal use in small settings and for one-on-one conversations. All can be used with or without hearing aids or a cochlear implant. ALD systems for large facilities include hearing loop systems, frequency-modulated (FM) systems, and infrared systems.

Hearing Loop Installed. Switch hearing aid to T-coil, www.hearingloop.org

This logo informs people that a public area is looped. Source: HearingLoop.org

Hearing loop (or induction loop) systems use electromagnetic energy to transmit sound. A hearing loop system involves four parts:

  • A sound source, such as a public address system, microphone, or home TV or telephone
  • An amplifier
  • A thin loop of wire that encircles a room or branches out beneath carpeting
  • A receiver worn in the ears or as a headset

Amplified sound travels through the loop and creates an electromagnetic field that is picked up directly by a hearing loop receiver or a telecoil (see sidebar), a miniature wireless receiver that is built into many hearing aids and cochlear implants. To pick up the signal, a listener must be wearing the receiver and be within or near the loop. Because the sound is picked up directly by the receiver, the sound is much clearer, without as much of the competing background noise associated with many listening environments. Some loop systems are portable, making it possible for people with hearing loss to improve their listening environments, as needed, as they proceed with their daily activities. A hearing loop can be connected to a public address system, a television, or any other audio source. For those who don’t have hearing aids with embedded telecoils, portable loop receivers are also available.

What’s a telecoil?

A telecoil, also called a t-coil, is a coil of wire that is installed inside many hearing aids and cochlear implants to act as a miniature wireless receiver. It was originally designed to make sounds clearer to a listener over the telephone. It also is used with a variety of other assistive listening devices, such as hearing loop (or induction loop) systems, FM systems, infrared systems, and personal amplifiers.

The telecoil works by receiving an electromagnetic signal from the hearing loop and then turning it back into sound within the hearing aid or cochlear implant. This process eliminates much of the distracting background noise and delivers sound customized for one’s own need. For people who are hard-of-hearing who do not have a telecoil-equipped hearing aid or cochlear implant, loop receivers with headsets can provide similar benefits but without the customized or “corrected sound” feature that matches one’s hearing loss pattern.

Many cochlear implants have a telecoil built into the sound processor, or can use an external telecoil accessory with both hearing aid compatible telephones and public loop systems. A simple switch or programming maneuver performed by the user activates this function.

See the NIDCD fact sheet Hearing Aids for more information.

FM systems use radio signals to transmit amplified sounds. They are often used in classrooms, where the instructor wears a small microphone connected to a transmitter and the student wears the receiver, which is tuned to a specific frequency, or channel. People who have a telecoil inside their hearing aid or cochlear implant may also wear a wire around the neck (called a neckloop) or behind their aid or implant (called a silhouette inductor) to convert the signal into magnetic signals that can be picked up directly by the telecoil. FM systems can transmit signals up to 300 feet and are able to be used in many public places. However, because radio signals are able to penetrate walls, listeners in one room may need to listen to a different channel than those in another room to avoid receiving mixed signals. Personal FM systems operate in the same way as larger scale systems and can be used to help people with hearing loss to follow one-on-one conversations.

Infrared systems use infrared light to transmit sound. A transmitter converts sound into a light signal and beams it to a receiver that is worn by a listener. The receiver decodes the infrared signal back to sound. As with FM systems, people whose hearing aids or cochlear implants have a telecoil may also wear a neckloop or silhouette inductor to convert the infrared signal into a magnetic signal, which can be picked up through their telecoil. Unlike induction loop or FM systems, the infrared signal cannot pass through walls, making it particularly useful in courtrooms, where confidential information is often discussed, and in buildings where competing signals can be a problem, such as classrooms or movie theaters. However, infrared systems cannot be used in environments with too many competing light sources, such as outdoors or in strongly lit rooms.

Personal amplifiers are useful in places in which the above systems are unavailable or when watching TV, being outdoors, or traveling in a car. About the size of a cell phone, these devices increase sound levels and reduce background noise for a listener. Some have directional microphones that can be angled toward a speaker or other source of sound. As with other ALDs, the amplified sound can be picked up by a receiver that the listener is wearing, either as a headset or as earbuds.

The simplest AAC device is a picture board or touch screen that uses pictures or symbols of typical items and activities that make up a person’s daily life. For example, a person might touch the image of a glass to ask for a drink. Many picture boards can be customized and expanded based on a person’s age, education, occupation, and interests.

Keyboards, touch screens, and sometimes a person’s limited speech may be used to communicate desired words. Some devices employ a text display. The display panel typically faces outward so that two people can exchange information while facing each other. Spelling and word prediction software can make it faster and easier to enter information.

Speech-generating devices go one step further by translating words or pictures into speech. Some models allow users to choose from several different voices, such as male or female, child or adult, and even some regional accents. Some devices employ a vocabulary of prerecorded words while others have an unlimited vocabulary, synthesizing speech as words are typed in. Software programs that convert personal computers into speaking devices are also available.

For many years, people with hearing loss have used text telephone or telecommunications devices, called TTY or TDD machines, to communicate by phone. This same technology also benefits people with speech difficulties. A TTY machine consists of a typewriter keyboard that displays typed conversations onto a readout panel or printed on paper. Callers will either type messages to each other over the system or, if a call recipient does not have a TTY machine, use the national toll-free telecommunications relay service at 711 to communicate. (See Telecommunications Relay Services for more information.) Through the relay service, a communications assistant serves as a bridge between two callers, reading typed messages aloud to the person with hearing while transcribing what’s spoken into type for the person with hearing loss.

With today’s new electronic communication devices, however, TTY machines have almost become a thing of the past. People can place phone calls through the telecommunications relay service using almost any device with a keypad, including a laptop, personal digital assistant, and cell phone. Text messaging has also become a popular method of communication, skipping the relay service altogether.

Another system uses voice recognition software and an extensive library of video clips depicting American Sign Language to translate a signer’s words into text or computer-generated speech in real time. It is also able to translate spoken words back into sign language or text.

Finally, for people with mild to moderate hearing loss, captioned telephones allow you to carry on a spoken conversation, while providing a transcript of the other person’s words on a readout panel or computer screen as back-up.

Alerting or alarm devices use sound, light, vibrations, or a combination of these techniques to let someone know when a particular event is occurring. Clocks and wake-up alarm systems allow a person to choose to wake up to flashing lights, horns, or a gentle shaking.

Visual alert signalers monitor a variety of household devices and other sounds, such as doorbells and telephones. When the phone rings, the visual alert signaler will be activated and will vibrate or flash a light to let people know. In addition, remote receivers placed around the house can alert a person from any room. Portable vibrating pagers can let parents and caretakers know when a baby is crying. Some baby monitoring devices analyze a baby’s cry and light up a picture to indicate if the baby sounds hungry, bored, or sleepy.

The National Institute on Deafness and Other Communication Disorders (NIDCD) funds research into several areas of assistive technology, such as those described below.

  • Improved devices for people with hearing loss NIDCD-funded researchers are developing devices that help people with varying degrees of hearing loss communicate with others. One team has developed a portable device in which two or more users type messages to each other that can be displayed simultaneously in real time. Another team is designing an ALD that amplifies and enhances speech for a group of individuals who are conversing in a noisy environment.
  • More natural synthesized speech NIDCD-sponsored scientists are also developing a personalized text-to-speech synthesis system that synthesizes speech that is more intelligible and natural sounding to be incorporated in speech-generating devices. Individuals who are at risk of losing their speaking ability can prerecord their own speech, which is then converted into their personal synthetic voice.
  • Brain–computer interface research A relatively new and exciting area of study is called brain–computer interface research. NIDCD-funded scientists are studying how neural signals in a person’s brain can be translated by a computer to help someone communicate. For example, people with amyotrophic lateral sclerosis (ALS, or Lou Gehrig’s disease) or brainstem stroke lose their ability to move their arms, legs, or body. They can also become locked-in, where they are not able to express words, even though they are able to think and reason normally. By implanting electrodes on the brain’s motor cortex, some researchers are studying how a person who is locked-in can control communication software and type out words simply by imagining the movement of his or her hand. Other researchers are attempting to develop a prosthetic device that will be able to translate a person’s thoughts into synthesized words and sentences. Another group is developing a wireless device that monitors brain activity that is triggered by visual stimulation. In this way, people who are locked-in can call for help during an emergency by staring at a designated spot on the device

The NIDCD maintains a directory of organizations that provide information on the normal and disordered processes of hearing, balance, taste, smell, voice, speech, and language. 

Use the following keywords to help you search for organizations that can answer questions and provide printed or electronic information on assistive devices:

  • Assistive technology
  • Assistive listening device
  • Augmentative and alternative communication

For more information, contact us at:

NIDCD Information Clearinghouse 1 Communication Avenue Bethesda, MD 20892-3456 Toll-free voice: (800) 241-1044 Toll-free TTY: (800) 241-1055 Email: [email protected]

NIH Publication No. 11-7672 December 2011

*Note: PDF files require a viewer such as the free Adobe Reader .

University Library, University of Illinois at Urbana-Champaign

University of Illinois Library Wordmark

Speech Disorders: Common Assistive Technologies

  • Popular Literature
  • Web Resources
  • Reference Resources
  • Academic Resources
  • Common Assistive Technologies

What are assistive technologies?

The Technology Related Assistance to Individuals with Disabilities Act of 1988 described an assistive technology device as "any item, piece of equipment, or product system, whether acquired commercially off the shelf, modified, or customized, that is used to increase, maintain, or improve functional capabilities of individuals with disabilities."

Assistive technologies can be "high tech" and "low tech:" from canes and lever doorknobs to voice recognition software and augmentative communication devices (speech generating devices).

Augmentative and Alternative Communication Devices

Augmentative and alternative communication (AAC) i s the use of symbols, aids, strategies, and techniques to enhance the communication process. This includes sign language, various communication boards, and both manual and electronic  devices help those who have trouble with communication.

Some examples of AAC include: 

1. Unaided communication systems  – Rely on the user's body to convey messages. Examples include gestures, body language, and/or sign language. One advantage of unaided communication communication systems is that it does not require any technology beyond the person's body.

2.  Low-tech AAC  - Any type of aid that does not require batteries or electricity. This includes things like a simple pen and paper to write messages on, as well as pictures boards, that can be carried to aid communication. On picture boards, users can point to images, words, pictures, drawings, or letters in order to communicate their message. The pointing might be done with the user's hands, other body parts, eye gaze, or a pointer held in the hands or mouth. 

3.  High-tech AAC  - Any aid that requires electricity or batters. This includes specialized devices, software, smartphone applications, electronic communication boards, and keyboards. Many high-tech AAC devices are Speech Generating Devices, which means they can produce digitized speech when the user either types a message or presses on images, words, or letters.

Below are just some examples of what AACs can look like

speech and language impairment assistive technology

Left to Right: Images 1, 2, 4, 5 7, ©  User:Poule /  Wikimedia Commons  /  CC-BY-SA-3.0 . Image 8  ©  User Joxerrazabala  /  Wikimedia Commons  /  CC-BY-SA-3.0 . Image 3   © User:pennstatelive  /  flickr /  CC BY-NC-ND 2.0 .  Image 6 in public domain.

Electronic Fluency Devices

Electronic Fluency Devices are devices intended to help improve the fluency of people with stutters. They do this by playing the sound of the user's own voice back into their ear, slightly altered. 

There are two main types of Electronic Fluency Devices :

Delayed Auditory Feedback (DAF)-  Delayed Auditory Feedback devices play the user's voice back delayed by a fraction of a second. DAF devices may resemble hearing aids or headphones with a microphone. There are also apps that can use DAF on phone calls.

Frequency Altered Feedback (FAF)-  Frequency Altered Feedback devices are similar to DAF but rather than delaying the user hearing their own voice, they change the pitch at which the user hears their own voice.

Employees With Speech Disorders

  • Accommodations for Employees with Speech Disorders The Job Accommodation Center provides guidance on workplace accommodations and disability employment issues. This page lists resources for effective accommodations for those with speech disorders.

Finding Assistive Technology

  • EduTopia: Assistive Technology Discover websites, blog posts, articles, and videos related to understanding, selecting, and assessing assistive technology.

Mobile Applications

  • Dragon Dictation App Dragon Dictation app can be used to turn speech into text, as well as text to speech. This app is only available for iOS.
  • TTSReader This app is available through Google Play. It instantly reads out loud any text with natural sounding voices. You can either enter text or provide a website address that you want it to read.
  • DAF Professional Delayed Auditory Feedback (DAF) app for iPhone or Android
  • Stamurai: Stuttering Therapy Speech therapy app that includes DAF, reading practice, and other practice techniques.
  • Smarty Ears Apps A variety of apps for assistance with several different speech disorders
  • << Previous: Academic Resources
  • Last Updated: Nov 9, 2023 9:11 AM
  • URL: https://guides.library.illinois.edu/c.php?g=613892

Job Accommodation Network

Assistive Technology Solutions for Employees with Speech Impairments

From the desk of teresa goddard, m.s., lead consultant – assistive technology services.

Have you ever wondered how a person who hears but doesn’t speak uses the phone? Of course there are TTY and speech-to-speech relay services, but what if relay methods don’t meet the employee’s needs? What if the nature of the work requires a more direct and confidential method of communicating?

There are devices for phone and face-to-face communication that are designed for individuals who do not speak at all or who find speaking very challenging called AAC (Alternative and Augmentative Communication). AAC devices, also called speech-generating devices, are an example of a type of technology that can be used by individuals who have difficulty speaking. JAN has general information about AAC devices as well as information about AAC with telephone access .

Ideally when AAC is being considered, a speech language pathologist with expertise in AAC would be involved. The American Speech-Language-Hearing Association has a special interest group of professionals working in AAC.

It may be worthwhile to see if the State AT project in your area would be able to at a minimum demo some products. In some states they may also be able to perform assessments. State AT projects are funded under the Technology-related Assistance for Individuals with Disabilities Act and support consumer-driven state plans for the delivery of AT.

While JAN does not endorse or recommend specific vendors and products, there are some specific products worth noting:

  • Tobii Dynavox is an example of a vendor that has some products that may be of interest. For example, Tobii Dynavox has the Lightwriter for mobile calling. 
  • Zygo carries an adaptor that can be used to connect speech-generating devices to an office phone. It is also possible to use many AAC/speech-generating devices via a speakerphone. Learn more at http://aac.textspeak.com/demos/ .
  • The sound quality and naturalness of synthetic speech is often a concern, particularly for new users of AAC. There is now a company specializing in customized synthetic voices .
  • Sometimes employees may benefit from merely having their natural voice amplified .

For more information, contact JAN directly. 

phone console

Add Page to MyJAN

  • Back issues

speech and language impairment assistive technology

Home » Communication Skills » What Assistive Technology for Speech and Language Disorders Are Available and How do They Work?

What Assistive Technology for Speech and Language Disorders Are Available and How do They Work?

By   Donnesa McPherson, AAS

September 27, 2023

There are many different types of assistive technology for speech and language disorders available on the market today. With the range of needs, the technology can help with communicating with others, hearing what others are saying, and emergencies.

There are many other uses for this technology ranging from non verbal autism and ranging across all the other communication disorders that there are.

Assistive technology that is available

These are the main types of assistive technology available:

Augmentative and alternative communication (AAC):

These help people with speech and language impairments with language skills and communication. Examples can range from visual aids which include sign language, communication boards, all the way to speech generating devices.

Assistive listening devices (ALD):

These create amplified sound that will transmit sound to the individual and help cut distracting background noise. This would include hearing aids and personal amplifiers to inner cochlear implants that help improve sound transmission to the individual.

Devices that alert

These are devices that make loud sounds and can connect to the telephone or can be a part of an alarm system that can produce a light signal or other alert to the individual to let them know something is happening.

Infrared systems

These systems are worn by the individual and use alerting devices connected by using infrared light to amplify sounds. The use of infrared systems cannot transmit through walls making it a good choice when private and sensitive information is being shared because it is a closed system that stays within the hearing aids or inner cochlear implant.

Personal amplifiers

These help cut down on unwanted background noise when some of the other systems may be unavailable like in a car. Devices that are about cell phone size help increase sound while cutting back on unwanted background sounds.

Hearing loop systems

These are also known as induction loop systems where the transmitter converts sound through electromagnetic energy that has four main parts.

Four parts of a hearing loop system include:

  • A central source (microphone, television, etc.)
  • Sound converter or amplifier
  • An array of thin wires that are placed around a room or under the carpet or flooring
  • Receiver (headset, etc.)

These systems are able to spread amplified sound through radio signals. These can be used in larger areas, like a presentation, where the presenter uses specialized devices, like a microphone, and the individual has a receiver on a special channel to listen to the speech.

Which device is the best?

That question is best answered and is dependent on the individual and why they are using the device. Since the devices range from about the size of a cell phone size that the individual can carry up to specialized devices and software programs for those with speech difficulties and hearing loss, it is dependent on why the devices are being used.

This can seem like quite a large decision but can be helped by speech language pathologists, an occupational therapist, and/or your child’s doctor. These professionals are able to make referrals and know what needs to be considered when making this choice.

There are also augmentative and alternative communication (AAC) evaluations that can evaluate an individual’s skill level and needs. These evaluations can be a key point to consider when making this decision.

Takeaways and key points

There are so many different devices available that help support individuals with speech disorders and hearing loss. The needs and services that these technologies provide have a significant range and can be covered by an individual’s insurance and/or school.

I would recommend talking to a speech-language pathologist, occupational therapist, or your child’s doctor for recommendations of devices and what they think could benefit the individual based on their needs.

There are AAC evaluations that can be referred and could benefit your child and help aid the search for the device that would best support the individual. These evaluations will take into account what your child does, their skill level, and many other aspects.

Evaluations and referrals are a key and important part that can help pinpoint what device is going to work the best. Also what device will provide the most support and allow for success because the child’s lifestyle and skill level were taken into consideration when searching for the best device.

In conclusion

It is always important to keep communication open between you and your child’s doctor and any other professionals that make up the team the helps with development and support for the individual.

Autism Parenting Magazine does not endorse or promote certain devices, therapies, or services. Those decisions are best made by the individual’s parent and/or guardian and the child’s doctor.

It also helps to connect with other parents and professionals through support groups . Those support groups can be in person or online, social media is another great place to look.

There are so many options and opinions to consider. As long as the information and input are coming from those people that you trust and have had to make the same decisions, it can definitely be beneficial to both the parent and/or guardian and the individual that needs the support device.

It can also help to have training and work with individuals that have had the professional development and experience with the devices and talking to and recommending them. There are typically a myriad of opportunities for parents to receive additional training in their area and can be found in doctor’s or therapist’s offices, at support groups, online and social media. etc.

As always, double check the source of any training or recommendations you receive with your child’s doctor and therapists. Keeping that communication loop open is so important and can make decisions like choosing the device that’s right for your child a lot easier.

Hobbs, K. (2021). Assistive Communication Devices for Children with Autism. https://www.autismparentingmagazine.com/assistive-technology-autism/

National Institute on Deafness and Other Communication disorders. (2019). Assistive Devices for People with Hearing, Voice, Speech, or Language Disorders . https://www.nidcd.nih.gov/health/assistive-devices-people-hearing-voice-speech-or-language-disorders

Support Autism Parenting Magazine

We hope you enjoyed this article. In order to support us to create more helpful information like this, please consider purchasing a subscription to Autism Parenting Magazine.

Download our FREE guide on the best Autism Resources for Parents

Where shall we send the PDF?

Enter you email address below to  download your FREE guide & receive top autism parenting tips direct to your inbox

Privacy Policy

Related Articles

How to Communicate with a Nonverbal Autistic Child

9 tips on how to introduce yourself to a child with autism, simple tips for family communication in autism, communication boards for autistic children, three hacks for improving communication with autistic children, 5 tips for choosing the best toys for late talkers, how to help babies and toddlers understand and use gestures, why an autism emotion chart can be beneficial, how to increase functional communication at home, joint attention in children with autism spectrum disorder, the link between visual discrimination and autism, what do you mean i’m not communicating, privacy overview.

Get a FR E E   issue 

of the magazine & top autism parenting tips to your inbox

We respect the privacy of your email address and will never sell or rent your details.

Autism Parenting Magazine

Where shall we send it?

Enter your email address below to get a free issue of the magazine & top autism tips direct your inbox

Consider a Career with Beyond Therapy! From part-time positions to sign-on bonuses, we could be exactly what you’re looking for!

speech and language impairment assistive technology

  • Mission & Core Values
  • Philanthropy
  • Our Brand Family
  • Sell Your Practice
  • Partnership Program
  • Professional Development

Part of the Upstream Rehab Family of Care

  • What To Expect
  • Registration Forms
  • Share Your Story
  • Good Faith Estimate
  • Pediatric Occupational Therapy
  • Pediatric Physical Therapy
  • Pediatric Speech Therapy

Assistive Technology in Speech Therapy

For some patients, an alternate form of communication is best for them and will help them be more functional participants in their environments. Augmentative and Alternative Communication (AAC) devices are frequently used in our sessions to establish the best functional speech outcome for our patients. These devices range from low tech to high tech.

A low-tech option may be a picture or symbol system known as Picture Exchange Communication System (PECS). This system can be used to express wants/needs or as a scheduler for kids who have difficulty with transitions. ( https://pecsusa.com/pecs/ )

speech and language impairment assistive technology

A high tech option that is used in our clinic is the Language Acquisition through Motor Planning (LAMP) Words for Life. This software is downloaded onto an iPad and gives patients access to approximately 4,000 words. These words can be used at the single word level or combined to make phrases and sentences. This software is able to be customized to the skill level of the child. ( https://www.prentrom.com/prc_advantage/lamp-words-for-life-language-system )

speech and language impairment assistive technology

These low and high tech AAC options are incorporated in our typical speech therapy sessions in conjunction with verbal language models. A frequent question that parents ask in regards to use of an AAC device is “will this stop my child from learning how to speak?”. Research has shown that using AAC does not stop an AAC user from learning to speak. There are several other options for AAC that can be explored to find the best fit for the child.

“The Impact of Augmentative and Alternative Communication Intervention on the Speech Production of Individuals With Developmental Disabilities: A Research Review” by Diane C. Millar, Janice C. Light and Ralf W. Schlosser in Journal of Speech, Language, and Hearing Research, April 2006, Vol. 49, 248-264. doi:10.1044/1092-4388(2006/021)

“Effects of Augmentative and Alternative Communication Intervention on Speech Production in Children With Autism: A Systematic Review” by Ralf W. Schlosser and Oliver Wendt in American Journal of Speech-Language Pathology, August 2008, Vol. 17, 212-230. doi:10.1044/1058-0360(2008/021)

Presley Napier, M.S., CCC-SLP Beyond Therapy for Kids- Hattiesburg Clinic Director/Speech Language Pathologist

Our Locations

  • Share Story
  • Constitution
  • Mailing List

Welcome to the homepage of SIG-SLPAT!

The purpose of this SIG is to promote the application of speech and language technologies to the field of assistive technologies , especially for people whose impairments may affect their abilities to communicate.

Speech and Language Processing for Assistive Technologies is an inherently interdisciplinary endeavor that attracts researchers from such diverse fields as natural language processing (NLP) , speech processing , speech pathology , augmentative and alternative communication (AAC) , human-computer interaction (HCI) , and accessibility . SIG-SLPAT works to bring all of these disciplines and researchers together toward a common goal. Together, we can be more effective than the sum of our parts.

Latest News and Upcoming Events

We are pleased to announce the 4th Clarity Workshop for Machine Learning Challenges for Hearing Aids, to be co-organised with SIG-SLPAT and collocated with Interspeech 2023, Dublin, Ireland. This will be the first in-person Clarity event, following the 1st , 2nd and 3rd Clarity workshops that were held as online events.

View all events...

Clarity Workshop

  • International edition
  • Australia edition
  • Europe edition

How technology is changing speech and language therapy

From robots that play peekaboo, to speech recognition software that analyses TV shows, tech is being used to aid human communication

Kaspar

Speech and communication skills are at the heart of human relationships – without them we couldn’t share ideas and emotions. But technology is carving out a special role in boosting those skills. Pioneering research shows just how machines are helping people to make themselves understood.

Here we look at three projects where a range of academic specialists and industry partners have come together to develop and widen access to their innovations.

Kaspar the robot: helping children with autism communicate

Meet Kaspar: he can be talked to, tickled, stroked, played with and you can even prod and poke him and he won’t run away. Kaspar, developed at the University of Hertfordshire by a team under professor Kerstin Dautenhahn, is a child-like talking robot with a simplified human face and moveable limbs and features. He’s designed to help children with autism develop essential social skills through games such as peekaboo and learning activities.

Kaspar, the size of a small child, was “born” back in 2005 and has been developed since thanks to funding raised by the university. The multi-disciplinary scope of the project, spanning robotics, psychology, assistive technology and autism therapy, harnesses technology to assist communication. But this broad approach means it falls between research council stools and misses out on their grants, says Dautenhahn.

Initially, Kaspar has been used to help children in schools under the supervision of researchers. In the latest phase of the research, redesigned, wireless and more personalised versions of the little robot – controlled using a tablet - will to go out directly to schools and families in the next few weeks.

Parents and teachers will play games such as encouraging autistic youngsters to mimic and discuss different facial expressions, or even to pinch him and discuss why he cries out and looks sad, recording the results for the Hertfordshire team to analyse. “This is a new field study phase where Kaspar will go out into the world without the helping hand of researchers,” says Dautenhahn, whose work has combined both academic research – including collaboration with psychologists and clinicians - and the nuts and bolts of developing the robot as a potential mass product.

It is Kaspar’s highly predictable, simplified interactions that appeal to autistic children who may be overwhelmed by the complexity of everyday human communication, she believes.

“His simplicity appeals to children, and the fact that they can respond to him in their own time. If you are silent for 60 seconds, Kaspar won’t mind – he doesn’t make judgments.”

Alongside analysis of the new phase results, Dautenhahn and her team are seeking an investor to help give wider access to Kaspar at an affordable cost for schools and families.

Computers helping stroke patients with chronic speech impairments

For stroke sufferers, damage to speech can be a common and debilitating consequence, often requiring long periods of therapy. A new computerised treatment developed at Sheffield University and now used by healthcare teams in a range of countries helps patients treat themselves at home by following a carefully-staged programme designed to gradually rebuild speech and the layers of connections that lie beneath it.

The research, led by Sandra Whiteside and professor Rosemary Varley (now at UCL), rests on a new and controversial understanding of the way speech works. While traditional theories said speech output was put together sound by sound, the two academics developed an alternative model, arguing that speech in fact depends on plans stored in our brains for words and whole phrases.

Based on this theory, the women devised a computerised treatment – Sheffield Word, known as Sword – giving stroke and brain damage patients high-intensity stimulation through images, words and sounds to try to kick-start damaged nerve systems and get them speaking again.

“We are trying to connect up the whole system,” says Whiteside. “Rather than just focusing on getting people to speak from the outside, you are training up their perceptive system, bombarding it with all kinds of stimuli before you get them to produce speech.”

A patient using the software might start by hearing a word, such as “cat”, and matching it with an image from a choice of four on screen. A subsequent stage would see them imagining the sound of the word, then later imagining producing that sound. “Speech, the last tough hurdle, comes right at the end,” Whiteside explains.

With results showing significant improvements for patients with chronic speech impairments, the Sheffield team are now mining the data from a Bupa-funded study of 50 patients. Funding for this research area has “always been very difficult”, says Whiteside, but over 200 copies of the programme have also been sold, mainly to NHS trusts.

Dr Who darlek

Speech recognition: shrinking the gap between human and machine

When humans listen to speech, we’re extraordinarily forgiving: we’ll make out words against the sound of crashing waves or engine noise, filter out a single voice amid the hubbub of a football crowd and cope with unfamiliar accents and rushed or emotional speech.

While giant leaps have been made both in speech recognition technologies (converting audio to text) and speech synthesis over recent decades, machines have yet to catch up with our ability to interact fast and naturally, constantly learning and adapting as we do so.

Research being conducted by scientists at the universities of Cambridge, Edinburgh and Sheffield seeks to shrink the performance gap between machine and human, aiming to make speech technologies more usable and natural. The five-year, £6.2m project funded by the Engineering and Physical Sciences Research Council is intended to develop a new generation of the technology, opening up a wide range of potential new applications.

“We want to take the idea of adaptation in speech technology and take it significantly further than it has ever been used before,” says professor Phil Woodland of Cambridge’s department of engineering. Currently, Woodland explains, machines – which work by breaking down speech into sounds and calculating the probabilities of likely words and sentences - have required significant amounts of data from individual speakers in order to learn sufficiently to recognise their speech, entailing a long process of “training”.

“Our aim now is to improve the technology to allow rapid adaptation and unsupervised or lightly supervised training,” says Woodland, whose team have worked with a range of industry partners including Google, Nuance and IBM. “We want to leverage data from different situations.”

Thanks to the BBC, the researchers have been able to test their work on several months of the broadcaster’s output. While transcribing clearly articulated news reports is straightforward, piecing together speech in an episode of Dr Who – with its unearthly noises and music – or fast, excited commentary in sports programmes is more challenging.

The same technology is also being used by the team to develop personalised speech synthesis systems, designed to reflect a user’s individual accent and expressions.

Join the higher education network for more comment, analysis and job opportunities, direct to your inbox. Follow us on Twitter @gdnhighered .

  • Universities
  • Higher education

Comments (…)

Most viewed.

U.S. flag

An official website of the United States government

The .gov means it’s official. Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

The site is secure. The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

  • Publications
  • Account settings

Preview improvements coming to the PMC website in October 2024. Learn More or Try it out now .

  • Advanced Search
  • Journal List
  • Front Artif Intell

Logo of frontai

Cutting-edge communication and learning assistive technologies for disabled children: An artificial intelligence perspective

Katerina zdravkova.

1 Faculty of Computer Science and Engineering, Ss. Cyril and Methodius University in Skopje, Skopje, North Macedonia

Venera Krasniqi

Fisnik dalipi.

2 Department of Informatics, Faculty of Technology, Linnaeus University, Växjö, Sweden

Mexhid Ferati

In this study we provide an in-depth review and analysis of the impact of artificial intelligence (AI) components and solutions that support the development of cutting-edge assistive technologies for children with special needs. Various disabilities are addressed and the most recent assistive technologies that enhance communication and education of disabled children, as well as the AI technologies that have enabled their development, are presented. The paper summarizes with an AI perspective on future assistive technologies and ethical concerns arising from the use of such cutting-edge communication and learning technologies for children with disabilities.

Introduction

According to the WHO ( 2021 ), more than a billion people worldwide experience some form of disability including almost 240 million children whose well-being is endangered. As also highlighted in a report of UNICEF ( 2021 ), the ageing of the world population, new illnesses and an escalating trend of chronic diseases further increase the amount of disabled people. They just need a little help to live and work independently and with dignity.

Aiming to assist loved ones, and often themselves, people started designing low-tech devices many centuries ago (Robitaille, 2010 ). Since the early sixteenth century, people with mobility difficulties have been using the walking cane as a tool that assists them to walk and stand stably (Amato, 2004 ). Although wheeled seats and furniture have been used for transporting disabled people since sixth century BCE, mass production of wheelchairs started in 1933, when the paraplegic Everest and his friend Jennings designed a metal foldable wheelchair that used the X-brace design they patented as a “construction for collapsible invalids' wheelchairs” (Woods and Watson, 2004 ).

Glasses for partially sighted people were invented in the seventeenth century (Rosenthal, 1996 ; Lee, 2013 ). At the same time, ear trumpets, passive amplifiers that collect sound waves, and direct them into the ear channel, replaced the cupped hand, the popular method for hard of hearing used since Roman Emperor Hadrian's era (Valentinuzzi, 2020 ). These two sensory tools are the origins of modern assistive devices and technologies that support a more independent life for people with disabilities (Robitaille, 2010 ).

Blind people were unable to independently learn and study until the beginning of the nineteenth century. At that time, Louis Braille, who was blind since his early childhood, modified the military purpose language created by Charles Barbier de la Serre and invented the Braille alphabet, which is still among the most widely used reading media for the blind (Jiménez et al., 2009 ). In 1935, Agatha Christie's novel The Murder of Roger Ackroyd was recorded, initiating the transition of the printed toward the so called talking books (Philips, 2007 ). Electronic text-to-speech synthesizers were unveiled just a few years later, enabling blind people to consume written content by leveraging their enhanced hearing ability (Ohna, 2010 ).

The first electronic assistive technology ever developed was the Akouphone, a portable hearing device with carbon microphone and earphones (Kenefick, 2009 ). Although communication based on hand movements is mentioned in Plato's Cratylus, a Socratic' dialogue from his middle period as early as the fifth century BCE (Cratylus, 2022 ), the first sign language in wide use was the Old French Sign language, which originated in the eighteenth century and became the basis of American Sign language (Reagan, 2021 ).

All these developments and tools significantly improved the condition of people with disabilities in their daily communications, although their comprehensive education was still challenging. To ensure the fundamental human right to education and avoid the discrimination due to disability, in 2018 WHO launched the global cooperation on assistive technology (GATE) initiative (Boot et al., 2017 ). The only goal of GATE is “to improve access to high-quality affordable assistive products globally” (WHO, 2018 ).

Artificial Intelligence (AI) is the driving force of most assistive products, supporting people with different disabilities to keep and improve their education and everyday activities (Zdravkova, 2022 ). It enabled the creation of simulated environments that also include augmented and virtual reality. AI-supported tools improve visual tracking skills, help students with social disabilities and improve time-management skills.

The advantage of AI technologies over non-AI technologies used to date is the speed and precision they provide in analyzing and deciphering complex communication, expression, and visual behaviors. For example, applications that incorporate gesture-based text prediction in conjunction with AI are very useful for categorizing the most likely words from gestures and transforming them into meaningful sentences that support people with hearing impairments (Cheng and Lai, 2020 ). Other examples include machine and deep learning, which can improve substantially EEG-based brain-computer interfaces that help people who are unable to move gain independence (Sakti et al., 2021 ).

In this study, we present the findings of a narrative literature review that collected relevant literature published during the last ten years (Baumeister, 1997 ). The first step in this method pertains to conducting a preliminary search of the literature. For this reason, we constructed a search string to query the Scopus research database. In this endeavor, we have included papers only written in English that highlight evidence related to the impact of AI on modern assistive technologies for children with disabilities. As a next step, we analyzed and refined the explored topics from the relevant papers by subsequently identifying the employed AI technique to support and enhance the functionality of the assistive technologies. After a careful review, we classified the AI techniques into four different clusters, namely, augmentative and alternative communication (AAC), machine and deep learning (ML and DL), natural language processing (NLP), and Conversational AI.

The remainder of the paper is organized as follows: in the next section, a variety of disabilities will be explored and the cutting-edge assistive technologies that support communication and education of young children will be announced in line with AI technologies that enabled their creation. This section will be followed with a brief explanation of AI algorithms and techniques behind the announced assistive technologies including the futuristic ones. The paper will conclude with an AI perspective of future assistive technologies and ethical concerns that arise from the use of cutting-edge communication and learning assistive technologies intended for disabled children.

Cutting edge assistive technologies

Unhindered communication is the key prerequisite of quality education (Dhawan, 2020 ). If a student cannot listen to what a teacher presents and school mates talk about, or cannot see the visual content that supports the lectures and the assignments, then the effect of instructional behavior exhibited even by the most skilled teachers is reduced. Lack of attention, cognitive and intellectual preparedness to comprehend the school curriculum is an additional problem. Sometimes, perfect sight, hearing and intellectual abilities are obstructed by motor disabilities, which slow down or disable students' ability to write. Cutting-edge assistive technologies can mitigate many of the above-mentioned problems (Shinohara and Wobbrock, 2011 ).

According to American Speech-Language-Hearing Association (ASHA), communication disorders cover the impairments “in the ability to receive, send, process, and comprehend concepts or verbal, non-verbal and graphic symbol systems” (American Speech-Language-Hearing Association, 1993 ). They are classified into speech disorders, language disorders, hearing disorders, and central auditory processing disorders (American Speech-Language-Hearing Association, 1993 ).

The modern assistive technologies empowered by AI can significantly contribute to achieving important goals to support and provide new possibilities for children with disabilities. In Figure 1 we illustrate such goals, which include improved communication, inclusive education, enhanced accessibility, intellectual preparedness , and culminating with independent life .

An external file that holds a picture, illustration, etc.
Object name is frai-05-970430-g0001.jpg

AI enabled goals to support children with disabilities.

As mentioned earlier, the AI techniques were classified into four broad clusters: AI methods and algorithms that support AAC; ML and DL; NLP; and Conversational AI – speech and voice recognition, which are further divided into silent speech interface (SSI), speech recognition (SR), visual speech recognition, and voice recognition. These four clusters should not be mixed with the AI goals that according to Russel and Norvig ( http://aima.cs.berkeley.edu/ ) embrace: problem solving; reasoning; knowledge representation; planning; machine learning; communication, perceiving and acting. Deep learning and natural language processing are important AI subsections and tools, while conversational AI is an umbrella term of so called affective computing, which is an interdisciplinary field that relies on AI. AAC is not AI per se. Similarly to affective computing, it heavily relies on sophisticated AI. Although AI heterogeneous, the division was made according to the needs of assistive technologies, thus they are a symbiosis of AI methods and major trends in assistive technologies.

Table 1 lists on the first column the senses or the human ability being affected. For each sense then a disorder is listed (column two) followed by the appropriate assistive technologies (column three) designed to easier overcome the challenges associated with the disorder. In the table's last two columns, we list AI techniques and AI algorithms and methods respective of the listed assistive technologies. The main criteria when selecting the listed disorders was their impact on impaired communication and the availability of Scopus studies describing the appropriate assistive technologies. If multiple studies introduced the same assistive application, priority was given to the one that in more detail represents an implementation of the AI technologies, algorithms, and methods.

Assistive technologies aligned to impaired communication disorders.

The following AI algorithms and methods were used to develop the assistive technologies shown in the table: adaptive optimization based on artificial neural networks (AOANN), augmented reality (AR), Bayesian learning, Bidirectional Encoder Representations from Transformers (BERT), Bidirectional long short-term memory (BLSTM), Convolutional neural networks (CNN), Deep neural networks (DNN), Gausian Markov model, Haptic communication (HC), Hidden Markov model (HMM), Hybrid CNN and RNN, k-Nearest Neighbors, Multi-layer perceptron (MLP) classifier, Naïve Bayes, Pronunciation verification, Recurrent neural network (RNN), Spatiotemporal convolutional neural network (SCNN), Speaker recognition (SpR), Speech-to-Braille, Speech-to-text (STT), Support vector machines (SVM), TensorFlow, Text-to-speech (TTS) and Voice activity detection (VAD).

AI algorithms presented in the table were selected in three stages:

  • a) Determination of the most frequent communication disorders covered by American Speech-Language-Hearing Association (ASHA, https://www.asha.org ), mental disorders from World Health Organization ( https://www.who.int ), as well as learning difficulties listed in the UNESCO learning portal ( https://learningportal.iiep.unesco.org/ );
  • b) Selection of the relevant scholar articles that present AI-based assistive technologies created to support this communication and learning disorders;
  • c) Thorough examination and extraction of AI methods, algorithms and techniques and their clustering.

Next subsection introduces them briefly by presenting a short explanation of each of the implemented AI algorithms and the assistive technologies that were developed by implementing those algorithms. Then, AI-based assistive technologies from all four clusters are also introduced and illustrated.

AI algorithms and assistive technologies for communication and learning impaired students

The list of the algorithms in the previous section was alphabetically ordered. The introduction of AI algorithms within the subsection will start with artificial neural networks, which are the most widely used computational model in machine learning and the heart of deep learning. The list will continue with the learning algorithms, then it will introduce speech and text conversions and end with haptic technology.

Since the mid 1980s, artificial neural networks have become a very powerful model that enables excellent prediction and learning of many patterns that cannot be explicitly presented. According to Pai ( 2020 ), three types of neural networks contribute to deep learning: artificial neural networks (ANN), CNN, and RNN. ANNs are feedforward networks capable of learning nonlinear functions. Their limitations can be overcome by creating building blocks (CNN) or by adding a recurrent connection on the hidden state (RNN). CNNs are particularly valuable for image processing, while RNNs solve problems with time series, text and audio data (Pai, 2020 ). Combination of CNNs and RNNs, sometimes called hybrid CNN and RNN proved its efficiency for word prediction (Goulart et al., 2018 ). Spatiotemporal convolutional neural networks are a special type of CNNs capable of extracting spatial-temporal features. They are very efficient for sign language recognition (Li et al., 2022 ). More complex neural networks with many hidden layers that employ more sophisticated mathematical models belong to deep neural networks.

Bidirectional Encoder Representations from Transformers (BERT) is a language representation model that is trained on unlabeled data over different pre-training tasks using multi-layer bidirectional transformer encoders (Devlin et al., 2018 ). BERT is embedded into Google Search for over 70 languages. It has been successfully used in several assistive technologies, particularly for speech recognition (Brunner et al., 2020 ) and speech completion (Tsunematsu et al., 2020 ). Model performance of classification problems of sequential data can be improved by BLSTM. It is particularly useful for developing assistive technologies for people with visual impairment (Wahidin et al., 2018 ).

Artificial neural networks adaptive optimization is an effective classifier that was used to predict the disfluencies in speech signals of people with stuttered speech (Manjula et al., 2019 ). It adaptively optimizes network architecture using the artificial fish swarm optimization method, which implements stochastic search (Manjula et al., 2019 ).

k-Nearest Neighbors is non-parametric, supervised learning classifier, which is widely used for computing the distances from the test example to all stored examples. It was abundantly used for classification and regression in the voice recognition systems (Ali et al., 2021 ).

Support vector machines are supervised learning models used for pattern classification (Cortes and Vapnik, 1995 ). They were used for developing the smart voice conversational assistant (Lokitha et al., 2022 ).

The term Bayesian learning has been interchangeably used with Bayesian inference. It produces a probability distribution using the Bayes' theorem to predict the value of an unknown quantity (Neal, 2012 ). Naïve Bayes classifiers are a simple class of Bayesian networks capable of efficient classification and prediction. Bayesian learning and naïve Bayes have been successfully used to develop BridgeApp, an assistive mobile application that assists communication between people that are deaf and mute (Samonte et al., 2019 ).

Hidden Markov model is a statistical Markov model that supports modeling of an observable process using unobservable states. Gausian hidden Markov models expect that the observation probability distribution is Gausian (or normal). These two finite-state models establish correlations between adjacent symbols, domains, or events, which is crucial for speech recognition.

Speaker, speech and text recognition are crucial to enable smooth communication with and among people with more severe hearing and visual impairment. Speaker and speech recognition identify words spoken aloud and convert them into readable text presented with written or Braille alphabet (Benzeghiba et al., 2007 ). They implement almost all AI algorithms to develop various assistive technologies, such as AVA, Jaws or RogerVoice (Zdravkova and Krasniqi, 2021 ). Assessment of speech in TabbyTalks (Shahin et al., 2015 ) implements (VAD, Sohn et al., 1999 ).

Augmented reality (AR) is an umbrella term that embraces interactive experience by integrating 3D virtual objects into a 3D environment in real time (Azuma, 1997 ). AR has been widely used in various assistive technologies, including the popular live captioning system AVA ( https://www.ava.me/ ), which enables deaf and hard-of-hearing to read the spoken lectures.

People with dual sensory loss can rely on haptic communication (HC), that enables tactile communication and interaction via the sense of touch (Ozioko et al., 2018 ). Many assistive technologies use the open source software library TensorFlow library ( www.tensorflow.org ), which was developed by Google.

AI-based assistive technologies

In all the presented assistive technologies, AI has been abundantly used. Augmentative and alternative communication methods proved their significant role in at least half of them. Assistive technologies related to hearing or visual deficiency are developed using neural natural language processing algorithms, which are a symbiosis between natural language processing and deep learning. The predominance of neural and deep networks is also obvious, proving that new assistive technologies are machine learning powered.

AAC is the best assistive remedy for intelligibility, which is mildly or more severely disrupted by speech disorders, including: aphasia, apraxia of speech, articulation disorder, cluttering, dysarthria and stuttering (Kent, 2000 ). It can also be affected by cognitive problems, such as autism spectrum disorder (ASD), dyslexia and Down syndrome (Deb et al., 2022 ; Krasniqi et al., 2022 ), and by motor disabilities, for example cerebral palsy, multiple sclerosis, and Parkinson's disease (Stipancic and Tjaden, 2022 ). Assistive devices and technologies that are used to improve intelligibility are among the most prominent AAC devices and technologies (Norrie et al., 2021 ). Although the success of high-tech AAC is still limited, mainly due to “infrastructure, policy, and recruitment deficits”, their advancement is inevitable and they will soon “serve as mediator between teacher, aided communicator, and their assistive technology” (Norrie et al., 2021 ). It is expected that in the near future, AAC devices will be combined with non-invasive methods of access to the brain-computer interface, which will revolutionize communication (Luo et al., 2022 ). However, without powerful machine learning and neural imaging technology, this transformation will never become true.

Within most of the reviewed devices and technologies, machine learning is developed alongside deep learning. It is particularly powerful in the technologies related to speech and sound production, as well as with spoken and written language understanding (Jobanputra et al., 2019 ). Neural network brain-computer interfaces (NNBCI) have a potential to reduce disability by translating neural activity into control of assistive devices (Schwemmer et al., 2018 ).

Speech and voice recognition will never be possible without artificial intelligence (Singh et al., 2018 ). All modern deep learning techniques, including BERT (Brunner et al., 2020 ) contribute to better speech, voice and speech recognition (Amberkar et al., 2018 ). Non-invasive brain-computer interfaces emerge in this area, leading to much better performance compared to traditional systems that process auditory and visual information (Brumberg et al., 2018 ). United with deep learning methods and architectures, they “boost classification performance” of algorithms for computer vision and natural language processing (Singh et al., 2018 ).

Most of the tools introduced in this section have been developed to accommodate the needs of students starting from elementary education to the college level. On some occasions, they can be applicable even for elderly people. While the concept of AI technologies presented so far was mainly related to applications in the education sector, AI has also the potential to improve health and well-being of elderly people. Some forms of AI assistive technologies such as autonomous robots, AI-enabled health applications, voice-activated devices and intelligent homes could tackle the key aging related challenges.

Cutting-edge technologies should be carefully designed and should consider privacy and content. While children with disabilities are more familiar with the use of phones, which facilitate the design and implementation of AI technologies, more mature people need assistance for an independent life. To provide AI technologies to elderly people, designers and programmers of these technologies should implement some considerations tailored to their needs: ensure the participation of elderly people for development of AI technologies, cross-age data collection, investments in digital infrastructure, increased research to understand new uses of AI and how to avoid bias (WHO, 2022 ).

To conclude, cutting-edge technologies significantly improve communication and learning of people with disabilities, and particularly of young children who were born with technology and do not hesitate to use them without any effort and resistance. Next section will try to prove this claim by researching more thoroughly the four clusters announced in the introduction of this section.

AI technologies that support communication and learning assistive technologies

Augmentative and alternative communication.

Certain people with disabilities cannot use speech as their primary means of communication and need therefore to find an alternative way or specific techniques to express themselves. The idea of Augmented and Alternative Communication (AAC) is to use all the abilities that a person has, in order to compensate for the impairment of the verbal communication capacity (Chirvasiu and Simion-Blândă, 2018 ). In other words, the AAC system provides effective communication to maximize quality of life. There exist various types of AAC that can be chosen depending on the individual's skill and communication needs (Beukelman et al., 2013 ). The ACC systems are classified as unaided and aided. Unaided systems do not require a physical aid or tool (e.g., facial expressions, sign language). Aided systems, on the other hand use materials or tech devices which are categorized as:

  • – Low -tech devices (symbol boards, choice cards, communication books)
  • – High-tech devices (AAC apps on mobile devices, speech-generating devices or communication devices)

The rapid development of AI has recently opened up new ways to address more and more complex challenges, such as for instance, helping people with complex communication needs to overcome barriers (Delipetrev et al., 2022 ).

AI powerful tools have the capability of transforming AAC systems such as low-tech with words and symbols and high-tech with computers that employ a human voice for output (Beukelman et al., 2013 ). Below, AI technologies embedded in the AAC are presented with an analysis of how these tools are being developed and deployed to meet the diverse needs of users.

As part of the UNICEF's Innovation Fund Investments in Skills and Connectivity (OTTAA Project: AI Algorithms for Assistive Communications, 2022 ) a platform (OTTAA project) is developed, which is considered to be the first augmentative and alternative communication (AAC) platform that uses a combination of powerful AI algorithms (NLP and ML) and pictogram-based communication code to create sentences and communicate effectively. OTTAA platform allows speech-impaired people to communicate and express themselves using a simple three-tap interaction. Using appropriate pictogram-based communication and AI algorithms children with disabilities will have the opportunity to communicate better and faster (OTTAA Project: Accessible Communication for Children with Disabilities, 2022 ). In order to encourage more diversity and options for users, the algorithm is trained constantly by analyzing more than 1.8 million sentences previously created by other users. OTTAA is an open source platform that encourages people to participate in improving the source code, in this way it creates an environment where everyone feels that they are contributing.

Image recognition technologies are considered pivotal in inclusive education to make learning accessible and effective. GoVisual app is a program converting photos and videos into literacy and communication opportunities on an iPhone or iPad (GoVisual™, 2022 ). This innovative approach combines computer vision using the image recognition technology (collecting photographs and videos), the NLP tools to help in story creation, and finally the ML to help identify objects and shapes for ease of programming (Tintarev et al., 2016 ). Combination of these three techniques creates a potential of independence and self-determination for children with disabilities in their school environment.

HearMeOut app is an application which incorporates both gesture-based text prediction and pictogram-based augmentative and alternative communication using AI. The application uses Natural Language Conversation (Srivastava, 2021 ) enabling the disabled users to engage in conversation using Speech Recognition. It also uses a word level sign language dataset (Li et al., 2022 ) to categorize the most probable words from the gestures of the impaired person captured using a camera and then transforms them into the most meaningful and possible sentence using state-of-the-art algorithms. Another positive aspect to this approach is the security concerns; the application does not store the user data because of the continuous process of input and output without storage, which minimizes the chances of any data leakage.

“Fluent” is an AI Augmented Writing Tool which assists persons who stutter to identify words that they might struggle pronouncing and presents a set of alternative words which have similar meaning but are easier to pronounce. The overall landscape of AI-based writing tools is typically comprised of NLP based software systems (Ghai and Mueller, 2021 ) as in this app but in addition it uses AL (active learning, that is the subset of machine learning; Settles, 2011 ) to identify whether it can help learn the unique phonetic patterns that an individual might struggle pronouncing. The app does not intend to improve the stutter condition but helps camouflage it.

Machine and deep learning

The recent increase in computing capabilities has enabled ML algorithms to further enhance the functionalities of assistive technologies. The incorporation of ML into eye tracking technology can contribute to the development of smarter assistive systems for people with disabilities (Koester and Arthanat, 2018 ; Yaneva et al., 2020 ).

In a study conducted by Valliappan et al. ( 2020 ), ML is leveraged to demonstrate accurate smartphone-based eye tracking without any additional hardware. The study results highlight the utility of smartphone-based gaze for detecting reading comprehension difficulty and confirms findings from previous studies on oculomotor tasks. Another work was conducted by Deepika and Murugesan ( 2015 ) to facilitate the interactions between computers and people with special needs using eye tracking technology. The performance accuracy of the proposed system under good lightning conditions was 97%.

Additional research efforts are concentrated on people with motor disabilities for a hands-free computer interaction (Šumak et al., 2019 ), to measure the variation between fixations and saccades using K-means analysis (König and Buffalo, 2014 ), and for training purposes to control eye gaze via VR (Zhang and Hansen, 2019 ).

Research advances in machine and deep learning have also contributed to improved electroencephalogram (EEG) decoding and target identification accuracy. In this perspective, visual evoked potential (VEP) based brain-computer interface (BCI) systems are widely explored, mainly due to low user training rate (Waytowich et al., 2016 ). One research involving people with motor and speech disabilities to evaluate a new monitor for generating VEP for daily BCI applications is conducted by Maymandi et al. ( 2021 ). The target identification in this study was performed using DNN. Moreover, DNN have become a useful approach to improve classification performance of BCI systems using EEG signals (Kwak et al., 2017 ; Craik et al., 2019 ).

A framework for brain electrical activity-based VEP biometrics is proposed by Palaniappan and Mandic ( 2007 ). In this work, in order to improve the classification accuracy, authors utilized k-Nearest Neighbors (kNN), Elma Neural Network (ENN) classifiers and 10-fold cross validation classification.

Video accessibility is of paramount importance for blind and visually impaired individuals for education and other purposes. Computer vision applications show promising results on removing the accessibility barriers, especially toward helping blind people to better sense the visual world. The most recent applications of ML in computer vision are object detection and classification and extraction of relevant information from images or videos.

A machine learning based approach to video description by automating video text generation and scene segmentation is proposed by Yuksel et al. ( 2020 ). The quality of the video descriptions generated through this system compared to the human-only condition resulted in being rated higher by blind and visually impaired users. A multimodal comprehensive accessibility framework to generate accessible text and tactile graphics for visually impaired people is suggested by Cavazos Quero et al. ( 2021 ). The framework uses machine learning, i.e., image classification technique to classify various kinds of graphics and applies simplification methods to the graphics category. Recently, interactive machine learning (IML) is utilized to support interface design through workshops with disabled users (Katan et al., 2015 ). This work demonstrates IML's potential significance as a design tool, expediting the design process by allowing the swift mapping of participant observations into prototypes.

Natural language processing techniques

Screen and text magnifiers are very useful solutions for low vision people, enabling them to read and adjust text. They zoom in the whole screen or a selected section of the screen without any AI techniques utilized. On the other hand, screen readers, Braille displays, and speech recognition software assistive technologies for blind people are completely AI based (Choi et al., 2019 ).

Screen readers are a compulsory part of all the popular learning management systems: Blackboard Learn, Brightspace D2L, Canvas LMS and Moodle (Zdravkova and Krasniqi, 2021 ). JAWS is embedded in all of them with Chrome, NVDA, TalkBack, ORCA and VoiceOver as alternative text-to-speech plugins (Oh et al., 2021 ). As a standalone application or part of Web applications, screen readers support image and touchscreen accessibility (Oh et al., 2021 ). Apart from speech, they enable non-speech audio, vibration, tactile and force feedback (Oh et al., 2021 ).

Screen readers consist of two components: optical character recognition (OCR) that recognizes text, images and mathematical expressions; and text-to-speech (TTS), which delivers that content in the form of speech (Suzuki et al., 2004 ). They can be additionally enhanced by a machine translator enabling language localization (Suzuki et al., 2004 ).

OCR passes through several phases, three of which are AI-powered: image pre-processing that removes the potential distortions and transforms the image into light and dark areas; intelligent character recognition that compares scanned characters with the learned ones; and post-processing that corrects the errors (Chaudhuri et al., 2017 ). Past OCR algorithms for text recognition were realized with pattern recognition and ML techniques (Rao et al., 2016 ). Recent algorithms unite the following soft computing constituents: fuzzy sets, artificial neural networks, genetic algorithms, and rough sets (Rao et al., 2016 ).

Image recognition is enhanced by image captioning tools, which generate textual description of visually presented objects using template-based, retrieval-based and neural networks-based methods (Wang et al., 2020 ). Template-based method is a statistical modeling method that uses HMM, Maximum Entropy Markov Models, and Conditional Random Fields to recognize and machine learn the patterns (Wang et al., 2020 ). The retrieval-based method measures the visual similarity between a new image and an already interpreted image and generates a human-level sentence (Wang et al., 2020 ).

Recognition of mathematical expressions and formulas, including the handwritten ones consists of character recognition and structure recognition (Veres et al., 2019 ). Both recognition tasks depend on various ML methods, such as Bayesian inference, fuzzy logic, and neural networks (Veres et al., 2019 ).

At the end of the OCR part, information is stored as data or text waiting to be converted into voice. The conversion is done by TTS systems, which is a pure natural language processing task (Mache et al., 2015 ). It consists of text analysis, phonetic analysis, prosodic analysis, followed by speech synthesis (Adam, 2020 ). Deep neural networks are the most frequently used methods of modern TTS systems that successfully predict the acoustic feature parameters for speech synthesis (Adam, 2020 ).

Screen readers, which deliver content into speech or auditory signals, are not suitable for deafblind people. The best alternative are text to Braille translators, which “interpret letters and figures through a tactile display” (Gote et al., 2020 ). Refreshable Braille displays are fully supported by Blackboard Learn and partially supported by Brightspace D2L (Hsu, 2020 ). They use AI techniques and methods only during OCR phase (Gote et al., 2020 ). In contrast, AI is the key factor in the opposite direction: from Braille to text (Hsu, 2020 ). The system presented in this paper employs a convolutional neural network model for converting a line of Braille into text; a ratio character segmentation algorithm to enable image segmentation; and optical Braille recognition to convert Braille images into text (Hsu, 2020 ). AI impact for speech recognition will be in more detail explored in the next subsection.

Non-signers, i.e., people who are not familiar with sign language can communicate with deaf people who speak using the translators of sign language into text or speech (Truong et al., 2016 ). These translators predominantly use ML algorithms to find the correct sign, like convolutional and recurrent neural networks (Bendarkar et al., 2021 ) or deep learning (Bantupalli and Xie, 2018 ). A very promising human-machine interface (HMI) device are communication gloves, which have sensors that interpret the motions of sign languages into natural language combining virtual and augmented reality with AAC (Ozioko and Dahiya, 2022 ). Ozioko and Dahiya ( 2022 ) review many of them, for example, Robotic Alphabet (RALPH), CyberGlove, PneuGlove, 5DT Data Glove and Cyberglove, the last two achieving a recognition accuracy higher than 90%. Apart from purely mechanical interpretation of sign language, several research teams started interpreting facial expressions of people using sign language (Cardoso et al., 2020 ; Silva et al., 2020 ). A standard CNN and hybrid CNN+LST were successfully used to translate facial expressions in Brazilian Sign Language Libras (Silva et al., 2020 ). All these technologies abundantly use almost all the AI algorithms and methods, including NLP essentials, which are their driving force (Cardoso et al., 2020 ).

Text-to-speech and speech-to-text system preferences and extensions, for example Mercury Reader, Voice Typing and Co-Writer Universal are designed for different operating systems and are compatible with different browsers, including (Dawson et al., 2019 ). They are frequently used by gifted students who are frustrated due to their dyslexia (Dawson et al., 2019 ). Mobile applications like ReadandWrite and Speak It! Voice Dream Weaver and libraries, such as Bookshare, Audible are helpful to students with reading and writing disorders like dyslexia (Dawson et al., 2019 ). They benefit from word prediction too. Word prediction is completely AI powered and it implements various approaches. For example, assistive technology for children with cerebral palsy is based on hidden Markov models (Jordan et al., 2020 ), a successfully commercialized mobile on-device system that applies deep learning (Yu et al., 2018 ), whereas context-based word prediction is achieved with naïve Bayes that incorporates latent semantic analyses (Goulart et al., 2018 ).

Although the effect of AI-based conversational agents on people with disabilities or special needs is rather controversial, they are “widely used to support people services, decision-making and training in various domains” (Federici et al., 2020 ).

Voice and speech recognition

As in many other scenarios that involve people with disabilities, AI and various machine learning algorithms show promising results in challenges associated with voice and speech recognition, speech identification, and speech-to-text service applications.

One type of speech disability is the childhood apraxia of speech (CAS), which treatment involves direct therapy sessions with a speech language pathologist. Such sessions must happen during longer periods, which put high demand on time allocation of pathologists. Moreover, many children needing such one-on-one sessions live in rural areas and expenses associated with therapy sessions prevent many children from getting the required support early on Theodoros ( 2008 ) and Theodoros and Russell ( 2008 ).

Technology in general, and AI and ML in particular, help in enabling children with challenges to receive satisfactory treatment in their home, which makes it a time- and cost-effective solution. One such solution is shown in a study by Parnandi et al. ( 2013 ), where a child's progress is assessed through the therapist assigning speech exercises to the child, which then are analyzed using AI algorithms and an assessment is given back to the therapist. In a similar study, further details show how the AI automatically identifies three types of anomalous patterns that are associated with CAS: delays in sound production, incorrect pronunciation of phonemes, and inconsistent lexical stress (Shahin et al., 2015 ). Especially issues related to measuring the inconsistent lexical stress are addressed using deep neural network-based classification tools (McKechnie et al., 2021 ). Such a tool is beneficial for both diagnosis and treatment by using the Convolutional Neural Network (CNN) model to identify linguistic units that affect the speech intelligibility (Abderrazek et al., 2020 ) and voice recognition and production (Lee et al., 2021 ). The latter study also utilized techniques to acquire bio signals from muscle activity, brain activity, and articulatory activity in order to improve the accuracy.

Deep-learning algorithms are also used with people who stutter, which is a speech disorder that is manifested by an addition of involuntary pauses or repetition of sounds (Sheikh et al., 2021 ). Using a real-time application, the system records a person's voice, then it identifies and removes stammers by improving speech flow, and then finally produces a speech that is clean from stuttering. The speech flow is improved by implementing an amplitude threshold produced by the neural network model (Mishra et al., 2021 ).

One type of technology that helps with voice and speech disorders is the STT service. Such services help in maintaining a satisfactory conversation between people living with such disabilities, by capturing their voice and transcribing it into written text that can be read by another person (Seebun and Nagowah, 2020 ). A form of such disability is also the Functional Speech Disorder (FSD), which is the inability to correctly learn to pronounce specific sounds, such as “s, z, r, l, and th”. Study by Itagi et al. ( 2019 ) shows how Random Forest Classifier performs better than other algorithms, such as, Fuzzy Decision Tree and Logistic Regression, when detecting and correcting in real-time FSD cases. These services benefit from the Natural Language Processing (NLP) applied in the STT, which utilizes Google Speech API to convert spoken words into text (Seebun and Nagowah, 2020 ).

AI-based assistive technologies and remote education

COVID pandemic was a great challenge for all the students, particularly those who have some communication and learning disabilities. Urgent need to transform traditional in-class education to remote education was an inspiration for many AI researchers to start creating cutting-edge assistive technologies and support massive inclusiveness at all levels of education. AI-based assistive technologies played a significant role in supporting them to learn and study remotely (Zdravkova and Krasniqi, 2021 ; Zdravkova et al., 2022 ).

The most widely used operating systems: Windows, MacOS, Android, Linux and Ubuntu provide some or all accessibility features, among which: screen readers, personal assistants, switch controls and voice access and control (Zdravkova et al., 2022 ).

Learning management systems, such as: Blackboard Ally, Brightspace D2L, Canvas and Moodle have full or partial conformance with WCAG 2.1 (WCAG, 2018 ). WCAG 2.1, abbreviated from Web Content Accessibility Guidelines version 2.1, is a referenceable ISO technical standard in the form of guidelines and resources that ensures web and mobile accessibility. All LMSs have various embedded screen reader tools, JAWS being common for all, increasing their WCAG 2.1 compliance. Blackboard Ally enables speech recognition via screen reader Read Speaker, while Brightspace D2L uses Dragon Inspection. These two learning management systems provide the opportunity to present learning content with a Braille display (Zdravkova and Krasniqi, 2021 ).

Video-teleconferencing tools, including the most widely used Zoom, Google Meet, MS Teams, BigBlueButton and Blackboard Collaborate have many features supporting students with motor, vision and hearing impairment. They all incorporate screen reader JAWS, as well as different AI-based plugins (Zdravkova and Krasniqi, 2021 ).

Massive open online courses (MOOCs), for example Coursera, edX, MIT OpenCourseWare, and OpenLearning offer various accommodations for students with hearing impairments in the form of multilingual subtitles, and transmission of page text toward a Braille display device (Zdravkova and Krasniqi, 2021 ).

Socially responsible universities in the developed countries have abundantly used most of the assistive features intended for hearing and visually impaired students for decades (Zdravkova and Krasniqi, 2021 ).

AI perspective of future assistive technologies

Many children with speech, hearing, and cognitive challenges have limited communication and access to speech-activated gadgets. However, rapidly advancing AI research is opening the way for the creation of new tools to aid in the resolution of these communication issues.

AI has already shown that it has the ability to transform special education and improve results for students with impairments in a variety of ways. Children with ASD who have trouble understanding others' emotions have benefited from AI-driven applications and robots that assist them practice emotion identification and other social skills. Moreover, AI has influenced the creation of algorithms that can aid in the identification of ASD, specific learning difficulties (dyslexia, dysgraphia, and dyscalculia), and attention-deficit/hyperactivity disorder in students (ADHD). For students with disabilities, AI-enhanced therapies have included error analysis to inform instruction and tailored feedback in spelling and maths.

Despite these gains, gaps in AI research for children with impairments, such as AI for students with intellectual and developmental disabilities, remain to be persistent. Because many of these children have numerous disabilities and/or major health concerns, this is an especially vital area of future work. Children with intellectual and developmental disabilities who also have hearing loss or vision impairment, for example, have additional difficulties. Hearing loss and other health difficulties, such as heart issues, are common among Down syndrome students. AI allows for the integration of health data across multiple applications, hence improving the quality of life for these children by promoting independence. This constellation of solutions can aid in the management of student information and the communication of health information among instructors, physicians, and caregivers.

AI algorithms using big data struggle to deal with the individual uniqueness of disabled people (Wald, 2021 ). There are currently two major issues that prevent the AI use in clinical decision-making in such cases, i.e., a finite amount of labeled data to train the algorithms, and deep neural network models' black-box nature. We believe these issues may be solved in one of two ways. To begin with, self-supervised representation learning has lately been applied to the development of meaningful dense representation from small chunks of data. Furthermore, reinforcement learning paradigms can learn to optimize in any defined environment using an exploration-exploitation paradigm. Second, explainable artificial intelligence methods can be used to enhance the decision-making transparency and trust by creating meta-information that elucidate why and how a decision was reached, while also recommending the factors that influenced the decision the most. This will allow researchers to concentrate on precision in intervention studies and tailored treatment models, while AI algorithms handle the data collection and analysis process.

Moreover, BCI systems for vision impaired people that use steady state visual evoked potentials to stimulate brain electrical activity that enables communication with or without gaze shifting are already a reality. Electronic retinal implants have already restored sight of few patients with degenerative retinal diseases. Cochlear implants successfully provide a sense of sound to hard of hearing and even to deaf people. New brain implants enable people to formulate words and sentences by using their thoughts supporting simple communication. Few years ago, these achievements seemed to be science fiction. With the current pace of technological development, BCI and brain implants that enhance human senses will soon become a reality enabling better inclusivity of people with disabilities.

Nevertheless, every cutting-edge technology is a double-edged sword. To paraphrase Norberg Winner (Bynum, 2017 ), new technologies, particularly brain implants “may be used for the benefit of humanity”, but they “may also be used to destroy humanity.”

First challenge of cutting-edge communication devices is undoubtedly their rather high price. For example, JAWS and ZoomText and ZoomText Fusion, which are the most widely used accessibility magnifiers and screen readers, have an annual price ranging from 80 to 160 US$ (Zdravkova, 2022 ). Assistive tools for hearing impaired students AVA and RogerVoice are slightly cheaper, but still not affordable to many (Zdravkova, 2022 ). On the other hand, the price of DaVinci Pro HD OCR and Logan ProxTalker exceeds 3000 US$, making them available to few highly privileged students (Zdravkova, 2022 ). If assistive tools are selectively used, they will amplify economic inequality, i.e., the gap between rich and poor. In some wealthier societies, the gender and racial gap might also increase, sacrificing girls and minority groups.

Second challenge is their impact on patients' physical and mental health. Still insufficiently tested communication devices might worsen the state of already feeble hearing or vision sensory organs risking to cause incorrigible deafness or blindness (Shanmugam and Marimuthu, 2021 ). Such problems might be a result of various reliability problems. This challenge raises the question of liability (Zdravkova, 2019 ). Although promising, deep brain stimulation reflects the “invasive nature of the intervention” (Cagnan et al., 2019 ). Another problem related to deep brain stimulation is related to anatomical and pathophysiological differences of people who will undergo the intervention, which can result in inconsistent clinical outcomes (Cagnan et al., 2019 ).

Next challenge is privacy. Many communication devices are wirelessly connected to medical institutions, either as part of research studies or for health monitoring purposes. The increasing trend of cyber-security threats during COVID pandemic disrupted healthcare institutions worldwide (Muthuppalaniappan and Stevenson, 2021 ). They affected many hospitals, medical research groups, and healthcare workers, but also the World Health Organization and national authorities of many countries (Muthuppalaniappan and Stevenson, 2021 ). To avoid data leaks, very strict legal privacy frameworks should be created to significantly increase the level of data protection in public health (WHO, 2021 ).

Final challenge is related to the social acceptability of new technologies (Koelle et al., 2018 ). According to this research novel technologies and applications “might create new threats, raise new concerns and increase social tension between users and non-users” (Koelle et al., 2018 ). Many societies are technology skeptical and their first reaction to cutting-edge technologies is full resistance. If officially approved, there will be many disabled people who will be concerned with their impact making a vicious circle, which might worsen the situation instead of improving it.

In order to avoid all the challenges mentioned above, innovation in research should be very responsible. All potential ethical and legal challenges should be anticipated on time, and their remedies should be carefully included into new cutting-edge assistive technologies by design.

Author contributions

KZ, VK, FD, and MF: conceptualization, formal analysis, investigation, and writing–original draft preparation. KZ and FD: methodology, writing–review and editing, and project administration. KZ: supervision. All authors have read and agreed to the published version of the manuscript.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

  • Abderrazek S., Fredouille C., Ghio A., Lalain M., Meunier C. (2020). “Towards interpreting deep learning models to understand loss of speech intelligibility in speech disorders - step 1: cnn model-based phone classification,” in Interspeech 2020 (Shanghai: ), 2522–2526. 10.21437/Interspeech.2020-2239 [ CrossRef ] [ Google Scholar ]
  • Adam E. E. B. (2020). Deep learning based NLP techniques in text to speech synthesis for communication recognition . J. Soft Comput. 2 , 209–215. 10.36548/jscp.2020.4.002 [ CrossRef ] [ Google Scholar ]
  • Ali A. T., Abdullah H. S., Fadhil M. N. (2021). Voice recognition system using machine learning techniques . Mater. Today Proceed . 10.1016/j.matpr.2021.04.075 [ CrossRef ] [ Google Scholar ]
  • Amato J. (2004). On Foot: A History of Walking . New York, NY: NYU Press. [ Google Scholar ]
  • Amberkar A., Awasarmol P., Deshmukh G., Dave P. (2018). “Speech recognition using recurrent neural networks,” in 2018 International Conference on Current Trends Towards Converging Technologies (Coimbatore: IEEE; ), 1–4. [ Google Scholar ]
  • American Speech-Language-Hearing Association (1993). Definitions of Communication Disorders and Variations. Rockville, MD. 10.1044/policy.RP1993-00208 Available online at: www.asha.org/policy (accessed May 21, 2022). [ CrossRef ] [ Google Scholar ]
  • Azuma R. T. (1997). A survey of augmented reality . Pres. Tel. Virtual Environ. 6 , 355–385. 10.1162/pres.1997.6.4.355 [ CrossRef ] [ Google Scholar ]
  • Bantupalli K., Xie Y. (2018). “American sign language recognition using deep learning and computer vision,” in 2018 IEEE International Conference on Big Data (Seattle, WA: IEEE; ), 4896–4899. [ Google Scholar ]
  • Baumeister R. F. (1997). Writing narrative literature reviews . Rev. Gen. Psychol. 1 , 311–320. 10.1037/1089-2680.1.3.311 [ CrossRef ] [ Google Scholar ]
  • Bendarkar D., Somase P., Rebari P., Paturkar R., Khan A. (2021). Web Based Recognition and Translation of American Sign Language with CNN and RNN . Vienna: International Association of Online Engineering. [ Google Scholar ]
  • Benzeghiba M., Mori D. E., Deroo R., Dupont O., Erbes S., Jouvet T., et al.. (2007). Automatic speech recognition and speech variability: a review . Speech Commun , 49 , 763–786. 10.1016/j.specom.2007.02.006 [ CrossRef ] [ Google Scholar ]
  • Beukelman D., Mirenda P., Beukelman D. (2013). Augmentative and Alternative Communication . Towson, MD: Paul H. Brookes Pub. [ Google Scholar ]
  • Boot F. H., Dinsmore J., Khasnabis C., MacLachlan M. (2017). Intellectual disability and assistive technology: opening the GATE wider . Front. Pub. Health 5 , 10. 10.3389/fpubh.2017.00010 [ PMC free article ] [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • Brumberg J. S., Pitt K. M., Burnison J. D. (2018). A noninvasive brain-computer interface for real-time speech synthesis: the importance of multimodal feedback . IEEE Trans. Neural Syst. Rehab. Eng. 26 , 874–881. 10.1109/TNSRE.2018.2808425 [ PMC free article ] [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • Brunner A., Tu N. D. T., Weimer L., Jannidis F. (2020). “To BERT or not to BERT-comparing contextual embeddings in a deep learning architecture for the automatic recognition of four types of speech, thought and writing representation,” in SwissText/KONVENS (Zurich) . [ Google Scholar ]
  • Bynum T. W. (2017). Ethical Challenges to Citizens of ‘The Automatic Age': Norbert Wiener on the Information Society. Computer Ethics . New York, NY: Routledge. [ Google Scholar ]
  • Cagnan H., Denison T., McIntyre C., Brown P. (2019). Emerging technologies for improved deep brain stimulation . Nat Biotechnol. 37 , 1024–1033. 10.1038/s41587-019-0244-6 [ PMC free article ] [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • Cardoso M. E. D. A., Freitas F. D. A., Barbosa F. V., Lima C. A. D. M., Peres S. M., Hung P. C. (2020). “Automatic segmentation of grammatical facial expressions in sign language: towards an inclusive communication experience,” in Proceedings of the 53rd Hawaii International Conference on System Sciences (Hawaii: ). [ Google Scholar ]
  • Cavazos Quero L., Bartolomé J. I., Cho J. (2021). Accessible visual artworks for blind and visually impaired people: comparing a multimodal approach with tactile graphics . Electronics 10 , 297. 10.3390/electronics10030297 [ CrossRef ] [ Google Scholar ]
  • Chaudhuri A., Mandaviya K., Badelia P., Ghosh S. K. (2017). Optical Character Recognition Systems for Different Languages with Soft Computing . Cham: Springer, 9–41. [ Google Scholar ]
  • Cheng S. C., Lai C. L. (2020). Facilitating learning for students with special needs: a review of technology-supported special education studies . J. Comput. Educ. 7 , 131–153. 10.1007/s40692-019-00150-8 [ CrossRef ] [ Google Scholar ]
  • Chirvasiu N., Simion-Blândă E. (2018). Alternative and augmentative communication in support of persons with language development retardation . Rev. Roman. Pentru Educ. Multidimension. 10 , 28. 10.18662/rrem/43 [ CrossRef ] [ Google Scholar ]
  • Choi J., Jung S., Park D. G., Choo J., Elmqvist N. (2019). Visualizing for the non-visual: enabling the visually impaired to use visualization . Comput. Graph. Forum 38 , 249–260. 10.1111/cgf.13686 [ CrossRef ] [ Google Scholar ]
  • Cortes C., Vapnik V. (1995). Support-vector networks . Machine Learn. 20 , 273–297. 10.1007/BF00994018 [ CrossRef ] [ Google Scholar ]
  • Craik A., He Y., Contreras-Vidal J. L. (2019). Deep learning for electroencephalogram (EEG) classification tasks: a review . J. Neural Eng. 16 , 031001. 10.1088/1741-2552/ab0ab5 [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • Cratylus (2022). Available online at: https://www.gutenberg.org/files/1616/1616-h/1616-h.htm (accessed April 25, 2022).
  • Daly M. (2022). “ColorCode: A bayesian approach to augmentative and alternative communication with two buttons,” in Ninth Workshop on Speech and Language Processing for Assistive Technologies (Dublin: ), 17–23. [ Google Scholar ]
  • Dawson K., Antonenko P., Lane H., Zhu J. (2019). Assistive technologies to support students with dyslexia . Teach. Excep. Children 51 , 226–239. 10.1177/0040059918794027 [ CrossRef ] [ Google Scholar ]
  • Deb S. S., Roy M., Bachmann C., Bertelli M. O. (2022). Specific Learning Disorders, Motor Disorders, and Communication Disorders. Textbook of Psychiatry for Intellectual Disability and Autism Spectrum Disorder . Cham: Springer, 483–511. [ Google Scholar ]
  • Deepika S. S., Murugesan G. (2015). “A novel approach for Human Computer Interface based on eye movements for disabled people,” in 2015 IEEE International Conference on Electrical, Computer and Communication Technologies (Coimbatore: IEEE; ), 1–3. [ Google Scholar ]
  • Delipetrev B., Tsinaraki C., Kostic U. (2022). Historical Evolution of Artificial Intelligence. JRC Publications Repository . Available online at: https://publications.jrc.ec.europa.eu/repository/handle/JRC120469 (accessed May 28, 2022).
  • Devlin J., Chang M. W., Lee K., Toutanova K. (2018). Bert: Pre-training of deep bidirectional transformers for language understanding . arXiv [Preprint] . arXiv: 1810.04805. Available online at: https://arxiv.org/abs/1810.04805
  • Dhawan S. (2020). Online learning: a panacea in the time of COVID-19 crisis . J. Educ. Technol. Syst. 49 , 5–22. 10.1177/0047239520934018 [ CrossRef ] [ Google Scholar ]
  • Federici S., Filippis D. E., Mele M. L., Borsci M. L., Bracalenti S., Gaudino M., et al.. (2020). Inside pandora's box: a systematic review of the assessment of the perceived quality of chatbots for people with disabilities or special needs . Disability Rehab. Assis. Technol. 15 , 832–837. 10.1080/17483107.2020.1775313 [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • Ghai B., Mueller K. (2021). “Fluent: An AI Augmented Writing Tool for People who Stutter,” in 23rd International ACM SIGACCESS Conference On Computers And Accessibility . 10.1145/3441852.3471211 [ CrossRef ] [ Google Scholar ]
  • Gote A., Kulkarni T., Jha S., Gupta S. (2020). “A review of literature on braille tab and the underlying technology,” in 2020 5th International Conference on Devices, Circuits and Systems (Coimbatore: IEEE; ), 333–335. [ Google Scholar ]
  • Goulart H. X., Tosi M. D. L., Gonalves D. S., Maia R. F., Wachs-Lopes G. A. (2018). Hybrid model for word prediction using naive bayes and latent information . arXiv [Preprint]. arXiv:1803.00985. Available online at: http://hdl.handle.net/10993/52024 (accessed May 28, 2022).
  • GoVisual™ (2022). Attainmentcompany.com. Available online at: https://www.attainmentcompany.com/govisual (accessed May 20, 2022).
  • Hsu B. M. (2020). Braille recognition for reducing asymmetric communication between the blind and non-blind . Symmetry 12 , 1069. 10.3390/sym12071069 [ CrossRef ] [ Google Scholar ]
  • Itagi A., Baby C. J., Rout S., Bharath K. P., Karthik R., Rajesh Kumar M. (2019). Lisp Detection and Correction Based on Feature Extraction and Random Forest Classifier in Microelectronics, Electromagnetics and Telecommunications . Singapore: Springer, 55–64. [ Google Scholar ]
  • Jiménez J., Olea J., Torres J., Alonso I., Harder D., Fischer K., et al.. (2009). Biography of louis braille and invention of the Braille alphabet . Surv. Ophthalmol. 54 , 142–149. 10.1016/j.survophthal.2008.10.006 [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • Jobanputra C., Bavishi J., Doshi N. (2019). Human activity recognition: a survey . Proc. Comput. Sci. 155 , 698–703. 10.1016/j.procs.2019.08.100 [ CrossRef ] [ Google Scholar ]
  • Jordan M., Guilherme N. N. N., Alceu B., Percy N. (2020). Virtual keyboard with the prediction of words for children with cerebral palsy . Comput. Methods Prog. Biomed. 192 , 105402. 10.1016/j.cmpb.2020.105402 [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • Katan S., Grierson M., Fiebrink R. (2015). “Using interactive machine learning to support interface development through workshops with disabled people,” in Proceedings of the 33rd Annual ACM Conference on Human Factors In Computing Systems (Seoul: ACM; ), 251–254. [ Google Scholar ]
  • Kenefick J. A. (2009). Hearing aid innovation: 100+ years later . Volta Rev. 109 , 33. 10.17955/tvr.109.1.comm [ CrossRef ] [ Google Scholar ]
  • Kent R. D. (2000). Research on speech motor control and its disorders: a review and prospective . J. Commun. Disorders 33 , 391–428. 10.1016/S0021-9924(00)00023-X [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • Koelle M., Boll S., Olsson T., Williamson J., Profita H., Kane S., Mitchell R. (2018). “(Un acceptable!?! re-thinking the social acceptability of emerging technologies,” in Extended Abstracts of the 2018 CHI Conference on Human Factors in Computing Systems (Montreal, QC: ), 1–8. [ Google Scholar ]
  • Koester H. H., Arthanat S. (2018). Text entry rate of access interfaces used by people with physical disabilities: a systematic review . Assistive Technol. 30 , 151–163. 10.1080/10400435.2017.1291544 [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • König S. D., Buffalo E. A. (2014). A nonparametric method for detecting fixations and saccades using cluster analysis: removing the need for arbitrary thresholds . J. Neurosci. Methods 227 , 121–131. 10.1016/j.jneumeth.2014.01.032 [ PMC free article ] [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • Krasniqi V., Zdravkova K., Dalipi F. (2022). Impact of assistive technologies to inclusive education and independent life of down syndrome persons: a systematic literature review and research agenda . Sustainability 14 , 4630. 10.3390/su14084630 [ CrossRef ] [ Google Scholar ]
  • Kumar L. A., Renuka D. K., Rose S. L., Wartana I. M. (2022). Deep learning based assistive technology on audio visual speech recognition for hearing impaired . Int. J. Cognit. Comput. Eng. 3 , 24–30. 10.1016/j.ijcce.2022.01.003 [ CrossRef ] [ Google Scholar ]
  • Kwak N. S., Müller K. R., Lee S. W. (2017). A convolutional neural network for steady state visual evoked potential classification under ambulatory environment . PloS ONE 12 , e0172578. 10.1371/journal.pone.0172578 [ PMC free article ] [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • Lee K. (2013). Design transition of eyeglasses . J. Fashion Bus. 17 , 90–106. 10.12940/jfb.2013.17.4.90 [ CrossRef ] [ Google Scholar ]
  • Lee W., Seong J. J., Ozlu B., Shim B. S., Marakhimov A., Lee S., et al.. (2021). Biosignal sensors and deep learning-based speech recognition: a review . Sensors 21 , 1399. 10.3390/s21041399 [ PMC free article ] [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • Li D., Opazo C., Yu X., Li H. (2022). “Word-level deep sign language recognition from video: A new large-scale dataset and methods comparison,” in Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (Snowmass Village, CO: IEEE; ), 1459–1469. [ Google Scholar ]
  • Lokitha T., Iswarya R., Archana A., Kumar A., Sasikala S. (2022). “Smart voice assistance for speech disabled and paralyzed people,” in 2022 International Conference on Computer Communication and Informatics (Coimbatore: IEEE; ), 1–5. [ Google Scholar ]
  • Luo S., Rabbani Q., Crone N. E. (2022). Brain-computer interface: applications to speech decoding and synthesis to augment communication . Neurotherapeutics 19 , 263–273. 10.1007/s13311-022-01190-2 [ PMC free article ] [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • Mache S. R., Baheti M. R., Mahender C. N. (2015). Review on text-to-speech synthesizer . Int. J. Adv. Res. Comput. Commun. Engg. 4 , 54–59. 10.17148/IJARCCE.2015.4812 [ CrossRef ] [ Google Scholar ]
  • Manjula G., Shivakumar M., Geetha Y. V. (2019). “Adaptive optimization based neural network for classification of stuttered speech,” in Proceedings of the 3rd International Conference on Cryptography, Security and Privacy (Kuala Lumpur: ), 93–98. [ Google Scholar ]
  • Maymandi H., Perez Benitez J. L., Gallegos-Funes F., Perez Benitez J. A. (2021). A novel monitor for practical brain-computer interface applications based on visual evoked potential . Brain-Comput. Interfaces 8 , 1–13. 10.1080/2326263X.2021.1900032 [ CrossRef ] [ Google Scholar ]
  • McKechnie J., Shahin M., Ahmed B., McCabe P., Arciuli J., Ballard K. J. (2021). An automated lexical stress classification tool for assessing dysprosody in childhood apraxia of speech . Brain Sci. 11 , 1408. 10.3390/brainsci11111408 [ PMC free article ] [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • Mishra N., Gupta A., Vathana D. (2021). Optimization of stammering in speech recognition applications . Int. J. Speech Technol. 24 , 679–685. 10.1007/s10772-021-09828-w [ CrossRef ] [ Google Scholar ]
  • Mulfari D., Meoni G., Marini M., Fanucci L. (2021). Machine learning assistive application for users with speech disorders . App. Soft Comput. 103 , 107147. 10.1016/j.asoc.2021.107147 [ CrossRef ] [ Google Scholar ]
  • Muthuppalaniappan M., Stevenson K. (2021). Healthcare cyber-attacks and the COVID-19 pandemic: an urgent threat to global health . Int. J. Q. Health Care 33, mzaa117. 10.1093/intqhc/mzaa117 [ PMC free article ] [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • Neal R. M. (2012). Bayesian Learning for Neural Networks, Vol. 118. Berlin: Springer Science and Business Media. [ Google Scholar ]
  • Norrie C. S., Waller A., Hannah E. F. (2021). Establishing context: AAC device adoption and support in a special-education setting . ACM Trans. Comput. Hum. Int. (TOCHI) 28 , 1–30. 10.1145/3446205 [ CrossRef ] [ Google Scholar ]
  • Oh U., Joh H., Lee Y. (2021). Image accessibility for screen reader users: a systematic review and a road map . Electronics 10 , 953. 10.3390/electronics10080953 [ CrossRef ] [ Google Scholar ]
  • Ohna S. E. (2010). Open your eyes: deaf studies talking . Scand. J. Disability Res. 12 , 141–146. 10.1080/15017410902992775 [ CrossRef ] [ Google Scholar ]
  • OTTAA Project: Accessible Communication for Children with Disabilities (2022). UNICEF innovation fund. Unicefinnovationfund.org . Available online at: https://www.unicefinnovationfund.org/broadcast/updates/ottaa-project-accessible-communication-children-disabilities (accessed April 11, 2022).
  • OTTAA Project: AI Algorithms for Assistive Communications (2022). Unicef.org . Available online at: https://www.unicef.org/innovation/innovation-fund-ottaa-project (accessed April 4, 2022).
  • Ozioko O., Dahiya R. (2022). Smart tactile gloves for haptic interaction, communication, and rehabilitation . Adv. Intell. Syst. 4 , 2100091. 10.1002/aisy.202100091 [ CrossRef ] [ Google Scholar ]
  • Ozioko O., Hersh M., Dahiya R. (2018). “Inductance-based flexible pressure sensor for assistive gloves,” in 2018 IEEE SENSORS (New Delhi: IEEE; ), 1–4. [ Google Scholar ]
  • Pai A. (2020). CNN vs. RNN vs. ANN - Analyzing 3 Types of Neural Networks in Deep Learning. Analytics Vidhya, Feb, 17 . Available online at: https://www.analyticsvidhya.com/blog/2020/02/cnn-vs-rnn-vs-mlp-analyzing-3-types-of-neural-networks-in-deep-learning/ (accessed May 23, 2022).
  • Palaniappan R., Mandic D. P. (2007). Biometrics from brain electrical activity: a machine learning approach . IEEE Trans. Pattern Anal. Machine Int. 29 , 738–742. 10.1109/TPAMI.2007.1013 [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • Parnandi A., Karappa V., Son Y., Shahin M., McKechnie J., Ballard K., et al.. (2013). “Architecture of an automated therapy tool for childhood apraxia of speech,” in Proceedings of the 15th International ACM SIGACCESS Conference on Computers and Accessibility . ACM, Bellevue, Washington, DC, 1–8. [ Google Scholar ]
  • Philips D. (2007). Talking books: the encounter of literature and technology in the audio book . Convergence 13 , 293–306. 10.1177/1354856507079180 [ CrossRef ] [ Google Scholar ]
  • Rao N. V., Sastry A. S. C. S., Chakravarthy A. S. N., Kalyanchakravarthi P. (2016). Optical character recognition technique algorithms . J. Theor. Appl. Infm. Technol. 83. [ Google Scholar ]
  • Reagan T. (2021). Historical linguistics and the case for sign language families . Sign Lang. Stud. 21 , 427–454. 10.1353/sls.2021.0006 [ CrossRef ] [ Google Scholar ]
  • Ridha A. M., Shehieb W. (2021). “Assistive technology for hearing-impaired and deaf students utilizing augmented reality,” in 2021 IEEE Canadian Conference on Electrical and Computer Engineering (Montreal, QC: IEEE; ), 1–5. [ Google Scholar ]
  • Robitaille S. (2010). The Illustrated Guide to Assistive Technology and Devices: Tools and Gadgets for Living Independently: Easyread Super Large 18pt Edition . Available online at: ReadHowYouWant.com (accessed April 29, 2022).
  • Rosenthal J. W. (1996). Spectacles and Other Vision Aids: A History and Guide to Collecting . Novato, CA: Norman Publishing. [ Google Scholar ]
  • Sakti W., Anam K., Utomo S., Marhaenanto B., Nahela S. (2021). “Artificial intelligence IoT based EEG application using deep learning for movement classification,” in 2021 8th International Conference on Electrical Engineering, Computer Science and Informatics (Samarang: IEEE; ), 192–196. [ Google Scholar ]
  • Samonte M. J. C., Gazmin R. A., Soriano J. D. S., Valencia M. N. O. (2019). “BridgeApp: An assistive mobile communication application for the deaf and mute,” in 2019 International Conference on Information and Communication Technology Convergence (Jeju: IEEE; ), 1310–1315. [ Google Scholar ]
  • Schwemmer M. A., Skomrock N. D., Sederberg P. B., Ting J. E., Sharma G., Bockbrader M. A., et al.. (2018). Meeting brain–computer interface user performance expectations using a deep neural network decoding framework . Nat. Med. 24 , 1669–1676. 10.1038/s41591-018-0171-y [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • Seebun G. R., Nagowah L. (2020). “Let's talk: An assistive mobile technology for hearing and speech impaired persons,” in 2020 3rd International Conference on Emerging Trends in Electrical, Electronic and Communications Engineering (Balaclava: IEEE; ), 210–215. [ Google Scholar ]
  • Settles B. (2011). “From theories to queries: Active learning in practice,” in Conjunction with AISTATS 2010: Proceedings of Machine Learning Research. Available online at: https://proceedings.mlr.press/v16/settles11a.html (accessed June 02, 2022).
  • Shahin M., Ahmed B., Parnandi A., Karappa V., McKechnie J., Ballard K. J., et al.. (2015). Tabby talks: an automated tool for the assessment of childhood apraxia of speech . Speech Commun. 70 , 49–64. 10.1016/j.specom.2015.04.002 [ CrossRef ] [ Google Scholar ]
  • Shanmugam A. K., Marimuthu R. (2021). A critical analysis and review of assistive technology: advancements, laws, and impact on improving the rehabilitation of dysarthric patients . Handb. Dec. Supp. Syst. Neurol. Disorders 15 , 263–281. 10.1016/B978-0-12-822271-3.00001-3 [ CrossRef ] [ Google Scholar ]
  • Sheikh S. A., Sahidullah M., Hirsch F., Ouni S. (2021). Machine learning for stuttering identification: Review, challenges & future directions . arXiv [Preprint] . arXiv: 2107.04057. 10.48550/arXiv.2107.04057 [ CrossRef ] [ Google Scholar ]
  • Shinohara K., Wobbrock J. O. (2011). “In the shadow of misperception: assistive technology use and social interactions,” in Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (Vancouver, BC: ), 705–714. [ Google Scholar ]
  • Silva E. P. D., Costa P. D. P., Kumada K. M. O., Martino J. M. D., Florentino G. A. (2020). Recognition of Affective and Grammatical Facial Expressions: a Study for Brazilian Sign Language in European Conference on Computer Vision . Cham: Springer, 218–236. [ Google Scholar ]
  • Singh A. P., Nath R., Kumar S. (2018). “A survey: Speech recognition approaches and techniques,” in 2018 5th IEEE Uttar Pradesh Section International Conference on Electrical, Electronics and Computer Engineering (Gorakhpur: IEEE; ), 1–4. [ Google Scholar ]
  • Sohn J., Kim N. S., Sung W. (1999). A statistical model-based voice activity detection . IEEE Sign. Process Lett. 6 , 1–3. 10.1109/97.736233 [ CrossRef ] [ Google Scholar ]
  • Srivastava H. (2021). Using NLP techniques for enhancing augmentative and alternative communication applications . IJSRP. 11 , 243–245. 10.29322/IJSRP.11.02.2021.p11027 [ CrossRef ] [ Google Scholar ]
  • Stipancic K. L., Tjaden K. (2022). Minimally detectable change of speech intelligibility in speakers with multiple sclerosis and Parkinson's disease . J. Speech Lang. Hearing Res. 65 , 1858–1866. 10.1044/2022_JSLHR-21-00648 [ PMC free article ] [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • Šumak B., Špindler M., Debeljak M., Heričko M., Pušnik M. (2019). An empirical evaluation of a hands-free computer interaction for users with motor disabilities . J. Biomed. Inf. 96 , 103249. 10.1016/j.jbi.2019.103249 [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • Suzuki M., Kanahori T., Ohtake N., Yamaguchi K. (2004). “An integrated OCR software for mathematical documents and its output with accessibility,” in International Conference on Computers for Handicapped Persons . Berlin, Heidelberg, 648–655. [ Google Scholar ]
  • Theil A., Buchweitz L., Gay J., Lindell E., Guo L., Persson N. K., et al.. (2020). “Tactile board: a multimodal augmentative and alternative communication device for individuals with Deafblindness,” in 19th International Conference on Mobile and Ubiquitous Multimedia , 223–228. [ Google Scholar ]
  • Theodoros D., Russell T. (2008). Telerehabilitation: Current perspectives . Stud. Health Technol. Inform . 131 , 191–209. [ PubMed ] [ Google Scholar ]
  • Theodoros D., Russell T. (2008). Telerehabilitation: current perspectives . Stud. Health Technol. Inform . 131 , 191–209. [ PubMed ] [ Google Scholar ]
  • Tintarev N., Reiter E., Black R., Waller A., Reddington J. (2016). Personal storytelling: using natural language generation for children with complex communication needs, in the wild . Int. J. Hum. Comput. Stu. 93 , 1–16. 10.1016/j.ijhcs.2016.04.005 [ CrossRef ] [ Google Scholar ]
  • Truong V. N., Yang C. K., Tran Q. V. (2016). “A translator for American sign language to text and speech,” in 2016 IEEE 5th Global Conference on Consumer Electronics . IEEE, 1–2. [ Google Scholar ]
  • Tsunematsu K., Effendi J., Sakti S., Nakamura S. (2020). “Neural speech completion,” in INTERSPEECH 2020, 21st Annual Conference of the International Speech Communication Association (Shanghai: Interspeech; ), 2742–2746. 10.21437/Interspeech.2020-2110 [ CrossRef ] [ Google Scholar ]
  • UNICEF (2021). Nearly 240 million children with disabilities around the world, UNICEF's most comprehensive statistical analysis finds . Available online at: https://www.unicef.org/press-releases/nearly-240-million-children-disabilities-around-world-unicefs-most-comprehensive (accessed March 1, 2022).
  • Valentinuzzi M. E. (2020). Hearing Aid history: from ear trumpets to digital technology . IEEE Pulse 11 , 33–36. 10.1109/MPULS.2020.3023833 [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • Valliappan N., Dai N., Steinberg E., He J., Rogers K., Ramachandran V., Navalpakkam V. (2020). Accelerating eye movement research via accurate and affordable smartphone eye tracking . Nat. Commun. 11 , 1–12. 10.1038/s41467-020-18360-5 [ PMC free article ] [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • Veres O., Rishnyak I., Rishniak H. (2019). “Application of methods of machine learning for the recognition of mathematical expressions,” in COLINS 2019 , 378–389. [ Google Scholar ]
  • Wahidin H., Waycott J., Baker S. (2018). “The challenges in adopting assistive technologies in the workplace for people with visual impairment,” in Proceedings of the 30th Australian Conference on Computer-Human Interaction , 432–442. [ Google Scholar ]
  • Wald M. (2021). AI data-driven personalisation and disability inclusion . Front. Artif. Int. 117 , 571955. 10.3389/frai.2020.571955 [ PMC free article ] [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • Wang J., Wang W., Wang L., Wang Z., Feng D. D., Tan T., et al.. (2020). Learning visual relationship and context-aware attention for image captioning . Pattern Recog. 98 , 107075. 10.1016/j.patcog.2019.107075 [ CrossRef ] [ Google Scholar ]
  • Waytowich N. R., Yamani Y., Krusienski D. J. (2016). Optimization of checkerboard spatial frequencies for steady-state visual evoked potential brain–computer interfaces . IEEE Trans. Neural Syst. Rehab. Eng. 25 , 557–565. 10.1109/TNSRE.2016.2601013 [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • WCAG (2018). Web content accessibility guidelines (WCAG) overview . Available online at: https://www.w3.org/WAI/standards-guidelines/wcag/ (accessed June 17, 2022).
  • WHO (2018). Global Cooperation on Assistive Technology (GATE), Available online at: https://www.who.int/news-room/feature-stories/detail/global-cooperation-on-assistive-technology-(gate) (accessed August 25, 2022).
  • WHO (2021). Ensuring artificial intelligence (AI) technologies for health benefit older people, Available online at: https://www.who.int/news/item/09-02-2022-ensuring-artificial-intelligence-(ai)-technologies-for-health-benefit-older-people (accessed August 25, 2022).
  • WHO (2022). Disability and health, Available online at: https://www.who.int/news-room/fact-sheets/detail/disability-and-health (accessed August 11, 2022).
  • Woods B., Watson N. (2004). The social and technological history of wheelchairs . Int. J. Ther. Rehab. 11 , 407–410. 10.12968/ijtr.2004.11.9.19587 [ CrossRef ] [ Google Scholar ]
  • Yaneva V., Eraslan S., Yesilada Y., Mitkov R. (2020). Detecting high-functioning autism in adults using eye tracking and machine learning . IEEE Trans. Neural Syst. Rehab. Eng. 28 , 1254–1261. 10.1109/TNSRE.2020.2991675 [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • Yu S., Kulkarni N., Lee H., Kim J. (2018). “On-device neural language model based word prediction,” in Proceedings of the 27th International Conference on Computational Linguistics: System Demonstrations , 128–131. [ Google Scholar ]
  • Yuksel B. F., Kim S. J., Jin S. J., Lee J. J., Fazli P., Mathur U., Miele J. A. (2020). “Increasing video accessibility for visually impaired users with human-in-the-loop machine learning,” in Extended Abstracts of the 2020 CHI Conference on Human Factors in Computing Systems (Honolulu, HI: ), 1–9. [ Google Scholar ]
  • Zdravkova K. (2019). Reconsidering human dignity in the new era . New Ideas Psychol. 54 , 112–117. 10.1016/j.newideapsych.2018.12.004 [ CrossRef ] [ Google Scholar ]
  • Zdravkova K. (2022). The potential of Artificial Intelligence for Assistive Technology in Education, Handbook on Intelligent Techniques in the Educational Process, Vol 1 . Cham: Springer. [ Google Scholar ]
  • Zdravkova K., Dalipi F., Krasniqi V. (2022). Remote education trajectories for learners with special needs during COVID-19 outbreak: An accessibility analysis of the learning platforms . Int. J. Emerg. Technol. Learn . 17. 10.3991/ijet.v17i21.32401 [ CrossRef ] [ Google Scholar ]
  • Zdravkova K., Krasniqi V. (2021). “Inclusive Higher Education during the COVID-19 Pandemic,” in 2021 44th International Convention on Information, Communication and Electronic Technology (MIPRO) . IEEE , 833–836. [ Google Scholar ]
  • Zhang G., Hansen J. P. (2019). “Accessible control of telepresence robots based on eye tracking,” in Proceedings of the 11th ACM Symposium on Eye Tracking Research and Applications (Denver, CO: ACM; ), 1–3. [ Google Scholar ]

Swarthmore College - ITS Blog

Swarthmore College – ITS Blog

Illustration of woman sitting at desk with laptop in front of her. There are icons of a microphone, speaker, and film strip surrounding her.

Assistive Technology Tools: Speech-to-text

This is part of a series of Assistive Technology (AT) Tools posts about tools anyone can use on their devices without downloading anything additional.

Have you ever been stopped at a light but feel you need to respond to the text from your kid? Or maybe you have one hand amess with oil and herbs but you need to pull up the recipe to remember that last ingredient before placing the dish in the oven? Or perhaps you have a broken wrist and still need to respond to the 82 emails you received today?

Speech-to-text options on your mobile, tablet, and desktop/laptop devices can help you in these—and many other—situations.

Speech-to-text options allow you to type without touching the device.

Many of the more robust speech-to-text tools were originally built and inspired by folks who use speech as an input for all functions on their devices. We will not discuss that software here, but if that is something you need, please reach out to support.swarthmore.edu to inquire.

Dictation on desktops

  • From the Apple menu, 
  •  Select System Settings ,
  • then click Keyboard in the sidebar.
  • Go to Dictation on the right, 
  • then toggle it On . A microphone will now appear as a text entry option.

See: Dictate messages and documents on Mac – Apple Support

Windows logo key + h

See: Use voice typing to talk instead of type on your PC – Microsoft Suppor t

Dictation on mobile devices

Android: talk to text.

  • On your Android phone or tablet, install Gboard .
  • Open any app that you can type with, such as Gmail or Keep.
  • Tap an area where you can enter text. A microphone will now appear as a text entry option.

See: Type with your voice – Android – Gboard Help

iOS: Dictation Keyboard

  • Open Settings .
  • Select General ,
  • Then select Keyboard .
  • Scroll down.
  • Toggle Enable Dictation . A microphone will now appear as a text entry option.

See: Dictate text on iPhone – Apple Support

Do you have feedback?

  • are other tools you use for dictation;
  • are errors in the directions;
  • or there aren’t directions for your device or operating system (OS),

let us know! Please email [email protected] .

Acknowledgements

This series has been inspired by previous blog posts by Corrine Schoeb, including:

  • Live transcription now available in zoom
  • Built-in reading tools: Firefox, Safari
  • Online tools to help with eye fatigue
  • Tips for reducing eye strain
  • NVDA simplified
  • VoiceOver simplified

and by Swarthmore’s ITS Diversity, Equity, and Inclusion Committee’s recent presentation on disability inclusion and accessibility. Thanks to Mark Davis, Sean O’Donnell , Ashley Turner, and Doug Willen for the thought and collaboration you put into your presentations and that has inspired this post!

Share this:

IMAGES

  1. UPS Foundation Supports NJEDDA Assistive Technolgy for Special

    speech and language impairment assistive technology

  2. Augmentative Communication & Assistive Technology Clinic| Franciscan

    speech and language impairment assistive technology

  3. Assistive Technology for Deaf/Hard of Hearing Children Erin Turner

    speech and language impairment assistive technology

  4. 5 Life-Changing Assistive Technologies for the Visually Impaired

    speech and language impairment assistive technology

  5. Assistive Technology in Speech Therapy

    speech and language impairment assistive technology

  6. Assistive Technology (Speech Devices)

    speech and language impairment assistive technology

VIDEO

  1. A case Study of Speech and Language Impairment

  2. Every speech delay is not Autism it can be a delayed milestone or language impairment #autism #adhd

  3. Understanding Assistive Technology for blind people

  4. Speech-to-Speech (STS)

  5. Speech and Language Impairment (Instructional Considerations)

  6. Vision Assistance Technologies

COMMENTS

  1. Assistive Devices for People with Hearing, Voice, Speech, or Language

    Assistive listening devices (ALDs) help amplify the sounds you want to hear, especially where there's a lot of background noise. ALDs can be used with a hearing aid or cochlear implant to help a wearer hear certain sounds better. Augmentative and alternative communication (AAC) devices help people with communication disorders to express ...

  2. LibGuides: Speech Disorders: Common Assistive Technologies

    This guide provides resources about speech disorders. The Technology Related Assistance to Individuals with Disabilities Act of 1988 described an assistive technology device as "any item, piece of equipment, or product system, whether acquired commercially off the shelf, modified, or customized, that is used to increase, maintain, or improve functional capabilities of individuals with ...

  3. Augmentative and Alternative Communication (AAC)

    People of all ages can use AAC if they have trouble with speech or language skills. Augmentative means to add to someone's speech. Alternative means to be used instead of speech. Some people use AAC throughout their life. Others may use AAC only for a short time, like when they have surgery and can't talk. There are a lot of different types ...

  4. Hearing Assistive Technology

    The American Speech-Language-Hearing Association (ASHA) is the national professional, scientific, and credentialing association for 234,000 members, certificate holders, and affiliates who are audiologists; speech-language pathologists; speech, language, and hearing scientists; audiology and speech-language pathology assistants; and students.

  5. Augmentative and Alternative Communication (AAC)

    Augmentative and alternative communication (AAC) is an area of clinical practice that supplements or compensates for impairments in speech-language production and/or comprehension, including spoken and written modes of communication.AAC falls under the broader umbrella of assistive technology, or the use of any equipment, tool, or strategy to improve functional daily living in individuals with ...

  6. Speech and Language Impairment

    Speech and language impairment are basic categories that might be drawn in issues of communication involve hearing, speech, language, ... Assistive technology (AT) can also be very helpful to students, especially those whose physical conditions make communication difficult. Each student's IEP team will need to consider if the student would ...

  7. Top Assistive Technology for Speech Difficulties

    Top Assistive Technology for Speech Difficulties. These assistive apps and sites help students with speech difficulties or disabilities communicate. Many leverage the mobility and visual, touch-based interfaces of phones and tablets and use sound and text-to-speech to give kids a voice. There are also picks that help students express themselves ...

  8. Assistive Technology for Cognition

    The assistive technology literature describes a wide variety of aids, ranging from low-tech tools designed for single-task guidance to highly technical devices that compensate for cognitive impairments across environments and task domains. Table 1 [PDF] lists sample tools categorized by complexity and target task.

  9. Assistive technology

    Overview. Assistive technology is an umbrella term for assistive products and their related systems and services. Assistive products help maintain or improve an individual's functioning related to cognition, communication, hearing, mobility, self-care and vision, thus enabling their health, well-being, inclusion and participation.

  10. Assistive Technology Solutions for Employees with Speech Impairments

    AAC devices, also called speech-generating devices, are an example of a type of technology that can be used by individuals who have difficulty speaking. JAN has general information about AAC devices as well as information about AAC with telephone access. Ideally when AAC is being considered, a speech language pathologist with expertise in AAC ...

  11. What Assistive Technology for Speech and Language Disorders Are

    Assistive technology that is available. These are the main types of assistive technology available: Augmentative and alternative communication (AAC): These help people with speech and language impairments with language skills and communication.

  12. Advances in Specific Language Impairment Research and Intervention: An

    Under the leadership of Margaret Rogers, Chief Staff Officer for Science and Research at the American Speech-Language-Hearing Association (ASHA), there is an annual research forum offered at the time of the Annual Convention, funded by competitive grant support provided by the National Institute on Deafness and Other Communicative Disorders (NIDCD) and documented by follow-up publications ...

  13. A Systematic Review of Research on Augmentative and Alternative

    AMERICAN JOURNAL OF SPEECH-LANGUAGE PATHOLOGY (AJSLP) JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH (JSLHR) ... Spoken and written language relationships in language/learning-impaired and normally achieving school-age children. Journal of ... Assistive Technology, 14(7), 710-731.

  14. What is text-to-speech technology (TTS)?

    Text-to-speech (TTS) technology reads aloud digital text — the words on computers, smartphones, and tablets. TTS can help people who struggle with reading. There are TTS tools available for nearly every digital device. Text-to-speech (TTS) is a type of assistive technology that reads digital text aloud. It's sometimes called "read aloud ...

  15. Assistive Technology in Speech Therapy

    Assistive Technology in Speech Therapy. Feb 8 . For some patients, an alternate form of communication is best for them and will help them be more functional participants in their environments. ... A Systematic Review" by Ralf W. Schlosser and Oliver Wendt in American Journal of Speech-Language Pathology, August 2008, Vol. 17, 212-230. doi:10. ...

  16. Speech and Language processing as assistive technologies

    The purpose of this SIG is to promote the application of speech and language technologies to the field of assistive technologies, especially for people whose impairments may affect their abilities to communicate. Speech and Language Processing for Assistive Technologies is an inherently interdisciplinary endeavor that attracts researchers from ...

  17. Speech and language impairment

    Speech and language impairment are basic categories that might be drawn in issues of communication involve hearing, speech, language, and fluency. ... These devices are equipped with assistive technology features that enable the user to express themself, interact with peers, and be able to participate in all aspects of life. ...

  18. Assistive Listening Technology

    All HATS, large and small, frequency modulated (FM), infrared (IR), or induction loop systems (IL), are based on the same principle: they all bridge the distance between the sound source and the listener. They are all capable of considerably enhancing a hearing-impaired person's speech perception.

  19. Speech and Language Processing for Assistive Technologies (SLPAT)

    Welcome to the homepage of SIG-SLPAT! The purpose of this SIG is to promote the application of speech and language technologies to the field of assistive technologies, especially for people whose impairments may affect their abilities to communicate.. Speech and Language Processing for Assistive Technologies is an inherently interdisciplinary endeavor that attracts researchers from such ...

  20. Text-to-Speech Technology: What It Is and How It Works

    Understood. Text-to-speech (TTS) is a type of assistive technology that reads digital text aloud. It's sometimes called "read aloud" technology. TTS can take words on a computer or other digital device and convert them into audio. TTS is very helpful for kids who struggle with reading, but it can also help kids with writing and editing ...

  21. A scoping review on the use of speech-to-text technology for

    Several systematic reviews and meta-analyses have been conducted on studies of assistive technology for pupils with learning impairments [Citation 41-46], and reviews on assistive technology to support learners who struggle with reading and writing [Citation 47]. However, no previous literature reviews have focused on the use of STT among ...

  22. How technology is changing speech and language therapy

    Research being conducted by scientists at the universities of Cambridge, Edinburgh and Sheffield seeks to shrink the performance gap between machine and human, aiming to make speech technologies ...

  23. Cutting-edge communication and learning assistive technologies for

    Cutting edge assistive technologies. Unhindered communication is the key prerequisite of quality education (Dhawan, 2020).If a student cannot listen to what a teacher presents and school mates talk about, or cannot see the visual content that supports the lectures and the assignments, then the effect of instructional behavior exhibited even by the most skilled teachers is reduced.

  24. Assistive Technology Tools: Speech-to-text

    Speech-to-text options on your mobile, tablet, and desktop/laptop devices can help you in these—and many other—situations. Speech-to-text options allow you to type without touching the device. Many of the more robust speech-to-text tools were originally built and inspired by folks who use speech as an input for all functions on their devices.

  25. Ethopolitical media: Organizing Assistive Technology, disability and

    In this paper we prompt a re-reading of Assistive Technology (AT) as a media system that organizes disability in the framework of digital health-care and the platform society. ... Admon-Rick G (2014) Impaired encoding: Calculating, ordering, and the 'disability percentages' classification system. ... (2022) Reuniting speech-impaired people ...

  26. How AI is Revolutionizing Assistive Technology

    Speech-to-text or voice recognition technology allows students to write and revise without traditional pen and paper, significantly enhancing their learning experience. High-quality AI, incorporating Natural Language Processing (NLP) and machine learning algorithms, improves the accuracy of speech recognition and word predictability, thereby ...