AI helps US woman who had a stroke 18 years ago to speak again through an avatar

Ms Ann Johnson had a paralysing stroke in 2005 that took away her ability to speak. PHOTO: NYTIMES

SAN FRANCISCO - At Ms Ann Johnson’s wedding reception 20 years ago, her gift for speech was vividly evident. In an ebullient 15-minute toast, she joked that she had run down the aisle, wondered if the ceremony program should have said “flutist” or “flautist” and acknowledged that she was “hogging the mic”.

Just two years later, Ms Johnson - then a 30-year-old teacher, volleyball coach and mother of an infant - had a cataclysmic stroke that paralysed her and left her unable to talk.

On Wednesday, scientists reported a remarkable advance towards helping her, and other patients, speak again.

In a milestone of neuroscience and artificial intelligence, implanted electrodes decoded Ms Johnson’s brain signals as she silently tried to say sentences. Technology converted her brain signals into written and vocalised language and enabled an avatar on a computer screen to speak the words and display smiles, pursed lips and other expressions.

The research, published in the journal Nature, demonstrates the first time spoken words and facial expressions have been directly synthesised from brain signals, experts say. Ms Johnson chose the avatar, a face resembling hers, and researchers used her wedding toast to develop the avatar’s voice.

“We’re just trying to restore who people are,” said the team’s leader, Dr Edward Chang, the chair of neurological surgery at the University of California, San Francisco.

“It let me feel like I was a whole person again,” Ms Johnson, now 48, wrote me.

The goal is to help people who cannot speak because of strokes or conditions such as cerebral palsy and Lou Gehrig’s disease (or amyotrophic lateral sclerosis).

To work, Ms Johnson’s implant must be connected by cable from her head to a computer, but her team and others are developing wireless versions. Eventually, researchers hope, people who have lost speech may converse in real time through computerised pictures of themselves that convey tone, inflection and emotions such as joy and anger.

“What’s quite exciting is that just from the surface of the brain, the investigators were able to get out pretty good information about these different features of communication,” said Dr Parag Patil, a neurosurgeon and biomedical engineer at the University of Michigan, who was asked by Nature to review the study before publication.

Ms Johnson’s experience reflects the field’s fast-paced progress. Just two years ago, the same team published research in which a paralysed man, who went by the nickname Pancho, used a simpler implant and algorithm to produce 50 basic words such as “hello” and “hungry” that were displayed as text on a computer after he tried to say them.

Ms Johnson’s implant has nearly twice as many electrodes, increasing its ability to detect brain signals from speech-related sensory and motor processes linked to the mouth, lips, jaw, tongue and larynx. Researchers trained the sophisticated AI to recognsze not individual words, but phonemes, or sound units such as “ow” and “ah” that can ultimately form any word.

“It’s like an alphabet of speech sounds,” said Mr David Moses, the project manager.

While Pancho’s system produced 15 to 18 words per minute, Ms Johnson’s rate was 78 using a much larger vocabulary list. Typical conversational speech is about 160 words per minute.

When researchers began working with her, they didn’t expect to try the avatar or audio. But the promising results were “a huge green light to say, ‘OK, let’s try the harder stuff; let’s just go for it,’” Mr Moses said.

A letter board which Ms Ann Johnson sometimes uses to communicate at home. PHOTO: NYTIMES

They programmed an algorithm to decode brain activity into audio waveforms, producing vocalised speech, said Mr Kaylo Littlejohn, a graduate student at the University of California, Berkeley, and one of the study’s lead authors, along with Moses, Sean Metzger, Alex Silva and Margaret Seaton.

“Speech has a lot of information that is not well preserved by just text, like intonation, pitch, expression,” Mr Littlejohn said.

Working with a company that produces facial animation, researchers programmed the avatar with data on muscle movements. Ms Johnson then tried to make facial expressions for happy, sad and surprised, each at high, medium and low intensity. She also tried to make various jaw, tongue and lip movements. Her decoded brain signals were conveyed on the avatar’s face.

Through the avatar, she said, “I think you are wonderful,” and, “What do you think of my artificial voice?”

“Hearing a voice similar to your own is emotional,” Ms Johnson told the researchers.

She and her husband, Mr William Johnson, a postal worker, even engaged in conversation. She said through the avatar: “Do not make me laugh.” He asked how she was feeling about the Toronto Blue Jays’ chances. “Anything is possible,” she replied.

Mr William Johnson feeds his wife Ann Johnson, who had a paralyzing stroke in 2005 that took away her ability to speak, at home. PHOTO: NYTIMES

The field is moving so quickly that experts believe federally approved wireless versions might be available within the next decade. Different methods might be optimal for certain patients.

On Wednesday, Nature also published another team’s study involving electrodes implanted deeper in the brain, detecting activity of individual neurons, said Dr Jaimie Henderson, a professor of neurosurgery at Stanford University and the team’s leader, who was motivated by his childhood experience of watching his father lose speech after an accident. He said their method might be more precise but less stable because specific neurons’ firing patterns can shift.

Their system decoded sentences at 62 words per minute that the participant, Pat Bennett, 68, who has ALS, tried to say from a large vocabulary. That study didn’t include an avatar or sound decoding.

Both studies used predictive language models to help guess words in sentences. The systems don’t just match words but are “figuring out new language patterns” as they improve their recognition of participants’ neural activity, said Ms Melanie Fried-Oken, an expert in speech-language assistive technology at Oregon Health & Science University, who consulted on the Stanford study.

Neither approach was completely accurate. When using large vocabulary sets, they incorrectly decoded individual words about one-quarter of the time.

For example, when Ms Johnson tried to say, “Maybe we lost them,” the system decoded, “Maybe we that name.” But in nearly half of her sentences, it correctly deciphered every word.

Researchers found that people on a crowdsourcing platform could correctly interpret the avatar’s facial expressions most of the time. Interpreting what the voice said was harder, so the team is developing a prediction algorithm to improve that. “Our speaking avatar is just at the starting point,” Dr Chang said.

Ms Ann Johnson looks at the album from her wedding, which occurred just two years before her stroke. PHOTO: NYTIMES

Ms Johnson contacted Dr Chang in 2021, the day after her husband showed her my article about Pancho, the paralyzed man the researchers had helped. Chang said he initially discouraged her because she lived in Saskatchewan, Canada, far from his lab in San Francisco, but “she was persistent”.

Mr William Johnson, 48, arranged to work part time. “Ann’s always supported me to do what I’ve wanted,” including leading his postal union local, he said. “So I just thought it was important to be able to support her in this.”

She started participating last September. Traveling to California takes them three days in a van packed with equipment, including a lift to transfer her between wheelchair and bed. They rent an apartment there, where researchers conduct their experiments to make it easier for her. The Johnsons, who raise money online and in their community to pay for travel and rent for the multiyear study, spend weeks in California, returning home between research phases.

Ms Johnson communicated with me in emails composed with the more rudimentary assistive system she uses at home. She wears eyeglasses affixed with a reflective dot that she aims at letters and words on a computer screen.

It’s slow, allowing her to generate only 14 words per minute. But it’s faster than the only other way she can communicate at home: using a plastic letter board, a method Mr Johnson described as “her just trying to show me which letter she’s trying to try to look at and then me trying to figure out what she’s trying to say.”

Ms Johnson was teaching high school math, health and physical education, and coaching volleyball and basketball when she had her brainstem stroke while warming up to play volleyball. After a year in a hospital and a rehabilitation facility, she came home to her 10-year-old stepson and her 23-month-old daughter, who has now grown up without any memory of hearing her mother speak, Mr Johnson said.

“Not being able to hug and kiss my children hurt so bad, but it was my reality,” Ms Johnson wrote. “The real nail in the coffin was being told I couldn’t have more children.”

Caregiver Cheryl Ruddell and Ms Ann Johnson at the latter’s home in Regina, Saskatchewan, Canada. PHOTO: NYTIMES

For five years after the stroke, she was terrified. “I thought I would die at any moment,” she wrote, adding, “The part of my brain that wasn’t frozen knew I needed help, but how would I communicate?”

Gradually, her doggedness resurfaced. Initially, “my face muscles didn’t work at all,” she wrote, but after about five years, she could smile at will.

At first when she started making emotional expressions with the avatar, “I felt silly, but I like feeling like I have an expressive face again,” she wrote, adding that the exercises also enabled her to move the left side of her forehead for the first time.

She has gained something else, too. After the stroke, “it hurt so bad when I lost everything,” she wrote. “I told myself that I was never again going to put myself in line for that disappointment again.”

Now “I feel like I have a job again,” she wrote.

Besides, the technology makes her imagine being in “Star Wars”: “I have kind of gotten used to having my mind blown.” NYTIMES

Join ST's Telegram channel and get the latest breaking news delivered to you.