For parents Delbert and Sanaz Whetter a crying baby is a particularly difficult challenge. The Whetters are deaf, so when they’re in another room they rely on cameras and remote noise-monitors to help keep an eye on their two children, one of whom is an infant. But those technologies, while helpful, have limitations. “We have a child who is talkative. The alerts go off, but there is no way to distinguish between loud talking noises and crying,” Delbert said. But now the Whetters have a new tool — an innovative app developed by UCLA researchers — that will help.
Led by Ariana Anderson, assistant professor in residence of psychiatry and biobehavioral sciences in the David Geffen School of Medicine at UCLA, a team of scientists created Chatterbaby, which employs artificial intelligence to not only tell parents when their infant is crying, but also help determine why their baby is upset.
“I realized that the cries of my third baby were remarkably similar to the cries of the first two,” said Anderson, who is also a statistician in the UCLA Semel Institute for Neuroscience and Human Behavior. “As a mother, you instinctively know what your child is trying to tell you simply by listening to how they cry, even if you can’t see them. As a statistician, I thought, ‘Can we train an algorithm to do what my ears as a parent can do automatically?’ The answer was yes.”
In creating the app, Anderson and a team of researchers uploaded audio samples of more than 2,000 infant cries. Next, they built and tested algorithms that translated cries into three categories: pain, hunger and fussiness. The algorithms correctly flagged the type of cry, such as pain cries from babies receiving vaccinations or ear-piercings, more than 90 percent of the time.
“For the first time, we can confirm that our baby is crying, and then learn with a great deal of certainty what he’s crying about,” Delbert Whetter said. “That an app can do this is really amazing.”
The app, now available for free on iPhone and Android devices and at Chatterbaby.org, allows parents to record and upload their babies’ cries, which are then analyzed using artificial intelligence. Scientists hope that a new crying-pattern study launched through the app will provide insight into whether certain patterns can later be associated with specific infant development disorders, such as autism.
In addition to helping parents with hearing loss or who are deaf, like the Whetters, the smartphone app is also designed to help new parents, who may not yet understand what their baby is trying to communicate. The app may also help some women with postpartum depression because research shows they may have more difficulty than women without depression discerning the meaning of their babies’ cries.
But it’s the baby who may benefit the most from the technology. Studies show that faster rates of appropriate parental response to crying can facilitate language development in children. A device that helps deaf and hearing parents respond more quickly and accurately to vocal cues may boost language development.
“The program looks at the types of frequencies that are in the cry and at the different patterns of sounds and silence,” Anderson said. “For example, when you hear a cry that has a long period of silence in it, it’s more likely that the baby is fussy. But when babies are in pain, the cries typically have louder, longer bursts and there’s very little silence between sounds.”
Anderson hopes information in the database will help identify other patterns or associations related to infant development, including autism spectrum disorder. Deaf children have over twice the rate of autism compared to hearing children. Research has shown that babies at risk for autism show abnormal cry patterns even before they are diagnosed.