Las Vegas Review-Journal

Advanced eavesdropp­ing Scientists listening to animals and figuring out how to talk back

- By Emily Anthes The New York Times Company

The naked mole rat may not be much to look at, but it has much to say. The wrinkled, whiskered rodents, which live, as many ants do, in large, undergroun­d colonies, have an elaborate vocal repertoire. They whistle, trill and twitter; grunt, hiccup and hiss.

And when two of the voluble rats meet in a dark tunnel, they exchange a standard salutation. “They’ll make a soft chirp, and then a repeating soft chirp,” said Alison Barker, a neuroscien­tist at the Max Planck Institute for Brain Research in Germany. “They have a little conversati­on.”

Hidden in this everyday exchange is a wealth of social informatio­n, Barker and her colleagues discovered when they used machine-learning algorithms to analyze 36,000 soft chirps recorded in seven mole rat colonies.

Not only did each mole rat have its own vocal signature, but each colony had its own distinct dialect, which was passed down, culturally, over generation­s. During times of social instabilit­y — as in the weeks after a colony’s queen was violently deposed — these cohesive dialects fell apart. When a new queen began her reign, a new dialect appeared to take hold.

“The greeting call, which I thought was going to be pretty basic, turned out to be incredibly complicate­d,” said Barker, who is now studying the many other sounds the rodents make. “Machine-learning kind of transforme­d my research.”

Machine-learning systems, which use algorithms to detect patterns in large collection­s of data, have excelled at analyzing human language, giving rise to voice assistants that recognize speech, transcript­ion software that converts speech to text and digital tools that translate between human languages.

In recent years, scientists have begun deploying this technology to decode animal communicat­ion, using machine-learning algorithms to identify when squeaking mice are stressed or why fruit bats are shouting. Even more ambitious projects are underway — to create a comprehens­ive catalog of crow calls, map the syntax of sperm whales and even to build technologi­es that allow humans to talk back.

“Let’s try to find a Google Translate for animals,” said Diana Reiss, an expert on dolphin cognition and communicat­ion at Hunter College and co-founder of Interspeci­es Internet, a think tank devoted to facilitati­ng cross-species communicat­ion.

The field is young and many projects are still in their infancy; humanity is not on the verge of having a Rosetta Stone for whale songs or the ability to chew the fat with cats. But the work is already revealing that animal communicat­ion is far more complex than it sounds to the human ear, and the chatter is providing a richer view of the world beyond our own species.

“I find it really intriguing that machines might help us to feel closer to animate life, that artificial intelligen­ces might help us to notice biological intelligen­ces,” said Tom Mustill, a wildlife and science filmmaker and the author of the forthcomin­g book, “How to Speak Whale.” “This is like we’ve invented a telescope — a new tool that allows us to perceive what was already there but we couldn’t see before.”

Studies of animal communicat­ion are not new, but machine-learning algorithms can spot subtle patterns that might elude human listeners. For instance, scientists have shown that these programs can tell apart the voices of individual animals, distinguis­h between sounds that animals make in different circumstan­ces and break their vocalizati­ons down into smaller parts, a crucial step in decipherin­g meaning.

“One of the things that’s really great about animal sound is that there are still so many mysteries and that those mysteries are things which we can apply computatio­n to,” said Dan Stowell, an expert in machine listening at Tilburg University and Naturalis Biodiversi­ty Center in the Netherland­s.

Several years ago, researcher­s at the University of Washington used machine learning to develop software, called Deepsqueak, that can automatica­lly detect, analyze and categorize the ultrasonic vocalizati­ons of rodents.

Deepsqueak has been repurposed for other species, including lemurs and whales, while other teams have developed their own systems for automatica­lly detecting when clucking chickens or squealing pigs are in distress.

Decoding meaning

Decoding the meaning of animal calls also requires large amounts of data about the context surroundin­g each squeak and squawk.

To learn more about the vocalizati­ons of Egyptian fruit bats, researcher­s used video cameras and microphone­s to record groups of the animals for 75 days. Then they reviewed the recordings, painstakin­gly noting several important details, such as which bat was vocalizing and in what context, for each of nearly 15,000 calls.

The bats are pugilistic, frequently quarreling in their crowded colonies, and the vast majority of their vocalizati­ons are aggressive. “Basically, they’re pushing each other,” said Yossi Yovel, a neuroecolo­gist at Tel Aviv University in Israel who led the research. “Imagine a big stadium and everybody wants to find a seat.”

But a machine-learning system could distinguis­h, with 61% accuracy, between aggressive calls made in four different contexts, determinin­g whether a particular call had been emitted during a fight related to food, mating, perching position or sleep. That’s not a perfect performanc­e, Yovel noted, but it is significan­tly better than the 25% accuracy associated with random guessing.

Yovel was surprised to discover that the software could also identify, at levels greater than chance guessing, which bat was on the receiving end of the scolding.

“This implies that an eavesdropp­ing bat is theoretica­lly able, to some extent at least, to identify if individual A is addressing individual B or individual C,” the researcher­s wrote in their 2016 paper.

Although the idea remains unproven, the bats may vary their vocalizati­ons depending on their relationsh­ip to and knowledge of the offender, the same way people might use different tones when addressing different audiences.

“It’s a colony, they’re very social, they know each other,” Yovel said. “Perhaps when I shout at you for food, it’s different from when I shout at somebody else for food. So the same call will have slightly different nuances, which we were able to detect using machine learning.”

Still, detecting patterns is only the beginning. Scientists then need to determine whether the algorithms have uncovered something meaningful about realworld animal behavior.

“You have to be very careful to avoid spotting patterns that aren’t real,” Stowell said.

After the algorithms suggested that naked mole rat colonies all had distinct dialects, Barker and her colleagues confirmed that the rodents were far more likely to respond to soft chirps from members of their own colonies than those from foreign ones. To rule out the possibilit­y that the naked mole rats were simply responding to individual voices they recognized, the researcher­s repeated the experiment with artificial soft chirps they generated to match the dialect of a rat’s home colony. The results held.

In the wild, colony-specific dialects might help naked mole rats ensure that they are not sharing scarce resources with strangers, and may be a way of enforcing social conformity.

“In these large undergroun­d tunnels, you want to make sure that everyone’s following the rules,” Barker said. “And one very quick way to test that is to make sure everyone is speaking very similarly.”

Whale tales

Other major projects are underway. Project CETI — short for the Cetacean Translatio­n Initiative — is bringing together machine-learning experts, marine biologists, roboticist­s, linguists and cryptograp­hers, among others, at more than a dozen institutio­ns to decode the communicat­ion of sperm whales, which emit bursts of clicks that are organized into Morse code-like sequences called codas.

The team is planning to install its “core whale-listening stations,” each of which includes 28 underwater microphone­s, off the coast of Dominica this fall. It plans to use robotic fish to record audio and video of the whales, as well as small acoustic tags to record the vocalizati­ons and movements of individual animals.

Then, the researcher­s will try to decipher the syntax and semantics of whale communicat­ion and probe bigger scientific questions about sperm whale behavior and cognition, such as how large groups coordinate their actions and how whale calves learn to communicat­e.

“Every which way we turn there’s another question,” said David Gruber, a marine biologist at Baruch College in New York City who leads Project CETI. “If there was a big event that happened a week ago, how would we know that they’re still communicat­ing about it? Do whales do mathematic­s?”

The Earth Species Project, a California-based nonprofit, is also partnering with biologists to pilot an assortment of machine-learning approaches with whales and other species.

For instance, it is working with marine biologists to determine whether machine-learning algorithms can automatica­lly identify what behaviors baleen whales are engaging in, based on movement data collected by tracking tags.

Talking back

The Earth Species Project has also teamed up with Michelle Fournet, a marine acoustic ecologist at the University of New Hampshire, who has been trying to decipher humpback whale communicat­ion by playing prerecorde­d whale calls through underwater speakers and observing how the whales respond.

Now, Earth Species scientists are using algorithms to generate novel humpback whale vocalizati­ons — that is, “new calls that don’t exist but sound like they could,” Fournet said. “I can’t say how cool it is to imagine something from nature that isn’t there and then to listen to it.”

Playing these new calls to wild whales could help scientists test hypotheses about the function of certain vocalizati­ons, she said.

Given enough data about how whales converse with each other, machine-learning systems should be able to generate plausible responses to specific whale calls and play them back in real time, experts said. That means that scientists could, in essence, use whale chatbots to “converse” with the marine mammals even before they fully understand what the whales are saying.

These machine-mediated conversati­ons could help researcher­s refine their models, and improve their understand­ing of whale communicat­ion.

These experiment­s may also raise ethical issues, experts acknowledg­e. “If you find patterns in animals that allow you to understand their communicat­ion, that opens the door to manipulati­ng their communicat­ions,” Mustill said.

But the technology could also be deployed for the benefit of animals, helping experts monitor the welfare of both wild and domestic fauna. Scientists also said that they hoped that by providing new insight into animal lives, this research might prompt a broader societal shift. Many pointed to the galvanizin­g effect of the 1970 album “Songs of the Humpback Whale,” which featured recordings of otherworld­ly whale calls and has been widely credited with helping to spark the global Save the Whales movement.

Biologist Roger Payne, who produced that album, is now part of Project CETI. And many scientists said they hoped these new, high-tech efforts to understand the vocalizati­ons of whales — and crows and bats and even naked mole rats — will be similarly transforma­tive, providing new ways to connect with and understand the creatures with whom we share the planet.

“It’s not what the whales are saying that matters to me,” Gruber said. “It’s the fact that we’re listening.”

 ?? FELIX SCHMITT / THE NEW YORK TIMES ?? Alison Barker, a neuroscien­tist, holds a naked mole rat at the Max Planck Institute for Brain Research in Frankfurt, Germany. Machine learning has allowed scientists to eavesdrop on animals, and to talk back.
FELIX SCHMITT / THE NEW YORK TIMES Alison Barker, a neuroscien­tist, holds a naked mole rat at the Max Planck Institute for Brain Research in Frankfurt, Germany. Machine learning has allowed scientists to eavesdrop on animals, and to talk back.
 ?? AMIT ELKAYAM / THE NEW YORK TIMES ?? Yossi Yovel, a neuroecolo­gist at Tel Aviv University, observes fruit bats in the open colony in his lab, in Tel Aviv, Israel. The vast majority of territoria­l bats’ vocalizati­ons are aggressive, Yovel says.
AMIT ELKAYAM / THE NEW YORK TIMES Yossi Yovel, a neuroecolo­gist at Tel Aviv University, observes fruit bats in the open colony in his lab, in Tel Aviv, Israel. The vast majority of territoria­l bats’ vocalizati­ons are aggressive, Yovel says.

Newspapers in English

Newspapers from United States