Tag: vocal learning

  • Dolphin Signature Whistles: The Evidence That Bottlenose Dolphins Have Names

    Within the first few months of life, every bottlenose dolphin develops a unique acoustic signal — a specific pattern of frequency modulations that no other dolphin in its community produces. This isn’t a generic call. It isn’t a species-wide sound. It’s an individually distinctive whistle that the dolphin will use, with minor variations, for the rest of its life. Other dolphins learn it, remember it, and — critically — copy it to get that specific individual’s attention. When researchers at the University of St Andrews played recordings of a dolphin’s own signature whistle through an underwater speaker, the dolphin called back. When they played the signature whistle of an unfamiliar dolphin, it didn’t respond. When they played the whistle of a known associate, it didn’t respond. The animal reacted specifically and exclusively to hearing its own “name” — as if someone had called it across a room.

    That 2013 study, published in PNAS by Stephanie King and Vincent Janik, was the first experimental demonstration that a nonhuman mammal uses learned vocal labels to address specific individuals. The implications were immediate and significant: dolphins don’t just have identity signals the way a dog has a distinctive bark. They have signals that function referentially — labels that other dolphins can produce to mean “you, specifically.” That’s not a contact call. That’s a name.

    How the system works

    Signature whistles were first described by Melba and David Caldwell in the 1960s. It took decades of fieldwork — particularly from the Sarasota Dolphin Research Program in Florida, which has tracked individual dolphins since 1970 — to establish how the system operates. An infant dolphin develops its signature whistle during the first few months of life through vocal learning. The calf doesn’t inherit a whistle genetically. It listens to the whistles in its environment and constructs its own, typically by copying a whistle it heard rarely and then modifying it into something unique. The result is an individually distinctive signal that encodes identity independently of voice features — the acoustic equivalent of a name written on a nametag rather than recognized by the sound of someone’s voice.

    This independence from voice cues is the detail that makes the naming analogy hold. Janik, Sayigh, and Wells demonstrated in a 2006 PNAS study that dolphins extract identity information from signature whistles even when all voice features have been removed from the recording. They synthesized whistles using computer-generated tones that preserved only the frequency contour — the shape of the whistle — and stripped everything that would tell the listener who was producing it. The dolphins still recognized the whistles. They responded preferentially to the synthetic versions of whistles belonging to individuals they knew. The contour alone carries the identity. That’s not how most animals recognize each other. Most species rely on voice cues — the timbre, the resonance, the characteristics of the individual’s vocal apparatus. Dolphins evolved a system where the pattern is the identity, not the voice. That’s structurally closer to how human names work than anything else documented in animal communication.

    Copying as addressing

    Dolphins don’t just produce their own signature whistles. They copy each other’s. King and colleagues showed in 2013 that copying occurs almost exclusively between animals with close social bonds — mothers and calves, allied males — and typically happens when the animals are separated and apparently trying to reunite. One pair of allied males was recorded copying each other’s whistles 12 years apart, preserving the fine acoustic details across more than a decade. Signature whistles make up roughly 50 percent of all whistles a dolphin produces, making them by far the most common sound in the repertoire.

    The copying is selective and precise but not exact. When a dolphin copies another’s whistle, it introduces minor but consistent modifications — subtle enough to preserve the referential content (whose whistle this is) while potentially marking it as a copy rather than the original. This is a nuance researchers are still working to understand. It’s possible the modifications function like quotation marks — a way of saying “I’m producing your name” rather than “I am you.” If that interpretation holds, it would mean dolphins are not just labeling individuals but doing so with a meta-communicative marker that distinguishes original production from quotation. That’s a level of communicative sophistication that, as of 2026, hasn’t been fully confirmed but also hasn’t been ruled out.

    Male bottlenose dolphins in Shark Bay, Australia, retain individual vocal labels even within multi-level alliance structures — coalitions of two to three males that cooperate to herd females, embedded within larger super-alliances of up to 14 males. King and colleagues published in Current Biology in 2018 that allied males maintain their individually distinctive signature whistles rather than converging on a shared group call, which is what you’d expect if the whistles served a group-identity function. The fact that they don’t converge — that each male keeps his own whistle even within a tightly bonded coalition — supports the interpretation that the whistles are individual labels, not team jerseys.

    Motherese

    In 2023, a study published in PNAS by Sayigh and colleagues from the Sarasota Dolphin Research Program demonstrated something that stopped a lot of people scrolling: dolphin mothers modify their signature whistles when their calves are present. The modifications — shifts to higher maximum frequencies — parallel the acoustic changes human parents make when speaking to infants, the phenomenon known as “motherese” or infant-directed speech. Human motherese involves higher pitch, wider pitch range, and exaggerated intonation. Dolphin motherese involves higher-frequency whistles with extended contours. Same function, different species, different medium.

    The finding matters because it suggests that the modification isn’t a side effect of arousal or environment — mothers don’t shift their whistles when other dolphins are present, only when their own calves are nearby. The adjustment is calf-directed. Whether it serves the same developmental function as human motherese — facilitating attention, bonding, and potentially vocal learning — remains an open question. But the structural parallel is hard to dismiss.

    Beyond signature whistles

    The most recent advance — a 2025 preprint from Sayigh, Janik, and the Sarasota team — moves past signature whistles entirely into territory that may prove even more significant. Having catalogued the signature whistles of most individuals in a community of 170 dolphins, the researchers are now documenting “non-signature whistles” — stereotyped whistle types that are not individually distinctive but are shared across multiple animals. They’ve identified 22 shared non-signature whistle types so far, two of which have been produced by at least 25 and 35 different dolphins respectively. If signature whistles are names, non-signature whistles may be something closer to words — shared acoustic signals with community-wide meaning rather than individual identity. Playback experiments filmed with drones are underway to determine what these shared whistles mean and how dolphins respond to them. The work was selected as a finalist for the Coller-Dolittle competition, which features non-invasive approaches to studying animal communication.

    Deep-learning classifiers are also being developed to automate signature whistle identification — a task that previously required expert human listeners to visually compare spectrograms. Jensen and colleagues published methods in 2024 for training neural networks to classify signature whistles from field recordings, which could turn the Sarasota whistle database into a passive population-monitoring tool. Hydrophone networks throughout Sarasota Bay could, in principle, track individual dolphins by their whistles the way cell towers track phones by their signals.

    The comparative picture

    Dolphins are no longer alone in the naming evidence. In 2024, a study published in Nature Ecology & Evolution demonstrated that African elephants address one another with individually specific name-like calls — not by copying, as dolphins do, but by producing arbitrary learned labels, which is structurally even closer to how human names work. A separate 2024 study in Science showed vocal labeling in marmoset primates. The evidence for animal naming has gone from a single-species curiosity to a cross-taxon pattern in two years.

    But dolphins remain the most extensively documented case, with 50 years of signature whistle research, a longitudinal dataset spanning decades of known individuals, and a level of experimental rigor — playback studies with synthetic whistles, controlled for voice cues, replicated across wild and captive populations — that the elephant and marmoset findings don’t yet match. The combination of vocal learning — the rare ability to hear a sound and reproduce it, shared by dolphins, parrots, songbirds, hummingbirds, bats, and humans but absent in most mammals — with the social complexity of fission-fusion groups, where individuals constantly separate and reunite, created the evolutionary pressure for a labeling system. When you can’t see your allies in murky water, you need a way to call them by something more specific than “hey.”

    The question the field is converging on isn’t whether dolphins have names. The evidence for that is now robust. The question is how much further the communication system extends beyond naming — whether the shared non-signature whistles represent a rudimentary vocabulary, whether the modifications during copying carry grammatical information, and whether the dolphin communication system has more structure than we’ve been able to decode. The Neurozoology course covers dolphin signature whistles alongside octopus distributed cognition, corvid tool use and funerary behavior, and electroreception in sharks and platypuses — the full catalog of neural capabilities that evolution produced outside the human lineage, most of which we didn’t know existed until someone thought to look.

  • Vocal Learning: Why Parrots and Songbirds Can Imitate Sounds and Most Animals Can’t

    A gorilla named Koko understood roughly 1,000 signs and could comprehend an estimated 2,000 English words. She never produced a single one of them vocally. Not one syllable. Meanwhile, a pet budgerigar — a parrot with a brain that weighs about 2 grams — can learn to produce over 400 human words, combine them into novel sentences, and match the pitch and rhythm of its owner’s voice closely enough that visitors mistake the bird for a person in the next room. A dog understands “sit” in English, Spanish, and Japanese, but can’t say any of them back. The dog’s problem isn’t intelligence. It’s hardware. Or more precisely, it’s wiring — the neural connections between the brain and the vocal organ that make imitation of heard sounds physically possible. That wiring exists in roughly nine groups of animals on Earth. Everything else is locked out.

    Vocal learning — the ability to hear a sound, form a memory of it, and then reproduce it using your vocal organ — is one of the rarest traits in the animal kingdom. Among the approximately 40,000 species of vertebrates, only three groups of birds (songbirds, parrots, and hummingbirds) and at least six groups of mammals (humans, cetaceans, bats, elephants, seals, and possibly mice at a rudimentary level) have it. Your dog, your cat, every primate except you, every reptile, every amphibian, and the vast majority of birds — pigeons, chickens, hawks, penguins — are vocal non-learners. They produce sounds. Some of those sounds are complex and serve specific functions. But they can’t hear something new and copy it. Their vocalizations are innate, hardwired, and essentially identical across every member of the species regardless of what they’ve heard. A chicken raised in total silence sounds like every other chicken. A zebra finch raised in total silence sounds like nothing — because its song isn’t preloaded. It has to learn.

    The brain pathway that most animals don’t have

    Erich Jarvis, a neurobiologist at Rockefeller University and Howard Hughes Medical Institute investigator, has spent three decades mapping the neural architecture of vocal learning across species. His work, and the work of dozens of collaborating labs, has converged on a finding that is as elegant as it is strange: vocal learning birds — songbirds, parrots, and hummingbirds — each have seven specialized forebrain vocal nuclei that are active when they produce learned vocalizations. These nuclei are organized into two pathways: a posterior pathway that handles vocal production and an anterior pathway that handles vocal learning and modification. Non-vocal-learning birds — chickens, pigeons, quail — have auditory processing regions that look similar, but they completely lack the seven specialized vocal nuclei. The hardware for hearing is universal. The hardware for imitating what you hear is not.

    Three of those seven nuclei sit in nearly identical brain locations across all three vocal learning bird groups, despite the fact that songbirds, parrots, and hummingbirds are only distantly related. Their last common ancestor lived around the time of the mass extinction that killed the dinosaurs, roughly 66 million years ago. The implication — supported by the 2014 consortium that sequenced 48 bird genomes — is that each group evolved vocal learning independently. Three separate lineages arrived at the same neural solution through convergent evolution. The brain didn’t inherit vocal learning from a shared ancestor. It reinvented it at least three times.

    Jarvis’s hypothesis for how this happened is what he calls brain pathway duplication. Every vertebrate has motor learning circuits — neural pathways that connect the cortex to the brainstem to control body movements like walking, reaching, and grasping. In vocal learners, these motor learning circuits appear to have been duplicated through some genetic mutation, and the duplicate copy got wired to the vocal organ instead of to the limbs. The new pathway gave the brain direct cortical control over the muscles of the larynx (in mammals) or syrinx (in birds) — the kind of fine motor control that allows you to shape a vowel or hit a pitch. Non-vocal-learners don’t have this direct connection. Their cortex can control their hands, their legs, their facial muscles. It just can’t reach the voice box. A chimpanzee can make a “raspberry” with its lips because it has voluntary control over lip muscles. It can produce some clicking sounds with its tongue. But it cannot voluntarily modulate the muscles of its larynx to produce imitated speech. The wires aren’t there.

    Why parrots and songbirds aren’t doing the same thing

    Songbirds and parrots both have the seven vocal nuclei. Both learn from hearing. Both require practice. But a 2023 Current Biology study from Zhilei Zhao and colleagues at Cornell revealed that the two groups use their neural pathways in fundamentally different ways — a finding that has significant implications for understanding human speech.

    In songbirds like the zebra finch, the anterior forebrain pathway handles learning during a critical juvenile period, and the posterior pathway handles production in adulthood. You can temporarily inactivate a zebra finch’s anterior pathway and its song stays intact — the bird keeps singing normally because the posterior pathway runs the show once the song is learned. It’s like removing the driving instructor from the car after the student has passed the test.

    Parrots are different. When Zhao’s team inactivated the anterior pathway in budgerigars, the birds could still vocalize, but their calls lost their individually unique acoustic signatures — the personal vocal “fingerprints” that budgerigars use to identify each other. The anterior pathway in parrots isn’t just for learning. It’s for producing individually distinctive vocalizations in real time. The driving instructor never leaves the car.

    This makes sense when you consider the behavioral ecology. A zebra finch learns one song during adolescence and sings essentially the same song for life — females prefer consistency, so the evolutionary pressure rewards a clean separation between the learning pathway and the production pathway. A budgerigar, by contrast, continuously learns new contact calls throughout its life, actively imitates the calls of flockmates and potential mates, and modifies its vocal output depending on social context. The parrot needs its learning circuitry online during production because it’s never done learning. Zhao’s conclusion: to learn continuously and vocalize flexibly like humans do, parrots evolved brain mechanisms distinct from those in songbirds. This distinction — open-ended versus closed-ended vocal learning — may be the key variable that separates species that merely learn a fixed repertoire from species that use vocal learning for flexible, lifelong communication. Humans and parrots sit on one side of that divide. Zebra finches sit on the other.

    FoxP2: the gene that connects bird song to human speech

    In 2001, researchers identified a point mutation in the gene FOXP2 as the cause of an inherited speech and language disorder in a British family known as “KE.” Affected members couldn’t produce fluent speech — they had severe difficulty sequencing the complex mouth movements required for words, a condition called developmental verbal dyspraxia. Their comprehension was less impaired than their production. The gene was a transcription factor, meaning it regulates the activity of hundreds of other genes downstream.

    The discovery sent researchers to the songbird brain. FoxP2 turned out to be heavily expressed in Area X, a striatal nucleus in the songbird forebrain that is essential for vocal learning — and it’s the same basal ganglia region that’s abnormal in humans with FOXP2 mutations. When zebra finches sing, FoxP2 protein levels in Area X drop. This decline isn’t a malfunction — it’s the mechanism. The decrease in FoxP2 triggers coordinated changes in the activity of thousands of other genes, functioning like a conductor signaling an orchestra. UCLA neurobiologist Stephanie White’s lab showed that when you use gene therapy techniques to prevent FoxP2 from declining during singing, the birds fail to learn their song. The molecular version of practice makes perfect requires FoxP2 to cycle between high and low levels as the bird practices. Lock it in place, and learning stalls.

    FoxP2 comes in a long and a short isoform, in both birds and humans. The long version regulates other genes. Disrupting the long version impairs learning. Disrupting the short version, surprisingly, doesn’t affect learning — but it changes the variability of the song, making renditions more stereotyped. The two isoforms appear to control different aspects of vocal output: one governs learning, the other governs flexibility. White’s lab identified entire suites of genes whose coordinated activity during the critical period correlates with song learning in juveniles — patterns that disappear as the bird ages and the critical period closes. Many of these same gene networks are active in human brain regions associated with speech development.

    A 2021 Nature Communications study went further, demonstrating that FoxP2 knockdown in adult songbirds disrupts the fluent initiation and termination of song and impairs syllable sequencing — closely paralleling the speech sequencing deficits seen in humans with FOXP2 mutations. The mechanism involves an imbalance in dopamine receptor expression across basal ganglia pathways, connecting vocal learning directly to the dopaminergic reward circuitry that drives motor skill refinement in mammals.

    The continuum problem

    The binary framing — vocal learner or not — is useful but increasingly inadequate. Jarvis and others have argued that vocal learning exists on a continuum rather than as an all-or-nothing trait. Mice produce ultrasonic vocalizations that show some features of learning, though far less robust than songbirds. Some non-human primates can modify the amplitude and timing of their calls based on social context, even if they can’t imitate new sounds. The variable isn’t whether an animal has vocal learning circuitry. It’s how much of it, and how strongly it’s connected.

    In advanced vocal learners — humans, parrots, songbirds — hundreds of neural projections connect the cortex directly to the brainstem motor neurons controlling the vocal organ. In non-learners like chickens, those direct connections don’t exist. In mice, a few weak connections exist. The difference between vocal learning and non-learning may be a matter of connection density rather than a categorical presence or absence of circuitry. Jarvis has described this as finding that vocal learning is “more continuous” than previously assumed — not a binary switch that flipped in a few lucky species, but a trait that exists in rudimentary form across many vertebrates and got amplified, through gene duplication and pathway expansion, in the lineages that needed it.

    The critical period adds another dimension. Zebra finches learn during a juvenile window and crystallize their song by adulthood. Canaries reopen their critical period seasonally, modifying their song each breeding season — and their FoxP2 expression in Area X fluctuates accordingly, with higher levels during periods of vocal instability. Humans learn language most efficiently before puberty but retain some vocal learning capacity throughout life. Budgerigars appear to learn continuously. The duration of the critical period, not just the presence of vocal learning circuits, determines how flexibly a species can deploy the ability.

    Why it matters beyond birds

    The reason the NIH funds songbird vocal learning research isn’t ornithological curiosity. It’s because songbirds are the best animal model for human speech development and its disorders. About 8 percent of American children have some form of speech or language disorder. Over 3 million Americans stutter. Childhood apraxia of speech — the condition caused by FOXP2 mutations — affects the sequencing of mouth movements required for fluent speech. The genetic pathways identified in songbird Area X, including FoxP2’s downstream gene networks, are conserved across species and represent potential pharmacological targets for speech therapy interventions that don’t currently exist.

    The convergent evolution story is what makes the research tractable. Humans and songbirds last shared a common ancestor roughly 300 million years ago — before dinosaurs, before mammals, before birds. The fact that both independently evolved the same neural architecture, using many of the same genes, for the same behavioral function means that the underlying biological logic of vocal learning is deeply constrained. There aren’t many ways to build a brain that imitates sounds. Evolution found the solution, and it found it repeatedly, across lineages separated by hundreds of millions of years, using the same molecular toolkit. The songbird isn’t a metaphor for human speech. It’s a parallel implementation of the same engineering problem, running on the same genetic software, arrived at independently because the problem only has a few solutions.

    We cover the neuroscience of vocal learning — from the FoxP2 pathway to the critical period to what birdsong reveals about human speech disorders — across our Neurozoology course, where the question isn’t just which animals can learn to talk, but what talking requires a brain to do.