| eLife | |
| Low-dimensional learned feature spaces quantify individual and group differences in vocal repertoires | |
| John Pearson1  Jack Goffinet2  Richard Mooney3  Samuel Brudner3  | |
| [1] Center for Cognitive Neurobiology, Duke University, Durham, United States;Department of Neurobiology, Duke University, Durham, United States;Department of Biostatistics & Bioinformatics, Duke University, Durham, United States;Department of Electrical and Computer Engineering, Duke University, Durham, United States;Department of Computer Science, Duke University, Durham, United States;Center for Cognitive Neurobiology, Duke University, Durham, United States;Department of Neurobiology, Duke University, Durham, United States;Department of Neurobiology, Duke University, Durham, United States; | |
| 关键词: zebra finch; autoencoder; statistics; Mouse; Other; | |
| DOI : 10.7554/eLife.67855 | |
| 来源: eLife Sciences Publications, Ltd | |
PDF
|
|
【 摘 要 】
Increases in the scale and complexity of behavioral data pose an increasing challenge for data analysis. A common strategy involves replacing entire behaviors with small numbers of handpicked, domain-specific features, but this approach suffers from several crucial limitations. For example, handpicked features may miss important dimensions of variability, and correlations among them complicate statistical testing. Here, by contrast, we apply the variational autoencoder (VAE), an unsupervised learning method, to learn features directly from data and quantify the vocal behavior of two model species: the laboratory mouse and the zebra finch. The VAE converges on a parsimonious representation that outperforms handpicked features on a variety of common analysis tasks, enables the measurement of moment-by-moment vocal variability on the timescale of tens of milliseconds in the zebra finch, provides strong evidence that mouse ultrasonic vocalizations do not cluster as is commonly believed, and captures the similarity of tutor and pupil birdsong with qualitatively higher fidelity than previous approaches. In all, we demonstrate the utility of modern unsupervised learning approaches to the quantification of complex and high-dimensional vocal behavior.
【 授权许可】
CC BY
【 预 览 】
| Files | Size | Format | View |
|---|---|---|---|
| RO202106217608736ZK.pdf | 6368KB |
PDF