We release the **E**xpressive **A**nechoic **R**ecordings of **S**peech (EARS) dataset.
If you use the dataset or any derivative of it, please cite our [Paper](https://arxiv.org/)
```
@inproceedings{richter2024ears,
title={EARS: An Anechoic Fullband Speech Dataset Benchmarked for Speech Enhancement and Dereverberation},
author={Richter, Julius and Wu, Yi-Chiao and Krenn, Steven and Welker, Simon and Lay, Bunlong and Watanabe, Shinjii and Richard, Alexander and Gerkmann, Timo},
booktitle={Interspeech},
year={2024}
}
```
## Highlights
* **100 h** of speech data from **107 speakers**
* high-quality recordings at **48 kHz** in an anechoic chamber
* **high speaker diversity** with speakers from different ethnicities and age range from 18 to 75 years
* **full dynamic range** of human speech, ranging from whispering to yelling
* 18 minutes of **freeform monologues** per speaker
* sentence reading in **7 different reading styles** (regular, loud, whisper, high pitch, low pitch, fast, slow)
* emotional reading and freeform tasks covering **22 different emotions** for each speaker