| The RAVDESS is a validated multimodal database of emotional speech and song. The database is gender balanced consisting of 24 professional actors, vocalizing lexically-matched statements in a neutral North American accent. Speech includes calm, happy, sad, angry, fearful, surprise, and disgust expressions, and song contains calm, happy, sad, angry, and fearful emotions. Each _expression_ is produced at two levels of emotional intensity, with an additional neutral _expression_. All conditions are available in face-and-voice, face-only, and voice-only formats. The set of 7356 recordings were each rated 10 times on emotional validity, intensity, and genuineness. Ratings were provided by 247 individuals who were characteristic of untrained research participants from North America.
Livingstone, S. R., & Russo, F. A. (2018). The Ryerson Audio-Visual Database of Emotional Speech and Song (RAVDESS): A dynamic, multimodal set of facial and vocal expressions in North American English. PloS one, 13(5), e0196391.
The database has been widely used in psychological, neural, and computational studies of emotion. An acoustic analysis of the RAVDESS was recently reported here:
Major, Devon P., and Monita Chatterjee. "Acoustic analyses of the RAVDESS corpus of emotional stimuli." JASA Express Letters 6.2 (2026).
Best, Frank
________________________________ Frank A. Russo, PhD Professor (Full), Psychology, Faculty of Arts, Toronto Metropolitan University Professor (Status), Faculties of Music and Medicine, University of Toronto Adjunct Scientist, KITE, Toronto Rehabilitation Institute, University Health Network Scientific Director, SMART Lab Scientific Director, SingWell Project Chief Science Officer, LUCID Therapeutics Tel: 01-416-979-5000, x. 552647 (office), x. 554989 (lab) russo@xxxxxxxxxxxx
On Feb 26, 2026, at 04:13, Jochem Rieger <jochem.rieger@xxxxxxxxxxxxxxxx> wrote:
GAUDIE is an extensively validated German naturalistic auditory speech database with positive, neutral, and negative speech sequences available for non-profit academic research purposes. It comprises 37 audio speech sequences with a total duration of 92 minutes. The database is described in Lingelbach et al 2023
Lingelbach, K., Vukelić, M. & Rieger, J.W. GAUDIE: Development, validation, and exploration of a naturalistic German AUDItory Emotional database. Behav Res 56, 2049–2063 (2024). https://doi.org/10.3758/s13428-023-02135-z
Best,
Jochem
On 25.02.26 15:46, Morgan, Shae wrote:
ACHTUNG! Diese E-Mail kommt von Extern! WARNING! This email originated off-campus.
If the stimuli can have speech vs nonverbal or just emotional sounds, there are lots of databases available:
Toronto Emotional Speech Set (Dupuis & Pichora Fuller) Morgan Emotional Speech Set (Morgan, 2019) Ryerson Audio-Visual Database of Emotional Speech and Song (RAVDESS; Livingstone and Russo 2018)
They're all different in terms of which variables they code/don't code for. RAVDESS has song and speech, limited sentence variability (2 sentence frames), but many talkers and emotional expressions in high and low intensity. MESS has more semantic variability and target words (making it suitable for word and emotion recognition studies, masking, intelligibility, etc.), but fewer talkers (3 female 3 male) recorded at a single intensity for 4 emotion groups. TESS has few talkers, but includes age as a factor, a good number of emotional expressions.
Table 1 in Morgan & LaPaugh (2025) has a summary of commonly used emotional speech databases with citations for each - you can read up on the methods to see which database would suit your needs!
Morgan, S. D., & LaPaugh, B. (2025). Methodological Stimulus Considerations for Auditory Emotion Recognition Test Design. Journal of Speech, Language, and Hearing Research, 68(3), 1209-1224.
Cheers!
| Associate Professor Program Director, Audiology |
CAUTION: This email originated from outside of our organization. Do not click links, open attachments, or respond unless you recognize the sender's email address and know the contents are safe.
Best wishes Sophie
Prof Sophie Scott CBE Director, Institute of Cognitive Neuroscience, UCL, 17 Queen Square, London WC1N 3AZ 020 7679 1144 (office) 07881853586 (mobile)
⚠ Caution: External sender Hello all,
I am in need of a set of freely available emotions vocalisations for a study I am doing on detecting emotions in the human voice.
I am aware of the Montreal Affective Voices. I am also aware of the International Affective Digitized Sounds (IADS-E) database.
Is there any reason to choose one over the other or anybody have any other suggestions?
Thanks in advance,
Maria
Professor Maria Uther, C.Psychol., CSci., AFBPsS Honorary Professor University of Wolverhampton
--
Jochem Rieger
Prof. Dr. rer. nat. habil.
Applied Neurocognitive Psychology
DFG Center for Open and Reproducible Neuroscience Tools
COST Action INDoS CA24161
DFG RTG 2783 Neuromodulation of motor and cognitive function
Carl-von-Ossietzky University Oldenburg
Phone: +49 (0)441 798 4533
Web: https://uol.de/en/applied-neurocognitive-psychology
github: https://github.com/ANCPLabOldenburg
https://www.indos-costaction.eu/
|