scispace - formally typeset
F

Florian Eyben

Researcher at Technische Universität München

Publications -  158
Citations -  14521

Florian Eyben is an academic researcher from Technische Universität München. The author has contributed to research in topics: Recurrent neural network & Affective computing. The author has an hindex of 48, co-authored 147 publications receiving 11995 citations. Previous affiliations of Florian Eyben include Google.

Papers
More filters
Proceedings ArticleDOI

Opensmile: the munich versatile and fast open-source audio feature extractor

TL;DR: The openSMILE feature extraction toolkit is introduced, which unites feature extraction algorithms from the speech processing and the Music Information Retrieval communities and has a modular, component based architecture which makes extensions via plug-ins easy.
Proceedings ArticleDOI

Recent developments in openSMILE, the munich open-source multimedia feature extractor

TL;DR: OpenSMILE 2.0 as mentioned in this paper unifies feature extraction paradigms from speech, music, and general sound events with basic video features for multi-modal processing, allowing for time synchronization of parameters, on-line incremental processing as well as off-line and batch processing, and the extraction of statistical functionals (feature summaries).
Journal ArticleDOI

The Geneva Minimalistic Acoustic Parameter Set (GeMAPS) for Voice Research and Affective Computing

TL;DR: A basic standard acoustic parameter set for various areas of automatic voice analysis, such as paralinguistic or clinical speech analysis, is proposed and intended to provide a common baseline for evaluation of future research and eliminate differences caused by varying parameter sets or even different implementations of the same parameters.
Proceedings ArticleDOI

AVEC 2013: the continuous audio/visual emotion and depression recognition challenge

TL;DR: The third Audio-Visual Emotion recognition Challenge (AVEC 2013) has two goals logically organised as sub-challenges: the first is to predict the continuous values of the affective dimensions valence and arousal at each moment in time, and the second is to Predict the value of a single depression indicator for each recording in the dataset.