< Terug naar vorige pagina

Publicatie

In search of state and trait emotion markers in mobile-sensed language: Field study

Tijdschriftbijdrage - Tijdschriftartikel

BACKGROUND: Emotions and mood are important for overall well-being. Therefore, the search for continuous, effortless emotion prediction methods is an important field of study. Mobile sensing provides a promising tool and can capture one of the most telling signs of emotion: language. OBJECTIVE: The aim of this study is to examine the separate and combined predictive value of mobile-sensed language data sources for detecting both momentary emotional experience as well as global individual differences in emotional traits and depression. METHODS: In a 2-week experience sampling method study, we collected self-reported emotion ratings and voice recordings 10 times a day, continuous keyboard activity, and trait depression severity. We correlated state and trait emotions and depression and language, distinguishing between speech content (spoken words), speech form (voice acoustics), writing content (written words), and writing form (typing dynamics). We also investigated how well these features predicted state and trait emotions using cross-validation to select features and a hold-out set for validation. RESULTS: Overall, the reported emotions and mobile-sensed language demonstrated weak correlations. The most significant correlations were found between speech content and state emotions and between speech form and state emotions, ranging up to 0.25. Speech content provided the best predictions for state emotions. None of the trait emotion-language correlations remained significant after correction. Among the emotions studied, valence and happiness displayed the most significant correlations and the highest predictive performance. CONCLUSIONS: Although using mobile-sensed language as an emotion marker shows some promise, correlations and predictive R2 values are low.
Tijdschrift: JMIR Mental Health
ISSN: 2368-7959
Issue: 2
Volume: 9
Pagina's: 1 - 22
Jaar van publicatie:2022
Toegankelijkheid:Open