Search papers, labs, and topics across Lattice.
University of Southern California
9
0
8
19
Smiling during traumatic recollection not only occurs in moments of distress but actively enhances emotional recovery and narrative coherence.
Training speaker diarization models solely on adult speech leads to surprisingly poor performance on children and older adults, but a simple multi-age training strategy can fix it.
Acoustic and phonetic NACs encode accent in fundamentally different ways, with implications for how we interpret and manipulate these representations.
You can reliably decode frustration from facial muscle activity, even when people aren't speaking aloud.
Wearable sensors and speech AI can now unobtrusively reveal the hidden communication dynamics driving hospital caregiver workload and stress.
Speech tokenizers, despite being crucial for multimodal LLMs, primarily capture phonetic information, creating a semantic mismatch with text-derived semantics that hinders performance.
Achieve accent-specific speech synthesis without any accented training data by cleverly combining phonological rules with multilingual TTS.
Control the accent of your TTS output without needing any accented training data, by transferring accent characteristics from other languages.
A new multimodal dataset links brain activity, muscle activation, and articulation in speech, opening doors to understanding the causal chain of speech production.