Search papers, labs, and topics across Lattice.
The paper introduces VoxCare, a wearable audio sensing system for capturing communication behaviors of hospital professionals in real-world settings. VoxCare performs on-device acoustic feature extraction and uses a speech foundation model-guided teacher-student framework to identify speech activity and derive behavioral measures like communication frequency, duration, and vocal arousal. Analysis of data collected with VoxCare reveals correlations between communication patterns and workload/stress levels across different shifts and working units.
Wearable sensors and speech AI can now unobtrusively reveal the hidden communication dynamics driving hospital caregiver workload and stress.
Healthcare professionals work in complex, high-stakes environments where effective communication is critical for care delivery, team coordination, and individual well-being. However, communication activity in everyday clinical settings remains challenging to measure and largely unexplored in human behavioral research. We present VoxCare, a scalable egocentric wearable audio sensing and computing system that captures natural communication behaviors of hospital professionals in real-world settings without storing raw audio. VoxCare performs real-time, on-device acoustic feature extraction and applies a speech foundation model-guided teacher-student framework to identify foreground speech activity. From these features, VoxCare derives interpretable behavioral measures of communication frequency, duration, and vocal arousal. Our analyses reveal how, when, and how often clinicians communicate across different shifts and working units, and suggest that communication activity reflects underlying workload and stress. By enabling continuous assessment of communication patterns in everyday contexts, this study provides data-driven approaches to understand the behaviors of healthcare providers and ultimately improve healthcare delivery.