BU Researchers Develop Computational Tools to Safeguard Privacy Without Degrading Voice-Based Cognitive Markers

BU Researchers Develop Computational Tools to Safeguard Privacy Without Degrading Voice-Based Cognitive Markers

Digital voice recordings contain valuable information that can indicate an individual’s cognitive health, offering a non-invasive and efficient method for assessment. Research has demonstrated that digital voice measures can detect early signs of cognitive decline by analyzing features such as speech rate, articulation, pitch variation and pauses, which may signal cognitive impairment when deviating from normative patterns.

However, voice data introduces privacy challenges due to the personally identifiable information embedded in recordings, such as gender, accent and emotional state, as well as more subtle speech characteristics that can uniquely identify individuals. These risks are amplified when voice data is processed by automated systems, raising concerns about reidentification and potential misuse of data.

In a new study, researchers from Boston University Chobanian & Avedisian School of Medicine have introduced a computational framework that applies pitch-shifting, a sound recording technique that changes the pitch of a sound, either raising or lowering it, to protect speaker identity while preserving acoustic features essential for cognitive assessment.

“By leveraging techniques such as pitch-shifting as a means of voice obfuscation, we demonstrated the ability to mitigate privacy risks while preserving the diagnostic value of acoustic features,” explained corresponding author Vijaya B. Kolachalama, PhD, FAHA, associate professor of medicine.

Using data from the Framingham Heart Study (FHS) and DementiaBank Delaware (DBD), the researchers applied pitch-shifting at different levels and incorporated additional transformations, such as time-scale modifications and noise addition, to alter vocal characteristics to responses to neuropsychological tests. They then assessed speaker obfuscation via equal error rate and diagnostic utility through the classification accuracy of machine learning models distinguishing cognitive states: normal cognition (NC), mild cognitive impairment (MCI) and dementia (DE).

Using obfuscated speech files, the computational framework was able to accurately determine NC, MCI and DE differentiation in 62% of the FHS dataset and 63% of the DBD dataset.

According to the researchers, this work contributes to the ethical and practical integration of voice data in medical analyses, emphasizing the importance of protecting patient privacy while maintaining the integrity of cognitive health assessments. “These findings pave the way for developing standardized, privacy-centric guidelines for future applications of voice-based assessments in clinical and research settings,” adds Kolachalama, who also is an associate professor of computer science, affiliate faculty of Hariri Institute for Computing and a founding member of the Faculty of Computing & Data Sciences at Boston University.

No Comments Yet

Leave a Reply

Your email address will not be published.