Increasing the Translational Efficacy of Intracortical Speech Brain-Computer Interfaces

Placeholder Show Content

Abstract/Contents

Abstract

Vocal aphasia, a condition characterized by impaired ability to vocalize speech, affects an estimated two million individuals (Ivanova, 2022). A recent study by Willett et al. from the Neuroprosthetics Translational Laboratory (NPTL) has introduced Brain-Computer Interfaces (BCIs) as a novel approach to mitigating the effects of aphasia. In this study, researchers collected intracortical EEG data from a patient and developed a recurrent neural network (RNN) to decode the neural signals into words, boasting a word error rate of just 18.8%.
Building upon this foundation, our paper presents the application of two deep state space machine learning models (dSSMs)—the S5 and Mamba—to improve the accuracy of this task. These model types have gained significant popularity in recent years and have outperformed traditional RNN models in a variety of long time series tasks, motivating their use in our study (Cirone, 2024). After considerable re-engineering efforts to fine-tune our models', our Mamba model achieved a word error rate of 16.7%. This achievement emphasizes the concrete benefits that BCIs could offer to individuals with speech impairments.
Moreover, an essential part of our research was the novel implementation of various machine learning techniques, from the NPTL's Speech-BCI model pipeline, in JAX, an open-source Python library that is particularly well-suited for the intricacies of dSSMs. This endeavor has resulted in the creation of multiple JAX-based functions and classes, which will facilitate the development of other dSSMs in the future. Moving forward, I hope to further enhance the accuracy and utility of our models and investigate the neural representations underlying speech in our models hopefully providing insight into how speech is processed in the brain.

Description

Type of resource text
Date modified May 14, 2024; May 23, 2024
Publication date May 14, 2024; May 3, 2024

Creators/Contributors

Author Kounga, Maxwell
Author Zoltowki, David
Author Dong Lee, Hyun
Author Linderman, Scott

Subjects

Subject Brain-computer interfaces
Subject Machine Learning
Genre Text
Genre Thesis

Bibliographic information

Related item
DOI https://doi.org/10.25740/dq040ft5931, https://doi.org/10.25740/dq040ft5931
Location https://purl.stanford.edu/dq040ft5931

Access conditions

Use and reproduction
User agrees that, where applicable, content will not be used to identify or to otherwise infringe the privacy or confidentiality rights of individuals. Content distributed via the Stanford Digital Repository may be subject to additional license and use restrictions applied by the depositor.
License
This work is licensed under a Creative Commons Attribution Non Commercial 4.0 International license (CC BY-NC).

Preferred citation

Preferred citation
Kounga, M. (2024). Increasing the Translational Efficacy of Intracortical Speech Brain-Computer Interfaces. Stanford Digital Repository. Available at https://purl.stanford.edu/dq040ft5931. https://doi.org/10.25740/dq040ft5931.

Collection

Undergraduate Theses, School of Engineering

View other items in this collection in SearchWorks

Contact information

Also listed in

Loading usage metrics...