Improving brain-to-text models utilizing recurrent neural networks and transfer learning for neuroprosthetic speech

dc.catalogadorpva
dc.contributor.advisorRodríguez Fernández, María
dc.contributor.authorValle Araya, Carlos
dc.contributor.otherPontificia Universidad Católica de Chile. Instituto de Ingeniería Biológica y Médica
dc.date2024-05-02
dc.date.accessioned2023-10-13T16:15:13Z
dc.date.issued2023
dc.date.updated2023-10-12T15:53:35Z
dc.descriptionTesis (Ph.D. in Biological and Medical Engineering)--Pontificia Universidad Católica de Chile, 2023
dc.description.abstractThis PhD thesis focuses on the decoding of silent speech from electroencephalogram (EEG) signals using Brain-Computer Interfaces (BCIs) and transfer learning tech- niques. Silent speech, also known as imagined speech, involves the generation of neural patterns related to speech without audible sound production. Decoding silent speech from EEG signals is challenging due to the noisy and low-resolution nature of the signals. While previous studies have achieved varying levels of success in classify- ing individual words or phonemes from EEG signals, the translation of full sentences remains intricate. The motivation for this research stems from the potential of BCIs to restore com- munication abilities for individuals with speech impairments caused by neuromuscular disorders. Existing speech decoding methods, particularly for silent speech, heavily rely on neural networks and require large amounts of data, which is both costly and time-intensive to record. Transfer learning is proposed as a solution to address this data limitation in silent speech decoding. The research encompasses three main studies: subject-independent sentence de- coding, decoding individual words from continuous imagined speech, and applying Connectionist Temporal Classification (CTC) loss for EEG-based speech decoding. The findings highlight the potential of Deep Neural Networks (DNNs) in decoding speech signals across individuals, while also shedding light on the challenges inherent in silent speech production. Additionally, this thesis introduces the “Large Spanish EEG” dataset, facilitating further advancements in the field. Overall, this thesis demonstrates the feasibility and potential of deep learning techniques and transfer learning in EEG-based speech decoding. The research not only paves the way for future exploration but also brings us closer to the development of robust and real-world applicable tools for silent speech decoding. Ultimately, these advancements hold the promise of enhancing the quality of life for individuals with speech and language impairments.
dc.description.funderANID Chile
dc.description.version2024-05-02
dc.fechaingreso.objetodigital2023-10-12
dc.format.extent125 páginas
dc.fuente.origenAutoarchivo
dc.identifier.doi10.7764/tesisUC/BIO/75112
dc.identifier.urihttps://doi.org/10.7764/tesisUC/BIO/75112
dc.identifier.urihttps://repositorio.uc.cl/handle/11534/75112
dc.information.autorucInstituto de Ingeniería Biológica y Médica ; Rodríguez Fernández, María ; 0000-0003-1966-2920 ; 1031920
dc.information.autorucInstituto de Ingeniería Biológica y Médica ; Valle Araya, Carlos ; S/I ; 233198
dc.language.isoen
dc.nota.accesocontenido completo
dc.rightsacceso abierto
dc.subject.ddc610
dc.subject.deweyMedicina y saludes_ES
dc.subject.ods03 Good health and well-being
dc.subject.odspa03 Salud y bienestar
dc.titleImproving brain-to-text models utilizing recurrent neural networks and transfer learning for neuroprosthetic speeches_ES
dc.typetesis doctoral
sipa.codpersvinculados1031920
sipa.codpersvinculados233198
Files
Original bundle
Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
main_con_logo.pdf
Size:
6.92 MB
Format:
Adobe Portable Document Format
Description:
License bundle
Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
license.txt
Size:
1.98 KB
Format:
Item-specific license agreed upon to submission
Description: