Human Cognition: Decoding Perceived, Attended, Imagined Acoustic Events and Human-Robot Interfaces The Team • • • • • • • • Adriano Claro Monteiro Alain de Cheveign Anahita Mehta Byron Galbraith Dimitra Emmanouilidou Edmund Lalor Deniz Erdogmus Jim O’Sullivan • • • • • • Mehmet Ozdas Lakshmi Krishnan Malcolm Slaney Mike Crosse Nima Mesgarani Jose L Pepe ContrerasVidal • Shihab Shamma • Thusitha Chandrapala The Goal • To determine a reliable measure of imagined audition using electroencephalography (EEG). • To use this measure to communicate. What types of imagined audition? • Speech: – Short (~3-4s) sentences • “The whole maritime population of Europe and America.” • “Twinkle-twinkle little star.” • “London bridge is falling down, falling down, falling down.” • Music – Short (~3-4s) phrases • Imperial March from Star Wars. • Simple sequence of tones. • Steady-State Auditory Stimulation – 20 s trials • Broadband signal amplitude modulated at 4 or 6 Hz The Experiment • 64 – channel EEG system (Brain Vision LLC – thanks!) • 500 samples/s • Each “trial” consisted of the presentation of the actual auditory stimulus (“perceived” condition) followed (2 s later) by the subject imagining hearing that stimulus again (“imagined” condition). The Experiment • Careful control of experimental timing. • Perceived...2s... Imagined...2 s x 5 ... Break... next stimulus 4, 3, 2, 1, + Data Analysis - Preprocessing • Filtering • Independent Component Analysis (ICA) • Time-Shift Denoising Source Separation (DSS) – Looks for reproducibility over stimulus repetitions Data Analysis: Hypothesis-driven. • The hypothesis: – EEG recorded while people listen to (actual) speech varies in a way that relates to the amplitude envelope of the presented (actual) speech. – EEG recorded while people IMAGINE speech will vary in a way that relates to the amplitude envelope of the IMAGINED speech. Data Analysis: Hypothesis-driven. • Phase consistency over trials... • EEG from same sentence imagined over several trials should show consistent phase variations. • EEG from different imagined sentences should not show consistent phase variations. Data Analysis: Hypothesis-driven. Actual speech Consistency in theta (4-8Hz) band Imagined speech Consistency in alpha (8-14Hz) band Data Analysis: Hypothesis-driven. Data Analysis: Hypothesis-driven. • Red line – perceived music • Green line – imagined music Data Analysis - Decoding Data Analysis - Decoding Original Reconstruction London’s Bridge 1 Twinkle Twinkle 1.2 0.9 1 0.8 0.7 0.8 0.6 0.6 0.5 0.4 0.4 0.3 0.2 0.2 0 0.1 0 0 20 40 60 80 100 120 r = 0.30, p = 3e-5 140 160 180 200 -0.2 0 20 40 60 80 100 120 140 r = 0.19, p = 0.01 160 180 200 Data Analysis - SSAEP Data Analysis - SSAEP Perceived 4Hz 6Hz Imagined Data Analysis • Data Mining/Machine Learning Approaches: Data Analysis • Data Mining/Machine Learning Approaches: SVM Classifier Input Class Labels EEG data (channels × time) : 𝑒 𝑡 1 𝐸𝐸𝐺 = ⋮ 𝑒(𝑡)64 ⋮ Concatenate channels: 𝐸𝐸𝐺 = 𝑒 𝑡 1 1 1 1 1 1 0 0 … 𝑒(𝑡)64 1 0 0 0 0 0 1 Group N trials: 𝐸𝐸𝐺1 𝑋= ⋮ 𝐸𝐸𝐺𝑁 Input covariance matrix: 𝐶𝑋 = 𝑋𝑋 𝑇 Predicted Labels 0 0 1 0 1 ⋮ 1 1 0 1 1 SVM Classifier Results Decoding imagined speech and music: Mean DA = 87% Mean DA = 90% Mean DA = 90% DCT Processing Chain DSS Output (Look for repeatability) Raw EEG Signal (500Hz data) Mean input1 DCT Output (Reduce dimensionality) Mean DSS result for out1 DCT Model for class 1 0.5 10 2 1 20 4 30 1.5 6 40 8 50 60 2 10 200 400 600 800 1000 1200 1400 1600 1800 200 400 Mean input2 600 800 1000 1200 1400 1600 1800 2.5 1 2 3 Mean DSS result for out2 4 5 6 7 8 9 10 8 9 10 DCT Model for class 2 0.5 10 2 1 20 4 30 1.5 6 40 50 8 60 10 200 400 600 800 1000 1200 1400 1600 1800 2 200 400 600 800 1000 1200 1400 1600 1800 2.5 1 2 3 4 5 6 7 Percentage accuracy DCT Classification Performance Data Analysis • Data Mining/Machine Learning Approaches: – Linear Discriminant Analysis on Different Frequency Bands Music vs Speech Speech 1 vs Speech 2 Music 1 vs Music 2 Speech vs Rest Music vs Rest - results ~ 50 – 66% Summary • Both hypothesis drive and machine-learning approaches indicate that it is possible to decode/classify imagined audition • Many very encouraging results that align with our original hypothesis • More data needed!! • In a controlled environment!! • To be continued...
© Copyright 2025 Paperzz