Cross-Correlations and Cleaning Up Data

Cross-Correlations and
Cleaning Up Data
Jessica Ferguson
1



Senior Computer Science Major
English: Creative Writing Minor
Pacific University
Project Aims: HSD Project



Aim 1: Collecting and transcribing spoken
language data
Aim 2: Automatically deriving features from
spoken language samples
Aim 3: Characterizing features derived from
Aim 2
3
My Task



Falls under Aim 1
Improving the quality of the recordings in the
corpus
Reducing noise to give clearer speech
4
Subjects

Currently enrolled in studies at the Layton Aging
and Alzheimer’s Disease Center at OHSU
Individuals over 90
 Individuals with Mild Cognitive Impairment (MCI)

Test Battery





Wechsler Logical Memory I/II (Story Recall)
Category Fluency (Fruits, States)
Picture Description Task
Autobiographical reflections
Conversational Speech
6
Recording Setup


Same for all sessions
Four different microphones set up

Tests administered by examiner
7
Characteristics of Recordings


Similarly-shaped waves
Shifted horizontally
8
Sample Waves
9
Shifting Files


Shifting files is relatively easy
But how far to shift?
10
Close-up of Comments Files
11

Observed shift:
380
 320
 315

12
Calculating Shift – Cross-Correlation

Cross-correlation: a measure of how similar one
signal is to another
To calculate: split the file into overlapping windows
 Take windows of the same length in another file
 Multiply them together

13
Cross-Correlation Cont.




The window we multiply it by in the other file
keeps getting moved by one sample (1/16 msec)
If corresponding values have the same sign, they
contribute positively
If one is negative and the other is positive, they
contribute negatively
We take the highest value from the range
14
Issues with Cross-Correlation

With original parameters:
Window length: 1280 samples
 Lag: -400 to 400 samples
 For one value: 1280 * 800 = 512,000



One value every 10 msec: 100 values per second
of file correlated
This gets unmanageable very quickly
15
Time Under Original Parameters

Correlate 1.5s of files: up to 20 minutes
Relatively high accuracy, but impractical

Task: Reduce time while maintaining accuracy

16
Optimizing Parameters

Parameters that could be adjusted:
Window Size
 Lag
 Number of correlations (how much of the file gets
correlated)

17
Window Size

Initial parameters were 200 msec
Decreasing below 80 msec resulted in unacceptable
loss of accuracy
 Runtime was improved but not significantly enough

18
Number of Correlations



Unfortunately, correlations are not always
perfect
We take the mode of the correlations produced
n = 150 was the minimum, and still had a high
error rate
19
Lag

Recall the sound wave images from before:
20
Lag cont.



Assume that these are representative
Lag values should all be between 300-400
samples (18-25 msec)
Add this to previous improvements:

Runtime for one set of four files decreases to about
5-6 min
21
Other Benefits

If the assumption holds:
Error from optimal value decreases
 Max. error decreases from 50 msec to 6msec

22
Original File

Taken from a picture description task
23
Shifted File

The same file, but correlated and shifted
24
Acknowledgements



Paul Hosom and Brian Roark
Fellow Interns
Everyone who has made me welcome at CSLU
Questions?