EEG - Normative Data Base.doc

(853 KB) Pobierz
http://www

http://www.appliedneuroscience.com/LSNDBWEB.htm

 

SENSITIVITY & SPECIFICITY OF THE NEUROGUIDE EEG  NORMATIVE DATABASE:

Validation and Clinical Correlation

 

 

R.W. Thatcher, Ph.D., R.A. Walker, B.A.

C.J. Biver, Ph.D., D. M. North, M.A. and R. Curtin, B.A.

 

 

We each had different contributions as a team to compute and test these databases. Robert Thatcher was the program director and P.I., Rebecca Walker, Richard Curtin, Duane North and Dr. Carl Biver were the database managers and analysists and colleague scientists who spent years analyzing these data (Walker and Thatcher = 22 years and no one for less than 5 years).  Dr. Biver was responsible for writing of IDL and C programs for the Interpolations, sliding averages, digital signal processing and, with Duane North, for the cross-validation procedures.  Rebecca Walker computed the means and standard deviations of EEG values in the database and grouped and tabulated subjects, and oversaw the organization of the clinical database and Richard Curtin helped edit and archive data and organize the databases.   Duane North, Rebecca Walker and Richard Curtin were the data analyzers for the computation of sliding averages and for writing Excel programs to compute and tabulate the Skewness and Kurtosis and the parametric and non-parametric sensitivity and specificity statistics in the cross-validation studies as well as the clinical correlations.    

 

 

 

Abstract

 

              The digital Electroencephalogram (EEG) was recorded from19 scalp locations from 625  screened and evaluated normal individuals ranging in age from 2 months to 82 years.  After editing to remove artifact; one year to 5 year groupings were selected to produce different  average age groups..  Estimates of Gaussian distributions and logarithmic transforms of the digital EEG were used to establish approximate Gaussian distributions when necessary for different variables and age groupings.   The sensitivity of the lifespan database was determined by Gaussian Cross-Validation for any selection of age range in which the average percentage of Z scores ± 2 st. dev. equals approximately 2.3% and the average percentage for ± 3 st. dev. equals approximately .13%.   It was hypothesized that measures of Gaussian cross-validation of Z scores is a common metric by which the statistical sensitivity of any normative database for any age grouping can be calculated.   This notion was tested by computing eyes closed and eyes open Average Reference and Current Source Density norms and independently cross-validating and comparing to the Linked Ears norms.   The results indicate that age dependent Digital EEG normative databases are reliable and stable and behave like different Gaussian lenses that spatially focus the Electroencephalogram.   Clinical correlations of a normative database are determined by content validation and correlation with neuropsychological test scores and discriminate accuracy.    Non-parametric statistics were presented as an important aid to establish the alpha level necessary to reject a hypothesis and to estimate Type I and Type II errors, especially when there are multiple comparisons of an individual’s EEG to any normative EEG database. 

 

 

1.0- Introduction

              There are many potential uses of a normative EEG database among the most important being a statistical “guess” as to the “error rate” or to the probability of finding a particular patient’s EEG measure within a reference normal population. [1][1]  Most other uses of a reference EEG database also involve statistics and the same statistics that all of modern clinical medicine relies upon.   For example, null hypothesis testing, measures of reliability, sensitivity, power, predictive validity, content validity, etc.  all depend on specific assumptions and statistical procedures. 

              Predictive accuracy and error rates depend on the data that make up a given EEG database and the statistics of the database.  The statistical foundations of the scientific method were visited by the Supreme Court in Daubert, 1993 regarding admissibility of scientific evidence.    The Four Daubert Factors for scientific standards of admissibility in Federal Courts were: 1-  hypothesis testing, 2- error estimates of reliability and validity, 3- peer reviewed publications and 4- general acceptance (Mahle, 2001) [2][2].    These four Daubert factors for several EEG normative database have already been met.  The minimal standards are publication of: 1- inclusion/exclusion criteria, 2- methods to remove artifact and adequate sample sizes per age groups, 3-  demographic representativeness  (e.g., balanced gender, ethnicity, socioeconomic status, etc.), 4-  means and standard deviations as being normally distributed or “Gaussian” including Gaussian Cross-Validation and, 5- Content validity by correlations with Neuropsychological test scores and school achievement scores, etc. as validation.  Predictive validity is determined by regression and classification statistics.   Predictive validity relates to the classification accuracy, clinical severity, clinical outcome, etc. estimates.    The sensitivity and specificity of any EEG database is directly proportional to its adherence to the established statistical principals in the history of statistics (Hayes, 1973). 

 

1.1 – General Method to Produce a Valid Normative EEG Database

              Figure 1 is an illustration of a step by step procedure by which any normative EEG database can be validated and sensitivities calculated.   The left side of the figure is the edited and artifact clean and reliable Digital EEG Time Series which may be re-referenced or re-Montaged, which is then analyzed in either the time domain or the frequency domain.

Fig. 1 - Illustration of the step by step procedure to Gaussian cross-validate and then validate by correlations with clinical measures in order to estimate the predictive and content validity of any EEG normative database.  The feedback connections between Gaussian Cross Validation and the means and standard deviations refers to transforms to approximate Gaussian if the non-transformed data is less Gaussian (see section 6).   The Clinical Correlation and Validation arrow to the Montage stage represents repetition of clinical validation to a different montage or reference or condition such as eyes open, active tasks, eyes closed, etc.  to the adjustments and understanding of the experimental design(s) (see sections 6 to 8).

             

              The selected normal subjects are grouped by age with sufficiently large sample size and the means and standard deviations of the EEG time series and/or Frequency domain analyses are computed for each age group.   Transforms are applied to approximate a Gaussian distribution of the EEG measures that comprise the means.   Once approximation to Gaussian is completed, then Z scores are computed for each subject in the database and leave one[3][3] out Gaussian Cross-Validation is computed in order to arrive at an optimum Gaussian Cross-validation sensitivity.    Finally the Gaussian validated norms are subjected to content and predictive validation procedures such as correlation with Neuropsychological test scores and intelligence, etc. and also discriminant analyses and neural networks and outcome statistics, etc.   The content validations are with respect to clinical measures such as intelligence, neuropsychological test scores, school achievement, clinical outcomes, etc.  The predictive validations are with respect to the discriminative, statistical or neural network clinical classification accuracy.  Both parametric and non-parametric statistics are used to determine the content and predictive validity of a normative EEG database..

 

1.2- Example of a Normative EEG Database and the Procedure in Section 1.1

              An example of the step-by-step procedure in Figure 1 to produce a validated normative digital EEG database will be provided to show how any normative reference database can be constructed to meet measurable standards of reliability and validity. The Steps in Figure 1 can be repeated for different selections of subjects

 

2.0 - Subject and Variable Selection

              Nineteen (19) channels of EEG and a EOG (Electro-Oculogram) channel, a two hour battery of evoked potential tests and active challenges, psychometric tests, dietary evaluations, anthrometric measurements, demographic and trace element measurements from a population of 1,015 rural and urban children were collected (Thatcher et al, 1983; 1987; Thatcher, 1997).  The principal goal of this project was to evaluate the effects of environmental toxins on child development and to determine the extent to which good or poor diets may ameliorate or exacerbate the deleterious effects of environmental toxins.  Two data acquisition centers were established, one at the rural University of Maryland Eastern Shore campus and one at the urban campus of the University of Maryland School of Medicine in Baltimore, Maryland.  Identical data acquisition systems were built and calibrated, a staff was trained using uniform procedures and a clinical and psychometric protocol were utilized in the recruitment of normal subjects.  The total of 1,015 subjects ranging in age from 2 months to 82 years were tested during the period from 1979 to 1987.  Of these subjects, 564 met the criteria of normalcy and were included in the normative reference database (Thatcher et al, 1987; Thatcher 1997).   In 2000 the original digital EEG was revisited and a different selection of individuals was selected that also spanned the same interval from 2 months to 82 years and included 61 additional adult subjects to give rise to a total sample size of 625 subjects.  The expanded selection contained more individuals between the ages of 25 and 55 years of age.

              Figure 2 shows the number of subjects per year in the normative EEG lifespan database.  It can be seen that the largest number of subjects are in the younger ages (e.g., 1 to 14 years, N = 470) when the EEG is changing most rapidly.  As mentioned

Fig. 2 - The number of subjects per year in the Lifespan EEG reference normative database.  The database is a “life-span” database with the 2 months of age being the youngest subject and 82.3 years of age being the oldest subject.  This figure shows the number of subjects constituting mean values which range from a mean of .5 years to 62.6 years of age and constituting a total number of subjects = 625.

 

previously, a proportionately smaller number of subjects represents the adult age range from 14 to 83 years (N = 155).   Fifteen one-year groupings of subjects were computed with reasonable sample sizes from birth to 15 years of age.   Thirteen out of the 15 one year age groups have N > 20 with the largest sample size at age 3 to 4 years, N = 45.   The smallest one year sample size was between age 2 and 3 when N = 16.

              For each subject, original selections of the original digital EEG occurred by different artifact procedures involving the use of NeuroGuide editing selections in 2001.   Original arrangements of coherence, phase, amplitude asymmetry and relative power also occurred when comparing the database to previous publications and the 1988 copyright (Thatcher et al, 1987; Thatcher, 1988; Thatcher, 1997).   Although different selections of digital EEG values and different arrangements of the original digital EEG have occurred since 1987, nonetheless, the Gaussian validations and sensitivities of the previous databases and the current 2001-2002 database were all similar and equally valid and Gaussian distributed within a  90% to 99% range depending on the measure.   The original digital EEG and subjects and neuropsychological test scores that were measured from 1979 to 1987 are the same.

 

3.0- Inclusion/Exclusion Criteria, Demographics and Gender

              Details of the neuropsychological testing, demographic and sampling of the normative 1987 EEG database were previously published in Thatcher et al (1983; 1986; 1987) and Thatcher (1997).  Some but not all of the 61 adults added in 2000 - 2001 were given neuropsychological tests and other evaluations to help determine “normalcy”, however, all of the subjects were interviewed and filled out a history and neurological questionnaire.   All of the 61 added adults were gainfully employed as professors, graduate students, and other successfully employed adults without a history of neurological problems.                 Normalcy for the age range from 2 months to 18 years was determined by one or more exclusion/inclusion criteria: 1- a neurological history questionnaire given to the child’s parents and/or filled out by each subject, 2- psychometric evaluation of I.Q., and/or school achievement, 3- for children the teacher and class room performance as determined by school grades and teacher reports and presence of environmental toxins such as lead or cadmium.  A Neurological questionnaire was obtained from all of the adult subjects >18 years of age and those in which information was available about a history of problems as an adult were excluded.   . 

 

3.1- Intelligence and School Achievement Criteria:

              Psychometric, demographic and socioeconomic status measures were obtained from each child, adolescent and for some of the adults.  Different psychometric tests were administered depending upon the age of the child.  There is little reliability in the I.Q. tests of infants, however, when possible the infant's Apgar Score was obtained and the Vineland Social Maturity Scale test was administered (age birth to 2 years, 4 months).  From age 2 years to 3.99 years, the McCarthy Intelligence Scale Test was administered, from age 4.0 years to 5.99 years the Weschler Preschool and Primary Scale of Intelligence (WIPPSI) test was administered, from age 6.0 years to 16.99 years the Wechsler Intelligence Scale for Children (WISCR, 1972) was administered and from age 17.0 years to adulthood the Wechsler Adult Intelligence Scale test (WAIS) was administered.  In addition to Intelligence Tests, the Wide Range School Achievement test (WRAT) was administered to the school age children and grade cards were obtained from the public school systems.  Finally, a variety of neuropsychological tests were administered including the pegboard test of skilled motor movements, the Stott, Moyes and Henderson Test of Motor Impairment (MIT) and a eight item laterality test (see Thatcher et al, 1982; 1983 for further details).

      The criteria for entry into the normative database for those subjects given I.Q. tests and school achievement tests were:

1- A Full Scale I.Q. > 70.

2- WRAT School Achievement Scores > 89 on at least two subtests (i.e., reading, spelling, arithmetic) or demonstrated success in these subjects.

3- A grade point average of 'C' or better in the major academic classes (e.g., English, mathematics, science, social studies and history).

 

3.2-  Demographic Characteristics:

              It is important that the demographic mixture of males and females, different ethnic groups and socioeconomic status be reasonably representative of expected North American clientele.  The normative EEG database is made up of 58.9% males, 41.1% females, 71.4% whites, 24.2% blacks and 3.2% oriental.  Socioeconomic status (SES) was measured by the Hollingshead four factor scale (Hollingshead, Four factor Index of Social Status, 1975).  (see Thatcher et al, 1983 for details).

 

3.3 - Time of Day and Other Miscellaneous Factors

              There are many uncontrollable factors that influence the frequency spectrum of the EEG.  In general these factors are all confounded, and it would require an enormously expensive and large sample size to control each factor individually.  Even if one could control each factor, such experimental control would preclude the practical use of a database since each patient’s EEG would have to be acquired in a precisely matching manner.   Statistical randomization is one of the best methods to deal with these uncontrollable and miscellaneous factors.  Statistical randomization of a database involves randomly varying time of day of EEG acquisition, time between food intake and EEG acquisition, food content and EEG acquisition, etc. across ages, sex and demographics.  Because these factors are confounded with each other, randomization with a sufficient sample size will result in increased variance but, nonetheless, convergence toward a gaussian distribution.  Such convergence, even in the face of increased variance, still allows quantitative comparisons to be made and false positive and false negative error rates (i.e., sensitivity) to be calculated.  The method of statistical randomization of miscellaneous factors was used in the Matousek & Petersen, Thatcher, John and Duffy EEG normative databases (John et al, 1988; Thatcher et al, 1989; Duffy et al, 1994).

 

4.0 – Digital Electroencephalographic Recording Procedures

              EEG was recorded  and digitized at a rate of 100 Hz from the 19 leads of the International 10/20 system of electrode placement referenced to linked ear lobes and one bipolar EOG lead (Electrooculogram) (i.e., a total of 20 channels). (Thatcher et al, 1983; 1986; 1987, Thatcher 1997).  When head size was amenable, the data were acquired using a stretchable electrode cap (Electrocap International, Inc.).  When head sizes were either too small or too large for the electrocap, then the electrophysiological data were acquired by applying standard silver disk Grass electrodes.   Amplifiers were calibrated using sine wave calibration signals and standardized procedures and a permanent record made before and after each test session.  The frequency response of the amplifiers was approximately 3db down at 0.5 Hz and 30 Hz.  Impedance was measured and recorded for each electrode and efforts were made to obtain impedance measures less than 10K ohms (most of the impedance’s were < 5k ohms) for all subjects. 

 

4.1 – Artifact Removal and Quality Control Procedures

              EEG recording lengths varied from 58.6 seconds to 40 minutes.  Artifact rejection involved using the NeuroGuide editing procedures in which a 1 to 2 second template of “clean” or “artifact free” EEG was selected.   This template was then used to compute matching amplitudes of EEG using a flexible criteria of equal amplitudes to amplitudes that are 1.25 or 1.5 times larger in amplitude.   The decision as to which clean EEG sample multiplier to use was determined by the length of the sample 58.6 seconds as a minimum, visual inspection of the digital EEG and when split-half reliability > 0.97.   After multiple visual inspections and selection of “clean” EEG samples the edited samples varied in length from 58.6 seconds to 142.4 seconds.   Average split-half reliability = 0.982 for the selected EEG in the database.   Care was taken to inspect the EEG from each subject in order to eliminate “drowsiness” or other state changes in the EEG which may have been present in the longer EEG recording sessions.  No evidence of sharp waves or epileptogenic events were present in any of the EEG records.

 

4.2- Re-Montage to the Surface Laplacian and Average Reference

              The average reference involved summing the voltages across all 19 leads for each time point and dividing this value into the microvolt digital value from each lead at each time point.  This procedure produced a digital EEG time series that was then submitted to the same age groupings and Power Spectral analyses and the same Gaussian normative evaluations as for Linked ears.   See Figure 1.

              The reference free surface Laplacian or current source density (CSD) was computed using the spherical harmonic Fourier expansion of the EEG scalp potentials  to estimate the current source density (CSD) directed at right angles to the surface of the scalp in the vicinity of each scalp location (Pascual-Marqui et al., 1988).  The CSD is the second spatial derivative or Laplacian of the scalp electrical potentials which is independent of the linked ear reference itself.  The Laplacian is reference free in that it is only dependent upon the electrical potential gradients surrounding each electrode.  The Laplacian transform also produces a new digital EEG time series of estimates of current source density in microamperes, that were also submitted to the same age groupings Spectral Analyses (see Figure 1).

 

4.3 - Complex Demodulation Computations

              The mathematical details of both the FFT and complex demodulation are adequately described in Otnes and Enochson, (1977);  Bendat and Piersol, (1981).   The EEG norms use both the complex demodulation and the FFT so that users can compare and contrast both methods in the same subject or application.  Complex demodulation is a time domain digital method of spectral analysis whereas the fast Fourier transform (FFT) is a frequency domain method.  These two methods are related by the fact they both involve sines and cosines and both operate in the complex domain and in this way represent the same mathematical descriptions of the power spectrum.  The advantage of complex demodulation is that it is a  time domain  method and less sensitive to artifact and it does not require even integers of the power of 2 as does the FFT.   The FFT integrates frequency over the entire epoch length and requires windowing functions which can dramatically affect the power values whereas complex demodulation does not require windowing (Otnes and Enochson, 1972; 1978).   Complex demodulation was computed for the linked ears and eyes closed condition.   Future analyses are being considered for the computation of complex demodulation for average reference and the Laplacian estimate of current source density for eyes open and closed conditions.  However, due to the large amount of data and the large number of computations, the FFT may be the preferred method to conduct these analyses.

             

4.4 – FFT Linked Ears, Average Reference and Laplacian

              The 100 samples per second digital EEG were cubic-spline interpolated to 128 samples per second using standard procedures (Press, 1994).  The second step was to high pass filter the EEG at 40 Hz to eliminate any possible splice artifact that may have been produced by the short segment NeuroGuide editing method described in section 4.1.   The third step was to compute the FFT Power Spectral Density.   Four second epochs were used to compute the FFT Power Spectral Density thus producing 0.5 Hz resolution and a Hanning window was used for each 4 second epoch computation.  The 75% sliding window method of Kaiser and Sterman (2001) was used to compute the FFT normative database for linked ears, average reference and Laplacian estimator of current source density (CSD) in which successive four second epochs were advanced by 500 millisecond steps in order to minimize the effects of the FFT windowing procedure.  The FFT Power Spectral Density and the 256 point and 2 second epochs produced a total of 61 frequency values in uv2 /Hz from 0 to 30 Hz in 0.5 Hz increments. 

              This procedure was repeated for linked ears, average reference and Laplacian digital values for both the eyes closed and eyes open conditions, thus producing for a given subject a total of six different 61 point FFT power spectral density values.  These values were then used to compute means and standard deviations for different age groups as described in the next section (5.0) 

5.0 – Amplifier and Digital Matching

 

The frequency characteristics of all amplifiers differ to some extent, especially in the < 3 Hz  and > 20 Hz frequency range and there are no universal standards that all EEG amplifier manufacturers must abide by.   Therefore, amplifier filter and gain characteristics must be equilibrated to the amplifier gains and frequency characteristics of the normative EEG amplifiers that acquired the EEG in the first place.  A simple method to accomplish this is to inject into each amplifier system microvolt sine waves from 0 to 40 Hz in 1 Hz steps and at three different amplitudes.  The ratio of the frequency response characteristics between the normative EEG amplifiers and the amplifier characteristics by which EEG was measured from a patient can be used as equilibration factors to approximately match the norms.  There are some frequencies that are so severely attenuated by the amplifier filters that equilibration to the normative database amplifiers will not be able to recover the signal.  For example, rations of > 5.0 will significantly amplify the noise of the amplifiers where little or no EEG signal is present and render the Z scores invalid. 

              It should be kept in mind, that even with matching of amplifier characteristics within 3 to 5% error the enormous variability in skull thickness effects the amplitude and frequency characteristics of the EEG itself far more than slight differences in amplifier characteristics.   For example, the human skull is on the average 80 times less conductive than the brain and scalp.   Therefore, an individual with a 10% thinner skull may be result in a 800% change in EEG amplitude across all frequencies. 

 

6.0- Statistical Foundations: Gaussian Distributions

              The Gaussian or Normal distribution is a non-linear function that looks like a ideal bell shaped curve and provides a probability distribution which is symmetrical about its mean.   Skewness and kurtosis are measures of the symmetry and peakedness, respectively of the gaussian distribution.  In the ideal case of the Gaussian distribution skewness and kurtosis = 0.  In the real world of data sampling distributions skewness and kurtosis = 0 is never achieved and, therefore, some reasonable standard of deviation from the ideal is needed in order to determine the approximation of a distribution to Gaussian.   In the case of the Lifespan EEG Database we used the criteria of approximation as a reasonable measure of Gaussian distribution.   The most serious type of deviation from normality is "Skewness" or a unsymmetrical distribution about the mean (e.g., a tail to the left or right of the mean), while the second form of deviation from normality "Kurtosis" is the amount of peakedness in the distribution, which is not as serious an offense since the variance is symmetrical about the mean (mean = median).  However, it is preferable to attempt to achieve normality as best as one can to insure unbiased estimates of error.  The primary reason to achieve "Normality" is that the sensitivity of any normative database is determined directly by the shape of the sampling distribution.   In a normal distribution, for example, one would expect that 5% of the samples will be equal to or greater than ± 2 standard deviations and approximately .13 % ± 3 SD.

              It is important to note that automatic and blindly applied transformations of EEG measures does not insure improved normality of the sampling distribution.  For example, it is simple to demonstrate that while some transformations may improve the normality of distributions, these same transforms can also degrade the normality of the distributions.  For example, table I shows the effects of transforms on the distributions of the various EEG variables in the Lifespan EEG reference normative database.   The “No Transform” column shows the deviation from Gaussian for the untransformed or raw

 

Table I: Gaussian Distribution of the EEG Normative Database

 

EEG

Measure

Skewness

Kurtosis

No Transform

Transformed

No Transform

Transformed

...
Zgłoś jeśli naruszono regulamin