RESEARCH ARTICLE


Auditory Evoked Potential Response and Hearing Loss: A Review



M. P Paulraj*, 1, Kamalraj Subramaniam 1, Sazali Bin Yaccob 1, Abdul H. Bin Adom 1, C. R Hema2
1 Department of Mechatronic Engineering, Universiti Malaysia Perlis, Malaysia
2 Faculty of Engineering, Karpagam University, India


Article Metrics

CrossRef Citations:
36
Total Statistics:

Full-Text HTML Views: 2197
Abstract HTML Views: 513
PDF Downloads: 414
Total Views/Downloads: 3124
Unique Statistics:

Full-Text HTML Views: 1161
Abstract HTML Views: 363
PDF Downloads: 342
Total Views/Downloads: 1866



Creative Commons License
© Paulraj et al.; Licensee Bentham Open.

open-access license: This is an open access article licensed under the terms of the Creative Commons Attribution Non-Commercial License (http://creativecommons.org/licenses/by-nc/3.0/) which permits unrestricted, non-commercial use, distribution and reproduction in any medium, provided the work is properly cited.

* Address correspondence to this author at the The School of Mechatronic Engineering, Universiti Malaysia Perlis, Malaysia; Tel: +60164353674; Fax: +6049885167; E-mail: kamalrajece@gmail.com; kamalrajece@yahoo.co.in


Abstract

Hypoacusis is the most prevalent sensory disability in the world and consequently, it can lead to impede speech in human beings. One best approach to tackle this issue is to conduct early and effective hearing screening test using Electroencephalogram (EEG). EEG based hearing threshold level determination is most suitable for persons who lack verbal communication and behavioral response to sound stimulation. Auditory evoked potential (AEP) is a type of EEG signal emanated from the brain scalp by an acoustical stimulus. The goal of this review is to assess the current state of knowledge in estimating the hearing threshold levels based on AEP response. AEP response reflects the auditory ability level of an individual. An intelligent hearing perception level system enables to examine and determine the functional integrity of the auditory system. Systematic evaluation of EEG based hearing perception level system predicting the hearing loss in newborns, infants and multiple handicaps will be a priority of interest for future research.

Keywords: Auditory brainstem response, Auditory evoked potential, Electroencephalogram, Hearing loss, Hearing perception, neural networks.



1. INTRODUCTION

Hearing loss is the most prevalent sensory disability in the world. Over 275 million people in the world are differentially abled hearing level. Hearing impairment survey was conducted in different countries, and reported that 0.5 per cent of newborns have the onset of sensorineural hearing disorder [1]. Consequences of hearing loss include speech disorder, communication disorder, delay in language understanding, educational disadvantage, social separation and stigmatization [2, 3]. Universal newborn hearing screening test consists of transient evoked oto-acoustic emissions (TEOAE) and automated auditory brainstem response (AABR). Auditory brainstem response (ABR) is an electrical potential signal emanated from the scalp of the brain by presenting a sound stimulus to assess the functioning of auditory neuropathy by using electroencephalography [4]. EEG is a non-invasive clinical tool used for diagnosing brain diseases and also useful for both physiological research and medical applications. Vital brain activities from newborns to adults can be monitored using EEG signals. The somatosensory stimuli associated to visual and auditory reflect the hearing ability level [5-7]. EEG based hearing threshold estimation is also more ideal for persons who lack to provide the consistent or reliable behavioral response to the acoustic stimuli.

An ad-hoc intelligent hearing level assessment system is essential for determing the threshold level of a person for newborns, infants and multiple handicaps. The technological advancement in electronics and digital signal processing techniques enable the implementation of intelligent hearing level assessment system using the EEG as a reliable measurement technology.

This article presents a review on the recent physiological experiments that examined the effectiveness of auditory evoked potential in determining the hearing threshold level.This paper is organized as follows. In section 2 , the fundamentals of event related potential and evoked potential are given. In the following section, hearing threshold level estimation from auditory evoked potential signal is presented. Finally, the paper is concluded with discussion and conclusion.

2. NEUROPHYSIOLOGICAL SIGNAL

The electroencephalography is a non-invasive technique widely used in diagnosing many neurological diseases and problems associated with brain dynamics. The EEG signal is a clear indicator of the electrical activity of the brain and contains useful information about the brain state. Brain signals resulting from steady state tasks are briefly explained in subsections as follows.

2.1. Event Related Potential (ERP)

ERPs are task oriented potentials, spatio temporal patterns of brain signal, occurring to an event or task with respect to applied stimulus time. The transition from a disordered state to an ordered state of a brain in response to synchronisation of certain task or event, gives rise to event related potential [8, 9]. ERP refelcts the activity originating within brain and is phase locked to the stimulus onset [10, 11]. ERP provides a powerful tool for objective assessment of cognitive status and clinical studies of brain functions such as attention, memory and language.

2.2. Evoked Potential (EP)

Evoked potentials are commonly occurred in response to a physical stimulus. The physical stimuli are patterns of band energy received by the senses and their corresponding sensory receptors convert this energy into nerve impulse to the brain. The nerve impulses are interpreted in the cerebral cortex as sensations. These sensations are evoked by delivering auditory stimuli such as tone burst or click stimuli. Evoked potential components comprise both exogeneous and endogeneous components of the brain. Evoked potential components are generally fall into two categories, namely, Visual Evoked potential (VEP), and Auditory Evoked Potential (AEP).

2.3. Visual Evoked Potential

VEP is an electrical signal emanated from the brain while a visual stimulus is presented to the subject in a time locked manner. The VEP can be used as a diagnostic tool to detect ocular diseases in patients with visually impaired [12]. VEP response can also be used to detect eye diseases like glaucoma, diabetic retinopathy, multiple sclerosis, ocular hypertension, loss of peripheral (side) vision, macular degeneration and color blindness [13, 14].

2.4. Auditory Evoked Potential

AEP is an electrical signal elicited from the brain while an auditory stimulus is presented in a time-locked manner. AEP signal consists of reproducible positive or negative peaks, latency, amplitude and behavioral correlation. AEPs are much smaller in amplitude compared to the EEG signals [15, 16]. AEP signals can be classified as either transient or steady-state. The AEP signal emanated while perceiving an audio stimuli with slow rate to avoid overlap of the immediate stimuli response and the corresponding evoked potentials are known as transient AEP. The AEP signal emanated while perceiving an audio stimuli with fast rate to induce overlap of individual responses and known as steady-state AEP [17]. ABR comprises the early portion of (0-12 millisecond) of AEPs. ABR is composed of several waves and peaks, known as Jewett waves. The ABR waves or peaks are normally labeled using Roman numerals I-VII. Waves I, III and IV are generally considered as clinically significant [18]. ABRs are used for diagnosis and localization of pathologies affecting brainstem pathways. ABRs are widely employed to test auditory function, and used as an important diagnostic tool in infant hearing screening [19].

Middle latency Auditory Evoked Potential (MLAEP) comprises of (8-50 millisecond) AEPS. Middle latency components were observed to be elicited when the subject was not perceived in the direction of attention to the auditory stimulus [20]. Directing attention of the subject to perceive the auditory stimulus has been reported to enhance the N1 (90 millisecond), P2 (170 millisecond) components of the evoked potential response [21]. MLAEP is useful in the objective assessment of auditory functions in patient and applications which include monitoring the depth of anesthesia, localization of thalamocortical lesions and assessment of the effectiveness of electrical stimulation for cochlear implantation [19, 22].

Mismatch negativity (MMN) comprises of portion of (200-400 millisecond) AEPs. MMN is elicited when flow of identical pure tone (standard stimulus) sounds followed by ‘deviant’ sound. The MMN is one of the genetically earliest perception related responses recorded over the scalp [23]. An MMN peaking at (200-400 millisecond) has been elicited by burst tone frequency change and phonemic vowel change in newborns [24]. The MMN phenomenon is valuable when investigating subjects with their feedback responses are unavailable or unreliable, such as infants. The MMN is a component of brain event related potential that helps to understand the brain process forming the biological substrate of central auditory perception and various forms of auditory memory [25, 26]. Rita ceponiene et al. [27] investigated the mismatch negativity component of event related potential of newborns and their attentiveness to sound duration and frequency. Neonates possess effective neural mechanism to perceive auditory stimuli frequency and duration discrimination mechanisms as indexed by MMN. They have reported that, MMN can be used as index level in indicating the level of hearing ability of neonates.

3. HEARING THRESHOLD ESTIMATION FROM AEP

In this paper, some of the peer reviewed papers were analyzed to estimate the hearing threshold level using perceived EEG signals. In earlier studies, the normal hearing person was subjected to various levels of acoustic stimuli and their corresponding auditory responses were recorded. Auditory evoked potential response reflects the hearing perception level of an individual. The most of the reviewed authors [28-34] investigated in determining the hearing threshold level by identifying the ABR peaks and few authors [47-51] had analyzed entire AEP signal (see Table 1).

Table 1.

Summary of AEP studies.


Authors (Year) Sample size & Stimulus Duration Stimuli Frequency Range (Hz) Stimulus Levels Analysis method/ Feature Extraction Reports/Results
T. W. Picton et al. (1973)  N= 12 ear, 1 sec None 60 dB lowered by fixed 5 dB. The baseline was determined for peak amplitude and latency of the attend and the ignore conditions Potential significance in amplitude was detected between attend and ignore conditions of the stimuli
E. Delgada et al. (1994)  N=24 ear, HL= 11 ear, 10 msec None 10-70 dB at an increment of 10 dB. The peak identification system used a combined matched filtering and rule based system approach Precision for peak V for normal subject was 96 % compared to 82.3 % for hearing impaired
Barrie. W. Jervis et al. (1983) N=3 ear, 100 ms 1 K 40 , 70 dB Energy levels in pre and post stimulus was compared by means of paired t-test. AEP was due to phase reordering and contains additive energy in harmonic component
Robert Boston et al. (1981) N=14 ear, 10 msec 100, 500,1000 30, 50, 70, 80 dB The bias, variance, and mean square error were determined from Hamming window spectral estimates Peak occurs at lowest stimulus intensities, at which wave V is clearly defined.
J. Wilson et al. (1999) N= 240 ear, 0.1 msec None 0, 10, 30, 50, 70, 90 dB ABR waves time and frequency domain features were extracted Both time and frequency domain features of ABR wave reflects change to subject age and gender
J. Wilson et al. (1999) N= 240 ear, 0.1 msec None 90 dB A Daubechies 5 wavelet was employed to discrete analysis the ABR wave Multi resolution wavelet analysis reflects change to subject age and gender
Ulrich Hoppe (2001) N= 22 ear, 300 msec 500, 1 K, 2 K, 4 K 20, 40,60, 80 dB Wavelet features were extracted and then statistical test was conducted on classification The proposed automatic detector finds the response as identical human experts
Robert Boston et al. (1989) N= 14 ear, HL= 16 ear, 1.5 ms None None A rule based expert system with heuristic criteria to identify peak V The proposed system identifies when response was present but not as effective with no response present
Walker et al. (1983) N = 4 ear, 28 clicks/sec None 25, 45, 65 dB Matched filter was employed to detect peak V Matched filter system reduces computational time from 30 min to 5-10 min for a hearing test
Tapio Gronofors. (1993) N= 44 ear, 0.1 clicks/msec None 60, 70, 90 dB Mutlifilters and an attributed automation were used to identify peak V The proposed method identifies 80 per cent of peak V
Authors (Year) Sample size & Stimulus Duration Stimuli Frequency Range (Hz) Stimulus Levels Analysis method/ Feature Extraction Reports/Results
Sudirman et al. (2009) N= 4 ear, 10 sec 40, 500, 5000, 15000 None AEP signal was analyzed by fast fourier transform Feed -forward neural network was used to classify hearing level based on brain signals.
Masumi Kogure et al. (2010) N= 10 ear, 500 msec 440 None Gradient of the wave characteristics were analyzed to classify target and non-target trails. Hearing level perception in targeted stimulus was 65.4-76.3 %, while for non- targeted stimulus was 66.4-70.5 %
Maryam Ravan et al. (2011) 58 newborns, 300 msec 294, 784 70 dB First order and second order moment sequences of wavelet were identified as features Brain neurological developments for hearing was not associated with the age difference in newborns
Sriraam (2012) N= 16 ear, 10 sec None None Two time- frequency domain features were extracted and classified using neural network Applied features has a classification accuracy of 65.3-100 %

3.1. Potential Significance of ABR Wave

Picton et al. [18] examined the effects of attention to the auditory evoked potentials in humans. When the standard auditory stimuli perceived with attention, there was a significant increase in the N1 (90 msec) P2 (170 msec) components and further peak was evoked near 450 msec to the perceived signal. The detected signals based on stimuli condition were used to index the hearing perception level of a subject. A remarkable distinct series of wave (I-VI), after the clicks stimulus was observed and it was first reported by Jewett and Williston. The observed auditory response waveform was consistent and detectable in all subjects. At first, this series of an auditory waveform was known as Jewett wave and latter as auditory evoked potential response or auditory evoked potential. Auditory brainstem responses (ABRs) comprise the early portion of (0-12 millisecond) of AEPs. Waves I, III and IV are generally considered as clinically significant [28]. Barrie W. Jervis et al. [29] had shown that the evidence of AEP was due to phase reordering and further contains additive energy in each harmonic component. This was established by employing angular statistics techniques.

The post stimulus latency and amplitudes of the defined peaks characterize the response of the ABR waveform. The significance of ABR waveform is that it does not require any attention or feedback response from the subject under test. ABR signals are used for diagnosis and localization of pathologies affecting brainstem pathways. ABR signals are widely employed to test auditory integrity and they are used as an important diagnostic tool in infant hearing screening test [30]. The change in amplitude and latency of ABR waveform occurs besides age, sex, test ear and vigilance of the subject [31-33]. Delgada et al. [28] have proposed the complete automated system for ABR response identification and waveform recognition. The analysis portion was divided into: (i) peak identification and labeling, (ii) ABR interpretation. When the threshold level was above 20 dB hearing loss, the subjects were flagged as having a form of hearing loss. When the peak intensity level was greater than 4.40 ms the subjects were flagged as having hearing pathologies. Hall et al. [34] reported that spectral compositions of ABR waveform contain valuable and unique prognostic information than their time domain counterparts. Arnaud jacquin et al. [35], concentrated on combining a signal adaptive denoising techniques based on complex wavelets with figure of signal quality (Fsp) to denoise Brainstem Auditory Evoked Response (BAER) signals. This proposed technique increases the detection rate of AEP response waves, which determines the presence of hearing loss of a patient in real time.

3.2. Clinical Importance of Peak V of ABR

Roberston et al. [19] described the evaluating alternation of techniques for estimating the spectrum of the ABR signal. The mid-frequency and high-frequency components of the ABR waves are most important for determining the latencies, and threshold estimation. Peak V has a primary importance in identifying the hearing level of a person, because this peak V clearly occurs only at the lowest sound stimulus intensity. Walker et al. [36] proposed a matched filter system to predict the peak V latency which is significant in estimating the hearing loss. Tapio et al. [37] proposed a method with multi-filters and an attributed automation system was used to identify the peak V. Strauss et al. [38] had proposed the fast detection of wave V in ABR using single sweep analysis with hybrid supervised system. This proposed work provides 100 % sensitivity and 90 % specificity in terms of identifying the peak V for normal hearing persons.

Andrew et al. [39], proposed that ABR is one of the most widely used auditory evoked potentials for determining thresholds. A method to estimate the ABR peak counts which is directly associated in determining the threshold level of a patient was proposed. The proposed algorithm finds the peaks of clinical interest specifically peak V, which is instrumental in discriminating normal subjects and hearing loss subjects on the basis of first and second order derivatives. A desired ABR peaks were extracted using a narrow band FIR filter with a zero phase shift. This proposed method demonstrates that neural network can be used to estimate latency of ABR peaks [40]. Rushaidin et al. [41] estimated the peak instantaneous energy of peak V of ABR wave as a feature and discriminated the normal and abnormal hearing persons on the basis of their derived threshold values. The normal ABR wave is shown in Fig. (1).

Fig. (1).

Normal ABR signal (presence of peak V visible).


Fig. (2).

AEP based hearing perception level detection system. HPL- Hearing perception level.


3.3. Expert Rule Based System Detection of Peak V

Robert Boston et al. [30], proposed an expert decision support system for interpretation of brainstem auditory evoked potential response. The prototype system consists of 36 rules. 13 rules were framed in order to find the presence of a neural response and 10 rules to identify a peak as peak V. This proposed rule based system was not as effective in identifying ABR waveform with no response was present. D. Alpsan et al. [42], proposed that Feedforward neural network was employed to detect the brainstem auditory evoked potential response and no response signals. The maximum classification of 75.6 % was reported in discriminating the “response” and “no response” classes. R. Sanchez et al. [43], extracted individual features set and combination of features set from brainstem auditory evoked potential (BAEP) were given as an input feature vectors to linear discriminate function and artificial neural networks. The maximum classification accuracy of 97.2 %, and 98.85 % was reported using linear discriminate functions and artificial neural networks, respectively. Edwige Vannier et al. [44] proposed a brainstem auditory evoked potential detection method in the time domain based on supervised pattern recognition. The pattern of normal BAEP was used based on cross correlation with a template. The accuracy of detecting the BAEP pattern was 90 % and determination of threshold level is with a mean error of 5 dB. Nurettin et al. [45], proposed automated recognition of ABR waveform to detect the hearing threshold of a person. In this method, amplitude values, discrete cosine transform coefficients, and discrete wavelet transform coefficients were extracted from the ABR waveform classified using support vector machine and a classification accuracy of 97.7 % was reported. Jose Antonio et al. [46] constructed and developed an EEG auditory evoked potential data acquisition system (EEG-ITM03) and used to determine the hearing impairment of a patient. The EEG- ITM03 can be used under non-noise controlled conditions. AEP data from their corresponding channels are acquired and processed the signal to identify the presence/absence of hypoacusia.

3.4. Complete Characteristic Analysis of Recorded AEP

The authors [47-51] addressed the problem by analyzing the complete waveform of AEPs instead of searching the defined peaks. The interpretation of wave (IV, V) is more complex and even small changes in the waveform can lead to large changes in their latency and amplitudes of wave IV and V. The recorded AEP signal contains the deterministic and the stochastic components. Researchers have focused their attention on the changes of the structure of AEP signal or characteristic of AEP signal in determining the hearing threshold level of a person. Shangkai et al. [47] had estimated the AR models for the AEP records of normal hearing and abnormal hearing persons. The estimated hearing thresholds by model parameters were very different dynamic structures from normal and abnormal persons with hearing loss problems. The estimated hearing threshold by parametric model holds good agreement with the audiologist assessment. Sudirman et al. [48] investigated EEG based hearing ability identification level using artificial intelligence. AEP signal was recorded for 10 seconds and collected from temporal lobes (T3, T4, T5, and T6) of the brain. The signals were analyzed using FFT. The extracted features were trained using gradient descendant algorithm with momentum. The feed-forward neural network was used to classify the hearing level based on brain signals. Emre et al. [8] used the continuous time wavelet entropy of auditory evoked potential and characterized by the relative energy in the EEG frequency bands. Continuous time wavelet gives detailed information of AEP response of brain which determines threshold level of a person.

Masumi Kogure et al. [49], in their experiment, used sound localization system in which subjects were requested to listen to a sound cue and answer in the relative direction. Direction of stimuli plays a vital role in terms of subject perceiving the sound intensity level. It has been concluded that the EEG signal can be used to elicit the sound source direction. Maryam Ravan et al. [50] demonstrated that machine learning algorithms can be used to classify individual subjects using auditory evoked potential. The wavelet coefficient features of the evoked potentials were selected using greedy algorithm. The extracted features were fed into various machine learning algorithms such as multilayer perceptron neural network, support vector, and fuzzy C means clustering [52]. Sriraam extracted appropriate features from the recorded auditory evoked potential signals to differentiate the hearing perception based on target stimulus and non-stimulus. Applied two time domain features, spike rhythmicity, autoregression Levinson method two frequency domain features, power spectral density (AR Burg method), power spectral density (Yule-Walker method) have a classification accuracy of 65.3-100 % for normal hearing subjects.

4. DISCUSSION

The reviewed research papers illustrate that hearing threshold of a person can be estimated by analyzing different types of AEP signals. EEG based hearing threshold level determination system is suitable for persons who lack verbal communication (newborns, infants) and muscle movements (neuro muscular disorder).

4.1. Hearing Perception Loss in Newborns

Recently, most of the countries have adapted universal newborns hearing screening test to screen newborns using EEG. The screening tests for the newborns are usually conducted using transient oto-acoustic emissions and auditory brainstem response. AEPs are much smaller than EEG signal in terms of amplitude. Auditory brainstem response comprises the early portion (0-12 msec) of AEPs. The first seven peaks of auditory brain response waves are labeled as peak I-VII. The peak waves I, III, V are generally considered to be clinically significant. The main characteristics of ABR are inter-peak latency and their amplitude. A level of confidence is associated with identification of peaks (I, III, V) in response to sound stimulation. Peak V is mainly used to identify the hearing threshold level of newborns.

4.2. Issues with Existing ABR Methods in Clinics

The detection of hearing responses at threshold levels of different frequencies and sound pressure levels were not examined by current existing methods that are being used in clinics to test newborns hearing threshold level. References of human experts are always required in difficult cases. Further, the accuracy of the threshold determination depends on the relevance of the stimulation intensities. Boo et al. [53], in their research work, have compared the sensitivity and specificity of detecting sensorineural hearing loss using AABR and Oto acoustic emission (OAE) in newborns. Based on their reports, AABR finds the advantage in detection of sensorineural hearing loss in newborns. Their study also showed that both methods were not optimal in detecting hearing loss for mass screening of newborns.

4.3. AEP Hearing Perception Level Estimation

AEP response reflects the auditory ability level of an individual. Stimulating a repetitive click-sound of specific frequency at different stimulus intensity levels (20 dB, 25 dB, 30 dB, 40 dB, 50 dB and 70 dB) was able to determine the hearing sound perception level of a person. The hearing perception level of a subject can be determined by selecting suitable features from the AEP response signal. With further investigation and interpretation of the AEP signal, it is feasible to discriminate the different hearing perception levels of a person. The block diagram of the proposed AEP based hearing perception detection system is shown in Fig. (2).

4.4. AEP based Frequency Specific Hearing Estimation

AEP response reflects the auditory ability level of an individual with respect to frequency and sound pressure level. Stimulating a repetitive click-sound of different frequency at fixed stimulus intensity level enables to determine the hearing frequency perception level of a person. This proposed method is able to detect person with low frequency hearing loss, mid-frequency hearing loss and high-frequency hearing loss.

In order to develop a successful common method to determine the auditory loss in all persons including multiple handicaps, and infants, number of issues need to be addressed:

(i) Establishing a standard hearing threshold protocol; (b) setting an appropriate hearing threshold criterion. Each of these issues, discussed below, present an opportunity for further research in the area.

The standard procedure for recording auditory evoked potentials can be formulated. This provides researchers to form a common platform to discuss the various AEP processing techniques and research findings with their counterparts. The important factors to be considered while recording the AEP signals are: 1) Electrodes and Electrode placement; 2) Type of stimulus, stimuli levels; 3) Stimuli frequency range; 4) Sound proof room; 5) Artifact and noise; 6) Selecting suitable features and machine learning algorithms. From the literature review, it has been observed that most of the researchers have employed a two channel EEG data acquisition system. Thus, it provides an opportunity to explore the potentially significant electrode locations emanating AEP signals. Based on the potentially significant areas, minimum number of electrode channels can be selected.

4.5. Hearing Threshold Criterion Setting

To distinguish the normal or hearing impaired state, one or more criteria are required. These criteria should yield both high sensitivity and specificity. The selected parameters should indicate the levels of hearing impaired from no impairment to profound impairment. Specific frequency with sound intensity level determines the objective threshold level. The issue of sex difference was neglected in almost all studies, because recorded AEPs does not show any marked differentiation between males and females. The multiple criteria may be necessary in predicting the specific type of hearing impairment based on AEPs for a single individual. The standalone intelligent hearing system based on EEG by incorporating octave band frequency can be designed and developed to determine the hearing threshold of a person. The efficient intelligent hearing level system may require the setting of multiple criteria and the use of multiple measures.

CONCLUSION

In this paper, we have reviewed the current state of the knowledge of auditory evoked potential used as a significant tool in determining the threshold levels of a person. To conduct AEP response test, a type of special attention or cooperation from the patients are not expected. AEPs are reliable indicators for hearing level of a person. AEPs can be used to detect and estimate the hearing levels from severe to profound hearing impairment. The AEP hearing test may also be used to cross-validate and examine to prevent persons who pretend as a normal hearing person in conventional pure tone test method for their social and economic benefits. AEPs signals can be analyzed in time-frequency domain in order to extract underlying detailed information about the auditory pathway. Ultimately, it appears that a standalone hearing level system based on EEG which can detect hearing loss for all including neonates, infants and multiple handicaps is needed essentially to help improve their quality of life.

CONFLICT OF INTEREST

The authors confirm that this article content has no conflict of interest.

ACKNOWLEDGEMENTS

The authors would like to thank Brigedier Jeneral Dato' Professor Dr. Kamarudin Hussin, the Vice Chancellor, University of Malaysia Perlis for his continuous support and encouragement. The authors are thankful and would like to acknowledge the Fundamental Research Grant Scheme (FRGS) grant: 9003-00278 by Ministry of Higher Education, Malaysia.

References

[1] International Day for Ear and Hearing, 2012.
[2] C. Mather, A. Smith, and M. Concha, "Global burden of hearing loss", In: C. Mather, and M.F. Doris, Eds., Global Burden of Diseases: 2004 Update.. WHO Press Geneva, 2008, pp. 1-30.
[3] R.S. Rao, M.A. Subramanyam, N.S. Nair, and B. Rajashekhar, "Hearing impairment and ear diseases among children of school entry age in rural South India", Int. J. Pediatr. Otorhinolaryngol., vol. 64, no. 2, pp. 105-110, 2002.
[4] B.O. Olusanya, L.M. Luxon, and S.L. Wirz, "Benefits and challenges of newborn hearing screening for developing countries", Int. J. Pediatr. Otorhinolaryngol., vol. 68, no. 3, pp. 287-305, 2004.
[5] H. Preissl, C.L. Lowery, and H. Eswaran, "Fetal magnetoencephalography: current progress and trends", Exp. Neurol., vol. 190, suppl. Suppl. 1, pp. S28-S36, 2004.
[6] S. Vanhatalo, and K. Kaila, "Development of neonatal EEG activity: from phenomenology to physiology", Semin. Fetal Neonatal Med., vol. 11, no. 6, pp. 471-478, 2006.
[7] T.D. Cummins, S. Finnigan, and J. Ros, "Theta power is reduced in healthy cognitive aging", Int. J. Psychophysiol., vol. 66, no. 1, pp. 10-17, 2007.
[8] M.E. Cek, M. Ozgoren, and F.A. Savaci, "Continuous time wavelet entropy of auditory evoked potentials", Comput. Biol. Med., vol. 40, no. 1, pp. 90-96, 2010.
[9] D.P. Subha, P.K. Joseph, R. Acharya U, and C.M. Lim, "EEG signal analysis: a survey", J. Med. Syst., vol. 34, no. 2, pp. 195-212, 2010.
[10] K.S. Luck, Introduction to the Event Related Potential Technique. 2005.
[11] C. Başar-Eroglu, V. Kolev, B. Ritter, F. Aksu, and E. Başar, "EEG, auditory evoked potentials and evoked rhythmicities in three-year-old children", Int. J. Neurosci., vol. 75, no. 3-4, pp. 239-255, 1994.
[12] S. Sokol, "Visually evoked potentials: theory, techniques and clinical applications", Surv. Ophthalmol., vol. 21, no. 1, pp. 18-44, 1976.
[13] Y.H. Shih, Z.J. Huang, and C.E. Chang, "Color pattern-reversal visual evoked potential in eyes with ocular hypertension and primary open-angle glaucoma", Doc. Ophthalmol., vol. 77, no. 3, pp. 193-200, 1991.
[14] D. Regan, and D. Neima, "Visual fatigue and visual evoked potentials in multiple sclerosis, glaucoma, ocular hypertension and Parkinson’s disease", J. Neurol. Neurosurg. Psychiatry, vol. 47, no. 7, pp. 673-678, 1984.
[15] T.W. Picton, Human Auditory Evoked Potentials.. Plural Publishing: USA, 2010.
[16] T.W. Picton, M.S. John, D.W. Purcell, and G. Plourde, "Human auditory steady-state responses: the effects of recording technique and state of arousal", Anesth. Analg., vol. 97, no. 5, pp. 1396-1402, 2003.
[17] G. Plourde, "Auditory evoked potentials", Best Pract. Res. Clin. Anaesthesiol., vol. 20, no. 1, pp. 129-139, 2006.
[18] T.W. Picton, and S.A. Hillyard, "Human auditory evoked potentials. II. Effects of attention", Electroencephalogr. Clin. Neurophysiol., vol. 36, no. 2, pp. 191-199, 1974.
[19] J.R. Boston, "Spectra of auditory brainstem responses and spontaneous EEG", IEEE Trans. Biomed. Eng., vol. 28, no. 4, pp. 334-341, 1981.
[20] X.S. Zhang, R.J. Roy, D. Schwender, and M. Daunderer, "Discrimination of anesthetic states using midlatency auditory evoked potential and artificial neural networks", Proceeding of IEEE International Conference on Engineering in Medicine and Biology Society, 2000pp. 1383-86
[21] M.I. Mendel, and R. Goldstein, "Stability of the early components of the averaged electroencephalic response", J. Speech Hear. Res., vol. 12, no. 2, pp. 351-361, 1969.
[22] R.T. Wilkinson, and H.C. Morlock, "Auditory evoked response and reaction time", Electroencephalogr. Clin. Neurophysiol., vol. 23, no. 1, pp. 50-56, 1967.
[23] R. Näätänen, P. Paavilainen, T. Rinne, and K. Alho, "The mismatch negativity (MMN) in basic research of central auditory processing: a review", Clin. Neurophysiol., vol. 118, no. 12, pp. 2544-2590, 2007.
[24] R. Näätänen, and I. Winkler, "The concept of auditory stimulus representation in cognitive neuroscience", Psychol. Bull., vol. 125, no. 6, pp. 826-859, 1999.
[25] P. Korpilahti, Electrophysiological correlates of auditory perception in normal and language impaired children, 1996.
[26] P.H. Leppanen, K. Eklund, and M. Lyytinen, "Event related brain potentials to change in rapidly presented acoustic stimuli in newborns", J. Develop. Neurophysiol., vol. 13, pp. 175-04, 1997.
[27] J.R. Copeniene, E. Kushneremko, V. Fellman, M. Renlund, K. Suominen, and R. Noatanen, "Event related potential features indexing central auditory discrimination by newborns", J. Cogn. Brain Res., vol. 13, pp. 107-113, 2002.
[28] E. Delgada, and O. Ozdamar, "Automated auditory brainstem response interpretation", IEEE Eng. Med. Biol., vol. 13, pp. 227-237, 1994.
[29] B.W. Jervis, M.J. Nichols, T.E. Johnson, E. Allen, and N.R. Hudson, "A fundamental investigation of the composition of auditory evoked potentials", IEEE Trans. Biomed. Eng., vol. 30, no. 1, pp. 43-50, 1983.
[30] J.R. Boston, "Automated interpretation of brainstem auditory evoked potentials: a prototype system", IEEE Trans. Biomed. Eng., vol. 36, no. 5, pp. 528-532, 1989.
[31] W.J. Wilson, and F. Aghdasi, "Fast Fourier transform analysis of the auditory brainstem response: Effects of stimulus intensity and subject age, gender, test Ear", Proceeding of IEEE International Conference on Africon, vol. vol. 1, 1999pp. 285-90
[32] W.J. Wilson, and F. Aghdasi, "Discrete wavelet transform analysis of the auditory brainstem response: Effects of stimulus Intensity and subject age, gender, Test Ear", Proceeding of IEEE International Conference on Africon, vol. vol. 1, 1999pp. 291-96
[33] U. Hoppe, S. Weiss, R.W. Stewart, and U. Eysholdt, "An automatic sequential recognition method for cortical auditory evoked potentials", IEEE Trans. Biomed. Eng., vol. 48, no. 2, pp. 154-164, 2001.
[34] J.W. Hall III, "Auditory brain stem response spectral content in comatose head-injured patients", Ear Hear., vol. 7, no. 6, pp. 383-389, 1986.
[35] E.C. Jacquin, and E.R. John, "Optimal denoising of brainstem auditory evoked response for automatic peak identification and brainstem assessment", Proceeding of IEEE International Conference on Engineering in Medicine and Biology Society, vol. vol. 1, 2006pp. 1723-126
[36] W. Woodworth, S. Reisman, and A.B. Fontaine, "The detection of auditory evoked responses using a matched filter", IEEE Trans. Biomed. Eng., vol. 30, no. 7, pp. 369-376, 1983.
[37] T. Grönfors, "Peak identification of auditory brainstem responses with multi-filters and attributed automaton", Comput. Methods Programs Biomed., vol. 40, no. 2, pp. 83-87, 1993.
[38] D.J. Strauss, W. Delb, P.K. Plinkert, and H. Schmidt, "Fast detection of wave V in ABRs using a smart single sweep analysis system", Proceeding of IEEE International Conference on Engineering in Medicine and Biology Society, vol. vol. 1, 2004pp. 458-61
[39] P. Bradley, and W.J. Wilson, "Automated analysis of the auditory brainstem response", Proceeding of IEEE International Conference on Intelligent Sensors, Sensor Networks and Information Processing, 2004pp. 541-45
[40] J. Tian, M. Juhola, and T. Grönfors, "Latency estimation of auditory brainstem response by neural networks", Artif. Intell. Med., vol. 10, no. 2, pp. 115-128, 1997.
[41] M.M. Rushaidin, "Sh-Hussain Salleh, Tan Tian Swee, J. M. Najeb and A. Arooj, “Wave V detection using instantaneous energy of auditory brainstem response signal", Am. J. Appl. Sci., vol. 6, pp. 1669-1674, 2009.
[42] D. Alpsan, and O. Ozdamar, "Brainstem auditory evoked potential classification by backpropagation networks", Proceeding of IEEE International Joint Conference on Neural Networks, vol. vol. 2, 1991pp. 1266-1271
[43] R. Sánchez, A. Riquenes, and M. Pérez-Abalo, "Automatic detection of auditory brainstem responses using feature vectors", Int. J. Biomed. Comput., vol. 39, no. 3, pp. 287-297, 1995.
[44] E. Vannier, O. Adam, and J.F. Motsch, "Objective detection of brainstem auditory evoked potentials with a priori information from higher presentation levels", Artif. Intell. Med., vol. 25, no. 3, pp. 283-301, 2002.
[45] N. Acir, O. Ozdamar, and C. Guzelis, "Automatic classification of auditory brainstem responses using SVM-based feature selection algorithm for threshold detection", Eng. Appl. Artif. Intell., vol. 19, pp. 209-218, 2006.
[46] J.A. Gnecchi, and L.R. Lara, "Design and construction of an EEG data acquisition system for measurement of auditory evoked potential", Proceeding of IEEE International Conference on Electronics, Robotics and Automotive Mechanics, 2008pp. 547-552
[47] S.K. Gao, and M.H. Loew, "An autoregressive model of the BAEP signal for hearing-threshold testing", IEEE Trans. Biomed. Eng., vol. 33, no. 6, pp. 560-565, 1986.
[48] R. Sudirman, and S.C. Seow, "Electroencephalographic based hearing identification using back propagation algorithm", Proceeding of IEEE International Conference on Science and technology for Humanity, 2009pp. 991-95
[49] M. Kogure, S. Matszaki, and Y. Wada, "Investigation of brain computer interface that apply sound evoked event related potentials",
[50] M. Ravan, J.P. Reilly, L.J. Trainor, and A. Khodayari-Rostamabad, "A machine learning approach for distinguishing age of infants using auditory evoked potentials", Clin. Neurophysiol., vol. 122, no. 11, pp. 2139-2150, 2011.
[51] N. Sriraam, "EEG based automated detection of auditory loss: A pilot study", Expert Syst. Appl., vol. 39, pp. 723-731, 2012.
[52] N.Y. Boo, A.J. Rohani, and A. Asma, "Detection of sensorineural hearing loss using automated auditory brainstem-evoked response and transient-evoked otoacoustic emission in term neonates with severe hyperbilirubinaemia", Singapore Med. J., vol. 49, no. 3, pp. 209-214, 2008.