Leigh Anne Clevenger Pace University, DPS ’16

Slides:



Advertisements
Similar presentations
Voiceprint System Development Design, implement, test unique voiceprint biometric system Research Day Presentation, May 3 rd 2013 Rahul Raj (Team Lead),
Advertisements

Masters Presentation at Griffith University Master of Computer and Information Engineering Magnus Nilsson
Franz de Leon, Kirk Martinez Web and Internet Science Group  School of Electronics and Computer Science  University of Southampton {fadl1d09,
Improvement of Audio Capture in Handheld Devices through Digital Filtering Problem Microphones in handheld devices are of low quality to reduce cost. This.
Department of Computer Science University of California, San Diego
Introduction The aim the project is to analyse non real time EEG (Electroencephalogram) signal using different mathematical models in Matlab to predict.
Speaker Recognition Sharat.S.Chikkerur Center for Unified Biometrics and Sensors
Content-Based Classification, Search & Retrieval of Audio Erling Wold, Thom Blum, Douglas Keislar, James Wheaton Presented By: Adelle C. Knight.
Toward Semantic Indexing and Retrieval Using Hierarchical Audio Models Wei-Ta Chu, Wen-Huang Cheng, Jane Yung-Jen Hsu and Ja-LingWu Multimedia Systems,
SIGNAL PROCESSING TECHNIQUES USED FOR THE ANALYSIS OF ACOUSTIC SIGNALS FROM HEART AND LUNGS TO DETECT PULMONARY EDEMA 1 Pratibha Sharma Electrical, Computer.
1 Speech Parametrisation Compact encoding of information in speech Accentuates important info –Attempts to eliminate irrelevant information Accentuates.
LYU0103 Speech Recognition Techniques for Digital Video Library Supervisor : Prof Michael R. Lyu Students: Gao Zheng Hong Lei Mo.
Speaker Adaptation for Vowel Classification
A PRESENTATION BY SHAMALEE DESHPANDE
Authors: Anastasis Kounoudes, Anixi Antonakoudi, Vasilis Kekatos
Advisor: Prof. Tony Jebara
Representing Acoustic Information
Self-Calibrating Audio Signal Equalization Greg Burns Wade Lindsey Kevin McLanahan Jack Samet.
Audio Retrieval David Kauchak cs458 Fall Administrative Assignment 4 Two parts Midterm Average:52.8 Median:52 High:57 In-class “quiz”: 11/13.
EE513 Audio Signals and Systems Statistical Pattern Classification Kevin D. Donohue Electrical and Computer Engineering University of Kentucky.
GCT731 Fall 2014 Topics in Music Technology - Music Information Retrieval Overview of MIR Systems Audio and Music Representations (Part 1) 1.
Eng. Shady Yehia El-Mashad
Age and Gender Classification using Modulation Cepstrum Jitendra Ajmera (presented by Christian Müller) Speaker Odyssey 2008.
SoundSense by Andrius Andrijauskas. Introduction  Today’s mobile phones come with various embedded sensors such as GPS, WiFi, compass, etc.  Arguably,
Classification of place of articulation in unvoiced stops with spectro-temporal surface modeling V. Karjigi , P. Rao Dept. of Electrical Engineering,
VBS Documentation and Implementation The full standard initiative is located at Quick description Standard manual.
Implementing a Speech Recognition System on a GPU using CUDA
Evaluation of Speaker Recognition Algorithms. Speaker Recognition Speech Recognition and Speaker Recognition speaker recognition performance is dependent.
Jacob Zurasky ECE5526 – Spring 2011
Supervisor: Dr. Eddie Jones Co-supervisor: Dr Martin Glavin Electronic Engineering Department Final Year Project 2008/09 Development of a Speaker Recognition/Verification.
Multimodal Information Analysis for Emotion Recognition
Signature with Text-Dependent and Text-Independent Speech for Robust Identity Verification B. Ly-Van*, R. Blouet**, S. Renouard** S. Garcia-Salicetti*,
Jun-Won Suh Intelligent Electronic Systems Human and Systems Engineering Department of Electrical and Computer Engineering Speaker Verification System.
Chapter 5: Speech Recognition An example of a speech recognition system Speech recognition techniques Ch5., v.5b1.
Speaker Recognition by Habib ur Rehman Abdul Basit CENTER FOR ADVANCED STUDIES IN ENGINERING Digital Signal Processing ( Term Project )
Speech Recognition Feature Extraction. Speech recognition simplified block diagram Speech Capture Speech Capture Feature Extraction Feature Extraction.
Look who’s talking? Project 3.1 Yannick Thimister Han van Venrooij Bob Verlinden Project DKE Maastricht University.
Speaker Identification by Combining MFCC and Phase Information Longbiao Wang (Nagaoka University of Technologyh, Japan) Seiichi Nakagawa (Toyohashi University.
Singer Similarity Doug Van Nort MUMT 611. Goal Determine Singer / Vocalist based on extracted features of audio signal Classify audio files based on singer.
RCC-Mean Subtraction Robust Feature and Compare Various Feature based Methods for Robust Speech Recognition in presence of Telephone Noise Amin Fazel Sharif.
Chapter 7 Speech Recognition Framework  7.1 The main form and application of speech recognition  7.2 The main factors of speech recognition  7.3 The.
1 Electrical and Computer Engineering Binghamton University, State University of New York Electrical and Computer Engineering Binghamton University, State.
ADAPTIVE BABY MONITORING SYSTEM Team 56 Michael Qiu, Luis Ramirez, Yueyang Lin ECE 445 Senior Design May 3, 2016.
BIOMETRICS VOICE RECOGNITION. Meaning Bios : LifeMetron : Measure Bios : LifeMetron : Measure Biometrics are used to identify the input sample when compared.
Speech Processing Dr. Veton Këpuska, FIT Jacob Zurasky, FIT.
Emotional Intelligence Vivian Tseng, Matt Palmer, Jonathan Fouk Group #41.
CS 445/656 Computer & New Media
Heart Sound Biometrics for Continual User Authentication
Speech Processing AEGIS RET All-Hands Meeting
ARTIFICIAL NEURAL NETWORKS
Speech Processing AEGIS RET All-Hands Meeting
Spoken Digit Recognition
Presentation on Artificial Neural Network Based Pathological Voice Classification Using MFCC Features Presenter: Subash Chandra Pakhrin 072MSI616 MSC in.
Biometric Classification of Heart Sound for Active User Authentication and Clinical Applications Leigh Anne H. Clevenger Pace University, Computer Science.
Sharat.S.Chikkerur S.Anand Mantravadi Rajeev.K.Srinivasan
3. Applications to Speaker Verification
Wearable Sensors: Creating Research Projects for New Technologies
Neuro-Fuzzy and Soft Computing for Speaker Recognition (語者辨識)
Ala’a Spaih Abeer Abu-Hantash Directed by Dr.Allam Mousa
EE513 Audio Signals and Systems
Audio and Speech Computers & New Media.
Digital Systems: Hardware Organization and Design
AUDIO SURVEILLANCE SYSTEMS: SUSPICIOUS SOUND RECOGNITION
Presenter: Simon de Leon Date: March 2, 2006 Course: MUMT611
John H.L. Hansen & Taufiq Al Babba Hasan
Presenter: Shih-Hsiang(士翔)
Measuring the Similarity of Rhythmic Patterns
Keyword Spotting Dynamic Time Warping
Combination of Feature and Channel Compensation (1/2)
Practical Hidden Voice Attacks against Speech and Speaker Recognition Systems NDSS 2019 Hadi Abdullah, Washington Garcia, Christian Peeters, Patrick.
Presentation transcript:

Biometric Classification of Heart Sound for Continual User Authentication and Clinical Applications Leigh Anne Clevenger Pace University, DPS ’16 leighanne.clevenger@gmail.com Advisor: Dr. Charles C. Tappert

Heart Sound Biometrics - Mobile Device User Authentication Time = 0; Authenticate? Yes No No Access Time = 0+ to 30 min; Authenticate? Yes No 11/13/2018 L.A. Clevenger No Access – return to fingerprint authentication

Heart Sound Biometrics – Mobile Device Clinical Applications Time = 0; Authenticate? Yes No No Notification Sent Time = 0+ to 30 min; Authenticate? Notify Cardiologist or 911 Yes No 11/13/2018 L.A. Clevenger

Diagram of Heart and Typical Waveform https://www.youtube.com/watch?v=eBIWoLNzngI http://www.heart-valve-surgery.com/heart-surgery-blog/2010/05/30/video-tricuspid-valve-repair/ 11/13/2018 L.A. Clevenger

Phao, 2007: Heart Sound as a Biometric First paper in this research area, published in Pattern Recognition User authentication and identification accuracy rates 96% - 99%, 1000 data samples on their own dataset Welch Allyn Meditron electronic stethoscope, sampling rate of 2 KHz and 16 bits. 0.5 sec windows 11/13/2018 L.A. Clevenger

Spadaccini and Beritelli, 2007-2013: HSCT11 Heart Sound Evaluation Spadaccini and Beritelli provide a 200 person, two recordings each, public database of normal heart sounds (HSCT11), and use the S1 and S2 features for user authentication Biometric Analysis Modality Signal split into 4 second windows S1 and S2 identified, then window endpoints are connected Apply Chirp z-Transform (CZT), allowing high resolution analysis of narrow frequency bands Features extracted by Mel-Frequency Cepstral Coefficients (MFCC). Added to the feature vector First-to-Second Ratio (FSR) as the average power ratio of S1 to S2. Structural and Statistical Classification Structural – distance between input signal and stored template, using cepstral distance and an amplifying factor. Equal Error Rate (EER) of 36.86% Statistically – uses Gaussian Mixture Model – Universal Background Model (GMM-UBM) recognition technique. EER of 13.66% 11/13/2018 L.A. Clevenger

Spadaccini and Beritelli: HSCT11 Heart Sound Evaluation 11/13/2018 L.A. Clevenger

Zhao, 2013: Marginal Spectrum Analysis-based System Zhao et al use their own data, plus the HSCT11 dataset, for user authentication Database – 40 participants, 280 heart sounds. 0.256 sec windows Results Marginal Spectrum Analysis (MSA) – recognition rate of 94% Fourier spectrum (traditional algorithm) – recognition rate of 84% MSA on HSCT11 database – recognition rate of 92% 11/13/2018 L.A. Clevenger

Heart Sound Biometric Analysis Select waveforms from database Preprocess to find S1-S2 pairs; Apply noise reduction if needed Create data frames Generate Feature Vectors for each frame Calculate matching score Train and test the system, and evaluate classifier results 11/13/2018 L.A. Clevenger

Pre-processing to improve authentication accuracy Spadaccini split input signal into time windows S1 and S2 identified, placed in separate time windows, window endpoints are connected Chirp z-Transform (CZT) is applied, allowing high resolution analysis of narrow frequency bands Zhao, Sun, Abo-Zahhad De-noise using selected wavelets: improved Empirical Mode Decomposition (EMD) wavelet algorithm and the Shannon energy envelope algorithm, Discrete Wavelet Transformation (DWT) 11/13/2018 L.A. Clevenger

FFT and Cepstrum Generation 11/13/2018 L.A. Clevenger

Mel Frequency Cepstral Coefficient Generation Mel filterbank: maps frequency to mel scale – linear below 1kHz, log above Cepstrum is spectrum of log of the mel spectrum Result is 12 cepstral coefficients per frame, and 1 value for energy – the sum over time of the power of the samples in the frame Additional data for improved authentication accuracy can be appended to the MFCC feature vector: FSR, wave average, wave average differences, etc 11/13/2018 L.A. Clevenger

Innovative Technologies supporting Heart Sound Biometrics 11/13/2018 L.A. Clevenger

Demo of Heart Sound Input Visualization and audio using Audacity ThinkLabs Digital Stethoscope, HSCT11 Data iStethoscope Pro iPhone app, Heart Challenge A Data 3M Littmann Digital Stethoscope, Medscape Data 11/13/2018 L.A. Clevenger

Research Plans – Next Steps Investigate literature and current regulations for cardiac clinical applications scenarios, to identify opportunities and challenges Continue to create features for heart sound, using clinical features of 1st and 2nd beats, and additional features such as First-to-Second beat ratio. Use the HSCT-11 public heart sound database and scilab scripts. Incorporate methodologies from related research Implement machine learning algorithms for training and testing the system 11/13/2018 L.A. Clevenger

Backup 11/13/2018 L.A. Clevenger

Fundamentals of heart sound Heart function Is nonlinear Is periodic Has time-varying characteristics Heart rate variability (HRV) is non-stationary – it is effected by breathing, variation in blood flow (cardic-hemodynamics), and the environment Has randomness: palpitations - skipped or added beats; arrhythmia – change in beating patterns Typical heart sound measurement facts Range from 0-600 Hz S1 - First heart sound between 50-150 Hz, closure of the tricuspid and mitral valves S2 - Second heart sound between 50-200 Hz, closure of the aortic and pulmonary valves Second peak 250-350 Hz Research for user authentication often ignores 3rd and 4th heart sound as “murmur” In research, sampling frequencies used for measurement vary, some examples are 6400 Hz and 11025 Hz 11/13/2018 L.A. Clevenger

MFCC Generation Widely used for speech and speaker recognition Most often used feature extraction algorithm in heart sound biometric literature Results in 13 element feature vector Generate using *.wav files from Beritelli database and m scripts with scilab (free version of matlab) Scripts from http://www.ee.columbia.edu/~dpwe/resources/matlab/rastamat Resources D. Jurafsky, J. H. Martin, “Speech and Language Processing” http://practicalcryptography.com/miscellaneous/machine-learning/guide-mel-frequency-cepstral-coefficients-mfccs/ 11/13/2018 L.A. Clevenger

Public Heartsound Databases and Digital Stethoscope Sensors #signals recording length (s) sensor used frequency HSCT-11 [16,17] 206 60 ThinkLabs 11025 Hz, 16 bits per sample WelchAllyn [18] 150 30 Not available 16 kHz Heartchallenge A [19] 176 1 to 30 iStethoscope Pro 30 kHz to 44 kHz Heartchallenge B [19] 656 DigiScope 4 kHz Primer [20] 23 8 kHz to 44 kHz Texasheart [21] 22 8 kHz Medscape [22] 7 Littmann Stethoscope 10 kHz Stethoscope sensor Reference ThinkLabs Rhythm Digital Electronic Stethoscope http://www.thinklabs.com/   iStethoscope Pro iOS app https://itunes.apple.com/us/app/istethoscope-pro/id322110006?mt=8 DigiScope http://digiscope.diagnozit.com/ 3M Littmann Electronic Stethoscope Model 3200 http://www.3m.com/3M/en_US/company-us/all-3m-products/~/3M-Littmann-Electronic-Stethoscope-Model-3200?N=5002385+8707795+8707798+8711017+8711096+8711500+8711724+8727094+3293188392&rt=rud 11/13/2018 L.A. Clevenger