Breath Detection from a Microphone Using Machine Learning
This project investigates and implements various artificial intelligence techniques for the real-time detection of breath sounds using audio data captured via a computer microphone. The primary objective is to develop and compare methodologies to identify distinct phases of breathing, namely inhala...
Saved in:
Main Authors: | , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
Gdańsk University of Technology
2025-01-01
|
Series: | TASK Quarterly |
Subjects: | |
Online Access: | https://journal.mostwiedzy.pl/TASKQuarterly/article/view/3381 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
_version_ | 1841553370358218752 |
---|---|
author | Tomasz Sankowski Piotr Sulewski Jan Walczak Aleksandra Bruska |
author_facet | Tomasz Sankowski Piotr Sulewski Jan Walczak Aleksandra Bruska |
author_sort | Tomasz Sankowski |
collection | DOAJ |
description |
This project investigates and implements various artificial intelligence techniques for the real-time detection of breath sounds using audio data captured via a computer microphone. The primary objective is to develop and compare methodologies to identify distinct phases of breathing, namely inhalation, exhalation, and the silent intervals between breaths, in order to determine the most accurate, efficient, and practical approach.
The study explores three innovative approaches:
1. VGGish Model for Feature Extraction and Classification with Random Forest: This method utilizes the VGGish model to extract sound feature vectors, followed by classification using a random forest classifier.
2. Spectrogram Classification Using Convolutional Neural Networks: This approach involves classifying spectrograms of half-second or quarter-second audio segments using a convolutional neural network adapted for image classification tasks.
3. Mel-Frequency Cepstral Coefficients (MFCC) for Feature Extraction and Neural Network Classification: This method employs MFCCs as set of sound features for classification using a neural network.
The experimental results show that methods 1 and 3 achieved an accuracy of 87% in the test data, while method 2 achieved an accuracy of 83%. The dataset comprised approximately 1,000 recordings of inhalations, exhalations, and
silences between breaths, collected using four different microphones and recorded by three different individuals.
All implementations and training data are available on a public GitHub repository: github.com/tomaszsankowski/Breathing-
Classification.
|
format | Article |
id | doaj-art-448d6cbd140d4961b182e85b435781c9 |
institution | Kabale University |
issn | 1428-6394 |
language | English |
publishDate | 2025-01-01 |
publisher | Gdańsk University of Technology |
record_format | Article |
series | TASK Quarterly |
spelling | doaj-art-448d6cbd140d4961b182e85b435781c92025-01-09T10:23:55ZengGdańsk University of TechnologyTASK Quarterly1428-63942025-01-01272Breath Detection from a Microphone Using Machine LearningTomasz Sankowski0Piotr Sulewski1Jan Walczak2Aleksandra Bruska3Gdańsk University of Technology, Faculty of Electronics, Telecommunications and InformaticsGdańsk University of Technology, Faculty of Electronics, Telecommunications and InformaticsGdańsk University of Technology, Faculty of Electronics, Telecommunications and InformaticsGdańsk University of Technology, Faculty of Electronics, Telecommunications and Informatics This project investigates and implements various artificial intelligence techniques for the real-time detection of breath sounds using audio data captured via a computer microphone. The primary objective is to develop and compare methodologies to identify distinct phases of breathing, namely inhalation, exhalation, and the silent intervals between breaths, in order to determine the most accurate, efficient, and practical approach. The study explores three innovative approaches: 1. VGGish Model for Feature Extraction and Classification with Random Forest: This method utilizes the VGGish model to extract sound feature vectors, followed by classification using a random forest classifier. 2. Spectrogram Classification Using Convolutional Neural Networks: This approach involves classifying spectrograms of half-second or quarter-second audio segments using a convolutional neural network adapted for image classification tasks. 3. Mel-Frequency Cepstral Coefficients (MFCC) for Feature Extraction and Neural Network Classification: This method employs MFCCs as set of sound features for classification using a neural network. The experimental results show that methods 1 and 3 achieved an accuracy of 87% in the test data, while method 2 achieved an accuracy of 83%. The dataset comprised approximately 1,000 recordings of inhalations, exhalations, and silences between breaths, collected using four different microphones and recorded by three different individuals. All implementations and training data are available on a public GitHub repository: github.com/tomaszsankowski/Breathing- Classification. https://journal.mostwiedzy.pl/TASKQuarterly/article/view/3381audio classificationbreath detectionbreath analysis |
spellingShingle | Tomasz Sankowski Piotr Sulewski Jan Walczak Aleksandra Bruska Breath Detection from a Microphone Using Machine Learning TASK Quarterly audio classification breath detection breath analysis |
title | Breath Detection from a Microphone Using Machine Learning |
title_full | Breath Detection from a Microphone Using Machine Learning |
title_fullStr | Breath Detection from a Microphone Using Machine Learning |
title_full_unstemmed | Breath Detection from a Microphone Using Machine Learning |
title_short | Breath Detection from a Microphone Using Machine Learning |
title_sort | breath detection from a microphone using machine learning |
topic | audio classification breath detection breath analysis |
url | https://journal.mostwiedzy.pl/TASKQuarterly/article/view/3381 |
work_keys_str_mv | AT tomaszsankowski breathdetectionfromamicrophoneusingmachinelearning AT piotrsulewski breathdetectionfromamicrophoneusingmachinelearning AT janwalczak breathdetectionfromamicrophoneusingmachinelearning AT aleksandrabruska breathdetectionfromamicrophoneusingmachinelearning |