Datasets
Standard Dataset
Respiratory and Drug Actuation Dataset
- Citation Author(s):
- Submitted by:
- Stavros Nousias
- Last updated:
- Fri, 06/17/2022 - 04:50
- DOI:
- 10.21227/g84h-ma24
- Data Format:
- Links:
- License:
Abstract
Asthma is a common, usually long-term respiratory disease with negative impact on society and the economy worldwide. Treatment involves using medical devices (inhalers) that distribute medicationto the airways, and its efficiency depends on the precision of the inhalation technique. Health monitoring systems equipped with sensors and embedded with sound signal detection enable the recognition of drug actuation and could be powerful tools for reliable audio content analysis. The RDA Suite includes a set of tools for audio processing, feature extraction and classification and is provided along with a dataset consisting of respiratory and drug actuation sounds. The classification models in RDA are implemented based on conventional and advanced machine learning and deep network architectures. This study provides a comparative evaluation of the implemented approaches, examines potential improvements and discusses challenges and future tendencies. The central aim of this research is to identify associations between high-level classification labels and low-level features extracted from audio clips of different semantic activities. We investigate the clinical applicability of different audio-based signal processing methods for assessing medication adherence. The dataset consists of recordings acquired in anacoustically controlled setting, free of ambient indoor environment noise, at the University of Patras. Three subjects, who were familiar with the inhaler technique, participated in the study. The participants were instructed to use the inhaler, as typically performed in a clinical procedure. Foreach and every participant informed consent was obtained.During breathing and drug actuation, the audio signals were acquired by a microphone attached to the inhalation de-vice, communicating with a mobile phone via Bluetooth.The addition of the adherence monitoring device did not impact the normal functioning of the inhaler, which had afull placebo canister. In total, 370 audio files were recorded with a different duration each, containing an entire inhaleruse case, with respiratory flow ranging on 180-240 L/min.Each audio recording was sampled with a 8KHz sampling frequency, as a mono channel WAV file, at 8-bit depth.The audio recordings were segmented and annotated by a human specialist into inhaler actuation, exhalation, inhalationand environmental noise. The obtained segments (of non-mixed states) were of variable length and, for some methods, were further segmented into frames of fixed length for the purposes of feature extraction.The constructed database overall consisted of 193 drug actuation segments, 319 inhalation and 620 exhalation segments and 505 noise segments, ready to be used for audio sound recognition using different sets of features.
The central aim of this research is to identify associations between high-level classification labels and low-level features extracted from audio clips of different semantic activities. We investigate the clinical applicability of different audio-based signal processing methods for assessing medication adherence. The dataset consists of recordings acquired in anacoustically controlled setting, free of ambient indoor environment noise, at the University of Patras. Three subjects, who were familiar with the inhaler technique, participated in the study. The participants were instructed to use the inhaler, as typically performed in a clinical procedure. Foreach and every participant informed consent was obtained.During breathing and drug actuation, the audio signals were acquired by a microphone attached to the inhalation de-vice, communicating with a mobile phone via Bluetooth.The addition of the adherence monitoring device did not impact the normal functioning of the inhaler, which had afull placebo canister. In total, 370 audio files were recorded with a different duration each, containing an entire inhaleruse case, with respiratory flow ranging on 180-240 L/min.Each audio recording was sampled with a 8KHz sampling frequency, as a mono channel WAV file, at 8-bit depth.The audio recordings were segmented and annotated by a human specialist into inhaler actuation, exhalation, inhalationand environmental noise. The obtained segments (of non-mixed states) were of variable length and, for some methods, were further segmented into frames of fixed length for the purposes of feature extraction.The constructed database overall consisted of 193 drug actuation segments, 319 inhalation and 620 exhalation segments and 505 noise segments, ready to be used for audio sound recognition using different sets of features.
Generic format:
Filename, Class, Sample index at the beginning of the acoustic event, Sample index at the end of the acoustic event
Example:
- rec2018-01-22_17h41m33.475s.wav,Exhale,6015,17437
- rec2018-01-22_17h41m33.475s.wav,Inhale,20840,31655
- rec2018-01-22_17h41m33.475s.wav,Drug,31898,37610
- rec2018-01-22_17h41m33.475s.wav,Exhale,43686,59969
- rec2018-01-22_17h41m49.809s.wav,Inhale,5043,17316
- rec2018-01-22_17h41m49.809s.wav,Drug,18288,24364
- rec2018-01-22_17h41m49.809s.wav,Exhale,31412,46724
- rec2018-01-22_17h42m07.718s.wav,Exhale,303,9782
- rec2018-01-22_17h42m07.718s.wav,Inhale,16951,28010
Training and testing machine learning models with 10-fold cross validation:
python evaluate_ml_models.py
Training and testing LSTM models with 10-fold cross validation:
python evaluate_lstm.py
Training and testing CNN models with 10-fold cross validation:
python evaluate_cnn.py
Dataset Files
- Contains the dataset data.zip (149.82 MB)
- Contains the scripts required to evaluate the dataset scripts.zip (7.62 MB)