Automated classification of cough sounds has increased in importance due to factors such as the worldwide COVID-19 pandemic. To train such classification models requires a large dataset of cough sounds; however, it remains challenging to find sufficient expert-labelled training data. This thesis explores a novel form of audio data augmentation, where training cough sounds are corrupted with varying levels of reverberation and Gaussian noise. The combination of noise and reverberation is more effective than both traditional image-based augmentation techniques and either noise or reverberation alone, leading to near-human accuracy on a wet vs. dry cough classification task using a ResNet18 model across two cough datasets. Alignment between the training and testing environments is examined using the Speech-Commands audio dataset. While models trained with the closest reverb and noise level to the test environment gave the best results, the proposed audio augmentation technique produces models with robust performance across test environments.Finally, I would like to thank my family -your unconditional love and encouragement has been the fuel that powered me through this experience. To my wife, Fariha, thank you for your love and patience, especially the last months of the program. To my sister and family, appi (Rumpa), cb (Tinku), Rehan & Faraz, thank you for always being there to cheer me up and encourage me every time I needed it. To my mother and father, thank you for your love, kindness, patience, and your belief in me.v
Table of ContentsCopyright .