Digital media has a massive presence in the modern world, and it significantly impacts kids’ intellectual, cognitive, ethical, and social development. It is nearly impossible to isolate kids from digital media. Therefore, adult content on mobile applications should be avoided by children. Although mobile operating systems provide parental controls, handling such rules is impossible for illiterate people. Consequently, kids may download and use adults’ mobile applications. Mobile applications for adults often publish age group information to distinguish user segments that can be used to automate the downloading process. Sustainable Development Goal (SDG) #4 emphasizes inclusivity and equitability in terms of quality of education and the facilitation of conditions for the promotion of lifelong learning for everyone. The current study can be counted as being in line with SDG#4, as it proposes a machine-learning-based approach to the prediction of the suitability of mobile applications for kids. The approach first leverages natural language processing (NLP) techniques to preprocess user reviews of mobile applications. Second, it performs feature engineering based on the given bag of words (BOW), e.g., abusive words, and constructs a feature vector for each mobile app. Finally, it trains and tests a machine learning algorithm on the given feature vectors. To evaluate the proposed approach, we leverage the 10-fold cross-validation technique. The results of the 10-fold cross-validation indicate that the proposed solution is significant. The average results of the exploited metrics (precision, recall, and F1-score) are 92.76%, 99.33%, and 95.93%, respectively.
Environmental sound classification (ESC) involves the process of distinguishing an audio stream associated with numerous environmental sounds. Some common aspects such as the framework difference, overlapping of different sound events, and the presence of various sound sources during recording make the ESC task much more complicated and complex. This research is to propose a deep learning model to improve the recognition rate of environmental sounds and reduce the model training time under limited computation resources. In this research, the performance of transformer and convolutional neural networks (CNN) are investigated. Seven audio features, chromagram, Mel-spectrogram, tonnetz, Mel-Frequency Cepstral Coefficients (MFCCs), delta MFCCs, delta-delta MFCCs and spectral contrast, are extracted from the UrbanSound8K, ESC-50, and ESC-10, databases. Moreover, this research also employed three data enhancement methods, namely, white noise, pitch tuning, and time stretch to reduce the risk of overfitting issue due to the limited audio clips. The evaluation of various experiments demonstrates that the best performance was achieved by the proposed transformer model using seven audio features on enhanced database. For UrbanSound8K, ESC-50, and ESC-10, the highest attained accuracies are 0.98, 0.94, and 0.97 respectively. The experimental results reveal that the proposed technique can achieve the best performance for ESC problems.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
customersupport@researchsolutions.com
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.
Copyright © 2025 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.