Q-learning is arguably one of the most applied representative reinforcement learning approaches and one of the off-policy strategies. Since the emergence of Q-learning, many studies have described its uses in reinforcement learning and artificial intelligence problems. However, there is an information gap as to how these powerful algorithms can be leveraged and incorporated into general artificial intelligence workflow. Early Q-learning algorithms were unsatisfactory in several aspects and covered a narrow range of applications. It has also been observed that sometimes, this rather powerful algorithm learns unrealistically and overestimates the action values hence abating the overall performance. Recently with the general advances of machine learning, more variants of Q-learning like Deep Q-learning which combines basic Q learning with deep neural networks have been discovered and applied extensively. In this paper, we thoroughly explain how Q-learning evolved by unraveling the mathematical complexities behind it as well its flow from reinforcement learning family of algorithms. Improved variants are fully described, and we categorize Q-learning algorithms into single-agent and multi-agent approaches. Finally, we thoroughly investigate up-to-date research trends and key applications that leverage Q-learning algorithms.
There is a need to extract meaningful information from big data, classify it into different categories, and predict end-user behavior or emotions. Large amounts of data are generated from various sources such as social media and websites. Text classification is a representative research topic in the field of natural-language processing that categorizes unstructured text data into meaningful categorical classes. The long short-term memory (LSTM) model and the convolutional neural network for sentence classification produce accurate results and have been recently used in various natural-language processing (NLP) tasks. Convolutional neural network (CNN) models use convolutional layers and maximum pooling or max-overtime pooling layers to extract higher-level features, while LSTM models can capture long-term dependencies between word sequences hence are better used for text classification. However, even with the hybrid approach that leverages the powers of these two deep-learning models, the number of features to remember for classification remains huge, hence hindering the training process. In this study, we propose an attention-based Bi-LSTM+CNN hybrid model that capitalize on the advantages of LSTM and CNN with an additional attention mechanism. We trained the model using the Internet Movie Database (IMDB) movie review data to evaluate the performance of the proposed model, and the test results showed that the proposed hybrid attention Bi-LSTM+CNN model produces more accurate classification results, as well as higher recall and F1 scores, than individual multi-layer perceptron (MLP), CNN or LSTM models as well as the hybrid models.
Big web data from sources including online news and Twitter are good resources for investigating deep learning. However, collected news articles and tweets almost certainly contain data unnecessary for learning, and this disturbs accurate learning. This paper explores the performance of word2vec Convolutional Neural Networks (CNNs) to classify news articles and tweets into related and unrelated ones. Using two word embedding algorithms of word2vec, Continuous Bag-of-Word (CBOW) and Skip-gram, we constructed CNN with the CBOW model and CNN with the Skip-gram model. We measured the classification accuracy of CNN with CBOW, CNN with Skip-gram, and CNN without word2vec models for real news articles and tweets. The experimental results indicated that word2vec significantly improved the accuracy of the classification model. The accuracy of the CBOW model was higher and more stable when compared to that of the Skip-gram model. The CBOW model exhibited better performance on news articles, and the Skip-gram model exhibited better performance on tweets. Specifically, CNN with word2vec models was more effective on news articles when compared to that on tweets because news articles are typically more uniform when compared to tweets.
Mihail L. Sichitiu reduce energy consumption in WSN MAC protocols.Energy efficiency of the MAC protocol is a key design factor for wireless sensor networks (WSNs). Due to the importance of the problem, a number of energy efficient MAC protocols have been developedfor WSNs. Preamblesampling based MAC protocols (e.g., B-MAC and X-MAC) have overheads due to their preambles, and are inefficient at large wakeup intervals. SCP-MAC, a very energy efficient scheduling MAC protocol, minimizes the preamble by combining preamble sampling and scheduling techniques; however, it does not prevent energy loss due to overhearing; in addition, due to its synchronization procedure, it results in increased contention and delay. In this paper, we present an energy efficient MAC protocol for WSNs that avoids overhearing and reduces contention and delay by asynchronously scheduling the wakeup time ofneighboring nodes. To validate our design and analysis, we implement the proposed scheme on the MicaZ platform. Experimental results show that AS-MAC considerably reduces energy consumption, packet loss and delay when compared with SCP-MAC.
Fingerprint-based wireless indoor positioning approaches are widely used for location-based services because wireless signals, such as Wi-Fi and Bluetooth, are currently pervasive in indoor spaces. The working principle of fingerprinting technology is to collect the fingerprints from an indoor environment, such as a room or a building, in advance, create a fingerprint map, and use this map to estimate the user's current location. The fingerprinting technology is associated with a high level of accuracy and reliability. However, the fingerprint map must be entirely re-created, not only when the Wi-Fi access points are added, modified, or removed, but also when the interior features, such as walls or even furniture, are changed, owing to the nature of the wireless signals. Many researchers have realized the problems in the fingerprinting technology and are conducting studies to address them. In this paper, we review the indoor positioning technologies that do not require the construction of offline fingerprint maps. We categorize them into simultaneous localization and mapping; inter/extrapolation; and crowdsourcing-based technologies, and describe their algorithms and characteristics, including advantages and disadvantages. We compare them in terms of our own parameters: accuracy, calculation time, versatility, robustness, security, and participation. Finally, we present the future research direction of the indoor positioning techniques. We believe that this paper provides valuable information on recent indoor localization technologies without offline fingerprinting map construction.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
customersupport@researchsolutions.com
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.