BackgroundThe recent rise in popularity and scale of social networking services (SNSs) has resulted in an increasing need for SNS-based information extraction systems. A popular application of SNS data is health surveillance for predicting an outbreak of epidemics by detecting diseases from text messages posted on SNS platforms. Such applications share the following logic: they incorporate SNS users as social sensors. These social sensor–based approaches also share a common problem: SNS-based surveillance are much more reliable if sufficient numbers of users are active, and small or inactive populations produce inconsistent results.ObjectiveThis study proposes a novel approach to estimate the trend of patient numbers using indirect information covering both urban areas and rural areas within the posts.MethodsWe presented a TRAP model by embedding both direct information and indirect information. A collection of tweets spanning 3 years (7 million influenza-related tweets in Japanese) was used to evaluate the model. Both direct information and indirect information that mention other places were used. As indirect information is less reliable (too noisy or too old) than direct information, the indirect information data were not used directly and were considered as inhibiting direct information. For example, when indirect information appeared often, it was considered as signifying that everyone already had a known disease, leading to a small amount of direct information.ResultsThe estimation performance of our approach was evaluated using the correlation coefficient between the number of influenza cases as the gold standard values and the estimated values by the proposed models. The results revealed that the baseline model (BASELINE+NLP) shows .36 and that the proposed model (TRAP+NLP) improved the accuracy (.70, +.34 points).ConclusionsThe proposed approach by which the indirect information inhibits direct information exhibited improved estimation performance not only in rural cities but also in urban cities, which demonstrated the effectiveness of the proposed method consisting of a TRAP model and natural language processing (NLP) classification.
Abstract. Existing search engines contain the picture of the Web from the past and their ranking algorithms are based on data crawled some time ago. However, a user requires not only relevant but also fresh information. We have developed a method for adjusting the ranking of search engine results from the point of view of page freshness and relevance. It uses an algorithm that postprocesses search engine results based on the changed contents of the pages. By analyzing archived versions of web pages we estimate temporal qualities of pages, that is, general freshness and relevance of the page to the query topic over certain time frames. For the top quality web pages, their content differences between past snapshots of the pages indexed by a search engine and their present versions are analyzed. Basing on these differences the algorithm assigns new ranks to the web pages without the need to maintain a constantly updated index of web documents.
Wireless sensor networks (WSNs) are among the most popular wireless technologies for sensor communication purposes nowadays. Usually, WSNs are developed for specific applications, either monitoring purposes or tracking purposes, for indoor or outdoor environments, where limited battery power is a main challenge. To overcome this problem, many routing protocols have been proposed through the last few years. Nevertheless, the extension of the network lifetime in consideration of the sensors capacities remains an open issue. In this paper, to achieve more efficient and reliable protocols according to current application scenarios, two well-known energy efficient protocols, i.e., Low-Energy Adaptive Clustering hierarchy (LEACH) and Energy–Efficient Sensor Routing (EESR), are redesigned considering neural networks. Specifically, to improve results in terms of energy efficiency, a Levenberg–Marquardt neural network (LMNN) is integrated. Furthermore, in order to improve the performance, a sub-cluster LEACH-derived protocol is also proposed. Simulation results show that the Sub-LEACH with LMNN outperformed its competitors in energy efficiency. In addition, the end-to-end delay was evaluated, and Sub-LEACH protocol proved to be the best among existing strategies. Moreover, an intrusion detection system (IDS) has been proposed for anomaly detection based on the support vector machine (SVM) approach for optimal feature selection. Results showed a 96.15% accuracy—again outperforming existing IDS models. Therefore, satisfactory results in terms of energy efficiency, end-to-end delay and anomaly detection analysis were attained.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
customersupport@researchsolutions.com
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.