A data stream mining is relevant issue in the context of information gain. The data arrived are large in amount limitless and high rate with time impractical to stock, excavating and testimony at identical measure of time to retrieve intelligence. Sliding window model utilized for frequent pattern mining data stream mining emphasis on recent data and minimum space consumed. In past algorithm window measurement change was steady to concept variation when stagnant and gets smaller when the concept variation happens. Renewed frequent patterns are moderately kept in the current concept whereas the stable transaction is moved out of window. Panes steadily combined to window and performing unnecessary mining for frequent itemsets, conduct is diminishing. Based on the sliding window model the new algorithm named KF_FSW (Kalman Filter based Flexible Sliding Window Model) which utilize Kalman filter function for prediction and measurement approach. The prediction and measurement method is done on basis of already existing information as measure. Thus coagulating the error for accurate position of behavior variation in window size fluctuates in streamed database. Test on standard dataset reports that proposed algorithm coagulates less number of windows for mining and even predicting efficiently the number of count for change ratio captured by occurring change variation.
Mining frequent itemsets over data stream has been challenging task. The incoming data from various sources like ecommerce website, click streams, text, audio, weather forecasting etc. are massive unbounded and high speed that it is impractical to store all, process and scan complete data at the same time to extract information. While processing memory and time are the main parameters must be minimum consumed. Thus the paper provides different algorithms for mining over static and dynamic data also known as data stream.
The word BIG DATA is nothing but huge amount of data generated through all the sources including social networking sites like facebook, twitter, Intstagram etc. this data sometimes may be repetitive that is same person can have record in more than one databases, whereas it is belonging to a single person so those records should be merged. Also sometimes a situation may occur where you need entire history of a person in this case record linkage will make it possible. Many Researches has been done for efficiently linking the records as record linking is becoming important day-by-day since it increases the quality of the data. In this research we are going to focus on algorithm for efficiently linking the records and keeping records secure. The software called FEBRL is used for comparing our algorithm efficiency with previously defined algorithms.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
customersupport@researchsolutions.com
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.