Cloud computing is a modern technology for dealing with large-scale data. The Cloud has been used to process the selection and placement of replications on a large scale. Most previous studies concerning replication used mathematical models, and few studies focused on artificial intelligence (AI). The Artificial Bee Colony (ABC) is a member of the family of swarm intelligence based algorithms. It simulates bee direction to the final route and has been proven to be effective for optimization. In this paper, we present the different costs and shortest route sides in the Cloud with regard to replication and its placement between data centers (DCs) through Multi-Objective Optimization (MOO) and evaluate the cost distance by using the knapsack problem. ABC has been used to solve shortest route and lower cost problems to identify the best selection for replication placement, according to the distance or shortest routes and lower costs that the knapsack approach has used to solve these problems. Multi-objective optimization with the artificial bee colony (MOABC) algorithm can be used to achieve highest efficiency and lowest costs in the proposed system. MOABC can find an optimal solution for the best placement of data replicas according to the minimum distance and the number of data transmissions, affording low cost with the knapsack approach and availability of data replication.Low cost and fast access are characteristics that guide the shortest route in the CloudSim implementation as well. The experimental results show that the proposed MOABC is more efficient and effective for the best placement of replications than compared algorithms. INDEX TERMS Cloud computing, multi-objective optimization, artificial bee colony, replication, cloudsim, and knapsack problem.
Today, the Web is the largest source of information worldwide. There is currently a strong trend for decision-making applications such as Data Warehousing (DW) and Business Intelligence (BI) to move onto the Web, especially in the cloud. Integrating data into DW/BI applications is a critical and time-consuming task. To make better decisions in DW/BI applications, next generation data integration poses new requirements to data integration systems, over those posed by traditional data integration. In this paper, we propose a generic, metadata-based, service-oriented, and event-driven approach for integrating Web data timely and autonomously. Beside handling data heterogeneity, distribution and interoperability, our approach satisfies near real-time requirements and realize active data integration. For this sake, we design and develop a framework that utilizes Web standards (e.g., XML and Web services) for tackling data heterogeneity, distribution and interoperability issues. Moreover, our framework utilizes Active XML (AXML) to warehouse passive data as well as services to integrate active and dynamic data on-the-fly. AXML embedded services and changes detection services ensure near real-time data integration. Furthermore, the idea of integrating Web data actively and autonomously revolves around mining events logged by the data integration environment. Therefore, we propose an incremental XML-based algorithm for mining association rules from logged events. Then, we define active rules dynamically upon mined data to automate and reactivate
In recent years, cloud computing research, specifically data replication techniques and their applications, has been growing. If the replicas number is raised and put in multiple positions, it will be expensive to maintain the data usability, performance and stability of the application systems. In this paper, two bio-inspired algorithms were proposed to improve both selection and placement of data replicas in the cloud environment. The suggested algorithms for dynamic data replication are multi-objective particle swarm optimization (MO-PSO) and ant colony optimization (MO-ACO). The first suggested algorithm, i.e., MO-PSO, is employed to obtain the best selected data replica depend on the most frequent one. However, the second suggested algorithm, i.e., MO-ACO, is employed to obtain the best data replica placement depend on the shortest distance, and the replicas availability. A simulation of the suggested strategy was carried out using CloudSim. Each data center (DC) includes hosts with set of virtual machines (VMs). The data replication order is determined at random from a thousand cloudlets. All replication files are randomly distributed in the proposed architecture. The performance of suggested techniques was evaluated against several approaches including: Adaptive Replica Dynamic Strategy (ARDS), Enhance Fast Spread (EFS), Genetic Algorithm (GA), Replica Selection and Placement (RSP), Popular File Replication First (PFRF), and Dynamic Cost-aware Re-replication and Re-balancing Strategy (DCR2S). The simulation results prove that MOPSO gives improved data replication compared against other algorithms. Additionally, MOACO realizes higher data availability, lower cost, and less bandwidth consumption compared with other algorithms.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
customersupport@researchsolutions.com
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.