Article Fingerprint
ReserarchID
CSTSDEWSLFE
While replicating data over a decentralized Peer-to-Peer (P2P) network, transactions broadcasting updates arising from different peers run simultaneously so that a destination peer replica can be updated concurrently, that always causes transaction and data conflicts. Moreover, during data migration, connectivity interruption and network overload corrupt running transactions so that destination peers can experience duplicated data or improper data or missing data, hence replicas remain inconsistent. Different methodological approaches have been combined to solve these problems: the audit log technique to capture the changes made to data; the algorithmic method to design and analyse algorithms and the statistical method to analyse the performance of new algorithms and to design prediction models of the execution time based on other parameters. A Graphical User Interface software as prototype, have been designed with C #, to implement these new algorithms to obtain a database synchronizer-mediator. A stream of experiments, showed that the new algorithms were effective. So, the hypothesis according to which “The execution time of replication and reconciliation transactions totally depends on independent factors.” has been confirmed.
Katembo Kituta Ezechiel. 2019. \u201cMediation of Lazy Update Propagation in a Replicated Database over a Decentralized P2P Architecture\u201d. Global Journal of Computer Science and Technology - C: Software & Data Engineering GJCST-C Volume 19 (GJCST Volume 19 Issue C3): .
Crossref Journal DOI 10.17406/gjcst
Print ISSN 0975-4350
e-ISSN 0975-4172
Explore published articles in an immersive Augmented Reality environment. Our platform converts research papers into interactive 3D books, allowing readers to view and interact with content using AR and VR compatible devices.
Your published article is automatically converted into a realistic 3D book. Flip through pages and read research papers in a more engaging and interactive format.
Total Score: 103
Country: India
Subject: Global Journal of Computer Science and Technology - C: Software & Data Engineering
Authors: Katembo Kituta Ezéchiel, Shri Kant, Ruchi Agarwal (PhD/Dr. count: 0)
View Count (all-time): 286
Total Views (Real + Logic): 4849
Total Downloads (simulated): 1246
Publish Date: 2019 12, Mon
Monthly Totals (Real + Logic):
This paper attempted to assess the attitudes of students in
Advances in technology have created the potential for a new
Inclusion has become a priority on the global educational agenda,
While replicating data over a decentralized Peer-to-Peer (P2P) network, transactions broadcasting updates arising from different peers run simultaneously so that a destination peer replica can be updated concurrently, that always causes transaction and data conflicts. Moreover, during data migration, connectivity interruption and network overload corrupt running transactions so that destination peers can experience duplicated data or improper data or missing data, hence replicas remain inconsistent. Different methodological approaches have been combined to solve these problems: the audit log technique to capture the changes made to data; the algorithmic method to design and analyse algorithms and the statistical method to analyse the performance of new algorithms and to design prediction models of the execution time based on other parameters. A Graphical User Interface software as prototype, have been designed with C #, to implement these new algorithms to obtain a database synchronizer-mediator. A stream of experiments, showed that the new algorithms were effective. So, the hypothesis according to which “The execution time of replication and reconciliation transactions totally depends on independent factors.” has been confirmed.
Lorem ipsum dolor sit amet, consectetur adipiscing elit. Ut elit tellus, luctus nec ullamcorper mattis, pulvinar dapibus leo.