Affiliation:
1. Department of Computer Science and Business Systems Sri Sairam Engineering College Chennai India
2. Department of Computer Science and Engineering Vel Tech Rangarajan Dr. Sagunthala R&D Institute of Science and Technology Chennai India
3. Department of Data Science and Business Systems SRM Institute of Science & Technology Chennai Tamil Nadu India
Abstract
SummaryThe Hadoop distributed file system (HDFS) effectively manages data by segmenting it into blocks distributed across DataNodes in its cluster. While default block sizes in Hadoop 2.x and 1.x are 128 and 64 MB, respectively, they can be customized for larger files. HDFS ensures data reliability by replicating blocks across multiple DataNodes, but this can introduce high latency in cloud storage during heavy network traffic, particularly in big data processing. To address this, we introduce swarm intelligence with a chaotic leader and a salp (SI‐CLS) optimization algorithm. This algorithm reduces network traffic between racks in HDFS by optimizing block distribution. The SI‐CLS algorithm calculates a fitness value for each block, aiming to increase data availability and reduce latency. Performance metrics, including latency, data availability, and load balancing, indicate the effectiveness of SI‐CLS. Adopting this algorithm enhances HDFS performance, ensuring better data availability and lower latency, resulting in improved system reliability.
Cited by
3 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献