The Hadoop Distributed File System

Authors

  • ANMOL KUMAR TAUNK
  • Amit Kumar Parmar
  • Rajesh Srivastav

Keywords:

Hadoop, HDFS, distributed file system

Abstract

The Hadoop Distributed File System (HDFS) is designed to store very large data sets reliably, and to stream those data sets at high bandwidth to user applications. In a large cluster, thousands of servers both host directly attached storage and execute user application tasks. By distributing storage and computation across many servers, the resource can grow with demand while remaining economical at every size. We describe the architecture of HDFS and report on experience using HDFS to manage 25 petabytes of enterprise data at Yahoo.

 

Downloads

Published

2013-10-18

How to Cite

TAUNK, A. K., Parmar, A. K., & Srivastav, R. (2013). The Hadoop Distributed File System. International Journal of Computer (IJC), 8(1), 8–15. Retrieved from https://ijcjournal.org/index.php/InternationalJournalOfComputer/article/view/191

Issue

Section

Articles