Author(s): Amol Mahadev Kadam, Pradip K. Deshmukh, Prakash B. Dhainje
Published in: International Journal of Engineering Research & Technology
License: This work is licensed under a Creative Commons Attribution 4.0 International License.
Volume/Issue: Volume. 4 - Issue. 05 , May - 2015
When a dataset exceeds the storage capacity of a single physical machine, it becomes require to divide it across a number of separate machines. File systems that manage the storage over a network of machines are called distributed file system. Hadoop meets with a distributed file system called Hadoop Distributed File System (HDFS). HDFS is a file system designed for storing huge files with streaming data access patterns, running on clusters of commodity hardware. HDFS files are hundreds of gigabytes or in terabytes in size. There are Hadoop clusters running currently that store petabytes of data. HDFS is built around the most efficient data processing patterns is a write-once, read-many time patterns.
Number of Citations for this article: Data not Available
7 Paper(s) Found related to your topic: