Hadoop File System was developed using distributed file system
design. It is run on commodity hardware. Unlike other distributed
systems, HDFS is highly faulttolerant and designed using low-cost
hardware.
HDFS holds very large amount of data and provides easier access. To
store such huge data, the files are stored across multiple machines.
These files are stored in redundant fashion to rescue the system from
possible data losses in case of failure. HDFS also makes applications
available to parallel processing.
No comments:
Post a Comment