How to Configure Replication Factor and Block Size for HDFS?

By Ravi Karamsetty | September 1, 2014 |

Hadoop Distributed File System (HDFS) stores files as data blocks and distributes these blocks across the entire cluster. As HDFS was designed to be fault-tolerant and to run on commodity hardware, blocks are replicated a number of times to ensure high data availability. The replication factor is a property that can be set in the … Continue reading “How to Configure Replication Factor and Block Size for HDFS?”

Read More