HDFS works with commodity hardware (systems with average configurations) that has high chances of getting crashed any time. Thus, to make the entire system highly fault-tolerant, HDFS replicates and stores data in different places. Any data on HDFS gets stored at atleast 3 different locations. So, even if one of them is corrupted and the other is unavailable for some time for any reason, then data can be accessed from the third one. Hence, there is no chance of losing the data. This replication factor helps us to attain the feature of Hadoop called Fault Tolerant.
TOP 100 HADOOP INTERVIEW QUESTIONS ANSWERS PDF, REAL TIME HADOOP INTERVIEW QUESTIONS GATHERED FROM EXPERTS, TOP 100 BIG DATA INTERVIEW QUESTIONS, HADOOP ONLINE QUIZ QUESTIONS, BIG DATA MCQS, HADOOP OBJECTIVE TYPE QUESTIONS AND ANSWERS
Home » Unlabelled » Replication causes data redundancy then why is pursued in HDFS? | Hadoop Questions
Subscribe to:
Post Comments (Atom)
No comments:
Post a Comment