I am trying to understand the relationship between the HDFS file-system block size and the underlying physical file-system block size.
As per my understanding hdfs is just a virtual file-system which stores the actual data on the underlying physical file-system. The HDFS block size in hadoop 2 is 128 MB; however in most of the linux based file-systems the block size is 4 KB.
My questions:
Q1) When a HDFS block is written to the actual file-system, does it write to multiple block of the underlying file-system? That is for a single HDFS blocks, it has to write to 128 * 1024 KB / 4 KB --> 32,768 blocks?
Q2) If above is correct, doesn't it involve lot of seeks on the disk head? Isn't it time consuming process? How is Hadoop to make this process efficiently?
Can anyone help me understand this?