Hadoop Version: 2.7.3 Datanodes: 32 BlockSize: 512m Replication: 3
I Cleared all the data in HDFS using the command
hdfs dfs -rm -r /*
After clearing the HDFS cluster, few datanodes still showing the Block pool used
, though blocks are zero.
The output of hdfs fsck /
is as follows
Connecting to namenode via http://ip-10-0-2-7:50070/fsck?ugi=ubuntu&path=%2F
FSCK started by ubuntu (auth:SIMPLE) from /10.0.2.7 for path / at Tue Jul 18 04:34:19 UTC 2017
Status: HEALTHY
Total size: 0 B
Total dirs: 1
Total files: 0
Total symlinks: 0
Total blocks (validated): 0
Minimally replicated blocks: 0
Over-replicated blocks: 0
Under-replicated blocks: 0
Mis-replicated blocks: 0
Default replication factor: 3
Average block replication: 0.0
Corrupt blocks: 0
Missing replicas: 0
Number of data-nodes: 32
Number of racks: 1
FSCK ended at Tue Jul 18 04:34:19 UTC 2017 in 1 milliseconds
The filesystem under path '/' is HEALTHY
I know the block deletion is asynchronous process, but the datanodes which are having the Block Pool Used
not set to 0, taking a lot of time to release the blocks.
Can anybody help me figure out, why on those servers, block deletion is slow, or any other issue because of which the blocks are being used.