I am currently running a cluster with 2 nodes. 1 Node is master/slave and the other one is just slave. I have a file and I set the block size to half the size of that file. Then I do
hdfs dfs -put file /
File gets copied to the HDFS no problem, but when I check the HDFS site, I see that both the blocks that was created is in one datanode (the blocks are on the datanode whereI used the -put command). I even tried to call the balancer script but both the blocks are still on the same datanode.
I need the data to be evenly spread out (as much as possible) between all nodes.
Am I missing something here?
hdfs dfs -ls /file
? – jlliagrehdfs dfs -ls /file
? – jlliagre