I'm struggling distributing my HBase rows in a proper way for several map tasks. My aim is to split my scan via row key and distribute a set of rows each to a map job.
As far as now I am only able to define a scan where my mappers get always one row at a time. But that is not what I want - I need the map-input set-wise.
So is there a possibility to split-up my HBase table resp. the scan into n sets of rows, which are then input for n mappers?
I am not looking for a solution to start a MapReduce job writing n files and another MapReduce job for reading them back again as text input for getting these sets.
Thanks in advance!