At the moment we could pass in folders/files as input for MapReduce job, but what I wanted to know if we could 'cat' data from HDFS (hdfs dfs -cat file.txt) and pass that as MapReduce job input?
0
votes
1 Answers
0
votes
No, you could not do that with MapReduce.
Alternatives that can do that are: SparkStreaming and Storm.