I am going through the Mahout in Action book and trying out the distributed recommendation engine example. But instead of the Wikipedia dataset I am using a small dataset for my example.
input.txt->
1,15
1,13
1,12
1,10
2,11
2,15
2,20
2,17
2,12
3,10
user.txt->
3
where input.txt and user.txt is of the form user,item
and input.txt represent the items user has purchased in the past and user.txt represent the current purchases
When I copy these 2 files in hdfs and run
hadoop jar mahout-core-0.5-job.jar \ org.apache.mahout.cf.taste.hadoop.item.RecommenderJob \ -Dmapred.input.dir=input/input.txt \ -Dmapred.output.dir=output --usersFile input/user.txt --booleanData
The map reduce runs properly. However when I check the output in bin/hadoop fs -cat output/
part-r-00000
I find an empty file.
Can someone explain me what’s going wrong? If I were to understand correctly the Recommender Job should have built an item to item similarity matrix, multiply it with the user-item matrix(from user.txt) and produce the result.
Need some help understanding. I am using Mahout 0.5 and hadoop 1.2 on a single node. I hope its not an issue of version compatibility.
EDIT
I get an answer if I change the user.txt to
2
or
1