I have a hadoop cluster with 4 nodes(1master, 3slaves). And I create some hive tables from files stored in hdfs. Then I configure mysql as the hive metastore and copy the hive-site.xml file inside conf folder of spark.
To install spark, I just download and extract spark in the master node. And after copy the hive-site.xml inside spark conf folder, I start spark with spark-shell command
. Its needed to install in slave nodes also?
Im asking this because, Im executing with success spark sql queries like below, but if I try to acess the cluster manager default page in localhost:8080, it shows "Unable to connect". So it seems that spark sql is working fine, but without any cluster manager working, this is possible??
var hiveContext = new org.apache.spark.sql.hive.HiveContext(sc)
query = hiveContext.sql("select * from customers");
query.show()
master:8080
spark-shell
that you have started? have you set the master with this command? – user1314742