I have a hive table which is partitioned on many countries. I want to load specific partition data to my dataframe, as shown below:
df=spark.read.orc("/apps/hive/warehouse/emp.db/partition_load_table").where('country="NCL"' && 'county="RUS"')
It's giving me an error, though I was able to load for single partition.
below is my directory structure in hdfs
/apps/hive/warehouse/emp.db/partition_load_table/country=NCL
df=spark.read.orc("/apps/hive/warehouse/emp.db/partition_load_table").where('country="NCL"')