I am trying to join two pyspark dataframes like this
joined = df.join(df1,on=["date"],how='left').select([col('df.'+xx) for xx in df.columns] + [col('df1.daily_net_payment_sum'),col('df1.daily_net_payment_avg')])
But it results in
An error was encountered:
"cannot resolve '`df.cust_no`' given input columns:
Seems to me like I am unable to reference column by their dataframe/table name. Using spark 2.4.7
Any ideas appreciated