I've created a DataFrame which I would like to write / export next to my Azure DataLake Gen2 in Tables (need to create new Table for this).
In the future I will also need to update this Azure DL Gen2 Table with new DataFrames.
In Azure Databricks I've created a connection Azure Databricks -> Azure DataLake to see my my files:
Appreciate help how to write it in spark / pyspark.
Thank you!