We are using Databricks and its SQL DW connector to load data into Synapse. I have a dataset with 10 000 rows and 40 columns. It takes 7 minutes!
Loading same dataset using Data Factory with Polybase and staging option takes 27 seconds. Same with bulk copy.
What could be wrong? Am I missing some configuration? Or is this business as usual?
Connection configuration:
df_insert.write .format("com.databricks.spark.sqldw") .option("url", sqlDwUrlSmall) .option("dbtable", t_insert) .option( "forward_spark_azure_storage_credentials","True") .option("tempdir", tempDir) .option("maxStrLength", maxStrLength) .mode("append") .save()