0
votes

That’s what I am currently trying to do in Data Flow:

  1. Read table from SQL DWH (FileName, ProductID, MachinesCount, UsersCount, LastUsed)
  2. In my Sink settings, I have set “File name option” to “As data in column” and gave FileName column.

In my Sink dataset(Azure Blob Storage), I have specified container to “referencedata” and folderPath to “mostused/accounts/newaccounts”. Now whenever Data Flow runs it just writes data on root of Blob Container which is “referencedata” and ignoring folderPath.

Extra information that could be useful: When Data Flow is actually running its writing data inside “mostused/accounts/newaccounts” just like how Azure Databricks does but then moves it to root of Blob Container. I think the logic that copies Azure Databricks multi partitioned files into more human readable files is buggy and its moving files to root of Blob container.

1

1 Answers

2
votes

Found the solution. When choosing “As data in column”, you need to specify the folder path from the container. So if your filename was file.json and you wish to write it in directory “mostused/accounts/newaccounts”, you need to have the column value be “mostused/accounts/newaccounts/file.json”.

I understand this is a confusing experience and Microsoft is working to improve it.