Within my Azure Data Lake directory ingest
, I have timestamp-named subdirectories:
ingest/1510395023/file.tsv
ingest/1510425023/file.tsv
ingest/1510546210/file.tsv
...
In an Azure Data Factory v2 copy activity (Data Lake to SQL Server), how can I filter the Data Lake directories to those greater than a specified timestamp?
Assume I will record the already-copied timestamps into a SQL Server, and on the next pipeline run, I only want to copy incremental new directories based on a Lookup Activity.
Data Factory documentation references logical functions, but I don't understand how to implement them in this scenario.
Alternatively: If a logical comparison filter is available on the file name, as opposed to the directories, that would be helpful as well.
Note: I want to load in historical data as well, so only filtering a timestamp greater or equal to now
is insufficient.