I created a (once run) DF (V2) pipeline to load files (.lta.gz) from a SFTP server into an azure blob to get historical data. Worked beautifully. Every day there will be several new files on the SFTP server (which cannot be manipulated or deleted). So I want to create an incremental load pipeline which checks daily for new files - if so ---> copy new files.
Does anyone have any tips for me how to achieve this?
Thanks for using Data Factory!
To incrementally load newly generated files on SFTP server, you can leverage the GetMetadata activity to retrieve the LastModifiedDate property: https://learn.microsoft.com/en-us/azure/data-factory/control-flow-get-metadata-activity
Essentially you author a pipeline containing the following activities:
Have fun building data integration flows using Data Factory!
If you love us? You can donate to us via Paypal or buy me a coffee so we can maintain and grow! Thank you!
Donate Us With