I've created a few simple pipelines in data factory with different kinds of linked services and datasets. Manually, by writting jsons. But none works.
Pipeline created by wizard works and have Ready status. All manual pipelines have status Waiting: Dataset dependencies. What does it mean and how can I fix it?
Completion - If activity E has a Completion dependency condition on activity A, activity E will run regardless of the completion status of activity A.
The Azure Storage and Azure SQL Database linked services contain connection strings that Data Factory uses at runtime to connect to your Azure Storage and Azure SQL Database, respectively. The Azure Blob dataset specifies the blob container and blob folder that contains the input blobs in your Blob storage.
Data Factory supports three types of activities: data movement activities, data transformation activities, and control activities.
Also, need to set external property of upstream dataset to true
"external": true,
From the documentation:
Some scenarios require the ability to produce output at a frequency different than frequencies of one or more inputs. Data factory supports modeling these scenarios.
. . .
The output slice for every day depends on 24 hourly slices from input dataset. Data factory computes these dependencies automatically by figuring out the input data slices that fall in the same time period as the output slice to be produced. If any of the .. input slices is not available (due to processing happening in an activity upstream that produces the that slice for example) data factory will wait for the input slice to be ready before kicking off the daily activity run.
So you're doing something where your input data slices aren't available to the data factory. Unfortunately, without you posting your pipeline JSON, it's hard to diagnose the issue.
If you love us? You can donate to us via Paypal or buy me a coffee so we can maintain and grow! Thank you!
Donate Us With