Logo Questions Linux Laravel Mysql Ubuntu Git Menu
 

Azure: Waiting Dataset dependencies in manual created pipelines

I've created a few simple pipelines in data factory with different kinds of linked services and datasets. Manually, by writting jsons. But none works.

Pipeline created by wizard works and have Ready status. All manual pipelines have status Waiting: Dataset dependencies. What does it mean and how can I fix it?

like image 493
user6140587 Avatar asked Jun 30 '16 12:06

user6140587


People also ask

Which execution dependency guarantees that the next activity will be executed regardless of the status of the previous activity?

Completion - If activity E has a Completion dependency condition on activity A, activity E will run regardless of the completion status of activity A.

What connects an Azure data factory activity to a dataset?

The Azure Storage and Azure SQL Database linked services contain connection strings that Data Factory uses at runtime to connect to your Azure Storage and Azure SQL Database, respectively. The Azure Blob dataset specifies the blob container and blob folder that contains the input blobs in your Blob storage.

Which 3 types of activities can you run in Microsoft Azure data Factory?

Data Factory supports three types of activities: data movement activities, data transformation activities, and control activities.


2 Answers

Also, need to set external property of upstream dataset to true

    "external": true,
like image 110
Kez Nwichi Avatar answered Dec 27 '22 04:12

Kez Nwichi


From the documentation:

Some scenarios require the ability to produce output at a frequency different than frequencies of one or more inputs. Data factory supports modeling these scenarios.

. . .

The output slice for every day depends on 24 hourly slices from input dataset. Data factory computes these dependencies automatically by figuring out the input data slices that fall in the same time period as the output slice to be produced. If any of the .. input slices is not available (due to processing happening in an activity upstream that produces the that slice for example) data factory will wait for the input slice to be ready before kicking off the daily activity run.

So you're doing something where your input data slices aren't available to the data factory. Unfortunately, without you posting your pipeline JSON, it's hard to diagnose the issue.

like image 40
Kyle Hale Avatar answered Dec 27 '22 04:12

Kyle Hale