How do I chain multiple Google Cloud DataPrep flows?
up vote
1
down vote
favorite
I've created two Flows in Cloud DataPrep - the first outputs to a BigQuery table and also creates a reference dataset. The second flow takes the reference dataset and processes it further before outputting to a second BigQuery table.
Is it possible to schedule these two Flows to run sequentially? At the moment I have to estimate the time taken for the first flow, and schedule the second one to run XX minutes after the first.
Either a recipe in the first flow to trigger the second, or a way to schedule them in sequence would be ideal.
This question shows how to create a reference dataset, but doesn't explain if it's possible to run them automatically/sequentially.
google-cloud-dataprep
add a comment |
up vote
1
down vote
favorite
I've created two Flows in Cloud DataPrep - the first outputs to a BigQuery table and also creates a reference dataset. The second flow takes the reference dataset and processes it further before outputting to a second BigQuery table.
Is it possible to schedule these two Flows to run sequentially? At the moment I have to estimate the time taken for the first flow, and schedule the second one to run XX minutes after the first.
Either a recipe in the first flow to trigger the second, or a way to schedule them in sequence would be ideal.
This question shows how to create a reference dataset, but doesn't explain if it's possible to run them automatically/sequentially.
google-cloud-dataprep
add a comment |
up vote
1
down vote
favorite
up vote
1
down vote
favorite
I've created two Flows in Cloud DataPrep - the first outputs to a BigQuery table and also creates a reference dataset. The second flow takes the reference dataset and processes it further before outputting to a second BigQuery table.
Is it possible to schedule these two Flows to run sequentially? At the moment I have to estimate the time taken for the first flow, and schedule the second one to run XX minutes after the first.
Either a recipe in the first flow to trigger the second, or a way to schedule them in sequence would be ideal.
This question shows how to create a reference dataset, but doesn't explain if it's possible to run them automatically/sequentially.
google-cloud-dataprep
I've created two Flows in Cloud DataPrep - the first outputs to a BigQuery table and also creates a reference dataset. The second flow takes the reference dataset and processes it further before outputting to a second BigQuery table.
Is it possible to schedule these two Flows to run sequentially? At the moment I have to estimate the time taken for the first flow, and schedule the second one to run XX minutes after the first.
Either a recipe in the first flow to trigger the second, or a way to schedule them in sequence would be ideal.
This question shows how to create a reference dataset, but doesn't explain if it's possible to run them automatically/sequentially.
google-cloud-dataprep
google-cloud-dataprep
asked Nov 9 at 10:53
angusham
234
234
add a comment |
add a comment |
1 Answer
1
active
oldest
votes
up vote
1
down vote
accepted
The documentation for reference datasets sort of alludes to the behaviour, but could be clearer.
- The second flow that uses the first flow as reference dataset WILL RUN the first flow job and therefore use updated data for flow 2.
BUT the export to BigQuery would not perform the export from flow 1.
The easiest solution, but not necessarily what you want is that can choose to perform the BigQuery export from Flow 1 in Flow 2, i.e. have a simple recipe that does nothing but host the export job.
The solution I use, where recipes/jobs are stable, is to run the resultant Dataflow jobs rather than use the Dataprep tool itself for execution: Run Job on Cloud Dataflow
You have a few options for scheduling the two Dataflow jobs. (Cloud Scheduler could be a good one - its a new project and I'm looking at replacing custom solution I have right now hosted as Cloud Functions)
In terms of running them in sequence, however, BigQuery doesn't have update event that could trigger Job2 so you can either schedule them long enough apart and hope Job 1 finishes soon enough, OR you can poll the BigQuery table metadata to see if the modified date changes.
add a comment |
1 Answer
1
active
oldest
votes
1 Answer
1
active
oldest
votes
active
oldest
votes
active
oldest
votes
up vote
1
down vote
accepted
The documentation for reference datasets sort of alludes to the behaviour, but could be clearer.
- The second flow that uses the first flow as reference dataset WILL RUN the first flow job and therefore use updated data for flow 2.
BUT the export to BigQuery would not perform the export from flow 1.
The easiest solution, but not necessarily what you want is that can choose to perform the BigQuery export from Flow 1 in Flow 2, i.e. have a simple recipe that does nothing but host the export job.
The solution I use, where recipes/jobs are stable, is to run the resultant Dataflow jobs rather than use the Dataprep tool itself for execution: Run Job on Cloud Dataflow
You have a few options for scheduling the two Dataflow jobs. (Cloud Scheduler could be a good one - its a new project and I'm looking at replacing custom solution I have right now hosted as Cloud Functions)
In terms of running them in sequence, however, BigQuery doesn't have update event that could trigger Job2 so you can either schedule them long enough apart and hope Job 1 finishes soon enough, OR you can poll the BigQuery table metadata to see if the modified date changes.
add a comment |
up vote
1
down vote
accepted
The documentation for reference datasets sort of alludes to the behaviour, but could be clearer.
- The second flow that uses the first flow as reference dataset WILL RUN the first flow job and therefore use updated data for flow 2.
BUT the export to BigQuery would not perform the export from flow 1.
The easiest solution, but not necessarily what you want is that can choose to perform the BigQuery export from Flow 1 in Flow 2, i.e. have a simple recipe that does nothing but host the export job.
The solution I use, where recipes/jobs are stable, is to run the resultant Dataflow jobs rather than use the Dataprep tool itself for execution: Run Job on Cloud Dataflow
You have a few options for scheduling the two Dataflow jobs. (Cloud Scheduler could be a good one - its a new project and I'm looking at replacing custom solution I have right now hosted as Cloud Functions)
In terms of running them in sequence, however, BigQuery doesn't have update event that could trigger Job2 so you can either schedule them long enough apart and hope Job 1 finishes soon enough, OR you can poll the BigQuery table metadata to see if the modified date changes.
add a comment |
up vote
1
down vote
accepted
up vote
1
down vote
accepted
The documentation for reference datasets sort of alludes to the behaviour, but could be clearer.
- The second flow that uses the first flow as reference dataset WILL RUN the first flow job and therefore use updated data for flow 2.
BUT the export to BigQuery would not perform the export from flow 1.
The easiest solution, but not necessarily what you want is that can choose to perform the BigQuery export from Flow 1 in Flow 2, i.e. have a simple recipe that does nothing but host the export job.
The solution I use, where recipes/jobs are stable, is to run the resultant Dataflow jobs rather than use the Dataprep tool itself for execution: Run Job on Cloud Dataflow
You have a few options for scheduling the two Dataflow jobs. (Cloud Scheduler could be a good one - its a new project and I'm looking at replacing custom solution I have right now hosted as Cloud Functions)
In terms of running them in sequence, however, BigQuery doesn't have update event that could trigger Job2 so you can either schedule them long enough apart and hope Job 1 finishes soon enough, OR you can poll the BigQuery table metadata to see if the modified date changes.
The documentation for reference datasets sort of alludes to the behaviour, but could be clearer.
- The second flow that uses the first flow as reference dataset WILL RUN the first flow job and therefore use updated data for flow 2.
BUT the export to BigQuery would not perform the export from flow 1.
The easiest solution, but not necessarily what you want is that can choose to perform the BigQuery export from Flow 1 in Flow 2, i.e. have a simple recipe that does nothing but host the export job.
The solution I use, where recipes/jobs are stable, is to run the resultant Dataflow jobs rather than use the Dataprep tool itself for execution: Run Job on Cloud Dataflow
You have a few options for scheduling the two Dataflow jobs. (Cloud Scheduler could be a good one - its a new project and I'm looking at replacing custom solution I have right now hosted as Cloud Functions)
In terms of running them in sequence, however, BigQuery doesn't have update event that could trigger Job2 so you can either schedule them long enough apart and hope Job 1 finishes soon enough, OR you can poll the BigQuery table metadata to see if the modified date changes.
answered Nov 16 at 14:19
Jonathon
2,99531940
2,99531940
add a comment |
add a comment |
Thanks for contributing an answer to Stack Overflow!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
To learn more, see our tips on writing great answers.
Some of your past answers have not been well-received, and you're in danger of being blocked from answering.
Please pay close attention to the following guidance:
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstackoverflow.com%2fquestions%2f53224329%2fhow-do-i-chain-multiple-google-cloud-dataprep-flows%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown