Dear Stackoverflow member,
I am trying to use Airflow to schedule a DAG that launch a cluster, run some spark jobs then terminated itself. I tried to follow this example:
https://insight.io/github.com/apache/incubator-airflow/blob/master/airflow/contrib/example_dags/example_emr_job_flow_automatic_steps.py
It gave me the error: airflow.exceptions.AirflowException: The conn_id emr_default
isn't defined
I probably need to have emr_default
set up (please correct me if this is not the case). Does anyone have try this before and give me some hints on how to set it up? Airflow website doesn't provide too much on emr connection: http://pythonhosted.org/airflow/configuration.html?highlight=connection#connections
Thanks!
On the airflow UI, go to Admin -> Connections tab, and edit the values for emr_default row.
As@Him said, you can edit in airflow.
but if you can't find emr_default
, you can go to Admin
-> Connections
-> Create
to add any connection you want
If you love us? You can donate to us via Paypal or buy me a coffee so we can maintain and grow! Thank you!
Donate Us With