You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Airflow Webserver: Airflow Webserver role is used to start the Airflow Web UI. Webserver role can be deployed on more than instances. However, they will be the same and can be used for backup purposes.
70
-
71
-
Airflow Scheduler: Airflow Scheduler role is used to schedule the Airflow jobs. This is limited to one instance to reduce the risk of duplicate jobs.
91
+
Webserver: Airflow Webserver role runs the Airflow Web UI. Webserver role can be deployed on more than instances. However, they will be the same and can be used for backup purposes.
72
92
73
-
Airflow Worker: Airflow Worker role picks jobs from RabbitMQ and executed them on the nodes. Multiple instances can be deployed.
93
+
Scheduler: Airflow Scheduler role is used to schedule the Airflow jobs. This is limited to one instance to reduce the risk of duplicate jobs.
74
94
95
+
<<<<<<< HEAD
75
96
Airflow Flower: Airflow Flower is used to monitor celery clusters. Multiple instances are supported
97
+
=======
98
+
Worker: Airflow Worker role picks jobs from the Scheduler and executes them. Multiple instances can be deployed.
76
99
77
-
Kerberos: Kerberos is used to enable Kerberos protocol for the Airflow. It internally executes `airflow kerberos`. An external Kerberos Distribution Center must be setup. Multiple instances can be setup for load balancing purposes.
100
+
Flower Webserver: Flower Webserver role is used to monitor Celery clusters. Celery allows for the expansion of Worker Only one instance is needed.
101
+
>>>>>>> 11b39ebe4b384d6ad1c1735474123e5af2c8676a
78
102
79
-
Gateway: The purpose of the gateway role is to write the configurations from the configurations tab into the airflow.cfg file. This is done through the update_cfg.sh file which is executed from the scriptRunner within the gateway role.
103
+
Kerberos: Airflow Kerberos role is used to enable Kerberos protocol for the other Airflow roles and for DAGs. This role should exist on each host with an Airflow Worker role.
104
+
105
+
Gateway: The purpose of the gateway role is to make the configuration available to CLI clients.
80
106
81
107
## Using the Airflow binary:
82
108
Here are some of the examples of Airflow commands:
@@ -101,46 +127,35 @@ The DAG file has to be copied to `dags_folder` directory within all the nodes. I
101
127
In order to enable authentication for the Airflow Web UI check the "Enable Airflow Authentication" option. You can create Airflow users using one of two options below.
102
128
103
129
### Creating Airflow Users using UI:
104
-
1. Navigate to Airflow CSD. In the configurations page, enter the Airflow Username, Airflow Email, Airflow Password you want to create.
105
-
2. Deploy the client configurations to create the Airflow user.
130
+
One way to add Airflow users to the database is using the `airflow-mkuser` script. Users can be added as follows:
131
+
132
+
1. Navigate to Airflow WebUI.
133
+
2. In the Admin dropdown choose Users.
134
+
3. Choose Create and enter the username, email, and password you want to create.
106
135
107
136
Note: Although the last created user shows up in the Airflow configurations, you can still use the previously created users.
108
137
109
-
### Using mkuser.sh
110
-
Another way to add Airflow users is using the `mkuser.sh` script. Users can be added as follows:
111
-
1. Navigate to the current working directory of the CSD under `/var/run/cloudera-scm-agent/process`
112
-
2. Export PYTHONPATH and AIRFLOW_HOME environment variables. By default these are:
138
+
### Using airflow-mkuser
139
+
Another way to add Airflow users to the database is using the `airflow-mkuser` script. Users can be added as follows:
Update the `version` file before running `make dist` if creating a new release.
142
156
143
157
## Limitations:
158
+
<<<<<<< HEAD
144
159
1. The IP address of the RabbitMQ instance has to be manually entered during installation configuration.
145
160
2. After deploying configurations, there is no alert or warning that the specific roles needs to be restarted.
146
161
3. Only 'airflow.contrib.auth.backends.password_auth' mechanism is supported for Airflow user authentication.
@@ -149,6 +164,14 @@ jar -cvf AIRFLOW-1.0.0.jar -C src/ .
149
164
1. Build RabbitMQ parcel.
150
165
2. Test Database connection.
151
166
3. Add the support for more Airflow user authentication methods.
167
+
=======
168
+
1. After deploying configurations, there is no alert or warning that the specific roles needs to be restarted.
169
+
2. Only 'airflow.contrib.auth.backends.password_auth' mechanism is supported for Airflow user authentication.
170
+
171
+
## Future work:
172
+
1. Test Database connection.
173
+
2. Add the support for more Airflow user authentication methods.
174
+
>>>>>>> 11b39ebe4b384d6ad1c1735474123e5af2c8676a
152
175
153
176
## Known Errors:
154
177
@@ -157,7 +180,8 @@ jar -cvf AIRFLOW-1.0.0.jar -C src/ .
157
180
Upon many deployments, you may face an error called 'Markup file already exists' while trying to stop a role and the process never stops. In that case, stop the process using the "Abort" command and navigate to `/var/run/cloudera-scm-agent/process` and delete all the `GracefulRoleStopRunner` directories.
0 commit comments