-
Notifications
You must be signed in to change notification settings - Fork 1
/
Copy pathETL_toll_data.py
executable file
·70 lines (60 loc) · 2.2 KB
/
ETL_toll_data.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
from datetime import timedelta
from airflow import DAG
from airflow.operators.bash_operator import BashOperator
from airflow.utils.dates import days_ago
#Defining DAG arguments
default_args = {
'owner':'Malims',
'start_date': days_ago(0),
'email':['[email protected]'],
'email_on_failure': True,
'email_on_retry':True,
'reties':1,
'retry_delay':timedelta(minutes=5),
}
#Define the DAG
dag = DAG(
dag_id='ETL_toll_data',
default_args=default_args,
description='Apache Airflow Traffic Data',
schedule_interval=timedelta(days=1),
)
#Define the tasks
#download data from url
download_data=BashOperator(
task_id='download_data',
bash_command='wget "https://github.com/malimsZen/Airflow-ETL_Toll_Data/raw/main/tolldata.tgz" -O ~/Zen/Airflow-ETL_Toll_Data/tolldata.tgz',
dag=dag,
)
#define the first task names unzip_data
unzip_data = BashOperator(
task_id='unzip_data',
bash_command = 'tar zxvf ~/Zen/Airflow-ETL_Toll_Data/tolldata.tgz',
dag=dag,
)
#Task to extract data from csv file
extract_data_from_csv = BashOperator(
task_id = 'extract_data_from_csv',
bash_command = 'cut -d"," -f1,2,3,4 ~/Zen/Airflow-ETL_Toll_Data/vehicle-data.csv > ~/Zen/Airflow-ETL_Toll_Data/vehicle-data.csv',
dag=dag,
)
#Task to extract data from tsv file.
extract_data_from_tsv = BashOperator(
task_id='extract_data_from_tsv',
bash_command='cut -d$"\t" -f5,6,7 ~/Zen/Airflow-ETL_Toll_Data/tollplaza-data.tsv > ~/Zen/Airflow-ETL_Toll_Data/tollplaza-data.csv',
dag=dag,
)
#Task to extract data from fixed width file.
extract_data_from_fixed_width = BashOperator(
task_id='extract_data_from_fixed_width',
bash_command='cut -d " " -f 6-7 ~/Zen/Airflow-ETL_Toll_Data/payment-data.txt > ~/Zen/Airflow-ETL_Toll_Data/payment-data.csv',
dag=dag,
)
#Task to consolidate data
consolidate_data=BashOperator(
task_id='consolidate_data',
bash_command='paste ~/Zen/Airflow-ETL_Toll_Data/vehicle-data.csv >> ~/Zen/Airflow-ETL_Toll_Data/tollplaza-data.csv >> ~/Zen/Airflow-ETL_Toll_Data/payment-data.csv',
dag=dag,
)
#Task pipeline
download_data >> unzip_data >> extract_data_from_csv >> extract_data_from_tsv >> extract_data_from_fixed_width >> consolidate_data