forked from GoogleCloudPlatform/professional-services
-
Notifications
You must be signed in to change notification settings - Fork 0
/
main.py
59 lines (49 loc) · 2.43 KB
/
main.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
# Copyright 2021 Google LLC
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import argparse
import apache_beam as beam
from apache_beam.io.gcp import bigquery
from apache_beam.options.pipeline_options import PipelineOptions
from apache_beam.options.pipeline_options import SetupOptions
from pipeline.beam_classes.preprocess import PreprocessingTransform
from pipeline.beam_classes.extract import ExtractDataTransform
import logging
from pipeline.utils.utils import read_abbreviations_local
def run():
parser = argparse.ArgumentParser()
parser.add_argument("--input-csv", required=True)
parser.add_argument("--results-bq-table", required=True)
parser.add_argument("--errors-bq-table", required=True)
app_args, pipeline_args = parser.parse_known_args()
pipeline_options = PipelineOptions(pipeline_args)
pipeline_options.view_as(SetupOptions).save_main_session = True
# Read and parse side inputs.
abbreviations_dict = read_abbreviations_local('resources/abbreviations.txt')
logging.info("Read abbreviations dataset. Number of elements: {}".format(len(abbreviations_dict)))
with beam.Pipeline(options=pipeline_options) as p:
# ETL: Extract.
parsed_records, parsing_errors = p | "Extract and Parse" >> ExtractDataTransform(app_args.input_csv)
results = parsed_records | "Clean and calculate" >> PreprocessingTransform(abbreviations_dict)
# ETL: Load.
results | "Write results to BigQuery" >> beam.io.WriteToBigQuery(
table=app_args.results_bq_table,
create_disposition=bigquery.BigQueryDisposition.CREATE_NEVER,
write_disposition=bigquery.BigQueryDisposition.WRITE_TRUNCATE)
parsing_errors | "Write errors to BigQuery" >> beam.io.WriteToBigQuery(
table=app_args.errors_bq_table,
create_disposition=bigquery.BigQueryDisposition.CREATE_NEVER,
write_disposition=bigquery.BigQueryDisposition.WRITE_TRUNCATE)
if __name__ == '__main__':
logging.getLogger().setLevel(logging.INFO)
run()