Job Description: Mandatory Skills:
- Professional experience in Data Engineering using Google Big Query (GBQ) and Google Cloud Platform (GCP) data sets and building data pipelines.
- Hands on and deep experience working with Google Data Products (e.g. Big Query, Dataflow, Data proc, Data prep, Cloud Composer, Airflow, DAG etc.).
- Python Programming Expert Py Spark, Pandas.
- Experience in Airflow (Create DAG, Configure the variables in Airflow, Scheduling).
- Big Data technologies and solutions (Spark, Hadoop, Hive, MapReduce) and multiple scripting and languages (YAML, Python).
- Experience in DBT to create the lineage in GCP.
- Worked in Dev-Sec-Ops (CICD) environment.
- Design and develop the ETL ELT framework using Big Query, Expertise in Big Query concepts like Nested Queries, Clustering, Partitioning, etc.
- Experience in Data Integration, Data Transformation, Data Quality and Data Lineage tools.
- Should be able to automate the data load from Big Query using APIs or scripting language.
- E2E Data Engineering and Lifecycle (including non-functional requirements and operations) management.
- E2E Solution Design skills - Prototyping, Usability testing and data visualization literacy.
- Experience with SQL and NoSQL modern data stores.