What are the responsibilities and job description for the GCP Data Engineer - Middletown Township, NJ (Onsite) position at Caliber Smart?
Job Title: GCP Data Engineer
Duration: 12 Months
Location: Middletown Township, NJ (Onsite)
Responsibilities
Contribute to the migration of legacy data warehouse to a google cloud-based data warehouse for a Telecom Major..
Collaborate with Data Product Managers, Data Architects to design, implement, and deliver successful data solutions
Help architect data pipelines for the underlying data warehouse and data marts
Design and develop very complex ETL pipelines in Google cloud Data environments.
Our legacy tech stack includes Teradata and new tech stack includes GCP Cloud Data Technologies like BigQuery and Airflow and languages include SQL , Python
Maintain detailed documentation of your work and changes to support data quality and data governance
Support QA and UAT data testing activities
Support Deployment activities to higher environments
Ensure high operational efficiency and quality of your solutions to meet SLAs and support commitment to our customers (Data Science, Data Analytics teams)
Be an active participant and advocate of agile/scrum practice to ensure health and process improvements for your team
Basic Qualifications
8 years of data engineering experience developing large data pipelines in very complex environments
Very Strong SQL skills and ability to build Very complex transformation data pipelines using custom ETL framework in Google BigQuery environment
Exposure to Teradata and ability to understand complex Teradata BTEQ scripts
String Python programming skills
Strong Skills on build Airflow Jobs and Debug issues
Ability to Optimize the Query in BigQuery
Hands-on experience on Google Cloud data Technologies (GCS , BigQuery, Dataflow, Pub sub, Data Fusion , Cloud Function)
Preferred Qualifications
Experience with cloud data warehouse technology BigQuery.
Nice to have experience with Cloud technologies like GCP (GCS , Data Proc, Pub/sub, Data flow, Data Fusion, Cloud Function)
Nice to have exposure to Teradata
Solid experience with Job Orchestration Tools like Airflow and ability to build complex Jobs.
Writing and maintaining large Data Pipelines using Custom ETL framework
Ability to Automate Jobs using Python
Familiarity with Data Modeling techniques and Data Warehousing standard methodologies and practices
Very good experience with Code Version control repository like Github
Good Scripting skills, including Bash scripting and Python
Familiar with Scrum and Agile methodologies
Problem solver with strong attention to detail and excellent analytical and communication skills
Ability to work in Onsite / Offshore model and able to lead a Team.
Best Regards,
Vishal
vishal@smartcaliber.com
www.smartcaliber.com
Truth Lies in Heart
Duration: 12 Months
Location: Middletown Township, NJ (Onsite)
Responsibilities
Contribute to the migration of legacy data warehouse to a google cloud-based data warehouse for a Telecom Major..
Collaborate with Data Product Managers, Data Architects to design, implement, and deliver successful data solutions
Help architect data pipelines for the underlying data warehouse and data marts
Design and develop very complex ETL pipelines in Google cloud Data environments.
Our legacy tech stack includes Teradata and new tech stack includes GCP Cloud Data Technologies like BigQuery and Airflow and languages include SQL , Python
Maintain detailed documentation of your work and changes to support data quality and data governance
Support QA and UAT data testing activities
Support Deployment activities to higher environments
Ensure high operational efficiency and quality of your solutions to meet SLAs and support commitment to our customers (Data Science, Data Analytics teams)
Be an active participant and advocate of agile/scrum practice to ensure health and process improvements for your team
Basic Qualifications
8 years of data engineering experience developing large data pipelines in very complex environments
Very Strong SQL skills and ability to build Very complex transformation data pipelines using custom ETL framework in Google BigQuery environment
Exposure to Teradata and ability to understand complex Teradata BTEQ scripts
String Python programming skills
Strong Skills on build Airflow Jobs and Debug issues
Ability to Optimize the Query in BigQuery
Hands-on experience on Google Cloud data Technologies (GCS , BigQuery, Dataflow, Pub sub, Data Fusion , Cloud Function)
Preferred Qualifications
Experience with cloud data warehouse technology BigQuery.
Nice to have experience with Cloud technologies like GCP (GCS , Data Proc, Pub/sub, Data flow, Data Fusion, Cloud Function)
Nice to have exposure to Teradata
Solid experience with Job Orchestration Tools like Airflow and ability to build complex Jobs.
Writing and maintaining large Data Pipelines using Custom ETL framework
Ability to Automate Jobs using Python
Familiarity with Data Modeling techniques and Data Warehousing standard methodologies and practices
Very good experience with Code Version control repository like Github
Good Scripting skills, including Bash scripting and Python
Familiar with Scrum and Agile methodologies
Problem solver with strong attention to detail and excellent analytical and communication skills
Ability to work in Onsite / Offshore model and able to lead a Team.
Best Regards,
Vishal
vishal@smartcaliber.com
www.smartcaliber.com
Truth Lies in Heart