What are the responsibilities and job description for the Google Cloud Platform Data Architect @ Nashville, TN (Onsite) position at Sage IT Inc?
Job Details
Primary Skill -
- Google Cloud Platform Data engineering services Big Query , Airflow, Data Proc , Data Flow
- Programming Language Python / Pyspark
- Data warehousing knowledge and good with SQL skills . Working experience on data migration from On-premise to Cloud
Roles and Responsibilities:
Data Engineering, Design and Development:
Collaborate with cross-functional teams to design and implement scalable and reliable systems on Google Cloud Platform considering optimal performance, security, and cost-effectiveness.
Build Data ingestion pipelines to extract data from various sources (Azure Blob, Azure SQL, Flat files, Semi structure sources, AWS S3) into the data warehouse in Google Cloud Platform.
Utilize Google Cloud Platform services to build robust and scalable data solutions.
Design, develop, and maintain data pipelines and implement data architecture on Google Cloud Platform using services such as Big Query, Cloud Storage, and Cloud Composer.
Expertise in the tools and technology that helps in the process of data collection, cleaning, transforming, and modelling data to achieve useful information.
Leveraging Google Cloud Platform capabilities and technologies for migrating existing databases to cloud.
Collaborate with cross-functional teams to understand data requirements and implement scalable solutions.
Implement and optimize Big Query tables and Complex SQL queries for efficient data retrieval, performance, and efficiency.
Experience in Data Migration from On-premises Database to Big query and experience in BQ conversion
Experience and knowledge in building data pipelines and scheduling using Cloud Composer (Airflow) and data and file transformation using Python
EDW (Enterprise Data Warehouse) and Data Model Designing:
Experience with Data modelling, Data warehousing and ETL processes
Work closely with Business and analysts to design and implement data models for effective data representation and analysis.
Ensure data models meet industry standards and compliance requirements in the health-care domain.
Contribute to the design and development of the enterprise data warehouse architecture.
Implement best practices for data storage, retrieval, and security within the EDW.
Health-care Domain Knowledge:
Apply domain-specific knowledge to ensure that data solutions comply with health-care industry regulations and standards.
Stay updated on industry trends and advancements in health-care data management.
Collaboration:
Work collaboratively with cross-functional teams, including Business Teams, analysts, and software engineers, to deliver integrated and effective data solutions.
Participate in code reviews and provide constructive feedback to team members
Qualifications (Required Technical skills/Experience):
1.Bachelor's degree in computer science, Information Technology, or a related field.
2. Proficiency in Google Cloud Platform and in-depth knowledge of Google Cloud Platform services, Big Query, Cloud Functions and Composer
3. Strong programming skills in Data Engineering Python
4. Experience with data modeling, SQL, and EDW design.
6. Excellent problem-solving and analytical skills.
7.Strong communication and collaboration skills.
8. Proficiency in version control systems, particularly Git
9. Strong understanding of Datawarehouse concepts, data lakes