What are the responsibilities and job description for the Senior Data Ops Engineer position at hackajob?
hackajob is collaborating with GSK to connect them with exceptional tech professionals for this role.
https://youtu.be/OBOCZaA1HPE
The Onyx Research Data Tech organization is GSK’s Research data ecosystem which has the capability to bring together, analyze, and power the exploration of data at scale. We partner with scientists across GSK to define and understand their challenges and develop tailored solutions that meet their needs. The goal is to ensure scientists have the right data and insights when they need it to give them a better starting point for and accelerate medical discovery. Ultimately, this helps us get ahead of disease in more predictive and powerful ways.
Onyx Is a Full-stack Shop Consisting Of Product And Portfolio Leadership, Data Engineering, Infrastructure And DevOps, Data / Metadata / Knowledge Platforms, And AI/ML And Analysis Platforms, All Geared Toward
Sr. Data Ops Engineers takes full ownership of delivering high-performing, high-impact biomedical and scientific data ops products and services, from a description of a pattern that customer Data Engineers are trying to use all the way through to final delivery (and ongoing monitoring and operations) of a templated project and all associated automation. They are standard-bearers for software engineering and quality coding practices within the team and are expected to mentor more junior engineers; they may even coordinate the work of more junior engineers on a large project. They devise useful metrics for ensuring their services are meeting customer demand and having an impact and iterate to deliver and improve on those metrics in an agile fashion.
Key Responsibilities
Our teams are building modern, cloud-native, DevOps-first systems for standardizing and templatizing biomedical and scientific data engineering for both internal and external stakeholders and the selected candidate will be responsible for:
Basic Qualifications
We are looking for professionals with these required skills to achieve our goals:
If you have the following characteristics, it would be a plus:
https://youtu.be/OBOCZaA1HPE
The Onyx Research Data Tech organization is GSK’s Research data ecosystem which has the capability to bring together, analyze, and power the exploration of data at scale. We partner with scientists across GSK to define and understand their challenges and develop tailored solutions that meet their needs. The goal is to ensure scientists have the right data and insights when they need it to give them a better starting point for and accelerate medical discovery. Ultimately, this helps us get ahead of disease in more predictive and powerful ways.
Onyx Is a Full-stack Shop Consisting Of Product And Portfolio Leadership, Data Engineering, Infrastructure And DevOps, Data / Metadata / Knowledge Platforms, And AI/ML And Analysis Platforms, All Geared Toward
- Building a next-generation, metadata- and automation-driven data experience for GSK’s scientists, engineers, and decision-makers, increasing productivity and reducing time spent on “data mechanics”
- Providing best-in-class AI/ML and data analysis environments to accelerate our predictive capabilities and attract top-tier talent
- Aggressively engineering our data at scale, as one unified asset, to unlock the value of our unique collection of data and predictions in real-time
Sr. Data Ops Engineers takes full ownership of delivering high-performing, high-impact biomedical and scientific data ops products and services, from a description of a pattern that customer Data Engineers are trying to use all the way through to final delivery (and ongoing monitoring and operations) of a templated project and all associated automation. They are standard-bearers for software engineering and quality coding practices within the team and are expected to mentor more junior engineers; they may even coordinate the work of more junior engineers on a large project. They devise useful metrics for ensuring their services are meeting customer demand and having an impact and iterate to deliver and improve on those metrics in an agile fashion.
Key Responsibilities
Our teams are building modern, cloud-native, DevOps-first systems for standardizing and templatizing biomedical and scientific data engineering for both internal and external stakeholders and the selected candidate will be responsible for:
- Design, build, and operate tools, services, workflows, etc that deliver high value through the solution to key business problems
- Be responsible for development of key components of a scalable research data platform supporting GSK scientists along with their collaborators
- Represent Onyx within the wider scientific community across internal and external collaborations
- Partner with Infra and DevOps team where modifications to underlying tools (e.g. infrastructure as code, Cloud Ops, DevOps, logging / alerting) are needed to serve new use-cases, and to ensure operations are planned
- Consult scientific users on application scalability to petabytes of data by having a deep understanding of software engineering, algorithms, and underlying hardware infrastructure and their impact on performance.
- Produce well-engineered software, including appropriate automated test suites, technical documentation, and operational strategy
- Ensure consistent application of platform abstractions to ensure quality and consistency with respect to logging and lineage
- Be fully versed in coding best practices and ways of working, and participates in code reviews and partnering to improve the team’s standards
- Provide leadership to team members to help others get the job done right.
Basic Qualifications
We are looking for professionals with these required skills to achieve our goals:
- Bachelor’s degree in Computer Science, Software Engineering or related field.
- 6 years of relevant work experience
- Cloud experience (e.g., AWS, Google Cloud, Azure, Kubernetes)
- Experience with DevOps principles and tools (e.g. GitOps, Azure DevOps, GitHub Actions, GitFlow ...),
- Experience with data Governance
- Programming experience in Python. Scala or Go
If you have the following characteristics, it would be a plus:
- Experience managing external engagements, technical architecture forums etc.
- Experience with agile software development
- Experience in workflow orchestration with tools such as Argo Workflow, Airflow, and scientific workflow tools such as Nextflow, Snakemake, VisTrails, or Cromwell
- Experience with specialized data architecture (e.g. optimizing physical layout for access patterns, including bloom filters, optimizing against self-describing formats such as ORC or Parquet, etc)
- Demonstrated experience building reusable components on top of the CNCF ecosystem including Kubernetes (or similar ecosystem)
- Establish scalable, automated processes for data engineering teams across GSK
- Thought leader and partner with wider Onyx data engineering teams to advise on implementation and best practices
- Observability (monitoring, alerting, logging, tracing, ...)
- Embedding agile software engineering (task/issue management, testing, documentation, software development lifecycle, source control, )
Salary : $154,000 - $215,000