What are the responsibilities and job description for the Senior Data Engineer, Analytics position at CoreWeave?
About the Team
The Data Engineering team builds foundational datasets and analytics services that enable BI and data science across CoreWeave. We seek to democratize insights and foster a culture where data-driven decision-making thrives at every level.
About the Role
We’re seeking a skilled Senior Data Engineer to lead the development of foundational data models that empower our Business Intelligence Engineers, analysts, and data scientists to efficiently work with and gain insights from our data. This role will own the creation and maintenance of star and snowflake schemas within our lakehouse environment and set the standards for dimensional modeling best practices. The engineer will also create and optimize key datasets and metrics essential to tracking business health.
Responsibilities
- Develop and maintain data models, including star and snowflake schemas, to support analytical needs across the organization.
- Establish and enforce best practices for dimensional modeling in our Lakehouse.
- Engineer and optimize data storage using analytical table/file formats (e.g., Iceberg, Parquet, Avro, ORC).
- Partner with BI, analytics, and data science teams to design datasets that accurately reflect business metrics.
- Tune and optimize data in MPP databases such as StarRocks, Snowflake, BigQuery, or Redshift.
- Collaborate on data workflows using Airflow, building and managing pipelines that power our analytical infrastructure.
- Ensure efficient processing of large datasets through distributed computing frameworks like Spark or Flink.
Qualifications
- You thrive in a fast-paced, complex, work environment and you love tackling hard problems.
- Hands-on experience applying Kimball modeling principles to large datasets.
- Expertise in working with analytical table/file formats, including Iceberg, Parquet, Avro, and ORC.
- Proven experience optimizing MPP databases (StarRocks, Snowflake, BigQuery, Redshift).
- 5 years of programming experience in Python or Scala.
- Advanced SQL skills, with a strong ability to write, optimize, and debug complex queries.
- Hands-on experience with Airflow for batch orchestration distributed computing frameworks like Spark or Flink.
Our compensation reflects the cost of labor across several US geographic markets. The base pay for this position ranges from $175,000-$205,000. Pay is based on a number of factors including market location and may vary depending on job-related knowledge, skills, and experience.
Salary : $175,000 - $205,000