Job Description
Job Description
Essential Duties and Responsibilities :
- Data Pipeline Development :
Design, build, and maintain scalable ETL / ELT pipelines to process large datasets efficiently.
Leverage Python for scripting and orchestration tasks.Database Management :Develop and optimize queries and schemas in ClickHouse and SQL databases.
Ensure performance for high-volume data operations, including tuning for performance, correctness and efficient storage.Support data integration efforts combining ClickHouse, MS SQL Server and Databricks.Cloud Infrastructure :Deploy and manage data workflows and applications on Azure cloud services, Docker and Python orchestration tools.
Implement solutions for data storage, processing, and analytics using Azure-native and on-prem solutions.Data Orchestration and Automation :Use Python-based orchestration tools (e.g., Apache Airflow, Dagster, or Prefect) to schedule and monitor workflows.
Automate repetitive tasks and maintain robust workflows with fault-tolerance and scalability.Manage containerized applications for deployments and CI / CD pipelines.Collaborate with cross-functional teams to utilize Databricks for big data processing and advanced analytics.Documentation and Best Practices :Document processes, configurations, and troubleshooting guides.
Enforce best practices for coding, testing, and deployment across data engineering tasksTechnical Skills :
Strong Proficiency in Python for scripting, data manipulation, and orchestration.Strong knowledge of SQL database management and advanced SQL skills.Experience with Azure services (e.g., Azure Databricks, Blob Storage).Hands-on experience with Docker for containerization.Familiarity with Grafana for system monitoring and visualization.Working knowledge of Databricks for big data and machine learning workflows.Education and / or Work Experience Requirements :
Bachelor’s or Master’s degree in Computer Science, Data Engineering, or a related field.Strong problem-solving skills and a deep understanding of data architecture principles.Ability to manage multiple priorities and work effectively in a collaborative environment.Excellent communication and documentation skills.Preferred Qualifications :
Experience with ClickHouse, OLAP DatabasesKnowledge of distributed systems and data modeling best practices.Experience with PV modeling tools and techniques (e.g., PVSyst, PVlib, SAM, and / or similar software)Experience with AVEVA and IBM MaximoPhysical Requirements / Work Environment : Work is performed in an office environment. On occasion, the position may require an in-person site visit. Must be able to sit, stand, bend, twist and lift up to 20 pounds.
This job description is intended to describe the essential job functions of this position and is not intended to be an all-inclusive statement of job responsibilities.
I have read, understand and am able to perform the duties within the job description. I have received a copy of this form.