What are the responsibilities and job description for the Hadoop /Spark Developer with strong Java position at Centraprise?
Location: Charlotte, NC / Newark, DE / Dallas, TX (Onsite from Day 1)
Job Description:
Job Summary:
- Seeking talented and experienced Hadoop and Spark Developer with strong Java expertise to join data engineering team.
- The ideal candidate will have a solid understanding of big data technologies, hands-on experience with the Hadoop ecosystem, and the ability to build and optimize data pipelines and processing systems using Spark and Java.
Key Responsibilities:
- Develop, test, and deploy scalable big data solutions using Hadoop and Spark.
- Write efficient and optimized code in Java to process large datasets.
- Design and implement batch and real-time data processing pipelines using Spark.
- Monitor, troubleshoot, and enhance the performance of Spark jobs.
- Work closely with cross functional teams to integrate big data solutions into existing systems.
- Debug and resolve complex technical issues related to distributed computing.
- Collaborate on system architecture and contribute to technical design discussions.
Required Skills:
- Strong expertise in Java, with experience in writing optimized, high-performance code.
- Solid experience in Hadoop ecosystem (HDFS, Hive, Apache Spark (RDD, Dataframe, Dataset, Spark SQL, Spark Streaming).
- Proficiency in designing and building ETL pipelines for big data processing.
- Experience with query optimization and data manipulation using SQL-based technologies like Hive or Impala.
- Hands on experience with Git or similar version control systems.
- Strong understanding of Linux/Unix based environments for development and deployment.
Preferred Skills:
- Experience with Apache Kafka.
- Exposure to DevOps practices, including CI/CD pipelines.
- Knowledge of Python or Scala is a plus.
"Centraprise is an equal opportunity employer. Applicants must be authorized to work in the U.S. U.S. citizens and Green Card holders are strongly encouraged to apply."