Job Summary:
We are seeking a skilled Java Spark Developer to join our data engineering team. You will be responsible for building and optimizing large-scale data processing applications using Apache Spark and Java. The ideal candidate has hands-on experience with big data ecosystems, is detail-oriented, and enjoys working on complex, high-volume systems.
Key Responsibilities:
-
Develop and maintain scalable data processing pipelines using Apache Spark and Java
-
Optimize Spark jobs for performance, reliability, and scalability
-
Integrate with various data sources including HDFS, S3, Kafka, and relational databases
-
Collaborate with data scientists, analysts, and other engineers to understand data requirements
-
Ensure data quality and integrity across ETL processes
-
Implement logging, monitoring, and alerting for Spark jobs
-
Participate in code reviews and maintain high-quality code standards
-
Document system design, processes, and architecture
—