Job Title: Java Developer with Data Analytics (Scala & Databricks)
Location: Boston, MA
PP is mandatory
Job Summary:
We are looking for a Java Developer with strong exposure to Big Data, Scala, and Databricks to work on large-scale data engineering and analytics initiatives. The role requires blending Java-based backend logic with modern cloud-based data pipelines.
Key Responsibilities:
• Develop backend components and microservices using Java (Spring Boot / REST APIs).
• Build scalable data processing pipelines using Databricks and Apache Spark (Scala/PySpark).
• Integrate APIs and real-time data ingestion flows with Kafka, Azure Data Lake, or AWS S3.
• Work with data teams to implement data models, transformations, and cleansing logic.
• Optimize Spark jobs for performance, cost, and reliability in Databricks.
• Collaborate with DevOps to deploy scalable data and microservices solutions using CI/CD pipelines.
• Write efficient SQL queries and scripts for data validation and transformation.
Must-Have Skills:
• 4–8 years of hands-on experience in Java development (Spring Boot / RESTful APIs).
• Strong command of Scala, particularly for Spark-based data processing.
• Deep experience with Databricks, including Notebooks, Delta Lake, and Spark optimizations.
• Experience with distributed data systems (HDFS, Hive, Spark).
• Familiarity with cloud platforms – preferably Azure or AWS.
• Strong skills in SQL, data modelling, and structured/semi-structured data handling.
• Good understanding of version control (Git) and Agile methodologies.