Big Data Support Engineer (Scala/Spark)
Experience: 2+ Years
Location: Kochi, Kerala
Notice Period- Immediate to 30 Days
Role Overview
We are looking for a Big Data Support Engineer with a strong foundation in functional programming and distributed computing. You will be responsible for building, maintaining, and optimizing large-scale data processing pipelines. The ideal candidate has hands-on experience in Scala and Spark to transform raw data into actionable insights.
Key Responsibilities
Data Pipeline Development: Design and develop scalable ETL/ELT pipelines using Spark or Scala.
Data Processing: Process structured and unstructured data at scale, ensuring high performance and data accuracy.
Optimization: Debug and tune Spark jobs for performance, memory management, and resource utilization.
Collaboration: Work closely with Data Architects and Analysts to understand requirements and implement data models.
Technical Requirements
Core Skills: Minimum 2 years of professional experience in Big Data technologies.
Programming: Strong proficiency in Scala (preferred) or Spark.
Frameworks: Extensive experience with Apache Spark (Spark SQL, Spark Streaming) is preferred
Ecosystem: Familiarity with Hadoop, Hive, or Presto.
Storage: Experience with file formats like Parquet, Avro, or ORC.
Cloud: Good to have exposure to AWS (Glue, EMR), Azure (Databricks), or GCP.