Java+Big Data
Work exp - 5 to 7 yrs
Key Responsibilities
Design, develop, and maintain Java-based data processing applications.
Build and optimize Big Data pipelines using technologies such as Hadoop, Spark, Hive, HBase, Kafka etc.
Work on ETL processes , data ingestion, data transformation, and data quality improvements.
Implement and optimize distributed data processing solutions for large datasets.
Integrate Java applications with Big Data ecosystems (HDFS, Hive, Kafka, etc.).
Develop REST APIs and microservices using Java / Spring Boot for data access and processing.
Troubleshoot performance issues across complex data pipelines.
Collaborate with data engineers, architects, and analysts to understand business requirements.
Ensure best practices in coding, version control, CI / CD, and cloud-based deployments.
Write clean, maintainable, and scalable code following industry standards.
Required Skills
Java Skills
Strong proficiency in Core Java , OOPS, Collections, Multithreading
Experience with Spring / Spring Boot
Hands-on experience in building scalable backend systems
Good understanding of REST APIs and microservices
Big Data Skills
Experience with Hadoop ecosystem – HDFS, YARN, MapReduce
Strong knowledge of Apache Spark (RDD, DataFrame, Spark SQL)
Experience with Hive / Impala for querying large datasets
Knowledge of Kafka or other messaging / streaming platforms
Experience with NoSQL databases (HBase, Cassandra, MongoDB)
Additional Good-to-Have Skills
Experience with Airflow , NiFi, or similar orchestration tools
Exposure to AWS / Azure / GCP Big Data services
Familiarity with Docker / Kubernetes
Knowledge of data modeling, data warehousing concepts
Performance tuning and optimization for distributed systems
Software Engineer Java • Baddi, Himachal Pradesh, India