Description : Responsibilities :
- Design and develop robust data pipelines for ingestion, transformation, and storage.
- Implement data integration between various structured and unstructured data sources.
- Work with Java-based frameworks to develop scalable backend systems for data management.
- Optimize data systems for performance, reliability, and cost efficiency.
- Collaborate with analysts, data scientists, and application developers to deliver reliable datasets.
- Ensure data governance, quality, and security across all environments.
- Develop APIs and services for data access and analytics.
Technical Skills :
Proficient in Core Java, Java 8+, Spring Boot, and RESTful APIs.Strong understanding of SQL / NoSQL databases such as MySQL, PostgreSQL, MongoDB, Cassandra.Hands-on experience with Big Data tools : Apache Spark, Kafka, Hadoop, or Flink.Proficiency in ETL / ELT processes and data modeling.Experience working with cloud platforms (AWS, GCP, or Azure).Familiar with version control (Git) and CI / CD pipelines.Strong understanding of data structures, algorithms, and performance tuning.Preferred Qualifications :
Exposure to microservices architecture and containerization (Docker, Kubernetes).Experience with Python or Scala for data scripting.Working knowledge of Airflow, Databricks, or similar orchestration tools.Excellent problem-solving and communication skills.Educational Qualification :
Bachelors or Masters in Computer Science, Software Engineering, or related discipline.(ref : hirist.tech)