Key Responsibilities :
Data Engineering & Development :
- Design, build, and maintain ETL / ELT pipelines using Hadoop ecosystem tools
- Write complex Hive queries for data transformation and analysis
- Work with HDFS for storage and efficient data access
- Develop UNIX shell scripts for automation of jobs and workflows
- Optimize SQL queries for performance and scalability
Data Processing & Integration :
Process large volumes of structured and semi-structured dataIntegrate data from various sources into Hadoop-based data lakesWork with cross-functional teams to understand data requirements and deliver solutionsMonitoring, Maintenance & Quality :
Monitor and troubleshoot production data pipelinesEnsure data quality, integrity, and consistency across systemsSupport data ingestion, batch processing, and job scheduling (e.g., Oozie, Airflow)Required Skills and Qualifications :
Bachelor's degree in Computer Science, IT, or related field3–7 years of hands-on experience with Big Data toolsStrong expertise in :Hadoop Distributed File System (HDFS)Hive (querying, optimization, partitioning)SQL (advanced queries, joins, aggregations)UNIX / Linux shell scriptingGood understanding of data warehousing concepts and large-scale data processingSkills Required
Unix, Linux, Oozie, Airflow, Hadoop