About the Role :
We are looking for a highly skilled Data Engineer to join our clients Big Data team. The ideal candidate will have deep expertise in Hadoop, Spark, and modern data ecosystems, and will be responsible for designing, building, and optimizing data pipelines that power advanced analytics and large-scale processing.
You will collaborate closely with cross-functional teams, take ownership of key deliverables, and contribute to the evolution of data infrastructure and best practices within a fast-paced environment.
Key Responsibilities :
- Design, develop, and maintain robust big data pipelines and ETL workflows.
- Work extensively on Hadoop (Cloudera), Spark, Hive, Pig, and Impala to manage and analyze large-scale data sets.
- Optimize data processes for performance, scalability, and reliability.
- Collaborate with data scientists, analysts, and other engineers to deliver end-to-end data solutions.
- Troubleshoot and resolve complex data, networking, and software-related issues.
- Contribute to architecture design, data modeling, and integration strategies.
- Ensure best practices in data security, governance, and quality.
- Plan and organize work effectively, ensuring on-time delivery of assigned tasks.
- Mentor junior team members and support collective project goals.
Required Skills and Experience :
7- 10 years of hands-on experience as a Big Data Engineer.Strong proficiency in Hadoop ecosystem tools Cloudera, Spark, Hive, Pig, Impala.Solid programming background with one or more of the following : Java, C++, Python, R, Ruby, PHP, or Linux scripting.Proven ability to design and maintain data lakes, warehouses, and distributed systems.Expertise in data querying, data transformation, and performance tuning.Experience in troubleshooting complex networking and software integration issues.Strong analytical, problem-solving, and documentation skills.Excellent interpersonal and communication skills with a collaborative mindset.Preferred Qualifications :
Prior experience working with cloud-based data platforms (AWS, Azure, or GCP).Familiarity with DevOps practices and CI / CD pipelines for data deployments.Understanding of data governance frameworks and metadata management.Personal Attributes :
Self-motivated and proactive in driving projects to completion.Exceptional organizational and multitasking abilities.Strong commitment to quality, efficiency, and teamwork.Ability to thrive in a fast-paced, dynamic remote environment.Additional Information :
Contract Type : 6-Month Contract (extendable based on performance and project needs)Work Setup : Fully RemoteEngagement Type : Individual Contributor(ref : hirist.tech)