Ready to build the future with AI?
At Genpact, we don’t just keep up with technology—we set the pace. AI and digital innovation are redefining industries, and we’re leading the charge. Genpact’s AI Gigafactory, our industry-first accelerator, is an example of how we’re scaling advanced technology solutions to help global enterprises work smarter, grow faster, and transform at scale. From large-scale models to agentic AI, our breakthrough solutions tackle companies’ most complex challenges.
If you thrive in a fast-moving, innovation-driven environment, love building and deploying cutting-edge AI solutions, and want to push the boundaries of what’s possible, this is your moment.
Genpact (NYSE : G) is an advanced technology services and solutions company that delivers lasting value for leading enterprises globally. Through our deep business knowledge, operational excellence, and cutting-edge solutions – we help companies across industries get ahead and stay ahead. Powered by curiosity, courage, and innovation, our teams implement data, technology, and AI to create tomorrow, today. Get to know us at genpact.com and on LinkedIn, X, YouTube, and Facebook.
Inviting applications for the role of Principal Consultant - Big Data Admin
This key role demands a highly motivated individual with a strong background in Computer Science / Software Engineering. You are meticulous, thorough and possess excellent communication skills to engage with all levels of our stakeholders. A self-starter, you are up-to-speed with the latest developments in the tech world.
Responsibilities :
- Lead CDP platform upgrades and migrations, with strong hands-on execution and documentation from planning to go-live.
- Administer and tune Hadoop ecosystem services :
o Core : HDFS, YARN, Hive, Hue, Impala, Sqoop, Oozie
o Streaming : Apache Kafka (broker / topic ops), Apache Flink (streaming jobs)
o NoSQL / Query : HBase, Phoenix
o Security : Kerberos, Ranger, LDAP, TLS
Manage Cribl Stream deployments : build, configure, secure, and optimize data routing pipelines.Monitor and optimize platform performance using Cloudera Manager, NewRelic, BigPanda, Prometheus, Grafana, or any other observability tools.Design and implement backup, recovery, HA, and DR strategies for critical data infrastructure.Automate platform operations using Python, Bash / Shell, Scala, and CI / CD workflows.Work cross-functionally with Data Engineers, DevOps, InfoSec, and Cloud Engineering teams to support data pipeline reliability and scalability.Manage deployments using Docker, Kubernetes, Jenkins, Bitbucket, and optionally Ansible or GitOps practices.Support and maintain cloud-native or hybrid deployments, especially in GCP (Anthos) environments.Produce and maintain robust architecture documentation, runbooks, and operational SOPs.Qualifications we seek in you!
Minimum qualifications :
Bachelor's / Graduation / Equivalent : BE / B- Tech, MCA, MBAExcellent Communication skill and effectively interact with business userAbility to interact with business as well as technical teamsexperience in Big Data infrastructure, administration, and operations.Proven Cloudera CDP (7.x) experience, including production-grade migrations (7.1.6 to 7.1.9+).Deep expertise in :o Apache Spark – job tuning, executor / resource optimization
o Apache Kafka – security (SASL_SSL, GSSAPI), scaling, topic lifecycle management
o Apache Flink – real-time stream processing in HA environments
o Cribl Stream – full-lifecycle management and observability integration
o HBase & Phoenix – schema evolution, read / write tuning, replication
Scripting & Automation : Proficient in Python, Shell (Bash), and optionally ScalaSecurity-first mindset : Working knowledge of Kerberos, Ranger policies, LDAP integration, and TLS configuration.DevOps Experience : Hands-on with Docker, Kubernetes, Jenkins, Bitbucket, and monitoring tools like Grafana / Prometheus.Comfortable supporting large-scale, multi-tenant environments and production on-call rotationsPreferred qualifications :
Very good written and presentation / verbal communication skills with experience of customer interfacing role.In-depth requirement understanding skills with good analytical and problem-solving ability, interpersonal efficiency, and positive attitude.Cloudera Certified Administrator (CCA) or equivalent industry certification.Experience with BD on-prem, cloud and hybrid data infrastructure, particularly Google Cloud Platform (GCP) and Anthos clusters.Why join Genpact?
Lead AI-first transformation – Build and scale AI solutions that redefine industriesMake an impact – Drive change for global enterprises and solve business challenges that matterAccelerate your career—Gain hands-on experience, world-class training, mentorship, and AI certifications to advance your skillsGrow with the best – Learn from top engineers, data scientists, and AI experts in a dynamic, fast-moving workplaceCommitted to ethical AI – Work in an environment where governance, transparency, and security are at the core of everything we buildThrive in a values-driven culture – Our courage, curiosity, and incisiveness - built on a foundation of integrity and inclusion - allow your ideas to fuel progressCome join the 140,000+ coders, tech shapers, and growth makers at Genpact and take your career in the only direction that matters : Up.
Let’s build tomorrow together.