Talent.com
This job offer is not available in your country.
Senior Big Data Engineer - ETL / Data Warehousing

Senior Big Data Engineer - ETL / Data Warehousing

BCFORWARD INDIA TECHNOLOGIES PRIVATE LIMITEDGurugram
30+ days ago
Job description

We are seeking a highly skilled Senior Big Data Engineer with 6+ years of relevant experience in Big Data, ETL / Data Warehousing, GCP, and Java.

The ideal candidate will design, develop, and optimize scalable, cloud-based real-time data pipelines and REST APIs using Java frameworks.

You will play a vital role in enabling data-driven decision-making by delivering innovative and reliable data solutions in collaboration with business and technical teams.

Key Responsibilities :

  • Design, build, and maintain automated, scalable, and reliable Big Data pipelines on Google Cloud Platform (GCP) leveraging tools such as BigQuery, Dataproc, Dataflow, and Pub / Sub
  • Develop and optimize ETL workflows for batch and real-time data processing using custom Python, Scala, or Cloud Data Functions
  • Design and implement RESTful APIs and microservices using Java and Spring Boot frameworks, incorporating security standards such as JWT / OAuth
  • Manage data storage solutions across relational databases including MySQL, PostgreSQL, MongoDB, and Cloud SQL
  • Collaborate with cross-functional teams including data analysts, business stakeholders, and DevOps engineers to deliver data solutions aligned with business needs
  • Implement CI / CD pipelines for data engineering workflows using tools like Cloud Build and GitHub Actions
  • Monitor and troubleshoot production data pipelines with Cloud Monitoring and Cloud Logging to ensure high availability and performance
  • Continuously explore and prototype emerging GCP technologies and third-party tools to enhance data infrastructure and analytics capabilities

Required Skills & Qualifications :

  • 6+ years of hands-on experience in Big Data engineering, ETL, and Data Warehousing
  • Strong expertise with GCP Big Data ecosystem : BigQuery, Dataproc, Dataflow, Pub / Sub, Cloud Storage, Compute Engine
  • Proficient in Java programming and experience building REST APIs using Spring Boot framework
  • Solid SQL skills and experience with complex queries and data transformations
  • Familiarity with container orchestration and deployment technologies such as Kubernetes (GKE) is a plus
  • Experience with CI / CD processes and DevOps tools for automated build, test, and deployment pipelines
  • Understanding of backend system design, database operations, security, and API authentication
  • Excellent problem-solving skills with a creative and innovative approach to data challenges
  • Strong collaboration and communication skills to work effectively across teams and with non-technical stakeholders
  • Bachelors degree in Computer Science, Engineering, or a related field
  • Preferred Skills :

  • Proficiency with Python or Scala for data pipeline development
  • Experience with monitoring and logging tools like Cloud Monitoring and Cloud Logging
  • Knowledge of containerization and cloud infrastructure management
  • Familiarity with message-driven architectures and event ingestion frameworks
  • Agile development experience
  • (ref : hirist.tech)

    Create a job alert for this search

    Big Data Engineer • Gurugram