Job Title : Senior Data Architect
Location : Bangalore / Chandigarh
Job Type : Full-time
Experience : 10+ years
Job Summary are looking for an experienced Data Architect to lead the design, development, and optimization of our modern data infrastructure. The ideal candidate will have deep expertise in big data platforms, data lakes, lakehouse architectures, and hands-on experience with modern tools such as Spark clusters, PySpark, Apache Iceberg, the Nessie catalog, and Apache Airflow.
This role will be pivotal in evolving our data platform, including database migrations, scalable data pipelines, and governance-ready architectures for both analytical and operational use cases.
Key Responsibilities :
- Design and implement scalable and reliable data architectures for real-time and batch processing systems
- Evaluate and recommend data tools, frameworks, and infrastructure aligned with company goals
- Develop and optimize complex ETL / ELT pipelines using PySpark and Apache Airflow
- Architect and manage data lakes using Spark on Apache Iceberg and Nessie catalog for
versioned and governed data workflows
Perform data analysis, data profiling, data quality improvements, and data modelingLead database migration efforts, including planning, execution, and optimizationDefine and enforce data engineering best practices, data governance standards, and schemaevolution strategies
Collaborate cross-functionally with data scientists, analysts, platform engineers, and business Skills & Qualifications :10+ years of experience in data architecture, data engineering, data security, datagovernance, and big data platforms
Deep understanding of trade-offs between managed services and open-source data stacktools, including cost, scalability, operational overhead, flexibility, and vendor lock-in
Strong hands-on experience with PySpark for writing data pipelines and distributed data processingProven expertise with Apache Iceberg, Apache Hudi, and the Nessie catalog for modern table formats and versioned data catalogsExperience in scaling and managing Elasticsearch and PostgreSQL clustersStrong experience with Apache Airflow for workflow orchestration (or equivalent tools)Demonstrated success in database migration projects across multiple cloud providersAbility to perform deep data analysis and compare datasets between systemsExperience handling 100s of terabytes of data or moreProficiency in SQL, data modeling, and performance tuningExcellent communication and presentation skills, with the ability to lead technicalconversations
Nice to Have :
Experience in Sales, Marketing, and CRM domains, especially with Accounts and Contacts dataKnowledge in AI and vector databases.Exposure to streaming data frameworks (Kafka, Flink, etc.)Ability to support analytics and reporting initiativesWhy Join Us :
Work on cutting-edge data architectures using modern open-source technologiesBe part of a team transforming data operations and analytics at scaleOpportunity to architect high-impact systems from the ground upJoin a collaborative, innovation-driven culture(ref : hirist.tech)