Design and implement scalable batch and real-time data pipelines using Apache Iceberg, AWS services, and streaming technologies bi-directional data synch mechanisms.
Lead data migration projects from Oracle to cloud-native data lakes and warehouses.
Architect and manage real-time data ingress and egress using tools like Kafka, Kinesis, or Flink.
Build and maintain data abstraction layers (e.g., APIs, semantic layers) to enable secure and simplified data access.
Optimize data storage, partitioning, and schema evolution strategies in Iceberg.
Leverage Presto / Trino for federated querying and interactive analytics across heterogeneous data sources.
Collaborate with business, data scientists, analysts, and engineers to deliver high quality, reliable data solutions.
Implement robust monitoring, alerting, and data quality frameworks.
Required Qualifications :
Proven expertise in data migration from Oracle, including schema translation and performance optimization.
Proficiency in SQL, Python, and Spark.
Experience with Presto or Trino for distributed querying and federated data access.
Familiarity with building data abstraction mechanisms ,API, Stored Procedures.
Experience with CI / CD, infrastructure-as-code (Terraform, CloudFormation), and DevOps practices.
Strong understanding of data governance, security, and compliance in cloud environments.