Company overview
360DigiTMG is a pioneering EdTech organization known for upskilling professionals in Data Science, AI, and emerging technologies through industry-aligned certification programs. With Hyderabad as its main India office, the company focuses on practical, market-relevant training that prepares learners for real-world roles in a dynamic business environment.
Key Responsibilities
- Deliver multi‑cloud data engineering training (fundamentals to advanced) with a primary focus on Azure and exposure to AWS and GCP.
- Teach batch, streaming, and CDC pipelines, as well as modern Lakehouse architectures.
- Build and demonstrate real‑world use cases in Azure, AWS, GCP, Databricks, and Snowflake.
- Conduct hands‑on labs, code reviews, architecture walkthroughs, and best‑practice sessions.
- Guide students on data modeling, orchestration, monitoring, and CI / CD for data pipelines.
- Mentor students for interviews, technical assessments, and industry project execution.
- Design and deliver modules on Apache Kafka covering fundamentals, real‑time streaming use cases, producers / consumers, topics / partitions, consumer groups, and integration with cloud data platforms and Spark.
- Implement Kafka‑based streaming pipelines end‑to‑end, including reliability, scalability, and monitoring best practices.
- Introduce and apply big data processing frameworks such as Apache Spark (with strong focus on PySpark) for ETL / ELT, batch and streaming workloads.
- Teach PySpark for data transformations, optimization, and integration with Delta Lake, Kafka, and cloud storage.
- Use Apache Airflow (and other orchestrators) to design, schedule, and monitor data workflows and demonstrate DAG best practices.
- Incorporate dbt into the modern data stack to teach SQL‑based transformations, modular data modeling, testing, and documentation in cloud DWH / Lakehouse environments.
- Continuously update course content, labs, and projects to reflect evolving best practices in cloud, big data, and Lakehouse ecosystems.
Core Technical Expertise
Cloud : Azure (primary), with working exposure to AWS and GCP.Azure : Azure Synapse Analytics, Azure Data Factory, Azure Databricks, Delta Lake, Unity Catalog, Event Hubs, ADLS Gen2, Key Vault, private links.Streaming & Messaging : Apache Kafka (core architecture, producers / consumers, consumer groups, partitions, delivery semantics), Kafka integrations with Spark / Databricks and cloud services.Big Data & Compute : Apache Spark with strong PySpark, exposure to big data ecosystems and performance tuning.Lakehouse & DWH : Databricks Lakehouse, Snowflake (Streams, Tasks, governance, time travel), exposure to BigQuery / Redshift.Data Modeling : Star schema, Snowflake schema, SCD‑2, Data Vault 2.0.Orchestration : Azure Data Factory, AWS Glue Workflows, Apache Airflow, Databricks Workflows.Transformation Layer : dbt for modular SQL transformations, testing, and documentation on top of DWH / Lakehouse.CI / CD : Azure DevOps / GitHub Actions, secret management, and deployment workflows for data pipelines and analytics code.Preferred Experience
Hands‑on experience with Lakehouse and DWH platforms in production environments.Experience mentoring, teaching, or delivering corporate training in data engineering, cloud, and big data technologies.Experience building and maintaining Kafka‑based streaming pipelines and Spark / PySpark workloads.Experience using Airflow and / or dbt in real‑world projects.Relevant cloud and / or Kafka / Spark certifications are an added advantage.Outcome
Students gain real‑world, job‑ready skills in :
Cloud data pipelines (batch, ELT, CDC) across Azure and other clouds.Lakehouse architectures with Databricks, Delta Lake, and Snowflake.Real‑time streaming with Kafka and Spark / PySpark.Big data frameworks and orchestration using Airflow and other tools.Analytics engineering and transformation best practices using dbt.Governance, CI / CD, monitoring, and end‑to‑end project execution, along with strong interview preparation tailored to data engineering and cloud roles.