Job Title : Databricks Engineer
Experience : 7.5+ Years
Location : Remote (IST Shift)
Contract Type : Long-Term
Overview
We are seeking an experienced Databricks Engineer with 7.5+ years of expertise in database development, ETL, and cloud-based data solutions . The ideal candidate should be highly skilled in SQL, Databricks, PySpark, and cloud technologies , with proven experience in designing scalable data workflows, optimizing performance, and delivering enterprise-grade data solutions.
Responsibilities
- Design, develop, and optimize database solutions for enterprise applications.
- Build and manage ETL pipelines and large-scale data processing workflows using Databricks and PySpark .
- Implement data modeling, schema design, and performance tuning for high-volume data systems.
- Develop and maintain cloud-based workflows with AWS (S3, Redshift, Lambda, EC2).
- Automate ETL processes and improve system efficiency through workflow automation .
- Perform query optimization, indexing, and stored procedures for database performance improvements.
- Lead data migration and warehousing projects , ensuring smooth data integration.
- Develop dashboards and analytics using Power BI and Google Data Studio.
Requirements
7.5+ years of hands-on experience in data engineering, ETL, and databases .Proven expertise in Databricks, PySpark, and Snowflake .Strong knowledge of AWS cloud services (S3, Lambda, EC2).Experience with Active Batch or similar scheduling tools .Excellent problem-solving, analytical, and communication skills.Ability to work in fast-paced, remote environments .Key Skills
Databases : Oracle PL / SQL, MySQL, SQL Server, Redshift, PostgreSQL, Netezza
Programming & Tools : Python, Unix Shell Scripting, PySpark, Databricks, SnapLogic, Snowflake, SQL Developer, PL / SQL Developer, MySQL Workbench, DBeaver, PyCharm
Cloud & Scheduling : AWS S3, AWS Lambda, AWS EC2, Active Batch
Data Analytics : Power BI, Google Data Studio