About the Company
We are seeking an accomplished Data Solutions Architect with deep experience in designing, implementing, and optimizing modern cloud data platforms. The ideal candidate will expertly leverage Snowflake, Databricks, DBT, Airflow, Python, PySpark, and have hands-on exposure to advanced AI / ML and Generative AI (GenAI) solutions in real-world enterprise environments. You will partner with business and tech leaders to deliver scalable, innovative data products, architectures, and analytics capabilities.
About the Role
We are looking for a Data Solutions Architect who will be responsible for architecting and implementing scalable data platforms and solutions.
Responsibilities
- Architect and implement scalable data platforms, pipelines, and solutions using Snowflake, Databricks, DBT, Airflow, Python, and PySpark
- Lead cloud data migration projects, designing robust, secure, and reusable analytics architectures
- Develop and optimize ELT / ETL workflows for batch and real-time processing
- Design dimensional models, data lakes, data marts, and advanced analytics infrastructures
- Oversee and mentor data engineering teams in coding standards, CI / CD, and best practices
- Integrate AI / ML models and GenAI capabilities into data products and workflows (e.g., predictive analytics, intelligent automation)
- Collaborate with business stakeholders to define requirements, translating needs into technical architectures
- Manage platform cost optimization, data quality / governance, and performance tuning
- Evaluate and recommend emerging tech, tools, and architectures to drive innovation
Qualifications
Bachelor’s or Master’s degree in Computer Science, Data Engineering, Information Systems, or related field7+ years of professional experience in data architecture, engineering, or analytics rolesAdvanced hands-on experience with Snowflake (data warehousing, modeling, optimization)Deep expertise in Databricks (Spark, ML workflows, lakehouse architectures) and PySparkProficient with DBT for data transformation and modelingExpert in building robust workflows with Apache AirflowStrong programming skills in Python (including data and ML libraries)Proven track record implementing AI / ML solutions with at least basic exposure to GenAI (LLMs, prompt engineering, AI API integration)Cloud platform experience (Azure, AWS, or GCP)Excellent problem-solving, communication, and stakeholder management skillsRequired Skills
Experience with Terraform, CI / CD for data engineeringFamiliarity with data governance, privacy, and security best practicesPrior experience in client-facing solution architect or enterprise delivery rolesCertifications in Snowflake, Databricks, Python, or cloud platformsPublished work or demonstrated portfolio of data / AI projectsUnderstanding of modern API integration and real-time analyticsEqual Opportunity Statement
We are committed to diversity and inclusivity.
```