Job Description
About Koantek :
Koantek is a Databricks Pure-Play Elite Partner, helping enterprises modernize faster and unlock the full power of Data and AI. Backed by Databricks Ventures and honored as a six-time Databricks Partner of the Year, we enable global enterprises to modernize at speed, operationalize AI, and realize the full value of their data. Our deep expertise spans industries such as healthcare, financial services, retail, and SaaS, delivering end-to-end solutions from rapid prototyping to production-scale AI deployments. We deliver tailored solutions that enable businesses to leverage data for growth and innovation. Our team of experts utilizes deep industry knowledge combined with cutting-edge technologies, tools, and methodologies to drive impactful results. By partnering with clients across a diverse range of industries—from emerging startups to established enterprises—we help them uncover new opportunities and achieve a competitive advantage in the digital age.
About the Role
As a Solutions Architect at Koantek, you will collaborate with customers to design scalable data architectures utilizing Databricks technology and services. The RSA at Koantek builds secure, highly scalable big data solutions to achieve tangible, data-driven outcomes all the while keeping simplicity and operational effectiveness in mind. Leveraging your technical expertise and business acumen, you will navigate complex technology discussions, showcasing the value of the Databricks platform throughout the sales process. Working alongside Account Executives, you will engage with customers' technical leaders, including architects, engineers, and operations teams, aiming to become a trusted advisor who
delivers concrete outcomes. This role collaborates with teammates, product teams, and cross-functional project teams to lead the adoption and integration of the Databricks Platform into the enterprise ecosystem and AWS / Azure / GCP architecture.
The impact you will have :
- Develop Account Strategies : Work with Sales and other essential partners to
develop strategies for your assigned accounts to grow their usage of the Databricks
platform.
Establish Architecture Standards : Establish the Databricks Lakehouse architecture as the standard data architecture for customers through excellent technical account planning.Demonstrate Value : Build and present reference architectures and demo applications to help prospects understand how Databricks can be used to achieve their goals and land new use cases.Capture Technical Wins : Consult on big data architectures, data engineering pipelines, and data science / machine learning projects to prove out Databricks technology for strategic customer projects. Validate integrations with cloud servicesand other third-party applications.
Promote Open-Source Projects : Become an expert in and promote Databricks-inspired open-source projects (Spark, Delta Lake, MLflow) across developer
communities through meetups, conferences, and webinars.
Requirements
Technical Expertise :
Experience translating a customer's business needs to technology solutions, including establishing buy-in with essential customer stakeholders at all levels of the business.Experienced at designing, architecting, and presenting data systems for customers and managing the delivery of production solutions of those data architectures.Projects delivered with hands-on experience in development on databricksExpert-level knowledge of data frameworks, data lakes and open-source projects such as Apache Spark, MLflow, and Delta LakeExpert-level hands-on coding experience in Spark / Scala, Python or PysparkIn depth understanding of Spark Architecture including Spark Core, Spark SQL, Data Frames, Spark Streaming, RDD caching, Spark MLibT / event-driven / microservices in the cloudDeep experience with distributed computing with spark with knowledge of spark runtimeExperience with private and public cloud architectures, pros / cons, and migration considerations.Extensive hands-on experience implementing data migration and data processing using AWS / Azure / GCP servicesFamiliarity with CI / CD for production deploymentsFamiliarity with optimization for performance and scalabilityCompleted data engineering professional certification and required classesSQL Proficiency : Fluent in SQL and database technologyEducational Background
Degree in a quantitative discipline (Computer Science, Applied Mathematics, Operations Research.Relevant certifications (e.g., Databricks certifications, AWS / Azure / GCP AI / ML certifications) are a plus. Workplace FlexibilityOn-site presence at customer locations MAY be required based on the project and business needs. Candidates should be willing and able to travel for short or medium-term assignments when necessary.