Job description
The Role
You'll maintain our data platform, working with pioneering AWS technologies and industry-standard tools like Databricks to build robust, scalable data pipelines and integrations. This isn't just about writing code – it's about creating systems that transform raw data into actionable insights and provisioning of data to core applications. You'll work on a rotation base with another engineer to provide 16 hours of coverage to the distributed team.
Responsibilities :
- Maintain and improve IT systems infrastructure, ensuring flawless operation and integration
- Develop, implement, and optimize data pipelines and integrations using industry-standard tools
- Automate data acquisition processes and optimize data delivery for real-time analytics
- Collaborate with cross-functional teams across various time zones to ensure seamless operation
- Provide timely updates and recommendations on operational issues and improvements
- Develop customized SQL queries for database solutions and ad hoc requests
- Ensure data integrity and fidelity across all systems
Requirements :
Bachelor's degree in computer science or information technology3+ years of experience in IT, including 2 years in data integration and pipeline development using AWS S3, Redshift, Athena ecosystemExtensive experience with SQL Queries, Python development, and common Python librariesProficiency in PySpark / Databricks Delta Lake and AWS Cloud data integrationProven ability to manipulate, process, and extract value from large datasetsExcellent communication skills and strong leadership capabilitiesHigh energy level and dedication to Thermo Fisher Scientific's 4i valuesPreferred Qualifications :
Experience in Life Sciences or Service organizationsFamiliarity with data mining, data science, and predictive analyticsExperience with visualization tools such as Microsoft Power BISkills Required
Pyspark, Sql, Python