Position Summary
This is the Requisition for Employee Referrals Campaign and JD is Generic.
We are looking for Associates with 5+ years of experience in delivering solutions around Data Engineering, Big data analytics and data lakes, MDM, BI, and data visualization. Experienced to Integrate and standardize structured and unstructured data to enable faster insights using cloud technology. Enabling data-driven insights across the enterprise.
Job Responsibilities
- He / she should be able to design implement and deliver complex Data Warehousing / Data Lake, Cloud Data Management, and Data Integration project assignments.
- Technical Design and Development – Expertise in any of the following skills.
- Any ETL tools (Informatica, Talend, Matillion, Data Stage), and hosting technologies like the AWS stack (Redshift, EC2) is mandatory.
- Any BI tools among Tablau, Qlik & Power BI and MSTR.
- Informatica MDM, Customer Data Management.
- Expert knowledge of SQL with the capability to performance tune complex SQL queries in tradition and distributed RDDMS systems is must.
- Experience across Python, PySpark and Unix / Linux Shell Scripting.
- Project Management is must to have. Should be able create simple to complex project plans in Microsoft Project Plan and think in advance about potential risks and mitigation plans as per project plan.
- Task Management – Should be able to onboard team on the project plan and delegate tasks to accomplish milestones as per plan. Should be comfortable in discussing and prioritizing work items with team members in an onshore-offshore model.
- Handle Client Relationship – Manage client communication and client expectations independently or with support of reporting manager. Should be able to deliver results back to the Client as per plan. Should have excellent communication skills.
Education
Bachelor of Technology
Master's Equivalent - Engineering
Work Experience
Overall, 3- 7 years of relevant experience in Data Warehousing, Data management projects with some experience in the Pharma domain.
We are hiring for following roles across Data management tech stacks -
Data Engineer - Advanced knowledge of PySpark ,python, pandas, numpy frameworks.
Minimum 3 years of extensive experience in design, build and deployment of Spark / Pyspark - for data integration.Deep experience in developing data processing tasks using pySpark such as reading data from external sources, merge data, perform data enrichment and load in to target data destinationsCreate Spark jobs for data transformation and aggregationAWS Infra - 5-8 years of experience & should have hands-on experience working on strong expertise in AWS IAM, EKS (deep expertise), S3, EC2, and Cost Monitoring.
Up-to-date with recent AWS service trends and best practicesKubernetes / EKS cluster setup, scaling, monitoring, troubleshooting & Karpenter implementationExperience with Terraform for infrastructure automationSkills Required
Pyspark, Sql, Aws, Python, Microsoft Project, Informatica Mdm