Key Responsibilities :
- Collaborate with Team Members and Clients : Work with coworkers and clients to clarify the requirements for data-related tasks and projects.
- Develop Data Infrastructure : Conceptualize and design robust infrastructures to facilitate easy access and analysis of big data.
- Optimize Existing Frameworks : Refactor and improve current data frameworks for enhanced performance and scalability.
- Test Data Structures : Ensure the designed data structures are fit for use by conducting rigorous testing.
- Prepare Raw Data : Clean, process, and prepare raw data to be utilized by data scientists for further analysis and modeling.
- Error Detection and Correction : Identify, troubleshoot, and correct errors in the data processing or storage pipeline.
- Data Backup and Accessibility : Ensure that all work is regularly backed up and is easily accessible by relevant coworkers or teams.
- Stay Current with Industry Trends : Keep up-to-date with the latest industry standards, technologies, and best practices that can enhance the quality of data processing and management.
Skills Required
Hadoop, Spark, Cloud, Python, Java, Aws