Data Software Engineer - Spark, Python, (AWS ,Kafka or Azure Databricks or GCP)
Job Description :
- 5-12 Years of in Big Data & Data related technology experience
- Expert level understanding of distributed computing principles
- Expert level knowledge and experience in Apache Spark
- Hands on programming with Python
- Proficiency with Hadoop v2, Map Reduce, HDFS, Sqoop
- Experience with building stream-processing systems, using technologies such as Apache Storm or Spark-Streaming
- Experience with messaging systems, such as Kafka or RabbitMQ
- Good understanding of Big Data querying tools, such as Hive, and Impala
- Experience with integration of data from multiple data sources such as RDBMS (SQL Server, Oracle), ERP, Files
- Good understanding of SQL queries, joins, stored procedures, relational schemas
- Experience with NoSQL databases, such as HBase, Cassandra, MongoDB
- Knowledge of ETL techniques and frameworks
- Performance tuning of Spark Jobs
- Experience with native Cloud data services AWS or AZURE Databricks,GCP.
- Ability to lead a team efficiently
- Experience with designing and implementing Big data solutions
- Practitioner of AGILE methodology