Job Description :
- 5.5- 7.5 Years of in Big Data & Data related technology experience
- Expert level understanding of distributed computing principles
- Expert level knowledge and experience in Apache Spark
- Hands on programming with Python
- Proficiency with Hadoop v2, Map Reduce, HDFS, Sqoop
- Experience with building stream-processing systems, using technologies such as Apache Storm or Spark-Streaming
- Experience with messaging systems, such as Kafka or RabbitMQ
- Good understanding of Big Data querying tools, such as Hive, and Impala
- Experience with integration of data from multiple data sources such as RDBMS (SQL Server, Oracle), ERP, Files
- Good understanding of SQL queries, joins, stored procedures, relational schemas
- Experience with NoSQL databases, such as HBase, Cassandra, MongoDB
- Knowledge of ETL techniques and frameworks
- Performance tuning of Spark Jobs
- Experience with native Cloud data services AWS or AZURE Databricks,GCP.
- Ability to lead a team efficiently
- Experience with designing and implementing Big data solutions
- Practitioner of AGILE methodology
NOTE : 5 Days working from Baner / Kharadi location and 3- 11 pm ist shift (shift allowance will be given).