Job descriptionBuilding efficient storage for structured and unstructured dataTransform and aggregate the data using data processor technologiesDeveloping and deploying distributed computing Big Data applications using Open Source frameworks like Apache Spark, Apex, Flink, Nifi, and Kafka on AWS CloudUtilizing programming languages like Java, Scala, Python, and Open Source RDBMS and NoSQL databases and Cloud-based data warehousing services such as RedshiftUsing Hadoop modules such as YARN & MapReduce, and related Apache projects such as Hive, Hbase, Pig, and CassandraLeveraging DevOps techniques and practices like Continuous Integration, Continuous Deployment, Test Automation, Build Automation, and Test Driven Development to enablethe rapid delivery of working code utilizing tools like Jenkins, Maven, Nexus, Chef, Terraform, Ruby, Git, and Docker