Only those candidates can apply who:
1. are available for full time (in-office) internship
2. can start the internship between 24th Apr'19 and 24th May'19
3. are available for duration of 4 months
4. have relevant skills and interests
* Women wanting to start/restart their career can also apply.
1. Experience with AWS/Azure services stack
2. Hands-on experience with Hadoop and big data stack using Python, Java or Scala
3. Software development experience, ideally in Big Data technologies
4. Experience with SQL RDBMS like SQL Server, Oracle, and MySQL
5. Experience working with large data sets, experience working with distributed computing a plus (Map/Reduce, Hadoop, Hive, Apache Spark, etc.)
6. Experience in Extract data from multiple structured and unstructured feeds by building and maintaining scalable ETL pipelines on distributed software systems
7. Programming experience in Java, Scala, and Python. Python a plus
8. Knowledge of Hadoop or similar data processing frameworks (such as EMR, Hadoop, Spark) and a good understanding of optimization techniques
9. Understanding of NoSQL data stores, messaging or pub-sub queuing systems and data processing frameworks
10. Working knowledge of computer algorithms
OR