Only those candidates can apply who:
1. are available for full time (in-office) internship
2. can start the internship between 19th May'19 and 22nd Jun'19
3. are available for duration of 4 months
4. have relevant skills and interests
* Women wanting to start/restart their career can also apply.
1. Should have experience with AWS/Azure services stack
2. Should have hands-on experience with Hadoop and big data stack using Python, Java or Scala
3. Should have software development experience, ideally in big data technologies
4. Should have experience with SQL RDBMS like SQL Server, Oracle, and MySQL
5. Should have experience working with large data sets and distributed computing (Map/Reduce, Hadoop, Hive, Apache Spark, etc.)
6. Should have experience in extracting data from multiple structured and unstructured feeds by building and maintaining scalable ETL pipelines on distributed software systems
7. Should have programming experience in Java, Scala, and Python (would be an advantage)
8. Should have knowledge of Hadoop or similar data processing frameworks (such as EMR, Hadoop, Spark) and a good understanding of optimization techniques
9. Should have an understanding of NoSQL data stores, messaging or pub-sub queuing systems and data processing frameworks
10. Should have working knowledge of computer algorithms
OR