Job Description:
Must-Have: Hadoop, Spark, Pyspark, Hive, Scala
Responsibility:
Ingest data from disparate sources (Structured, unstructured and semi-structured) and develop ETL jobs using the above skills.
Do impact analysis and come up with estimates.
Take responsibility for end-to-end deliverable.