What you will do
- Curate, build and transform raw data in scalable information
- Plan long-term strategies and projects for junior engineers
- Keep a closer look on various projects and monitor the progress
- Carry on smooth collaborations with the sales and engineering teams
- Provide management mentorship for junior team members and set the tone for holistic growth
- Ensure everyone is on the same page and taking ownership of their respective projects
- Make extracted data accessible so that organisations can use it to optimise performance
What you will need
- Bachelors /Master's in Computer Science
- At least 2 to 5 years of professional experience
- 1+ years of experience working with Big Data systems involving Kafka, Apache Spark/EMR, Hive/Impala, Delta lake, Presto and Airflow
- 1+ years of hands-on experience in implementation and performance tuning Spark/Deltalake/Presto implementations for terabytes scale
- Experience developing software code in one or more programming languages (Java, Scala, Python)
- 1+ years of experience in building streaming solutions using Apache Flink/Spark Streaming/Samza
- Good understanding of AWS BigData stack
- Good to Have
- Having been a contributor/commitor to the big data stack
- Data modeling experience and end-to-end data pipelining experience
- Knowledge of OLAP data cubes and BI tools
- Working knowledge of ELK, Redis and MYSQL
- Knowledge of RxJava, Java Springboot and Microservices architecture