Expected Technical Skills:
- 5+ years of experience in building and optimizing data pipelines and architectures in high availability environments.
- Strength in SQL, data modeling, and ETL development.
- Ability to extract data from multiple sources using tools such as Kafka, Sqoop and Nifi.
- Demonstrated experience of working in the Hadoop ecosystem using Parquet and Avro.
- Experience in working on databases such as Postgres and MongoDB.
- Ability to write complex Hive queries.
- Data transformation using Spark.
- Experience on pipelining tools such as Oozie and Airflow.
- Experience on AWS tools such as EMR, RDS, and Redshift will be an added advantage.