Employer Active
• Experienced in exploring, debugging and data quality issues and defining data quality rules across complex data pipelines
• Experience working with any one programming language (Spark/Scala, Python/Panda/PySpark)
• Good knowledge of SQL and write complex queries
• Exposure to working on data pipelines with scalable ETL jobs across from ingestion to data lake, data warehousing and data marts
• Build strong relationships and liaise with multiple scrum teams
Good to have:
• Understanding of distributed systems
• Experience on Hadoop (HDFS, Hive, HBase)
• Experience on Azure (ADF, ADLS, Databricks)
• Exposure to Snowflake
Full Time