Required Skills: DataProc, Azure, pySpark, Bigdata, Scala, Python, Java
Job Description
Experience working in Bigdata, Spark, Scala, Python/Java
Working on cloud experience, deploying containerized applications, building data pipelines using GCP
DataProc/Azure Databricks using pySpark.
Own code quality and experience in writing extensive unit tests.
Excellent written communication.
SQL Coding.