Vacancy expired!
- At least 4 years of experience developing in Python, Spark, Scala, Java
- At least 3 years’ experience in one of the following Cloud technologies: Amazon Web Services (AWS), Google
- Bachelor’s degree
- Experience working with different Databases and understanding of data concepts (including data warehousing, data lake patterns, structured and unstructured data)
- Implementation and tuning experience specifically using Amazon Elastic Map Reduce (EMR)
- Implementing AWS services in a variety of distributed computing, enterprise environments.
- Experience writing automated unit, integration, regression, performance, and acceptance tests
- Develop data governance related services using Python, Spark, Java, and AWS
- Develop data filtering, transformational and loading requirements
- Define and execute ETLs using Python and Spark among other Data technologies
- Determine appropriate translations and validations between source data and target databases
- Implement business logic to cleanse & transform data
- Design and implement appropriate error handling procedures
- Develop project, documentation, and storage standards in conjunction with data architects
- Monitor performance, troubleshoot and tune ETL processes as appropriate using tools like CloudWatch in the AWS ecosystem.