Vacancy expired!
- Enhance and maintain our current data pipelines and associated infrastructure.
- Assemble large, moderately complex data sets that meet functional / non-functional business requirements.
- Engage with different stakeholder teams to troubleshoot various database systems.
- Build and maintain tools that monitor data and system health.
- Identify, design, and implement internal process improvements to optimize production of results and enable cost savings.
- Performance tune and optimize data pipeline on Spark.
- Create and maintain documentation describing data catalog and data objects.
- Bachelor’s degree in computer science, an engineering field, or equivalent work experience in an engineering field.
- 3 years of experience with data engineering/ETL ecosystem, such as Palantir Foundry, Spark, Informatica, SAP BODS, OBIEE.
- Experience with data engineering/ETL ecosystem, such as Palantir Foundry, Spark, Informatica, SAP BODS, OBIEE.
- Database design fundamentals.
- Experience with Python, Pandas and APIs.
- Knowledge of Time Series data set development.
- Demonstrated commitment to teamwork and enabling others.
- Proven ability to translate business desires into technical requirements.
- Ability to communicate with various stakeholders and leadership.
- Ability to break down ambiguous problems.
- Experience with Scikit Learn, PySpark or equivalent big data processing framework, CI/CD tool.
- Experience with an infrastructure as code tool, writing production-level code, writing health checks, unit tests, integration tests, schema validations.
- Familiarity with cloud computing security fundamentals.
- Experience with the Palantir Foundry platform.
- Experience working with data scientists and machine learning engineers.
- Familiarity with model deployment.
- Front end tools: PowerBi, Tableau.
- ID: #49374183
- State: California Oakland 94601 Oakland USA
- City: Oakland
- Salary: Depends on Experience
- Job type: Permanent
- Showed: 2023-02-27
- Deadline: 2023-04-25
- Category: Et cetera