Data Engineer

19 May 2024

Vacancy expired!

Job Description

Data Engineer

Job Description:
  • At Clients, the community is what we are and it's what we do.
  • It's what makes us different.
  • To create the best ride for all, we start in our own community by creating an open, inclusive, and diverse organization where all team members are recognized for what they bring.
  • Here at our client, Data is the only way we make decisions.
  • It is the core of our business, helping us create a transportation experience for our customers and providing insights into the effectiveness of our product launch & features.
  • As a Data Engineer at our client, you will be a part of an early stage team that builds the data transport, collection, and storage, and exposes services that make data a first-class citizen at our client.
  • We are looking for a Data Engineer to build a scalable data platform.
  • You'll have ownership of our core data pipeline that powers our client's top-line metrics;
  • You will also use data expertise to help evolve data models in several components of the data stack;
  • You will help architect, build, and launch scalable data pipelines to support our client's growing data processing and analytics needs.
  • Your efforts will allow access to business and user behavior insights, using huge amounts of our client data to fuel several teams such as Analytics, Data Science, Marketplace, and many others.
  • You will report to a Data Engineering Manager.

Job Responsibilities:
  • Owner of the core company data pipeline, responsible for scaling up data processing flow to meet the rapid data growth at our client
  • Evolve data model and data schema based on business and engineering needs
  • Implement systems tracking data quality and consistency
  • Develop tools supporting self-service data pipeline management (ETL)
  • SQL and MapReduce job tuning to improve data processing performance

Required Experience:
  • 5+ years of relevant professional experience
  • Experience with Hadoop (or similar) Ecosystem (MapReduce, Yarn, HDFS, Hive, Spark, Presto, Pig, HBase, Parquet)
  • Proficient in at least one of the SQL languages (MySQL, PostgreSQL, SqlServer, Oracle)
  • Good understanding of SQL Engine and able to conduct advanced performance tuning
  • Strong skills in a scripting language (Python, Ruby, Bash)
  • 1+ years of experience with workflow management tools (Airflow, Oozie, Azkaban, UC4)
  • Comfortable working directly with data analytics to bridge our client's business goals with data engineering
#tech