Vacancy expired!
- Lead a global team with footprints in North America, Europe, and Asia
- Lead and guide the development team on project execution to align with business vision and objectives while defining best practices and code standards
- Share responsibility for architecting and implementing the next generation of Data Platform Architecture for Data Pipeline and Distributed Computation.
- Collaborate with various partners to understand multi-functional requirements and convert them into reusable service components
- Define and drive the data modeling solution for Data Warehousing and distributed data storage including best practices and design standards.
- Participate in architecting, designing and implementation a scalable computation and data distribution platform.
- Identify, design, and implement internal process improvements and share with the relevant technology organization.
- Develop and mentor other team members in design and development while ensuring accurate development estimates on projects and tasks.
- Establish and enforce CI/CD, code review, design standards, best development practices.
- Motivate data engineers to deliver top quality, supportable products that continue to raise the bar higher.
- Work with data scientists to develop data ready tools to support their job.
- Identify, investigate, and resolve data discrepancies by finding the root cause of issues; work with partners across various cross-functional teams to prevent future occurrences.
- Understand existing systems and resolve operations issues while working with other support staff located across the globe.
- Automate manual ingestion processes and optimize data delivery ensuring SLAs are met.
- Design, maintain, and own the Data Infrastructure. Work with infrastructure teams on re-designing environment for greater scalability.
- Be up to date with the latest tech trends in the big-data space and recommend them as needed.
- You will be accountable for managing high-quality data exposed for internal and external consumption by downstream users and applications.
- 10+ years of hands-on experience in Data Engineering or Software Engineering.
- Experience leading or mentoring a team and drive the execution.
- 7+ years of experience with data transformation and computation leveraging OO languages with Python for data transformation (Core Python, Pandas and pySpark) or Java/J2EE/Spring architecture design and development
- 5+ years experience with SPARK (pySpark), performance tuning and scaling.
- 5+ years using distributed eco-systems like Hadoop, Hive, etc. Proficiency on bucketing, partitioning, tuning and different file storages (like S3) and formats (ORC, PARQUET & AVRO).
- Experience with design and actual implementation of Data Warehouse or OLAP system or Data Lake
- Experience with ETL tools and Workflow management systems like Airflow, Luigi, NiFi, Kilo.
- Extensive experience and advanced knowledge of SQL and relational databases (e.g., MS SQL Server, MySQL, Postgres), data modeling, stored procedures, and complex queries. .
- B.S. / M.S. degree in Computer Science, Engineering, or a related discipline.
- Experience with stream-processing systems: Storm, Spark-Streaming, Kafka
- Experience with distributed databases and query engines like Snowflake or Presto is huge plus
- Experience with API development in Python or Java
- Experience with containerization architecture: Docker and Kubernetes
- Experience with cached databases (e.g., Ignite)
- Knowledge of any Graph Databases
- Experience with Scala
- Any experience with Cloud platform such as Azure, AWS or Google Cloud Platform is huge PLUS!
- ID: #23795807
- State: Washington Seattle-tacoma 98101 Seattle-tacoma USA
- City: Seattle-tacoma
- Salary: USD TBD TBD
- Job type: Permanent
- Showed: 2021-12-03
- Deadline: 2022-01-31
- Category: Sales