Vacancy expired!
Hello! We, (Miracle Software Systems, Inc), are actively looking for"Data Engineer"to work with one of our direct client at Novi,MI,USA location. Below the Job description:
Requirement DetailsPosition : Senior Data EngineerDuration: Full TimeLocation: Novi, MichiganPrimary Skills: Google Cloud Platform, Hadoop, Spark, KafkaDescription:- This position will be part of the Data Factory Enterprise Data Ingestion (EDI) product group using tools like Attunity, Hive, Hbase, Spark, Google Cloud Platform, etc. The product team"s objective is to ingest data from hundreds of sources within the company to the EDI Hadoop and Google Cloud Platform environment and perform transformations to make it usable for data scientists in GDIA.
- This position will require an individual who has a strong background with multiple database technologies, who is process oriented and has knowledge Hadoop and Google Cloud Platform environment. This role is part of a highly dynamic team supporting GDIA activities.
- Experience with big data tools: Hadoop, Spark, Kafka, etc. -Experience with relational SQL databases,
- Experience with Google Cloud Platform based Big Data deployments (Batch/Real-Time) leveraging Big Query, Google Cloud Storage, Pub-Sub, Data Fusion, Dataflow, Dataproc, Cloud functions etc.
- Experience building and optimizing "big data" data pipelines, architectures and data sets. Experience performing root cause analysis on internal and external data and processes to answer specific business questions and identify opportunities for improvement.
- Strong analytic skills related to working with unstructured datasets.
- Build processes supporting data transformation, data structures, metadata, dependency and workload management.
- Experience supporting and working with cross-functional teams in agile mode.
- Experience with object-oriented/object function scripting languages: Python, Java
- 10+ years of application development experience 3+ years of Google Cloud Platform experience 4+ years of experience with large-scale solutioning and operationalization of data warehouses and/or data lakes