Notes: Data engineer will be working on analytics, lifecycle management and observability, fitting into the existing team and program and working closely with the WW team to drive automation and infrastructure as code.
Position Summary
Responsible for developing and operationalizing data pipelines to make data available for consumption (reports and advanced analytics). This includes data ingestion, data transformation, data validation / quality, data visualization, data pipeline optimization, orchestration; and engaging with DevOps Engineers during CI / CD. The role requires grounding in programming and SQL, followed by expertise in data storage, modeling, cloud, data warehousing, and data lakes.
The Data Engineer is responsible for data across the IS Sustainability stack. This is a new team and will be fast paced, highly visible, supporting the business goals by being an industry leader in this space. This role is focused on data engineering to build and deliver automated data pipelines from a variety of internal and external data sources. The Data Engineer will partner with DevOps Engineers, product owners, engineering and data platform teams to design, build, test and automate data pipelines that are relied upon across the company as the single source of truth.
We are looking for a highly motivated, customer focused professional who wants to work in a fun and highly collaborative environment and willing to be the player as well as coach for the data science community within the organization.
*This position will be filled onsite in Issaquah, WA
Job Duties/Essential Functions
- Develop and operationalize data pipelines to create enterprise certified data sets that are made available for consumption (BI, Advanced analytics, APIs/Services).
- Identifies ways to improve data reliability, efficiency and quality of data management.
- Works with area SMEs to design data pipelines and recommends ongoing optimization of data storage, data ingestion, data quality and orchestration.
- Identifies, designs, and implements internal process improvements: automating manual processes, optimizing data delivery.
- Identifies ways to improve data reliability, efficiency and quality of data management.
- Conducts ad-hoc data retrieval for business reports and dashboards.
- Assesses the integrity of data from multiple sources.
- Designs, develops, & implements ETL/ELT processes using Informatica Intelligent Cloud Services (IICS) and Azure Data Factory (ADF).
- Uses Azure services such as Databricks, Azure SQL DW (Synapse), Data Lake Storage, Azure Event Hub, Cosmos, Delta-Lake to improve and speed up delivery of our data products and services.
- Develop and implement PowerBI reports and applications
- Implements big data and NoSQL solutions by developing scalable data processing platforms to drive high-value insights to the organization.
- Leads the analysis by applying statistics, machine learning and analytic approaches to predict and optimize business outcomes.
- Designs and builds ML/DL models to solve business problems.
- Correctly frames a problem and comes up with a hypothesis.
- Communicates technical concepts to non-technical audiences both in written and verbal form.
- Regular and reliable workplace attendance at your assigned location.
Experience, Skills, Education & Licenses/Certifications
Required:
- Experience engineering and operationalizing data pipelines with large and complex datasets.
- Hands-on experience with Informatica PowerCenter and/or IICS.
- Experience with Cribl, Confluent/Kafka, Big Panda.
- Experience working with Cloud technologies such as ADLS, Azure Databricks, Spark, Azure Synapse, Cosmos DB and other big data technologies.
- Extensive experience working with various data sources (DB2, SQL, Oracle, flat files (csv, delimited), APIs, XML, JSON.
- Advanced SQL skills required. Solid understanding of relational databases and business data; ability to write complex SQL queries against a variety of data sources.
- Experience with Data Modeling, ETL, and Data Warehousing.
- Strong understanding of database storage concepts (data lake, relational databases, NoSQL, Graph, data warehousing).
- Experience in delivering business insights using advanced statistical and machine learning models and visualization.
- Proficiency in working with diverse databases and other data sources.
- Experience with Git / Azure DevOps.
- Experience delivering data solutions through agile software development methodologies.
- Graduate degree in Computer Science, Data Science, and Statistics/Mathematics or related field.
Recommended:
- Azure, GCP Certifications.
- Experience implementing data integration techniques such as event/message-based integration (Kafka, Azure Event Hub), ETL.
- Exposure to the retail industry.
- Experience with UC4 Job Scheduler.
- Strong proficiency in Machine Learning, Statistical and Reporting tools (Python, R, SQL, PowerBI).
- Knowledge of Deep Learning and Neural Networks, and its applications.
- Strong experience working in Cloud (Azure, GCP) based analytics platform.
- Knowledge of Agile software development.
- Experience in software development.
- Excellent verbal and written communication skills.
“Tekgence is an equal opportunity employer. Applicants must be authorized to work in the U.S.A. Citizens and Green Card holders are strongly encouraged to apply.”