Data Engineer/ETL/Data ware House engineer

New York Technology Partners
New York Technology Partners

Job Overview

Seeking a Scientific Data Application Engineer with rich experience building Data Centric Solutions. We enable science by leading and contributing to full software lifecycle architecture, design, implementation, testing and deployment of Scientific Data Applications that add proven value and impact to the scientific operations of Research Development at Client. Our current Data Platform Must have Pyspark, Python, Airflow In this role, a typical day might include the following Develop and maintain scalable data pipelines for new Data Product support continuing increases in data volume and complexity Collaborate with Data SMEs and Business Users to improve data models, growing data accessibility and promoting data-driven decision making Define Data models building Data product pipeline as script in PySpark Define and drive adoption of standard methodologies in code health, testing, and maintainability Perform data analysis required to solve data related issues, root cause analysis and assisting with resolution of data issues Act as an experienced technical resource for Data Application Solution Maintain strong expertise and knowledge of current and emerging technologies and products evaluating AWS Services, Opensource Vendor tools and implementing new technologies Mentor on approach and execution of solutions, coaching on technologies, and establishing a team-wide comprehension of solution capabilities and direction This job might be for you if You enjoy working in a fast-paced environment, supporting innovative scientists doing ground breaking science to discover lifesaving medicines for people with serious medical conditions. You have a collaborative, agile approach to drive creative and efficient use of technology to accelerate drug discovery, and you enjoy aligning multiple perspectives into an agreed course of action You are comfortable with the above technology landscape but will execute a technology roadmap that aligns with business priorities, scales for the future, and disrupts when vital to continually improve and deliver value. To be considered for the role, you must have a BS or MS degree in Computer Science as well as 7 years of Data pipeline development experience with Spark (PySpark), HIVE Airflow. 5 years of relevant experience with schema design and dimensional data modeling, 3 years of Datawarehouse BI Tools (Spotfire preferable) experience and Hadoop Platform experience of any sizescale. Proven track record in managing and communicating BI, Data Lake, Datawarehouse concepts to different partners. Ability to adapt to new languages, frameworks, technologies as the need arises. Demonstrable skill in collaborating with other IT teams, vendor partners, internal collaborators, and architects to deliver solutions advocacy for a DevOps approach to the software development lifecycle. Biotechnology Pharmaceutical Working Experience.

View More
Job Detail
Shortlist Never pay anyone for job application test or interview.