Waltham, Massachusetts, USA
14 days ago
Scientist III, Data Sciences

Work Schedule

Other

Environmental Conditions

Office

Job Description

COMPANY:                           Thermo Fisher Scientific Inc.

LOCATION:                           168 Third Avenue, Waltham, MA 02451

TITLE:                                    Scientist III, Data Sciences

HOURS:                                 Monday to Friday, 8:00 am to 5:00 pm

DUTIES:                                 •           Develop scalable data pipelines and builds out new API integrations to support continuing increases in data volume and complexity.
•           Own and deliver Projects/Enhancements associated with Data platform solutions.
•           Develop solutions using Pyspark/EMR, SQL and databases, AWS Athena, S3, Redshift, AWS API Gateway, Lambda, Glue and other Data Engineering technologies.
•           Implement solutions using AWS and other cloud platform tools, including GitHub, Jenkins, Terraform, Jira, and Confluence.
•           Follow agile development methodologies to deliver solutions and product features by following DevOps, Data Ops and Dev Sec Ops practices.
•           Propose Data load optimizations and continuously implement to improve the performance of the Data loads.
•           Identify, design, and implement internal process improvements, including automating manual processes, optimizing data delivery, and re-designing infrastructure for greater scalability.
•           Keep data separated and secure across through multiple data centers and AWS regions.
•           Be available and participate in on-call schedule to address critical operational incidents and business requests. 
•           Meet and exceed BI Operational SLAs for Service Now incidents, Major Incidents, xMatters alerts, Employee Experience Metrics and BI application/process availability metrics.
Travel: Up to 10% domestic travel required. Telecommuting Employee: Reports to company headquarters in Waltham, MA. Can work remotely or telecommute.                                                                                                                                   

REQUIREMENTS:                 MINIMUM Education Requirement: Master’s degree in Computer Science, Mechanical Engineering, or related field of study. MINIMUM Experience Requirement: 1 year of IT experience as a Data Engineer, or related experience. Alternative Education and Experience Requirement:  Bachelor’s degree in Computer Science, Mechanical Engineering, or related field of study plus 3 years of IT experience as a Data Engineer, or related experience.  

Required knowledge or experience with:
•           Informatica Power center 10.4, Oracle R12, TOAD, SFDC, Data Warehouse Administration, UNIX, Win SCP, Windows 7, Linux, Informatica Power Exchange, Oracle 11g, and Flat Files;
•           SQL,PL/SQL, No SQL, and Bigdata;
•           Databricks, Data/Delta lake, Oracle, or AWS Redshift type relational databases;
•           Databricks/Spark-based Data Engineering Pipeline development;
•           Python-based data integration and pipeline development;
•           Data lake and Delta lake experience with AWS Glue and Athena;
•           AWS Cloud on data integration with Apache Spark, Glue, Kafka, Elastic Search, Lambda, S3, Redshift, RDS, MongoDB/DynamoDB ecosystems;
•           Python development, including in pySpark in AWS Cloud environment;
•           Python and common python libraries;
•           Analytical experience with database in writing complex queries, query optimization, debugging, user defined functions, views, and indexes;
•           Source control systems, including Git and Jenkins build and continuous integration tools;
•           Development methodology;
•           Writing functional and technical design specifications; and
•           Resolving complex data integration problems.

#LI-DNI

Confirm your E-mail: Send Email