Work Schedule
OtherEnvironmental Conditions
OfficeJob Description
COMPANY: Thermo Fisher Scientific Inc.
LOCATION: 168 Third Avenue, Waltham, MA 02451
TITLE: Scientist III, Data Sciences
HOURS: Monday to Friday, 8:00 am to 5:00 pm
DUTIES: • Develop scalable data pipelines and builds out new API integrations to support continuing increases in data volume and complexity.
• Own and deliver Projects/Enhancements associated with Data platform solutions.
• Develop solutions using Pyspark/EMR, SQL and databases, AWS Athena, S3, Redshift, AWS API Gateway, Lambda, Glue and other Data Engineering technologies.
• Implement solutions using AWS and other cloud platform tools, including GitHub, Jenkins, Terraform, Jira, and Confluence.
• Follow agile development methodologies to deliver solutions and product features by following DevOps, Data Ops and Dev Sec Ops practices.
• Propose Data load optimizations and continuously implement to improve the performance of the Data loads.
• Identify, design, and implement internal process improvements, including automating manual processes, optimizing data delivery, and re-designing infrastructure for greater scalability.
• Keep data separated and secure across through multiple data centers and AWS regions.
• Be available and participate in on-call schedule to address critical operational incidents and business requests.
• Meet and exceed BI Operational SLAs for Service Now incidents, Major Incidents, xMatters alerts, Employee Experience Metrics and BI application/process availability metrics.
Travel: Up to 10% domestic travel required. Telecommuting Employee: Reports to company headquarters in Waltham, MA. Can work remotely or telecommute.
REQUIREMENTS: MINIMUM Education Requirement: Master’s degree in Computer Science, Mechanical Engineering, or related field of study. MINIMUM Experience Requirement: 1 year of IT experience as a Data Engineer, or related experience. Alternative Education and Experience Requirement: Bachelor’s degree in Computer Science, Mechanical Engineering, or related field of study plus 3 years of IT experience as a Data Engineer, or related experience.
Required knowledge or experience with:
• Informatica Power center 10.4, Oracle R12, TOAD, SFDC, Data Warehouse Administration, UNIX, Win SCP, Windows 7, Linux, Informatica Power Exchange, Oracle 11g, and Flat Files;
• SQL,PL/SQL, No SQL, and Bigdata;
• Databricks, Data/Delta lake, Oracle, or AWS Redshift type relational databases;
• Databricks/Spark-based Data Engineering Pipeline development;
• Python-based data integration and pipeline development;
• Data lake and Delta lake experience with AWS Glue and Athena;
• AWS Cloud on data integration with Apache Spark, Glue, Kafka, Elastic Search, Lambda, S3, Redshift, RDS, MongoDB/DynamoDB ecosystems;
• Python development, including in pySpark in AWS Cloud environment;
• Python and common python libraries;
• Analytical experience with database in writing complex queries, query optimization, debugging, user defined functions, views, and indexes;
• Source control systems, including Git and Jenkins build and continuous integration tools;
• Development methodology;
• Writing functional and technical design specifications; and
• Resolving complex data integration problems.
#LI-DNI