Software Engineer/ETL; SQL; Cloud - ONSITE Nashville, TN
Open Systems Technologies
Title: Software Engineer/ETL; SQL; Cloud - ONSITE Nashville, TN
Duties:
Provides expert content/professional leadership on complex Engineering assignments/projects. Designs and develops a consolidated, conformed enterprise data warehouse and data lake which store all critical data across Customer, Provider, Claims, Client and Benefits data. Designs, develops and implements methods, processes, tools and analyses to sift through large amounts of data stored in a data warehouse or data mart to find relationships and patterns. Participates in the delivery of the definitive enterprise information environment that enables strategic decision-making capabilities across enterprise via an analytics and reporting. Exercises considerable creativity, foresight, and judgment in conceiving, planning, and delivering initiatives. Uses deep professional knowledge and acumen to advise functional leaders. Focuses on providing thought leadership within Information Management but works on broader projects, which require understanding of wider business. Recognized internally as a subject matter expert.
Skills:
.
Design, develop and maintain ETL platforms for various business use cases which are fault tolerant, highly distributed and robust.
Analyze large sets of structured and semi structured data for business analytics and ETL design.
Translate business needs and vision into roadmap, project deliverables and organization strategies.
Design and implement ETL solutions using leveraging cloud native platforms.
Collaborate with analytics and business teams to design data models that feed business intelligence tools, increasing data accessibility and encouraging data driven solutions.
Experience with data modelling techniques for cloud data stores and on prem databases like Teradata, Teradata Vantage (TDV)etc
Preferred working experience in ETL development in Teradata vantage and data migration from on prem to Teradata vantage.
Proficiency in SQL, relational and non-relational databases, query optimization and data modelling.
Experience with source code control systems like Gitlab.
Experience with large scale distributed relational and NoSQL database systems.
Good experience on designing and developing data pipelines for data ingestion and transformation using Spark.
Distributed computing experience using Python, Pyspark.
Good understanding of spark framework and spark architecture.
Experience working in Cloud based big data infrastructure.
Excellent in trouble shooting the performance and data skew issues.
Must have good understanding of spark run time metrics and tune applications based on metrics.
Deep knowledge in partitioning, bucketing concepts of data ingestion.
Good understanding of AWS services like Glue, Athena, S3, Lambda, Cloud formation.
Preferred working knowledge on the implementation of datalake ETL using AWS glue, Databricks etc.
Technologies:
Pyspark, Python, AWS services, Teradata Vantage, CI/CD technologies, Terraform, SQL.
Duties:
Provides expert content/professional leadership on complex Engineering assignments/projects. Designs and develops a consolidated, conformed enterprise data warehouse and data lake which store all critical data across Customer, Provider, Claims, Client and Benefits data. Designs, develops and implements methods, processes, tools and analyses to sift through large amounts of data stored in a data warehouse or data mart to find relationships and patterns. Participates in the delivery of the definitive enterprise information environment that enables strategic decision-making capabilities across enterprise via an analytics and reporting. Exercises considerable creativity, foresight, and judgment in conceiving, planning, and delivering initiatives. Uses deep professional knowledge and acumen to advise functional leaders. Focuses on providing thought leadership within Information Management but works on broader projects, which require understanding of wider business. Recognized internally as a subject matter expert.
Skills:
.
Design, develop and maintain ETL platforms for various business use cases which are fault tolerant, highly distributed and robust.
Analyze large sets of structured and semi structured data for business analytics and ETL design.
Translate business needs and vision into roadmap, project deliverables and organization strategies.
Design and implement ETL solutions using leveraging cloud native platforms.
Collaborate with analytics and business teams to design data models that feed business intelligence tools, increasing data accessibility and encouraging data driven solutions.
Experience with data modelling techniques for cloud data stores and on prem databases like Teradata, Teradata Vantage (TDV)etc
Preferred working experience in ETL development in Teradata vantage and data migration from on prem to Teradata vantage.
Proficiency in SQL, relational and non-relational databases, query optimization and data modelling.
Experience with source code control systems like Gitlab.
Experience with large scale distributed relational and NoSQL database systems.
Good experience on designing and developing data pipelines for data ingestion and transformation using Spark.
Distributed computing experience using Python, Pyspark.
Good understanding of spark framework and spark architecture.
Experience working in Cloud based big data infrastructure.
Excellent in trouble shooting the performance and data skew issues.
Must have good understanding of spark run time metrics and tune applications based on metrics.
Deep knowledge in partitioning, bucketing concepts of data ingestion.
Good understanding of AWS services like Glue, Athena, S3, Lambda, Cloud formation.
Preferred working knowledge on the implementation of datalake ETL using AWS glue, Databricks etc.
Technologies:
Pyspark, Python, AWS services, Teradata Vantage, CI/CD technologies, Terraform, SQL.
Confirm your E-mail: Send Email
All Jobs from Open Systems Technologies