BANGALORE, IND
1 day ago
Data Engineer-Data Platforms-AWS
**Introduction** In this role, you'll work in one of our IBM Consulting Client Innovation Centers (Delivery Centers), where we deliver deep technical and industry expertise to a wide range of public and private sector clients around the world. Our delivery centers offer our clients locally based skills and technical expertise to drive innovation and adoption of new technology. A career in IBM Consulting is rooted by long-term relationships and close collaboration with clients across the globe. You'll work with visionaries across multiple industries to improve the hybrid cloud and AI journey for the most innovative and valuable companies in the world. Your ability to accelerate impact and make meaningful change for your clients is enabled by our strategic partner ecosystem and our robust technology platforms across the IBM portfolio; including Software and Red Hat. Curiosity and a constant quest for knowledge serve as the foundation to success in IBM Consulting. In your role, you'll be encouraged to challenge the norm, investigate ideas outside of your role, and come up with creative solutions resulting in ground breaking impact for a wide network of clients. Our culture of evolution and empathy centers on long-term career growth and development opportunities in an environment that embraces your unique skills and experience **Your role and responsibilities** * Design, implement, and manage large-scale data processing systems using Big Data Technologies such as Hadoop, Apache Spark, and Hive. * Develop and manage our database infrastructure based on Relational Database Management Systems (RDBMS), with strong expertise in SQL. * Utilize scheduling tools like Airflow, Control M, or shell scripting to automate data pipelines and workflows. * Write efficient code in Python and/or Scala for data manipulation and processing tasks. * Leverage AWS services including S3, Redshift, and EMR to create scalable, cost-effective data storage and processing solutions **Required technical and professional expertise** * Proficiency in Big Data Technologies, including Hadoop, Apache Spark, and Hive. * Strong understanding of AWS services, particularly S3, Redshift, and EMR. * Deep expertise in RDBMS and SQL, with a proven track record in database management and query optimization. * Experience using scheduling tools such as Airflow, Control M, or shell scripting. * Practical experience in Python and/or Scala programming languages **Preferred technical and professional experience** * Knowledge of Core Java (1.8 preferred) is highly desired Excellent communication skills and a willing attitude towards learning. * Solid experience in Linux and shell scripting. Experience with PySpark or Spark is nice to haveFamiliarity with DevOps tools including Bamboo, JIRA, Git, Confluence, and Bitbucket is nice to have * Experience in data modelling, data quality assurance, and load assurance is a nice-to-have
Confirm your E-mail: Send Email