This role will be responsible for analyzing business and system requirements and define optimal data strategy of fulfilling them. The role holder will be designing and implementing data acquisition strategies and they will own and extend the business’s data pipeline through the collection, storage, processing, and transformation of large data-sets. This role will be responsible of leading the data architecture in the region LatAm, responsible of a small team to implement this strategy. To be successful in this role you will need to have excellent analytic skills related to working with structured, semi structured and unstructured datasets, Expert SQL knowledge and experience with relational as well as no-SQL database systems. You must have expert level Programming skills in Python/PySpark, and it would be beneficial if you have also experience in Java/Scala programming.
Requirements\nUniversity Degree in Computer Science, Information Systems, Statistics, or related field Expertise with Data Lake/Big Data Projects implementation in Cloud (preferably MS Azure) and/or On-premise platforms Experience with Cloud\nAzure technology stack: ADLS Gen2, Databricks EventHub, Stream Analytics, Synapse Analytics, AKS, Key Vault On Premise: Spark, HDFS, Hive, Hadoop distributions (Cloudera or MapR), Kafka, Airflow (or any other scheduler)\nExperience with designing and building lakehouse architectures in Parquet/Delta and Synapse Serverless or Databricks SQL (knowledge of Unity Catalogue is big plus)\nAbility to develop, maintain and distribute the code in modularized fashion Working experience with DevOps framework\nVery good understanding of Software Development Lifecycle, source code management, code reviews, etc.\nExperience in performing root cause analysis on internal and external data and processes to answer specific business questions and identify opportunities for improvement\nExperience in building processes supporting data transformation, data structures, metadata, dependency, and workload management\nAbility to collaborate across different teams/geographies/stakeholders/levels of seniority Energetic, enthusiastic, and results-oriented personality\nCustomer focus with an eye on continuous improvement\nMotivation and ability to perform as a consultant in data engineering projects\nAbility to work independently but also within a team Leadership skills to develop team’s strategy & a team of 3 people Strong will to overcome the complexities involved in developing and supporting data pipelines Agile mindset Language Requirements English\nFluent spoken and written Spanish – Nice to have\nKey Responsabilities\nDevelopment & Lead data engineers (3 people)\nBuilding scalable, performant, supportable and reliable data pipelines you must enjoy optimizing data systems as well as building them from the ground up\nSetting up new and monitoring of existing metrics, analyzing data, and in cooperation with other Data & Analytics team members to identifying and implementing system and process improvements\nSupporting collection of the metadata into our data catalogue system where all available, ensuring data is maintained and cataloged accurately Defining and implementing DevOps framework using CI/CD\nWorking closely with big data architects, data analysts, data warehouse engineers, and data scientists\nSupporting and promoting Agile way of working using SCRUM framework\n