Results-driven Data Engineer at Wizeline with expertise in ETL development and data pipeline design. Successfully implemented scalable data platforms, enhancing performance and integration of diverse data sources. Proficient in Spark and Python, with strong data governance skills, fostering collaboration across teams to achieve high-quality data management solutions.
Earnest (Feb 2022 - Present):
During my time at Earnest, I was responsible for driving key data transformation projects and enhancing the company's data infrastructure. I primarily focused on leveraging DBT to design and implement complex data transformations that optimized data workflows and ensured scalability. Additionally, I led critical database migration projects, ensuring smooth transitions and minimizing any operational disruptions.
Key responsibilities included:
Azteca Bank (Dec 2021 - Feb 2022):
During my time at Azteca Bank, I was responsible for developing and optimizing automated reporting solutions using cloud-native technologies. I focused on integrating AWS Lambda, Step Functions, and RDS PostgreSQL to create scalable, reliable, and efficient reporting pipelines that supported critical business operations.
Key responsibilities included:
Mercado libre (Feb 2021 - Dec 2021):
At Mercado Libre, I was part of the data engineering team responsible for migrating analytical workloads from Teradata to BigQuery. My focus was on translating and optimizing SQL queries to align with BigQuery’s architecture and best practices. I collaborated closely with stakeholders to validate query results, ensure data consistency, and improve overall performance during the transition.
Key responsibilities included:
FactSet (Oct 2020 - Feb 2021):
I was part of the data engineering team responsible for generating large-scale financial reports by leveraging Apache Spark and Scala on AWS EMR. I developed and optimized scalable data pipelines that processed complex datasets efficiently, enabling timely and accurate reporting to support business analytics and client deliverables.
GEPP (Sep 2020 - Oct 2020):
I worked on building and maintaining automated reporting workflows using PySpark on AWS EMR. My focus was on processing operational and sales data from multiple sources, ensuring data quality and consistency. I collaborated closely with analytics teams to deliver reliable reports aligned with key business metrics, helping drive strategic insights across the organization.
Autodesk (Aug 2020):
At Autodesk, I was part of a cross-functional team responsible for evaluating Qubole and Astronomer as orchestration solutions for data pipelines. My role involved testing real workflows across both platforms, assessing performance, scalability, and integration with existing tools. I helped document findings and technical trade-offs to guide the selection of the most suitable platform for long-term data orchestration needs.
Lumiata (Oct 2019 - Aug 2020):
At Lumiata, I was part of a data engineering team responsible for designing and implementing multiple dataflow processes to support analytics and machine learning initiatives. I contributed to building scalable and reliable pipelines that automated data ingestion, transformation, and delivery across various sources, helping ensure data quality and timely access for downstream teams.
ETL development
Data pipeline design
Data modeling
Data warehousing
SQL expertise
Data migration
Data governance
Spark
Python
EKS
Terraform
DBT
AWS EMR