|

Hello, I'm Sergio!

I enjoy figuring things out in messy, real-world situations, where trade-offs shape structure and governance, and the goal is to keep things clear and manageable.

(There's more stuff down here btw 👇)

About

I like building systems that are practical, scalable, and easy for teams to work with.

Snapshot

Over my career, I have worked on multiple international projects involving owning multiple end-to-end batch and streaming ETL pipelines, data governance, internal tooling, REST APIs and regulatory reporting for financial products operating across multiple countries.

5+

years building software and data systems

60+

end-to-end delivered & maintained data pipelines

12k+

PII columns governed with policy-driven masking

35%

runtime reduction on optimized data jobs

Core Stack

The tools I use most often.

  • Scala
  • Python
  • Spark
  • SQL
  • AWS
  • Databricks
  • Terraform
  • Airflow
  • TypeScript
You can find more of my skills in the Skills section.

Production first

I care about clean architecture, but I care even more about systems that are stable, observable, and easy to operate once they reach production.

Team amplifier

I like building things that help the team move faster too, whether that means better tooling, clearer structure, or less manual work.

Governance without drama

PII, compliance, and reporting work should be straightforward to maintain. I try to make that kind of work reliable, repeatable, and hard to break by accident.

Experience

Most of my work has been in backend engineering, data platforms, and internal tooling.

Clara

Jan 2023 — Present

Data Engineer

Worked on backend features, integrations, and web application flows as part of my first professional software role.

  • Collaborated in building and maintaining the data lakehouse of a large expense management solution serving Brazil, Mexico, and Colombia; designed and implemented high-performance ETL pipelines using Scala, Python, Spark, AWS, Databricks, Terraform and Airflow.
  • Architected and implemented a declarative PII masking system using Databricks Unity Catalog ABAC policies and Terraform, replacing legacy write-time masking with query-layer enforcement. Designed config-as-code YAML definitions, SQL masking UDFs, and conditional masking strategies (person vs. company) across 12,000+ PII findings in 5,500+ tables for Brazil, Mexico, and Colombia.
  • Delivered ETL pipelines under Medallion Architecture (Bronze/Silver/Gold) for multiple data providers, including DMS migrations and AutoLoader integrations.
  • Built webhook-based processing flows triggered via AWS API Gateway endpoints, using AWS Lambda to asynchronously process onboarding events and ingest structured payloads to S3.
  • Implemented regulatory reporting pipelines for Bacen (Brazilian Central Bank) and AML (Anti-Money Laundering) reports for government institutions in Mexico and Brazil, as well as transaction reports for credit companies in Colombia and Brazil.
  • Optimized data processing jobs with date partition reading, reducing execution times by 35%, saving costs while ensuring data quality.
  • Independently initiated and implemented multiple projects to improve the work experience of the Data team, including a customized Slack notification bot for the service desk, an automated ETL project template generator, and a VS Code extension for DLT SQL with full IDE support (go-to-definition, autocomplete, rename/refactor, diagnostics) for the team's 1,500+ SQL model codebase.
  • Scala
  • Python
  • Spark
  • AWS
  • Databricks
  • Terraform
  • Airflow

Itaú Unibanco

Oct 2022 — Jan 2023

Business Intelligence Analyst

Built dashboards and ETL flows for an operations team that needed clearer visibility into process efficiency and costs.

  • Developed dashboards and their ETL pipelines to bring data-driven insights to the operations of the Printing and Tangible Efficiency Team of the largest bank in Latam analyzing business processes and collecting and interpreting data from various sources.
  • Utilized tools such as Python for data extraction, cleaning, and transformation, and Excel for data analysis and creating dashboards.
  • Identified pain points and presented actionable recommendations to improve KPIS related to cost-effectiveness.
  • Python
  • Excel
  • ETL
  • Dashboarding

Itaú Unibanco

Jan 2022 — Oct 2022

Software Engineer

Worked on automation, logistics systems, and operational analytics for banking processes that depended on fewer manual steps and better visibility.

  • Performed process automations, system improvements and data analysis for the largest bank in LATAM; utilized SQL, Python and C#.
  • Developed process automations for the Tangible Logistics team, working on projects such as Invoice Delivery, Credit Card Password Delivery, Card Retention, and Invoicing Controls, monitoring service desk and automation availability indicators, and generating dashboards for analysis and daily reports.
  • Conducted data processing and database generation for analysis utilizing SQL and Python and integration with dashboards created on Excel.
  • SQL
  • Python
  • C#
  • Automation

Nova Singular

Jan 2017 — Jul 2017

Full Stack Developer

Worked on backend features, integrations, and web application flows as part of my first professional software role.

  • Responsible for building and maintaining the server side of websites. Worked with database management and on the implementation of several different processes, including login systems, form validation and website internationalization. I was also responsible for the integration of these processes with the front-end.
  • Node.js
  • Web Development
  • Databases

Selected Wins

A few examples of the kinds of problems I have worked on.

Masking policies with fewer sharp edges

Moved sensitive-data masking from write time to query time, which made governance more flexible and much easier to manage consistently.

12,000+ columns governed across 5,500+ tables

Regulatory reporting without spreadsheet archaeology

Built pipelines for Bacen, AML, and country-specific transaction reporting so compliance workflows could rely on reproducible code and trusted data.

Brazil, Mexico, and Colombia reporting flows

Developer tooling for a large SQL codebase

Created a VS Code extension for DLT SQL with go-to-definition, autocomplete, rename/refactor, and diagnostics across a 1,500+ SQL model codebase.

Better IDE support for a large SQL-heavy workflow

Webhook ingestion built for reliability

Implemented AWS-powered event ingestion flows with API Gateway, Lambda, and S3 to process onboarding data asynchronously and reliably.

Faster integrations and cleaner event processing

Toolbox

This is the stack behind most of the work I've done.

Not an old dog, though - I can learn new tricks!

Languages

  • Scala
  • Python
  • SQL
  • TypeScript
  • C#
  • Node.js

Data and Platform

  • Spark
  • Databricks
  • Airflow
  • AWS
  • Lakehouse
  • ETL
  • Batch and stream pipelines

Backend and Delivery

  • API integrations
  • Webhook ingestion
  • Lambda
  • S3
  • Terraform
  • Observability
  • IaC

Governance and Reporting

  • PII masking
  • Unity Catalog ABAC
  • Regulatory reporting
  • AML pipelines
  • Data quality

Contact

Have a data platform challenge, a backend project, or just a good engineering problem?

Send me a note. I like talking about why things break and how we can make them easier to live with.