Logo
Amelia – SQL, Airflow, Python, experts in Lemon.io

Amelia

From Brazilflag

Data Engineer|Strong senior

Amelia – SQL, Airflow, Python

Amelia is an accomplished Data Engineer with in-depth expertise in streaming and batch data architectures, including hands-on experience with Kafka, Spark, and Airflow. She has demonstrated ownership of production systems, architectural decision-making, and end-to-end pipeline design on AWS and GCP. Amelia also brings experience as a Tech Lead and a strong track record in both startup and enterprise environments, such as Tripadvisor.

16 years of commercial experience in
Advertising
Business intelligence
Consulting services
Data analytics
Tourism
PaaS
Virtual assistants
Main technologies
SQL
8 years
Airflow
5 years
Python
6 years
AWS
4 years
Additional skills
PySpark
Web scraping
Snowflake
OpenAI
FastAPI
PostgreSQL
LLM
n8n
BigQuery
GCP
ETL
API Gateway
Apache Airflow
Git
Docker
Linux
Maven
NativeScript
Distributed Systems
Java
AWS CloudFormation
AWS Lambda
Amazon SNS
Apache Spark
Amazon S3
Terraform
Apache Kafka
Redshift
DBT
Direct hire
Possible
Ready to get matched with vetted developers fast?
Let’s get started today!

Experience Highlights

Data Engineer & Solution Architect
Oct 2025 - Dec 20252 months
Project Overview

It's a data and conversational automation platform designed to ingest data daily from an external API, process both historical and incremental datasets, and expose insights through a WhatsApp bot powered by an LLM.

Responsibilities:
  • Designed the end-to-end data architecture.
  • Implemented scalable ETL pipelines with Apache Airflow, processing a 24-month historical backfill and daily incremental updates.
  • Modeled data to support both analytical and conversational queries.
  • Integrated a WhatsApp chatbot with an LLM via n8n, ensuring reliability, observability, and cost-efficient MVP operation.
Project Tech stack:
Apache Airflow
PostgreSQL
API Gateway
n8n
LLM
ETL
GCP
BigQuery
Distributed Data Engineer
Mar 2025 - Sep 20255 months
Project Overview

It's a custom Trino connector for a proprietary big data and AI-oriented database, enabling distributed SQL query execution and metadata federation across large-scale datasets.

Responsibilities:
  • Designed and implemented core connector components for schema discovery, query planning, and predicate pushdown.
  • Optimized performance for handling large datasets.
  • Ensured full compatibility with the Trino SPI.
  • Collaborated with stakeholders to align technical solutions with production requirements.
Project Tech stack:
Java
SQL
Distributed Systems
Docker
Linux
Maven
Git
NativeScript
Senior Data Engineer
Jun 2024 - Dec 20246 months
Project Overview

It's a cloud-native data platform on AWS, supporting both batch and analytical workloads. The platform enabled data ingestion, transformation, governance, and visualization, providing high-quality datasets and insights for business decision-making.

Responsibilities:
  • Built and maintained scalable data pipelines.
  • Implemented data modeling and governance strategies.
  • Optimized analytical queries using Iceberg and Athena.
  • Delivered interactive data visualizations to support business stakeholders.
Project Tech stack:
AWS
AWS Lambda
Amazon S3
Amazon SNS
AWS CloudFormation
Apache Spark
PySpark
Python
Docker
Senior Data Engineer
Aug 2023 - Jun 20249 months
Project Overview

It's a robust data platform handling both real-time streaming and batch processing workloads.

Responsibilities:
  • Implemented and optimized streaming and batch data pipelines.
  • Improved data reliability, scalability, and performance.
  • Managed infrastructure as code using Terraform.
  • Collaborated with cross-functional teams to deliver robust data solutions.
Project Tech stack:
AWS
Amazon S3
Redshift
Apache Airflow
Apache Kafka
Apache Spark
PySpark
Terraform
Python
Docker
Senior Data Engineer
Dec 2021 - Jul 20231 year 6 months
Project Overview

It's a large-scale enterprise data platform following data mesh principles. The platform enables domain-oriented data ownership, scalable data products, and both batch and streaming analytics across multiple business units.

Responsibilities:
  • Contributed to the design and implementation of a data mesh architecture.
  • Built scalable and maintainable data pipelines.
  • Enhanced observability and system reliability.
  • Supported multiple teams in delivering high-quality data products.
Project Tech stack:
AWS
Amazon S3
Apache Airflow
Apache Kafka
Apache Spark
PySpark
DBT
Terraform
Python
Docker
Java

Education

2014
Computer Science
Bachelor's degree

Languages

English
Advanced

Hire Amelia or someone with similar qualifications in days
All developers are ready for interview and are are just waiting for your requestdream dev illustration
Copyright © 2026 lemon.io. All rights reserved.