Logo
Luis – Machine learning, Python, AWS, experts in Lemon.io

Luis

From Mexico (UTC-6)

flag
Machine Learning EngineerSenior
AI Engineer
9 years of commercial experience
AI
Banking
Healthcare
Machine learning
Media
Marketplace
Trade
AI software
CRM
Enterprise software
HRMS
NLP software

Luis – Machine learning, Python, AWS

Meet Luis, a skilled Machine Learning Engineer with over 7 years of experience in distributed environments. He excels in Apache Beam, Dataflow pipeline design patterns, Kubeflow pipeline design and deployment, Airflow orchestration, Spark optimization and streaming, Kafka, and developing recommendation engines. Luis has extensive experience designing, deploying, monitoring, and maintaining ML models in production and building MLOps pipeline cycles. Fluent in English and holding a degree in Computer Science, Luis stands to be a valuable asset to various teams!

Main technologies
Machine learning
7 years
Python
8 years
Additional skills
AWS
Apache Airflow
BeautifulSoup
Bash
Apache Spark
Big Data
Apache Kafka
BigQuery
GCP
FastAPI
Docker
Helm
Flask
GitHub Actions
Grafana
Jenkins
Kubernetes
JSON
NLP
Linux
NumPy
Keras
PySpark
Pandas
SQLAlchemy
OOP
R
Scala
scikit-learn
Tensorflow
YAML
.NET
C
DevOps
DigitalOcean
ElasticSearch
GPT-3
Ansible
Cassandra
Golang
Kibana
Prometheus
Terraform
Neo4j
Integration testing
AI
Kafka
SQL
PostgreSQL
SQLite
Yarn
NoSQL
GoogleAPI
Salesforce
Apache
MongoDB
ETL
LLM
Direct hire
Possible
Ready to get matched with vetted developers fast?
Let’s get started today!

Experience Highlights

Lead Machine Learning Engineer
May 2022 - Ongoing3 years 2 months
Project Overview

Senior software engineer for MLOps, Machine Learning infrastructure, Machine Learning engineering, feature extraction, streaming/batch data pipelines preprocessing, vector data bases, data lakes optimization and inference endpoints design.

Responsibilities:

● Developed 'BobAI' by fine-tuning the Home Depot text using the davinci-003 text model. Implemented FastAPI with Uvicorn for LLM interactions and utilized Airflow -> Kubeflow DSL -> Apache Beam pipeline for continuous model refinement with curated text. Employed prompt engineering techniques with a mix of blend factual and task-based queries, resulting in semantically structured responses without requiring additional handler code. ● Developed the 'Shoppable Images' architecture using Apache Beam, PyTorch, and Kubeflow pipelines. ● Designed workflow orchestration DAGs with Apache Airflow, BigQuery, and Google Cloud Storage (GCS). ● Developed Apache Beam pipelines for ETL/ELT and machine learning purposes. ● Utilized Dataflow pipelines for distributed data processing. ● Incorporated pre-trained TorchVision models for image recognition. ● Implemented cloud functions for event-driven mechanisms.

Project Tech stack:
Kubernetes
AWS
AWS CloudFormation
Apache Spark
Git
Confluence
Docker
GCP
Vertex AI
AWS SageMaker
Apache Airflow
Apache Hadoop
Apache Kafka
NumPy
Python
Pandas
FastAPI
AI
Aiohttp
BigQuery
Senior Data Engineer
May 2021 - Apr 202211 months
Project Overview

Senior data engineer for Spark .NET optimization data pipelines and build Data Lake.

Responsibilities:

● Improved Spark .NET job performance by 36% for the FA-Import project. ● Conducted proof of concept (PoC) for migrating SQL Server external DB to a data lake using S3, leveraging DMS, Glue jobs, Delta Spark, and Athena. ● Proficient in profiling and diagnosing Spark components to identify critical vulnerabilities. ● Conducted scientific research to identify critical breaking points in Spark .NET applications. ● Skilled in Spark optimization techniques and fine-tuning dimensioning.

Project Tech stack:
Integration testing
Unit testing
CI
Scrum
Jira
Confluence
Dependency Injection
Design system
SQL
PostgreSQL
Agile
.NET
Apache Spark
AWS
AWS Lambda
Machine Learning Engineer
Mar 2020 - Apr 20211 year 1 month
Project Overview

Software engineer for ML infrastructure, ML engineering and data science.

Responsibilities:

● Implemented Kubeflow pipelines architecture using Vertex AI; ● Refactored ETL pipelines replacing KFP v1 with KFP v2 (ephemeral components); ● Implemented model metadata tracker and data lineage with Kubeflow ● Designed the 'Prism' torch model pipeline in Kubeflow and Vertex AI. ● Implemented services for managing data processing pipeline on Kubeflow version 1. ● Redesigned and automated a full cycle running ETL process in Kubeflow and GCP environments. ● Successfully debugged and redesigned an ETL project on Kubeflow. ● Migrated Kubeflow components from KFP v1 to KFP v2. ● Created and designed lightweight components. ● Developed event-driven functions using Cloud Functions as sensors for detecting bucket updates

Project Tech stack:
Python
Tensorflow
GCP
GCP Compute Engine
Kubernetes
Docker
Git
GitHub
GitHub Actions
Bash
Data Science
Machine leaning
Scikit-learn
NumPy
BigQuery
Big Data
PySpark
Jenkins
Unit testing
Integration testing
Apache Airflow
SQLAlchemy
SQL
SQLite
PostgreSQL
Jira
Confluence
Agile
Design system
Data Modeling
Senior Data Scientist
Jul 2016 - Feb 20203 years 7 months
Project Overview

Development of a new reference architecture and business applications that accelerate and enable a better customer experience, adapting the bank's products to real people.

Responsibilities:

● Optimized Spark data-mining pipeline for improved performance. ● Enhanced efficiency of Spark ML and Mllib jobs through optimization techniques. ● Engaged in Machine Learning engineering activities for distributed systems. ● Carried out responsibilities as a Data Scientist and Machine Learning engineer. ● Developed an Investment product type logit engine. ● Created 'M&Ms' (Minimum & Maximums) left-over balance z-score engine. ● Developed a Markov transition predictor for outlier states engine. ● Implemented Convolution filter for time series smoothing. ● Developed an engine for generating suitable trends for investment plans. ● Implemented and modeled a 360° financial health classification engine 'Check-Up'. ● Conducted research and analysis on Catalyst and Tungsten components for internal documentation. ● Managed task creation, prioritization, and automatic assignment. ● Conducted code reviews to ensure high-quality code standards. ● Collaborated with team members to discuss and review technical solutions. ● Organized local SCRUM meetings for effective project coordination. ● Worked closely with development and DevOps teams. ● Automated day-to-day operations for increased efficiency.

Project Tech stack:
Apache Kafka
Kafka
Apache Spark
PySpark
Apache Hadoop
Docker
NumPy
Python
Scala
Algorithms and Data Structures
Jenkins
IDEA
Prometheus
Grafana
Kibana
Kubernetes
Yarn
Agile
Akka
Scrum
Kanban
Big Data
Bash
BitBucket
Git
Linux
Ubuntu
AI
Data Science
Machine leaning
ElasticSearch
Dependency Injection
Design system
Confluence
Jira
SQL
NoSQL
Lead Data Analytics
May 2015 - May 20161 year
Project Overview

The company provides Flexible Human Capital, Headhunting, Payroll Processing Solutions (PPO), Business & Process Solutions (BPS), Recruitment Process Outsourcing (RPO), and Supplier Management (CWO) services.

Responsibilities:

● Lead pricing and revenue models & Automated recurrent data mining pipelines with R; ● Implemented regression models for pricing prediction; ● Researched time series from historical sales factors; ● Created new sales strategies based on data analytics and hypothesis testing; ● Implemented a solver for optimizing pricing factors; ● Salesforce analytics with RForcecom library for classification models; ● Implemented training internal moocs for Salesforce adoption.

Project Tech stack:
PostgreSQL
SQL
R
Salesforce
Salesforce Commerce Cloud
Python
Algorithms and Data Structures
Visual Basic (VB)
MATLAB

Education

2020
Computer Science
Bachelor of Engineering
2023
MSc Data Science
Master of Science

Languages

English
Advanced

Hire Luis or someone with similar qualifications in days
All developers are ready for interview and are are just waiting for your requestdream dev illustration
Copyright © 2025 lemon.io. All rights reserved.