Vasco
From Germany (GMT+2)
9 years of commercial experience
Lemon.io stats
Vasco – AWS, Python, OpenAI
For the recent 8 years, Vasco has been involved heavily in tackling challenges within data science and AI engineering domains. His contributions could not be underestimated: from IoT to AI products of various sizes - Vasco elevated businesses' potentials to next levels.
With senior-level hard skills polished with solid soft skills, Vasco checklists the must-haves for an exceptional asset to any project!
Main technologies
Additional skills
Ready to start
15.01.2024Direct hire
Potentially possibleExperience Highlights
Lead Data Scientist
A digital advertising platform focused on optimizing campaign performance and maximizing ROI. It leveraged advanced algorithms and data-driven insights to deliver targeted ads across multiple channels. It helped businesses achieve greater reach, engagement, and efficiency in their advertising efforts.
Vasco was responsible for the following:
- ideated and developed AI-driven product solutions tailored to client needs, leveraging LangChain and LlamaIndex for LLM application development and knowledge integration;
- designed and deployed Retrieval Augmented Generation (RAG) pipelines to improve information retrieval and contextual responses, using Pinecone and Weaviate for vector search optimization;
- built and fine-tuned Large Language Models (LLMs) for chatbot applications, utilizing OpenAI APIs, prompt engineering, and generative AI frameworks;
- created scalable RESTful APIs with FastAPI to deploy AI and ML services, ensuring seamless integration with existing systems;
- deployed generative AI models in production, focusing on quantization, inference optimization, and deployment using PyTorch, TensorFlow, and Hugging Face Transformers;
- applied computer vision techniques for creative tagging and performance analysis, leveraging CLIP, Deep Learning, and Keras, driving campaign ROI improvements;
- designed and implemented MLOps workflows, including monitoring, versioning, and scaling AI systems with Kubernetes, Docker, and cloud platforms like AWS EKS, GCP, and Azure;
- developed robust data pipelines for real-time analytics and machine learning workflows, integrating BigQuery, DBT, and Databricks to enhance performance and scalability;
- integrated vector databases (e.g., Pinecone, Weaviate) and embeddings for advanced semantic search and retrieval in AI-driven applications.
Lead Data Scientist
A solution that automates the generation of AdWords text and optimizes bidding strategies for recruitment campaigns. Designed to enhance performance, it streamlines ad creation while ensuring cost-effective delivery to the right audience. This tool helps employers maximize their recruitment ROI with minimal effort.
Vasco's main responsibilities included:
- developed and architected very complex real-time ML systems multiple times to production as a team;
- developed and managed an automated bidding model optimized to maximize revenue/margin, generating a 6% uplift in spend and a decrease in CPA;
- developed and managed an automated text generation models (keywords, ad copy) using classic NLP and LLM techniques (self-hosting and external, prompt engineering, persuasion techniques);
- addressed third-party cookie deprecation by using probabilistic clustering in full-funnel attribution, using Spark to pipeline data to enhance conversion tracking and data accuracy;
- leveraged causal inference (double robust learners — econml) to refine conversion tracking, boosting accuracy and stakeholder confidence on top of clustering output;
- developed, A/B tested, and deployed with Docker/Kubernetes a linear regression-based algorithm, achieving a 3.7% YoY revenue increase and enhancing pacing KPIs by 35%;
- successfully led A/B testing for new product features, driving improvement in core company KPIs;
- collaborated with cross-functional teams to enable data-driven decision-making and stakeholder buy-in; recruited elements for the DS team.
Lead Data Scientist
A solution that automated the generation of AdWords text and optimized bidding strategies for recruitment campaigns. Designed to enhance performance, it streamlined ad creation while ensuring cost-effective delivery to the right audience. This tool helped employers maximize their recruitment ROI with minimal effort.
Vasco carried out the following responsibilities:
- created scalable RESTful APIs with FastAPI to deploy AI and ML services, ensuring seamless integration with existing systems;
- deployed generative AI models in production, focusing on quantization, inference optimization, and deployment using PyTorch, TensorFlow, and Hugging Face Transformers;
- applied computer vision techniques for creative tagging and performance analysis, leveraging CLIP, Deep Learning, and Keras, driving campaign ROI improvements;
- designed and implemented MLOps workflows, including monitoring, versioning, and scaling AI systems with Kubernetes, Docker, and cloud platforms like AWS EKS, GCP, and Azure;
- developed robust data pipelines for real-time analytics and machine learning workflows, integrating BigQuery, DBT, and Databricks to enhance performance and scalability;
- integrated vector databases (e.g., Pinecone, Weaviate) and embeddings for advanced semantic search and retrieval in AI- driven applications.
Data Scientist
A technology-driven platform offered advanced solutions for data analysis, visualization, and decision-making. It combined cutting-edge tools with intuitive interfaces to empower businesses to transform complex data into actionable insights. The platform supported a wide range of industries, enabling smarter strategies and enhanced outcomes.
Among Vasco's contributions, there were the following:
- developed a system to optimize portfolio risk: key risk KPIs (implied volatility, maximum drawdown) decreased on average 7.6%;
- delivered a system for option pricing using deep learning, leading to an improvement in average daily returns of 2.7%;
- managed end-to-end data processing, enhancing system performance through effective sourcing, preprocessing, and partitioning for model training and inference;
- utilized NLP techniques to analyze tweets, creating a feature store (embedding) for machine learning models that enhanced stock market understanding;
- developed a back-and-forward testing framework.