Logo
Fernando – AI agent orchestration, Multi-agent systems architecture, AI telemetry, experts in Lemon.io

Fernando

From Brazil (UTC-3)flag

AI Agent Architect|Senior
AI Engineer|Senior

Fernando – AI agent orchestration, Multi-agent systems architecture, AI telemetry

Fernando is a senior AI Agent Architect with 8 years of experience and deep expertise in Python, LLMs, multi-agent systems architecture, AI agent orchestration, and RAG. He has led end-to-end delivery of AI-driven platforms in healthtech and legal domains, demonstrating strong product judgment, stakeholder communication, and technical ownership.

8 years of commercial experience in
AI
Healthtech
Legal tech
Productivity
UI/UX
B2B
B2C
AI software
Mobile apps
SaaS
Web development
Software development
Main technologies
AI agent orchestration
2 years
Multi-agent systems architecture
2 years
AI telemetry
2 years
Python
7 years
LLM
2 years
Additional skills
RAG
MCP
LLM orchestration
AI agent development
Vector Databases
LangGraph
LangChain
Prompt engineering
OpenAI
Anthropic
Direct hire
Possible
Ready to get matched with vetted developers fast?
Let’s get started today!

Experience Highlights

Senior AI Engineer
Feb 2026 - May 20263 months
Project Overview

A multi-agent orchestration layer built on top of Claude Code, managing multiple simultaneous LLM agent instances coordinating on development tasks. The system handles autonomous agent decision-making, real-time streaming, and a permission-gated execution model that gates tool use behind explicit user approval policies. The project explores production patterns for agentic systems beyond single-agent prompting, focusing on observability, controlled autonomy, and reliability in long-running multi-agent workflows.

Responsibilities:
  • Architected and built a multi-agent orchestration layer coordinating multiple LLM agent instances in parallel;
  • Designed an autonomous agent for monitoring development sessions and deciding when to act or escalate;
  • Applied advanced prompt engineering to improve tool selection accuracy and agent decision quality;
  • Developed real-time WebSocket streaming for agent outputs with backpressure and connection lifecycle handling;
  • Implemented permission-gated tool execution with explicit approval policies for sensitive actions;
  • Built async backend infrastructure with FastAPI and SQLAlchemy for concurrent agent session management;
  • Integrated speech-to-text and text-to-speech layers for voice-driven agent interaction;
  • Packaged the system for reproducible deployment with Docker.
Project Tech stack:
Python
FastAPI
SQLAlchemy
Claude Code
Claude API
Pydantic
WebSocket
Docker
Multi-agent systems architecture
Tech Lead / AI & Full-Stack Engineer
Feb 2025 - Mar 20261 year 1 month
Project Overview

A healthtech platform for pediatricians in Brazil that combines a specialized clinical forum, AI-driven decision support, and intelligent case management. It bridges the gap between static medical content platforms and real clinical workflows by integrating AI directly into forum discussions, with autonomous participation, RAG over historical cases, and clinical modules for diagnostic analysis, medical chat, and case triage. The main challenge was creating AI that clinicians would trust and adopt in a high-stakes domain, while balancing technical reliability with product judgment.

Project gallery:
Portfolio example for Conecped by Fernando, Tech Lead - AI & Software Engineer
Portfolio example for Conecped by Fernando, Tech Lead - AI & Software Engineer
Portfolio example for Conecped by Fernando, Tech Lead - AI & Software Engineer
Portfolio example for Conecped by Fernando, Tech Lead - AI & Software Engineer
Responsibilities:
  • Led end-to-end architecture and delivery of the platform, owning AI integration, backend, and product decisions;
  • Designed and implemented AI clinical modules (diagnostic analysis, intelligent medical chat, case triage) with structured prompt engineering, chain-of-thought reasoning, output validation, and safety guardrails for medical content;
  • Refactored the AI layer through LangChain to standardize prompt templates, chains, and structured output parsing across all modules;
  • Built an autonomous AI agent that monitors forum discussions via Cloud Functions, used as an async message queue, applying contribution heuristics to decide when to act versus stay silent;
  • Implemented a RAG pipeline over the forum's historical case base with embedding pipelines and vector search, grounding AI responses in past clinical discussions and surfacing similar cases at posting time;
  • Implemented subscription-based access via Stripe Billing, role-based authentication, and normalized transactional data modeling;
  • Built CI/CD pipelines with unit and integration testing, plus performance observability;
  • Shaped the product UX of AI participation, moving away from generic AI-style responses toward concise doctor-to-doctor communication patterns.
Project Tech stack:
Python
LangChain
RAG
LLM integration
Vector Databases
Prompt engineering
Node.js
Typescript
React
Flutter
Firebase
Stripe API
CI
CD
Lead AI Engineer / Co-founder
Nov 2025 - Mar 20264 months
Project Overview

An on-premises RAG assistant for law firms that cannot send confidential client data to external LLM providers. The product runs entirely locally, including a hosted LLM and a retrieval stack tuned for legal documents, and supports the full workflow from ingestion and semantic chunking to embedding generation, vector search, and retrieval-aware response generation. The main challenge was delivering production-grade RAG infrastructure under strict data isolation requirements, with no external API dependencies in the critical path.

Responsibilities:
  • Co-founded the project and led technical design end-to-end, including the full RAG architecture;
  • Designed and implemented the embedding generation pipeline tailored to legal document structure;
  • Built the vector search layer on Qdrant with semantic chunking strategies optimized for legal content;
  • Deployed a locally hosted LLM (Qwen3-32B) on Apple Silicon, evaluating model trade-offs against task and latency requirements;
  • Implemented retrieval-aware prompt engineering with citation grounding, ensuring responses reference specific source passages;
  • Integrated MCP-based document ingestion from Google Drive, enabling automated and structured document onboarding;
  • Architected the system to operate fully air-gapped, meeting the data isolation requirements of legal clients;
  • Defined product positioning and go-to-market approach in partnership with stakeholders, including hardware selection and operational cost modeling.
Project Tech stack:
Python
Qdrant
LLM
AI API integration
Vector Databases
RAG
Prompt engineering
MCP
FastMCP
Senior AI & Full-Stack Engineer
Jun 2024 - Jun 20251 year
Project Overview

A no-code landing page builder that lets users describe what they need in natural language and generates fully structured, editable landing pages. It was built to turn LLM output into reliable, deterministic UI content that the editor can consume directly, with validation, fallbacks, and idempotent updates instead of free-form text. The product was designed and delivered as a single-founder effort, from the AI generation pipeline to the editor and deployment infrastructure.

Project gallery:
Portfolio example for Kreat.me by Fernando, Senior AI & Software Engineer
Portfolio example for Kreat.me by Fernando, Senior AI & Software Engineer
Portfolio example for Kreat.me by Fernando, Senior AI & Software Engineer
Responsibilities:
  • Built the frontend and backend of the platform, from initial concept to deployed platform;
  • Designed the product strategy, user experience, and technical architecture end-to-end;
  • Engineered the LLM layer that translates natural language input into deterministic JSON/YAML mapped to UI components (layout, copy, CTAs, forms);
  • Implemented validation, fallback handling, content scoring, and idempotent updates to ensure reliable generation across varied user inputs;
  • Defined the structured output contract between the LLM and the rendering layer, enforcing schema validity end-to-end;
  • Built the editing flow that allows end users to iterate on AI-generated pages without breaking structural integrity;
  • Set up automated deployment pipelines with GitHub Actions for reliable releases;
  • Balanced model selection, prompt design, and latency to maintain responsive UX during interactive AI generation.
Project Tech stack:
Python
React
Flutter WEB
LLM
Prompt engineering
Data Structures
JavaScript
Typescript
Node.js
Firebase
GitHub Actions

Education

2022
Electronics and Automation Engineering
Master's degree.

Languages

French
Advanced
Portuguese
Advanced
English
Advanced

Hire Fernando or someone with similar qualifications in days
All developers are ready for interview and are are just waiting for your requestdream dev illustration
Copyright © 2026 lemon.io. All rights reserved.