// ai architecture & technical advisory

I build AI systems that don't fall apart in production.

Most AI projects look great in demos. They collapse when real users arrive. I help CTOs, founders, and engineering teams design and ship AI systems that actually hold up — with the architecture, reliability, and documentation to prove it.

25+
years building production systems
3,000+
active installations of my AI plugin
#1
ByteDance Global Coze AI Challenge
40%
support ticket reduction with RAG chatbot
// what I keep hearing

Does any of this sound familiar?

These are not edge cases. They're the default outcome when AI projects skip architecture and go straight to implementation.

"It worked perfectly in the demo."

Your LLM system performed beautifully in controlled conditions. Production traffic, messy real user inputs, and unexpected edge cases exposed the gap between prototype and product.

"Our RAG keeps returning wrong answers."

Retrieval-Augmented Generation is not plug-and-play. Chunking strategy, embedding model choice, re-ranking, and context window management all require deliberate architecture decisions.

"We have AI in the roadmap but no one who can own it."

Your engineering team is strong, but nobody has shipped a production LLM system before. You need senior guidance without the cost and delay of a full-time hire.

"Our automation breaks every time something changes."

Agent workflows and automation pipelines built without proper error handling, observability, and fallback logic fail silently — and team confidence erodes every time.

"Hallucinations are killing our credibility."

Unconstrained LLM outputs in customer-facing flows are a trust problem. Grounding, output validation, and feedback loops are not optional — they're architectural requirements.

"We're scaling and the system is buckling."

What worked at 100 requests per day breaks at 10,000. Token limits, latency, costs, and rate limits need to be designed for — not discovered in production.

// what I actually do

Strategic AI architecture for systems that have to work.

01

RAG Architecture & Design

End-to-end retrieval pipeline design: data ingestion, chunking strategy, vector store selection, embedding optimization, re-ranking, and hallucination containment.

02

LLM System Design

Prompt engineering at the architecture level, model selection, context management, output validation, and reliability patterns for production LLM deployments.

03

Agent Workflow Architecture

Multi-agent orchestration, tool use patterns, human-in-the-loop design, and error recovery systems that don't require constant babysitting.

04

AI Infrastructure Decisions

Infrastructure trade-offs: self-hosted vs. API-based models, caching strategies, cost optimization, latency budgets, and monitoring setup.

05

Architecture Reviews

Structured review of your existing AI implementation — finding failure modes, architectural debt, and scalability blockers before they hit production.

06

Technical Advisory & Fractional Leadership

Ongoing strategic guidance for engineering teams and leadership: architecture decisions, vendor evaluations, team upskilling, and AI roadmap ownership.

// consulting engagements

How we can work together.

Every engagement starts with understanding the actual problem — not the reported symptom. Scope, timeline, and deliverables are defined before any work begins.

architecture

AI Architecture Assessment

A structured deep-dive into your current or planned AI system. I identify architectural gaps, reliability risks, and scalability blockers — and deliver a written report with prioritized recommendations you can act on immediately.

  • Codebase and architecture review
  • Written assessment report with prioritized findings
  • Architecture decision recommendations
  • One follow-up session to walk through findings
Fixed scope · 1–2 weeksStart a conversation
implementation

Production AI System Build

Hands-on architecture and implementation of a production AI system: RAG pipeline, LLM integration, agent workflow, or automation infrastructure. I build it, document it, and hand it off working.

  • Architecture design and implementation
  • Integration with your existing stack
  • Error handling and observability setup
  • Full documentation and handover
Project-based · 4–12 weeksStart a conversation
advisory

Fractional AI Architect

Ongoing strategic and technical guidance embedded in your team. I own AI architecture decisions, review implementations, advise on vendor and model choices, and serve as the senior technical voice your team needs to ship confidently.

  • Weekly architecture reviews and guidance
  • Async availability for technical decisions
  • Quarterly architecture roadmap sessions
  • Team office hours and design reviews
Monthly retainer · OngoingStart a conversation
rescue

AI Project Rescue

Your AI project is stuck, broken, or about to go live with known problems. I come in, diagnose what went wrong, and build the path forward — whether that means fixing what's there or re-architecting the critical parts.

  • Root cause analysis and written diagnosis
  • Immediate stabilization recommendations
  • Prioritized remediation roadmap
  • Optional hands-on remediation
Fixed scope · 1–3 weeksStart a conversation
// what goes wrong

Why most AI projects fail in production.

After two years of building and reviewing AI implementations across dozens of companies, the failure patterns are consistent. None of them are mysterious.

01

Architecture decisions made by the wrong people at the wrong time.

LLM system architecture gets decided by developers under sprint pressure instead of by someone with production experience. By the time the problems surface, the architecture is already baked in.

02

Retrieval isn't treated as an engineering problem.

Most RAG implementations use default chunking, default embeddings, and no re-ranking. The result is a system that returns plausible-sounding wrong answers with confidence. Fixing this after the fact is expensive.

03

No observability, no feedback loops.

Teams deploy LLM features with no way to measure whether they're working. Without logging, evaluation pipelines, and user feedback mechanisms, you can't improve what you can't see.

04

Prompt engineering treated as a magic input, not a design surface.

A good prompt is a system specification. When prompts are written ad-hoc and not maintained like code, they drift, break with model updates, and become impossible to debug systematically.

05

No plan for what happens when the LLM is wrong.

Production AI systems need graceful degradation, output validation, and fallback logic. Systems built without these patterns fail loudly or — worse — silently, in ways users notice before you do.

Related reading: I wrote a detailed breakdown of this pattern: Why AI Projects Fail After the Demo.

// how I work

Structured. Documented. Direct.

I've built AI systems for production environments for long enough to know where most engagements break down — and it's rarely the technology. It's unclear scope, missing documentation, and no definition of done.

Phase 01

Diagnosis

Before I write a line of code or a single recommendation, I need to understand what's actually broken. This means access to the codebase, system logs, and an honest conversation about what's been tried and what failed.

Phase 02

Architecture

I define what we're building, what's explicitly out of scope, and what "done" looks like. Architecture decisions are written down before implementation starts. No sliding scope.

Phase 03

Build

Implementation with the error states, edge cases, and observability that production environments demand. I work in focused blocks — not in continuous Slack threads.

Phase 04

Handover

Everything I build gets documented as part of the deliverable: architecture decisions, integration guides, runbooks. The goal is a system that works without me. If you need me again, it's for the next problem.

I'm remote-first, based in São Paulo (UTC-3). I communicate through structured written updates, not constant availability. You'll always know what I'm working on and when it'll be ready.

Read: How I Work →
// track record

Production AI systems, not presentation slides.

I don't have case studies with projected outcomes. These are things that shipped, that ran on real traffic, and that are still running.

25+
years in production systems engineering
3,000+
active WP-AutoInsight plugin installations
17,000+
developer followers on Dev.to
40%
support ticket reduction with production RAG chatbot

Selected work

Paulo has outstanding skills to organize and communicate demands in situations that seem absolutely chaotic. He has a lot of technical knowledge and is able to communicate efficiently with both technical and lay people.
Gus FuneChief Operating Officer, Courate
Paulo is a great sysadmin. Every website and blog that Paulo has taken care of never crashed, even during traffic spikes with thousands of visits. The main reason for keeping him was not just his technical competence, but the fact that he is one of the most reliable people I've met in my entire life.
Edney SousaCEO, Interney

Companies I've worked with

Ovolo Hotels
Ola Cabs
Porto Seguro
Mojo Nomad
SportTechie
Colette Baron-Reid
Interney
Namu Cursos

Recognition

  • 1st Place — ByteDance Global Coze AI Challenge (2024)
  • Technical book published — 4.6★ on Amazon (still in print)
  • 17,000+ developer followers on Dev.to

Speaking & Teaching

  • Campus Party Brasil — Speaker (2009, 2010, 2011, 2012)
  • Sebrae Empreendedor — Speaker, Belém (2010)
  • Senac Franca — Instructor (2009)
  • Apadi — WordPress instruction (2010–2013)
  • ComSchool — WordPress instruction (2014–2016)
// faq

Common questions.

What is an AI architecture consultant?

+
An AI architecture consultant is a technical advisor who designs the systems, infrastructure, and decision frameworks for production AI deployments. This is different from AI development work: I define how systems should be built — model selection, retrieval strategy, data pipelines, failover logic, observability — and either implement directly or guide the team that does.

How is this different from hiring an AI development agency?

+
Agencies build. I architect. The difference matters in production. Agencies implement against a spec; I help you define the spec itself — which models to use, which retrieval approach to take, where the failure modes are, and what your architecture needs to look like before a line of code is written. I'm the person you bring in to prevent the problems that agencies are hired to fix.

What does a fractional AI architect actually do?

+
A fractional AI architect is a senior technical leader embedded part-time in your team. In practice: I own AI architecture decisions, review implementations before they ship, advise on model and vendor choices, conduct design reviews with your engineers, and serve as the senior technical voice in conversations with leadership. You get strategic AI leadership without a full-time executive hire.

My team already has LLM experience. Why would I need this?

+
Having developers who can call an LLM API is different from having architecture designed for production reliability. Most LLM experience is prompt engineering and API integration — which is valuable, but it doesn't cover RAG pipeline design, context management at scale, output validation frameworks, observability setups, or failure mode analysis. I fill the gap between "we can build an LLM feature" and "we can ship it and trust it."

What does a RAG architecture review cover?

+
A RAG architecture review covers the full retrieval pipeline: data ingestion strategy, document chunking approach, embedding model selection, vector store configuration, retrieval accuracy (precision and recall), re-ranking setup, prompt construction, context window management, hallucination containment patterns, and evaluation methodology. I deliver a written report with prioritized findings and specific recommendations.

Do you work with teams outside Brazil?

+
Yes. All my consulting work is remote. I'm based in São Paulo, Brazil (UTC-3), but work with clients across North America, Europe, and Asia. Timezone overlap for synchronous calls is generally manageable with a scheduling buffer.

How long does an AI architecture assessment take?

+
Typically one to two weeks from access to deliverable. This includes codebase review, architecture analysis, one clarification session, and the written report. For larger or more complex systems, scope can be extended accordingly.

What if our AI project is already broken?

+
That's what the AI Project Rescue engagement is for. I diagnose what went wrong, separate architectural problems from implementation problems, and deliver a prioritized roadmap for stabilization. If you need hands-on remediation as well, that can be scoped as an additional engagement.
// let's talk

Your AI project deserves production-grade architecture.

If you're building AI systems that need to work reliably — under real traffic, with real users, with real consequences for failure — I can help you get there. Start with a short conversation about where you are and what you're trying to solve.

No pitch decks. No discovery call with a sales team. You talk to me.