Nairobi · Production Systems · Africa-First

Real AI & Data Systems, Built Under African Constraints.

Most technical content assumes stable power, cheap GPUs, and fast internet. I document what actually works in production — systems engineered in Kenya, tested against the constraints most tutorials never mention.

Read by engineers building real systems.

Scroll

Trending

From the Field

Systems analysis, architecture deep dives, production postmortems, and applied research. Written from experience, not speculation.

Building a RAG System That Survives Production
In Progress

Building a RAG System That Survives Production

14 min read·Mar 2025·AI / ML

Beyond the demo: chunking strategy, hybrid retrieval, latency budgets, and why your embedding model matters more than your vector DB.

Coming soon
Hallucination as an Engineering Problem: H-Score
In Progress

Hallucination as an Engineering Problem: H-Score

13 min read·Apr 2025·AI / ML

Most teams treat hallucination as a model quality issue. I built a metric to measure it as a system property.

Coming soon
When RL Meets RAG: Systems That Know What They Don't Know
In Progress

When RL Meets RAG: Systems That Know What They Don't Know

15 min read·Apr 2025·AI / ML

Combining reinforcement learning with retrieval to build enterprise AI that adapts without hallucinating.

Coming soon
How M-Pesa Processes 10 Million Transactions a Day
In Progress

How M-Pesa Processes 10 Million Transactions a Day

12 min read·Mar 2025·Production Systems

Inside the queuing, validation, and settlement layers keeping East Africa's most critical financial rail running under load.

Coming soon
Inside a Real-Time Fraud Detection System
In Progress

Inside a Real-Time Fraud Detection System

10 min read·Mar 2025·Production Systems

Architecture breakdown of a streaming fraud detection system: feature engineering, model serving, and the false positive problem.

Coming soon
Why Our Vector Search Was Slow: A Postmortem
In Progress

Why Our Vector Search Was Slow: A Postmortem

8 min read·Apr 2025·Production Systems

HNSW index misconfiguration, chunk boundary issues, and the query pattern we never tested. A debugging session, documented.

Coming soon
Designing for Intermittent Connectivity
In Progress

Designing for Intermittent Connectivity

9 min read·Mar 2025·Architecture

What most guides skip: building data pipelines that handle 4-hour connectivity gaps without data loss or downstream inconsistency.

Coming soon
Building AI for Environments That Don't Fit the Defaults
In Progress

Building AI for Environments That Don't Fit the Defaults

11 min read·Apr 2025·Architecture

What changes when you're building AI systems for hard cost ceilings, unreliable APIs, and users whose defaults differ from Silicon Valley's.

Coming soon
Three Ways Our ML Model Broke in Deployment
In Progress

Three Ways Our ML Model Broke in Deployment

10 min read·Feb 2025·MLOps

Data drift, silent failures, and a misconfigured feature store. A postmortem on what we missed in testing.

Coming soon
The Hidden Cost of Switching Embedding Models
In Progress

The Hidden Cost of Switching Embedding Models

7 min read·Mar 2025·MLOps

We changed our embedding model mid-project. Here's what broke, what we didn't anticipate, and what re-indexing at scale actually costs.

Coming soon

2 published · 10 in progress

Ongoing

Signature Series

A 3-part production series. Read in order. Each part builds on the last.

PART 01

Signature Series

Ongoing

How It Actually Runs

Real systems, explained without the marketing diagram.

Every system that runs at scale has a gap between the architecture diagram and what's actually deployed. This series tears down real systems — payments infrastructure, data platforms, ML pipelines — and explains exactly how they work in production, constraints included.

System internalsArchitecture tradeoffsAfrican deployment contexts

In this series

01

How M-Pesa Handles 10 Million Transactions a Day

02

Inside a Real-Time Fraud Detection Pipeline

03

How Recommendation Engines Work at Low Data Volume

Starting soon

PART 02

Signature Series

Ongoing

If I Built This Today

Retrospective system design. What experience actually changes.

Looking back at systems I actually built in production. Not a highlight reel — the real decisions I made, why I made them given the constraints at the time, and what I would do completely differently with what I know now.

Decision archaeologyConstraint-driven tradeoffsHard-won redesigns

In this series

01

The Data Pipeline I Regret Building

02

Our First ML System: Every Wrong Assumption, Documented

03

Rebuilding a Feature Store After It Failed at Scale

Starting soon

PART 03

Signature Series

Ongoing

What Breaks in Production

Honest postmortems. No blame, no hand-waving.

Production failures documented in full: what the symptoms were, what we initially diagnosed, what was actually wrong, and what changed so it would not happen again. Written as engineering records, not incident theater.

Real failure modesRoot cause analysisWhat changed afterward

In this series

01

The Silent Data Loss We Found Three Months Late

02

When Our Model Passed All Tests and Failed in Production

03

The Latency Regression We Shipped to 100k Users

Starting soon

Start with Part 01 and read in sequence for the full picture.

Applied System Design

Research

Working Paper2024

A Reliability-First Framework Combining Adaptive AI with Hallucination Safeguards

Autonomous AI agents in enterprise create a specific tension: the system needs to adapt to new information, but it also needs to produce outputs you can stake a business decision on. I designed a framework that addresses both — combining Reinforcement Learning, Retrieval-Augmented Generation, and uncertainty quantification. The core contribution is the H-Score, a metric I built to measure hallucination risk as a system property across IT operations, healthcare, and logistics environments.

H-Score

A metric for measuring hallucination risk in enterprise AI systems

CWAS

Confidence-Weighted Action Selection: RL that knows when to be cautious

REAC

A benchmark corpus for testing enterprise AI agents against real workflows

Reinforcement LearningRAGHallucination MitigationEnterprise AIAI Safety
Read Full Paper

Experiments · Prototypes · Real Implementations

System Lab

Real systems built under real constraints. Each entry documents the problem, the design decisions, and what actually happened — including what failed. Not a portfolio. An engineering record.

QdrantBGE-smallFastAPIPython
SQLiteDuckDBAirflowPython
PythonSentence TransformersNLI ModelsPrometheus
FaustIsolation ForestRedisFastAPI

About the Author

Sam Odongo

Sam Odongo

Senior Data Engineer · Applied Data Scientist

Open to remote roles

I build AI and data systems where most architectures fail. Low bandwidth, unstable infrastructure, real users, real constraints.

Most technical content assumes stable power, cheap GPUs, and fast internet. I work in Nairobi, Kenya, where none of those are guaranteed. That gap between theory and production is where I live and where I write from.

I hold an MSc in Artificial Intelligence and have spent years building production data systems across fraud detection, real-time streaming, RAG pipelines, and MLOps infrastructure. Every system I document here has been tested against constraints most tutorials never mention.

If you are building systems that have to work in the real world, not just on a benchmark, this is where you should be.

Newsletter

Join engineers learning how to build AI systems that don't break in production.

Real systems. Real constraints. Written from Nairobi, tested against the infrastructure most tutorials never mention. No spam. No theory. Unsubscribe anytime.

Written by a practitioner, not a marketer.