Intelligence Is a Stack

Everyone talks about AI like it’s one thing. A model. A chatbot. A magic box.

It’s not. Intelligence — the kind that actually works in production — is a stack. Layers of engineering, each doing one job, each depending on the one below it. Break any layer, the whole thing falls apart.

The Stack

Here’s how I think about it:

┌─────────────────────────────────┐
│  6. Application Layer           │  ← Products, UIs, user-facing systems
├─────────────────────────────────┤
│  5. System Layer                │  ← Pipelines, APIs, orchestration
├─────────────────────────────────┤
│  4. Reasoning Layer             │  ← LLM inference, chain-of-thought
├─────────────────────────────────┤
│  3. Representation Layer        │  ← Embeddings, vector spaces
├─────────────────────────────────┤
│  2. Model Layer                 │  ← Neural networks, training
├─────────────────────────────────┤
│  1. Computation Layer           │  ← GPUs, TPUs, hardware
└─────────────────────────────────┘

This is not a metaphor. This is how real AI systems are built.

Layer 1: Computation

No GPU, no party. Every intelligent system starts with raw compute. CUDA cores, tensor cores, memory bandwidth. People underestimate how much of AI progress is just hardware getting faster.

You can have the best architecture in the world. If you can’t fit it in VRAM, it doesn’t exist.

Layer 2: Model

This is where most of the hype lives. Transformers, diffusion models, CNNs. The layer everyone obsesses over.

But here’s the thing — a model sitting in a Jupyter notebook is not a product. It’s a science experiment. The model layer is necessary but wildly insufficient.

Layer 3: Representation

Embeddings are how machines see the world. Turn text into vectors. Turn images into vectors. Turn everything into geometry.

# This single line is doing more than most people realize
embedding = model.encode("what is intelligence?")
# Result: a 768-dimensional vector that captures meaning
# Not words. Not syntax. Meaning.

This layer is what makes retrieval, similarity search, and clustering possible. Without good representations, your RAG system is just keyword search with extra steps.

Layer 4: Reasoning

LLM inference. Chain-of-thought. The layer that makes people think machines can “think.”

They can’t. But they can do something useful: pattern completion at scale. Given enough context, a large language model can produce outputs that look like reasoning. Sometimes it even is reasoning — but you can never be sure which time is which.

The dangerous part isn’t that LLMs can’t reason. It’s that they sometimes can, and you don’t know when.

Layer 5: System

This is where I spend most of my time. The system layer.

Kafka for event streaming. gRPC for low-latency communication. Triton for model serving. Docker for isolation. Jenkins for CI/CD. MinIO for storage.

# A typical inference pipeline
services:
  gateway:
    image: api-gateway:latest
    depends_on: [model-server, queue]

  model-server:
    image: tritonserver:latest
    deploy:
      resources:
        reservations:
          devices:
            - capabilities: [gpu]

  queue:
    image: kafka:latest
    # Because you never serve models synchronously in production
    # unless you enjoy being paged at 3 AM

Most AI failures are not model failures. They’re system failures. The model works fine — the pipeline around it doesn’t.

Layer 6: Application

The part the user sees. A chat interface. A medical imaging tool. A recommendation engine.

This layer is deceptively simple-looking. But it has to handle:

Why This Matters

When someone says “we’re building an AI product,” I ask: which layers are you actually building?

Most startups are building Layer 6 on top of someone else’s Layers 1-5. That’s fine — you don’t need to manufacture GPUs to build a useful product. But you should know which layers you own and which layers own you.

What you controlWhat you understand
More layers → more differentiationFewer layers → more dependency
More complexity → more hiringLess complexity → more vendor risk

The intelligence stack is not about doing everything. It’s about knowing where you sit in it.


This blog is about the full stack. Not just the model. Not just the infra. The whole thing — from GPU to user interface, from training loop to production deploy.

That’s what intelligence engineering actually looks like.