How we think about AI.
A plain-language essay on what we mean when we say "AI integration" — what we build, what we refuse, and what we measure.
Most "AI integrations" we see in the wild are a large language model bolted onto a product roadmap as decoration. A press release, a chatbot, a logo on the homepage. That is not integration. That is signage.
Real integration starts with a boring question. What decision, made by a human inside your business today, could be made better, faster, or at a scale that your team cannot sustain? Then we trace the data in, the data out, and what happens when the model is wrong — because it will be.
Start with the boring question.
We do not begin with the model. We begin with the workflow. We sit with the people who do the thing, we watch them do it, and we write down the judgement calls they make without thinking. Most of those calls are not AI problems. Some of them are.
The ones that are AI problems usually have three things in common: they are repetitive, they draw on unstructured information, and they are hard to fully automate but expensive to do by hand. That is the sweet spot.
Plan for the day the model lies.
Every Semantic Kernel we ship comes with a plan for the day it is confidently wrong. That means guardrails, fallback paths, an audit trail a human can read, and an escalation route back to a person. It also means calibrating the model's confidence honestly — a system that says "I do not know" is worth more than one that bluffs.
A model that answers in plain language should also admit uncertainty in plain language. We design for both.Outcomes, measured in money.
No AI project is finished until we can point at the line in the P&L where the integration shows up. That might be reduced handling time, higher first-contact resolution, fewer manual reviews, or a product we could not have shipped at all without it. What it is not: "users feel more engaged".
We do not take on AI work that cannot be measured. If we cannot agree on the metric before we start, the integration is not real — it is theatre.
What this means in practice.
We build with Semantic Kernel because it lets us put production code, traceable prompts, and domain tools in the same place. We use retrieval-augmented generation rather than fine-tuned models wherever possible, because the customer's data changes and the model should not need to be retrained every time. And we document the failure modes before the demo.
If this matches how you want to ship AI, tell us what you are building.