An empirical study of the natural-language understanding system behind InHouse America's search bar — its architecture, intent resolution, conversational memory, and the design choices that let it answer messy human queries with precision.
The AI Brain is the natural-language understanding (NLU) engine behind every search on InHouse America. Where a traditional search bar performs lexical matching against a product index, the AI Brain treats each query as a linguistic event — parsing intent, resolving ambiguity, binding context from prior turns, and routing the resolved meaning to the correct downstream subsystem (catalog search, Pricing Feature, content lookup, or guided navigation). Across 18,900 synthetic queries generated by our internal test harness, the Brain correctly classified user intent in 97.4% of cases, resolved follow-up references ("more like that," "cheaper ones") in 94.1%, and reduced zero-result results by 71% compared to a lexical baseline. This paper documents its architecture, design signals, and measured behavior.
Real shoppers don't speak in keywords. They speak in fragments, comparisons, and follow-ups: "something cheaper," "the blue one," "ladies version," "what about under twenty." A lexical search bar sees these as noise. The AI Brain sees them as structured intent over a shared context.
Three observations from our internal test scenarios (Nov 2025 – Apr 2026) drove the design:
The AI Brain exists so that a shopper can think out loud and still land on the right product.
The Brain runs as a four-stage pipeline. Each stage is independently observable and replaceable, which is what allows the system to evolve without regressing prior behavior.
raw query ─► [1] Normalize ─► [2] Parse ─► [3] Resolve ─► [4] Route ─► subsystem
│ │ │ │
│ │ │ └─ catalog | pricing | content | guide
│ │ └─ context binding, coreference, slot fill
│ └─ tokens, entities, modifiers, intent candidates
└─ casing, unicode, voice→text fixups, profanity scrub
The Brain recognizes nine top-level intents. Each carries a confidence score; below 0.55 the router escalates to a clarifying suggestion rather than guessing.
| Intent | Example query | Routed to |
|---|---|---|
| Find product | "navy crewneck" | Catalog search |
| Budget search | "under twenty dollars" | Pricing Feature |
| Compare | "cheaper ones," "smaller size" | Catalog ranker (re-rank) |
| Refine | "in blue," "long sleeve" | Catalog ranker (filter) |
| Recommend | "something for my dad" | Recommender |
| Locate | "where's my order" | Account / orders |
| How-to | "how do returns work" | Content lookup |
| Greeting / chitchat | "hi," "thanks" | Acknowledge, no search |
| Out-of-scope | "what's the weather" | Polite refusal |
Each session carries a short-lived memory object: the last category, last price tier, last comparator, and last result set. The memory is bounded (oldest entries decay after five turns) and is the substrate that makes follow-ups work.
memory = {
category: "ladies-shorts",
price_tier: { mode: "ceiling", value: 30 },
last_results: [sku_a, sku_b, sku_c, ...],
comparator: null,
turn: 4
}
When the user types "cheaper ones," the resolver consults memory.price_tier, lowers the ceiling by one tier (e.g. $30 → $20), and re-issues the query against the same category. No re-typing required.
We evaluated the Brain on 18,900 synthetic queries produced by our internal test harness, covering scenarios constructed between January and April 2026. Queries were stratified across typed (62%), voice (29%), and follow-up (9%) inputs. Three metrics were measured:
An automated regression suite re-validated 1,400 randomly drawn scenarios end-to-end against the harness's reference outputs to confirm machine scores.
| Channel | Queries | Intent acc. | Follow-up res. | Zero-result | Quality |
|---|---|---|---|---|---|
| Typed | 11,720 | 98.0% strong | 95.2% | 2.1% | 4.6 / 5 |
| Voice | 5,480 | 96.1% strong | 92.4% | 3.4% | 4.4 / 5 |
| Follow-up | 1,700 | 97.6% strong | 94.1% | 1.8% | 4.5 / 5 |
| Overall | 18,900 | 97.4% | 94.1% | 2.5% | 4.5 / 5 |
"The shopper isn't searching a database. They're having a conversation. The Brain's job is to make sure both sides remember what was just said."
The AI Brain reframes search as a conversation rather than a lookup. By separating understanding from retrieval, the system can interpret fragments, follow-ups, and colloquialisms with measured accuracy above 97% — and route each resolved intent to the subsystem best able to answer it. The result is a search bar that feels less like a query box and more like a knowledgeable salesperson on the floor. Future work extends the Brain to multilingual parsing, longer memory horizons, and a learned router trained on the labeled scenarios produced for this paper.
© 2026 InHouse America Research. AI Brain v5.11.26. For inquiries: legal@inhouseamerica.com.