A complete technical analysis, behavioral taxonomy, and seller optimization playbook — built on the A9 patent that documents the dual-pipeline system classifying shopper intent and generating the next recommended query before anyone types a letter.
For five years, most sellers have optimized Amazon listings one way. Keywords in titles, bullets, and backend fields. A9 rewarded relevance through lexical matching, and that model worked. Past tense.
Amazon's search engine appears to operate on two layers: keyword retrieval and intent prediction. Patent 12,561,383 documents that second layer — a system that classifies shopper behavior and predicts what they will search next, built on top of keyword matching infrastructure that still drives product discovery. Keywords remain foundational. But a second system now operates on top of keywords, and most sellers have never encountered documentation of how that system works.
Filed by A9 and assigned to Amazon Technologies, patent 12,561,383 documents a dual-model machine learning system that classifies customer intent from search query sequences and generates predicted next-queries aligned with that intent. A system here doesn't wait for a customer to type. Prediction happens first. Results get precomputed and cached for near-instant delivery.
What does the patent actually cover? Not chatbots. Not conversational AI. It covers a search bar — autocomplete suggestions, "related searches," "customers also searched for," and recommended query refinements appearing when millions of shoppers type into Amazon's search box every day. Foundational infrastructure powering predictive search bar recommendations across Amazon's entire marketplace.
Architecture here operates on six exhaustive intent categories: Specification, Generalization, Equivalence, Substitution, Complement, and Irrelevant. Every query reformulation a customer makes gets classified into one of those six types. I call that taxonomy an Intent Hexagon, where every query transition maps to a vertex and every product competes within one or more vertices.
Amazon Search now operates on intent prediction in addition to keyword matching. Sellers who optimize only for keywords are missing a critical second layer — the intent classification system that determines which queries the search bar recommends next.
The patent reveals a fundamental architectural decision: Amazon runs two separate machine learning models in sequence. One classifies why a customer changed their search query. The other generates the next query the customer is likely to type — constrained by that classified intent. These are not the same model.
BERT converts query text into dense vector representations. Self-attention lets every token attend to every other token simultaneously, capturing full bidirectional context in one pass. "Running shoes" and "jogging sneakers" produce similar embeddings. Keywords still matter — but the system now understands meaning, not just tokens.
Each pair of consecutive queries gets classified into one of six intent categories via multi-class cross-entropy loss. A softmax probability distribution across the six categories picks the dominant intent for the transition. "Laptops" → "gaming laptops" = Specification. "iPhone 15" → "smartphones" = Generalization.
Masked self-attention runs over historical intent sequences with six parallel attention heads — one per intent type. Six simultaneous interpretations of every customer's search history are maintained until generation. Products visible across multiple intent heads gain broader recommendation exposure.
A Variational Autoencoder captures the statistical distribution of query transitions and samples from it. Three loss functions (reconstruction, KL divergence, uniformity) tune the system to produce 3–5 diverse, relevant, stable query suggestions — including queries the customer would not have generated independently.
Intent classification operates on top of retrieval. Keywords drive retrieval. Intent classification drives which queries the search bar suggests next. Sellers optimizing for keywords are doing necessary work — on one layer.
Every time a customer modifies a search query, the system classifies that modification into one of six exhaustive, mutually exclusive intent categories. Not a spectrum. A taxonomy. Classification operates on the difference between consecutive queries, never on individual queries in isolation. The same keyword carries entirely different intent depending on what preceded it.
Shoppers add keywords, introduce brand names, or specify colors and materials. They have converged. Purchase intent runs highest during Specification.
Keyword removal signals dissatisfaction. Shoppers who broaden search terms are voting against current results — expanding scope because what they found feels too narrow or irrelevant.
Highest conversion intent on Amazon. Exact model numbers, brand-specific product names, SKU-level terms. No discovery happening. A retrieval mission.
Replacing one brand name or characteristic with another. When shoppers enter Substitution mode, the search bar proactively recommends additional alternatives not yet considered.
Phone cases after phones. Chargers after laptops. Keywords related to, but functionally different from, previous results. Products in the Complement window receive highest algorithmic lift immediately after a purchase event on a related primary product.
A nuclear option. Keywords completely unrelated to prior queries trigger Irrelevant — and the patent language runs precise: the system "may exclude" prior queries from the prediction pipeline entirely. One Irrelevant classification resets the customer's intent context.
Classifications here don't run binary. The system computes a softmax probability vector across all six intent types for every reformulation. A transition might register 60% Specification, 25% Complement, and 15% Generalization simultaneously. Products can match multiple intents at once, and optimizing for a spread of plausible intents hedges visibility across a probability distribution.
Context functions as the variable. Keywords stay constant. "Nike shoes" after "Adidas shoes" registers as Substitution. "Nike shoes" after "shoes" registers as Specification. Identical keyword. Completely different intent.
Ten walkthroughs of real shopping journeys. Each traces a multi-query session, showing which intent the system assigns at every transition — and what that classification means for sellers competing at each stage.
Five named dynamics that govern how the system translates shopper behavior into predicted intent and generated queries. Each is documented or strongly implied by patent text.
When a customer purchases or adds to cart, the system significantly increases the predicted probability of Complement intent for the next search. The customer who just bought a phone won't search for more phones — the system begins generating phone-case and accessory queries before the customer types a single character. If you sell accessories, your maximum algorithmic visibility window opens the instant a customer converts on a related primary product. The window is architecturally engineered.
Patent 12,561,383 illustrates a pattern with an example of four Specification queries. After approximately four consecutive Specifications without conversion, probability increases that the customer will switch to Substitution — actively seeking alternatives to everything narrowed toward. Recommended queries begin surfacing competitors. Brands dominating a Specification sequence but failing to convert inside that window face increasing probability that the system predicts Substitution intent next.
When the system classifies a reformulation as Irrelevant, all queries before that classification may get excluded from subsequent intent counts. Prior session context may no longer inform predictions. Sellers who lost a customer to an Irrelevant shift cannot rely on session continuity for recovery because the prediction pipeline may operate without benefit of prior behavioral signals.
Patent text explicitly enumerates three tiers of interaction: navigation to a product page, addition to cart or list, and initiation of purchase. No collapsing into a single "engagement" metric. Each tier carries different weight in the intent prediction model. A page view nudges. A cart-add shifts. A purchase inverts.
An identical search query produces different intent predictions for different customers. Patent text enumerates five metadata layers enriching every query before encoding: human language indication, geographic characteristics, demographic data, hardware and software components of the device, and network identifiers indicative of geolocation. No single "search result page" exists. Millions exist, personalized at the encoding layer.
Six technical details enumerated in the patent that shape how the system behaves in production.
Voice queries convert via speech-to-text. Image searches use recognition to extract text. All modalities reduce to text embeddings before intent analysis. Voice-search optimization is keyword optimization under the hood.
Search results for high-probability predicted queries get precomputed and cached. Products in precomputed result sets benefit from reduced latency — a structural advantage distinct from listing quality.
Recommended queries may contain terms a customer was unaware of and unlikely to type manually. Amazon actively expands customer vocabulary. Distinctive features, novel materials, and unique attributes can surface even when no one explicitly searches for them.
Intent is inferred by comparing consecutive queries — what keywords were added, removed, or substituted. The system measures the manner of change as its primary input, not the content of any individual query.
Predicted queries appear in two contexts: inside the search interface (autocomplete) and alongside results pages. Two visibility channels, same prediction model, different placement windows in the shopper's journey.
The sequence of queries is bounded by a session — a temporal grouping of activity. Cross-session context is not guaranteed. A customer returning hours later may start with a clean slate. Session boundaries are prediction boundaries.
The Intent-Optimized Listing Framework. Four listing surfaces, each addressing one or more intent types. Every edit traces back to a classification the system performs internally.
Leads with core category term (Generalization capture). Follows with primary modifier (Specification capture). Brand name placed where Equivalence traffic rewards recognition. The first 80 characters do most of the ranking work.
Covers: Generalization ("shoes"), Specification ("women's tennis shoes"), Irrelevant filter ("professional," "court-optimized").
Bullet 1 filters Irrelevant traffic and confirms fit inside three seconds. Bullets 2–3 carry Specification depth — features backed by specific evidence. Bullet 4 runs Substitution defense — explicit reasons to choose over alternatives. Bullet 5 opens Complement hooks — "pairs well with" ecosystem references.
Recommended starting mix: ~60% Specification (long-tail modifiers), ~15% Generalization (core category), ~15% Equivalence / Substitution (alternatives and competitor terms used descriptively), ~10% Complement (related-product language). Tune from Search Query Performance data.
(1) Quick relevance confirmation. (2) Specification details and features. (3) Equivalence / Substitution positioning vs. alternatives — comparison modules. (4) Complement ecosystem showing what pairs. Every module does one job the hexagon names.
A9's classifier lives deep inside Amazon infrastructure. Sellers never see the labels — only downstream effects. Intent Mapping of the Intent Hexagon converts A9's machine-learning taxonomy into a seller-facing lens: six transition types become six coverage dimensions a seller can score, diagnose, and rewrite toward.
Pull a Helium 10 Cerebro or Magnet export. Filter by search-volume threshold. Separate terms into five input categories: core category, phrase variants, modifiers, use cases, and adjacent needs.
A Classifier subagent loads reference material on the six types plus worked examples. Every keyword receives a dominant intent label. Ambiguous keywords also receive a secondary label. Output: a classification table ready for journey modeling.
A Journey Modeler produces 3–5 realistic flows specific to the product. Each flow names a starting query, a sequence of transitions, and a likely ending query. Flows describe how a shopper probably moves through a session.
A Vulnerability Diagnostician walks the current listing through every modeled journey. Output lists where the listing loses relevance, why, and which page element causes the loss.
Every diagnosed vulnerability is mapped to a specific listing surface — title, bullets, hero image, secondary images, A+ content, backend terms, brand store. No fix floats without an anchor.
A rewrite plan executes edits traceable to identified vulnerabilities. Intent Mapping produces strategy. Execution happens downstream inside whatever copy pipeline the seller runs. Scope boundary stays clean.
A Scorer applies a rubric across six coverage dimensions. Scale runs 1–10 per dimension, 60 total. Grading bands translate the score into Poor, Developing, Strong, or Optimized.
Real shopper journeys repeat in predictable patterns. Six archetypes drawn from my analysis of Helium 10 keyword graphs alongside A9's reformulation patterns.
Broad to narrow to known-item. "laptop" → "gaming laptop" → "gaming laptop RTX 4090" → "Alienware m18 R2." A listing prepared for retrieval but not for narrowing loses visibility at step two.
Specification narrows too far, reverses course, either back to Generalization or sideways to Substitution. Failed refinement kills conversion if the listing cannot reappear at the broader query.
Shopper searches for a known item, commits mentally, begins building around the purchase. "AirPods Pro 2" → "AirPods Pro 2 case" → "Apple USB-C adapter." Brand store architecture drives ecosystem capture.
Narrow, reject candidate, compare alternatives, then add a complementary item after selection. A+ content earns decisive weight during the comparison phase.
First narrowing path stalls. Shopper backs up and narrows along a different axis. Broad title coverage plus multiple feature axes in bullets let a listing ride recovery paths without disappearing.
Session ends or pivots to an unrelated category. Strong bullet 1 and a clean hero image filter wrong shoppers early. Abandonment drops conversion metrics when a listing fails to filter.
Patent 12,561,383 classifies transitions. Intent Mapping adds a diagnostic layer on top: what happens when a listing cannot survive a given transition. Six failure modes, one per intent type.
Listing retrieves on broad terms but vanishes as shoppers add modifiers. Title carries core category and no meaningful attributes. Bullets lack progressive feature depth. Images fail to prove attribute claims.
Listing ranks only on long-tail modifiers. Strip a qualifier and the listing disappears. Hero image fails to confirm broad category fit. First half of title over-specializes before establishing category.
Product sold under a name shoppers do not actually type. Model numbers missing from title. Variant language inconsistent across bullets, images, and A+. Exact-name traffic routes elsewhere.
Comparison shoppers arrive, scan for differentiation, find none, select a competitor. A+ content lacks comparison modules. Bullet 4 repeats features instead of arguing choice rationale.
Adjacent search surfaces exist but the listing does not participate. No ecosystem visibility. Brand store fails to route Complement traffic. A+ comparison modules do not link related SKUs.
Listing attracts wrong shoppers. Title ambiguity produces impressions outside category fit. CTR and conversion erode. Organic rank degrades downstream.
Theory without a worked example leaves a seller reading instead of operating. Here is Intent Mapping run end-to-end against a real product category — a mid-tier 32 oz insulated water bottle.
The setup. Mid-tier brand. Stainless steel, double-wall vacuum, leak-proof lid, eight colors. Seller pulls a Cerebro export off a leading competitor ASIN. 1,847 keywords with volume data. After filtering under 300 monthly searches, 412 working terms remain, separated into the five input categories: 38 core, 174 phrase variants, 86 modifiers, 77 use cases, 37 adjacent needs.
Classification. "Insulated water bottle" → Specification (from a Generalization starting point). "Stainless steel water bottle" → Specification. "Hydro Flask 32 oz" → Equivalence. "Yeti Rambler" (after Hydro Flask) → Substitution. "Water bottle brush" → Complement. "Gym water bottle for women" → Specification with Use Case tag.
Journey modeling. Four realistic flows emerge — a Refinement Path, a Comparison Path across Hydro Flask / Yeti / Stanley, a Known-Item-to-Ecosystem path, and a Recovery Path where "insulated bottle leak proof for kids" returns thin results and the shopper broadens.
Diagnosis. The current listing holds Specification depth but over-commits to brand before establishing category — blocking Generalization recapture mid-session. No comparison modules in A+ content, so Substitution shoppers have no anchor. No ecosystem cross-linking in brand store. Bullet 1 buries kid-safe language behind feature lists.
Rewrite map. Title leads with category term before brand. A+ content adds a comparison module built around Hydro Flask / Yeti / Stanley attributes. Brand store adds an "everything for your bottle" landing page with lids, brushes, sleeves. Bullet 1 moves kid-safe language forward.
Before rewrite vs. post-rewrite across the six coverage dimensions. A separate example — a dog harness listing, mid-tier brand, moderate rank. Audit reveals the following findings.
Rewrite targets. Bring Retrieval Precision and Comparison Defense into the 7 range. Add Complement ecosystem through a brand-store rebuild. Move the listing from 30 (Developing) to 45 (Strong) over two rewrite cycles. Water-bottle case moved 25 → 47; post-rewrite conversion climbed 19% inside 60 days across a five-SKU portfolio.
The rubric does not reward perfection. A commodity product cannot earn a 10 on Comparison Defense because the product has limited differentiation. Optimized band sits at 55+ because most listings plateau before 55 without either deep product innovation or sustained content investment.
Classification looks easy in a reference guide and harder under live keywords. Five confusion patterns show up across most audits.
"Gaming laptop RTX 4090 16 inch" feels specific enough to read as Equivalence. Test fails. Equivalence requires a named known item, not a narrow description. A shopper typing "Alienware m18 R2" has committed. Describing narrows a candidate set. Naming retrieves one. Specification gets served feature-led content. Equivalence gets served exact-name retrieval content. Misclassifying collapses the two rails.
When a shopper types a different brand, sellers sometimes log that as narrowing. "Nike shoes" → "Adidas shoes" adds no new attribute. Original candidate gets replaced. Same need, new candidate. Substitution. Treating the transition as Specification misses a comparison moment and underbuilds A+ content that should carry competitive positioning.
A harness becomes a leash. A camera becomes a tripod. Related products feel like substitutions because of surface similarity. Test: does the primary product stay wanted? Complement keeps the primary wanted and adds a companion. Substitution replaces the primary entirely. Misclassifying collapses cross-sell opportunity.
"Gaming laptop RTX 4090" → "laptop" looks like a wide swing. Underlying need stays alive — shopper still wants a laptop. Generalization, not Irrelevant. "Gaming laptop RTX 4090" → "hiking boots" breaks need entirely. That is Irrelevant. Distinction drives whether a listing stays in consideration or gets dropped.
Sellers sometimes discard Irrelevant keywords as junk. The patent treats Irrelevant as a session reset. Prior context dies. The seller needs a filter strategy. Bullet 1 and hero image carry decisive weight because a cold-start shopper landing on a listing has no accumulated intent context. The listing must earn relevance from zero inside three seconds.
Most Amazon sellers optimize for only one layer of a two-layer system. They write keyword-dense titles, manage backend search terms, bid on exact-match phrases. That work remains necessary because keywords still drive retrieval. But patent 12,561,383 documents a second layer operating on top of keyword matching.
Patent 12,561,383 documents a system that classifies intent on top of keyword retrieval. Each query reformulation gets classified into one of six Intent Hexagon types. Sessions accumulate behavioral signal stacks. Purchase events significantly increase the probability of Complement predictions. Sustained Specification sequences without conversion increase the probability of Substitution predictions. An algorithm doesn't guess. Classification runs with trained precision, and training has drawn on Amazon's entire behavioral corpus.
Sellers who understand the architecture hold a structural advantage. They can engineer visibility at the precise moment a customer's intent shifts toward their product type. They can position complementary products in an inversion window. They can build semantic richness that survives BERT encoding. They can optimize across a probability distribution rather than for a single keyword.
The optimization target has shifted. A system classifies intent, predicts transitions, generates queries, and precomputes results — all before a customer types a next search. Sellers who understand the named dynamics documented here will capture customers the system already steers toward them.
Keywords help a product get retrieved. Intent Mapping helps a product survive whatever move a shopper makes next. Sellers who understand their position within the Intent Hexagon can engineer visibility at the exact moment a customer's intent shifts toward their product type.
The writer consumes the packet and executes title, bullets, A+ copy, backend terms. Intent labels propagate through copy so the title surfaces Generalization and Specification terms, bullet 4 carries Substitution defense language, and bullet 5 surfaces Complement hooks pulled from the adjacency tier. Seller gets strategy, then copy, then a scorecard that tracks movement across rewrite cycles.
The full 17-page grant. Flip through it below — page by page, spread by spread. Filed by A9.com LLC, assigned to Amazon Technologies, issued under the B1 designation (granted without prior publication, so the printed cover date is the grant date).


The full Intent Mapping method — Truth Card builder, 6-type classifier, journey modeler, vulnerability diagnostician, scorer, and synthesizer — packaged as an installable Claude skill. Drop the zip into Claude Code or Cowork and it triggers automatically whenever you paste keywords or ask for a listing audit.
intent-mapping.zipintent-mapping/ folder~/.claude/skills/, Cowork: upload via Skills settings)The same seven-step Intent Mapping method packaged as a custom GPT. Paste your keyword export and product description — the GPT runs the full classify → journey → diagnose → score → rewrite loop inside ChatGPT. No install, no local setup.
Intent Mapping GPT in ChatGPT