Executive AI Dinner hosted by Kumo - Austin, April 8

Register here
3Ranked Recommendation · Search

Search Ranking

For each user's search query, which products should rank highest?

Book a demo and get a free trial of the full platform: research agent, fine-tune capabilities, and forward-deployed engineer support.

By submitting, you accept the Terms and Privacy Policy.

Loved by data scientists, ML engineers & CXOs at

Catalina Logo

A real-world example

For each user's search query, which products should rank highest?

Default search engines rank by text relevance and popularity. Two users searching "running shoes" see the same results even though one is a trail runner and the other runs on pavement. Click-through rates on search results average 15-20% when they could be 35-50% with personalization. For an ecommerce site doing $1B in GMV, search drives 40% of revenue — a 20% improvement in search conversion is worth $80M annually.

Quick answer

Personalized search ranking re-orders product search results based on each user's purchase history, browsing behavior, and similarity to other users. Two users searching 'running shoes' see different results because one is a trail runner and the other runs on pavement. Graph-based re-ranking improves search conversion by 20-40% over default text-relevance ranking.

Approaches compared

4 ways to solve this problem

1. Text Relevance (BM25 / TF-IDF)

Rank search results by text match between the query and product titles/descriptions. The default for most ecommerce search engines (Elasticsearch, Solr).

Best for

Queries where text relevance is the primary signal and personalization is not needed (e.g., searching for a specific SKU or brand).

Watch out for

Everyone searching 'running shoes' sees the same results. A trail runner and a road runner have completely different intent, but text relevance cannot distinguish them. Click-through rates sit at 15-20% when they could be 35-50%.

2. Popularity-Boosted Search

Boost search results by global popularity metrics: best-selling, highest-rated, most-reviewed. A simple improvement over pure text relevance.

Best for

High-traffic searches where popular products genuinely satisfy most users. Reduces the risk of surfacing obscure low-quality results.

Watch out for

Still not personalized. Popular products are not always the right products for each user. A bestselling road running shoe is irrelevant to a trail runner, regardless of its popularity.

3. Learning to Rank (LambdaMART, RankNet)

Train a ranking model on features: text relevance, click-through rate, conversion rate, price, and user features. The current industry standard for personalized search.

Best for

Teams with ML infrastructure and large click-stream datasets. Captures non-linear feature interactions for ranking.

Watch out for

Requires extensive feature engineering: user-product affinity scores, category preference vectors, price sensitivity bins. Each new signal (return rate, browse depth, wishlist data) requires a new feature pipeline. The model treats each user-product pair independently, missing cross-user click patterns.

4. KumoRFM (Graph Neural Networks on Relational Data)

Builds a graph connecting users, searches, clicks, products, and purchases. Re-ranks results by learning from the full relational structure: purchase history, click patterns of similar users, location affinity, and return rates. No feature engineering required.

Best for

Ecommerce sites with rich user behavior data and diverse product catalogs where personalization significantly changes result relevance.

Watch out for

The graph advantage is largest for ambiguous queries ('running shoes,' 'dress') where user intent varies widely. For specific queries ('Nike Air Max 90 size 10'), text relevance alone is usually sufficient.

Key metric: Graph-based search re-ranking improves conversion by 20-40%. RelBench benchmark: 76.71 vs 62.44 for flat baselines, with the largest gains on ambiguous queries where user intent varies.

Why relational data changes the answer

User U001 (outdoor enthusiast, Denver) searches 'running shoes.' Default search returns the same bestselling road running shoes shown to everyone. But the relational graph reveals that U001 bought 3 trail products in 6 months, 67% of users with similar purchase history clicked on the TrailMax Pro GTX, and Denver users have a trail product affinity score of 0.82. The click position bias correction also matters: U001 clicked on P401 at position 3, indicating high intent (they scrolled past 2 other products to click it).

These personalization signals span multiple tables. Purchase history is in PURCHASES. Click patterns require the CLICKS table joined to SEARCHES. Location affinity comes from the USERS table cross-referenced with PRODUCTS. Cross-user click patterns require aggregating behavior across similar users. A learning-to-rank model would need an engineer to pre-compute all of these as features. The graph neural network discovers them automatically and learns how they interact. On the RelBench benchmark, graph-based models score 76.71 vs 62.44 for flat-table baselines. For search ranking specifically, the improvement translates to 20-40% higher search conversion rates because the model surfaces the right products for each user, not just the globally popular ones.

Default search ranking is like a concierge who gives every hotel guest the same restaurant recommendation. Personalized graph-based ranking is like a concierge who remembers that you ordered vegan at the hotel restaurant, your travel companion mentioned a gluten allergy, and guests with similar preferences loved the farm-to-table spot on Fifth Street. The same question ('good restaurant nearby?') gets a completely different answer based on relational context.

How KumoRFM solves this

Relational intelligence for true personalization

Kumo re-ranks search results by learning from the full relational graph of user behavior, product attributes, and cross-user click patterns. It discovers that trail runners who search "running shoes" click on different products than road runners — and uses purchase history, return patterns, and graph neighborhood signals to personalize rankings. The model captures that users who bought hydration packs and trail GPS devices should see trail shoes first.

From data to predictions

See the full pipeline in action

Connect your tables, write a PQL query, and get predictions with built-in explainability — all in minutes, not months.

1

Your data

The relational tables Kumo learns from

USERS

user_idsegmentlocation
U001outdoor_enthusiastDenver, CO
U002casual_fitnessMiami, FL
U003competitive_runnerBoston, MA

SEARCHES

search_iduser_idquerytimestamp
S001U001running shoes2025-02-20
S002U002running shoes2025-02-20
S003U003lightweight trainers2025-02-21

CLICKS

click_idsearch_idproduct_idpositiontimestamp
CL001S001P40132025-02-20
CL002S001P40872025-02-20
CL003S002P40212025-02-20

PRODUCTS

product_idnamecategoryprice
P401TrailMax Pro GTXTrail Running159.99
P402StreetRunner LiteRoad Running89.99
P408Summit Ridge TrainerTrail Running139.99
2

Write your PQL query

Describe what to predict in 2–3 lines — Kumo handles the rest

PQL
PREDICT LIST_DISTINCT(CLICKS.PRODUCT_ID, 0, 7, days)
RANK TOP 20
FOR EACH USERS.USER_ID
3

Prediction output

Every entity gets a score, updated continuously

USER_IDCLASSSCORETIMESTAMP
U001P4010.912025-03-12
U001P4080.872025-03-12
U002P4020.842025-03-12
4

Understand why

Every prediction includes feature attributions — no black boxes

User U001 (outdoor_enthusiast, Denver, CO)

Predicted: P401 (TrailMax Pro GTX) ranked #1 — score 0.91

Top contributing features

Past trail running purchases

3 trail products in 6 months

31% attribution

Graph neighbors clicked P401

67% of similar users clicked

27% attribution

Location affinity (mountain region)

Trail product affinity 0.82

20% attribution

Click position bias correction

Clicked at position 3 (high intent)

14% attribution

Return rate for user segment

0.04 (low returns)

8% attribution

Feature attributions are computed automatically for every prediction. No separate tooling required. Learn more about Kumo explainability

Frequently asked questions

Common questions about search ranking

How much does personalized search improve conversion?

Personalized search ranking improves search conversion rates by 20-40% compared to text-relevance-only ranking. For ecommerce sites with $1B+ GMV where search drives 40% of revenue, this translates to $50-80M in incremental annual revenue.

Does personalized search require user login?

Login improves personalization, but it is not required. Session-level behavior (clicks in the current session, products viewed, search refinements) provides enough signal for meaningful personalization. Graph models connect anonymous session behavior to patterns from logged-in users with similar interactions.

How do you handle position bias in search data?

Products shown at position 1 get clicked more often regardless of relevance (position bias). Graph models correct for this by learning the relationship between click position and true relevance. A click at position 7 is a stronger relevance signal than a click at position 1 because the user had to scroll past 6 alternatives.

Can personalized search work for small catalogs?

The benefit scales with catalog size and query ambiguity. A catalog with 50 products does not need much re-ranking. A catalog with 50,000 products where 'running shoes' returns 200 results benefits enormously from personalization. The bigger the result set and the more ambiguous the query, the higher the impact.

Bottom line: 20-40% improvement in search conversion rate. For ecommerce sites with $1B+ GMV, personalized search ranking drives $50-80M in incremental annual revenue.

Topics covered

search ranking AIpersonalized search resultssearch relevance optimizationecommerce search rankinggraph neural network searchKumoRFMpredictive query languagesearch conversion optimizationuser intent predictionproduct search personalizationlearning to rankrelational search signals

One Platform. One Model. Infinite Predictions.

KumoRFM

Relational Foundation Model

Turn structured relational data into predictions in seconds. KumoRFM delivers zero-shot predictions that rival months of traditional data science. No training, feature engineering, or infrastructure required. Just connect your data and start predicting.

For critical use cases, fine-tune KumoRFM on your data using the Kumo platform and Research Agent for 30%+ higher accuracy than traditional models.

Book a demo and get a free trial of the full platform: research agent, fine-tune capabilities, and forward-deployed engineer support.