Personalization in Local Search Personalization of Content Ranking

  • Slides: 47
Download presentation
Personalization in Local Search Personalization of Content Ranking in the Context of Local Search

Personalization in Local Search Personalization of Content Ranking in the Context of Local Search Philip O’Brien, Xiao Luo, Tony Abou-Assaleh, Weizheng Gao, Shujie Li Research Department, Genie. Knows. com September 17, 2009

About Genie. Knows. com Based in Halifax, Nova Scotia, Canada Established in 1999 ~35

About Genie. Knows. com Based in Halifax, Nova Scotia, Canada Established in 1999 ~35 People Online Advertising Network - 100 to 150 million searches per day Search Engines (local, health, games) Content Portals 2009 -09 -17 2

2009 -09 -17 3

2009 -09 -17 3

About Tony Abou-Assaleh Director of Research at Genie. Knows - Since 2006 - Build

About Tony Abou-Assaleh Director of Research at Genie. Knows - Since 2006 - Build search engines - Other internal R&D initiatives Lecturer at Brock University, St. Catharines, Canada - 2005 – 2006 GNU grep official maintainer 2009 -09 -17 4

Agenda Introduction Related Work Our Approach Experiments Conclusion & Future Work 2009 -09 -17

Agenda Introduction Related Work Our Approach Experiments Conclusion & Future Work 2009 -09 -17 5

Agenda Introduction Related Work Our Approach Experiments Conclusion & Future Work 2009 -09 -17

Agenda Introduction Related Work Our Approach Experiments Conclusion & Future Work 2009 -09 -17 6

Introduction Local Search - What? Why? Personalization - What? How? Why? Assumptions Objectives 2009

Introduction Local Search - What? Why? Personalization - What? How? Why? Assumptions Objectives 2009 -09 -17 7

What is Local Search? Local Search vs. Business Directory Contains: - Internet Yellow Pages

What is Local Search? Local Search vs. Business Directory Contains: - Internet Yellow Pages (IYP) Business Directory Enhanced business listings Map Ratings and Reviews Articles and editorials Pictures and rich media Social Networking 2009 -09 -17 8

2009 -09 -17 9

2009 -09 -17 9

Why Local Search? Good for end users Good for businesses Good for our company

Why Local Search? Good for end users Good for businesses Good for our company Interesting research problems No market leader Could be the next big thing 2009 -09 -17 10

What is Personalization? No personalization: - Everybody gets the same results Personalization: - User

What is Personalization? No personalization: - Everybody gets the same results Personalization: - User may see different results Personalization vs. customization 2009 -09 -17 11

What to Personalize? Ranking Snippets Presentation Collection Recommendations 2009 -09 -17 12

What to Personalize? Ranking Snippets Presentation Collection Recommendations 2009 -09 -17 12

How to Personalize? Search history Click history User profiles – interests Collaborative filtering 2009

How to Personalize? Search history Click history User profiles – interests Collaborative filtering 2009 -09 -17 13

Why Personalization? One size does not fit all Ambiguity of short queries Improve per-user

Why Personalization? One size does not fit all Ambiguity of short queries Improve per-user precision Improve user experience Targeted advertising $$$ 2009 -09 -17 14

Assumptions Interests are location dependent Long-term interests Implicit relevance feedback Relevance in location dependent

Assumptions Interests are location dependent Long-term interests Implicit relevance feedback Relevance in location dependent Relevance is category dependent User cooperation Single-user personalization 2009 -09 -17 15

Objectives General framework for personalization of spatialkeyword queries User profile representation Personalized ranking Improve

Objectives General framework for personalization of spatialkeyword queries User profile representation Personalized ranking Improve over baseline system 2009 -09 -17 16

Agenda Introduction Related Work Our Approach Experiments Conclusion & Future Work 2009 -09 -17

Agenda Introduction Related Work Our Approach Experiments Conclusion & Future Work 2009 -09 -17 17

Related Work User Profile Modeling Personalized Ranking 2009 -09 -17 18

Related Work User Profile Modeling Personalized Ranking 2009 -09 -17 18

User Profile Modeling Topic based (Liu et al, 2002) - Vector of interests -

User Profile Modeling Topic based (Liu et al, 2002) - Vector of interests - Explicit: how to collect data? - Implicit: relevance feedback Click based (Li et al, 2008) - Implicit feedback from click through data - Require a lot of data Ontological profiles (Sieg et al, 2007) Hierarchical representations (Huete et al, 2008) 2009 -09 -17 19

Personalized Ranking Web, desktop, and enterprise search Local search? Strategies: - Implicit Clicks as

Personalized Ranking Web, desktop, and enterprise search Local search? Strategies: - Implicit Clicks as relevance feedback Query topic identification Collaborative filtering Learning algorithms 2009 -09 -17 20

Agenda Introduction Related Work Our Approach Experiments Conclusion & Future Work 2009 -09 -17

Agenda Introduction Related Work Our Approach Experiments Conclusion & Future Work 2009 -09 -17 21

Our Approach Problem formulation Ranking Function Decomposition Business Features User Profile User Interest Function

Our Approach Problem formulation Ranking Function Decomposition Business Features User Profile User Interest Function Business-specific Preference Function 2009 -09 -17 22

Problem Formulation Query: keywords + spatial (geographic) context Ranking function: Relevant Results ✕ User

Problem Formulation Query: keywords + spatial (geographic) context Ranking function: Relevant Results ✕ User Profiles ✕ Location Real Number Online personalized ranking: - Optimization of an objective function over rank scores 2009 -09 -17 23

Ranking Function Decomposition Final rank = weighted combination of: - Baseline rank - User

Ranking Function Decomposition Final rank = weighted combination of: - Baseline rank - User rank - Business rank 2009 -09 -17 24

Ranking Function Decomposition Final rank = weighted combination of: - Baseline rank - User

Ranking Function Decomposition Final rank = weighted combination of: - Baseline rank - User rank - Business rank 2009 -09 -17 25

Baseline Rank Okapi BM 25 F on textual fields Distance from query centre Other

Baseline Rank Okapi BM 25 F on textual fields Distance from query centre Other non-textual features 2009 -09 -17 26

Business Features List of categories - 18 top level, 275 second level Terms -

Business Features List of categories - 18 top level, 275 second level Terms - Vector-space model Location - Geocoded address Meta data - Year established, number of employees, languages, etc. 2009 -09 -17 27

User Profile Local Profile - For each geographic region (city) - For each category

User Profile Local Profile - For each geographic region (city) - For each category - Needs at least 1 query Global Profile - Aggregation of local profiles - Used for new city and category combination 2009 -09 -17 28

Local Profile Category interest score - Fraction of queries in this category - Fraction

Local Profile Category interest score - Fraction of queries in this category - Fraction of clicks in this category Number of queries Terms vector-space model Clicks (business, timestamp) 2009 -09 -17 29

Global Profile Estimated global category interest score - Aggregated over all cities Weighted combination

Global Profile Estimated global category interest score - Aggregated over all cities Weighted combination of interest scores Weights derived from query volume Estimated using a Dirichlet Distribution 2009 -09 -17 30

Ranking Function Decomposition Final rank = weighted combination of: - Baseline rank - User

Ranking Function Decomposition Final rank = weighted combination of: - Baseline rank - User rank - Business rank 2009 -09 -17 31

User Interest Function Rank (business, user, query) = Category interest score ✕ Term similarity

User Interest Function Rank (business, user, query) = Category interest score ✕ Term similarity ✕ Click count Averaged over all categories of the business Term similarity: cosine similarity Click count: capture navigational queries 2009 -09 -17 32

Ranking Function Decomposition Final rank = weighted combination of: - Baseline rank - User

Ranking Function Decomposition Final rank = weighted combination of: - Baseline rank - User rank - Business rank 2009 -09 -17 33

Business-specific Preference Function Rank (business, user, city, category) = Sum of query dependent click

Business-specific Preference Function Rank (business, user, city, category) = Sum of query dependent click scores + Sum of query independent click scores Click scores are time discounted - 1 year windows - 1 week intervals Parameter to control relative importance of querydependency 2009 -09 -17 34

Agenda Introduction Related Work Our Approach Experiments Conclusion & Future Work 2009 -09 -17

Agenda Introduction Related Work Our Approach Experiments Conclusion & Future Work 2009 -09 -17 35

Experiments Data Procedure Results Discussion 2009 -09 -17 36

Experiments Data Procedure Results Discussion 2009 -09 -17 36

Data 22 Million businesses 30 participants Only 12 with sufficient queries 2388 queries 1653

Data 22 Million businesses 30 participants Only 12 with sufficient queries 2388 queries 1653 unique queries 2009 -09 -17 37

Procedure Types of tasks: - Navigational, browsing, information seeking 5 -point explicit relevance feedback

Procedure Types of tasks: - Navigational, browsing, information seeking 5 -point explicit relevance feedback Ranking algorithm - Baseline vs. personalized Alternates every 2 minutes Identical interface No bootstrapping phase 2009 -09 -17 38

Results Measures: - Mean Average Precision – MAP - Mean Reciprocal Rank – MRR

Results Measures: - Mean Average Precision – MAP - Mean Reciprocal Rank – MRR - Normalized Discounted Cumulative Gain – n. DCG 2009 -09 -17 39

Results 2009 -09 -17 40

Results 2009 -09 -17 40

Results Welch two-sample t-test: - Significant improvement - MAP: 95% confidence, p=0. 04113 -

Results Welch two-sample t-test: - Significant improvement - MAP: 95% confidence, p=0. 04113 - MRR: 95% confidence, p=0. 02192 2009 -09 -17 41

Results n. DCG@10 16 randomly selected queries Not significant 2009 -09 -17 42

Results n. DCG@10 16 randomly selected queries Not significant 2009 -09 -17 42

Agenda Introduction Related Work Our Approach Experiments Conclusion & Future Work 2009 -09 -17

Agenda Introduction Related Work Our Approach Experiments Conclusion & Future Work 2009 -09 -17 43

Contributions Personalization framework for spatial-keyword queries Model for user profiles Local and global profiles

Contributions Personalization framework for spatial-keyword queries Model for user profiles Local and global profiles Address data sparseness problem Personalized ranking function - Interests, clicks, terms Empirical evaluation - Significant improvement over the baseline system 2009 -09 -17 44

Future Work Modeling of short-term interests Modeling of recurring interests “Learning to Rank” algorithms

Future Work Modeling of short-term interests Modeling of recurring interests “Learning to Rank” algorithms Multi-user personalization - Recommender system Incorporate on www. genieknows. com 2009 -09 -17 45

Thanks you! http: //www. genieknows. com http: //tony. abou-assaleh. net taa@genieknows. com @tony_aa 2009

Thanks you! http: //www. genieknows. com http: //tony. abou-assaleh. net taa@genieknows. com @tony_aa 2009 -09 -17 46

Questions Can I access your data? Did you do parameter tuning? Did users try

Questions Can I access your data? Did you do parameter tuning? Did users try to test/cheat the system? What is the computational complexity? Any confounding variables? 2009 -09 -17 47