№ | Слайд | Текст |
1 |
 |
Joint Relevance and Freshness Learning From Clickthroughs for NewsSearch Hongning Wang+, Anlei Dong*, Lihong Li*, Yi Chang*, Evgeniy Gabrilovich* +CS@UIUC *Yahoo! Labs |
2 |
 |
Relevance v.s. FreshnessRelevance Topical relatedness Metric: tf*idf, BM25, Language Model Freshness Temporal closeness Metric: age, elapsed time Trade-off Serve for user’s information need |
3 |
 |
Freshness is Important for News Search“Apple Company” @ Oct. 4, 2011 |
4 |
 |
Freshness is Important for News Search“Apple Company” @ Oct. 5, 2011 |
5 |
 |
Understand User’s Information NeedUser’s emphasis on relevance/freshness varies Breaking news queries Prefer latest news reports – freshness driven E.g., “apple company” Newsworthy queries Prefer high coverage and authority news reports – relevance driven E.g., “bin laden death” |
6 |
 |
Understand User’s Information NeedUser’s emphasis on relevance/freshness varies |
7 |
 |
Assess User’s Information NeedUnsupervised integration [Efron 2011, Li 2003] Limited on timestamps Editor’s judgment [Dong 2010, Dai 2011] Expensive for timely annotation Inadequate to recover end-user’s information need |
8 |
 |
Manipulate Editor’s AnnotationFreshness-demoted relevance Rule-based hard demotion [Dong 2010] E.g., if the result is somewhat outdated, it should be demoted by one grade (e.g., from excellent to good) |
9 |
 |
User’s Judgment on Relevance and FreshnessUser’s browsing behavior |
10 |
 |
Joint Relevance and Freshness LearningJRFL: (Relevance, Freshness) -> Click Query => trade-off URL => relevance/freshness |
11 |
 |
Joint Relevance and Freshness LearningModel formalization Latent |
12 |
 |
Joint Relevance and Freshness LearningLinear instantiation Associative property Relevance/Freshness model learning Query model learning |
13 |
 |
Joint Relevance and Freshness LearningCoordinate descent for JRFL Randomly initialize , and set Repeat until converge Update Relevance/Freshness models: Update Query model: Return the final model Convex programming |
14 |
 |
Temporal FeaturesURL freshness features Identify freshness from content analysis |
15 |
 |
Temporal FeaturesQuery freshness features Capture latent preference |
16 |
 |
Experiment ResultsData sets Two months’ Yahoo! News Search sessions Normal bucket: top 10 positions Random bucket [Li 2011] Randomly shuffled top 4 positions Unbiased evaluation corpus Editor’s judgment: 1 day’s query log Preference pair selection [Joachims 2005] Click > Skip above Click > Skip next Ordered by Pearson’s value |
17 |
 |
Experiment ResultsData sets Statistics |
18 |
 |
Analysis of JRFLConvergence Train/Test sets: 90k/60k preference pairs Varying initial query weight (a) Object Function Value Update |
19 |
 |
Analysis of JRFLConvergence Train/Test sets: 90k/60k preference pairs Varying initial query weight (b) Pairwise Error Rate Update |
20 |
 |
Analysis of JRFLConvergence Train/Test sets: 90k/60k preference pairs Varying initial query weight (c) Query Weight Update |
21 |
 |
Analysis of JRFLFeature weight learning |
22 |
 |
Analysis of JRFLRelevance and Freshness Learning Baseline: GBRank trained on Dong et al.’s relevance/freshness annotation set Testing corpus: editor’s one day annotation set |
23 |
 |
Analysis of JRFLQuery weight analysis |
24 |
 |
Analysis of JRFLQuery weight analysis Query length differs in relevance/freshness driven queries significantly |
25 |
 |
Quantitative ComparisonRanking performance Random bucket clicks |
26 |
 |
Quantitative ComparisonRanking performance Normal clicks |
27 |
 |
Quantitative ComparisonRanking performance Editorial annotations |
28 |
 |
Qualitative ComparisonCTR distribution revisit |
29 |
 |
ConclusionsJoint Relevance and Freshness Learning Query-specific preference Learning from query logs Temporal features Future work Personalized retrieval Broad spectral of user’s information need E.g., trustworthiness, opinion |
30 |
 |
References[Efron 2011] M. Efron and G. Golovchinsky. Estimation methods for ranking recent information. In SIGIR, pages 495–504, 2011. [Li 2003] X. Li and W. Croft. Time-based language models. In CIKM, pages 469–475, 2003. [Dong 2010] A. Dong, Y. Chang, Z. Zheng, G. Mishne, J. Bai, R. Zhang, K. Buchner, C. Liao, and F. Diaz. Towards recency ranking in web search. In WSDM, pages 11–20, 2010. [Dai 2011] N. Dai, M. Shokouhi, and B. D. Davison. Learning to rank for freshness and relevance. In SIGIR, pages 95–104, 2011. [Li 2011] L. Li, W. Chu, J. Langford, and X. Wang. Unbiased offline evaluation of contextual-bandit-based news article recommendation algorithms. In Proceedings of ACM WSDM '11, pages 297–306, 2011. [Joachims 2005] T. Joachims, L. Granka, B. Pan, H. Hembrooke, and G. Gay. Accurately interpreting clickthrough data as implicit feedback. In SIGIR, pages 154–161, 2005. |
31 |
 |
Thank youQ&A |
«Joint Relevance and Freshness Learning From Clickthroughs for News Search» |
http://900igr.net/prezentacija/anglijskij-jazyk/joint-relevance-and-freshness-learning-from-clickthroughs-for-news-search-137925.html