Title of article :
Learning to rank using smoothing methods for language modeling
Author/Authors :
Yuan Lin، نويسنده , , Hongfei Lin، نويسنده , , Kan Xu، نويسنده , , Xiaoling Sun، نويسنده ,
Issue Information :
ماهنامه با شماره پیاپی سال 2013
Pages :
11
From page :
818
To page :
828
Abstract :
The central issue in language model estimation is smoothing, which is a technique for avoiding zero probability estimation problem and overcoming data sparsity. There are three representative smoothing methods: Jelinek-Mercer (JM) method; Bayesian smoothing using Dirichlet priors (Dir) method; and absolute discounting (Dis) method, whose parameters are usually estimated empirically. Previous research in information retrieval (IR) on smoothing parameter estimation tends to select a single value from optional values for the collection, but it may not be appropriate for all the queries. The effectiveness of all the optional values should be considered to improve the ranking performance. Recently, learning to rank has become an effective approach to optimize the ranking accuracy by merging the existing retrieval methods. In this article, the smoothing methods for language modeling in information retrieval (LMIR) with different parameters are treated as different retrieval methods, then a learning to rank approach to learn a ranking model based on the features extracted by smoothing methods is presented. In the process of learning, the effectiveness of all the optional smoothing parameters is taken into account for all queries. The experimental results on the Learning to Rank for Information Retrieval (LETOR) LETOR3.0 and LETOR4.0 data sets show that our approach is effective in improving the performance of LMIR.
Keywords :
Information retrieval , Searching , Machine learning
Journal title :
Journal of the American Society for Information Science and Technology
Serial Year :
2013
Journal title :
Journal of the American Society for Information Science and Technology
Record number :
994845
Link To Document :
بازگشت