Files in this item
Files | Description | Format |
---|---|---|
application/pdf ![]() ![]() | (no description provided) |
Description
Title: | Reinforced co-learning for semi-supervised ranking |
Author(s): | He, Shibi |
Advisor(s): | Peng, Jian |
Department / Program: | Computer Science |
Discipline: | Computer Science |
Degree Granting Institution: | University of Illinois at Urbana-Champaign |
Degree: | M.S. |
Genre: | Thesis |
Subject(s): | Learning to rank
Semi-supervised Learning Reinforcement Learning Machine Learning |
Abstract: | Learning to rank is vital to information retrieval and recommendation systems. Directly optimizing the listwise evaluation measure such as normalized discounted cumulative gain (NDCG) is an advanced way to learn a ranking model. However, this is only suited for training data with effective labels. In real applications, we are more often faced with the semi-supervised setting that only a partial set of data has labels. In this paper, we propose a co-learning strategy for the semi-supervised ranking problem. Our model has two modules: the classifier module and the reinforcement ranker module. Given a query, the classifier module is trained to classify whether a document is relevant or not. The reinforcement ranker module is trained to give relevance scores on the basis of treating ranking problems as Markov decision processes (MDP). We name our approach "reinforced co-learning" because the two modules are iteratively optimized and affect each other while training. When training the classifier module, we use the reinforcement module to give every candidate a relevance score and sample lower scored documents as irrelevant samples (negative samples). Likewise, in order to train the reinforcement ranker module, we use the classifier module to predict labels in the sequence in order to calculate the combined rewards. The linkage between the two modules is also reflected in the network structure. We add the feature sharing layer, which enables the classifier to distill its intermediate representations to the learning of reinforcement ranker module. Extensive experiments and ablation studies show that both our co-learning strategy and feature sharing can improve semi-supervised ranking problems. |
Issue Date: | 2018-12-14 |
Type: | Thesis |
URI: | http://hdl.handle.net/2142/102868 |
Rights Information: | Copyright 2018 Shibi He |
Date Available in IDEALS: | 2019-02-07 |
Date Deposited: | 2018-12 |
This item appears in the following Collection(s)
-
Dissertations and Theses - Computer Science
Dissertations and Theses from the Dept. of Computer Science -
Graduate Dissertations and Theses at Illinois
Graduate Theses and Dissertations at Illinois