
Efficient regularized least-squares algorithms for conditional ranking on relational data
- Author
- Tapio Pahikkala, Antti Airola, Michiel Stock (UGent) , Bernard De Baets (UGent) and Willem Waegeman (UGent)
- Organization
- Abstract
- In domains like bioinformatics, information retrieval and social network analysis, one can find learning tasks where the goal consists of inferring a ranking of objects, conditioned on a particular target object. We present a general kernel framework for learning conditional rankings from various types of relational data, where rankings can be conditioned on unseen data objects. We propose efficient algorithms for conditional ranking by optimizing squared regression and ranking loss functions. We show theoretically, that learning with the ranking loss is likely to generalize better than with the regression loss. Further, we prove that symmetry or reciprocity properties of relations can be efficiently enforced in the learned models. Experiments on synthetic and real-world data illustrate that the proposed methods deliver state-of-the-art performance in terms of predictive power and computational efficiency. Moreover, we also show empirically that incorporating symmetry or reciprocity properties can improve the generalization performance.
- Keywords
- KERNEL, SUPPORT VECTOR MACHINES, PREFERENCES, EXAMPLES, INVERSE, TREES, TIME, SVMS, Regularized least-squares, Kernel methods, Learning to rank, Symmetric relations, Reciprocal relations, NETWORK INFERENCE, MODEL
Downloads
-
(...).pdf
- full text
- |
- UGent only
- |
- |
- 1.85 MB
Citation
Please use this url to cite or link to this publication: http://hdl.handle.net/1854/LU-4167267
- MLA
- Pahikkala, Tapio, et al. “Efficient Regularized Least-Squares Algorithms for Conditional Ranking on Relational Data.” MACHINE LEARNING, vol. 93, no. 2–3, 2013, pp. 321–56, doi:10.1007/s10994-013-5354-7.
- APA
- Pahikkala, T., Airola, A., Stock, M., De Baets, B., & Waegeman, W. (2013). Efficient regularized least-squares algorithms for conditional ranking on relational data. MACHINE LEARNING, 93(2–3), 321–356. https://doi.org/10.1007/s10994-013-5354-7
- Chicago author-date
- Pahikkala, Tapio, Antti Airola, Michiel Stock, Bernard De Baets, and Willem Waegeman. 2013. “Efficient Regularized Least-Squares Algorithms for Conditional Ranking on Relational Data.” MACHINE LEARNING 93 (2–3): 321–56. https://doi.org/10.1007/s10994-013-5354-7.
- Chicago author-date (all authors)
- Pahikkala, Tapio, Antti Airola, Michiel Stock, Bernard De Baets, and Willem Waegeman. 2013. “Efficient Regularized Least-Squares Algorithms for Conditional Ranking on Relational Data.” MACHINE LEARNING 93 (2–3): 321–356. doi:10.1007/s10994-013-5354-7.
- Vancouver
- 1.Pahikkala T, Airola A, Stock M, De Baets B, Waegeman W. Efficient regularized least-squares algorithms for conditional ranking on relational data. MACHINE LEARNING. 2013;93(2–3):321–56.
- IEEE
- [1]T. Pahikkala, A. Airola, M. Stock, B. De Baets, and W. Waegeman, “Efficient regularized least-squares algorithms for conditional ranking on relational data,” MACHINE LEARNING, vol. 93, no. 2–3, pp. 321–356, 2013.
@article{4167267, abstract = {{In domains like bioinformatics, information retrieval and social network analysis, one can find learning tasks where the goal consists of inferring a ranking of objects, conditioned on a particular target object. We present a general kernel framework for learning conditional rankings from various types of relational data, where rankings can be conditioned on unseen data objects. We propose efficient algorithms for conditional ranking by optimizing squared regression and ranking loss functions. We show theoretically, that learning with the ranking loss is likely to generalize better than with the regression loss. Further, we prove that symmetry or reciprocity properties of relations can be efficiently enforced in the learned models. Experiments on synthetic and real-world data illustrate that the proposed methods deliver state-of-the-art performance in terms of predictive power and computational efficiency. Moreover, we also show empirically that incorporating symmetry or reciprocity properties can improve the generalization performance.}}, author = {{Pahikkala, Tapio and Airola, Antti and Stock, Michiel and De Baets, Bernard and Waegeman, Willem}}, issn = {{0885-6125}}, journal = {{MACHINE LEARNING}}, keywords = {{KERNEL,SUPPORT VECTOR MACHINES,PREFERENCES,EXAMPLES,INVERSE,TREES,TIME,SVMS,Regularized least-squares,Kernel methods,Learning to rank,Symmetric relations,Reciprocal relations,NETWORK INFERENCE,MODEL}}, language = {{eng}}, number = {{2-3}}, pages = {{321--356}}, title = {{Efficient regularized least-squares algorithms for conditional ranking on relational data}}, url = {{http://doi.org/10.1007/s10994-013-5354-7}}, volume = {{93}}, year = {{2013}}, }
- Altmetric
- View in Altmetric
- Web of Science
- Times cited: