A4 Vertaisarvioitu artikkeli konferenssijulkaisussa

Fast Gradient Computation for Learning with Tensor Product Kernels and Sparse Training Labels




TekijätTapio Pahikkala

ToimittajaPasi Fränti, Gavin Brown, Marco Loog, Francisco Escolano, Marcello Pelillo

Konferenssin vakiintunut nimiJoint IAPR International Workshop, S+SSPR 2014

Julkaisuvuosi2014

JournalLecture Notes in Computer Science

Kokoomateoksen nimiStructural, Syntactic, and Statistical Pattern Recognition (S+SSPR 2014)

Sarjan nimiLecture Notes in Computer Science

Vuosikerta8621

Aloitussivu123

Lopetussivu132

Sivujen määrä10

ISBN978-3-662-44414-6

ISSN0302-9743

DOIhttps://doi.org/10.1007/978-3-662-44415-3_13


Tiivistelmä

Supervised learning with pair-input data has recently become one of the most intensively studied topics in pattern recognition literature, and its applications are numerous, including, for example, collaborative filtering, information retrieval, and drug-target interaction prediction. Regularized least-squares (RLS) is a kernel-based learning algorithm that, together with tensor product kernels, is a successful tool for solving pair-input learning problems, especially the ones in which the aim is to generalize to new types of inputs not encountered in during the training phase. The training of tensor kernel RLS models for pair-input problems has been traditionally accelerated with the so-called vec-trick. We show that it can be further accelerated by taking advantage of the sparsity of the training labels. This speed improvement is demonstrated in a running time experiment and the applicability of the algorithm in a practical problem of predicting drug-target interactions.



Ladattava julkaisu

This is an electronic reprint of the original article.
This reprint may differ from the original in pagination and typographic detail. Please cite the original version.





Last updated on 2024-26-11 at 21:17