Cardiff University | Prifysgol Caerdydd ORCA
Online Research @ Cardiff 
WelshClear Cookie - decide language by browser settings

Optimal quadrature-sparsification for integral operator approximation

Gauthier, Bertrand and Suykens, Johan A.K. 2018. Optimal quadrature-sparsification for integral operator approximation. SIAM Journal on Scientific Computing 40 (5) , A3636-A3674. 10.1137/17M1123614

[img]
Preview
PDF - Accepted Post-Print Version
Download (2MB) | Preview

Abstract

The design of sparse quadratures for the approximation of integral operators related to symmetric positive-semidefinite kernels is addressed. Particular emphasis is placed on the approximation of the main eigenpairs of an initial operator and on the assessment of the approximation accuracy. Special attention is drawn to the design of sparse quadratures with support included in fixed finite sets of points (that is, quadrature-sparsification), this framework encompassing the approximation of kernel matrices. For a given kernel, the accuracy of a quadrature approximation is assessed through the squared Hilbert--Schmidt norm (for operators acting on the underlying reproducing kernel Hilbert space) of the difference between the integral operators related to the initial and approximate measures; by analogy with the notion of kernel discrepancy, the underlying criterion is referred to as the squared-kernel discrepancy between the two measures. In the quadrature-sparsification framework, sparsity of the approximate quadrature is promoted through the introduction of an $\ell^{1}$-type penalization, and the computation of a penalized squared-kernel-discrepancy-optimal approximation then consists in a convex quadratic minimization problem; such quadratic programs can in particular be interpreted as the Lagrange dual formulations of distorted one-class support-vector machines related to the squared kernel. Error bounds on the induced spectral approximations are derived, and the connection between penalization, sparsity, and accuracy of the spectral approximation is investigated. Numerical strategies for solving large-scale penalized squared-kernel-discrepancy minimization problems are discussed, and the efficiency of the approach is illustrated by a series of examples. In particular, the ability of the proposed methodology to lead to accurate approximations of the main eigenpairs of kernel matrices related to large-scale datasets is demonstrated.

Item Type: Article
Date Type: Publication
Status: Published
Schools: Mathematics
Subjects: Q Science > QA Mathematics
Publisher: Society for Industrial and Applied Mathematics
ISSN: 1064-8275
Date of First Compliant Deposit: 10 July 2018
Date of Acceptance: 5 July 2018
Last Modified: 28 Jun 2019 16:08
URI: http://orca.cf.ac.uk/id/eprint/113057

Actions (repository staff only)

Edit Item Edit Item

Downloads

Downloads per month over past year

View more statistics