ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1802.07382
66
15

Generic Coreset for Scalable Learning of Monotonic Kernels: Logistic Regression, Sigmoid and more

21 February 2018
Elad Tolochinsky
Ibrahim Jubran
Dan Feldman
    CLL
ArXivPDFHTML
Abstract

Coreset (or core-set) is a small weighted \emph{subset} QQQ of an input set PPP with respect to a given \emph{monotonic} function f:R→Rf:\mathbb{R}\to\mathbb{R}f:R→R that \emph{provably} approximates its fitting loss ∑p∈Pf(p⋅x)\sum_{p\in P}f(p\cdot x)∑p∈P​f(p⋅x) to \emph{any} given x∈Rdx\in\mathbb{R}^dx∈Rd. Using QQQ we can obtain approximation of x∗x^*x∗ that minimizes this loss, by running \emph{existing} optimization algorithms on QQQ. In this work we provide: (i) A lower bound which proves that there are sets with no coresets smaller than n=∣P∣n=|P|n=∣P∣ for general monotonic loss functions. (ii) A proof that, under a natural assumption that holds e.g. for logistic regression and the sigmoid activation functions, a small coreset exists for \emph{any} input PPP. (iii) A generic coreset construction algorithm that computes such a small coreset QQQ in O(nd+nlog⁡n)O(nd+n\log n)O(nd+nlogn) time, and (iv) Experimental results which demonstrate that our coresets are effective and are much smaller in practice than predicted in theory.

View on arXiv
Comments on this paper