The search functionality is under construction.
The search functionality is under construction.

Fast Iterative Mining Using Sparsity-Inducing Loss Functions

Hiroto SAIGO, Hisashi KASHIMA, Koji TSUDA

  • Full Text Views

    0

  • Cite this

Summary :

Apriori-based mining algorithms enumerate frequent patterns efficiently, but the resulting large number of patterns makes it difficult to directly apply subsequent learning tasks. Recently, efficient iterative methods are proposed for mining discriminative patterns for classification and regression. These methods iteratively execute discriminative pattern mining algorithm and update example weights to emphasize on examples which received large errors in the previous iteration. In this paper, we study a family of loss functions that induces sparsity on example weights. Most of the resulting example weights become zeros, so we can eliminate those examples from discriminative pattern mining, leading to a significant decrease in search space and time. In computational experiments we compare and evaluate various loss functions in terms of the amount of sparsity induced and resulting speed-up obtained.

Publication
IEICE TRANSACTIONS on Information Vol.E96-D No.8 pp.1766-1773
Publication Date
2013/08/01
Publicized
Online ISSN
1745-1361
DOI
10.1587/transinf.E96.D.1766
Type of Manuscript
PAPER
Category
Pattern Recognition

Authors

Hiroto SAIGO
  Kyushu Institute of Technology
Hisashi KASHIMA
  University of Tokyo
Koji TSUDA
  Advanced Industrial Science and Technology

Keyword