The search functionality is under construction.
The search functionality is under construction.

Keyword Search Result

[Keyword] empirical Bayes(2hit)

1-2hit
  • Empirical Bayes Estimation for L1 Regularization: A Detailed Analysis in the One-Parameter Lasso Model

    Tsukasa YOSHIDA  Kazuho WATANABE  

     
    PAPER-Machine learning

      Vol:
    E101-A No:12
      Page(s):
    2184-2191

    Lasso regression based on the L1 regularization is one of the most popular sparse estimation methods. It is often required to set appropriately in advance the regularization parameter that determines the degree of regularization. Although the empirical Bayes approach provides an effective method to estimate the regularization parameter, its solution has yet to be fully investigated in the lasso regression model. In this study, we analyze the empirical Bayes estimator of the one-parameter model of lasso regression and show its uniqueness and its properties. Furthermore, we compare this estimator with that of the variational approximation, and its accuracy is evaluated.

  • Generalization Performance of Subspace Bayes Approach in Linear Neural Networks

    Shinichi NAKAJIMA  Sumio WATANABE  

     
    PAPER-Algorithm Theory

      Vol:
    E89-D No:3
      Page(s):
    1128-1138

    In unidentifiable models, the Bayes estimation has the advantage of generalization performance over the maximum likelihood estimation. However, accurate approximation of the posterior distribution requires huge computational costs. In this paper, we consider an alternative approximation method, which we call a subspace Bayes approach. A subspace Bayes approach is an empirical Bayes approach where a part of the parameters are regarded as hyperparameters. Consequently, in some three-layer models, this approach requires much less computational costs than Markov chain Monte Carlo methods. We show that, in three-layer linear neural networks, a subspace Bayes approach is asymptotically equivalent to a positive-part James-Stein type shrinkage estimation, and theoretically clarify its generalization error and training error. We also discuss the domination over the maximum likelihood estimation and the relation to the variational Bayes approach.