The search functionality is under construction.

Keyword Search Result

[Keyword] learning automata(4hit)

1-4hit
  • An Improved Model of Ant Colony Optimization Using a Novel Pheromone Update Strategy

    Pooia LALBAKHSH  Bahram ZAERI  Ali LALBAKHSH  

     
    PAPER-Fundamentals of Information Systems

      Vol:
    E96-D No:11
      Page(s):
    2309-2318

    The paper introduces a novel pheromone update strategy to improve the functionality of ant colony optimization algorithms. This modification tries to extend the search area by an optimistic reinforcement strategy in which not only the most desirable sub-solution is reinforced in each step, but some of the other partial solutions with acceptable levels of optimality are also favored. therefore, it improves the desire for the other potential solutions to be selected by the following artificial ants towards a more exhaustive algorithm by increasing the overall exploration. The modifications can be adopted in all ant-based optimization algorithms; however, this paper focuses on two static problems of travelling salesman problem and classification rule mining. To work on these challenging problems we considered two ACO algorithms of ACS (Ant Colony System) and AntMiner 3.0 and modified their pheromone update strategy. As shown by simulation experiments, the novel pheromone update method can improve the behavior of both algorithms regarding almost all the performance evaluation metrics.

  • Inertial Estimator Learning Automata

    Junqi ZHANG  Lina NI  Chen XIE  Shangce GAO  Zheng TANG  

     
    PAPER-Numerical Analysis and Optimization

      Vol:
    E95-A No:6
      Page(s):
    1041-1048

    This paper presents an inertial estimator learning automata scheme by which both the short-term and long-term perspectives of the environment can be incorporated in the stochastic estimator – the long term information crystallized in terms of the running reward-probability estimates, and the short term information used by considering whether the most recent response was a reward or a penalty. Thus, when the short-term perspective is considered, the stochastic estimator becomes pertinent in the context of the estimator algorithms. The proposed automata employ an inertial weight estimator as the short-term perspective to achieve a rapid and accurate convergence when operating in stationary random environments. According to the proposed inertial estimator scheme, the estimates of the reward probabilities of actions are affected by the last response from environment. In this way, actions that have gotten the positive response from environment in the short time, have the opportunity to be estimated as “optimal”, to increase their choice probability and consequently, to be selected. The estimates become more reliable and consequently, the automaton rapidly and accurately converges to the optimal action. The asymptotic behavior of the proposed scheme is analyzed and it is proved to be ε-optimal in every stationary random environment. Extensive simulation results indicate that the proposed algorithm converges faster than the traditional stochastic-estimator-based S ERI scheme, and the deterministic-estimator-based DGPA and DPRI schemes when operating in stationary random environments.

  • A New Learning Algorithm for the Hierarchical Structure Learning Automata Operating in the General Multiteacher Environment

    Norio BABA  Yoshio MOGAMI  

     
    PAPER-Automata and Formal Language Theory

      Vol:
    E87-D No:5
      Page(s):
    1208-1213

    Learning behaviors of hierarchically structured stochastic automata operating in a general nonstationary multiteacher environment are considered. It is shown that convergence with probability 1 to the optimal path is ensured by a new learning algorithm which is an extended form of the relative reward strength algorithm. Several computer simulation results confirm the effectiveness of the proposed algorithm.

  • Introducing an Adaptive VLR Algorithm Using Learning Automata for Multilayer Perceptron

    Behbood MASHOUFI  Mohammad Bagher MENHAJ  Sayed A. MOTAMEDI  Mohammad R. MEYBODI  

     
    PAPER-Algorithms

      Vol:
    E86-D No:3
      Page(s):
    594-609

    One of the biggest limitations of BP algorithm is its low rate of convergence. The Variable Learning Rate (VLR) algorithm represents one of the well-known techniques that enhance the performance of the BP. Because the VLR parameters have important influence on its performance, we use learning automata (LA) to adjust them. The proposed algorithm named Adaptive Variable Learning Rate (AVLR) algorithm dynamically tunes the VLR parameters by learning automata according to the error changes. Simulation results on some practical problems such as sinusoidal function approximation, nonlinear system identification, phoneme recognition, Persian printed letter recognition helped us better to judge the merit of the proposed AVLR method.