The search functionality is under construction.

Keyword Search Result

[Keyword] gradient ascent learning(4hit)

1-4hit
  • An Elastic Net Learning Algorithm for Edge Linking of Images

    Jiahai WANG  Zheng TANG  Qiping CAO  Xinshun XU  

     
    PAPER-Neural Networks and Bioengineering

      Vol:
    E86-A No:11
      Page(s):
    2879-2886

    Edge linking is a fundamental computer vision task, yet presents difficulties arising from the lack of information in the image. Viewed as a constrained optimization problem, it is NP hard-being isomorphic to the classical Traveling Salesman Problem. This paper proposes a gradient ascent learning algorithm of the elastic net approach for edge linking of images. The learning algorithm has two phases: an elastic net phase, and a gradient ascent phase. The elastic net phase minimizes the path through the edge points. The procedure is equivalent to gradient descent of an energy function, and leads to a local minimum of energy that represents a good solution to the problem. Once the elastic net gets stuck in local minima, the gradient ascent phase attempts to fill up the valley by modifying parameters in a gradient ascent direction of the energy function. Thus, these two phases are repeated until the elastic net gets out of local minima and produces the shortest or better contour through edge points. We test the algorithm on a set of artificial images devised with the aim of demonstrating the sort of features that may occur in real images. For all problems, the systems are shown to be capable of escaping from the elastic net local minima and producing more meaningful contours than the original elastic net.

  • A Gradient Ascent Learning Algorithm for Elastic Nets

    Zheng TANG  Jia Hai WANG  Qi Ping CAO  

     
    PAPER-Neural Networks and Bioengineering

      Vol:
    E86-A No:4
      Page(s):
    940-945

    This paper proposes a gradient ascent learning algorithm for the elastic net approach to the Traveling Salesman Problem (TSP). The learning model has two phases: an elastic net phase, and a gradient ascent phase. The elastic net phase is equivalent to gradient descent of an energy function, and leads to a local minimum of energy that represents a good solution to the problem. Once the elastic net gets stuck in local minima, the gradient ascent phase attempts to fill up the valley by modifying parameters in a gradient ascent direction of the energy function. Thus, these two phases are iterated until the elastic net gets out of local minima. We test the algorithm on many randomly generated travel salesman problems up to 100 cities. For all problems, the systems are shown to be capable of escaping from the elastic net local minima and generating shorter tour than the original elastic net.

  • A Near-Optimum Parallel Algorithm for Bipartite Subgraph Problem Using the Hopfield Neural Network Learning

    Rong-Long WANG  Zheng TANG  Qi-Ping CAO  

     
    PAPER-Neural Networks and Bioengineering

      Vol:
    E85-A No:2
      Page(s):
    497-504

    A near-optimum parallel algorithm for bipartite subgraph problem using gradient ascent learning algorithm of the Hopfield neural networks is presented. This parallel algorithm, uses the Hopfield neural network updating to get a near-maximum bipartite subgraph and then performs gradient ascent learning on the Hopfield network to help the network escape from the state of the near-maximum bipartite subgraph until the state of the maximum bipartite subgraph or better one is obtained. A large number of instances have been simulated to verify the proposed algorithm, with the simulation result showing that our algorithm finds the solution quality is superior to that of best existing parallel algorithm. We also test the proposed algorithm on maximum cut problem. The simulation results also show the effectiveness of this algorithm.

  • A Hopfield Network Learning Algorithm for Graph Planarization

    Zheng TANG  Rong Long WANG  Qi Ping CAO  

     
    LETTER-Neural Networks and Bioengineering

      Vol:
    E84-A No:7
      Page(s):
    1799-1802

    A gradient ascent learning algorithm of the Hopfield neural networks for graph planarization is presented. This learning algorithm uses the Hopfield neural network to get a near-maximal planar subgraph, and increases the energy by modifying parameters in a gradient ascent direction to help the network escape from the state of the near-maximal planar subgraph to the state of the maximal planar subgraph or better one. The proposed algorithm is applied to several graphs up to 150 vertices and 1064 edges. The performance of our algorithm is compared with that of Takefuji/Lee's method. Simulation results show that the proposed algorithm is much better than Takefuji/Lee's method in terms of the solution quality for every tested graph.