Abstract | ||
---|---|---|
The Particle Swarm Optimizer (PSO) has previously been used to train neural networks and generally met with success. The advantage of the PSO over many of the other optimization algorithms is its relative simplicity and quick convergence. But those particles collapse so quickly that it exits a potentially dangerous property: stagnation, which state would make it impossible to arrive at the global optimum, even a local optimum. The ecological and physical universal laws enlighten us to improve the PSO algorithm. We introduce a concept, swarm's survival density, into PSO for balancing the gravity and repulsion forces between two particles. A modified algorithm, survival density particle swarm optimization (SDPSO) is proposed for neural network training in this paper. Then it is applied to benchmark function minimization problems and neural network training for benchmark dataset classification problems. The experimental results illustrate its efficiency. |
Year | DOI | Venue |
---|---|---|
2004 | 10.1007/978-3-540-28647-9_56 | ADVANCES IN NEURAL NETWORKS - ISNN 2004, PT 1 |
Keywords | Field | DocType |
neural network | Particle swarm optimization,Convergence (routing),Mathematical optimization,Swarm behaviour,Computer science,Local optimum,Global optimum,Multi-swarm optimization,Artificial intelligence,Artificial neural network,Machine learning,Metaheuristic | Conference |
Volume | ISSN | Citations |
3173 | 0302-9743 | 1 |
PageRank | References | Authors |
0.37 | 3 | 5 |
Name | Order | Citations | PageRank |
---|---|---|---|
Hongbo Liu | 1 | 1426 | 105.95 |
Bo Li | 2 | 8 | 1.22 |
Xiu-kun Wang | 3 | 45 | 8.99 |
Ye Ji | 4 | 1 | 0.37 |
Yiyuan Tang | 5 | 82 | 8.30 |