TY - GEN
T1 - Reinforced snap-drift learning for proxylet selection in active computer networks
AU - Lee, Sin Wee
AU - Palmer-Brown, Dominic
AU - Roadknight, Chris
PY - 2004
Y1 - 2004
N2 - A new continuous learning method is applied to the problem of optimizing the selection of services in response to user requests in an active computer network simulation environment. The learning is an enhanced version of the 'snap-drift' algorithm, a hybrid form of learning that employs the complementary modes: fast, minimalist (snap) learning; and slower drift (towards the input patterns) learning, in a nonstationary environment where new patterns are continually introduced. Snap is based on Adaptive Resonance Theory, and drift on Learning Vector Quantization. The new algorithm swaps its learning style between the two modes of self-organisation when declining performance levels are received, but maintains the same learning style during episodes of improved performance. Performance updates occur at the end of each epoch. Reinforcement also occurs by maintaining successful adaptations, since learning is enabled with a probability that increases with declining performance. The method is capable of rapidly re-learning and is used in the design of a modular neural network system, Performanceguided Adaptive Resonance Theory. Simulations demonstrate the learning is stable, effective and able to discover alternative solutions in response to new performance requirements and significant changes in the stream of input patterns.
AB - A new continuous learning method is applied to the problem of optimizing the selection of services in response to user requests in an active computer network simulation environment. The learning is an enhanced version of the 'snap-drift' algorithm, a hybrid form of learning that employs the complementary modes: fast, minimalist (snap) learning; and slower drift (towards the input patterns) learning, in a nonstationary environment where new patterns are continually introduced. Snap is based on Adaptive Resonance Theory, and drift on Learning Vector Quantization. The new algorithm swaps its learning style between the two modes of self-organisation when declining performance levels are received, but maintains the same learning style during episodes of improved performance. Performance updates occur at the end of each epoch. Reinforcement also occurs by maintaining successful adaptations, since learning is enabled with a probability that increases with declining performance. The method is capable of rapidly re-learning and is used in the design of a modular neural network system, Performanceguided Adaptive Resonance Theory. Simulations demonstrate the learning is stable, effective and able to discover alternative solutions in response to new performance requirements and significant changes in the stream of input patterns.
UR - http://www.scopus.com/inward/record.url?scp=10944240542&partnerID=8YFLogxK
U2 - 10.1109/IJCNN.2004.1380185
DO - 10.1109/IJCNN.2004.1380185
M3 - Conference contribution
AN - SCOPUS:10944240542
SN - 0780383591
T3 - IEEE International Conference on Neural Networks - Conference Proceedings
SP - 1545
EP - 1550
BT - 2004 IEEE International Joint Conference on Neural Networks - Proceedings
T2 - 2004 IEEE International Joint Conference on Neural Networks - Proceedings
Y2 - 25 July 2004 through 29 July 2004
ER -