TY - JOUR
T1 - Evolution of reinforcement learning in foraging bees
T2 - A simple explanation for risk averse behavior
AU - Niv, Yael
AU - Joel, Daphna
AU - Meilijson, Isaac
AU - Ruppin, Eytan
PY - 2002
Y1 - 2002
N2 - Reinforcement learning is a fundamental process by which organisms learn to achieve goals from their interactions with the environment. We use evolutionary computation techniques to derive (near-)optimal neuronal learning rules in a simple neural network model of decision-making in simulated bumblebees foraging for nectar. The resulting bees exhibit efficient reinforcement learning. The evolved synaptic plasticity dynamics give rise to varying exploration/exploitation levels and to the well-documented foraging strategy of risk aversion. This behavior is shown to emerge directly from optimal reinforcement learning, providing a biologically founded, parsimonious and novel explanation of risk-averse behavior.
AB - Reinforcement learning is a fundamental process by which organisms learn to achieve goals from their interactions with the environment. We use evolutionary computation techniques to derive (near-)optimal neuronal learning rules in a simple neural network model of decision-making in simulated bumblebees foraging for nectar. The resulting bees exhibit efficient reinforcement learning. The evolved synaptic plasticity dynamics give rise to varying exploration/exploitation levels and to the well-documented foraging strategy of risk aversion. This behavior is shown to emerge directly from optimal reinforcement learning, providing a biologically founded, parsimonious and novel explanation of risk-averse behavior.
KW - Bumble bees
KW - Evolutionary computation
KW - Exploration/exploitation tradeoff
KW - Reinforcement learning
KW - Risk aversion
UR - http://www.scopus.com/inward/record.url?scp=17744419133&partnerID=8YFLogxK
U2 - 10.1016/S0925-2312(02)00496-4
DO - 10.1016/S0925-2312(02)00496-4
M3 - ???researchoutput.researchoutputtypes.contributiontojournal.article???
AN - SCOPUS:17744419133
SN - 0925-2312
VL - 44-46
SP - 951
EP - 956
JO - Neurocomputing
JF - Neurocomputing
ER -