LOGIN TO YOUR ACCOUNT

Username
Password
Remember Me
Or use your Academic/Social account:

CREATE AN ACCOUNT

Or use your Academic/Social account:

Congratulations!

You have just completed your registration at OpenAire.

Before you can login to the site, you will need to activate your account. An e-mail will be sent to you with the proper instructions.

Important!

Please note that this site is currently undergoing Beta testing.
Any new content you create is not guaranteed to be present to the final version of the site upon release.

Thank you for your patience,
OpenAire Dev Team.

Close This Message

CREATE AN ACCOUNT

Name:
Username:
Password:
Verify Password:
E-mail:
Verify E-mail:
*All Fields Are Required.
Please Verify You Are Human:
fbtwitterlinkedinvimeoflicker grey 14rssslideshare1
Teichmann, J.; Broom, M.; Alonso, E. (2014)
Publisher: Elsevier
Languages: English
Types: Article
Subjects: QA, QH
An experience-based aversive learning model of foraging behaviour in uncertain environments is presented. We use Q-learning as a model-free implementation of Temporal difference learning motivated by growing evidence for neural correlates in natural reinforcement settings. The predator has the choice of including an aposematic prey in its diet or to forage on alternative food sources. We show how the predator's foraging behaviour and energy intake depend on toxicity of the defended prey and the presence of Batesian mimics. We introduce the precondition of exploration of the action space for successful aversion formation and show how it predicts foraging behaviour in the presence of conflicting rewards which is conditionally suboptimal in a fixed environment but allows better adaptation in changing environments.
  • The results below are discovered through our pilot algorithms. Let us know how we are doing!

    • 4 5 6 7 8 TD{ezrror [2] S. Yachi and M. Higashi, “The evolution of warning signals,” Nature, vol. 394, no. 6696, pp. 882-884, 1998.
    • [3] M. Broom, M. Speed, and G. Ruxton, “Evolutionarily stable defence and signalling of that defence,” Journal of Theoretical Biology, vol. 242, pp. 32- 34, 2006.
    • [4] O. Leimar, M. Enquist, and B. Sillen-Tullberg, “Evolutionary stability of aposematic coloration and prey unprofitability: A theoretical analysis,” American Society of Naturalists, vol. 128, pp. 469-490, 1986.
    • [5] T. J. Lee, M. P. Speed, and P. A. Stephens, “Honest signaling and the uses of prey coloration,” American Society of Naturalists, vol. 178, pp. E1-E9, 2011.
    • [6] N. M. Marples, D. J. Kelly, and R. J. Thomas, “Perspective: The evolution of warning coloration is not paradoxical,” Evolution, vol. 59, no. 5, pp. 933- 940, 2005.
    • [7] M. P. Speed, “Warning signals, receiver psychology and predator memory,” Animal Behaviour, vol. 60, no. 3, pp. 269 - 278, 2000.
    • [8] K. Svádová, A. Exnerová, P. Štys, E. Landová, J. Valenta, A. Fučíková, and R. Socha, “Role of different colours of aposematic insects in learning, memory and generalization of naïve bird predators,” Animal Behaviour, vol. 77, no. 2, pp. 327 - 336, 2009.
    • [9] J. Skelhorn and C. Rowe, “Prey palatability influences predator learning and memory,” Animal Behaviour, vol. 71, no. 5, pp. 1111 - 1118, 2006.
    • 289 [10] A. N. Johnston and T. H. Burne, “Aposematic colouration enhances mem290 ory formation in domestic chicks trained in a weak passive avoidance learn291 ing paradigm,” Brain Research Bulletin, vol. 76, no. 3, pp. 313 - 316, 2008.
    • 292 [11] M. Speed and G. Ruxton, “Aposematism: what should our starting point 293 be?,” Proceedings of the Royal Society B: Biological Sciences, vol. 272, 294 no. 1561, pp. 431-438, 2005.
    • 295 [12] C. Barnett, M. Bateson, and C. Rowe, “State-dependent decision making: 296 educated predators strategically trade off the costs and benefits of con297 suming aposematic prey,” Behavioral Ecology, vol. 18, no. 4, pp. 645-651, 298 2007.
    • 299 [13] T. N. Sherratt, “State-dependent risk-taking by predators in systems with 300 defended prey,” Oikos, vol. 103, no. 1, pp. 93-100, 2003.
    • TD{ezrror 421 Q 0 422 s_k s_0 423 WHILE learning DO 424 a_k (s_k; Q) 425 s_(k + 1) f (s_k; a_k) 426 Q(s_k; a_k) Q(s_k; a_k) + 427 max _a Q(s_(k + 1); a) 428 s_k s_(k + 1) 429
  • No related research data.
  • No similar publications.

Share - Bookmark

Cite this article