Ground-nesting insects could use visual tracking for monitoring nest position during learning flights
Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
MetadataShow full item record
Please cite this item using this persistent URLhttp://hdl.handle.net/11693/26830
Ants, bees and wasps are central place foragers. They leave their nests to forage and routinely return to their home-base. Most are guided by memories of the visual panorama and the visual appearance of the local nest environment when pinpointing their nest. These memories are acquired during highly structured learning walks or flights that are performed when leaving the nest for the first time or whenever the insects had difficulties finding the nest during their previous return. Ground-nesting bees and wasps perform such learning flights daily when they depart for the first time. During these flights, the insects turn back to face the nest entrance and subsequently back away from the nest while flying along ever increasing arcs that are centred on the nest. Flying along these arcs, the insects counter-turn in such a way that the nest entrance is always seen in the frontal visual field at slightly lateral positions. Here we asked how the insects may achieve keeping track of the nest entrance location given that it is a small, inconspicuous hole in the ground, surrounded by complex natural structures that undergo unpredictable perspective transformations as the insect pivots around the area and gains distance from it. We reconstructed the natural visual scene experienced by wasps and bees during their learning flights and applied a number of template-based tracking methods to these image sequences. We find that tracking with a fixed template fails very quickly in the course of a learning flight, but that continuously updating the template allowed us to reliably estimate nest direction in reconstructed image sequences. This is true even for later sections of learning flights when the insects are so far away from the nest that they cannot resolve the nest entrance as a visual feature. We discuss why visual goal-anchoring is likely to be important during the acquisition of visual-spatial memories and describe experiments to test whether insects indeed update nest-related templates during their learning flights. © 2014 Springer International Publishing Switzerland.
- Conference Paper 2294
Showing items related by title, author, creator and subject.
Aman, Aytek (Bilkent University, 2014)Augmented reality (AR) is the enhancement of real scenes with virtual entities. It is used to enhance user experience and interaction in various ways. Educational applications, architectural visualizations, military ...
Multiperson tracking with a network of ultrawideband radar sensors based on gaussian mixture PHD filters Gulmezoglu, B.; Guldogan, M.B.; Gezici, S. (Institute of Electrical and Electronics Engineers Inc., 2015)In this paper, we investigate the use of Gaussian mixture probability hypothesis density filters for multiple person tracking using ultrawideband (UWB) radar sensors in an indoor environment. An experimental setup consisting ...
Demir H.S.; Cetin A.E. (IEEE Computer Society, 2016)This paper presents a novel infrared (IR) object tracking algorithm based on the co-difference matrix. Extraction of co-difference features is similar to the well known covariance method except that the vector product ...