Text Box 1. Motion-Driven Image Sampling in Morphodynamic Active Vision. This text box graphically illustrates two fundamental sampling mechanisms that enhance vision through motion, using Drosophila as a model organism: (A) local sampling at the level of individual photoreceptors (“single-pixel”) and (B) global sampling across the entire retinal matrix (whole-matrix). These interactive processes, which jointly affect the eyes’ spatiotemporal resolution, stereoscopic range and adaptive capabilities, likely co-evolved to optimise visual perception and behaviour in dynamic natural environments.
Because compound eyes extend from the rigid head exoskeleton, appearing stationary to an outside observer, the
prima facie is that their inner workings would also be immobile[
52,
53,
75]. Therefore, as the eyes’ ommatidial faceting sets their photoreceptor spacing, the influential static theory of compound eye optics postulates that insects can only see a “pixelated” low-resolution image of the world. According to this traditional static viewpoint, the ommatidial grid limits the granularity of the retinal image and visual acuity. Resolving two stationary objects requires at least three photoreceptors, and this task becomes more challenging when objects are in motion, further reducing visual acuity. The presumed characteristics associated with small static compound eyes, including large receptive fields, slow integration times, and spatial separation of photoreceptors, commonly attributed to spherical geometry, contribute to motion blur that impairs the detailed resolution of moving objects within the visual field[
52]. As a result, male
Drosophila relying on coarse visual information face a real dilemma in distinguishing between a receptive female fly and a hungry spider. To accurately differentiate, the male must closely approach the subject to detect distinguishing characteristics such as body shape, colour patterns, or movements. In this context, the difference between sex and death may hinge on an invisible line.
Recent studies on
Drosophila have challenged the notion that fixed factors such as photoreceptor spacing, integration time, and receptive field size solely determine visual acuity[
13,
15]. Instead, these characteristics are dynamically regulated by photoreceptor photomechanics[
13,
14,
15], leading to significant improvements in vision through morphodynamic processes. In the following subsections, we begin by explaining how microsaccadic movements of photoreceptors enable hyperacute image sampling (
Figure 2), phasic image contrast enhancement (
Figure 3), information maximisation during saccadic behaviours (
Figure 4), hyperacute stereovision (
Figure 5), and antialiased vision (
Figure 6). We then relate these predominantly
local image sampling dynamics to the
global movements of the retina, head, and entire animal in goal-oriented visual behaviours. Finally, we discuss the generic benefits of neural morphodynamics. Through specific examples, we explore how morphodynamic information sampling and processing adapt to maximise information allocation in neural channels (
Figure 7). We also link multiscale observations with
Gedankenexperiments to envision how these ultrafast phasic processes synchronise the brain’s neural representation of the external world with its dynamic changes (
Figure 8), thereby enhancing cognitive abilities and efficiency. Some of these concepts related to neural computations, intelligence, and future technologies are further explored in Text Boxes 2–4.
Figure 2.
Photomechanical photoreceptor microsaccades enhance insect vision through adaptive compound eye optics. (
A) High-speed infrared deep-pseudopupil microscopy [
14,
15] uncovers the intricate movement dynamics and specific directions of light-induced photoreceptor microsaccades across the compound eyes in living
Drosophila. Fully immobilising the flies inside a pipette tip minimises whole-retina movements, allowing one to record photoreceptor microsaccade dynamics in isolation[
14,
15]. (
B) During a microsaccade within an ommatidium, the R1-R7/8 photoreceptors undergo rapid axial (inward) contraction and sideways movement along the R1-R2-R3 direction, executing a complex piston motion[
14,
15]. Meanwhile, the lens positioned above them, as an integral component of the rigid exoskeleton, remains stationary[
13]. (
C) When a moving light stimulus, such as two bright dots, traverses a photoreceptor’s (shown for R5) receptive field (RF), the photoreceptor rapidly contracts away from the lens, causing the RF to narrow[
13,
15]. Simultaneously, the photoreceptor’s swift sideways movement, aided by the lens acting as an optical lever, results in the RF moving in the opposite direction (of about 40-60°/s, illustrated here for movement with or against the stimuli). As a result, in a morphodynamic compound eye, the photoreceptor responses (depicted by blue and red traces) can detect finer and faster changes in moving stimuli than what the previous static compound eye theory predicts (represented by black traces). (
D) Microsaccades result from photomechanical processes involving refractory photon sampling dynamics within the 30,000 microvilli [
8,
13,
15,
61], which comprise the light-sensitive part of a photoreceptor known as the rhabdomere. Each microvillus encompasses the complete phototransduction cascade, enabling the conversion of successful photon captures into elementary responses called quantum bumps. This photomechanical refractory sampling mechanism empowers photoreceptors to consistently estimate changes in environmental light contrast across a wide logarithmic intensity range. The intracellularly recorded morphodynamic quantal information sampling and processing (represented by dark blue traces) can be accurately simulated under various light conditions using biophysically realistic stochastic photoreceptor sampling models (illustrated by cyan traces) [
13,
15,
76]. (
E)
Drosophila photoreceptor microsaccades shift their rhabdomeres sideways by around 1-1.5 µm (maximum < 2 µm), resulting in receptive field movements of approximately 3-4.5° in the visual space. The receptive field half-widths of R1-6 photoreceptors cover the entire visual space, ranging from 4.5-6°. By limiting the micro-scanning to the interommatidial angle,
Drosophila integrates a neural image that surpasses the optical limits of its compound eyes. Honeybee photoreceptor microsaccades shift their receptive fields by < 1°, smaller than the average receptive field half-width (~1.8°) at the front of the eye. This active sampling strategy in honeybees is similar to
Drosophila and suggests that honeybee vision also surpasses the static pixelation limit of its compound eyes [
14]. Data are modified from the cited papers.
Figure 2.
Photomechanical photoreceptor microsaccades enhance insect vision through adaptive compound eye optics. (
A) High-speed infrared deep-pseudopupil microscopy [
14,
15] uncovers the intricate movement dynamics and specific directions of light-induced photoreceptor microsaccades across the compound eyes in living
Drosophila. Fully immobilising the flies inside a pipette tip minimises whole-retina movements, allowing one to record photoreceptor microsaccade dynamics in isolation[
14,
15]. (
B) During a microsaccade within an ommatidium, the R1-R7/8 photoreceptors undergo rapid axial (inward) contraction and sideways movement along the R1-R2-R3 direction, executing a complex piston motion[
14,
15]. Meanwhile, the lens positioned above them, as an integral component of the rigid exoskeleton, remains stationary[
13]. (
C) When a moving light stimulus, such as two bright dots, traverses a photoreceptor’s (shown for R5) receptive field (RF), the photoreceptor rapidly contracts away from the lens, causing the RF to narrow[
13,
15]. Simultaneously, the photoreceptor’s swift sideways movement, aided by the lens acting as an optical lever, results in the RF moving in the opposite direction (of about 40-60°/s, illustrated here for movement with or against the stimuli). As a result, in a morphodynamic compound eye, the photoreceptor responses (depicted by blue and red traces) can detect finer and faster changes in moving stimuli than what the previous static compound eye theory predicts (represented by black traces). (
D) Microsaccades result from photomechanical processes involving refractory photon sampling dynamics within the 30,000 microvilli [
8,
13,
15,
61], which comprise the light-sensitive part of a photoreceptor known as the rhabdomere. Each microvillus encompasses the complete phototransduction cascade, enabling the conversion of successful photon captures into elementary responses called quantum bumps. This photomechanical refractory sampling mechanism empowers photoreceptors to consistently estimate changes in environmental light contrast across a wide logarithmic intensity range. The intracellularly recorded morphodynamic quantal information sampling and processing (represented by dark blue traces) can be accurately simulated under various light conditions using biophysically realistic stochastic photoreceptor sampling models (illustrated by cyan traces) [
13,
15,
76]. (
E)
Drosophila photoreceptor microsaccades shift their rhabdomeres sideways by around 1-1.5 µm (maximum < 2 µm), resulting in receptive field movements of approximately 3-4.5° in the visual space. The receptive field half-widths of R1-6 photoreceptors cover the entire visual space, ranging from 4.5-6°. By limiting the micro-scanning to the interommatidial angle,
Drosophila integrates a neural image that surpasses the optical limits of its compound eyes. Honeybee photoreceptor microsaccades shift their receptive fields by < 1°, smaller than the average receptive field half-width (~1.8°) at the front of the eye. This active sampling strategy in honeybees is similar to
Drosophila and suggests that honeybee vision also surpasses the static pixelation limit of its compound eyes [
14]. Data are modified from the cited papers.
Figure 3.
Saccadic Turns and Fixation Periods Enhance Information Extraction in Drosophila. (
A) A representative walking trajectory of a fruit fly [
67]. (B) Angular velocity and yaw of the recorded walk. (
C) A 360° natural scene utilised to generate three distinct time series of light intensity[
13]. The dotted white line indicates the intensity plane employed during the walk. The blue trace represents a light intensity over time generated by overlaying the walking fly’s yaw dynamics (A-B) onto the scene. The red trace corresponds to the time series of light intensity obtained by scanning the scene at the median velocity of the walk (linear: 63.3°/s). The grey trace depicts the time series of light intensity obtained using shuffled walking velocities. Brief saccades and longer fixation periods introduce burst-like patterns to the light input. (
D) These light intensity time series were employed as stimuli in intracellular photoreceptor recordings and simulations using a biophysically realistic stochastic photoreceptor model. Both the recordings and simulations showed that saccadic viewing enhances information transmission in R1-6 photoreceptors, indicating that this mechanism has evolved with refractory photon sampling to maximise information capture from natural scenes[
13]. Immobilising the flies (their head, proboscis and thorax) with beeswax[87,88] in a conical holder minimises whole-retina movements[
13,
14,
15], enabling high signal-to-noise recording conditions to study photoreceptors’ voltage responses to dynamic light stimulation[
13]. Data are modified from the cited papers.
Figure 4.
The mirror-symmetric ommatidial photoreceptor arrangement and morphodynamics of the left and right eyes enhance detection of moving objects during visual behaviours. (A) The photoreceptor rhabdomere patterns (as indicated by their rotating orientation directions: yellow and green arrows) of the ommatidial left and right eyes (inset images) exhibit horizontal and ventral mirror symmetry, forming a concentrically expanding diamond shape[
14,
15,106]. (B) When a moving object, such as a fly, enters the receptive fields (RFs) of the corresponding frontal left and right photoreceptors (indicated by red and blue beams), the resulting light intensity changes cause the photoreceptors to contract mirror-symmetrically. (C) The half-widths of the frontal left and right eye R6 photoreceptors’ RFs (disks), projected 5 mm away from the eyes[
15]. Red circles represent the RFs of neighbouring photoreceptors in the left visual field, blue in the right. (D) Contraction (light-on) moves R1-R7/8 photoreceptors (left) in R3-R2-R1 direction (fast-phase), recoil (light-off) returns them in opposite R1-R2-R3 direction (slow-phase)[
14,
15]. The corresponding fast-phase (centre) and slow-phase (right) RF vector maps. (E) The fast-phase RF map compared to the forward flying fly’s optic flow field (centre), as experienced with the fly head upright[
15]. Their difference is shown right. The fast-phase matches the ground flow (light yellow pixels), while the opposite slow-phase (dark yellow pixels) matches the sky flow[
15]. (F) During yaw rotation, the mirror-symmetric movement of the photoreceptor RFs in the left and right eyes enhances the binocular contrast differences in the surrounding environment (sample visualisation as panel E). Immobilising the flies inside a pipette tip, as was done for these recordings, minimises whole-retina movements, allowing for the isolated study of photoreceptor microsaccade dynamics[
14,
15]. Data are modified from the cited papers.
Figure 5.
Drosophila visual behaviours exhibit hyperacute 3D vision, aligning with morphodynamic compound eye modelling.
(A) Drosophila compound eyes’ depth perception constraints and the computations for morphodynamic triangulation of object depth (z)[15]. k is the distance between the corresponding left and right eye photoreceptors, and t is their time-delay. t
c is the time-delay between the neighbouring photoreceptors in the same eye. The left eye is represented by the red receptive field (RFs), while the right eye is represented by the blue RF. Simulated voltage responses (top) of three morphodynamically responding R6-photoreceptors when a 1.7° x 1.7° object (orange) moves across their overlapping RFs at a speed of 50°/s and a distance of 25 mm. The corresponding binocular cross-correlations (bottom), which represents the depth information, likely occur in the retinotopically organised neural cartridges of the lobula optic lobe, where location-specific ipsi- and contralateral photoreceptor information is pooled (green LC14 neuron[
15]). Time delays between the maximum correlations (vertical lines) and the moment the object crosses the RF centre of the left R6-photoreceptor (vertical dashed line). (
B) In neural superposition wiring[111], the R1-6 photoreceptors originating from six neighbouring ommatidia sample a moving stimulus (orange dot). Their overlapping receptive fields (RFs; coloured rings) swiftly bounce along their predetermined microsaccade directions (coloured arrows; see also
Figure 4D) as the photoreceptors transmit information to large monopolar cells (LMC, specifically L1-L3, with L2 shown) and the lamina amacrine cells. While R7/8 photoreceptors share some information with R1 and R6 through gap junctions[105] R7/8 establish synapses in the medulla. Simulations reveal the superpositional R1-R7/8s’ voltage responses (coloured traces) with their phase differences when a 1.7° x 1.7° dot traverses their receptive fields at 100°/s (orange dot). 2-photon imaging of L2 terminals’ Ca
2+-responses to a dynamically narrowing black-and-white grid that moves in different directions shows L2 monopolar cells generating hyperacute (<5°; cf.
Figure 2B-C,E) responses along the same microsaccade movement axis (coloured arrows) of the superpositioned photoreceptors that feed information to them (cf.
Figure 4). (
C) In a visual learning experiment, a tethered, head-immobilised
Drosophila flies in a flight simulator. The fly was positioned at the centre of a panoramic arena to prevent it from perceiving motion parallax cues[
15]. The arena features two hyperacute dots placed 180° apart and two equally sized 3D pins positioned perpendicular to the dots. The fly generates subtle yaw torque signals to indicate its intention to turn left or right, allowing it to explore the visual objects within the arena. These signals are used to rotate the arena in the opposite direction of the fly’s intended turns, establishing a synchronised feedback loop. During the training phase, a heat punishment signal is associated with either the dot or 3D pin stimulus, smaller than an ommatidial pixel at this distance, delivered through an infrared laser. After training, without any heat punishment, the extent to which the fly has learned to avoid the tested stimulus is measured. Flies with normal binocular vision (above) exhibit significant learning scores, indicating their ability to see the dots and the pins as different objects. However, flies with monocular vision (one eye painted black, middle) or mutants that exhibit lateral photoreceptor microsaccades only in one eye (below) cannot learn this task. These results show that
Drosophila has hyperacute stereovision[
15]. Notably, this flight simulator-based setup did not allow simultaneous monitoring of photoreceptor microsaccades and whole-retina movements, both likely crucial to
Drosophila stereovision and the observed visual behaviours. Data are modified from the cited papers.
Figure 6.
Stochasticity and variations in the ommatidial photoreceptor grid structure and function combat spatiotemporal aliasing in morphodynamic information sampling and processing. (A)
Drosophila R1-R7/8 photoreceptors are differently sized and asymmetrically positioned[
13,
15], forming different numbers of synapses with interneurons[
3] (L1-L4). Moreover, R7y and R7p receptors’ colour sensitivity[115] establishes a random-like sampling matrix, consistent with anti-aliasing sampling[
13,116]. The inset shows similar randomisation for the macaque retina[117] (red, green and blue cones) (B) Demonstration of how a random sampling matrix eliminates aliasing[
13]. An original sin(x
2 + y
2) image in 0.1 resolution. Under-sampling this image with 0.2 resolution by a regular sampling matrix leads to aliasing: ghost rings appear (pink square), which the nervous system cannot differentiate from the original real image. Sampling the original image with a 0.2 resolution random matrix loses some of its fine resolution due to broadband noise, but sampling is aliasing-free. (C) In the flight simulator optomotor paradigm, a tethered head-fixed
Drosophila robustly responds to hyperacute stimuli (tested from ~0.5° to ~4° wavelengths) for different velocities (tested from 30°/s to 500°/s). However, flies show a response reversal to 45°/s rotating 6.4°-stripe panorama. In contrast, monocular flies, with one eye painted black, do not reverse their optomotor responses, indicating that the reversal response is not induced by spatial aliasing[
15]. Notably, this flight simulator-based setup did not allow for the simultaneous monitoring of photoreceptor microsaccades and whole-retina movements, both of which must contribute to the flies’ optomotor behaviour. (D) The compound eyes’ active stereo information sampling integrates body, head movements and global retina movements with local photomechanical photoreceptor microsaccades. Data are modified from the cited papers.
Figure 6.
Stochasticity and variations in the ommatidial photoreceptor grid structure and function combat spatiotemporal aliasing in morphodynamic information sampling and processing. (A)
Drosophila R1-R7/8 photoreceptors are differently sized and asymmetrically positioned[
13,
15], forming different numbers of synapses with interneurons[
3] (L1-L4). Moreover, R7y and R7p receptors’ colour sensitivity[115] establishes a random-like sampling matrix, consistent with anti-aliasing sampling[
13,116]. The inset shows similar randomisation for the macaque retina[117] (red, green and blue cones) (B) Demonstration of how a random sampling matrix eliminates aliasing[
13]. An original sin(x
2 + y
2) image in 0.1 resolution. Under-sampling this image with 0.2 resolution by a regular sampling matrix leads to aliasing: ghost rings appear (pink square), which the nervous system cannot differentiate from the original real image. Sampling the original image with a 0.2 resolution random matrix loses some of its fine resolution due to broadband noise, but sampling is aliasing-free. (C) In the flight simulator optomotor paradigm, a tethered head-fixed
Drosophila robustly responds to hyperacute stimuli (tested from ~0.5° to ~4° wavelengths) for different velocities (tested from 30°/s to 500°/s). However, flies show a response reversal to 45°/s rotating 6.4°-stripe panorama. In contrast, monocular flies, with one eye painted black, do not reverse their optomotor responses, indicating that the reversal response is not induced by spatial aliasing[
15]. Notably, this flight simulator-based setup did not allow for the simultaneous monitoring of photoreceptor microsaccades and whole-retina movements, both of which must contribute to the flies’ optomotor behaviour. (D) The compound eyes’ active stereo information sampling integrates body, head movements and global retina movements with local photomechanical photoreceptor microsaccades. Data are modified from the cited papers.
Figure 7.
Pre- and postsynaptic morphodynamic sampling adapt to optimise information allocation in neural channels. (
A) Adaptation enhances sensory information flow over time. R1–6 photoreceptor (above) and LMC voltage responses (below), as recorded intracellularly from
Drosophila compound eyes in vivo, to a repeated naturalistic stimulus pattern, NS[
45]. The recordings show how these neurons’ information allocation changes over time (for 1
st, 2
nd and 20
th s of stimulation). The LMC voltage modulation grows rapidly over time, whereas the photoreceptor output changes less, indicating that most adaptation in the phototransduction occurs within the first second. Between these traces are their probability and the joint probability density functions (“hot” colours denote high probability). Notably, the mean synaptic gain increases dynamically as presented by the shape of join probability; white lines highlight its steepening slope during repetitive NS[
45]. (
B) LMC output sensitises dynamically[
45]: its probability density flattens and widens over time (arrows; from blue to green), causing a time-dependent upwards trend in standard deviation (SD). Simultaneously, its frequency distribution changes. Because both its low- (up arrow) and high-frequency (up right) content increases while R1-6 output is less affected, the synapse allocates information more evenly within the LMC bandwidth over time. (
C) Left: Signal-to-noise ratio (SNR) of
Drosophila R1-6 photoreceptor responses to 20 Hz (red), 100 Hz (yellow), and 500 Hz (blue) saccade-like contrast bursts[
13]. SNR increases with contrast (right) and reaches its maximum value (~6,000) for 20 Hz bursts (red, left), while 100 Hz bursts (yellow) exhibit the broadest frequency range. Right: Information transfer rate comparisons between photoreceptor recordings and stochastic model simulations for saccadic light bursts and Gaussian white noise stimuli of varying bandwidths[
13]. The estimated information rates from both recorded and simulated data closely correspond across the entire range of encoding tested. This indicates that the morphodynamic refractory sampling (as performed by 30,000 microvilli) generates the most information-rich responses to saccadic burst stimulation. (
D) Adaptation to repetitive naturalistic stimulation shows phasic scale-invariance to pattern speed. 10,000 points-long naturalistic stimulus sequence (NS) was presented and repeated at different playback velocities, lasting from 20 s (0.5 kHz) to 333 ms (30kHz)[
45]. The corresponding intracellular photoreceptor (top trace) and LMC (middle trace) voltage responses are shown. The coloured sections highlight stimulus-specific playback velocities used during continuous recording. (
E) The time-normalised shapes of the photoreceptor (above) and LMC (below) responses depict similar aspects of the stimulus, regardless of the playback velocity used (ranging from 0.5 to 30 kHz)[
45]. The changes in the naturalistic stimulus speed, which follow the time-scale invariance of 1/f statistics, maintain the power within the frequency range of LMC responses relatively consistent. Consequently, LMCs can integrate similar size responses (contrast constancy) for the same stimulus pattern, irrespective of its speed[
45]. These responses are predicted to drive generation of self-similar (scalable) action potential representations of the visual stimuli in central neurons. Data are modified from the cited papers.
Figure 7.
Pre- and postsynaptic morphodynamic sampling adapt to optimise information allocation in neural channels. (
A) Adaptation enhances sensory information flow over time. R1–6 photoreceptor (above) and LMC voltage responses (below), as recorded intracellularly from
Drosophila compound eyes in vivo, to a repeated naturalistic stimulus pattern, NS[
45]. The recordings show how these neurons’ information allocation changes over time (for 1
st, 2
nd and 20
th s of stimulation). The LMC voltage modulation grows rapidly over time, whereas the photoreceptor output changes less, indicating that most adaptation in the phototransduction occurs within the first second. Between these traces are their probability and the joint probability density functions (“hot” colours denote high probability). Notably, the mean synaptic gain increases dynamically as presented by the shape of join probability; white lines highlight its steepening slope during repetitive NS[
45]. (
B) LMC output sensitises dynamically[
45]: its probability density flattens and widens over time (arrows; from blue to green), causing a time-dependent upwards trend in standard deviation (SD). Simultaneously, its frequency distribution changes. Because both its low- (up arrow) and high-frequency (up right) content increases while R1-6 output is less affected, the synapse allocates information more evenly within the LMC bandwidth over time. (
C) Left: Signal-to-noise ratio (SNR) of
Drosophila R1-6 photoreceptor responses to 20 Hz (red), 100 Hz (yellow), and 500 Hz (blue) saccade-like contrast bursts[
13]. SNR increases with contrast (right) and reaches its maximum value (~6,000) for 20 Hz bursts (red, left), while 100 Hz bursts (yellow) exhibit the broadest frequency range. Right: Information transfer rate comparisons between photoreceptor recordings and stochastic model simulations for saccadic light bursts and Gaussian white noise stimuli of varying bandwidths[
13]. The estimated information rates from both recorded and simulated data closely correspond across the entire range of encoding tested. This indicates that the morphodynamic refractory sampling (as performed by 30,000 microvilli) generates the most information-rich responses to saccadic burst stimulation. (
D) Adaptation to repetitive naturalistic stimulation shows phasic scale-invariance to pattern speed. 10,000 points-long naturalistic stimulus sequence (NS) was presented and repeated at different playback velocities, lasting from 20 s (0.5 kHz) to 333 ms (30kHz)[
45]. The corresponding intracellular photoreceptor (top trace) and LMC (middle trace) voltage responses are shown. The coloured sections highlight stimulus-specific playback velocities used during continuous recording. (
E) The time-normalised shapes of the photoreceptor (above) and LMC (below) responses depict similar aspects of the stimulus, regardless of the playback velocity used (ranging from 0.5 to 30 kHz)[
45]. The changes in the naturalistic stimulus speed, which follow the time-scale invariance of 1/f statistics, maintain the power within the frequency range of LMC responses relatively consistent. Consequently, LMCs can integrate similar size responses (contrast constancy) for the same stimulus pattern, irrespective of its speed[
45]. These responses are predicted to drive generation of self-similar (scalable) action potential representations of the visual stimuli in central neurons. Data are modified from the cited papers.
Figure 8.
Synchronised minimal delay brain activity. (
A) A
Drosophila has three electrodes inserted into its brain: right (E1) and left (E2) lobula/lobula plate optic lobes and reference (Ref). It flies in a flight simulator seeing identical scenes of black and white stripes on its left and right[
64]. When the scenes are still, the fly flies straight, and the right and left optic lobes show little activity; only a sporadic spike and the local field potentials (LFPs) are flat (E2, blue; E1, red traces). When the scenes start to sweep to the opposing directions, it takes less than 20 ms (yellow bar) for the optic lobes to respond to these visual stimuli (first spikes, and dips in LFPs). Interestingly, separate intracellular photoreceptor and large monopolar cell (LMC) recordings to 10 ms light pulse shows comparable time delays, peaking on average at 15 ms and 10 ms, respectively. Given that lobula and lobula plate neurons, which generate the observed spike and LFP patterns, are at least three synapses away from photoreceptors, the neural responses at different processing layers (retina, lamina, lobula plate) are closely synchronised, indicating minimal delays. Even though the fly brain has already received the visual information about the moving scenes, the fly makes little adjustments in its flight path, and the yaw torque remains flat. Only after minimum of 210 ms of stimulation, the fly finally chooses the left stimulus by attempting to turn left (dotted line), seen as intensifying yaw torque (downward). (
B) Brief high-intensity X-ray pulses activate
Drosophila photoreceptors[
15], causing photomechanical photoreceptor microsaccades across the eyes (characteristic retina movement). Virtually simultaneously, also other parts of the brain move, shown for lamina, Medulla and Central brain. (
C) During 2-photon imaging, L2-monopolar cell terminals can show mechanical jitter (grey noisy trace) that is synchronised with moving stimulus[
15] (vertical stripes). (
D)
Drosophila brain networks likely utilise multiple synchronised morphodynamic neural pathways to integrate a continuously adjusted, combinatorial, and distributed neural representation of a lemon, leading to its coherent and distinct object perception. Data are modified from the cited papers.
Figure 8.
Synchronised minimal delay brain activity. (
A) A
Drosophila has three electrodes inserted into its brain: right (E1) and left (E2) lobula/lobula plate optic lobes and reference (Ref). It flies in a flight simulator seeing identical scenes of black and white stripes on its left and right[
64]. When the scenes are still, the fly flies straight, and the right and left optic lobes show little activity; only a sporadic spike and the local field potentials (LFPs) are flat (E2, blue; E1, red traces). When the scenes start to sweep to the opposing directions, it takes less than 20 ms (yellow bar) for the optic lobes to respond to these visual stimuli (first spikes, and dips in LFPs). Interestingly, separate intracellular photoreceptor and large monopolar cell (LMC) recordings to 10 ms light pulse shows comparable time delays, peaking on average at 15 ms and 10 ms, respectively. Given that lobula and lobula plate neurons, which generate the observed spike and LFP patterns, are at least three synapses away from photoreceptors, the neural responses at different processing layers (retina, lamina, lobula plate) are closely synchronised, indicating minimal delays. Even though the fly brain has already received the visual information about the moving scenes, the fly makes little adjustments in its flight path, and the yaw torque remains flat. Only after minimum of 210 ms of stimulation, the fly finally chooses the left stimulus by attempting to turn left (dotted line), seen as intensifying yaw torque (downward). (
B) Brief high-intensity X-ray pulses activate
Drosophila photoreceptors[
15], causing photomechanical photoreceptor microsaccades across the eyes (characteristic retina movement). Virtually simultaneously, also other parts of the brain move, shown for lamina, Medulla and Central brain. (
C) During 2-photon imaging, L2-monopolar cell terminals can show mechanical jitter (grey noisy trace) that is synchronised with moving stimulus[
15] (vertical stripes). (
D)
Drosophila brain networks likely utilise multiple synchronised morphodynamic neural pathways to integrate a continuously adjusted, combinatorial, and distributed neural representation of a lemon, leading to its coherent and distinct object perception. Data are modified from the cited papers.
2.2. Microsaccades Are Photomechanical Adaptations in Phototransduction
Drosophila photoreceptors exhibit a distinctive toothbrush-like morphology characterised by their “bristled” light-sensitive structures known as rhabdomeres. In the outer photoreceptors (R1-6), there are approximately 30,000 bristles, called microvilli, which act as photon sampling units (
Figure 2D) [
8,
13,
15]. These microvilli collectively function as a waveguide, capturing light information across the photoreceptor’s receptive field[
14,
15]. Each microvillus compartmentalises the complete set of phototransduction cascade reactions[
61], contributing to the refractive index and waveguide properties of the rhabdomere[
77]. The phototransduction reactions within rhabdomeric microvilli of insect photoreceptors generate ultra-fast contractions of the whole rhabdomere caused by the PLC-mediated cleavage of PIP
2 headgroups (InsP3) from the microvillar membrane[
8,
61]. These photomechanics rapidly adjust the photoreceptor, enabling it to dynamically adapt its light input as the receptive field reshapes and interacts with the surrounding environment. Because photoreceptor microsaccades directly result from phototransduction reactions [
8,
13,
15,
61], they are an inevitable consequence of compound eye vision. Without microsaccades, insects with microvillar photoreceptors would be blind [
8,
13,
15,
61].
Insects possess an impressively rapid vision, operating approximately 3 to 15 times faster than our own. This remarkable ability stems from the microvilli’s swift conversion of captured photons into brief unitary responses (
Figure 2D; also known as quantum bumps[
61]) and their ability to generate photomechanical micromovements[
8,
13] (
Figure 2C). Moreover, the size and speed of microsaccades adapt to the microvilli population’s refractory photon sampling dynamics[
13,
76] (
Figure 2D). As light intensity increases, both the quantum efficiency and duration of photoreceptors’ quantum bumps decrease[
76,
78], resulting in more transient microsaccades[
13,
15]. These adaptations extend the dynamic range of vision[
76,
79] and enhance the detection of environmental contrast changes[
13,
80], making visual objects highly noticeable under various lighting conditions. Consequently,
Drosophila can perceive moving objects across a wide range of velocities and light intensities, surpassing the resolution limits of the static eye’s pixelation by 4-10 times (
Figure 2E; the average inter-ommatidial angle, φ ≈ 5°)[
13,
15].
Morphodynamic adaptations involving photoreceptor microvilli play a crucial role in insect vision by enabling rapid and efficient visual information processing. These adaptations lead to contrast-normalised (
Figure 2D) and more phasic photoreceptor responses, achieved through significantly reduced integration time [
13,
80,
81]. Evolution further refines these dynamics to match species-specific visual needs (
Figure 2E). For example, honeybee microsaccades are smaller than those of
Drosophila[
14], corresponding to the positioning of honeybee photoreceptors farther away from the ommatidium lenses. Consequently, reducing the receptive field size and interommatidial angles in honeybees is likely an adaptation that allows optimal image resolution during scanning[
14]. Similarly, fast-flying flies such as houseflies and blowflies, characterised by a higher density of ommatidia in their eyes, are expected to exhibit smaller and faster photoreceptor microsaccades compared to slower-flying
Drosophila with fewer and less densely packed ommatidia[
15]. This adaptation enables the fast-flying flies to capture visual information with higher velocity[
76,
80,
82,
83] and resolution, albeit at a higher metabolic cost[
80].
2.3. Microsaccades Maximise Information during Saccadic Behaviours
Photoreceptors’ microsaccadic sampling likely evolved to align with animals’ saccadic behaviours, maximising visual information capture[
13,
15]. Saccades are utilised by insects and humans to explore their environment (
Figure 1I–L), followed by fixation periods where the gaze remains relatively still[
34]. Previously, it was believed that detailed information was only sampled during fixation, as photoreceptors were thought to have slow integration times, causing image blurring during saccades[
34]. However, fixation intervals can lead to perceptual fading through powerful adaptation, reducing visual information and potentially limiting perception to average light levels [
13,
84,
85]. Therefore, to maximise information capture, fixation durations and saccade speeds should dynamically adapt to the statistical properties of the natural environment[
13]. This sampling strategy would enable animals to efficiently adjust their behavioural speed and movement patterns in diverse environments, optimising vision - for example, moving slowly in darkness and faster in daylight[
13].
To investigate this theory, researchers studied the body yaw velocities of walking fruit flies[
67] to sample light intensity information from natural images[
13] (
Figure 3). They found that saccadic viewing of these images improved the photoreceptors’ information capture compared to linear or shuffled velocity walks[
13]. This improvement was attributed to saccadic viewing generating bursty high-contrast stimulation, maximising the photoreceptors’ ability to gather information. Specifically, the photomechanical and refractory phototransduction reactions of
Drosophila R1-6 photoreceptors, associated with motion vision[
86], were found to be finely tuned to saccadic behaviour for sampling quantal light information, enabling them to capture 2-to-4-times more information in a given time compared to previous estimates[
13,
78].
Further analysis, utilising multiscale biophysical modelling[
81], investigated the stochastic refractory photon sampling by 30,000 microvilli[
13]. For readers interested in more details, Text Box 2 graphically illustrates the basic principles of stochastic quantal refractory sampling. The findings revealed that the improved information capture during saccadic viewing can be attributed to the interspersed fixation intervals[
13,
79]. When fixating on darker objects, which alleviates microvilli refractoriness, photoreceptors can sample more information from transient light changes, capturing larger photon rate variations[
13]. The combined effect of photomechanical photoreceptor movements and refractory sampling worked synergistically to enhance spatial acuity, reduce motion blur during saccades, facilitate adaptation during gaze fixation, and emphasise instances when visual objects crossed a photoreceptor’s receptive field. Consequently, the encoding of high-resolution spatial information was achieved through the temporal mechanisms induced by physical motion[
13].
Text Box 2. Visualising Refractory Quantal Computations. By utilising powerful multi-scale morphodynamic neural models[13,15,76], we can predict and analyse the generation and integration of voltage responses during morphodynamic quantal refractory sampling and compare these simulations to actual intracellular recordings for similar stimulation[
13,
15,
76]. This approach, combined with information-theoretical analyses[
76,
80,88,89], allows us to explain how phasic response waveforms arise from ultrafast movements and estimate the signal-to-noise ratio and information transfer rate of the neural responses. Importantly, these methods are applicable for studying the morphodynamic functions of any neural circuit. To illustrate the analytic power of this approach, we present a simple example: an intracellular recording (whole-cell voltage response) of dark-adapted
Drosophila photoreceptors (
C) to a bright light pulse. See also Figure 2D which shows morphodynamic simulations of how a photoreceptor responds to two dots crossing its receptive field from east to west and west to east directions.
Text Box 2. Visualising Refractory Quantal Computations. By utilising powerful multi-scale morphodynamic neural models[13,15,76], we can predict and analyse the generation and integration of voltage responses during morphodynamic quantal refractory sampling and compare these simulations to actual intracellular recordings for similar stimulation[
13,
15,
76]. This approach, combined with information-theoretical analyses[
76,
80,88,89], allows us to explain how phasic response waveforms arise from ultrafast movements and estimate the signal-to-noise ratio and information transfer rate of the neural responses. Importantly, these methods are applicable for studying the morphodynamic functions of any neural circuit. To illustrate the analytic power of this approach, we present a simple example: an intracellular recording (whole-cell voltage response) of dark-adapted
Drosophila photoreceptors (
C) to a bright light pulse. See also Figure 2D which shows morphodynamic simulations of how a photoreceptor responds to two dots crossing its receptive field from east to west and west to east directions.
An insect photoreceptor’s sampling units – e.g., 30,000 microvilli in a fruit fly or 90,000 in a blowfly R1-6 - count photons as variable samples (quantum bumps) and sum these up into a macroscopic voltage response, generating a reliable estimate of the encountered light stimulus. For clarity, visualise the light pulse as a consistent flow of photons, or golden balls, over time (A). The quantum bumps that the photons elicit in individual microvilli can be thought of as silver coins of various sizes (B). The photoreceptor persistently counts these “coins” produced by its microvilli, thus generating a dynamically changing macroscopic response (C, depicted as a blue trace). These basic counting rules[90] shape the photoreceptor response:
Each microvillus can produce only one quantum bump at a time [
76,91,92,93].
After producing a quantum bump, a microvillus becomes refractory for up to 300 ms (in Drosophila R1-6 photoreceptors at 25°C) and cannot respond to other photons[91,94,95].
Quantum bumps from all microvilli sum up the macroscopic response[
76,91,92,93,96].
Microvilli availability sets a photoreceptor’s maximum sample rate (quantum bump production rate), adapting its macroscopic response to a light stimulus[
76,93].
Global Ca
2+ accumulation and membrane voltage affect samples of all microvilli. These global feedbacks strengthen with brightening light to reduce the size and duration of quantum bumps, adapting the macroscopic response[
78,88,97,98].
Adaptation in macroscopic response (
C) to continuous light (
A) is mostly caused by a reduction in the number and size of quantum bumps over time (
B). When the stimulus starts, a large portion of the microvilli is simultaneously activated (
A i and
B i), but they subsequently enter a refractory state (
A ii and
B ii). This means that a smaller fraction of microvilli can respond to the following photons in the stimulus until more microvilli become available again. As a result, the number of activated microvilli initially peaks and then rapidly declines, eventually settling into a steady state (
A iii and
B iii) as the balance between photon arrivals and refractory periods is achieved. If all quantum bumps were identical, the macroscopic current would simply reflect the number of activated microvilli based on the photon rate, resulting in a steady-state response. Light-induced current also exhibits a decaying trend towards lower plateau levels. This is because quantum bumps adapt to brighter backgrounds (
A iii and
B iii), becoming smaller and briefer[
78,88]. This adaptation is caused by global negative feedback, Ca
2+-dependent inhibition of microvillar phototransduction reactions[97,98,99,100,101]. Additionally, the concurrent increase in membrane voltage compresses responses by reducing the electromotive force for the light-induced current across all microvilli[
76]. Together, these adaptive dynamics enhance phasic photoreceptor responses, similar to encoding phase congruency[102].
The signal-to-noise ratio and rate of information transfer increase with the average sampling rate, which is the average number of samples per unit time. Thus, the more samples that make up the macroscopic response to a given light pattern, the higher its information transfer rate. However, with more photons being counted by a photoreceptor at brightening stimulation, information about saccadic light patterns of natural scenes in its responses first increases and then approaches a constant rate. This is because:
(a) When more microvilli are in a refractory state, more photons fail to generate quantum bumps. As quantum efficiency drops, the equilibrium between used and available microvilli approaches a constant (maximum) quantum bump production rate (sample rate). This process effectively performs division, scaling logarithmic changes in photon rates into macroscopic voltage responses with consistent size and waveforms, thereby maintaining contrast constancy[
76,
80,90].
(b) Once global Ca2+ and voltage feedbacks saturate, they cannot make quantum bumps any smaller and briefer with increasing brightness.
(c) After initial acceleration from the dark-adapted state, quantum bump latency distribution remains practically invariable in different light-adaptation states[88].
Therefore, when sample rate modulation (a) and sample integration dynamics (b and c) of the macroscopic voltage responses settle (at intensities >10
5 photons/s in
Drosophila R1-6 photoreceptors, allocation of visual information in the photoreceptor’s amplitude and frequency range becomes nearly invariable[
76,
80,103]. Correspondingly, stochastic simulation closely predicts measured responses and rates of information transfer[
76,
79,
80]. Notably, when the microvilli usage reaches a midpoint (~50 % level), the information rate encoded by the macroscopic responses to natural light intensity time series saturates[
76]. This is presumably because sample rate modulation to light increments and decrements – which in the macroscopic response code for the number of different stimulus patterns[89] - saturate. Quantum bump size, if invariable, does not affect the information transfer rate – as long as the quantum bumps are briefer than the stimulus changes they encode. Thus, like any other filter, a fixed bump waveform affects signal and noise equally (Data Processing theorem[
76,89]). But varying quantum bump size adds noise; when this variation is adaptive (memory-based), less noise is added[
76,89].
In summary, insect photoreceptors count photons through microvilli, integrate the responses, and adapt their macroscopic response based on the basic counting rules and global feedback mechanisms. The information transfer rate increases with the average sampling rate but eventually reaches a constant rate as the brightness of the stimulus increases. The size of the quantum bumps affects noise levels, with adaptive variation reducing noise.
These discoveries underscore the crucial link between an animal’s adaptation in utilising movements across different scales, ranging from nanoscale molecular dynamics to microscopic brain morphodynamics, to maximise visual information capture and acuity[
13]. The new understanding from the
Drosophila studies is that contrary to popular assumptions, neither saccades[
52] nor fixations[
84] hinder the vision. Instead, they work together to enhance visual perception, highlighting the complementary nature of these active sampling movement patterns[
13].
2.4. Left and Right Eyes’ Mirror-Symmetric Microsaccades Phase-Enhance Moving Objects
When
Drosophila encounters moving objects in natural environments, its left and right eye photoreceptor pairs generate microsaccadic movements that synchronise their receptive field scanning in opposite directions (
Figure 4)[
13,
15]. To quantitatively analyse these morphodynamics, researchers utilised a custom-designed high-speed microscope system[
14], tailored explicitly for recording photoreceptor movements within insect compound eyes; an early prototype of this instrument is shown in
Figure 2A, while Video 1 demonstrates these experiments. Using infrared illumination, which flies cannot see[
14,
15,104,105], the positions and orientations of photoreceptors in both eyes were measured, revealing mirror-symmetric angular orientations between the eyes and symmetry within each eye (
Figure 4A). It was discovered that a single point in space within the frontal region, where receptive fields overlap (
Figure 4B), is detected by at least 16 photoreceptors, eight in each[
14,
15]. This highly ordered mirror-symmetric rhabdomere organisation, leading to massively over-complete tiling of the eyes’ binocular visual fields[
15] (
Figure 4C), challenges the historical belief that small insect eyes, such as those of
Drosophila, are optically too coarse and closely positioned to support stereovision[
52].
By selectively stimulating the rhabdomeres with targeted light flashes, researchers determined the specific photomechanical contraction directions for each eye’s location (
Figure 4D). Analysis of the resulting microsaccades enabled the construction of a 3D-vector map encompassing the frontal and peripheral areas of the eyes. These microsaccades exhibited mirror symmetry between the eyes and aligned with the rotation axis of the R1-R2-R3 photoreceptor of each ommatidium (
Figure 4D, left), indicating that the photoreceptors’ movement directions were predetermined during development (
Figure 4A)[
14,
15]. Strikingly, the 3D-vector map representing the movements of the corresponding photoreceptor receptive fields (
Figure 4D) coincides with the optic flow-field generated by the fly’s forward thrust (
Figure 4E)[
14,
15]. This alignment provides microsaccade-enhanced detection and resolution of moving objects (cf.
Figure 4C) across the extensive visual fields of the eyes (approximately 360°), suggesting an evolutionary optimisation of fly vision for this intriguing capability.
The microsaccadic receptive field movements comprise a fast phase (
Figure 4D, left) aligned with the flow-field direction during light-on (
Figure 4D, middle), followed by a slower phase in the opposite direction during light-off (
Figure 4D, right). When a fly is in forward flight with an upright head (
Figure 5E, left and middle), the fast and slow phases reach equilibrium (
Figure 4E, right). The fast phase represents “ground-flow,” while the slower phase represents “sky-flow.” In the presence of real-world structures, locomotion enhances sampling through a push-pull mechanism. Photoreceptors transition between fast and slow phases, thereby collectively improving neural resolution over time[
15] (
Figure 2C). Fast microsaccades are expected to aid in resolving intricate visual clutter, whereas slower microsaccades enhance the perception of the surrounding landscape and sky[
15]. Moreover, this eye-location-dependent orientation-tuned bidirectional visual object enhancement makes any moving object deviating from the prevailing self-motion-induced optic flow field stand out. Insect brains likely utilise the resulting phasic neural image contrast differences to detect or track predator movements or conspecifics across the eyes’ visual fields. For example, this mechanism could help a honeybee drone spot and track the queen amidst a competing drone swarm[107], enabling efficient approach and social interaction.
Rotation (yaw) (
Figure 4F, left and middle) further enhances binocular contrasts (
Figure 4F, right), with one eye’s phases synchronised with field rotation while the other eye’s phases exhibit the reverse pattern[
15]. Many insects, including bees and wasps, engage in elaborately patterned learning or homing flights, involving fast saccadic turns and bursty repetitive wave-like scanning motion when leaving their nest or food sources[108,109] (
Figure 4G). Given the mirror-symmetricity and ultrafast photoreceptor microsaccades of bee eyes[
14], these flight patterns are expected to drive enhanced binocular detection of behaviourally relevant objects, landmarks, and patterns, utilising the phasic differences in microsaccadic visual information sampling between the two eyes[
13,
15]. Thus, learning flight behaviours might make effective use of optic-flow-tuned and morphodynamically enhanced binocular vision, enabling insects to navigate and return to their desired locations successfully.
2.5. Mirror-Symmetric Microsaccades Enable Hyperacute Stereovision
Crucially,
Drosophila uses mirror-symmetric microsaccades to sample the three-dimensional visual world, enabling the extraction of depth information (
Figure 5). This process entails comparing the resulting morphodynamically sampled neural images from its left and right eye photoreceptors[
15]. The disparities in x- and y-coordinates between corresponding “pixels” provide insights into scene depth. In response to light intensity changes, the left and right eye photoreceptors contract mirror-symmetrically, narrowing and sliding their receptive fields in opposing directions, thus shaping neural responses (
Figure 5A; also see
Figure 2C)[
13,
15]. By cross-correlating these photomechanical responses between neighbouring ommatidia, the
Drosophila brain is predicted to achieve a reliable stereovision range spanning from less than 1 mm to approximately 150 mm[
15]. The crucial aspect lies in utilising the responses’ phase differences as temporal cues for perceiving 3D space (
Figure 5A,B). Furthermore, researchers assessed if a static
Drosophila eye model with immobile photoreceptors could discern depth[
15]. These calculations indicate that the lack of scanning activity by the immobile photoreceptors and the small distance between the eye (
Figure 5A, k = 440 μm) would only enable a significantly reduced depth perception range, underlining the physical and evolutionary advantages of moving photoreceptors in depth perception.
Furthermore, optical calculations using the Fourier beam propagation[
15,110] - which models in reverse how light beams pass through the photoreceptor rhabdomeres and the ommatidium lens into the visual space - have confirmed and expanded upon Pick’s earlier and often overlooked discovery[
63]. This analysis reveals that the receptive fields of R1-6 photoreceptors from neighbouring fly ommatidia, which feed information to the first visual interneurons (Large Monopolar Cells, LMCs), do not overlap perfectly. Instead, due to variations in the sizes of R1-6 rhabdomeres, their distances from the ommatidial centre, and the non-spherical shape of the eye, their receptive fields tile a small area in the visual space over-completely in neural superposition[
15,
63] (
Figure 5B; see also
Figure 4C). In living flies, this situation becomes more complex and interesting as these receptive fields move and narrow independently, as illustrated through computer simulations in Video 2, following the morphodynamic rules of their photoreceptor microsaccades[
14,
15] (
Figure 5B and Text Box 1A). Consequently, this coordinated morphodynamic sampling motion is reflected in the orientation-sensitive hyperacute LMC responses, as observed in high-speed calcium imaging of L2 monopolar cell axon terminals[
15] (
Figure 5B).
Behavioural experiments in a flight simulator verified that
Drosophila possesses hyperacute stereovision[
15] (
Figure 5C). Tethered head-fixed flies were presented with 1-4° 3D and 2D objects, smaller than their eyes’ average interommatidial angle (cf.
Figure 2E). Notably, the flies exhibited a preference for fixating on the minute 3D objects, providing support for the new morphodynamic sampling theory of hyperacute stereovision.
In subsequent learning experiments, the flies underwent training to avoid specific stimuli, successfully showing the ability to discriminate between small (<< 4°) equal-contrast 3D and 2D objects. Interestingly, because of their immobilised heads, flies could not rely on motion parallax signals during learning, meaning the discrimination relied solely on the eyes’ image disparity signals. Flies with one eye painted over failed to learn the stimuli. Moreover, it was discovered that rescuing R1-6 or R7/R8 photoreceptors in blind
norpAP24 mutants made their microsaccades’ lateral (sideways) component more vulnerable to mechanical stress or developmental issues, with ∼10% of these mutants displaying microsaccades only monocularly[
15]. However, both eyes showed a characteristic electroretinogram response, indicating intact phototransduction and axial microsaccade movement. Flies with normal lateral microsaccades learned to distinguish hyperacute 3D pins from 2D dots and the standard 2D T vs. Ʇ patterns, though less effectively than wild-type flies, showing that R1-6 input suffices for hyperacute stereovision but that R7/R8s also play a role. Conversely, mutants with monocular sideways microsaccades failed to learn 3D objects or 2D patterns, indicating that misaligned binocular sampling impairs 3D perception and learning. R7/R8 rescued
norpAP24 and
ninaE8 mutants confirmed that inner photoreceptors contribute to hyperacute stereopsis.
These results firmly establish the significance of binocular mirror-symmetric photoreceptor microsaccades in sampling 3D information and that both R1-6 (associated with motion vision[
86]) and R7/R8 (associated with colour vision[112]) photoreceptor classes contribute to hyperacute stereopsis. The findings provide compelling evidence that mirror-symmetric microsaccadic sampling, as a form of ultrafast neural morphodynamics, is necessary for hyperacute stereovision in
Drosophila[
15].
2.6. Microsaccade Variability Combats Aliasing
The heterogeneous nature of the fly’s retinal sampling matrix - characterised by varying rhabdomere sizes[
13], random distributions of visual pigments[113], variations in photoreceptor synapse numbers[
3] (
Figure 6A), the stochastic integration of single photon responses[
76,
78,88] (quantum bumps)[
13,
79,
81] and stochastic variability in microsaccade waveforms[
13,
14,
15] - eliminates spatiotemporal aliasing[
13,
15] (
Figure 6B), enabling reliable transmission of visual information. This reliable encoding from variable samples aligns with the earlier examples (
Figure 2C–D,
Figure 3D and
Figure 5B) and touches on Francis Galton’s idea of
vox populi[114]: “The mean of variable samples, reported independently by honest observers, provides the best estimate of the witnessed event”[90]. Consequently, the morphodynamic information sampling theory[
13,
15] challenges previous assumptions of
static compound eyes[
52], which suggested that the ommatidial grid of immobile photoreceptors structurally limits spatial resolution, rendering the eyes susceptible to undersampling the visual world and prone to aliasing[
52].
Supporting the new morphodynamic theory[
13,
15], tethered head-fixed
Drosophila exhibit robust optomotor behaviour in a flight simulator system (
Figure 6C). The flies generated yaw torque responses, represented by the blue traces, indicating their intention to follow the left or right turns of the stripe panorama. These responses are believed to be a manifestation of an innate visuomotor reflex aimed at minimising retinal image slippage[
52,118]. Consistent with
Drosophila’s hyperacute ability to differentiate small 3D and 2D objects[
15], as shown earlier in
Figure 5C, the tested flies reliably responded to rotating panoramic black-and-white stripe scenes with hyperacute resolution, tested down to 0.5° resolution[
13,
15]. This resolution is about ten times finer than the eyes’ average interommatidial angle (cf.
Figure 2E), significantly surpassing the explanatory power of the traditional
static compound eye theory[
52], which predicts 4°-5° maximum resolvability.
However, when exposed to slowly rotating 6.4-10° black-and-white stripe waveforms, a head-fixed tethered
Drosophila displays reversals in its optomotor flight behaviour[
15] (
Figure 6C). Previously, this optomotor reversal was thought to result from the static ommatidial grid spatially aliasing the sampled panoramic stripe pattern due to the stimulus wavelength being approximately twice the eyes’ average interommatidial angle. Upon further analysis, the previous interpretation of these reversals as a sign of aliasing[
35,
52] is contested. Optomotor reversals primarily occur at 40-60°/s stimulus velocities, matching the speed of the left and right eyes’ mirror-symmetric photoreceptor microsaccades[
15] (
Figure 6C; cf.
Figure 2C). As a result, one eye’s moving photoreceptors are more likely to be locked onto the rotating scene than those in the other eye, which move against the stimulus rotation. This discrepancy creates an imbalance that the fly’s brain may interpret as the stimulus rotating in the opposite direction[
15].
Notably, the optomotor behaviour returns to normal when the tested fly has monocular vision (with one eye covered) and during faster rotations[
15] or finer stripe pattern waveforms[
13,
15] (
Figure 6C). Therefore, the abnormal optomotor reversal, which arises under somewhat abnormal and specific stimulus conditions when tested with head-fixed and position-constrained flies, must reflect high-order processing of binocular information and cannot be attributed to spatial sampling aliasing that is velocity and left-vs-right eye independent[
15].
2.7. Multiple Layers of Active Sampling vs Simple Motion Detection Models
In addition to photoreceptor microsaccades, insects possess intraocular muscles capable of orchestrating coordinated oscillations of the entire photoreceptor array, encompassing the entire retina[
13,
15,
35,119] (
Figure 6D). This global motion has been proposed as a means to achieve super-resolution[120,121], but not for stereopsis. While the muscle movements alter the light patterns reaching the eyes, leading to the occurrence of photoreceptor microsaccades, it is the combination of local microsaccades and global retina movements, which include any body and head movements[
33,
55,109,122] (
Figure 6D), that collectively govern the active sampling of stereoscopic information by the eyes[
13,
14,
15].
The
Drosophila brain effectively integrates depth and motion computations using mirror-symmetrically moving image disparity signals from its binocular vision channels[
15]. During goal-oriented visual behaviours, coordinated muscle-induced vergence movements of the left and right retinae[
35], a phenomenon also observed in larger flies walking on a trackball[
36,
37], likely further extend the stereo range by drawing bordering photoreceptors into the eye’s binocular region (cf. Text Box 1B iv). Interestingly, in fully immobilised
Drosophila, which rarely shows these retinal movements[
13,
14,
15], the photoreceptor microsaccade amplitudes characteristically fluctuate more during repeated light stimuli than the corresponding intracellular voltage responses[
13,
15]. This suggests that, in addition to retinal movements, the fly brain might exert top-down control over retinal muscle tone and tension, thereby modulating the lateral range of photomechanical microsaccades through retinal strain adjustments. This interaction between retinal muscles and photoreceptor microsaccades could ultimately facilitate attentive accommodation, allowing the fly to precisely focus its hyperacute gaze on specific visual targets, analogous to how vertebrate lens eyes use ciliary muscles to fine-tune focus[123].
Conversely, by maximally tensing or relaxing the retinal muscles - and thus the retinae - a fly might be able to fully suppress the microsaccades’ lateral movement, as suggested by studies involving optogenetic activation or genetic inactivation of retinal motor neurons[
35]. While photomechanical microsaccades are robust and occur without muscle involvement, as observed in dissociated photoreceptors in a Petri dish[
13], their lateral movement range can be physically constrained by increasing the stiffness of the surrounding medium. For example, in
spam-mutant eyes, where the open rhabdom of R1-8 photoreceptors reverts to an ancestral fused rhabdom state[124,125], microsaccade kinematics are similar to those in wild-type photoreceptors, but their displacement range is reduced due to the increased structural stiffness of the fused rhabdom[
14]. If the maximally tensing or relaxing of the retinal muscles were linked to top-down synaptic inhibition of photoreceptor signals - potentially mediated by centrifugal GABAergic C2/C3 fibres[126] from the brain that innervate the photoreceptors[
3] - this centralised visual information suppression (“closing the eyes”) could serve to minimise environmental interference and the eyes’ energy consumption during sleep.
These findings and new ideas about fast and complex motion-based interactions in visual information sampling and processing challenge the traditional view that insect brains rely on low-resolution input from static eyes for high-order computations. For instance, the motion detection ideals of reduced input-output systems[
86,127,128], such as Hassenstein-Reichardt[129] and Barlow-Levick[130] models, require updates to incorporate ultrafast morphodynamics[
13]
-[
15], retinal muscle movements[
35] and state-dependent synaptic processing[
64,
70,
73,
74,131]. The updates are crucial as these processes actively shape neural responses, perception, and behaviours[
64], providing essential ingredients for hyperacute attentive 3D vision[
13,
15,
48,
50] and intrinsic decision-making[
13,
15,132,133,134] that occur in a closed-loop interaction with the changing environment.
Accumulating evidence, consistent with the idea that brains reduce uncertainty by synchronously integrating multisensory information[135], further suggests that object colour and shape information partially streams through the same channels previously thought to be solely for motion information[105]. Consequently, individual neurons within these channels should engage in multiple parallel processing tasks[105], adapting in a phasic and goal-oriented manner. These emerging concepts challenge oversimplified input-output models of insect vision, highlighting the importance of complex interactions between local ultrafast neural morphodynamics and global active vision strategies in perception and behaviour.