Seeing the future: Predictive control in neural models of ocular accommodation by Read JCA, Kaspiris-Rousellis C, Wood TS, Wu B, Vlaskamp BNS, Schor CM, ReadKaspirisRousellisWoodWuVlaskampSchor2022.pdf (5.4 MiB) - Ocular accommodation is the process of adjusting the eye's crystalline lens so as to bring the retinal image into sharp focus. The major stimulus to accommodation is therefore retinal defocus, and in essence, the job of accommodative control is to send a signal to the ciliary muscle which will minimize the magnitude of defocus. In this article, we first provide a tutorial introduction to control theory to aid vision scientists without this background. We then present a unified model of accommodative control that explains properties of the accommodative response for a wide range of accommodative stimuli. Following previous work, we conclude that most aspects of accommodation are well explained by dual integral control, with a “fast” or “phasic” integrator enabling response to rapid changes in demand, which hands over control to a “slow” or “tonic” integrator which maintains the response to steady demand. Control is complicated by the sensorimotor latencies within the system, which delay both information about defocus and the accommodation changes made in response, and by the sluggish response of the motor plant. These can be overcome by incorporating a Smith predictor, whereby the system predicts the delayed sensory consequences of its own motor actions. For the first time, we show that critically-damped dual integral control with a Smith predictor accounts for adaptation effects as well as for the gain and phase for sinusoidal oscillations in demand. In addition, we propose a novel proportional-control signal to account for the power spectrum of accommodative microfluctuations during steady fixation, which may be important in hunting for optimal focus, and for the nonlinear resonance observed for low-amplitude, high-frequency input.
CNV-Net: Segmentation, Classification and Activity Score Measurement of Choroidal Neovascularization (CNV) Using Optical Coherence Tomography Angiography (OCTA) by Vali M, Nazari B, Sadri S, Khalili Pour E, Riazi-Esfahani H, Faghihi H, Ebrahimiadib N, Azizkhani M, Innes W, Steel DH, Hurlbert AC, Read JCA, Kafieh R , ValiNazariSadriEAReadKafieh2023.pdf (4.7 MiB) - This paper aims to present an artificial intelligence-based algorithm for the automated segmentation of Choroidal Neovascularization (CNV) areas and to identify the presence or absence of CNV activity criteria (branching, peripheral arcade, dark halo, shape, loop and anastomoses) in OCTA images. Methods: This retrospective and cross-sectional study includes 130 OCTA images from 101 patients with treatment-naïve CNV. At baseline, OCTA volumes of 6 × 6 mm2 were obtained to develop an AI-based algorithm to evaluate the CNV activity based on five activity criteria, including tiny branching vessels, anastomoses and loops, peripheral arcades, and perilesional hypointense halos. The proposed algorithm comprises two steps. The first block includes the pre-processing and segmentation of CNVs in OCTA images using a modified U-Net network. The second block consists of five binary classification networks, each implemented with various models from scratch, and using transfer learning from pre-trained networks. Results: The proposed segmentation network yielded an averaged Dice coefficient of 0.86. The individual classifiers corresponding to the five activity criteria (branch, peripheral arcade, dark halo, shape, loop, and anastomoses) showed accuracies of 0.84, 0.81, 0.86, 0.85, and 0.82, respectively. The AI-based algorithm potentially allows the reliable detection and segmentation of CNV from OCTA alone, without the need for imaging with contrast agents. The evaluation of the activity criteria in CNV lesions obtains acceptable results, and this algorithm could enable the objective, repeatable assessment of CNV features.
(This article belongs to the Special Issue Advances in Optical Coherence Tomography Angiography)
Peripheral Flicker Fusion at High Luminance: Beyond the Ferry–Porter Law by Fernandez-Alonso M, Innes W, Read JCA, FernandezAlonsoInnesRead2023.pdf (0.9 MiB) - The relationship between luminous intensity and the maximum frequency of flicker that can be detected defines the limits of the temporal-resolving ability of the human visual system, and characterizing it has important theoretical and practical applications; particularly for determining the optimal refresh rate for visual displays that would avoid the visibility of flicker and other temporal artifacts. Previous research has shown that this relationship is best described by the Ferry–Porter law, which states that critical flicker fusion (CFF) increases as a linear function of log retinal illuminance. The existing experimental data showed that this law holds for a wide range of stimuli and up to 10,000 Trolands; however, beyond this, it was not clear if the CFF continued to increase linearly or if the function saturated. Our aim was to extend the experimental data available to higher light intensities than previously reported in the literature. For this, we measured the peripheral CFF at a range of illuminances over six orders of magnitude. Our results showed that for up to 104 Trolands, the data conformed to the Ferry–Porter law with a similar slope, as previously established for this eccentricity; however, at higher intensities, the CFF function flattens and saturates at ~90 Hz for a target size of 5.7 degrees, and at ~100 Hz for a target of 10 degrees of angular size. These experimental results could prove valuable for the design of brighter visual displays and illumination sources that are temporally modulated.
New Approaches to 3D Vision by Linton P, Morgan MJ, Read JCA, Vishwanath D, Creem-Regehr SH, Domini F, LintonMorganReadVishwanathCreemRegehrDomini2022.pdf (3.2 MiB) - New approaches to 3D vision are enabling new advances in artificial intelligence and autonomous vehicles, a better understanding of how animals navigate the 3D world, and new insights into human perception in virtual and augmented reality. Whilst traditional approaches to 3D vision in computer vision (SLAM: simultaneous localization and mapping), animal navigation (cognitive maps), and human vision (optimal cue integration) start from the assumption that the aim of 3D vision is to provide an accurate 3D model of the world, the new approaches to 3D vision explored in this issue challenge this assumption. Instead, they investigate the possibility that computer vision, animal navigation, and human vision can rely on partial or distorted models or no model at all. This issue also highlights the implications for artificial intelligence, autonomous vehicles, human perception in virtual and augmented reality, and the treatment of visual disorders, all of which are explored by individual articles.
This article is part of a discussion meeting issue ‘New approaches to 3D vision’.
Stereopsis without correspondence by Read JCA, Read2022.pdf (1.9 MiB) - Stereopsis has traditionally been considered a complex visual ability, restricted
to large-brained animals. The discovery in the 1980s that insects, too, have
stereopsis, therefore, challenged theories of stereopsis. How can such simple
brains see in three dimensions? A likely answer is that insect stereopsis has
evolved to produce simple behaviour, such as orienting towards the closer
of two objects or triggering a strike when prey comes within range. Scientific
thinking about stereopsis has been unduly anthropomorphic, for example
assuming that stereopsis must require binocular fusion or a solution of the
stereo correspondence problem. In fact, useful behaviour can be produced
with very basic stereoscopic algorithms which make no attempt to achieve
fusion or correspondence, or to produce even a coarse map of depth across
the visual field. This may explain why some aspects of insect stereopsis
seem poorly designed from an engineering point of view: for example,
paying no attention to whether interocular contrast or velocities match. Such
algorithms demonstrably work well enough in practice for their species, and
may prove useful in particular autonomous applications.
This article is part of a discussion meeting issue ‘New approaches to 3D
vision’.
Predicting attitudes towards screening for neurodegenerative diseases using OCT and artificial intelligence: Findings from a literature review by Nichol BAB, Hurlbert AC, Read JCA, NicholHurlbertRead2022.pdf (0.2 MiB)
Synthetic OCT Data Generation to Enhance the Performance of Diagnostic Models for Neurodegenerative Diseases by Danesh H, Steel DH, Hogg J, Ashtari F, Innes WF, Bacardit, J, Hurlbert A, Read JCA, Kafieh R, DaneshMaghooliDehghaniKafieh2021.pdf (3.1 MiB) - Purpose: Optical coherence tomography (OCT) has recently emerged as a source for powerful biomarkers in neurodegenerative diseases such as multiple sclerosis (MS) and neuromyelitis optica (NMO). The application of machine learning techniques to the analysis of OCT data has enabled automatic extraction of information with potential to aid the timely diagnosis of neurodegenerative diseases. These algorithms require large amounts of labeled data, but few such OCT data sets are available now.
Methods: To address this challenge, here we propose a synthetic data generation method yielding a tailored augmentation of three-dimensional (3D) OCT data and preserving differences between control and disease data. A 3D active shape model is used to produce synthetic retinal layer boundaries, simulating data from healthy controls (HCs) as well as from patients with MS or NMO.
Results: To evaluate the generated data, retinal thickness maps are extracted and evaluated under a broad range of quality metrics. The results show that the proposed model can generate realistic-appearing synthetic maps. Quantitatively, the image histograms of the synthetic thickness maps agree with the real thickness maps, and the cross-correlations between synthetic and real maps are also high. Finally, we use the generated data as an augmentation technique to train stronger diagnostic models than those using only the real data.
Conclusions: This approach provides valuable data augmentation, which can help overcome key bottlenecks of limited data.
Translational Relevance: By addressing the challenge posed by limited data, the proposed method helps apply machine learning methods to diagnose neurodegenerative diseases from retinal imaging.
A computational model of stereoscopic prey capture in praying mantises by O’Keeffe J, Yap SH, Llamas-Cornejo I, Nityananda V, Read JCA, OKeeffeYapLlamasCornejoNityanandaRead2022.pdf (4.2 MiB) - We present a simple model which can account for the stereoscopic sensitivity of praying mantis predatory strikes. The model consists of a single “disparity sensor”: a binocular neuron sensitive to stereoscopic disparity and thus to distance from the animal. The model is based closely on the known behavioural and neurophysiological properties of mantis stereopsis. The monocular inputs to the neuron reflect temporal change and are insensitive to contrast sign, making the sensor insensitive to interocular correlation. The monocular receptive fields have a excitatory centre and inhibitory surround, making them tuned to size. The disparity sensor combines inputs from the two eyes linearly, applies a threshold and then an exponent output nonlinearity. The activity of the sensor represents the model mantis’s instantaneous probability of striking. We integrate this over the stimulus duration to obtain the expected number of strikes in response to moving targets with different stereoscopic disparity, size and vertical disparity. We optimised the parameters of the model so as to bring its predictions into agreement with our empirical data on mean strike rate as a function of stimulus size and disparity. The model proves capable of reproducing the relatively broad tuning to size and narrow tuning to stereoscopic disparity seen in mantis striking behaviour. Although the model has only a single centre-surround receptive field in each eye, it displays qualitatively the same interaction between size and disparity as we observed in real mantids: the preferred size increases as simulated prey distance increases beyond the preferred distance. We show that this occurs because of a stereoscopic “false match” between the leading edge of the stimulus in one eye and its trailing edge in the other; further work will be required to find whether such false matches occur in real mantises. Importantly, the model also displays realistic responses to stimuli with vertical disparity and to pairs of identical stimuli offering a “ghost match”, despite not being fitted to these data. This is the first image-computable model of insect stereopsis, and reproduces key features of both neurophysiology and striking behaviour.
Binocular Vision and Stereopsis Across the Animal Kingdom by Read JCA, Read2021.pdf (3.4 MiB) - Most animals have at least some binocular overlap, i.e., a region of space that is viewed by both eyes. This reduces the overall visual field and raises the problem of combining two views of the world, seen from different vantage points, into a coherent whole. However, binocular vision also offers many potential advantages, including increased ability to see around obstacles and increased contrast sensitivity. One particularly interesting use for binocular vision is comparing information from both eyes to derive information about depth. There are many different ways in which this might be done, but in this review, I refer to them all under the general heading of stereopsis. This review examines the different possible uses of binocular vision and stereopsis and compares what is currently known about the neural basis of stereopsis in different taxa. Studying different animals helps us break free of preconceptions stemming from the way that stereopsis operates in human vision and provides new insights into the different possible forms of stereopsis.
Reduced surround suppression in monocular motion perception by Arranz-Paraiso S, Read JCA, Serrano-Pedraza I, ArranzParaisoReadSerranoPedraza2021.pdf (0.7 MiB) - Motion discrimination of large stimuli is impaired at high contrast and short durations. This psychophysical result has been linked with the center-surround suppression found in neurons of area MT. Recent physiology results have shown that most frontoparallel MT cells respond more strongly to binocular than to monocular stimulation. Here we measured the surround suppression strength under binocular and monocular viewing. Thirty-nine participants took part in two experiments: (a) where the nonstimulated eye viewed a blank field of the same luminance (n = 8) and (b) where it was occluded with a patch (n = 31). In both experiments, we measured duration thresholds for small (1 deg diameter) and large (7 deg) drifting gratings of 1 cpd with 85% contrast. For each subject, a Motion Suppression Index (MSI) was computed by subtracting the duration thresholds in logarithmic units of the large minus the small stimulus. Results were similar in both experiments. Combining the MSI of both experiments, we found that the strength of suppression for binocular condition (MSIbinocular = 0.249 ± 0.126 log10 (ms)) is 1.79 times higher than under monocular viewing (MSImonocular = 0.139 ± 0.137 log10 (ms)). This increase is too high to be explained by the higher perceived contrast of binocular stimuli and offers a new way of testing whether MT neurons account for surround suppression. Potentially, differences in surround suppression reported in clinical populations may reflect altered binocular processing.