Back in April I attended the excellent NETI meeting at UT Austin. Here’s the group photo:
Check out our guide to mantis breeding, by top insect technician Adam Simmons.
I have just been asked for “a succinct explanation of da Vinci stereopsis”. I googled in the hope of finding one, but couldn’t, so thought I’d put one up here.
Leonardo da Vinci didn’t quite realise that stereoscopic depth perception was a thing, but he did explain in his “Treatise on Painting” that a given object occludes different parts of the background when viewed from the left eye as compared to the right eye. “Da Vinci Stereopsis” now refers to depth perception based on the occlusion geometry in the two eyes. The term was introduced by Nakayama and Shimojo in a 1990 paper.
Consider the left-hand figure below. Both eyes see a large black rectangular object, but the right eye also sees a black bar to its right. Most observers, seeing these images, experience a weak sense that the bar is further away than the rectangle. This is because of the geometry shown in the figure. The left eye doesn’t see the bar because it’s hidden from view (“occluded”) by the nearer black rectangle.
Conversely, in the right-hand figure, the bar is only visible in the left eye, again to the right of the rectangle. Now, most people will report a weak sense that the bar is closer than the rectangle. This is because these retinal images could be accounted for by the scene shown at the top of the figure: the bar is technical seen by both eyes, but in the right eye it appears on top of the rectangle. Both objects are black and so the bar is invisible in the right eye.
Many vision scientists think that da Vinci stereopsis is a separate form of stereo vision that is not based on disparity (the separation between the images of the same object as seen in left and right eye). The argument is that because the bar is only visible in one eye’s image, a disparity cannot be defined.
Our latest paper on mantis stereopsis has just come out in Current Biology. Briefly, we find that mantis stereopsis operates very differently from humans’: it is based on temporal change, and does not require the images to be correlated.
We made a video abstract to explain the paper’s key findings and significance:
Lab members Chris Kaspiris-Rousellis, from Greece, and Maydel Fernandez Alonso, from Cuba, are Early Stage Researchers on the European Training Network on Full-Parallax Imaging. Here, they explain their projects.
Sid, Maydel, Chris and I had an excellent time at the VSS meeting last month. Thanks to Ignacio for this snap of my talk on “When invisible noise obscures the signal: the consequences of nonlinearity in motion detection.” Sid gave a great talk on his work “Modeling response variability in disparity-selective cells.”
I was giving a talk recently about my work on viewer experience with stereoscopic 3D television, and an audience member asked a good question, which was: Was there any relationship between people complaining of adverse effects and whether they routinely wore prescription spectacles? Such people are wearing two pairs of glasses to view S3D, which might be more uncomfortable, but equally they are already used to wearing glasses so might be less bothered than your average person who is wearing glasses only to view 3D.
We didn’t put anything about that in the papers, but I dug out the data and had a look. I haven’t done the stats, but it seems pretty clear there’s no effect of glasses. First, here is Fig 7 from Read & Bohr 2014:
And here is a version split up by whether or not participants usually wore glasses (in each pair of bars, the left-hand bar is for people who wore contacts or no correction, and the right-hand bar is for people who wore glasses).
In the graph, it looks as if there’s a striking difference in the “fake 3D passive” case, but really that’s to do with the small number of participants – we have 1/17 people without glasses reporting adverse effects, compared to 3/15 in the people with glasses. So if just one person changed their answer, it would look much less impressive. Since the effect isn’t seen in the other groups, I think it’s probably just a blip.
Averaging over all participants who wore 3D glasses (ie excluding only those in the true 2D group), the numbers are as follows:
n total = 311
n reporting adverse effects = 64 (21%)
n who habitually wear glasses = 117 (38%)
of whom n reporting adverse effects = 21 (18%)
n who do not habitually wear glasses = 194 (62%)
of whom n reporting adverse effects = 43 (22%)
Read JCA, Bohr I ( 2014 )
User experience while viewing stereoscopic 3D television
Journal press release "Good news for couch potatoes".
This question comes up occasionally and I was just recently asked a similar question by email, so I thought it would be a good idea to do a blog post that everyone can see. Although there’s a great article on this here: http://mentalfloss.com/uk/biology/30542/your-eyes-see-everything-upside-down
First off, the image of the world projected onto our retina is upside. This is just a consequence of geometry. This image from the Wikipedia article on pinhole cameras shows this nicely:
Our eye is more sophisticated than a pinhole camera — it has a lens so it can collect light over the whole of our pupil and bring it to a focus on our retina — but that isn’t important here. The retinal image is still upside-down. So why don’t we see the world upside-down?
One way of answering that is to point out that our eyes don’t, actually, “see” anything at all. Seeing happens in the brain. All your brain needs to know is the relationship between which photoreceptors are receiving the light, and where the object is in the world. We’ve learnt that if we want to touch an object whose image appears at the bottom of our eye, we usually have to raise our hands up (in the direction of our shoulders) while extending them, not move them down (towards our feet). So long as we know the correct mapping, it doesn’t actually matter where on the eye the information is.
My former colleague Dr Catherine O’Hanlon, now at Aberystwyth, and I have just published a paper on an interesting effect we found in small children. The roots of this study go back 7 years to when my son was two and I was reading a picture book with him. I was asking him to find various colours, and I noticed that he struggled to find colours that were in the background of the picture, like “blue” in a seascape or “yellow” in a beach scene. Apparently, he automatically ignored the background and attended only to colours in foreground objects. I thought that was an interesting phenomenon, and asked my developmental-psychology colleague Catherine if this was well known in small children. She said it hadn’t been reported, so we went ahead and tested more children to see if it was just my offspring or more widespread. It turns out that many two- and three-year-olds show this effect. They seem to assume that you must be talking about objects, even when you ask them something neutral like “Can you find green?”.
We think this bias helps young children acquire language.