In cognitive neuroscience, we’re interested in what guides human attention. We distinguish between influences from high-level cognition (e.g. current goals), and low-level visual features. There are highly sophisticated models of how visual features such as intensity, colour, and movement guide human attention. Computerised implementations of these models allow computers to mimic human eye movements. Turns out Taylor Swift’s amazing videos are an excellent example!
Traditional English DPhil programmes end in a viva. This is an examination by one internal and one external examiner, who critically assess the candidate and their thesis over the course of several hours. Mine was today, and I am delighted to announce that I passed! Many thanks to my wonderful …
Today was Python Day at the MRC Cognition and Brain Sciences Unit. Download all the required files and scripts here.
Tobii is a major player in the eye-tracking world, selling devices to customers in business and science. Today, Tobii has made a major step towards supporting open science by adding support for its new SDK in PyGaze (and by extension in OpenSesame). You can review the code on GitHub, download the PyGaze package as a zip, or download a full WinPython version. Details below, and more info on downloading here.
European Conference on Eye Movements The European Conference on Eye Movements is a biennial event for scientists and engineers who research or use eye movements in their work. The 2017 edition will be hosted in Wuppertal, Germany, and is preceded by several workshops. One of these is Edwin Dalmaijer’s “Learn …
Gazepoint is a relatively small player on the eye-tracking market. They sell two devices: the 60 Hz GP3 at a price of $695, and the 150 Hz GP3 HD at $1995 (both of those prices exclude VAT and shipping). Because of its relatively low price, the basic GP3 is an appealing model for researchers on a budget. As of today, PyGaze supports Gazepoint’s trackers through their OpenGaze API. Download the new code from GitHub, and have fun!
That is one confusing title! The point is this: When light reaches your eyes, you’re not immediately aware of that. It takes some time for your visual system to process the light, and to translate it into something the rest of your brain can work with. When that’s done, you consciously ‘see’. In a new paper, we show that the process of becoming aware of what you see, is affected by how large an object is. With an oversimplified example: If light bounces of a puppy, into your eyes, it takes a fraction of a second for you to become aware of the puppy. And it takes a fraction of a second longer if it’s a fat puppy.
This morning, the EyeTribe announced via an email to their customers that they would stop the development of their products. The particular reason is rather vague (“we’ve decided to go in a different direction with our technology“), and researchers across the board are not happy. The EyeTribe was the only real option for cheap eye tracking: It was great for demonstrations, for pupillometry and fixation control, it had a very elegant API, and the hardware was great for how much you paid for it. Best of all: It didn’t come with the restrictive licenses that almost all of the EyeTribe’s competitors use to milk their customers for more money. I, for one, am sad about the loss of this great company.