Touch, feel, see and hear the data

February 14, 2014 by Anthony King

It is now possible to sense scientific data as a means to deal with the mountains of information we face in our environment by applying subconscious processing to big data analysis

Imagine that could be transposed into a tactile experience. This is precisely what the CEEDs project, funded by the EU, promises. It uses integrated technologies to support human experience, when attempting to make sense of very large datasets. Jonathan Freeman Professor of Psychology at Goldsmith University of London, UK, talks to about how this project can help present data better, depending on the feedback participants provide to data from their environment, via close monitoring of their explicit responses—such as eye movement, for example—and their inner reactions, like .

What inspired you to get involved in data representation?

I felt there was a disjoint between our online and offline experience. For example, when shopping online or searching for a product—maybe a pair of jeans—the webpage on which you land, receives information about previous searches in all your cookies. Then, it can make inferences about you and target content appropriately. In the physical environment of a shop, there just isn't that level of insight and information provided to the environment. One big driver is to ask whether there are ways we can serve content that better suits a user's needs. And it does not have to be in a commercial environment.

What solution do you suggest?

We realised that humans have to deal with all this data. The problem is that our ability to analyse and understand it is a massive bottleneck. At the same time, the brain does an awful lot of processing that is not being used and that we are not consciously aware of. Nor does it figure in our behaviour. Our idea was therefore to marry the two and apply human subconscious processing to the analysis of big data.

Could you provide a specific example of how this could be of benefit?

Take a scientist analysing say a huge neuroscience dataset in our project experience induction machine. We apply measurements that tell us whether they are getting fatigued or overloaded with information. If that's the case, the system does one of two things. It either changes the visualisations to simplify them so as to reduce the cognitive load [a measure of brain workout], thus keeping the user less stressed and more able to focus. Or it will implicitly guide the person to areas of the data representation that are not as heavy in information. We can use subliminal clues to do this, such as arrows that flash so quickly that people are not aware of them.

Part of your approach involves watching the watchers use data. So what kind of technology do you rely on?

We devised an immersive setup, where the user is subjected to constant monitoring. We use a Kinect [motion sensing device] to track people's postural and body response. A glove tracks people's hand movements in more detail and measures galvanic skin responses, which is a measure of stress. We have an eye tracker that tells the user where about in the data they are focusing. It also looks at their pupil to see how dilated they are, as a measure of their cognitive work rate. In parallel, a camera system analyses facial expressions and a voice analysis system measures the emotional characteristics of their voice. Finally, users can wear the vest we developed under the project in this mixed reality space, called the CEEDs eXperience Induction Machine (XIM), which measures their heart rate.

Is the visual part of the project important?

Visualisation technologies in the experience induction machine are important as people are in an immersive 3D environment. But the representations that we use for the data are not just visual. There are also audio representations for data: spatialisation of audio and sonification of data so that users can hear the data. For example, the activity that normally flows in part of the brain can be represented so that more activity can sound louder or higher pitched to neuroscientists looking at these flows. There are also tactile actuators in the glove that allow users to grab data and feel feedback in their fingertips.

What is so novel about this approach?

The Kinect is available. But never before has anyone put all these components in one place to trial them together. And never before has this advanced set of sensors been put together with the goal of addressing optimising the human understanding of . This is novel, cutting edge and ambitious. It is not simple product development. This is about pushing the boundaries and taking risks.

Who will this technology be useful to?

Initially those who deal with massive data sets such as economists, neuroscientists and data analysts will benefit. But people will also benefit. We are all bombarded with information. There are going to be real benefits for people in having systems that response to your implicit cues as a consumer or person. It does not have to be in a consumption context.

Could you provide an example of application outside commercial applications?

Imagine you are an archaeologist working in the field and you come across a piece of pottery. You look at it and say it comes from the 4th century and from such and such object. It takes years of experience for an archaeologist to be able to do that. In our project, what we are doing is measuring how expert archaeologists look at objects and evaluate them. Then, we feed this interpretation into a database to speed up potential matching of pottery pieces. It makes the machine better, speeding up the predictive search powers of technology.

Explore further: Beam me to my meeting!

More information:

Related Stories

Beam me to my meeting!

October 29, 2012

Forget about crackly lines or blurry webcams. Video conferencing has just got a whole lot better. By combining robotics, video and a host of other sensor and display technologies, European scientists can now virtually 'beam' ...

New framework for 3D tele-immersive social networking

November 8, 2013

It's Friday night, you're exhausted after a long week in the office. You're not going to leave the house so you could either watch TV or spend a few hours catching up on your social networks. But how about doing both? Thanks ...

Active learning model for computer predictions

December 3, 2013

Computers serve as powerful tools for categorizing, displaying and searching data. They're tools we continually upgrade to improve processing power, both locally and via the cloud. But computers are simply the medium for ...

Recommended for you

Intelligent robots threaten millions of jobs

February 14, 2016

Advances in artificial intelligence will soon lead to robots that are capable of nearly everything humans do, threatening tens of millions of jobs in the coming 30 years, experts warned Saturday.

Tiny diatoms boast enormous strength

February 8, 2016

Diatoms are single-celled algae organisms, around 30 to 100 millionths of a meter in diameter, that are ubiquitous throughout the oceans. These creatures are encased within a hard shell shaped like a wide, flattened cylinder—like ...

Search engines will know what you want ... sooner

February 8, 2016

If you enter "Oklahoma" in a search engine, you might get a travelogue, news about the oil industry, Oklahoma State football scores or an article on Rodgers and Hammerstein musicals. What appears at the top of the list might ...


Please sign in to add a comment. Registration is free, and takes less than a minute. Read more

Click here to reset your password.
Sign in to get notified via email when new comments are made.