Tag Archives: Sriram Subramanian

Skin as a touchscreen (“smart” hands)

An April 11, 2016 news item on phys.org highlights some research presented at the IEEE (Institute of Electrical and Electronics Engineers) Haptics (touch) Symposium 2016,

Using your skin as a touchscreen has been brought a step closer after UK scientists successfully created tactile sensations on the palm using ultrasound sent through the hand.

The University of Sussex-led study – funded by the Nokia Research Centre and the European Research Council – is the first to find a way for users to feel what they are doing when interacting with displays projected on their hand.

This solves one of the biggest challenges for technology companies who see the human body, particularly the hand, as the ideal display extension for the next generation of smartwatches and other smart devices.

Current ideas rely on vibrations or pins, which both need contact with the palm to work, interrupting the display.

However, this new innovation, called SkinHaptics, sends sensations to the palm from the other side of the hand, leaving the palm free to display the screen.

An April 11, 2016 University of Sussex press release (also on EurekAlert) by James Hakmer, which originated the news item, provides more detail,

The device uses ‘time-reversal’ processing to send ultrasound waves through the hand. This technique is effectively like ripples in water but in reverse – the waves become more targeted as they travel through the hand, ending at a precise point on the palm.

It draws on a rapidly growing field of technology called haptics, which is the science of applying touch sensation and control to interaction with computers and technology.

Professor Sriram Subramanian, who leads the research team at the University of Sussex, says that technologies will inevitably need to engage other senses, such as touch, as we enter what designers are calling an ‘eye-free’ age of technology.

He says: “Wearables are already big business and will only get bigger. But as we wear technology more, it gets smaller and we look at it less, and therefore multisensory capabilities become much more important.

“If you imagine you are on your bike and want to change the volume control on your smartwatch, the interaction space on the watch is very small. So companies are looking at how to extend this space to the hand of the user.

“What we offer people is the ability to feel their actions when they are interacting with the hand.”

The findings were presented at the IEEE Haptics Symposium [April 8 – 11] 2016 in Philadelphia, USA, by the study’s co-author Dr Daniel Spelmezan, a research assistant in the Interact Lab.

There is a video of the work (I was not able to activate sound, if there is any accompanying this video),

The consequence of watching this silent video was that I found the whole thing somewhat mysterious.

Shapeshifting on demand but no stretching yet: morphees

This research (Morphees) is from Bristol University where researchers have created prototypes for shapeshifting mobile devices,

A high-fidelity prototype using projection and tracking on wood tiles that are actuated with thin shape-memory alloy wires [downloaded from http://www.bris.ac.uk/news/2013/9332.html/]

A high-fidelity prototype using projection and tracking on wood tiles that are actuated with thin shape-memory alloy wires [downloaded from http://www.bris.ac.uk/news/2013/9332.html/]

The Apr. 28, 2013 news release on EurekAlert provides more detail,

The research, led by Dr Anne Roudaut and Professor Sriram Subramanian, from the University of Bristol’s Department of Computer Science, have used ‘shape resolution’ to compare the resolution of six prototypes the team have built using the latest technologies in shape changing material, such as shape memory alloy and electro active polymer.

One example of a device is the team’s concept of Morphees, self-actuated flexible mobile devices that can change shape on-demand to better fit the many services they are likely to support.

The team believe Morphees will be the next generation of mobile devices, where users can download applications that embed a dedicated form factor, for instance the “stress ball app” that collapses the device in on itself or the “game app” that makes it adopt a console-like shape.

Dr Anne Roudaut, Research Assistant in the Department of Computer Science’s Bristol Interaction and Graphics group, said: “The interesting thing about our work is that we are a step towards enabling our mobile devices to change shape on-demand. Imagine downloading a game application on the app-store and that the mobile phone would shape-shift into a console-like shape in order to help the device to be grasped properly. The device could also transform into a sphere to serve as a stress ball, or bend itself to hide the screen when a password is being typed so passers-by can’t see private information.”

By comparing the shape resolution of their prototypes, the researchers have created insights to help designers towards creating high shape resolution Morphees.

In the future the team hope to build higher shape resolution Morphees by investigating the flexibility of materials. They are also interested in exploring other kinds of deformations that the prototypes did not explore, such as porosity and stretchability.

Here’s the video where the researchers demonstrate their morphees,


The work will be presented at ACM CHI 2013, sometime between Saturday 27 April to Thursday 2 May 2013, in Paris, France. For those who’d like to see the paper which will be presented, here’s a link to it,

Morphees: Toward High “Shape Resolution” in Self-Actuated Flexible Mobile Devices by
Anne Roudaut, Abhijit Karnik, Markus Löchtefeld, and Sriram Subramanian

After reading the news release and watching the video, I am reminded of the ‘morph’ concept, a shapeshifting, wearable device proposed by Cambridge University and Nokia. Last I wrote about that project, they had announced a stretchable skin, as per my Nov. 7, 2011 posting.

For those who are interested in what ACM CHI 2013 is all about, from the home page,

The ACM SIGCHI Conference on Human Factors in Computing Systems is the premier international conference on human-computer interaction. CHI 2013 is about changing perspectives: we draw from the constantly changing perspectives of the diverse CHI community and beyond, but we also change perspectives, offering new visions of people interacting with technology. The conference is multidisciplinary, drawing from science, engineering and design, with contributions from research and industry in 15 different venues. CHI brings together students and experts from over 60 countries, representing different cultures and different application areas, whose diverse perspectives influence each other.

CHI 2013 is located in vibrant Paris, France, the most visited city in the world. The conference will be held at the Palais de Congrès de Paris. First in Europe in research and development, with the highest concentration of higher education students in Europe, Paris is a world-class center for business and culture, with over 3800 historical monuments.The Louvre’s pyramid captures the spirit of CHI’13, offering diverse perspectives on design and technology, contrasting the old and new. The simple glass sides reveal inner complexity, sometimes transparent, sometimes reflecting the people and buildings that surround it, in the constantly
changing Paris light.

CHI 2013 welcomes works addressing research on all aspects of human-computer interaction (HCI), as well as case studies of interactive system designs, innovative proof-of-concept, and presentations by experts on the latest challenges and innovations in the field. In addition to a long-standing focus on professionals in design, engineering, management, and user experience; this year’s conference has made special efforts to serve communities in the areas of: design, management, engineering, user experience, arts, sustainability, children, games and health. We look forward to seeing you at CHI 2013 in Paris!

As I recall, ACM stands for Association of Computing Machinery, CHI stands for computer-human interface, and SIG stands for Special Interest Group.

ETA May 13, 2013: I meant to do this two weeks ago (Apr. 30,2013), ah well. Roel Vertegaal and his team at Canada’s Queen’s University introduced something called a MorePhone, which can curl up and change shape, at the CHI 2013. From the Apr. 30, 2013 news release on EurekAlert*,

Researchers at Queen’s University’s Human Media Lab have developed a new smartphone – called MorePhone – which can morph its shape to give users a silent yet visual cue of an incoming phone call, text message or email.

“This is another step in the direction of radically new interaction techniques afforded by smartphones based on thin film, flexible display technologies” says Roel Vertegaal (School of Computing), director of the Human Media Lab at Queen’s University who developed the flexible PaperPhone and PaperTab.

“Users are familiar with hearing their phone ring or feeling it vibrates in silent mode. One of the problems with current silent forms of notification is that users often miss notifications when not holding their phone. With MorePhone, they can leave their smartphone on the table and observe visual shape changes when someone is trying to contact them.”

MorePhone is not a traditional smartphone. It is made of a thin, flexible electrophoretic display manufactured by Plastic Logic – a British company and a world leader in plastic electronics. Sandwiched beneath the display are a number of shape memory alloy wires that contract when the phone notifies the user. This allows the phone to either curl either its entire body, or up to three individual corners. Each corner can be tailored to convey a particular message. For example, users can set the top right corner of the MorePhone to bend when receiving a text message, and the bottom right corner when receiving an email. Corners can also repeatedly bend up and down to convey messages of greater urgency.

I have written about Vertegaal and his team’s ‘paper’ devices previously. The most recent piece is this Jan. 9, 2013 posting, Canada’s Queen’s University strikes again with its ‘paper’ devices. You can find out more about Plastic Logic here.

*’Eurkealert’ changed to ‘EurekAlert’ on Feb. 17, 2016.