Dr. Roel Vertegaal’s team at Queen’s University (Ontario, Canada) has developed a flexible smartphone that allows uses to page through a document as if it were paper. A Feb. 16, 2016 news item on phys.org makes the announcement,
Researchers at Queen’s University’s Human Media Lab have developed the world’s first full-colour, high-resolution and wireless flexible smartphone to combine multitouch with bend input. The phone, which they have named ReFlex, allows users to experience physical tactile feedback when interacting with their apps through bend gestures.
“This represents a completely new way of physical interaction with flexible smartphones” says Roel Vertegaal (School of Computing), director of the Human Media Lab at Queen’s University.
“When this smartphone is bent down on the right, pages flip through the fingers from right to left, just like they would in a book. More extreme bends speed up the page flips. Users can feel the sensation of the page moving through their fingertips via a detailed vibration of the phone. This allows eyes-free navigation, making it easier for users to keep track of where they are in a document.”
ReFlex is based on a high definition 720p LG Display Flexible OLED touch screen powered by an Android 4.4 “KitKat” board mounted to the side of the display. Bend sensors behind the display sense the force with which a user bends the screen, which is made available to apps for use as input. ReFlex also features a voice coil that allows the phone to simulate forces and friction through highly detailed vibrations of the display. Combined with the passive force feedback felt when bending the display, this allows for a highly realistic simulation of physical forces when interacting with virtual objects.
“This allows for the most accurate physical simulation of interacting with virtual data possible on a smartphone today,” says Dr. Vertegaal. “When a user plays the “Angry Birds” game with ReFlex, they bend the screen to stretch the sling shot. As the rubber band expands, users experience vibrations that simulate those of a real stretching rubber band. When released, the band snaps, sending a jolt through the phone and sending the bird flying across the screen.”
Dr. Vertegaal thinks bendable, flexible smartphones will be in the hands of consumers within five years. Queen’s researchers will unveil [have unveiled] the ReFlex prototype at the tenth anniversary Conference on Tangible Embedded and Embodied Interaction (TEI) in Eindhoven, The Netherlands on February 17 . The annual forum is the world’s premier conference on tangible human-computer interaction.
This research was support by Immersion Canada Inc. and the Natural Sciences and Engineering Research Council of Canada (NSERC).
The researchers have provided a video showing what ReFlex can do,
Here’s a citation and a link to a PDF of the paper being presented at the conference,
For anyone who noticed that in addition to the government funding (NSERC), part of this group’s funding came from Immersion Canada Inc., here’s the parent company’s website (It’s a US company with some sort of Canadian presence).
I last wrote about Vertegaal’s work in an April 29, 2013 posting (scroll down about 70% of the way to find information about the group’s shapeshifting MorePhone).
The research, led by Dr Anne Roudaut and Professor Sriram Subramanian, from the University of Bristol’s Department of Computer Science, have used ‘shape resolution’ to compare the resolution of six prototypes the team have built using the latest technologies in shape changing material, such as shape memory alloy and electro active polymer.
One example of a device is the team’s concept of Morphees, self-actuated flexible mobile devices that can change shape on-demand to better fit the many services they are likely to support.
The team believe Morphees will be the next generation of mobile devices, where users can download applications that embed a dedicated form factor, for instance the “stress ball app” that collapses the device in on itself or the “game app” that makes it adopt a console-like shape.
Dr Anne Roudaut, Research Assistant in the Department of Computer Science’s Bristol Interaction and Graphics group, said: “The interesting thing about our work is that we are a step towards enabling our mobile devices to change shape on-demand. Imagine downloading a game application on the app-store and that the mobile phone would shape-shift into a console-like shape in order to help the device to be grasped properly. The device could also transform into a sphere to serve as a stress ball, or bend itself to hide the screen when a password is being typed so passers-by can’t see private information.”
By comparing the shape resolution of their prototypes, the researchers have created insights to help designers towards creating high shape resolution Morphees.
In the future the team hope to build higher shape resolution Morphees by investigating the flexibility of materials. They are also interested in exploring other kinds of deformations that the prototypes did not explore, such as porosity and stretchability.
Here’s the video where the researchers demonstrate their morphees,
The work will be presented at ACM CHI 2013, sometime between Saturday 27 April to Thursday 2 May 2013, in Paris, France. For those who’d like to see the paper which will be presented, here’s a link to it,
After reading the news release and watching the video, I am reminded of the ‘morph’ concept, a shapeshifting, wearable device proposed by Cambridge University and Nokia. Last I wrote about that project, they had announced a stretchable skin, as per my Nov. 7, 2011 posting.
For those who are interested in what ACM CHI 2013 is all about, from the home page,
The ACM SIGCHI Conference on Human Factors in Computing Systems is the premier international conference on human-computer interaction. CHI 2013 is about changing perspectives: we draw from the constantly changing perspectives of the diverse CHI community and beyond, but we also change perspectives, offering new visions of people interacting with technology. The conference is multidisciplinary, drawing from science, engineering and design, with contributions from research and industry in 15 different venues. CHI brings together students and experts from over 60 countries, representing different cultures and different application areas, whose diverse perspectives influence each other.
CHI 2013 is located in vibrant Paris, France, the most visited city in the world. The conference will be held at the Palais de Congrès de Paris. First in Europe in research and development, with the highest concentration of higher education students in Europe, Paris is a world-class center for business and culture, with over 3800 historical monuments.The Louvre’s pyramid captures the spirit of CHI’13, offering diverse perspectives on design and technology, contrasting the old and new. The simple glass sides reveal inner complexity, sometimes transparent, sometimes reflecting the people and buildings that surround it, in the constantly
changing Paris light.
CHI 2013 welcomes works addressing research on all aspects of human-computer interaction (HCI), as well as case studies of interactive system designs, innovative proof-of-concept, and presentations by experts on the latest challenges and innovations in the field. In addition to a long-standing focus on professionals in design, engineering, management, and user experience; this year’s conference has made special efforts to serve communities in the areas of: design, management, engineering, user experience, arts, sustainability, children, games and health. We look forward to seeing you at CHI 2013 in Paris!
As I recall, ACM stands for Association of Computing Machinery, CHI stands for computer-human interface, and SIG stands for Special Interest Group.
ETA May 13, 2013: I meant to do this two weeks ago (Apr. 30,2013), ah well. Roel Vertegaal and his team at Canada’s Queen’s University introduced something called a MorePhone, which can curl up and change shape, at the CHI 2013. From the Apr. 30, 2013 news release on EurekAlert*,
Researchers at Queen’s University’s Human Media Lab have developed a new smartphone – called MorePhone – which can morph its shape to give users a silent yet visual cue of an incoming phone call, text message or email.
“This is another step in the direction of radically new interaction techniques afforded by smartphones based on thin film, flexible display technologies” says Roel Vertegaal (School of Computing), director of the Human Media Lab at Queen’s University who developed the flexible PaperPhone and PaperTab.
“Users are familiar with hearing their phone ring or feeling it vibrates in silent mode. One of the problems with current silent forms of notification is that users often miss notifications when not holding their phone. With MorePhone, they can leave their smartphone on the table and observe visual shape changes when someone is trying to contact them.”
MorePhone is not a traditional smartphone. It is made of a thin, flexible electrophoretic display manufactured by Plastic Logic – a British company and a world leader in plastic electronics. Sandwiched beneath the display are a number of shape memory alloy wires that contract when the phone notifies the user. This allows the phone to either curl either its entire body, or up to three individual corners. Each corner can be tailored to convey a particular message. For example, users can set the top right corner of the MorePhone to bend when receiving a text message, and the bottom right corner when receiving an email. Corners can also repeatedly bend up and down to convey messages of greater urgency.
Not all science research and breakthroughs require massive investments of money, sometimes all you need is a home DVD burner as this Feb. 19, 2013 news release on EurekAlert notes,
While the demand for ever-smaller electronic devices has spurred the miniaturization of a variety of technologies, one area has lagged behind in this downsizing revolution: energy-storage units, such as batteries and capacitors.
Now, Richard Kaner, a member of the California NanoSystems Institute at UCLA and a professor of chemistry and biochemistry, and Maher El-Kady, a graduate student in Kaner’s laboratory, may have changed the game.
The UCLA researchers have developed a groundbreaking technique that uses a DVD burner to fabricate micro-scale graphene-based supercapacitors — devices that can charge and discharge a hundred to a thousand times faster than standard batteries. These micro-supercapacitors, made from a one-atom–thick layer of graphitic carbon, can be easily manufactured and readily integrated into small devices such as next-generation pacemakers.
The new cost-effective fabrication method, described in a study published this week in the journal Nature Communications, holds promise for the mass production of these supercapacitors, which have the potential to transform electronics and other fields.
“Traditional methods for the fabrication of micro-supercapacitors involve labor-intensive lithographic techniques that have proven difficult for building cost-effective devices, thus limiting their commercial application,” El-Kady said. “Instead, we used a consumer-grade LightScribe DVD burner to produce graphene micro-supercapacitors over large areas at a fraction of the cost of traditional devices. [emphasis mine] Using this technique, we have been able to produce more than 100 micro-supercapacitors on a single disc in less than 30 minutes, using inexpensive materials.”
The process of miniaturization often relies on flattening technology, making devices thinner and more like a geometric plane that has only two dimensions. In developing their new micro-supercapacitor, Kaner and El-Kady used a two-dimensional sheet of carbon, known as graphene, which only has the thickness of a single atom in the third dimension.
Kaner and El-Kady took advantage of a new structural design during the fabrication. For any supercapacitor to be effective, two separated electrodes have to be positioned so that the available surface area between them is maximized. This allows the supercapacitor to store a greater charge. A previous design stacked the layers of graphene serving as electrodes, like the slices of bread on a sandwich. While this design was functional, however, it was not compatible with integrated circuits.
In their new design, the researchers placed the electrodes side by side using an interdigitated pattern, akin to interwoven fingers. This helped to maximize the accessible surface area available for each of the two electrodes while also reducing the path over which ions in the electrolyte would need to diffuse. As a result, the new supercapacitors have more charge capacity and rate capability than their stacked counterparts.
Interestingly, the researchers found that by placing more electrodes per unit area, they boosted the micro-supercapacitor’s ability to store even more charge.
Kaner and El-Kady were able to fabricate these intricate supercapacitors using an affordable and scalable technique that they had developed earlier. They glued a layer of plastic onto the surface of a DVD and then coated the plastic with a layer of graphite oxide. Then, they simply inserted the coated disc into a commercially available LightScribe optical drive — traditionally used to label DVDs — and took advantage of the drive’s own laser to create the interdigitated pattern. The laser scribing is so precise that none of the “interwoven fingers” touch each other, which would short-circuit the supercapacitor.
“To label discs using LightScribe, the surface of the disc is coated with a reactive dye that changes color on exposure to the laser light. Instead of printing on this specialized coating, our approach is to coat the disc with a film of graphite oxide, which then can be directly printed on,” Kaner said. “We previously found an unusual photo-thermal effect in which graphite oxide absorbs the laser light and is converted into graphene in a similar fashion to the commercial LightScribe process. With the precision of the laser, the drive renders the computer-designed pattern onto the graphite oxide film to produce the desired graphene circuits.”
“The process is straightforward, cost-effective and can be done at home,” El-Kady said. “One only needs a DVD burner and graphite oxide dispersion in water, which is commercially available at a moderate cost.”
The new micro-supercapacitors are also highly bendable and twistable, making them potentially useful as energy-storage devices in flexible electronics like roll-up displays and TVs, e-paper, and even wearable electronics.
The reference to e-paper and roll-up displays calls to mind work being done at Queen’s University (Kingston, Canada) and Roel Vertegaal’s work on bendable, flexible phones and computers (my Jan. 9, 2013 posting). Could this work on micro-supercapacitors have an impact on that work?
Here’s an image (supplied by UCLA) of the micro-supercapacitors ,
Kaner and El-Kady’s micro-supercapacitors
UCLA has also supplied a video of Kaner and El-Kady discussing their work,
Roel Vertegaal at Queen’s University (Ontario, Canada) has released a ‘paper’ tablet. Like the bendable, flexible ‘paper’ phone he presented at the CHI 2011 meeting in Vancouver, Canada (my May 12, 2011 posting), this tablet offers some intriguing possibilities but is tethered. The Jan. 9, 2013 news item on phys.org provides more information about the new ‘paper’ device (Note: Links have been removed),
Watch out tablet lovers – a flexible paper computer developed at Queen’s University in collaboration with Plastic Logic and Intel Labs will revolutionize the way people work with tablets and computers.
The PaperTab tablet looks and feels just like a sheet of paper. However, it is fully interactive with a flexible, high-resolution 10.7-inch plastic display developed by Plastic Logic and a flexible touchscreen. It is powered by the second generation I5 Core processor developed by Intel.
Vertegaal and his team have produced a video demonstrating their ‘paper’ tablet/computer:
“Using several PaperTabs makes it much easier to work with multiple documents,” says Roel Vertegaal, Director of Queen’s University’s Human Media Lab. “Within five to ten years, most computers, from ultra-notebooks to tablets, will look and feel just like these sheets of printed color paper.”
“We are actively exploring disruptive user experiences. The ‘PaperTab’ project, developed by the Human Media Lab at Queen’s University and Plastic Logic, demonstrates novel interactions powered by Intel processors that could potentially delight tablet users in the future,” says Intel’s Experience Design Lead Research Scientist, Ryan Brotman.
PaperTab’s intuitive interface allows users to create a larger drawing or display surface by placing two or more PaperTabs side by side. PaperTab emulates the natural handling of multiple sheets of paper. It can file and display thousands of paper documents, replacing the need for a computer monitor and stacks of papers or printouts.
Unlike traditional tablets, PaperTabs keep track of their location relative to each other, and to the user, providing a seamless experience across all apps, as if they were physical computer windows.
“Plastic Logic’s flexible plastic displays allow a natural human interaction with electronic paper, being lighter, thinner and more robust compared with today’s standard glass-based displays. This is just one example of the innovative revolutionary design approaches enabled by flexible displays,” explains Indro Mukerjee, CEO of Plastic Logic.
The partners are saying that ‘paper’ tablets may be on the market in foreseeable future according to Emma Wollacott’s Jan. 8, 2013 article for TG Daily,
The bendy tablet has been coming for quite a while now, but a version to be shown off today at CES [Consumer Electronics Show] could be ready for the market within three years, say its creators.
CEATEC (Cutting Edge IT [Information Technology] and Electronics Comprehensive Exhibition) Japan, Oct.4-8, 2011 is a large technology fair being held in Chiba, near Tokyo. Some 800 companies are showcasing their latest and greatest according to the Oct. 4, 2011 news item on physorg.com,
Around 600 firms unveiled their innovations at the Combined Exhibition of Advanced Technologies (Ceatec) exhibition in Chiba, near Tokyo, expected to draw 200,000 visitors during its five-day run, organisers said.
The impact of Japan’s March 11 earthquake, tsunami and nuclear disaster gave added resonance to technologies on display, particularly those aimed at improving urban infrastructure and energy efficiency.
State-of-the-art radiation counters and power-saving technologies are in high demand after Japan’s disasters sparked fears over contamination and led to power shortages, requiring cuts to energy consumption this summer.
Japanese telecom giant NTT [Nippon Telegraph and Telephone] DoCoMo showed off a smartphone with changeable sensor-embedded shells that can detect bad breath, vital body signs and even be used to measure background radiation levels.
One item that particularly interested me is a transparent organic film from Murata Manufacturing. From the news item,
Electronics parts maker Murata Manufacturing unveiled devices using a newly developed transparent organic film that can deliver instructions via twisting motions or pressure.
One of its gadgets, a light-powered plate called the Leaf Grip Remote Controller, has no buttons but is instead operated by the user bending and twisting it.
Another application of the film is as a touch panel which responds to left-right and up-down finger swipes but also senses how strongly it is being pressed, unlike conventional touchscreen glass used on smartphones.
“Currently we give commands two-dimensionally on touch panels in smartphones and tablet computers but this invention would give us another dimension — how hard they are pressed,” Murata spokesman Kazuhisa Mashita said.
“This could enable users to scroll screens slowly by touching the screen lightly and move images faster by pressing it harder,” he told AFP [Agence France-Presse] ahead of the exhibition.
Earlier this year when CHI (computer-human interface) 2011 was taking place in Vancouver, Canada, I wrote about Roel Vertegaal and his team’s work on their PaperPhone and bending and twisting gestures (May 12, 2011 posting).
Bending and twisting a flexible screen doesn’t seem all that complicated but when you think about making those gestures meaningful, i. e., ‘slowing a screen image by pressing more softly’, you realize just how much effort and thought are required for features, that if successful, will not be noticed.
For anyone who’s not familiar with the Morph, it’s an idea that Nokia and the University of Cambridge’s Nanoscience Centre have been working on for the last few years. Originally announced as a type of flexible phone that you could wrap around your wrist, the Morph is now called a concept. Here’s an animation illustrating some of the concepts which include flexibility and self-cleaning,
There have been very few announcements of any kind about the Morph or the technology that will support this concept. A few months ago, they did make an announcement about researching graphene as a means of actualizing the concept (noted in my May 6, 2011 posting [scroll down about 1/2 way]).
Interestingly the latest research published on graphene and the flexible, transparent screens that are necessary to making something like the Morph a reality has come from a lab at Rice University. From the August 1, 2011 news item on Nanowerk,
The lab of Rice chemist James Tour lab has created thin films that could revolutionize touch-screen displays, solar panels and LED lighting. The research was reported in the online edition of ACS Nano (“Rational Design of Hybrid Graphene Films for High-Performance Transparent Electrodes”).
Flexible, see-through video screens may be the “killer app” that finally puts graphene — the highly touted single-atom-thick form of carbon — into the commercial spotlight once and for all, Tour said. Combined with other flexible, transparent electronic components being developed at Rice and elsewhere, the breakthrough could lead to computers that wrap around the wrist and solar cells that wrap around just about anything. [emphasis mine]
The lab’s hybrid graphene film is a strong candidate to replace indium tin oxide (ITO), a commercial product widely used as a transparent, conductive coating. It’s the essential element in virtually all flat-panel displays, including touch screens on smart phones and iPads, and is part of organic light-emitting diodes (OLEDs) and solar cells.
Here’s James Tour and Yu Zhu, the paper’s lead author, explaining how the flexible screen was developed,
There are other flexible screens and competitors to the Morph notably the PaperPhone mentioned in my May 6,2011 posting (scroll down about 2/3 of the way) and in my May 12, 2011 posting featuring an interview with Roel Vertegaal of Queen’s University, Ontario, Canada, about the PaperPhone. (We did not discuss the role that graphene might or might not play in the development of the Paperphone’s screens.)
I wonder what impact this work at Rice will have not only for the Morph and the PaperPhone but on the European Union’s pathfinder research competition (the prize is $1B Euros), mentioned in my June 13, 2011 posting about graphene (scroll down about 1/3 of the way). Graphene is one of the research areas being considered for the prize.
The cookie gambit started on a dare when Tour mentioned at a meeting that his lab had produced graphene from table sugar.
“I said we could grow it from any carbon source — for example, a Girl Scout cookie, because Girl Scout Cookies were being served at the time,” Tour recalled. “So one of the people in the room said, ‘Yes, please do it. … Let’s see that happen.'”
Members of Girl Scouts of America Troop 25080 came to Rice’s Smalley Institute for Nanoscale Science and Technology to see the process. Rice graduate students Gedeng Ruan, lead author of the paper, and Zhengzong Sun calculated that at the then-commercial rate for pristine graphene — $250 for a two-inch square — a box of traditional Girl Scout shortbread cookies could turn a $15 billion profit.
Here’s the full reference for this second paper,
Gedeng Ruan, Zhengzong Sun, Zhiwei Peng, James M. Tour. Growth of Graphene from Food, Insects and Waste. ACS Nano, 2011; 110729113834087 DOI: 10.1021/nn202625c
This story has it all: military, patents, international competition and cooperation, sex (well, not according to the academics but I think it’s possible), and a bizarre device – the PaperPhone (last mentioned in my May 6, 2011 posting on Human-Computer Interfaces).
“If you want to know what technologies people will be using 10 years in the future, talk to the people who’ve been working on a lab project for 10 years,” said Dr. Roel Vertegaal, Director of the Human Media Lab at Queen’s University in Kingston, Ontario. By the way, 10 years is roughly the length of time Vertegaal and his team have been working on a flexible/bendable phone/computer and he believes that it will be another five to 10 years before the device is available commercially.
PaperPhone consists of an Arizona State University Flexible Display Center 3.7” Bloodhound flexible electrophoretic display, augmented with a layer of 5 Flexpoint 2” bidirectional bend sensors. The prototype is driven by an E Ink Broadsheet AM 300 Kit featuring a Gumstix processor. The prototype has a refresh rate of 780 ms for a typical full screen gray scale image.
An Arduino microcontroller obtains data from the Flexpoint bend sensors at a frequency of 20 Hz. Figure 2 shows the back of the display, with the bend sensor configuration mounted on a flexible printed circuit (FPC) of our own design. We built the FPC by printing its design on DuPont Pyralux flexible circuit material with a solid ink printer, then etching the result to obtain a fully functional flexible circuit substrate. PaperPhone is not fully wireless. This is because of the supporting rigid electronics that are required to drive the display. A single, thin cable bundle connects the AM300 and Arduino hardware to the display and sensors. This design maximizes the flexibility and mobility of the display, while keeping its weight to a minimum. The AM300 and Arduino are connected to a laptop running a Max 5 patch that processes sensor data, performs bend gesture recognition and sends images to the display. p. 3
It may look ungainly but it represents a significant step forward for the technology as this team (composed of researchers from Queen’s University, Arizona State University, and E Ink Corporation) appears to have produced the only working prototype in the world for a personal portable flexible device that will let you make phone calls, play music, read a book, and more by bending it. As they continue to develop the product, the device will become wireless.
The PaperPhone and the research about ‘bending’, i.e., the kinds of bending gestures people would find easiest and most intuitive to use when activating the device, were presented in Vancouver in an early session at the CHI 2011 Conference where I got a chance to speak to Dr. Vertegaal and his team.
Amongst other nuggets, I found out the US Department of Defense (not DARPA [Defense Advanced Research Projects Agency] oddly enough) has provided funding for the project. Military interest is focused on the device’s low energy requirements, lowlight screen, and light weight in addition to its potential ability to be folded up and carried like a piece of paper (i. e., it could mould itself to fit a number of tight spaces) as opposed to the rigid, ungiving borders of a standard mobile device. Of course, all of these factors are quite attractive to consumers too.
As is imperative these days, the ‘bends’ that activate the device have been patented and Vertegaal is in the process of developing a startup company that will bring this device and others to market. Queen’s University has an ‘industrial transfer’ office (they probably call it something else) which is assisting him with the startup.
There is international interest in the PaperPhone that is collaborative and competitive. Vertegaal’s team at Queen’s is partnered with a team at Arizona State University led by Dr. Winslow Burleson, professor in the Computer Systems Engineering and the Arts, Media, and Engineering graduate program and with Michael McCreary, Vice President Research & Development of E Ink Corporation representing an industry partner.
On the competitive side of things, the UK’s University of Cambridge and the Finnish Nokia Research Centre have been working on the Morph which as I noted in my May 6, 2011 posting still seems to be more concept than project.
Vertegaal noted that the idea of a flexible screen is not new and that North American companies have gone bankrupt trying to bring the screens to market. These days, you have to go to Taiwan for industrial production of flexible screens such as the PaperPhone’s.
One of my last questions to the team was about pornography. (In the early days of the Internet [which had its origins in military research], there were only two industries that made money online, pornography and gambling. The gambling opportunities seem pretty similar to what we already enjoy.) After an amused response, the consensus was that like gambling it’s highly unlikely a flexible phone could lend itself to anything new in the field of pornography. Personally, I’m not convinced about that one.
So there you have a case study for innovation. Work considered bleeding edge 10 years ago is now cutting edge and, in the next five to 10 years, that work will be become a consumer product. Along the way you have military investment, international collaboration and competition, failure and success, and, possibly, sex.
This time I’ve decided to explore a few of the human/computer interface stories I’ve run across lately. So this posting is largely speculative and rambling as I’m not driving towards a conclusion.
My first item is a May 3, 2011 news item on physorg.com. It concerns an art installation at Rensselaer Polytechnic Institute, The Ascent. From the news item,
A team of Rensselaer Polytechnic Institute students has created a system that pairs an EEG headset with a 3-D theatrical flying harness, allowing users to “fly” by controlling their thoughts. The “Infinity Simulator” will make its debut with an art installation [The Ascent] in which participants rise into the air – and trigger light, sound, and video effects – by calming their thoughts.
I’ve seen this a few times and it still absolutely blows me away.
If you should be near Rensselaer on May 12, 2011, you could have a chance to fly using your own thoughtpower, a harness, and an EEG helmet. From the event webpage,
Come ride The Ascent, a playful mash-up of theatrics, gaming and mind-control. The Ascent is a live-action, theatrical ride experience created for almost anyone to try. Individual riders wear an EEG headset, which reads brainwaves, along with a waist harness, and by marshaling their calm, focus, and concentration, try to levitate themselves thirty feet into the air as a small audience watches from below. The experience is full of obstacles-as a rider ascends via the power of concentration, sound and light also respond to brain activity, creating a storm of stimuli that conspires to distract the rider from achieving the goal: levitating into “transcendence.” The paradox is that in order to succeed, you need to release your desire for achievement, and contend with what might be the biggest obstacle: yourself.
Theater Artist and Experience Designer Yehuda Duenyas (XXXY) presents his MFA Thesis project The Ascent, and its operating platform the Infinity System, a new user driven experience created specifically for EMPAC’s automated rigging system.
The Infinity System is a new platform and user interface for 3D flying which combines aspects of thrill-ride, live-action video game, and interactive installation.
Using a unique and intuitive interface, the Infinity System uses 3D rigging to move bodies creatively through space, while employing wearable sensors to manipulate audio and visual content.
Like a live-action stunt-show crossed with a video game, the user is given the superhuman ability to safely and freely fly, leap, bound, flip, run up walls, fall from great heights, swoop, buzz, drop, soar, and otherwise creatively defy gravity.
“The effect is nothing short of movie magic.” – Sean Hollister, Engadget
Here’s a brief description of the technology behind this ‘Ascent’ (from the news item on physorg.com),
Ten computer programs running simultaneously link the commercially available EEG headset to the computer-controlled 3-D flying harness and various theater systems, said Todd. [Michael Todd, a Rensselaer 2010 graduate in computer science]
Within the theater, the rigging – including the harness – is controlled by a Stage Tech NOMAD console; lights are controlled by an ION console running MIDI show control; sound through MAX/MSP; and video through Isadora and Jitter. The “Infinity Simulator,” a series of three C programs written by Todd, acts as intermediary between the headset and the theater systems, connecting and conveying all input and output.
“We’ve built a software system on top of the rigging control board and now have control of it through an iPad, and since we have the iPad control, we can have anything control it,” said Duenyas. “The ‘Infinity Simulator’ is the center; everything talks to the ‘Infinity Simulator.’”
This May 3, 2011 article (Mystery Man Gives Mind-Reading Tech More Early Cash Than Facebook, Google Combined) by Kit Eaton on Fast Company also concerns itself with a brain/computer interface. From the article,
Imagine the money that could be made by a drug company that accurately predicted and treated the onset of Alzheimer’s before any symptoms surfaced. That may give us an idea why NeuroVigil, a company specializing in non-invasive, wireless brain-recording tech, just got a cash injection that puts it at a valuation “twice the combined seed valuations of Google’s and Facebook’s first rounds,” according to a company announcement
NeuroVigil’s key product at the moment is the iBrain, a slim device in a flexible head-cap that’s designed to be worn for continuous EEG monitoring of a patient’s brain function–mainly during sleep. It’s non-invasive, and replaces older technology that could only access these kind of brain functions via critically implanted electrodes actually on the brain itself. The idea is, first, to record how brain function changes over time, perhaps as a particular combination of drugs is administered or to help diagnose particular brain pathologies–such as epilepsy.
But the other half of the potentailly lucrative equation is the ability to analyze the trove of data coming from iBrain. And that’s where NeuroVigil’s SPEARS algorithm enters the picture. Not only is the company simplifying collection of brain data with a device that can be relatively comfortably worn during all sorts of tasks–sleeping, driving, watching advertising–but the combination of iBrain and SPEARS multiplies the efficiency of data analysis [emphasis mine].
I assume it’s the notion of combining the two technologies (iBrian and SPEARS) that spawned the ‘mind-reading’ part of this article’s title. The technology could be used for early detection and diagnosis, as well as, other possibilities as Eaton notes,
It’s also possible it could develop its technology into non-medicinal uses such as human-computer interfaces–in an earlier announcement, NeuroVigil noted, “We plan to make these kinds of devices available to the transportation industry, biofeedback, and defense. Applications regarding pandemics and bioterrorism are being considered but cannot be shared in this format.” And there’s even a popular line of kid’s toys that use an essentially similar technique, powered by NeuroSky sensors–themselves destined for future uses as games console controllers or even input devices for computers.
What these two technologies have in common is that, in some fashion or other, they have (shy of implanting a computer chip) a relatively direct interface with our brains, which means (to me anyway) a very different relationship between humans and computers.
In the next couple of items I’m going to profile a couple of very similar to each other technologies that allow for more traditional human/computer interactions, one of which I’ve posted about previously, the Nokia Morph (most recently in my Sept. 29, 2010 posting).
It was first introduced as a type of flexible phone with other capabilities. Since then, they seem to have elaborated on those capabilities. Here’s a description of what they now call the ‘Morph concept’ in a [ETA May 12, 2011: inserted correct link information] May 4, 2011 news item on Nanowerk,
Morph is a joint nanotechnology concept developed by Nokia Research Center (NRC) and the University of Cambridge (UK). Morph is a concept that demonstrates how future mobile devices might be stretchable and flexible, allowing the user to transform their mobile device into radically different shapes. It demonstrates the ultimate functionality that nanotechnology might be capable of delivering: flexible materials, transparent electronics and self-cleaning surfaces.
Morph, will act as a gateway. It will connect the user to the local environment as well as the global internet. It is an attentive device that adapts to the context – it shapes according to the context. The device can change its form from rigid to flexible and stretchable. Buttons of the user interface can grow up from a flat surface when needed. User will never have to worry about the battery life. It is a device that will help us in our everyday life, to keep our self connected and in shape. It is one significant piece of a system that will help us to look after the environment.
Without the new materials, i.e. new structures enabled by the novel materials and manufacturing methods it would be impossible to build Morph kind of device. Graphene has an important role in different components of the new device and the ecosystem needed to make the gateway and context awareness possible in an energy efficient way.
Graphene will enable evolution of the current technology e.g. continuation of the ever increasing computing power when the performance of the computing would require sub nanometer scale transistors by using conventional materials.
For someone who’s been following news of the Morph for the last few years, this news item doesn’t give you any new information. Still, it’s nice to be reminded of the Morph project. Here’s a video produced by the University of Cambridge that illustrates some of the project’s hopes for the Morph concept,
While the folks at the Nokia Research Centre and University of Cambridge have been working on their project, it appears the team at the Human Media Lab at the School of Computing at Queen’s University (Kingston, Ontario, Canada) in cooperation with a team from Arizona State University and E Ink Corporation have been able to produce a prototype of something remarkably similar, albeit with fewer functions. The PaperPhone is being introduced at the Association of Computing Machinery’s CHI 2011 (Computer Human Interaction) conference in Vancouver, Canada next Tuesday, May 10, 2011.
The world’s first interactive paper computer is set to revolutionize the world of interactive computing.
“This is the future. Everything is going to look and feel like this within five years,” says creator Roel Vertegaal, the director of Queen’s University Human Media Lab,. “This computer looks, feels and operates like a small sheet of interactive paper. You interact with it by bending it into a cell phone, flipping the corner to turn pages, or writing on it with a pen.”
The smartphone prototype, called PaperPhone is best described as a flexible iPhone – it does everything a smartphone does, like store books, play music or make phone calls. But its display consists of a 9.5 cm diagonal thin film flexible E Ink display. The flexible form of the display makes it much more portable that any current mobile computer: it will shape with your pocket.
For anyone who knows the novel, it’s very Diamond Age (by Neal Stephenson). On a more technical note, I would have liked more information about the display’s technology. What is E Ink using? Graphene? Carbon nanotubes?
(That does not look like to paper to me but I suppose you could call it ‘paperlike’.)
In reviewing all these news items, it seems to me there are two themes, the computer as bodywear and the computer as an extension of our thoughts. Both of these are more intimate relationships, the latter far more so than the former, than we’ve had with the computer till now. If any of you have any thoughts on this, please do leave a comment as I would be delighted to engage on some discussion about this.
You can get more information about the Association of Computing Machinery’s CHI 2011 (Computer Human Interaction) conference where Dr. Vertegaal will be presenting here.
You can find more about Dr. Vertegaal and the Human Media Lab at Queen’s University here.
The academic paper being presented at the Vancouver conference is here.
Also, if you are interested in the hardware end of things, you can check out E Ink Corporation, the company that partnered with the team from Queen’s and Arizona State University to create the PaperPhone. Interestingly, E Ink is a spin off company from the Massachusetts Institute of Technology (MIT).