Tag Archives: wearable sensors

If only AI had a brain (a Wizard of Oz reference?)

The title, which I’ve borrowed from the news release, is the only Wizard of Oz reference that I can find but it works so well, you don’t really need anything more.

Moving onto the news, a July 23, 2018 news item on phys.org announces new work on developing an artificial synapse (Note: A link has been removed),

Digital computation has rendered nearly all forms of analog computation obsolete since as far back as the 1950s. However, there is one major exception that rivals the computational power of the most advanced digital devices: the human brain.

The human brain is a dense network of neurons. Each neuron is connected to tens of thousands of others, and they use synapses to fire information back and forth constantly. With each exchange, the brain modulates these connections to create efficient pathways in direct response to the surrounding environment. Digital computers live in a world of ones and zeros. They perform tasks sequentially, following each step of their algorithms in a fixed order.

A team of researchers from Pitt’s [University of Pittsburgh] Swanson School of Engineering have developed an “artificial synapse” that does not process information like a digital computer but rather mimics the analog way the human brain completes tasks. Led by Feng Xiong, assistant professor of electrical and computer engineering, the researchers published their results in the recent issue of the journal Advanced Materials (DOI: 10.1002/adma.201802353). His Pitt co-authors include Mohammad Sharbati (first author), Yanhao Du, Jorge Torres, Nolan Ardolino, and Minhee Yun.

A July 23, 2018 University of Pittsburgh Swanson School of Engineering news release (also on EurekAlert), which originated the news item, provides further information,

“The analog nature and massive parallelism of the brain are partly why humans can outperform even the most powerful computers when it comes to higher order cognitive functions such as voice recognition or pattern recognition in complex and varied data sets,” explains Dr. Xiong.

An emerging field called “neuromorphic computing” focuses on the design of computational hardware inspired by the human brain. Dr. Xiong and his team built graphene-based artificial synapses in a two-dimensional honeycomb configuration of carbon atoms. Graphene’s conductive properties allowed the researchers to finely tune its electrical conductance, which is the strength of the synaptic connection or the synaptic weight. The graphene synapse demonstrated excellent energy efficiency, just like biological synapses.

In the recent resurgence of artificial intelligence, computers can already replicate the brain in certain ways, but it takes about a dozen digital devices to mimic one analog synapse. The human brain has hundreds of trillions of synapses for transmitting information, so building a brain with digital devices is seemingly impossible, or at the very least, not scalable. Xiong Lab’s approach provides a possible route for the hardware implementation of large-scale artificial neural networks.

According to Dr. Xiong, artificial neural networks based on the current CMOS (complementary metal-oxide semiconductor) technology will always have limited functionality in terms of energy efficiency, scalability, and packing density. “It is really important we develop new device concepts for synaptic electronics that are analog in nature, energy-efficient, scalable, and suitable for large-scale integrations,” he says. “Our graphene synapse seems to check all the boxes on these requirements so far.”

With graphene’s inherent flexibility and excellent mechanical properties, these graphene-based neural networks can be employed in flexible and wearable electronics to enable computation at the “edge of the internet”–places where computing devices such as sensors make contact with the physical world.

“By empowering even a rudimentary level of intelligence in wearable electronics and sensors, we can track our health with smart sensors, provide preventive care and timely diagnostics, monitor plants growth and identify possible pest issues, and regulate and optimize the manufacturing process–significantly improving the overall productivity and quality of life in our society,” Dr. Xiong says.

The development of an artificial brain that functions like the analog human brain still requires a number of breakthroughs. Researchers need to find the right configurations to optimize these new artificial synapses. They will need to make them compatible with an array of other devices to form neural networks, and they will need to ensure that all of the artificial synapses in a large-scale neural network behave in the same exact manner. Despite the challenges, Dr. Xiong says he’s optimistic about the direction they’re headed.

“We are pretty excited about this progress since it can potentially lead to the energy-efficient, hardware implementation of neuromorphic computing, which is currently carried out in power-intensive GPU clusters. The low-power trait of our artificial synapse and its flexible nature make it a suitable candidate for any kind of A.I. device, which would revolutionize our lives, perhaps even more than the digital revolution we’ve seen over the past few decades,” Dr. Xiong says.

There is a visual representation of this artificial synapse,

Caption: Pitt engineers built a graphene-based artificial synapse in a two-dimensional, honeycomb configuration of carbon atoms that demonstrated excellent energy efficiency comparable to biological synapses Credit: Swanson School of Engineering

Here’s a link to and a citation for the paper,

Low‐Power, Electrochemically Tunable Graphene Synapses for Neuromorphic Computing by Mohammad Taghi Sharbati, Yanhao Du, Jorge Torres, Nolan D. Ardolino, Minhee Yun, Feng Xiong. Advanced Materials DOP: https://doi.org/10.1002/adma.201802353 First published [online]: 23 July 2018

This paper is behind a paywall.

I did look at the paper and if I understand it rightly, this approach is different from the memristor-based approaches that I have so often featured here. More than that I cannot say.

Finally, the Wizard of Oz song ‘If I Only Had a Brain’,

I’ve got my eye on you (tissue paper sensors from the University of Washington [state])

University of Washington graduate student, Jinyuan Zhang, demonstrates how wearable sensors can track eye movement. Dennis R. Wise/University of Washington

A February 13, 2018 news item on phys.org announces a technology that can transform tissue paper into a sensor,

University of Washington engineers have turned tissue paper – similar to toilet tissue – into a new kind of wearable sensor that can detect a pulse, a blink of an eye and other human movement. The sensor is light, flexible and inexpensive, with potential applications in health care, entertainment and robotics.

A February 12, 2018 University of Washington news release (also on EurekAlert) by Jackson Holtz, which originated the news item, provides more detail (Note: Links have been removed),

The technology, described in a paper published in January in the journal Advanced Materials Technologies, shows that by tearing tissue paper that’s loaded with nanocomposites and breaking the paper’s fibers, the paper acts as a sensor. It can detect a heartbeat, finger force, finger movement, eyeball movement and more, said Jae-Hyun Chung, a UW associate professor of mechanical engineering and senior author of the research.

Finger sensor

University of Washington graduate student, Jinyuan Zhang, demonstrates how a wearable sensor can measure finger pressure.Dennis R. Wise/University of Washington

“The major innovation is a disposable wearable sensor made with cheap tissue paper,” said Chung. “When we break the specimen, it will work as a sensor.”

These small, Band Aid-sized sensors could have a variety of applications in various fields. For example, monitoring a person’s gait or the movement of their eyes can be used to inspect brain function or a game player’s actions. The sensor could track how a special-needs child walks in a home test, sparing the child the need for hospital visits. Or the sensors could be used in occupational therapy for seniors.

“They can use these sensors and after one-time use, they can be thrown away,” said Chung.

In their research, the scientists used paper similar to toilet tissue. The paper – nothing more than conventional paper towels – is then doused with carbon nanotube-laced water. Carbon nanotubes are tiny materials that create electrical conductivity. Each piece of tissue paper has both horizontal and vertical fibers, so when the paper is torn, the direction of the tear informs the sensor of what’s happened. To trace eye movement, they’re attached to a person’s reading glasses.

For now, the work has been contained to a laboratory, and researchers are hoping to find a suitable commercial use. A provisional patent was filed in December 2017.

The study was funded partially by Samsung Research America through the Think Tank Team Award.

Foot sensor

University of Washington mechanical engineering undergraduate, Yared Shella, demonstrates how foot pressure is measured with a wearable sensor.Dennis R. Wise/University of Washington

Here’s a link to and a citation for the paper,

Fracture-Induced Mechanoelectrical Sensitivities of Paper-Based Nanocomposites by Jinyuan Zhang, Gil-Yong Lee, Chiew Cerwyn, Jinkyu Yang, Fabrice Fondjo, Jong-Hoon Kim, Minoru Taya, Dayong Gao, and Jae-Hyun Chung. Advanced Materials Technologies Vol. 3 Issue 1 DOI: 10.1002/admt.201700266 Version of Record online: 26 JAN 2018

© 2018 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim

This paper is behind a paywall.

Humans can distinguish molecular differences by touch

Yesterday, in my December 18, 2017 post about medieval textiles, I posed the question, “How did medieval artisans create nanoscale and microscale gilding when they couldn’t see it?” I realized afterwards that an answer to that question might be in this December 13, 2017 news item on ScienceDaily,

How sensitive is the human sense of touch? Sensitive enough to feel the difference between surfaces that differ by just a single layer of molecules, a team of researchers at the University of California San Diego has shown.

“This is the greatest tactile sensitivity that has ever been shown in humans,” said Darren Lipomi, a professor of nanoengineering and member of the Center for Wearable Sensors at the UC San Diego Jacobs School of Engineering, who led the interdisciplinary project with V. S. Ramachandran, director of the Center for Brain and Cognition and distinguished professor in the Department of Psychology at UC San Diego.

So perhaps those medieval artisans were able to feel the difference before it could be seen in the textiles they were producing?

Getting back to the matter at hand, a December 13, 2017 University of California at San Diego (UCSD) news release (also on EurekAlert) by Liezel Labios offers more detail about the work,

Humans can easily feel the difference between many everyday surfaces such as glass, metal, wood and plastic. That’s because these surfaces have different textures or draw heat away from the finger at different rates. But UC San Diego researchers wondered, if they kept all these large-scale effects equal and changed only the topmost layer of molecules, could humans still detect the difference using their sense of touch? And if so, how?

Researchers say this fundamental knowledge will be useful for developing electronic skin, prosthetics that can feel, advanced haptic technology for virtual and augmented reality and more.

Unsophisticated haptic technologies exist in the form of rumble packs in video game controllers or smartphones that shake, Lipomi added. “But reproducing realistic tactile sensations is difficult because we don’t yet fully understand the basic ways in which materials interact with the sense of touch.”

“Today’s technologies allow us to see and hear what’s happening, but we can’t feel it,” said Cody Carpenter, a nanoengineering Ph.D. student at UC San Diego and co-first author of the study. “We have state-of-the-art speakers, phones and high-resolution screens that are visually and aurally engaging, but what’s missing is the sense of touch. Adding that ingredient is a driving force behind this work.”

This study is the first to combine materials science and psychophysics to understand how humans perceive touch. “Receptors processing sensations from our skin are phylogenetically the most ancient, but far from being primitive they have had time to evolve extraordinarily subtle strategies for discerning surfaces—whether a lover’s caress or a tickle or the raw tactile feel of metal, wood, paper, etc. This study is one of the first to demonstrate the range of sophistication and exquisite sensitivity of tactile sensations. It paves the way, perhaps, for a whole new approach to tactile psychophysics,” Ramachandran said.

Super-Sensitive Touch

In a paper published in Materials Horizons, UC San Diego researchers tested whether human subjects could distinguish—by dragging or tapping a finger across the surface—between smooth silicon wafers that differed only in their single topmost layer of molecules. One surface was a single oxidized layer made mostly of oxygen atoms. The other was a single Teflon-like layer made of fluorine and carbon atoms. Both surfaces looked identical and felt similar enough that some subjects could not differentiate between them at all.

According to the researchers, human subjects can feel these differences because of a phenomenon known as stick-slip friction, which is the jerking motion that occurs when two objects at rest start to slide against each other. This phenomenon is responsible for the musical notes played by running a wet finger along the rim of a wine glass, the sound of a squeaky door hinge or the noise of a stopping train. In this case, each surface has a different stick-slip frequency due to the identity of the molecules in the topmost layer.

In one test, 15 subjects were tasked with feeling three surfaces and identifying the one surface that differed from the other two. Subjects correctly identified the differences 71 percent of the time.

In another test, subjects were given three different strips of silicon wafer, each strip containing a different sequence of 8 patches of oxidized and Teflon-like surfaces. Each sequence represented an 8-digit string of 0s and 1s, which encoded for a particular letter in the ASCII alphabet. Subjects were asked to “read” these sequences by dragging a finger from one end of the strip to the other and noting which patches in the sequence were the oxidized surfaces and which were the Teflon-like surfaces. In this experiment, 10 out of 11 subjects decoded the bits needed to spell the word “Lab” (with the correct upper and lowercase letters) more than 50 percent of the time. Subjects spent an average of 4.5 minutes to decode each letter.

“A human may be slower than a nanobit per second in terms of reading digital information, but this experiment shows a potentially neat way to do chemical communications using our sense of touch instead of sight,” Lipomi said.

Basic Model of Touch

The researchers also found that these surfaces can be differentiated depending on how fast the finger drags and how much force it applies across the surface. The researchers modeled the touch experiments using a “mock finger,” a finger-like device made of an organic polymer that’s connected by a spring to a force sensor. The mock finger was dragged across the different surfaces using multiple combinations of force and swiping velocity. The researchers plotted the data and found that the surfaces could be distinguished given certain combinations of velocity and force. Meanwhile, other combinations made the surfaces indistinguishable from each other.

“Our results reveal a remarkable human ability to quickly home in on the right combinations of forces and swiping velocities required to feel the difference between these surfaces. They don’t need to reconstruct an entire matrix of data points one by one as we did in our experiments,” Lipomi said.

“It’s also interesting that the mock finger device, which doesn’t have anything resembling the hundreds of nerves in our skin, has just one force sensor and is still able to get the information needed to feel the difference in these surfaces. This tells us it’s not just the mechanoreceptors in the skin, but receptors in the ligaments, knuckles, wrist, elbow and shoulder that could be enabling humans to sense minute differences using touch,” he added.

This work was supported by member companies of the Center for Wearable Sensors at UC San Diego: Samsung, Dexcom, Sabic, Cubic, Qualcomm and Honda.

For those who prefer their news by video,

Here’s a link to and a citation for the paper,

Human ability to discriminate surface chemistry by touch by Cody W. Carpenter, Charles Dhong, Nicholas B. Root, Daniel Rodriquez, Emily E. Abdo, Kyle Skelil, Mohammad A. Alkhadra, Julian Ramírez, Vilayanur S. Ramachandran and Darren J. Lipomi. Mater. Horiz., 2018, Advance Article DOI: 10.1039/C7MH00800G

This paper is open access but you do need to have opened a free account on the website.