Tag Archives: audio speakers

Shape-changing speaker (aka acoustic swarms) for sound control

To alleviate any concerns, these swarms are not kin to Michael Crichton’s swarms in his 2002 novel, Prey or his 2011 novel, Micro (published after his death).

A September 21, 2023 news item on ScienceDaily announces this ‘acoustic swarm’ research,

In virtual meetings, it’s easy to keep people from talking over each other. Someone just hits mute. But for the most part, this ability doesn’t translate easily to recording in-person gatherings. In a bustling cafe, there are no buttons to silence the table beside you.

The ability to locate and control sound — isolating one person talking from a specific location in a crowded room, for instance — has challenged researchers, especially without visual cues from cameras.

A team led by researchers at the University of Washington has developed a shape-changing smart speaker, which uses self-deploying microphones to divide rooms into speech zones and track the positions of individual speakers. With the help of the team’s deep-learning algorithms, the system lets users mute certain areas or separate simultaneous conversations, even if two adjacent people have similar voices. Like a fleet of Roombas, each about an inch in diameter, the microphones automatically deploy from, and then return to, a charging station. This allows the system to be moved between environments and set up automatically. In a conference room meeting, for instance, such a system might be deployed instead of a central microphone, allowing better control of in-room audio.

The team published its findings Sept. 21 [2023] in Nature Communications.

A September 21, 2023 University of Washington (state) news release (also on EurekAlert), which originated the news item, delves further into the work, Note: Links have been removed,

“If I close my eyes and there are 10 people talking in a room, I have no idea who’s saying what and where they are in the room exactly. That’s extremely hard for the human brain to process. Until now, it’s also been difficult for technology,” said co-lead author Malek Itani, a UW doctoral student in the Paul G. Allen School of Computer Science & Engineering. “For the first time, using what we’re calling a robotic ‘acoustic swarm,’ we’re able to track the positions of multiple people talking in a room and separate their speech.”

Previous research on robot swarms has required using overhead or on-device cameras, projectors or special surfaces. The UW team’s system is the first to accurately distribute a robot swarm using only sound.

The team’s prototype consists of seven small robots that spread themselves across tables of various sizes. As they move from their charger, each robot emits a high frequency sound, like a bat navigating, using this frequency and other sensors to avoid obstacles and move around without falling off the table. The automatic deployment allows the robots to place themselves for maximum accuracy, permitting greater sound control than if a person set them. The robots disperse as far from each other as possible since greater distances make differentiating and locating people speaking easier. Today’s consumer smart speakers have multiple microphones, but clustered on the same device, they’re too close to allow for this system’s mute and active zones.

“If I have one microphone a foot away from me, and another microphone two feet away, my voice will arrive at the microphone that’s a foot away first. If someone else is closer to the microphone that’s two feet away, their voice will arrive there first,” said co-lead author Tuochao Chen, a UW doctoral student in the Allen School. “We developed neural networks that use these time-delayed signals to separate what each person is saying and track their positions in a space. So you can have four people having two conversations and isolate any of the four voices and locate each of the voices in a room.”

The team tested the robots in offices, living rooms and kitchens with groups of three to five people speaking. Across all these environments, the system could discern different voices within 1.6 feet (50 centimeters) of each other 90% of the time, without prior information about the number of speakers. The system was able to process three seconds of audio in 1.82 seconds on average — fast enough for live streaming, though a bit too long for real-time communications such as video calls.

As the technology progresses, researchers say, acoustic swarms might be deployed in smart homes to better differentiate people talking with smart speakers. That could potentially allow only people sitting on a couch, in an “active zone,” to vocally control a TV, for example.

Researchers plan to eventually make microphone robots that can move around rooms, instead of being limited to tables. The team is also investigating whether the speakers can emit sounds that allow for real-world mute and active zones, so people in different parts of a room can hear different audio. The current study is another step toward science fiction technologies, such as the “cone of silence” in “Get Smart” and“Dune,” the authors write.

Of course, any technology that evokes comparison to fictional spy tools will raise questions of privacy. Researchers acknowledge the potential for misuse, so they have included guards against this: The microphones navigate with sound, not an onboard camera like other similar systems. The robots are easily visible and their lights blink when they’re active. Instead of processing the audio in the cloud, as most smart speakers do, the acoustic swarms process all the audio locally, as a privacy constraint. And even though some people’s first thoughts may be about surveillance, the system can be used for the opposite, the team says.

“It has the potential to actually benefit privacy, beyond what current smart speakers allow,” Itani said. “I can say, ‘Don’t record anything around my desk,’ and our system will create a bubble 3 feet around me. Nothing in this bubble would be recorded. Or if two groups are speaking beside each other and one group is having a private conversation, while the other group is recording, one conversation can be in a mute zone, and it will remain private.”

Takuya Yoshioka, a principal research manager at Microsoft, is a co-author on this paper, and Shyam Gollakota, a professor in the Allen School, is a senior author. The research was funded by a Moore Inventor Fellow award.

Two of the paper`s authors, Malek Itani and Tuochao Chen, have written a ‘Behind the Paper’ article for Nature.com’s Electrical and Electronic Engineering Community, from their September 21, 2023 posting,

Sound is a versatile medium. In addition to being one of the primary means of communication for us humans, it serves numerous purposes for organisms across the animal kingdom. Particularly, many animals use sound to localize themselves and navigate in their environment. Bats, for example, emit ultrasonic sound pulses to move around and find food in the dark. Similar behavior can be observed in Beluga whales to avoid obstacles and locate one other.

Various animals also have a tendency to cluster together into swarms, forming a unit greater than the sum of its parts. Famously, bees agglomerate into swarms to more efficiently search for a new colony. Birds flock to evade predators. These behaviors have caught the attention of scientists for quite some time, inspiring a handful of models for crowd control, optimization and even robotics. 

A key challenge in building robot swarms for practical purposes is the ability for the robots to localize themselves, not just within the swarm, but also relative to other important landmarks. …

Here’s a link to and a citation for the paper,

Creating speech zones with self-distributing acoustic swarms by Malek Itani, Tuochao Chen, Takuya Yoshioka & Shyamnath Gollakota. Nature Communications volume 14, Article number: 5684 (2023) DOI: https://doi.org/10.1038/s41467-023-40869-8 Published: 21 September 2023

This paper is open access.

Printing paper loudspeakers

When I was working on my undergraduate communications degree, we spent a fair chunk of time discussing the printed word; this introduction (below in the excerpt) brings back memories. I am going to start with an excerpt from the study (link and citation to follow at the end of this post) before moving on to the news item and press release. It’s a good introduction (Note Links have been removed),

For a long time, paper has been used as storing medium for written information only [emphasis mine]. In combination with the development of printing technologies, it became one of the most relevant materials as information could be reproduced multiple times and brought to millions of people in a simple, cheap, and fast way. However, with the digital revolution the end of paper has been forecasted.

However, paper still has its big advantages. The yearly production is still huge with over 400 million tons worldwide[1] for a wide application range going much beyond conventional books, newspapers, packages, or sanitary products. It is a natural light‐weight, flexible, recyclable, multi‐functional material making it an ideal candidate as part of novel electronic devices, especially based on printed electronics.[2] During the last decade, a wide variety of electronic functionalities have been demonstrated with paper as the common substrate platform. It has been used as basis for organic circuits,[3] microwave and digital electronics,[4] sensors,[5-7] actuators,[8, 9] and many more.

My first posting about this work from Chemnitz University of Technology with paper, loudspeakers, and printed electronics was a May 4, 2012 posting.

Enough of that trip down memory lane, a January 26, 2021 news item on Nanowerk announces research into printing loudspeakers onto roll-to-roll printed paper,

If the Institute for Print and Media Technology at Chemnitz University of Technology [Germany] has its way, many loudspeakers of the future will not only be as thin as paper, but will also sound impressive. This is a reality in the laboratories of the Chemnitz researchers, who back in 2015 developed the multiple award-winning T-Book – a large-format illustrated book equipped with printed electronics. If you turn a page, it begins to sound through a speaker invisibly located inside the sheet of paper.

“The T-Book was and is a milestone in the development of printed electronics, but development is continuing all the time,” says Prof. Dr. Arved C. Hübler, under whose leadership this technology trend, which is becoming increasingly important worldwide, has been driven forward for more than 20 years.

A January 26, 2021 Chemnitz University of Technology press release by Mario Steinebach/Translator: Chelsea Burris, which originated the news item, delves further into the topic,

From single-sheet production to roll-to-roll printing

Five years ago, the sonorous paper loudspeakers from Chemnitz were still manufactured in a semi-automatic single-sheet production process. In this process, ordinary paper or foils are printed with two layers of a conductive organic polymer as electrodes. A piezoelectric layer is sandwiched between them as the active element, which causes the paper or film to vibrate. Loud and clear sound is produced by air displacement. The two sides of the speaker paper can be printed in color. Since this was only possible in individual sheets in limited formats, the efficiency of this relatively slow manufacturing process is very low. That’s why researchers at the Institute of Print and Media Technology have been looking for a new way towards cost-effective mass production since May 2017.

The aim of their latest project, roll-to-roll printed speaker paper (T-Paper for short), was therefore to convert sheet production into roll production. “Researchers from the fields of print media technology, chemistry, physics, acoustics, electrical engineering, and economics from six nations developed a continuous, highly productive, and reliable roll production of loudspeaker webs,” reports project manager Georg C. Schmidt. Not only did they use the roll-to-roll (R2R) printing process for this, but they also developed inline technologies for other process steps, such as the lamination of functional layers. “This allows electronics to be embedded in the paper – invisibly and protected,” says Hübler. In addition, he says, inline polarization of piezoelectric polymer layers has been achieved for the first time and complete inline process monitoring of the printed functional layers is possible. The final project results were published in the renowned journal Advanced Materials in January 2021.

Long and lightweight paper loudspeaker webs for museums, the advertising industry, and Industry 4.0

The potential of loudspeaker paper was extended to other areas of application in the T-Paper project. For example, meter-long loudspeaker installations can now be manufactured in web form or as a circle (T-RING). “In our T-RING prototype, an almost four-meter-long track with 56 individual loudspeakers was connected to form seven segments and shaped into a circle, making a 360° surround sound installation possible,” says Schmidt. The speaker track, including printed circuitry, weighs just 150 grams and consists of 90 percent conventional paper that can be printed in color on both sides. “This means that low-cost infotainment solutions are now possible in museums, at trade shows and in the advertising industry, for example. In public buildings, for example, very homogeneous sound reinforcement of long stretches such as corridors is possible. But the process technology itself could also become interesting for other areas, such as the production of inline measurement systems for Industry 4.0,” says the project manager, looking to the future.

The T-Paper project was funded by the Federal Ministry of Education and Research from 2017 to 2020 with 1.37 million euros as part of the Validation of the technological and societal innovation potential of scientific research – VIP+ funding measure.

Here’s a link to and a citation for the paper,

Paper‐Embedded Roll‐to‐Roll Mass Printed Piezoelectric Transducers by Georg C. Schmidt, Pramul M. Panicker, Xunlin Qiu, Aravindan J. Benjamin, Ricardo A. Quintana Soler, Issac Wils, Arved C. Hübler. Advanced Materials DOI: https://doi.org/10.1002/adma.202006437 First published: 18 January 2021

This paper is open access.

For anyone curious about the T-Paper project, you can find it here.

Graphene speakers for mobile devices?

Credit: American Chemical Society

Credit: American Chemical Society

This speaker doesn’t look its ready for the marketplace yet but it does possess a rather nifty component, graphene aerogel. A Sept. 7, 2016 news item on phys.org briefly describes the scientists’ accomplishment,

Graphene has been hailed as a wonder material since it was first made more than a decade ago. It’s showing up in an increasing number of products, including coatings, sports equipment and even light bulbs. Now scientists are one step closer to making graphene audio speakers for mobile devices. They report in the journal ACS Applied Materials & Interfaces a simple way to fabricate once-elusive thermoacoustic speakers using the ultra-thin material.

A Sept. 7, 2016 American Chemical Society (ACS) news release (also on EurekAlert), which originated the news item, offers an explanation of how these new speakers differ from conventional speakers and a brief description of graphene aerogel (I once saw aerogel described as ‘solid smoke’),

Conventional speakers today rely on many mechanical parts that vibrate to create sound and must be encased in an acoustic cavity — essentially, in a box. But this approach complicates manufacturing and limits where listeners can put their speakers. Scientists have been pursuing ways around this by turning to a principle conceived of more than a century ago: thermoacoustics, the production of sound by rapidly heating and cooling a material rather than through vibrations. Science has caught up to this concept largely thanks to the development of graphene, which is highly conductive and durable. Some efforts to make graphene speakers have succeeded, but making them en masse would be challenging. Jung-Woo Choi, Byungjin Cho, Sang Ouk Kim and colleagues at Korea Advanced Institute of Science and Technology (KAIST) wanted to come up with a simpler approach.

The researchers developed a two-step (freeze-drying and reduction/doping) method for making a sound-emitting graphene aerogel. An array of 16 of these aerogels comprised a speaker that could operate on 40 Watts of power with a sound quality comparable to that of other graphene-based sound systems. The researchers say their fabrication method is practical and could lend itself to mass production for use in mobile devices and other applications. Because the speaker is thin and doesn’t vibrate, it could fit snugly against walls and even curved surfaces.

Here’s a link to and citation for the paper,

Application of N-Doped Three-Dimensional Reduced Graphene Oxide Aerogel to Thin Film Loudspeaker by Choong Sun Kim, Kyung Eun Lee, Jung-Min Lee, Sang Ouk Kim, Byung Jin Cho, and Jung-Woo Choi. ACS Appl. Mater. Interfaces, 2016, 8 (34), pp 22295–22300 DOI: 10.1021/acsami.6b03618 Publication Date (Web): August 17, 2016

Copyright © 2016 American Chemical Society

This paper is behind a paywall.