Researchers Create First-Ever Personalised Sound Projector With £10 Webcam

A University of Sussex research team has demonstrated the first sound projector that can track a moving individually and deliver an acoustic message as they move, to a high-profile tech and media conference in LA.

Dr. Gianluca Memoli and his colleagues demonstrated what they believe to be the world's first sound projector with an autozoom objective in a talk at the 46th International Conference and Exhibition on Computer Graphics & Interactive Techniques (SIGGRAPH 2019) this week.

Dr. Memoli, Lecturer in Novel Interfaces and Interactions at the University of Sussex's School of Engineering and Informatics who led the research, said: "By designing acoustic materials at a scale smaller than the wavelength of the sound to create thin acoustic lenses, the sky is the limit in new potential acoustic applications.

"Centuries of optical design can now be applied to acoustics. We believe this technology can be harnessed for plenty of positive applications including personalized alarm messages in a crowd, immersive experiences without headphones, the audio equivalent of special effects."

The system works with an in-house face-tracking software which is used to pilot an Arduino-controlled acoustic telescope to focus sound on a moving target. The low-cost camera is able to track a person and command the distance between two acoustic lenses, delivering a sphere of sound around 6cm in diameter in front of the target, which then responds to the individual's movement.

Joshua Kybett, the second-year undergraduate at Sussex who designed the tracking, adds: "Since acoustic lenses can be 3D-printed for only £100, we wanted a tracking technique that worked on a similarly low budget. With a £10 webcam, this is one-tenth of standard tracking systems.

"In addition, our method has been designed to require user consent in order to function. This requirement ensures the technology cannot be used intrusively, nor deliver sound to an unwilling audience."

Thomas Graham, the research fellow in the School of Engineering and Informatics who run the measurements and the simulations, says: "In our study, we were inspired by autozoom cameras that extend their objectives to match the distance of a target.

"We used a very similar system, with even the same mechanical sound of the motor. I believe our work is also the first step towards hand-held, low-cost acoustic cameras."

The research team is now working to expand the capabilities of the system beyond tracking for just one direction and over one octave, to ensure it can be scaled up to cover most speech and basic melodies and eventually to deliver a full piece of music.

Arash PourYazdan, who designed the electronics, said: "SIGGRAPH is a place where emerging and futuristic ideas are discussed. This is the conference where entertainment giants such as Disney, Marvel, and Microsoft meet to share their visions: it was the perfect place for us to demonstrate how we think sound might be managed in the future." 

© 2024 iTech Post All rights reserved. Do not reproduce without permission.

Tags

Company from iTechPost

More from iTechPost