Wearable devices such as smartwatches, TWS earbuds and AR glasses that can combine sound recognition and movement detection will open up opportunities to deliver powerful new and improved user experiences. These experiences range from a more granular understanding of physical activity to helping people navigate or discover the world around them.

The combination of sound recognition and movement is the subject of a new patent, that was granted to us this week.

Sometimes the readings from motion sensors suffer from ambiguity when it comes to guessing what people are doing. However, if we want our smart devices to be more helpful and act as frictionless, hands-free, assistive interfaces, then compact, embedded AI needs to understand what people are doing as precisely as possible. In turn, this triggers the right kind of user experience at the right moment and in the proper context.

As humans, we infer the context around us by integrating multiple sensory perceptions, and we do that so naturally that we even forget that we are doing it. The sounds around us provide rich information that enables us to better understand context.

By empowering consumer devices to combine multiple senses, consumers benefit from products that can do more on their behalf, whether the application is health, wellbeing, convenience, safety or entertainment.

Our patent covers multiple use cases, including:

  • Your wearable device can more accurately detect the sport you are playing by combining movement and sound (e.g. sprinting vs soccer, tennis vs squash)
  • Your smartwatch can detect whether you are washing your hands by combining its understanding of the movement of your hands and the sound of running water
  • Your AR glasses can combine your movement and the environmental sounds around you to anticipate your needs better, keep you safe or deliver context-relevant information about your location or activity.

Sound recognition is an essential piece of machine perception that was missing from the perceptual AI puzzle until we overcame the significant challenges presented by this specialised branch of AI. As a result, product designers now see how essential the sense of hearing is to context recognition and the role it plays alongside other types of AI. This is because in addition to enabling valuable and unique applications in its own right, sound recognition enhances the value of other sensory inputs by providing critical contextual cues.

I’m very excited by the wide range of new user experiences that this type of ‘sound +’ sensor fusion will bring to consumer electronics.

*****

Like this? You can subscribe to our blog and receive an alert every time we publish an announcement, a comment on the industry or something more technical. 

 

 About Audio Analytic 

Audio Analytic is the pioneer of AI sound recognition technology. The company is on a mission to give machines a compact sense of hearing. This empowers them with the ability to react to the world around us, helping satisfy our entertainment, safety, security, wellbeing, convenience, and communication needs across a huge range of consumer products.

Audio Analytic’s ai3™ and ai3-nano™ sound recognition software enables device manufacturers to equip products at the edge with the ability to recognize and automatically respond to our growing list of sounds and acoustic scenes.

We are using our own and third party cookies which track your use of our website to help give you the best experience. By continuing, we’ll assume that you are happy to receive all cookies on our website.

You can check what cookies we use and how to manage them here and you read our privacy policy here.

Accept and close
>