Scalable, high-performance ML toolboxes designed to deliver a compact and broad sense of hearing

Our technology allows us to lead the sound recognition field both from a performance and compactness point of view.

Compared to Apple’s iOS accessibility feature, our technology delivers 56x better classification performance.

And when compared with MobileNet (a popular network architecture used for compact mobile and embedded ML applications) our network architecture requires 144x fewer FLOPS (floating point operations per second).

Data: The bedrock of good audio ML

To teach our technology to accurately model sounds, we have to expose it to high quality, real-world data. Quantity matters, but high-performance is also about relevance and diversity. Read more in our whitepaper.

We record sounds either in our dedicated anechoic Sound Labs, through our network of volunteers, or via our dedicated data collection teams. It means that we know everything about the data we use to train our models.

Our approaches to data collection and management are also in-line with global legal and ethical best practice, which means that our customers can confidently deploy our technology into the global market. Read more in our whitepaper.

Alexandria™ contains 30,000,000 labelled recordings across 1,000 sound classes, and 400,000,000 meta data points.

Alexandria™ – The world’s largest, commercially exploitable audio dataset

Alexandria™ provides us with a rich platform from which to build and manage the data necessary for giving machines a compact sense of hearing.

By leveraging the vast amount of data in Alexandria™ and by using our Acoustically Smart Temporal Augmentation techniques, we can construct acoustically accurate polyphonic soundscapes that represent a wide range of real-world scenarios which would be practically impossible to collect.

AuditoryNET™ – Sophisticated and optimized DNNs built for hearing

AuditoryNET™ is a range of compact, topologically-optimized DNNs, associated training frameworks and toolboxes designed specifically for the formal ML tasks within sound recognition (sound event detection, scene recognition and tagging).

Giving machines a broad sense of hearing beyond speech and music presents a number of specific challenges. This has led to many breakthroughs including:

Comparing network size and computational load

To illustrate the compactness of our technology we’ve compared one of our models from AuditoryNET™ (recognizing six sound classes) with two network architectures used in the research arena: MobileNet (recognizing six sound classes) and ResNet (recognizing 10 sound classes). Both are familiar to a broad range of people working in machine learning and both selected network architectures are also often used by groups entering the annual academic DCASE Challenge for sound recognition tasks.

ai3™ and ai3-nano™ – Proven, ultra-compact and flexible software platforms

Our customers license ai3™ and ai3-nano™ which give a wide range of consumer products a sense of hearing and require as little as 40kB of RAM and ~1mA of power.

Its flexible, scalable architecture means that it can:

  • Be adapted to customer hardware
  • Support technologies such as our patented post-biasing, which can adapt to contextual information based on other sensor data
  • Take advantage of SoC acceleration.

Find out more

Polyphonic Sound Detection Score

First published and presented at ICASSP 2020, the Polyphonic Sound Detection Score (PSDS) is an industry-standard evaluation framework and metric for polyphonic sound recognition systems designed by Audio Analytic and based on our extensive expertise in real-world sound recognition.

PSDS solves the fundamental shortcomings of previous evaluation approaches, and we believe that the wider sound recognition community will benefit from our expertise: feel free to access the source code and benefit from using PSDS in your own work.

Take a look

Our patents

We have a portfolio of 31 patents (14 granted, 13 published, 4 pending), covering the uses of sound recognition in products as well as the specialist techniques that we use to give machines a sense of hearing.

The Cortex M0+ challenge: How low can we go?

So we set ourselves a really complicated hardware challenge with exciting potential…. to see if we can embed our class-leading software on the M0+ based chip.

Latest news

Audio Analytic named in CB Insights AI 100 List of Most Innovative Artificial Intelligence Startups

Read more

Latest news

Why real sounds matter for machine learning

Read more

Latest news

Patent US20200329330 – Improvements in sound reproduction

Read more

We’re proud to sponsor DCASE 2021

Read more

Consumer research shows strong demand for smart(er)phones that react to the world around them

Read more

We are using our own and third party cookies which track your use of our website to help give you the best experience. By continuing, we’ll assume that you are happy to receive all cookies on our website.

You can check what cookies we use and how to manage them here and you read our privacy policy here.

Accept and close