EspAudioSensor: Difference between revisions
(70 intermediate revisions by the same user not shown) | |||
Line 10: | Line 10: | ||
This project is about creating an audio level meter, for example as an environmental noise measuring node in a citizen science project or as a standalone "decibel meter". | This project is about creating an audio level meter, for example as an environmental noise measuring node in a citizen science project or as a standalone "decibel meter". | ||
Marcel Meek built a sound sensor based on an ESP32 board with built-in LoRa-chip. | |||
It samples audio from a digital I2S microphone, analyses it into sub-bands, then forwards it over LoRaWAN. | |||
His code can be found here: https://github.com/meekm/LoRaSoundkit | |||
From TheThingsNetwork, the data is picked up by a kind of 'integration' that sends it to sensor.community. | |||
This is handled by my [https://github.com/bertrik/sensor-data-bridge sensor-data-bridge noise forwarder] Java project. | |||
RIVM noise measuring project: | |||
* https://www.samenmetenaanluchtkwaliteit.nl/nieuws/citizen-science-geluidmetertest-op-rivm-terrein | |||
I would like to build a case like that! | |||
Github links to the project of Marcel Meek / TTN Apeldoorn: | |||
* https://github.com/TTNApeldoorn/sound-sensor | |||
* https://github.com/meekm/LoRaSoundkit | |||
Links back: | |||
* https://iotassistant.io/esp32/smart-door-bell-noise-meter-using-fft-esp32/ | |||
* https://en.usini.eu/digital_microphone_on_esp32/ | |||
=== TODO === | |||
* figure out which LoRaSoundkit is "current" (most up-to-date) -> this one: https://github.com/meekm/LoRaSoundkit | |||
* figure out alternative pinning for LoRaSoundkit so it works for both TTGO and Heltec boards | |||
* put it together! | |||
=== Ideas === | |||
==== Use interface similar to casper.aero ==== | |||
See https://rtm.flighttracking.casper.aero/ | |||
==== Bluetooth interface ==== | |||
The ESP32 supports bluetooth as an RF technology. It would be nice to be able to quickly see the current/most recently measured data. This can be shown on the display. Perhaps even more convenient, this can be exposed over bluetooth for display on a mobile phone. Are there android apps that can interface with bluetooth? | |||
The Bluetooth specification reserves 0x27C3 as the measurement unit code for sound pressure (dB) | |||
and 0x2BE4 as the noise characteristic. | |||
How the BLE indicate and notify work, and the difference between them: https://openlabpro.com/guide/ble-notify-on-esp32-controller/ | |||
The characteristic 'noise' is not one of the characteristics permitted for the 'environmental sensing service', see | |||
https://bitbucket.org/bluetooth-SIG/public/src/main/assigned_numbers/profiles_and_services/ess/ess_permitted_characteristics.yaml | |||
so this probably has to be specified in a custom service | |||
== Measuring audio as a citizen science project == | == Measuring audio as a citizen science project == | ||
Line 39: | Line 53: | ||
* A small box containing a microphone outside your house measures environmental sounds (traffic, etc), for example it takes a 1 second audio recording every 10 seconds | * A small box containing a microphone outside your house measures environmental sounds (traffic, etc), for example it takes a 1 second audio recording every 10 seconds | ||
* A spectral analysis is made of the audio (separating it in different frequency bands), calculating sound intensity for each individual band | * A spectral analysis is made of the audio (separating it in different frequency bands), calculating sound intensity for each individual band | ||
* Every 5 minutes, the measured sound intensity is sent to a central server on the internet | * Every 5 minutes, the measured sound intensity is sent to a central server on the internet. Because we only communicate intensities, this does not reveal audio fragments (conversations for example). | ||
* The central server takes the measured intensities and can do corrections, like a microphone specific correction, or a correction to apply A-weighting | * The central server takes the measured intensities and can do corrections, like a microphone specific correction, or a correction to apply A-weighting | ||
* The central server | * The central server forwards the measurements to sensor.community, from there they are automatically picked up by RIVM | ||
External things to investigate: | External things to investigate: | ||
* How is audio being evaluated in general, are there norms for it? | * How is audio being evaluated in general, are there norms for it? | ||
** how often should audio be sampled? 1 second/minute? 10 seconds/min? | ** how often should audio be sampled? 1 second/minute? 10 seconds/min? How long should the audio fragment be? | ||
** is it a valid assumption to use full octaves? or should the spectrum be divided in smaller parts? | ** is it a valid assumption to use full octaves? or should the spectrum be divided in smaller parts? | ||
* https://www.rijksoverheid.nl/onderwerpen/geluidsoverlast/geluidsoverlast-in-de-wet | * https://www.rijksoverheid.nl/onderwerpen/geluidsoverlast/geluidsoverlast-in-de-wet | ||
== Theory == | == Theory == | ||
The | The audio spectrum is divided up into octaves and the total energy in each octave is calculated. | ||
We can then easily apply sensor/housing specific corrections, do A weighting, etc. | We can then easily apply sensor/housing specific corrections, do A weighting, etc. | ||
Line 108: | Line 120: | ||
|} | |} | ||
[ | [https://www.nti-audio.com/en/support/know-how/frequency-weightings-for-sound-level-measurements List of A-weighting coefficients] | ||
The sampling rate is chosen such that the FFT bands align exactly with the bands as defined in the A-norm. | The sampling rate is chosen such that the FFT bands align exactly with the bands as defined in the A-norm. | ||
Line 121: | Line 133: | ||
Investigation into FFT libraries: | Investigation into FFT libraries: | ||
* [https://github.com/kosme/arduinoFFT arduinoFFT] | * [https://github.com/kosme/arduinoFFT arduinoFFT] does all calculations in double, severely limiting the audio buffer size. For each sample, you need two doubles, so that's 16 bytes per sample. I can use about 6000 samples in the audio buffer. Using floats gives me twice the audio buffer size. | ||
* Alternatively, [https://github.com/fakufaku/esp32-fft esp32-fft] looks nice, it uses floats instead of doubles and is optimised for esp32, but it uses malloc internally ... Also I can't use it as an Arduino library because it doesn't have the arduino library structure (with a src dir, examples dir, library.json file, library.properties file, etc) Maybe I can still use it by copying it in my sketch. | * Alternatively, [https://github.com/fakufaku/esp32-fft esp32-fft] looks nice, it uses floats instead of doubles and is optimised for esp32, but it uses malloc internally ... Also I can't use it as an Arduino library because it doesn't have the arduino library structure (with a src dir, examples dir, library.json file, library.properties file, etc) Maybe I can still use it by copying it in my sketch. | ||
Line 127: | Line 139: | ||
* Overview of octave bands https://www.engineeringtoolbox.com/octave-bands-frequency-limits-d_1602.html | * Overview of octave bands https://www.engineeringtoolbox.com/octave-bands-frequency-limits-d_1602.html | ||
* http://www.robinscheibler.org/2017/12/12/esp32-fft.html | * http://www.robinscheibler.org/2017/12/12/esp32-fft.html | ||
-> I went with ArduinoFFT, modified it to use float instead of double | |||
=== Decibel meters === | === Decibel meters === | ||
Line 146: | Line 160: | ||
Myself, I ordered a couple of [https://aliexpress.com/item/INMP441/32960945048.html INMP441 microphones] from Aliexpress. | Myself, I ordered a couple of [https://aliexpress.com/item/INMP441/32960945048.html INMP441 microphones] from Aliexpress. | ||
The microphone is connected to the microcontroller as follows: | Ideally, we come up with a pinout that allows both the TTGO and Heltect boards to be used with essentially the same firmware. | ||
=== NodeMCU ESP32 === | |||
My initial prototype used an NodeMCU ESP32. | |||
The microphone is connected to the microcontroller as follows (old): | |||
* INMP441 GND to ESP32 GND | * INMP441 GND to ESP32 GND | ||
* INMP441 VDD to ESP32 3.3V | * INMP441 VDD to ESP32 3.3V | ||
Line 159: | Line 178: | ||
The I2S clock signal is 64 times higher than the sample clock, so at a sample rate of 44100 Hz, this means 2.8 MHz. | The I2S clock signal is 64 times higher than the sample clock, so at a sample rate of 44100 Hz, this means 2.8 MHz. | ||
This might be a bit high for a random wire, probably we should keep this connection short. | This might be a bit high for a random wire, probably we should keep this connection short. | ||
At a sample rate of 22627 Hz, this results in an I2S clock of 1.45 MHz. | |||
=== Heltec LoRa32 v2 === | |||
Preliminary, for the LoRaSoundKit: | |||
{| class="wikitable" | |||
! ESP32 !! INMP441 !! Remark | |||
|- | |||
| GND || GND || - | |||
|- | |||
| 3.3V || VDD || - | |||
|- | |||
| 35 || SD || conflicts with LoRa DIO1! | |||
|- | |||
| 14 (A16) || SCK || - | |||
|- | |||
| 15 || WS || - | |||
|- | |||
| GND || L/R || - | |||
|} | |||
=== Microphones === | |||
* INMP441: cheap microphone | |||
* SPH0645: untested | |||
* <s>ICS-43434: expensive, used in the DNMS</s> can apparently be found more cheaply now, for like $3 | |||
=== Casing === | |||
[[File:NoiseCasing.jpg|thumb|right|250px|Casing prototype]] | |||
My own initial attempt: lasdoor + wartel + 5/8 inch pvc pipe on the right. | |||
Some inspiration can be found in the "DNMS" sensor: | |||
https://sensor.community/en/sensors/dnms/ | |||
RIVM did a comparison of sound meters, they appear to use some sort of PVC tube: | |||
https://www.samenmetenaanluchtkwaliteit.nl/nieuws/citizen-science-geluidmetertest-op-rivm-terrein | |||
== Software == | == Software == | ||
=== Original experimental code === | |||
My original experimental initial code for sampling audio from the digital microphone, calculating a frequency spectrum and displaying it, can be found [https://github.com/bertrik/NoiseLevel on github]. | |||
This code is in its current form not usable as a citizen science audio meter. | |||
What the software | What the software demonstrates: | ||
* Take an audio measurement from | * Take an audio measurement from a digital microphone at a regular interval. | ||
* On the recorded audio, perform a 4096-point real->complex FFT with a windowing function (flat-top | * On the recorded audio, perform a 4096-point real->complex FFT with a windowing function (flat-top). | ||
* Calculate power for each FFT-bin (Im-squared + Re-squared) and sum up bins per octave. | * Calculate power for each FFT-bin (Im-squared + Re-squared) and sum up bins per octave. | ||
* | * Display as a kind of ASCII-art 'waterfall'. | ||
=== Citizen science code === | |||
==== firmware ==== | |||
Marcel Meek has written an Arduino firmware that can actually be used in a citizen science sensor: | |||
https://github.com/meekm/LoRaSoundkit | |||
==== backend code ==== | |||
I have written a forwarder application in Java that takes the data from that sensor from TheThingsNetwork and forwards it to sensor.community: | |||
<s>https://github.com/bertrik/noiseforwarder</s> -> is now integrated in my sensor-data-bridge. | |||
=== Improvement ideas === | |||
To improve on the sound sensor of Marcel Meek, I intend to change the following: | |||
== | * Minor fixes on the build system: | ||
** make sure it compiles without a link error caused by duplicate hal_init() function | |||
** fix minor compiler errors, like "warning: no return statement in function returning non-void [-Wreturn-type]" | |||
* Make it work on a Heltec LoRa32 v2 board | |||
** This could be just adding a platformio configuration | |||
** Might also require remapping of the pins, since pin 35 (IIS data in) is already used for LoRa DIO1 on the Heltec LoRa32 v2 | |||
* Avoid burn-in of the OLED display | |||
** Currently the OLED continuously shows the most recent measurement data, the OLED is known to suffer from burn-in, making the display unreadable after a couple of months | |||
** The plan is to add a timeout, e.g. turn off display after 2 minutes, reactivate by button-press | |||
* Add firmware update interface | |||
** Internal webpage over WiFi that allows you to select a bin file and update the firmware over-the-air | |||
* Add easier commissioning of LoRaWAN parameters | |||
** Currently the node shows credentials using very small characters | |||
** Nice solution would be to make this machine-readable, like a QR code on the display with TTN credentials as long as it has not successfully joined TTN | |||
* Allow local reading of the most currently measured noise value | |||
** Add a bluetooth BLE GATT interface with the most recently measured dBa value | |||
** Maybe a sensor.community-style web page data.json file? | |||
== Use as a citizen science sound/noise sensor == | |||
The sensor.community node has support for a "DNMS" sound sensor, see https://sensor.community/en/sensors/dnms/ | The sensor.community node has support for a "DNMS" sound sensor, see https://sensor.community/en/sensors/dnms/ | ||
We emulate the network protocol used by a sensor.community node for DNMS, sending a "LAeq", minimum and maximum value. | |||
=== DNMS === | |||
In my opinion, the DNMS module has the following disadvantages: | In my opinion, the DNMS module has the following disadvantages: | ||
* the module produces a single value with a minimum and a maximum value (not a spectrum), no spectral corrections are possible | * the module produces a single value with a minimum and a maximum value (not a spectrum), no spectral corrections are possible | ||
* the communication interface is I2C, where the sound sensor is I2C slave, which is more difficult to implement than (for example) a 9600 bps serial interface | * the communication interface is I2C, where the sound sensor is I2C slave, which is more difficult to implement than (for example) a 9600 bps serial interface | ||
* the communication protocol is a bit weird, sends an 8-bit CRC for every 16-bit data word | * the communication protocol is a bit weird, e.g. sends an 8-bit CRC for every 16-bit data word | ||
* it needs a custom PCB | |||
* building the firmware is unnecessary complex | |||
It would be nice to be compatible with this sensor, so an ESP32 audio sensor could be a drop-in replacement for the "official" DNMS module. | It would be nice to be compatible with this sensor, so an ESP32 audio sensor could be a drop-in replacement for the "official" DNMS module. | ||
Line 201: | Line 286: | ||
*** send the DNMS_CMD_READ_LEQ command, this results in three values: a, a_min and a_max | *** send the DNMS_CMD_READ_LEQ command, this results in three values: a, a_min and a_max | ||
*** if an error occurred, send the DNMS_CMD_RESET command | *** if an error occurred, send the DNMS_CMD_RESET command | ||
Conclusion: Use of the I2C slave protocol makes it very difficult to use an ESP32 to emulate a DNMS sensor. | |||
=== Sensor.community protocol === | === Sensor.community protocol === | ||
Line 213: | Line 300: | ||
** "software_version", description of the software version | ** "software_version", description of the software version | ||
** "sensordatavalues", list of structures: | ** "sensordatavalues", list of structures: | ||
*** "value_type" = "noise_LAeq", "value" = | *** "value_type" = "noise_LAeq", "value" = value in dB | ||
*** "value_type" = "noise_LA_min", "value" = | *** "value_type" = "noise_LA_min", "value" = value in dB | ||
*** "value_type" = "noise_LA_max", "value" = | *** "value_type" = "noise_LA_max", "value" = value in dB | ||
Links: | Links: | ||
* https://github.com/hbitter/DNMS github page | * https://github.com/hbitter/DNMS github page |
Latest revision as of 08:15, 17 July 2023
Project ESP audio sensor | |
---|---|
ESP32-based audio sensor | |
Status | In progress |
Contact | bertrik |
Last Update | 2023-07-17 |
Introduction
This project is about creating an audio level meter, for example as an environmental noise measuring node in a citizen science project or as a standalone "decibel meter".
Marcel Meek built a sound sensor based on an ESP32 board with built-in LoRa-chip. It samples audio from a digital I2S microphone, analyses it into sub-bands, then forwards it over LoRaWAN. His code can be found here: https://github.com/meekm/LoRaSoundkit
From TheThingsNetwork, the data is picked up by a kind of 'integration' that sends it to sensor.community. This is handled by my sensor-data-bridge noise forwarder Java project.
RIVM noise measuring project:
I would like to build a case like that!
Github links to the project of Marcel Meek / TTN Apeldoorn:
Links back:
- https://iotassistant.io/esp32/smart-door-bell-noise-meter-using-fft-esp32/
- https://en.usini.eu/digital_microphone_on_esp32/
TODO
- figure out which LoRaSoundkit is "current" (most up-to-date) -> this one: https://github.com/meekm/LoRaSoundkit
- figure out alternative pinning for LoRaSoundkit so it works for both TTGO and Heltec boards
- put it together!
Ideas
Use interface similar to casper.aero
See https://rtm.flighttracking.casper.aero/
Bluetooth interface
The ESP32 supports bluetooth as an RF technology. It would be nice to be able to quickly see the current/most recently measured data. This can be shown on the display. Perhaps even more convenient, this can be exposed over bluetooth for display on a mobile phone. Are there android apps that can interface with bluetooth?
The Bluetooth specification reserves 0x27C3 as the measurement unit code for sound pressure (dB) and 0x2BE4 as the noise characteristic.
How the BLE indicate and notify work, and the difference between them: https://openlabpro.com/guide/ble-notify-on-esp32-controller/
The characteristic 'noise' is not one of the characteristics permitted for the 'environmental sensing service', see https://bitbucket.org/bluetooth-SIG/public/src/main/assigned_numbers/profiles_and_services/ess/ess_permitted_characteristics.yaml so this probably has to be specified in a custom service
Measuring audio as a citizen science project
The concept is the following:
- A small box containing a microphone outside your house measures environmental sounds (traffic, etc), for example it takes a 1 second audio recording every 10 seconds
- A spectral analysis is made of the audio (separating it in different frequency bands), calculating sound intensity for each individual band
- Every 5 minutes, the measured sound intensity is sent to a central server on the internet. Because we only communicate intensities, this does not reveal audio fragments (conversations for example).
- The central server takes the measured intensities and can do corrections, like a microphone specific correction, or a correction to apply A-weighting
- The central server forwards the measurements to sensor.community, from there they are automatically picked up by RIVM
External things to investigate:
- How is audio being evaluated in general, are there norms for it?
- how often should audio be sampled? 1 second/minute? 10 seconds/min? How long should the audio fragment be?
- is it a valid assumption to use full octaves? or should the spectrum be divided in smaller parts?
- https://www.rijksoverheid.nl/onderwerpen/geluidsoverlast/geluidsoverlast-in-de-wet
Theory
The audio spectrum is divided up into octaves and the total energy in each octave is calculated. We can then easily apply sensor/housing specific corrections, do A weighting, etc.
A weighting
Subjective audio levels are generally calculated on a logarithmic scale in dB using "A-weighting". A-weighting calculates a subjective loudness level from a physical loudness, applying a correction factor for each (part of an) octave band.
Octave start | Octave center | Octave end | Remark |
---|---|---|---|
- | - | 22627 Hz | sample rate |
5657 Hz | 8000 Hz | 11314 Hz | |
2828 Hz | 4000 Hz | 5657 Hz | |
1414 Hz | 2000 Hz | 2828 Hz | |
707 Hz | 1000 Hz | 1414 Hz | |
354 Hz | 500 Hz | 707 Hz | |
177 Hz | 250 Hz | 354 Hz | |
88 Hz | 125 Hz | 177 Hz | |
44 Hz | 63 Hz | 88 Hz | |
22 Hz | 31 Hz | 44 Hz |
List of A-weighting coefficients
The sampling rate is chosen such that the FFT bands align exactly with the bands as defined in the A-norm.
FFT
The energy in each octave is calculated by applying an FFT (fast fourier transform) on the audio data. The FFT takes in real values and outputs complex values. The intensity in each octave band is by summing the energy in a set of FFT 'bins'. The energy in each bin is calculated as the real part squared plus the imaginary part squared. The division of the audio spectrum for the FFT is chosen so it matches the octaves used in the A-weighting as described above. I plan to use the 'flat-top' window because it has good properties for measuring power levels.
Investigation into FFT libraries:
- arduinoFFT does all calculations in double, severely limiting the audio buffer size. For each sample, you need two doubles, so that's 16 bytes per sample. I can use about 6000 samples in the audio buffer. Using floats gives me twice the audio buffer size.
- Alternatively, esp32-fft looks nice, it uses floats instead of doubles and is optimised for esp32, but it uses malloc internally ... Also I can't use it as an Arduino library because it doesn't have the arduino library structure (with a src dir, examples dir, library.json file, library.properties file, etc) Maybe I can still use it by copying it in my sketch.
Links:
- Overview of octave bands https://www.engineeringtoolbox.com/octave-bands-frequency-limits-d_1602.html
- http://www.robinscheibler.org/2017/12/12/esp32-fft.html
-> I went with ArduinoFFT, modified it to use float instead of double
Decibel meters
Commercially available meters:
Fairly typical specs:
- dynamic range: 30 - 130 dB
- accuracy: 1.5-2 dB
- frequency range: 31.5 Hz - 8 kHz (!)
- norm: EN 61672-1
Hardware
The physical device consists of:
- an ESP32, it has an I2S digital audio input for sampling data from a digital microphone and a WiFi interface to communicate things to the internet. It has more internal RAM than an ESP8266 (for example), this helps to take a larger audio sample and do analysis on it. It is still comparatively simple, cheap, easy to flash and powerful enough to do communication over WiFi and do audio analysis.
- a digital I2S microphone, like the INMP441 (datasheet)
Waag society uses the Invensense ICS4342 microphone in their kit 2.1. Myself, I ordered a couple of INMP441 microphones from Aliexpress.
Ideally, we come up with a pinout that allows both the TTGO and Heltect boards to be used with essentially the same firmware.
NodeMCU ESP32
My initial prototype used an NodeMCU ESP32.
The microphone is connected to the microcontroller as follows (old):
- INMP441 GND to ESP32 GND
- INMP441 VDD to ESP32 3.3V
- INMP441 SD to ESP32 A4/32
- INMP441 SCK to ESP32 A16/14
- INMP441 WS to ESP32 15
- INMP441 L/R to ESP32 GND
The connection carries only digital signals (max 3 MHz or so). No sensitive analog electronics are needed, the microphone and the microcontroller are simply connected using "dupont" wire.
The I2S clock signal is 64 times higher than the sample clock, so at a sample rate of 44100 Hz, this means 2.8 MHz. This might be a bit high for a random wire, probably we should keep this connection short. At a sample rate of 22627 Hz, this results in an I2S clock of 1.45 MHz.
Heltec LoRa32 v2
Preliminary, for the LoRaSoundKit:
ESP32 | INMP441 | Remark |
---|---|---|
GND | GND | - |
3.3V | VDD | - |
35 | SD | conflicts with LoRa DIO1! |
14 (A16) | SCK | - |
15 | WS | - |
GND | L/R | - |
Microphones
- INMP441: cheap microphone
- SPH0645: untested
ICS-43434: expensive, used in the DNMScan apparently be found more cheaply now, for like $3
Casing
My own initial attempt: lasdoor + wartel + 5/8 inch pvc pipe on the right.
Some inspiration can be found in the "DNMS" sensor: https://sensor.community/en/sensors/dnms/
RIVM did a comparison of sound meters, they appear to use some sort of PVC tube: https://www.samenmetenaanluchtkwaliteit.nl/nieuws/citizen-science-geluidmetertest-op-rivm-terrein
Software
Original experimental code
My original experimental initial code for sampling audio from the digital microphone, calculating a frequency spectrum and displaying it, can be found on github. This code is in its current form not usable as a citizen science audio meter.
What the software demonstrates:
- Take an audio measurement from a digital microphone at a regular interval.
- On the recorded audio, perform a 4096-point real->complex FFT with a windowing function (flat-top).
- Calculate power for each FFT-bin (Im-squared + Re-squared) and sum up bins per octave.
- Display as a kind of ASCII-art 'waterfall'.
Citizen science code
firmware
Marcel Meek has written an Arduino firmware that can actually be used in a citizen science sensor: https://github.com/meekm/LoRaSoundkit
backend code
I have written a forwarder application in Java that takes the data from that sensor from TheThingsNetwork and forwards it to sensor.community:
https://github.com/bertrik/noiseforwarder -> is now integrated in my sensor-data-bridge.
Improvement ideas
To improve on the sound sensor of Marcel Meek, I intend to change the following:
- Minor fixes on the build system:
- make sure it compiles without a link error caused by duplicate hal_init() function
- fix minor compiler errors, like "warning: no return statement in function returning non-void [-Wreturn-type]"
- Make it work on a Heltec LoRa32 v2 board
- This could be just adding a platformio configuration
- Might also require remapping of the pins, since pin 35 (IIS data in) is already used for LoRa DIO1 on the Heltec LoRa32 v2
- Avoid burn-in of the OLED display
- Currently the OLED continuously shows the most recent measurement data, the OLED is known to suffer from burn-in, making the display unreadable after a couple of months
- The plan is to add a timeout, e.g. turn off display after 2 minutes, reactivate by button-press
- Add firmware update interface
- Internal webpage over WiFi that allows you to select a bin file and update the firmware over-the-air
- Add easier commissioning of LoRaWAN parameters
- Currently the node shows credentials using very small characters
- Nice solution would be to make this machine-readable, like a QR code on the display with TTN credentials as long as it has not successfully joined TTN
- Allow local reading of the most currently measured noise value
- Add a bluetooth BLE GATT interface with the most recently measured dBa value
- Maybe a sensor.community-style web page data.json file?
Use as a citizen science sound/noise sensor
The sensor.community node has support for a "DNMS" sound sensor, see https://sensor.community/en/sensors/dnms/
We emulate the network protocol used by a sensor.community node for DNMS, sending a "LAeq", minimum and maximum value.
DNMS
In my opinion, the DNMS module has the following disadvantages:
- the module produces a single value with a minimum and a maximum value (not a spectrum), no spectral corrections are possible
- the communication interface is I2C, where the sound sensor is I2C slave, which is more difficult to implement than (for example) a 9600 bps serial interface
- the communication protocol is a bit weird, e.g. sends an 8-bit CRC for every 16-bit data word
- it needs a custom PCB
- building the firmware is unnecessary complex
It would be nice to be compatible with this sensor, so an ESP32 audio sensor could be a drop-in replacement for the "official" DNMS module.
DNMS Protocol:
- as reverse engineered from the sensor.community firmware:
- device is at i2c address 0x55 (7-bit or 8-bit address?)
- commands exchanged are 16-bit
- 1 DNMS_CMD_RESET
- 2 DNMS_CMD_READ_VERSION
- 3 DNMS_CMD_CALCULATE_LEQ
- 4 DNMS_CMD_READ_DATA_READY
- 5 DNMS_CMD_READ_LEQ
- a command looks like this:
- 16-bit (big endian) command id
- a variable number of 16-bit data words, each with an 8-bit CRC
- getting the data from the DNMS:
- send the DNMS_CMD_CALCULATE_LEQ command
- repeat DNMS_CMD_READ_DATA_READY command until it indicates that a measurement is available
- send the DNMS_CMD_READ_LEQ command, this results in three values: a, a_min and a_max
- if an error occurred, send the DNMS_CMD_RESET command
Conclusion: Use of the I2C slave protocol makes it very difficult to use an ESP32 to emulate a DNMS sensor.
Sensor.community protocol
The protocol is basically the same as the particulate matter concentration protocol, with different "value_type" fields in the JSON.
- REST with JSON
- HTTP POST request to "https://api.sensor.community" with path "/v1/push-sensor-data/"
- the POST contains the following HTTP headers:
- HTTP header "X-Sensor" contains the unique hardware id, for an ESP32-board, an id of type "esp32-<number>" would be appropriate
- HTTP header "X-Pin" with value 15
- HTTP header "Content-type" with value "application/json"
- the POST body contains the following fields:
- "software_version", description of the software version
- "sensordatavalues", list of structures:
- "value_type" = "noise_LAeq", "value" = value in dB
- "value_type" = "noise_LA_min", "value" = value in dB
- "value_type" = "noise_LA_max", "value" = value in dB
Links:
- https://github.com/hbitter/DNMS github page