Skip to content

Commit 1017851

Browse files
committed
App note content update & preview gen retrigger
1 parent dc9798e commit 1017851

File tree

1 file changed

+16
-9
lines changed
  • content/hardware/06.nicla/boards/nicla-voice/tutorials/glass-break-detector

1 file changed

+16
-9
lines changed

content/hardware/06.nicla/boards/nicla-voice/tutorials/glass-break-detector/content.md

Lines changed: 16 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -98,7 +98,7 @@ This dataset provides a broad collection of audio samples designed for rare soun
9898
- **Isolated Events:** 474 unique events across three target categories: baby crying, **glass breaking**, and gunshot.
9999
- **Background Recordings:** 1121 of 30 second recordings of ambient sounds totaling 9 hours and 20 minutes of audio.
100100

101-
For this application note, we focus on glass-breaking sounds. From this dataset, *300 audio samples* are used for training and *54 samples* for testing. These samples represent two main categories:
101+
For this application note, we focus on glass-breaking sounds. From this dataset, *313 audio samples* are used for training and *54 samples* for testing. These samples represent two main categories:
102102

103103
- Glass-breaking events
104104
- Non-glass-breaking ambient noise
@@ -123,27 +123,31 @@ Additionally, Edge Impulse allows you to collect data directly from your devices
123123

124124
### Data Processing
125125

126-
In this step, you will use an audio processing block optimized for the NDP120 processor found on the Nicla Voice. This block uses a *Fast Fourier Transform (FFT)* to convert audio data from the time domain to the frequency domain, helping in extracting meaningful features for classification.
126+
In this step, you will use an audio processing block specifically optimized for the NDP120 processor integrated into the Nicla Voice. This block uses a *Fast Fourier Transform (FFT)* to convert audio signals from the time domain into the frequency domain, extracting meaningful features for classification.
127127

128-
Dimensionality reduction techniques visualize the features, describing different clusters for glass-breaking and non-glass-breaking sounds.
128+
Dimensionality reduction techniques help visualize and cluster features, effectively distinguishing between glass-breaking and non-glass-breaking sounds.
129129

130-
The image below shows the time series and audio processing block configuration in Edge Impulse. Configuring parameters like window size and frequency is important for extracting meaningful audio data features. It can be found within the **Create Impulse** section in the Edge Impulse platform.
130+
The first image below shows the time series configuration and audio processing block in Edge Impulse. Configurable parameters, such as window size and frequency, are important in extracting the most meaningful features for audio data classification.
131+
132+
These configurations are available in the **Create Impulse** section of the Edge Impulse platform.
131133

132134
![Signal Processing Configuration](assets/edge-impulse-config.png)
133135

134-
The processing block provides configurable parameters such as window size and feature scaling, allowing you to fine-tune signal processing. This application note emphasizes the ease with which these configurations can be adjusted.
136+
The processing block includes customizable settings such as window size and feature scaling, allowing users to fine-tune signal processing based on specific application requirements. This flexibility provides optimal feature extraction and improved classification accuracy.
135137

136138
![Feature Generation and Visualization](assets/edge-generation-visualization.png)
137139

138-
In the image above, you can see the feature generation output. The data is visualized in clusters, which helps you understand the different features of the glass-breaking and non-glass-breaking classes. It can be found within the **Parameters** tab from **Syntiant** section in the Edge Impulse platform.
140+
The second image shows the output of feature generation. The data visualization reveals clusters corresponding to glass-breaking and non-glass-breaking audio samples.
141+
142+
This helps us understand how well the system can separate these two types of sounds, which can be accessed under the **Parameters** tab of the **Syntiant** section in Edge Impulse.
139143

140144
### Model Training
141145

142146
You will use a dense neural network architecture that is well-suited for audio classification tasks. The configuration ensures compatibility with the hardware, using a 968 millisecond audio window for processing requirements.
143147

144148
Edge Impulse’s Eon tuner can help automatically optimize model parameters, simplifying the process for you without extensive machine learning expertise.
145149

146-
The model achieved an accuracy of *89%* during training, showing its capability to differentiate between glass-breaking and background noises. This can be found within the **Classifier** section in the Edge Impulse platform.
150+
The model achieved an accuracy of *89 %* during training, showing its capability to differentiate between glass-breaking and background noises. This can be found within the **Classifier** section in the Edge Impulse platform.
147151

148152
![Model Testing Results](assets/edge-model-test-result.png)
149153

@@ -162,9 +166,12 @@ After training, the model is evaluated using a test dataset that includes previo
162166

163167
![Model Testing Results](assets/edge-model-test.png)
164168

165-
The image above shows the testing results with an accuracy of **99.12%**. This can be found within the **Model testing** section in the Edge Impulse platform.
169+
The image above shows the testing results with an accuracy of **99.12 %**. This can be found within the **Model testing** section in the Edge Impulse platform.
170+
171+
The confusion matrix highlights the model's capability to classify audio samples correctly while identifying potential areas for improvement:
166172

167-
The confusion matrix highlights the model's capability to classify audio samples correctly while identifying potential areas for improvement.
173+
- Correctly classified "glassbreak" events: 94.4 %
174+
- Correctly classified "unknown" samples: 100 %
168175

169176
Misclassifications mainly happened for short-duration glass-breaking sounds, suggesting the need for additional data or refinement of the processing pipeline to improve robustness. Regular updates to the dataset with new samples can address such limitations and improve performance.
170177

0 commit comments

Comments
 (0)