You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: tensorflow-to-larod-artpec8/README.md
+2-2Lines changed: 2 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -71,7 +71,7 @@ If your machine doesn't have the hardware requisites, like not enough GPU to tra
71
71
72
72
### The example model
73
73
74
-
In this tutorial, we'll train a simple model with one input and two outputs. The input to the model is a FP32 RGB image scaled to the [0, 1] range and of shape `(480, 270, 3)`.
74
+
In this tutorial, we'll train a simple model with one input and two outputs. The input to the model is a FP32 RGB image scaled to the [0, 1] range and of shape `(256, 256, 3)`.
75
75
The output of the model are two separate tensors of shape `(1,)`, representing the model's confidences for the presence of `person` and `car`. The outputs are configured as such, and not as one tensor with a SoftMax activation, in order to demonstrate how to use multiple outputs.
76
76
However, the general process of making a camera-compatible model is the same irrespective of the dimensions or number of inputs or outputs.
77
77
@@ -84,7 +84,7 @@ In order to produce a model with BatchNormalization layers that are fused with t
84
84
Specifically, the convolutional layers need to not use bias, e.g., for Keras Conv2D layers have the `use_bias=False` parameter set, and the layer order needs to be: `convolutional layer -> batch normalization -> relu`.
85
85
This will "fold" , or "fuse", the batch normalization, which increases performance.
86
86
87
-
The pre-trained model is trained on the MS COCO 2017 **training** dataset, which is significantly larger than the supplied MS COCO 2017 **validation** dataset. After training it for 8 epochs and fine-tuning the model with quantization for 4 epochs, it achieves around 85% validation accuracy on both the people output and the car output with 6.6 million parameters. This model is saved in the frozen graph format in the `/env/output_models` directory.
87
+
To replicate the model training used for the model in [vdo-larod](../vdo-larod/), utilize the MS COCO 2017 **training** dataset, which is significantly larger than the provided MS COCO 2017 **validation** dataset. After training for 12 epochs and fine-tuning the model with quantization for 1 epoch, it achieves good accuracy on both people and cars.
Copy file name to clipboardExpand all lines: vdo-larod/README.md
+77-48Lines changed: 77 additions & 48 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -33,8 +33,8 @@ See the manifest.json.* files to change the configuration on chip, image size, n
33
33
34
34
## Which backends and models are supported?
35
35
36
-
Unless you modify the app to your own needs you should only use our pretrained model that takes 480x270 (256x256 for Ambarella CV25 and Google TPU) RGB (interleaved or planar) images as input,
37
-
and that outputs an array of 2 confidence scores of person and car in the format of `float32`.
36
+
Unless you modify the app to your own needs you should only use our pretrained model that takes 256x256 RGB (interleaved or planar) images as input,
37
+
and that outputs an array of 2 confidence scores of person and car in the format of `uint8`.
38
38
39
39
You can run the example with any inference backend as long as you can provide it with a model as described above.
40
40
@@ -280,67 +280,96 @@ In previous larod versions, the chip was referred to as a number instead of a st
0 commit comments