Application Example
This section shows you a simple application with hardware in the loop using the dAIEdge-VLab Python API.
dAIEdge-VLab API Demo Video
What does this application do ?
In this application example, we will train a model with Tensorflow, prepare a validation dataset and execute the trained model on a remote device. Finally, we will analyze the results given by the remote device.
The following diagram shows the pipeline of this application :
sequenceDiagram participant User participant dAIEdge-VLab participant Target User->>User: Prepare Dataset User->>User: Train mdel User->>dAIEdge-VLab: Model, Dataset, Target, Runtime dAIEdge-VLab->>Target: Model, Dataset Target->>Target: Perfom the inferences Target->>dAIEdge-VLab: report, logs, raw outpout dAIEdge-VLab->>User: report, logs, raw outpout User->>User: Format raw output User->>User: Compute accuracy metrics
Main actions :
- Training of the model on the user end
- Prepareing of the dataset on the user end
- Running the model on the remote device
- Analyzing the resutls
What problem will be solved ?
In this simple application, we will train a model to recogize numbers from an image. We will use the MNIST dataset as based for the training and testing of the model. The image bellow illustrates the kind of images that compose the dataset. Our model will simply take an image as input and output a prediction of the number that was on the image.
The output of the model will be one hot encoded. This means that we will have one output node per class we want to detect. In this case this is 10 classes as they are 10 numbers from 0 to 9. Knowing the input and ouput shape is important for the preparation of the dataset we will provide to the dAIEdge-VLab.
Prepare the envionement
For this application example you will need to have a Python environement wiht tensoflow
, sklearn
and dAIEdgeVLabAPI
installed.
Use the step-by-step user guide to install dAIEdgeVLabAPI
. Use the following commandes to install the other depedencies:
pip install tensorflow
pip install scikit-learn
Implement Application
The following section shows a simple but complete application to train a model on the MNIST dataset and validate the model on a remote target.
Define the model architecture
To train a simple model that will recogize the number on an image we can use a simple but effective neural network structure. The model will be composed of :
- The input layer made of
28x28x1
nodes - A Conv2D (32 filters,
3x3
kernel, ReLU activation) layer - A MaxPooling2D (
2x2
pool size) layer - A Conv2D (64 filters,
3x3
kernel, ReLU activation) - A axPooling2D (
2x2
pool size) layer - A Flatten layer : Converts 2D feature maps into a 1D vector for the dense layer.
- A Dense (128 neurons, ReLU activation) layer
- A Dense (10 neurons, softmax activation) layer : The output layer
Prepare the different datasets
From the MNIST dataset we create 3 sub-datasets :
- A train dataset (used to train the model)
- A validation dataset (used to test the model during training)
- A test dataset (used to test the model on the remote target)
The following code prepares the different datasets:
import numpy as np
import tensorflow as tf
from tensorflow import keras
from sklearn.model_selection import train_test_split
from dAIEdgeVLabAPI import dAIEdgeVLabAPI
# Define test dataset name
DATASET = "dataset_mnist.bin"
# Load MNIST dataset
(x_train, y_train), (x_test, y_test) = keras.datasets.mnist.load_data()
# Normalize pixel values to [0,1]
x_train, x_test = x_train / 255.0, x_test / 255.0
# Expand dimensions for CNN input (from (28,28) to (28,28,1))
x_train = x_train[..., tf.newaxis]
x_test = x_test[..., tf.newaxis].astype(np.float32)
# Split training set into training (90%) and validation (10%) sets
x_train, x_val, y_train, y_val = train_test_split(x_train, y_train, test_size=0.1, random_state=42)
# Print dataset sizes
print(f"Training set: {x_train.shape[0]} samples")
print(f"Validation set: {x_val.shape[0]} samples")
print(f"Test set: {x_test.shape[0]} samples")
# Save the test dataset in binary format
x_test.tofile(DATASET)
Train the model
The following code defines the network architecture and train the model with the training and validation dataset. Finally, we export and save the model in a tflite
format :
# Define model filename
MODEL = "mnist_model.tflite"
model = keras.Sequential([
keras.layers.Conv2D(32, kernel_size=(3,3), activation='relu', input_shape=(28,28,1)),
keras.layers.MaxPooling2D(pool_size=(2,2)),
keras.layers.Conv2D(64, kernel_size=(3,3), activation='relu'),
keras.layers.MaxPooling2D(pool_size=(2,2)),
keras.layers.Flatten(),
keras.layers.Dense(128, activation='relu'),
keras.layers.Dense(10, activation='softmax') # 10 output classes (digits 0-9)
])
# Compile the model
model.compile(optimizer='adam',
loss='sparse_categorical_crossentropy',
metrics=['accuracy'])
# Train the model
model.fit(x_train, y_train, epochs=5, validation_data=(x_val, y_val))
# Convert the model to TensorFlow Lite format
converter = tf.lite.TFLiteConverter.from_keras_model(model)
tflite_model = converter.convert()
# Save the model to a .tflite file
with open(MODEL, 'wb') as f:
f.write(tflite_model)
Add hardware in the loop
Now we will connect to the dAIEdge-VLab and start a benchmark with the newly trained model and test dataset :
# Define hardware
TARGET = "rpi5"
RUNTIME = "tflite"
# Connect to dAIEdge-VLab
api = dAIEdgeVLabAPI("setup.yaml")
# Upload dataset
api.uploadDataset(DATASET)
# Start benchmark
id = api.startBenchmark(
target = TARGET,
runtime = RUNTIME,
model_path = MODEL,
dataset = DATASET
)
print(f"Benchmark started for {TARGET} - {RUNTIME}")
# Wait for the results
result = api.waitBenchmarkResult(id, verbose=True)
# Delete the dataset from the dAIEdge-VLab server
api.deleteDataset(DATASET)
Analyze the results
Once we get the result back from the target, we can analyze the output of the model. This allows us to get the accuracy of the model once run on the device. We can also retrieve the mean inference time :
nb_samples = len(x_test)
# Convert the binary file to an array of float32, then reshape it as (nb_samples, 10)
output_array = np.frombuffer(result["raw_output"], dtype=np.float32)
output_array = output_array.reshape(nb_samples, 10)
# Find the max argument in each nb_samples array of 10 float values
predictions = np.argmax(output_array, axis=1)
# Compute the accuracy of the prediction
accuracy = np.mean(predictions == y_test)
print("Number of inference:", result["report"]["nb_inference"])
print("Mean inference time us:", result["report"]["inference_latency"]["mean"])
print("Accuracy:", accuracy)
Expected output
You can download the full Python script here. Once you run it, you should see something like :
...
Number of inference: 10000
Mean inference time us: 202.7421925
Accuracy: 0.9913