Uploaded on Oct 19, 2021
The Model Zoo is a repository which includes pre-trained, pre-compiled models and a full building and evaluation environment. Its aim is to provide developers with quick functionality to hit the ground running.
Hailo Model Zoo
Hailo Model Zoo
What is a Model Zoo?
The Model Zoo is a repository which includes
pre-trained, pre-compiled models and a full
building and evaluation environment. Its aim
is to provide developers with quick
functionality to hit the ground running.
With the goal of making it as easy as
possible for developers to get up and
running with the Hailo-8™ AI processor,
we have built our own open Hailo Model
Zoo repository, which includes:
Pre-trained models – a large
selection, demonstrating both
the versatility and high AI
performance of the Hailo-8.
Pre-configured build flows – validated and
optimized flows to take each network
from an ONNX/TensorFlow model to a
deployable binary, incorporating models
evaluation and performance analysis.
To allow a “fine print-free” evaluation we focused on
populating hailo model zoo with popular models taken
from open-source repositories without modification
and trained on publicly available datasets. A link to
the model source was also added in case users would
like to adjust the model, for instance by training it on
their own custom dataset.
How It Works
The Model Zoo employs the Hailo Dataflow Compiler for a full flow
from a pre-trained model (ckpt/ONNX) to a final Hailo Executable
Format (HEF) that can be executed on the Hailo-8. To that end, the
Hailo Model Zoo provides users with the following functions:
Parse: translate Tensorflow/ONNX model
into Hailo’s internal representation, which
includes the network topology and the
original weights of the model. The output
of this stage is an Hailo Archive file
(HAR).
Profile: a report that includes a the
expected model’s performance on the
Hailo-8 including FPS, latency, power
consumption and a full breakdown for each
layer in HTML format.
Quantize: optimize the model for runtime by
converting it from full precision to limited integer bit
precision (4/8/16) while minimizing accuracy
degradation. This stage includes several algorithms
that optimize performance by assuring the accuracy
of the quantized model. The output of this stage is a
Hailo Archive file that includes the quantized weights.
Evaluate: evaluate the model accuracy
on common datasets (e.g., ImageNet,
COCO). Evaluation can be done on the
full precision model and on the quantized
model using our numeric emulator or the
Hailo-8.
Compile: compile the quantized
model to generate a Hailo
Executable Format (HEF) file which
can be deployed on the Hailo-8 chip.
Comments