►
From YouTube: 004 ONNX 20211021 Wang ONNX Intel Neural Compressor A Scalable Quantization Tool for ONNX Models
Description
LF AI & Data Day - ONNX Community Meeting, October 21, 2021
IntelĀ® Neural Compressor: A Scalable Quantization Tool for ONNX Models
Speaker: Mengni Wang (Intel)
A
A
A
It
is
built
above
deep
learning
frameworks,
including
tensorflow,
python,
onyx
and
mxnet,
and
can
yield
a
compressed
framework
framework
model
for
deployment
continuation
said
it
unified,
different
connection,
apis
of
different
deep
learning
frameworks
and
provides
an
out-turning
mechanism
to
help
users
figure
out
the
best
low
precision
solution
on
intel
hardware
rapidly.
A
A
A
A
A
A
A
It
can
make
great
great
contributions
to
productivity
until
now,
resonant
50,
vdd,
16
and
the
shuffle
net
v2
have
been
upstreamed
to
onyx
model
zoo,
and
we
got
some
positive
feedback
from
mother
zoo
owner.
He
said
into
eight
res:
15
is
the
first
contest
model
for
onyx
model
zoom,
and
it
shows
a
great
performance
improvement.
A
This
is
our
contribute
contribution
plan.
We
plan
to
contribute
to
more
context
models
in
the
future,
which
will
enrich
the
diversity
and
the
skill
of
intent
models
in
onyx
model
zoo.
Our
goal
is
all
enable
the
fp32
models
in
onyx
model
2
would
have
corresponding
context
models
through
inter
neural
compressor.