►
Description
LF AI & Data Day - ONNX Community Meeting - October 21, 2021
Emma Ning (Microsoft)
ONNX Runtime Web for In Browser Inference
A
A
A
A
A
A
A
Native
rx
runtime
supports
including
four
onix
operator
coverage,
quantized
onix
model,
as
well
as
mini
audix
runtime,
build
onyx
runtime
web
also
utilize,
maltese,
writing
and
simd
in
web
assembly
to
further
accelerate
model
inferencing
taking
mobile
at
v2.
As
an
example
in
this
table,
the
cpu
inference
performance
can
be
accelerated
by
3.4
times
with
three
thread
of
these
two
threads,
together
with
a
simd
enabled
comparing
to
the
pure
web
assembly
without
enabling
these
two
features.
A
All
right
operator,
support
and
platform
compatibility
are
two
important
factors
for
ai
development,
as
we
discussed
before
since
the
whole
onix
runtime
cpu
engine
is
built
into
web
assembly
backend.
All
the
onix
operators
are
supported
by
web
assembly.
Backend
webgl
backhand
only
supports
a
subset
of
onyx
operators.
A
A
A
By
doing
that,
we
can
provide
a
consistent
development
experience
for
server-side
and
client-side,
influencing
with
onyx
runtime
to
demonstrate
you
know
web
machine
learning
capability
with
onyx
runtime
web.
We
build
up
on
its
runtime
web
demo
website,
where
you
can
see
several
interesting
in
a
browser
vision.
Scenarios
powered
by
image
models
here
is
an
example
of
running
mobile
net
model.
In
a
browser
you
can
choose
different
backend
for
cpu
or
for
gpu.
A
Secondly,
there
are
still
a
lot
of
opportunities
to
further
optimize
onyx
runtime
web,
including
both
performance
and
memory
consumption,
for
example.
Webmn
is
one
promising
technique
on
its
runtime
web
could
leverage
in
the
future
some
experimental
results
have
already
showed
very
promising
performance
gains.