►
From YouTube: Node.js Benchmarking WG Meeting 07-09-2019
Description
B
A
B
A
We
talked
about
multiple
ways:
we
thought,
because
there
are,
we
have
so
many
of
those
micro
benchmarks.
We
thought
you
could
take
the
subset,
we
could
run
either
weekly
basis
and
we
could
run
all
of
them,
maybe
once
a
month
kind
of
thing,
so
there's
some
kind
of
proposal,
some
kind
of
purpose
that
we
need
or
approach.
We
need
to
cover
micro
benchmarks
in
the
in
this
guarding.
C
That
is,
can
we
can
we
can
we
do
like
a
like
a
random
sample
of
benchmarks
like
a
time-limited
random
sample,
and
then
you
know
if
we
limit
it
to
like
two
hours
or
something
like
that,
and
then
you
know
we
can.
We
can
just
run
certain
benchmarks
at
certain
times
and
we
just
choose
randomly,
and
so
so
you
know
we
can
run
that
more
often.
C
That
way-
and
you
know
the
the
data
points
would
be
a
little
sparse
err,
but
at
least
on
average
all
the
benchmarks
would
be
covered,
except
maybe
buffer,
because
buffer
takes
a
long
time,
no
matter,
no
matter
how
rarely
you
run
it,
it
always
takes
a
long
time.
That
was
my
only
thought
that
I
had
on
this
topic
and
I'm
wondering
what
you
guys
think
about
that
I.
B
B
B
C
B
C
I
mean
yeah,
I,
guess
I,
guess
you
know
we
could.
We
could
take
like
the
lazy
evaluation
approach.
You
know
and
just
have
charts
and
cache
them
and
and-
and
you
know,
generate
them
on
the
fly
so
so
for
things
that
are
never
needed,
except
once
it
takes
that
person
a
long
time
to
to
see
their
chart,
but
but
the
data
would
be
there
right.
If
we
did
this
random
sampling,
we
would
just
be
collecting
data
but
not
generating
any
charts
right.
C
A
C
Yeah
yeah,
yeah,
I,
guess
I,
guess
you
know
what
are
people
interested
in
right?
I
mean
they,
they
just
made
the
change
and
they
want
to
know
how
it
affects
things
or
or
a
change.
The
thing
like
like
you
know,
maybe
maybe
they
want
to
see
before
a
relief.
You
know
how
the
change
has
has
affected
performance
right.
So
so
you
know
if
we're
on.
C
If
we
run
the
random
sample
like
every
night,
you
know,
then
then
chances
are
that
before
released
is
made
at
least
two
or
three
samples
will
have
been
collected
of
the
benchmark.
That
is
relevant
to
them.
You
know,
and-
and
so
you
know
they
can,
they
can
have
a
look.
They
go
in
the
chart
is
generated
for
their
particular
case
and,
and
you
know
they
can
check.
You
know
what
happened.
You
know
there.
There
are
a
few
data
points
from
like
three
months
ago,
and
then
you
know
there
are.
C
A
C
See,
that's
that
that's
a
those
are
yeah
classifying
the
benchmarks
like
that
is
is
a
tough
problem.
Right
I
mean
we're
having
the
same
problem
with
the
buffer
benchmark,
all
right,
there's
one
PR
that
we
were
submitted
to
try
to
cut
back
on
the
number
of
buffer
benchmarks
and-
and
we
still
couldn't
decide
like
he,
even
even
at
the
TSC
level.
It's
like
okay,
but
you
know
I.
We
are
going
to
be
missing
some
stuff.
You
know
because
the
benchmark
doesn't
doesn't
strictly
eliminate
duplicates.
C
You
know
it
does
eliminate
some
benchmarks,
and
so
you
know
we
would
be
losing
perspective
and
so
arguing
as
to
you
know
whether
this
particular
aspect
to
benchmarking,
a
particular
aspect
is
important
or
not
important
or
how
important
it
is.
You
know
from
zero
to
five.
It's
it's
tough
too.
It's
tough
to
classify!
C
It's
going
to
be
that
conversation
repeated
over
and
over
and
over
for
all
the
different
benchmarks,
so
so
I
I
suspect
that
coming
up
without
what
the
classification
might
be
complicated,
you
know
and
a
slow-moving
process,
and
we
need
to
consult
a
lot
of
people
and
a
lot
of
people.
We
need
to
agree
that,
okay,
this
benchmark
is
good
to
have,
but
it's
not
critical,
you
know
and
making
this
decision
for
every
single
benchmark
you
know,
might
take
a
long
time.
So
so
you
know
that's
why
I
was
thinking
like.
C
If
we
do
a
random
sample,
then
it
doesn't
matter
how
important
the
benchmark
is.
Eventually
we
will
get
to
it,
you
know
and
and
that,
because
what
is
the
limiting
factor
the
limiting
factor
is
is
is
is
CPU
time
right,
because
you
know
we
don't
want
to
hog.
We
don't
want
to
hog
the
whole
system,
just
for
our
benchmarks,
that
you
know
that's
one
in
the
spectrum
and
the
other
end
of
the
spectrum.
Is
you
run
these
benchmarks
only
as
needed?
C
But
then,
of
course
you
don't
have
perspective
right
like
you
need
to
know
how
the
how
the
performance
of
any
given
micro
benchmark
changes
over
time
right
and
so
so
I
think.
The
in-between
solution
is,
you
know,
give
us
as
much
CPU
time
as
you
can
and
and
we'll
try
to
distribute
it
as
randomly
as
possible
across
the
benchmarks
and
and
and
get
some
numbers.
You
know
and
and
then
there's
a
good
chance,
not
a
perfect
chance,
but
there's
a
good
chance
that
you
will.
C
You
will
get
the
perspective
that
you
need
once
you
look
at
the
numbers
when
you
need
to
you
know
the
numbers
may
not
be
there
in
which
case
you're
like
dang
okay,
can
we
do
a
run
for
this
particular
commit
ID?
Now
you
know,
because
I
really
need
to
have
this
before
before
an
after
picture,
but
normally
we
will
run
a
set
of
benchmarks
that
fits
within
the
CPU
time
that
we
have
allocated.
C
B
C
D
C
A
C
Yes,
yes,
yeah
I
think
that's
a
good
idea.
You
know
what
granularity
are
we
looking
at
here
like
we're?
Not
gonna
like
we're
gonna
treat
each
file
as
a
separate
benchmark
right
like
we're
not
talking
about
like
individual
parameters
like
this
parameter
is
more
important.
You
know,
testing
with
with
512
byte
chunks
is
more
important
than
testing
with
1024
bytes
chunks.
I
mean
we're
not
going
in
that
kind
of
level
of
granularity
right,
I.
A
B
A
A
C
A
B
In
a
random,
rather
than
a
random
portion,
just
do
the
first
tenth,
second
tenth
or
twentieth
or
whatever
and
over
time,
it'll
all
fill
in
right,
because
really
you
don't
want
to,
is
it
worthwhile
having
two
for
one
and
zero
for
another?
It
may
be
better
just
to
have
one
for
everything
and
to
run
it.
C
B
C
B
C
C
A
B
A
So
that
well
I
think
in
the
top
12
category
fifteen
categories
we
have,
we
can
take
one
run
sequence
pick
one.
So
that
way,
maybe
we
can
cover
all
the
inputs
it
has
already
in
the
files.
So
we
don't
have
to
pick
and
choose
a
number.
We
can
just
run
that
one
okay,
I
could
am
gonna
run
buffer
tomorrow.
Maybe
data
view,
whatever
sequence,
we
have
yeah.
C
C
C
C
B
B
A
So
I
believe
I
know
Gabriel.
Somehow
it
shows
his
unmute,
but
I
think
we
should
put
a
proposal
like
that.
Okay,
we
are
going
to
yeah
stops
categories.
We
are
going
to
run
one
a
day
and
then
kind
of
cover,
all
of
them
or
maybe
a
two
weeks
period,
if
not
within
three
days
or
four
days.
That
is
still
good
enough
yeah,
and
if
someone
wants
to
go
detail
at
least
we
have
a
mechanism
but
infrastructure
to
run
this
benchmark
in
the
our
benchmarking
setup
right
now
we
don't
have
anything
yeah.
B
B
A
B
B
B
So
yeah
I
mean
today
we,
you
know
the
existing
ones.
We
have
actually
post
individual
results
to
my
sequel,
yeah
and
but
I.
Think
in
this
case
I'm
not
sure
that
that's
the
best
way
like
because
we'd
have
to
extract
all
the
data
points
individually,
push
them
out.
It
might
be
better
to
somehow
like
store
the
files
and
then
because
there's
already
like
I,
think
a
comparison
you
can,
you
can
use
the
the
run.
That
says
show
me
the
deltas
between
two
two
runs
yeah.
C
B
If
we
could
somehow
reuse
that
that
says,
you
know
from
the
data
points
the
most
recent
data
points
we
you
know,
maybe
it
takes
three
or
four
or
five
six
whatever
days.
If
we
could
pull
that
into
one
thing,
you
could
compare
against
or
something
and
then
you
know
that
basically
would
let
you
say,
download
a
file.
You
could
say
we'll
compare
against
this
and
you
know
if
you
could
run
a
subset
of
the
buffer
things
and
say,
compare
against
what
we've
got
in
the
database,
but.
C
I'm
wrong:
the
the
comparison
algorithm
actually
establishes
not
only
uses
uses
like
a
mean
and
standard
deviation
to
compare
and
to
also
establish
relevancy
right.
So
so
so
so
then,
if
we
do
that
and
that,
then
that's
even
more
expensive,
because
Marya
stablishing
a
single
data
point.
It
then
runs
the
benchmark
multiple
times
right
to
establish
a
standard,
deviation
and
a
mean
you
need.
You
need
multiple
run.
C
B
C
I
mean
the
truth
is
the
truth
is
to
be
able
to
run
that
tool
on
every
benchmark
would
be
ideal
even
more
so
because
you
know,
if
you,
if
you,
if
you
run
it
once
because
you
know
it,
takes
six
days
to
run
everything
and
you
run
one
benchmark
once
and
you
get
a
value
for
that.
I
mean
that
could
be
the
outlier
for
all.
C
B
A
C
A
C
The
relevant
group,
I
guess
I'll,
have
to
look
at
it,
who
touched
the
benchmarks,
the
various
different
benchmarks
and
and
I
just
sort
of
mentioned
them
and
and
see.
If
they,
you
know
collaborators
come
and
go
so
they
may
no
longer
be
active
in
that
area.
That
kind
of
thing
so
it'll
be
a
little
bit
of
a
forensic
task,
but
yeah
I
guess
I
can
try
to
find
some
cycles
for
this
yeah.
A
A
D
D
Sure
this
specifically
talked
about
so
I
have
followed
the
this
data
and
I'm
trying
here
to
share
the
screen.
I,
don't
know
how
to
do
that
share
the
screen
share,
share
screen
yeah,
so
here
the
here
are
some
some
dates
when
I
tracked
the
data
and
here
the
master,
the
tenex
version
and
so
on.
So
what
I
observe
it
at
a
certain
stage,
all
the
data
was
stagnant
for
more
than
a
week,
so
things
did
not
change
in
on
17
of
June
I
stopped
this
tracking,
because
I
didn't
know.
D
We
have
to
continue
not
so
it's
a
long
time
when
the
numbers
didn't
change
today,
I
have
looked
and
I
saw
some
strange
numbers,
so
mastering
techniques
have
the
same.
The
same
performance
I
would
say
that
it's
a
little
bit
strange.
The
only
difference
that
I
saw
I
saw
here
in
this
in
this
column.
So
so
they
are
here.
Operations
per
second
latency
footprint
before
Road
and
foot
through
in
after
load.
A
Yeah
I
mean
it
makes
sense
for
sure
so.
10X
master
and
10x.
Only
question
is
why
we
don't
have
number
for
the
12,
8,
X
and
canary,
build
and
then
also
at
the
same
time.
Why
do
we
have
LOB's
per
second,
but
we
don't
have
late.
We
have
latency
for
a
takes,
but
we
don't
have
the
ops
per
second
for
e-text
version.
It
dot,
X,
I,.
D
A
D
A
The
normally
what
I
used
to
do
is
whenever
I
see
something
like
this
I
would
try
to
reproduce
on
my
local
machine,
for
example,
I
can
take
if
you
clone
the
whole
benchmarking
repo,
you
will
find
all
the
scripts.
You
need
to
run
the
no
DCIS
and
through
all
this
various
versions
of
node,
and
see
whether
you
can
reproduce
any
of
them.
Okay,.