►
From YouTube: SIG Instrumentation 20220623
Description
SIG Instrumentation Bi-Weekly Meeting June 23rd 2022
A
A
Welcome
to
today's
edition
of
sea
instrumentation
by
weekly
meeting
today
is
the
23rd
of
June.
It
seems
that
we
have
a
couple
of
factual
details
on
our
list
for
today.
Let
me
show
my
screen.
Oh,
can
you
give
me
the
permissions.
A
Can
you
see
it
okay,
yep?
So
the
first
topic
is
mine.
So
basically,
I
wanted
to
talk
about
a
problem
that
has
come
up
a
few
times
during,
like
reviewing
of
PRS
so
essentially
like
when
people
are
introducing
new
metrics.
A
A
Eight
positions
problem
for
a
lot
of
them
and
I
keep
having
question
of
like
what's
the
best
practices
to
like
initialize
a
metric.
What's
the
best
way
to
register
them,
and
things
like
that
and
through
the
look
at
base
like
there
is
no
real
standout
and
like
on
the
way
to
do
it
like
everybody
does.
It
is
on
the
way
and
we
keep
having
like
more
and
more
Global
metrics,
as
well
as
more
and
more
metrics
that
are
registered
into
the
global.
A
The
Legacy
register
and
I
was
wondering
that,
if
it
would
make
sense
to
standardize
that
in
order
to
make
it
cleaners
all
over
the
code
base,
so
that,
like
the
end
goal,
would
be
maybe
to
get
away
from
this
Legacy
registry
and
from
all
these
globals
that
we
have,
but
maybe
a
first
step
would
be
just
to
like
have
a
good
standard.
Whether
we
have
like
I.
Don't
know.
Example,
for
instance,
that
we
could
share
to
the
other
developers
to
give
them
like
some
idea
of
how
they
should
write
their
code.
B
I
think
that
definitely
makes
sense.
Would
this
look
like
sort
of
a
helper
library
that
people
should
use
to
register
things
or.
A
That
would
serve
as
an
example
of
how
it
can
be
done
either
like
for
me,
there
could
be
two
example,
one
which
essentially
relies
on
the
Legacy
registry
because,
like
in
a
lot
of
areas
in
the
code
base,
we
cannot
move
away
from
it
that
easily,
but
we
can
prepare
the
code
in
a
way
that
it
will
be
easier
to
migrate
away
from
the
Legacy
registry
at
some
point
and
another
example.
That
would
be
like
kind
of
more
like
the
new
way,
with
the
cube
registry
and
stuff
like
that.
A
So
it
would
be
just
like
some
quick
example
to
guide
them
through
how
this
can
be
actually
done,
because
at
the
end
of
the
day
like,
if
we
really
want
to
create
the
standards
and
have
it
like
applied
to
the
whole
code
base,
then
we
will
need
like
some
avistatic
analysis
and
that
will
require
like
too
much
time.
I.
Think
parties
for
now.
A
Because
the
the
more
essentially
the
more
people
are
adding
metrics
in
a
wrong
way,
the
other
it
will
be
to
move
away
from
the
global.
So
in
the
Legacy
registry
as
a
whole.
A
B
A
I
can
try
to
work
on
that.
I
just
wanted
to
gather
some
feedback
to
see
like
if
any
other
has
any
subs
on
that
and
yeah.
If
anyone
want
to
look
into
that
with
me,
that
could
be
a
good
way
also
to
see
How
Matrix
registration
works
in
the
kubernetes
code
base
and
what
other
challenges.
B
A
A
All
right,
if
we
don't
have
anything
more
on
the
topic
we
can
switch
to
social
trees,
is
not
here
today.
So
we
will
switch
to
geocent
Elementary's
exponential
Instagram
description
that
we
had
two
weeks
ago
and
now
it's
a
continuation
because
I
essentially
David
and
me
attended
the
open
Telemetry
meeting
yesterday.
That
was
discussing
this
kind
of
topic
with
the
committee's
team
and
at
the
end
of
this
meeting.
A
So
they
are
kind
of
going
into
the
Primitive
way
of
like
implementing
these
new
histograms,
and
so
they
are
having
like
some
kind
of
Middle
Ground
now
today,
and
we
have
a
discussion
with
David,
where
we
kind
of
wanted
to
kind
of
explore
what
would
be
the
future
for
us
in
kubernetes
with
this
music
histogram
and
what
would
be
the
challenges
essentially
like
one
of
the
main
issues
that
we
kind
of
we
didn't
know
like
what
could
happen
with.
A
It
is
the
cardinality
because,
with
normal
Instagram,
we
already
see
like
some
of
the
histograms,
with
only
like
10
buckets
already
of
like
so
much
Cardiology
like
they
generate
so
many
time
series.
So
what
would
happen
if
we
have
like
this
new
committees,
exponential
Instagram
that
have
like
hundreds
of
buckets
and
I
discussed
that
with
one
of
the
Committees
maintainer
and
apparently
so?
A
First,
the
way
it
will
be
stored
in
the
times
3
database
is
different,
and
so
essentially,
instead
of
having
the
buckets
as
a
dimension
of
the
metrics
like
we
used
to
have
so
one
bucket
would
be
one
type
seven.
Now
the
buckets
are
part
of
the
Matrix
in
the
tsdb,
so
they
are
not
counted
up
Dimension
anymore,
so
that
will
reduce
the
more
usage
of
The
Matrix
by
a
hole
like
we
won't
have
to.
We
take
that
into
accounting
Academy
anymore.
A
But
what
I
was
told
is
that
if
the
metric
were
already
posing
a
problem
in
terms
of
cardinality
with
The
Advocates
before
with
a
new
implementation,
that
would
still
be
the
case
like
we
would
still
need
to
worry
about
the
increasing
number
of
buckets.
That
might
happen,
and
that
may
still
be
an
issue
even
though,
like
this
has
been
optimized
but
they've
looked
into
adding
limits
per
histogram,
so
very
strong.
A
You
can
configure
a
limit
and
say
well,
I,
don't
want
this
histogram
to
have
more
than
100
buckets
and
if
it
has
more
than
the
scale
will
go
higher
and
the
buckets
will
be
reduced
in
the
position
as
well.
So
they
always
thought
about
like
some
kind
of
mechanism
to
prevent
that.
B
A
B
Well,
so
so
let
me
explain
what
I
understand
the
migration
path
to
the
and
then
so
that
the
idea
was
like
from
our
perspective.
Let's
say
we
have
a
histogram
for
API
server,
latency
or
something
right
and
yeah.
We
take
that
histogram
and
keep
the
same
name
the
same
labels
and
stuff,
but
change
it
from
a
regular
histogram
to
an
exponential,
histogram
yep
right,
supposedly,
according
to
what
the
Prometheus
folks
have
said.
If
you
query
it
with
a
Prometheus
server
that
doesn't
support
exponential
histograms
it'll
return,
a
fixed
bucket,
histogram
representation
of
that
histogram.
B
B
If,
like,
let's
say,
we
had
10
buckets
before,
if
we
were
able
to
limit
it
to
10
buckets
afterwards,
especially
if
it
applied
to
the
fixed
bucket
representation
or
the
thick
sorry,
the
fix
yeah,
the
fixed
budget
representation
of
it,
then
that
might
be
useful
for
us,
because
we
could
like
we.
The
bucket
boundaries
would
still
change
yeah.
B
A
B
A
fixed
bucket
histogram
to
an
exponential
bucket
histogram
at
query
time
without
it
changing
any
of
the
data
like
they
want
that
to
be
seamless,
but
it
does
mean
that
if
we
switch
it
over
in
our
code,
we're
basically
like
there's.
No
there's
no
like
way
of
easing
that
transition
in
code
for
us,
which
I
think
is
going
to
make
it
difficult.
B
From
the
format
perspective,
as
long
as
they
can
still
get
a
fixed
bucket,
histogram
out,
I
think
it's
like
we
could
maybe
call
it
backwards
compatible.
The
only
question
for
me
again
is
the
number
of
buckets:
if
we're
now
giving
them
100
buckets
and
they
have
to
parse
all
of
them,
then
like
they're,
gonna
notice,.
A
Yeah
yeah
I
think
that's
difficult,
but
we
still
have
a
lot
of
time
actually
like
to
think
about
that,
because
they
are
only
at
the
stage
of
POC
and
for
now,
like
the
just
trying
things
around
optimizing.
The
algorithm
and
things
like
that,
so
yeah
I
see
a
long
way
before
I
didn't
see
prediction
but
yeah
the
sooner.
We
have
like
an
idea
of
what
would
be
the
challenges.
The
result
would
be
I
guess.
B
A
Yeah
yeah-
that
sounds
great,
this
one's
great.
Let's
do
that
I
can
stop
working
on
it
and
we
can
show
our
concerns.
Yeah
increase.
It.
C
Awesome
good
yeah,
yes,
so
yeah
it's
authority
to
the
aps
over
metric
yeah.
So
it's
a
noun
issue,
I
think
yeah.
It's
the
cardinality
is
very
high
yeah.
It's
about
yeah
yeah,
so
I'm
wondering
whether
we
can't
use
trees
to
replace,
replace
that
metric
I,
don't
know
whether
it's
possible.
So
basically,
the
idea
is
that
we
use
choices
and
the
one
simplify
the
metric
so
that
metric
with
the
same
name
but
only
captures
the
high
on
the
high
latency
cases
and
a
drop
out
labels.
C
A
C
A
Are
using
all
of
them
today
and
even
like
six
scalability,
is
relying
on
that
to
build
their
slos
and
to
test
that,
like
in
their
test
by
plane
that
even
like
with
a
hundred
node,
we
are
not
going
above.
Our
slos.
C
Yeah
so
basically
I
want
to
have
an
aggregate
metric
and
that
algorithmetric
only
captures
the
high
latency
cases
for
low
latency.
Another
value
yeah
because
for
alerts
I
think
it
only
cares
about.
Is
the
high
latency
for
low
latency
I
think
we
just
can't
know
it.
If
one
we
want
to
know
more
details,
we
can
use
the
EPS
over
treating
Trace
data
to
say
the
details
and
we
can
also
drop
some
labels.
C
C
And
yeah.
B
So
what
what
you're
proposing
sounds
actually
a
lot
just
like
changing
the
bucket
structure
right
to
focus
on.
C
A
As
well,
because
even
like,
even
with
this
metric
and
as
available
as
it
is,
you
still
need
logs
or
traces
in
order
to
debug
any
kind
of
latency
issue
because
well
latency
show
are
complex,
and
if
it's
not
like,
if
your
Oslo
is
going
down,
not
because
of
errors,
and
it's
because
of
the
latency,
then
it
can
be
anything
in
your
cluster
because,
like
the
API
server
is
at
the
center
of
everything
would
be
a
good
way
to
do
it.
A
C
Yeah
yeah
I
will
look
like
look
into
the
new
new
new
Hitler
and
scratcher
later
to
know
more
yeah
I
think
it
can.
Can
we
do
some
some,
the
Carnation,
Maybe
Yeah,
so
basically
I
think
the.
A
Even
though
it
do
it
takes
a
lot
of
meaning
to
me
and
like
to
I
guess
a
lot
of
people.
This
metric
is
like
the
the
most
important
one
out
of
all
of
the
metrics
that
the
kubernetes
exposes.
So
we
kind
of
went
with
the
idea
that,
even
though
it's
taking
a
lot
of
memory,
considering
the
amount
of
things
we
can
do
with
it,
we
can
afford
the
fact
that
it
takes
I.
Don't
know
like
twenty
percent
of
the
time
series,
but
yeah
yeah.
C
A
B
Your
metrics
should
cover
100
of
the
case,
so
you
have
high
quality
signals
and
then
traces
are
for
debugging
particular
cases.
I
think
exemplars
will
help
a
lot
with
this.
B
There
there
is
a
school
of
thought
that
says
that
you
should
turn
on
a
hundred
percent
tracing
and
then
generate
metrics
from
your
traces,
but
the
end
of
the
day.
You
need
the
metrics
anyways,
so
the
route
that
you
take
to
get
there
isn't
particularly
important
and
I
think
we
already
have
this
established.
So
we
should
stick
with.
A
It
yeah
yeah
I,
think
I
wrote
today's
more
like
going
from
metrics
to
logs
to
traces
in
that
direction
and
not
like
from
traces
to
Netflix
so
and
going
like
thinking
that
that
could
be
and
like
reflecting
on.
That
would
be
all
because
everything
is
built
that
way
today,
but
yeah.
A
So
the
idea
that
I
had
and
I
discussed
with
you
David,
something
was
to
add,
like
this
examples
to
this
particular
metric,
like
the
one
with
the
latency
and
also
like
a
trace
ID
propagated
into
the
LG
clouds,
so
that
at
least
there
is
a
way
to
Via
this
latency
metric
to
go
deeper
into
like
what's
happening
really,
because
it
doesn't
give
you
much
info
like
you
only
know
that
some
of
the
requests
are
slow
when
you
are
using
it,
which
is
great
like
for
you
as
solo
and
for
what
you
promise
to
your
customer
and
stuff
like
that.
A
A
So
I
guess
we
can
get
four
minutes
back
of
our
time.
See
you
everyone
on
the
next
call
right.