►
From YouTube: SIG Instrumentation 20210624
Description
SIG Instrumentation Bi-Weekly Meeting June 24th 2021
B
All
right,
I
think
the
recording's
on
welcome
everyone.
The
this
is
the
sick.
Instrumentation
meeting
today
is
june
24th
and
we
have
a
couple
of
items
on
the
agenda
and
I
think
elena,
do
you
want
to
take
the
first
one
or
do
you
just
want
me
to
make
the
announcement.
B
C
Watch
one
I
threw
this
on
the
agenda
and
it's
mostly
to
I
guess,
pastor
david,
because
we
only
have
two
caps
targeted
for
this
release,
so
one
of
them
is
the
structured
logging
stuff
and
that's
being
handled
by
working
group,
structured
logging,
so
we're
just
deferring
to
them.
So
the
other
is
tracing.
That's
our
problem
so,
and
I
saw
that
david's
pr,
I
think,
is
ready
for
review.
C
So
I
just
wanted
to
make
sure
that
we
had
it
on
the
agenda
that
we
made
sure
that
we've
got
reviewers
and
approvers
assigned
from
the
institution
side
that
if
we
need
them
from
other
sigs,
we
know
who
they
are.
So
we
can
go
pester
them
because
I
know
this
is
like
slipped
now
like
three
releases
or
something
like
that,
and
that
doesn't
feel
great.
So.
D
I
think
we
definitely
have
good
coverage
from
the
instrumentation
side.
We
have
lily's
review
that
frederick's
reviewed
it
and
han's
also
taken
a
couple
passes.
I
think
alana
did
you
review
it
as
well.
E
B
D
D
Minutes:
okay,
with
liggett's
review.
E
Yeah
yeah,
I
think
I
think,
he's
happy
to
have
anyone
review
it.
That's
not
himself.
D
D
D
So
one
we
need
to
add
the
interceptors,
so
it's
sort
of
like
the
grpc
version
of
adding
an
http
handler
or
using
yeah,
and
then
because
we
don't
want
to
use
globals
in
the
api
server.
We
need
to
plumb
the
tracer
provider.
All
the
way
through
so
there'll
be
a
bunch
of
functions
that
it
gets
added
to
and
then
we're
just
going
to
use
the
standard,
grpc
interceptors.
E
A
D
Yeah,
so
look
for
those.
I
hope
that
those
will
be
less
contentious
than
the
first
one
has
been
since
they're
doing
fairly
straightforward
things
and
they're
there's
not
as
many
things
to
have
opinions
about.
I
think
so.
Yeah
I'll
send
those
around
yeah
yep.
That's
what
I've
got.
I
also
signed
up
to
write
a
blog
and,
if
you'd
like
to
help,
I
know
I've
been
doing
most
of
this,
so
I'm
going
to
assume
that
I'll
write
the
whole
thing,
but
if
anyone's
interested
and
wants
to
collaborate
on
I'm
happy
to
work
together,.
E
B
Cool
okay,
cool
anything
else
on
on
that
note,
if
not,
then
I
think
this
is
lily's
point,
but
I
think
this
is
continuing
continuation
of
our
discussion
from
last
week.
I
think.
A
Yeah
I
just
I
remembered
we
started
a
discussion
I
think
in
triage,
and
then
we
were
like,
let's
park
it
for
this
one
I
forget
which
metric
it
was,
but
it
was.
C
G
Let
me
find
it
hold
on
it's
just
so
hard
to
find
stuff
on
github.
I
get.
A
B
Found
it
through
tree,
I
think.
E
A
G
G
C
C
E
E
E
B
So
I
I
like
this,
I,
what
I
would
ask
for
is
if
we
promote
a
metric
like
this,
could
we
have
some
sort
of
a
signal
of
these
kind
of
requests?
Failing
right,
because,
ultimately,
that's
what
we
like
what
it?
What
is
it
that
we
want
to
detect
with
this
metric
right,
and
I
I
guess
it's
something
to
the
extent
of
their
being
interrupted
for
some
reason
or
right
like
what
is
the.
E
B
B
So
we
would
need
a
metric
for
like
watches
started,
I
guess
or
something
like
that,
because
that
is
the
thing
that
you're
worried
about
right.
If
there's
a
high
increase
in
watches
being
opened
or
the
total
length
of
a
watch
being
short.
G
Let
me
actually
look
at
the
magic.
What
what
does
the
metric
actually
look
like.
B
C
So
maybe
to
summarize,
there
were
a
bunch
of
issues
that
we
saw
right.
So
one
is
like
so
issue.
One
is
we
have
the
duplicated
metrics
where,
like
one
is
kind
of
a
subset
of
the
other
and
then
another?
Is
that
the
name
of
the
one
that
is
the
superset
has
gauge
in
it?
So
it
like
doesn't.
A
C
With
the
metrics
naming
policy,
so
we
can't
make
it
stable.
Was
there
anything
else
that
was
a
problem
here.
B
B
B
C
We
don't
have
a,
we,
don't
have
anything
to
say.
We
want
something
like
this
and
a
plan
to
promote
it
to
stable.
I
would
say
I
guess,
there's
also
sort
of
like
issue
four,
which
is
that
we
don't
own
this
metric,
it's
owned
by
api
machinery,
so.
B
C
So
I
guess
like
given
this
issue
list.
Does
this
give
us
a
clear
roadmap
like
we
know
that
we
have
one
that's
a
subset
of
the
other,
so
we
want
to
get
rid
of
the
subset
metric.
We
know
that
there's
a
violation
of
the
naming
policy,
so
we
want
to
fix
that
and
then
we
want
to
make
it
stable.
Like
does.
That
is
that
enough
of
a
proposal.
A
I
think
that
one
thing
that
I'm
a
bit
confused
by
is,
we
haven't
really
fully
agreed
on
what
makes
it
stable,
at
least
judging
by
the
previous
discussion
right,
what
what
makes
a
metric
be
stable,
because
last
time
I
think
we
discussed
like
if
it's
a
an
alert,
metric
or
a
dashboard
metrics,
then
we
should
make
it
stable,
but
we
haven't
really
fully
decided.
Have
we
I.
E
I
work
backwards.
You
know
I
work
from
slos
like
metrics
that
are
used
in
slos
or.
E
Ones,
right,
metrics
that
are
commonly
used
in
alerts.
A
Yeah,
maybe
we
can
just
document
that
somewhere,
yeah
yeah.
F
C
Sorry
sorry
lily
for
this
one,
I
would
say
that,
like
it's,
probably
a
fair
thing
to
promote
to
a
stable
metric
in
the
sense
that
like
if
we
ask
a
question
of
how
do
we
use
this
to
like
quantify
an
actual
symptom
that
an
operator
can
respond
to,
we
can
say:
oh
look,
the
tail
latency
is
really
high,
that's
going
to
be
user
impacting
and
so
like
they're
going
to
see
that.
So
that
would
be
a
problem.
B
Yeah,
I
I
feel
I
I
think
we
all
agree
on
the
intention.
I'm
not
sure
this
metric
either
of
those
metrics
actually
give
us
that
today,
and
so
I
think,
that's
that's
the
only
gap
that
we
haven't
actually
figured
out,
but
we
all
agree
on
the
intention.
So
I
think
that's
just
the
thing
that
we
need
to
figure
out.
E
B
E
I
I
do
have
I
have
a
plan.
My
plan
is
to
make
it
almost
parallel
to
request
the
request
metrics,
which
there
are
two
of
them,
there's
the
total
and
there's
the
seconds,
but
in
this
case
like
we
probably
didn't
even
need
the
total,
because
the
count
is
already
in
the
in
the
histogram.
But
in
this
case
it's
a
gauge.
We
actually
do
need
to
yeah.
B
C
An
action
to
put
together
a
proposal
to
drive
like
first
like
coming
up
with
a
replacement,
metric
and
deprecating
the
other
two.
E
B
In
hindsight,
I'm
not
sure
alpha
was
ever
a
good
name,
because
many
metrics
just
are
never
gonna
be
promoted
right
and
it
feels
like
alpha
seems
wrong
like
giving
it
that
name
sounds
like
they're.
It
will
be
promoted
at
one
point
or
another
yeah.
A
E
C
Will
say
for
like
from
the
enhancement
planning
perspective,
it's
a
problem
right
now
that
we
don't
have
beta,
because
if
people
add
metrics
as
part
of
an
alpha
implementation
and
they
want
to
graduate
their
feature
to
beta,
there's
no
corresponding
change
promoting
those
metrics.
So
then,
all
of
a
sudden,
it's
like
they
get
promoted
to
stable
at
the
last
minute,
and
there
was
no
instrumentation
review
in
between
and
at
that
point
it
might
be
too
late
for
us
to
give
feedback
on
a
thing.
C
B
So
I
think
what
ncd
did
is
very
similar
to
what
how
I
think
about
a
lot
of
the
metrics.
I
think
etcd
and
then
correct
me.
If
I'm
wrong,
I
think
ncd
has
a
specific
debug
prefix,
where
they
say
these
are
metrics
that
are
that
are
kind
of
bound
to
the
code
surrounding
it,
and
so,
when
this
code
changes
inevitably
these
metrics
will
go
away,
probably
or
change
significantly,
and
I
I
I
would
like
to
have
something
like
this
to
signify
this.
There
is
no
promotion
process
for
this
metric.
H
E
B
Like
I
mean
I
I
I
like
debug
metrics,
I
don't
know
I,
but
that
could
just
be
my
predisposition
of
having
worked
with
ncd
for
like
five
years
right,
but
I
guess
what
I
was
also
trying
to
say.
On
the
same
note,
I
think
the
majority
of
metrics
that
we
have
are
actually
on
that
note
and
the
debug
minority
are
actually
the
slo
ones
and
those
should
absolutely
be
on
their
path
towards
graduation.
B
B
Internal
I'm
fine
with
internal
as
well.
It
kind
of
reflects
the
same
notion
and
then
the
other
thing
was
beta,
which
I
guess
we
still
need
to
figure
out.
D
E
Yeah
well,
we
would
be
changing
the
existing
behavior,
so
yeah.
We
would
have
to
have
a
flag
that
disables
debug
metrics.
Otherwise
we
could
break
people.
C
We're
keeping
an
eye
on
time,
because
we
have
one
more
thing
on
the
agenda:
do
we
have
somebody
to
own
drive,
put
together
a
proposal
for
what
we
want
to
see
here,
because
I
don't
have
a
great
idea
of
what
we
want
or
who's
gonna
make
that
happen.
B
In
it,
I
I'm
happy
to
flesh
out
the
details
of
each
of
the
steps,
because
the
more
I
think
about
it
now,
if
we
introduce
a
debug
type,
then
I'm
not
sure
a
beta
type
is
actually
still
needed,
because
then
the
alpha
metric
is
actually
the
type
that
we
are
intending
to
eventually
graduate,
which
was
the
thing
that
we
wanted.
This.
C
Is
why
I
think
that
we
should
keep
alpha
as
alpha
and
promote
all
the
alpha
stuff
to
beta,
but
I
think
we
just
need
a
proposal,
so
I'm
gonna
say
in
the
notes,
han
and
frederick,
you
can
work
on
that.
H
H
So
let
me
talk
about
what
I
cropped
up
last
week,
so
han
said
that
he
really
likes
globals,
so
I
think
I
am
trying
to
get
rid
of
one
hello.
I
don't
know
if
you
heard
about
this.
G
H
H
Yeah,
it
really
is
not
like
one
line
global
that
became
pretty
pretty
big
so
insured
to
to
really
graduate
alternative
login
formats.
We
need
a
way
to
to
like
configure
and
manage
and
provide
users.
H
Make
give
users
like
assurance
that
this,
in
this
alternative
format,
like
has
the
same
features
works?
The
same
is
like
has
the
like
will
not
break
you
like?
Basically
will
not
break
you
so
for
that
we
need
to
have
comparable
features
that
are
provided
by
k,
lock
and
the
the
here,
mostly
json
format,
as
we
are
developing
it
further.
H
So
we
brought
with
the
122
we
implemented.
I
think
one
feature
that
was
missing:
what
was
color
information
so
source
code?
So
now
you
can
you
get
the
color
information
in
in
your
logs,
which
is
great,
because
there
are
many
complaints,
and
so
we
know
that
it's
important,
but
we
we
want
to
to
graduate
it.
We
we
we
need
to
make
sure
that
k-lock
and
k-log
and
the
json
doesn't
like
go
into
different
ways
that
we
have
a
core
functionality
that
is
delivered
by
both
the
solution
and
people
can
add
more
solutions.
H
We
don't
over
restrict,
so
we
can
add
more
alternative
implementation
of
logging.
Where
one
example,
I
think
what
someone
someone
asks
about
asked
about
integration,
a
journal
for
cubelet
like
instead
of
having
a
text
in
your
journal,
as
most
people
use
systemd
for
your
cubelet.
It
would
be
great
if
you
just
get
structure
logging
within
your
cubelet
internal,
so
yeah.
H
The
proposal-
basically
it's
over
for
now
over
like
aggressively,
gives
you,
let's
duplicate
everything
that
doesn't
serve
the
original
purpose
of
logging
and
and
all,
but
only
deprecate
it
within
carbonite's
components
and
leave
the
k-lock
people
like
to
their
own.
But
we
care
about
the
those
components
I
wanted
to
first
go
through
through
sig
instrumentation
for
your
feedback
and
then
possibly
to
seek
arch.
Have
you
talked?
Have
you
talked
to
jordan
about
this
jordan?
I
talked
so
I
could
think,
but
for
now
no
no.
E
H
I
know
is
there
any
reason
that
we
could
not
treat
those
flags
are
as
any
other
flag
as
we
are
not
breaking
caleb
losers,
which
is
apparently
a
huge
community
of
people
that
use
k-lock?
We
just
want
to
deprecate
in
in
kubernetes
in
the
core
components,
those
flags
that
we
think
are
not
needed
and
we
have
already
processed
for
defecation
like
we
can
comply
to
it
and
the
I
often
first
question:
are
we
okay
with
like
dropping
flux
from
k,
lock
or
changing
the
flags
for
k?
Look.
H
We
already
did
at
some
point
change
the
flux.
So
basically,
kubernetes
went
through
standardizing
the
format
how
flux
use
hyphens
versus
underscores
and
which
we
drop.
People
like.
E
E
Talks
about
this
is
that,
like
basically
k,
log
uses
very
commonly
used
flag
names
that
collide
with
other
loggers,
and
so
basically,
you
can't
test
two
loggers
simultaneously,
because
you
have
flag
collisions.
C
E
C
A
quick
question
because
I'm
looking
at
the
flags
that
are
being
proposed
for
deprecation
and
it's
not
clear
to
me
why
they
should
be
proposed
for
deprecation
for
json,
like
looking
at
things
like
log
file
max
size,
for
like
log
rotation
like
json
logs
on
disk.
You
know,
like
you're
gonna,
have
a
full
serializable
log
line
and
then,
like
you
know,
that's
when
you
would
drunk
it,
it
wouldn't
like
truncated
halfway
through
the
line.
C
So
I'm
I'm
not
sure
why
that
would
be
considered
not
applicable
similar
for,
like
writing
to
standard
error
versus
standard
out
like
this
is
a
different
file
descriptor.
Why
does
it
matter
where
we're
writing
the
json.
H
So,
okay
about
maybe
starting
from
files
so
like
at
the
beginning
of
the
proposal,
I'll
just
stay
like
I,
I
just
I.
I
said
that
we
should
comply
to
some
basic.
You
know
idea
or
have
some
set
of
rules
that
we
want
to
comply
and
having
application
card
about
logging
that
gives
people
so
much
weight
or.
H
Way
to
use
it
and
abuse
it
to
to
add
more
features
like
like
writing
to
files
changing
frequency
of
flashing
and
syncing
to
disk.
I
know
I
just
read
only
you
know
recently
that
apparently
k
log
syncs
every
five
seconds
or
kubernetes
syncs
syncs
to
disk
like
calls
a
full
sync
on
all
your
disk,
that
a
sync
on
x,
the
free
format.
If
you
have
linux
syncs,
all
the
disk
on
data
all
for
disk
doesn't
work
for
single
files,
so
this
is
basically
just
to
comply
to.
H
Those
flags
would
be
a
nightmare
to
implement
and
we
need
to
adjust
the
core.
If
we
just
standardize
those
three
flags,
it
should
be
enough
to
get
people.
I
understand
like
it
would
be
useful,
for
I
know,
love
rotation,
but
there
are
already
so
so
many
tools
that
can
like
give
you
the
same
experience
and
from
what
I
know
like
at
some
point.
C
So
I
have
a
request
because
that's
not
documented
in
the
issue,
so
my
request
would
be.
Can
we
make
sure
that
that
and
any
motivating
information
for
like
why
each
flag
we
want
to
deprecate
because
we
may
need
to
make
a
case-by-case
call
on
them?
So
if
we
could
include
like
sort
of
a
table
list
thing
with
like
we
don't
want
this,
because
it's
hard
for
these
reasons
and
so
on.
I
think
that
would
be
helpful
in
trying
to
make
a
decision
because
it
sounds
like
we're.
Gonna
break
a
lot
of
backwards.
C
Compatibility
potentially
like
I
know
I've
seen
lots
of
bugs
in
terms
of
like
you
know,
cubelet
like
not
possibly
like
supporting
one
of
these
flags
properly.
So
I
think
people
care.
H
H
B
We
are
also
two
minutes
over
time,
so
I
think
that's
a
good
idea
to
take
that
offline
is
there.
We
are
two
minutes
three
minutes
over
time.
Is
there
anything
else
that
people
want
to
say
last
word,
or
should
we
adjourn
and
see
y'all
next
time
I
don't
like
globals.
I
was
joking
just
wanted
to
make
that
to
you
in
the
recording
yeah.
Just
just
for
the
record.
B
All
right
in
that
case
have
a
wonderful
local
time.
Everyone
bye.