►
From YouTube: SIG Instrumentation 20201210
Description
SIG Instrumentation Dec 10th 2020
B
Yeah
I
can,
I
can
take
it.
Okay,
welcome
everyone.
This
is
the
consultation
meeting
today
is
december
10th,
and
this
is
actually
our
last
instrumentation
meeting
for
the
year.
So,
let's
make
it
a
good
one,
okay.
So
the
first
point
we
have
on
the
agenda
is
continue
discussing
on
metrics
naming
policy.
C
This
is
about
the
whether
to
use
cube,
prefix
metrics
from
components
other
than
cubestate
metrics,
and
I
think
we.
C
B
Personally,
I
think
that's,
okay,
I
think
that's
just
fine,
but
it's
inconsistent.
A
C
I
think
the
main
follow-up
is
that
we
should
update
our
instrumentation
guidelines
such
that
it's
consistent
going
forward,
so
I
believe
actually.
C
And
she's
not
in
attendance
today,
but
I
think
let's
differ,
maybe
yeah,
I'm
okay
with
that
we
can
defer
it
until
the
next
year.
Okay,
I
also
think
if
someone
wants
to
open
a
change
to
the
guidelines,
that
might
also
be
a
good
place
to
discuss
in
the
meantime,.
B
Okay,
I
guess:
if
someone
wants
to
go
ahead
and
open
those
changes,
they
can
do
that.
I
would
say
for
now
we
can
go
on
with
our
next
with
our
next
point.
So
we've
got.
We
don't
have
anyone
tagged
on
this,
but
what's
going
on
with
stablemetrics
lots
of
bugs
against
website
for
the
list
being
empty.
B
A
A
Yeah
the
plan
is
is
to
to
to
elevate
some
of
these
like
entity.
Objects
should
probably
be
stable
121..
We
could
flip
that
one
I've
been
talking
to
some
people,
api
machinery
who
might
be
interested
in
taking
up
the
api
server
processing
time
for
requesting
latencies,
because
right
now,
all
of
our
request.
Latency
metrics,
include
webhook
processing
time,
which
makes
it
on
slowable,
because
we
don't
control
webhooks.
B
I
think
that's
kind
of
silly
like
an
slow
intended
to
to
show
user
experience
right.
So
if
we
don't
include
it,
then
that
kind
of
defeats,
the
purpose
of
setting
an
slo
in
that.
A
B
Me
try
to
get
up.
I
I
think
that
that
that
kind
of
thing
is
up
for
a
provider
to
define
and
if
they
have
a
definition
like
that,
then
they
need
to
enforce.
I
don't
know
resource
quota
or
whatever.
A
Let
me
find
the
thing
I
can
find
this:
let's
certainly
let's
move
on
and
then
I'll
I'll
find
the
link.
B
But
in
general
I
think,
like
request
and
duration
metrics
should
we
should
consider
it
to
make
to
to
graduate
them
to
to
stable.
I
think
there
are
a
number
of
other
ones,
especially
those
that
are
important
for
setting
slos,
like
I
do
think,
should
be
stable.
I'm
not
sure
I
entirely
agree
with
etsy
object
cons.
A
D
I'm
going
back
maybe
to
original
automatic
about
matrix
stabilization
or
like
list
of
stable
metrics.
I
was
thinking
about
this
and
like
the
main
problem,
is
that
we
don't
have
six
like
actively
having
any
motivation
like
to
to
go
forward,
and
it's
not
our
like,
also
job
to
to
to
to
go
run
after
them
and
chase
them.
But
there
was
a
music
created
like
for
a
secret
group.
D
Group
for
reliability-
and
I
was
interested
in
talking
with
maybe
why
tech
or
overall,
like
yeah
vertically,
if
they
can
motivate
as
part
of
definition
of
reliability,
motivates
introducing
observability
or
some
sort
of
stable
signals
that
everyone
can
depend
on
to
provide
reliability
and
drive
the
the
stabilization
of
metrics.
For
this
I
know
that's.
D
Sense,
yeah.
That
was
like
only
very
barely
like
scratchy
first
idea
and
maybe
discussed
it
with
someone
else
yeah,
but.
B
Yeah
this
sounds
pretty
reasonable
to
me
like
I
I
do.
I
do
think
we
need
to
graduate
metrics
that
are
important
for
slos,
but
I
don't
think
we
should
create
stability
for
things
that
are
internal
implementation.
Details.
It's
like
that's
kind
of
like
yeah.
That's
how
that's
how
I
see
it
and
like
and
lcd
objects
count.
I
th
iron,
I
agree
is
important,
but
I
don't
think
that
it
qualifies
as
an
slo
in
itself.
B
So
anyways,
that's
how
I
think
about
it.
A
What
are
people
using
for
charts
and
alerts
that
we
cannot
make
changes
without
breaking
people
right
like
and,
if
you
think
about
like
the
ones
that
are
the
most
important
like
yes,
api
requests,
latencies,
obviously
slo,
because
latency
thing
std
object
counts
is
more
kubernetes
specific,
but
it
is.
A
Not
that
bad,
because
you
have
like
it's
it's
bounded
to
the
number
of
crds
in
a
specific
cluster,
and
you
you
know
for
for
red
hat,
I
would
think
right.
It
is
very
important,
especially
with
openshift
stuff
right
like
so
I'm
actually
surprised
so.
B
I
mean
I
maybe
someone
who
actually
still
works
at
redhead
can
answer
this,
but
as
far
as
I'm
aware,
there's
no
alerting
on
this
on
this
metric,
but
it
is
being
reported
to
red
hat
so
that
red
head
knows
about
the
number
of
lcd
objects
in
each
in
each
cluster.
E
Yes,
we
expanded
it
a
bit
more
to
include
a
bit
more
information
we
send,
but
we
don't
alert
on
it.
No
from
what
I
just
remember
off
the
top
of
my
head.
B
Precisely
because
the
the
like
red
head
mostly
follows
the
the
rule
of
like
symptom-based,
alerting
right,
and
that
is
100
percent
a
cause.
The
number
of
ncd
objects,
it's
very
possible
that
there's
an
entirely
different
cause
that
is
causing
the
api
latency
to
be
high
right
and
that's
why
we
want
to
alert
on
the
api
latency,
not
to
say
that
the
alerting
on
density
objects
also
is
wrong,
that
that
makes
sense.
But
the
combination
is
really
what
like
leads
you
to
the
to
the
root
cause
right.
E
Yeah,
exactly,
I
think,
matthias
can
speak
about
the
slo
alert
I
think
he
joined.
We
used
the
ones
that.
E
In
the
high
defined
in
the
kubernetes
mixin,
so
we
use
that
in
red
hat
as
well.
So
that's
the
thing
that
would
make
sense
to
be
stable
from
our
perspective,
at
least
but
yeah.
I
agree.
I
mean
the
fcd
metrics
as
well.
The
object
count,
but
for
different
reasons,
I
guess.
A
E
Us
yeah,
sorry
for
us,
like
our
clusters,
can
be
very
varying
in
size.
So
what
is
large
number
of
accounts
of
objects
is
different
for
each
customer
right,
so
it
doesn't
make
that
much
sense
right
so
for
for
us,
the
causes
are
not
something
we
alert
on.
A
D
D
Yeah,
so
I
just
wanted
to
to
some
to
tell
you
if
you
want
to
define
the
slo,
you
usually
want
to
have
some
boundaries
like
you
cannot
have.
I
know
unlimited
have
like
infinite
boundary
about
okay.
We
support
all
possible
number
of
lcd
objects
within
this
slo.
You,
you
give
some
boundaries
like
open
source,
scalability
envelopes
like
this
number
of
objects,
this
the
type-
and
this
is
supported-
and
this
is
within
slr.
It's
not
part
of
what
we
monitor
tesla,
it's
something
that
we
decide
if
slo
can
cover
such
cluster
or
such.
A
B
I
guess
I
mean
we
never
we
never
discussed
any
any
rules
like
that
right,
so
it
is,
it
is
totally
up
for
discussion.
The
what
I
would
I
just
worry
about
is
that
if
we
accept
something
like
cd
objects
counts
that
that
kind
of
opens
like
sort
of
pandora's
box
and
everybody
wants
all
the
metrics
to
be
stable.
All
of
a
sudden.
B
That's
just
what
I
worry
about,
and
I
agree
like
I
guess
that's
what
I've
been
saying
all
the
entire
time
now
that,
like
slo
ones,
definitely
need
to
be
stable.
After
that
it
becomes
more
vague.
C
A
Broadly
yeah,
and
so
from
the
api
machinery
side,
I'm
saying
I
am
personally
planning
on
making
a
city
object
counts
as
stable
metric,
because
it's
important
to
me
and
I
I
will
bring
it
up
dave,
I'm
sure
I
feel
like
the
other
people.
There
also
think
it's
important.
E
I
don't
disagree
with
making
things
stable,
but
I'm
curious:
how
do
we
so
and
do
we
have
any
versioning
as
well
like
if
we
in
the
future
want
to
go
to
version
two
but
still
keep
the
previous
metric
or
something
along
those
lines,
or
do
we
just
we'll
always
keep
it
once
we
make
it
stable
and
just
introduce
another
metric
in
case?
Something
goes
wrong.
A
That
makes
it
you'd
have
to
do
yeah.
You
would
have
to
do
the
latter,
because
otherwise
you
could
potentially
bring
congestion.
B
I
I
guess
what
I
first
of
all.
Yes,
component
owners
can
propose
stable
metrics,
but
the
process
does
say
that
we
do
need
to
approve
them.
But,
however,
the
things
that
the
things
that
I'm
kind
of
try
was
trying
to
get
at
that,
I
don't
think,
can
ever
be
stable
and
fcd
is
like
mostly
does.
B
This
category
are
things
that
are
implementation,
specific
right,
where
the
implementation
might
evolve
over
time
in
some
specific
component
and
those
kinds
of
metrics
I
don't
think
can
ever
be
stable.
I
I
realize
this.
This
is
like
we're
so
pretty
far
in
with
etsy.
That
probably
ncd
will
always
be
there,
but
no.
A
I
mean
there's
kind,
there's
kind,
there's
you
know,
there's
k3s
right:
they
have,
they
have
the
storage
adapter
with
a
simple
light.
So
actually
that's
a
great
argument.
That's
a
that's
a
great
argument
for
before
we
turn
it
to
stable.
We
should
maybe
just
have
an
a
storage.
Internal
storage
object
account.
It's
a
resource.
Resource-Based
object,
account.
B
Yeah
that
I
think
that
that
I
could,
I
could
agree
more
with
that.
We
have
an
a
resource
object.
Count:
okay,
instead
of
an
etsy
object,
count.
A
B
That
sounds
good
great
seems
like
we're
like
okay,
we
have
10
more
minutes
left
and
we've
got
two
more
topics:
lily.
Take
it
away.
E
What
did
I
write?
I
guess
we
recently
more
recently,
we've
had
in
keepsake
metrics.
We
had
a
bunch
of
issues
for
feature
requests
around
metrics.
We
don't
necessarily
expose
so
things
that
are
not
one-to-one
mapping
to
the
kubernetes
api
and
I
think,
since
we
had
so
many
and
and
a
lot
of
them
do
make
sense,
there
is
nothing
wrong
with
them.
We
were
thinking
of
either
creating
a
new
sub-project
of
sick
instrumentation
for
those.
E
A
E
Brought
it
here
but
like
the
thing
is:
either
we
create
a
new
project
that
is
not
part
of
cube,
state
metrics
and
as
part
of
like
sig
instrumentation,
that
exposes
those
metrics
that
we
want,
or
we
add
it
as
a
new
metrics
endpoint
as
part
of
cubestate
metrics,
so
that
you
enable
via
flag.
I
think
that
was
the
main
discussion
we
did
on
the
issue.
A
E
E
Yeah
we
had
those
as
well
those
ideas
and
we
had
them
mainly
around
customer
resources,
but
not
sort
of.
E
Yeah
around
custom
resources
makes
sense,
but
around
the
specific
metrics,
which
are
calculations
that
we
know
should
make
should
be
there,
but
we
just
can't
map
it
correctly
would
not
be
like
they're
they're
common
for
every
user
right.
So
it
would
make
sense
to
be
done
by
us,
for
everyone
essentially
go
ahead.
B
Yeah,
I
I
think
that
makes
sense.
I
think
something
along
I
I
I'm
happy
with
trying
this
within
cubside
metrics
as
a
separate
metrics
endpoint.
B
I
I
don't
get
the
impression
that
we
could
make
use
of
the
optimizations
that
we
did
for
the
like
normal
metrics
endpoint,
that
we
have
on
coupe
state
metrics,
because
those
are
pretty
specific
to
the
current
model
like
aggregations,
won't
work,
and
that's
specifically
what
this
user
was
requesting.
B
E
Coop
state
metrics
yeah
that,
like
I,
I
don't
necessarily
care
about
the
implementation
details
but
having
those
common
calculations
done
by
either
us
or
whoever,
whichever
component
metrics,
we're
exposing
so
that
the
people
who
have
the
knowledge
how
those
metrics
should
be
calculated
can
make
those
I'm
happy
with
whether
it's
in
the
code
or
if
it's
configuration,
I'm
fine
with
either.
B
Kind
of
both
like
some
of
some
some
aggregations
that
users
are
asking
for,
would
just
cause
wild
amounts
of
cardinality
that
are
completely
useless
unless
you're
doing
exactly
this
aggregation
so
pre-calculating.
This
would
save
a
ton
of
resources
which
would
make
sense.
That's
so
yes,
kind
of
both.
B
That's
kind
of
the
point
right
like
it's
extremely
cheap
to
do
this,
like
within
group
state,
metrics,
yeah
yeah
for
sure.
E
Yeah
we
had
a
bunch
of
them
that
we
closed
before
because,
like
the
first
paragraph
is
we
stick
to
the
kubernetes
api
one-on-one
and
we
do
the
mapping
so
yeah.
I
think
the
only
thing
is
I'm
happy
to
have
this
part
of
keepsake
metrics,
but
for
crds,
whether
we
want
to
put
it
under
this
or
keep
state
metrics,
I'm
not
sure,
but
we've
been
getting
a
lot
of
requests
for
crd
metrics.
A
Yeah,
I
mean
the
reason
why
I
was
asking
about
the
working
group
was.
I
actually
have
some
interest
in
extending
keepsake
metrics
and
I
was
wondering
what
the
forum
was
for
it
and
since
we
yeah
there's
just
a
lot
possibly
to
discuss
and
yeah.
E
Yeah,
I'm
happy
to
set
that
up.
Currently,
it's
just
opening
an
issue
really
and
just
discussing
there
between
the
maintainers
but
yeah.
We
can
do
that
as
well.
We
can
have
a
separate
working
group,
there's
definitely
a
lot
of
things
that
we
can
do,
especially
if
we
accept
the
new
aggregated
format.
I
think
there's
a
lot
of
nice
metrics
we
can
make
so
yeah
sounds
good.
We
can
discuss
that.
B
Yeah,
the
the
reason
why
I
was
mentioning
like
configuration
for
the
aggregations
is,
I
feel,
like
we're,
never
going
to
be
able
to
make
everyone
happy
with
the
chosen
aggregations
right.
It's
almost
like
doing
queries,
so
that's
why
this
I
feel
like
it
has
to
be
configurable,
but
it
makes
the
whole
problem
a
lot.
More
complicated
yeah
but
yeah
like
group
set
metrics
is
a
sick,
instrumentation
sub
project.
So
I
don't
see
why
we
can't
discuss
it
here.
Actually,.
A
Well,
mostly
because
we
have
four
minutes
left
and
the
extension
thing
would
be
a
a
lot.
It
could
easily
eat
up
the
entire
time.
B
F
B
I
I
would
say
we
can
start
it
within
coop
state
metrics.
If
we
realize
this
may
outgrow
group
state
metrics,
I
think
we
can
still
create
a
new
sub
project.
Should
we.
B
E
D
D
E
Oh
no,
I
just
wanted
to
say
that
like
for
us,
we
just
emailed
the
scalability
and
we're
writing
a
new
test
for
this,
because
it's
quite
unique,
I
think,
to
keep
state
metrics
but
yeah.
That's
all
sorry!.
A
Merrick,
were
you
talking
about
cubesat
metrics?
Were
you
talking
about
the
slo.
D
Thing
mean
keepsake
metrics,
like
I
just
wanted
to
make
sure,
because
this
is
something
that
I
was
also
looking
at
metric
server
just
to
to
make
sure
that
we
have
consistent
approach,
that
that
was
discussed
with
sex
guardians.
If
you
have
tests
like,
I
would
be
happy.
I
would
also
want
to
maybe
look
at
them,
because
this
is
something
that
I'm
looking
at.
B
We
haven't
actually
written
them
yet,
but
six
scalability
kind
of
instructed
us
where
those
would
go.
Yeah.
E
B
Okay
looks
like
we
have
discussed
everything
on
our
agenda
for
today
and
we've
got
one
minute
left.
I
guess
we
can
get
give
everybody
that
one
minute
left
it's
over.
So
time's
up.
This
was
our
last
meeting
for
the
year,
so
everybody
have
a
wonderful
holidays
and
a
happy
new
year
see
you
all
next
year.