►
From YouTube: July 2019 :: Ceph Developer Monthly
Description
Monthly developer meeting for the coordination of Ceph project development. http://tracker.ceph.com/projects/ceph/wiki/Planning
A
B
A
B
B
There's
example:
output
here
that
shows
so
I
set
the
the
warning
level
to
one
microsecond,
so
anything
over
one.
It's
going
to
report
basically
on
all
my
connections,
so
we
can
see
that
the
back
interface
had
a
long,
long
ping
time
and
the
longest
one
is
shown
in
the
summary
and
same
thing
with
the
front
interface
and
the.
B
Showing
you
which
connections
on
which
interfaces
and
then
I
changed
them
algorithm,
but
possibly
improving
right
now
is
showing
if,
as
long
as
it's
going
from
largest
to
smallest,
no
smallest
to
largest
meaning,
it's
been
lately
been
going
okay,
so
oh
sorry,
I
should
have
prefaced
this
with.
We
currently
do
one
minute,
five
minute
and
fifteen
minute
intervals
and
get
the
average
ping
time
across
those
those
intervals.
D
E
D
A
B
B
B
B
There
might
be
more
ways
of
sort
of
sorting
this
such
that
you
could
see
that
oh,
this
switch
is
failing
because
it's
only
say
the
the
rack
to
rack
communication-
that's
going
slow,
but
we
don't
have
that
right.
Now,
if
you
query
an
OSD
I
change
this
Josh
from
what
you
might
have
seen
before,
so
that
there's
consistency
with
the
manager
output.
So
if
you
query
an
OSD
you'll
just
see
every
heartbeat
interval
average
that
it's
maintaining
to
every
other
OSD
I
put
a
zero
here,
so
it
would
dump
everything
out.
B
A
B
B
B
So
something
that
we
talked
about
was
instead
of
averaging
over
the
interval,
is
to
get
the
worst,
the
worst
ping
during
the
interval
instead.
Now,
if
anybody
knows
more
about
networking
hardware,
networking
failures
that
could
suggest
whether
the
average
is
better
or
the
maximum
or
something
else
as
we're
monitoring
during
these
intervals,
that
would
be
wait
to
hear.
B
Okay,
so
would
you
say
that
it
should
be
the
so
I'm
getting
I'm
keeping
track
of
one
minute
data
and
then
I
average
that
into
the
last
five
minutes,
and
then
the
last
15
so
is
is
five-minute
maximum?
Is
that
the
the
maximum
of
each
of
those
minutes
or
is
it
the
average
of
the
maximums
of
the
five
one-minute
maximums?
G
G
E
G
B
B
B
G
E
B
B
B
E
D
E
E
E
E
B
Know,
fortunately,
it's
not
exactly
six
seconds,
it's
a
random
value
between
0.5
and
5.9,
I
think
based
on
the
six
seconds
yeah.
So
it
isn't
a
precise
minute
either
right,
because
when
the
next
heartbeat
triggers
it
sees
whether
a
minute
has
gone
by.
But
at
that
point
obviously
it
could
be.
You
know
I'm
in
in
five
seconds,
but.
B
And
I
I
did
change
the
config
in
my
book
in
my
pull
request
to
a
dev
option,
and
so
you
know
you,
you
would
completely
make
this
mechanism
invalid.
If
you
change,
if
you
change
the
interval
to
you,
know
greater
than
30
seconds
or
greater
than
for
sure,
obviously
greater
than
60
seconds,
so
I
changed
it
to
a
dev
option
and
it
has
a
restricted
range.
Even
if
ya
the
max
is
60,
so
I
guess
it
would
get.
One
trigger
per
would
only
get
one
trigger
per
minute.
B
Yeah,
okay,.
E
Unless
they
set
the
option
in
like
a
specific
monitor
or
only
in
the
Mon
section
or
something
like
that,
but
right
you're,
not
the
only
one
they're
a
whole
bunch
of
options
that
are
like
that
or
the
naming
is
sort
of
using
yeah.
B
Well,
yeah
I
called
it
Mon
well,
because
the
Mon
produces
the
warnings,
even
though
it's
part
of
the
dump
output
mechanism,
it's
it
is
for
the
warning.
So
that's
why
I
did
call
it
Mon
warned
so
if,
from
now
on,
we
can
kind
of
assume
that
configs
are,
for
the
most
part,
always
the
same
on
all
all
types
and
and
in
this
case
it's
nothing
critical
or
anything.
If
it
it's.
G
B
A
B
Okay
and
oh,
so
a
big
item
is
if
we
handled
load
averages
in
the.
If
we
added
load
averages
to
the
heartbeats,
then
we
could
kind
of
see
that
oh
well,
this
guy
is
slow
to
respond
to
a
ping,
but
hey
we
can
see
his
load
is
high,
and
so
he
it
probably
wasn't
the
network
now
I,
don't
know
what
the
math
is
gonna
be
on
that
you
know
how
much
load,
how
much
I
mean.
It's
probably
gonna,
be
tough
because
of
how
many
CPUs.
A
B
B
G
B
Right
I
mean
you
know
now
that
I've
been
playing
with
this
I'm,
almost
wondering.
If
we
even
care
about
us
having
the
OSD
report
his
his
heartbeats
I
mean,
maybe
you
could
have
an
option
to
the
manager,
one
that
just
says
only
only
show
me
OSD
zero
I
mean
there's
a
bit
of
a
delay
of
him
flushing
to
the
manager,
but
it
well.
E
G
B
And
then,
obviously
we
have
that
information
or
multiple
pieces
of
information,
so
Oh
ooh.
This
is
why
the
OSD
would
want
to
collect
load,
because
then
he
could
say:
OSD
choose
heavily
loaded,
I'm,
not
going
to
kill
him
after
the
20
seconds.
We,
it
would
have
more
flexibility.
There,
I'm
gonna,
give
him
extra
time.
B
G
A
Yeah
reports
it
down
and
then
the
monitor
can
size
whether
it's
actually
mark
it
down
or
not,
and
the
monitor
already
is
already
adjusting
the
grace
period
based
on
past
reports
that
have
been
false.
So
there
could
be
more
inputs
that,
like
load,
average
or
I,
guess
suggesting
and
Chad
just
get
disk
utilization.
H
Yeah
from
the
support
point
I
mean
we
have
seen
so
much
of
these
disk
utilization.
We
always
see
like
like
when
we
try
to
troubleshoot
flow
requests
or
like
performance
issues.
We
try
to
see
like
if
it
is
because
of
network
or
because
network
is
the
worst
is
not
responding
because
it
was
the
disk
itself
is
so
much
busy.
H
So
if
we
get
the
disk
utilization
in
this
septum
and
manager,
output
from
OSD
and
then
one
column
as
the
percentage
utilization,
it
will
be
very
much
clear
that
ok,
this
was
ji,
has
highest
this
utilization
as
compared
to
the
others
Oh,
because
loader
average
would
be
global
to
that
node.
But
the
dispute
realization
would
be
per
OS
D
yeah.
G
E
A
G
D
If
we,
if
we
start
adjusting
the
crease
based
on
certain
node
conditions
or
rescue,
ties
a
tional
conditions,
there
should
be
a
limit
as
to
how
much
we're
adjusting
at
some
point
we
should
probably
not
auto
adjust.
There
should
be
a
hard
to
stop
at
some
point.
Maybe
like
do
it
to
our
iteration,
so
three
iterations
and
then
maybe
report
it
and
then
let
the
user
decide
whether.
G
E
A
Shirt:
it's
that
one
thought
about
the
naming.
These
are
just
attracting
steeping
times
in
the
future.
Maybe
you
want
to
do
some
kind
of
network
monitoring
among
other
demons
too.
So
should
we
have
like
OST
in
the
name
of
these
things.
B
E
E
E
Okay,
so
the
idea
would
be
that
you
know
a
number
of
alerts
they
come
and
go.
You
could
additionally
have
a
mute
that
the
monitor
will
record
and
the
mute
is
associated
with
one
of
those
codes
and
it
has
a
time
to
live.
E
I
guess.
One
question
is:
if
don't
specify
time
to
live,
does
it
just
meet
up
forever
or
do
you
and
force
the
maximum
time
to
live
good,
go
either
way
and
beats
specify
which
the
late
going
mute
and
what
the?
How
long
will
I
meet
it
and
it'll
then
just
go
at
the
go.
Look
at
the
current
alert
that
matches
that
and
it'll
also
to
track
us
some
additional
stuff
to
match
that
particular
instance
of
that
alert.
E
We
could
have
an
a
wild
card.
That'll
meet
everything
I'm,
not
sure.
If
that's
hopeful
or
not,
it
might
be
nice
if
you're
on
the
CLI.
But
it's
also
like.
Why
would
you
meet
everything
without
knowing
exactly
waiting
meeting
so
who
knows?
I
have
a
strong
opinion
on
that
one,
but
once
the
once,
you
have
things
needed,
everything
you
ever
looked
at
will
also
tell
you
that
things
are
alert
or
muted,
so
that
you
are
aware
of
it
so
stuff.
E
B
E
So
basically,
the
rest
of
this
pad
is
all
about
how
to
deal
with
that,
because
each
Health
Alert
a
little
difference,
but
at
the
like
phase,
zero
would
just
be
to
match
the
description
just
to
get
the
rest
of
the
pieces
in
place.
But
the
real
challenge
is
around:
when
done
you
unmute
things
and
I
think
I
think
we're
just
going
to
need
a
bunch
of
different
modes,
because
each
will
Health
Alert
pays
a
little
differently
so,
for
example,
OSD
down
it'll
say
like
Oh
Steve's
down.
E
E
But
if
maybe,
if
those
two
O's
T's
resolved
and
then
new
OS
D
went
down,
then
maybe
you
do
so
Paul.
You
actually
want
to
look
at
like
the
set
of
which
O's
DS
it
is
and
as
long
as
that
shrinks,
then
it's
fine
and
if
the
sec
rose
and
it
unmutes
other
things
like
you,
know,
degraded
and
there's
a
percentage
there's
a
number
that
you
want
to
like
decrease
versus
increase.
Maybe
we
just
look
at
the
detail
items.
E
Each
dollar
has
a
bunch
of
details
that
actually
say
like
which
of
those
d
has
failed.
Well
as
long
as
those
like
as
long
as
the
detail
item
goes
away,
then
it's
fine,
but
if
a
new
one
appears
then
it
wouldn't
be
fine,
but
even
that
still
the
tricky,
because
there's
always
a
cap
on
those
there's
always
like
a
maximum
is
30.
E
I,
think
it
is
after
when
it
just
says
plus,
like
you
know,
700
other
rows
to
use
or
something
like
that
and
sort
of
bambbles
that
so
it's
a
little
bit
a
little
bit
tricky
but
I
think
initially
the
idea
would
be
to
just
have
a
couple
different
modes
and
then
probably
have
a
one,
that's
default,
and
then,
in
the
code
say
like
this
particular
Health,
Alert
use
this
other
mode.
They
can
sort
of
use
something
that's
appropriate.
You
might
need
to
go
back
and
adjust
exactly
have
what
the
details
strings.
E
E
That's
the
detail
to
make
that
a
little
bit
more
structured,
though
you
know,
have
like
a
some
scalar
value
that
might
be
used
for
some
health
alerts,
like
you
know,
0.7%,
degraded
or
something
or
maybe
have
a
set
of
integers
for
the
OSD
one.
Where,
like
you,
actually
have
the
set
of
ID's
that
are
dam
or
or
whatever
it
is
the
set
of
Fiji's.
Maybe
something
like
that.
E
Go
back
and
figure
out
what
exactly
what
structured
elements
might
be
appropriated
so
that
they
can
be
a
little
more
precise,
but
I
think
kind
of
that
depends
on
if
we
can,
if
we
can
cover
most
of
the
health
alerts
in
a
pretty
decent
way
with
something
pretty
simple.
It's
just
like
a
handful
of
different
modes
for
a
meeting.
Then
that
might
be
good
enough
to
that.
Could
you
encourage
way.
A
E
Those
values
that
number
is
current,
just
constantly
changing
and
presumably
slowly
ramping
down
to
zero
and
so
ideally
feel
to
mute
it
long
as
it
continues
to
decrease.
But
maybe
there
was
an
easier
way
to
do
that,
like
maybe
we
just
like
parse
a
floating-point
value
out
of
the
string
or
something
in
one
of
the
modes
and
then
use
that,
like
a
decreasing
rule
or
something
maybe
a
quick
hack
is
good
enough.
I.
A
Guess
you
also
talked
about
potentially
letting
you
specify
thresholds.
That
would
be
things
like
this
words.
A
percentage
said
that
English
between
health
war
and
health
care
or
health-
okay,
maybe
like
a
1%
degree
to
this,
is
okay
in
your
cluster.
But
20%
is
a
warning
and
50%
is
an
error.
I
think
mm-hmm.
E
Yeah
yeah
I
think
I
think
those
are
like
per
health
alert,
adding
additional
options
to
parameterize
what
the
behavior
should
be.
A
lot
of
the
alerts
already
have
one
or
more
settings
that
sort
of
control
when
to
learn
whether
to
alert
at
all
or
what
the
threshold
is
that
a
warning
state
is,
and
so
we
could,
you
know,
go
through
one
by
one
and
say
like:
are
there
ways
that
we'd
like
to
modify
the
behavior
of
this
particular
helpful
Earth?
D
F
E
E
H
B
B
E
E
B
E
E
B
D
E
D
E
H
E
D
H
A
E
H
E
B
E
B
B
E
F
B
B
B
E
E
E
E
Okay,
well,
let's,
let's
start
with
that
and
then
play
with
it
and
see
I
think
the
real
interesting
part
is
going
to
be
when,
during
this
phase,
where
you
look
at
each
health
code-
and
you
look
at
what
the
messages
look
like
and
see
how
many
modes
out
to
be
implemented-
hopefully
not
very
many.
No
people
provide
decent
behavior
but
find
out.
B
E
A
So,
in
terms
of
disks
being
too
busy,
we
have
the
suicide
timeouts
right
now,
which
ends
up
making
the
Westies
crash
and
can
kind
of
be
confusing
the
users
since
they
see
you,
know
Steve
crashing,
not
really
necessarily
understand.
Why
that's
one
thing
that
we
could
improve
they're,
just
making
that
have
some
clear,
clear
message
in
the
log.
E
That
is
a
2d,
because
I
wasn't
sure
how
to
do
it.
Health
names
or
brushes
and
I
think
it's
basically
because
I
didn't
know
how
to
mute
them.
So
if
you
do
crash
LS,
you
can
see
all
the
crashes
like
historically
forever.
So
the
question
is
like
you
to
Health
Alert.
If
there's
a
new
crash
and
if
so,
for
how
long
do
you
burn
for
before
your?
It
goes
away.
E
Yeah
I,
don't
know,
or
do
you
like
delete
it?
Do
you
like
have
an
act
like
I
act,
acknowledged
that
this
crashed
and
I've
looked
at
it
and
therefore
the
health
Gordon
goes
away
or
do
you
need
it?
I
wasn't
really
sure
what
the
what
the
right
metric
was.
I
didn't
delete
the
crashes
or
something
they
can,
but
you
kind
of
don't
want
to
delete
them
right.
You
just
want
to
like
you'd
like
to
have
a
record
of
them
all
and
they
want
to
make
sure
they
all
get
phoned
home
and
oh
I,
think.
B
A
E
Kind
of
I
mean,
if
you
want
to
tell
the
operator,
that's
something
bad
happened,
and
you
want
to
tell
the
developers
I
think
that
happened
and
they're
think
they're
disjoint
right.
Just
because
it's
been
phone
home
and
we'll
eventually
maybe
find
out
about
it
doesn't
mean
they
don't
want
to
tell
the
operator.
E
A
Yes,
we're
going
back
to
the
disks
being
too
busy
we're
actually
failing.
We
might
also
capture
some
other
metrics,
like
things
like
IQ
depth
at
various
levels,
about
mr.
level
or
at
the
OST
level,
and
make
sure
that
we
were
exposing
those
through
foreigners
or
potentially
before
those
as
metadata
to
the
manager
or
your
new
faces.
E
G
E
A
A
E
A
A
E
One
that
bothers
me,
I,
can't
think
of
anything
else
that
we
just
want
only
generate
raw
data
for,
but
the
thing
that
bothers
me
is
that
if
an
OSD
fails
because
it
gets
an
I/o
error
and
it
crashes-
that's
logged
in
the
crash
report,
but
nothing
ever
looks
at
that,
and
the
cluster
doesn't
know
that
it
failed
because
the
device
failed.
It
just
knows
that
the
OSD
is
doing.
Yes,
that's
a
good!
That's
in
case
we
were
kind
of
crashed,
a
twisty
and
don't
report
anything
any
evident.
E
E
Failed
somehow,
but
even
then
it's
like
what
is
balance
of
the
whole
thing
with
all
the
failure
protection
stuff.
That's
like
what
what
has
failed
mean.
Is
it
that
you
got
one
eye
or
once
or
is
it
that
you
got
too
many
of
them
or
because
what
with
blue
storm
most
of
the
time
when
you
get
an
I/o
err
on
read,
we
can
fix
it
like
will
use
the
replica
to
repair,
and
when
that
happens
like
we
don't
do
anything.
E
E
B
E
B
E
A
I
guess
right
now
we're
kind
of
just
asserting
and
it
could
be
easy
to
understand
if
we
didn't
just
do
this
an
effort
there,
but
had
a
message
in
the
logs
I
mean
it's
called
like
exit,
zero
or
sorry
exit
earlier
and
didn't
can
really
confusing
temper
of
all
the
vlogs
back-trace
information
like
that
was
time
out
and
then
understand
and
that
what
that
means,
or
even
getting
the
IO
error
and
not
understand
that
means
the
disk
is
filling
so.
E
G
E
E
E
A
Let's
play
lost,
but
four
discs
now
I'm,
so
calming
about
like
as
ICP.
You
talk
a
little
bit
of
already
about
this
trying
to
track
CPU
load
and
you
just
wait
for
padding
that
the
log
periodically
so
that
we
have
a
record
of
what's
going
on
if
the
user
doesn't
have
other
monitoring
setup
and
that
way,
if
we
do
like
see,
you
know,
Steve
suicide
time
matter,
have
a
some
other
issue.
E
E
E
A
A
A
E
G
E
E
A
Yeah,
so
beyond
the
basic
physical
resources
in
the
system,
I
just
wanted
to
add
some
more
specific
warnings
for
things
that
we
do
know
about.
Like
not
databases.
Thank
you.
Large
I
think
the
tricky
part
here
is
figuring
out
what
a
good
threshold
is
for
fast
or
it
might
have
been
like
20
gigabytes,
but
for
a
blue
store.
It
seems
to
be
in
the
hundreds
petabytes
range.
It's
gonna
vary
pretty
heavily
I.
Think,
based
on
the
workload
mm-hmm.
A
A
E
A
E
E
E
So
we
could
have
worn
it
like
0.5
I,
guess,
that's
device,
size,
I'm,
just
I'm,
just
nervous
about
saying,
like
100,
gigs
or
picking
a
number,
because
this
can
very
similar
time
for
food.
It's
kind
of
vary
based
on
device
size
and
the
device
speed,
and
they
do
you
have
any
other
things,
but
at
least
the
ratio.
That's
like
I
know
an
issue
doesn't
matter,
you
don't
be
able
to.
A
Okay,
they
it's
not
so
useful
in.
If
we
can't
figure
it
I
got
a
reasonably
useful
value.
That's
like
what's
most
workloads,
yeah
I
think
this.
A
So
last,
but
here
was
about
time
trying
to
kind
of
analyze
with
what's
happening
in
the
cluster
or
the
human.
Yet
do
the
analysis
based
on
a
bit
more
data,
so
Michael
KITT
has
written
some
scripts
for
parsing
the
stuff
that
log
and
trying,
and
what
kind
of
bring
this
into
a
kind
of
spreadsheet
format.
We
can
look
at
things
over
time
or
in
provost
e
basis,
try
to
see
if
their
psychic
events
that
are
correlated
with
still
requests
to
the
log.
A
E
A
H
H
So
earlier
it
used
to
have
like
lot
of
information
from
the
I
of
point
of
view,
recovery,
I
of
point
of
view,
deep,
scrubbing
states,
whatever
slow
requests
are
coming,
how
much
time
they
are
taking
how
many
number
of
OSD,
how
much
time
deep
scrub
is
checking
like
these
kind
of
staffs.
That's
found
the
Questor
logs
and
I
think
he
might
go.
Kid
has
written
a
script
which
takes
some
interval
and
it
just
drops
it
to
CSV
file
and
rest
I.
H
A
H
Like
report
like,
if
the
deep
scrubbed
is
running
so
like
reporting,
what
is
an
average
deep
scrub
timing
for
each
each
interval
like
if
you're
taking
one
minute
two
minutes?
If
there
is
some
impact
from
the
like,
because
of
deepest
curves,
there's
an
impact
in
a
slow
rate
request
because
of
deep
something
like
that.
H
H
And
this
is
all
to
check
if
cluster
is
reporting
slow
request.
What
is
the
reason
behind
it?
Is
it
deep
scrub?
Is
it
the
disturb
busy
because
of
some
other,
it
is
split
like
earlier?
It
used
to
be
file
story
split
now,
with
blue
stores.
We
can
think
of
like
PBS
is
doing
some
spillover
or
or
maybe
some
compaction
is
running,
which
is
taking
lot
of
time,
which
is
causing
a
CPU
busy
or
something.
A
H
I
think
what
I
understood
from
David
disclination
didn't
we
combine
this
Network
thing,
and
this
I
think
there
will
be
lot
wave,
a
better
information.
What
we
will
have
for
troubleshooting
as
compared
to
now
I
mean
networking
and
the
more
information
from
the
slow
request,
point
of
view.
I
think
will
give
us
a
lot
of
data.
What
about
sorting.
D
H
E
A
A
E
A
E
E
I've
Leonard
is
is
if,
if
there
should
be
a
way
to
like
capture
the
log
for
a
specific
PG,
whenever
there
is
like
a
an
annoying
bug
and
something
goes
wrong
and
those
he's
crashing
or
a
good
song
during
pairing
or
whatever
like
we
have
to
turn
on
the
whole
log
for
the
host
e.
But
it's
just
one
PG
that
we're
looking
at
yeah.
E
It's
like
the
PG
is
stuck
in
a
particular
state
and
we
can't
figure
out
why
from
the
query,
and
so
we
have
to
go
turn
on
log.
So
we
get
that
one
PG
and
make
a
repair,
and
they
see
wonder
if
we,
if
it
makes
sense
to
be
able
to
set
a
logging
level
on
a
per
PG
basis,
I
kind
of
like
the
way
you
can
do
it
for
OSD,
but
for
PG.
E
You
can
figure
out
in
that
way
that
do
one
thing
or,
alternatively,
if
all
of
the,
like
all
the
log
messages
out
of
the
peering
state
machine
like
clearing
state,
basically
should
be
retained
and
memory
have
a
different,
have
a
high
level
of
memory
retention,
so
that,
if
did
you
hit
a
crash
like
you?
Have
it
like
it's
up
to
like
debug
level
10,
at
least
that.
A
A
E
G
G
E
E
Think
about
because
when
when
people
are
like,
when
QA
is
like,
we
want
a
tool
that
was
gonna
go
gather
the
logs
to
crash
like
what
you
really
want
is
to
reproduce
it
with
debug
OC
20
and
then
go
grep
out
that
PG
out
of,
like
whatever
the
appropriate
OSD,
is
like.
If
there
was
a
way
to
automate
that
maybe
it'd
be
really
nice.
I.
E
E
E
C
E
C
E
E
H
E
E
That's
actually
useful
and
there's
like
there's
like
the
crashed
part
of
it,
parts
of
it
that
you
just
have
to
like
extract
the
actual
dump,
make
them
separate
records
and
put
him
in
a
one
particular
table
or
database
or
whatever,
and
then
the
telemetry
is
similar
where
you
have
to
like.
You
have
to
associate
them
with
a
particular
cluster,
and
you
really
won't
care
about
the
most
recent
report
for
a
single
cluster
and
so
on.
E
E
Archiving
them
yeah
they
just
can't
like
bury
them
yet
and
incidentally,
there's
a
I
had
an
RFC
request,
love
to
hear
comments
on
that's,
basically
adding
an
ax
concept
of
a
channel
to
the
filmer
tree,
though
there
would
be
the
basic
channel
which
it's
like
the
basic
cluster
stats.
What
version
you're
running,
how
many
OS
DS
you
have
how
much
data
you're
storing
that
sort
of
thing
there'd
be
the
crash
channel,
which
is
all
the
crash
dumps
and
then
eventually
we'd
add
the
device
channel,
which
is
all
the
device,
health,
metric,
telemetry
and
the
theory.
E
E
Yeah,
it's
it's
a
little
bit
weird,
because
the
the
telemetry
and
crashes
are
gonna
go
into
one
database
just
like
for
the
stuff
clusters,
the
Vice
Health
metrics.
The
intention
is
for
that
to
go
into
a
different
data
set,
that's
not
necessarily
just
stuff.
It
might
actually
be
phoning
home
to
a
different
location.
That
I
was
trying
to
like
just
keep
it
simple,
so
they're
just
channels-
and
you
turn
them
on
and
off
sort
of
independent
of
what
the
mechanism
is,
that
they
actually
get
I.
E
E
E
E
B
A
D
Maybe
when
they
do
the
upgrade
or
using
the
dashboard,
they'll
have
to
first
fill
that
out
whether
you
want
it
done
this
module
on
after
the
upgrade,
and
if
they
say
no,
you
capture
the
reason
you
should
have
options
like
ABC.
Why
did
you
turn
it
off
exactly
so
that
they
make
a
choice
during
upgrade
and
that's
how
most
pieces
of
software
these
days
are
made
so
yeah.
E
Okay,
it's
doctor
ones
about
that,
all
right
anything
else.
We
should
discuss
we're
ready
for
bed,
pepper.