►
From YouTube: Technical Oversight Committee 2021/01/08
Description
Istio's Technical Oversight Committee for January 8th, 2021.
Topics:
- Which versions of Kubernetes to support in Istio 1.9
- Current Test Flake status
- Update on docs working group
A
A
A
A
So,
if
you
haven't
already
any
features
you
think
needs
to
work
upon
in
this
year,
please
add
them
in
the
in
the
draft
of
istio
2021,
together
with
the
priority,
the
stages,
the
user
benefit
and
the
contributors.
So
that
way,
once
we
are
ready
in
two
to
three
weeks,
we
can
have
it
review
with
the
toc
folks.
C
And
see
what
we
had
all
right,
all
right,
we
have
the
I
guess
jacob
or
sam.
Do
you
want
to
talk
about
the
upgrade
work
group
charter.
E
Yeah,
so
we
got
approval
from
the
environments
group
as
well
as
testing
release.
Josh
wanted
to
talk
to
neraj
kind
of
a
little
bit
more
about
that.
So
hopefully
they've
discussed
that.
G
F
Yeah,
okay,
well,
we'll
sneak
up
all
right,
and
I
think
part
of
it
also
is
I
want
to
know.
Is
this
a
long-lived
working
group,
or
is
this
more
like?
You
know
a
faction
that
that
lives
long
enough
for
for
upgrades
to
be
better
and
then
it
just
merges
back
into
one
of
the
existing
working
groups.
E
E
That
is
the
expectation,
but
but
obviously,
if
there
are
other
concerns,
then
or
if
people
think
that
it
is
a
worthwhile
endeavor,
then
I'm
imagining.
We
could
live
long.
Yeah.
F
Frankly,
my
other
concern
is
just
whether
we
have
enough
people
working
in
this
working
group.
So
that's
what
I
wanted
to
talk
to
narash
about.
H
Yeah,
I
guess
one
question
I
have
is
for
the
one
nine
book
item.
It
would
be
great
if
we
can
link
up
to
the
issues
or
maybe
have
a
umbrella
work
item.
So,
for
instance,
you
know
eric
and
I
have
been
having
discussions
on
how
we
can
help
as
part
of
one
night
effort.
So
we
try
to
link
issues
that
we
are
going
to
eric
is
going
to
help
us
on
from
the
scenarios
that
matters
most
to
us.
So
I
I
noticed
there's
a
lot
of
issue.
H
E
Okay,
yeah:
that
is
an
action
item
that
I
can
bring
up
and
and
try
to
have
something
a
little
bit
more
formal.
I
A
E
Jacob
we'll
talk
about
supporting
versions
yeah,
so
this
hasn't
been
raised.
Unfortunately,
I
didn't
talk
about
this
in
environments,
but
typically
we
bump
up
the
kubernetes
supported
versions,
so
1
8,
supported
116
through
119
120,
was
released
in
december.
So
I'm
just
looking
for
verification
that
that
we
are
intending
to
support
120
as
well.
As
you
know,
117,
you
know
through
it
for
1.9.
C
Topic
right
as
a
whole,
like
we've,
had
questions
about
this
in
public
forums.
Mitch
has
given
a
presentation
about
this,
so
I
guess
there's
a
question.
E
C
Anyone
else
feel
free
to
correct
me
if
I
start
to
reading
correct
factually,
which
would
mean
that
we
would,
if
we
stayed
with
n
minus
two,
we
would
fall.
One
of
supported,
kubernetes
release
would
fall
out
of
the
window,
which
is
probably
not
a
good
idea
for
the
project.
C
C
C
I
C
Right,
there's
three
questions
right
one:
when
do
we
start
to
qualify
120,
do
we
drop
116
right?
We
can
treat
those
as
independent
questions
and
then
what
you
know.
What
release
would
we
consider
to
be
this
sweet
spot
right,
which
probably
is
not
120
right?
So
if
we
were
going
to
do
manual,
qualification
or
other
things
that
were
not
automated
or
repeatable,
which
release
would
we
recommend
to
be
used
for
manual
qualification
of
anything
or
even
like
day-to-day
development
work.
C
C
I
L
L
L
There
is
like
this
minor
issue
where
their
pod
termination
they
have
some
regression.
So
it
takes
super
long
for
pods
to
terminate
and
our
test
time
out
so
we're
working
on
getting
that
fixed
before
we
adopt
it
in
our
ci.
But
fundamentally,
there's
like
it.
I
It's
broken
in
kubernetes,
so
so
john,
I
know
in
120
there
was
a
lot
of
work
done
in
the
kubernetes
community
for
dual
support
for
ipv4
and
ipv6.
Does
that
impact
make
our
life
better
easier,
easier
harder?
Do
you
know.
L
C
C
Things
which
I
don't
think
are
that
impactful
for
us
either
or
that
they
don't
really
represent
a
change.
So
it
sounds
like
getting
the
automation.
Working
is
just
the
major
thing
here.
J
C
It's
a
good
question.
I
don't
necessarily
know
if
we
have
to
answer
that
now
or
even
so,
we
could
choose
to
refine
that
right.
So
there
was
the.
I
brought
up
the
point
of
sweet
spot
testing
right
where,
if
yes
clearly,
I
think
we
would
like
the
long-running
tests
to
run
against
every
known,
supported
release.
I
I
mean
it
looks
like
when
we
had
this
when
we
were
discussing
the
kubernetes
version
matrix
last
time.
We
weren't
sure,
if
increasing
this
support
is
necessarily
going
to
help
our
cause,
or
maybe
I'm
misplacing,
my
facts
here.
Can
somebody
remind
me:
will
this
help
if
you
make
it
for
going
forward
and
then
maybe
five
going
forward.
K
N
L
If,
if
there's
nothing
going
on
it's,
it's
super
easy,
it
is
literally
zero
time
I
mean
other
than
you
know
the
test
runs,
but
that
doesn't
really
impact
anything.
But
then,
if
we
decide
that
we
need
some
feature,
that's
only
supported
in
117
plus
at
some
point.
Then
it
could
become.
You
know
extremely
hard.
F
How,
if
we
say,
m
minus
four,
how
hard
is
it
to
change
that
later,
if
we
run
into
something
really
difficult.
L
L
Maybe
that's
not
reliable
enough
for
users.
I
don't
know.
C
I'm
not
sure
if
that
really
helps
right.
That's.
F
I
F
C
C
That's
the
reason
why
kubernetes
went
to
n
minus
three:
it's
like
I
can't
afford
to
do
more
than
one
upgrade
a
year
right
that
that
problem
or
two
upgrades
a
year,
and
so
with
that
policy
in
place
like
I'm
always
for
three
months,
you're
going
to
be
out
of
support
with
istio,
right
or
or
something
of
that
built
right.
That's
why
kubernetes
extended
their
window.
G
So
that
there
is
also
the
other
side
of
upgrade,
which
is
someone
trying
to
kind
of
leapfrog,
several
store
versions
and
now
they're
forced
to
upgrade
their
kubernetes
version,
which
they
hadn't
planned.
G
C
C
I
think
we'd
probably
all
agree.
The
kubernetes
upgrade
and
or
support
window
is
less
problematic
than
the
steel
upgrade
and
support
window.
C
C
C
And
then
there's
one
of
the
vendors
right.
So
if.
K
Yeah,
I
I
think
I
agree
with
john
that
it
makes
sense
to
make
these
decisions
in
an
ad
hoc
manner
that
we
have
a
minimum
bar
that
we
support
for
kubernetes
and
at
each
release
we
evaluate
how
much
more
than
that
minimum
bar
we
can
support.
I
think
that's
going
to
be
great
for
our
users,
and
it
also
keeps
us
from
tying
the
project
to
a
policy
that
would
might
prevent
us
from
moving
forward.
G
C
Yeah,
okay,
so
I
think
what
we're
saying
is
we
will
bump,
but
we
should
evaluate
and
and
if
we
can
carry
116-
or
at
least
we
should
have
a
positive
indication
about
why
we
shouldn't
carry
it.
I
Do
we
do
we
know
if
the
current
vendors
or
the
hosted
pro
kubuntu
providers
what's
their?
Where
are
they
at
are
most
of
them
at
116.?.
O
H
I
think
they
fall
out
of
support.
We
we
auto
upgrade
for
fixes
like
from
116
15
to
116,
16
or
17,
but
the
major
ones
they
have
to
click
on
the
button
to
update
like
major
miner.
C
All
right,
so
we
need
an
action
item
to
go,
follow
up
and
see
what
the
the
vendor
states
are
or
are
likely
to
be.
By
the
time
we
hit
the
release.
H
F
F
C
F
Actually,
sorry,
are
we
n
minus
three
right
now,
yeah,
okay
sounds
good
in
terms
of
having
a
cap.
C
Well,
a
reasonable
way
to
define
the
cap
is
the
minimum
kubernetes
release,
supported
by
one
of
the
major
four
or
five
vendors
right,
so
iks
openshift,
aks,
eks
gke
and
use
that
as
the
rubric
right
or
maybe
some
other
thing,
but
in
general
like
that's
what
we
would
look
at
by
the
time
of
the
release
right.
So
we
can
make
this
decision
late
right,
because
this
is
right.
We're
not
going
to
use
this
for
manual
testing
right,
we're
only
going
to
use
it
for
any
automation.
That's.
J
C
J
C
J
Is
was
end
of
life?
I
think
in
september.
C
Okay,
yeah,
then,
then,
we're
we're
we're
still
n
minus
two.
Until
1.9
becomes
the
tail
of
n
minus
three
one:
nine
sorry
119,
and
then
we
can
choose
to
be
one
and
or
two
more
than
that,
based
on
what
the
vendors
have.
C
B
I
I
C
O
We
I
don't
know
if
this
was
a
comment
from
from
jones,
that
we
test
in
115.
Even
can
we
take
a
more
pragmatic
approach
and
when
it's
becoming
expensive
to
support
a
release,
we
drop
it
and
we
just
leave
it
there
I
mean.
Maybe
we
can
say
that
officially
supported
it's
n
minus
three,
but
hey,
it
was
2015..
L
C
K
O
O
C
E
C
C
F
Okay,
so
what's
the
summary,
it's
it's
n
minus
three
at
a
minimum!
Sorry
supported
n
minus
three
at
a
minimum.
We
might
go
up
to
n
minus
four
in
a
case
by
case
basis
and
we
may
qualify
beyond
that,
but
we
wouldn't
support
it.
C
Yeah,
okay,
and
when,
when
kubernetes
goes
to
n
minus
three,
yes,
then
we'll
be
n
minus
three,
but
that
policy
will
stay
the
same.
I
C
J
C
Necessarily
want
to
start
trying
to
like
describe
the
rules,
the
policy
I
think
it's
simpler
to
just
say
what
the
supported
releases
are
and
say
if
you
care
about
these
kubernetes,
if
you
care
about
over
uber
90s
releases.
Well,
here's
the
ones
we've
tested
against
these
ones,
but
these
are
now
no
longer
supported
right.
We'll
just
find
some
way
to
describe
the
meaning
without
having
to
describe
the
rules
that
we
use
to
generate
it.
I
D
So
we
actually
do
document
already
are
in
release
cadence
we
document
how
often
release
and
how
often
they
are
supported.
It
is
woefully
out
of
date,
so
we
should
probably
update
it.
We
can
either
get
rid
of
it
and
just
say:
look
on
our
supported
releases
page
for
the
support
of
things,
but
we
might
want
to
actually
reword
it
to
be
up
to
date.
C
Yeah
I
mean
right
now
on
the
release.
Cadence
page
even
has
an
lts
section,
yep,
okay,
that
is
what
flyover
did
yeah
and
jacob.
Can
you
guys
take
on
you're
getting
really
rid
of
the
release,
cadence
page
and
just
rolling
it
into
the
supportive
releases
page
sure
I
can
do
that?
Is
this
approved
by
the
toc?
I
think
so
I
think
so.
Yeah.
C
Okay,
thanks
for
bringing
that
up
test
late,
dashboard.
C
K
K
So
there
were
questions
about
test
flakes
and
the
dashboard
was
partially
functional
for
several
months
there.
It's
back
up
to
speed.
It's
automatically
updating
you
can
see
flakes
as
of
january
5th
right
now.
It
does
update
about
once
a
day.
So
our
flake
rate,
if
we
remove
our
remove
the
date
filter
here
and
look
back
historically
for
all
time.
This
chart
shows
the
ratio
of
prs
which
do
not
experience
a
flake
to
those
that
do
experience
a
flake
and
in
december,
21
of
all
prs
experienced
at
least
one
test
flake.
K
The
goal
that
we
had
set
as
a
project
way
back
in
the
code
mob
days
was
five
percent.
We've
never
actually
achieved
that
goal,
but
we
did
get
pretty
close
in
april.
We
were
at
just
seven
and
three
quarters
percent,
so
we
do
see
a
decent
regression
and
flakiness,
which
is
borne
out
by
I
think
john
howard's
experience
and
others
in
the
project.
K
The
two
themes,
I'll
call
out
is,
we
see
a
lot
of
the
multi-cluster
suites
are
creating
a
decent
amount
of
flakiness
right
now.
This
is
not
a
multi-cluster
suite
here.
This
is
a
multi-cluster
suite.
This
is
also
a
multi-cluster
suite.
The
other
theme
is
that
telemetry
here's,
the
telemetry
multi-cluster
suite
here's
the
telemetry
non-multi-cluster
suite,
so
those
are
kind
of
the
two
biggest
areas
of
change
that
I
see
over
the
last
few
months
in
terms
of
increase
in
flakiness,
for
anyone
who
is
interested
in
identifying
flakes
and
finding
them
and
root
causing
them.
K
This
section
up
here
gives
you
a
list
of
recent
flakes
the
20
most
recent
flakes
and
if
you
filter
by
test
name,
you
can
get
to
the
most
recent
flakes
in
your
test
suite.
So
it's
a
good
launching
point
for
troubleshooting
and
identifying
root
causes
and
I
hope
to
see
us
get
back
down
in
the
in
the
seven
to
five
range
soon.
K
A
So
one
more
question
with
that:
josh
mitch.
Thank
you
for
picking
it
up.
I
know
it
was
down
for
a
while,
and
so
thank
you
for
working
on
it
same
thing
with
josh
asked
is
the
maintenance.
I
know
there
was
a
plans
of
moving
that
to
tnr.
A
K
So
the
idea
was
that
I
needed
to
get
it
back
up
to
functioning
before
I
could
hand
it
off,
we've
still
not
really
identified
a
future
owner
for
it.
I
have
do
have
one
volunteer
from
the
community
who
heard
about
it
in
the
test
and
release
group
this
week
and
he's
coming
up
to
speed
on
operations.
K
Ideally
I'd
like
to
have
about
three
of
us
who
are
able
to
help,
contribute
and
maintain
this
mostly
for
upkeep,
although
if
there's
interest
in
adding
features,
there's
certainly
a
good
number
of
features
that
could
be
added
to
make
this
more
useful
over
time.
So
I
am
looking
to
tnr
for
a
few
more
volunteers.
C
Getting
back
to
josh's
question
about
how
do
we
get
this,
you
know
into
developers
faces
as
part
of
their
normal
flow?
Is
there.
C
C
K
This
system,
as
it
is
implemented
today,
is
completely
disconnected
from
prow.
It
reads:
prow's,
output
on
a
periodic
basis.
Currently
it's
at
once
every
24
hours,
the
engprod
team
had
committed
to
integrating
this
with
prow
and
then
our
end
fraud
representative
left
the
team
and
I
believe
the
commitment
left
with
him.
G
So
I
think
I
think,
though,
before
that
right
we
like
the
so
the
person
that
experiences
the
flake
is
not
actually
the
person
that
or
the
people
that
are,
that
should
be
fixing
the
flakes
right.
So
for,
for
example,
I
guess
unfortunately
here
telemetry
has
has
lots
of
flakes,
so
it's
so
the
so
I
think
it
it's
really
up
to
the
the
people
that
maintain
these
subsystems
to
know
about
this
dashboard
and
and
go
there
not
so
much
the
people
experiencing
the
flakes.
K
H
H
If
I
don't
own
those
tests,
I
don't
even
want
to
look
at
them
and
then,
if
this
one
is
a
valid
failure,
you
know
I
could
actually
look
into
the
logs
to
see
how
my
pr
could
introduce
that,
because
today
I
don't
have
that
knowledge.
What
I
do
is
I
retest
and
then
see
the
same
tests
come
back
right,
because
why
would
I
spend
time
to
look
at
the
logs
when
I
can
just
spend
one
second
to
have
the
ci
system
re-test
everything.
L
And
I
think
part
of
the
problem
is
that
you
don't
know
if
the
test
is
flaky
until
you
rerun
it
and
it
passes.
Otherwise
it
could
just
be
your
pr
is
bad
and
making
the
chest
fail
and
yeah
what
you
are
doing
also
by
retesting
it.
I'm
not
saying
you
shouldn't
do
this
and
I
think
everyone,
including
myself,
does
but
that's
part
of
the
reason
why
we
have
flaky
tests,
because
someone
submits
a
pr
and
they're
introducing
a
new
flaky
aspect
and
it
fails
and
they
re-test
it
and
then
passes
and
emerges.
L
So
it's
it's
great.
It's
it's
a
complicated.
H
L
L
When
you
try
it,
we
should
open
an
issue
that
says
foo
bar
test
is
flaky,
try
and
find
out
who
owns
that
test
and
then,
when
you
see
a
failure
on
your
pr
that
you
don't
expect
is
yours,
you
should
search
the
issues
and
if
that
one
has
an
open
issue
that
this
is
a
known
flaky
test,
it's
probably
fine
to
retest
it.
If
not,
then
you
should
look
at
either.
Did
my
pr
cause
it
or
maybe
I
should
open
an
issue
that
this
is
flaky,
so
we
can
resolve
this
now.
H
K
L
That
previously,
when
we
had
a
bunch
of
flakes,
it
was
because
we
had
bad
tests
or
bad
infrastructure.
Recently,
a
lot
of
these
test
failures
are
from
real
bugs
in
the
code.
So
all
of
these
unit
test
failures,
I
think
they're
all
related
to
a
bug,
actually
a
collection
of
bugs
we
just
fixed
and
the
multi-cluster
ones.
Most
of
them
are
a
bug
and
envoy,
so
this
is
actually
in
some
ways.
K
I
do
also
want
to
call
out
that
often
times
the
root
cause
of
a
given
set
of
flakiness
is
not
necessarily
belong
to
the
team
that
owns
the
suite.
K
So
I
know
that
doug
looked
a
little
bit
at
the
integ
or
in
the
telemetry
failures,
and
he
seemed
to
think
that
there
was
some
sort
of
underlying
infrastructure
problem
associated
with
them.
So
while
we
do
rely
on
the
working
groups
to
at
least
begin
troubleshooting
and
identifying
a
root
cause,
responsibility
will
probably
pass
off
from
one
work
group
to
another.
C
H
K
Nice
to
tell
us
my
guess
on
that
is
that
it's
simply
no
noise
in
the
signal.
K
One
thing
that
we
do
I
do
remember
experiencing
back
in
2019
when
we
had
a
very
big
problem
with
flakes,
is
that
running
more
tests
concurrently
tends
to
cause
a
higher
rate
of
flakiness.
I
don't
know
if
that's
the
case
today.
I
can't
prove
that
today,
but,
given
that
this
dashboard
is
updated
through
january
5th
and
january
5th
was
the
first
workday
of
the
new
year
for
googlers
and
the
second
for,
I
think
most
other
people.
I
think
that's
what
we're
seeing
we
should
see
in
another
week.
L
That
may
have
a
a
role
in
it,
but
we
made
a
bunch
of
changes
to
the
tests.
We
fixed
I'm
very
confident
that
the
majority
of
the
unit
test
flakes
are
fixed,
I'm
close
to
having
it
set
up
where
it's
just
running
the
unit
test
in
a
loop
overnight
and
so
far
it's
going
pretty
well
for
for
the
test.
Now,
steven
and
doug
also
made
some
big
fixes,
like
the
multi-cluster
one.
We
used
to
run
five
clusters,
and
now
we
run
three
so
there's
three
fifths
as
many
opportunities
to
flake,
for
example.
K
It
is,
I
would
love
to
be
able
to
get
something
that
shows
a
day-by-day
flakiness
rate
for
people
like
john
who
are
actively
working
on
flakiness
of
various
suites,
to
be
able
to
see
their
results
in
more
real
time.
The
month-to-month
granularity
is
not
ideal,
but
again
we
don't
have
any
hours
allocated
to
that
in
the
coming
months.
L
Okay,
so
we
do,
if
you
go
on
the
toc,
I
posted
a
link.
There
is
a
spreadsheet
that
I
periodically
update,
which
is
looking
at
something
that's
a
different
metric
but
similar,
which
is
the
number
of
post-submit
jobs
that
succeeded
for
pr
and
so
that
we
have,
you
know,
per
merge,
pull
request
granularity,
but
it's
measuring
something
different,
because
in
your
case,
this
dashboard
only
shows
a
pull
request.
That's
not
merged
that
fails
and
then
passes,
whereas
the
other
case
is
just
running
the
test
once
and
if
it
fails,
it's
marked
as
failed.
I
I
I
think
the
other
thing
is
just
some
education
around
how
we
can
file
bugs
and
not
ignore
until
and
just
rerun
the
test
until
your
test
passes
and
pr
as
much
right
is
there
anything
else.
I
think
you
do
now.
F
I
mean
I'm
thinking
out
loud
her
like
what
about
I
mean.
There's.
Also.
This
idea
I
think
louis
was
bringing
up
earlier.
I've
tried
to
incorporate
into
you
know,
pull
request
reviews,
so
that
would
be
great
great
avenue
to
explore.
I
can
imagine
doing
a
kind
of
a
you
know.
Sometimes
I
think
github
does
this,
but
they're
tools
that
try
to
yeah
so
when
github
tries
to
automatically
suggest
reviewers
for
for
pr
they
look.
P
I
P
C
All
kinds
of
things
google
has
all
kinds
of
properties
internally
for
the
moment
right
now,
I
would
rather,
I
think,
rather
than
diluting
the
effort,
just
focus
on,
let's
just
deflate
as
aggressively
as
we
can
sure.
So.
The
other
stuff
is
quite
expensive
to
implement
and
yeah
it.
If
we
can
stay
under
a
very
low
percentage
of
plates,
but
then
it
probably
has
almost
no
value.
K
C
L
E
I
L
C
C
C
C
F
C
F
I
I
C
Right,
I
I
understand
this
is
not
going
to
happen
overnight
right
and
I
understand
that
there's
a
middle
ground
between
these
two
things
right,
but
we
want
to
get
to
the
point
that
the
first
like
we
can
obviously
keep
retest.
But
we
want
to
get
to
the
point.
If
you
see
a
flight
right,
we
want
the
default
behavior
to
be
okay.
I'm
gonna
go
and
harass
the
person
who
owns
the
test
right.
C
C
Let's
let
the
working
group
lead
process
kick
in
and
let's
let
them
have
an
objective
right.
That's
the
right
form
to
take
ownership
ownership,
as
as
john
mentioned,
is
the
thing
we
need
to
work
on
first
and
and
then
yeah.
Maybe
once
we've
done
that,
then
by
all
means
we
have
to
start
to
have
all
kinds
of
behavioral.
K
So
I
do
have
one
question
about
the
objective
itself:
when
we
did
the
code
move
thing
back
in
2019,
we
set
an
objective
of
five
percent
or
fewer
of
prs
experience,
a
test
click
that
sounds
really
great,
but
we've
never
actually
achieved
that
in
the
history
of
the
project
like
since
we
started
using
prow
to
run
tests,
which
is
as
far
back
as
this
data
runs,
we've
never
had
that
that
high
or
low
level
of
flakiness
rather
do.
K
We
want
to
adjust
that
goal
and
have
it
be
the
advantage
of
adjusting
it,
I
think,
is
we
can
say
hey
if
we're
above
pick
an
arbitrary
number.
So
let's
say
10
percent
of
prs
are
experienced
on
a
flake.
Maybe
we
need
to
double
down
our
efforts
on
minimizing
flakiness.
Maybe
we
need
a
little
bit
more
of
a
hey
work.
Group
leads
no.
We
should
de-prioritize
some
other
work
so
that
we
can
get
these
flakes
fixed,
whereas
if
we're
under
ten
percent,
maybe
that's
acceptable.
C
Better,
so
having
a
goal
for
199
to
be
under
five
percent
of
test
fights-
and
you
know
using
the
work
leads
meeting-
is
the
vehicle
to
hold
ourselves
accountable
for
that
and
raising
that
to
the
toc?
If
there's
an
issue
seems
perfectly
reasonable
right
now
that
we
have
visibility
right,
because
we
didn't
have
good
john,
had
visibility
by
other.
C
I
Okay
looks
like
that's
doug's.
This
year's
objective
he's
already
taken
care
of
ten
of
them.
C
C
Talk,
I
agree.
There
are
issues
with
distributed
ownership
right
I
mean
that's,
that's
why
we
want
to
bring
them
into
these
forums
right
where
the
leads
come
together
or
the
toc
right,
so
that
we
can
wrangle
those
things.
If
there's
a
thorny,
flaky
issue
that
requires
multiple
owners,
then
that's
the
right
forum
and
we
should
decide
that
and
this
tool
provides
the
right
kind
of
insight
for
that
right.
N
N
I
mean
it's
not
quicker,
necessarily
to
always
ask
me
to
fix
a
test
right,
oh
yeah.
So
that's
that's.
C
C
K
To
use
like
this
system
does
not
have
any
false
positives,
it
does
have
false
negatives,
so
we
could
write
a
system
that
says
hey
if
if
we
observed
a
flake
in
your
tests
that
we've
never
observed
before-
or
maybe
we
just
haven't
observed
in
the
last
10
days-
maybe
we
need
to
block
your
pr
until
that
flakiness
is
resolved.
L
L
L
It's
it's
not.
C
Well,
the
the
root
closing
thing
was
just
looking
at
pr
history
and
test
results.
I'm
just
doing
correlations.
I
C
Like
so,
I
suspect
john's
thing
about
sorry
go
ahead.
Let
you
finish
right.
I
mean
one
thing
at
google
and
like
that
we
do.
Is
you
know
you?
You
can
indicate
that
a
test
is
sensitive
or
potentially
flaky,
because
it's
doing
something
big
and
nebulous,
and
so
the
test
infrastructure
just
runs
it
a
bunch.
C
F
Five
john's
joke,
which
I
took
as
a
real
suggestion
and
loved.
You
know
this
just
remove
retest.
I
wonder
if
a
variant
of
that
might
actually
work,
which
is
you
can
only
retest
so
many
times.
L
L
Of
it's
very
hard,
though,
like
these
a
lot
of
the
multi-cluster
flakes
are
due
to
like
major
issues
with
multi-cluster
that
one
person
can't
just
reasonably
fix,
and
it
would
also
be
turned
off
all
the
multi-cluster
tests
entirely.
So
we're
just
kind
of
stuck
in
a
medium-term
crappy
state
where
the
multi-cluster
tests
are
kind
of
flaky.
Okay,.
C
I
C
Yep,
we'll
we'll
save
the
the
retest
coinbase
market
for
later.