►
From YouTube: Kubernetes SIG Testing - 2020-12-01
Description
A
Hey
everybody
today
is
tuesday
december
1st.
This
is
the
kubernetes
sig
testing
bi-weekly
meeting.
I
am
your
host
aaron
of
sig
beard.
This
meeting
is
adhering.
We
adhere
to
the
kubernetes
code
of
conduct
in
this
meeting,
which
basically
means
we're
going
to
be
our
very
best
selves
and
be
kind
to
each
other.
A
A
A
B
B
It's
it's
just
a
thought.
I
think
the
two
of
us
had,
but
I
don't
know
if
we've
been
in
an
action,
live
action
item
for
it
or
an
issue
for
us
or
if
we
want
to
just
park
this
for
now
and
just
I
might
try
and
dig.
A
It
out
sure
I
don't
have
it
handy
right
now,
but
yeah.
This
is
something
we're
pretty
interested
in.
I
think
I
feel
like
there's
a
lot
of
stuff
in
the
testing
for
repo,
that
it
would
be
helpful
to
try
and
separate
some
things
out.
So
we've
talked
about
the
idea
of,
for
example,
trying
to
move
proud
into
its
own
repo,
or
the
other
thought
was
moving
config
files
into
their
own
repo.
A
Some
of
those
things,
I
think,
are
kind
of
tangled
up
with
the
way
that
prow
auto
deploys
itself,
but
I'm
just
trying
to
describe.
I
think,
like
the
challenge
that
the
challenge
is
for,
why,
like
it,
it
hasn't
been
done
already
not
to
say
that,
like,
I
think
anybody
here
if
they
were
willing
to
could
draft
a
plan
on
how
to
do
that.
Migration
and
we'd
be
supportive
of
moving
forward
with
that.
A
Okay,
so
I'll
put
something
in
the
notes
just
to
see.
If
I
can
dig
up
that
issue
for
you
to
take
a
look
at,
and
maybe
that
about
it
in
terms
of
plans
for
next
year,
the
next
time,
okay
next
item
was
looking
for
advice
or
thoughts
or
comments
on
modifying
the
e2e
test.
Binary
is
wilson
here.
C
Yes,
present:
hey
so
hi
folks,
I'm
wilson,
I'm
looking
to
add
a
flag
to
the
e2s
e2e
dot
test
binary,
which
would
improve
the
user
experience
in
overriding
registries
for
e2e
tests.
C
The
idea
is
that
today
we
have
environment,
variable,
cube,
test
repo
list
to
override
image
registries
for
people
testing
or
running
conformance
tests
in
air
gap,
environment
and
custom
registries,
and
the
way
that
we
can
know
or
build
the
config
file
for
that
is
through
looking
at
the
code
and
I'd
like
to
propose
that
there
is
a
way
for
us
to
override
it
and
provide
a
valid
configuration
without
looking
at
the
code.
C
D
C
Oh
the
case
gcrio,
yes,
sir
yeah,
so
I
think
there
is
an
effort
to
effort
or
a
proposal
to
move.
C
The
registries
that's
being
used
for
for
e2e
testing
to
be
consolidated
in
kids,
gcr
io.
I
linked
the
issue
in
my
comment,
which
I
think
I
can
paste
on
the
chat
as
well.
So.
C
Sorry
yeah,
so
what
we
do
is
that
we'll
have
to
extract
sorry
we'll
have
to
pull
down
the
images
from
docker
hub
or
any
other
registries,
which
has
the
images
that
is
being
used
for
e2e
test
re-push
them
to
a
registry.
That's
typically
local.
That's
just
like
a
registry,
that's
accessible
by
the
air
gap.
Cluster
then
configure
that
to
the
registry
to
the
local
registry.
D
D
A
Yeah,
I
sorry
for
taking
so
long
to
first
respond
on
the
issue.
I
started
trying
to
investigate
what
we
do
with
images
today,
because
so
basically,
I
feel
like
folks
from
vmware
got
real
interested
in
air
gap
testing
a
number
of
years
ago,
and
I
felt
like
I
don't
think
I
was
a
I
was.
I
didn't
review
the
particular
piece
of
code
that
added
this
registry
configuration
to
e2b
test,
but
it
seemed
like
that's
likely
where
that
came
from,
and
I
thought
that
was
sort
of
sufficient.
A
So
when
it
comes
to
starting
to
build
scaffolding
to
like
start
to
smooth
out
that
workflow
like
I
was
trying
to
express
in
the
so
sorry,
let
me
back
up.
I
guess
I
had
a
couple
concerns
number.
E
A
Like
the
e2e
test,
binary
changing
it
to
not
run
tests,
I
wasn't
sure
if
we
had
prior
for
that-
and
we
do
there's
like
a
version
flag
and
there's
also
a
list
conformance
tests
flag
which
causes
the
test
to
print
stuff
out
and
then
not
actually
run
tests.
A
So
it's
like
okay,
we've
got
prior
art,
there's
also
one
that
lists
all
the
images
that
are
used,
and
I
really
I
was
hoping
that
we
could
like
just
use
the
list
of
images
and
that
would
be
sufficient
for
you
to
do
air
gapping,
because
you
know
which
images
you
have
to
like
put
into
your
aircraft
registries.
A
But
the
problem
is
we
have
these
like
weird
variable
names
that
correspond
to
those
registries
and
like
in
my
ideal
world.
We
just
have
a
list
of
like
registries
to
search
and
replace
so
like
change,
case.gcr.io
to
my
awesome,
local
registry
or
whatever.
So
that's
kind
of
is
kind
of
my.
A
My
only
knit
is
like
because,
like
wilson
was
saying,
like
you
know,
looking
at
the
the
docker
library
images
that
we
depend
on
we're
going
to
want
to
move
those
to
a
different
registry-
and
there
are
a
couple
other
gcr
dot,
io
registries-
that
we
want
to
move
to
case.gcr.io,
and
so
that
makes
me
start
thinking
about.
Like
do
we
need
to
version
this
registry
list,
and
so
I
just
want
to
be
like
mindful
of
attempting
to
create
any
kind
of
api
around
the
etv
test
thing
right.
A
So
if
we
think
of
it
in
just
like
registries,
that
you
have
to
search
or
place
that
that
feels
a
lot
easier.
So
I
was
trying
to
propose
a
way
where
we
could
probably
support
both
the
old,
like
variable
names,
that
you
have
to
inspect
the
code
to
know
and
the
registries
that
you
could.
You
could
just
get
by
looking
at
the
existing
list
of
images
because,
as
we
add
images
for
new
tests,
you're
gonna
have
to
look
at
that
list
of
images
anyway,
to
figure
out
what
you
need
to
mirror.
A
It
also
sent
me
down
this
long
rabbit
hole.
Apparently
I
discovered
that
our
storage
e2e
tests
do
some
crazy
stuff,
where,
like
they
load
in,
manifests
and
then
dynamically
patch
images
in
the
manifests
on
the
fly
based
on
the
registry
names,
so
that
was
that
was
fun
anyway.
Like
sorry,
my
tldr
is
like
I'm
supportive
of
it.
I
dropped
an
issue
in
the
comments
describing
why,
and
I
see
your
your
follow-up
questions
wilson,
so
I'll
I'll
respond.
There.
A
That,
okay
yeah
thanks
for
thanks
for
bringing
that
up.
Okay
next
up,
I
wanted
to
hand
off
to
grant
grant.
I
assume
you
might
need
to
share
your
screen
so.
F
If
everyone
can
see
this
guy,
it
should
just
be
the
this
metrics
issue
over
the
thanksgiving
break.
I
took
all
of
the
or
a
good
portion
of
the
metrics
in
our
metrics,
like
subdirectory,
or
the
queries
and
added
them
as
data
sources
to
data
studio
and
then
just
put
together
like
a
little
sample
report
for
things
like
yeah
our
commit
consistency.
F
This
is
a
little
bit
more
human
readable
and
I
just
wanted
some
people's
input
on
the
issue
of
what
metrics
they
might
want
to
see
from
the
metrics.
What
might
be
useful
and
where
this
should
live.
It's
embeddable-
and
I
was
thinking
that
maybe
triage
like
a
new
page
added
to
triage,
might
be
a
good
spot
for
it.
But
I
was
hoping
I
could
embed
it
in
just
the
metrics
readme
itself,
but
it
seems
like
github
doesn't
support
iframes
as
inside
the
markdown,
but
it
does
automatic.
F
F
B
F
B
Yeah
I'd
be
keen
to
help
you
out
on
that.
If
you
wanted,
but.
B
B
F
F
I
am
not
sure
I
was
assuming
since
it's
embeddable,
it
can
be
available
publicly
without
any
access
rights,
but
I
have
I
can
test
it
inside,
like
a
incognito
or
something.
B
I've
been
talking
with
grant
on
this
in
the
past
day
or
two,
and
one
of
the
questions
I
have
is
whether
or
not
I
could
be
given
access
to
the
bigquery
console
in
order
to
run
queries.
I
because
that's
something
I
might
be
interested
in
doing
for
other
work
and
for
ci
signal
related
work.
E
A
So
you
you
can
do
that
today,
yeah,
if
you
have
a
so
like.
I
did
this
with
my
personal
account
before
I
worked
for
google
I
just
signed.
I
just
like
signed
up
for
google
cloud
or
something,
but
I
never
actually
set
up
any
building
information
so
that
I
could
use
the
free
tier,
yeah
you're
able
to
query
up
to.
I
think
it's
up
to
a
terabyte
of
data
from
gubernator
without
having
to
like
or
from
from
bigquery
without.
B
Having
to
pay
anything
okay,
yeah,
okay,
so
alrighty,
so,
okay
right,
I
I
didn't
even
think
about
cost,
but
but
yeah,
but
let's
yeah.
I
might
I've
tried
to
do
that,
but
I'm
I
may
have
there
could
be
user
error
there.
You
know
on
that
so,
but
but
I
was
getting
permissions
this
year
when
I
tried
to
run
queries
and
query
editor.
E
A
Add
you
specifically
to
that
project
so
like
I
would
like
to
see
us
migrate,
that
data
set
to
sort
of
the
community-owned
google
cloud
organization
where
you've
added
tons
of
people
to
be
able
to
do
stuff.
A
A
B
A
I
would
really
love
for
people
to
get
curious
and
explore
that
data,
because
I,
I
think,
there's
a
lot
of
useful
stuff
there.
This
used
to
be
available
via
thing
called
telodrome,
which
was
basically
a
grafana
dashboard
that
did
not
query
bigquery
directly
but
queried
something
else
that
bigqueries
were
periodically
dumped
into.
A
So
that's
where
my
question
of
anonymous
access
came
from
because
we
used
to
have
anonymous
access
there.
The
the
other
thought
I
I
have,
I
guess,
would
be
the
ability
to
either
change
time
ranges
on
the
the
metrics
dashboard
or
have
certain
dashboards
that
are
updated
more
frequently.
I
guess
because
the
way
I
personally
used
telodrome,
I
don't
know
if
anybody
else
did
was
one.
A
A
I
want
to
be
able
to
surface
the
the
issues
that
should
most
urgently
be
dealt
with
like
top
flakes
and
stuff
or
tests
that
suddenly
start
you
know
their
duration,
suddenly
changes
or
their
failure
rate
suddenly
changes,
or
something
like
that,
and
so
today
people
go
to
test
grid
to
look
for
that
sort
of
stuff
which
is
updated
like
every
15
minutes.
20
minutes,
it
kind
of
depends
on
how
frequently
the
updater
runs,
and
I
would
love
to
see
like
the
dashboards
generated
from
those
bigquery
metrics
to
update
at
a
similar
cadence.
A
F
That's
definitely
possible,
I
think
we
can
go
down
to
like
even
30
minutes
update
cycle
and
if
I
pipe
through
the
like
finish
times
of
jobs
on
the
queries,
then
we
can
actually
filter
with
like
a
little
calendar.
A
Cool
thank
you
for
sharing
that
with
us
great.
I'm
super
excited
to
see
all
that
data
visible
again.
A
A
So
somebody
here
has
a
question
about
what
can
be
done
to
get
image
pushes
happening
again.
G
Whoever
has
a
question
I
have
the
answer:
there's
a
request.
I've
placed
in
chat,
I
think,
when
we
merge
that
it
should
be
okay
for
the
most
part,
recently
antonio
submitted
a
pull
request
which
basically
split
the
the
personal
job
into
individual
image
job.
G
So
it's
a
lot
more
atomic
in
that
sense.
So,
even
if
one
image
fails,
the
others
should
still
function
properly,
which
is
great,
and
this
prequels
that
I
linked.
Basically,
we
need
that,
because,
in
order
for
docker
build
decks
to
properly
build
multi-arc
images,
it
has
to
call
that
register
that
sh
with
the
persistent
flag.
This
is
like
it
really
has
to
do
that,
and
we've
included
qmo
and
camera
binaries
into
the
gcp
docker
gcloud
image,
which
is
being
used
by
google
cloud
builder
to
build
the
test
images.
G
So
now,
when
you're,
actually
using
that
image
to
to
build
old
images,
it
has
all
the
necessary
dependencies
and
requirements
to
build
all
the
images,
including
windows
ones,
including
arm
images,
s
390x.
So
on
so
forth,
I've
tried
it
multiple
times,
even
double
rasa
systems,
with
nothing
on
them
after
reboots.
So
I
had
no
lingering
cpu
flags
set,
so
it
worked
for
me.
G
If
you
really
want
to
you,
can
test
it
yourself
in
the
pull
request.
I
also
have
a
paste
with
that
and
the
first
line
is
the
command
I
use
to
build
the
image.
G
G
G
But
after
that,
I
think
we
have
to
trigger
the
image
jobs
individually
because
before
it
used
to
be
perf,
the
entire
folder
on
the
kubernetes
test
images
and
after
this
we'll
have
to
either
get
someone
to
trigger
those
jobs
or
commit
something
to
those
images
to
get
the
possibility
automatically
running
afterwards.
A
Yes,
I
have
a
separate
issue
for
migrating
us
to
all
of
the
images
that
are
being
built
via
that
in
gcd.
I
will
paste
it
in
chat,
I
think
so
so.
Essentially
my
short
answer
is
like:
can
we
wait
until
the
release
is
cut?
A
It
kind
of
we
passed
test
freeze,
and
so
at
the
moment
we
should
really
be
looking
to
either
fix
broken
tests
or
fix
regressions
that
have
been
introduced
or
revert
stuff,
and
I
feel
like
I.
I
appreciate
that
rob
is
here
on
behalf
of
the
ci
signal
like
I.
I
don't
think
we
should
be
changing
the
images
that
are
used
for
tests
at
this
point.
A
The
current
release
schedule
that
I'm
aware
of
is
that
we
plan
on
cutting
the
release
tuesday
of
next
week,
and
we
start
getting
to
a
point
where
people
are
going
to
want
to
see
like
continuous
passes
against
the
same
sha
or
the
same
yeah
sha
of
kubernetes,
to
make
sure
that,
like
it's
stable
and
we're
not
seeing
flakes
and
stuff
I'm
trying
to
keep
in
mind
that
the
scale
tests
that
we
run
take
a
while,
and
so,
while
in
theory,
this
shouldn't
break
anything
because
it's
just
changing
how
stuff
gets
built.
A
H
I
actually
think
a
little
different.
I
think
we
should
just
not
bump
the
images
we
use
and
test.
I
think
we
should
get
the
fix
for
images
being
able
to
build
in.
I
think
we're
gonna
wind
up
wanting
to
cherry
pick
that
back
anyhow
and
people
have
been.
You
know
on
their
hands
waiting
for
this.
H
H
I
do
think
it
makes
sense,
perhaps
to
not
to
not
change
what
images
we
run
and
test
this
late.
But
if
the
release
team
is
comfortable
with
it,
I
think
we
should
shift
the
fixes
to
just
being
able
to
build
them.
H
A
A
Them
right,
so
I
also
acknowledge,
and
I'm
just
one
voice
here.
I
really
think
it's.
It
should
be
the
release
team's
call.
So
I
think
this
maybe
is
a
discussion
best
had
in
the
sick
release
slack
channel
to
work
out.
I.
H
A
So
if
the
release
team's
cool
with
it
like
and
we're
all
cautious
about,
you
know
what
we
what
we
push
and
we
make
very
sure
we
revert
stuff
if
it's
broken
and
we
respect
when
the
release
team
decides,
they
really
don't
want
to
see
anything
else,
land
and
kubernetes
sure,
because
I
do
know
a
lot
of
stuff
we
want
to
do
here.
I
I
I
was
wondering
part
of
my
questions
where
we
had
made
some
changes
to
an
image
that
was
required
for
a
test
of
ours
and
and
we
had
it
where
this
image
wasn't
pushed,
and
I
wasn't
sure
of
the
process,
the
normal
process
for
triggering
it
when
it
failed,
and
I
got
different
different
suggestions
depending
on
the
the
group
that
I
popped
into,
I
did
get
a
suggestion
from
testing
ops
to
create
a
pr
that
changed
a
file
in
order
to
at
the
right
level
to
trigger
that,
and
that
didn't
quit
feel
quite
the
right
approach,
and
it
was
also
mentioned
by
someone
else
that
wasn't
maybe
the
right
approach
and
then
I
was
also
asked
to
trigger,
via
asking
the
and
for
on
call
to
manually,
trigger
the
job,
and
so
my
question,
I'm
totally
cool
with
waiting
until
test
freeze.
I
This
isn't
a
major
thing
it
just.
It
did
block
us
from
from
releasing
more
coverage
this
release,
but
I
would
love
to
know
what's
the
best
practice.
As
we
start
to
add
supporting
test
infrastructure
to
our
images
when
we
need
to
bump
this
release,
was
it
just?
Was
this
a
one-off
and
we
don't
have
to
worry
about
it
too
much
or
is
there
a
process
going
forward
where
we
would
need
to
do
something?
I'm
not
sure
what
that
is.
A
Let
me
think
about
this,
so
there's
the
proud
configuration
file
that
says
like
build
for
master
or
build
from
a
release
branch,
and
then,
when
prow
turns
that
configuration
into
a
proud
job
crd,
it
resolves
that
to
hash
or
sorry
a
sha,
and
then
that
is
what
is
used
to
run
the
job,
and
so,
if
you
want
to
re-trigger
the
job
by
clicking
the
retry
button
via
in
browse
ui,
what
that
does
is
basically
just
say:
hey,
I
see
this
crd.
A
Could
you
please
like
create
another
projob
for
it,
so
it
uses
the
same
shop.
It
doesn't
try
to
refresh
hey
this
says
it's
using
master,
so
I
should
go
and
use
head
for
master
right.
That
also
depends
upon
how
long
that
proud
job
crd
stays
in
that
kubernetes
clusters
registry,
which
I
think
it
times
out
within
48
hours.
At
this
moment,
it's
sort
of
dependent
on
like
how
much
traffic
pro
is
getting
and
stuff
different
pro
installations.
Have
it
tuned
differently.
A
A
A
So
I
I'm
open
to
ideas
on
how
to
make
that
process
better,
typically
like
changing
how
long
it
takes
for
the
like.
We
could
change
the
life
cycle,
the
crd,
but
that
kind
of
requires
tuning
based
on
the
performance
and
traffic
that
the
cluster
handles.
A
A
But
I
think
we
as
a
community
would
really
appreciate
that
ability,
because
we
are
getting
a
lot
more
dependent
upon
post,
submits
and
it's
kind
of
there's
kind
of,
like
the
broader
question
of
like
making
sure
post
submits,
are
reliable
and
flake
free.
Like
we've
already
got
feta,
fatabot
set
up
to
go
and
just
spam
retest
on
pr's
that
have
flaky
jobs
right
and
just
keeps
rerunning
the
pre-submits
until
they
pass.
A
I
One
thing
that
I
had
difficulty
connecting
the
dots
with
on
post
submits
is
where
to
find
them
sort
of
we
would
I
I
don't
know
whether
that
the
who
has
permission
to
see
the
retry
and
also
how,
because,
because
finding
that
particular
pre-submit
that
failed
and
seeing
the
job
logs
wasn't
super
obvious
to
us,
and
that
may
be
a
documentation
thing
or
maybe
we're
looking
in
the
wrong
area,
but
post-submit
failures,
weren't
obvious,
and
definitely
when
you're
talking
about
the
proud,
are
the
our
friends
bot
coming
back
and
checking
to
retry
the
pre-submits.
A
Yeah,
so
I
don't
want
to-
I
want
to
stitch
while
I'm
going
to
search
for
stuff,
but
I
just
want
to
comment
on
that
real
quick.
So
there
is
an
open
issue
somewhere
in
test.
Infra
that
talks
about.
Could
we
make
post
submit
reporting
better?
So,
like
you
know
how,
when
you
update
a
job
config,
and
you
merge
that
pr
you'll
see
a
comment
from
prow,
that's
like
hey.
I
deployed
these
yaml
files
which
equate
to
that
job.
A
A
What
we
have
right
now
is
that
post
submits
will
post
the
status
to
to
the
commit
status
context
the
catch
is
they
post
it
to
the
merge
commit.
A
So
if
I
open
up
a
pr
that
changes
an
image
and
then
I
merge
that
the
the
merge
commit
is
what
will
eventually
have
a
little
green
check
mark
or
a
little
red
x,
depending
on
whether
the
post
submit
passed.
J
Oh
sorry,
one
additional
comment
to
this
topic
of
possums
and
actually
also
periodic
reporting,
because
they
have
the
same
problem.
What
we've
been
doing
in
openshift
is
we've
been
using
a
slack
reporter
that
will
basically
for
a
given
set
of
jobs
report
to
a
slack
channel
of
ours.
If
they
fail
because
for
success,
we
don't
really
care,
there's
nothing
to
do.
A
There's
we
also
have
the
job
set
up
to
email
alerts
to
an
email
address.
If
you
take
a
look
at
the
job
config,
I
think
it's
kubernetes
sync
testing
alerts
gets
email
alerts
for
any
of
the
test,
image
jobs
that
fail,
and
so
you
could
get
alerts
that
way
too.
Instead
of
slack.
G
I
think
I
I
suggested
some
time
ago
that
we
could
maybe
add
some
commands
to
rebuild
a
certain
image
and
that
would
allow
a
group
of
people
to
use
that
command
to
manually
trigger
the
rebuild
for
an
image.
G
A
Yeah,
so
I
don't
know,
I
personally
would
be
less
interested
in
solving
this
problem
specifically
for
image
building
jobs,
and
I
try
to
look
at
it
at
the
abstraction
level
of
like
how
do
I
trigger
a
periodic
or
a
post
submit
on
demand
and
I'm
yeah,
I'm
not.
I
guess
I
don't
have
an
idea
off
the
top
of
my
head
there,
but
I'm
not
sure
like
adding
a
slash,
build
command
would
make
a
lot
of
sense.
A
We
already
have
the
trigger
plug-in,
which
is
what
you
can
use
to
slash
test
whatever,
and
it's
a
pre-submit
job
name.
I
guess
we
could.
We
could
talk
about
looking
into
just
adding
the
ability
to
trigger
post,
submits
or
ci
jobs
from
that,
but
I'm
sure
we'd
want
to
take
a
good
long.
Look
at
how
we
authorize
that
stuff.
A
I
don't
know
alvaro
how
does
openshift
handle
like
letting
people
trigger
periodics
and
post
submits
arbitrarily.
J
Yeah,
the
same
also
via
this
v1
authentication
group
setting
you
can
have
on
the
jobs
but
yeah
the
the
issue
we
also
have
is
that
often
people,
especially
for
post
submits,
want
the
latest
version,
because
not
only
the
sha.
Actually,
it's
also
the
job
configuration
itself
because
of
that
change,
it
also
won't
get
picked
up.
J
A
Yeah,
so
I
think
there
are
a
lot
of
great
ideas
here.
If
you
want
to
open
up
issues,
so
we
can
keep
track
of
them.
I
think
this
is
something
anybody
could
work
on.
That
would
be
appreciated
by
everybody.
Oh,
that
would
be
super
cool.
I
I'll
put
together
a
ticket
for
at
least
for
discussion,
some
approaches
for
on-demand
periodic
and
post-event
jobs,
because
it
does.
It
impacts
a
lot
of
the
areas
that
we're
that
we
spend
a
lot
of
time
in.
A
Okay,
sean,
I
think
I
saw
you
pop
up.
Oh
did
you
want
to
talk
to
pre-submits
in
test
grid.
K
All
right,
one
of
my.
K
Isn't
working
I'll
deal
with
that
later?
Yes,
I
want
to
talk
about
pre-submits
in
test
grid.
K
There
is
an
centralized
issue
for
putting
pre-submits
in
test
grid
for
almost
all
repos
and
modules
and,
with
the
exception
of
a
short,
allow
list,
this
is
enforced
with
our
test
grid
like
repo
specific
testing,
some
users
of
test
grid
like
to
have
their
pre-submits
up
and
some
think
that
it's
noisy
and
flaky
and
that
they
don't
want
it.
Aaron
commented
on
this
thread
kind
of
explaining
why
we
would
want
to
have
pre-submits
visible
on
test
grid.
A
So
not
to
put
you
on
the
spot
rob,
but
what
do
you?
How
does
ci
signal
keep
track
of
like
how
or
how's
this
pre-submit
behaving.
B
A
The
the
verified
job
failed
on
this
pr.
How
do
you
answer
the
question?
Is
this
failing
on
a
bunch
of
pr's
like?
Did
it
get
way
more
flaky,
all
of
a
sudden.
B
Well,
well
in
in
in
terms
of
our
focus
on
ci
signal,
I'm
not
really
looking
at
very
well,
I'm
looking
for
verify
master
is
is
generally
is
generally
pretty
good
like
it
hasn't,
come
across
hasn't
come
across
my
desk
in
a
long
time,
and
that's
the
only
one
that
I'm
looking
at
and
master
blocking
and
then,
if
I
look
at
master
forming.
A
Right
but
those
are
so
my
point
is
those
are
all
those
are
all
periodic
jobs
they
run
on
whatever
is
in
master
or
in
a
release.
Branch.
B
Do
you
keep
track
of
like
the
merge
blocking
jobs
and
not
really?
No,
because
because
our
focus
is
master
blocking
and
mastering
forming
those
those
two
dashboards.
B
Expectation
is:
is
this
we're
dealing
with
after
the
merge
in
terms
of
monitoring
the
signal
for
the
release?
Does
that
make
sense.
A
Okay,
yeah
so
who
keeps
track
of
whether
the
merged
blocking
jobs
are
asking
your
failing.
I.
H
That
I
mean,
do
you
mean
like,
like
failure
rate
or
something
it's
a
little
hard
to
like
I've,
tried
to
monitor
that
in
the
past
it's
a
little
hard
to
poke
around
for
like
piers
that
just
don't
build,
and
sometimes
people
are
spamming,
pairs
that
don't
build,
and
then
the
failure
rate
goes
through
the
right
and
doesn't
really
mean
anything.
So
the
heat
map
on
product
h.I.o
is
pretty
helpful
because
you
can
visually
filter
out
really
short
runs,
but
I
don't
think
there's
any
like
hard
tool
for
like
how
you
would
monitor
the
head.
A
H
So
I
look
at
the
I
look
at
the
pass
rate
versus
some
of
the
other
pre-submits
that
are
similar
as
a
control,
and
I
look
at
if
I
can
visually
see
a
cluster
of
failures.
H
That
is
not
that's
over
some
minimum
run
time
and
maybe
spot
check
those
the
quick
failures
to
make
sure
that
they
are
just
build
fills-
and
I
do
this
like
just
annually
like
I
did-
vary
sometimes
when
things
have
been
pretty
stable,
I'm
not
checking
as
often,
but
if
we've
you
know
in
the
past,
we've
had
a
rougher
patch.
It
might
be
something
I
do
daily
here
or
some
more
than
daily,
okay
kind
of
like
ti
signal,
but
just
for
just
for
what
I'm
working
on.
A
Yeah
I'll
just
share
like
my
personal
experience,
but
I
recognize
I'm
just
one
person
I
feel
like
there
are
a
few
heroes
who
are
in
charge
of
kind
of
like
noticing
that
certain
pre-submits
started
failing
or
started
flaking
significantly
more
in
kubernetes,
and
I
think
the
general
process
I
have
seen
is
that
people
who
review
a
lot
of
prs
start
to
notice
that,
like
the
same
job,
is
failing
across
multiple
pr's,
and
so
they
start
to
ask
the
question
like
hey:
has
anybody
filed
an
issue
that
this
job
is
failing
or
not,
and
it
can
take
like
a
couple
hours
before
we,
as
a
community
notice
that,
like
one
of
the
merge
blocking
jobs,
is
failing
or
flaking
a
lot
and
then
it
usually
falls
to
like
ben
or
sometimes
me
or
dan,
from
ci
signal
or
often
ligit,
to
like
go
figure
out?
A
What
what
introduced
the
regression
or
what
broke
things
or
often
times
there
are
questions
about
like
hey,
did
something
in
the
test:
infra
change
that
was
there
some
environmental
issue
that
caused
this
job
to
start.
Failing
generally
in
those
scenarios,
you
want
to
start
looking
at
like
okay,
so
can
we
pinpoint,
when
this
job
over
time
started
to
fail
or
like
started
to
look
a
lot
worse
and
so
for
jobs
that
do
report
to
test
grid?
A
I
can
just
go
to
pastries
and
take
a
look
at
that
view,
but
for
jobs
that
don't.
I
have
to
know
that
I
could
also
go
through
and
click
on
go
to.
Products.I
owe,
and
I
could
search
for
the
pre-submit
job
name
and
then
I
could
click
onto
one
of
those
runs,
and
then
I
could
click
on
job
history
and
then
I
could
start
to
like
scroll.
B
H
Have
one
for
blocking
kubernetes
pre-submits?
I
I
found
it's
not
very
useful
for
two
reasons:
one
on
test
grid,
it's
harder
to
filter
out
the
noise
of
like
oh,
this
is
just
a
failed
build
from
like
code
that
doesn't
build
you're
gonna
need
to
click
through
to
the
logs
anyhow,
and
I
guess,
because
you
can't
see
the
overall
time
on
the
on
the
main
grid
view
and
the
other
one
is
the
test.
H
Script
just
doesn't
perform
super
well
when
you
have
like
a
ton
of
data,
so
a
lot
of
the
pre-submits
have
like
two:
they
have
too
much
running.
I've
even
ran
into
problems
with
that
we've
like
owned
the
front
end
from
trying
to
load
some
of
the
submit
dashboard
pages.
B
Yeah
not
to
deal
with
the
discussion
park
this
if
we
want
one
of
the
distinct
one
of
the
things
I'd
like
to
see
through
the
ci
infrastructure,
is
to
to
make
the
distinction
between
build
and
test
a
bit
more
clear,
so
so
that
we
would
have
some
shot
at
a
at,
maybe
pruning
that
data
and
making
those
distinctions
so
that
yeah
we're
not
mega
damning
the
front
end
as
it
were.
H
A
So,
just
to
finish
my
thought
there,
like,
like
my
two
conferences,
it's
all
good
like
I
don't
want
to
be
the
only
one
speaking
I
really
do
appreciate
I
just
want
to
so.
My
principal
concern
is
like
we
force
all
post,
submits
and
periodics
to
report
to
test
group,
but
we
don't
do
that
for
presidents.
A
So
at
the
moment,
people
need
to
know
that
there's
one
way
like
they
can
go
to
test
grade
for
all
their
post
submits
and
periodics,
but
they
have
to
go
this
other
way
or
some
pre-submits,
depending
on
whether
or
not
they've
been
added
to
that
period.
A
That
was
just
my
thing.
It's
like
I!
This
is
from
years
ago.
I
added
this
test
because
I
was
like
well.
I
don't
want
to
solve
this
problem
for
all
like.
I
don't
want
to
be
the
guy
who
goes
and
changes
everybody
else's
jobs
for
them,
so
instead
I'll
create
like
an
exemption
list
of
like
well.
A
A
H
I
found
it
really
hard
as
well
to
get
test
grid
alerting
working
well
for
pre-submit
because
of
the
fact
that
you're
going
to
see
pass
unfail
just
from
pr
noise.
But
I
do
think
that
something
like
grant's
work
on
the
on
the
business
dashboard
is
where
we
might
be
able
to
say.
Okay,
let's
have
people
other
than
heroic.
Looking
at
how
reliable
our
appraisals
are
and
have
actual
metrics,
instead
of
trying
to
like
squint
at
a
job
or
something.
K
Yeah,
I
can't
really
recommend
setting
up
alerts
for
pre-submits
in
general,
which
is,
I
guess,
kind
of
where
my
confusion
on
adding
them
to
test
grid
came
from,
but
adding
them
to
test
grid
doesn't
necessarily
require
an
alert.
It
doesn't
set
up
an
alert
for
you.
It
doesn't
set
up
stale,
alerting
or
any
of
those
kinds
of
things
that
wouldn't
necessarily
make
sense
in
a
pre-submit
context.
It
just
kind
of
displays
it
in
the
test
grid
format
for
up
to
15
days
back
or
I
guess
until
it
runs
out
of
memory.
K
So
like
from
that
perspective,
I
can
see
like
kind
of
a
contrabex
reason
for
adding
them
like
adding
those
pre-submits
like,
if
you
add
the
pre-submits
and
have
a
test
that
requires
them
to
add
a
pre-you
know,
add
them
to
test
grid.
Then
you
can
just
say:
oh
every
proud,
job's
on
test
grid
they're
all
there.
K
So
you
can
go
to
test
grid
and
look
at
a
history.
There's
another
place.
You
can
look
to
also
get
a
jobs
running
history,
so
maybe
this
is
also
a
question
for
like
contrabax
as
well
as
testing,
because
I
think
that's
kind
of
the
crux
on
it.
H
And
not
not
all
jobs
are
going
to
happen
and
for
jobs
that
don't
it
it
is
like
not
actually
significantly
more
useful
than
just
like
native
display,
which
will
search
for
the
job
name.
Probably,
actually
I
don't
know.
Oh,
I
do
also
want
to
add
a
quick
assignment
in
a
perfect
world.
I
actually
have
a
reporting
enabled
first
imprisonment
like
to
me
personally,
because
I
would
like
to
know-
and
that's
a
little
bit
less
to
just
be
like.
H
H
But
if
that
was
a
thing
I
I
would
absolutely
use
it
and
I'm
trying
to
use
it,
as
is
not
surprisingly,
it's
not
very,
not
very
helpful
tuning
up
the
number
of
jobs
and
the
reasons
it
helps
a
bit
to
filter
out
for
it
because,
like
you,
can
have
an
alert
for
when
it
just
starts
failing
continuously,
which
does
happen
occasionally
like
an
infrastructure
outage
or
something.
But
you
need
to
turn
it
pretty
high,
so
it
takes
a
while
to
notice.
A
I
want
to
be
cognizant
of
the
fact
that
we're
five
minutes
over
time
and
I
appreciate
everybody
for
spending
the
extra
time
with
us,
so
I
don't
know
if
we
actually
came
to
a
resolution.
The
last
thing,
like
I
said
people
feel
like
it's
it's
cluttering
things
up.
If
there's
a
there
is
a
cost
to
it.
A
A
Anyway,
thank
you
all
for,
for
showing
up.
A
I
just
want
to
put
out
throw
out
there
the
idea
that
well
we'll
probably
be
meeting
again
in
two
weeks,
but
then
I
think
the
next
meeting
after
that
will
likely
fall
really
close
to
the
new
year
when
I
suspect
a
number
of
people
will
be
available,
so
we
might
want
to
consider
canceling
that
so
I
was
thinking
either
like
next
meeting
or
the
first
meeting
that
we
have
in
the
new
year.
We
could
kind
of
get
together
and
talk
about
sort
of
what
our
what
our
goals
are
for
the
year.
A
What
are
some
of
the
bigger
boulders?
We
want
to
move
forward,
just
the
thought
and
I'll
send
this
out
to
the
mailings
too.
So,
thank
you
all
for
your
time.
I
hope
you
all
have.