►
From YouTube: 2021-04-01 Kubernetes SIG Scalability Meeting
Description
Agenda and meeting notes - https://docs.google.com/document/d/1hEpf25qifVWztaeZPFmjNiJvPo-5JX1z0LSvvVY5G2g/edit?ts=5d1e2a5b
A
Okay,
hello,
everyone:
this
is
six
scalability
meeting
february
18th.
A
We
don't
seem
to
have
anything
at
the
agenda.
Yet,
to
be
honest,
I'm
like
I
was
mostly
busy
last
weeks
with
like
feature
freeze
and
production
readiness,
and
I'm
still
recovering
from
that.
B
B
Hey
waytech,
I
I
can
go
first.
So
so
are
you
able
to
hear
me?
B
Yes,
yes,
okay,
yeah
yeah,
so
I
had
a
question
about
like
specifically
with
respect
to
endpoint
slices,
so
so
now
that
we
are
actually
splitting
those
endpoints
into
different
blocks
like
do
you
know
how
we
are
like
measuring
or
like
if
we
plan
to
measure
the
like
the
network,
programming,
latencies
and
stuff
with
with
that
like?
How
are
we
going
to
measure
p99?
A
It
seems
I
was
muted,
so
no,
I
think
the
definition
doesn't
doesn't
really
change.
The
definition
was
basically
per
pod,
more
or
less
so
that
doesn't
really
matter.
If
we
are
using
end
points
or
endpoint
slices,
it
doesn't
change
anything.
B
A
We
are
basically
measuring
from
from
when
pod
was
turned
into
being
ready
or
not
ready
to
when
it's
reflected
in
ip
tables.
A
Okay.
Well,
I
think
it
like
yeah.
It
doesn't
have
to
be
ip
tables,
but
I
guess
we
didn't
implement
it
for
others
than
ip
tables.
I
might
be
wrong,
but
that's
my
feeling,
if
I
remember
correctly.
C
Hi,
I
have
a
good
question
around
infrastructure
for
scale
testing,
so
there
is
a
perv
dash
project
inside
perf
test
repository.
I
wonder
how
actively
it
is
developed.
I
I
see
there
is
a
public
installation
of
it
and
I
was
wondering
how
feasible
it
is
to
have
it
locally
installed.
A
It
is
definitely
feasible.
We
are,
for
example,
in
google.
We
are
running
our
own
instance
too
internally,
so
it's
it's
generally
quite
easy
to
set
it
up
regarding
how
actively
it's
being
developed,
it's
not
really
being
very
actively
developed.
What
we
are
thinking
about
is
we
are
thinking
about
switching
at
least
some
parts
of
our
tooling,
in
particular,
of
perth
dash
to
to
something
what
is
called
mako.
A
A
For
load,
generation
and
and
and
stuff
like
that,
doesn't
really
work
well
for
our
purposes,
but
the
depart
from
like
gathering
results,
processing
them,
detecting,
regressions
and
and
analyzing
data,
and
all
that
stuff
is
actually
pretty
generic
and
we
ideally
would
like
to
migrate
to
that.
That's
why
we
are
not
really
investing
into
dash,
but
it
like.
We
don't
have
much
capacity
to
do
that.
A
C
Got
it
thank
you
so
so
mainly
you
publish
the
results
to
test
grid
which
is
part
of
sick
testing
and
really
part
of
testing
infrastructure,
and
there
is
perth
dash
to
be
replaced
by
maca.
It
does
that
conclude
a
set
of
visualization
tooling
around
it
or
other
more
projects.
A
No,
I
think
that
that's
that's
it.
Okay,.
A
A
The
marco
yeah,
I
think,
let
me
let
me
find
it
yeah.
D
Yeah,
this
is
the
first
time
attending
this
meeting,
I'm
I'm
wilson
and
I'm
from
by
dance,
and
actually
we
are
working
on
the
schedule
like
cluster
scalability
projects
and
we
are
trying
to
you
know
like
set
up
cluster
like
around,
like
10
10k
to
100k
nodes.
That's
the
the
target
is
very
ambitious,
but
we
are
working
on
this
so
yeah.
D
That's
why
I
talk
with
different,
like
you
know,
cognitive
teams
and
see
if
we
can
get
some,
you
know
ideas
or
get
some
of
our
like
idea,
get
some
ideas
and
also
get
or
like
the
optimization.
D
You
know
like
get
into
the
open
source
or
contribute
back
to
the
open
source
projects,
or
something
like
this.
Okay,
yeah
nice
to.
C
B
Hey,
I
can
go
ahead
with
another
question,
so
this
is.
This
is
more
like
curiosity
and
I
I
probably
haven't
fully
thought
they
thought
about
this
idea
yet,
but
so
so
like
when
you're
sending
a
request
when
we're
reading
something
from
hcd
right
from
like
when
we
do
a
regular
read
without
any
resource
version,
and
the
list
list
call
goes
to
hcd.
B
So
in
such
cases
like
today,
like
the
response,
the
whole
response
comes
back
from
hcd
right.
Is
it?
Is
it
possible,
let's
say
to
not
do
that
and
just
return?
For
example,
a
resource
version
of
what
the
response
would
have
been
would
have
had
and
can
the
can
the
api
server
if
it
has
a
if
it
has
a
cache,
which
is
at
least
at
that
resource
version,
then
it
can
just
serve
from
that
is
that
is
that
something
which.
A
B
A
A
Let
me
find
it
here,
I
think
I
will.
I
will
probably
add
it
to
the
notes
for
the
agenda
at
our
agenda
notes,
but
I
think
I
can
probably
yeah
you
can
keep
discussing
stuff.
I
will
probably
find
it
pretty
quickly.
B
Okay,
yeah,
so
that's
one
I
mean
so
I
had
a
question
around
that.
B
Probably
you
guys
would
have
already
discussed
it
on
that
thread
so
like
as
in
if
you're
going
to
serve
from
cash,
which
is
at
least
at
that
resource
version
or
higher,
then,
is
that
going
to
cause
problems
when
there
are
like
two,
two
apis
server
instances
right
and
then
like
once,
let's
say
you're
requesting
for
rvx,
but
one
is
on,
let's
say
x,
plus
five
and
the
other
is
on
x,
plus
10,
and
the
first
call
goes
to
the
second
one.
B
A
Now,
I
think
what
we,
what
we
were
thinking
about
is
a
slightly
different
version
of
it,
which
is
basically
you
send
the
request
to
hcd,
but
not
for
the
value,
but
just
for
its
resource
version.
You
can
do
that.
So
give
me
a
resource
version
of
that
particular
object
and
what
you
basically
do
on
api
server
is
take
the
object
with
that
research
version.
You
need
to
you.
You
need
to
have
a
quite
specialized
cache
for
that,
but
it's
doable.
A
Oh
so
you're
saying
some
kind
of
a
quorum
read
so
you
you
said
so.
What
you
do
is
like
you
send
a
request
to
at
cd
for
the
resource
version,
which
is
a
quorum
break.
We
are,
I
think
we
are
right
or
I'm
sure
we
are
by
default.
Using
quorum
reads
to
nxt
or
api
server
by
default
is
sending
quorum
rates.
A
So
you
know
what
exact
version
of
the
of
the
object
is
from
the
fcd,
but
you
don't
send
the
data
themselves
from
a
cd
and
you
don't
have
to
like
deserialize
them
and
stuff
like
that,
because
you
already
have
them
in
cash
hopefully,
and
you
you,
basically
save
that
dc
realization
and
send
and
sending
that
data.
Basically,
I
see
so.
B
So
so
just
digging
a
bit
deeper.
So
if,
if
that,
if
let's
say
you
got
this
resource
version,
but
your
cache
does
not
have
information
at
that,
it
has
something
which
is
further
into
the
future.
A
Yeah,
it
doesn't
work
now,
like
you
need
to
significantly
you
pretty
much
need
to
re-implement
the
whole
whole
stuff,
like
the
re-implement.
The
watch
question
completely
different
way.
This
is
the
link.
This
is
where
the
discussion
about
that
is
starting.
D
Okay,
I
saw
from
the
aps
server
it's
not
like
it's
using
linearized
linearizable,
like
request
to
the
active
activity
server,
so
I
saw
thought
it
was
just
like
doing
one
read
and
then
you
know
like
send
a
request
to
any
one
of
them,
but
but
specify
it's
like
you
know
by
default.
I
think
it's
linearizable
read,
but
I'm
not
sure
like
I.
I
didn't
check
like
that.
I
didn't
see
the
detail
for
the
corner.
Read
so,
do
you
guys
have
any
more
information
about
this.
D
Oh
client
level
right,
okay,
okay,
I'll,
go
back
and
double
check
that.
But
I
from
what
I've
seen
I
thought
it
was
just
doing.
The
linear
rise,
linearized,
lowest
linearizable
are
like
requests
so
that
like,
when
you
send
to
the
onenote,
then
the
activity
server
likes.
The
follower
will
request
a
leader
and
then
request
requested
the
I
mean
send
the
request
to
the
leader
and
then
later,
when
the
leader
make
sure
that
all
the
com,
all
the
like
the
current
like
id
applied,
then
the
follower
return.
A
Yeah
yeah,
I
don't
know,
to
be
honest,
I
don't
know
the
exact
decades
of
raft
and
like
how
it's
it's
the
implemented
underneath,
but
but
yeah
and
conceptually
it's
a
quorum
right
here.
D
Okay,
so
we
are
doing
it
on
the
client.
We
are
not
doing
anything
specific
right,
so
just
using
a
client
and
we're
doing
the
like
quantum
ray.
Is
that
what
you're
saying
yes,
okay?
Okay,
I
see
that
yeah.
I
think
it's
probably
just
using
something
like
using
the
thing
that
I
probably
just
said
like
the
internally
by
the
activity
server
to
do
the
do
the
quorum
read.
I
think.
D
Yeah
we
discussed
about
the
the
apk
server
and
activity
server,
so
I
think
we
now
have
actually
in
internal
in
our
company.
We
are
actually
thinking
about
like,
like
the
how
to
make
the
activity
server
like
see
like
do
the
like
do
the
sharding.
So
what
we
all
plan
is
actually
because,
like
for
the
one
activity
cluster,
the
it's
like,
basically,
it's
it
can
serve
like
many
reads,
but
for
the
right
it's
like
for
a
single
cluster
is
limited.
D
So
what
we
want
to
do
with
a
as
well
cluster
increases
we
want
to
have
the
like.
The
we
saw
that
the
opponent
like
the
activities
of
an
aps
over
here
it
becomes
a
bottleneck.
So
we
are
what
we
are
planning
actually
to
do
the
to
the
sharding
for
the
for
the
aps:
server
like
backhand,
so
that
means
like
for
the
api
server
instead
of
having
one
activity
cluster
it
can
like
use
several
activity
cluster
and
like
do
the
key
charting.
D
I
know
that
in
the
eps
server
is
already
it
can
already
for
the
different
names.
I
think
for
different
resources
using
different
active
activity
cluster,
but
it's
actually,
we
saw
the
the
number
of
as
all
cluster
size
increases
significantly.
The
number
of
paws
itself
is
very
significant,
so
what
we
want?
Actually
we
want
to
have
the
like
the
resources
distributed
like
evenly
like
amount
and
also
the
load
distributed
evenly
among
different
activity
clusters.
D
So
actually
I
in
the
community
I
discussed
about
this
before,
but
the
what
they
suggest
me
is
actually
file
like
file
kp,
so
that
people,
you
know
like-
can
discover,
discuss
about
this
and
well
actually
from
from
my
initial
design.
D
Actually,
we
we
want
to
make
the
resource
version,
and
since
it's
like
a
string
inside
the
eps
servers,
what
we
are
planning
to
do
for
the
different
activity,
cluster
backend,
we
have
a
resource
version
and
then
for
the
api
server
like
the
resource
version
string,
we
actually
contains
a
vector
which
is
actually
the
different
activity.
Back-End
cluster
version,
like
the
combination,
make
it
into
one
string.
So
in
this
way
we
know
the
different
resource
versions
for
like
on
the
different
activity:
back-end
clusters.
D
Yes,
that's
a
war
initial
plans
but
since
like
since
aps,
server
is
not
using
like
not
required
to
use
any
like
distributed
transaction
or
something.
So
I
think
like
because
api
server
is
just
using
some
of
the
basics,
such
as
like
basic
functionality
such
as,
like
the
transaction
simple,
you
know
like
update
and
delay
operation,
also
the
watch
functionality.
So
I
think
probably
this
might
be
possible
from
the
like.
A
So
we've
been
considering
that
like
couple
years
ago
and
they
were
like
you-
can
probably
try
looking
into
the
repository,
I
remember
those
discussions
happening
in
the
past.
There
were
couple
things
that
I
can't
remember
exactly.
I
would
need
to
think
about
it.
More
like
watch
is
certainly
the
non-trivial
here
how
to
how
to
make
it
basically.
A
Yeah,
if
watch
is
definitely
one
of
the
things
that
we
need
to
keep,
we
need
to
pay
a
lot
of
attention
to.
The
other
thing
is
like
how
complicated
the
exchange
will
be,
and
the
other
thing
that
is
probably
well.
It
depends
on
like
the
exact
setup,
but
the
if
you
know
up
front
like
how
many
shards
you
want
to
have
that
sounds
much
simpler,
but
it
sounds
like
much
simpler
project
problem
than
if
you
would
like
to
allow
resharding
later.
Also.
D
Nicely
yeah
recharging,
definitely
a
issue.
You
need
some
coordinator,
you
know
or
like
some
something
to
you
know
like
help
in
in
in
the
during
the.
D
B
Repository
hey
one
last
question,
so
do
you
think
it
makes
sense
to
allow
if
I
don't
know
if
the
client
already
does
my
understanding
is
it
doesn't
if
we
can
actually
ask
for
a
non-quorum
read
through
a
client
if
we
don't
actually
care
about
the
results
results
being
linearized?
A
We
so
if
you
don't,
if
you
don't
care
about
the
most
fresh
request,
you
can
simply
opt
in
for
watch
question
right,
so
setting
resource
version,
zero.
B
A
A
B
A
All
it
may
always
happen
right,
so
not
even
too
old
actually
and
to.
B
B
B
Because
this
is
actually
causing
problems,
so
I
mean
I
didn't
give
the
background
about
this.
Sorry
so
so,
like
we
were
observing
this
on
some
some
of
our
like
clusters
that
we
manage
so
what's
happening
is
like
all
those
which
are
sending
requests
to
hcd
they're
by
default
becoming
quorum
and
the
client
actually
to
be
honest,
is
probably
doesn't
care
that
much
about
being
quorum.
B
A
Yeah,
I
think
the
question
is
like
which
direction
we
would
like
to
go.
If
we
want
to
go
if
we
want
to
proceed
with
the
direction
that,
like
we've,
been
discussing
with
clayton,
which
I
mentioned
about
about
this,
like
re
kind
of
redesigning,
watch
cache,
I
think
it
may
be
possible
to
just
implement
pagination
there.
Basically
and
then
you
would
be
able
to
to
paginate
from
the
cache
directly.