►
From YouTube: NUG meeting Feb 2021
Description
Video from our monthly NUG meeting on Feb 18, 2021. Topic of the day was "Making the most of Slurm at NERSC" with Shahzeb Siddiqui from NERSC User Engagement Group
A
All
right,
so
people
who
have
been
to
a
few
of
these
now
will
be
familiar
with
this.
Our
plan
is
for
quite
an
interactive
meeting.
We've
got
somewhat
more
than
50
people,
which
potentially
could
get
yeah
a
little
bit
noisy,
but
you
know
that's:
okay,
we'll
we'll
start
off
reasonably
informally
and
if,
if
if
we
need
to
we'll
go
to
a
slightly
more
formal
q,
a
part
but
really
the
key
point
here
is
this
is
not
meant
to
be
just
a
presentation.
This
is
a
interactive
meeting.
A
It's
a
forum,
an
opportunity
to
share
things.
Please
participate
yeah,
unmute
yourself
and
speak
when
you
have
something
to
say
if
it
starts
getting
noisy,
we'll
ask
people
to
raise
their
hand
first
yeah,
but
otherwise,
so
like
that,
if
you
haven't
already
seen,
we
have
a
nurse
user
slack
and
there
is
a
hashtag
webinars
channel
there.
That's
another
good
place
to
ask
questions
and
continue
the
conversation
one
of
the
yeah,
the
nice
things
about.
That
is
that
the
chatter
there
result
is
retained
beyond
the
end
of
the
zoom
meeting.
A
The
slides,
in
fact,
are
already
up
on
the
web
page
associated
with
this
meeting,
which
is
under
www.nurse.gov
under
events.
You
can
use
those
slides
to
find
the
the
link
and,
in
fact,
if
you
haven't
already
got
this
slack
link,
you
can
post
it
in
the
chat
here.
A
A
So
our
agenda
follows:
follows
our
normal
agenda
for
these
meetings,
we'll
start
out
with
a
win
of
the
month,
which
is
an
opportunity
to
tell
about
success
stories.
A
Today
I
learned,
which
is
an
opportunity
to
talk
about
things
that
didn't
go
as
smoothly,
but
that,
maybe
you
know
each
other
can
learn
something
from
we'll
have
a
section
for
announcements
and
calls
for
participation
and
then
we'll
go
into
our
topic
of
the
day,
which
is
around
slurm
and
shazam
siddiqui
from
nest's
user
engagement
group
will
give
us
some
tips
on
how
to
work
with
slurm
to
get
you
know
the
the
most
outcome
for
the
least
wall
clock
or
the
the
least
curating
time
and
the
least
cost.
A
So,
let's
start
out
with
win
of
the
month,
so
the
purpose
of
this
segment
is
to
show
off
an
achievement
or
shout
out
somebody
else's
achievement,
and
it
doesn't
have
to
be
big.
It
can
be
a
fairly
big
thing
like
having
a
paper
accepted
or
a
relatively
small
thing,
like
solving
a
bug
that
had
you
know,
kept
you
stumped
for
a
few
days.
A
This
is
also
a
good
opportunity
to
tell
about
some
scientific
achievements.
We
know
that
nurse
users
are
doing
some.
You
know
pretty
amazing
things.
We
don't
always
know
the
details
of
what
those
things
are,
but
we
do
really
like
hearing
about
it
and
we
think
it's
inspiring,
for
you
know
us
and
each
other
as
well
in
particular
nurse
hands
out
some
awards
every
year
for
high
impact
scientific
achievement
and
for
innovative
use
of
high
performance
computing,
and
you
know
maybe
you're
doing
some
work.
A
That
is
a
candidate
for
those
awards,
and
you
know
we'd
love
to
hear
about
here.
On
that
note,
actually,
the
early
career
nominations
for
those
awards
are
due
this
week.
We'll
have
an
announcement
about
that
fairly
shortly,
but
just
so
we're
we're
especially
keen
to
hear
about
that.
We've
got
one
more
day
before
the
deadline
for
those
so
yeah
any
high
impact,
scientific
work
or
innovative
uses
of
high
performance
computing
that
you
either
are
doing
or
know
of
you
know
please
nominate
by
the
end
of
this
week.
A
Oh
something
I
forgot
to
mention
earlier
is
we
are
recording
this
session
and
the
video
will
be
available
on
the
website
afterwards.
A
So
maybe
people
are
a
little
shy.
Oh
I'll
kick
one
off.
I
had
something
that
I
I
was
pretty
pleased
with
how
it
came
out
working
with
one
of
our
users,
who
is
who
has
a
high
throughput
computing
type
workflow,
which
and
and
this
particular
workflow
has
an
extra
challenge
in
the
individual
tasks,
mpi
tasks
and
multiple
mpi
tasks
across
you
know,
per
node
mpi
task
across
many.
Many
nodes
are
a
little
bit
tricky
because
most
of
the
workflow
tools
use
the
mpi
infrastructure
to
manage
the
workflow
and
so
running.
A
An
mpi
task
inside
of
an
mpi
job,
essentially
inside
of
a
larger
workflow,
has
a
few
challenges
and
there
is
kind
of
a
known.
You
know,
at
least
in
principle,
solution
which
is
to
use
shifter
containers
so
create
a
docker
container
on
on
the
laptop
and
set
up
a
the
mpi
workflow
to
happen.
Sort
of
you
know
locally
to
a
node
within
that
container
and
then
use
shifter
to
run
that
container
in
parallel
multiple
times
across
many
nodes
of
corey,
and
you
know
it
took
a
it
took
a
few.
A
A
But
you
know,
after
a
a
a
couple
of
days
of
sort
of
yeah
backwards
and
forwards,
we
yeah
we're
able
to
make
a
a
working
workflow
of
it
and
that
that
will
make
its
way
onto
a
or
a
variation
of
that
will
make
its
way
as
an
example
on
our
documentation
pages
in
the
in
the
not
too
distant
future.
But
you
know
I
was.
I
was
pretty
chuffed
about
that
yeah
anybody
got
any
other
stories
they'd
like
to.
A
C
We
have
one
not
kind
of
wind
of
the
month
but
more
like
ireland
stuff.
That
might
be.
You
know
interesting
to
share,
so
I
was
running
I'll,
try
to
run
globally
high
resolution
simulations
in
the
context
of
climate
simulations,
but
again
io
has
been
taking
quite
a
lot
of
chunks
with
time,
but
I
just
realized
so
again
again
concerning
the
file
size,
ranging
from
the
tens
of
gigabyte
200
gigabyte
as
a
model
output.
C
I
didn't
realize
that
I
have
not
set
properly
the
the
last
file
stripping
striping
before
yeah,
for
example,
in
the
end
of
the
simulations,
the
model
write,
restart
file.
That's
you
know
more
than
100
gigabytes,
but
it
taking
in
the
worst
time
about
two
hours
to
just
write
this
file.
But
then,
after
I
you
know
set
the
five
striping
to
the
the
medium.
I
believe
then.
Actually
it
reduced
like
10
minutes
or
maybe
20
minutes.
C
D
A
C
C
Improvement
exactly
and
then
also
it's
nice
thing
is
once
I
set
this
property
for
the
directory,
then
all
the
files
underneath
you
know
written
after
that.
It's
it's
inherit
that
property.
So
all
the
output
files
is,
it's
naturally
striped
medium
size
by
just
sitting
at
the
top
directory.
So
that's
quite
handy.
Yes,.
A
Yeah
good
work,
nice,
nice
outcome.
C
Yeah,
and
maybe
another
one
that
I
might
want
to
share
is
that
so
this
exercise
was
part
of
my
preparation
for
running
production.
High
resolution
simulation,
so
we
are
testing
different
number
of
nodes
or
different
io
settings
or
compile
optimizations,
and
one
thing
I
so
I
was
in
the
end
of
the
exercise.
I
was
calculating
okay,
so
at
the
end
of
one
year,
our
own
one
year
how
many
simulation
years
can
we
get?
So
that's
the
you
know,
purpose
of
this
exercise,
so
that's
I
just
realized
doing
that.
C
Not
just
increasing
number
of
nodes
to
get
some
improvement
in
the
model.
Throughput
does
not
necessarily
leads
to
much
many
more
simulation
years
to
be
done
in
one
year,
because
let
me
can
I
share
one
like
a
graph
file
on
the
on
the
chat.
I
need
a
google
spreadsheet
to
schedule.
Statistics
for
the
average
q
wait
time
for
corey
knight
landing
on
the
past
one
year
and
were
you.
C
Actually,
I
shared
on
the
chat,
but
I
don't
know
if
everyone
can
open
up.
Maybe
is
it
in
the
in
the.
A
A
C
A
C
C
E
C
C
That's
yeah,
that's
one
reason,
but
also
you
know
I
found
about
20
to
30
percent
of
increase
or
decrease
in
estimation
time
going
from
like
100
north
to
200,
but
that
improvement
actually
very
overwhelmed
by
the
wait
time
difference
because
wait
time
difference
is
more
than
50
percent.
C
So
once
we
I
take
into
consideration
those
things
to
get
how
many
years
of
simulation
we
can
get
in
in
a
given
year.
Actually
100
nose
is
better
than
using.
200
knows
I
have.
I
have
to
go
up
to
actually
1020
photos
or
more
to
actually
to
to
get
a
better
annual
simulation
throughput.
A
Learn
to
you
know,
minimize
queue,
wait
time
and
and
find
optimal
points,
and
that's
a
good
consideration
that
looking
for
the
sweet
spot,
including
the
q,
wait
time
yeah
the
size
and
shape
of
your
job
is
a
worthwhile
exercise.
C
E
A
A
Any
other
wins
or
stories
to
share.
F
So
can
I
actually
ask
a
quick
question
to
quiche
because
I
had
sort
of
this
fred
avantos
for
what
we
are
doing-
and
I
was
curious
in
in
your
table-
is:
is
this
wait
time?
F
So
let
me
try
to
understand,
is
the
number
of
node
is
the
wait
time
the
average
wait
time
per
node
or
or
certain
your
job,
the
request
and
and
nodes.
C
Now
this
is
a
wait
time.
So,
if
you
use
a
request,
you
know
knows
between
64
127
and
it
depends
on
the
month,
but
you
on
average
you,
for
example,
if
it's
in
2020
january,
you
have
to
wait
almost
50
hours,
an.
F
C
C
A
Okay,
thank
you.
So
so
that's
some
very
interesting
tips
and
we're
very
much
getting
that.
That's
that's
very
much
in
line
with
our
topic
of
the
day.
So
what
do
I
do?
Is
I
pause
this
conversation
until
we
get
up
to
that
section
and
I
think
we'll
have
an
opportunity
to
to
go
into
some
more
depth
in
detail
and
and
shows
I've
just
got
some
some
background
information
about
why
this
is
the
case
as
well.
A
Okay,
great
so
time
is
actually
getting
along,
so
we
might
move
on
to
our
next
section
which,
which
we've
sort
of
started
covering
already
here,
which
is
kind
of
the
the
flip
side
of
win
of
the
month,
which
is
today,
I
learned
and
the
the
thinking
behind
this
is
that
you
know
research
happens
by
experimenting
and
you
know
getting
things
wrong
a
few
times
until
we
get
them
right
and
yeah.
A
I
think
we
can
learn
from
not
only
our
own
sort
of
yeah
points
that
we
got
stuck
in
and
tripped
up,
but
the
the
things
that
were
challenging
for
other
people,
but
this
doesn't
have
to
be
things
that
didn't
work.
That's
also
an
opportunity
to
call
out
you
know:
resources
you've
stumbled
across,
for
instance,
or
discovered
recently
that
might
be
valuable
to
other
nurse.
A
Users
so
coach's
tip
just
before
about
you
know
the
the
difference
that
setting
striping
made
to
his
code
is
a
good
example
here.
A
A
Well,
we've
had
a
a
few
discussions,
I
think
of
things
that
we
learned,
so
we
can
step
on
to
the
next
item.
So
our
next
segment
is
about
announcements
and
calls
for
participation.
We
have
a
a
couple
of
announcements
from
nurse
side,
but
this
is
also
yeah.
This
is
a
general
user
forum
and
we're
also
keen
to
hear
about
you
know,
conferences,
events
and
so
on
that
you
know
of
or
perhaps
organizing
or
particular
you're
contributing
to
that
would
be
valuable
to
other
nurse
users.
A
From
from
nurse
side,
you
hopefully
saw
this
already
in
your
weekly
email,
etc,
but
the
nurse
early
career
nominations
for
hpc
achievement
awards
due
at
the
end
of
this
week,
that
is
to
say,
bye
tomorrow.
A
So
we
have
two
categories
for
these
awards.
One
is
for
high
impact
scientific
achievement
which
recognizes
work
that
has
or
is
expected
to
have.
You
know
an
exceptional
impact
on
scientific
understanding,
engineering,
design
for
scientific
facilities
or
a
broader
societal
problem.
A
Yeah
we've
got
a,
I
guess.
Yeah
society
has
a
fairly
big
invisible
target
in
the
pandemic
at
the
moment.
For
that,
the
other
award
is
for
innovative
uses
of
high
performance
computing,
which
recognizes
research,
researchers
who
have
used
nurse
resources
in
innovative
ways
to
solve
a
significant
problem
or
have
come
up
with
a
new
methodology
that
might
have
a
large
scientific
impact.
A
So
this
can
include
things
like
you
know,
using
hpc
in
a
field
where
it
hasn't
previously
been
used
or
used
terribly
much
or
combining
computing
data,
networking
and
edge
services
to
do
something
new.
You
know
you
know
in
a
domain
where
hpc
is
already
in
use,
so
for
the
early
career
eligibility.
This
is
I
take
it
particularly
at
users
using
nurse
resources
who
early
in
their
career,
so
so
postdocs
or
or
nurse
users
who
have
received
their
degree.
A
So
you
can
nominate
these
people
by.
I
think,
if
you're
checking
the
weekly
email,
there
is
probably
a
link
or
you
can
send
us
a
ticket
with
the
topic
being
nomination
for
award
and
yeah.
Give
us
a
point
of
to
who
you
know
of
that.
You
would
be
a
a
good
recipient
for
one
of
these
awards.
A
So
I
think
that's
the
only
announcement
we
have
at
the
moment
on
nurse's
side.
Apart
from
the
there
are
a
few
other
announcements
actually
of
the
things
going
on
in
there
that
are
in
the
weekly
email.
A
Do
does
our
user
community
here
have
any
announcements
or
events
that
would
be
good
to
know
about.
A
A
Organization,
so
if
we
don't
have
anything
else
in
the
way
of
announcements,
we
can
go
on
to
our
topic
of
the
day
and
I'd
like
to
introduce
shazam,
siddiqui
who's
part
of
our
user
engagement
group
here
and
has
a
lot
of
experience
with
sloan
and
is
going
to
walk
us
through
some
of
the
details
of.
I
guess
your.
A
What
slum
is
doing
underneath
and
you
know
how
it
works
at
nurse
and
how
you
can
use
that
information
to
get
more
out
of
at
a
slum
and
spend
you
know
less
time
in
the
queue
if
you'd
like
to
just
say
next,
I
can
click
the
slide
through
to
the
next
one,
when
you're
ready.
G
Thanks
yeah,
you
can
start
okay,
so
yeah.
This
is
going
to
be
a
quick
recap
on
you
know,
for
you
guys
to
get
up
to
speed
on
slurm
and
get
some
best
practices
using
our
cluster.
So
you
know,
as
you
may
log
you
know,
use
our
cluster.
You
typically
log
in
to
our
nodes.
You
know
please
note
that
these
login
nodes
they're
shared
with
many
users
right,
so
these
are
not
meant
for
computational
resource.
G
What
you
would
typically
do
is,
you
know,
specify
your
s
patch
directives
and
a
and
you
know,
replace
the
number
of
nodes
and
you
know,
run
your
script.
What
what
happens
in
the
back
end
is
your
script.
Actually,
you
know
is
processed
by
this
alarm
server.
It
processes
the
script
and
then
finds
the
allocated
nodes,
and
you
know
fulfills
your
request.
G
So
on
corey
we
have
like
several
login
nodes.
Those
are,
you
know,
those
icons
in
blue
and
sometimes
some
of
our
login
nodes
are
also
used
as
compute
resources
right
when
a
job
gets
allocated,
you
may
you
may
get
allocated
to
one
node,
which
will
be
this
yellow
icon.
If
you
have
multiple
nodes,
then
you
know,
slurm
will
do
that
for
you.
G
That
could
be
those
red
icons
and
you
know,
typically
at
the
end
of
the
job,
you
will
get
an
output
or
error
file
on
the
saved
on
the
disk
right
and
just
one
thing
to
also
notice.
You
know
we
have
cues
that
allow
us
to
submit
exclusive
node.
So
you
can.
You
can
do
that,
but
we
have
like,
for
instance,
a
shared
queue
which
is
shared
with
multiple
users.
G
Okay,
so
when
you
submit
a
job
either
through
s
patch
or
yeah,
basically
what
happens
is
the
job
will
go
through
several
job
states?
The
first
is
the
you
know,
append
and
salaam
will
try
to
figure
out.
You
know
when
it
needs
to
get
the
resources,
so
it
will
be
in
this
state.
It
will
once
it's
done
from
pending.
You
will
configure
the
nodes
that
are
required
to
run
the
job,
and
then
your
job
will
actually
be
in
a
running
state.
G
G
And
finally,
if
the
job
either
completes
successfully
will
be
completed
or
most
of
the
other
time,
you
would
see
it
failed
if
it's
the
fail,
if
you
run
out
of,
for
instance,
like
timeout
or
something
like
that,
it
would
also
show
up
as
failed,
but
if
you
cancel
a
job,
you
know
that
will
show
up
now,
as
during
the
process
of
this
job
lifetime,
you
can
actually
monitor
this
job
through
either
sqrs
control.
G
One
thing
to
note:
is
you
use
sq
and
s
control
for
the
lifetime
of
the
job
from
pending
to
completed,
but
you
can
use
sacct
to
query
historical
jobs.
Right,
didn't
know
that
you
know
using
these
commands
frequently
will
ping
our
slurm
server
so
doing
too
much
request.
You
know
on
like
a
loop
can
actually
impact
a
server,
so
it's
it's
generally
prohibited
right.
G
G
G
The
idea
is
that
backfilling
allows
us
to
increase
throughput
of
the
of
our
jobs,
and
you
know
backfilling
what
what
what
it,
what
the
intent
is
to
allow
jobs
that
are
like
lower
priority,
like
short
running
jobs,
to
be
filled
in
wherever
there
is
gaps
into
the
cluster.
G
Typically
short
running
jobs
are,
like
you
know,
smaller,
like
really
small
boxes
like
either
low
run
time
or
a
low
number
of
nodes.
If
you
see
like
in
this
diagram,
you
should
look
at
it
from
the
top
down,
so
jobs
that
are
submitted
long
time
ago.
G
Long
rectangulars
are,
like
you
know,
small
jobs
with
long
running
time
or,
like
you
know,
horizontal
rectangulars
are
like
like
accessing
the
full
node,
like
you
know,
thousand
a
thousand
nodes,
but
like
it's
a
very
short
running
job.
These
are
generally
hard
for.
G
This
is
bad
for
scheduling,
so
this
alarm
will
try
to
accommodate
this
and
it
will
put
it
into
the
scheduler,
but
what
backfilling
does
it
allows
for
short
and
small
running
jobs
to
be
run
earlier
in
the
schedule
like,
for
instance,
the
blue
and
the
purple
one
they
get
submitted
earlier,
so
that
we
improve
the
throughput,
and
you
know
without
backfilling
it
would
just
generally
just
do
first
in
first
serve
and
that's
not
going
to
be
optimal.
G
Is
you
know
if
you,
if
you
have
a
job,
you
don't
need
to
guess
the
wall
clock
time
you
can
say
something
like
the
minimum
amount
of
time
that
you
need
like,
for
instance,
if
you
have
a
job
that
runs,
let's
say
six
hours,
but
you
want
to
make
it
this
time
so
that
sloan
will
pick
it
up
more
quickly.
You
can
say,
let's
say
a
time
limit
of
like
let's
say
three
hours
in
this
example.
G
You
may
have
a
job
that
you
don't
know
how
much
time
it
takes,
but
you
expect
that
three
hours
is
sufficient
with
min
mean
time,
sloan
will
actually
look
at
it
and
say:
okay,
I
have
a
three
three
hour
time
limit
and
I
will
actually
fit
that
in
in
this
diagram.
G
If
you
see
the
whole
box
is
actually
blue,
that's
six
hours,
but
the
dark
blue
could
be
the
the
three
hours
and
then
it
will
just
try
to
fit
it
in
to
the
scheduler.
One
thing
to
note
is
that
some
of
our
q
policies
require
a
min
time.
For
instance,
flex
skew
is
one
of
them,
so
it's
just
one
thing
to
be
noted.
A
A
G
Yeah,
I'm
just
looking
exactly
so
according
to
what
it
does.
It's
set
the
minimum
limit
on
the
job
allocation.
So
the
time
is,
it
allows
to
be
executed
earlier.
G
It
doesn't
change
the
the
actual
job
time
itself,
so
it
improves
the
back.
The
backfilling
scheduling
algorithm
so
like
allows
scheduler
to
see
that
you
know
this
job,
which
is
supposed
to
be,
let's
say
six
hours.
It
doesn't
need
all
that
time
just
needs.
Let's
say
the
amount
up
to
min
time,
which
may
be
like
two
or
three
hours
and
it
will
schedule
a
job
ahead.
K
A
So
the
the
time
gets
adjusted
at
the
time
that
the
job
starts.
So
if
there's
a
gap
in
the
schedule
of
say
three
and
a
half
hours
and
you
have
a
min
time
of
two
hours
in
a
six
hour
job,
it
will
adjust
the
time
of
your
job
to
three
and
a
half
hours
to
exactly
fill
that
gap.
B
A
It
it
adjusts
the
time
at
the
time
that
the
job
starts.
So
it
has
a
schedule,
for
instance,
that
it
knows
that
this
wide
job
is
expected
to.
You
know
have
nodes
available
to
it
to
start
say
in
in
three
and
a
half
hours
from
now.
A
A
Okay,
sorry,
I
kind
of
interrupted
there.
Would
you
like
to
continue
this
okay.
G
Next
slide,
okay,
so
in
terms
of
what
we
have
and
the
cues
that
we
have
typically
you
will.
You
will
use
a
regular
clue
for
most
of
your
workload
right
and
you
know
so
in
terms
of
the
the
cost.
You
know
regular,
crucial
piece
of
food
it.
I
think
I
believe
it
has.
I
think
48
hours
is
a
time
limit,
so
it
should
be
sufficient
for
most
of
your
workload.
G
If
you
need,
for
instance,
some
emergency
workload,
premium
queue
is,
is
the
way
to
go,
it
will
submit,
it
will
schedule
the
job
much
faster,
but
you
know
it.
It
is
more
expensive
right.
Premium
queue
is
also
special,
so
yeah
q
in
in
the
sense
that
not
all
user
automatic
access
to
it,
the
your
pi
will
have
to
grant
you
access
to
this,
and
also
the
charge
factor
gets
changed
once
you
reach.
I
believe
two
percent
is
our
is
the
rate
so
just
be
mindful
of
that.
G
Only
use
this
queue
when
you
really
need
to
debug
is
a
really
good
queue
if
you
want
to
just
submit
a
job
just
for
debugging
purpose,
so
I
think
it's
got
a
30
minute
time.
So
it's
it's.
G
You
get
a
good
turnaround
time.
If
you
need
interactive
access
to
a
node,
then
use
the
s
the
interactive
queue,
so
one
thing
to
know
is
you
need
to
use
sl
up
s
patch
is
not
going
to
work.
G
A
log
qos:
this
is
good.
If
you
have
workload,
that
really
is
not
that
important.
You
don't
mind
having
a
long
kind
of
wait
time,
so
you
you
can
use
that
and
it's
also
very
cheap.
The
flex
q
is
is
good.
If
you
support
like
flexible
wall
time,
this
queue
requires
you
to
have
the
min
time
or
time
main
option
set
to
at
least
two
hours.
It
won't
accept
the
job.
If
you
don't
set
this,
so
it
has
to
be
less
than
two
hours
and
it's
only
available
on
k.
G
L-
and
this
is
good
for
if
you
have
your
job
supports
like
restarts
or
if
you
wanted
to
be
able
to
start
if
the
job,
for
instance
gets
killed,
if
you're
going
to
use
a
shared
queue,
this
is
shared
with
other
users.
So
when
you,
when
you
submit
a
job,
you
get
into
a
node,
keep
in
mind
that
you
will
be
sharing
this
note
with
other
users.
G
So
this
is
good
if
you,
if
you
just
want
to
get
your
jobs
done,
but
you
don't
care
about
in
the
case
of
like
performance,
you
don't
need
exclusive
node.
If
you
do
need
an
exclusive
node,
then
you
should
use
something
else
or
just
use
the
exclusive
option:
the
overrun
queue.
This
is.
This
is
good
if
you
have
a
zero
project,
account
balance
and
you
still
need
to
submit
jobs.
G
Otherwise
you
can't
use
this
queue
and
the
real
view
is
only
for
a
special
purpose.
If
you
need
real
like
immediate
access,
so
yep
and
you
can
take
the
take
a
look
at
the
link
below
for
the
qos
next
slide.
G
Okay,
so,
as
I
mentioned,
you
know
you
most
of
your
workload
should
be
going
through
the
regular
queue.
You
know.
We
also
have
the
premium
queue.
That's
for
you
know,
like
you,
know,
immediate
kind
of
needs
like
if
you
have
some
kind
of
conference
that
you
need
to
submit
something
and
only
have
like
a
week
or
so
you
know
you
could
use
the
premium
queue
or
so
for
ins.
One
thing
to
note
is
the
flat
skill.
G
You
know
it
is
we
we
do
discount,
I
think
75
on
and
it's
only
on
knl
right
and
you
know
because
it's
on
node
you,
you
can
actually
pretty
much
on
the
node
itself.
So
that's
pretty
good
and
yeah.
G
The
large
kit,
like
if
you're
submitting
like
large
jobs
on
known
nodes,
then
we
also
do
discount
like
up
to
10
20
foods
with
a
50
percent
discount
so
that
it's
good
to
know
if,
if
you,
if
you
want
to
submit
like
a
larger
workload
and
and
this
is
available
on
the
regular
regular
queue,
okay
and
yep
so
just
to
summarize,
you
know,
I
think
one
thing.
One
thing
that
you
may
learn
is
you
know
you
is
the
scheduling
them.
Do
we
use
backfilling?
G
So
if
you
have
short
running
jobs,
you
know
just
try
to
use,
let's
say
mint
or
time
mission
to
get
your
jobs
too
quickly.
You
know
we
support
several
cues.
So
pick
the
right.
Cue
and
that's
you
know,
use
a
flex
queue
when
you
need
it
will
save
you
money
for
sure
users
can
if
you're
gonna
submit
large
number
of
jobs
up
to
1024.
A
Thanks
joseph
so
so
we've
got
about
five
minutes
or
so
for
q
a
and
I
see
actually
canoe
priya,
I'm
not
sure.
If
I'm
pronouncing
it
correctly
has
a
has
asked
a
question
in
the
chat.
If
you
ask
for
four
nodes
using
the
interactive
queue,
can
you
do
this
split
between
two
different
jobs?
I
think
I
want
to
clarify
that
question.
Do
you
mean
running
different
s
runs
together
or
do
you
mean
request
two
jobs
of
two
nodes?
Each.
A
Yes,
you
can
do
that.
We
have
in
the
examples
page
of
our
docs,
which
is,
I
might
put
a
put
a
link
to
it
in
there
in
a
chat
in
a
moment
you
can-
and
you
can
do
this
interactively
as
well
start
an
s
run
in
the
background
where,
for
one
of
the
jobs
and
then
start
the
other
one,
what
you'll
probably
want
to
do
is
actually
I
I
assume
that
you're
looking
at,
for
instance,
comparative
debugging,
that's
correct
yep
as
the
use
case.
A
A
A
Are
there
any
other
questions
that
people
would
like
to
ask.
B
So
this
is
kind
of
related
to
what
the
earlier
speaker
was
talking
about.
You
know
if
you
have
something
that
you
expect,
can
you
backfill
reasonably?
Well,
so
the
you
know,
total
number
of
core
hours
for
your
job
is
is
modest
right,
but
you
know,
essentially
you
know
if
you
can
divide
the
job
up
in
a
way
such
that
you
know,
you
essentially
have
a
rectangle
in
you
know
time
node
space
right.
B
B
A
Yeah,
as
a
general
rule
short
time
is
much
easier
to
to
find
gaps
for
even
even
quite
wide.
Gaps
are
relatively
easy
to
find
compared
to
long
gaps.
A
A
B
D
D
So
just
digging
up
okay.
A
So
I'm
just
posting
now
in
the
webinars
channel,
one
of
our
docs
page
is,
is
example
job
scripts,
and
this
is
a
great
resource
actually
well.
We
hope
it's
a
great
resource
for
examples
of
lots
of
different
use.
Cases
which
includes
multiple
simultaneous
jobs
for
I'll
need
to
dig
a
little
further
to
find
the
link
specifically
for
ssh
into
a
node.
A
A
So
there
are
a
couple
of
ways
of
doing
this:
you
can
use
job
arrays,
which
use
a
single
job
script,
to
manage
a
whole
series
of
jobs
with
a
single
s
batch
or
you
can
ask
batch
a
whole
lot
of
individual
jobs.
M
A
The
request
kind
of
acts
as
a
single
job
and
then
basically
the
the
individual
jobs,
get
kind
of
pulled
off
the
request.
So
my
understanding-
and
maybe
shazeed,
has
more
information
about
this.
Is
that
the
request
will
age.
So
so
you
know
the
entire
request
will
reach.
You
know
a
kind
of
you
know
the
same
priority,
and
then
you
know
when
slums
scanning
it
it
will
pull
off.
You
know,
however
many
jobs.
It
can
start
at
the
moment,
or
at
least
the
next
job
that
it
can
start
at
the
moment.
G
Yes,
so
one
thing
that
that
we
didn't
discuss
today
was
the
how
house
learn:
does
job
priority
so
currently
we
we
use
this
learn
feature.
It's
called
multi-factor
priority
when
salary
actually
figures
out
which
job
actually
needs
to
be
scheduled.
It
does
it
based
on
priority
and
there
are
multiple
factors.
G
What
goes
into
job
priority,
and
one
of
the
factor
is
age,
the
the
age
of
the
job
length,
right
as
it
sits
in
the
queue
and
then
also
when
it's
actually
running,
as
you
may
know,
if
it's
a
longer,
if
it's
a
job,
that's
waiting
in
the
cube
for
a
longer
time,
then
the
age
factor
will
grow.
That
means
that
slurm
is
trying
to
figure
out
that
this
job
is
in
the
queue
for
a
long
time.
It
needs
to
get
scheduled
right
same
thing.
G
If
the
job
is
running
for
a
longer
time,
then
that's
not
good
too.
You
can
take
a
look
at
the
link.
I
I
think
one
of
the
commands
that
you
could
use
to
actually
get
job.
Priority
is
espio,
but
we
haven't
documented
that,
but
there
is
some
useful
things
that
you
can
get
out
of
it.
G
If
you
think
that
would
be
useful,
I
think
we
can.
We
can
try
to
put
some
more
documentation
into
that.
One
thing
to
note
is
that
espio
works
only
for
pending
jobs.
So
if
you
have
a
job
that's
pending
and
you
want
to
know
the
priority
of
pending
jobs,
then
that
could
help
you
can
try
to
see
your
job
and
then
also
sorted
by
the
queues
and
it.
F
G
G
A
Yeah,
I
was
gonna,
say
yeah,
so
espresso
gives
you
the
breakdown
of
what
are
the
components
of
your
job's
priority.
I
think
the
the
total
number
can
actually
be
found
in
sq
as
well.
A
Yes,
so
we
have
a
couple.
More
questions
have
come
up.
I
might
address
them
in
reverse
order,
because
of
basically
how
how
easy
easily
answered
they
are
so
ronnie
asks.
Do
specific
projects
have
different
priority
and
the
answer
is,
for
the
most
part,
no
different
projects
have
different
allocations,
which
you
know,
I
guess,
influences
how
many
jobs
they
can
run
over
the
entire
year.
A
But
priority
is
all
the
same.
There
is
kind
of
an
exception
in
that
projects
can
request
access
to
the
real
time
queue
and
that's
generally,
for
sort
of
special
cases
such
as
you
know,
needing
needing
to
synchronize
with
time
on
a
on
an
instrument
somewhere.
For
instance,
you
know
super
facility
type
work
and
real-time
queue.
Jobs
have
super
high
priority,
but
that's
kind
of
a
special
case
in
the
in
the
normal
course
of
things.
A
You
know
one
project
and
another
project
have
the
same
priority
in
the
queue
I
see
says:
joseph's
posted
an
example
of
what
the
esprit
output
looks
like.
A
When
my
pronunciation
correctly
asks,
can
we
explain
a
bit
about
the
denial
of
service
problem?
A
I
guess
you
mean
about
too
many.
Too
many
s
runs
or
too
many
sqs
requests.
A
Yes,
sir
yeah,
so
this
is
this
is
because
corey
is
quite
a
large
system
with
quite
a
lot
of
users,
but
the
the
scheduling,
I
guess,
is
a
centralized
service,
and
so
you
know
the
the
slum
daemon
basically
has
to
you
know:
answer
requests
as
well
as
do
the
scheduling
and
if
it
gets
too
many
requests,
you
know
it's,
it's
multi-threaded,
it
scales
reasonably
well,
but
it's
it's
not
omnipotent
and
it
is
entirely
possible
to
overwhelm
it
with
requests
and
they
your
one
easy
way
to
do.
A
That
is
by
calling
you
know,
sqs
or
sram
in
a
loop,
particularly
with
no
no
sort
of
a
sleep
between
them.
Yeah,
the
the
demon
will
just
sort
of
get
so
many
requests
coming
from
so
many
different
directions
that
it
starts
to
drown
under
them
a
little
bit.
So
that's
why
we
ask
users
not
to
do
that.
A
Certainly,
never,
I
think
deliberately
it's
mostly
by
mostly
when
this
happens
it's
by
users
who
have
a
a
genuine
use
case.
You
know,
I
need
to
start
a
thousand
things
in
my
high
throughput
workflow
and
they
do
the
kind
of
obvious
standing
answer
or
make
a
loop
that
runs
s
run.
A
Which,
unfortunately,
has
this
side
effect
of
hitting
the
scheduler
very
very
hard?
So
as
far
as
I'm
aware,
I
I
don't
know
of
any
malicious
attacks
like
this,
but
it
has
definitely
accidentally
happened.
A
And
oftentimes,
if
you
see
slow
response
from
slurm,
if
you
yo,
if
you
type
sqs
and
it
seems
to
take
ages
to
come
back,
it's
probably
because
slum
is
dealing
with
a
lot
of
requests.
A
There
was
another
question
further
up:
oh
for
jobs
that
finish
within
two
hours
on
k.
L.
Is
it
better
to
use
the
flex
cube?
So
that's
an
interesting
question.
If,
if
the
job
finishes
within
two
hours,
chances
are
it'll
run
fine
in
the
normal
queue,
you'll
get
started,
sort
of
straight
away.
A
A
You
know
you
could
also,
for
instance,
you
know
if
you've
got
a
job,
that
yeah
will
run
up
to
two
hours
but
can
run
one
hour
chunks.
You
could
use
the
flexq
for
that
as
well.
A
We're
actually
getting
close
to
the
top
of
the
hour,
and
I
think
we've
covered
the
questions
that
have
come
in
in
the
chat.
We
might
be
able
to
continue
with
a
q
a
session
sort
of
at
the
at
the
end
of
the
meeting.
For
those
who
are
still,
you
know
around
and
available,
but
what
we
might
do
now
is
flip
through
the
last
couple
of
items
in
our
agenda
so
that
we
can
finish
the
sort
of
formal
part
of
the
meeting
before
12
o'clock,
so
the
next
one's
a
fairly
quick
and
easy
one.
A
What's
coming
up
next,
we
are
always
looking
for
topic,
requests,
suggestions
or
better
still
nominations
and
volunteers
to
host
a
topic
of
the
day.
This
is
a
a
great
opportunity
if
you
want
to
do
a
kind
of
a
relatively
short.
You
know,
lightning
talk,
kind
of
level,
overview
of
some
interesting
work
that
you're
doing
using
nurse
resources
and
some
tips
that
you've
learned
that
might
help
other
users.
You
know
we'd
love
to
hear
about
it.
A
Last
month's
numbers,
so
our
availability
was,
was
pretty
high.
In
january
we
had,
you
can
see
this
black
era
was
the
scheduled
maintenance
that
happened
during
the
allocation
year
transition.
A
We
did
have
a
very
short
schedule,
unscheduled
outage
later
in
the
month
other
than
that
corey's
scheduled
availability
was
pretty
close
to
100
and
the
storage
systems
scheduled
availability
was
at
100.
That
was
good
news.
Corey's
utilization
was
nicely
high.
Things
like
the
flex.
A
Q
are
actually
really
helping
us
here,
because
you
know
an
increasing
number
of
users
are
making
their
jobs
more
easily
fit
into
gaps,
we're
not
getting
very
much
time
of
empty
nodes,
sitting
idle
because
they're
waiting
for
other
nodes
to
be
unavailable
for
some
job,
we're
able
to
make
good
use
of
them.
A
So
that's
that's
great
to
see
we
have
a
target
of
25
percent
of
the
workload
on
corey
being
jobs
that
need
a
system
of
corey
scale,
basically
large
jobs,
things
needing
more
than
more
than
a
thousand
nodes,
and
you
know
it's
good
to
see
that
corey
is
being
well
used
for
this
use
case.
So
we
have
a
25
target
and
over
40
of
our
workload
in
january
was
these
large
jobs.
A
Tickets
are
coming
in
at
a
slightly
faster
rate
last
month
than
than
what
we
closed
them.
So
we
have
a
current
backlog
as
of
a
couple
of
weeks
ago,
of
about
620
tickets,
and
that
is
all
of
the
formal
part
of
the
meeting.
If
people
are
interested
in
sticking
around
a
little
longer
to
chat
about
tips
on
using
slurm,
I
think
I'm
available
for
a
little
longer.
I'm
not
sure
if
I'm
not
sure
what
she's
ed's
calendar
looks
like.
Are
you
available
to
stick
around
for
another
five
or
ten
minutes?
A
Okay,
so
it
might
be
only
only
a
few
minutes
then
and
then
we'll
probably
need
to
drop,
but
I
I
believe
we
have
a
few
other
nurse
people
on
the
line
as
well
so
between
us.
We
can
probably
answer
some
questions
and
I
wouldn't
be
surprised
if
some
of
our
users
online
are
also
quite
experienced
slum
users
and
can
participate
in
that
contribute.
A
Answers
other
than
that.
Yes,
thank
you
all
for
joining
us,
we'll
post
the
recording
on
the
web
page
reasonably
soon
after
the
meeting
and
look
forward
to
seeing
you
all
again
next
month,.
E
A
A
So
as
people
are
exiting,
were
there
any
other
questions
that
people
had
about
making
making
the
best
use
of
slab.
C
Ask
a
follow-up
questions
about
this
flex:
queue
sure
for
the
shorter
jobs.
I
was
just
curious.
If
people
you
know,
run
shorter
jobs
like
snowy
hours
and
use
flex
q
in
not
intentional
way
just
to
get
the
cheaper.
You
know
charge
factor.
C
A
Don't
think
we
actually
have
a
mechanism
in
place
to
prevent
or
disallow
it
at
the
moment,
but
it
is
entirely
likely
that
that
would
happen
if,
if
we
do
start
to
see
it
being
used,
you
know
as
a
as
a
way
of
getting
a
discount.
C
A
Yes,
it's
that's
a
very
good
discount,
because
so
so
I
guess
perhaps
we
can
talk
a
little
bit
about
the
intent
here
and,
and
you
know
what
are
the
yeah?
A
What
are
the
the
the
needs
and
goals
that
we're
balancing
off
so
so
one
important
thing
to
keep
in
mind
is
that
for
the
most
part,
nurse
actually
doesn't
allocate
the
hours,
so
the
department
of
energy
program
managers
allocate
the
hours
for
for
projects,
and
so
the
hours
allocation
should
then,
at
least
in
theory
reflect
the
priority
that
the
doe
has
for
yeah
as
well
as
I
guess,
the
the
needs
in
that
you
know
for
for
different
research,
and
so
we
don't
want
to
kind
of
undermine
that.
A
You
know
if,
if,
if
the
program
managers
really
want
a
lot
of
the
time
to
be
spent
in
a
particular
area
of
research
or
a
lot
of
resources
to
be
dedicated
to
a
particular
area
of
research,
we
don't
want
to
risk
undermining
that
through
you
know,
discounts
and
and
and
so
on,
like
that,
on
the
other
hand,
we
don't
want
the
system
to
go
to
waste
so
yeah.
A
We
would
much
rather
have
jobs
that
can
fill
gaps,
fill
those
gaps
and
do
useful
research
yeah,
rather
than
letting
those
jobs
sit
idle,
and
so
this
is
what
drives
having
some
of
these
discounts
for,
for
things
like
the
flex
queue
is
that
by
encouraging
people
to
to
work
towards
making
their
jobs
flexible
and
to
create
jobs
that
work
well
with
the
scheduler
to
sort
of
you
know
optimize
the
experience
for
everybody,
you
know,
there's
there's
benefits
all
around,
and
so
that
is
a
you
know:
a
motivation
behind
putting
a
large
discount
there.
A
C
A
So
so,
if
your
purpose
is
that
you
have
yeah
the
the
flex
q,
will
you
know
it's
it's
filling
gaps,
but
it
is
a
a
lower
priority
than
others.
There's
also
a
low
priority
queue
which
is
actually
starts
slightly
higher
than
the
flex
queue,
but
has
a
50
discount.
A
So
if
you,
if
you
have
a
project
that
is,
you
know
not
rich
in
terms
of
nursing
hours,
but
the
workload
doesn't
have
high
urgency,
then
you
know
you
can
use
the
low
q
to
to
stretch
your
scholars
further.
E
Yes,
oh,
I
see
another
a
couple
of
questions.
A
A
So
this
is,
this
is
kind
of
a
challenge
and
the
the
best
and
preferred
solution
is
if
that
workload
can
be
made
to
support
stopping
and
starting
so
that
it
can
be
broken
up
into
smaller
chunks.
You
can
actually
use.
A
We
have
a
documentation
page
about
variable
time,
jobs
where
we've
got
some
example,
scripts,
pretty
much
for
setting
up
a
long
workload
to
use
the
flex
queue
to
break
it
up
into
chunks
and
then
resubmit
itself
that
you
know
when
it
reaches
the
end
of
one
time
chunk
to
sort
of
automatically
do
long-running
jobs.
But
that
does
require
the
job
to
be
capable
of
checkpointing.
A
A
Longer
running
jobs
make
it
more
difficult
for
scheduling
things
like
maintenance
or,
if
you
know,
if
the
system
becomes
unavailable,
there's
a
there's
a
higher.
I
guess
probability
of
of
interrupt
that
way.
A
A
Which,
for
you
know
for
jobs
that
don't
have
their
own
built-in
checkpoint
and
restart
capability
in
a
lot
of
cases,
even
for
mpi
jobs?
This
can
allow
the
job
to
be
stopped.
You
know
within
48
hours,
even
running
the
flex
queue
for
a
few
hours
at
a
time
and
automatically
re-chewed
rejuvened.
A
A
C
I
was
just
curious
one
more
question
about
you
know
the
table
I
made
out
of
this
webpage
too.
I
I
wonder
if
you
guys
have
any
idea
about
the
variability
within
each
group
or
each
column.
I
made
like,
for
example,
between
in
the
64
to
108
28
notes
among
that
group
like
how
much
variability.
A
E
Screen
with
that,
let's
share.
C
A
So
so
I
have
to
admit
this
is
not
something
I've
looked
into,
and
I
I
don't
know
if
anybody
else
said
nurse,
it's
entirely
possible.
Somebody
else
has
looked
a
little
more
than
more
into
it
than
what
I
have.
A
There
is
more
variability
than
I
expected
to
see
so,
which
it
might
just
imply
that
there
is
a
different
workload
at
different
times,
something
that's
a
fairly
easy.
I
guess
statistical
trap
to
trip
up
on
with
the
wait
time
charts
is.
A
There
is
a
second
chart
on
that
page
showing
the
number
of
jobs,
and
in
some
cases,
when
you
see
a
you
know
the
the
wait
time
might
be
particularly
high
or
particularly
no
low
for
a
certain
category.
But
then,
when
you
look
it's
because
during
the
period
there
was
only
a
very
small
number
of
jobs,
you
know
five
or
ten
jobs
that
requested
that
sort
of
yeah
that
particular
shape,
and
so
these,
statistically
probably
not
too
significant
jobs,
maybe
kind
of
adding
a
lot
of.
If
you
like
false
variability.
C
A
Okay
yeah,
so
this
is
oh,
and
this
is
particularly
by
a
number
of
nodes.
That's
a
that's
a
point
as
well
yeah.
Then
the
number
of
hours
will
add
sort
of
a
second
dimension
to
it.
C
Yeah,
but
I
didn't
yeah
that
I
don't
see
other
tables
in
this
particular
web
page
too,
that
the
you
know
numbers
as
a
function
of
hours
requested
only
graphically
it's.
It
is
represented.
I
believe.
A
Oh
as
in
as
in
getting
getting
the
underlying
table
behind
this
chart,
yeah
yeah
yeah
that
information
should
be
available,
but
I
don't
think
we
like,
specifically,
you
know,
publish
it.
So
it's
not
it's
not
that
it's
hidden.
It's
just
not
actually
actively
published
right.
A
C
A
A
Although
this
is
just
by
nodes
and
not
necessarily
by
time,
weighted
right,
if
I
sorry
wall
time
requested
so
yeah,
you
will
sometimes
find
that
yeah
here.
Some
of
these
you
know
in
this
square
there
was
only
six
jobs,
and
so
that
you
know
is,
is
contributing
to
you
know
one
of
these
boxes
here,
because
it's
not
very
many
jobs,
it's
not
statistically
fantastic.
A
So
it
is
something
to
have
in
mind
when
you're.
Looking
at
this
another
really
useful
one.
I
find
that
in
a
way
zooms
out
a
little
bit
and-
and
I
think
can
be
a
bit
helpful
because
of
that
is
this
backlog,
and
what
this
is
showing
is
the
total
amount
of
queued
work
in
terms
of
you
know
all
of
corey
days.
A
A
Hopefully,
if
it's
a
little
bit
short
it'll,
be
able
to
backfill
and
fill
and
and
start
a
lot
sooner
than
that,
you
will
notice
that
k.
L
jobs
have
a
a
much
lower
backlog,
they're
they're
in
the
kind
of
the
two-day
range
compared
to
10
days
for
haswell,
and
that
is
driven
partly
by
the
fact
that
we
have
five
times
as
many
k,
n
l
nodes
right.
A
A
C
C
A
Nodes,
okay,
I
see
got
it
so
so
the
numbers
for
corey
are
obviously
a
lot
bigger
because
there's
some
somewhere
in
excess
of
12
000
nodes,
and
so
you
know
was
it
there's
about
10
000,
k,
l
nodes,
so
there's
sort
of
you
know,
2.2
days
times,
10
000
nodes
worth
of
work
currently
queued
for
corey
knl.
C
D
I
I
E
N
So
sqs,
I
will
remove
that
column
now
before.
What
we
had
is
that
it
tells
you
it's
gonna
be
starting
three
days.
Actually,
it's
not
it's
going
to
start
in
three
days.
It's
it's
about
its
priority
is
going
to
be
reaching
the
threshold
that
scheduler
will
consider
it.
So
this
is
for
the
the
backfield
pass,
but
then
there's
also
when
you
see
it,
it
happens
immediately.
N
It's
because
it's
backfilled,
so
the
column
tells
you
it's
going
to
reach
the
threshold
to
for
the
scheduler
to
consider
its
regular
path
in
how
many
days,
because
we
have
a
threshold
set
there,
but
then
all
the
jobs
in
there.
If
it's
short
enough
small
enough
and
you
can
get
a
opportunity
to
get
back
filled,
it
will
run
quickly.
N
Again
that
you're,
this
is
what
this
plot
is
going
to
tell
us.
We
have
in
our
configuration
actual
numbers.
So
if
you
submit
a
job
in
a
regular
queue,
you
come
in
with
a
start
priority,
and
then
we
have
it.
It
takes
you
three
days
to
start
to
be
scheduled,
but
actually
recently
we
removed
that
wait
time.
The
the
original
we
had
this
as
manual
wait
time
for
a
regular
job
to
be
scheduled.
N
Our
a
new
scheme
does
not
need
to
have
this
up
manual
weight
to
originally
we
had
some
some
related
bugs
that
without
this
weight
there
were
issues
related
to
the
system
over
utilization
and
newer
version
does
not
need
this
weight
anymore.
So,
basically,
a
a
job
submit
it
already
reaches
threshold.
So
that's
why
we
remove
that
column,
you're,
not
seeing
it
in
sqs
anymore,
something
like
that.
N
I
But
the
backbone
now,
returning
to
that
diagram,
the
backlog
just
means
that
this
is
how
much
work
is
scheduled
and
waiting
to
be
done,
but
it
doesn't
mean
that
that's
how
exactly
how
long
it
will
take
these
jobs
to
finish
many
of
them
could
ensure
them.
Many
of
them
could
arrow
out
and
etc.
Right.
N
The
backlog
is
like,
however,
many
workload
is
in
the
system
right
now,
new
jobs
after
that
like
right
now,
you
could
still
finish
maybe
before
that,
because
there
are
some
jobs
submitted
in
the
priority
lower
than
yours,
and
there
are
also
some
users
have
more
than
two
jobs
in
the
queue
and
they
they're
big
enough
they're,
not
fitting
to
a
backfield.
A
Jobs
so
expanding
on
that
a
little
bit
so
so
your
job
goes
into
the
queue
at
some
point
with
a
like.
Your
priority
basically
increases
over
time.
The
the
change
that
helen
was
talking
about
where
you
used
to
see
like
a
three-day
wait
was
because
qregular
used
to
start
out
down
here
and
it
would
take.
A
You
know
three
days
to
get
up
to
this
line,
whereas
now
it
starts
right
at
the
top,
but
only
two
jobs
can
increase
past
this
line
at
a
time
per
user
and
above
here
this
is
the
this
is
the
part
of
the
queue
that
slurm
spends
a
lot
of
time,
trying
to
find
a
place
to
schedule
the
job
to
start
and
then
for
everything
below
that
it
does
a
quick
run
through
and
just
ask.
Can
I
start
this
job
right
now.
A
So
we're
almost
at
12
30
now
and
it
might
be
time
to
close
out
the
meeting
thanks
all
for
sticking
around
for
further
discussion
and
yeah
thanks
jose
again,
although
I
think
he's
dropped
out
for
for
telling
us
about
slum
and
we'll
look
forward
to
seeing
you
all
again
next
month.