►
From YouTube: Grafana Tempo Community Call 2023-02-09
Description
Tempo 2.0 discussion and celebration!
Join our next Tempo community call: https://docs.google.com/document/d/1yGsI6ywU-PxZBjmq3p3vAXr9g5yBXSDk4NU8LGo8qeY/edit
Learn more at https://grafana.com and if all of this looks like fun, feel invited to see if there’s a role that fits you at https://grafana.com/about/careers/
A
Okay,
all
right
all
right
welcome
everyone.
This
is
a
pretty
I
think
this
is
an
exciting
Community
called
This
is
the
first
one
we've
had
since
releasing
Tempo
2.0.
So
this
is
great
yeah.
Thank
you.
Everyone
for
coming.
A
We
shared
the
Google
Doc
in
the
slack
Channel,
and
so,
if
you
haven't
opened
that
you
know
we
kind
of
have
a
light
agenda
in
there,
but
you're
welcome
to
add
anything
to
it.
You
know
feedback
or
issues
or
discussion
topics.
Anything
we
love
to
hear
it
so
otherwise
I
can
go
ahead
and
start
off
yeah,
we'll
just
talk
about
Tempo
2.0.
So
all
right,
so
we're
all
really
excited
about
this.
A
This
represents
months
and
months
of
work
by
dozens
of
people,
huge
effort,
but
we
think
it's
a
great
release.
Best
release
of
tempo
tons
of
new
features
in
there
kind
of
like
a
very
large
changes
to
how
Tempo
Works
internally.
So
we're
really
happy
about
that
yeah.
So
the
big
feature
in
here
is
probably
Trace.
Ql,
that's
really
cool
I
mean
we
knew.
A
We
wanted
a
query
language
for
a
long
time
and
but
we
kind
of
needed
the
new
back
end
storage,
to
kind
of
like
power
that
so
the
existing
one.
What
just
wasn't
going
to
be
efficient
enough,
so
the
new
back
end
is
the
parquet.
So
that's
we,
you
know
that
was
actually
released
in
1.5
and
I
know.
A
Some
people
were
interested
enough
to
go
ahead
and
try
that
and
that
was
really
cool
and
we,
of
course
we
were
using
it
ourselves
internally
for
a
long
time
several
months,
but
it's
finally
the
default.
It's
we
feel
like
it's
production,
ready
we've
been
using
it
in
behind
behind
graphonic
Cloud
for
months
now,
so
we
feel
pretty
good
yeah.
So
that's
turned
on
by
default
and
so
is
traceql.
A
So
yeah
I,
guess
I
just
want
to
say
yeah.
We
should
just
thank
you
to
the
whole
team
and
everyone
and
all
Community
contributors
and
everyone.
So
it
was
a
huge
effort.
Yeah
Zach
yeah
cool,
so
there's
a
couple
links
here
in
the
documents:
there's
a
blog
post
kind
of
like
from
Joe
about
Tempo,
2.0
kind
of
covers
like
the
new
features
and
the
basics.
A
There
are
some
breaking
changes
so
when
you're
upgrading
I
think
we
have
a
really
good
upgrade
guide
in
the
documentation,
so
kind
of
cover
the
braking
changes
and
then
other
things
like
settings
defaults
that
have
changed.
You
know
that
you
may
need
to
be
aware
of
yeah.
So
and
of
course
we
have
a
patch
release
upcoming,
so
there's
a
couple
kind
of
smaller
issues
and
things
that
we
weren't
able
to
get
there
in
time.
So
yeah
there's
also
a
link
here
to
the
201,
Milestone
and
I.
A
Don't
think
there's
anything
critical
in
there,
except
for
there
are
some
people
that
are
impacted
and
I.
Think
if
you
are
impacted
you
know
it.
So
that's
good!
So
yep,
oh
yeah,
we
got
some
activity
in
the
document
here,
yeah
and
then
a
couple
days
ago
there
was
another
blog
post,
just
kind
of
like
getting
to
know
Trace
ql,
so
that
link
is
in
there
too,
so
kind
of
like
just
a
good
way
to
get
into
the
basics
of
it
and
get
up
and
running
cool
yeah.
A
So
I
guess
I'd
like
to
do
kind
of
a
survey
here
like
is
anyone
running
Tempo
2.0
yet
or
have
you
been
using
parquet
or
Trace
ql
yeah
cool.
A
Do
you
have
any
any
feedback
or
thoughts
or
like
that
you'd
like
to
share.
B
Yeah,
so
we
with
the
version
1.5
we
enable
parquet,
and
now
we
have
it
with
the
2.0
and
I
can
say
that
at
least
in
our
in
gestures,
I
see
that
is
working
better.
So,
like
performance
improved
there
I
used
to
see
the
compactors
working
better
I,
don't
know
if
I
had
some
issues
before
but
now
I
see,
I
have
the
block
list
like
better
and
so
far
like
it
just
feels
more
just
feel
faster.
Yeah
I,
like
it
yeah.
A
Yeah
so
the
version
that
was
in
1.5
was
stable,
but
the
performance
wasn't
where
we
wanted
it
to
be
yeah,
and
so
that
was
a
big
focus
of
effort
between
1.5
and
2.0
yeah
yeah,
especially
getting
it
to
run
better
at
higher
scales
and
yeah,
like
better
use
of
memory,
handling
large
traces.
Things
like
that,
yeah
cool.
A
Cool
yeah
deployed
V2
this
morning:
okay,
cool,
great
yeah,
okay,
that's
great
yeah,
except
for
the
breaking
changes
and
things
yeah
I
think
it
was
pretty
good.
A
All
right,
cool
yeah,
so
okay
mentioned
about
resource
uses,
yeah,
that's
true,
so
3x,
so
we
were
estimating,
maybe
about
50
higher
is
what
we've
been
seeing
internally
compared
to
well.
This
is
with
Tempo
2.0,
so
in
1.5
with
parquet
it
was
definitely
higher.
A
3X
is
definitely
more
than
we
were
expecting.
Yeah,
oh
on
the
S3
storage,
yeah
yeah,
so
but
for
CPU
and
memory
I
think
we're.
You
know
we're
kind
of
seeing
50
more
resources,
but
I
think
it's
worth
it
right
for
the
new
features
like
yeah
I.
Think-
and
it's
still
you
know,
Tempo
is
still
very
cost
effective
as
far
for
a
back
end
and
it's
I
think
it's
still.
It
was
a
good
choice:
yeah
cool
yeah,
yeah.
A
Really!
This
is
all
we
kind
of
wanted
to
cover
and
we
can
talk
a
little
bit
more
about
maybe
some
of
these
things
that
are
in
here,
but
this
was
the
big
news.
This
is
the
first
Community
call
since
Tempo
2.0.
B
I
have
seen
that
using
this
ql
now
I
can't
query
like
longer
hour.
Hours
like
before,
I
was
limit,
so
we
have.
There
is
a
settings
that
you
can
say:
okay,
I
can
query
like
one
hour
or
three
hours
or
so
bye
go
in
one
go,
but
now
I
do
like
sometimes
like.
Okay,
look
in
the
last
six
hour,
but
my
default
was
one
hour.
B
A
A
Yeah
yeah,
just
V2,
wasn't
efficient
enough,
so
it
had
to
do
a
lot
of
work
to
search
through.
So
one
hour
is
quite
a
lot
of
work,
but
on
parquet
with
it's
addressing
the
columns
and
also
like
the
internal
metadata,
you
can
skip
over
things
a
lot
more
efficiently,
yeah
yeah,
actually
24
hours
is
probably
you
know,
that's
something
that
probably
could
be
higher
in
the
future,
like
I
think
yeah.
Maybe
that
works
for
some
cases.
But
you
know,
as
we
work
as
we
improve
the
things.
D
Yeah
I
had
a
question
about
yeah
about
how
to
handle
like
scaling.
Up,
Tempo
queries,
I
noticed
that
which
I
skew
all
the
the
the
member
usage
can
really
Spike
up,
and
you
know
when
I
throw
a
lot
of
simple
queries
at
it.
It
really
helps
to
to
handle
some
some
larger,
expensive,
Trace
ql
queries,
but
is
there
like
a
recommended
way
on
how
to
scale
this
up?
Like
I
noticed
there
was
some
documentation
about
using
like
Lambda
functions
or
you
know.
D
Other
sort
of
you
know
scaling
functions
like
that
or
it's
like
an
auto
scaler
sufficient
most
of
the
time
just
trying
to
figure
out
yeah
what
are
some
like,
cost-effective
ways
or
any
like
recommendations.
The
team
might
have
for
handling
that.
A
Sure,
yeah
yeah,
so
there
are
a
lot
of
knobs
that
to
turn
with
all
this
new
stuff.
So
one
might
be
the
traceql
query.
A
So
if
you
just
like
a
short
recommendation,
would
be
use
scoped
attributes,
if
you
can
so
in
that
case,
that
would
be
like
instead
of
dot,
http.statuscode
span.hb
dot
status
code
is
more
precise
and
if
you're,
following
the
semantic
conventions,
that's
where
the
HTTP
status
code
would
be,
and
so
that
when
you,
when
you
do
that,
that
lets
Tempo
know
only
to
look
at
the
span
level
attribute
column
for
that
one
versus
searching
everywhere,
which
is
the
resource
level
and
the
span.
A
So
there
are
ways
to
like
make
the
query
a
little
bit
more
efficient,
and
these
are
just
early
things
and
kind
of
maybe
you
know,
is
not
intended,
but
this
is
something
that
works,
so
maybe
we
can
work
up,
make
it
easier
in
the
future.
Vice
versa.
Dot
service.name,
you
know,
would
be
kind
of
like
looking
at
the
resource
service
named
I.
A
Think
that's
pretty
common,
but
if
you
did
resource.service.net
and
that
would
be
efficient
for
the
same
reasons,
we
can
only
have
to
look
at
the
resources
which
are
way
faster
as
far
as
scaling
queries,
yeah,
so
Tempo
is
kind
of
unique,
so
we
have
serverless
and
we
use
that
ourselves
and
yeah
to
achieve
like
the
scale
that
we
need.
So
that's
a
really
good
solution
like
depending
on
your
load
and
it's
cost
effective
too
right.
A
So
because
the
serverless,
you
know
the
queries,
maybe
are
infrequent
enough
like
make
serverless
a
lot
more
cost
effective
than
running
queries.
A
So
if
you
haven't
done
that
I
think
that
would
provide
instant
scale
right
as
far
as
memory
I
think
our
serverless
functions
run
with
two
gigabytes
of
memory:
yeah,
it
seems
to
be
okay,
yeah
yeah,
so
I
mean
you
can
run
thousands
of
serverless
functions
and
we
do,
and
so
that's
kind
of
like
an
instant
way
to
scale.
But
it
does
take
some
operational
setup.
A
Yeah
you're
right
I
guess
when
we've
seen
Courier
spikes
too
and
I,
don't
think
that's
usually
from
searching,
maybe
large,
pulling
back
large
traces
can
can
be
a
thing
as
far
as
query
or
is
sure
you
can
scale
them
up
and
run
a
couple,
hundreds
of
PODS
quarters
and
things
like
that,
and
they
actually
have
more
consistent
performance
that
we've
seen
than
serverless
functions
just
as
far
as
startup
time
and
tail
latency.
A
But
as
far
as,
if
there's
anything
to
do
about
yeah
I'm,
not
sure.
If
there's
any
Hey
can
anyone
else.
Think
of
like
any
specifics
or
best
practices
for
scaling
the
query
or
pods
specifically.
E
I
mean
the
serverless
is
a
good
start,
but
yeah
as
Marty
said
it's
it's
a
little
more
complex
to
set
up
I
know.
We've
got
some
items
that
we're
looking
at
in
the
next
quarter
to
help
smooth
out
the
query
path,
but
I
don't
actually
know
if
that's
going
to
help
performance
necessarily
yeah
interesting,
add
more
of
them.
D
A
Query
shards
so
I'll
type
this
in
here,
but
I
think
that's
the
setting
name.
So
that's
particularly
impactful
too.
It
should
be
equal
to
or
greater
than
the
number
of
PODS
you
have
or
yeah.
So
if
this
is
smaller
than
the
number
of
PODS,
that
means
it
won't
actually
scale
up
enough
to
make
use
of
every
pod.
D
Oh
interesting
and
you're,
referring
to
like
Tempo
query
or
pods
yeah
yeah,
would
you
recommend
Amanda?
Was
it
called?
We
have
the
query
front
end
as
well.
Is
that
something
that
needs
to
be
scaled
up?
Similarly,
I
haven't
noticed
as
many
spikes
there,
but
and
I'm
not
quite
sure
how
it's
all
wired
up,
but
is
that
something
that
needs
to
be
skilled
as
well.
A
No,
not
really
just
a
couple
plots.
There
is
all
you
need,
and
that
can
go
quite
far
so
they're
just
kind
of
coordinators
and
proxies
they
don't
do
a
lot
of
the
heavy
lifting
I.
Think
like
two
query
front
ends.
Just
for
high
availability
can
run
a
couple
hundred
queer
pods,
just
fine
yeah,
gotcha
yeah.
A
You
know,
there's
just
a
lot
of
tunables
for
performance
row
group
size
like
that's:
a
setting
maximum
block
size,
maximum
traces
per
block,
there's
kind
of
like
a
long
list
of
things
to
check
but
yeah
for
just
that
instant
scale.
I
would
probably
go
serverless.
D
Sounds
good,
yeah
yeah?
Is
there
any
documentation
on
like
how
to
you
know
tune
these
knobs
or
or
maybe?
Is
there
a
need
for
that?
Maybe
like
like,
like
here's,
some
recommended
settings
or
like
a
a
guide,
perhaps
on
tuning
that
or
maybe
that's
kind
of
hard
to
do,
perhaps
because
it's
maybe
variable
on
everyone's
depending
on
everyone's
situation,
yeah.
A
I
think
we
did,
we
did
change
some
defaults
in
2.0.
I.
Think
that
we've
found
work
better,
like
some
of
those
defaults
were
kind
of
getting
outdated,
yeah.
It's
kind
of
a
tough
tough
thing
that
variability
in
traffic
does
make
it
hard
to
choose
one
set
of
settings.
Yeah.
F
A
So
it
the
the
group
I
in
Trace
uo
is
yeah.
Maybe,
like
we
don't
know
everything.
F
A
It
would
be
able
to
do,
but
it
would
be
able
to
like
count
the
number
of
times
each
query
was
used
in
a
trace
and
I'm,
not
sure
that
we
have
anything
in
language
about
like
just
ranking
them
like
show
me
the
top
five,
but
it
could
just
show
you
the
count
for
each
one,
but
what
also
might
work
is
metrics
generator.
A
So
this
is
a
feature
that
has
been
in
tempo
for
a
while,
so
we
can
generate
metrics
from
spans,
it's
a
separate
component
in
tempo
and
you
can
customize
the
attributes
that
are
turned
into
metrics
Dimensions
there
yeah,
so
you,
if
you
your
attribute,
was
like
db.query.
You
could
add
that
as
a
custom,
Dimension,
I
think
to
the
span
metrics
and
depending
on
how
long
that
query
is.
Maybe
that
might
not
be
great
I,
don't
know.
Mario
knows
more
about
this
area.
Yeah.
G
Yeah
I
will
link
the
metrics
generator
docs,
so
everyone
can
they
commit
to
look
but
yeah
in
a
sense
with
this
component,
you
can
derive
metrics
from
your
as
patents
to
get
this
pan
level
metrics
and
additionally,
like
you,
can
extract
any
attributes
that
are
present
in
this
pans
asymmetric
labels.
So
if
you
have
a
an
an
attribute
in
European
studies,
SQL
statements
or
something
of
the
like,
you
could
distract
it
and
and
have
it
as
metric
and
then
operate
with
it.
G
Just
as
regular
metrics
and
yeah,
you
could
do
top
five
SQL
statements,
the
only
caveat
or
limitation
is
this.
Is
this
pan
level
so
I?
Think
for
the
for
your
initial
example
would
work
perfectly
since
this
would
be
like
a
span
level,
question
or
yeah
question
that
you
want
to
ask,
but
yeah
I
think
it's
a
different
approach
to
getting
metrics
out
of
choices
and
so
far
as
worked
well
based,
simple
and
useful.
G
No,
that's
the
limitation.
I
was
referring
to
this.
This
is
Spam
level.
It
doesn't
record
metrics
with
knowledge
of
the
entire
Trace.
It's
Perth
pattern.
Only
yeah.
A
A
A
Okay,
I
guess
yeah
otherwise,
like
we
can
wrap
up
unless
there's
any
other
topics,
we'd
like
to
discuss
sure
yeah,
Lucas.
C
Yeah
is
there
a
way
to
filter
what
the
Distributors
send
to
the
metrics
generator?
Like
can
I
say,
like
hey,
I
only
want
this
service
to
go.
C
G
Yeah,
so
this
is
an
item
that
we're
actively
working
on
yeah
I
didn't
respond
because
I
was
trying
to
look
for
the
issue
to
give
the
most
accurate
information
of
the
state
of
that
so
yeah
there's
something
that
we
definitely
intend
on
doing.
Okay,
I.
G
If
you
want
to
subscribe
to
updates
but
yeah,
essentially,
we
just
want
to
filter
down
what
we
send
to
the
metrics
generator,
although
right
now,
if
you
want
to
cut
down
on
a
cardinality
or
the
metrics
that
you're
sending
there's
still
workarounds
that
you
could
apply
like
since
we're
embed
in
the
Prometheus
agent
in
in
The
Matrix
generator
for
the
sending
metrics
via
remote
right,
you
could
apply
relabel
configs
to
drop
labels
to
drop
metrics
and
to
do
all
the
magic
that
we
label
reliable.
G
Configs
do
yeah.
C
Yeah,
that's
good.
I
was
just
curious
because
we
do
like
I
forget
why
I
think
it's
like
one
and
a
half
or
two
terabytes
of
traces
a
day,
but
we
only
actually
care
about
the
APM
statistics
on
just
a
handful
of
those
services.
So
it's
kind
of
it's
a
lot
less
workload
for
the
metrics
generator
to
just
not
have
to
be
sent
that
at
all,
when
I
was
using
another
APM
tool,
I
was
trying
to
do
the
filtering
like
at
the
hotel
agent.
C
You
know,
send
everything
to
Tempo
and
then
only
send
a
subset
to
this
other
tool
that
I
had
building
the
APM
stuff
and
so
now
I'm
trying
to
switch
from
that
APM
tool
to
Tempo
for
APM.
So
I
was
just
wondering
if
there's
a
way
to
do
it,
but
it's
not
a
huge
deal,
just
mainly
if
it's
on
the
roadmaps
right.
G
Yeah
makes
sense,
then,
hopefully
we'll
have
a
explicit
solution.
Just
filtering
in
the
metric
generates
itself
for
now
yeah.
If
we
want
to
investigate
yeah,
you
can
still
apply
reliable
configs
to
the
Prometheus
agent,
and
this
will
do
all
kinds
of
magic
like
Drop
in
specific
metrics,
specific
labels,
and
it
I
mean
it
will
still
eat
into
memory
and
CPU
and
the
metric
generator.
But
since
you're
not
writing
them
you're
still
saving
a
lot
in
your
metrics
shortage
yeah.
G
If
maybe
this
is
too
deep
to
get
into
reliable
configs
if
you're
not
familiar
with
them,
I
want
to
chat.
Maybe
we
can
move
the
condensation
to
the
public
slack
or
someone
else.
Do
yeah
talk
more
in
depth.
A
Okay,
all
right
well,
I
guess
we
can
wrap
it
up
again.
Happy
to
see
all
the
new
faces.
Definitely
welcome,
and
this
was
yep
really
excited
about
Tempo
2.0
check
it
out
and
I.
Guess
we'll
see
you
next
month,
okay,
bye,
everybody.