►
From YouTube: Get Started with Processing Delta Tables Using AWS Glue, Amazon Athena, and Amazon Redshift
Description
There are a lot of use cases of Delta tables on AWS. Noritaka Sekiyama, Principal Big Data Architect at AWS Glue, will demonstrate how to get started with processing Delta tables on Amazon S3 using AWS Glue, Amazon Athena, and Amazon Redshift on Tuesday, November 8, 2022 at 4:00PM PDT.
Learn more about Delta Lake: https://delta.io/
Noritaka Sekiyama: https://www.linkedin.com/in/moomindani/
Denny Lee: https://www.linkedin.com/in/dennyglee/
Join us on Slack: https://go.delta.io/slack
Delta Lake Releases: https://github.com/delta-io/delta/releases
A
And
YouTube
channels,
but
if
you're
wondering
what
session
you
are,
we
propped
it
right
on
screen
we're
talking
about
getting
started
with
processing,
Delta
tables
using
AWS,
glue,
Athena
and
redshift,
but,
like
I,
said
we're
going
to
take
a
couple
minutes
just
to
get
our
YouTube
channels
and
our
LinkedIn
channels
set
up
so
in
the
interim
by
all
means.
A
If
you
have
questions,
please
use
the
Q
a
and
if
you
want
to
tell
us
a
little
bit
on
where
you're,
based
out
of
where
you're
from
by
all
means
go
ahead
and
chime
in
inside
the
chat
on
where
you're,
based
on,
for
example,
my
name
is
Denny
and
I'm
based
out
of
Seattle
Washington,
specifically
Kirkland,
but
Seattle
area.
How
about
you
Nori?
Where
are
you
based
out
of.
A
So
we
got
a
nice
Global
audience
today,
awesome
so
based
on
so
again,
norian
Tokyo
myself
and
Seattle
yeah
for
the
for
those
who
are
chatting
away
go
ahead
and
tell
us
where
you're
based
out
of.
A
Excellent
we've
got
Steven
based
out
of
Wicklow,
I,
think
Ireland,
so
that's
really
cool
all
right
and
oh
Alexandra
Egypt.
Now
that's
pretty
cool!
Oh
man
I
like
that.
Okay,
let's
see
sorry,
we've
got
the
linked.
A
We've
got
LinkedIn
all
set
up,
and
so
which
is
really
cool
and
looks
like
now.
We've
got
YouTube
set
up,
so
we
are
set
all
right,
perfect,
well,
welcome
everybody
again,
we'll
go
ahead
and
restart.
So
now
that
everybody
is
is
online
in
terms
of
all
the
LinkedIn
and
the
YouTube
and
the
zoom
channels
hi
there.
A
My
name
is
Denny
Lee
I'm,
with
databricks
and
I'm
a
Delta
Lake,
committer
and
I'm
really
really
really
happy
to
introduce
Nori
he's
from
AWS
he's
gonna
introduce
himself
a
little
bit,
but
he's
going
to
be
talking
today
about
getting
started
with
processing,
Delta
tables
with
AWS
glue,
Athena
and
redshift
before
we
dive
into
those
particulars,
I
did
want
to
do
a
little
housekeeping.
A
So
if
you
do
have
questions
for
those
of
you
who
have
dialed
in
using
Zoom,
just
use
the
Q,
a
okay
for
those
who
of
you
who
are
doing
LinkedIn
by
all
means,
go
ahead
and
put
your
chat
or
your
questions
in
the
LinkedIn
and
the
same
for
the
folks
who
are
actually
on
YouTube
I'll
chime
in
on
the
background
I'm
going
to
mute
myself
when
Nori
goes
ahead
and
speaks
so
I'll
do
my
best
to
answer
various
questions
but
by
the
same
token,
here's
here's.
A
How
we're
gonna
go
here,
we'll
make
sure
that
to
answer
the
questions
live
once
he
Nori
has
done
the
main
section
so
without
further
Ado.
Let
me
just
go
shift
over
to
Nori
going
to
introduce
yourself
and
dive
into
our
today's
session.
Please.
B
Sure,
hello,
everyone
I'm
super
excited
to
to
talk
about
theater,
Recon
AWS,
using
group,
Athena
and
redshift
in
this
session.
I'm
Nori
principal
Big,
Data
architect
on
the
AWS
group
team.
I
am
responsible
for
building
software
artifacts
that
help
customers
to
build
data
rakes
on
AWS
Cloud.
Recently
I
have
published
a
book
about
AWS
group
for
today's
webinar.
We
will
be
giving
away
a
few
copies
of
this
book
to
three
of
the
lucky
winners.
B
B
B
As
many
of
you
already
know,
Delta
rake
is
an
open
source
project
that
helps
Implement
modern
data
array
card
texture
commonly
built
on
Amazon,
S3
or
other
Cloud
strategies
with
deterrick.
You
can
achieve
AC
transactions,
time
travel,
queries
CDC,
including
abstract
and
other
common
use
cases
on
the
crowd.
B
Aws
group
is
a
serverless
data
integration
service
on
AWS.
It
gives
you
scalable
data
integration,
ending
unified
data,
governance,
ability
to
connect
to
various
data
source
and
so
on
this
database
group
you
can
easily
integrate
data
across
multiple
data
stores
and
transform
and
enrich
data
and
make
it
queryable
from
different
analytics
and
email
applications.
B
Why
AWS
group
is
good
for
the
rake
he
has
five
years.
First
AWS
group
make
data
preparation
and
data
integration
simpler,
faster
and
cheaper.
Second
AWS
do
provide
the
powerful
servers
data
integration
capability
for
all
Enterprises
using
the
AWS
ecosystem;
third,
that
both
AWS
screw
jobs
and
chloras
support,
Delta
Lake
and
then
break
formation,
AWS
rake
formation.
The
sub
is
for
centralized
access
control,
which
also
supports
the
the
direct
tables
foreign.
B
B
B
B
B
This
is
a
very
good
for
simple
use
cases
such
as
the
data
movement,
data,
Transformations
and
so
on.
B
B
B
B
B
B
Okay,
let
me
start
the
demo.
Okay
now
I
will
create
a
new
jifton
notebook
job
from
the
existing
iPhone
notebook
file.
This
is
available
in
the
public
GitHub
repository.
Then
I
will
be
giving
some
names
for
job
name.
Then
setting
the
IAM
role
when
I
provided
this
information,
then
group
studio
will
try
to
spin
off
the
serverless,
managed
Jupiter
notebook
interface.
B
So
it's
loading
now
or
it
will
be
spinning
up
within
several
seconds
after
that,
I'll
be
explaining
about
how
to
oh,
how
the
notebook
experience
with
AWS
group
and
data
rate
look
like
okay.
Now
the
notebook
is
ready,
so
I
will
make
the
small
changes
to
or
make
it
available
for.
Marketplace
connector
then
run
it.
Then
the
options
are
configured
for
this
notebook
session
now
or
let
me
give
some
bucket
name.
B
I
have
this
time
I
am
using
Sydney
region,
so
I
am
passing
the
my
bucket
name
that
is
located
in
a
cylinder
region.
The
one
I
run
this
cell,
then
the
internally.
It
is
spinning
up
the
ephemeral
group
cluster,
while
waiting
for
that
I
I
will
be
explaining
about
how
what
we
are
doing
on
this
notebook.
The
first
I
will
be
creating
five
sample
records
using
spark
or
data
frame
API
and
then
create
a
Delta
rig
table
or
using
this
sample
record.
B
B
That's
pretty
much
everything
for
all
this
notebook
demo
and
in
the
upset
phase.
I
will
use
merge
command
for
that
now.
The
ephemeral
cluster
is
already
set
up
so
I'm
running
some
cleanup
phase.
For
this
then
or
said
next
cell
or
I'll
be
creating
five
sample
records.
B
This
table
all
shows
the
fictional
or
product
inventory
that
has
the
product
name,
product
price
and
category.
Okay,
then
right
now,
I
am
writing
into
a
screen
location
using
Delta
table.
So
once
this
server
has
been
completed,
then
the
data
Delta
files
I
mean
pocket
files
with
the
transaction.
Rook
and
metadata
will
be
located
on
your
Amazon
S3
bucket.
A
Folks
that
are
actually
having
problems
posting
comments
on
LinkedIn.
We
notice
that
there's
a
problem
there
please
go
ahead
and
I'm
going
to
repost
on
LinkedIn,
under
my
personal
account
for
you
to
chime
into
the
Delta
user
slack,
and
if
you
can
just
ask
your
question
in
events
and
or
you
can
also
join
us
directly
through
the
zoom
link,
so
I'm
actually,
actually
you
know
what
I'm
gonna
do
that
well
Nori,
why
you
go
ahead
and
take
care
of
that
I'm
gonna
go
ahead
and
send
out
the
zoom
link.
B
Thanks
thanks:
okay,
let
me
resume
the
demo
there
and
now
the
data
is
already
written
into
S3
and
Delta
table
is
visible
in
the
cutter.
Now
let
me
read
from
Delta
table
yep
five
records
are
shown
up
in
the
table,
so
in
this
spark
SQL
you
can
see
the
content
of
that
table.
Now
this
is
the
basic
operation.
Now
I
will
be
trying
to
insert
the
new
record
like
product
ID
6.
The
process
name
is
pen.
B
B
The
next
one
is
a
pretty
much
very
important.
It
simulates
absurd
operations
on
data
Lake,
one
record
will
be
inserted.
Newly
one
record
will
be
updated
or
absurd.
I'm
using
merge
query
on
data
Delta
table.
Then
after
this
query
finished,
then
you
can
see
the
one
new
record,
one
new
record
for
product
ID
8
and
one
existing
record
product
id1.
The
price
has
been
changed
successfully.
B
B
B
We
have
seven
sample
notebooks
available
on
GitHub.
You
can
easily
use
any
of
the
samples
to
understand
how
it
works
and
simulate
this
demo,
three
notebooks
out
of
seven
notebooks
are
related
to
the
rhetoric.
B
Okay,
in
this
demo,
I
have
demonstrated
how
the
notebook
experience
will
create
in
the
next
ripe
demo,
I'll
be
explaining
how
the
group
study
or
visual
editor
experience
with
rank.
Let
me
show
my.
C
B
That's
right,
so
this
is
the
AWS
group
console
and
from
now
on,
I
am
demonstrating
the
glue
Studio
experience.
So
let
me
open
Guru
Studio.
B
And
to
use
a
div
Drake
as
I
explained
first,
you
will
be.
You
will
need
to
satisfy
the
marketplace
connector
that
is
designed
for
data
Lake.
The
directory
Marketplace
connector
is
located
here.
So
if
you
view
this
product,
then
you
will
see
the
wizard
to
subscribe.
This
connector.
B
Once
we've
you
finished
the
this
wizard,
then
you
will
have
the
you
will
be
able
to
use
the
Delta
rig
on
Google
studio.
Let
me
go
back.
C
B
By
the
way,
as
you
can
see,
there
are
so
many
connectors,
so
if
you
want
to
bring
some
data
from
somewhere
to
the
redirect,
you
can
use
this.
If
you
want
to
move
the
some
data
on
data
Lake
to
other
place,
you
can
also
use
this
system.
B
Okay,
now
let
me
create
a
new
visual
job.
Okay,
so
in
this
demo,
I
read
from
some
public
S3
packet
and
write
into
or
my
S3
bucket,
using
Delta
rig,
so
the
sources
standard
S3
the
target
target,
is
the
Delta
rig.
So
let
me
do
like
this.
Okay,
then
the
template
will
be
shown
up.
Let
me
name
it
there
with
the
rake
visual
okay
in
the
at
the
source.
You
can
specify
a
three
location
or
you
are
existing
catalog
table
in
this
demo.
B
B
Okay,
this
is
the
Kobe
90
data
set
okay,
I.
Probably
mapping
is
very
simple
or
transformation
that
will
map
your
crumbs
into
the
different
names
or
different
types
of
whatever.
Then
after
that,
I'll
be
using
Delta
rig
to
store
this
table
for
this
I
need
to
pass
one
extra
parameter
to
show
the
location.
C
C
B
Okay,
now
that's
it!
Okay,
now
I
save
this
job.
Then
let
me
run
this,
and
now
all
the
girls
of
system
is
trying
to
spin
up
the
server
cluster
in
the
internal
backend.
Then
I
will
be
demonstrating
how
it
works
with
Delta
Lake.
B
A
B
A
That's
really
cool
so,
and
do
you
then
specify
which
inside
because
I
noticed
all
the
configurations
and
the
scripts?
Do
you
then
basically
specify
in
the
script
or
in
the
initial
job?
Where
which
version
of
spark
you
want
to
play
with
for
sake
argument,
thank.
B
A
B
You,
okay
now
the
Adobe
is
still
running,
but
let
me
explain
what
is
happening
right
now,
so
you
can
see
startup
time
8
seconds.
It
means
Guru
was
able
to
spin
up
the
10,
node
or
spark
cluster
10
node
I
mean
here
10
dpu.
It
means
10
dot.
10
nodes
spark
cluster
within
eight
seconds,
then,
after
that
we
are
running
some
script
and
it
is
executing
right
now
succeeded.
So
start
type.
Time
is
seven
seconds
and
execution
time
is
1
minute
and
43
seconds.
B
Okay,
now
at
the
destination.
Let
me
open
S3
console
because
I
am
writing
reading
from
public
bucket
and
writing
it.
To
my
bucket,
let
me
verify
that
S3
pass
group
rate
formation,
demo
Us
best
tool.
This
is
the
bucket
S3
bucket
name.
B
Then
The
Rake
video
demo.
Now
you
are
seeing
these
raw
files
located
on
S3.
This
is
the
main
pocket
file,
or
this
is
the
transaction
row
that
is
required
for
the
return.
Okay,
so
the
you
you,
as
you
can
see,
we
were
able
to
ingest
into
Delta
rake
format
on
S3
I
will
use
this
the
direct
table
or
for
the
further
demo.
B
Thank
you.
The
next
step
is
so
once
we
ingest
this
Delta
table
on
my
S3
bucket
I
want
to
query
this
table
or
from
different
engines
like
Athena
redshift
like
that,
and
then
in
the
next
demo,
we
will
demonstrate
to
query
Delta
tables
using
Amazon,
Athena
and
redshift.
B
Before
doing
that,
let
me
explain
a
little
bit
more
about
the.
What
is
the
Delta
table
and
what
is
the
New
Concept
of
manifestable
in
in
the
previous
demo,
I
used
AWS
screw
with
apart
spark
to
read
I
write
Delta
table.
Actually
it
was
that
Delta
table
created
in
the
about
spark,
so
it
is
a
native
Delta
table
in
the
next
two
demo.
I
want
to
query
Delta
table
from
Amazon
Athena
and
press
shift
to
make
the
Delta
table
queryable
for
those
engines.
B
It
is
important
to
create
a
manifest
table
which
is
based
on
the
simmering
text
format.
The
standard
way
to
create
a
manifest
table
is
to
run
generate
command,
so
you
can,
as
you
can
see
here
the
example
of
the
generate
command.
If
you
run
this
command
in
your
spark
cluster
with
data
rig
library,
then
it
will
automatically
populate
the
same
link
based
manifest
table.
This
is
one
standard
way
that
is
natively
supported
in
the
Drake.
B
Another
positive
way
on
AWS
group
platform
is
to
use
AWS
Google.
To
do
the
same
thing,
it
simplifies
the
generating
manifestable
operation
and
you
can
easily
schedule
that
to
or
sync
these
or
tables
in
the
specific
period
of
time,
I
will
choose
this
way
for
Simplicity
for
the
next
demo.
Okay,
then,
in
the
next
demo
or
I,
will
be
demonstrating
chlorine
Delta
tables
using
Google
Chrome
is
a
component
to
automatically
create
a
table
definition
from
actual
data.
B
You
you
can
you
do
not
need
to
or
configure
any
of
the
table
schema.
A
Google
Chrome
will
be
creating
the
table.
Definition
on
behalf
of
you
for
Delta
rig.
It
also
has
the
extra
capability
to
Auto
generate
simulink
based
manifest
file,
so
the
Manifest
type,
the
Manifest
table,
manifest
file
enables
Athena
and
receive
Spectrum
to
query
the
Delta
table.
A
No
worries
yeah
I
was
just
about
to
ask
you
see
if
he
can
go.
Let's,
let's
go
show
that
part
again,
because
actually
we
got
a
lot
of
questions
both
on
okay,
linked
on
the
Q,
a
and
also
on
the
LinkedIn,
actually
exactly
about
how
does
glue
work
with
the
Delta
tables
and
the
crawler.
So,
yes,
please
go
ahead
and.
A
B
So
this
is
that
that's
right
deck
I
wanted
to
show.
So
there
are
two
type
of
the
tables,
a
native
Delta
tables
and
the
Manifest
table.
So
in
the
glue
I
used
a
native
data
table
because
I
use
Apache
spark
ecosystem
to
create
a
Delta
table,
but
for
next
demo,
I
will
need
to
create
manifest
space
table.
Now
that
is
based
on
simmering
text
format
to
make
it
equivalent
from
Athena
and
press
shift.
Usually
you
will
need
to
do
this
operation
or
generate
command.
B
B
B
The
three
paths
is
this:
one:
S3
console
has
the
link
to
create
the
S3
path.
So
let
me
use
this
way
then
copy
paste,
the
location.
This
is
the
S3
location.
That
is
something
I
use
in
a
previous
demo.
So
in
the
visual
you
in
the
interface
I
ingested
or
the
sample
data
set
into
this
location.
Now,
let
me
close
this
location
using
chlora.
B
The
important
thing
is
here
so
I
need
to
check
enable
write
manifest
to
make
it
clearable
from
both
Latin
and
redshift
specter.
Okay,
now
this
location
and
this
time
I
need
only
one
data
source,
then
next
other
than
that
I
need
to
pass.
The
IEM
role
to
this
program
then
choose
the.
B
C
B
Okay,
now
the
chlora
is
trying
to
read
from
the
S3
location
in,
for
the
schema
then
create
a
table
definition
and
also
the
Manifest
file
that
is
based
on
simulink
text
format.
B
A
C
C
B
B
B
And
you
can
see
the
table
content
here,
so
this
is
very
quickly,
therefore,
but
you
can,
as
you
can
see,
that
I
think
I
was
able
to
query
the
data
table
or
that
is
based
on
similar
exist
format
created
by
group
chlora.
This
is
a
sample
record
coming
from
the
public
S3
a
bucket
okay.
As
you
can
see,
this
is
queryable
from
Auto
Athena,
but
the
same
table
is
curable
from
Amazon
redshift,
so
you
can
choose
any
of
the
these
query
engine
based
on
your
requirement.
B
B
Here
we
have.
Oh
query,
editor
button
two,
so
this
is
connecting
to
the
let's
use
serverless
cluster.
B
B
B
Foreign
is
shown
up
in
the
redshift
console
too
so,
but
based
on
the
requirement,
you
can
choose
any
of
Amazon,
Tina,
Amazon,
great
shift
and
any
other
engines
as
ever
Yep.
This
is
a
pretty
much
everything
I
prepared
for
today's
demo
or
so,
as
you
can
run
from
this
demo,
you
can
easily
create
a
table.
Definition
from
group
Rover
then
use
the
Amazon
redshift
Amazon
Tina
to
query
the
Delta
table.
A
Awesome
Nori:
this
is
great.
I
love
the
fact
that
you're
showing
how
easy
it
is
to
use
glue
with
Athena
and
red
Chef
to
go
ahead
and
work
with
Delta
tables.
We
got
a
bunch
of
questions,
I'm
actually
gonna,
even
though
I
answer
some
of
those
questions
I'm
actually
going
to
go
back
to
it.
But
I
did
want
to
ask
the
audience
that
if
you
do
have
questions,
please
chime
in
to
LinkedIn
Zoom
q,
a
or
YouTube
I'm
actually
going
to
be
looking
at
all
three.
A
So
you're
probably
wondering
why
I'm
zooming
back
and
forth
It's,
because
I
got
two
monitors.
I
have
to
go.
Look
at
all
the
questions,
but
I'm
going
to
start
with
the
ones
some
of
the
ones
that
people
have
already
asked
just
to
provide
a
little
context
for
everybody,
okay,
sure
and
oh
by
the
way,
I
know
you're
in
the
Q
a
section,
but
I
did
think
you
won.
You
had
two
additional
slides,
one
on
re
invent
and
one
on
the
book.
Why.
A
B
Okay,
so
let
me
add
two
more
thread,
so
today's
demo
is
based
on
the
past
blog
post
that
is
authored
by
me
and
the
other
co-authors.
So
if
you
are
interested,
you
can
go
and
visit
this
blog
post
and
check
out
any
of
them
to
try
the
same
instruction
in
your
account,
and
we
have
reinvent
that
it's
a
religious
event
in
AWS
that
is
coming
soon
in
the
AWS
re
invent
or
but
that
happens
at
the
end
of
this
month.
I
have
two
sessions
there.
B
So
if
you
are
interested
or
if
you
are
coming,
then
please
come
to
my
session
and
the
red
stock
there
yeah
that's
the
two
two
thread:
I
have.
A
No
problem:
that's
perfect!
Okay,
all
right!
So
let
me
let
me
start
by
asking
some
of
the
questions
that,
like
the
that
that
came
through
okay,
let's
think,
oh
right.
For
example,
one
of
the
questions
that
actually
had
to
do
with
blue
in
the
Crawlers
is
that
basically,
it
is
the
is
there
crawler
running
in
the
background
to
collect
the
info
on
the
Delta
tables.
What
you
demoed
basically
was
it
running,
so
I
presume
that
you
can
basically
have
this
continue
running
as
their
data
is
being
loaded
into
the
Delta
table.
B
That's
good
good
question,
so
Gloria
can
be
run
by
a
manual
operation
or
a
schedule
or
event
driven
basis.
If
you
want
to
or
keep
the
your
Delta
manifest
fire
Very
Fresh,
then
you
can
configure
a
very
short
interval
right
there,
every
five
seconds,
five,
five
minutes
or
whatever.
If
you
want
to
or
the
Randy
crore
for
the
all
the
operations
you
have,
then
you
can
configure
S3
event
with
secluda
so
that
you
can
reflect
all
the
changes
to
the
table
definition.
So
it
will
be
a
better
for
the
real-time
use
case.
A
Perfect
so
then
related
to
that,
in
terms
of
the
real
that
sort
of
real-time
use
case-ish,
typically
with
the
Manifest
right.
That's
not
it's
you're
not
going
to
be
able
to
refresh
that
fast
enough.
Potentially,
so
the
question
I'm
wondering
about
is
that
it
does
glue
automatically
take
into
account
like
if
the
Manifest
gets
refreshed.
Then
it's
automatically
refreshes
in
the
glue
catalog
as
well
or
is
it
just
really
part
of
the
job
itself.
A
Gotcha,
so
so,
then,
basically
what
happens
just
to
clarify,
because
I
actually
did
answer
a
question
that
I
wasn't
complete
for
some
folks
is
that
you
can
actually
Pro
read
data,
that's
streamed
into
a
Delta
table,
provided
that
basically
you're
allowing
glue
to
basically
manage
the
Manifest
and
manage
the
update
of
the
schema.
So
that
way,
when
you're
querying
the
data
through
glue
right,
it
is
the
one
that
defines
what
the
scheme
is.
Is
that
a
correct
assertion?
Yes,.
B
A
So,
with
the
capabilities
I'm
going
to
switch
over
to
LinkedIn
and
I'm,
going
to
switch
over
to
zoom
and
I'll,
go
back
so
sorry
to
YouTube,
then
I'll
go
back
to
zoom
so
from
LinkedIn.
Miguel
has
the
question
so
with
this
capacity
that
we're
seeing
right
now
is
Lake
formation
security
also
available
for
this
as
well.
Yes,.
B
Yeah,
the
information
access
control
is
already
integrated
with
the
Delta
recruiter,
so
you
can
configure
that
any
of
the
table,
level,
Chrome
level
and
row
level
or
security
using
break
formation.
A
B
So
country
country
Athena
can
create
a
manifest
table,
so
it
is
a
just
a
snapshot
of
the
Delta
table
so
yeah,
that's
the
current
behavior.
A
C
A
B
Yeah,
as
you
can
imagine,
we
have
so
many
interesting,
exciting
announcements
at
three
invents,
so
looking
forward
to
seeing
you
at
the
beginning.
A
Exactly
exactly
all
right
perfect,
there
is
another
question
from
Sandeep
from
this
one's
back
to
the
our
Zoom
here
it
so
he
originally
asked
is
the
the
visual
ETL
able
to
generate
optimized
code,
optimize
spark
code
and
I'm.
Just
curious
like?
Is
that
something
that
the
that
the
glue
Studio
takes
care
of
as
well
or
is?
This
is
more
like
you're
sort
of
responsible
for
going
and
doing
that
yourself.
B
Yeah,
so
Guru
Studio,
visual
user
interface
will
automatically
create
a
spark,
a
bunch
of
spark
code.
So
let
me
show
you
how
it
works.
Perfect!
That's
right!
Let
me
go
back
to
Google
studio.
B
Okay,
this
is
the
console
I
have
used
for
the
demo
and
this
one.
This
one
is
a
new
table
created
in
the
demo.
But
if
you
click
this
script
tab,
then
you
will
see
the
auto
generated
code
that
is
coming
from
the
this
visual
dag.
So
if
you
write
this
feature
that
then
Google
studio
will
populate
this
script
automatically.
B
A
Cool
all
right,
I
think
you
definitely
answer
that
question.
Another
question
is
the:
if
you're
using
since
you're,
using
an
S3
bucket
here,
can
you
actually
use
a
Kafka
topic
as
a
source
as
well
when
you're
using
glue
Studio?
Yes,.
B
A
Perfect-
and
you
know,
we've
got
time
for
one
more
question,
and
so
this
is
also
a
glue,
specific
question,
so
loving
the
fact
that
we're
actually
giving
some
love
to
our
friends
in
AWS
on
glue
are
there
mats
asking
the
question:
are
there
any
plans
to
support
newer
versions
of
spark
past
3.1,
since
the
that
basically
does
limit
Delta
Lake
to
version
1.0
right?
So,
in
other
words,
for
sake
argument
are
you
know,
support
spark
3.3
which
allows
us
to
do
spark
2.1,
for
example?
B
B
A
B
A
Yes,
and
so
so,
in
other
words,
we're
not
trying
to
not
answer
your
questions,
ladies
and
gentlemen,
we,
but
we
are
calling
out
they're,
going
to
be
some
really
great
announcements
at
the
end
of
the
month.
Part
of
the
reason
why
we
did
the
session
now
and
I'm
sure,
Nori
and
I
can
have
a
soft
food
session.
After
remember,
we
can
update
some
of
our
answers
for
y'all.
Sure,
yes,
perfect!
Well,
hey
I!
Think
that's
it
for
today.
A
So
in
terms
of
questions,
my
apologies,
if
I'm
not
able
to
answer
all
of
your
questions
but
I
think
we
covered
most
of
them,
but
in
case
you
can't
just
go
ahead
and
join
us
at
go.delta,
dot,
IO
slack
and
just
ask
your
questions
there,
because
we're
actually
online
on
on
that
slack.
Answering
questions
as
well,
so
saying
that
Nori
I
really
appreciate
you
taking
the
time
to
answer
everybody's
questions
and
I.
Think
I
want
to
that's
it
for
today
and
oh,
yes,
perfect
leave
it
at
the
screen.
A
If
you
want
to
ask
Nori
questions
directly
as
well,
you
can
ping
him
directly
on
GitHub,
Twitter
and
Linkedin,
but,
like
I
said,
you
can
also
join
us.
Go
dot,
Delta,
dot,
IO
slack!
If
you
want
to
just
ask
a
bunch
of
Delta
Lake
questions,
but
hopefully
you
guys
enjoyed
this
awesome
session,
showcasing
glue,
Athena
and
rest
shift
with
Delta
Lake
and
with
that
Dory.
Thank
you
very
much.
Thank.