►
From YouTube: WEBRTC WG interim March 15, 2022
Description
See also meeting minutes https://www.w3.org/2022/03/15-webrtc-minutes.html
02:44 TPAC 2022
06:02 WebRTC-SVC
13:11 WebRTC-Extensions
32:50 Avoiding the “Hall of Mirrors”
53:56 Display Surface Hints
1:05:06 getViewportMedia update
1:10:38 MediaCapture Extensions proposals
B
Okay,
so
this
is
the
march
15
2022
meeting
of
the
w3c
webrtc
working
group,
the
group
of
bytes
by
the
w3c
patent
policy
and
only
companies
and
people
that
are
listed
on
this
status
webpage
are
allowed
to
make
substantive
contributions.
B
So
for
the
future
we
are
going
to
meet
also
on
april
19th
and
may
17th
at
the
same
time,
and
that
is
up
on
the
w3c
to
webrtc
working
group
wiki.
Okay,
a
little
bit
about
the
meeting
the
slides
have
been
published
to
the
wiki.
We
do
need
a
scribe
to
figure
out
when
we
make
decisions
and
write
that
down.
Do
we
have
someone,
volunteering.
B
Okay,
thank
you
dom.
As
you
heard,
the
meetings
being
recorded
and
the
recording
will
be
made
public
a
little
bit
about
the
code
of
conduct.
We
do
operate
under
the
w3c
code
of
ethics
and
professional
conduct,
we're
all
passionate
about
improving
rubber
to
see.
But
let's
try
to
keep
it
cordial
and
professional
a
little
bit
about
the
meeting
tips.
C
B
Canceling
speaker
phone
wait
for
the
microphone
access
to
be
granted
before
speaking
and
state
your
full
name.
So
we
can
get
you
in
the
minutes.
B
We
probably
will
use
the
pull
mechanism
today,
so
to
figure
out
the
sense
of
the
room
or
just
get
input
from
people
a
little
bit
about
document
status,
just
because
something's
hosted
in
wc
repo
doesn't
imply
adoption
by
the
working
group.
That
requires
a
call
for
adoption
as
you've
seen
on
the
mailing
list.
Editors
drafts
do
not
represent
working
group
consensus,
but
the
working
group
drafts
do
once
they're
confirmed
by
the
call
for
consensus,
it's
possible
to
merge
prs
that
lack
incense
is
if
a
note's
attached.
B
That
indicates
that
okay,
so
now
for
our
first
poll-
and
I
guess
we'll
need
a
little
help
from
harold
and
the
question
is:
are
you
considering
attending
in
person
the
tpac
2022
and
the
possible
answers?
Are
yes?
No
and
you
have
no
idea.
A
Let
me
introduce
quickly
why
the
question
is
here,
so
you
know
that
the
past
two
tea
packs
due
to
the
pandemic
have
been
purely
virtual.
With
some
of
the
pandemic
impacts
diminishing
in
some
region.
There
is
discussion
about
having
a
hybrid
tea
park
this
year
in
vancouver
week
of
september.
A
The
question
is
whether
there
would
be
enough
people
to
justify
all
the
work
and
expenses
needed
to
get
such
a
meeting
set
up
and
running
in
september,
and
so
I
guess,
we've
been
asked
by
the
event
organizers,
whether
or
not
we
think
people
would
show
up
and
to
what
extent.
So
this
is
just
trying
to
gauge
how
many
of
us
feel
today
that
they
would
likely
come
likely
not
come,
or
they
are
simply
too
much
in
the
dark
to
express
a
useful
opinion.
D
D
F
B
B
Okay,
all
right
so
for
discussion
today
the
agenda
as
much
as
I
talked
about-
and
this
is
roughly
the
time
allocation,
so
we
we
to
try
to
keep
to
the
time
I
guess
I'll
be
talking.
So
maybe
harold
you'll
have
to
ring
me
in
if
I
go
over
okay,
so
here
are
the
pr's
and
issues
we're
going
to
talk
about
one
in
weber,
csvc
and
two
and
robert
c
extensions.
B
So
in
webrtc,
svc
issue.
68
relates
to
the
behavior
of
get
parameters,
and
specifically
the
text
that
was
there
was
unclear
about
renegotiation,
so
it
talks
about
before
negotiation
as
completed
and
then
after
and
the
question
is
okay.
What
does
this
mean
about
renegotiation?
Renegotiation,
like
you
did
an
initial
negotiation.
Now
you're
renegotiating?
B
You
know,
it
seems
to
say
it's
kind
of
unclear
what
what
happens
during
a
renegotiation
and
what
you
know
what
values
you
get
back.
B
So
what
I
would
like
to
propose
here
I
mean
to
the
extent
you
can
look
at
this
and
give
any
reaction
is
the
following
text,
where
it's
clarified
that
we're
really
talking
about
initial
negotiation
and
and
after
the
initial
negotiation.
So
basically,
what
I
think
happens
is
after
initial
negotiation.
Get
parameters
will
always
return
the
currently
configured
scalability
mode
for
each
encoding,
and
that
that
includes
so,
if
you
do,
an
initial
negotiation
and
renegotiating
you're
still
gonna
get
the
currently
configured
scalability
mode,
and
so
this
is.
B
This
is
the
clarification
that
I
I'd
like
to
put
into
the
spec.
So
basically,
before
the
initial
negotiation
has
completed,
you
get
the
scalability
mode
that
was
last
set
by
and
transceiver
and
set
parameters.
It
may
not
be
the
one
that
you'll
see
after
negotiation
is
completed,
because
you
know
the
codec
might
not
support
the
mode
that
you
put
in
and
if
you
don't
provide
anything
then
or
if
it
wasn't
successfully
set,
then
you
don't
get
you
don't
get
a
value
for
that
encoding.
B
B
G
G
B
Well,
yes,
I
think
what
you're
talking
about
yanivaris?
Let's
say
you
called
set
codec
preferences
to
change
the
preference
order,
like
that's
a
a
pretty
interesting
one
right
and
now
your
renegotiation
renegotiating.
Like
here's
a
weird
example,
you
were
using
vp8
and
you
got
you
were
using,
for
example,
l1
t2.
Now
you
renegotiate
to
h264,
which
doesn't
support
any
scalability
happens.
B
So
that's
kind
of
a
very
concrete
example
where
you,
after
you
finish
the
vp8.
Your
get
parameters
is
returning
l1
t2.
Now
you
call
psychotic
preferences
at
what
point
do
things
change.
G
Yes,
even
without
get
calling
said
codex
preferences,
I
think
there's
a
question
of
that.
Get
parameters
may
return
different
information
depending
on
whether
a
negotiation
renegotiation
is
currently
happening
or
not,
because
the
remote
offer
could
can
well
wait
if
you
have
a
local
offer
that
might
affect
the.
If
you
just
changed
something
in
the
local
offer.
B
Right
so
harold,
can
we
just
talk
about
that
specific
case
that
I
mentioned
switching
bpa
for
h264?
B
So
what
do
you
think
happens
during
the
renegotiation
before
I?
I
get
confirmation
of
the
change
to
h264,
I'm
still
going
to
get
l1
t2
back
right.
B
Right
yeah,
that's
what
I
would
expect
and
I
wouldn't
yes,
I
mean
that's
what
I
think
this
text
is
trying
to
say,
which
is
just
you
know,
while
you're
in
progress
nothing's
changing
right.
Just
because
I
called
set
code
at
preferences,
I
sent
an
offer
nothing's
changing,
I'm
still
getting
the
I'm
still
acting
as
though
it
was
vp8
and
l1
t2
as
soon
as
so.
I
think.
If,
if
what
we're
saying
is
correct,
then
then
the
text
does
make
sense
right
how
about
after
the
initiation,
it's
always
the
currently
configured.
One.
B
And
so,
like
you
said
when
it's
when
you're
actually
sending
h264
it
switches
to
l1
t1
anyway,
I
think
yanivar.
If
you
could
write
up
your
concern
in
on
in
issue
68,
we
can
kind
of
go
through
it
in
more
detail
to
see
if
there's
a
problem
here,
okay
issue
68
on
in
weber,
csvc,
okay,
yeah,
just
try
to
trying
to
figure
out
if
there's
just
we'll,
try
to
go
into
more
detail
and
see
if
there's
a
problem.
B
Okay,
so
I
think
that's
the
resolution
is
janet-
will
continue
the
discussion
in
in
the
issue.
Okay,
so
now
we're
on
issue
98.
This
is
we've
now
switched
over
to
webrtc
extensions,
and
this
issue
is
relating
to
disabling
hardware
acceleration,
so
fippo
provided
a
long
list
of
hardware,
acceleration
implementation
bugs
that
we've
experienced,
and
it's
really
quite
quite
an
incredible
little
list,
and
I
won't
bore
you
with
all
of
the
details,
but
basically
hardware,
accelerations
changes
are
very
hard
to
test.
B
They
often
create
problems,
and
so
the
question
is
whether
we
we
can
provide
a
way
to
disable
hardware
acceleration.
An
example
of
this
is
in
web
codecs.
Today,
for
example,
you
in
the
video
encoder
config
dictionary,
there's
a
hardware
acceleration
member
and
it
can
take
a
value
of
no
preference,
prefer
hardware
preferred
software,
at
least
as
it's
been
implemented.
This
is
actually
not
really
a
hint.
The
spec
says
it's
a
hint,
but
it's
not
really
so,
if
you
say
prefer
hardware,
you're
gonna
get
harder
or
robust
prefer
software
again
software
robust.
B
B
When
I
was
thinking
about
set
parameters,
it
seemed
like
this
wasn't
necessarily
a
great
idea,
because
it
would
have
the
limitation
that
you,
as
we
said
in
weber,
cpc.
You
can't
change
the
envelope
negotiated
by
offer
answer
within
set
parameters,
and
the
problem
is,
if
you
had
a
hardware,
only
codec
like
a
higher
profile
of
av1
or
something
that
could
only
be
done
in
hardware,
and
you
try
to
disable
hardware.
Acceleration
you'd
have
to
throw
there
or
same
for
software.
If
it
was
software
only
you
couldn't
say.
B
B
So
I
have
some
concerns
about
whether
we,
whether
set
parameters,
makes
any
sense
for
this.
The
other
way
to
go
about
it
would
be
to
try
to
do
it
in
set
coded
preferences,
so
it
would
basically
change
your
negotiation
preference.
You
then
potentially
would
disable
some
profiles
like,
if
you
said,
prefer
software,
and
there
were
a
few
profiles
that
were
only
in
hardware.
Those
would
disappear
from
create
offer
from
the
sdp,
then
you'd,
renegotiate,
and
so
whatever
you
got
would
be
something
you
could
actually
support,
and
some
of
these
complexities
wouldn't
happen.
B
So
one
specific
idea
is
to
add
a
hardware
acceleration
member
to
the
rtc
rtp
coda
capability
dictionary
and
that,
as
I
said,
would
influence
the
codec
and
profile
combinations
that
would
show
up
and
create
author
and
great
answer,
and
if,
if
you
set
to
something
to
prefer
software,
you
wouldn't
get
any
profiles
that
depended
only
on
harvard
and
then
the
other
related
question
is
what
would
happen
in
rtc
rtp
coder
capability?
How
would
you
discover
this
and
media
capabilities
api?
B
I
guess
there's
issue
185,
which
is
allowing
you
to
retrieve
the
codec
capability
from
media
capabilities.
I
guess
johannes
can
comment
on
on
how
that's
going
and
then
the
question
would
be.
If,
if
we
did,
that,
would
you
have
a
hardware
acceleration
member
be
returned,
or
would
you
just
do
what
we
do
today,
which
is
return,
smooth
and
powerful
and
then
supported,
and
that
would
be
enough
info.
D
So
I
have
yandiva
on
the
queue
and
dom
has
raised
his
hand.
A
So
I'll
go
further,
I
guess
the
first
question
is:
do
we
really
expect
developers
to
be
keeping
track
of
this
hardware,
implementation
bugs
and
so
to
be
doing
the
job
of
saying
yes
or
no?
I
want
hardware
acceleration,
wouldn't
it
be
more
efficient
if
browsers
were
in
practice
doing
that
if
they
know
that
hardware
is
buggy
again,
I
it
feels
hard
to
be
saying
that
all
developers
need
to
keep
track
of
all
the
way.
Hardware
can
go
wrong.
B
Well,
this
came
up
because
this
stuff
is
very
hard
to
test
and
it's
actually
slipped
through
and
disabled.
You
know
large
scale,
implementations.
A
H
I
want
to
add
also
that
sometimes
decoders
work,
fine,
unless
you
are
paired
with
a
specific
encoder
which
might
not
be
very
common.
So
in
some
situations
that
might
you
might
not
have
any
problems,
and
you
just
want
to
disable
the
encoder
on
very
specific
cases,
but
not
just
say,
I'm
going
to
disable
decoder
for
everyone,
because
right
there.
I
H
Be
someone
with
a
broken?
We
have
an
encoder
of
a
generator
stream
that
cannot
be
decoded.
B
Yeah,
so
that's
kind
of
the
next
question
florence
is
like
if
you.
C
G
Yes,
so
for
set
parameters,
I
do
believe
we
have
read
only
properties
and
we
have
ways
to
to
to
say
when
prop
produce
can
be
said
or
not.
So
I
don't
my
concern
with
putting
it
in
rtc.
Rtp
coded
capability
is
that
it's
actually
information.
That's
returned
from
get
capabilities,
but.
B
B
Acceleration
field
yeah,
it
would
just
be
three
profiles
you
have
well.
Actually
I
wouldn't
expect
the
question.
The
other
question
is:
would
you
even
return
this
at
all
and
that's
a
separate
thing
you,
you
might
not
actually
return
it
from
get
capabilities.
G
It
doesn't
seem
to
fit
well
because
the
way
you've
shown
the
the
web
ideal
here-
you're
you're,
adding
a
hardware,
acceleration
preference
to
information
about
a
codec,
and
it's
not
the
codec.
That
has
a
preference,
but
it's
and
also
we
just
spent
a
large
amount
of
time
moving.
Some
of
the
fingerprinting
concerns
over
the
media
session.
So
and
we
were
talked
about
retiring
the
ability
to
detect
hardware
in
webrtc,
so
we've.
B
G
Yes,
but
we've
lumped
in
the
the
sensitive
information
into
the
existing
category
of
media
sessions
to
get
it
out
of
the
webrtc
working
group
domain
right.
So
now
that
we're
reintroducing,
I
want
to
make
sure
we
don't
reintroduce
potential
concern
that
people
might
have
again
if,
unless
there's
a
good
reason-
and
maybe
it
would
be
better-
it
doesn't
seem
necessary
to
include
that
information
here
if
we
could
instead
phrase
it
as
a
preference
which
I
think
is
easier
to
do
in
set
parameters.
C
Yeah
thanks,
so
I
guess
the
way
I
understand
this
is
that
it
should
just
be
some
way
for
like
a
short
term.
Okay,
there
was
a
bug
in
the
browser
and
then
the
developer
could
then
disable
the
hardware
encoding
or
decoding
for
some
short
player
down
until
it's
fixed
in
the
browser.
C
So
I
think
from
that
sense
I
wonder
well
to
me
it
seems
like
it
maybe
doesn't
have
to
influence
that,
like
the
melee
capabilities
api,
what's
returned
from
that,
it's
more
like
okay!
This
is
something
instead
of
having
everything
everything
broken
you
get
like
like
some
kind
of
recovery
mode
or
something.
C
But
I
I
agree
also
that
it
seems
a
bit
hard
for
two
developers
to
use
it,
but
I
understand
the
concern
that
it
it's
very
tricky
to
test
it.
So
that's
a
sense.
I
understand
that
it
may.
E
D
D
D
B
D
D
F
D
C
C
Yeah
so
that
you
quickly
can
get
out
this
kind
of
information.
Okay.
Now
we
have
a
certain
bargain
in
this.
For
this
particular
hardware,
encoder.
D
J
Yeah
we
do
here
so
I
say
that
yes,
the
the
block
list
which
is
used
in
chrome,
something
like
that
gpu
driver
puglits,
just
json.
J
Those
are
the
ones
which
we
use,
give
information
to
our
driver
teams,
because
most
of
this
issues
are
driver
issues,
as
you
know,
and
our
driver
teams
basically
look
at
that
json
file
inside
the
chrome
browser
and
try
to
fix
a
platform
by
platform.
So
yeah.
B
Okay,
I
think
we're
running
out
of
time.
Do
we
have
a
kind
of
recommended
resolution
other
than
to
go
to
github
and
continue
to
noodle
on
it?
There.
D
Seems
to
me
we
don't
have
a
resolution.
We
have
a
couple
of
approaches
beyond
the
ones
that
were
that
have
already
been
outlined,
so.
B
Okay,
so
I'll
give
the
floor
to
you,
harold
for
issue
99.
D
Yep
so
now
I'm
not
watching
the
queue
anymore.
So
this
is
about
the
rtc
rtp
header
extension
capabilities
that
we're
hoping
that
we
can
get
out
to
our
shipping
soon
so
scenario:
you
have
an
implementation,
it
supports
snazzy
extension
number
five.
It
provides
dancing
videos
by
default,
not
seven
offers,
since
the
browser
is
capable
of
it.
It's
listed
and
can
build
this,
but
when
you
create
an
offer,
it's
not
there
and
you
don't
know
why.
D
C
D
B
Well,
bernard
yeah,
I
I
think
it's
a
convenience
in
the
in
the
use
case.
You
describe
right
there
and
there
will.
There
will
be
scenarios
like
this,
where
you
don't
want
to
set
it
on
by
default,
and
yet
you
would
like
to
understand
why.
B
D
I
think
you
can
shame
this
by
doing
doing
some
some
negotiation
creating
an
offer
and
dancing
this
and
doing
a
dance
by
with
a
throwaway
pair
connection,
so
that
you
can
actually
query
the
result
of
negotiate
negotiating
it.
D
A
D
A
I
guess
my
sense
is
that
if
there
is
a
big
demand
for
it,
it
sounds
cheap
to
do
if
there
is
no
demand
for
it.
This
sounds
expensive
to
do
so.
Maybe
it's
a
matter
of
getting
the
voice
of
developers
to
say
how
big
an
issue
this
is
for
now
and
like
the
idea,
I
think,
could
be
brought
up
later.
If
it's
not
a
narrow
problem,
and
if
it
is,
then
we
can
iterate.
A
D
C
B
E
Perfect,
so
I'm
here
to
talk
about
the
hall
of
mirrors,
which
is
what
happens
when
you
capture
a
surface,
draw
it
back
to
the
surface
from
which
you're
capturing
it
and
capture
it.
Recursively
like
that,
so
some
of
us
have
already
tried,
for
example,
sharing
a
single
monitor
and
you're.
E
You
see
a
preview
of
the
monitor
or
a
window
or
a
tab,
and
specifically,
I'm
mostly
thinking
about
tab,
but
this
applies
to
all
three
of
those
next
slide,
please
so,
as
mentioned
you
can
get
into
that
for
in
several
ways
and
the
problems
that
this
creates
are
that
it
confuses
the
local
user.
It
confuses
the
remote
and
remote
users
and
it
can
potentially
even
produce
mic
hull,
which
is
when
you
get
a
feedback
loop
on
your
mic
and
or
on.
You
know,
on
the
whatever
you're
capturing
and
nobody
benefits
right.
E
It
usually
happens
by
mistake,
except
when
it's
not
actually
a
mistake,
but
then,
during
during
video
conferencing,
if
you
capture
the
current
tab,
that's
usually
a
mistake.
Next
slide,
please
just
a
second
award
before
that.
So
one
of
the
reasons
that
I'm
focusing
on
tab
is
that
if
you're
capturing
the
current
win
the
current
monitor,
then
that
is
usually
not
a
mistake
right.
Usually
the
intention
is
okay,
you
capture
the
entire
monitor,
but
then
you
quickly
switch
away
from
the
video
conference
to
something
else
and
that's
a
bit
more.
E
That's
a
more
interesting
case,
a
bit
more
a
bit
more
delicate.
So
I'm
only
talking
about
when
you
capture
the
current
tab.
E
So
one
solution
that
you
could
think
of
is
you
could
say:
okay,
so
how
about
the
user
agent
just
don't
offer
the
current
tab,
and
I
argue
that
this
would
be
a
mistake,
because
there
are
legitimate
applications
that
offer
the
user
to
capture
the
current
tab
and
that
is
actually
desirable.
So,
for
example,
if
you
want
to
take
a
screenshot
of
the
current
application
and
file
feedback,
I
know
there
are
some
applications
that
are
experimenting
exactly
with
that.
E
So
you
would
want
to
allow
the
user
to
capture
the
current
tab.
Also,
if
you're
recording
something
like,
for
example,
if
google
me
didn't
offer
recording
on
the
cloud,
one
of
us
might
have
recorded
this
entire
meeting,
using
using
current
tab
capture
and
so
long
as
we
didn't
actually
that
the
person
did
not
preview,
the
whatever
they
were
capturing
back
to
the
screen.
E
And
I
think
that
if
we
remember
the
display
media
gets
this
dictionary
and
we
can
extend
the
dictionary
with
another
member,
it
can
be
include
current
tab.
It
can
exclude
current
tab.
It
could
be
a
couple
of
other
shapes
that
we
can
discuss
in
soon.
It
can
have
a
default
volume
and
it
doesn't.
But
the
point
is
it's
going
to
be
a
control
that
allows
you
to
hint
to
the
user
agent,
whether
you
want
to
see
the
current
tab
as
one
of
the
options
or
not
and
next
slide
please.
E
So
then
we
come
to
the
inevitable
security
discussion,
and
that
is
that
whenever
we
think
about
influencing
user
selection,
a
lot
of
red
red
lights,
turn
on
and
star
and
ominous
music
starts
playing,
and
we
need
to
ask
if
there's
a
reasonability
this
time.
Is
that
something
that
we
should
worry
about
this
time?
And
I
would
argue
that
in
this
case
this
is
not
actually
problematic,
because
the
problem
with
influencing
user
selection
comes
in
two
flavors.
E
So
the
other
problem
is
when
you,
when
you
influence
the
user
towards
something
that
might
not
be
under
your
control,
but
is
either
inherently
dangerous,
like
maybe
it's
the
entire
screen,
or
maybe
it's
something
sensitive
to
begin
with,
like
a
specific
tab
with
your
bank
account
that
might
not
be
under
your
control,
but
you
won't
get
that,
and
I
argue
that
in
this
very
limited
case
of
saying
I
don't
want
to
capture
the
current
tab,
you're
actually
moving
the
user
away
from
those
two
cases,
in
which
case
the
usual
concerns
we
have
about
influencing
user
choice.
E
So,
then,
assuming
that
we,
assuming
that
we
agree
on
all
of
this,
which
is
an
assumption
that
we
will
soon
test,
we
would
have
to
ask
the
question
okay,
but
what
is
the
default
value
gonna
be?
And
after
a
couple
of
discussions
on
github,
I
think
that
maybe
the
best
thing
to
do
is
just
to
say:
you
know
what
there
is
no
default
value.
E
This
is
a
hint
each
user
agent
will
choose
to
interpret
its
the
lack
of
it
as
they
want,
which
is
the
current
case
by
the
way
right
now,
there
is
not
insane
whether
the
user
agent
must
include
the
current
tab
or
not,
and
we
won't
be
changing
that
next
slide.
Please
and
then
some
foreshadowing,
like
we
won't
jump
directly
into
this.
E
But
if
we
manage
to
make
some
progress
on
this
right
now,
then
maybe
we
can
even
expand
the
scope
and
say:
okay,
do
we
want
something
similar
for
excluding
the
current
window
and
excluding
the
current
screen,
but
this
is
probably
something
that
we
would
want
to
potentially
get
to
if
we
have
time
so
if
you
could
go
one
back,
one
or
two
slides
and
then
I'll
mute
myself
and
listen
for
people
to
people.
G
Yeah,
sorry,
I'm
actually
not
sure
what
we're
deciding
here
I
mean
there's
an
issue.
This
is
issue
209
that
has
more
detail
and.
E
Here
what
I'm
proposing
as
one
slide
backwards?
Please
exactly
so.
I
suggest
that
we
add
either
include
current
tab
or
exclude
current
tab.
It
is
not
important
to
me
which
one
it
is
so
long
as
we
don't
have
a
default
value
and
that
one
serves
as
a
hint.
It
basically
says
user
tells
the
user
agent
that
the
application
is
not
interested
or
is
interested
in
potentially
getting
the
current
tab.
E
G
I'm
listening,
okay
yeah,
so
I
I
like
this
api
other
than
I
think
the
default
need
should
be
false
because
I
think
that's
the
of
all
the
sites
that
use
screen
capture
I
think
most
sites
would
rather
let
me
back
up.
I
don't
actually
think
this
is
about
avoiding
the
hall
of
mirrors.
I
think
that's
a
symptom
that
could
be
mitigated
with
other
things.
G
Like
you
and
mentioned,
you
know,
having
user
agents
block
that
the
view
of
that
or
blur
it
or
do
other
tricks
with
the
video,
but
I
think
the
use
case
is
actually
that
picking
your
own
tab
is
often
undesirable
in
many
applications
like
in
in
this
meeting
here,
module
the
self-recording
thing,
which
I
think
isn't
the
primary
use
case
there,
but
I
think
long
term.
G
We
want
self
capture
to
be
get
viewport
media
and
get
display
media
capture
of
everything
else,
and
I
think
most
sites
would
be
fine
with
that.
So
I
would
argue
that
it's
actually
desirable
to
so.
The
question
comes
down
to.
Yes,
there
are
some
sites
that
that
want
self-capture
to
be
part
of
the
options,
and
I
they
should
be
allowed
to
express
that.
So,
I
think
include
current
tab
is
the
right
boolean
for
that,
but
I
I
think
the
default
is
also
important.
G
G
Do
the
opposite
of
what
the
web
developer
thinks
they're
doing.
So
so,
if
we
wanted
to
have
a
default
behavior
of
true,
which
I
don't.
E
E
We
agree
on
the
on
the
point
of
the
true
and
that's
why
it
could
either
be
include
or
exclude
current,
but
we
need
to
discuss
that
right.
Yes,
I'm
sorry
did
you
want
to
say
something.
E
E
Okay,
so,
first
of
all
about
default.
True,
thank
you
very
much
for
bringing
it
to
my
attention
and,
yes,
I
think
that
we
agreed
there
about
un's
suggestion
of
obfuscation,
as
he
called
it.
That
applies
more
to
whole
screen
capture,
but
not
for
current
screen
capture,
a
current
tab
capture,
because,
if
you're
capturing
the
current
tab-
and
you
obfuscate
that
then
you're
capturing
nothing.
So
that's
a
bit
less
interesting,
so
specifically
for
current
tab,
you
are
correct
that
hall
of
mirrors
is
not
the
only
case.
This
is
more
of
a
symptom.
E
The
case
is
that
we
are
capturing
something
that
you're
not
actually
situated
to
handle
right
like,
for
example,
if
you
want
just
to
work,
we
understand
so,
given
that
you
like
this
api
and
that
the
only
thing
that
we
potentially
disagree
about
is
the
default.
E
I
would
say
that
we
should
probably
not
take
get
viewport
media
into
account
when
making
this
decision,
because
we
have
no
idea
how
long
it's
going
to
take
for
it
to
be
adopted
by
the
w3c
implemented
and
then
adopted
by
web
developers,
and
there
is
a
significant
risk
of
it
being
adopted
by
web
developers,
because
it
requires
two
different
mechanisms
which
are
not
terribly
common
nowadays.
So
it
could
be
that
not
many
web
developers
would
be
able
to
use
that
api.
E
So
I
don't
think
that
we
should
use
it
as
a
guiding
principle
just
yet.
I
do
think
that
we
should
avoid
breaking
current
applications
that
potentially
have
millions
of
users,
even
if
it's
only
for
two
days,
I
think
that's
undesirable,
and
I
think
that,
for
that
reason
we
should
just
not
have
a
default
value.
E
G
That
doesn't
change
anything
actually
because
it's
a
boolean
so
there's
going
to
be
a
default
behavior.
But.
G
To
clarify
that
I
can,
I
can
separate
this
from
what
implementation
they're
doing,
because
implementations
excuse
me,
user
agents
are
already
allowed
to
to
provide
any
choices.
They
want.
It's
totally
up.
The
specification
doesn't
add
any
limitations
there.
What
we're
talking
about
here,
I
think,
is
a
hint
from
the
application
and
the
shape
of
that
hint.
We
can
either
find
out
which
applications
want
the
current
tab
or
we
can
find
out
which
applications
don't
want.
G
The
current
tab,
independent
of
how
the
browsers
work-
and
I
argue
that
out
of
100
applications,
I
think
99-
would
not
want
expect
or
have
any
use
for
capture
of
the
current
tab
and
one
application
would
have
used
for
it.
So
I
think,
would
be
better
to
have
the
default
behavior
applied
to
the
99
and
have
the
one
application
include
specify
include
current
tab.
Please
true.
E
I
I
don't
know
how
you
go
to
the
numbers
of
99
out
of
100.
I
think
that
if
we
spoke
to
one
of
those
applications
that
do
want
it,
they
would
give
different
estimations.
I
can
tell
you
that
I've
got
some
histograms
and
self-capture
happens
millions
of
times
per
year.
Presumably
a
significant
portion
of
that
is
not
accidental.
I
Yeah
so,
first
about
the
security
issues,
I
would
say
in
general
that
security
issues
related
to
tap
capture
in
the
current
spect.
The
current
spec
is
not
really
dealing
with
it.
It's
very
light,
and
it's
it's
start
it's
starting
to
to
feel
I
I'm
starting
to
feel
some
pain
there,
because
we
there
are
some
proposals
that
are
more
and
more
trying
to
control,
to
allow
webpage
to
control
what
the
browser
will
will
show.
I
So
what
the
user
will
pick
and
the
more
we
are
adding
in
terms
of
search
control,
the
more
we
need
to
provide
guidelines
to
precisely
tell
hey
that
catcher
is
tap.
Capture
is
very
specific,
so
you
need
to
deal
with
it
in
in
very,
very
good
ways,
and
we
we
already
need
to
to
provide
more
guidelines
there.
So
if
we
proceed
with
this
to
me,
we
in
parallel
really
need
to
provide
more
guidelines.
I
I
My
understanding,
though,
is
that
I'm
not
pretty
clear
about
it,
but
some
implementations
may
see
the
hint
and
then
they
might
remove
entirely
the
possibility
for
the
user
to
actually
select
the
tab
and
if
so,
that's
something
new,
because
currently,
as
I
understand
it,
the
hints
are
only
allowing
user
to
to
push
together
towards
what
is
the
more
meaningful
choice,
but
still
the
user,
with
a
few
clicks,
can
change
like?
Oh
no,
I
already
don't
want
to
tap
capture.
I
actually
want
window
capture
and
that's
still
something
feasible
with
the
current
hints
with
this
hint.
I
It's
not
clear
to
me
whether
we
will
change
this.
That's
at
some
point
I
want
to
highlight
as
well
and
and
third
related
to
hall
of
mirror.
I
don't
think
that
this
is
solving
whole
of
mirror
at
all.
Clearly,
it's
it's
it's
a
different
problem.
I
I
If
you
have
it
and
you're
not
happy
with
having
the
whole
current
tab,
because
it's
doing
a
half
mirror,
you
can
always
group
to
what
what
is
meaningful
in
your
page
or
ask
the
user
to
to
select
again,
which
would
be
really
sad.
So
I
really
think
that
for
hall
of
mirror,
we
should
keep
the
issue
open
and
pursue
it
as
a
follow-up,
and
there
might
be
some
efforts
or
some
some
energy
that
we
could
provide
there.
I
As
I
said,
according
the
use
case-
and
it's
if
it's
only
a
hint,
then
that's
fine
in
the
github
issue
or
whether
at
some
point
we
would
say
you
must
remove
pm3,
and
I
really
think
we
should
not
go
there
currently.
So
as
long
as
it's
certain,
then
what
a
clarification
would
be,
for
instance,
is
chrome
planning
to
remove
the
entry
all
together
if
the
boolean
flag
is
true.
If
that's
the
case,
maybe
we'll
go
in
with
interrupt
issues
at
some
point,
because
safari
might
not
do
the
same.
I
So
that's
that's
the
tiny
bit
there
where
I
would
be
a
little
cautious
about
this
api
but
interviews.
Why
is
that
fine.
E
In
order
to
answer
the
uncertainty
that
you've
just
voiced,
could
we
go
too
to
slide
forward
please
so
I'm
suggesting
the
bold
text
at
the
bottom,
and
that
is
most
definitely
a
hint,
and
I
understand
that
maybe
we
need
to
rephrase
it
a
bit.
But
what
do
you
think
about
this
general
direction?.
I
I
As
long
as
we
provide
more
security
guidelines
related
to
tap
capture
to
the
difference
between
current
tap
capture
and
over
tap
catches,
and
so
on,
I
would
I
will
be
supportive.
Is
it
possible
and
we
keep
the
issue
open
for
actually
solving
our
smurfs
as
well.
E
So,
of
course,
we
can
keep
the
issue
because
this
only
partially
addresses
it
mitigates,
so
we
can
definitely
do
more
work.
Is
it
possible
for
you
maybe
to
kick
start
the
effort
of
elaborating
security
concerns
that
you
worry
about.
I
I
think
that
the
chrome
team
and
the
chrome
ui
has
done
extensive
work
there.
So
it
would
be
good
if
you
could,
for
instance,
express
what
mitigations
chrome
did
for
tap
capture
because
you're
the
one
that
actually
did
the
implementation
and
you're
the
one
who
are
proposing
to
allow
a
webpage
to
more
more
easily
select
tabs,
which
are
which
have
their
own
issues.
I
So,
and
you
have
done
things
like,
for
instance,
if
the
page
is
navigating
away,
you're,
making
it
clear
in
origins
and
so
on,
and
we
which
are
really
great
information
and
so
you're
way
ahead
of
what
safari
is
doing,
because
safari
is
not
implementing
that
capture.
So
I
can
certainly
find
an
issue,
but
I
think
that
in
terms
of
foods
and
how
how
much
mitigations.
A
Yeah,
so
in
terms
of
boolean
default
values,
a
simple
way
if
this
is
an
issue,
is
to
move
to
a
new
value
which
happens
to
have
two
values,
so
I
don't
think
we
should
get
stuck
on
that
design.
Consideration
I'm
personally
supportive
of
this
hint
as
long
as
it
is
a
hint
really,
then
we
are
just
helping
the
ua
provide
the
best
smoothest
user
experience,
and
so
I'm
supportive
of
that
direction.
E
D
Referring
to
thank
you,
I
think
the
conclusion
of
this
point
was
that
we
we
will
continue
discussing.
It
seems
that
people
who
are
arguing
about
it
want
to
want
to
go
for
making
this
a
hint
and
not
total,
never
show
this
tab,
but
people
are
generally
positive
towards
something
like
this.
E
I
think
intel
was
supposed
to
be
in
between
okay,
I
see
but
meek
right.
I
see
you.
E
Okay,
I'll
try
to
do
it
in
two
minutes
anyway.
Why
not
so
we've
discussed
for
a
long
time
a
similar
issue
to
the
one
just
just
introduced.
So
let's
please
keep
them
separate,
okay.
So
what
everything
we've
discussed
right
now
separate
new
issue.
Another
problem
that
sometimes
happens
is
that
the
application
wants
to
hint
to
the
user
agent
that
it
is
especially
well
geared
to
handle
windows
or
tabs,
or
something
like
that
and
this
there
are
all
sorts
of
ways
that
we
could
do
that.
E
So
generally,
if
you
are
gonna
share
a
browser
better
to
share
a
tab,
and
there
were
a
lot
of
other
reasons
for
why
we
might
want
to
do
that
and
now,
after
nine
months
of
discussing
the
latest
iteration,
I
hope
that
we
could
give
birth
to
this
particular
proposal
and
because
we've
been
unable
to
really
completely
convince
each
other
about
which
way
to
implement
this.
I
suggest
that
we
compromise
on
the
average
of
everything
at
the
list
change.
E
So
in
my
mind,
this
is
to
a
not
introduced
any
new
mechanisms
to
use
the
current
mechanism
established
mechanisms,
whether
we
like
them
or
not,
namely
constraints.
I
think
this
is
also
good
for
web
developers
because
they
often
file
bugs
on
chrome,
saying:
hey,
this
constraint
doesn't
work,
so
obviously
they
already
expect
this
to
work.
So
that's
number
one
and
number
two
is:
I
think
that
we
should
keep
this
a
hint
and
stay
as
open-ended
about
how
this
hint
is
to
be
interpreted
by
the
user
agent.
I
Hint
is
fine,
I
think
that
there's
a
compromise
where
we
could
say
that
underneath
it's
a
constraint,
we
are
reusing
the
constraint
mechanism,
but
we
change
a
bit
of
a
web
ideal
that
is
exposed
so
that
we
remove
all
the
craft
that
constraints
bring
in
terms
of
web
idea,
and
that
would
allow,
for
instance,
whenever
you
are
using
exact.
I
E
So
you
mean,
do
you
mean
reject
if
it's
exact,
but
I
forgot
the
word
but
preferred.
I
E
I
D
I
would
strongly
strongly
propose
not
making
this
that
part
of
this
proposal,
but
that's
that's
because
I
hate
irregularities
and
to
we
have
yaniwa,
of
course,.
G
Yes,
I
agree
with
harold
on
this
one.
I
think
exact
is
already
a
tight
error
in
get
display
media
so
get
display.
Media
is
already.
I
think
we
did
a
good
job
of
narrowing
down
the
constraints
mechanism.
There's
no
advanced,
there's,
no
exact,
there's
no
min.
There
is
max
for
some
reasons.
G
So
since
it's
already
operating
on
a
reduced
version
of
constraints,
I
think
the
issues
that
you
mentioned
aren't
that
serious
severe
for
users
and
that
I
I
so
I
agree
with
using
the
display
surface
here,
because
it
already
exists,
it's
already
specified
and
if
implementations
haven't
implemented
it,
if
it's
not
showing
up
and
get
constraints,
it's
because
some
limitations
haven't
implemented
the
way
they
should.
So
I
think
this
is
a
good
use
of
that.
G
G
If
we
add
that
my
concern
is
that
applications
that
use
this
to
to
ask
for
a
monitor,
for
instance,
like
use
cases,
schools
where
teachers
want
to
see
the
entire
screen
of
the
desktop
of
the
user,
that
that
sort
of
that's
not
a
use
case,
I
feel,
is
something
we
should
be
providing
that's
a
level
of
control
that
is
not
conducive
to
the
best
interest
of
the
user,
which
I
in
this
case
will
be
the
student.
G
E
So
if
it
is
something
relatively
non-committed
like
should,
then
maybe
you
can
talk
about
this,
but
I
think
that
this
is
not
actually
helping,
because
user
agents
might
otherwise
so
currently
in
a
relatively
big
implementation,
chrome,
the
default
is
monitor
and
we
would
all
like
to
move
away
from
this,
and
an
off-ramp
would
be
to
allow
this
to
be
maintained,
as
you
know,
as
a
hint
that
slowly
gets
deprecated
if
there
isn't
enough.
Pushback.
E
We've
discussed
this
on
the
issue
and
I
feel
that
also,
if
we
just
keep
it
minimal,
we
don't
need
to
discuss,
monitor.
Firefox
will
not
respect
the
hint
for
monitor
and
hopefully,
chrome
can
one
day
stop
respecting
that
one
too,
and
at
that
point
we
can
either
add
this
language
or
just
keep
ignoring
it,
because
it
is
implicit
from
the
fact
that
it's
a
user
agent
decision,
whether
to
regard
the
hint
or
not,
and
I
think
that
would
be
an
easy
compromise
right.
So.
G
Well,
that's
why
I
propose
should
otherwise
I
would
propose.
Must
yes,
so
I
think
with
the
should
chrome
should
be
able
to
do
the
off-ramp
you
mentioned
right.
I
see
tommy's
on
the
queue.
A
Yeah
again,
this
is
a
hint.
So
what
ela
described
as
you
know,
you
can
do
it
or
you
cannot
do
it
like
firefox,
is
in
a
position
to
take
a
better
approach
by
default.
Personally,
I
feel
that's
good
enough.
I
I
don't
at
all
object
to
having
the
shoot,
but
with
a
hint
hinting
on
the
hint
feels
also
maybe
a
bit
too
much
like.
E
Yes
to
support
what
dom
just
said,
I
think
that
if
we
have
a
shoot
that
is
later
disregarded,
that
is,
that
does
not
serve
anybody
right.
It
just
confuses
web
developers,
so
I
think
that
we
would
be
better
off
without
it.
I
I
don't
I
disagree
there.
A
spec
is
used
to
it
by
implementers.
So
if
you're
a
new
implementer
and
there
will
be
new
implementations,
then
people
will
implement
it
and
will
allow
monitor
by
default
and
then
at
some
point
they
will
figure
out.
Oh
no,
then
the
specs
should
really
have
told
me
that
I
should
have
actually
ignored
it,
and
specs
are
also
good
for
that,
so
they
should
will
actually
solve
that
for
a
new
implementers.
So
chrome
is
an
existing
implementer.
So
this
does
not
really
care
about
the
sheet.
E
Interesting,
I
see
the
submarine
in
this
point.
Yes,.
E
I
think
how
about
the
following
compromise?
What
if
we
have
non-normative
language
explaining
to
implementers
that
a
lot
respecting
a
monitor
hint
comes
at
risk
and
that
it's
better
to
not
do
that.
G
I
E
Okay,
so
am
I
hearing
that
modulus
should
issue
everything
else.
We
agree
on.
E
The
way
I
understand
this,
please
correct
me
if
I'm
wrong.
If
we
take
the
current
text
and
just
add
the
point
of
should,
then
you
would
accept
this.
G
Oh
yes,
so
this
is
actually
not
an
ask
from
the
working
group.
It's
just
a
reminder
that
we
finally
have
a
pr
up
on
get
viewport
media.
So
we
would
like
to
do
a
call
for
adoption.
So
please
have
a
look
at
that
link
at
your
convenience
after
the
meeting.
Just
a
quick
recap:
it's
it
looks
very
much
like
it
display
media
except
it's
called
get
viewport
media,
it
returns.
G
So
there's
a
document
policy,
viewport
capture
required
document
policy,
viewport
capture
for
iframes,
there's
a
user
permission,
viewport
capture
and
a
permissions
policy
for
iframe,
which
is
also
called
viewportcapture,
and
then
this
also
requires
transient
activation
and
that's
the
same.
Privacy
indicator,
requirements
and
constraints,
video
and
audio.
As
get
display
media,
so
we
didn't
really
talk
about
audio,
but
I
I
feel
that
there's
really
no
reason
to
exclude
audio,
so
it
is
in
the
current
document.
So
any
questions
about
that
yeah,
two,
okay,.
I
Looking
at
the
web
ideal
display,
media
stream
constraints
is
probably
not
what
you
want.
You
probably
want
your
own
version
of
viewport
media
and
the
second
thing
is
about
audio
you're
capturing
the
tab.
So
the
question
then,
for
audio
is
whether
audio
will
be
the
system
audio
or
whether
it
will
be
the
current
tab
audio
and
that's
something
to
think
about.
G
Right
the
current
document
actually
says
it
must
not.
It
must
not
capture
system
level,
audio.
E
E
It
it
should
probably
do
even
more.
It
probably
should
not
even
capture
any
other
tab,
so
it's
not
just
system,
but
it
must
be
only
from
the
current
app.
I
think
that's
the
only
sensible
thing
to
do
actually.
G
Yeah,
so
that
is
very
much
the
intent
of
the
document
and
I'm
happy
to
clarify
that,
if
maybe.
G
Just
bear
in
mind
for
the
there's
there's
a
pr.
If
you
look
at
the
document
right
now,
I
think
it
does
say
html
capture,
that's
I
didn't
merge
issue
number
four,
so
hopefully
that'll
be
merged
this
thursday
or
before
before
call
for
adoption.
E
This
is
not
a
question
but
more
of
a
public
statement.
I
would
like
to
say
that
I
think
this
is
modulo.
The
specifics
right.
The
general
intent
of
this
is
awesome,
I'm
very
excited
by
this
work
and
I
hope
to
see
this
both
finalized
as
well
as
implemented.
E
I
just
want
to
raise
the
concern
that
we
don't
know
whether
it's
actually
going
to
be
adopted
by
web
developers,
so
I
think
that
we
should
stay
away
from
avoiding
other
progress
based
on
the
fact
that
this
is
upcoming,
and
I
would
also
like
us
to
keep
in
mind
that
if
adoption
does
not
happen,
we
might
need
to
relax
a
couple
of
the
requirements
which
might
be
difficult,
but
we
might
need
to
do
so
or
deprecate.
The
whole
thing.
I
So
related
to
that,
do
we
have
like?
Did
we
reach
out
to
some
web
developers
about
get
report
media
and,
in
particular,
the
constraints
related
to
cross-original
isolation
and
whether
that
would
not
be
that
would
not
slow
down
the
adoption
of
this
api.
E
G
Well,
I
agree
that
I
think
we're
taking
the
long
view
here,
so
I
I
think
we're
the
working
group
has
been
very
open
to
making
a
lot
of
changes
to
get
this
my
media
lately.
G
Sorry
to
your
comment,
you
went
about
this
display
media
constraints.
I
think
that's
mostly
editorial,
but
yes
right
now,
it
display
media
constraints
fit
the
bill,
but
we
can,
of
course,
if
there's
changes
needed
for
that.
We
can
duplicate
that
if.
G
B
All
right,
okay,
we're
a
little
bit
ahead
of
time,
but
we're
going
to
hand
the
floor
over.
I
guess
is
it
you
ritchie
or
yeah,
okay,
right.
J
Okay,
so
yeah
hi,
all
we
spoke
about
some
of
the
webrtc
and
the
features
at
tpac,
2021
and
again
briefly
in
the
november
interim
meeting.
So
now
that
the
initial
peers
have
been
out
for
a
while,
hopefully
this
audience
had
the
time
to
go
through
the
details.
So
we
have
a
lot
of
to
discuss.
So,
let's
deep
dive
into
the
face
detection
next
slide,
please
yeah!
J
Well,
here's
a
snapshot
of
the
ideal
for
the
face
detection
proposal.
I
have
linked
to
the
full
proposal
in
the
notes
section.
Actually,
if
you
click
there,
we
want
a
way
to
do
face:
detection
natively
on
browsers.
Instead
of
you
know,
cloud-based
solutions
and
the
ml
framework
way
to
unlock
specific,
like
client
capabilities.
Last
time
we
demoed
face
detection
on
chrome
os,
and
there
were
few
remarks
so
number
one
bernard
her
harold
uen
wanted
face
detection
to
be
anchored
to
video
frame
defined
in
web
codex
instead
of
the
media
stream
track.
J
I
think
the
pr
now
reflects
that
number
two,
the
boundary
box
issue,
I
think
harold
asked-
was
to
make
the
api
a
bit
more
general
and
forward-looking
something
for
the
future
as
well.
Not
getting
fixated
only
on
the
present,
specifically,
they
ask,
was
to
return
something
like
a
mask.
Instead
of
a
rectangle,
we
tried
to
reason
a
bit
and
even
though
right
now
there
is
no
platform
api
supporting
a
mask.
J
We
try
to
accommodate
this
request
by
returning
a
contour.
The
number
of
points
describing
the
contour
can
be
user
defined
like
in
that
phase,
detection,
num,
contour
points,
settings
and
implementation
wise
right
now.
Obviously,
we
default
to
a
four
point.
Rectangle
I
mean
discussions
with
camera.
Driver
teams
across
orgs
actually
have
revealed
that
underlying
face
detection
algorithms
do
detect
those
points,
but
the
main
pain
point
has
been
standardizing.
The
number
of
counts
of
this
point,
so
that's
why
they
are
not
yet
putting
up
a
standardized
platform
api.
J
The
next
task
was
something
very
similar
like
since
a
few
frameworks
do
return
face
mesh
like
the
tensorflow
returns,
a
468
landmark
face
mesh.
Can
we
have
that
also,
so
we
have
reworked
it.
The
face
detection
mode
to
non-present,
contour
and
mesh
again
mesh
is
not
possible
right
away
on
the
platform,
but
I
think
like
for
the
sake
of
extensibility.
J
Number
four
was
face
expressions.
It
did
not
get
much
support
from
the
audience
last
time.
So
I
guess
it
was
on
everybody's
mind,
but
tim
had
voiced
that
face
expressions
was
more
subjective
and
there's
a
concern
about
the
expression
detection
going
wrong.
We
have
removed
them
from
the
pr
totally,
maybe
maybe
in
future
we
could
add
blink
and
smite
only
the
other,
the
entire
list
I
kept.
Maybe
it's
not
yet
ready
for
prime
time
but
as
of
now
in
the
pr
I
have
removed
the
entire
expressions
we
can
add
later.
J
If
things
improve,
number
five
ask
was
to
make
sure
face:
detection
works
with
the
transform
stream.
So
we
put
up
a
example.
I
think
it's
in
the
next
slide.
If
we
could
go
to
the
next
slide,
yeah
we
put
up
an
example
mostly
to
show
how
face
depiction
and
transform
team
and
workers
can
work
together
and
yeah.
And
lastly,
I
think
the
last
point
which
I
remember
is,
I
think
I
said
I
will
bring
up
some
pnp
numbers
to
validate
some.
You
know
claims.
J
Unfortunately,
I
could
not
get
the
permission
for
such
official
numbers
on
a
public
forum,
but
let's
say
unofficially,
we
can
say
that
around
it's
half
of
the
time
like
a
half
of
power,
is
needed
for
phase
detection
or
we
have
done
only
on
chrome
os
system
so
compared
it
versus
media
pipe.
J
Obviously,
I
before
chrome
up
streaming
we'll
be
sharing
the
numbers
proper,
proper
numbers,
so
I
mean
from
we
just
quickly
hacked
and
got
the
power
consumption
as
say
two
wattage
for
this
total
system
power,
while
doing
the
phase
detection
our
poc,
and
when
we
did
the
same
resolution
on
media
pipe,
it
was
around
3.25
to
3.5
voltage
yeah,
but
you
will
bring
better
like
official
numbers
soon.
I
Yeah,
I
have
some
some
questions
in
general.
I
think
that
it's
good
to
expose
it
in
video
frame.
It
would
also
be
good
to
expose
this
kind
of
metadata
to
request
video
frame
callback
metadata
as
well,
so
that
you
can
do
that
with
canvas
as
well
as
if
you
have
a
stream
of
video
frames,
so
that
that's
something
that
I
would
hope
is
not
controversial,
and
we
could
do
that.
I
So
I
would
think
that
all
these
constraints
would
not
allow
exactly
actually-
and
my
general
question
would
be
so
that
there
seemed
to
be
like
several
switches
and
a
possibility
to
provide
parameters
to
the
camera
so
that
they
can
tune
their
algorithms
and
so
on,
and
I
was
wondering
whether
in
general
camera
I
was
wondering
why
we
need
like
several
switches
and
not
just
have
one
single
switch
in
the
case
where
the
camera
is
either
exposing
data
or
not
exposing
data.
I
But
it's
not
like
trying
to
disable
an
algorithm,
re-enable,
an
algorithm
and
so
on.
So,
are
you
expecting,
based
on
these
constraints,
that
some
algorithms
will
kick
in
or
will
be
disabled
and
so
on
so
and
why
there
should
be
multiple
constraints,
or
would
it
be
fine
to
just
have
a
single
switch
telling
hey?
I
want
to
learn
about
face
detection,
so
please
provide
me
whatever
you're
grabbing
your
camera
sensor,.
J
I
It's
more
a
general
question
of
why
there
are
multiple
switches
and
whether
a
single
switch
would
not
be
simpler
and
good
enough
in
general.
And
then,
if
we
have
just
one
single
switch,
then
it's
up
to
the
web
application
to
read
what
is
available
and
do
what
it
can
with
the
provided
data
by
the
sensor.
J
Okay,
I
mean
let's
say
well,
for
example,
the
contour
points
right
now,
there's
no
not
such
support,
so
it
looks
a
bit
superfluous,
maybe
right
now,
but
just
for
like
to
accommodate
the
ask
about
future
and
all
those
things
we
added
that
way.
So.
J
I
I
mean
I
I
could
see.
I
don't
really
understand
why
face
dictation,
name
counterpoints
with
with
b4,
for
instance,
it
seems
like
your
webpage
and
you're
trying
to
provide
some
parameters
to
the
camera
sensor,
but
my
understanding
was
mostly
like
the
camera
sensor
is
anyway
doing
something.
It's
providing.
J
So
can
you
can
we
go
a
bit
previously?
Please.
J
Yeah,
what
I
mean
is
johan,
is
there
like?
Do
you
think
this
idea
looks
fine
to
you
or
you
think
this
is
too
much
information
or
a
light
too
too
much
superfluous
data.
I
J
Right
so
I
can
say
that,
right
now
the
contour
mesh
is
not
at
all
exposed,
but
we
added
these
kept
these
things
for
extensibility,
so
but
say,
landmarks
and
id
and
probability.
These
three
are
things
the
driver
uses
ids
specifically
for
face.
Tracking
probability
is
like
what
is
the
probability
that
it's
a
human
face
and
landmarks
is
like
left
like
nose
eyes,
so
these
three
are
totally
present
there,
the
contour
and
mesh,
not
yet
there,
but
they
are
working
on
it.
J
Let's
say
in
future,
so
we
wanted
to
keep
it
there,
but
because
there
was
ask
about
future
extensibility,
that's
why
we
kept
it.
There.
I
I
think,
but
so
maybe
that
that
means
that
enums
are
fine,
but
maybe
we
could
like
and
dictionaries
are
fine
as
well,
because
enums
and
dictionaries
are
extensible,
but
maybe
we
can
reduce
to
what
is
implementable
right
now
and
make
sure
that
what's
coming
comes
next
will
be
a
will
be
like
the
structure
we
are
using
will
be
extensible
enough
for
those
cases.
J
Yeah,
that
was
the
point
with
the
contour,
because
implementation
wise,
we
would,
if.
J
Number
of
contour
points
equal
to
four.
It
means
a
rectangle
which
is
implementable
right
now,
the
the
platform
api,
the
the
people
in
the
camera
drivers
in
different
organizations,
so
they
are
unable
to
standardize
the
number
of
points.
As
of
now,
that's
why
they
cannot
expose
that
as
a
platform
api,
because
everybody
has
to
let's
say
microsoft.
Google,
michael
everybody
has
to
agree
on
okay
to
have
a
good
algorithm
everything
we.
We
would
standardize
on
16
points,
but
that
agreement
is
not
yet
there.
J
So
so
that's
why
there
is
no
platform
api
for
the
contour
points,
but
hopefully
in
future,
it's
going
to
work
that
way,
but
you
were
actually
right
in
the
thinking
that
this,
the
camera
is
anyways
using
the
phase
detection
to
improve
its
three
algorithms
and
we
are
taking
those
things
from
the
stream.
So
that
way,
you
are
right.
B
So
the
reason
I
understand
this
is
the
way
it
is
is
because
what
you're
trying
to
do
through
the
supported
constraints
is
essentially
provide
and
capabilities
is,
is
to
provide
basic
parameters
for
the
algorithm
which
you
set
in
the
driver
and
then
essentially,
you
now
have
video
frame
dot
detected
faces
because
it's
it's
already
been
done
by
the
drivers
you
specified
is.
That
is
that
right.
J
J
B
Done
this
way,
as
opposed
to
having
like
a
promise
based
method
that
would
give
you
you
know
to
which
you
would
provide
the
parameters
is
essentially
it's.
It's
kind
of
it's
dependent
on
the
driver.
So
if
your
camera
driver
doesn't
support
this
you're
not
going
to
have
the
you're
not
going
to
have
to
basically.
J
No
yeah
the
promised
thing.
What,
if
we
do
through
the
promise
thing,
then
we
will
have
to
call
something
called
detect
phase
and
then
the
let's
say,
implementation
wise.
It
won't
be
great
right
because
it's
it's
going
to
call
again
what
the
driver
has
already
done.
B
J
J
Like
if
you
want
more
detail
on
that,
I
can
tell
at
least
for
the
windows
part.
What
happens
is
if
you
do
through
the
promised
way.
You'll
use
something
like
face
analysis,
whereas
if
you
use
this
way,
you'll
use
something
called
mf
like
micro,
mdft
stuff,
and
there
will
be
no
like.
J
If
I,
if
I
do
through
phase
analysis,
there
will
be
duplicate
cost
for
compute.
It
cannot
take
advantage
of
implementation.
Details
like
using
lower
resolution
for
video
handling,
sub
orientation
of
subjects
and
using
camera
roi
all
these
kind
of
things.
So
that's
why
this
way.
G
All
right,
I
think,
I'm
next
on
thecube,
so
I
I
think
you
answered
some
of
this
from
bernard's
question
and
that
this
is
sounds
like
it's
a
camera
api.
Not
I
mean
it's
on
media
stream
track
on
video
frame,
but
it
sounds
like
this
would
only
be
available
if
the
source
is
a
camera.
Is
that
right.
G
To
make
it
so
that's
what
I
thought
so
my
concern
there
is
that
there's
also
there's
another
effort
in
the
wicg,
which
is
the
accelerated,
shape,
detection
and
images
api.
So
I'm
wondering
how
this
yes,
my
concern
is
that
if,
if
that
effort
is
also
ongoing,
it
would.
J
J
You
can
call
it
multiple
times
inside
transform,
but
I
it's
not
an
efficient
way
to
do
this.
There
is
no
face
tracking
available,
which
means
I
mean
face
tracking
is,
is
a
very
handy
way
to
detect
across
frames
and
make
those
efficient.
J
J
So
I
don't
know
I
can
ask
riley
and
miguel
if
they're
planning
to
continue
it
because
intel
was
our
team
initially
put
up
the
chrome
patches
for
windows,
and
I
think
we
realized
that
this
is
a
much
better
like
in
terms
of
implementation.
J
This
is
a
better
way
which
will
give
at
least
the
perf
results,
much
better
so
and
shape
detections
phase
detector
is
not
yet
shaped
because
it
works
only
on
windows
as
of
now
and
now
and
android,
maybe
not
on
chrome,
os
so
and
the
work
is
stopped
right,
intel
actually
implemented.
The
windows
part.
B
J
D
D
D
Available
in
the
api,
I'm
still
not
happy
with
the
design
that
seems
to
be
totally
focused
on
on
making
this
an
access
api
for
for
hardware
or
drivers
based
resources.
Instead
of
making
this
this
representation
api
that
allows
us
to
say
in
this
video
frame,
you
have
the
following:
information
objects.
D
It's
kind
of
getting
a
little
bit
of
that
flavor,
but
with
the
the
all
the
dictionaries
and
so
on,
are
still
very
much
about.
We
have
to
configure
the
camera
driver
to
do
this
work
for
us
when
it
would
be
perfectly
acceptable
to
do
it
in
many
cases
to
do
it
ourselves
right.
So
I
was
a
bit
surprised
about
about
the
number
that
you
only
gained
the
factor
of
less
than
50
over
media
over
media
pipe.
D
I'd
like
to
see
this
being
proposed
as
a
proposal,
a
proposal
not
just
as
a
set
of
api
patches
that
is
having
the
explainer
having
the
use
cases
having
the
the
apis
having
examples
all
that
stuff
that
we
want
to
put
together
before
we
right
before.
We
say
that
this
is
yes.
This
is
the
right
api
to
change
to
have
so
I'm
ready.
J
Okay,
so,
firstly,
the
we
don't
need
to
configure
driver
anything.
We
put
this
thing's
contour
mesh.
Just
to
you
know,
for
extensibility,
there's
no
driver
configuration
needed
if
the.
If
the
camera
is
in
auto
mode,
it
automatically
happens.
So
this
is
mainly
for
like
getting
the
information
and
that
kind
of
thing
regarding
examples.
I
guess
you
have
had
a
look
at
the
pr.
Do
you
think
those
examples
were
okay
or
you
like?
D
A
Just
to
clarify,
I
guess
nobody
is
asking
you
to
improvise
an
answer
to
this.
The
the
typical
approach
that
has
been
taken
in
this
group
and
others
is
to
package
answers
to
this
kind
of
question
in
an
explainer
which
would
also
be
needed
for
a
tag
review
down
the
line,
and
I
guess
that's
I
mean
what
I'm
sensing
here
is
before
we
can
say
whether
this
is
striking
the
right
set
of
traders.
A
We
would
need
to
better
understand
which
use
cases
this
fulfills,
which
developers
are
asking
for
it
and
which,
which
of
the
many
options
that
you're
pushing
forward,
are
optional
or
necessary,
and
so
on.
So
I
guess
I'm
hearing
the
need
to
level
up
some
of
the
conversations
rather
than
necessarily
diving
in
into
the
detail
api
surface.
A
J
A
Way
of
bringing
this
proposal
for
review
by
the
group,
one
of
the
things
that
I'm
also
hearing
is
that
this
is
big
enough,
a
proposal
that
it
should
like
probably
be
a
spec
on
its
own
other
than
a
patch
on
a
patch,
which
is
what
media
capture
extensions
is
today
for
for
better
hours,
and
so
again
we
probably
need
to
figure
the
logistics
for
how
to
make
that
happen
as
smoothly
as
possible.
J
Okay,
but
specifically
from
like
from
this
group,
so
you
think
nobody
is
going
to
use
this
like.
Is
there
a
question
about
whether
it's
useful
or
not.
I
G
Yes,
sorry
to
skip
the
cube,
but
maybe
I
can
clarify
I.
I
worry
that
the
other
specification
was
dealing
with.
You
mentioned
some
benefits.
This
is
tied
to
video
frame,
which
I
agree
has
benefits
over
photo
and
images,
but
it
still
feels
like
tying
this
to
the
camera.
I
mean
the
camera
itself
provides
an
a
necessary
function.
I
mean
there's
no
way
to
simulate
the
camera
and
stop
well.
G
You
need
to
record
my
face,
but
as
far
as
this
feature,
it
sounds
like
it's
going
to
be
obsolete
in
five
or
ten
years,
because
machines
become
faster
and
then
a
lot
of
use.
Cases
I
think,
would
be
to
to
do
this
kind
of
processing
with
hardware
on
sources
other
than
the
camera,
and
that's
what
I'm
missing
here
I
feel
like
it
might
be
better
to
have
a.
G
I
think,
a
long
longer
view
api
would
not
focus
so
much
on
the
hardware
acceleration
in
the
being
in
the
driver,
but
that
this
would
be
the
way
to
express
what
javascript
wants
on
the
video
frame
that
could
be
done
by
the
user
agent.
For
example.
Right
I
mean.
J
Without
going
to
the
implementation
detail
about
exactly
if
it's
doing
on
camera
within
the
isp
in
the
cpu
or
vpu,
those
things
actually
are
abstracted
there,
so,
but
well,
okay,
so
so
hard
harold.
If
I
am
understanding
correctly,
you
want
basically
proof
that
background
blur
face
detection.
All
these
things
are
going
to
be
used
at
all
usable
or
I'd
like
to.
D
C
I
In
that
case
is
only
coming
from
the
camera,
so
it
all
makes
sense.
I
really
think
that
it's
a
good
point
that
you
define
an
api
that
is
useful
for
driver
generated
as
well
as
algorithm,
like
transform
stream,
like
you,
have
a
transform
stream
that
is
doing
phase
detection
and
you
get
some
metadata,
and
hopefully
we
should
have
the
same
metadata
coming
from
the
transform
stream
or
to
the
driver.
That's
that's
the
point
of
this
vc
fort.
I
If
we
are
not
able
to
get
that
or
we're
not
comfortable
having
that,
then
it
becomes
much
harder
to
sell
the
the
proposal.
J
Okay,
so
and
explainer
right,
okay,
yeah,
explainer
and
all
those
things
I
have
sort
of
ready
in
my
google
docs.
So
I
just
know
need
to
know
the
exact
location
where
to
put
them
I
can
off.
I
can
coordinate
with
dom
to
find
a
home
for
that.
I
have
the
data
sort
of
ready,
okay,
so
oh
coming.
G
If
I,
but
also
my
concerns,
were
about
the
api
itself
and
whether
this
is
the
right
shape
for
okay
and
whether
you
intend
to
solve
sources
other
than
camera,.
F
A
That
was
mentioned
in
december
that
I'm
hearing
is
still
a
concern
with
the
current
approach.
G
Oh,
I
mean,
even
if
even
if
browsers
today
won't
be
able
to
emulate
this
as
software
they
might
be
able
to
in
the
future.
So
at
least
if
the
api
is
in
the
right
shape,
we
don't
have
to
redefine
it.
J
Yeah,
I
can
okay,
I
I'll
add
some
data
to
that
right
in
the
explainer.
I
I'll
just
move,
because
I
see
there's
not
much
time
left
I'll.
J
J
As
of
now,
we
are
going
to
propose
just
the
background
blur,
because
the
replacement
again
not
many
platform
apis
are
there.
So,
of
course,
there
is
also
no
platform
api
right
now
to
control
the
blur
level,
but
obviously
there
are
many
frameworks
where
you
can
control,
so
I
kept
it
there,
but
it
it
will
hopefully
work
in
future,
but
yeah.
J
This
was
a
sort
of
simple
api,
like
so
platforms,
do
some
sort
of
in
in-stream
correction
and
the
example
we
present
later
on
you'll
see
that
the
users
can
either
opt
for
this
one,
this
api
or,
if
the,
if
their
platform,
supports
or
use
the
framework
if
they
want
to
differentiate
what
a
native
look
and
feel
like
like,
in
short,
suppose
you're
on
windows
the
way-
and
there
is
a
windows
app.
J
I
don't
know
if
you
are
using
teams
and
you
have
the
teams
chrome,
running
on
chrome
or
edge
on
on
windows
spectrum.
It
will
look
the
same
the
because
the
underlying
native
call
is
same.
Obviously,.
I
You
just
to
mention
that
on
some
platforms,
like
ios
macros,
there's
the
ability
for
the
user
to
switch
on
and
off
background
blur,
for
instance,
and
it's
outside
it's
fully
outside
of
the
control
of
the
web
application
and
it's
fully
dynamic.
I
It
does
not
so
basically,
web
application
could
not
unblur
right
if
the
user
decides
to
blur.
But
yes,
but
it.
C
I
Blur
the
web
application
could
blur
if
the
user
is
not
wanting
to
blur
through
the
os
settings.
That's
not
something
that
is
well
supported,
currently
by
constraints
because
constraints,
usually
you
have
a
camera
and
the
resolution.
The
highest
resolution
will
not
change
in
real
time.
So
maybe
we
need
some
api.
If
we
want
to
use
constraints
there,
we
might
need
some
apis
so
that,
for
instance,
a
web
application
knows
that
at
some
point
background
blur
was
changed.
I
The
user
decided
to
to
use
the
system
background
blur
and
the
web
equation
will
not
be
able
to
to
set
it
back
to
false,
for
instance,
and
some
api
is
missing,
there
either
event
based
or
like.
Maybe
just
an
even
saying,
hey
constraints
have
changed,
or
I
don't
know,
but
we
probably
need
to
to
find
something.
If
we
want
to
support
ios
platforms,
for
instance,.
G
Yeah,
it's
me
again
well,
I
think
this
is
actually
a
case
where
constraints
work
really.
Well,
I
mean
you
because
it
does
allow
the
application
to
state
it's
ideal
and
then
the
user
agent
can
still
override
it
and
that's
for
whether
an
application
should
be
allowed
to
specify
exactly
a
background
blur
or
not
that
we
can
always
discuss.
I
think,
but
I
generally
support
this
idea,
that
I
think
that.
I
I
don't
think
that
ideal
constraints
fully
support
everything
there,
because
you
you
can
it
can
be
set
after
the
track
is
created
or
after
apply
constraints
or
whatever.
So
we
we
don't
have
good
support
there
to
express
all
these
things
and
the
web
application
might
want
to
be
notified
that
the
change
of
background
blur
from
force
to
even
though
the
web
application
is
set
back,
one
block
to
force,
for
instance.
So
there
are
cases
that
we
we
need
some
api
there
to
to
expose
those
that's
performance.
J
I'm
I'm
just
thinking
whether
people
feel
it's.
I
mean
everybody
is
using
this
right.
So
if
we
can
provide
a
platforms
api,
I
guess
is
I
mean
I'm
trying
to
find.
Is
there
any
blockers,
apart
from
what
u.n
suggested.
B
J
J
But
there's
no,
it
yeah,
but
I
think
that
platform
teams
are
working
on
going
making
it
exposed
as.
I
Settable,
it's
it's
something
that
is
difficult
for
web
applications
to
to
set
because
they
might
not
know
the
actual
algorithm.
So
maybe
they
will
do
something
and
say:
oh
and
change
it
a
little
bit
or
provide
a
user
selection
and
so
on.
But
it
seems
like
advanced
and
advanced
case.
So
maybe
there's
a
two-step
proposal.
There.
First,
the
simple
thing
and
then
the
double
double
suitable
property.
G
Okay,
yeah,
I
understand
that
as
well.
I
mean,
I
think,
the
way
to
view
this
is
that
user
agents
may
provide
blurring,
allow
the
user
to
blur
their
camera
independent
of
this
api.
The
question
is
whether
user
agents
feel
any
value
and
letting
an
application
have
a
button
to
turn
it
on
and
off.
I
think,
and
I
could
see
an
implementation
go
either
way
and
say:
no.
G
This
is
the
user
blurred
this
it's
not
changeable
by
the
actual
application,
we're
not
going
to
expose
that
it
happened
at
all
or
we're
going
to
expose
it
and
not
be
able
to
change
it
or-
and
I
think
applications
should
be
able
to
read
this
and
and
if
see,
if
the
property
is
there,
they
could
experiment
with
turning
on
and
off
and
maybe
expose
a
button.
If
that
was
useful
but
yeah,
it
does
beg
the
question
of
how
much
blur.
D
Cases
where
it
was
very,
very
valuable
to
tell
that
the
user
had
been
been
manipulating
settings
that
were
supposed
to
be
helpful
in
drivers,
but
really
messed
the
things
up,
such
as
causing
double
double
echo,
detection
and
and
the
aec
controls
that
we're
fighting
each
other.
D
So
that
was
so
in.
In
that
case,
the
most
important
property
we
had
well.
The
most
important
control
we
had
was
the
ability
to
to
turn
platform
effects
off
and
the
most
and
the
the
the
second
most
important
feature
was
to
be
able
to
detect
that
platform
and
ex
effects
have
been
turned
on,
so
that
we
cannot
use
it
to
turn
them
off.
B
Well,
I
think,
actually,
rather
than
trying
to
dump
everything
into
the
last
minute,
I
think
we
maybe
should
talk
about
how
to
move
forward
in
general
because
there's
you
know
a
whole
bunch
of
stuff
here
we
didn't
get
to
so.
J
C
J
Just
bull,
I
just
want
to
get
a
feeling,
because
I
did
not
get
any
comments
from
on
the
pr
just
wanted
to
know
that
I
mean
these
are
all
these
are
the
features
you
will
see
working
on
facetime
every
everything
is
working
there
and
there
are
obviously
platform
it
it.
J
It
works
on
lighting
correction
works
on
meat
also,
these
days
right
so
yeah
and
what
we
are
trying
to
do
is
giving
give
meat
and
teams
and
and
anybody
the
web
apps
a
way
to
use
the
platform
features
directly
instead
of
running
the
framework
and
obviously
they
are
free
to
choose
both
it's
just
giving
another
option
and
suppose
meet
teams
fails.
The
blur
is
working
fine
on
windows,
they
can
choose
the
this
one
and
if
they
think
no,
they
can
keep
on.
J
B
Just
want
to
also
figure
out,
I
guess,
from
the
chair's
input
on
how
what
are
the
next
steps
here
for
the
rest
of
this?
Do
we
re
provide
time
in
april?
What
do
you
think
janibar
and
harold.
G
Oh
yeah,
I
forgot
why
I
was
on
the
queue,
but
I
don't
think
we
have
a
strong
interest
to
implement
at
the
moment.
So
I
know
mozilla's
position
on
the
similarly
named
shape.
Detection
api
is
that
we're
worried
about
complexity,
variations
in
support
within
operating
systems.
G
So
we
have
it
as
a
defer
at
the
moment
so,
but
I
I
think
the
advice
I
was
given
so
far
was
good.
Tomorrow.
B
B
D
For
for
the
face
detection,
we
have
a
pretty
solid.
D
J
You're
saying,
okay:
is
there
a
location
you
want
location
of
the
document
or
just
a
github?
My
personal
data
pick
a
github
and
I
get
that.
D
I
think
I
think
it
is
better
to
have
them
together.
I
mean
these
are,
if,
if
we
accept
the
idea
of
of
con
of
using
the
constraint
api
or
any
other
api
on
media
capture
to
control
camera
driver
settings,
then
having
one
one
big
one
document
with
all
the
camera
drivers,
I
think,
stinks
in
it.
H
G
Not
opposed
to
that,
but
I
could
also
see
that
media
capture
extensions
has
been
used
for
for
individual
constraints
in
the
past,
like
for
focal
length
and
that
kind
of
stuff.
So
there's
some.
I
To
me,
it
seems
really
different,
like
one
one
is
in
terms
of
complexity.
We
are
talking
about
a
constraint
which
is
boolean
or
not
compared
to
a
constraint,
plus
exposing
complex
data
with
potential
interaction
with
other
apis.
So
it's
not
all
the
same
thing
so,
and
I
think
that
one
could
progress
much
faster
than
the
others
as
well.
In
terms
of
implementations
for
some
platforms.
A
Right
so
we
drew,
I
guess
I
will
need
to
get
a
clear
input
from
the
chair,
so
the
ball
is
in
my
court.
I
I
do
think
there
is
clarity
that
we
want
an
explainer
for
face
detection.
Yes,
whether
the
three
or
four
additional
camera
driver
settings
sure
this
can
just.
C
A
Small
additions
to
media
capture,
extensions
or
something
else
together
that
I
will
have
to
come
back
to
you
sure
I
have.
I
have
all.