►
From YouTube: Multiple Fields in NuPIC
Description
with Dr. Subutai Ahmad
B
A
I'll
just
go
through
that
because
I
know
that's.
There
are
a
lot
of
questions
about
that
on
the
mailing
list
and
and
I
think
thinking
about
multiple
fields
in
no
big.
It's
it's
very
different
from
how
you
think
about
it.
Normally
in
machine
learning
so-
and
it's
often
important
in
practice
to
understand
that
so
Matt
asked
me
to
do
that.
It's
really
informal
and
maybe
not
too
long
and
it's
just
threw
together
some
notes.
So
my
agenda
is.
D
A
C
A
A
D
A
A
C
C
A
It's
kind
of
tough
to
do
too
much
with
2048
columns
and
in
general,
if
you
think
about
finding
the
common
spatial
patterns,
the
space
of
patterns
explodes
exponentially,
as
you
add,
more
and
more
fields,
and
so
the
in
general,
the
space
that
common
spatial
patterns
are
going
to
explore,
explode
exponentially
as
well.
So
it's
kind
of
hard
to
do
this
envision.
A
Of
columns,
only
looking
at
a
small
set
of
STRs
coming
in
another
nearby
set
of
columns
looks
at
another
small
set
of
STRs.
In
the
case
of
vision,
you
can
do
that
and
the
topology
really
restricts
the
space
that
a
set
of
columns
is
looking,
so
that
really
helps
in
the
problem
in
new
paint
that
we
use
it.
Today
we
have
2048
columns
and
typically
there's
no
topology,
so
the
more
fields
you
had
every
column
has
to
participate
in
understanding
the
full
richness
in
this
space,
and
it's
just
it.
A
E
A
Or
something
to
reduce
the
space,
but
we
find
this
is
kind
of
the
range
where
it
works:
okay
and
then,
if
you
as
you
add
more
fields,
so
even
if
you're
getting
two,
you
know
five
fields.
If
you
really
want
the
spatial
Pooler
to
learn
common
spatial
patterns
on
the
temple
memory,
to
learn,
common
sequences,
you're
going
to
also
need
more
data,
so
you
know
one
field,
you
need
less
data
and
then
you
know
five
fields.
You're
gonna
need
a
lot
more
data.
So
that's
another
confounding
factor,
often
in
practice.
A
You
don't
have
enough
data
to
really
handle
five
fields
properly,
even
though,
technically
that
you
could
here
so
that's
another
consideration.
So,
typically
in
the
the
way
we
have
the
typical
parameters
we
use
one
to
five
fields,
it's
kind
of
the
range
that
you
should
be
taking.
You
know
and
then
you've,
if
you
add
more
fields,
who
really
need
a
lot
more
data
for
it
to
learn
the
common
patterns
and
stuff
like
that.
A
F
A
F
A
This
kind
of
an
IT
and
in
our
anomaly
detection
we
typically
have
a
time
of
day,
as
well
as
the
scalar
metric
and
dad,
for
example.
So
this
range
is
pretty
okay,
one
and
two
I.
Don't
really
have
a
good
feel
from
five.
We
very
rarely
really
gone
to
that,
but
I'm
pretty
sure
you'll
need
a
lot
more
than
this.
A
A
D
A
A
A
All
of
the
fields,
but
from
the
classifier
standpoint
from
an
external
site,
one
is
predicting
this
one
field.
This
formulation
is
very
different
from
the
typical
machine
learning
formulation.
Typically,
the
machine
learning.
What
you
would
do
is
you
don't
have
a
concept
of
a
stream
of
data
you
just
have
you
know
you
have
your
prediction,
which
is
the
output
up
your
system
and
any
of
the
inputs
which
are
these
fields.
So
what
you're
trying
to
do
is
build
a
function
where
you
predict
your
predicted.
A
G
A
Time
here,
so
you
don't
really
know
this
stuff
until
it.
Actually
that
happens
at
the
same
time,
missus
so
you're,
making
predictions
of
everything
and
checking
against
those
predictions
and
so
on,
but
usually
in
machine
learning.
You
formulate
the
problems,
so
you
can
use
for
it's.
Let's
say
using
energy.
C
A
A
E
A
D
A
A
F
1
is
very
correlated
with
the
predicted
field.
You
already
have
all
the
information
in
the
history
to
make
this
prediction.
You
don't
need
that
other
field.
In
fact
it's
making
it
worse,
because
if
you
add
another
field,
you
get
back
to
these
issues
of
needing
more
training
data
to
really
understand
the
common
spatial
patterns
and
all
that,
because.
A
C
B
A
A
Always
bring
up
the
fact
that
things
like
temperature
are
very
necessary
and
machine
learning
to
create
a
good
predictive
model
of
energy
usage.
Whenever
we
run
the
HTM,
we've
got
good
or
better
results,
as
both
are
better
results.
We
never
this
one
never
ended
up
picking
up
temperature
and
they
were
always
puzzled
because
in
in
their.
G
A
Without
temperature,
because
of
this
reason,
if
you
think
about
it,
energy
energy
usage
is
correlated
with
temperature,
that's
it's
getting
warmer
and
your
energy
usage
is
going
higher
because
your
air
conditioner
is
being
used
more
you'll,
already
see
an
increase
in
the
energy
usage
here.
You'll
see
the
up
tectum
that
and.
A
A
C
A
A
A
F
A
F
A
A
Yeah
you
could
try
some
like
mutual
information.
Maybe
that
would
work
better
I
don't
know.
Maybe
if
there
is
oh,
but
you
have
to
always
include
the
predicted
field.
So
if
using
this,
if
you
look
at
the
mutual
information
with
the
next
point
in
time
versus
including
both
of
these,
if
you're
better
with
both,
maybe
you
haven't
really
explored
that,
like.
A
A
A
D
A
A
F
F
F
A
A
You're
doing
hourly
prediction
or
every
15
minutes,
that's
a
really
higher
order,
sequence,
and
so,
in
order
for
the
to
learn
that
you
need
tons
and
tons
of
data
so
adding
day
of
the
week,
can't
help.
But
that's
why
I
put
this
not
easily,
though
so
there
the
week
makes
it
much
easier
now
that
HTML
only
has
to
learn
much
shorter
sequences.
I
have.
B
C
B
A
A
A
E
A
A
C
A
A
A
A
This
is
a
really
little
picture
factor,
but
if
you're
doing
a
swarm
make
sure
you
do
a
large
swarm
instead
of
a
medium
swarm,
because
a
medium
swarm
only
looks
at
pairs
of
field
combinations
and
stops
there,
whereas
large
swarm
will
look
at
will
go
beyond
pairs,
and
so
very
often
we
see
that
times
then
and
the
value
will
will
help
a
lot.
And
so,
if
one
really
try
to
add
another
field
in
a
medium
swarm,
only
a
large
swarm
will
do
that.
Now,
large
one
will
take
a
lot
more
time.
A
A
C
B
C
A
D
B
A
A
Fields
at
one
time,
but
all
your
examples
feed
one
input
field
per
model,
wouldn't
it
be
better
to
have
one
model
in
many
fields
that
returns
and
overall
anomaly
score
across
all
the
input.
Well,
actually,
all
our
most
of
our
examples
use
two
fields
which
is
timestamp
n,
scalar
values.
That's
really
yeah,
but.
C
A
B
D
A
B
B
B
C
B
C
B
A
A
F
C
F
F
F
A
F
A
E
A
E
A
A
A
C
A
B
A
So
we
get
likelihood,
which
is
when
the
first
one
the
likelihood
is
doing,
is
looking
at
the
anomaly
scores
and
telling
you
it's
a
probability.
It's
telling
you
what
is
the
likelihood
that
this
anomaly
score
comes
from?
This
is
similar
to
what
we've
seen
before
with
anomaly
score.
So
if
the
metric
is.
A
Anomaly
score
is
always
high,
then
you
need
a
lot
of
very
high
likelihood
high
on
these
scores
in
the
road
to
get
a
high-end.
Omni
light
be
good.
So
that's
what
the
likelihood
is
is
a
probability
now.
The
nominees
are
really
rare.
So
these
probabilities,
the
threshold
we
often
said,
is
like
point:
nine,
nine,
nine,
nine,
nine,
nine
to
five
nines
or
something
like
one
in
a
10,000
chance
or
something
like
that
of
it
of
an
anomaly.
So
it's
really
hard
to
deal
with
those
numbers.
A
D
A
It's
easiest
to
look
at
the
log
likelihood.
You
could
just
say
point:
five
is
your
threshold:
you
can
do
it
if,
otherwise,
if
you
look
at
the
likelihood,
it's
really
hard
to
tell
in
a
chart
the
difference
between
0.99
99.99999,
so
so
you
might
want
to
do
something
like
god.
Let's
say
you
have
two
of
them
in
each
one.
Normally,
whatever
threshold
of
0.5,
you
could
add
them
up
and
have
a
threshold
of
plate
should
be
lower
than
one,
and
so
that
means,
if
both
of
them
are
four
nines
instead
of
five
nines.
A
C
A
G
G
A
C
D
C
A
I
get
one
more.
This
is
nothing
to
an
anomaly
models.
It's
a
completely
separate
question,
see
how
am
I
supposed
to
handle
data
that
comes
in
at
different
intervals,
when
swarming
or
feeding
data
into
nupoc
at
some
time,
stamps
I
don't
have
input
for
all
fields.
What
should
I
do
so,
I
think
the
you.
A
Just
to
copy
this
guy
over
say
easiest,
if
you
don't
have
any
other
information,
because
that's
basically
saying
constant
for
that
half
an
hour
and
you
could
have
a
more
complicated
model
of
this.
So
if
you
know
like,
for
example,
if
this
is
day
of
the
week,
let's
say
so-
and
this
is
hourly
data.
This
is
what
day
of
the
week
would
only
change
every
24
rows.
So
you
could.