最終更新日:2022/12/24
Specifically, our method contains three key stages: 1) each expression video clip is characterized as a spatial-temporal manifold (STM) formed by dense low-level features; 2) a Universal Manifold Model (UMM) is learned over all low-level features and represented as a set of local modes to statistically unify all the STMs. 3) the local modes on each STM can be instantiated by fitting to UMM, and the corresponding expressionlet is constructed by modeling the variations in each local mode.
音声機能が動作しない場合はこちらをご確認ください
編集履歴(0)
元となった例文
Specifically,
our
method
contains
three
key
stages:
1)
each
expression
video
clip
is
characterized
as
a
spatial-temporal
manifold
(STM)
formed
by
dense
low-level
features;
2)
a
Universal
Manifold
Model
(UMM)
is
learned
over
all
low-level
features
and
represented
as
a
set
of
local
modes
to
statistically
unify
all
the
STMs.
3)
the
local
modes
on
each
STM
can
be
instantiated
by
fitting
to
UMM,
and
the
corresponding
expressionlet
is
constructed
by
modeling
the
variations
in
each
local
mode.