Please enable JavaScript.
Coggle requires JavaScript to display documents.
Ogawa Et Al (2018), Favorite Video Classification Based On Multimodal…
Ogawa Et Al (2018), Favorite Video Classification Based On Multimodal Bidirectional LSTM
Model Architecture
RNN
LSTM
-
Used in this research to take advantage of long-term information found across frames in video clips as well as signals in EEG
-
Sequence-to-sequence
-
Used for problems like translation where your input is a sentence and your output is also a sentence
Bi-directional RNN
An RNN approach where information is propagated forward in time and then backwards in time before generating output
-
Input Layer
Composed of a series of vectors with 1024 dimension video features concatenated with 1024 dimension EEG features
-
-