To comprehend long duration videos, the deep video understanding (DVU) task is proposed to recognize interactions on scene level and relationships on movie level and answer questions on these two levels. In this paper, we propose a solution to the DVU task which applies joint learning of interaction and relationship prediction and multimodal feature fusion. Our solution handles the DVU task with three joint learning sub-tasks: scene sentiment classification, scene interaction recognition and super-scene video relationship recognition, all of which utilize text features, visual features and audio features, and predict representations in semantic space. Since sentiment, interaction and relationship are related to each other, we train a unified framework with joint learning. Then, we answer questions for video analysis in DVU according to the results of the three sub-tasks. We conduct experiments on the HLVU dataset to evaluate the effectiveness of our method.
Joint Learning for Relationship and Interaction Analysis in Video with Multimodal Feature Fusion
Beibei Zhang,Fan Yu,Yanxin Gao,Tongwei Ren,Gangshan Wu
Published 2021 in ACM Multimedia
ABSTRACT
PUBLICATION RECORD
- Publication year
2021
- Venue
ACM Multimedia
- Publication date
2021-10-17
- Fields of study
Computer Science
- Identifiers
- External record
- Source metadata
Semantic Scholar
CITATION MAP
EXTRACTION MAP
CLAIMS
- No claims are published for this paper.
CONCEPTS
- No concepts are published for this paper.
REFERENCES
Showing 1-21 of 21 references · Page 1 of 1
CITED BY
Showing 1-12 of 12 citing papers · Page 1 of 1