Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_cv 90% Match Research Paper Computer Vision Researchers,AI Engineers,Robotics Engineers,Security Analysts 3 weeks ago

Dual-Stream Alignment for Action Segmentation

computer-vision › video-understanding
📄 Abstract

Abstract: Action segmentation is a challenging yet active research area that involves identifying when and where specific actions occur in continuous video streams. Most existing work has focused on single-stream approaches that model the spatio-temporal aspects of frame sequences. However, recent research has shifted toward two-stream methods that learn action-wise features to enhance action segmentation performance. In this work, we propose the Dual-Stream Alignment Network (DSA Net) and investigate the impact of incorporating a second stream of learned action features to guide segmentation by capturing both action and action-transition cues. Communication between the two streams is facilitated by a Temporal Context (TC) block, which fuses complementary information using cross-attention and Quantum-based Action-Guided Modulation (Q-ActGM), enhancing the expressive power of the fused features. To the best of our knowledge, this is the first study to introduce a hybrid quantum-classical machine learning framework for action segmentation. Our primary objective is for the two streams (frame-wise and action-wise) to learn a shared feature space through feature alignment. This is encouraged by the proposed Dual-Stream Alignment Loss, which comprises three components: relational consistency, cross-level contrastive, and cycle-consistency reconstruction losses. Following prior work, we evaluate DSA Net on several diverse benchmark datasets: GTEA, Breakfast, 50Salads, and EgoProcel. We further demonstrate the effectiveness of each component through extensive ablation studies. Notably, DSA Net achieves state-of-the-art performance, significantly outperforming existing

Key Contributions

Proposes the Dual-Stream Alignment Network (DSA Net) for action segmentation, incorporating a second stream of learned action features to capture both action and transition cues. It utilizes a Temporal Context block with cross-attention and a novel Quantum-based Action-Guided Modulation (Q-ActGM) to fuse information from both streams effectively.

Business Value

Enables more accurate and detailed analysis of human actions in videos, leading to improved applications in security surveillance, human-robot interaction, and automated video summarization. This can enhance safety, efficiency, and user experience.