Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: With the increasing complexity of mobile device applications, these devices
are evolving toward high agility. This shift imposes new demands on mobile
sensing, particularly in achieving high-accuracy and low-latency. Event-based
vision has emerged as a disruptive paradigm, offering high temporal resolution
and low latency, making it well-suited for high-accuracy and low-latency
sensing tasks on high-agility platforms. However, the presence of substantial
noisy events, lack of stable, persistent semantic information, and large data
volume pose challenges for event-based data processing on resource-constrained
mobile devices. This paper surveys the literature from 2014 to 2025 and
presents a comprehensive overview of event-based mobile sensing, encompassing
its fundamental principles, event \textit{abstraction} methods,
\textit{algorithm} advancements, and both hardware and software
\textit{acceleration} strategies. We discuss key \textit{applications} of event
cameras in mobile sensing, including visual odometry, object tracking, optical
flow, and 3D reconstruction, while highlighting challenges associated with
event data processing, sensor fusion, and real-time deployment. Furthermore, we
outline future research directions, such as improving the event camera with
advanced optics, leveraging neuromorphic computing for efficient processing,
and integrating bio-inspired algorithms. To support ongoing research, we
provide an open-source \textit{Online Sheet} with recent developments. We hope
this survey serves as a reference, facilitating the adoption of event-based
vision across diverse applications.