TY - JOUR
T1 - Exploring contextual redundancy in improving object-based video coding for video sensor networks surveillance
AU - Tsai, Tsung Han
AU - Lin, Chung Yuan
N1 - Funding Information:
Manuscript received December 15, 2010; revised June 20, 2011 and October 17, 2011; accepted November 30, 2011. Date of publication December 21, 2011; date of current version May 11, 2012. This work was supported by the National Science Council, Taiwan, under Grant 100-2220-E-008-001. The associate editor coordinating the review of this manuscript and approving it for publication was Prof. Thinh Nguyen.
PY - 2012
Y1 - 2012
N2 - In recent years, intelligent video surveillance attempts to provide content analysis tools to understand and predict the actions via video sensor networks (VSN) for automated wide-area surveillance. In this emerging network, visual object data is transmitted through different devices to adapt to the needs of the specific content analysis task. Therefore, they raise a new challenge for video delivery: how to efficiently transmit visual object data to various devices such as storage device, content analysis server, and remote client server through the network. Object-based video encoder can be used to reduce transmission bandwidth with minor quality loss. However, the involved motion-compensated technique often leads to high computational complexity and consequently increases the cost of VSN. In this paper, contextual redundancy associated with background and foreground objects in a scene is explored. A scene analysis method is proposed to classify macroblocks (MBs) by type of contextual redundancy. The motion search is only performed on the specific type of context of MB which really involves salient motion. To facilitate the encoding by context of MB, an improved object-based coding architecture, namely dual-closed-loop encoder, is derived. It encodes the classified context of MB in an operational rate-distortion-optimized sense. The experimental results show that the proposed coding framework can achieve higher coding efficiency than MPEG-4 coding and related object-based coding approaches, while significantly reducing coding complexity.
AB - In recent years, intelligent video surveillance attempts to provide content analysis tools to understand and predict the actions via video sensor networks (VSN) for automated wide-area surveillance. In this emerging network, visual object data is transmitted through different devices to adapt to the needs of the specific content analysis task. Therefore, they raise a new challenge for video delivery: how to efficiently transmit visual object data to various devices such as storage device, content analysis server, and remote client server through the network. Object-based video encoder can be used to reduce transmission bandwidth with minor quality loss. However, the involved motion-compensated technique often leads to high computational complexity and consequently increases the cost of VSN. In this paper, contextual redundancy associated with background and foreground objects in a scene is explored. A scene analysis method is proposed to classify macroblocks (MBs) by type of contextual redundancy. The motion search is only performed on the specific type of context of MB which really involves salient motion. To facilitate the encoding by context of MB, an improved object-based coding architecture, namely dual-closed-loop encoder, is derived. It encodes the classified context of MB in an operational rate-distortion-optimized sense. The experimental results show that the proposed coding framework can achieve higher coding efficiency than MPEG-4 coding and related object-based coding approaches, while significantly reducing coding complexity.
KW - Contextual redundancy coding
KW - intelligent video surveillance
KW - object-based video coding
KW - operational rate-distortion theory
KW - visual sensor network
UR - http://www.scopus.com/inward/record.url?scp=84861139241&partnerID=8YFLogxK
U2 - 10.1109/TMM.2011.2180705
DO - 10.1109/TMM.2011.2180705
M3 - 期刊論文
AN - SCOPUS:84861139241
SN - 1520-9210
VL - 14
SP - 669
EP - 682
JO - IEEE Transactions on Multimedia
JF - IEEE Transactions on Multimedia
IS - 3 PART 2
M1 - 6111302
ER -