§ 瀏覽學位論文書目資料
  
系統識別號 U0002-0711201215583000
DOI 10.6846/TKU.2013.00220
論文名稱(中文) 基於低運算複雜度範圍樹之感興趣物件檢索技術應用於影片摘要系統
論文名稱(英文) An Efficient Region of Interest Retrieval Technique Based on Low-Complexity Range Tree for Video Synopsis System
第三語言論文名稱
校院名稱 淡江大學
系所名稱(中文) 電機工程學系碩士班
系所名稱(英文) Department of Electrical and Computer Engineering
外國學位學校名稱
外國學位學院名稱
外國學位研究所名稱
學年度 101
學期 1
出版年 102
研究生(中文) 謝吉芳
研究生(英文) Chi-Fang Hsieh
學號 699450044
學位類別 碩士
語言別 繁體中文
第二語言別
口試日期 2012-10-25
論文頁數 71頁
口試委員 指導教授 - 江正雄(chiang@ee.tku.edu.tw)
委員 - 夏至賢(chhsia@ee.tku.edu.tw)
委員 - 周建興(chchou@mail.tku.edu.tw)
委員 - 許明華(sheumh@yuntech.edu.tw)
關鍵字(中) 物件偵測
物件辨識
影片檢索系統
K-D樹
關鍵字(英) Object detection
object tracking
video retrieval system
low-complexity
第三語言關鍵字
學科別分類
中文摘要
隨著近年攝影機的發展,使得數位化及高解析度的監視系統日漸普及,也對保全系統造成革命性的衝擊,根據研究報告指出,在英國有四百多萬的監視攝影機分布於城市街道,人們可以藉由將攝影機擺在關鍵的地點,藉此對該區域做日以繼夜的監控。因此,數位監視攝影系統在現今的保全上扮演著愈來愈重要的角色,越來越多的監視攝影機被配置於各種場合以確保人員及財產的安全。然而,大量的數位監視攝影資訊管理並不容易,需要花費大量的人力來監視監控畫面,負責監視監控畫面的人員在長時間專注盯著螢幕的情況之下也會造成的體力與精神的衰減與考驗,往往導致安全維護效率的降低。
在本論文中,我們提出一個高效率的影片檢索系統,透過此系統,使用者感興趣的物件(Region of Interest, ROI)可以藉由電腦視覺中的物件偵測與物件辨識在長時間的影片中快速且有效率的被擷取出來,並且快速得索引與瀏覽。其利用了高斯混合模型(Gaussian Mixture Model, GMM)作為物件偵測的方法,為了讓使用者有效率的找出符合搜尋條件的物件,本文提出低複雜度範圍樹(Range Tree)來做為搜索的方法,其透過影片摘要(Video Abstraction)的技術,搜索到的物件將在短時間中播放,藉此達到快速檢索與瀏覽進而節省時間的目的。透過本系統,前處理物件擷取的部分FPS可以達到32,而搜索的的部分,時間複雜度則是從O(N)下降為O(logD-1N)。
英文摘要
With the development of surveillance, digital and high resolution surveillance become more and more popular, impacting security system. It is reported that in the UK alone there are 4.2 million security cameras covering city streets. More and more digital surveillances are placed in everywhere for safety and security. Therefore digital surveillance system plays indispensable role in security today. However, the great amount of video captured form digital surveillance is difficult to manage and retrieve. 
In this study, we propose an efficient video retrieval technique. With the system, the Region of Interest (ROI) could be extracted in long video effectively and user could browse it with quick and easy way. According to the characteristic of the object in foreground distribution for the real-world video sequence, this work employs Gaussian Mixture Model (GMM) is used for object detection. In order to let users, a new video synopsis search approach, low-complexity rang tree algorithm, is proposed to improve the search the objects matching the conditions effectively in this work. With the time and space redundancy-reducing technique of video synopsis, the objects searched by users could be display in very short time. Therefore wanted objects and events could be found out and displayed quickly without wasting time watching the part without ROI. For the test video sequences, it can have an accuracy rate of 95%  and achieve 32 Frame Per Second (FPS) of online phase in processing speed and time complexity of searching decrease from O(N) to O(logD-1N).
第三語言摘要
論文目次
中文摘要	I
英文摘要	II
內文目錄	III
圖表目錄	V

第一章  緒論	1
1.1 研究背景與動機	1
1.2 論文架構	3
第二章  相關技術	4
2.1物件偵測與追蹤系統	4
    2.1.1物件偵測	5
    2.1.2物件追蹤	8
2.2影片摘要相關研究	10
    2.2.1靜態摘要	11
    2.2.2動態摘要	12
    2.2.3其他	13

2.3影片檢索相關研究	15
2.4總結	17
第三章  事件偵測流程	19
3.1偵測前景物件	19
3.2色相轉換	20
3.3離散小波轉換	25
3.4高斯混合模型	28
3.5形態學補償	31
3.6標籤化	35
第四章  影片摘要檢索系統	40
4.1 概述	40
4.2 精簡影片摘要	40
4.3 物件檢索基於改良式範圍樹	47
第五章  實驗結果	61
第六章  結論	67
參考文獻	68

圖目錄

圖2.1高斯背景模型偵測法	7
圖2.2自適應性的快速移動估測流程	10
圖2.3影片摘要	15
圖2.4影片檢索系統	16
圖3.1 online phase流程示意圖	20
圖3.2色相轉換示意圖	21
圖3.3彩色影像轉換為灰階影像	21
圖3.4 RGB色彩方塊	22
圖3.5加色模式	22
圖3.6 YUV的色差信號與顏色對應關係	23
圖3.7以SMDWT降解析度示意圖	27
圖3.8高斯混合模型物件偵測流程	30
圖3.9膨脹運算示意圖	32
圖3.10侵蝕運算示意圖	33
圖3.11影像經由斷開運算之結果	35
圖3.12標籤化時掃描影像的方向	36
圖3.13標籤化時所檢查的近鄰點	37
圖3.14給予前景物件暫時的標籤值	38
圖3.15透過群組編號給定最後的標籤值	39
圖4.1影片摘要示意圖	41
圖4.2產生影片摘要流程圖	42
圖4.3閃頻效應	43
圖4.4閃頻效應產生的重疊	44
圖4.5 Dead Zone產生與物件貼合	45
圖4.6連續物件流	47
圖4.7檢索系統圖	49
圖4.8 二維樹狀結構	50
圖4.9 二維範圍樹的搜索示意圖	52
圖4.10節點所包含的子集合陣列	53
圖4.11搜尋效能圖	55
圖4.12特徵點範圍搜索	56
圖4.13搜尋方式	57
圖4.14系統操作流程	60
圖5.1影片摘要實驗結果截圖	63
圖5.2搜索效能曲線圖	65
圖5.3搜索比對	66























表目錄

表4.1樹狀結構的效能	50
表4.2改良範圍樹的效能	53
表5.1系統物件偵測的效能	56
表5.2系統影片摘要的壓縮效能	57
表5.3系統搜尋效能	59
參考文獻
[1]	J. Nam and A.H. Tewfik, “Video abstract of video,” IEEE Workshop on Multimedia Signal Processing , pp.117-122, 1999.
[2]	C. Pal and N. Jojic, “Interactive montages of sprites for indexing and summarizing security video,” IEEE Computer Society Conference on  Computer Vision and Pattern Recognition, vol. 2, pp. 1192, June 2005.
[3]	M. Furini, “Fast play: a novel feature for digital consumer video devices,” IEEE Transactions on Consumer Electronics, vol. 54, no. 2, pp. 513-520, May 2008.
[4]	M. A. Smith and T. Kanade, “Video skimming and characterization through the combination of image and language understanding techniques,” IEEE Proc. Computer Vision and Pattern Recognition, pp. 775-781, Jun. 1997.
[5]	Y. Zhuang, Y. Rui, T.-S. Huang, and S. Mehrotra, “Adaptive key frame extraction using unsupervised clustering,” IEEE Proc. Image Processing, vol.1, pp. 866-870, Oct. 1998.
[6]	C. Kim and J. N. Hwang, “Object-based video abstraction for video surveillance systems,” IEEE Transactions on Circuits and Systems for Video Technology, vol.12, no.12, pp. 1128- 1138, Dec. 2002.
[7]	D. Farin, W. Effelsberg and P.H.N. de With, “Robust clustering-based video-summarization with integration of domain-knowledge,” IEEE International Conference on Multimedia and Expo, vol.1, no., pp. 89- 92, 2002.
[8]	A. Rav-Acha, Y. Pritch, and S. Peleg, “Making a long video short: dynamic video synopsis,” IEEE Conference on Computer Vision and Pattern Recognition, pp. 435-441, June 2006.
[9]	Y. Pritch, A. Rav-Acha, and S. Peleg, “Nonchronological video synopsis and indexing,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 30, no. 11, pp. 1971-1984, Nov. 2008.
[10]	C.-H. Lin and Y.-C. Chen, “A LDWT-based surveillance video synopsis system using color-based SVM,” IPPR Conference on Computer Vision, Graphics, and Images Processing, Aug. 2011.
[11]	 J. Ren, P. Astheimer, D.D. Feng, “Real-time moving object detection under complex background,” International Symposium on Image and Signal Processing and Analysis, vol. 2, pp. 662-667, Sept. 2003.
[12]	 D. Fan, M. Cao and C. Lv, “An updating method of self-adaptive background for moving objects detection in video,” International Conference on Audio, Language and Image Processing, pp.1497-1501, July 2008.
[13]	 W. Zhang, M.-J. Wu, G. Wang and H. Yin, “An adaptive computational model for salient object detection,” IEEE Transactions on Multimedia, vol. 12, no. 4, pp. 300-316, June 2010.
[14]	 K. Chamnongthai, “Efficient particle filter using non-stationary gaussian based model,” International Conference on Electrical Engineering/Electronics, Computer, Telecommunications and Information Technology, pp.468-471, May. 2011.
[15]	 O. D. Nouar, G. Ali, and C. Raphael, “Improved object tracking with CamShift algorithm,” IEEE International Conference on Acoustics, Speech and Signal Processing, vol. 2, pp. II-657-II-660, May. 2006.
[16]	 C. Zhang and Y. Qiao, “An improved CamShift algorithm for target tracking in video surveillance,” IT&T Conference, 2009.
[17]	 C.-H. Hsia, Y.-J. Liou, Y.-J. Dai, and J.-S. Chiang, “Adaptive motion estimation based on CamShift algorithm for moving object tracking,” International Conference on Service and Interactive Robotics, pp. 316-321, Nov. 2011.
[18]	X. Zhu, A. K. Elmagarmid, X. Xue, L. Wu, A. C. Catlin, “InsightVideo: toward hierarchical video content organization for efficient browsing, summarization and retrieval,” IEEE Transactions on Multimedia, vol.7, no.4, pp. 648- 666, Aug. 2005.
[19]	 B. Sugandi, H. Kim, J. K. Tan, and S. Ishikawa, “Tracking of moving objects by using a low resolution image,” International Conference on Innovative Computing, Information and Control, pp. 408-408, Sept. 2007.
[20]	 C.-H. Hsia, J.-M. Guo, and J.-S. Chiang, “Improved Low-complexity algorithm for 2-D integer lifting-based discrete wavelet transform using symmetric mask-based scheme,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 19, pp. 1202-1208, Aug. 2009.
[21]	 F.-H. Cheng, and Y.-L. Chen, “Real time multiple objects tracking and identification based on discrete wavelet transform,” Pattern Recognition, vol. 39, no. 3, pp. 1126-1139, Jun. 2006.
[22]	 S.-D. Jean, C.-M. Liu, C.-C. Chang, and Z. Chen, “A new algorithm and its VLSI architecture design for connected component labeling,” International Symposium on Circuits and Systems, vol. 2, pp. 565-568, May 1994.
[23]	 W.-K. Chan and S.-Y. Chien, “Subword parallel architecture for connected componentlabeling and morphological operations,” IEEE Asia Pacific Conference on Circuits and Systems, pp. 936-939, Dec. 2006.
[24]	 K. Suzuki, I. Horiba, and N. Sugie, “Linear-time connected-component labeling based on sequential local operations,” Computer Vision and Image Understanding, vol. 89, pp. 1-23, Jan. 2003.
[25]	 H. Hedberg, F. Kristensen, and V. Owall, “Implementation of a labeling algorithm based on contour tracing with feature extraction,” International Symposium on Circuits and Systems, pp. 1101-1104, May 2007.
[26]	 L. He, Y. Chao and K. Suzuki, “A run-based two-scan labeling algorithm,” IEEE Transactions on Image Processing, vol. 17, pp. 749-756, May 2008.
[27]	 D. G. Lowe, “Distinctive Image features from scale-invariant keypoints,” International Journal of Computer Vision, vol. 60, no. 2, pp. 91-110, Nov 2004.
[28]	 C.-H. Hsia and J.-S. Chiang, “Real-time moving objects detection and tracking with direct LL-mask band scheme,” International Journal of Innovative Computing, Information and Control, vol. 8, no. 7(A), pp. 4451-4468, Jul. 2012.
論文全文使用權限
校內
校內紙本論文立即公開
同意電子論文全文授權校園內公開
校內電子論文立即公開
校外
同意授權
校外電子論文立即公開

如有問題,歡迎洽詢!
圖書館數位資訊組 (02)2621-5656 轉 2487 或 來信