研究生: |
郭珈妤 Guo, Karen |
---|---|
論文名稱: |
藉由學習稀疏性表達特徵的方式偵測圖中視覺顯著區域 Learning Sparse Feature Dictionary for Saliency Detection |
指導教授: |
陳煥宗
Chen, Hwann-Tzong |
口試委員: |
劉庭祿
Liu, Tyng-Luh 賴尚宏 Lai, Shang-Hong |
學位類別: |
碩士 Master |
系所名稱: |
電機資訊學院 - 資訊工程學系 Computer Science |
論文出版年: | 2012 |
畢業學年度: | 100 |
語文別: | 英文 |
論文頁數: | 22 |
中文關鍵詞: | 顯著性 、稀疏式 、字典 、描述圖片方式 、注視區域 、學習方法 |
外文關鍵詞: | Saliency, Sparse Coding, Dictionary, Feature, fixation, Learning |
相關次數: | 點閱:2 下載:0 |
分享至: |
查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報 |
圖中顯著性的偵測在計算機視覺研究領域中變得越來越流行。在這篇論文中,我們提出了一種新的方法來產生顯著性偵測的逼近結果圖。其基本思路是使用稀疏編碼係數當作是一種特徵來組成我們要的結果。我們的方法包括兩部分:訓練步驟和測試步驟。在訓練步驟,我們利用在圖片中找到的特徵以及利用人眼注視這張圖的資料來產生圖片相關的字典和如何從稀疏係數轉換成結果;在測試步驟中,給一張新的圖片,我們可以藉由基於特徵的字典得到其相應的稀疏編碼,然後生成結果。我們使用洗牌式AUC以及兩個圖片資料庫來評估我們的研究結果並證明我們的方法可以利用稀疏係數來學習和產生顯著性偵測的結果圖。
Saliency detection becomes more and more popular in computer vision research field. In this thesis we present a new method to generate the saliency map. The basic idea
is to use the sparse coding coefficients as features and find a way to reconstruct the sparse features into a saliency map. Our method consists of two parts: training step and testing step. In the training step, we use the features generated from images and the fixation values from ground-truth fixation map to train the feature-based
dictionary for the sparse coding and the fixation-based dictionary for converting the sparse coding to a saliency map. In the test step, given a new image, we can get
its corresponding sparse coding from the feature-based dictionary and then generate the result. We evaluate our results on two datasets with the shued AUC score and demonstrate that our method gives an efficient sparse coding learning and combination for saliency detection.
[1] Codruta Orniana Ancuti, Cosmin Ancuti, and Philippe Bekaert. Enhancing by
saliency-guided decolorization. In CVPR, pages 257{264, 2011.
[2] Shai Avidan and Ariel Shamir. Seam carving for content-aware image resizing.
ACM Trans. Graph., 26(3):10, 2007.
[3] Ali Borji and Laurent Itti. Exploiting local and global patch rarities for saliency
detection. In CVPR, 2012.
[4] Neil D. B. Bruce and John K. Tsotsos. Saliency based on information maximization.
In NIPS, 2005.
[5] Kai-Yueh Chang, Tyng-Luh Liu, Hwann-Tzong Chen, and Shang-Hong Lai. Fusing
generic objectness and visual saliency for salient object detection. In ICCV,
pages 914{921, 2011.
[6] Kai-Yueh Chang, Tyng-Luh Liu, and Shang-Hong Lai. From co-saliency to cosegmentation:
An ecient and fully unsupervised energy minimization model.
In CVPR, pages 2129{2136, 2011.
[7] Jonathan Harel, Christof Koch, and Pietro Perona. Graph-based visual saliency.
In NIPS, pages 545{552, 2006.
[8] Laurent Itti. Automatic foveation for video compression using a neurobiological
model of visual attention. IEEE Transactions on Image Processing, 13(10):1304{
1318, 2004.
21
[9] Laurent Itti, Christof Koch, and Ernst Niebur. A model of saliency-based visual
attention for rapid scene analysis. IEEE Trans. Pattern Anal. Mach. Intell.,
20(11):1254{1259, 1998.
[10] Tilke Judd, Krista A. Ehinger, Fredo Durand, and Antonio Torralba. Learning
to predict where humans look. In ICCV, pages 2106{2113, 2009.
[11] David G. Lowe. Distinctive image features from scale-invariant keypoints. In-
ternational Journal of Computer Vision, 60(2):91{110, 2004.
[12] Julien Mairal, Francis Bach, Jean Ponce, and Guillermo Sapiro. Online learning
for matrix factorization and sparse coding. Journal of Machine Learning
Research, 11:19{60, 2010.
[13] Federico Perazzi, Philipp Krahenbuhl, Yael Pritch, and Alexander Hornung.
Saliency lters: Contrast based ltering for salient region detection. In CVPR,
2012.
[14] Jingdong Wang, Long Quan, Jian Sun, Xiaoou Tang, and Heung-Yeung Shum.
Picture collage. In CVPR (1), pages 347{354, 2006.
[15] Meng Wang, Janusz Konrad, Prakash Ishwar, Kevin Jing, and Henry A. Rowley.
Image saliency: From intrinsic to extrinsic context. In CVPR, pages 417{424,
2011.
[16] Jimei Yang and Ming-Hsuan Yang. Top-down visual saliency via joint crf and
dictionary learning. In CVPR, 2012.
[17] L. Zhang, M. H. Tong, T. K. Marks, H. Shan, and G. W. Cottrell. SUN: A
Bayesian framework for saliency using natural statistics. Journal of Vision, 8:32{
32, 2008.