簡易檢索 / 詳目顯示

研究生: 王嚴璋
Wang, Yen-Chang
論文名稱: 利用光場相機所獲得之聚焦於不同距離的影像產生標籤圖及擴展景深
Generating Label Map and Extending Depth of Field from Different Focus Images Obtained by Means of Light Field Camera
指導教授: 陳永昌
Chen, Yung-Chang
口試委員: 賴尚宏
Lai, Shang-Hong
黃仲陵
Huang, Chung-Lin
學位類別: 碩士
Master
系所名稱: 電機資訊學院 - 電機工程學系
Department of Electrical Engineering
論文出版年: 2012
畢業學年度: 100
語文別: 英文
論文頁數: 55
中文關鍵詞: 光場相機聚焦景深標籤圖
外文關鍵詞: light field camera, focus, depth of field, label map
相關次數: 點閱:2下載:0
分享至:
查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報
  • 近來手持式光場相機的發明在攝影學上掀起了一場革命。雖然光場相機的概念在多年以前早被提出,但卻因為早期的設計過於笨重且不便於攜帶因此無法被廣泛的應用。有別於一般相機,透過其特殊的設計,在拍攝照片時光場相機除了紀錄光的強度之外,同時也記錄了光的角度。有了光的角度這項額外的資訊,我們可進行許多如數位變焦、改變視角及深度估測等應用,並且可於電腦視覺、計算機圖學、機械視覺上廣泛的使用。在本篇論文中,我們主要使用數位變焦的功能來獲得聚焦在不同深度的一連串影像。
    我們設計了一個能量函數,經由將此能量函數最小化可得到一張標籤圖,圖中每點的值代表該點在哪張影像中最為清晰。該能量函數的主要核心是以一個像元基礎(pixel-based)的自適應式聚焦估測(adaptive focus measure)為主,並輔以一個區域式(region-based)的Adaboost分類法。藉由這個能量函數,我們可以獲得比使用傳統聚焦測深的方法更為強健且適應性更高的結果。為了便於其他的應用,我們同時也利用這張標籤圖來產生一張虛擬的全聚焦圖。
    我們使用Lytro的手持式光場相機來捕捉現實世界的場景,並透過電腦將其重新聚焦在不同的深度以取得一連串的影像集合。接著我們利用上述的能量函數來對每個標籤給予每個像素不同的能量。最後,透過將此能量最小化來產生一張標籤圖及一張虛擬的全聚焦圖。


    Recently, the invention of hand-held light field camera raises a revolution in photography. We can record not only the intensity of light but also the direction of light by the light field camera. With the additional information, we could process many applications such as digital refocusing, moving observer, and depth estimation which can be utilized to computer vision, computer graphics, and machine vision. In this thesis, we mainly concentrate on the function of digital refocusing which can easily get series of images with different focal length.
    We design an energy function and minimize it to get a label map which represents the index of sharpest image for each pixel. The primary core of the energy function is a pixel-based AFM method and a region-based adaboost classification method is secondary. We get a more robust result than traditional depth from focus (DFF) method through this energy function. We also generate a virtual all-focus image for further applications by utilizing the label map.
    We us the Lytro light field camera to capture real world scene and refocus it to a set which contains several images with different focal length. For each pixel, we compute the cost to each label by applying the energy function described above. Finally, we generate a label map and a virtual all-focus image by our algorithm.

    Abstract i Table of Contents ii List of Figures iv List of Tables vi Chapter1 Introduction 1 1.1 Motivation 2 1.2 Thesis Organization 4 Chapter2 Related Work 5 2.1 Image Formation and Depth of Field 5 2.2 Depth from Focusing 9 2.3 Depth from Defocusing 11 2.4 Light Field Camera 13 Chapter3 Proposed Method 21 3.1 System Overview 21 3.2 Adaboost Classification 23 3.3 Adaptive Focus Measure 31 3.4 Energy Minimization and Re-labeling 34 3.5 Extending Depth of Field 40 Chapter4 Experimental Results 43 4.1 Lytro Light Field Camera 43 4.2 Adaboost Classification 45 4.3 Improvement from Adaptive Focus Measure 46 4.4 Results and Discussion 47 Chapter5 Conclusions and Future Works 52 5.1 Conclusions 52 5.2 Future Works 53 Reference 54

    [1] R. Ng, "Fourier slice photography," in ACM Trans. Graph (Proc. SIGGRAPH'05), vol. 24, no. 3, pp. 735-744, Jul. 2005.

    [2] R. Ng, M. Levoy, M. Br6dif, G. Duval, M. Horowitz, and P. Hanrahan, "Light field photography with a hand-held plenoptic camera," Stanford University CSTR 2005-02, 2005.

    [3] Ralph E. Jacobson, Sidney F. Ray, Geoffrey G. Atteridge, and Norman R. Axford, “The Manual of Photography: Photographic and Digital Imaging, 9th ed. Ed.”

    [4] C. Y. Wee and R. Paramesran. "Measure of image sharpness using eigenvalues." Inf. Sci., 177(12):2533–2552, 2007.

    [5] E.H. Adelson, , Bergen, J.R. (1991). "The plenoptic function and the elements of early vision", In Computation Models of Visual Processing, M. Landy and J.A. Movshon, eds., MIT Press, Cambridge, 1991, pp. 3–20.

    [6] M. Levoy, and P. Hanrahan, "Light Field Rendering", in ACM International Conference on Computer Graphics and Interactive Techniques, pp. 31-42, Aug. 1996

    [7] E. H. Adelson, and J. U. A. Wang, ”Single lens stereo with a plenoptic camera,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 14, no. 2, pp. 99-106, Feb. 1992

    [8] B. Wilburn, N. Joshi, V. Vaish, E. Talvala, E.Antunez, A. Barth, A. Adams, M. Horowitz, and M.Levoy, “High performance imaging using large camera arrays,” ACM Transactions on Graphics, vol. 24, no. 3, pp. 765-776, Jul. 2005.

    [9] C. Chen, Y. Lu, and M. Su, “Light field based digital refocusing using a DSLR camera with a pinhole array mask,” in IEEE International Conference on Acoustics, Speech, and Signal Processing, pp. 754-757, Mar. 2010.

    [10] H. Li, King N. Ngan, “Learning to Extract Focused Objects From Low DOF Images”, IEEE Transactions on circuits and system for video technology, vol. 21, no. 11, November 2011

    [11] J. Shi and J. Malik, ”Normalized Cuts and Image Segmentation,” IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 22, no. 8, pp. 888-905, Aug. 2000.

    [12] D. Comaniciu and P. Meer, “Mean Shift: A Robust Approach Toward Feature Space Analysis,” IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 24, no. 5, pp. 603-619, May 2002.

    [13] P. F. Felzenszwalb , D. P. Huttenlocher, “Efficient graph-based image segmentation,” Int. J. Comput. Vision, vol. 59, no. 2, pp. 167–181, Sep. 2004.

    [14] H. Li and K. N. Ngan, “Unsupervised video segmentation with low depth of field,” IEEE Trans. Circuits Syst. Video Technol., vol. 17, no. 12, pp. 1742–1751, Dec. 2007.

    [15] X. Wang and E. Grimson, “Spatial latent Dirichlet allocation,” in Proc. Neural Inform. Process. Syst. Conf., Dec. 2007, pp. 1577–1584.

    [16] S. Gould, R. Fulton, and D. Koller, “Decomposing a scene into geometric and semantically consistent regions,” in Proc. IEEE ICCV, Sep.–Oct. 2009, pp. 1–8.

    [17] P. Viola and M. Jones, “Rapid object detection using a boosted cascade of simple features,” in Proc. IEEE Comput. Soc. Conf. CVPR, vol. 1. Dec. 2001, pp. 511–518.

    [18] T. Aydin and Y. Akgul, “A new adaptive focus measure for shape from focus,” in "BMVC08,” (2008).

    [19] Carlo Tomasi and Roberto Manduchi. “Bilateral filtering for gray and color images.” In ICCV, pages 839–846, 1998.

    [20] Boykov, Y., Veksler, O., Zabih, R., 2001. "Fast approximate energy minimization via graph cuts." IEEE Trans. PAMI 23, 1222–1239.

    無法下載圖示 全文公開日期 本全文未授權公開 (校內網路)
    全文公開日期 本全文未授權公開 (校外網路)

    QR CODE