研究生: |
楊承翰 Yang, Cheng-Han |
---|---|
論文名稱: |
以靜態影像及動態影片為來源之浮現錯覺合成 Emergence Illusion Synthesis using Still Images and Dynamic Videos |
指導教授: |
朱宏國
Chu, Hung Kuo |
口試委員: |
李潤容
Lee, Ruen Rone 姚智原 Yao, Chih Yuan 莊永裕 Chuang, Yung Yu |
學位類別: |
碩士 Master |
系所名稱: |
電機資訊學院 - 資訊工程學系 Computer Science |
論文出版年: | 2016 |
畢業學年度: | 104 |
語文別: | 中文 |
論文頁數: | 60 |
中文關鍵詞: | 浮現 、錯覺影像 、非相片質感渲染 、生物運動 、完形心理學 、電腦識別 、影片追蹤 、驗證碼 |
外文關鍵詞: | Emergence, Illustration Art, Non-Photorealistic Rendering, Biological Motion, Pattern Recognition, Gestalttheorie, Video Tracking, CAPTCHA |
相關次數: | 點閱:1 下載:0 |
分享至: |
查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報 |
所謂浮現影像,是一種在一陣視覺雜訊中,人類能夠快速與自動的感知與歸納其中主體資訊的一種錯覺圖片。浮現影片則是浮現影像的動態延伸,在播放時人類僅能夠識別出影片中正在運動的物體,而該物體在靜止時則是一片雜亂無章的訊息。同時,浮現影像對於電腦的影像識別或影片追蹤技術皆有一定的干擾性,使得電腦無法識別該類影象或圖片,而此現象皆可以透過一些視覺心裡學的原則分析與歸納之。
近年來,電腦圖學興起一股藉由設計自動化演算法來重製大量錯覺藝術
的熱潮,將特別的錯覺藝術套用在照片、影片及動畫當中。本研究提出一套產生浮現像與浮現影片的系統,透過使用者輸入或是網路爬圖系統,即能自動並大量化的合成浮現影像與浮現影片。經過本研究的實驗證實,浮現影像具有成為新一代驗證碼(CAPTCHA)的潛力,亦能成為視覺心裡學中,研究生物視覺判別的生物運動(Biological Motion)理論的重要素材,另外,該類型的圖片及影片亦可成為非相片質感渲染技術中的一種錯覺風格類型。
Emerging image is the visual phenomenon allowing spectators to recognize the objects in a seemingly meaningless image by aggregating and perceiving information that is meaningful as a whole. Emerging video refers to emerging images in a dynamic range. While watching emerging video, the spectators can only distinguish the meaningful object when the clip is playing. Otherwise, what they see would be just some disorganized information. The ability of perceiving comprehensive information in either emerging image or video renders emergence an effective scheme to tell humans apart from machines, which can also been studied and analyzed through visual psychology.
Recently, in the field of Computer Graphics, designing the automation algorithms to reproduce the illusion of art and applying such special algorithms in photography, video and animations become rather prevalent. The thesis provides a relatively systematical and automatic way to generate and synthesize emerging image and video through the user input and the Web crawler. Moreover, as the thesis proved, it is believed that emerging image has the potential of being a promising and pioneering CAPTCHA system. Also, based on the thesis’ suggestion, emerging video can server as one of the important research materials while studying Biological Motion and as an illustration style of non-photorealistic rendering.
[1] Curtis L. Baker and Oliver J. Braudick. The basis of area and dot number effects in random
dot motion perception. Vision Research, pages 1253–1259, 1982.
[2] Niloy J. Mitra, Hung-Kuo Chu, Tong-Yee Lee, Lior Wolf, Hezy Yeshurun, and Daniel
Cohen-Or. Emerging images. ACM Trans. Graph. (Proc. SIGGRAPH Asia), 28(5):163:1–
163:8, 2009.
[3] Craig S. Kaplan and David H. Salesin. Escherization, 2000.
[4] Ran Gal, Olga Sorkine, Tiberiu Popa, Alla Sheffer, Daniel Cohen-or, and Tu Berlin. 3d
collage: Expressive non-realistic modeling. In In Proceedings of NPAR 2007, pages 7–14,
2007.
[5] T. Igarashi, N. Max, F. Sillion, Jong chul Yoon, In kwon Lee, and Henry Kang. A hiddenpicture
puzzles generator, 2008.
[6] Antonio Torralba. Hybrid images. ACM Transactions on Graphics (TOG, 25, 2006.
[7] Ming-Te Chi, Tong-Yee Lee, Yingge Qu, and Tien-Tsin Wong. Self-animating images:
Illusory motion using repeated asymmetric patterns. ACM Transactions on Graphics, 27
(3), 2008.
[8] Niloy J. Mitra and Mark Pauly. Shadow art. ACM Trans. Graph., 28(5):156:1–156:7,
December 2009.
[9] A. M. Treisman and G. Gelade. A feature-integration theory of attention. Cognitive Psychology,
12:97–136, 1980.
[10] Hung-Kuo Chu, Wei-Hsin Hsu, Niloy J. Mitra, Daniel Cohen-Or, Tien-Tsin Wong, and
Tong-Yee Lee. Camouflage images. ACM Trans. Graph. (Proc. SIGGRAPH), 29(4):51:1–
51:8, 2010.
[11] Carsten Rother, Vladimir Kolmogorov, and Andrew Blake. Grabcut -interactive foreground
extraction using iterated graph cuts. ACM TRANS. GRAPH, pages 309–314, 2004.
[12] M. M. Cheng, N. J. Mitra, X. Huang, P. H. S. Torr, and S. M. Hu. Global contrast based
salient region detection. IEEE Transactions on Pattern Analysis and Machine Intelligence,
37(3):569–582, March 2015.
[13] Radhakrishna Achanta, Appu Shaji, Kevin Smith, Aurelien Lucchi, Pascal Fua, and Sabine
Susstrunk. Slic superpixels compared to state-of-the-art superpixel methods. IEEE Trans.
Pattern Anal. Mach. Intell., 34(11):2274–2282, November 2012.
[14] M. Grundmann, V. Kwatra, M. Han, and I. Essa. Efficient hierarchical graph based video
segmentation. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR),
2010.
[15] Jason Chang, Donglai Wei, and John W. Fisher Iii. A video representation using temporal
superpixels, 2013.
[16] Berthold K. P. Horn and Brian G. Schunck. Determining optical flow, 1981.
[17] Anton Andriyenko, Konrad Schindler, and Stefan Roth. Discrete-continuous optimization
for multi-target tracking. In CVPR, 2012.
[18] Jerome Revaud, Zaid Harchaoui, Cordelia Schmid, Jerome Revaud, Zaid Harchaoui, and
Cordelia Schmid. Epicflow: Edge-preserving interpolation of correspondences for optical
flow, 2015.
[19] Liang Lin, Kun Zeng, Han Lv, Yizhou Wang, Yingqing Xu, and Song-Chun Zhu. Painterly
animation using video semantics and feature correspondence. In Proceedings of the 8th International
Symposium on Non-Photorealistic Animation and Rendering, NPAR ’10, New
York, NY, USA, 2010. ACM.
[20] Tinghuai Wang, John Collomosse, David Slatter, Phil Cheatle, and Darryl Greig. Video
stylization for digital ambient displays of home movies. In Proceedings of the 8th International
Symposium on Non-Photorealistic Animation and Rendering, NPAR ’10. ACM,
2010.
[21] Daniel Sýkora, Mirela Ben-Chen, Martin Čadík, Brian Whited, and Maryann Simmons.
Textoons: Practical texture mapping for hand-drawn cartoon animations. In Proceedings
of International Symposium on Non-photorealistic Animation and Rendering, pages 75–83,
2011.
[22] Robert D. Kalnins, Philip L. Davidson, Lee Markosian, and Adam Finkelstein. Coherent
stylized silhouettes. ACM Trans. Graph., July 2003.
[23] Pierre Bénard, Forrester Cole, Aleksey Golovinskiy, and Adam Finkelstein. Self-similar
texture for coherent line stylization. In NPAR 2010: Proceedings of the 8th International
Symposium on Non-photorealistic Animation and Rendering, June 2010.
[24] Pierre Bénard, Jingwan Lu, Forrester Cole, Adam Finkelstein, and Joëlle Thollot. Active
strokes: Coherent line stylization for animated 3d models. In Proceedings of the Symposium
on Non-Photorealistic Animation and Rendering, NPAR ’12. Eurographics Association,
2012.
[25] Nir Ben-Zvi, Jose Bento, Moshe Mahler, Jessica Hodgins, and Ariel Shamir. Line-drawing
video stylization. Computer Graphics Forum, page Article Accepted, 2015.
[26] S. M. Axstis. Phi movement as a subtraction process. Vision Research, 1970.
[27] S. M. Axstis. The perception of apparent motion. Scientific American, 1986.
[28] Sang-Hun Lee and Randolph Blake. Detection of temporal structure depends on spatial
structure. Vision Research, 1999.
[29] GUNNAR JOHANSSON. Visual perception of biological motion and a model for its analysis.
Perception and Psychophysics, 1973.
[30] Vicki Ahlström, Randolph Blake, and Ulf Ahlström. Perception of biological motion. Perception,
1997.
[31] Emily D. Grossman and Randolph Blake. Perception of coherent motion, biological motion
and form-from-motion under dim-light conditions. Vision Research, 1999.
[32] Emily D. Grossman, Lorella Battelli, and Pascual-leone Alvaro. Repetitive tms over posterior
sts disrupts perception of biological motion. Vision Research, 2005.
[33] Jan Eric Kyprianidis and Jürgen Döllner. Image abstraction by structure adaptive filtering.
In Poster at 6th Symposium on Non-Photorealistic Animation and Rendering (NPAR), 2008.
[34] Robert Bridson. Fast poisson disk sampling in arbitrary dimensions. In ACM SIGGRAPH
2007 Sketches, SIGGRAPH ’07. ACM, 2007. ISBN 978-1-4503-4726-6. doi: 10.1145/
1278780.1278807.
[35] Jing Liao, Mark Finch, and Hugues Hoppe. Fast computation of seamless video loops.
ACM Trans. Graph., 34(6):197:1–197:10, October 2015.
[36] Dorita HF Chang and Nikolaus F Troje. Characterizing global and local mechanisms in
biological motion perception. Journal of Vision, 9(5):8–8, 2009.
[37] Benjamin Watson, Alinda Friedman, and Aaron McGaffey. Measuring and predicting visual
fidelity. SIGGRAPH ’01, pages 213–220. ACM, 2001.
[38] Yangqing Jia, Evan Shelhamer, Jeff Donahue, Sergey Karayev, Jonathan Long, Ross Girshick,
Sergio Guadarrama, and Trevor Darrell. Caffe: Convolutional architecture for fast
feature embedding. arXiv preprint arXiv:1408.5093, 2014.
[39] Chao Ma, Jia-Bin Huang, Xiaokang Yang, and Ming-Hsuan Yang. Hierarchical convolutional
features for visual tracking. In Proceedings of the IEEE International Conference
on Computer Vision), 2015.