|
中文文獻 1.ArtLinks(2017)。馬塞爾·杜尚,挑戰藝術底線的先鋒藝術家。取自: https://kknews.cc/culture/5a3evnl.html 2.人工智慧(2018)。一幅AI畫像憑什麼賣300萬?取自:https://www.itread01.com/lyxcx.html 3.孔成(2013)。攝影技術對繪畫的影響。長江大學學報:社會科學版,36(6),188-189。 4.瓦爾特·本雅明(2002)。機械複製時代的藝術作品(王才勇譯)。北京市:中國城市出版社。 5.白茹意、郭小英、賈春花、耿海軍(2019)。繪畫圖像美學研究方法綜述。中國圖象圖形學報,24(11),1860-1881。 6.朱光潛(2006)。談美。台北市:五南圖書出版股份有限公司。 7.李博、郭琛、任慧(2018)。基於加權 K 近鄰演算法的抽象畫圖像情感分佈預測。中國傳媒大學學報:自然科學版,25(1),36-40。 8.李開複(2021)。AI賦能時代的創業。創新工廠講AI課,北京市:電子工業出版社。 9.李素超(2017)。當科技有了創造的溫度-- A.I.與藝術結合的機構實踐。取自:https://artouch.com/view/content-4066.html 10.李欣(2007)。數字藝術探議。天津美術學院學報,13(2),23。 11.范士誠,、莊明振、許峻誠(2013)。正面感性評價的造形構成與眼球凝視位置研究。設計學報,18(3),63-84。 12.房旭(2015)。論攝影中的西方繪畫藝術風格。藝術教育,(1),184。 13.林榮泰(2003)。人與機器的對話:科技始終來自於人性?取自:https://scitechvista.nat.gov.tw/c/sWUc.htm 14.林榮泰、李仙美(2015)。詩情畫意 – 仙雲之美習作經驗分享。新北市:國立臺灣藝術大學。 15.林雅惠(2009)。解構攝影的誕生對藝術及歷史的意義,文化研究@ 嶺南,16(1),4。 16.吳士新(2019)。科技發展,助力藝術創新。取自: https://baijiahao.baidu.com/s?id=1637818175322636157&wfr=spider&for=pc 17.洪銘駿(2016)。智慧型視覺回授控制機器人應用於彩色藝術繪畫(未出版之碩士論文)。國立臺灣大學電機工程學研究所,臺北市,臺灣。取自:https://hdl.handle.net/11296/9386wf 18.品途商業評論(2018)。微軟智慧機器人「小冰」展現填詞新技能。取自:https://kknews.cc/tech/on6bglq.html 19.貢布里希(1987),範景中、林夕、李本正 譯。藝術與錯覺。杭州市:浙江攝影出版社。 20.唐大崙、張文瑜(2007)。利用眼動追蹤法探索傳播研究。中華傳播學刊,(12),165-211。 21.唐稔為、劉啟和、譚浩(2021)。神經風格遷移模型綜述。電腦工程與應用,57(19),32-43。 22.陳淮源、張廣馳、陳高、周清峰.(2021)。基于深度學習的圖像風格遷移研究進展。計算机工程与應用,57(11),37-45。 23.陳淑環、韋玉科、徐樂、董曉華、溫坤哲(2019)。基于深度學習的圖像風格遷移研究總述。計算机應用研究,36(08),2250-2255。 24.許峻誠、陳韋呈(2017)。利用眼動追蹤技術探討版式設計與美感情緒之關聯性。藝術教育研究,(33),33-67。 25.許素朱(2019a)。人工智慧與藝術的競合創作。藝術家,535,240-245。 26.許素朱(2019b)。競爭或合作?人工智慧在人文社會的應用與影響探究。取自:https://nthuai2019.wixsite.com/ai-hss/main 27.量子位(2017)。AI通過了藝術創作圖靈測試,機器發展出了自己的美感。取自 https://kknews.cc/news/b48zovj.html 28.曾靖越(2018)。無縫空間的沈浸感:虛擬實境。國教新知,65(3),105-120。 29.葉衛平(2018)。藝術的內在發生。北京市:清華大學出版社。 30.劉潤坤(2017)。人工智慧取代藝術家?-從本體論視角看人工智慧藝術創作。民族藝術研究,30(2),71-76。 31.劉育成(2019)。機器學習(人工智慧)如何可能重新定義「創作」:「模糊性」作為創造性的來源之一。取自:https://www.digiarts.org.tw/DigiArts/DataBasePage/4_140864213871031/Chi 32.魯越、郭超、林懿倫、卓凡、王飛躍(2020)。繪畫藝術圖像的計算美學: 研究前沿與展望。自動化學報,46(11),2239-2259。 33.謝其昌(2015)。論油畫的材料發展談繪畫技法的形式與演變。美學與視覺藝術學刊,(7),35-50。
外文文獻 1.Alameda-Pineda, X., Ricci, E., Yan, Y., & Sebe, N. (2016). Recognizing emotions from abstract paintings using non-linear matrix completion. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (pp. 5240-5248). 2.Amirshahi, S. A., Hayn-Leichsenring, G. U., Denzler, J., & Redies, C. (2016). Color: A crucial factor for aesthetic quality assessment in a subjective dataset of paintings. arXiv preprint arXiv:1609.05583. 3.Arnheim, R. (1965). Art and visual perception: A psychology of the creative eye. Berkeley, CA: University of California Press. 4.Audry, S. (2021). Art in the Age of Machine Learning. Cambridge, MA: MIT Press. 5.Bar, Y., Levy, N., & Wolf, L. (2015). Classification of Artistic Styles Using Binarized Features Derived from a Deep Neural Network. In L. Agapito, M. M. Bronstein & C. Rother (Eds.). Computer Vision - ECCV 2014 Workshops. ECCV 2014. Lecture Notes in Computer Science, vol 8925 (pp. 71-84). Springer, Cham. https://doi.org/10.1007/978-3-319-16178-5_5 6.Barthes, R. (1968). Elements of semiology (Vol. 4). New York, NY: Hill & Wang. 7.Beardsley, M. C. (1981). Aesthetics, problems in the philosophy of criticism. Indianapolis, IN: Hackett Publishing. 8.Beatty, E. L., & Ball, L. J. (2011). Investigating exceptional poets to inform an understanding of the relationship between poetry and design. In Procedings of the Second Conference on Creativity and Innovation in Design - DESIRE ’11, 157-165. doi:10.1145/2079216.2079238 9.Beckett, W. (1994). The story of painting. London, UK: DK. 10.Bengio, Y. (2009). Learning deep architectures for AI. Found Trends Mach Learn, 2 (1), 1-127. doi:10.1561/2200000006 11.Berlyne, D. (1958). The influence of the albedo and complexity of stimuli on visual fixation in the human infant. British Journal of Psychology. 12.Brachmann, A., & Redies, C. (2017). Computational and experimental approaches to visual aesthetics. Frontiers in computational neuroscience, 11, 102. 13.CANGÖZ, B., OKTAY, B., KAYIRAN, N. R., & KARAMANOĞLU, S. E. (2021). Art Education and Expertise An Eye Tracking Study. Hacettepe Üniversitesi Edebiyat Fakültesi Dergisi, 38 (1), 259-271. 14.Chan, C.-S. (2015). Style and creativity in design (Vol. 17). Berlin, German: Springer. 15.Chatterjee, A. (2004). Prospects for a cognitive neuroscience of visual aesthetics. Bulletin of Psychology and the Arts, 4, 55–59. 16.Chen, C. (2020). Structure-emphasized Multimodal Style Transfer (Unpublished Master’s Thesis). Tokyo Institute of Technology, Tokyo, Japan. Retrieved from https://drive.google.com/file/d/1Y77Zy25gtuapEQCEfysoRmhiuEnc48Op/view 17.Chen, H., Zhao, L., Wang, Z., Zhang, H., Zuo, Z., Li, A., ... & Lu, D. (2021). Dualast: Dual style-learning networks for artistic style transfer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 872-881). 18.Chen, T. Q., & Schmidt, M. (2016). Fast patch-based style transfer of arbitrary style. arXiv preprint arXiv:1612.04337. 19.Chilvers, I., Zaczek, I., Welton, J., Bugler, C., & Mack, L. (2013). Art That Changed the World. London, UK: DK. 20.Condorovici, R. G., Florea, C., Vrânceanu, R., & Vertan, C. (2013). Perceptually-inspired artistic genre identification system in digitized painting collections. Image Analysis, 687-696. doi:10.1007/978-3-642-38886-6_64 21.Condorovici, R. G., Florea, C., Vrânceanu, R., & Vertan, C. (2013). Perceptually-inspired artistic genre identification system in digitized painting collections. In Scandinavian Conference on Image Analysis (pp. 687-696). Springer, Berlin, Heidelberg. 22.Castellotti, S., Scipioni, L., Mastandrea, S., & Del Viva, M. M. (2021). Pupil responses to implied motion in figurative and abstract paintings. Plos one, 16 (10), e0258490. 23.Dutton, D. (2013). Aesthetic universals. In The Routledge companion to aesthetics (pp. 289-299). Routledge. 24.Efros, A. A., & Freeman, W. T. (2001). Image quilting for texture synthesis and transfer. In L. Pocock (Ed.). Proceedings of the 28th annual conference on Computer graphics and interactive techniques (pp. 341-346). New York, NY: ACM. 25.Fan, Z. B., Li, Y. N., Yu, J., & Zhang, K. (2017). Visual complexity of chinese ink paintings. In Proceedings of the ACM Symposium on Applied Perception (pp. 1-8). 26.Fernie, E. (1995). Art history and its methods: A critical anthology. London: Phaidon. 27.Fiske, J. (2010). Introduction to communication studies. London, UK: Routledge. 28.Furst, C. J. (1971). Automatizing of visual attention. Perception & Psychophysics, 10 (2), 65-70. 29.Ganczarek, J., Pietras, K., & Rosiek, R. (2020). Perceived cognitive challenge predicts eye movements while viewing contemporary paintings. PsyCh Journal, 9 (4), 490-506. 30.Gao, Y., Wu, J., Lee, S., & Lin, R. (2019). Communication between artist and audience: A case study of creation journey. Cross-Cultural Design. Culture and Society, 33-44. doi:10.1007/978-3-030-22580-3_3 31.Gao, Y. J., Chen, L. Y., Lee, S., Lin, R., & Jin, Y. (2017, July). A study of communication in turning “poetry” into “painting”. In International Conference on Cross-Cultural Design (pp. 37-48). Springer, Cham. 32.Gardner, H. (1970). Children’s sensitivity to painting styles. Child Development, 41 (3), 813–821. doi:10.2307/1127226 33.Gatys, L. A., Ecker, A. S., & Bethge, M. (2015). A neural algorithm of artistic style. Retrieved from https://arxiv.org/abs/1508.06576v2. 34.Gatys, L. A., Ecker, A. S., Bethge, M., Hertzmann, A., & Shechtman, E. (2017). Controlling perceptual factors in neural style transfer. In 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (pp. 3985-3993), Honolulu, HI, US. doi:10.1109/cvpr.2017.397 35.Getlein, M., & Gilbert, R. (2008). Living with art. New York, NY: McGraw-Hill. 36.Gunsel, B., Sariel, S., & Icoglu, O. (2005). Content-based access to art paintings. In IEEE International Conference on Image Processing 2005 (Vol. 2, pp. II-558). IEEE. 37.Haeberli, P. (1990). Paint by numbers: Abstract image representations. In Proceedings of the 17th Annual Conference on Computer Graphics and Interactive Techniques (pp. 207-214). New York, NY: ACM Press. 38.Hageback, N., & Hedblom, D. (2021). AI for Arts. CRC Press. 39.Hagtvedt, H., Patrick, V. M., & Hagtvedt, R. (2008). The perception and evaluation of visual art. Empirical studies of the arts, 26 (2), 197-218. 40.Hertzmann, A. (2003). A survey of stroke-based rendering. IEEE Computer Graphics and Applications, 4. 41.Hertzmann, A. (2010, June). Non-photorealistic rendering and the science of art. In Proceedings of the 8th International Symposium on Non-Photorealistic Animation and Rendering (pp. 147-157). 42.Hertzmann, A. (2018). Can computers create art? Arts, 7 (2), 18. 43.Hinton, G. E., & Salakhutdinov, R. R. (2006). Reducing the dimensionality of data with neural networks. Science, 313 (5786), 504-507. doi:10.1126/science.1127647 44.Huang, X., & Belongie, S. (2017). Arbitrary style transfer in real-time with adaptive instance normalization. Retrieved from https://arxiv.org/abs/1703.06868 45.Jakobson, R., Pomorska, K., & Rudy, S. (1987). Language in literature. Cambridge, MA: Harvard University Press. 46.Jing, Y., Yang, Y., Feng, Z., Ye, J., Yu, Y., & Song, M. (2018). Neural Style Transfer: A Review. IEEE transactions on visualization and computer graphics. Retrieved from https://arxiv.org/abs/1705.04058v7 47.Johnson, J., Alahi, A., & Li, F. (2016). Perceptual losses for real-time style transfer and super-resolution. Retrieved from https://arxiv.org/abs/1603.08155 48.Joshi, B., Stewart, K., & Shapiro, D. (2017). Bringing impressionism to life with neural style transfer in come swim. Retrieved from https://arxiv.org/abs/1701.04928 49.Kang, D., Shim, H., & Yoon, K. (2018). A method for extracting emotion using colors comprise the painting image. Multimedia Tools and Applications, 77 (4), 4985-5002. 50.Karayev, S., Trentacoste, M., Han, H., Agarwala, A., Darrell, T., Hertzmann, A., & Winnemoeller, H. (2013). Recognizing image style. Retrieved from https://arxiv.org/abs/1311.3715 51.Karnewar, A., Kanawaday, A., Sawant, C., & Gupta, Y. (2017, June). Classification of abstract images using machine learning. In Proceedings of the 2017 International Conference on Deep Learning Technologies (pp. 36-40). 52.Kirsh, A., & Levenson, R. S. (2002). Seeing through paintings: Physical examination in art historical studies (Vol. 1). New Haven, CT: Yale University Press. 53.Kotovenko, D., Sanakoyeu, A., Ma, P., Lang, S., & Ommer, B. (2019). A content transformation block for image style transfer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 10032-10041). 54.Krejtz, K., Duchowski, A., Szmidt, T., Krejtz, I., González Perilli, F., Pires, A., Vilaro, A., & Villalobos, N. (2015). Gaze transition entropy. ACM Transactions on Applied Perception (TAP), 13 (1), 1-20. 55.Kurzweil, R. (2005). The singularity is near: When humans transcend biology. London, UK: Penguin. 56.Lakoff, G., & Johnson, M. (2003). Metaphors we live by. London, UK: University of Chicago Press. 57.Lawrie, E. (2019). Could a computer ever create better art than a human? Retrieved from https://www.bbc.com/news/business-47700701 (accessed on 1 April2022) 58.Lee, S. G., & Cha, E. Y. (2016). Style classification and visualization of art painting’s genre using self-organizing maps. Human-centric Computing and Information Sciences, 6 (1), 7. doi:10.1186/s13673-016-0063-4 59.Lewis, M. (2008). Evolutionary visual art and design. In J. Romero & P. Machado (Ed.). The art of artificial evolution (pp. 3-37). Berlin, German: Springer. 60.Li, C., & Wand, M. (2016). Combining markov random fields and convolutional neural networks for image synthesis. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 2479-2486). 61.Li, Y., Fang, C., Yang, J., Wang, Z., Lu, X., & Yang, M. H. (2017a). Universal style transfer via feature transforms. Retrieved from https://arxiv.org/abs/1705.08086 62.Li, Y., Fang, C., Yang, J., Wang, Z., Lu, X., & Yang, M. H. (2017). Universal style transfer via feature transforms. Advances in neural information processing systems, 30. 63.Lin, C.-L., Chen, J.-L., Chen, S.-J., & Lin, R. (2015). The cognition of turning poetry into painting. Journal of US-China Education Review B, 5 (8), 471-487. 64.Lin, R., Qian, F., Wu, J., Fang, W. T., & Jin, Y. (2017, July). A pilot study of communication matrix for evaluating artworks. In International Conference on Cross-Cultural Design (pp. 356-368). Springer, Cham. 65.Liu, S., Yang, J., Agaian, S. S., & Yuan, C. (2021). Novel features for art movement classification of portrait paintings. Image and Vision Computing, 108, 104121. 66.Louie, R., Coenen, A., Huang, C. Z., Terry, M., & Cai, C. J. (2020, April). Novice-AI music co-creation via AI-steering tools for deep generative models. In Proceedings of the 2020 CHI conference on human factors in computing systems (pp. 1-13). 67.Mackworth, N. H., & Morandi, A. J. (1967). The gaze selects informative details within pictures. Perception & psychophysics, 2 (11), 547-552. 68.Mazzone, M., & Elgammal, A. (2019). Art, creativity, and the potential of artificial intelligence. Arts, 8 (1), 26. doi:10.3390/arts8010026 69.McGuinness, D. E. (2016). Painting: Materials, Techniques, Styles, and Practice. New York, NY: Rosen Education Service. 70.Mishory, A. (2000). Art history: an introduction. Ra'anana, Israel: Open University of Israel. 71.Park, S. A., Yun, K., & Jeong, J. (2015). Reappraising abstract paintings after exposure to background information. PLoS One, 10 (5), e0124159. 72.Pasupa, K., Chatkamjuncharoen, P., Wuttilertdeshar, C., & Sugimoto, M. (2015, November). Using image features and eye tracking device to predict human emotions towards abstract images. In Image and Video Technology (pp. 419-430). Springer, Cham. 73.Sanakoyeu, A., Kotovenko, D., Lang, S., & Ommer, B. (2018). A style-aware content loss for real-time hd style transfer. In proceedings of the European conference on computer vision (ECCV) (pp. 698-714). 74.Sandoval, C., Pirogova, E., & Lech, M. (2019). Two-stage deep learning approach to the classification of fine-art paintings. IEEE Access, 7, 41770-41781. 75.Sartori, A., Şenyazar, B., Salah, A. A. A., Salah, A. A., & Sebe, N. (2015, September). Emotions in abstract art: does texture matter?. In International Conference on Image Analysis and Processing (pp. 671-682). Springer, Cham. 76.Shamir, L., Macura, T., Orlov, N., Eckley, D. M., & Goldberg, I. G. (2010). Impressionism, expressionism, surrealism: automated recognition of painters and schools of art. ACM Transactions on Applied Perception (TAP), 7 (2), 1-17. doi:10.1145/1670671.1670672 77.Sheng, L., Lin, Z., Shao, J., & Wang, X. (2018). Avatar-net: Multi-scale zero-shot style transfer by feature decoration. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (pp. 8242-8250). 78.Silverman, K. (1984). The subject of semiotics. New York, NY: Oxford University Press. 79.Sims, K. (1991). Artificial evolution for computer graphics. ACM SIGGRAPH Computer Graphics, 25 (4), 319-328. doi:10.1145/127719.122752 80.Špakov, O., & Miniotas, D. (2007). Visualization of eye gaze data using heat maps. Elektronika ir elektrotechnika, 74 (2), 55-58. 81.Spehr, M., Wallraven, C., & Fleming, R. W. (2009). Image statistics for clustering paintings according to their visual appearance. Computational Aesthetics 2009: Eurographics Workshop on Computational Aesthetics in Graphics, Visualization and Imaging (pp. 57-64). Eurographics. 82.Spence, C., Youssef, J., Michel, C., & Woods, A. (2019). Assessing the aesthetic oblique effect in painting and plating. International Journal of Gastronomy and Food Science, 17, 100168. 83.Steenberg, E. (2007). Visual aesthetic experience. The Journal of Aesthetic Education, 41 (2), 89-94. doi:10.1353/jae.2007.0018 84.Stork, D. G., & Coddington, J. (2008). Computer image analysis in the study of art. SPIE/IS&T, Bellingham, WA, 14-18. 85.Stork, D. G., Coddington, J., & Bentkowska-Kafel, A. (Eds.). (2010). Computer vision and image analysis of art. SPIE (Vol. 7531). 86.Tan, W. R., Chan, C. S., Aguirre, H. E., & Tanaka, K. (2016, September). Ceci n'est pas une pipe: A deep convolutional network for fine-art paintings classification. In 2016 IEEE international conference on image processing (ICIP) (pp. 3703-3707). IEEE. 87.Taylor, J. E. T., Witt, J. K., & Grimaldi, P. J. (2012). Uncovering the connection between artist and audience: Viewing painted brushstrokes evokes corresponding action representations in the observer. Cognition, 125 (1), 26-36. 88.Ulyanov, D., Lebedev, V., Vedaldi, A., & Lempitsky, V. S. (2016, June). Texture networks: Feed-forward synthesis of textures and stylized images. ICML,1 (2), 4. 89.Wang, F. Y., Zhang, J. J., Zheng, X., Wang, X., Yuan, Y., Dai, X., & Yang, L. (2016). Where does AlphaGo go: from church-turing thesis to AlphaGo thesis and beyond. IEEE/CAA Journal of Automatica Sinica, 3 (2), 113-120. doi:10.1109/jas.2016.7471613 90.Wang, X., Lyu, Y., Huang, J., Wang, Z., & Qin, J. (2021). Interactive Artistic Multi-style Transfer. International Journal of Computational Intelligence Systems, 14 (1), 1-13. 91.Whitfield, S. (1994). Fauvism. In N. Stangos (Ed.), Concepts of Modern Art: From Fauvism to Postmodernism, 11-29. London, UK: Thames and Hudson. 92.Wolfram Research, W. (2012). ImageDistance, language function https://reference.wolfram.com/language/ref/ImageDistance.html (updated 2016) 93.Yanulevskaya, V., Uijlings, J., Bruni, E., Sartori, A., Zamboni, E., Bacci, F., ... & Sebe, N. (2012, October). In the eye of the beholder: employing statistical analysis and eye tracking for analyzing abstract paintings. In Proceedings of the 20th ACM international conference on multimedia (pp. 349-358). 94.Zhang, H., & Dana, K. (2018). Multi-style generative network for real-time transfer. In Proceedings of the European Conference on Computer Vision (ECCV)Workshops. 95.Zhang, R., Isola, P., Efros, A. A., Shechtman, E., & Wang, O. (2018). The unreasonable effectiveness of deep features as a perceptual metric. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 586-595). 96.Zhang, Y., Fang, C., Wang, Y., Wang, Z., Lin, Z., Fu, Y., & Yang, J. (2019). Multimodal style transfer via graph cuts. Retrieved from https://arxiv.org/abs/1904.04443 97.Zhong, S.-h., Huang, X., & Xiao, Z. (2020). Fine-art painting classification via two-channel dual path networks. International Journal of Machine Learning and Cybernetics, 11 (1), 137-152. 98.Zhu, J. Y., Park, T., Isola, P., & Efros, A. A. (2017). Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE international conference on computer vision (pp. 2223-2232). 99.Zujovic, J., Gandy, L., Friedman, S., Pardo, B., & Pappas, T. N. (2009). Classifying paintings by artistic genre: An analysis of features & classifiers. 2009 IEEE International Workshop on Multimedia Signal Processing (pp. 1-5). doi:10.1109/mmsp.2009.5293271
|