参考文献/References:
[1] 柴刚. 解析国画作品中的意象世界[J]. 科技资讯,2009(20):197-198.
[2]雷莹. 浅析“情感”在国画创作中的艺术精神[J]. 大众文艺(学术版),2016(8):91.
[3]WRIGHT J,MA Y,MAIRAL J,et al. Sparse representation for computer vision and pattern recognition[J]. Proceedings of the IEEE,2010,98(6):1 031-1 044.
[4]SUN Y,FISHER R. Object-based visual attention for computer vision[J]. Artificial intelligence,2003,146(1):77-123.
[5]DATTA R. Semantics and aesthetics inference for image search:statistical leaning approaches[D]. University Park:The Pernnsylvania State University,2009.
[6]GUDIVADA V N. A geometry-based representation for efficient and effective retrieval of images by spatial similarity[J]. IEEE transactions of knowledge and data engineering,1998,10(3):504-512.
[7]HANJALIC,ALAN. Video and image retrieval beyond the cognitive level: the needs and possibilities[C]//Storage and Retrieval for Media Databases,San Jose,2001. Bellingham:Optical Engineering,2001:130-140.
[8]LECUN Y,BENGIO Y,HINTON G. Deep learning[J]. Nature,2015,521(7 553):436-444.
[9]HEMPHILL M. A note on adults’ color-emotion associations[J]. Journal of genetic psychology,1996,157(3):275-280.
[10]ELLIOT A J,MAIER M A,MOLLER A C,et al. Color and psychological functioning:the effect of red on performance attainment[J]. Journal of experimental psychology general,2007,136(1):154.
[11]陈俊杰. 图像情感语义分析技术[M]. 北京:电子工业出版社,2011:13-15.
[12]余英林,田菁,蔡志峰. 图像视觉感知信息的初步研究[J]. 电子学报,2001,29(10):1 373-1 375.
[13]WANG T,WU D J,COATES A,et al. End-to-end text recognition with convolutional neural networks[C]//International Conference on Pattern Recognition. Tsukuba Japan,2012. New York:IEEE,2012:3 304-3 308.
[14]ABDEL H O,MOHAMED A R,JIANG H,et al. Convolutional neural networks for speech recognition[J]. IEEE/ACM transactions on audio speech and language processing,2014,22(10):1 533-1 545.
[15]CECOTTI H,GR?SER A. Convolutional neural networks for P300 detection with application to brain-computer interfaces[J]. IEEE transactions on pattern analysis and machine intelligence,2011,33(3):433-45.
[16]NEBAUER C. Evaluation of convolutional neural networks for visual recognition[J]. IEEE transactions on neural networks,1998,9(4):685.
[17]BAI S. Growing random forest on deep convolutional neural networks for scene categorization[M]. London:Pergamon Press,2017:50-70.
[18]SUN M,SONG Z,JIANG X,et al. Learning pooling for convolutional neural network[J]. Neurocomputing,2017,224:96-104.
[19]YU S,JIA S,XU C. Convolutional neural networks for hyperspectral image classification[J]. Neurocomputing,2017,219:88-98.
[20]KUMAR A,KIM J,LYNDON D,et al. An ensemble of fine-tuned convolutional neural networks for medical image classification[J]. IEEE journal of biomedical and health informatics,2017,21(99):1.
[21]乐毅,王斌. 深度学习-Caffe之经典模型详解与实战[M]. 北京:电子工业出版社,2016:107-111.
[22]ZENG R,WU J,SHAO Z,et al. Quaternion softmax classifier[J]. Electronics letters,2014,50(25):1 929-1 931.
[23]CHEN X,YE Q,ZOU J,et al. Visual trajectory analysis via replicated softmax-based models[J]. Signal,image and video processing,2014,8(1):183-190.
[24]LEE J Y,KIM K. A feature-based approach to extracting machining features[J]. Computer-aided design,1998,30(13):1 019-1 035.
[25]CANDèS E J,LI X,MA Y,et al. Robust principal component analysis?[J]. Journal of the Acm,2009,58(3):11.
[26]XIE X,ZHENG W S,LAI J,et al. Normalization of face illumination based on large- and small-scale features[J]. IEEE transactions on image processing a publication of the IEEE signal processing society,2011,20(7):1 807-1 821.
[27]QI Z,TIAN Y,SHI Y. Robust twin support vector machine for pattern classification[J]. Pattern recognition,2013,46(1):305-316.
[28]QI Z,TIAN Y,SHI Y. Structural twin support vector machine for classification[J]. Knowledge-based systems,2013,43(2):74-81.
[29]SRIVASTAVA N,HINTON G,KRIZHEVSKY A,et al. Dropout:a simple way to prevent neural networks from overfitting[J]. Journal of machine learning research,2014,15(1):1 929-1 958.
[30]JR S E. Detecting and evaluating the impact of multidimensionality using item fit statistics and principal component analysis of residuals[J]. Journal of applied measurement,2002,3(2):205-231.