导师简介-钟微

作者:发布时间:2026-04-21

姓名:钟微

职称 硕(博)导:研究员 硕导

学硕招生专业 方向:信息与通信工程 / 智能视频处理

专硕招生专业 / 方向:人工智能  / 智能视听技术

主要研究领域:媒体认知计算

电子邮箱Emailwzhong@cuc.edu.cn

个人简介

钟微,中国传媒大学媒介音视频教育部重点实验室研究员,硕士生导师。2004年获得西安电子科技大学电子信息工程专业学士学位,2010年获得西安电子科技大学电路与系统专业博士学位。20117月至今就职于中国传媒大学媒介音视频教育部重点实验室,20247月被评为研究员,研究领域为媒体认知计算、脑电情感分析与计算。主持或主要参与国家自然科学基金重点/面上/青年项目、国家重点研发计划课题/子课题、以及省部级、横向等科研项目10余项;在IEEE Trans. MultimediaIEEE Trans. Signal ProcessingPattern RecognitionDisplaysNeurocomputingChinese Journal of Electronics等国内外期刊/会议上发表SCIEI检索论文40余篇,荣获IFTC 2018最佳论文奖、2019年广播影视科技优秀论文一等奖、ICME 2023-AIART Workshop最佳论文奖以及2025年第四届高新视频创新应用大赛二等奖。

承担的主要科研项目

[1] 国家自然科学基金面上项目,“情智耦合关系的感知与建模”,项目编号:6227145520231-202612月,在研,参与

[2] 国家重点研发计划项目子课题,“基于动态光场的6DOF全景视频制作技术研究”,项目编号:2021YFF0900501-1202112-202411月,已结项,主持

[3] 广东南方新媒体股份有限公司横向项目,“元宇宙在IPTV/OTT的行业发展与应用规划研究”,项目编号:HG2300520231-20252月,已结项,参与

[4] 国家自然科学基金重点项目,“历史影音资料音频修复方法与关键技术研究”,项目编号:6163101620171-202112月,已结项,参与

[5] 国家自然科学基金青年项目,“基于用户行为与媒体情感分析的存储资源贝叶斯分配策略研究”,项目编号:6180144020191-202112月,已结项,参与

[6] 中国电科装备预先研究项目,“视觉智能识别技术”,项目编号:6141B082311030520191-202012月,已结项,参与

[7] 国家自然科学基金面上项目,“基于广义建模理论的多原子库图像编码方法研究”,项目编号:6137119120141-201712月,已结项,参与

[8] 国家自然科学基金青年科学基金项目,“基于调制技术的多通道线性相位滤波器组构造及其在图像纹理分析中的应用”,项目编号:6110116620121-201412月,已结项,主持

[9] 国家自然科学基金重点项目,“基于统计模型与运动基的物体解析与视频重构方法”,项目编号:6083200420091-201212月,已结项,参与

[10] 国家科技支撑计划子课题,“立体电视技术体系架构及内容直播技术与应用研究”,项目编号:2012BAH39F0020131-201512月,已结项,参与

[11] 国家新闻出版广电总局科研项目,“电影虚拟化制作虚拟化资产管理研究及应用示范(一期)”,项目编号:2015-5320156-201512月,已结项,参与

[12] 校级工科规划项目,“视频情感内容分析研究”,项目编号:3132018XNG182420185-20205月,已结项,主持

[13] 中国传媒大学“优秀中青年教师培养工程”,项目编号:YXJS20153220153-20183月,已结项,主持

代表性学术成果

[1] Shuzhan Hu, Mingyu Li, Yang Liu, Weiwei Jiang, Bingrui Geng, Wei Zhong, Long Ye, “Audio-visual perceptual quality measurement via multi-perspective spatio-temporal EEG analysis,” Pattern Recognition, vol. 176, pp. 113156, Jan. 2026.

[2] Yefei Huang, Wei Zhong*, Fei Hu, Long Ye and Qin Zhang, “RPGCN-GDA: regionally progressive graph convolutional network with gender-sensitive domain adaptation for EEG emotion recognition,” IEEE International Conference on Multimedia and Expo AIART Workshop, Nantes, France, pp. 1-6, 2025.

[3] Mingyu Li, Shuzhan Hu, Yang Liu, Danjing Liu, Wei Zhong and Long Ye, “Brain-inspired audio quality assessment based on audio-EEG feature fusion,” The 22nd International Forum on Digital Multimedia Communication, Taiyuan, China, Nov. 20-22, 2025.

[4] Fei Hu, Yibo Ma, Wei Zhong*, Long Ye, Xinyan Yang, Li Fang and Qin Zhang, “A dataset and benchmark for 3D scene plausibility assessment,” IEEE Transactions on Multimedia, pp. 1-14, Jan. 2024.

[5] Ran Shuang, Zhong Wei*, Ma Lin, Duan Danting., Ye Long, Zhang Qin, “Mind to music: an EEG signal-driven real-time emotion music generation system,” International Journal of Intelligent Systems, Dec. 2024.

[6] Ming Gong, Wei Zhong*, Long Ye and Qin Zhang, “MISNet: multi-source information-shared EEG emotion recognition network with two-stream structure,” Frontiers in Neuroscience, vol. 18, 2024.

[7] Ran Shuang, Zhong Wei*, Duan Danting, Hu Fei, Ye Long, Zhang Qin, “EEG emotion recognition supervised by temporal features of video stimuli,” The 46th Engineering in Medicine and Biology Conference, Orlando, FL, USA, July 15-19, 2024.

[8] Ran Shuang, Zhong Wei*, Hu Fei, Ye Long, Zhang Qin, “Coarse-to-fine domain adaptation for cross-subject EEG emotion recognition with contrastive learning,” The 7th Chinese Conference on Pattern Recognition and Computer Vision, pp. 406-419, Urumqi, China, October 18-20, 2024.

[9] Ran Shuang, Zhong Wei*, Duan Danting, Ye Long, and Zhang Qin, “SSTM-IS: simplified STM method based on instance selection for real-time EEG emotion recognition,” Frontiers in Human Neuroscience, vol. 17, pp. 1132254, June 2023.

[10] Duan Danting, Zhong Wei, Ran Shuang, Ye Long, Zhang Qin, “A standardized database of Chinese emotional short videos based on age and gender differences,” PloS One, vol. 18, no. 3, pp. e0283573, 2023.

[11] Xiang Xu, Wei Zhong*, Yi Zou, Long Ye and Qin Zhang, “Differential melody generation based on time series prediction,” IEEE International Conference on Multimedia and Expo Workshops, Brisbane, Australia, July 10-14, 2023.

[12] Xin Ma, Wei Zhong*, Long Ye, and Qin Zhang, “Visually aligned sound generation via sound-producing motion parsing,” Neurocomputing, vol. 492, pp. 1-15, 2022.

[13] Lin Ma, Wei Zhong, Xin Ma, Long Ye, Qin Zhang, “Learning to generate emotional music correlated with music structure features,” Cognitive Computation and Systems, vol. 4, 2022.

[14] Wang Hongfei, Wei Zhong*, Lin Ma, Long Ye, Qin Zhang, “Emotional quality evaluation for generated music based on emotion recognition model,” IEEE International Conference on Multimedia and Expo Workshops, Taipei, Taiwan, July 18-22, 2022.

[15] Gong Ming, Zhong Wei*, Hu Jiayu, Ye Long, Zhang Qin, “Multi-source information-shared domain adaptation for EEG emotion recognition,” The 5th Chinese Conference on Pattern Recognition and Computer Vision, Part II, pp. 441-453, Shenzhen, China, November 4-7, 2022.

[16] Qinxin Lin, Zhong Wei, Long Ye, and Qin Zhang, “Image harmonization based on the semantic information of foreground human,” Displays, pp. 102132, 2021.

[17] Z. Zheng, W. Zhong, L. Ye, L. Fang, Q. Zhang, “Violent scene detection of film videos based on multi-task learning of temporal-spatial features,” IEEE Workshop on Artificial Intelligence for Art Creation, Tokyo, Japan, Sept. 2021.

[18] Li Fang, Wei Zhong, Long Ye, Ran Li and Qin Zhang, “Light field reconstruction with a hybrid sparse regularization-pseudo 4DCNN framework,” IEEE Access, vol. 8, pp. 171009-171020, 2020.

[19] Hu Fei, Long Ye, Wei Zhong, Li Fang, and Qin Zhang, “Deep auxiliary learning for point cloud generation,” IEEE Access, vol. 8, pp. 18538-18545, 2020.

[20] Xiaona Guo, Wei Zhong*, Long Ye, Li Fang, Yan Heng, Qin Zhang, “Global affective video content regression based on complementary audio-visual features,” The 26th International Conference on Multimedia Modeling, Daejeon, Korea, Jan. 5-8, 2020.

[21] Mingda Zhang, Wei Zhong, Long Ye, Li Fang and Qin Zhang, “Synchronous prediction of continuous affective video content based on multi-task Learning,” The 17th International Forum of Digital TV and Multimedia Communication, Shanghai, China, Dec. 3-4, 2020.

[22] Fei Hu, Long Ye, Wei Zhong, Li Fang, Yun Tie, Qin Zhang, “Semantic based autoencoder attention 3D reconstruction network,” Graphical Models, vol. 106, 2019.

[23] Kaiyang Xia, Wei Zhong, Long Ye, Li Fang, Qin Zhang and Yan Heng, “Fine-grained footstep image classification,” The 12th International Congress on Image and Signal Processing, Huaqiao, Suzhou, China, Oct. 19-21, 2019.

[24] Xiaona Guo, Wei Zhong, Long Ye, Li Fang and Qin Zhang, “Affective video content analysis based on two compact audio-visual features,” The 16th International Forum of Digital TV and Multimedia Communication, Shanghai, China, Sep. 19-20, 2019.

[25] Li Fang, Long Ye, Yun Tie, Wei Zhong, Qin Zhang, “Design of linear-phase nonsubsampled nonuniform directional filter bank with arbitrary directional partitioning,” Journal of Visual Communication and Image Representation, vol. 51, pp. 23-28, Feb. 2018.

[26] Wei Zhong, Kaiyang Xia, Li Fang, Long Ye and Qin Zhang, “Design of 2D checkboard nonuniform directional filter banks and its application to image nonlinear approximation,” The 15th International Forum of Digital TV and Multimedia Communication, Shanghai, China, Sep. 20-21, 2018.

[27] Wei Zhong, Ke Liu, Long Ye, Li Fang and Qin Zhang, “Value-based local connection scheduling algorithm for distributed video transcoding system,” IEEE International Conference on Natural Computation, Fuzzy Systems and Knowledge Discovery, Huangshan, China, July 28-30, 2018.

[28] Wei Zhong, Li Fang, Qin Zhang and Long Ye, “Design of oversampled nonuniform filter banks with arbitrary rational frequency partitioning,” Signal, Image and Video Processing, vol. 11, no. 4, pp. 689-696, May 2017.

[29] Li Fang, Wei Zhong and Qin Zhang, “Design of M-channel linear-phase nonuniform filter banks with arbitrary rational sampling factors,” IET Signal Processing, vol. 10, no. 2, pp. 106-114, April 2016.

[30] Wei Zhong, Guangming Shi, Xuemei Xie, Xuyang Chen, “Design of linear-phase nonuniform filter banks with partial cosine modulation,” IEEE Trans. Signal Processing, vol. 58, no. 6, pp. 3390-3395, Jun. 2010.

[31] Wei Zhong, Guangming Shi, Xuemei Xie, Xuyang Chen, “Design of M-channel uniform linear-phase filter banks with partial cosine modulation,” Chinese Journal of Electronics, vol. 18, no. 3, pp. 477-480, July 2009.

[32] 钟微, 林琴心, 胡飞, 叶龙, 张勤, “基于前景人物语义理解的图像调和方法及系统,专利号: ZL202111180926.X, 权公告日: 2025923.

[33] 钟微, 操奎, 叶龙, 方力, 张勤, “基于室内场景的对象和场景合成方法及系统,专利号: ZL202011313114.3, 授权公告日: 202479.

[34] 钟微, 马鑫, 张勤, 叶龙, 胡飞, 方力, “音效合成方法、装置、电子设备及存储介质,专利号: ZL202111547899.5, 授权公告日: 2023630.