杨双 女 硕导 中国科学院计算技术研究所
电子邮件: shuang.yang@ict.ac.cn
通信地址: 北京市海淀区科学院南路6号
邮政编码: 100190
课题组主页:视听语言感知与理解课题组,长期欢迎优秀实习生加入 (简历与自我介绍发上述邮箱);
研究领域
计算机视觉,模式识别,视听语言感知与理解
招生信息
招生专业
招生方向
工作经历
工作简历
出版信息
F Wang, S Yang, S Shan, X Chen. Cooperative Dual Attention for Audio-Visual Speech Enhancement with Visual Cues. British Machine Vision Conference (BMVC), Aberdeen, UK, Nov. 20-24, 2023.
B Xia, S Yang, S Shan, X Chen. UniLip: Learning Visual-Textual Mapping with Uni-Modal Data for Lip Reading. British Machine Vision Conference, Aberdeen, UK, Nov. 20-24, 2023.
S Luo, S Yang, S Shan, X Chen. Learning Separable Hidden Unit Contributions for Speaker-Adaptive Lip-Reading. British Machine Vision Conference, Aberdeen, UK, Nov. 20-24, 2023.
- D Feng, S Yang, S Shan, X Chen. Audio-Driven Deformation Flow for Effective Lip Reading. 26th International Conference on Pattern Recognition (ICPR), pp. 274-280, Aug. 21-25, 2022.
Y Zhang, S Liang, S Yang, X Liu, Z Wu and S Shan. ICTCAS-UCAS-TAL Submission to the AVA-ActiveSpeaker Task (The 1st Place). IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshop of the International Challenge on Activity Recognition (ActivityNet), 2021.
Y Zhang, S Liang, S Yang, X Liu, Z Wu, S Shan, X Chen. UniCon: Unified Context Network for Robust Active Speaker Detection. ACM International Conference on Multimedia (ACM Multimedia), pp. 3964-3972, Chengdu, China, Oct. 20-24, 2021.
D Feng, S Yang, S Shan. An Efficient Software for Building Lip Reading Models Without Pains. IEEE International Conference on Multimedia & Expo Workshops (ICMEW), pp. 1-2, Virtual Event, Jul. 5-9, 2021
- 《机器学习·应用视角》,机械工业出版社,合译,2020
- Mutual Information Maximization for Effective Lip Reading, X Zhao, S Yang, S Shan, X Chen, IEEE International Conference on Automatic Face and Gesture Recognition, 2020
Deformation Flow Based Two-Stream Network for Lip Reading, J Xiao, S Yang, Y Zhang, S Shan, X Chen, IEEE International Conference on Automatic Face and Gesture Recognition, 2020
Pseudo-Convolutional Policy Gradient for Sequence-to-Sequence Lip-Reading, M Luo, S Yang, S Shan, X Chen, IEEE International Conference on Automatic Face and Gesture Recognition, 2020
Can We Read Speech Beyond the Lips? Rethinking RoI Selection for Deep Visual Speech Recognition, Y Zhang, S Yang, J Xiao, S Shan, X Chen, IEEE International Conference on Automatic Face and Gesture Recognition, 2020
A Novel Pseudo Viewpoint based Holoscopic 3D Micro-Gesture Recognition,Y Liu, S Yang, H Meng, MR Swash, S Shan, ACM ICMI, 2020
Synchronous Bidirectional Learning for Multilingual Lip Reading, M Luo, S Yang, X Chen, Z Liu, S Shan, The British Machine Vision Conference (BMVC), 2020
LRW-1000: A naturally-distributed large-scale benchmark for lip reading in the wild, S Yang, Y Zhang, D Feng, M Yang, C Wang, J Xiao, K Long, S Shan, X Chen, IEEE International Conference on Automatic Face & Gesture Recognition, 2019
Multi-Task Learning for Audio-Visual Active Speaker Detection, YH Zhang, J Xiao, S Yang, S Shan, The ActivityNet Large-Scale Activity Recognition Challenge @ CVPR 2019,2019
TinyPoseNet: A Fast and Compact Deep Network for Robust Head Pose Estimation, S Li, L Wang, S Yang, Y Wang, C Wang, International Conference on Neural Information Processing, 2017
The Class-specific Oriented Attributes for Action Recognition, H Yang, B Wu, S Yang, C Yuan, W Hu, Chinese Association for Artificial Inteeligence, 2016
Hierarchical Bayesian Multiple Kernel Learning Based Feature Fusion for Action Recognition, W Sun, C Yuan, P Wang, S Yang, W Hu, Z Cai, Multimodal Pattern Recognition of Social Signals in Human-Computer-Interaction, 2016
Multi-feature max-margin hierarchical bayesian model for action recognition, S Yang, C Yuan, B Wu, W Hu, F Wang, Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, 2015
Human Action Recognition Based on Oriented Motion Salient Regions, B Wu, S Yang, C Yuan, W Hu, F Wang, Asian Conference on Computer Vision Workshop, 2014
A hierarchical model based on latent dirichlet allocation for action recognition, S Yang, C Yuan, W Hu, X Ding, International Conference on Pattern Recognition, 2014
Learning human actions by combining global dynamics and local appearance, G Luo, S Yang, G Tian, C Yuan, W Hu, SJ Maybank, IEEE transactions on pattern analysis and machine intelligence, 2014
Combining sparse appearance features and dense motion features via random forest for action detection,S Yang, C Yuan, H Wang, W Hu, IEEE International Conference on Acoustics, Speech and Signal Processing, 2013
Multi-task sparse learning with beta process prior for action recognition, C Yuan, W Hu, G Tian, S Yang, H Wang, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2013
Online Detection and Tracking Method of Foreign Substances in Ampoules in High-speed Pharmaceutical Lines, S Yang, Y Wang, Chinese Journal of Scientific Instrument, 2011
A Detection System for Impurity of Ampoule Injection Based on Machine-vision, S Yang, Y Wang, Optp-Electronic Engineering
科研活动
NSFC面上项目,项目负责人
NSFC青年基金项目,项目负责人
NSFC面上项目,子课题负责人
大中型企业委托项目,多模态识别技术合作项目,项目负责人
大中型企业委托项目,语音驱动的高质量说话视频合成,项目负责人
中科院计算所百星计划,项目负责人
科技创新2030-重大项目,开放场景下的行为识别,骨干成员
国家重点研发计划,下一代深度学习理论、方法与关键技术,骨干成员