Follow
Difei Gao
Difei Gao
National U. of Singapore; Institute of Computing Technology, Chinese Academy of Sciences
Verified email at nus.edu.sg
Title
Cited by
Cited by
Year
Multi-modal graph neural network for joint reasoning on vision and scene text
D Gao, K Li, R Wang, S Shan, X Chen
IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 12746 …, 2020
772020
Egocentric video-language pretraining
KQ Lin, AJ Wang, M Soldan, M Wray, R Yan, EZ Xu, D Gao, R Tu, W Zhao
Neural Information Processing Systems (NeurIPS) 2 (3), 2022
28*2022
CRIC: A vqa dataset for compositional reasoning on vision and commonsense
D Gao, R Wang, S Shan, X Chen
IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI), 2022
13*2022
Env-QA: A Video Question Answering Benchmark for Comprehensive Understanding of Dynamic Environments
D Gao, R Wang, Z Bai, X Chen
IEEE/CVF International Conference on Computer Vision (ICCV), 1675-1685, 2021
92021
GEB+: A Benchmark for Generic Event Boundary Captioning, Grounding and Retrieval
Y Wang, D Gao, L Yu, W Lei, M Feiszli, MZ Shou
European Conference on Computer Vision (ECCV), 2022
8*2022
AssistQ: Affordance-centric Question-driven Task Completion for Egocentric Assistant
B Wong, J Chen, Y Wu, SW Lei, D Mao, D Gao, MZ Shou
European Conference on Computer Vision (ECCV), 2022
82022
Learning to recognize visual concepts for visual question answering with structural label space
D Gao, R Wang, S Shan, X Chen
IEEE Journal of Selected Topics in Signal Processing (JSTSP) 14 (3), 494-505, 2020
62020
Symbolic replay: Scene graph as prompt for continual learning on vqa task
SW Lei*, D Gao*, JZ Wu, Y Wang, W Liu, M Zhang, MZ Shou
The AAAI Conference on Artificial Intelligence (AAAI), 2023
42023
CONE: An Efficient COarse-to-fiNE Alignment Framework for Long Video Temporal Grounding
Z Hou, W Zhong, L Ji, D Gao, K Yan, WK Chan, CW Ngo, Z Shou, N Duan
arXiv preprint arXiv:2209.10918, 2022
42022
Correlated warped Gaussian processes for gender-specific age estimation
D Gao, L Pan, R Liu, R Chen, M Xie
2015 IEEE International Conference on Image Processing (ICIP), 133-137, 2015
42015
KMIR: A Benchmark for Evaluating Knowledge Memorization, Identification and Reasoning Abilities of Language Models
D Gao, Y Jia, L Li, C Fu, Z Dou, H Jiang, X Zhang, L Chen, Z Cao
arXiv preprint arXiv:2202.13529, 2022
32022
Assistsr: Task-oriented video segment retrieval for personal AI assistant
W Lei, D Gao, Y Wang, D Mao, Z Liang, L Ran, MZ Shou
Findings of the Association for Computational Linguistics: EMNLP 2022, 319-338, 2022
22022
MIST: Multi-modal Iterative Spatial-Temporal Transformer for Long-form Video Question Answering
D Gao, L Zhou, L Ji, L Zhu, Y Yang, MZ Shou
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern …, 2023
12023
An Efficient COarse-to-fiNE Alignment Framework@ Ego4D Natural Language Queries Challenge 2022
Z Hou, W Zhong, L Ji, D Gao, K Yan, WK Chan, CW Ngo, Z Shou, N Duan
arXiv preprint arXiv:2211.08776, 2022
12022
Egocentric Video-Language Pretraining@ Ego4D Challenge 2022
KQ Lin, AJ Wang, M Soldan, M Wray, R Yan, EZ Xu, D Gao, R Tu, W Zhao, ...
arXiv preprint arXiv:2207.01622, 2022
12022
Mover: Mask and Recovery based Facial Part Consistency Aware Method for Deepfake Video Detection
J Hu, X Liao, D Gao, S Tsutsui, Q Wang, Z Qin, MZ Shou
arXiv preprint arXiv:2305.05943, 2023
2023
DeepfakeMAE: Facial Part Consistency Aware Masked Autoencoder for Deepfake Video Detection
J Hu, X Liao, D Gao, S Tsutsui, Z Qin, MZ Shou
arXiv preprint arXiv:2303.01740, 2023
2023
Affordance Grounding from Demonstration Video to Target Image
J Chen, D Gao, KQ Lin, MZ Shou
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern …, 2023
2023
Learning to Learn: How to Continuously Teach Humans and Machines
P Singh, Y Li, A Sikarwar, W Lei, D Gao, MB Talbot, Y Sun, MZ Shou, ...
arXiv preprint arXiv:2211.15470, 2022
2022
Visual Textbook Network: Watch Carefully before Answering Visual Questions.
D Gao, R Wang, S Shan, X Chen
British Machine Vision Conference (BMVC), 2017
2017
The system can't perform the operation now. Try again later.
Articles 1–20