Follow
Chaofan Tao
Title
Cited by
Cited by
Year
Compression of Generative Pre-trained Language Models via Quantization
C Tao, L Hou, W Zhang, L Shang, X Jiang, Q Liu, P Luo, N Wong
ACL 2022, 2022
912022
Dynamic and static context-aware lstm for multi-agent motion prediction
C Tao, Q Jiang, L Duan, P Luo
European Conference on Computer Vision, 547-563, 2020
662020
Upop: Unified and progressive pruning for compressing vision-language transformers
D Shi, C Tao, Y Jin, Z Yang, C Yuan, J Wang
International Conference on Machine Learning, 2023
332023
Mr-net: Exploiting mutual relation for visual relationship detection
Y Bin, Y Yang, C Tao, Z Huang, J Li, HT Shen
Proceedings of the AAAI Conference on Artificial Intelligence 33 (01), 8110-8117, 2019
292019
Structured pruning for efficient generative pre-trained language models
C Tao, L Hou, H Bai, J Wei, X Jiang, Q Liu, P Luo, N Wong
Findings of the Association for Computational Linguistics: ACL 2023, 10880-10895, 2023
282023
Scaling laws with vocabulary: Larger models deserve larger vocabularies
C Tao, Q Liu, L Dou, N Muennighoff, Z Wan, P Luo, M Lin, N Wong
arXiv preprint arXiv:2407.13623, 2024
182024
LiteVL: Efficient Video-Language Learning with Enhanced Spatial-Temporal Modeling
D Chen, C Tao, L Hou, L Shang, X Jiang, Q Liu
Conference on Empirical Methods in Natural Language Processing (EMNLP-2022), 2022
162022
Litegt: Efficient and lightweight graph transformers
C Chen, C Tao, N Wong
Proceedings of the 30th ACM International Conference on Information …, 2021
162021
Meit: Multi-modal electrocardiogram instruction tuning on large language models for report generation
Z Wan, C Liu, X Wang, C Tao, H Shen, Z Peng, J Fu, R Arcucci, H Yao, ...
arXiv preprint arXiv:2403.04945, 2024
152024
Rethinking kullback-leibler divergence in knowledge distillation for large language models
T Wu, C Tao, J Wang, R Yang, Z Zhao, N Wong
arXiv preprint arXiv:2404.02657, 2024
142024
Minimax entropy network: Learning category-invariant features for domain adaptation
C Tao, F Lv, L Duan, M Wu
arXiv preprint arXiv:1904.09601, 2019
142019
Crossget: Cross-guided ensemble of tokens for accelerating vision-language transformers
D Shi, C Tao, A Rao, Z Yang, C Yuan, J Wang
arXiv preprint arXiv:2305.17455, 2023
112023
FAT: Frequency-aware transformation for bridging full-precision and low-precision deep representations
C Tao, R Lin, Q Chen, Z Zhang, P Luo, N Wong
IEEE Transactions on Neural Networks and Learning Systems 35 (2), 2640-2654, 2022
11*2022
RoboCodeX: Multimodal Code Generation for Robotic Behavior Synthesis
Y Mu, J Chen, Q Zhang, S Chen, Q Yu, C Ge, R Chen, Z Liang, M Hu, ...
arXiv preprint arXiv:2402.16117, 2024
102024
Frequency Regularization for Improving Adversarial Robustness
B Huang, C Tao, R Lin, N Wong
AAAI 2023 workshop, 2022
10*2022
Dybit: Dynamic bit-precision numbers for efficient quantized neural network inference
J Zhou, J Wu, Y Gao, Y Ding, C Tao, B Li, F Tu, KT Cheng, HKH So, ...
IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems, 2023
92023
D2O: Dynamic Discriminative Operations for Efficient Generative Inference of Large Language Models
Z Wan, X Wu, Y Zhang, Y Xin, C Tao, Z Zhu, X Wang, S Luo, J Xiong, ...
arXiv preprint arXiv:2406.13035, 2024
72024
Source-free domain adaptation with unrestricted source hypothesis
J He, L Wu, C Tao, F Lv
Pattern Recognition 149, 110246, 2024
72024
Batmann: A binarized-all-through memory-augmented neural network for efficient in-memory computing
Y Ren, R Lin, J Ran, C Liu, C Tao, Z Wang, C Li, N Wong
2021 IEEE 14th International Conference on ASIC (ASICON), 1-4, 2021
22021
Autoregressive Models in Vision: A Survey
J Xiong, G Liu, L Huang, C Wu, T Wu, Y Mu, Y Yao, H Shen, Z Wan, ...
arXiv preprint arXiv:2411.05902, 2024
12024
The system can't perform the operation now. Try again later.
Articles 1–20