Follow
Zheng Che
Zheng Che
Verified email at google.com
Title
Cited by
Cited by
Year
Synthesizer: Rethinking self-attention for transformer models
Y Tay, D Bahri, D Metzler, DC Juan, Z Zhao, C Zheng
International conference on machine learning, 10183-10192, 2021
1662021
Structformer: Joint unsupervised induction of dependency and constituency structure from masked language modeling
Y Shen, Y Tay, C Zheng, D Bahri, D Metzler, A Courville
arXiv preprint arXiv:2012.00857, 2020
172020
Choppy: Cut transformer for ranked list truncation
D Bahri, Y Tay, C Zheng, D Metzler, A Tomkins
Proceedings of the 43rd International ACM SIGIR Conference on Research and …, 2020
82020
Reverse engineering configurations of neural text generation models
Y Tay, D Bahri, C Zheng, C Brunk, D Metzler, A Tomkins
arXiv preprint arXiv:2004.06201, 2020
82020
Synthesizer: Rethinking self-attention in transformer models. arXiv 2020
Y Tay, D Bahri, D Metzler, D Juan, Z Zhao, C Zheng
arXiv preprint arXiv:2005.00743 2, 2020
62020
Generative models are unsupervised predictors of page quality: A colossal-scale study
D Bahri, Y Tay, C Zheng, D Metzler, C Brunk, A Tomkins
arXiv preprint arXiv:2008.13533, 2020
22020
Surprise: Result List Truncation via Extreme Value Theory
D Bahri, C Zheng, Y Tay, D Metzler, A Tomkins
arXiv preprint arXiv:2010.09797, 2020
2020
The system can't perform the operation now. Try again later.
Articles 1–7