Zihang Dai
Zihang Dai
Unknown affiliation
No verified email - Homepage
Cited by
Cited by
Xlnet: Generalized autoregressive pretraining for language understanding
Z Yang, Z Dai, Y Yang, J Carbonell, R Salakhutdinov, QV Le
arXiv preprint arXiv:1906.08237, 2019
Transformer-xl: Attentive language models beyond a fixed-length context
Z Dai, Z Yang, Y Yang, J Carbonell, QV Le, R Salakhutdinov
arXiv preprint arXiv:1901.02860, 2019
Unsupervised data augmentation for consistency training
Q Xie, Z Dai, E Hovy, MT Luong, QV Le
arXiv preprint arXiv:1904.12848, 2019
Good semi-supervised learning that requires a bad gan
Z Dai, Z Yang, F Yang, WW Cohen, R Salakhutdinov
arXiv preprint arXiv:1705.09783, 2017
Breaking the softmax bottleneck: A high-rank RNN language model
Z Yang, Z Dai, R Salakhutdinov, WW Cohen
arXiv preprint arXiv:1711.03953, 2017
Unsupervised data augmentation
Q Xie, Z Dai, E Hovy, MT Luong, QV Le
arXiv preprint arXiv:1904.12848 2 (6), 2019
Controllable invariance through adversarial feature learning
Q Xie, Z Dai, Y Du, E Hovy, G Neubig
arXiv preprint arXiv:1705.11122, 2017
Cfo: Conditional focused neural question answering with large-scale knowledge bases
Z Dai, L Li, W Xu
arXiv preprint arXiv:1606.01994, 2016
Characterizing and avoiding negative transfer
Z Wang, Z Dai, B Póczos, J Carbonell
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern …, 2019
Calibrating energy-based generative adversarial networks
Z Dai, A Almahairi, P Bachman, E Hovy, A Courville
arXiv preprint arXiv:1702.01691, 2017
Switchout: an efficient data augmentation algorithm for neural machine translation
X Wang, H Pham, Z Dai, G Neubig
arXiv preprint arXiv:1808.07512, 2018
An interpretable knowledge transfer model for knowledge base completion
Q Xie, X Ma, Z Dai, E Hovy
arXiv preprint arXiv:1704.05908, 2017
Meta pseudo labels
H Pham, Z Dai, Q Xie, MT Luong, QV Le
arXiv preprint arXiv:2003.10580, 2020
Large-scale cloze test dataset created by teachers
Q Xie, G Lai, Z Dai, E Hovy
arXiv preprint arXiv:1711.03225, 2017
Funnel-transformer: Filtering out sequential redundancy for efficient language processing
Z Dai, G Lai, Y Yang, QV Le
arXiv preprint arXiv:2006.03236, 2020
Systems and methods for human inspired simple question answering (HISQA)
L Li, Z Dai, W Xu
US Patent 10,606,846, 2020
Transformer-xl: Language modeling with longer-term dependency
Z Dai, Z Yang, Y Yang, WW Cohen, J Carbonell, QV Le, R Salakhutdinov
Fast and simple mixture of softmaxes with bpe and hybrid-lightrnn for language generation
X Kong, Q Xie, Z Dai, E Hovy
Proceedings of the AAAI Conference on Artificial Intelligence 33 (01), 6626-6633, 2019
A mutual information maximization perspective of language representation learning
L Kong, CM d'Autume, W Ling, L Yu, Z Dai, D Yogatama
arXiv preprint arXiv:1910.08350, 2019
Wiki-40B: Multilingual Language Model Dataset
M Guo, Z Dai, D Vrandečić, R Al-Rfou
Proceedings of The 12th Language Resources and Evaluation Conference, 2440-2452, 2020
The system can't perform the operation now. Try again later.
Articles 1–20