Intermediate-task transfer learning with pretrained models for natural language understanding: When and why does it work? Y Pruksachatkun, J Phang, H Liu, PM Htut, X Zhang, RY Pang, C Vania, ... arXiv preprint arXiv:2005.00628, 2020 | 185 | 2020 |
BLiMP: The benchmark of linguistic minimal pairs for English A Warstadt, A Parrish, H Liu, A Mohananey, W Peng, SF Wang, ... Transactions of the Association for Computational Linguistics 8, 377-392, 2020 | 171 | 2020 |
Investigating BERT's knowledge of language: five analysis methods with NPIs A Warstadt, Y Cao, I Grosu, W Peng, H Blix, Y Nie, A Alsop, S Bordia, ... arXiv preprint arXiv:1909.02597, 2019 | 97 | 2019 |
jiant: A software toolkit for research on general-purpose text understanding models Y Pruksachatkun, P Yeres, H Liu, J Phang, PM Htut, A Wang, I Tenney, ... arXiv preprint arXiv:2003.02249, 2020 | 78* | 2020 |
Few-shot parameter-efficient fine-tuning is better and cheaper than in-context learning H Liu, D Tam, M Muqeeth, J Mohta, T Huang, M Bansal, CA Raffel Advances in Neural Information Processing Systems 35, 1950-1965, 2022 | 61 | 2022 |
English intermediate-task training improves zero-shot cross-lingual transfer too J Phang, I Calixto, PM Htut, Y Pruksachatkun, H Liu, C Vania, K Kann, ... arXiv preprint arXiv:2005.13013, 2020 | 56 | 2020 |
Learning which features matter: RoBERTa acquires a preference for linguistic generalizations (eventually) A Warstadt, Y Zhang, HS Li, H Liu, SR Bowman arXiv preprint arXiv:2010.05358, 2020 | 54 | 2020 |
Counterfactually-augmented SNLI training data does not yield better generalization than unaugmented data W Huang, H Liu, SR Bowman arXiv preprint arXiv:2010.04762, 2020 | 23 | 2020 |
Comparing test sets with item response theory C Vania, PM Htut, W Huang, D Mungra, RY Pang, J Phang, H Liu, K Cho, ... arXiv preprint arXiv:2106.00840, 2021 | 15 | 2021 |
Precise task formalization matters in Winograd schema evaluations H Liu, W Huang, DA Mungra, SR Bowman arXiv preprint arXiv:2010.04043, 2020 | 9 | 2020 |
Memd: A diversity-promoting learning framework for short-text conversation M Zou, X Li, H Liu, ZH Deng Proceedings of the 27th International Conference on Computational …, 2018 | 5 | 2018 |
Fine-tuned transformers show clusters of similar representations across layers J Phang, H Liu, SR Bowman arXiv preprint arXiv:2109.08406, 2021 | 4 | 2021 |
Retrieving Relevant and Diverse Image from Social Media Images. X Chen, H Liu, ZH Deng, Y Yang MediaEval, 2015 | 3 | 2015 |