Distilling task-specific knowledge from bert into simple neural networks R Tang*, Y Lu*, L Liu*, L Mou, O Vechtomova, J Lin arXiv preprint arXiv:1903.12136, 2019 | 465 | 2019 |
Generative Adversarial Network for Abstractive Text Summarization L Liu, Y Lu, M Yang, Q Qu, J Zhu, H Li AAAI Conference on Artificial Intelligence, 2018 | 226 | 2018 |
PAQ: 65 million probably-asked questions and what you can do with them P Lewis, Y Wu, L Liu, P Minervini, H Küttler, A Piktus, P Stenetorp, ... Transactions of the Association for Computational Linguistics 9, 1098-1115, 2021 | 201 | 2021 |
What the DAAM: Interpreting Stable Diffusion Using Cross Attention R Tang*, L Liu*, A Pandey, Z Jiang, G Yang, K Kumar, J Lin, F Ture arXiv preprint arXiv:2210.04885, 2022 | 120 | 2022 |
Bridging the Gap Between Relevance Matching and Semantic Matching for Short Text Similarity Modeling J Rao, L Liu, Y Tay, W Yang, P Shi, J Lin EMNLP, 2019 | 86 | 2019 |
Neurips 2020 efficientqa competition: Systems, analyses and lessons learned S Min, J Boyd-Graber, C Alberti, D Chen, E Choi, M Collins, K Guu, ... NeurIPS 2020 Competition and Demonstration Track, 86-111, 2021 | 74 | 2021 |
When do flat minima optimizers work? J Kaddour, L Liu, R Silva, MJ Kusner Advances in Neural Information Processing Systems 35, 16577-16595, 2022 | 73 | 2022 |
Controllable abstractive dialogue summarization with sketch supervision CS Wu*, L Liu*, W Liu, P Stenetorp, C Xiong arXiv preprint arXiv:2105.14064, 2021 | 52 | 2021 |
Challenges in generalization in open domain question answering L Liu, P Lewis, S Riedel, P Stenetorp arXiv preprint arXiv:2109.01156, 2021 | 47 | 2021 |
Detecting" Smart" Spammers On Social Network: A Topic Model Approach L Liu, Y Lu, Y Luo, R Zhang, L Itti, J Lu Proceedings of the NAACL Student Research Workshop, 2016 | 47 | 2016 |
Mkd: a multi-task knowledge distillation approach for pretrained language models L Liu, H Wang, J Lin, R Socher, C Xiong arXiv preprint arXiv:1911.03588, 2019 | 45* | 2019 |
Multi-task knowledge distillation for language model L Liu, C Xiong US Patent 11,620,515, 2023 | 15 | 2023 |
Incorporating Contextual and Syntactic Structures Improves Semantic Similarity Modeling L Liu, W Yang, J Rao, R Tang, J Lin EMNLP, 2019 | 14 | 2019 |
Query expansion using contextual clue sampling with language models L Liu, M Li, J Lin, S Riedel, P Stenetorp arXiv preprint arXiv:2210.07093, 2022 | 12 | 2022 |
Distilling task-specific knowledge from bert into simple neural networks. arXiv 2019 R Tang, Y Lu, L Liu, L Mou, O Vechtomova, J Lin arXiv preprint arXiv:1903.12136, 1903 | 9 | 1903 |
Distilling taskspecific knowledge from BERT into simple neural networks. arXiv R Tang, Y Lu, L Liu, L Mou, O Vechtomova, J Lin arXiv preprint arXiv:1903.12136, 2019 | 7 | 2019 |
Coarse-to-fine abstractive dialogue summarization with controllable granularity CS Wu, W Liu, C Xiong, L Liu US Patent App. 17/159,625, 2022 | 6 | 2022 |
Scalable content-based analysis of images in web archives with tensorflow and the archives unleashed toolkit HW Yang, L Liu, I Milligan, N Ruest, J Lin 2019 ACM/IEEE Joint Conference on Digital Libraries (JCDL), 436-437, 2019 | 6 | 2019 |
Distilling task-specific knowledge from BERT into simple neural networks. CoRR abs/1903.12136 (2019) R Tang, Y Lu, L Liu, L Mou, O Vechtomova, J Lin URL: http://arxiv. org/abs, 1903 | 5 | 1903 |
Systems and methods for multi-scale pre-training with densely connected transformer L Liu, C Xiong US Patent 11,941,356, 2024 | | 2024 |