[1] Yixuan Su, Fangyu Liu, Zaiqiao Meng, Lei Shu, Ehsan Shareghi, and Nigel Collier. TaCL: Improving bert pre-training with token-aware contrastive learning, 2021
[2] Lee Xiong, Chenyan Xiong, Ye Li, Kwok-Fung Tang, Jialin Liu, Paul N. Bennett, Junaid Ahmed, and Arnold Overwijk. Approximate nearest neighbor negative contrastive learning for dense text retrieval. CoRR, abs/2007.00808, 2020.
[3] Rodrigo Nogueira and Kyunghyun Cho. Passage re-ranking with BERT. CoRR, abs/1901.04085, 2019.
[4] Jimmy Lin, Rodrigo Nogueira, and Andrew Yates. Pretrained transformers for text ranking: BERT and beyond. CoRR, abs/2010.06467, 2020.
[5] Sebastian Hofst ̈atter, Sheng-Chieh Lin, Jheng-Hong Yang, Jimmy Lin, and Allan Hanbury. Efficiently teaching an effective dense retriever with balanced topic-aware sampling. CoRR, abs/2104.06967, 2021
[6] Omar Khattab and Matei Zaharia. ColBERT: Efficient and Effective Passage Search via Contextualized Late Interaction over BERT, pages 39–48. Association for Computing Machinery, New York, NY, USA, 2020.
[7] Luyu Gao and Jamie Callan. Is your language model ready for dense representation fine-tuning? CoRR, abs/2104.08253, 2021
[8] Sebastian Hofst ̈atter, Sophia Althammer, Michael Schr ̈oder, Mete Sertkan, and Allan Hanbury. Improving efficient neural ranking models with cross-architecture knowledge distillation. CoRR, abs/2010.02666, 2020.
[9] Google. Google search understanding using BERT. https://blog.google/products/search/search-language-understanding-bert/.