CodeRetriever: A Large Scale Contrastive Pre-Training Method for Code Search.

  • Xiaonan Li ,
  • ,
  • Yelong Shen ,
  • Xipeng Qiu ,
  • Hang Zhang ,
  • Bolun Yao ,
  • Weizhen Qi ,
  • Daxin Jiang (姜大昕) ,
  • ,
  • Nan Duan

EMNLP 2022 |

In this paper, we propose the CodeRetriever model, which learns the function-level code semantic representations through large-scale code-text contrastive pre-training. We adopt two contrastive learning schemes in CodeRetriever: unimodal contrastive learning and bimodal contrastive learning. For unimodal contrastive learning, we design an unsupervised learning approach to build semantic-related code pairs based on the documentation and function name. For bimodal contrastive learning, we leverage the documentation and in-line comments of code to build code-text pairs. Both contrastive objectives can fully leverage large-scale code corpus for pre-training.
Extensive experimental results show that CodeRetriever achieves new state-of-the-art with significant improvement over existing code pre-trained models, on eleven domain/language-specific code search tasks with six programming languages in different code granularity (function-level, snippet-level and statement-level). These results demonstrate the effectiveness and robustness of CodeRetriever.