“…To substantiate this quality, we fine-tune prominent CodeLLMs on tasks that necessitate the involvement of both code and text, including code summarization, code search, and code generation. [Clement et al, 2020] 1 ≈ 7,700,000 -CoDesc [Hasan et al, 2021] 1 4,211,516 -CodeSearchNet [Husain et al, 2019] 6 2,326,976 4,125,470 CodeXGLUE CSN [Lu et al, 2021] 6 1,005,474 -Deepcom [Hu et al, 2020] 1 424,028 -CONCODE [Iyer et al, 2018b We then compare these models, which have been fine-tuned on The Vault, with those fine-tuned on CSN. The comparison is made using the same test datasets and commonly employed metrics, such as BLEU, MRR, and pass@k.…”