1 |
Code to Comment Translation: A Comparative Study on Model Effectiveness & Errors ...
|
|
|
|
Abstract:
Automated source code summarization is a popular software engineering research topic wherein machine translation models are employed to "translate" code snippets into relevant natural language descriptions. Most evaluations of such models are conducted using automatic reference-based metrics. However, given the relatively large semantic gap between programming languages and natural language, we argue that this line of research would benefit from a qualitative investigation into the various error modes of current state-of-the-art models. Therefore, in this work, we perform both a quantitative and qualitative comparison of three recently proposed source code summarization models. In our quantitative evaluation, we compare the models based on the smoothed BLEU-4, METEOR, and ROUGE-L machine translation metrics, and in our qualitative evaluation, we perform a manual open-coding of the most common errors committed by the models when compared to ground truth captions. Our investigation reveals new insights into ... : Accepted to the 2021 NLP4Prog Workshop co-located with The Joint Conference of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (ACL-IJCNLP 2021) ...
|
|
Keyword:
Artificial Intelligence cs.AI; Computation and Language cs.CL; FOS Computer and information sciences; Machine Learning cs.LG; Software Engineering cs.SE
|
|
URL: https://dx.doi.org/10.48550/arxiv.2106.08415 https://arxiv.org/abs/2106.08415
|
|
BASE
|
|
Hide details
|
|
2 |
Towards Minimal Supervision BERT-based Grammar Error Correction ...
|
|
|
|
BASE
|
|
Show details
|
|
3 |
It's not a Non-Issue: Negation as a Source of Error in Machine Translation ...
|
|
|
|
BASE
|
|
Show details
|
|
|
|