Fix-Filter-Fix: Intuitively Connect Any Models for Effective Bug Fixing
- Publication Type:
- Conference Proceeding
- Citation:
- EMNLP 2021 - 2021 Conference on Empirical Methods in Natural Language Processing, Proceedings, 2021, pp. 3495-3504
- Issue Date:
- 2021-01-01
Closed Access
Filename | Description | Size | |||
---|---|---|---|---|---|
2021.emnlp-main.282.pdf | Published version | 516.33 kB |
Copyright Clearance Process
- Recently Added
- In Progress
- Closed Access
This item is closed access and not available.
Locating and fixing bugs is a time-consuming task. Most neural machine translation (NMT) based approaches for automatically bug fixing lack generality and do not make full use of the rich information in the source code. In NMT-based bug fixing, we find some predicted code identical to the input buggy code (called unchanged fix) in NMT-based approaches due to high similarity between buggy and fixed code (e.g., the difference may only appear in one particular line). Obviously, unchanged fix is not the correct fix because it is the same as the buggy code that needs to be fixed. Based on these, we propose an intuitive yet effective general framework (called Fix-Filter-Fix or F3) for bug fixing. F3 connects models with our filter mechanism to filter out the last model's unchanged fix to the next. We propose an F3 theory that can quantitatively and accurately calculate the F3 lifting effect. To evaluate, we implement the Seq2Seq Transformer (ST) and the AST2Seq Transformer (AT) to form some basic F3 instances, called F3ST+AT and F3AT+ST. Comparing them with single model approaches and many model connection baselines across four datasets validates the effectiveness and generality of F3 and corroborates our findings and methodology.
Please use this identifier to cite or link to this item: