Abstract
Software defect prediction (SDP) is essential for ensuring software quality and reliability. Transformer architectures, known for their success in vision and NLP, offer significant potential for SDP. This study investigates the application of the AutoInt Transformer model, which excels in capturing complex feature interactions from numerical and categorical data, for predicting defects. AutoInt is evaluated on several SDP datasets and compared to advanced models like GRU, Tab-Net, FT-Transformer, and TabTransformer using standard evaluation metrics. Results demonstrate AutoInt's superior performance, further validated through effect size measurements. This research highlights the promise of transformer-based techniques for advancing SDP and enhancing software quality assurance.
| Original language | English |
|---|---|
| Pages (from-to) | 103-104 |
| Number of pages | 2 |
| Journal | Proceedings of the IEEE International Conference on Big Data and Smart Computing, BIGCOMP |
| Issue number | 2025 |
| DOIs | |
| State | Published - 2025 |
| Event | 2025 IEEE International Conference on Big Data and Smart Computing, BigComp 2025 - Kota Kinabalu, Malaysia Duration: 2025.02.9 → 2025.02.12 |
Keywords
- AutoInt
- Deep Learning
- Software Defect prediction
- Transformer
Quacquarelli Symonds(QS) Subject Topics
- Computer Science & Information Systems
- Data Science
Fingerprint
Dive into the research topics of 'Enhancing the Software Defect Prediction using AutoInt'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver