Skip to content

Latest commit

 

History

History
53 lines (35 loc) · 4 KB

File metadata and controls

53 lines (35 loc) · 4 KB

DOI License: CC BY 4.0

Predicting Unstable Software Benchmarks Using Static Source Code Features

This is the replication package of the paper "Predicting Unstable Software Benchmarks Using Static Source Code Features" authored by Christoph Laaber, Mikael Basmaci, and Pasquale Salza.

This replication package is also available on GitHub: sealuzh/benchmark-instability-prediction-replication-package.

Approach

approach/ contains all data and scripts of our approach, including feature extraction and combination, variability computation, and generating the resulting files for the machine learning part in study/. The approach's README contains detailed information.

Study

study/ contains the majority of the data and scripts to completely reproduce the study we conducted to evaluate our approach.

r_analyses/ contains the data and scripts for running the individual feature analysis of RQ 2 as well as the scripts for creating the scatter plot of Figure 2.

resources/variabilities_5_iterations.csv, resources/variabilities_10_iterations.csv, resources/variabilities_20_iterations.csv, and resources/variabilities_30_iterations.csv are the files containing the data we collected by running the benchmarks, with a number of iterations of 5, 10, 20, and 30, respectively.

Classification (RQ 1)

  • classification.py is a Python script to run all the experiments to train the machine learning models and evaluate their performance.
  • resources/classification_results.csv.xz is the outcome of the previous step, containing all the computed metrics for all the combinations of machine learning algorithm, number of iterations, threshold, and fold.
  • classification_study.ipynb is a Jupyter Notebook we used to study the prediction performance of our approach.

Feature Importance (RQ 2)

Individual Features

  • feature_importance.py is a Python script to run the permutation feature importance of individual features.
  • resources/feature_importance_mcc_results.csv is the outcome of the previous step, containing all MCC feature importances for each variability measure (RCIW Maritz-Jarrett, RCIW bootstrap, and RMAD) and each fold. We also provide feature importances for other prediction preformances metrics, i.e., AUC and F-measure.
  • individual_feature_importance.R is an R script to run the individual feature analyses and plot Figure 10. To plot the figure, run the R function run. To get statistics, run the R function run_individual_features_stats.

Feature Categories

For Jupyter Notebooks, we also provide the compiled HTML file with included output.

Contact

Contact information is available in the file CONTACT.md.