Skip to content

VeriFIT/smt-string-bench-results

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

SMT String Bench Results

Running the experimental evaluation

We assume you have VeriFIT/smt-bench set up and running on an evaluation server where you run the experimental evaluation (according to its instructions).

  1. Set up the Python virtual environment:

    python -m venv .venv
    source .venv/bin/activate
    pip install -r requirements.txt
  2. Get new tasks from the server with experimental results (change host, port, etc. if running from a different server):

    ./get_tasks_and_generate_csv.sh
  3. Process the results (choose one):

  • Run the Jupyter evaluation notebook eval.ipynb:
    • Set the correct tools and benchmarks to evaluate (set the version of NOODLER).
    • Run the first 4 cells to load the benchmarks and other cells based on your need.
  • Only prepare the results for manual processing:
    ./pyco_proc.py [options] <requested_tasks_file_with_results.tasks>
    Store the processed results and evaluate them manually.
  1. Exit the Python virtual environment:
    deactivate

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published