Skip to content

Latest commit

 

History

History

10-benchmark_interpretation

Benchmark interpretation

Various attempts at interpreting the results of the benchmark.

# script/folder description
1 📄variability.R The variability of the benchmarking results across dataset sources and trajectory types
2 📄dataset_sources.R Similarity of the results across dataset sources
3a 📄topology_statistics.R Calculate some statistics about the topologies predicted by the methods
3b 📄topology_complexity.R Compare the complexity of the topologies predicted by the methods
4 📄complementarity.R Try to find good combinations of methods given a set of datasets, which could indicate some complementarity between methods
5 📄benchmark_interpretation_figure.R Combine the different figures of this experiment into one
6 📄compare_timings.R Compare the predicted execution times and memory usages with the actual ones
📄helper-complementarity.R Helper plotting functions the complementarity experiment
📄scaling_fit_cv.R Test different models for scalability, using the running times from the benchmark as validation set