mirror of
https://github.com/catchorg/Catch2.git
synced 2024-11-22 13:26:10 +01:00
First commit of benchmarks for Catch runtime perf.
So far its very much a WIP with some problems that are known already and not very representative tests.
This commit is contained in:
parent
e3659cdddd
commit
3b7511e564
@ -7,6 +7,7 @@ set_property(GLOBAL PROPERTY USE_FOLDERS ON)
|
||||
# define some folders
|
||||
set(CATCH_DIR ${CMAKE_CURRENT_SOURCE_DIR})
|
||||
set(SELF_TEST_DIR ${CATCH_DIR}/projects/SelfTest)
|
||||
set(BENCHMARK_DIR ${CATCH_DIR}/projects/Benchmark)
|
||||
set(HEADER_DIR ${CATCH_DIR}/include)
|
||||
|
||||
if(USE_CPP11)
|
||||
@ -161,13 +162,20 @@ set(HEADERS
|
||||
${HEADER_DIR}/reporters/catch_reporter_xml.hpp
|
||||
)
|
||||
|
||||
set(BENCH_SOURCES
|
||||
${BENCHMARK_DIR}/BenchMain.cpp
|
||||
${BENCHMARK_DIR}/StringificationBench.cpp
|
||||
)
|
||||
|
||||
# Provide some groupings for IDEs
|
||||
SOURCE_GROUP("Tests" FILES ${TEST_SOURCES})
|
||||
SOURCE_GROUP("Surrogates" FILES ${IMPL_SOURCES})
|
||||
SOURCE_GROUP("Benchmarks" FILES ${BENCH_SOURCES})
|
||||
|
||||
# configure the executable
|
||||
include_directories(${HEADER_DIR})
|
||||
add_executable(SelfTest ${TEST_SOURCES} ${IMPL_SOURCES} ${HEADERS})
|
||||
add_executable(Benchmark ${BENCH_SOURCES} ${HEADERS})
|
||||
|
||||
# configure unit tests via CTest
|
||||
enable_testing()
|
||||
|
2
projects/Benchmark/BenchMain.cpp
Normal file
2
projects/Benchmark/BenchMain.cpp
Normal file
@ -0,0 +1,2 @@
|
||||
#define CATCH_CONFIG_MAIN
|
||||
#include "catch.hpp"
|
39
projects/Benchmark/StringificationBench.cpp
Normal file
39
projects/Benchmark/StringificationBench.cpp
Normal file
@ -0,0 +1,39 @@
|
||||
#include "catch.hpp"
|
||||
|
||||
#include <vector>
|
||||
|
||||
///////////////////////////////////////////////////////////////////////////////
|
||||
TEST_CASE("Successful tests -- REQUIRE", "[Success]") {
|
||||
const size_t sz = 1 * 1024 * 1024;
|
||||
|
||||
|
||||
std::vector<size_t> vec; vec.reserve(sz);
|
||||
for (size_t i = 0; i < sz; ++i){
|
||||
vec.push_back(i);
|
||||
REQUIRE(vec.back() == i);
|
||||
}
|
||||
}
|
||||
|
||||
///////////////////////////////////////////////////////////////////////////////
|
||||
TEST_CASE("Successful tests -- CHECK", "[Success]") {
|
||||
const size_t sz = 1 * 1024 * 1024;
|
||||
|
||||
|
||||
std::vector<size_t> vec; vec.reserve(sz);
|
||||
for (size_t i = 0; i < sz; ++i){
|
||||
vec.push_back(i);
|
||||
CHECK(vec.back() == i);
|
||||
}
|
||||
}
|
||||
|
||||
///////////////////////////////////////////////////////////////////////////////
|
||||
TEST_CASE("Unsuccessful tests -- CHECK", "[Failure]") {
|
||||
const size_t sz = 128 * 1024;
|
||||
|
||||
|
||||
std::vector<size_t> vec; vec.reserve(sz);
|
||||
for (size_t i = 0; i < sz; ++i){
|
||||
vec.push_back(i);
|
||||
CHECK(vec.size() == i);
|
||||
}
|
||||
}
|
4
projects/Benchmark/readme.txt
Normal file
4
projects/Benchmark/readme.txt
Normal file
@ -0,0 +1,4 @@
|
||||
This is very much a work in progress.
|
||||
The past results are standardized to a developer's machine,
|
||||
the benchmarking script is basic and there are only 3 benchmarks,
|
||||
but this should get better in time. For now, at least there is something to go by.
|
@ -0,0 +1,3 @@
|
||||
Successful tests -- CHECK: median: 3.38116 (s), stddev: 0.11567366292001534 (s)
|
||||
Successful tests -- REQUIRE: median: 3.479955 (s), stddev: 0.16295972890734556 (s)
|
||||
Unsuccessful tests -- CHECK: median: 1.966895 (s), stddev: 0.06323488524716572 (s)
|
56
scripts/runner.py
Normal file
56
scripts/runner.py
Normal file
@ -0,0 +1,56 @@
|
||||
#!/usr/bin/env python3
|
||||
|
||||
import subprocess, os, sys
|
||||
import xml.etree.ElementTree as ET
|
||||
from collections import defaultdict
|
||||
from statistics import median, stdev
|
||||
from datetime import datetime
|
||||
|
||||
def get_commit_hash():
|
||||
res = subprocess.run('git rev-parse HEAD'.split(), check=True, stdout=subprocess.PIPE, universal_newlines=True)
|
||||
return res.stdout.strip()
|
||||
|
||||
if len(sys.argv) < 2:
|
||||
print('Usage: {} benchmark-binary'.format(sys.argv[0]))
|
||||
exit(1)
|
||||
|
||||
|
||||
num_runs = 10
|
||||
data = defaultdict(list)
|
||||
|
||||
|
||||
def parse_file(file):
|
||||
|
||||
def recursive_search(node):
|
||||
if node.tag == 'TestCase':
|
||||
results = node.find('OverallResult')
|
||||
time = results.get('durationInSeconds')
|
||||
data[node.get('name')].append(float(time))
|
||||
elif node.tag in ('Group', 'Catch'):
|
||||
for child in node:
|
||||
recursive_search(child)
|
||||
|
||||
tree = ET.parse(file)
|
||||
recursive_search(tree.getroot())
|
||||
|
||||
def run_benchmarks(binary):
|
||||
call = [binary] + '-d yes -r xml -o'.split()
|
||||
for i in range(num_runs):
|
||||
file = 'temp{}.xml'.format(i)
|
||||
print('Run number {}'.format(i))
|
||||
subprocess.run(call + [file])
|
||||
parse_file(file)
|
||||
# Remove file right after parsing, because benchmark output can be big
|
||||
os.remove(file)
|
||||
|
||||
|
||||
# Run benchmarks
|
||||
run_benchmarks(sys.argv[1])
|
||||
|
||||
result_file = '{:%Y-%m-%dT%H-%M-%S}-{}.result'.format(datetime.now(), get_commit_hash())
|
||||
|
||||
|
||||
print('Writing results to {}'.format(result_file))
|
||||
with open(result_file, 'w') as file:
|
||||
for k in sorted(data):
|
||||
file.write('{}: median: {} (s), stddev: {} (s)\n'.format(k, median(data[k]), stdev(data[k])))
|
Loading…
Reference in New Issue
Block a user