mirror of
https://github.com/catchorg/Catch2.git
synced 2024-11-20 12:26:10 +01:00
78e33ce51f
Test case hashing includes tags and class name As the hasher involves more code now, it was split out into its own file and it got its own set of tests. Closes #2304
517 lines
21 KiB
CMake
517 lines
21 KiB
CMake
include(CatchMiscFunctions)
|
|
|
|
if (CATCH_BUILD_SURROGATES)
|
|
message(STATUS "Configuring targets for surrogate TUs")
|
|
|
|
# If the folder does not exist before we ask for output redirect to
|
|
# a file, it won't work.
|
|
file(MAKE_DIRECTORY ${CMAKE_CURRENT_BINARY_DIR}/surrogates)
|
|
|
|
# Creates target to generate the surrogate TU for provided header.
|
|
# Returns the path to the generated file.
|
|
function(createSurrogateFileTarget sourceHeader pathToFile)
|
|
set(pathPrefix ${PROJECT_SOURCE_DIR}/src)
|
|
|
|
file(RELATIVE_PATH includePath ${pathPrefix} ${sourceHeader})
|
|
|
|
get_filename_component(basicFileName "${sourceHeader}" NAME_WE)
|
|
|
|
set(surrogateFilePath ${CMAKE_CURRENT_BINARY_DIR}/surrogates/surrogate_${basicFileName}.cpp)
|
|
|
|
add_custom_command(
|
|
OUTPUT ${surrogateFilePath}
|
|
COMMAND cmake -E echo "\#include <${includePath}>" > "${surrogateFilePath}"
|
|
VERBATIM
|
|
)
|
|
|
|
set(${pathToFile} ${surrogateFilePath} PARENT_SCOPE)
|
|
endfunction()
|
|
|
|
# Extracts all non-helper (e.g. catch_all.hpp) headers from the
|
|
# Catch2 target, and returns them through the argument.
|
|
function(ExtractCatch2Headers OutArg)
|
|
get_target_property(targetSources Catch2 SOURCES)
|
|
foreach(Source ${targetSources})
|
|
string(REGEX MATCH "^.*\\.hpp$" isHeader ${Source})
|
|
string(REGEX MATCH "_all.hpp" isAllHeader ${Source})
|
|
if(isHeader AND NOT isAllHeader)
|
|
list(APPEND AllHeaders ${Source})
|
|
endif()
|
|
endforeach()
|
|
set(${OutArg} ${AllHeaders} PARENT_SCOPE)
|
|
endfunction()
|
|
|
|
|
|
ExtractCatch2Headers(mainHeaders)
|
|
|
|
if (NOT mainHeaders)
|
|
message(FATAL_ERROR "No headers in the main target were detected. Something is broken.")
|
|
endif()
|
|
|
|
foreach(header ${mainHeaders})
|
|
createSurrogateFileTarget(${header} pathToGeneratedFile)
|
|
list(APPEND surrogateFiles ${pathToGeneratedFile})
|
|
endforeach()
|
|
|
|
|
|
add_executable(Catch2SurrogateTarget
|
|
${surrogateFiles}
|
|
)
|
|
target_link_libraries(Catch2SurrogateTarget PRIVATE Catch2WithMain)
|
|
|
|
endif(CATCH_BUILD_SURROGATES)
|
|
|
|
####
|
|
# Temporary workaround for VS toolset changes in 2017
|
|
# We need to disable <UseFullPaths> property, but CMake doesn't support it
|
|
# until 3.13 (not yet released)
|
|
####
|
|
if (MSVC)
|
|
configure_file(${CATCH_DIR}/tools/misc/SelfTest.vcxproj.user
|
|
${CMAKE_BINARY_DIR}/tests
|
|
COPYONLY)
|
|
endif(MSVC) #Temporary workaround
|
|
|
|
|
|
# define the sources of the self test
|
|
# Please keep these ordered alphabetically
|
|
set(TEST_SOURCES
|
|
${SELF_TEST_DIR}/TestRegistrations.cpp
|
|
${SELF_TEST_DIR}/IntrospectiveTests/Clara.tests.cpp
|
|
${SELF_TEST_DIR}/IntrospectiveTests/CmdLine.tests.cpp
|
|
${SELF_TEST_DIR}/IntrospectiveTests/ColourImpl.tests.cpp
|
|
${SELF_TEST_DIR}/IntrospectiveTests/Details.tests.cpp
|
|
${SELF_TEST_DIR}/IntrospectiveTests/FloatingPoint.tests.cpp
|
|
${SELF_TEST_DIR}/IntrospectiveTests/GeneratorsImpl.tests.cpp
|
|
${SELF_TEST_DIR}/IntrospectiveTests/InternalBenchmark.tests.cpp
|
|
${SELF_TEST_DIR}/IntrospectiveTests/PartTracker.tests.cpp
|
|
${SELF_TEST_DIR}/IntrospectiveTests/RandomNumberGeneration.tests.cpp
|
|
${SELF_TEST_DIR}/IntrospectiveTests/Reporters.tests.cpp
|
|
${SELF_TEST_DIR}/IntrospectiveTests/Tag.tests.cpp
|
|
${SELF_TEST_DIR}/IntrospectiveTests/TestCaseInfoHasher.tests.cpp
|
|
${SELF_TEST_DIR}/IntrospectiveTests/TestSpecParser.tests.cpp
|
|
${SELF_TEST_DIR}/IntrospectiveTests/TextFlow.tests.cpp
|
|
${SELF_TEST_DIR}/IntrospectiveTests/Sharding.tests.cpp
|
|
${SELF_TEST_DIR}/IntrospectiveTests/Stream.tests.cpp
|
|
${SELF_TEST_DIR}/IntrospectiveTests/String.tests.cpp
|
|
${SELF_TEST_DIR}/IntrospectiveTests/StringManip.tests.cpp
|
|
${SELF_TEST_DIR}/IntrospectiveTests/Xml.tests.cpp
|
|
${SELF_TEST_DIR}/IntrospectiveTests/ToString.tests.cpp
|
|
${SELF_TEST_DIR}/IntrospectiveTests/UniquePtr.tests.cpp
|
|
${SELF_TEST_DIR}/TimingTests/Sleep.tests.cpp
|
|
${SELF_TEST_DIR}/UsageTests/Approx.tests.cpp
|
|
${SELF_TEST_DIR}/UsageTests/BDD.tests.cpp
|
|
${SELF_TEST_DIR}/UsageTests/Benchmark.tests.cpp
|
|
${SELF_TEST_DIR}/UsageTests/Class.tests.cpp
|
|
${SELF_TEST_DIR}/UsageTests/Compilation.tests.cpp
|
|
${SELF_TEST_DIR}/UsageTests/Condition.tests.cpp
|
|
${SELF_TEST_DIR}/UsageTests/Decomposition.tests.cpp
|
|
${SELF_TEST_DIR}/UsageTests/EnumToString.tests.cpp
|
|
${SELF_TEST_DIR}/UsageTests/Exception.tests.cpp
|
|
${SELF_TEST_DIR}/UsageTests/Generators.tests.cpp
|
|
${SELF_TEST_DIR}/UsageTests/Message.tests.cpp
|
|
${SELF_TEST_DIR}/UsageTests/Misc.tests.cpp
|
|
${SELF_TEST_DIR}/UsageTests/ToStringByte.tests.cpp
|
|
${SELF_TEST_DIR}/UsageTests/ToStringChrono.tests.cpp
|
|
${SELF_TEST_DIR}/UsageTests/ToStringGeneral.tests.cpp
|
|
${SELF_TEST_DIR}/UsageTests/ToStringOptional.tests.cpp
|
|
${SELF_TEST_DIR}/UsageTests/ToStringPair.tests.cpp
|
|
${SELF_TEST_DIR}/UsageTests/ToStringTuple.tests.cpp
|
|
${SELF_TEST_DIR}/UsageTests/ToStringVariant.tests.cpp
|
|
${SELF_TEST_DIR}/UsageTests/ToStringVector.tests.cpp
|
|
${SELF_TEST_DIR}/UsageTests/ToStringWhich.tests.cpp
|
|
${SELF_TEST_DIR}/UsageTests/Tricky.tests.cpp
|
|
${SELF_TEST_DIR}/UsageTests/VariadicMacros.tests.cpp
|
|
${SELF_TEST_DIR}/UsageTests/MatchersRanges.tests.cpp
|
|
${SELF_TEST_DIR}/UsageTests/Matchers.tests.cpp
|
|
)
|
|
|
|
|
|
# Specify the headers, too, so CLion recognises them as project files
|
|
set(HEADERS
|
|
${TOP_LEVEL_HEADERS}
|
|
${EXTERNAL_HEADERS}
|
|
${INTERNAL_HEADERS}
|
|
${REPORTER_HEADERS}
|
|
${BENCHMARK_HEADERS}
|
|
${BENCHMARK_SOURCES}
|
|
)
|
|
|
|
# Provide some groupings for IDEs
|
|
#SOURCE_GROUP("benchmark" FILES ${BENCHMARK_HEADERS} ${BENCHMARK_SOURCES})
|
|
#SOURCE_GROUP("Tests" FILES ${TEST_SOURCES})
|
|
|
|
include(CTest)
|
|
|
|
add_executable(SelfTest ${TEST_SOURCES})
|
|
target_link_libraries(SelfTest PRIVATE Catch2WithMain)
|
|
|
|
if (CATCH_ENABLE_COVERAGE)
|
|
set(ENABLE_COVERAGE ON CACHE BOOL "Enable coverage build." FORCE)
|
|
find_package(codecov)
|
|
add_coverage(SelfTest)
|
|
list(APPEND LCOV_REMOVE_PATTERNS "'/usr/*'")
|
|
coverage_evaluate()
|
|
endif()
|
|
|
|
# configure unit tests via CTest
|
|
add_test(NAME RunTests COMMAND $<TARGET_FILE:SelfTest> --order rand --rng-seed time)
|
|
set_tests_properties(RunTests PROPERTIES
|
|
FAIL_REGULAR_EXPRESSION "Filters:"
|
|
COST 60
|
|
)
|
|
|
|
# Because CTest does not allow us to check both return code _and_ expected
|
|
# output in one test, we run these commands twice. First time we check
|
|
# the output, the second time we check the exit code.
|
|
add_test(NAME List::Tests::Output COMMAND $<TARGET_FILE:SelfTest> --list-tests --verbosity high)
|
|
set_tests_properties(List::Tests::Output PROPERTIES
|
|
PASS_REGULAR_EXPRESSION "[0-9]+ test cases"
|
|
FAIL_REGULAR_EXPRESSION "Hidden Test"
|
|
)
|
|
# This should be equivalent to the old --list-test-names-only and be usable
|
|
# with --input-file.
|
|
add_test(NAME List::Tests::Quiet COMMAND $<TARGET_FILE:SelfTest> --list-tests --verbosity quiet)
|
|
# Sadly we cannot ask for start-of-line and end-of-line in a ctest regex,
|
|
# so we fail if we see space/tab at the start...
|
|
set_tests_properties(List::Tests::Quiet PROPERTIES
|
|
PASS_REGULAR_EXPRESSION "\"#1905 -- test spec parser properly clears internal state between compound tests\"[\r\n]"
|
|
FAIL_REGULAR_EXPRESSION "[ \t]\"#1905 -- test spec parser properly clears internal state between compound tests\""
|
|
)
|
|
add_test(NAME List::Tests::ExitCode COMMAND $<TARGET_FILE:SelfTest> --list-tests --verbosity high)
|
|
add_test(NAME List::Tests::XmlOutput COMMAND $<TARGET_FILE:SelfTest> --list-tests --verbosity high -r xml)
|
|
set_tests_properties(List::Tests::XmlOutput PROPERTIES
|
|
PASS_REGULAR_EXPRESSION "<Line>[0-9]+</Line>"
|
|
FAIL_REGULAR_EXPRESSION "[0-9]+ test cases"
|
|
)
|
|
|
|
add_test(NAME List::Tags::Output COMMAND $<TARGET_FILE:SelfTest> --list-tags)
|
|
set_tests_properties(List::Tags::Output PROPERTIES
|
|
PASS_REGULAR_EXPRESSION "[0-9]+ tags"
|
|
FAIL_REGULAR_EXPRESSION "\\[\\.\\]")
|
|
add_test(NAME List::Tags::ExitCode COMMAND $<TARGET_FILE:SelfTest> --list-tags)
|
|
add_test(NAME List::Tags::XmlOutput COMMAND $<TARGET_FILE:SelfTest> --list-tags -r xml)
|
|
set_tests_properties(List::Tags::XmlOutput PROPERTIES
|
|
PASS_REGULAR_EXPRESSION "<Count>18</Count>"
|
|
FAIL_REGULAR_EXPRESSION "[0-9]+ tags"
|
|
)
|
|
|
|
|
|
add_test(NAME List::Reporters::Output COMMAND $<TARGET_FILE:SelfTest> --list-reporters)
|
|
set_tests_properties(List::Reporters::Output PROPERTIES PASS_REGULAR_EXPRESSION "Available reporters:")
|
|
add_test(NAME List::Reporters::ExitCode COMMAND $<TARGET_FILE:SelfTest> --list-reporters)
|
|
add_test(NAME List::Reporters::XmlOutput COMMAND $<TARGET_FILE:SelfTest> --list-reporters -r xml)
|
|
set_tests_properties(List::Reporters::XmlOutput PROPERTIES
|
|
PASS_REGULAR_EXPRESSION "<Name>compact</Name>"
|
|
FAIL_REGULAR_EXPRESSION "Available reporters:"
|
|
)
|
|
|
|
|
|
|
|
add_test(NAME NoAssertions COMMAND $<TARGET_FILE:SelfTest> -w NoAssertions "An empty test with no assertions")
|
|
set_tests_properties(NoAssertions PROPERTIES PASS_REGULAR_EXPRESSION "No assertions in test case")
|
|
|
|
# We cannot combine a regular expression on output with return code check
|
|
# in one test, so we register two instead of making a checking script because
|
|
# the runtime overhead is small enough.
|
|
add_test(NAME TestSpecs::CombiningMatchingAndNonMatchingIsOk-1 COMMAND $<TARGET_FILE:SelfTest> Tracker, "___nonexistent_test___")
|
|
|
|
add_test(NAME TestSpecs::CombiningMatchingAndNonMatchingIsOk-2 COMMAND $<TARGET_FILE:SelfTest> Tracker, "___nonexistent_test___")
|
|
set_tests_properties(TestSpecs::CombiningMatchingAndNonMatchingIsOk-2 PROPERTIES
|
|
PASS_REGULAR_EXPRESSION "No test cases matched '___nonexistent_test___'"
|
|
FAIL_REGULAR_EXPRESSION "No tests ran"
|
|
)
|
|
|
|
add_test(NAME TestSpecs::NoMatchedTestsFail
|
|
COMMAND $<TARGET_FILE:SelfTest> "___nonexistent_test___"
|
|
)
|
|
set_tests_properties(TestSpecs::NoMatchedTestsFail
|
|
PROPERTIES
|
|
WILL_FAIL ON
|
|
)
|
|
add_test(NAME TestSpecs::OverrideFailureWithNoMatchedTests
|
|
COMMAND $<TARGET_FILE:SelfTest> "___nonexistent_test___" --allow-running-no-tests
|
|
)
|
|
|
|
add_test(NAME Warnings::UnmatchedTestSpecIsAccepted
|
|
COMMAND $<TARGET_FILE:SelfTest> Tracker --warn UnmatchedTestSpec
|
|
)
|
|
set_tests_properties(Warnings::UnmatchedTestSpecIsAccepted
|
|
PROPERTIES
|
|
FAIL_REGULAR_EXPRESSION "Unrecognised warning option: "
|
|
)
|
|
|
|
add_test(NAME Warnings::MultipleWarningsCanBeSpecified
|
|
COMMAND
|
|
$<TARGET_FILE:SelfTest> Tracker
|
|
--warn NoAssertions
|
|
--warn UnmatchedTestSpec
|
|
)
|
|
|
|
add_test(NAME TestSpecs::WarnUnmatchedTestSpecFailsWithUnmatchedTestSpec
|
|
COMMAND
|
|
$<TARGET_FILE:SelfTest> Tracker, "___nonexistent_test___" --warn UnmatchedTestSpec
|
|
)
|
|
set_tests_properties(TestSpecs::WarnUnmatchedTestSpecFailsWithUnmatchedTestSpec
|
|
PROPERTIES
|
|
WILL_FAIL ON
|
|
)
|
|
|
|
add_test(NAME UnmatchedOutputFilter COMMAND $<TARGET_FILE:SelfTest> [this-tag-does-not-exist])
|
|
set_tests_properties(UnmatchedOutputFilter
|
|
PROPERTIES
|
|
PASS_REGULAR_EXPRESSION "No test cases matched '\\[this-tag-does-not-exist\\]'"
|
|
)
|
|
|
|
add_test(NAME FilteredSection-1 COMMAND $<TARGET_FILE:SelfTest> \#1394 -c RunSection)
|
|
set_tests_properties(FilteredSection-1 PROPERTIES FAIL_REGULAR_EXPRESSION "No tests ran")
|
|
add_test(NAME FilteredSection-2 COMMAND $<TARGET_FILE:SelfTest> \#1394\ nested -c NestedRunSection -c s1)
|
|
set_tests_properties(FilteredSection-2 PROPERTIES FAIL_REGULAR_EXPRESSION "No tests ran")
|
|
|
|
add_test(
|
|
NAME
|
|
FilteredSection::GeneratorsDontCauseInfiniteLoop-1
|
|
COMMAND
|
|
$<TARGET_FILE:SelfTest> "#2025: original repro" -c "fov_0"
|
|
)
|
|
set_tests_properties(FilteredSection::GeneratorsDontCauseInfiniteLoop-1
|
|
PROPERTIES
|
|
PASS_REGULAR_EXPRESSION "inside with fov: 0" # This should happen
|
|
FAIL_REGULAR_EXPRESSION "inside with fov: 1" # This would mean there was no filtering
|
|
)
|
|
|
|
# GENERATE between filtered sections (both are selected)
|
|
add_test(
|
|
NAME
|
|
FilteredSection::GeneratorsDontCauseInfiniteLoop-2
|
|
COMMAND
|
|
$<TARGET_FILE:SelfTest> "#2025: same-level sections"
|
|
-c "A"
|
|
-c "B"
|
|
)
|
|
set_tests_properties(FilteredSection::GeneratorsDontCauseInfiniteLoop-2
|
|
PROPERTIES
|
|
PASS_REGULAR_EXPRESSION "All tests passed \\(4 assertions in 1 test case\\)"
|
|
)
|
|
|
|
# AppVeyor has a Python 2.7 in path, but doesn't have .py files as autorunnable
|
|
add_test(NAME ApprovalTests COMMAND ${PYTHON_EXECUTABLE} ${CATCH_DIR}/tools/scripts/approvalTests.py $<TARGET_FILE:SelfTest>)
|
|
set_tests_properties(ApprovalTests
|
|
PROPERTIES
|
|
FAIL_REGULAR_EXPRESSION "Results differed"
|
|
COST 120 # We know that this is either the most, or second most,
|
|
# expensive test in the test suite, so we give it high estimate for CI runs
|
|
)
|
|
|
|
add_test(NAME RegressionCheck-1670 COMMAND $<TARGET_FILE:SelfTest> "#1670 regression check" -c A -r compact)
|
|
set_tests_properties(RegressionCheck-1670 PROPERTIES PASS_REGULAR_EXPRESSION "Passed 1 test case with 2 assertions.")
|
|
|
|
add_test(NAME VersionCheck COMMAND $<TARGET_FILE:SelfTest> -h)
|
|
set_tests_properties(VersionCheck PROPERTIES PASS_REGULAR_EXPRESSION "Catch2 v${PROJECT_VERSION}")
|
|
|
|
add_test(NAME LibIdentityTest COMMAND $<TARGET_FILE:SelfTest> --libidentify)
|
|
set_tests_properties(LibIdentityTest PROPERTIES PASS_REGULAR_EXPRESSION "description: A Catch2 test executable")
|
|
|
|
add_test(NAME FilenameAsTagsTest COMMAND $<TARGET_FILE:SelfTest> -\# --list-tags)
|
|
set_tests_properties(FilenameAsTagsTest PROPERTIES PASS_REGULAR_EXPRESSION "\\[#Approx.tests\\]")
|
|
|
|
# Check that the filename tags can also be matched against (#2064)
|
|
add_test(NAME FilenameAsTagsMatching COMMAND $<TARGET_FILE:SelfTest> -\# --list-tags [\#Approx.tests])
|
|
set_tests_properties(FilenameAsTagsMatching
|
|
PROPERTIES
|
|
PASS_REGULAR_EXPRESSION "\\[#Approx.tests\\]"
|
|
# Avoids false positives by looking for start of line (newline) before the 0
|
|
FAIL_REGULAR_EXPRESSION "[\r\n]0 tag"
|
|
)
|
|
|
|
add_test(NAME EscapeSpecialCharactersInTestNames COMMAND $<TARGET_FILE:SelfTest> "Test with special\\, characters \"in name")
|
|
set_tests_properties(EscapeSpecialCharactersInTestNames PROPERTIES PASS_REGULAR_EXPRESSION "1 assertion in 1 test case")
|
|
|
|
add_test(NAME NegativeSpecNoHiddenTests COMMAND $<TARGET_FILE:SelfTest> --list-tests ~[approval])
|
|
set_tests_properties(NegativeSpecNoHiddenTests PROPERTIES FAIL_REGULAR_EXPRESSION "\\[\\.\\]")
|
|
|
|
add_test(NAME TestsInFile::SimpleSpecs COMMAND $<TARGET_FILE:SelfTest> "-f ${SELF_TEST_DIR}/Misc/plain-old-tests.input")
|
|
set_tests_properties(TestsInFile::SimpleSpecs PROPERTIES PASS_REGULAR_EXPRESSION "6 assertions in 2 test cases")
|
|
|
|
add_test(NAME TestsInFile::EscapeSpecialCharacters COMMAND $<TARGET_FILE:SelfTest> "-f ${SELF_TEST_DIR}/Misc/special-characters-in-file.input")
|
|
set_tests_properties(TestsInFile::EscapeSpecialCharacters PROPERTIES PASS_REGULAR_EXPRESSION "1 assertion in 1 test case")
|
|
|
|
add_test(NAME TestsInFile::InvalidTestNames-1 COMMAND $<TARGET_FILE:SelfTest> "-f ${SELF_TEST_DIR}/Misc/invalid-test-names.input")
|
|
set_tests_properties(TestsInFile::InvalidTestNames-1
|
|
PROPERTIES
|
|
PASS_REGULAR_EXPRESSION "Invalid Filter: \"Test with special, characters in \\\\\" name\""
|
|
FAIL_REGULAR_EXPRESSION "No tests ran"
|
|
)
|
|
|
|
add_test(NAME TagAlias COMMAND $<TARGET_FILE:SelfTest> [@tricky] --list-tests)
|
|
set_tests_properties(TagAlias PROPERTIES
|
|
PASS_REGULAR_EXPRESSION "[0-9]+ matching test cases"
|
|
FAIL_REGULAR_EXPRESSION "0 matching test cases"
|
|
)
|
|
|
|
add_test(NAME RandomTestOrdering COMMAND ${PYTHON_EXECUTABLE}
|
|
${CATCH_DIR}/tests/TestScripts/testRandomOrder.py $<TARGET_FILE:SelfTest>)
|
|
|
|
|
|
add_test(NAME CheckConvenienceHeaders
|
|
COMMAND
|
|
${PYTHON_EXECUTABLE} ${CATCH_DIR}/tools/scripts/checkConvenienceHeaders.py
|
|
)
|
|
|
|
|
|
add_test(NAME "Benchmarking::FailureReporting::OptimizedOut"
|
|
COMMAND
|
|
$<TARGET_FILE:SelfTest> "Failing benchmarks" -c "empty" -r xml
|
|
# This test only makes sense with the optimizer being enabled when
|
|
# the tests are being compiled.
|
|
CONFIGURATIONS Release
|
|
)
|
|
set_tests_properties("Benchmarking::FailureReporting::OptimizedOut"
|
|
PROPERTIES
|
|
PASS_REGULAR_EXPRESSION "could not measure benchmark\, maybe it was optimized away"
|
|
FAIL_REGULAR_EXPRESSION "successes=\"1\""
|
|
)
|
|
|
|
add_test(NAME "Benchmarking::FailureReporting::ThrowingBenchmark"
|
|
COMMAND
|
|
$<TARGET_FILE:SelfTest> "Failing benchmarks" -c "throw" -r xml
|
|
)
|
|
set_tests_properties("Benchmarking::FailureReporting::ThrowingBenchmark"
|
|
PROPERTIES
|
|
PASS_REGULAR_EXPRESSION "<failed message=\"just a plain literal"
|
|
FAIL_REGULAR_EXPRESSION "successes=\"1\""
|
|
)
|
|
|
|
add_test(NAME "Benchmarking::FailureReporting::FailedAssertion"
|
|
COMMAND
|
|
$<TARGET_FILE:SelfTest> "Failing benchmarks" -c "assert" -r xml
|
|
)
|
|
set_tests_properties("Benchmarking::FailureReporting::FailedAssertion"
|
|
PROPERTIES
|
|
PASS_REGULAR_EXPRESSION "<Expression success=\"false\""
|
|
FAIL_REGULAR_EXPRESSION "successes=\"1\""
|
|
)
|
|
|
|
add_test(NAME "Benchmarking::FailureReporting::FailMacro"
|
|
COMMAND
|
|
$<TARGET_FILE:SelfTest> "Failing benchmarks" -c "fail" -r xml
|
|
)
|
|
set_tests_properties("Benchmarking::FailureReporting::FailMacro"
|
|
PROPERTIES
|
|
PASS_REGULAR_EXPRESSION "This benchmark only fails\, nothing else"
|
|
FAIL_REGULAR_EXPRESSION "successes=\"1\""
|
|
)
|
|
|
|
add_test(NAME "Benchmarking::FailureReporting::ShouldFailIsRespected"
|
|
COMMAND
|
|
$<TARGET_FILE:SelfTest> "Failing benchmark respects should-fail"
|
|
)
|
|
set_tests_properties("Benchmarking::FailureReporting::ShouldFailIsRespected"
|
|
PROPERTIES
|
|
PASS_REGULAR_EXPRESSION "1 failed as expected"
|
|
)
|
|
|
|
add_test(NAME "ErrorHandling::InvalidTestSpecExitsEarly"
|
|
COMMAND
|
|
$<TARGET_FILE:SelfTest> "[aa,a]"
|
|
)
|
|
set_tests_properties("ErrorHandling::InvalidTestSpecExitsEarly"
|
|
PROPERTIES
|
|
PASS_REGULAR_EXPRESSION "Invalid Filter: \\[aa\,a\\]"
|
|
FAIL_REGULAR_EXPRESSION "No tests ran"
|
|
)
|
|
|
|
if (MSVC)
|
|
set(_NullFile "NUL")
|
|
else()
|
|
set(_NullFile "/dev/null")
|
|
endif()
|
|
|
|
# This test checks that there is nothing written out from the process,
|
|
# but if CMake is running the tests under Valgrind or similar tool, then
|
|
# that will write its own output to stdout and the test would fail.
|
|
if (NOT MEMORYCHECK_COMMAND)
|
|
add_test(NAME "MultiReporter::CapturingReportersDontPropagateStdOut"
|
|
COMMAND
|
|
$<TARGET_FILE:SelfTest> "Sends stuff to stdout and stderr"
|
|
--reporter xml::${_NullFile}
|
|
--reporter junit::${_NullFile}
|
|
)
|
|
set_tests_properties("MultiReporter::CapturingReportersDontPropagateStdOut"
|
|
PROPERTIES
|
|
FAIL_REGULAR_EXPRESSION ".+"
|
|
)
|
|
endif()
|
|
|
|
add_test(NAME "MultiReporter::NonCapturingReportersPropagateStdout"
|
|
COMMAND
|
|
$<TARGET_FILE:SelfTest> "Sends stuff to stdout and stderr"
|
|
--reporter xml::${_NullFile}
|
|
--reporter console::${_NullFile}
|
|
)
|
|
set_tests_properties("MultiReporter::NonCapturingReportersPropagateStdout"
|
|
PROPERTIES
|
|
PASS_REGULAR_EXPRESSION "A string sent to stderr via clog"
|
|
)
|
|
|
|
add_test(NAME "Outputs::DashAsOutLocationSendsOutputToStdout"
|
|
COMMAND
|
|
$<TARGET_FILE:SelfTest> "Factorials are computed"
|
|
--out=-
|
|
)
|
|
set_tests_properties("Outputs::DashAsOutLocationSendsOutputToStdout"
|
|
PROPERTIES
|
|
PASS_REGULAR_EXPRESSION "All tests passed \\(5 assertions in 1 test case\\)"
|
|
)
|
|
|
|
add_test(NAME "Reporters::DashAsLocationInReporterSpecSendsOutputToStdout"
|
|
COMMAND
|
|
$<TARGET_FILE:SelfTest> "Factorials are computed"
|
|
--reporter console::-
|
|
)
|
|
set_tests_properties("Reporters::DashAsLocationInReporterSpecSendsOutputToStdout"
|
|
PROPERTIES
|
|
PASS_REGULAR_EXPRESSION "All tests passed \\(5 assertions in 1 test case\\)"
|
|
)
|
|
|
|
add_test(NAME "Colours::ColourModeCanBeExplicitlySetToAnsi"
|
|
COMMAND
|
|
$<TARGET_FILE:SelfTest> "Factorials are computed"
|
|
--reporter console
|
|
--colour-mode ansi
|
|
)
|
|
set_tests_properties("Colours::ColourModeCanBeExplicitlySetToAnsi"
|
|
PROPERTIES
|
|
PASS_REGULAR_EXPRESSION "\\[1\;32mAll tests passed"
|
|
)
|
|
|
|
if (CATCH_ENABLE_CONFIGURE_TESTS)
|
|
add_test(NAME "CMakeConfig::DefaultReporter"
|
|
COMMAND
|
|
"${PYTHON_EXECUTABLE}" "${CMAKE_CURRENT_LIST_DIR}/TestScripts/testConfigureDefaultReporter.py" "${CATCH_DIR}" "${CMAKE_CURRENT_BINARY_DIR}"
|
|
)
|
|
add_test(NAME "CMakeConfig::Disable"
|
|
COMMAND
|
|
"${PYTHON_EXECUTABLE}" "${CMAKE_CURRENT_LIST_DIR}/TestScripts/testConfigureDisable.py" "${CATCH_DIR}" "${CMAKE_CURRENT_BINARY_DIR}"
|
|
)
|
|
add_test(NAME "CMakeConfig::DisableStringification"
|
|
COMMAND
|
|
"${PYTHON_EXECUTABLE}" "${CMAKE_CURRENT_LIST_DIR}/TestScripts/testConfigureDisableStringification.py" "${CATCH_DIR}" "${CMAKE_CURRENT_BINARY_DIR}"
|
|
)
|
|
add_test(NAME "CMakeConfig::ExperimentalRedirect"
|
|
COMMAND
|
|
"${PYTHON_EXECUTABLE}" "${CMAKE_CURRENT_LIST_DIR}/TestScripts/testConfigureExperimentalRedirect.py" "${CATCH_DIR}" "${CMAKE_CURRENT_BINARY_DIR}"
|
|
)
|
|
endif()
|
|
|
|
if (CATCH_USE_VALGRIND)
|
|
add_test(NAME ValgrindRunTests COMMAND valgrind --leak-check=full --error-exitcode=1 $<TARGET_FILE:SelfTest>)
|
|
add_test(NAME ValgrindListTests COMMAND valgrind --leak-check=full --error-exitcode=1 $<TARGET_FILE:SelfTest> --list-tests --verbosity high)
|
|
set_tests_properties(ValgrindListTests PROPERTIES PASS_REGULAR_EXPRESSION "definitely lost: 0 bytes in 0 blocks")
|
|
add_test(NAME ValgrindListTags COMMAND valgrind --leak-check=full --error-exitcode=1 $<TARGET_FILE:SelfTest> --list-tags)
|
|
set_tests_properties(ValgrindListTags PROPERTIES PASS_REGULAR_EXPRESSION "definitely lost: 0 bytes in 0 blocks")
|
|
endif()
|
|
|
|
list(APPEND CATCH_WARNING_TARGETS SelfTest)
|
|
set(CATCH_WARNING_TARGETS ${CATCH_WARNING_TARGETS} PARENT_SCOPE)
|