2022-02-07 00:05:11 +01:00
|
|
|
include(CatchMiscFunctions)
|
2018-06-11 10:48:10 +02:00
|
|
|
|
2021-02-15 23:56:56 +01:00
|
|
|
if (CATCH_BUILD_SURROGATES)
|
|
|
|
message(STATUS "Configuring targets for surrogate TUs")
|
|
|
|
|
|
|
|
# If the folder does not exist before we ask for output redirect to
|
|
|
|
# a file, it won't work.
|
|
|
|
file(MAKE_DIRECTORY ${CMAKE_CURRENT_BINARY_DIR}/surrogates)
|
|
|
|
|
|
|
|
# Creates target to generate the surrogate TU for provided header.
|
|
|
|
# Returns the path to the generated file.
|
|
|
|
function(createSurrogateFileTarget sourceHeader pathToFile)
|
|
|
|
set(pathPrefix ${PROJECT_SOURCE_DIR}/src)
|
|
|
|
|
|
|
|
file(RELATIVE_PATH includePath ${pathPrefix} ${sourceHeader})
|
|
|
|
|
|
|
|
get_filename_component(basicFileName "${sourceHeader}" NAME_WE)
|
|
|
|
|
|
|
|
set(surrogateFilePath ${CMAKE_CURRENT_BINARY_DIR}/surrogates/surrogate_${basicFileName}.cpp)
|
|
|
|
|
|
|
|
add_custom_command(
|
|
|
|
OUTPUT ${surrogateFilePath}
|
|
|
|
COMMAND cmake -E echo "\#include <${includePath}>" > "${surrogateFilePath}"
|
|
|
|
VERBATIM
|
|
|
|
)
|
|
|
|
|
|
|
|
set(${pathToFile} ${surrogateFilePath} PARENT_SCOPE)
|
|
|
|
endfunction()
|
|
|
|
|
|
|
|
# Extracts all non-helper (e.g. catch_all.hpp) headers from the
|
|
|
|
# Catch2 target, and returns them through the argument.
|
|
|
|
function(ExtractCatch2Headers OutArg)
|
|
|
|
get_target_property(targetSources Catch2 SOURCES)
|
|
|
|
foreach(Source ${targetSources})
|
|
|
|
string(REGEX MATCH "^.*\\.hpp$" isHeader ${Source})
|
|
|
|
string(REGEX MATCH "_all.hpp" isAllHeader ${Source})
|
|
|
|
if(isHeader AND NOT isAllHeader)
|
|
|
|
list(APPEND AllHeaders ${Source})
|
|
|
|
endif()
|
|
|
|
endforeach()
|
|
|
|
set(${OutArg} ${AllHeaders} PARENT_SCOPE)
|
|
|
|
endfunction()
|
|
|
|
|
|
|
|
|
|
|
|
ExtractCatch2Headers(mainHeaders)
|
|
|
|
|
|
|
|
if (NOT mainHeaders)
|
|
|
|
message(FATAL_ERROR "No headers in the main target were detected. Something is broken.")
|
|
|
|
endif()
|
|
|
|
|
|
|
|
foreach(header ${mainHeaders})
|
|
|
|
createSurrogateFileTarget(${header} pathToGeneratedFile)
|
|
|
|
list(APPEND surrogateFiles ${pathToGeneratedFile})
|
|
|
|
endforeach()
|
|
|
|
|
|
|
|
|
|
|
|
add_executable(Catch2SurrogateTarget
|
|
|
|
${surrogateFiles}
|
|
|
|
)
|
|
|
|
target_link_libraries(Catch2SurrogateTarget PRIVATE Catch2WithMain)
|
|
|
|
|
|
|
|
endif(CATCH_BUILD_SURROGATES)
|
|
|
|
|
2018-08-28 10:48:05 +02:00
|
|
|
####
|
|
|
|
# Temporary workaround for VS toolset changes in 2017
|
|
|
|
# We need to disable <UseFullPaths> property, but CMake doesn't support it
|
|
|
|
# until 3.13 (not yet released)
|
|
|
|
####
|
|
|
|
if (MSVC)
|
2019-12-06 11:40:53 +01:00
|
|
|
configure_file(${CATCH_DIR}/tools/misc/SelfTest.vcxproj.user
|
2019-12-05 16:00:20 +01:00
|
|
|
${CMAKE_BINARY_DIR}/tests
|
2018-08-28 10:48:05 +02:00
|
|
|
COPYONLY)
|
|
|
|
endif(MSVC) #Temporary workaround
|
|
|
|
|
|
|
|
|
2018-06-11 10:48:10 +02:00
|
|
|
# define the sources of the self test
|
|
|
|
# Please keep these ordered alphabetically
|
|
|
|
set(TEST_SOURCES
|
2020-05-07 16:24:05 +02:00
|
|
|
${SELF_TEST_DIR}/TestRegistrations.cpp
|
2023-02-05 16:40:38 +01:00
|
|
|
${SELF_TEST_DIR}/IntrospectiveTests/Algorithms.tests.cpp
|
2020-11-07 17:51:50 +01:00
|
|
|
${SELF_TEST_DIR}/IntrospectiveTests/Clara.tests.cpp
|
2018-06-11 10:48:10 +02:00
|
|
|
${SELF_TEST_DIR}/IntrospectiveTests/CmdLine.tests.cpp
|
2022-04-04 11:54:02 +02:00
|
|
|
${SELF_TEST_DIR}/IntrospectiveTests/CmdLineHelpers.tests.cpp
|
2022-03-24 00:09:27 +01:00
|
|
|
${SELF_TEST_DIR}/IntrospectiveTests/ColourImpl.tests.cpp
|
2019-08-09 10:50:53 +02:00
|
|
|
${SELF_TEST_DIR}/IntrospectiveTests/Details.tests.cpp
|
2021-07-26 22:01:04 +02:00
|
|
|
${SELF_TEST_DIR}/IntrospectiveTests/FloatingPoint.tests.cpp
|
2018-06-12 18:50:39 +02:00
|
|
|
${SELF_TEST_DIR}/IntrospectiveTests/GeneratorsImpl.tests.cpp
|
2019-04-23 23:41:13 +02:00
|
|
|
${SELF_TEST_DIR}/IntrospectiveTests/InternalBenchmark.tests.cpp
|
2022-10-20 15:06:26 +02:00
|
|
|
${SELF_TEST_DIR}/IntrospectiveTests/Parse.tests.cpp
|
2018-06-11 10:48:10 +02:00
|
|
|
${SELF_TEST_DIR}/IntrospectiveTests/PartTracker.tests.cpp
|
2019-10-06 21:47:54 +02:00
|
|
|
${SELF_TEST_DIR}/IntrospectiveTests/RandomNumberGeneration.tests.cpp
|
2021-01-26 17:53:59 +01:00
|
|
|
${SELF_TEST_DIR}/IntrospectiveTests/Reporters.tests.cpp
|
2019-03-29 10:48:56 +01:00
|
|
|
${SELF_TEST_DIR}/IntrospectiveTests/Tag.tests.cpp
|
2022-03-31 16:46:41 +02:00
|
|
|
${SELF_TEST_DIR}/IntrospectiveTests/TestCaseInfoHasher.tests.cpp
|
2022-10-26 21:03:49 +02:00
|
|
|
${SELF_TEST_DIR}/IntrospectiveTests/TestSpec.tests.cpp
|
2021-12-27 10:04:48 +01:00
|
|
|
${SELF_TEST_DIR}/IntrospectiveTests/TestSpecParser.tests.cpp
|
2021-11-01 22:51:17 +01:00
|
|
|
${SELF_TEST_DIR}/IntrospectiveTests/TextFlow.tests.cpp
|
2021-07-11 21:46:05 +02:00
|
|
|
${SELF_TEST_DIR}/IntrospectiveTests/Sharding.tests.cpp
|
2022-02-21 23:48:15 +01:00
|
|
|
${SELF_TEST_DIR}/IntrospectiveTests/Stream.tests.cpp
|
2018-06-11 10:48:10 +02:00
|
|
|
${SELF_TEST_DIR}/IntrospectiveTests/String.tests.cpp
|
2019-09-06 18:50:57 +02:00
|
|
|
${SELF_TEST_DIR}/IntrospectiveTests/StringManip.tests.cpp
|
2018-06-11 10:48:10 +02:00
|
|
|
${SELF_TEST_DIR}/IntrospectiveTests/Xml.tests.cpp
|
2022-11-01 08:48:23 +01:00
|
|
|
${SELF_TEST_DIR}/IntrospectiveTests/Traits.tests.cpp
|
2019-04-25 11:13:11 +02:00
|
|
|
${SELF_TEST_DIR}/IntrospectiveTests/ToString.tests.cpp
|
2020-05-24 23:41:02 +02:00
|
|
|
${SELF_TEST_DIR}/IntrospectiveTests/UniquePtr.tests.cpp
|
2022-10-28 00:16:38 +02:00
|
|
|
${SELF_TEST_DIR}/helpers/parse_test_spec.cpp
|
2020-04-13 14:34:27 +02:00
|
|
|
${SELF_TEST_DIR}/TimingTests/Sleep.tests.cpp
|
2018-06-11 10:48:10 +02:00
|
|
|
${SELF_TEST_DIR}/UsageTests/Approx.tests.cpp
|
|
|
|
${SELF_TEST_DIR}/UsageTests/BDD.tests.cpp
|
|
|
|
${SELF_TEST_DIR}/UsageTests/Benchmark.tests.cpp
|
|
|
|
${SELF_TEST_DIR}/UsageTests/Class.tests.cpp
|
|
|
|
${SELF_TEST_DIR}/UsageTests/Compilation.tests.cpp
|
|
|
|
${SELF_TEST_DIR}/UsageTests/Condition.tests.cpp
|
|
|
|
${SELF_TEST_DIR}/UsageTests/Decomposition.tests.cpp
|
|
|
|
${SELF_TEST_DIR}/UsageTests/EnumToString.tests.cpp
|
|
|
|
${SELF_TEST_DIR}/UsageTests/Exception.tests.cpp
|
2018-06-12 18:50:39 +02:00
|
|
|
${SELF_TEST_DIR}/UsageTests/Generators.tests.cpp
|
2018-06-11 10:48:10 +02:00
|
|
|
${SELF_TEST_DIR}/UsageTests/Message.tests.cpp
|
|
|
|
${SELF_TEST_DIR}/UsageTests/Misc.tests.cpp
|
2023-01-12 15:01:47 +01:00
|
|
|
${SELF_TEST_DIR}/UsageTests/Skip.tests.cpp
|
2019-07-13 14:47:56 +02:00
|
|
|
${SELF_TEST_DIR}/UsageTests/ToStringByte.tests.cpp
|
2018-06-11 10:48:10 +02:00
|
|
|
${SELF_TEST_DIR}/UsageTests/ToStringChrono.tests.cpp
|
|
|
|
${SELF_TEST_DIR}/UsageTests/ToStringGeneral.tests.cpp
|
2019-01-25 22:05:40 +01:00
|
|
|
${SELF_TEST_DIR}/UsageTests/ToStringOptional.tests.cpp
|
2018-06-11 10:48:10 +02:00
|
|
|
${SELF_TEST_DIR}/UsageTests/ToStringPair.tests.cpp
|
|
|
|
${SELF_TEST_DIR}/UsageTests/ToStringTuple.tests.cpp
|
2018-09-20 14:13:35 +02:00
|
|
|
${SELF_TEST_DIR}/UsageTests/ToStringVariant.tests.cpp
|
2018-06-11 10:48:10 +02:00
|
|
|
${SELF_TEST_DIR}/UsageTests/ToStringVector.tests.cpp
|
|
|
|
${SELF_TEST_DIR}/UsageTests/ToStringWhich.tests.cpp
|
|
|
|
${SELF_TEST_DIR}/UsageTests/Tricky.tests.cpp
|
|
|
|
${SELF_TEST_DIR}/UsageTests/VariadicMacros.tests.cpp
|
2020-03-02 14:25:43 +01:00
|
|
|
${SELF_TEST_DIR}/UsageTests/MatchersRanges.tests.cpp
|
2018-06-11 10:48:10 +02:00
|
|
|
${SELF_TEST_DIR}/UsageTests/Matchers.tests.cpp
|
|
|
|
)
|
|
|
|
|
2022-10-28 00:16:38 +02:00
|
|
|
set(TEST_HEADERS
|
|
|
|
${SELF_TEST_DIR}/helpers/parse_test_spec.hpp
|
2023-02-05 16:40:38 +01:00
|
|
|
${SELF_TEST_DIR}/helpers/range_test_helpers.hpp
|
2022-11-01 15:01:43 +01:00
|
|
|
${SELF_TEST_DIR}/helpers/type_with_lit_0_comparisons.hpp
|
2022-10-28 00:16:38 +02:00
|
|
|
)
|
|
|
|
|
2018-06-11 10:48:10 +02:00
|
|
|
|
|
|
|
# Specify the headers, too, so CLion recognises them as project files
|
|
|
|
set(HEADERS
|
|
|
|
${TOP_LEVEL_HEADERS}
|
|
|
|
${EXTERNAL_HEADERS}
|
|
|
|
${INTERNAL_HEADERS}
|
|
|
|
${REPORTER_HEADERS}
|
2019-12-18 17:28:19 +01:00
|
|
|
${BENCHMARK_HEADERS}
|
2019-06-15 11:14:24 +02:00
|
|
|
${BENCHMARK_SOURCES}
|
|
|
|
)
|
2018-06-11 10:48:10 +02:00
|
|
|
|
|
|
|
# Provide some groupings for IDEs
|
2019-11-30 20:45:34 +01:00
|
|
|
#SOURCE_GROUP("benchmark" FILES ${BENCHMARK_HEADERS} ${BENCHMARK_SOURCES})
|
|
|
|
#SOURCE_GROUP("Tests" FILES ${TEST_SOURCES})
|
2018-06-11 10:48:10 +02:00
|
|
|
|
|
|
|
include(CTest)
|
|
|
|
|
2022-10-28 00:16:38 +02:00
|
|
|
add_executable(SelfTest ${TEST_SOURCES} ${TEST_HEADERS})
|
|
|
|
target_include_directories(SelfTest PRIVATE ${SELF_TEST_DIR})
|
2020-05-07 16:24:05 +02:00
|
|
|
target_link_libraries(SelfTest PRIVATE Catch2WithMain)
|
2022-07-18 11:28:04 +02:00
|
|
|
if (BUILD_SHARED_LIBS AND WIN32)
|
|
|
|
add_custom_command(TARGET SelfTest PRE_LINK
|
|
|
|
COMMAND ${CMAKE_COMMAND} -E copy_if_different $<TARGET_FILE:Catch2>
|
|
|
|
$<TARGET_FILE:Catch2WithMain> $<TARGET_FILE_DIR:SelfTest>
|
|
|
|
)
|
|
|
|
endif()
|
2018-06-11 10:48:10 +02:00
|
|
|
|
|
|
|
if (CATCH_ENABLE_COVERAGE)
|
|
|
|
set(ENABLE_COVERAGE ON CACHE BOOL "Enable coverage build." FORCE)
|
|
|
|
find_package(codecov)
|
|
|
|
add_coverage(SelfTest)
|
|
|
|
list(APPEND LCOV_REMOVE_PATTERNS "'/usr/*'")
|
|
|
|
coverage_evaluate()
|
|
|
|
endif()
|
|
|
|
|
|
|
|
# configure unit tests via CTest
|
2020-05-03 09:49:59 +02:00
|
|
|
add_test(NAME RunTests COMMAND $<TARGET_FILE:SelfTest> --order rand --rng-seed time)
|
2019-04-16 23:49:22 +02:00
|
|
|
set_tests_properties(RunTests PROPERTIES
|
|
|
|
FAIL_REGULAR_EXPRESSION "Filters:"
|
2022-11-06 00:35:50 +01:00
|
|
|
COST 15
|
2019-04-16 23:49:22 +02:00
|
|
|
)
|
2018-06-11 10:48:10 +02:00
|
|
|
|
2019-06-16 16:12:47 +02:00
|
|
|
# Because CTest does not allow us to check both return code _and_ expected
|
|
|
|
# output in one test, we run these commands twice. First time we check
|
|
|
|
# the output, the second time we check the exit code.
|
|
|
|
add_test(NAME List::Tests::Output COMMAND $<TARGET_FILE:SelfTest> --list-tests --verbosity high)
|
|
|
|
set_tests_properties(List::Tests::Output PROPERTIES
|
2018-06-11 10:48:10 +02:00
|
|
|
PASS_REGULAR_EXPRESSION "[0-9]+ test cases"
|
|
|
|
FAIL_REGULAR_EXPRESSION "Hidden Test"
|
|
|
|
)
|
2020-10-20 15:09:48 +02:00
|
|
|
# This should be equivalent to the old --list-test-names-only and be usable
|
|
|
|
# with --input-file.
|
|
|
|
add_test(NAME List::Tests::Quiet COMMAND $<TARGET_FILE:SelfTest> --list-tests --verbosity quiet)
|
|
|
|
# Sadly we cannot ask for start-of-line and end-of-line in a ctest regex,
|
|
|
|
# so we fail if we see space/tab at the start...
|
|
|
|
set_tests_properties(List::Tests::Quiet PROPERTIES
|
|
|
|
PASS_REGULAR_EXPRESSION "\"#1905 -- test spec parser properly clears internal state between compound tests\"[\r\n]"
|
|
|
|
FAIL_REGULAR_EXPRESSION "[ \t]\"#1905 -- test spec parser properly clears internal state between compound tests\""
|
|
|
|
)
|
2019-06-16 16:12:47 +02:00
|
|
|
add_test(NAME List::Tests::ExitCode COMMAND $<TARGET_FILE:SelfTest> --list-tests --verbosity high)
|
2019-06-22 15:31:11 +02:00
|
|
|
add_test(NAME List::Tests::XmlOutput COMMAND $<TARGET_FILE:SelfTest> --list-tests --verbosity high -r xml)
|
|
|
|
set_tests_properties(List::Tests::XmlOutput PROPERTIES
|
|
|
|
PASS_REGULAR_EXPRESSION "<Line>[0-9]+</Line>"
|
|
|
|
FAIL_REGULAR_EXPRESSION "[0-9]+ test cases"
|
|
|
|
)
|
2018-06-11 10:48:10 +02:00
|
|
|
|
2019-06-16 16:12:47 +02:00
|
|
|
add_test(NAME List::Tags::Output COMMAND $<TARGET_FILE:SelfTest> --list-tags)
|
|
|
|
set_tests_properties(List::Tags::Output PROPERTIES
|
|
|
|
PASS_REGULAR_EXPRESSION "[0-9]+ tags"
|
2018-09-08 20:56:31 +02:00
|
|
|
FAIL_REGULAR_EXPRESSION "\\[\\.\\]")
|
2019-06-16 16:12:47 +02:00
|
|
|
add_test(NAME List::Tags::ExitCode COMMAND $<TARGET_FILE:SelfTest> --list-tags)
|
2019-06-22 15:31:11 +02:00
|
|
|
add_test(NAME List::Tags::XmlOutput COMMAND $<TARGET_FILE:SelfTest> --list-tags -r xml)
|
|
|
|
set_tests_properties(List::Tags::XmlOutput PROPERTIES
|
|
|
|
PASS_REGULAR_EXPRESSION "<Count>18</Count>"
|
|
|
|
FAIL_REGULAR_EXPRESSION "[0-9]+ tags"
|
|
|
|
)
|
|
|
|
|
2018-06-11 10:48:10 +02:00
|
|
|
|
2019-06-16 16:12:47 +02:00
|
|
|
add_test(NAME List::Reporters::Output COMMAND $<TARGET_FILE:SelfTest> --list-reporters)
|
|
|
|
set_tests_properties(List::Reporters::Output PROPERTIES PASS_REGULAR_EXPRESSION "Available reporters:")
|
|
|
|
add_test(NAME List::Reporters::ExitCode COMMAND $<TARGET_FILE:SelfTest> --list-reporters)
|
2019-06-22 15:31:11 +02:00
|
|
|
add_test(NAME List::Reporters::XmlOutput COMMAND $<TARGET_FILE:SelfTest> --list-reporters -r xml)
|
|
|
|
set_tests_properties(List::Reporters::XmlOutput PROPERTIES
|
|
|
|
PASS_REGULAR_EXPRESSION "<Name>compact</Name>"
|
|
|
|
FAIL_REGULAR_EXPRESSION "Available reporters:"
|
|
|
|
)
|
2018-06-11 10:48:10 +02:00
|
|
|
|
2022-05-10 20:00:36 +02:00
|
|
|
add_test(NAME List::Listeners::Output
|
|
|
|
COMMAND
|
|
|
|
$<TARGET_FILE:SelfTest> --list-listeners
|
|
|
|
)
|
|
|
|
set_tests_properties(List::Listeners::Output
|
|
|
|
PROPERTIES
|
|
|
|
PASS_REGULAR_EXPRESSION "Registered listeners:"
|
|
|
|
)
|
|
|
|
add_test(NAME List::Listeners::ExitCode
|
|
|
|
COMMAND
|
|
|
|
$<TARGET_FILE:SelfTest> --list-listeners
|
|
|
|
)
|
2022-05-10 20:52:53 +02:00
|
|
|
add_test(NAME List::Listeners::XmlOutput
|
|
|
|
COMMAND
|
|
|
|
$<TARGET_FILE:SelfTest>
|
|
|
|
--list-listeners
|
|
|
|
--reporter xml
|
|
|
|
)
|
|
|
|
set_tests_properties(List::Listeners::XmlOutput
|
|
|
|
PROPERTIES
|
|
|
|
PASS_REGULAR_EXPRESSION "<RegisteredListeners>"
|
|
|
|
FAIL_REGULAR_EXPRESSION "Registered listeners:"
|
|
|
|
)
|
2018-06-11 10:48:10 +02:00
|
|
|
|
2020-02-03 21:24:53 +01:00
|
|
|
add_test(NAME NoAssertions COMMAND $<TARGET_FILE:SelfTest> -w NoAssertions "An empty test with no assertions")
|
2018-06-11 10:48:10 +02:00
|
|
|
set_tests_properties(NoAssertions PROPERTIES PASS_REGULAR_EXPRESSION "No assertions in test case")
|
|
|
|
|
2021-12-13 15:15:23 +01:00
|
|
|
# We cannot combine a regular expression on output with return code check
|
|
|
|
# in one test, so we register two instead of making a checking script because
|
|
|
|
# the runtime overhead is small enough.
|
|
|
|
add_test(NAME TestSpecs::CombiningMatchingAndNonMatchingIsOk-1 COMMAND $<TARGET_FILE:SelfTest> Tracker, "___nonexistent_test___")
|
|
|
|
|
|
|
|
add_test(NAME TestSpecs::CombiningMatchingAndNonMatchingIsOk-2 COMMAND $<TARGET_FILE:SelfTest> Tracker, "___nonexistent_test___")
|
|
|
|
set_tests_properties(TestSpecs::CombiningMatchingAndNonMatchingIsOk-2 PROPERTIES
|
2022-10-27 17:10:28 +02:00
|
|
|
PASS_REGULAR_EXPRESSION "No test cases matched '\"___nonexistent_test___\"'"
|
2019-08-06 20:51:19 +02:00
|
|
|
FAIL_REGULAR_EXPRESSION "No tests ran"
|
|
|
|
)
|
|
|
|
|
2021-12-13 15:15:23 +01:00
|
|
|
add_test(NAME TestSpecs::NoMatchedTestsFail
|
|
|
|
COMMAND $<TARGET_FILE:SelfTest> "___nonexistent_test___"
|
|
|
|
)
|
|
|
|
set_tests_properties(TestSpecs::NoMatchedTestsFail
|
|
|
|
PROPERTIES
|
|
|
|
WILL_FAIL ON
|
|
|
|
)
|
|
|
|
add_test(NAME TestSpecs::OverrideFailureWithNoMatchedTests
|
|
|
|
COMMAND $<TARGET_FILE:SelfTest> "___nonexistent_test___" --allow-running-no-tests
|
|
|
|
)
|
2019-08-06 20:51:19 +02:00
|
|
|
|
2023-01-12 15:01:47 +01:00
|
|
|
add_test(NAME TestSpecs::OverrideAllSkipFailure
|
|
|
|
COMMAND $<TARGET_FILE:SelfTest> "tests can be skipped dynamically at runtime" --allow-running-no-tests
|
|
|
|
)
|
|
|
|
|
2022-10-27 19:11:20 +02:00
|
|
|
add_test(NAME TestSpecs::NonMatchingTestSpecIsRoundTrippable
|
|
|
|
COMMAND $<TARGET_FILE:SelfTest> Tracker, "this test does not exist" "[nor does this tag]"
|
|
|
|
)
|
|
|
|
set_tests_properties(TestSpecs::NonMatchingTestSpecIsRoundTrippable
|
|
|
|
PROPERTIES
|
|
|
|
PASS_REGULAR_EXPRESSION "No test cases matched '\"this test does not exist\" \\[nor does this tag\\]'"
|
|
|
|
)
|
|
|
|
|
2021-12-18 19:59:23 +01:00
|
|
|
add_test(NAME Warnings::UnmatchedTestSpecIsAccepted
|
|
|
|
COMMAND $<TARGET_FILE:SelfTest> Tracker --warn UnmatchedTestSpec
|
|
|
|
)
|
|
|
|
set_tests_properties(Warnings::UnmatchedTestSpecIsAccepted
|
|
|
|
PROPERTIES
|
|
|
|
FAIL_REGULAR_EXPRESSION "Unrecognised warning option: "
|
|
|
|
)
|
|
|
|
|
2021-12-18 20:48:43 +01:00
|
|
|
add_test(NAME Warnings::MultipleWarningsCanBeSpecified
|
|
|
|
COMMAND
|
|
|
|
$<TARGET_FILE:SelfTest> Tracker
|
|
|
|
--warn NoAssertions
|
|
|
|
--warn UnmatchedTestSpec
|
|
|
|
)
|
|
|
|
|
2021-12-18 19:59:23 +01:00
|
|
|
add_test(NAME TestSpecs::WarnUnmatchedTestSpecFailsWithUnmatchedTestSpec
|
|
|
|
COMMAND
|
|
|
|
$<TARGET_FILE:SelfTest> Tracker, "___nonexistent_test___" --warn UnmatchedTestSpec
|
|
|
|
)
|
|
|
|
set_tests_properties(TestSpecs::WarnUnmatchedTestSpecFailsWithUnmatchedTestSpec
|
|
|
|
PROPERTIES
|
|
|
|
WILL_FAIL ON
|
|
|
|
)
|
|
|
|
|
2021-12-13 15:15:23 +01:00
|
|
|
add_test(NAME UnmatchedOutputFilter COMMAND $<TARGET_FILE:SelfTest> [this-tag-does-not-exist])
|
2019-08-06 20:51:19 +02:00
|
|
|
set_tests_properties(UnmatchedOutputFilter
|
|
|
|
PROPERTIES
|
|
|
|
PASS_REGULAR_EXPRESSION "No test cases matched '\\[this-tag-does-not-exist\\]'"
|
|
|
|
)
|
2018-06-11 10:48:10 +02:00
|
|
|
|
2019-01-15 23:13:16 +01:00
|
|
|
add_test(NAME FilteredSection-1 COMMAND $<TARGET_FILE:SelfTest> \#1394 -c RunSection)
|
|
|
|
set_tests_properties(FilteredSection-1 PROPERTIES FAIL_REGULAR_EXPRESSION "No tests ran")
|
|
|
|
add_test(NAME FilteredSection-2 COMMAND $<TARGET_FILE:SelfTest> \#1394\ nested -c NestedRunSection -c s1)
|
|
|
|
set_tests_properties(FilteredSection-2 PROPERTIES FAIL_REGULAR_EXPRESSION "No tests ran")
|
|
|
|
|
Fix potential infinite loops in generators combined with section filter
The problem was that under specific circumstances, namely that none
of their children progressed, `GeneratorTracker` will not progress.
This was changed recently, to allow for code like this, where a
`SECTION` follows a `GENERATE` at the same level:
```cpp
SECTION("A") {}
auto a = GENERATE(1, 2);
SECTION("B") {}
```
However, this interacted badly with `SECTION` filters (`-c foo`),
as they could deactivate all `SECTION`s below a generator, and thus
stop it from progressing forever. This commit makes GeneratorTracker
check whether there are any filters active, and if they are, it checks
whether its section-children can ever run.
Fixes #2025
2020-10-22 14:20:47 +02:00
|
|
|
add_test(
|
|
|
|
NAME
|
|
|
|
FilteredSection::GeneratorsDontCauseInfiniteLoop-1
|
|
|
|
COMMAND
|
|
|
|
$<TARGET_FILE:SelfTest> "#2025: original repro" -c "fov_0"
|
|
|
|
)
|
|
|
|
set_tests_properties(FilteredSection::GeneratorsDontCauseInfiniteLoop-1
|
|
|
|
PROPERTIES
|
|
|
|
PASS_REGULAR_EXPRESSION "inside with fov: 0" # This should happen
|
|
|
|
FAIL_REGULAR_EXPRESSION "inside with fov: 1" # This would mean there was no filtering
|
|
|
|
)
|
|
|
|
|
|
|
|
# GENERATE between filtered sections (both are selected)
|
|
|
|
add_test(
|
|
|
|
NAME
|
|
|
|
FilteredSection::GeneratorsDontCauseInfiniteLoop-2
|
|
|
|
COMMAND
|
|
|
|
$<TARGET_FILE:SelfTest> "#2025: same-level sections"
|
|
|
|
-c "A"
|
|
|
|
-c "B"
|
2022-06-14 23:23:59 +02:00
|
|
|
--colour-mode none
|
Fix potential infinite loops in generators combined with section filter
The problem was that under specific circumstances, namely that none
of their children progressed, `GeneratorTracker` will not progress.
This was changed recently, to allow for code like this, where a
`SECTION` follows a `GENERATE` at the same level:
```cpp
SECTION("A") {}
auto a = GENERATE(1, 2);
SECTION("B") {}
```
However, this interacted badly with `SECTION` filters (`-c foo`),
as they could deactivate all `SECTION`s below a generator, and thus
stop it from progressing forever. This commit makes GeneratorTracker
check whether there are any filters active, and if they are, it checks
whether its section-children can ever run.
Fixes #2025
2020-10-22 14:20:47 +02:00
|
|
|
)
|
|
|
|
set_tests_properties(FilteredSection::GeneratorsDontCauseInfiniteLoop-2
|
|
|
|
PROPERTIES
|
|
|
|
PASS_REGULAR_EXPRESSION "All tests passed \\(4 assertions in 1 test case\\)"
|
|
|
|
)
|
|
|
|
|
2018-06-11 10:48:10 +02:00
|
|
|
# AppVeyor has a Python 2.7 in path, but doesn't have .py files as autorunnable
|
2022-11-06 00:11:32 +01:00
|
|
|
add_test(NAME ApprovalTests
|
|
|
|
COMMAND
|
|
|
|
${PYTHON_EXECUTABLE}
|
|
|
|
${CATCH_DIR}/tools/scripts/approvalTests.py
|
|
|
|
$<TARGET_FILE:SelfTest>
|
|
|
|
"${CMAKE_CURRENT_BINARY_DIR}"
|
|
|
|
)
|
|
|
|
|
2020-02-03 09:07:23 +01:00
|
|
|
set_tests_properties(ApprovalTests
|
|
|
|
PROPERTIES
|
|
|
|
FAIL_REGULAR_EXPRESSION "Results differed"
|
2022-11-06 00:35:50 +01:00
|
|
|
|
|
|
|
# This is the most expensive test in the basic test suite, so we give
|
|
|
|
# it high cost estimate so that CI runs it as one of the first ones,
|
|
|
|
# for better parallelization.
|
|
|
|
COST 30
|
2022-04-24 12:50:24 +02:00
|
|
|
LABELS "uses-python"
|
2020-02-03 09:07:23 +01:00
|
|
|
)
|
2018-06-11 10:48:10 +02:00
|
|
|
|
2019-06-30 00:00:41 +02:00
|
|
|
add_test(NAME RegressionCheck-1670 COMMAND $<TARGET_FILE:SelfTest> "#1670 regression check" -c A -r compact)
|
2022-10-24 14:17:15 +02:00
|
|
|
set_tests_properties(RegressionCheck-1670 PROPERTIES PASS_REGULAR_EXPRESSION "All tests passed \\(2 assertions in 1 test case\\)")
|
2019-01-15 23:13:16 +01:00
|
|
|
|
2019-08-03 20:16:46 +02:00
|
|
|
add_test(NAME VersionCheck COMMAND $<TARGET_FILE:SelfTest> -h)
|
2022-01-27 11:21:43 +01:00
|
|
|
set_tests_properties(VersionCheck PROPERTIES PASS_REGULAR_EXPRESSION "Catch2 v${PROJECT_VERSION}")
|
2019-08-03 20:16:46 +02:00
|
|
|
|
2019-09-07 20:36:12 +02:00
|
|
|
add_test(NAME LibIdentityTest COMMAND $<TARGET_FILE:SelfTest> --libidentify)
|
|
|
|
set_tests_properties(LibIdentityTest PROPERTIES PASS_REGULAR_EXPRESSION "description: A Catch2 test executable")
|
|
|
|
|
2019-09-07 20:43:14 +02:00
|
|
|
add_test(NAME FilenameAsTagsTest COMMAND $<TARGET_FILE:SelfTest> -\# --list-tags)
|
|
|
|
set_tests_properties(FilenameAsTagsTest PROPERTIES PASS_REGULAR_EXPRESSION "\\[#Approx.tests\\]")
|
|
|
|
|
2020-10-22 16:17:34 +02:00
|
|
|
# Check that the filename tags can also be matched against (#2064)
|
|
|
|
add_test(NAME FilenameAsTagsMatching COMMAND $<TARGET_FILE:SelfTest> -\# --list-tags [\#Approx.tests])
|
2020-10-23 23:02:09 +02:00
|
|
|
set_tests_properties(FilenameAsTagsMatching
|
|
|
|
PROPERTIES
|
|
|
|
PASS_REGULAR_EXPRESSION "\\[#Approx.tests\\]"
|
|
|
|
# Avoids false positives by looking for start of line (newline) before the 0
|
|
|
|
FAIL_REGULAR_EXPRESSION "[\r\n]0 tag"
|
|
|
|
)
|
2020-10-22 16:17:34 +02:00
|
|
|
|
2019-09-21 00:11:45 +02:00
|
|
|
add_test(NAME EscapeSpecialCharactersInTestNames COMMAND $<TARGET_FILE:SelfTest> "Test with special\\, characters \"in name")
|
|
|
|
set_tests_properties(EscapeSpecialCharactersInTestNames PROPERTIES PASS_REGULAR_EXPRESSION "1 assertion in 1 test case")
|
|
|
|
|
2019-06-22 20:26:03 +02:00
|
|
|
add_test(NAME NegativeSpecNoHiddenTests COMMAND $<TARGET_FILE:SelfTest> --list-tests ~[approval])
|
|
|
|
set_tests_properties(NegativeSpecNoHiddenTests PROPERTIES FAIL_REGULAR_EXPRESSION "\\[\\.\\]")
|
|
|
|
|
2019-12-05 16:00:20 +01:00
|
|
|
add_test(NAME TestsInFile::SimpleSpecs COMMAND $<TARGET_FILE:SelfTest> "-f ${SELF_TEST_DIR}/Misc/plain-old-tests.input")
|
2019-10-20 15:14:50 +02:00
|
|
|
set_tests_properties(TestsInFile::SimpleSpecs PROPERTIES PASS_REGULAR_EXPRESSION "6 assertions in 2 test cases")
|
2019-10-09 12:48:55 +02:00
|
|
|
|
2019-12-05 16:00:20 +01:00
|
|
|
add_test(NAME TestsInFile::EscapeSpecialCharacters COMMAND $<TARGET_FILE:SelfTest> "-f ${SELF_TEST_DIR}/Misc/special-characters-in-file.input")
|
2019-10-20 15:14:50 +02:00
|
|
|
set_tests_properties(TestsInFile::EscapeSpecialCharacters PROPERTIES PASS_REGULAR_EXPRESSION "1 assertion in 1 test case")
|
|
|
|
|
2019-12-05 16:00:20 +01:00
|
|
|
add_test(NAME TestsInFile::InvalidTestNames-1 COMMAND $<TARGET_FILE:SelfTest> "-f ${SELF_TEST_DIR}/Misc/invalid-test-names.input")
|
2021-12-12 15:18:50 +01:00
|
|
|
set_tests_properties(TestsInFile::InvalidTestNames-1
|
|
|
|
PROPERTIES
|
|
|
|
PASS_REGULAR_EXPRESSION "Invalid Filter: \"Test with special, characters in \\\\\" name\""
|
|
|
|
FAIL_REGULAR_EXPRESSION "No tests ran"
|
|
|
|
)
|
2019-10-09 12:48:55 +02:00
|
|
|
|
2019-12-18 17:30:41 +01:00
|
|
|
add_test(NAME TagAlias COMMAND $<TARGET_FILE:SelfTest> [@tricky] --list-tests)
|
|
|
|
set_tests_properties(TagAlias PROPERTIES
|
|
|
|
PASS_REGULAR_EXPRESSION "[0-9]+ matching test cases"
|
|
|
|
FAIL_REGULAR_EXPRESSION "0 matching test cases"
|
|
|
|
)
|
2019-08-03 20:16:46 +02:00
|
|
|
|
2020-04-10 23:25:13 +02:00
|
|
|
add_test(NAME RandomTestOrdering COMMAND ${PYTHON_EXECUTABLE}
|
|
|
|
${CATCH_DIR}/tests/TestScripts/testRandomOrder.py $<TARGET_FILE:SelfTest>)
|
2022-04-24 12:50:24 +02:00
|
|
|
set_tests_properties(RandomTestOrdering
|
|
|
|
PROPERTIES
|
|
|
|
LABELS "uses-python"
|
|
|
|
)
|
2021-07-11 21:46:05 +02:00
|
|
|
|
2020-05-06 14:57:55 +02:00
|
|
|
add_test(NAME CheckConvenienceHeaders
|
|
|
|
COMMAND
|
|
|
|
${PYTHON_EXECUTABLE} ${CATCH_DIR}/tools/scripts/checkConvenienceHeaders.py
|
|
|
|
)
|
2022-04-24 12:50:24 +02:00
|
|
|
set_tests_properties(CheckConvenienceHeaders
|
|
|
|
PROPERTIES
|
|
|
|
LABELS "uses-python"
|
|
|
|
)
|
2020-05-06 14:57:55 +02:00
|
|
|
|
2022-04-23 23:14:49 +02:00
|
|
|
add_test(NAME "Benchmarking::SkipBenchmarkMacros"
|
|
|
|
COMMAND
|
|
|
|
$<TARGET_FILE:SelfTest> "Skip benchmark macros"
|
|
|
|
--reporter console
|
|
|
|
--skip-benchmarks
|
2022-06-14 23:23:59 +02:00
|
|
|
--colour-mode none
|
2022-04-23 23:14:49 +02:00
|
|
|
)
|
|
|
|
set_tests_properties("Benchmarking::SkipBenchmarkMacros"
|
|
|
|
PROPERTIES
|
|
|
|
PASS_REGULAR_EXPRESSION "All tests passed \\(2 assertions in 1 test case\\)"
|
|
|
|
FAIL_REGULAR_EXPRESSION "benchmark name"
|
|
|
|
)
|
|
|
|
|
2021-06-09 20:48:58 +02:00
|
|
|
|
|
|
|
add_test(NAME "Benchmarking::FailureReporting::OptimizedOut"
|
|
|
|
COMMAND
|
|
|
|
$<TARGET_FILE:SelfTest> "Failing benchmarks" -c "empty" -r xml
|
|
|
|
# This test only makes sense with the optimizer being enabled when
|
|
|
|
# the tests are being compiled.
|
|
|
|
CONFIGURATIONS Release
|
|
|
|
)
|
|
|
|
set_tests_properties("Benchmarking::FailureReporting::OptimizedOut"
|
|
|
|
PROPERTIES
|
|
|
|
PASS_REGULAR_EXPRESSION "could not measure benchmark\, maybe it was optimized away"
|
|
|
|
FAIL_REGULAR_EXPRESSION "successes=\"1\""
|
|
|
|
)
|
|
|
|
|
|
|
|
add_test(NAME "Benchmarking::FailureReporting::ThrowingBenchmark"
|
|
|
|
COMMAND
|
|
|
|
$<TARGET_FILE:SelfTest> "Failing benchmarks" -c "throw" -r xml
|
|
|
|
)
|
|
|
|
set_tests_properties("Benchmarking::FailureReporting::ThrowingBenchmark"
|
|
|
|
PROPERTIES
|
|
|
|
PASS_REGULAR_EXPRESSION "<failed message=\"just a plain literal"
|
|
|
|
FAIL_REGULAR_EXPRESSION "successes=\"1\""
|
|
|
|
)
|
|
|
|
|
|
|
|
add_test(NAME "Benchmarking::FailureReporting::FailedAssertion"
|
|
|
|
COMMAND
|
|
|
|
$<TARGET_FILE:SelfTest> "Failing benchmarks" -c "assert" -r xml
|
|
|
|
)
|
|
|
|
set_tests_properties("Benchmarking::FailureReporting::FailedAssertion"
|
|
|
|
PROPERTIES
|
|
|
|
PASS_REGULAR_EXPRESSION "<Expression success=\"false\""
|
|
|
|
FAIL_REGULAR_EXPRESSION "successes=\"1\""
|
|
|
|
)
|
|
|
|
|
|
|
|
add_test(NAME "Benchmarking::FailureReporting::FailMacro"
|
|
|
|
COMMAND
|
|
|
|
$<TARGET_FILE:SelfTest> "Failing benchmarks" -c "fail" -r xml
|
|
|
|
)
|
|
|
|
set_tests_properties("Benchmarking::FailureReporting::FailMacro"
|
|
|
|
PROPERTIES
|
|
|
|
PASS_REGULAR_EXPRESSION "This benchmark only fails\, nothing else"
|
|
|
|
FAIL_REGULAR_EXPRESSION "successes=\"1\""
|
|
|
|
)
|
|
|
|
|
2021-08-21 00:06:31 +02:00
|
|
|
add_test(NAME "Benchmarking::FailureReporting::ShouldFailIsRespected"
|
|
|
|
COMMAND
|
|
|
|
$<TARGET_FILE:SelfTest> "Failing benchmark respects should-fail"
|
|
|
|
)
|
|
|
|
set_tests_properties("Benchmarking::FailureReporting::ShouldFailIsRespected"
|
|
|
|
PROPERTIES
|
|
|
|
PASS_REGULAR_EXPRESSION "1 failed as expected"
|
|
|
|
)
|
|
|
|
|
2021-12-12 15:18:50 +01:00
|
|
|
add_test(NAME "ErrorHandling::InvalidTestSpecExitsEarly"
|
|
|
|
COMMAND
|
|
|
|
$<TARGET_FILE:SelfTest> "[aa,a]"
|
|
|
|
)
|
|
|
|
set_tests_properties("ErrorHandling::InvalidTestSpecExitsEarly"
|
|
|
|
PROPERTIES
|
|
|
|
PASS_REGULAR_EXPRESSION "Invalid Filter: \\[aa\,a\\]"
|
|
|
|
FAIL_REGULAR_EXPRESSION "No tests ran"
|
|
|
|
)
|
|
|
|
|
2021-12-29 23:53:59 +01:00
|
|
|
if (MSVC)
|
|
|
|
set(_NullFile "NUL")
|
|
|
|
else()
|
|
|
|
set(_NullFile "/dev/null")
|
|
|
|
endif()
|
|
|
|
|
|
|
|
# This test checks that there is nothing written out from the process,
|
|
|
|
# but if CMake is running the tests under Valgrind or similar tool, then
|
|
|
|
# that will write its own output to stdout and the test would fail.
|
|
|
|
if (NOT MEMORYCHECK_COMMAND)
|
|
|
|
add_test(NAME "MultiReporter::CapturingReportersDontPropagateStdOut"
|
|
|
|
COMMAND
|
|
|
|
$<TARGET_FILE:SelfTest> "Sends stuff to stdout and stderr"
|
2022-04-04 21:07:04 +02:00
|
|
|
--reporter xml::out=${_NullFile}
|
|
|
|
--reporter junit::out=${_NullFile}
|
2021-12-29 23:53:59 +01:00
|
|
|
)
|
|
|
|
set_tests_properties("MultiReporter::CapturingReportersDontPropagateStdOut"
|
|
|
|
PROPERTIES
|
|
|
|
FAIL_REGULAR_EXPRESSION ".+"
|
|
|
|
)
|
|
|
|
endif()
|
|
|
|
|
|
|
|
add_test(NAME "MultiReporter::NonCapturingReportersPropagateStdout"
|
|
|
|
COMMAND
|
|
|
|
$<TARGET_FILE:SelfTest> "Sends stuff to stdout and stderr"
|
2022-04-04 21:07:04 +02:00
|
|
|
--reporter xml::out=${_NullFile}
|
|
|
|
--reporter console::out=${_NullFile}
|
2021-12-29 23:53:59 +01:00
|
|
|
)
|
|
|
|
set_tests_properties("MultiReporter::NonCapturingReportersPropagateStdout"
|
|
|
|
PROPERTIES
|
|
|
|
PASS_REGULAR_EXPRESSION "A string sent to stderr via clog"
|
|
|
|
)
|
|
|
|
|
2021-12-31 14:51:15 +01:00
|
|
|
add_test(NAME "Outputs::DashAsOutLocationSendsOutputToStdout"
|
|
|
|
COMMAND
|
|
|
|
$<TARGET_FILE:SelfTest> "Factorials are computed"
|
|
|
|
--out=-
|
2022-06-14 23:23:59 +02:00
|
|
|
--colour-mode none
|
2021-12-31 14:51:15 +01:00
|
|
|
)
|
|
|
|
set_tests_properties("Outputs::DashAsOutLocationSendsOutputToStdout"
|
|
|
|
PROPERTIES
|
|
|
|
PASS_REGULAR_EXPRESSION "All tests passed \\(5 assertions in 1 test case\\)"
|
|
|
|
)
|
|
|
|
|
|
|
|
add_test(NAME "Reporters::DashAsLocationInReporterSpecSendsOutputToStdout"
|
|
|
|
COMMAND
|
|
|
|
$<TARGET_FILE:SelfTest> "Factorials are computed"
|
2022-04-04 21:07:04 +02:00
|
|
|
--reporter console::out=-
|
2022-06-14 23:23:59 +02:00
|
|
|
--colour-mode none
|
2021-12-31 14:51:15 +01:00
|
|
|
)
|
|
|
|
set_tests_properties("Reporters::DashAsLocationInReporterSpecSendsOutputToStdout"
|
|
|
|
PROPERTIES
|
|
|
|
PASS_REGULAR_EXPRESSION "All tests passed \\(5 assertions in 1 test case\\)"
|
|
|
|
)
|
|
|
|
|
2022-04-04 21:07:04 +02:00
|
|
|
add_test(NAME "Reporters::ReporterSpecificColourOverridesDefaultColour"
|
|
|
|
COMMAND
|
|
|
|
$<TARGET_FILE:SelfTest> "Factorials are computed"
|
|
|
|
--reporter console::colour-mode=ansi
|
|
|
|
--colour-mode none
|
|
|
|
)
|
|
|
|
set_tests_properties("Reporters::ReporterSpecificColourOverridesDefaultColour"
|
|
|
|
PROPERTIES
|
|
|
|
PASS_REGULAR_EXPRESSION "\\[1\;32mAll tests passed"
|
|
|
|
)
|
|
|
|
|
|
|
|
add_test(NAME "Reporters::UnrecognizedOptionInSpecCausesError"
|
|
|
|
COMMAND
|
|
|
|
$<TARGET_FILE:SelfTest> "Factorials are computed"
|
|
|
|
--reporter console::bad-option=ansi
|
|
|
|
)
|
|
|
|
set_tests_properties("Reporters::UnrecognizedOptionInSpecCausesError"
|
|
|
|
PROPERTIES
|
|
|
|
WILL_FAIL ON
|
|
|
|
)
|
|
|
|
|
2022-03-27 23:35:41 +02:00
|
|
|
add_test(NAME "Colours::ColourModeCanBeExplicitlySetToAnsi"
|
|
|
|
COMMAND
|
|
|
|
$<TARGET_FILE:SelfTest> "Factorials are computed"
|
|
|
|
--reporter console
|
|
|
|
--colour-mode ansi
|
|
|
|
)
|
|
|
|
set_tests_properties("Colours::ColourModeCanBeExplicitlySetToAnsi"
|
|
|
|
PROPERTIES
|
|
|
|
PASS_REGULAR_EXPRESSION "\\[1\;32mAll tests passed"
|
|
|
|
)
|
|
|
|
|
2022-07-17 14:00:04 +02:00
|
|
|
add_test(NAME "Reporters::JUnit::NamespacesAreNormalized"
|
|
|
|
COMMAND
|
|
|
|
$<TARGET_FILE:SelfTest>
|
|
|
|
--reporter junit
|
|
|
|
"A TEST_CASE_METHOD testing junit classname normalization"
|
|
|
|
)
|
|
|
|
set_tests_properties("Reporters::JUnit::NamespacesAreNormalized"
|
|
|
|
PROPERTIES
|
|
|
|
PASS_REGULAR_EXPRESSION "testcase classname=\"SelfTest(\.exe)?\\.A\\.B\\.TestClass\""
|
|
|
|
)
|
|
|
|
|
2022-02-19 21:27:53 +01:00
|
|
|
if (CATCH_ENABLE_CONFIGURE_TESTS)
|
2022-04-10 21:55:52 +02:00
|
|
|
foreach(testName "DefaultReporter" "Disable" "DisableStringification"
|
|
|
|
"ExperimentalRedirect")
|
|
|
|
|
|
|
|
add_test(NAME "CMakeConfig::${testName}"
|
|
|
|
COMMAND
|
|
|
|
"${PYTHON_EXECUTABLE}" "${CMAKE_CURRENT_LIST_DIR}/TestScripts/testConfigure${testName}.py" "${CATCH_DIR}" "${CMAKE_CURRENT_BINARY_DIR}"
|
|
|
|
)
|
|
|
|
set_tests_properties("CMakeConfig::${testName}"
|
|
|
|
PROPERTIES
|
|
|
|
COST 240
|
2022-04-24 12:50:24 +02:00
|
|
|
LABELS "uses-python"
|
2022-04-10 21:55:52 +02:00
|
|
|
)
|
|
|
|
|
|
|
|
endforeach()
|
2022-02-19 21:27:53 +01:00
|
|
|
endif()
|
2021-08-21 00:06:31 +02:00
|
|
|
|
2023-06-14 16:30:56 +02:00
|
|
|
if (CATCH_ENABLE_CMAKE_HELPER_TESTS)
|
|
|
|
add_test(NAME "CMakeHelper::DiscoverTests"
|
|
|
|
COMMAND
|
|
|
|
"${PYTHON_EXECUTABLE}" "${CMAKE_CURRENT_LIST_DIR}/TestScripts/DiscoverTests/VerifyRegistration.py" "${CATCH_DIR}" "${CMAKE_CURRENT_BINARY_DIR}"
|
|
|
|
)
|
|
|
|
set_tests_properties("CMakeHelper::DiscoverTests"
|
|
|
|
PROPERTIES
|
|
|
|
COST 240
|
|
|
|
LABELS "uses-python"
|
|
|
|
)
|
|
|
|
endif()
|
|
|
|
|
2022-04-19 10:32:13 +02:00
|
|
|
foreach (reporterName # "Automake" - the simple .trs format does not support any kind of comments/metadata
|
|
|
|
"compact"
|
|
|
|
"console"
|
|
|
|
"JUnit"
|
|
|
|
"SonarQube"
|
|
|
|
"TAP"
|
|
|
|
# "TeamCity" - does not seem to support test suite-level metadata/comments
|
|
|
|
"XML")
|
2022-10-26 11:26:10 +02:00
|
|
|
|
|
|
|
add_test(NAME "Reporters:Filters:${reporterName}"
|
|
|
|
COMMAND
|
|
|
|
$<TARGET_FILE:SelfTest> [comparisons][string-case] "CaseInsensitiveLess is case insensitive"
|
|
|
|
--reporter ${reporterName}
|
|
|
|
)
|
2022-10-27 20:43:00 +02:00
|
|
|
# Different regex for these two reporters, because the commas end up xml-escaped
|
|
|
|
if (reporterName MATCHES "JUnit|XML")
|
|
|
|
set(testCaseNameFormat ""CaseInsensitiveLess is case insensitive"")
|
|
|
|
else()
|
|
|
|
set(testCaseNameFormat "\"CaseInsensitiveLess is case insensitive\"")
|
|
|
|
endif()
|
2022-10-26 11:26:10 +02:00
|
|
|
set_tests_properties("Reporters:Filters:${reporterName}"
|
|
|
|
PROPERTIES
|
2022-10-27 20:43:00 +02:00
|
|
|
PASS_REGULAR_EXPRESSION "[fF]ilters.+\\[comparisons\\] \\[string-case\\] ${testCaseNameFormat}"
|
2022-10-26 11:26:10 +02:00
|
|
|
)
|
|
|
|
|
2022-04-19 10:32:13 +02:00
|
|
|
add_test(NAME "Reporters:RngSeed:${reporterName}"
|
|
|
|
COMMAND
|
|
|
|
$<TARGET_FILE:SelfTest> "Factorials are computed"
|
|
|
|
--reporter ${reporterName}
|
|
|
|
--rng-seed 18181818
|
|
|
|
)
|
|
|
|
set_tests_properties("Reporters:RngSeed:${reporterName}"
|
|
|
|
PROPERTIES
|
|
|
|
PASS_REGULAR_EXPRESSION "18181818"
|
|
|
|
)
|
|
|
|
|
|
|
|
endforeach()
|
|
|
|
|
|
|
|
|
2020-02-21 21:15:45 +01:00
|
|
|
list(APPEND CATCH_WARNING_TARGETS SelfTest)
|
|
|
|
set(CATCH_WARNING_TARGETS ${CATCH_WARNING_TARGETS} PARENT_SCOPE)
|