/*- * Copyright (c) 2004 Nik Clayton * All rights reserved. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * 1. Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF * SUCH DAMAGE. */ /** * plan_tests - announce the number of tests you plan to run * @tests: the number of tests * * This should be the first call in your test program: it allows tracing * of failures which mean that not all tests are run. * * If you don't know how many tests will actually be run, assume all of them * and use skip() if you don't actually run some tests. * * Example: * plan_tests(13); */ void plan_tests(unsigned int tests); #if (!defined(__STDC_VERSION__) || __STDC_VERSION__ < 199901L) && !defined(__GNUC__) # error "Needs gcc or C99 compiler for variadic macros." #else /** * ok1 - Simple conditional test * @e: the expression which we expect to be true. * * This is the simplest kind of test: if the expression is true, the * test passes. The name of the test which is printed will simply be * file name, line number, and the expression itself. * * Example: * ok1(init_subsystem() == 1); */ # define ok1(e) ((e) ? \ _gen_result(1, __func__, __FILE__, __LINE__, "%s", #e) : \ _gen_result(0, __func__, __FILE__, __LINE__, "%s", #e)) /** * ok - Conditional test with a name * @e: the expression which we expect to be true. * @...: the printf-style name of the test. * * If the expression is true, the test passes. The name of the test will be * the filename, line number, and the printf-style string. This can be clearer * than simply the expression itself. * * Example: * ok1(init_subsystem() == 1); * ok(init_subsystem() == 0, "Second initialization should fail"); */ # define ok(e, ...) ((e) ? \ _gen_result(1, __func__, __FILE__, __LINE__, \ __VA_ARGS__) : \ _gen_result(0, __func__, __FILE__, __LINE__, \ __VA_ARGS__)) /** * pass - Note that a test passed * @...: the printf-style name of the test. * * For complicated code paths, it can be easiest to simply call pass() in one * branch and fail() in another. * * Example: * x = do_something(); * if (!checkable(x) || check_value(x)) * pass("do_something() returned a valid value"); * else * fail("do_something() returned an invalid value"); */ # define pass(...) ok(1, __VA_ARGS__) /** * fail - Note that a test failed * @...: the printf-style name of the test. * * For complicated code paths, it can be easiest to simply call pass() in one * branch and fail() in another. */ # define fail(...) ok(0, __VA_ARGS__) /* I don't find these to be useful. */ # define skip_if(cond, n, ...) \ if (cond) skip((n), __VA_ARGS__); \ else # define skip_start(test, n, ...) \ do { \ if((test)) { \ skip(n, __VA_ARGS__); \ continue; \ } # define skip_end } while(0) #ifndef PRINTF_ATTRIBUTE #ifdef __GNUC__ #define PRINTF_ATTRIBUTE(a1, a2) __attribute__ ((format (__printf__, a1, a2))) #else #define PRINTF_ATTRIBUTE(a1, a2) #endif #endif unsigned int _gen_result(int, const char *, char *, unsigned int, char *, ...) PRINTF_ATTRIBUTE(5, 6); /** * diag - print a diagnostic message (use instead of printf/fprintf) * @fmt: the format of the printf-style message * * diag ensures that the output will not be considered to be a test * result by the TAP test harness. It will append '\n' for you. * * Example: * diag("Now running complex tests"); */ void diag(char *fmt, ...) PRINTF_ATTRIBUTE(1, 2); /** * skip - print a diagnostic message (use instead of printf/fprintf) * @n: number of tests you're skipping. * @fmt: the format of the reason you're skipping the tests. * * Sometimes tests cannot be run because the test system lacks some feature: * you should explicitly document that you're skipping tests using skip(). * * From the Test::More documentation: * If it's something the user might not be able to do, use SKIP. This * includes optional modules that aren't installed, running under an OS that * doesn't have some feature (like fork() or symlinks), or maybe you need an * Internet connection and one isn't available. * * Example: * #ifdef HAVE_SOME_FEATURE * ok1(test_some_feature()); * #else * skip(1, "Don't have SOME_FEATURE"); * #endif */ void skip(unsigned int n, char *fmt, ...) PRINTF_ATTRIBUTE(2, 3); /** * todo_start - mark tests that you expect to fail. * @fmt: the reason they currently fail. * * It's extremely useful to write tests before you implement the matching fix * or features: surround these tests by todo_start()/todo_end(). These tests * will still be run, but with additional output that indicates that they are * expected to fail. * * This way, should a test start to succeed unexpectedly, tools like prove(1) * will indicate this and you can move the test out of the todo block. This * is much more useful than simply commenting out (or '#if 0') the tests. * * From the Test::More documentation: * If it's something the programmer hasn't done yet, use TODO. This is for * any code you haven't written yet, or bugs you have yet to fix, but want to * put tests in your testing script (always a good idea). * * Example: * todo_start("dwim() not returning true yet"); * ok(dwim(), "Did what the user wanted"); * todo_end(); */ void todo_start(char *fmt, ...) PRINTF_ATTRIBUTE(1, 2); /** * todo_end - end of tests you expect to fail. * * See todo_start(). */ void todo_end(void); /** * exit_status - the value that main should return. * * For maximum compatability your test program should return a particular exit * code (ie. 0 if all tests were run, and every test which was expected to * succeed succeeded). * * Example: * exit(exit_status()); */ int exit_status(void); /** * plan_no_plan - I have no idea how many tests I'm going to run. * * In some situations you may not know how many tests you will be running, or * you are developing your test program, and do not want to update the * plan_tests() call every time you make a change. For those situations use * plan_no_plan() instead of plan_tests(). It indicates to the test harness * that an indeterminate number of tests will be run. * * Remember, if you fail to plan, you plan to fail. * * Example: * plan_no_plan(); * while (random() % 2) * ok1(some_test()); * exit(exit_status()); */ void plan_no_plan(void); /** * plan_skip_all - Indicate that you will skip all tests. * @reason: the string indicating why you can't run any tests. * * If your test program detects at run time that some required functionality * is missing (for example, it relies on a database connection which is not * present, or a particular configuration option that has not been included * in the running kernel) use plan_skip_all() instead of plan_tests(). * * Example: * if (!have_some_feature) { * plan_skip_all("Need some_feature support"); * exit(exit_status()); * } * plan_tests(13); */ void plan_skip_all(char *reason); #endif /* C99 or gcc */