mirror of
https://gitlab.gnome.org/GNOME/glib.git
synced 2024-11-15 05:46:15 +01:00
b639687b60
Coverity CID: #1430603 Signed-off-by: Philip Withnall <withnall@endlessm.com>
4100 lines
122 KiB
C
4100 lines
122 KiB
C
/* GLib testing utilities
|
||
* Copyright (C) 2007 Imendio AB
|
||
* Authors: Tim Janik, Sven Herzberg
|
||
*
|
||
* This library is free software; you can redistribute it and/or
|
||
* modify it under the terms of the GNU Lesser General Public
|
||
* License as published by the Free Software Foundation; either
|
||
* version 2.1 of the License, or (at your option) any later version.
|
||
*
|
||
* This library is distributed in the hope that it will be useful,
|
||
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
||
* Lesser General Public License for more details.
|
||
*
|
||
* You should have received a copy of the GNU Lesser General Public
|
||
* License along with this library; if not, see <http://www.gnu.org/licenses/>.
|
||
*/
|
||
|
||
#include "config.h"
|
||
|
||
#include "gtestutils.h"
|
||
#include "gfileutils.h"
|
||
|
||
#include <sys/types.h>
|
||
#ifdef G_OS_UNIX
|
||
#include <sys/wait.h>
|
||
#include <sys/time.h>
|
||
#include <fcntl.h>
|
||
#include <unistd.h>
|
||
#endif
|
||
#include <string.h>
|
||
#include <stdlib.h>
|
||
#include <stdio.h>
|
||
#ifdef HAVE_SYS_RESOURCE_H
|
||
#include <sys/resource.h>
|
||
#endif
|
||
#ifdef G_OS_WIN32
|
||
#include <io.h>
|
||
#include <windows.h>
|
||
#endif
|
||
#include <errno.h>
|
||
#include <signal.h>
|
||
#ifdef HAVE_SYS_SELECT_H
|
||
#include <sys/select.h>
|
||
#endif /* HAVE_SYS_SELECT_H */
|
||
#include <glib/gstdio.h>
|
||
|
||
#include "gmain.h"
|
||
#include "gpattern.h"
|
||
#include "grand.h"
|
||
#include "gstrfuncs.h"
|
||
#include "gtimer.h"
|
||
#include "gslice.h"
|
||
#include "gspawn.h"
|
||
#include "glib-private.h"
|
||
#include "gutilsprivate.h"
|
||
|
||
|
||
/**
|
||
* SECTION:testing
|
||
* @title: Testing
|
||
* @short_description: a test framework
|
||
*
|
||
* GLib provides a framework for writing and maintaining unit tests
|
||
* in parallel to the code they are testing. The API is designed according
|
||
* to established concepts found in the other test frameworks (JUnit, NUnit,
|
||
* RUnit), which in turn is based on smalltalk unit testing concepts.
|
||
*
|
||
* - Test case: Tests (test methods) are grouped together with their
|
||
* fixture into test cases.
|
||
*
|
||
* - Fixture: A test fixture consists of fixture data and setup and
|
||
* teardown methods to establish the environment for the test
|
||
* functions. We use fresh fixtures, i.e. fixtures are newly set
|
||
* up and torn down around each test invocation to avoid dependencies
|
||
* between tests.
|
||
*
|
||
* - Test suite: Test cases can be grouped into test suites, to allow
|
||
* subsets of the available tests to be run. Test suites can be
|
||
* grouped into other test suites as well.
|
||
*
|
||
* The API is designed to handle creation and registration of test suites
|
||
* and test cases implicitly. A simple call like
|
||
* |[<!-- language="C" -->
|
||
* g_test_add_func ("/misc/assertions", test_assertions);
|
||
* ]|
|
||
* creates a test suite called "misc" with a single test case named
|
||
* "assertions", which consists of running the test_assertions function.
|
||
*
|
||
* In addition to the traditional g_assert_true(), the test framework provides
|
||
* an extended set of assertions for comparisons: g_assert_cmpfloat(),
|
||
* g_assert_cmpfloat_with_epsilon(), g_assert_cmpint(), g_assert_cmpuint(),
|
||
* g_assert_cmphex(), g_assert_cmpstr(), g_assert_cmpmem() and
|
||
* g_assert_cmpvariant(). The
|
||
* advantage of these variants over plain g_assert_true() is that the assertion
|
||
* messages can be more elaborate, and include the values of the compared
|
||
* entities.
|
||
*
|
||
* Note that g_assert() should not be used in unit tests, since it is a no-op
|
||
* when compiling with `G_DISABLE_ASSERT`. Use g_assert() in production code,
|
||
* and g_assert_true() in unit tests.
|
||
*
|
||
* A full example of creating a test suite with two tests using fixtures:
|
||
* |[<!-- language="C" -->
|
||
* #include <glib.h>
|
||
* #include <locale.h>
|
||
*
|
||
* typedef struct {
|
||
* MyObject *obj;
|
||
* OtherObject *helper;
|
||
* } MyObjectFixture;
|
||
*
|
||
* static void
|
||
* my_object_fixture_set_up (MyObjectFixture *fixture,
|
||
* gconstpointer user_data)
|
||
* {
|
||
* fixture->obj = my_object_new ();
|
||
* my_object_set_prop1 (fixture->obj, "some-value");
|
||
* my_object_do_some_complex_setup (fixture->obj, user_data);
|
||
*
|
||
* fixture->helper = other_object_new ();
|
||
* }
|
||
*
|
||
* static void
|
||
* my_object_fixture_tear_down (MyObjectFixture *fixture,
|
||
* gconstpointer user_data)
|
||
* {
|
||
* g_clear_object (&fixture->helper);
|
||
* g_clear_object (&fixture->obj);
|
||
* }
|
||
*
|
||
* static void
|
||
* test_my_object_test1 (MyObjectFixture *fixture,
|
||
* gconstpointer user_data)
|
||
* {
|
||
* g_assert_cmpstr (my_object_get_property (fixture->obj), ==, "initial-value");
|
||
* }
|
||
*
|
||
* static void
|
||
* test_my_object_test2 (MyObjectFixture *fixture,
|
||
* gconstpointer user_data)
|
||
* {
|
||
* my_object_do_some_work_using_helper (fixture->obj, fixture->helper);
|
||
* g_assert_cmpstr (my_object_get_property (fixture->obj), ==, "updated-value");
|
||
* }
|
||
*
|
||
* int
|
||
* main (int argc, char *argv[])
|
||
* {
|
||
* setlocale (LC_ALL, "");
|
||
*
|
||
* g_test_init (&argc, &argv, NULL);
|
||
*
|
||
* // Define the tests.
|
||
* g_test_add ("/my-object/test1", MyObjectFixture, "some-user-data",
|
||
* my_object_fixture_set_up, test_my_object_test1,
|
||
* my_object_fixture_tear_down);
|
||
* g_test_add ("/my-object/test2", MyObjectFixture, "some-user-data",
|
||
* my_object_fixture_set_up, test_my_object_test2,
|
||
* my_object_fixture_tear_down);
|
||
*
|
||
* return g_test_run ();
|
||
* }
|
||
* ]|
|
||
*
|
||
* ### Integrating GTest in your project
|
||
*
|
||
* If you are using the [Meson](http://mesonbuild.com) build system, you will
|
||
* typically use the provided `test()` primitive to call the test binaries,
|
||
* e.g.:
|
||
*
|
||
* |[<!-- language="plain" -->
|
||
* test(
|
||
* 'foo',
|
||
* executable('foo', 'foo.c', dependencies: deps),
|
||
* env: [
|
||
* 'G_TEST_SRCDIR=@0@'.format(meson.current_source_dir()),
|
||
* 'G_TEST_BUILDDIR=@0@'.format(meson.current_build_dir()),
|
||
* ],
|
||
* )
|
||
*
|
||
* test(
|
||
* 'bar',
|
||
* executable('bar', 'bar.c', dependencies: deps),
|
||
* env: [
|
||
* 'G_TEST_SRCDIR=@0@'.format(meson.current_source_dir()),
|
||
* 'G_TEST_BUILDDIR=@0@'.format(meson.current_build_dir()),
|
||
* ],
|
||
* )
|
||
* ]|
|
||
*
|
||
* If you are using Autotools, you're strongly encouraged to use the Automake
|
||
* [TAP](https://testanything.org/) harness; GLib provides template files for
|
||
* easily integrating with it:
|
||
*
|
||
* - [glib-tap.mk](https://gitlab.gnome.org/GNOME/glib/blob/glib-2-58/glib-tap.mk)
|
||
* - [tap-test](https://gitlab.gnome.org/GNOME/glib/blob/glib-2-58/tap-test)
|
||
* - [tap-driver.sh](https://gitlab.gnome.org/GNOME/glib/blob/glib-2-58/tap-driver.sh)
|
||
*
|
||
* You can copy these files in your own project's root directory, and then
|
||
* set up your `Makefile.am` file to reference them, for instance:
|
||
*
|
||
* |[<!-- language="plain" -->
|
||
* include $(top_srcdir)/glib-tap.mk
|
||
*
|
||
* # test binaries
|
||
* test_programs = \
|
||
* foo \
|
||
* bar
|
||
*
|
||
* # data distributed in the tarball
|
||
* dist_test_data = \
|
||
* foo.data.txt \
|
||
* bar.data.txt
|
||
*
|
||
* # data not distributed in the tarball
|
||
* test_data = \
|
||
* blah.data.txt
|
||
* ]|
|
||
*
|
||
* Make sure to distribute the TAP files, using something like the following
|
||
* in your top-level `Makefile.am`:
|
||
*
|
||
* |[<!-- language="plain" -->
|
||
* EXTRA_DIST += \
|
||
* tap-driver.sh \
|
||
* tap-test
|
||
* ]|
|
||
*
|
||
* `glib-tap.mk` will be distributed implicitly due to being included in a
|
||
* `Makefile.am`. All three files should be added to version control.
|
||
*
|
||
* If you don't have access to the Autotools TAP harness, you can use the
|
||
* [gtester][gtester] and [gtester-report][gtester-report] tools, and use
|
||
* the [glib.mk](https://gitlab.gnome.org/GNOME/glib/blob/glib-2-58/glib.mk)
|
||
* Automake template provided by GLib. Note, however, that since GLib 2.62,
|
||
* [gtester][gtester] and [gtester-report][gtester-report] have been deprecated
|
||
* in favour of using TAP. The `--tap` argument to tests is enabled by default
|
||
* as of GLib 2.62.
|
||
*/
|
||
|
||
/**
|
||
* g_test_initialized:
|
||
*
|
||
* Returns %TRUE if g_test_init() has been called.
|
||
*
|
||
* Returns: %TRUE if g_test_init() has been called.
|
||
*
|
||
* Since: 2.36
|
||
*/
|
||
|
||
/**
|
||
* g_test_quick:
|
||
*
|
||
* Returns %TRUE if tests are run in quick mode.
|
||
* Exactly one of g_test_quick() and g_test_slow() is active in any run;
|
||
* there is no "medium speed".
|
||
*
|
||
* By default, tests are run in quick mode. In tests that use
|
||
* g_test_init(), the options `-m quick`, `-m slow` and `-m thorough`
|
||
* can be used to change this.
|
||
*
|
||
* Returns: %TRUE if in quick mode
|
||
*/
|
||
|
||
/**
|
||
* g_test_slow:
|
||
*
|
||
* Returns %TRUE if tests are run in slow mode.
|
||
* Exactly one of g_test_quick() and g_test_slow() is active in any run;
|
||
* there is no "medium speed".
|
||
*
|
||
* By default, tests are run in quick mode. In tests that use
|
||
* g_test_init(), the options `-m quick`, `-m slow` and `-m thorough`
|
||
* can be used to change this.
|
||
*
|
||
* Returns: the opposite of g_test_quick()
|
||
*/
|
||
|
||
/**
|
||
* g_test_thorough:
|
||
*
|
||
* Returns %TRUE if tests are run in thorough mode, equivalent to
|
||
* g_test_slow().
|
||
*
|
||
* By default, tests are run in quick mode. In tests that use
|
||
* g_test_init(), the options `-m quick`, `-m slow` and `-m thorough`
|
||
* can be used to change this.
|
||
*
|
||
* Returns: the same thing as g_test_slow()
|
||
*/
|
||
|
||
/**
|
||
* g_test_perf:
|
||
*
|
||
* Returns %TRUE if tests are run in performance mode.
|
||
*
|
||
* By default, tests are run in quick mode. In tests that use
|
||
* g_test_init(), the option `-m perf` enables performance tests, while
|
||
* `-m quick` disables them.
|
||
*
|
||
* Returns: %TRUE if in performance mode
|
||
*/
|
||
|
||
/**
|
||
* g_test_undefined:
|
||
*
|
||
* Returns %TRUE if tests may provoke assertions and other formally-undefined
|
||
* behaviour, to verify that appropriate warnings are given. It might, in some
|
||
* cases, be useful to turn this off with if running tests under valgrind;
|
||
* in tests that use g_test_init(), the option `-m no-undefined` disables
|
||
* those tests, while `-m undefined` explicitly enables them (the default
|
||
* behaviour).
|
||
*
|
||
* Returns: %TRUE if tests may provoke programming errors
|
||
*/
|
||
|
||
/**
|
||
* g_test_verbose:
|
||
*
|
||
* Returns %TRUE if tests are run in verbose mode.
|
||
* In tests that use g_test_init(), the option `--verbose` enables this,
|
||
* while `-q` or `--quiet` disables it.
|
||
* The default is neither g_test_verbose() nor g_test_quiet().
|
||
*
|
||
* Returns: %TRUE if in verbose mode
|
||
*/
|
||
|
||
/**
|
||
* g_test_quiet:
|
||
*
|
||
* Returns %TRUE if tests are run in quiet mode.
|
||
* In tests that use g_test_init(), the option `-q` or `--quiet` enables
|
||
* this, while `--verbose` disables it.
|
||
* The default is neither g_test_verbose() nor g_test_quiet().
|
||
*
|
||
* Returns: %TRUE if in quiet mode
|
||
*/
|
||
|
||
/**
|
||
* g_test_queue_unref:
|
||
* @gobject: the object to unref
|
||
*
|
||
* Enqueue an object to be released with g_object_unref() during
|
||
* the next teardown phase. This is equivalent to calling
|
||
* g_test_queue_destroy() with a destroy callback of g_object_unref().
|
||
*
|
||
* Since: 2.16
|
||
*/
|
||
|
||
/**
|
||
* GTestSubprocessFlags:
|
||
* @G_TEST_SUBPROCESS_INHERIT_STDIN: If this flag is given, the child
|
||
* process will inherit the parent's stdin. Otherwise, the child's
|
||
* stdin is redirected to `/dev/null`.
|
||
* @G_TEST_SUBPROCESS_INHERIT_STDOUT: If this flag is given, the child
|
||
* process will inherit the parent's stdout. Otherwise, the child's
|
||
* stdout will not be visible, but it will be captured to allow
|
||
* later tests with g_test_trap_assert_stdout().
|
||
* @G_TEST_SUBPROCESS_INHERIT_STDERR: If this flag is given, the child
|
||
* process will inherit the parent's stderr. Otherwise, the child's
|
||
* stderr will not be visible, but it will be captured to allow
|
||
* later tests with g_test_trap_assert_stderr().
|
||
*
|
||
* Flags to pass to g_test_trap_subprocess() to control input and output.
|
||
*
|
||
* Note that in contrast with g_test_trap_fork(), the default is to
|
||
* not show stdout and stderr.
|
||
*/
|
||
|
||
/**
|
||
* g_test_trap_assert_passed:
|
||
*
|
||
* Assert that the last test subprocess passed.
|
||
* See g_test_trap_subprocess().
|
||
*
|
||
* Since: 2.16
|
||
*/
|
||
|
||
/**
|
||
* g_test_trap_assert_failed:
|
||
*
|
||
* Assert that the last test subprocess failed.
|
||
* See g_test_trap_subprocess().
|
||
*
|
||
* This is sometimes used to test situations that are formally considered to
|
||
* be undefined behaviour, like inputs that fail a g_return_if_fail()
|
||
* check. In these situations you should skip the entire test, including the
|
||
* call to g_test_trap_subprocess(), unless g_test_undefined() returns %TRUE
|
||
* to indicate that undefined behaviour may be tested.
|
||
*
|
||
* Since: 2.16
|
||
*/
|
||
|
||
/**
|
||
* g_test_trap_assert_stdout:
|
||
* @soutpattern: a glob-style [pattern][glib-Glob-style-pattern-matching]
|
||
*
|
||
* Assert that the stdout output of the last test subprocess matches
|
||
* @soutpattern. See g_test_trap_subprocess().
|
||
*
|
||
* Since: 2.16
|
||
*/
|
||
|
||
/**
|
||
* g_test_trap_assert_stdout_unmatched:
|
||
* @soutpattern: a glob-style [pattern][glib-Glob-style-pattern-matching]
|
||
*
|
||
* Assert that the stdout output of the last test subprocess
|
||
* does not match @soutpattern. See g_test_trap_subprocess().
|
||
*
|
||
* Since: 2.16
|
||
*/
|
||
|
||
/**
|
||
* g_test_trap_assert_stderr:
|
||
* @serrpattern: a glob-style [pattern][glib-Glob-style-pattern-matching]
|
||
*
|
||
* Assert that the stderr output of the last test subprocess
|
||
* matches @serrpattern. See g_test_trap_subprocess().
|
||
*
|
||
* This is sometimes used to test situations that are formally
|
||
* considered to be undefined behaviour, like code that hits a
|
||
* g_assert() or g_error(). In these situations you should skip the
|
||
* entire test, including the call to g_test_trap_subprocess(), unless
|
||
* g_test_undefined() returns %TRUE to indicate that undefined
|
||
* behaviour may be tested.
|
||
*
|
||
* Since: 2.16
|
||
*/
|
||
|
||
/**
|
||
* g_test_trap_assert_stderr_unmatched:
|
||
* @serrpattern: a glob-style [pattern][glib-Glob-style-pattern-matching]
|
||
*
|
||
* Assert that the stderr output of the last test subprocess
|
||
* does not match @serrpattern. See g_test_trap_subprocess().
|
||
*
|
||
* Since: 2.16
|
||
*/
|
||
|
||
/**
|
||
* g_test_rand_bit:
|
||
*
|
||
* Get a reproducible random bit (0 or 1), see g_test_rand_int()
|
||
* for details on test case random numbers.
|
||
*
|
||
* Since: 2.16
|
||
*/
|
||
|
||
/**
|
||
* g_assert:
|
||
* @expr: the expression to check
|
||
*
|
||
* Debugging macro to terminate the application if the assertion
|
||
* fails. If the assertion fails (i.e. the expression is not true),
|
||
* an error message is logged and the application is terminated.
|
||
*
|
||
* The macro can be turned off in final releases of code by defining
|
||
* `G_DISABLE_ASSERT` when compiling the application, so code must
|
||
* not depend on any side effects from @expr. Similarly, it must not be used
|
||
* in unit tests, otherwise the unit tests will be ineffective if compiled with
|
||
* `G_DISABLE_ASSERT`. Use g_assert_true() and related macros in unit tests
|
||
* instead.
|
||
*/
|
||
|
||
/**
|
||
* g_assert_not_reached:
|
||
*
|
||
* Debugging macro to terminate the application if it is ever
|
||
* reached. If it is reached, an error message is logged and the
|
||
* application is terminated.
|
||
*
|
||
* The macro can be turned off in final releases of code by defining
|
||
* `G_DISABLE_ASSERT` when compiling the application. Hence, it should not be
|
||
* used in unit tests, where assertions should always be effective.
|
||
*/
|
||
|
||
/**
|
||
* g_assert_true:
|
||
* @expr: the expression to check
|
||
*
|
||
* Debugging macro to check that an expression is true.
|
||
*
|
||
* If the assertion fails (i.e. the expression is not true),
|
||
* an error message is logged and the application is either
|
||
* terminated or the testcase marked as failed.
|
||
*
|
||
* Note that unlike g_assert(), this macro is unaffected by whether
|
||
* `G_DISABLE_ASSERT` is defined. Hence it should only be used in tests and,
|
||
* conversely, g_assert() should not be used in tests.
|
||
*
|
||
* See g_test_set_nonfatal_assertions().
|
||
*
|
||
* Since: 2.38
|
||
*/
|
||
|
||
/**
|
||
* g_assert_false:
|
||
* @expr: the expression to check
|
||
*
|
||
* Debugging macro to check an expression is false.
|
||
*
|
||
* If the assertion fails (i.e. the expression is not false),
|
||
* an error message is logged and the application is either
|
||
* terminated or the testcase marked as failed.
|
||
*
|
||
* Note that unlike g_assert(), this macro is unaffected by whether
|
||
* `G_DISABLE_ASSERT` is defined. Hence it should only be used in tests and,
|
||
* conversely, g_assert() should not be used in tests.
|
||
*
|
||
* See g_test_set_nonfatal_assertions().
|
||
*
|
||
* Since: 2.38
|
||
*/
|
||
|
||
/**
|
||
* g_assert_null:
|
||
* @expr: the expression to check
|
||
*
|
||
* Debugging macro to check an expression is %NULL.
|
||
*
|
||
* If the assertion fails (i.e. the expression is not %NULL),
|
||
* an error message is logged and the application is either
|
||
* terminated or the testcase marked as failed.
|
||
*
|
||
* Note that unlike g_assert(), this macro is unaffected by whether
|
||
* `G_DISABLE_ASSERT` is defined. Hence it should only be used in tests and,
|
||
* conversely, g_assert() should not be used in tests.
|
||
*
|
||
* See g_test_set_nonfatal_assertions().
|
||
*
|
||
* Since: 2.38
|
||
*/
|
||
|
||
/**
|
||
* g_assert_nonnull:
|
||
* @expr: the expression to check
|
||
*
|
||
* Debugging macro to check an expression is not %NULL.
|
||
*
|
||
* If the assertion fails (i.e. the expression is %NULL),
|
||
* an error message is logged and the application is either
|
||
* terminated or the testcase marked as failed.
|
||
*
|
||
* Note that unlike g_assert(), this macro is unaffected by whether
|
||
* `G_DISABLE_ASSERT` is defined. Hence it should only be used in tests and,
|
||
* conversely, g_assert() should not be used in tests.
|
||
*
|
||
* See g_test_set_nonfatal_assertions().
|
||
*
|
||
* Since: 2.40
|
||
*/
|
||
|
||
/**
|
||
* g_assert_cmpstr:
|
||
* @s1: a string (may be %NULL)
|
||
* @cmp: The comparison operator to use.
|
||
* One of `==`, `!=`, `<`, `>`, `<=`, `>=`.
|
||
* @s2: another string (may be %NULL)
|
||
*
|
||
* Debugging macro to compare two strings. If the comparison fails,
|
||
* an error message is logged and the application is either terminated
|
||
* or the testcase marked as failed.
|
||
* The strings are compared using g_strcmp0().
|
||
*
|
||
* The effect of `g_assert_cmpstr (s1, op, s2)` is
|
||
* the same as `g_assert_true (g_strcmp0 (s1, s2) op 0)`.
|
||
* The advantage of this macro is that it can produce a message that
|
||
* includes the actual values of @s1 and @s2.
|
||
*
|
||
* |[<!-- language="C" -->
|
||
* g_assert_cmpstr (mystring, ==, "fubar");
|
||
* ]|
|
||
*
|
||
* Since: 2.16
|
||
*/
|
||
|
||
/**
|
||
* g_assert_cmpint:
|
||
* @n1: an integer
|
||
* @cmp: The comparison operator to use.
|
||
* One of `==`, `!=`, `<`, `>`, `<=`, `>=`.
|
||
* @n2: another integer
|
||
*
|
||
* Debugging macro to compare two integers.
|
||
*
|
||
* The effect of `g_assert_cmpint (n1, op, n2)` is
|
||
* the same as `g_assert_true (n1 op n2)`. The advantage
|
||
* of this macro is that it can produce a message that includes the
|
||
* actual values of @n1 and @n2.
|
||
*
|
||
* Since: 2.16
|
||
*/
|
||
|
||
/**
|
||
* g_assert_cmpuint:
|
||
* @n1: an unsigned integer
|
||
* @cmp: The comparison operator to use.
|
||
* One of `==`, `!=`, `<`, `>`, `<=`, `>=`.
|
||
* @n2: another unsigned integer
|
||
*
|
||
* Debugging macro to compare two unsigned integers.
|
||
*
|
||
* The effect of `g_assert_cmpuint (n1, op, n2)` is
|
||
* the same as `g_assert_true (n1 op n2)`. The advantage
|
||
* of this macro is that it can produce a message that includes the
|
||
* actual values of @n1 and @n2.
|
||
*
|
||
* Since: 2.16
|
||
*/
|
||
|
||
/**
|
||
* g_assert_cmphex:
|
||
* @n1: an unsigned integer
|
||
* @cmp: The comparison operator to use.
|
||
* One of `==`, `!=`, `<`, `>`, `<=`, `>=`.
|
||
* @n2: another unsigned integer
|
||
*
|
||
* Debugging macro to compare to unsigned integers.
|
||
*
|
||
* This is a variant of g_assert_cmpuint() that displays the numbers
|
||
* in hexadecimal notation in the message.
|
||
*
|
||
* Since: 2.16
|
||
*/
|
||
|
||
/**
|
||
* g_assert_cmpfloat:
|
||
* @n1: a floating point number
|
||
* @cmp: The comparison operator to use.
|
||
* One of `==`, `!=`, `<`, `>`, `<=`, `>=`.
|
||
* @n2: another floating point number
|
||
*
|
||
* Debugging macro to compare two floating point numbers.
|
||
*
|
||
* The effect of `g_assert_cmpfloat (n1, op, n2)` is
|
||
* the same as `g_assert_true (n1 op n2)`. The advantage
|
||
* of this macro is that it can produce a message that includes the
|
||
* actual values of @n1 and @n2.
|
||
*
|
||
* Since: 2.16
|
||
*/
|
||
|
||
/**
|
||
* g_assert_cmpfloat_with_epsilon:
|
||
* @n1: a floating point number
|
||
* @n2: another floating point number
|
||
* @epsilon: a numeric value that expresses the expected tolerance
|
||
* between @n1 and @n2
|
||
*
|
||
* Debugging macro to compare two floating point numbers within an epsilon.
|
||
*
|
||
* The effect of `g_assert_cmpfloat_with_epsilon (n1, n2, epsilon)` is
|
||
* the same as `g_assert_true (abs (n1 - n2) < epsilon)`. The advantage
|
||
* of this macro is that it can produce a message that includes the
|
||
* actual values of @n1 and @n2.
|
||
*
|
||
* Since: 2.58
|
||
*/
|
||
|
||
/**
|
||
* g_assert_no_errno:
|
||
* @expr: the expression to check
|
||
*
|
||
* Debugging macro to check that an expression has a non-negative return value,
|
||
* as used by traditional POSIX functions (such as `rmdir()`) to indicate
|
||
* success.
|
||
*
|
||
* If the assertion fails (i.e. the @expr returns a negative value), an error
|
||
* message is logged and the testcase is marked as failed. The error message
|
||
* will contain the value of `errno` and its human-readable message from
|
||
* g_strerror().
|
||
*
|
||
* This macro will clear the value of `errno` before executing @expr.
|
||
*
|
||
* Since: 2.66
|
||
*/
|
||
|
||
/**
|
||
* g_assert_cmpmem:
|
||
* @m1: (nullable): pointer to a buffer
|
||
* @l1: length of @m1
|
||
* @m2: (nullable): pointer to another buffer
|
||
* @l2: length of @m2
|
||
*
|
||
* Debugging macro to compare memory regions. If the comparison fails,
|
||
* an error message is logged and the application is either terminated
|
||
* or the testcase marked as failed.
|
||
*
|
||
* The effect of `g_assert_cmpmem (m1, l1, m2, l2)` is
|
||
* the same as `g_assert_true (l1 == l2 && memcmp (m1, m2, l1) == 0)`.
|
||
* The advantage of this macro is that it can produce a message that
|
||
* includes the actual values of @l1 and @l2.
|
||
*
|
||
* @m1 may be %NULL if (and only if) @l1 is zero; similarly for @m2 and @l2.
|
||
*
|
||
* |[<!-- language="C" -->
|
||
* g_assert_cmpmem (buf->data, buf->len, expected, sizeof (expected));
|
||
* ]|
|
||
*
|
||
* Since: 2.46
|
||
*/
|
||
|
||
/**
|
||
* g_assert_cmpvariant:
|
||
* @v1: pointer to a #GVariant
|
||
* @v2: pointer to another #GVariant
|
||
*
|
||
* Debugging macro to compare two #GVariants. If the comparison fails,
|
||
* an error message is logged and the application is either terminated
|
||
* or the testcase marked as failed. The variants are compared using
|
||
* g_variant_equal().
|
||
*
|
||
* The effect of `g_assert_cmpvariant (v1, v2)` is the same as
|
||
* `g_assert_true (g_variant_equal (v1, v2))`. The advantage of this macro is
|
||
* that it can produce a message that includes the actual values of @v1 and @v2.
|
||
*
|
||
* Since: 2.60
|
||
*/
|
||
|
||
/**
|
||
* g_assert_no_error:
|
||
* @err: a #GError, possibly %NULL
|
||
*
|
||
* Debugging macro to check that a #GError is not set.
|
||
*
|
||
* The effect of `g_assert_no_error (err)` is
|
||
* the same as `g_assert_true (err == NULL)`. The advantage
|
||
* of this macro is that it can produce a message that includes
|
||
* the error message and code.
|
||
*
|
||
* Since: 2.20
|
||
*/
|
||
|
||
/**
|
||
* g_assert_error:
|
||
* @err: a #GError, possibly %NULL
|
||
* @dom: the expected error domain (a #GQuark)
|
||
* @c: the expected error code
|
||
*
|
||
* Debugging macro to check that a method has returned
|
||
* the correct #GError.
|
||
*
|
||
* The effect of `g_assert_error (err, dom, c)` is
|
||
* the same as `g_assert_true (err != NULL && err->domain
|
||
* == dom && err->code == c)`. The advantage of this
|
||
* macro is that it can produce a message that includes the incorrect
|
||
* error message and code.
|
||
*
|
||
* This can only be used to test for a specific error. If you want to
|
||
* test that @err is set, but don't care what it's set to, just use
|
||
* `g_assert_nonnull (err)`.
|
||
*
|
||
* Since: 2.20
|
||
*/
|
||
|
||
/**
|
||
* GTestCase:
|
||
*
|
||
* An opaque structure representing a test case.
|
||
*/
|
||
|
||
/**
|
||
* GTestSuite:
|
||
*
|
||
* An opaque structure representing a test suite.
|
||
*/
|
||
|
||
|
||
/* Global variable for storing assertion messages; this is the counterpart to
|
||
* glibc's (private) __abort_msg variable, and allows developers and crash
|
||
* analysis systems like Apport and ABRT to fish out assertion messages from
|
||
* core dumps, instead of having to catch them on screen output.
|
||
*/
|
||
GLIB_VAR char *__glib_assert_msg;
|
||
char *__glib_assert_msg = NULL;
|
||
|
||
/* --- constants --- */
|
||
#define G_TEST_STATUS_TIMED_OUT 1024
|
||
|
||
/* --- structures --- */
|
||
struct GTestCase
|
||
{
|
||
gchar *name;
|
||
guint fixture_size;
|
||
void (*fixture_setup) (void*, gconstpointer);
|
||
void (*fixture_test) (void*, gconstpointer);
|
||
void (*fixture_teardown) (void*, gconstpointer);
|
||
gpointer test_data;
|
||
};
|
||
struct GTestSuite
|
||
{
|
||
gchar *name;
|
||
GSList *suites;
|
||
GSList *cases;
|
||
};
|
||
typedef struct DestroyEntry DestroyEntry;
|
||
struct DestroyEntry
|
||
{
|
||
DestroyEntry *next;
|
||
GDestroyNotify destroy_func;
|
||
gpointer destroy_data;
|
||
};
|
||
|
||
/* --- prototypes --- */
|
||
static void test_run_seed (const gchar *rseed);
|
||
static void test_trap_clear (void);
|
||
static guint8* g_test_log_dump (GTestLogMsg *msg,
|
||
guint *len);
|
||
static void gtest_default_log_handler (const gchar *log_domain,
|
||
GLogLevelFlags log_level,
|
||
const gchar *message,
|
||
gpointer unused_data);
|
||
|
||
|
||
static const char * const g_test_result_names[] = {
|
||
"OK",
|
||
"SKIP",
|
||
"FAIL",
|
||
"TODO"
|
||
};
|
||
|
||
/* --- variables --- */
|
||
static int test_log_fd = -1;
|
||
static gboolean test_mode_fatal = TRUE;
|
||
static gboolean g_test_run_once = TRUE;
|
||
static gboolean test_isolate_dirs = FALSE;
|
||
static gchar *test_isolate_dirs_tmpdir = NULL;
|
||
static const gchar *test_tmpdir = NULL;
|
||
static gboolean test_run_list = FALSE;
|
||
static gchar *test_run_seedstr = NULL;
|
||
G_LOCK_DEFINE_STATIC (test_run_rand);
|
||
static GRand *test_run_rand = NULL;
|
||
static gchar *test_run_name = "";
|
||
static GSList **test_filename_free_list;
|
||
static guint test_run_forks = 0;
|
||
static guint test_run_count = 0;
|
||
static guint test_count = 0;
|
||
static guint test_skipped_count = 0;
|
||
static GTestResult test_run_success = G_TEST_RUN_FAILURE;
|
||
static gchar *test_run_msg = NULL;
|
||
static guint test_startup_skip_count = 0;
|
||
static GTimer *test_user_timer = NULL;
|
||
static double test_user_stamp = 0;
|
||
static GSList *test_paths = NULL;
|
||
static GSList *test_paths_skipped = NULL;
|
||
static GTestSuite *test_suite_root = NULL;
|
||
static int test_trap_last_status = 0; /* unmodified platform-specific status */
|
||
static GPid test_trap_last_pid = 0;
|
||
static char *test_trap_last_subprocess = NULL;
|
||
static char *test_trap_last_stdout = NULL;
|
||
static char *test_trap_last_stderr = NULL;
|
||
static char *test_uri_base = NULL;
|
||
static gboolean test_debug_log = FALSE;
|
||
static gboolean test_tap_log = TRUE; /* default to TAP as of GLib 2.62; see #1619; the non-TAP output mode is deprecated */
|
||
static gboolean test_nonfatal_assertions = FALSE;
|
||
static DestroyEntry *test_destroy_queue = NULL;
|
||
static char *test_argv0 = NULL;
|
||
static char *test_argv0_dirname;
|
||
static const char *test_disted_files_dir;
|
||
static const char *test_built_files_dir;
|
||
static char *test_initial_cwd = NULL;
|
||
static gboolean test_in_forked_child = FALSE;
|
||
static gboolean test_in_subprocess = FALSE;
|
||
static GTestConfig mutable_test_config_vars = {
|
||
FALSE, /* test_initialized */
|
||
TRUE, /* test_quick */
|
||
FALSE, /* test_perf */
|
||
FALSE, /* test_verbose */
|
||
FALSE, /* test_quiet */
|
||
TRUE, /* test_undefined */
|
||
};
|
||
const GTestConfig * const g_test_config_vars = &mutable_test_config_vars;
|
||
static gboolean no_g_set_prgname = FALSE;
|
||
|
||
/* --- functions --- */
|
||
const char*
|
||
g_test_log_type_name (GTestLogType log_type)
|
||
{
|
||
switch (log_type)
|
||
{
|
||
case G_TEST_LOG_NONE: return "none";
|
||
case G_TEST_LOG_ERROR: return "error";
|
||
case G_TEST_LOG_START_BINARY: return "binary";
|
||
case G_TEST_LOG_LIST_CASE: return "list";
|
||
case G_TEST_LOG_SKIP_CASE: return "skip";
|
||
case G_TEST_LOG_START_CASE: return "start";
|
||
case G_TEST_LOG_STOP_CASE: return "stop";
|
||
case G_TEST_LOG_MIN_RESULT: return "minperf";
|
||
case G_TEST_LOG_MAX_RESULT: return "maxperf";
|
||
case G_TEST_LOG_MESSAGE: return "message";
|
||
case G_TEST_LOG_START_SUITE: return "start suite";
|
||
case G_TEST_LOG_STOP_SUITE: return "stop suite";
|
||
}
|
||
return "???";
|
||
}
|
||
|
||
static void
|
||
g_test_log_send (guint n_bytes,
|
||
const guint8 *buffer)
|
||
{
|
||
if (test_log_fd >= 0)
|
||
{
|
||
int r;
|
||
do
|
||
r = write (test_log_fd, buffer, n_bytes);
|
||
while (r < 0 && errno == EINTR);
|
||
}
|
||
if (test_debug_log)
|
||
{
|
||
GTestLogBuffer *lbuffer = g_test_log_buffer_new ();
|
||
GTestLogMsg *msg;
|
||
guint ui;
|
||
g_test_log_buffer_push (lbuffer, n_bytes, buffer);
|
||
msg = g_test_log_buffer_pop (lbuffer);
|
||
g_warn_if_fail (msg != NULL);
|
||
g_warn_if_fail (lbuffer->data->len == 0);
|
||
g_test_log_buffer_free (lbuffer);
|
||
/* print message */
|
||
g_printerr ("{*LOG(%s)", g_test_log_type_name (msg->log_type));
|
||
for (ui = 0; ui < msg->n_strings; ui++)
|
||
g_printerr (":{%s}", msg->strings[ui]);
|
||
if (msg->n_nums)
|
||
{
|
||
g_printerr (":(");
|
||
for (ui = 0; ui < msg->n_nums; ui++)
|
||
{
|
||
if ((long double) (long) msg->nums[ui] == msg->nums[ui])
|
||
g_printerr ("%s%ld", ui ? ";" : "", (long) msg->nums[ui]);
|
||
else
|
||
g_printerr ("%s%.16g", ui ? ";" : "", (double) msg->nums[ui]);
|
||
}
|
||
g_printerr (")");
|
||
}
|
||
g_printerr (":LOG*}\n");
|
||
g_test_log_msg_free (msg);
|
||
}
|
||
}
|
||
|
||
static void
|
||
g_test_log (GTestLogType lbit,
|
||
const gchar *string1,
|
||
const gchar *string2,
|
||
guint n_args,
|
||
long double *largs)
|
||
{
|
||
GTestResult result;
|
||
gboolean fail;
|
||
GTestLogMsg msg;
|
||
gchar *astrings[3] = { NULL, NULL, NULL };
|
||
guint8 *dbuffer;
|
||
guint32 dbufferlen;
|
||
|
||
switch (lbit)
|
||
{
|
||
case G_TEST_LOG_START_BINARY:
|
||
if (test_tap_log)
|
||
g_print ("# random seed: %s\n", string2);
|
||
else if (g_test_verbose ())
|
||
g_print ("GTest: random seed: %s\n", string2);
|
||
break;
|
||
case G_TEST_LOG_START_SUITE:
|
||
if (test_tap_log)
|
||
{
|
||
/* We only print the TAP "plan" (1..n) ahead of time if we did
|
||
* not use the -p option to select specific tests to be run. */
|
||
if (string1[0] != 0)
|
||
g_print ("# Start of %s tests\n", string1);
|
||
else if (test_paths == NULL)
|
||
g_print ("1..%d\n", test_count);
|
||
}
|
||
break;
|
||
case G_TEST_LOG_STOP_SUITE:
|
||
if (test_tap_log)
|
||
{
|
||
/* If we didn't print the TAP "plan" at the beginning because
|
||
* we were using -p, we need to print how many tests we ran at
|
||
* the end instead. */
|
||
if (string1[0] != 0)
|
||
g_print ("# End of %s tests\n", string1);
|
||
else if (test_paths != NULL)
|
||
g_print ("1..%d\n", test_run_count);
|
||
}
|
||
break;
|
||
case G_TEST_LOG_STOP_CASE:
|
||
result = largs[0];
|
||
fail = result == G_TEST_RUN_FAILURE;
|
||
if (test_tap_log)
|
||
{
|
||
const gchar *ok;
|
||
|
||
/* The TAP representation for an expected failure starts with
|
||
* "not ok", even though it does not actually count as failing
|
||
* due to the use of the TODO directive. "ok # TODO" would mean
|
||
* a test that was expected to fail unexpectedly succeeded,
|
||
* for which GTestResult does not currently have a
|
||
* representation. */
|
||
if (fail || result == G_TEST_RUN_INCOMPLETE)
|
||
ok = "not ok";
|
||
else
|
||
ok = "ok";
|
||
|
||
g_print ("%s %d %s", ok, test_run_count, string1);
|
||
if (result == G_TEST_RUN_INCOMPLETE)
|
||
g_print (" # TODO %s\n", string2 ? string2 : "");
|
||
else if (result == G_TEST_RUN_SKIPPED)
|
||
g_print (" # SKIP %s\n", string2 ? string2 : "");
|
||
else
|
||
g_print ("\n");
|
||
}
|
||
else if (g_test_verbose ())
|
||
g_print ("GTest: result: %s\n", g_test_result_names[result]);
|
||
else if (!g_test_quiet ())
|
||
g_print ("%s\n", g_test_result_names[result]);
|
||
if (fail && test_mode_fatal)
|
||
{
|
||
if (test_tap_log)
|
||
g_print ("Bail out!\n");
|
||
g_abort ();
|
||
}
|
||
if (result == G_TEST_RUN_SKIPPED || result == G_TEST_RUN_INCOMPLETE)
|
||
test_skipped_count++;
|
||
break;
|
||
case G_TEST_LOG_SKIP_CASE:
|
||
if (test_tap_log)
|
||
g_print ("ok %d %s # SKIP\n", test_run_count, string1);
|
||
break;
|
||
case G_TEST_LOG_MIN_RESULT:
|
||
if (test_tap_log)
|
||
g_print ("# min perf: %s\n", string1);
|
||
else if (g_test_verbose ())
|
||
g_print ("(MINPERF:%s)\n", string1);
|
||
break;
|
||
case G_TEST_LOG_MAX_RESULT:
|
||
if (test_tap_log)
|
||
g_print ("# max perf: %s\n", string1);
|
||
else if (g_test_verbose ())
|
||
g_print ("(MAXPERF:%s)\n", string1);
|
||
break;
|
||
case G_TEST_LOG_MESSAGE:
|
||
if (test_tap_log)
|
||
g_print ("# %s\n", string1);
|
||
else if (g_test_verbose ())
|
||
g_print ("(MSG: %s)\n", string1);
|
||
break;
|
||
case G_TEST_LOG_ERROR:
|
||
if (test_tap_log)
|
||
g_print ("Bail out! %s\n", string1);
|
||
else if (g_test_verbose ())
|
||
g_print ("(ERROR: %s)\n", string1);
|
||
break;
|
||
default: ;
|
||
}
|
||
|
||
msg.log_type = lbit;
|
||
msg.n_strings = (string1 != NULL) + (string1 && string2);
|
||
msg.strings = astrings;
|
||
astrings[0] = (gchar*) string1;
|
||
astrings[1] = astrings[0] ? (gchar*) string2 : NULL;
|
||
msg.n_nums = n_args;
|
||
msg.nums = largs;
|
||
dbuffer = g_test_log_dump (&msg, &dbufferlen);
|
||
g_test_log_send (dbufferlen, dbuffer);
|
||
g_free (dbuffer);
|
||
|
||
switch (lbit)
|
||
{
|
||
case G_TEST_LOG_START_CASE:
|
||
if (test_tap_log)
|
||
;
|
||
else if (g_test_verbose ())
|
||
g_print ("GTest: run: %s\n", string1);
|
||
else if (!g_test_quiet ())
|
||
g_print ("%s: ", string1);
|
||
break;
|
||
default: ;
|
||
}
|
||
}
|
||
|
||
/* We intentionally parse the command line without GOptionContext
|
||
* because otherwise you would never be able to test it.
|
||
*/
|
||
static void
|
||
parse_args (gint *argc_p,
|
||
gchar ***argv_p)
|
||
{
|
||
guint argc = *argc_p;
|
||
gchar **argv = *argv_p;
|
||
guint i, e;
|
||
|
||
test_argv0 = argv[0];
|
||
test_initial_cwd = g_get_current_dir ();
|
||
|
||
/* parse known args */
|
||
for (i = 1; i < argc; i++)
|
||
{
|
||
if (strcmp (argv[i], "--g-fatal-warnings") == 0)
|
||
{
|
||
GLogLevelFlags fatal_mask = (GLogLevelFlags) g_log_set_always_fatal ((GLogLevelFlags) G_LOG_FATAL_MASK);
|
||
fatal_mask = (GLogLevelFlags) (fatal_mask | G_LOG_LEVEL_WARNING | G_LOG_LEVEL_CRITICAL);
|
||
g_log_set_always_fatal (fatal_mask);
|
||
argv[i] = NULL;
|
||
}
|
||
else if (strcmp (argv[i], "--keep-going") == 0 ||
|
||
strcmp (argv[i], "-k") == 0)
|
||
{
|
||
test_mode_fatal = FALSE;
|
||
argv[i] = NULL;
|
||
}
|
||
else if (strcmp (argv[i], "--debug-log") == 0)
|
||
{
|
||
test_debug_log = TRUE;
|
||
argv[i] = NULL;
|
||
}
|
||
else if (strcmp (argv[i], "--tap") == 0)
|
||
{
|
||
test_tap_log = TRUE;
|
||
argv[i] = NULL;
|
||
}
|
||
else if (strcmp ("--GTestLogFD", argv[i]) == 0 || strncmp ("--GTestLogFD=", argv[i], 13) == 0)
|
||
{
|
||
gchar *equal = argv[i] + 12;
|
||
if (*equal == '=')
|
||
test_log_fd = g_ascii_strtoull (equal + 1, NULL, 0);
|
||
else if (i + 1 < argc)
|
||
{
|
||
argv[i++] = NULL;
|
||
test_log_fd = g_ascii_strtoull (argv[i], NULL, 0);
|
||
}
|
||
argv[i] = NULL;
|
||
|
||
/* Force non-TAP output when using gtester */
|
||
test_tap_log = FALSE;
|
||
}
|
||
else if (strcmp ("--GTestSkipCount", argv[i]) == 0 || strncmp ("--GTestSkipCount=", argv[i], 17) == 0)
|
||
{
|
||
gchar *equal = argv[i] + 16;
|
||
if (*equal == '=')
|
||
test_startup_skip_count = g_ascii_strtoull (equal + 1, NULL, 0);
|
||
else if (i + 1 < argc)
|
||
{
|
||
argv[i++] = NULL;
|
||
test_startup_skip_count = g_ascii_strtoull (argv[i], NULL, 0);
|
||
}
|
||
argv[i] = NULL;
|
||
}
|
||
else if (strcmp ("--GTestSubprocess", argv[i]) == 0)
|
||
{
|
||
test_in_subprocess = TRUE;
|
||
/* We typically expect these child processes to crash, and some
|
||
* tests spawn a *lot* of them. Avoid spamming system crash
|
||
* collection programs such as systemd-coredump and abrt.
|
||
*/
|
||
#ifdef HAVE_SYS_RESOURCE_H
|
||
{
|
||
struct rlimit limit = { 0, 0 };
|
||
(void) setrlimit (RLIMIT_CORE, &limit);
|
||
}
|
||
#endif
|
||
argv[i] = NULL;
|
||
|
||
/* Force non-TAP output when spawning a subprocess, since people often
|
||
* test the stdout/stderr of the subprocess strictly */
|
||
test_tap_log = FALSE;
|
||
}
|
||
else if (strcmp ("-p", argv[i]) == 0 || strncmp ("-p=", argv[i], 3) == 0)
|
||
{
|
||
gchar *equal = argv[i] + 2;
|
||
if (*equal == '=')
|
||
test_paths = g_slist_prepend (test_paths, equal + 1);
|
||
else if (i + 1 < argc)
|
||
{
|
||
argv[i++] = NULL;
|
||
test_paths = g_slist_prepend (test_paths, argv[i]);
|
||
}
|
||
argv[i] = NULL;
|
||
}
|
||
else if (strcmp ("-s", argv[i]) == 0 || strncmp ("-s=", argv[i], 3) == 0)
|
||
{
|
||
gchar *equal = argv[i] + 2;
|
||
if (*equal == '=')
|
||
test_paths_skipped = g_slist_prepend (test_paths_skipped, equal + 1);
|
||
else if (i + 1 < argc)
|
||
{
|
||
argv[i++] = NULL;
|
||
test_paths_skipped = g_slist_prepend (test_paths_skipped, argv[i]);
|
||
}
|
||
argv[i] = NULL;
|
||
}
|
||
else if (strcmp ("-m", argv[i]) == 0 || strncmp ("-m=", argv[i], 3) == 0)
|
||
{
|
||
gchar *equal = argv[i] + 2;
|
||
const gchar *mode = "";
|
||
if (*equal == '=')
|
||
mode = equal + 1;
|
||
else if (i + 1 < argc)
|
||
{
|
||
argv[i++] = NULL;
|
||
mode = argv[i];
|
||
}
|
||
if (strcmp (mode, "perf") == 0)
|
||
mutable_test_config_vars.test_perf = TRUE;
|
||
else if (strcmp (mode, "slow") == 0)
|
||
mutable_test_config_vars.test_quick = FALSE;
|
||
else if (strcmp (mode, "thorough") == 0)
|
||
mutable_test_config_vars.test_quick = FALSE;
|
||
else if (strcmp (mode, "quick") == 0)
|
||
{
|
||
mutable_test_config_vars.test_quick = TRUE;
|
||
mutable_test_config_vars.test_perf = FALSE;
|
||
}
|
||
else if (strcmp (mode, "undefined") == 0)
|
||
mutable_test_config_vars.test_undefined = TRUE;
|
||
else if (strcmp (mode, "no-undefined") == 0)
|
||
mutable_test_config_vars.test_undefined = FALSE;
|
||
else
|
||
g_error ("unknown test mode: -m %s", mode);
|
||
argv[i] = NULL;
|
||
}
|
||
else if (strcmp ("-q", argv[i]) == 0 || strcmp ("--quiet", argv[i]) == 0)
|
||
{
|
||
mutable_test_config_vars.test_quiet = TRUE;
|
||
mutable_test_config_vars.test_verbose = FALSE;
|
||
argv[i] = NULL;
|
||
}
|
||
else if (strcmp ("--verbose", argv[i]) == 0)
|
||
{
|
||
mutable_test_config_vars.test_quiet = FALSE;
|
||
mutable_test_config_vars.test_verbose = TRUE;
|
||
argv[i] = NULL;
|
||
}
|
||
else if (strcmp ("-l", argv[i]) == 0)
|
||
{
|
||
test_run_list = TRUE;
|
||
argv[i] = NULL;
|
||
}
|
||
else if (strcmp ("--seed", argv[i]) == 0 || strncmp ("--seed=", argv[i], 7) == 0)
|
||
{
|
||
gchar *equal = argv[i] + 6;
|
||
if (*equal == '=')
|
||
test_run_seedstr = equal + 1;
|
||
else if (i + 1 < argc)
|
||
{
|
||
argv[i++] = NULL;
|
||
test_run_seedstr = argv[i];
|
||
}
|
||
argv[i] = NULL;
|
||
}
|
||
else if (strcmp ("-?", argv[i]) == 0 ||
|
||
strcmp ("-h", argv[i]) == 0 ||
|
||
strcmp ("--help", argv[i]) == 0)
|
||
{
|
||
printf ("Usage:\n"
|
||
" %s [OPTION...]\n\n"
|
||
"Help Options:\n"
|
||
" -h, --help Show help options\n\n"
|
||
"Test Options:\n"
|
||
" --g-fatal-warnings Make all warnings fatal\n"
|
||
" -l List test cases available in a test executable\n"
|
||
" -m {perf|slow|thorough|quick} Execute tests according to mode\n"
|
||
" -m {undefined|no-undefined} Execute tests according to mode\n"
|
||
" -p TESTPATH Only start test cases matching TESTPATH\n"
|
||
" -s TESTPATH Skip all tests matching TESTPATH\n"
|
||
" --seed=SEEDSTRING Start tests with random seed SEEDSTRING\n"
|
||
" --debug-log debug test logging output\n"
|
||
" -q, --quiet Run tests quietly\n"
|
||
" --verbose Run tests verbosely\n",
|
||
argv[0]);
|
||
exit (0);
|
||
}
|
||
}
|
||
|
||
/* We've been prepending to test_paths, but its order matters, so
|
||
* permute it */
|
||
test_paths = g_slist_reverse (test_paths);
|
||
|
||
/* collapse argv */
|
||
e = 1;
|
||
for (i = 1; i < argc; i++)
|
||
if (argv[i])
|
||
{
|
||
argv[e++] = argv[i];
|
||
if (i >= e)
|
||
argv[i] = NULL;
|
||
}
|
||
*argc_p = e;
|
||
}
|
||
|
||
/* A fairly naive `rm -rf` implementation to clean up after unit tests. */
|
||
static void
|
||
rm_rf (const gchar *path)
|
||
{
|
||
GDir *dir = NULL;
|
||
const gchar *entry;
|
||
|
||
dir = g_dir_open (path, 0, NULL);
|
||
if (dir == NULL)
|
||
{
|
||
/* Assume it’s a file. */
|
||
g_remove (path);
|
||
return;
|
||
}
|
||
|
||
while ((entry = g_dir_read_name (dir)) != NULL)
|
||
{
|
||
gchar *sub_path = g_build_filename (path, entry, NULL);
|
||
rm_rf (sub_path);
|
||
g_free (sub_path);
|
||
}
|
||
|
||
g_dir_close (dir);
|
||
|
||
g_rmdir (path);
|
||
}
|
||
|
||
/* Implement the %G_TEST_OPTION_ISOLATE_DIRS option, iff it’s enabled. Create
|
||
* a temporary directory for this unit test (disambiguated using @test_run_name)
|
||
* and use g_set_user_dirs() to point various XDG directories into it, without
|
||
* having to call setenv() in a process which potentially has threads running.
|
||
*
|
||
* Note that this is called for each unit test, and hence won’t have taken
|
||
* effect before g_test_run() is called in the unit test’s main(). Hence
|
||
* references to XDG variables in main() will not be using the temporary
|
||
* directory. */
|
||
static gboolean
|
||
test_do_isolate_dirs (GError **error)
|
||
{
|
||
gchar *subdir = NULL;
|
||
gchar *home_dir = NULL, *cache_dir = NULL, *config_dir = NULL;
|
||
gchar *data_dir = NULL, *runtime_dir = NULL;
|
||
gchar *config_dirs[3];
|
||
gchar *data_dirs[3];
|
||
|
||
if (!test_isolate_dirs)
|
||
return TRUE;
|
||
|
||
/* The @test_run_name includes the test suites, so may be several directories
|
||
* deep. Add a `.dirs` directory to contain all the paths we create, and
|
||
* guarantee none of them clash with test paths below the current one — test
|
||
* paths may not contain components starting with `.`. */
|
||
subdir = g_build_filename (test_tmpdir, test_run_name, ".dirs", NULL);
|
||
|
||
/* We have to create the runtime directory (because it must be bound to
|
||
* the session lifetime, which we consider to be the lifetime of the unit
|
||
* test for testing purposes — see
|
||
* https://standards.freedesktop.org/basedir-spec/basedir-spec-latest.html.
|
||
* We don’t need to create the other directories — the specification
|
||
* requires that client code create them if they don’t exist. Not creating
|
||
* them automatically is a good test of clients’ adherence to the spec
|
||
* and error handling of missing directories. */
|
||
runtime_dir = g_build_filename (subdir, "runtime", NULL);
|
||
if (g_mkdir_with_parents (runtime_dir, 0700) != 0)
|
||
{
|
||
gint saved_errno = errno;
|
||
g_set_error (error, G_FILE_ERROR, g_file_error_from_errno (saved_errno),
|
||
"Failed to create XDG_RUNTIME_DIR ‘%s’: %s",
|
||
runtime_dir, g_strerror (saved_errno));
|
||
g_free (runtime_dir);
|
||
g_free (subdir);
|
||
return FALSE;
|
||
}
|
||
|
||
home_dir = g_build_filename (subdir, "home", NULL);
|
||
cache_dir = g_build_filename (subdir, "cache", NULL);
|
||
config_dir = g_build_filename (subdir, "config", NULL);
|
||
data_dir = g_build_filename (subdir, "data", NULL);
|
||
|
||
config_dirs[0] = g_build_filename (subdir, "system-config1", NULL);
|
||
config_dirs[1] = g_build_filename (subdir, "system-config2", NULL);
|
||
config_dirs[2] = NULL;
|
||
|
||
data_dirs[0] = g_build_filename (subdir, "system-data1", NULL);
|
||
data_dirs[1] = g_build_filename (subdir, "system-data2", NULL);
|
||
data_dirs[2] = NULL;
|
||
|
||
/* Remember to update the documentation for %G_TEST_OPTION_ISOLATE_DIRS if
|
||
* this list changes. */
|
||
g_set_user_dirs ("HOME", home_dir,
|
||
"XDG_CACHE_HOME", cache_dir,
|
||
"XDG_CONFIG_DIRS", config_dirs,
|
||
"XDG_CONFIG_HOME", config_dir,
|
||
"XDG_DATA_DIRS", data_dirs,
|
||
"XDG_DATA_HOME", data_dir,
|
||
"XDG_RUNTIME_DIR", runtime_dir,
|
||
NULL);
|
||
|
||
g_free (runtime_dir);
|
||
g_free (data_dir);
|
||
g_free (config_dir);
|
||
g_free (cache_dir);
|
||
g_free (home_dir);
|
||
g_free (data_dirs[1]);
|
||
g_free (data_dirs[0]);
|
||
g_free (config_dirs[1]);
|
||
g_free (config_dirs[0]);
|
||
g_free (subdir);
|
||
|
||
return TRUE;
|
||
}
|
||
|
||
/* Clean up after test_do_isolate_dirs(). */
|
||
static void
|
||
test_rm_isolate_dirs (void)
|
||
{
|
||
gchar *subdir = NULL;
|
||
|
||
if (!test_isolate_dirs)
|
||
return;
|
||
|
||
subdir = g_build_filename (test_tmpdir, test_run_name, NULL);
|
||
rm_rf (subdir);
|
||
g_free (subdir);
|
||
}
|
||
|
||
/**
|
||
* g_test_init:
|
||
* @argc: Address of the @argc parameter of the main() function.
|
||
* Changed if any arguments were handled.
|
||
* @argv: Address of the @argv parameter of main().
|
||
* Any parameters understood by g_test_init() stripped before return.
|
||
* @...: %NULL-terminated list of special options, documented below.
|
||
*
|
||
* Initialize the GLib testing framework, e.g. by seeding the
|
||
* test random number generator, the name for g_get_prgname()
|
||
* and parsing test related command line args.
|
||
*
|
||
* So far, the following arguments are understood:
|
||
*
|
||
* - `-l`: List test cases available in a test executable.
|
||
* - `--seed=SEED`: Provide a random seed to reproduce test
|
||
* runs using random numbers.
|
||
* - `--verbose`: Run tests verbosely.
|
||
* - `-q`, `--quiet`: Run tests quietly.
|
||
* - `-p PATH`: Execute all tests matching the given path.
|
||
* - `-s PATH`: Skip all tests matching the given path.
|
||
* This can also be used to force a test to run that would otherwise
|
||
* be skipped (ie, a test whose name contains "/subprocess").
|
||
* - `-m {perf|slow|thorough|quick|undefined|no-undefined}`: Execute tests according to these test modes:
|
||
*
|
||
* `perf`: Performance tests, may take long and report results (off by default).
|
||
*
|
||
* `slow`, `thorough`: Slow and thorough tests, may take quite long and maximize coverage
|
||
* (off by default).
|
||
*
|
||
* `quick`: Quick tests, should run really quickly and give good coverage (the default).
|
||
*
|
||
* `undefined`: Tests for undefined behaviour, may provoke programming errors
|
||
* under g_test_trap_subprocess() or g_test_expect_message() to check
|
||
* that appropriate assertions or warnings are given (the default).
|
||
*
|
||
* `no-undefined`: Avoid tests for undefined behaviour
|
||
*
|
||
* - `--debug-log`: Debug test logging output.
|
||
*
|
||
* Options which can be passed to @... are:
|
||
*
|
||
* - `"no_g_set_prgname"`: Causes g_test_init() to not call g_set_prgname().
|
||
* - %G_TEST_OPTION_ISOLATE_DIRS: Creates a unique temporary directory for each
|
||
* unit test and uses g_set_user_dirs() to set XDG directories to point into
|
||
* that temporary directory for the duration of the unit test. See the
|
||
* documentation for %G_TEST_OPTION_ISOLATE_DIRS.
|
||
*
|
||
* Since 2.58, if tests are compiled with `G_DISABLE_ASSERT` defined,
|
||
* g_test_init() will print an error and exit. This is to prevent no-op tests
|
||
* from being executed, as g_assert() is commonly (erroneously) used in unit
|
||
* tests, and is a no-op when compiled with `G_DISABLE_ASSERT`. Ensure your
|
||
* tests are compiled without `G_DISABLE_ASSERT` defined.
|
||
*
|
||
* Since: 2.16
|
||
*/
|
||
void
|
||
(g_test_init) (int *argc,
|
||
char ***argv,
|
||
...)
|
||
{
|
||
static char seedstr[4 + 4 * 8 + 1];
|
||
va_list args;
|
||
gpointer option;
|
||
/* make warnings and criticals fatal for all test programs */
|
||
GLogLevelFlags fatal_mask = (GLogLevelFlags) g_log_set_always_fatal ((GLogLevelFlags) G_LOG_FATAL_MASK);
|
||
|
||
fatal_mask = (GLogLevelFlags) (fatal_mask | G_LOG_LEVEL_WARNING | G_LOG_LEVEL_CRITICAL);
|
||
g_log_set_always_fatal (fatal_mask);
|
||
/* check caller args */
|
||
g_return_if_fail (argc != NULL);
|
||
g_return_if_fail (argv != NULL);
|
||
g_return_if_fail (g_test_config_vars->test_initialized == FALSE);
|
||
mutable_test_config_vars.test_initialized = TRUE;
|
||
|
||
va_start (args, argv);
|
||
while ((option = va_arg (args, char *)))
|
||
{
|
||
if (g_strcmp0 (option, "no_g_set_prgname") == 0)
|
||
no_g_set_prgname = TRUE;
|
||
else if (g_strcmp0 (option, G_TEST_OPTION_ISOLATE_DIRS) == 0)
|
||
test_isolate_dirs = TRUE;
|
||
}
|
||
va_end (args);
|
||
|
||
/* setup random seed string */
|
||
g_snprintf (seedstr, sizeof (seedstr), "R02S%08x%08x%08x%08x", g_random_int(), g_random_int(), g_random_int(), g_random_int());
|
||
test_run_seedstr = seedstr;
|
||
|
||
/* parse args, sets up mode, changes seed, etc. */
|
||
parse_args (argc, argv);
|
||
|
||
if (!g_get_prgname() && !no_g_set_prgname)
|
||
g_set_prgname ((*argv)[0]);
|
||
|
||
/* Set up the temporary directory for isolating the test. We have to do this
|
||
* early, as we want the return values from g_get_user_data_dir() (and
|
||
* friends) to return subdirectories of the temporary directory throughout
|
||
* the setup function, test, and teardown function, for each unit test.
|
||
* See test_do_isolate_dirs().
|
||
*
|
||
* The directory is deleted at the bottom of g_test_run().
|
||
*
|
||
* Rather than setting the XDG_* environment variables we use a new
|
||
* G_TEST_TMPDIR variable which gives the top-level temporary directory. This
|
||
* allows test subprocesses to reuse the same temporary directory when
|
||
* g_test_init() is called in them. */
|
||
if (test_isolate_dirs)
|
||
{
|
||
if (g_getenv ("G_TEST_TMPDIR") == NULL)
|
||
{
|
||
gchar *test_prgname = NULL;
|
||
gchar *tmpl = NULL;
|
||
GError *local_error = NULL;
|
||
|
||
test_prgname = g_path_get_basename (g_get_prgname ());
|
||
if (*test_prgname == '\0')
|
||
{
|
||
g_free (test_prgname);
|
||
test_prgname = g_strdup ("unknown");
|
||
}
|
||
tmpl = g_strdup_printf ("test_%s_XXXXXX", test_prgname);
|
||
g_free (test_prgname);
|
||
|
||
test_isolate_dirs_tmpdir = g_dir_make_tmp (tmpl, &local_error);
|
||
if (local_error != NULL)
|
||
{
|
||
g_printerr ("%s: Failed to create temporary directory: %s\n",
|
||
(*argv)[0], local_error->message);
|
||
g_error_free (local_error);
|
||
exit (1);
|
||
}
|
||
g_free (tmpl);
|
||
|
||
/* Propagate the temporary directory to subprocesses. */
|
||
g_setenv ("G_TEST_TMPDIR", test_isolate_dirs_tmpdir, TRUE);
|
||
|
||
/* And clear the traditional environment variables so subprocesses
|
||
* spawned by the code under test can’t trash anything. If a test
|
||
* spawns a process, the test is responsible for propagating
|
||
* appropriate environment variables.
|
||
*
|
||
* We assume that any in-process code will use g_get_user_data_dir()
|
||
* and friends, rather than getenv() directly.
|
||
*
|
||
* We set them to ‘/dev/null’ as that should fairly obviously not
|
||
* accidentally work, and should be fairly greppable. */
|
||
{
|
||
const gchar *overridden_environment_variables[] =
|
||
{
|
||
"HOME",
|
||
"XDG_CACHE_HOME",
|
||
"XDG_CONFIG_DIRS",
|
||
"XDG_CONFIG_HOME",
|
||
"XDG_DATA_DIRS",
|
||
"XDG_DATA_HOME",
|
||
"XDG_RUNTIME_DIR",
|
||
};
|
||
gsize i;
|
||
|
||
for (i = 0; i < G_N_ELEMENTS (overridden_environment_variables); i++)
|
||
g_setenv (overridden_environment_variables[i], "/dev/null", TRUE);
|
||
}
|
||
}
|
||
|
||
/* Cache this for the remainder of this process’ lifetime. */
|
||
test_tmpdir = g_getenv ("G_TEST_TMPDIR");
|
||
}
|
||
|
||
/* verify GRand reliability, needed for reliable seeds */
|
||
if (1)
|
||
{
|
||
GRand *rg = g_rand_new_with_seed (0xc8c49fb6);
|
||
guint32 t1 = g_rand_int (rg), t2 = g_rand_int (rg), t3 = g_rand_int (rg), t4 = g_rand_int (rg);
|
||
/* g_print ("GRand-current: 0x%x 0x%x 0x%x 0x%x\n", t1, t2, t3, t4); */
|
||
if (t1 != 0xfab39f9b || t2 != 0xb948fb0e || t3 != 0x3d31be26 || t4 != 0x43a19d66)
|
||
g_warning ("random numbers are not GRand-2.2 compatible, seeds may be broken (check $G_RANDOM_VERSION)");
|
||
g_rand_free (rg);
|
||
}
|
||
|
||
/* check rand seed */
|
||
test_run_seed (test_run_seedstr);
|
||
|
||
/* report program start */
|
||
g_log_set_default_handler (gtest_default_log_handler, NULL);
|
||
g_test_log (G_TEST_LOG_START_BINARY, g_get_prgname(), test_run_seedstr, 0, NULL);
|
||
|
||
test_argv0_dirname = g_path_get_dirname (test_argv0);
|
||
|
||
/* Make sure we get the real dirname that the test was run from */
|
||
if (g_str_has_suffix (test_argv0_dirname, "/.libs"))
|
||
{
|
||
gchar *tmp;
|
||
tmp = g_path_get_dirname (test_argv0_dirname);
|
||
g_free (test_argv0_dirname);
|
||
test_argv0_dirname = tmp;
|
||
}
|
||
|
||
test_disted_files_dir = g_getenv ("G_TEST_SRCDIR");
|
||
if (!test_disted_files_dir)
|
||
test_disted_files_dir = test_argv0_dirname;
|
||
|
||
test_built_files_dir = g_getenv ("G_TEST_BUILDDIR");
|
||
if (!test_built_files_dir)
|
||
test_built_files_dir = test_argv0_dirname;
|
||
}
|
||
|
||
static void
|
||
test_run_seed (const gchar *rseed)
|
||
{
|
||
guint seed_failed = 0;
|
||
if (test_run_rand)
|
||
g_rand_free (test_run_rand);
|
||
test_run_rand = NULL;
|
||
while (strchr (" \t\v\r\n\f", *rseed))
|
||
rseed++;
|
||
if (strncmp (rseed, "R02S", 4) == 0) /* seed for random generator 02 (GRand-2.2) */
|
||
{
|
||
const char *s = rseed + 4;
|
||
if (strlen (s) >= 32) /* require 4 * 8 chars */
|
||
{
|
||
guint32 seedarray[4];
|
||
gchar *p, hexbuf[9] = { 0, };
|
||
memcpy (hexbuf, s + 0, 8);
|
||
seedarray[0] = g_ascii_strtoull (hexbuf, &p, 16);
|
||
seed_failed += p != NULL && *p != 0;
|
||
memcpy (hexbuf, s + 8, 8);
|
||
seedarray[1] = g_ascii_strtoull (hexbuf, &p, 16);
|
||
seed_failed += p != NULL && *p != 0;
|
||
memcpy (hexbuf, s + 16, 8);
|
||
seedarray[2] = g_ascii_strtoull (hexbuf, &p, 16);
|
||
seed_failed += p != NULL && *p != 0;
|
||
memcpy (hexbuf, s + 24, 8);
|
||
seedarray[3] = g_ascii_strtoull (hexbuf, &p, 16);
|
||
seed_failed += p != NULL && *p != 0;
|
||
if (!seed_failed)
|
||
{
|
||
test_run_rand = g_rand_new_with_seed_array (seedarray, 4);
|
||
return;
|
||
}
|
||
}
|
||
}
|
||
g_error ("Unknown or invalid random seed: %s", rseed);
|
||
}
|
||
|
||
/**
|
||
* g_test_rand_int:
|
||
*
|
||
* Get a reproducible random integer number.
|
||
*
|
||
* The random numbers generated by the g_test_rand_*() family of functions
|
||
* change with every new test program start, unless the --seed option is
|
||
* given when starting test programs.
|
||
*
|
||
* For individual test cases however, the random number generator is
|
||
* reseeded, to avoid dependencies between tests and to make --seed
|
||
* effective for all test cases.
|
||
*
|
||
* Returns: a random number from the seeded random number generator.
|
||
*
|
||
* Since: 2.16
|
||
*/
|
||
gint32
|
||
g_test_rand_int (void)
|
||
{
|
||
gint32 r;
|
||
|
||
G_LOCK (test_run_rand);
|
||
r = g_rand_int (test_run_rand);
|
||
G_UNLOCK (test_run_rand);
|
||
|
||
return r;
|
||
}
|
||
|
||
/**
|
||
* g_test_rand_int_range:
|
||
* @begin: the minimum value returned by this function
|
||
* @end: the smallest value not to be returned by this function
|
||
*
|
||
* Get a reproducible random integer number out of a specified range,
|
||
* see g_test_rand_int() for details on test case random numbers.
|
||
*
|
||
* Returns: a number with @begin <= number < @end.
|
||
*
|
||
* Since: 2.16
|
||
*/
|
||
gint32
|
||
g_test_rand_int_range (gint32 begin,
|
||
gint32 end)
|
||
{
|
||
gint32 r;
|
||
|
||
G_LOCK (test_run_rand);
|
||
r = g_rand_int_range (test_run_rand, begin, end);
|
||
G_UNLOCK (test_run_rand);
|
||
|
||
return r;
|
||
}
|
||
|
||
/**
|
||
* g_test_rand_double:
|
||
*
|
||
* Get a reproducible random floating point number,
|
||
* see g_test_rand_int() for details on test case random numbers.
|
||
*
|
||
* Returns: a random number from the seeded random number generator.
|
||
*
|
||
* Since: 2.16
|
||
*/
|
||
double
|
||
g_test_rand_double (void)
|
||
{
|
||
double r;
|
||
|
||
G_LOCK (test_run_rand);
|
||
r = g_rand_double (test_run_rand);
|
||
G_UNLOCK (test_run_rand);
|
||
|
||
return r;
|
||
}
|
||
|
||
/**
|
||
* g_test_rand_double_range:
|
||
* @range_start: the minimum value returned by this function
|
||
* @range_end: the minimum value not returned by this function
|
||
*
|
||
* Get a reproducible random floating pointer number out of a specified range,
|
||
* see g_test_rand_int() for details on test case random numbers.
|
||
*
|
||
* Returns: a number with @range_start <= number < @range_end.
|
||
*
|
||
* Since: 2.16
|
||
*/
|
||
double
|
||
g_test_rand_double_range (double range_start,
|
||
double range_end)
|
||
{
|
||
double r;
|
||
|
||
G_LOCK (test_run_rand);
|
||
r = g_rand_double_range (test_run_rand, range_start, range_end);
|
||
G_UNLOCK (test_run_rand);
|
||
|
||
return r;
|
||
}
|
||
|
||
/**
|
||
* g_test_timer_start:
|
||
*
|
||
* Start a timing test. Call g_test_timer_elapsed() when the task is supposed
|
||
* to be done. Call this function again to restart the timer.
|
||
*
|
||
* Since: 2.16
|
||
*/
|
||
void
|
||
g_test_timer_start (void)
|
||
{
|
||
if (!test_user_timer)
|
||
test_user_timer = g_timer_new();
|
||
test_user_stamp = 0;
|
||
g_timer_start (test_user_timer);
|
||
}
|
||
|
||
/**
|
||
* g_test_timer_elapsed:
|
||
*
|
||
* Get the time since the last start of the timer with g_test_timer_start().
|
||
*
|
||
* Returns: the time since the last start of the timer, as a double
|
||
*
|
||
* Since: 2.16
|
||
*/
|
||
double
|
||
g_test_timer_elapsed (void)
|
||
{
|
||
test_user_stamp = test_user_timer ? g_timer_elapsed (test_user_timer, NULL) : 0;
|
||
return test_user_stamp;
|
||
}
|
||
|
||
/**
|
||
* g_test_timer_last:
|
||
*
|
||
* Report the last result of g_test_timer_elapsed().
|
||
*
|
||
* Returns: the last result of g_test_timer_elapsed(), as a double
|
||
*
|
||
* Since: 2.16
|
||
*/
|
||
double
|
||
g_test_timer_last (void)
|
||
{
|
||
return test_user_stamp;
|
||
}
|
||
|
||
/**
|
||
* g_test_minimized_result:
|
||
* @minimized_quantity: the reported value
|
||
* @format: the format string of the report message
|
||
* @...: arguments to pass to the printf() function
|
||
*
|
||
* Report the result of a performance or measurement test.
|
||
* The test should generally strive to minimize the reported
|
||
* quantities (smaller values are better than larger ones),
|
||
* this and @minimized_quantity can determine sorting
|
||
* order for test result reports.
|
||
*
|
||
* Since: 2.16
|
||
*/
|
||
void
|
||
g_test_minimized_result (double minimized_quantity,
|
||
const char *format,
|
||
...)
|
||
{
|
||
long double largs = minimized_quantity;
|
||
gchar *buffer;
|
||
va_list args;
|
||
|
||
va_start (args, format);
|
||
buffer = g_strdup_vprintf (format, args);
|
||
va_end (args);
|
||
|
||
g_test_log (G_TEST_LOG_MIN_RESULT, buffer, NULL, 1, &largs);
|
||
g_free (buffer);
|
||
}
|
||
|
||
/**
|
||
* g_test_maximized_result:
|
||
* @maximized_quantity: the reported value
|
||
* @format: the format string of the report message
|
||
* @...: arguments to pass to the printf() function
|
||
*
|
||
* Report the result of a performance or measurement test.
|
||
* The test should generally strive to maximize the reported
|
||
* quantities (larger values are better than smaller ones),
|
||
* this and @maximized_quantity can determine sorting
|
||
* order for test result reports.
|
||
*
|
||
* Since: 2.16
|
||
*/
|
||
void
|
||
g_test_maximized_result (double maximized_quantity,
|
||
const char *format,
|
||
...)
|
||
{
|
||
long double largs = maximized_quantity;
|
||
gchar *buffer;
|
||
va_list args;
|
||
|
||
va_start (args, format);
|
||
buffer = g_strdup_vprintf (format, args);
|
||
va_end (args);
|
||
|
||
g_test_log (G_TEST_LOG_MAX_RESULT, buffer, NULL, 1, &largs);
|
||
g_free (buffer);
|
||
}
|
||
|
||
/**
|
||
* g_test_message:
|
||
* @format: the format string
|
||
* @...: printf-like arguments to @format
|
||
*
|
||
* Add a message to the test report.
|
||
*
|
||
* Since: 2.16
|
||
*/
|
||
void
|
||
g_test_message (const char *format,
|
||
...)
|
||
{
|
||
gchar *buffer;
|
||
va_list args;
|
||
|
||
va_start (args, format);
|
||
buffer = g_strdup_vprintf (format, args);
|
||
va_end (args);
|
||
|
||
g_test_log (G_TEST_LOG_MESSAGE, buffer, NULL, 0, NULL);
|
||
g_free (buffer);
|
||
}
|
||
|
||
/**
|
||
* g_test_bug_base:
|
||
* @uri_pattern: the base pattern for bug URIs
|
||
*
|
||
* Specify the base URI for bug reports.
|
||
*
|
||
* The base URI is used to construct bug report messages for
|
||
* g_test_message() when g_test_bug() is called.
|
||
* Calling this function outside of a test case sets the
|
||
* default base URI for all test cases. Calling it from within
|
||
* a test case changes the base URI for the scope of the test
|
||
* case only.
|
||
* Bug URIs are constructed by appending a bug specific URI
|
||
* portion to @uri_pattern, or by replacing the special string
|
||
* '\%s' within @uri_pattern if that is present.
|
||
*
|
||
* If g_test_bug_base() is not called, bug URIs are formed solely
|
||
* from the value provided by g_test_bug().
|
||
*
|
||
* Since: 2.16
|
||
*/
|
||
void
|
||
g_test_bug_base (const char *uri_pattern)
|
||
{
|
||
g_free (test_uri_base);
|
||
test_uri_base = g_strdup (uri_pattern);
|
||
}
|
||
|
||
/**
|
||
* g_test_bug:
|
||
* @bug_uri_snippet: Bug specific bug tracker URI portion.
|
||
*
|
||
* This function adds a message to test reports that
|
||
* associates a bug URI with a test case.
|
||
* Bug URIs are constructed from a base URI set with g_test_bug_base()
|
||
* and @bug_uri_snippet. If g_test_bug_base() has not been called, it is
|
||
* assumed to be the empty string, so a full URI can be provided to
|
||
* g_test_bug() instead.
|
||
*
|
||
* Since: 2.16
|
||
* See also: g_test_summary()
|
||
*/
|
||
void
|
||
g_test_bug (const char *bug_uri_snippet)
|
||
{
|
||
const char *c = NULL;
|
||
|
||
g_return_if_fail (bug_uri_snippet != NULL);
|
||
|
||
if (test_uri_base != NULL)
|
||
c = strstr (test_uri_base, "%s");
|
||
if (c)
|
||
{
|
||
char *b = g_strndup (test_uri_base, c - test_uri_base);
|
||
char *s = g_strconcat (b, bug_uri_snippet, c + 2, NULL);
|
||
g_free (b);
|
||
g_test_message ("Bug Reference: %s", s);
|
||
g_free (s);
|
||
}
|
||
else
|
||
g_test_message ("Bug Reference: %s%s",
|
||
test_uri_base ? test_uri_base : "", bug_uri_snippet);
|
||
}
|
||
|
||
/**
|
||
* g_test_summary:
|
||
* @summary: One or two sentences summarising what the test checks, and how it
|
||
* checks it.
|
||
*
|
||
* Set the summary for a test, which describes what the test checks, and how it
|
||
* goes about checking it. This may be included in test report output, and is
|
||
* useful documentation for anyone reading the source code or modifying a test
|
||
* in future. It must be a single line.
|
||
*
|
||
* This should be called at the top of a test function.
|
||
*
|
||
* For example:
|
||
* |[<!-- language="C" -->
|
||
* static void
|
||
* test_array_sort (void)
|
||
* {
|
||
* g_test_summary ("Test my_array_sort() sorts the array correctly and stably, "
|
||
* "including testing zero length and one-element arrays.");
|
||
*
|
||
* …
|
||
* }
|
||
* ]|
|
||
*
|
||
* Since: 2.62
|
||
* See also: g_test_bug()
|
||
*/
|
||
void
|
||
g_test_summary (const char *summary)
|
||
{
|
||
g_return_if_fail (summary != NULL);
|
||
g_return_if_fail (strchr (summary, '\n') == NULL);
|
||
g_return_if_fail (strchr (summary, '\r') == NULL);
|
||
|
||
g_test_message ("%s summary: %s", test_run_name, summary);
|
||
}
|
||
|
||
/**
|
||
* g_test_get_root:
|
||
*
|
||
* Get the toplevel test suite for the test path API.
|
||
*
|
||
* Returns: the toplevel #GTestSuite
|
||
*
|
||
* Since: 2.16
|
||
*/
|
||
GTestSuite*
|
||
g_test_get_root (void)
|
||
{
|
||
if (!test_suite_root)
|
||
{
|
||
test_suite_root = g_test_create_suite ("root");
|
||
g_free (test_suite_root->name);
|
||
test_suite_root->name = g_strdup ("");
|
||
}
|
||
|
||
return test_suite_root;
|
||
}
|
||
|
||
/**
|
||
* g_test_run:
|
||
*
|
||
* Runs all tests under the toplevel suite which can be retrieved
|
||
* with g_test_get_root(). Similar to g_test_run_suite(), the test
|
||
* cases to be run are filtered according to test path arguments
|
||
* (`-p testpath` and `-s testpath`) as parsed by g_test_init().
|
||
* g_test_run_suite() or g_test_run() may only be called once in a
|
||
* program.
|
||
*
|
||
* In general, the tests and sub-suites within each suite are run in
|
||
* the order in which they are defined. However, note that prior to
|
||
* GLib 2.36, there was a bug in the `g_test_add_*`
|
||
* functions which caused them to create multiple suites with the same
|
||
* name, meaning that if you created tests "/foo/simple",
|
||
* "/bar/simple", and "/foo/using-bar" in that order, they would get
|
||
* run in that order (since g_test_run() would run the first "/foo"
|
||
* suite, then the "/bar" suite, then the second "/foo" suite). As of
|
||
* 2.36, this bug is fixed, and adding the tests in that order would
|
||
* result in a running order of "/foo/simple", "/foo/using-bar",
|
||
* "/bar/simple". If this new ordering is sub-optimal (because it puts
|
||
* more-complicated tests before simpler ones, making it harder to
|
||
* figure out exactly what has failed), you can fix it by changing the
|
||
* test paths to group tests by suite in a way that will result in the
|
||
* desired running order. Eg, "/simple/foo", "/simple/bar",
|
||
* "/complex/foo-using-bar".
|
||
*
|
||
* However, you should never make the actual result of a test depend
|
||
* on the order that tests are run in. If you need to ensure that some
|
||
* particular code runs before or after a given test case, use
|
||
* g_test_add(), which lets you specify setup and teardown functions.
|
||
*
|
||
* If all tests are skipped or marked as incomplete (expected failures),
|
||
* this function will return 0 if producing TAP output, or 77 (treated
|
||
* as "skip test" by Automake) otherwise.
|
||
*
|
||
* Returns: 0 on success, 1 on failure (assuming it returns at all),
|
||
* 0 or 77 if all tests were skipped with g_test_skip() and/or
|
||
* g_test_incomplete()
|
||
*
|
||
* Since: 2.16
|
||
*/
|
||
int
|
||
g_test_run (void)
|
||
{
|
||
if (g_test_run_suite (g_test_get_root()) != 0)
|
||
return 1;
|
||
|
||
/* Clean up the temporary directory. */
|
||
if (test_isolate_dirs_tmpdir != NULL)
|
||
{
|
||
rm_rf (test_isolate_dirs_tmpdir);
|
||
g_free (test_isolate_dirs_tmpdir);
|
||
test_isolate_dirs_tmpdir = NULL;
|
||
}
|
||
|
||
/* 77 is special to Automake's default driver, but not Automake's TAP driver
|
||
* or Perl's prove(1) TAP driver. */
|
||
if (test_tap_log)
|
||
return 0;
|
||
|
||
if (test_run_count > 0 && test_run_count == test_skipped_count)
|
||
return 77;
|
||
else
|
||
return 0;
|
||
}
|
||
|
||
/**
|
||
* g_test_create_case:
|
||
* @test_name: the name for the test case
|
||
* @data_size: the size of the fixture data structure
|
||
* @test_data: test data argument for the test functions
|
||
* @data_setup: (scope async): the function to set up the fixture data
|
||
* @data_test: (scope async): the actual test function
|
||
* @data_teardown: (scope async): the function to teardown the fixture data
|
||
*
|
||
* Create a new #GTestCase, named @test_name.
|
||
*
|
||
* This API is fairly low level, and calling g_test_add() or g_test_add_func()
|
||
* is preferable.
|
||
*
|
||
* When this test is executed, a fixture structure of size @data_size
|
||
* will be automatically allocated and filled with zeros. Then @data_setup is
|
||
* called to initialize the fixture. After fixture setup, the actual test
|
||
* function @data_test is called. Once the test run completes, the
|
||
* fixture structure is torn down by calling @data_teardown and
|
||
* after that the memory is automatically released by the test framework.
|
||
*
|
||
* Splitting up a test run into fixture setup, test function and
|
||
* fixture teardown is most useful if the same fixture type is used for
|
||
* multiple tests. In this cases, g_test_create_case() will be
|
||
* called with the same type of fixture (the @data_size argument), but varying
|
||
* @test_name and @data_test arguments.
|
||
*
|
||
* Returns: a newly allocated #GTestCase.
|
||
*
|
||
* Since: 2.16
|
||
*/
|
||
GTestCase*
|
||
g_test_create_case (const char *test_name,
|
||
gsize data_size,
|
||
gconstpointer test_data,
|
||
GTestFixtureFunc data_setup,
|
||
GTestFixtureFunc data_test,
|
||
GTestFixtureFunc data_teardown)
|
||
{
|
||
GTestCase *tc;
|
||
|
||
g_return_val_if_fail (test_name != NULL, NULL);
|
||
g_return_val_if_fail (strchr (test_name, '/') == NULL, NULL);
|
||
g_return_val_if_fail (test_name[0] != 0, NULL);
|
||
g_return_val_if_fail (data_test != NULL, NULL);
|
||
|
||
tc = g_slice_new0 (GTestCase);
|
||
tc->name = g_strdup (test_name);
|
||
tc->test_data = (gpointer) test_data;
|
||
tc->fixture_size = data_size;
|
||
tc->fixture_setup = (void*) data_setup;
|
||
tc->fixture_test = (void*) data_test;
|
||
tc->fixture_teardown = (void*) data_teardown;
|
||
|
||
return tc;
|
||
}
|
||
|
||
static gint
|
||
find_suite (gconstpointer l, gconstpointer s)
|
||
{
|
||
const GTestSuite *suite = l;
|
||
const gchar *str = s;
|
||
|
||
return strcmp (suite->name, str);
|
||
}
|
||
|
||
static gint
|
||
find_case (gconstpointer l, gconstpointer s)
|
||
{
|
||
const GTestCase *tc = l;
|
||
const gchar *str = s;
|
||
|
||
return strcmp (tc->name, str);
|
||
}
|
||
|
||
/**
|
||
* GTestFixtureFunc:
|
||
* @fixture: (not nullable): the test fixture
|
||
* @user_data: the data provided when registering the test
|
||
*
|
||
* The type used for functions that operate on test fixtures. This is
|
||
* used for the fixture setup and teardown functions as well as for the
|
||
* testcases themselves.
|
||
*
|
||
* @user_data is a pointer to the data that was given when registering
|
||
* the test case.
|
||
*
|
||
* @fixture will be a pointer to the area of memory allocated by the
|
||
* test framework, of the size requested. If the requested size was
|
||
* zero then @fixture will be equal to @user_data.
|
||
*
|
||
* Since: 2.28
|
||
*/
|
||
void
|
||
g_test_add_vtable (const char *testpath,
|
||
gsize data_size,
|
||
gconstpointer test_data,
|
||
GTestFixtureFunc data_setup,
|
||
GTestFixtureFunc fixture_test_func,
|
||
GTestFixtureFunc data_teardown)
|
||
{
|
||
gchar **segments;
|
||
guint ui;
|
||
GTestSuite *suite;
|
||
|
||
g_return_if_fail (testpath != NULL);
|
||
g_return_if_fail (g_path_is_absolute (testpath));
|
||
g_return_if_fail (fixture_test_func != NULL);
|
||
g_return_if_fail (!test_isolate_dirs || strstr (testpath, "/.") == NULL);
|
||
|
||
suite = g_test_get_root();
|
||
segments = g_strsplit (testpath, "/", -1);
|
||
for (ui = 0; segments[ui] != NULL; ui++)
|
||
{
|
||
const char *seg = segments[ui];
|
||
gboolean islast = segments[ui + 1] == NULL;
|
||
if (islast && !seg[0])
|
||
g_error ("invalid test case path: %s", testpath);
|
||
else if (!seg[0])
|
||
continue; /* initial or duplicate slash */
|
||
else if (!islast)
|
||
{
|
||
GSList *l;
|
||
GTestSuite *csuite;
|
||
l = g_slist_find_custom (suite->suites, seg, find_suite);
|
||
if (l)
|
||
{
|
||
csuite = l->data;
|
||
}
|
||
else
|
||
{
|
||
csuite = g_test_create_suite (seg);
|
||
g_test_suite_add_suite (suite, csuite);
|
||
}
|
||
suite = csuite;
|
||
}
|
||
else /* islast */
|
||
{
|
||
GTestCase *tc;
|
||
|
||
if (g_slist_find_custom (suite->cases, seg, find_case))
|
||
g_error ("duplicate test case path: %s", testpath);
|
||
|
||
tc = g_test_create_case (seg, data_size, test_data, data_setup, fixture_test_func, data_teardown);
|
||
g_test_suite_add (suite, tc);
|
||
}
|
||
}
|
||
g_strfreev (segments);
|
||
}
|
||
|
||
/**
|
||
* g_test_fail:
|
||
*
|
||
* Indicates that a test failed. This function can be called
|
||
* multiple times from the same test. You can use this function
|
||
* if your test failed in a recoverable way.
|
||
*
|
||
* Do not use this function if the failure of a test could cause
|
||
* other tests to malfunction.
|
||
*
|
||
* Calling this function will not stop the test from running, you
|
||
* need to return from the test function yourself. So you can
|
||
* produce additional diagnostic messages or even continue running
|
||
* the test.
|
||
*
|
||
* If not called from inside a test, this function does nothing.
|
||
*
|
||
* Since: 2.30
|
||
**/
|
||
void
|
||
g_test_fail (void)
|
||
{
|
||
test_run_success = G_TEST_RUN_FAILURE;
|
||
}
|
||
|
||
/**
|
||
* g_test_incomplete:
|
||
* @msg: (nullable): explanation
|
||
*
|
||
* Indicates that a test failed because of some incomplete
|
||
* functionality. This function can be called multiple times
|
||
* from the same test.
|
||
*
|
||
* Calling this function will not stop the test from running, you
|
||
* need to return from the test function yourself. So you can
|
||
* produce additional diagnostic messages or even continue running
|
||
* the test.
|
||
*
|
||
* If not called from inside a test, this function does nothing.
|
||
*
|
||
* Since: 2.38
|
||
*/
|
||
void
|
||
g_test_incomplete (const gchar *msg)
|
||
{
|
||
test_run_success = G_TEST_RUN_INCOMPLETE;
|
||
g_free (test_run_msg);
|
||
test_run_msg = g_strdup (msg);
|
||
}
|
||
|
||
/**
|
||
* g_test_skip:
|
||
* @msg: (nullable): explanation
|
||
*
|
||
* Indicates that a test was skipped.
|
||
*
|
||
* Calling this function will not stop the test from running, you
|
||
* need to return from the test function yourself. So you can
|
||
* produce additional diagnostic messages or even continue running
|
||
* the test.
|
||
*
|
||
* If not called from inside a test, this function does nothing.
|
||
*
|
||
* Since: 2.38
|
||
*/
|
||
void
|
||
g_test_skip (const gchar *msg)
|
||
{
|
||
test_run_success = G_TEST_RUN_SKIPPED;
|
||
g_free (test_run_msg);
|
||
test_run_msg = g_strdup (msg);
|
||
}
|
||
|
||
/**
|
||
* g_test_failed:
|
||
*
|
||
* Returns whether a test has already failed. This will
|
||
* be the case when g_test_fail(), g_test_incomplete()
|
||
* or g_test_skip() have been called, but also if an
|
||
* assertion has failed.
|
||
*
|
||
* This can be useful to return early from a test if
|
||
* continuing after a failed assertion might be harmful.
|
||
*
|
||
* The return value of this function is only meaningful
|
||
* if it is called from inside a test function.
|
||
*
|
||
* Returns: %TRUE if the test has failed
|
||
*
|
||
* Since: 2.38
|
||
*/
|
||
gboolean
|
||
g_test_failed (void)
|
||
{
|
||
return test_run_success != G_TEST_RUN_SUCCESS;
|
||
}
|
||
|
||
/**
|
||
* g_test_set_nonfatal_assertions:
|
||
*
|
||
* Changes the behaviour of the various `g_assert_*()` macros,
|
||
* g_test_assert_expected_messages() and the various
|
||
* `g_test_trap_assert_*()` macros to not abort to program, but instead
|
||
* call g_test_fail() and continue. (This also changes the behavior of
|
||
* g_test_fail() so that it will not cause the test program to abort
|
||
* after completing the failed test.)
|
||
*
|
||
* Note that the g_assert_not_reached() and g_assert() macros are not
|
||
* affected by this.
|
||
*
|
||
* This function can only be called after g_test_init().
|
||
*
|
||
* Since: 2.38
|
||
*/
|
||
void
|
||
g_test_set_nonfatal_assertions (void)
|
||
{
|
||
if (!g_test_config_vars->test_initialized)
|
||
g_error ("g_test_set_nonfatal_assertions called without g_test_init");
|
||
test_nonfatal_assertions = TRUE;
|
||
test_mode_fatal = FALSE;
|
||
}
|
||
|
||
/**
|
||
* GTestFunc:
|
||
*
|
||
* The type used for test case functions.
|
||
*
|
||
* Since: 2.28
|
||
*/
|
||
|
||
/**
|
||
* g_test_add_func:
|
||
* @testpath: /-separated test case path name for the test.
|
||
* @test_func: (scope async): The test function to invoke for this test.
|
||
*
|
||
* Create a new test case, similar to g_test_create_case(). However
|
||
* the test is assumed to use no fixture, and test suites are automatically
|
||
* created on the fly and added to the root fixture, based on the
|
||
* slash-separated portions of @testpath.
|
||
*
|
||
* If @testpath includes the component "subprocess" anywhere in it,
|
||
* the test will be skipped by default, and only run if explicitly
|
||
* required via the `-p` command-line option or g_test_trap_subprocess().
|
||
*
|
||
* No component of @testpath may start with a dot (`.`) if the
|
||
* %G_TEST_OPTION_ISOLATE_DIRS option is being used; and it is recommended to
|
||
* do so even if it isn’t.
|
||
*
|
||
* Since: 2.16
|
||
*/
|
||
void
|
||
g_test_add_func (const char *testpath,
|
||
GTestFunc test_func)
|
||
{
|
||
g_return_if_fail (testpath != NULL);
|
||
g_return_if_fail (testpath[0] == '/');
|
||
g_return_if_fail (test_func != NULL);
|
||
g_test_add_vtable (testpath, 0, NULL, NULL, (GTestFixtureFunc) test_func, NULL);
|
||
}
|
||
|
||
/**
|
||
* GTestDataFunc:
|
||
* @user_data: the data provided when registering the test
|
||
*
|
||
* The type used for test case functions that take an extra pointer
|
||
* argument.
|
||
*
|
||
* Since: 2.28
|
||
*/
|
||
|
||
/**
|
||
* g_test_add_data_func:
|
||
* @testpath: /-separated test case path name for the test.
|
||
* @test_data: Test data argument for the test function.
|
||
* @test_func: (scope async): The test function to invoke for this test.
|
||
*
|
||
* Create a new test case, similar to g_test_create_case(). However
|
||
* the test is assumed to use no fixture, and test suites are automatically
|
||
* created on the fly and added to the root fixture, based on the
|
||
* slash-separated portions of @testpath. The @test_data argument
|
||
* will be passed as first argument to @test_func.
|
||
*
|
||
* If @testpath includes the component "subprocess" anywhere in it,
|
||
* the test will be skipped by default, and only run if explicitly
|
||
* required via the `-p` command-line option or g_test_trap_subprocess().
|
||
*
|
||
* No component of @testpath may start with a dot (`.`) if the
|
||
* %G_TEST_OPTION_ISOLATE_DIRS option is being used; and it is recommended to
|
||
* do so even if it isn’t.
|
||
*
|
||
* Since: 2.16
|
||
*/
|
||
void
|
||
g_test_add_data_func (const char *testpath,
|
||
gconstpointer test_data,
|
||
GTestDataFunc test_func)
|
||
{
|
||
g_return_if_fail (testpath != NULL);
|
||
g_return_if_fail (testpath[0] == '/');
|
||
g_return_if_fail (test_func != NULL);
|
||
|
||
g_test_add_vtable (testpath, 0, test_data, NULL, (GTestFixtureFunc) test_func, NULL);
|
||
}
|
||
|
||
/**
|
||
* g_test_add_data_func_full:
|
||
* @testpath: /-separated test case path name for the test.
|
||
* @test_data: Test data argument for the test function.
|
||
* @test_func: The test function to invoke for this test.
|
||
* @data_free_func: #GDestroyNotify for @test_data.
|
||
*
|
||
* Create a new test case, as with g_test_add_data_func(), but freeing
|
||
* @test_data after the test run is complete.
|
||
*
|
||
* Since: 2.34
|
||
*/
|
||
void
|
||
g_test_add_data_func_full (const char *testpath,
|
||
gpointer test_data,
|
||
GTestDataFunc test_func,
|
||
GDestroyNotify data_free_func)
|
||
{
|
||
g_return_if_fail (testpath != NULL);
|
||
g_return_if_fail (testpath[0] == '/');
|
||
g_return_if_fail (test_func != NULL);
|
||
|
||
g_test_add_vtable (testpath, 0, test_data, NULL,
|
||
(GTestFixtureFunc) test_func,
|
||
(GTestFixtureFunc) data_free_func);
|
||
}
|
||
|
||
static gboolean
|
||
g_test_suite_case_exists (GTestSuite *suite,
|
||
const char *test_path)
|
||
{
|
||
GSList *iter;
|
||
char *slash;
|
||
GTestCase *tc;
|
||
|
||
test_path++;
|
||
slash = strchr (test_path, '/');
|
||
|
||
if (slash)
|
||
{
|
||
for (iter = suite->suites; iter; iter = iter->next)
|
||
{
|
||
GTestSuite *child_suite = iter->data;
|
||
|
||
if (!strncmp (child_suite->name, test_path, slash - test_path))
|
||
if (g_test_suite_case_exists (child_suite, slash))
|
||
return TRUE;
|
||
}
|
||
}
|
||
else
|
||
{
|
||
for (iter = suite->cases; iter; iter = iter->next)
|
||
{
|
||
tc = iter->data;
|
||
if (!strcmp (tc->name, test_path))
|
||
return TRUE;
|
||
}
|
||
}
|
||
|
||
return FALSE;
|
||
}
|
||
|
||
/**
|
||
* g_test_create_suite:
|
||
* @suite_name: a name for the suite
|
||
*
|
||
* Create a new test suite with the name @suite_name.
|
||
*
|
||
* Returns: A newly allocated #GTestSuite instance.
|
||
*
|
||
* Since: 2.16
|
||
*/
|
||
GTestSuite*
|
||
g_test_create_suite (const char *suite_name)
|
||
{
|
||
GTestSuite *ts;
|
||
g_return_val_if_fail (suite_name != NULL, NULL);
|
||
g_return_val_if_fail (strchr (suite_name, '/') == NULL, NULL);
|
||
g_return_val_if_fail (suite_name[0] != 0, NULL);
|
||
ts = g_slice_new0 (GTestSuite);
|
||
ts->name = g_strdup (suite_name);
|
||
return ts;
|
||
}
|
||
|
||
/**
|
||
* g_test_suite_add:
|
||
* @suite: a #GTestSuite
|
||
* @test_case: a #GTestCase
|
||
*
|
||
* Adds @test_case to @suite.
|
||
*
|
||
* Since: 2.16
|
||
*/
|
||
void
|
||
g_test_suite_add (GTestSuite *suite,
|
||
GTestCase *test_case)
|
||
{
|
||
g_return_if_fail (suite != NULL);
|
||
g_return_if_fail (test_case != NULL);
|
||
|
||
suite->cases = g_slist_append (suite->cases, test_case);
|
||
}
|
||
|
||
/**
|
||
* g_test_suite_add_suite:
|
||
* @suite: a #GTestSuite
|
||
* @nestedsuite: another #GTestSuite
|
||
*
|
||
* Adds @nestedsuite to @suite.
|
||
*
|
||
* Since: 2.16
|
||
*/
|
||
void
|
||
g_test_suite_add_suite (GTestSuite *suite,
|
||
GTestSuite *nestedsuite)
|
||
{
|
||
g_return_if_fail (suite != NULL);
|
||
g_return_if_fail (nestedsuite != NULL);
|
||
|
||
suite->suites = g_slist_append (suite->suites, nestedsuite);
|
||
}
|
||
|
||
/**
|
||
* g_test_queue_free:
|
||
* @gfree_pointer: the pointer to be stored.
|
||
*
|
||
* Enqueue a pointer to be released with g_free() during the next
|
||
* teardown phase. This is equivalent to calling g_test_queue_destroy()
|
||
* with a destroy callback of g_free().
|
||
*
|
||
* Since: 2.16
|
||
*/
|
||
void
|
||
g_test_queue_free (gpointer gfree_pointer)
|
||
{
|
||
if (gfree_pointer)
|
||
g_test_queue_destroy (g_free, gfree_pointer);
|
||
}
|
||
|
||
/**
|
||
* g_test_queue_destroy:
|
||
* @destroy_func: Destroy callback for teardown phase.
|
||
* @destroy_data: Destroy callback data.
|
||
*
|
||
* This function enqueus a callback @destroy_func to be executed
|
||
* during the next test case teardown phase. This is most useful
|
||
* to auto destruct allocated test resources at the end of a test run.
|
||
* Resources are released in reverse queue order, that means enqueueing
|
||
* callback A before callback B will cause B() to be called before
|
||
* A() during teardown.
|
||
*
|
||
* Since: 2.16
|
||
*/
|
||
void
|
||
g_test_queue_destroy (GDestroyNotify destroy_func,
|
||
gpointer destroy_data)
|
||
{
|
||
DestroyEntry *dentry;
|
||
|
||
g_return_if_fail (destroy_func != NULL);
|
||
|
||
dentry = g_slice_new0 (DestroyEntry);
|
||
dentry->destroy_func = destroy_func;
|
||
dentry->destroy_data = destroy_data;
|
||
dentry->next = test_destroy_queue;
|
||
test_destroy_queue = dentry;
|
||
}
|
||
|
||
static gboolean
|
||
test_case_run (GTestCase *tc)
|
||
{
|
||
gchar *old_base = g_strdup (test_uri_base);
|
||
GSList **old_free_list, *filename_free_list = NULL;
|
||
gboolean success = G_TEST_RUN_SUCCESS;
|
||
|
||
old_free_list = test_filename_free_list;
|
||
test_filename_free_list = &filename_free_list;
|
||
|
||
if (++test_run_count <= test_startup_skip_count)
|
||
g_test_log (G_TEST_LOG_SKIP_CASE, test_run_name, NULL, 0, NULL);
|
||
else if (test_run_list)
|
||
{
|
||
g_print ("%s\n", test_run_name);
|
||
g_test_log (G_TEST_LOG_LIST_CASE, test_run_name, NULL, 0, NULL);
|
||
}
|
||
else
|
||
{
|
||
GTimer *test_run_timer = g_timer_new();
|
||
long double largs[3];
|
||
void *fixture;
|
||
g_test_log (G_TEST_LOG_START_CASE, test_run_name, NULL, 0, NULL);
|
||
test_run_forks = 0;
|
||
test_run_success = G_TEST_RUN_SUCCESS;
|
||
g_clear_pointer (&test_run_msg, g_free);
|
||
g_test_log_set_fatal_handler (NULL, NULL);
|
||
if (test_paths_skipped && g_slist_find_custom (test_paths_skipped, test_run_name, (GCompareFunc)g_strcmp0))
|
||
g_test_skip ("by request (-s option)");
|
||
else
|
||
{
|
||
GError *local_error = NULL;
|
||
|
||
if (!test_do_isolate_dirs (&local_error))
|
||
{
|
||
g_test_log (G_TEST_LOG_ERROR, local_error->message, NULL, 0, NULL);
|
||
g_test_fail ();
|
||
g_error_free (local_error);
|
||
}
|
||
else
|
||
{
|
||
g_timer_start (test_run_timer);
|
||
fixture = tc->fixture_size ? g_malloc0 (tc->fixture_size) : tc->test_data;
|
||
test_run_seed (test_run_seedstr);
|
||
if (tc->fixture_setup)
|
||
tc->fixture_setup (fixture, tc->test_data);
|
||
tc->fixture_test (fixture, tc->test_data);
|
||
test_trap_clear();
|
||
while (test_destroy_queue)
|
||
{
|
||
DestroyEntry *dentry = test_destroy_queue;
|
||
test_destroy_queue = dentry->next;
|
||
dentry->destroy_func (dentry->destroy_data);
|
||
g_slice_free (DestroyEntry, dentry);
|
||
}
|
||
if (tc->fixture_teardown)
|
||
tc->fixture_teardown (fixture, tc->test_data);
|
||
if (tc->fixture_size)
|
||
g_free (fixture);
|
||
g_timer_stop (test_run_timer);
|
||
}
|
||
|
||
test_rm_isolate_dirs ();
|
||
}
|
||
success = test_run_success;
|
||
test_run_success = G_TEST_RUN_FAILURE;
|
||
largs[0] = success; /* OK */
|
||
largs[1] = test_run_forks;
|
||
largs[2] = g_timer_elapsed (test_run_timer, NULL);
|
||
g_test_log (G_TEST_LOG_STOP_CASE, test_run_name, test_run_msg, G_N_ELEMENTS (largs), largs);
|
||
g_clear_pointer (&test_run_msg, g_free);
|
||
g_timer_destroy (test_run_timer);
|
||
}
|
||
|
||
g_slist_free_full (filename_free_list, g_free);
|
||
test_filename_free_list = old_free_list;
|
||
g_free (test_uri_base);
|
||
test_uri_base = old_base;
|
||
|
||
return (success == G_TEST_RUN_SUCCESS ||
|
||
success == G_TEST_RUN_SKIPPED ||
|
||
success == G_TEST_RUN_INCOMPLETE);
|
||
}
|
||
|
||
static gboolean
|
||
path_has_prefix (const char *path,
|
||
const char *prefix)
|
||
{
|
||
int prefix_len = strlen (prefix);
|
||
|
||
return (strncmp (path, prefix, prefix_len) == 0 &&
|
||
(path[prefix_len] == '\0' ||
|
||
path[prefix_len] == '/'));
|
||
}
|
||
|
||
static gboolean
|
||
test_should_run (const char *test_path,
|
||
const char *cmp_path)
|
||
{
|
||
if (strstr (test_run_name, "/subprocess"))
|
||
{
|
||
if (g_strcmp0 (test_path, cmp_path) == 0)
|
||
return TRUE;
|
||
|
||
if (g_test_verbose ())
|
||
g_print ("GTest: skipping: %s\n", test_run_name);
|
||
return FALSE;
|
||
}
|
||
|
||
return !cmp_path || path_has_prefix (test_path, cmp_path);
|
||
}
|
||
|
||
/* Recurse through @suite, running tests matching @path (or all tests
|
||
* if @path is %NULL).
|
||
*/
|
||
static int
|
||
g_test_run_suite_internal (GTestSuite *suite,
|
||
const char *path)
|
||
{
|
||
guint n_bad = 0;
|
||
gchar *old_name = test_run_name;
|
||
GSList *iter;
|
||
|
||
g_return_val_if_fail (suite != NULL, -1);
|
||
|
||
g_test_log (G_TEST_LOG_START_SUITE, suite->name, NULL, 0, NULL);
|
||
|
||
for (iter = suite->cases; iter; iter = iter->next)
|
||
{
|
||
GTestCase *tc = iter->data;
|
||
|
||
test_run_name = g_build_path ("/", old_name, tc->name, NULL);
|
||
if (test_should_run (test_run_name, path))
|
||
{
|
||
if (!test_case_run (tc))
|
||
n_bad++;
|
||
}
|
||
g_free (test_run_name);
|
||
}
|
||
|
||
for (iter = suite->suites; iter; iter = iter->next)
|
||
{
|
||
GTestSuite *ts = iter->data;
|
||
|
||
test_run_name = g_build_path ("/", old_name, ts->name, NULL);
|
||
if (!path || path_has_prefix (path, test_run_name))
|
||
n_bad += g_test_run_suite_internal (ts, path);
|
||
g_free (test_run_name);
|
||
}
|
||
|
||
test_run_name = old_name;
|
||
|
||
g_test_log (G_TEST_LOG_STOP_SUITE, suite->name, NULL, 0, NULL);
|
||
|
||
return n_bad;
|
||
}
|
||
|
||
static int
|
||
g_test_suite_count (GTestSuite *suite)
|
||
{
|
||
int n = 0;
|
||
GSList *iter;
|
||
|
||
g_return_val_if_fail (suite != NULL, -1);
|
||
|
||
for (iter = suite->cases; iter; iter = iter->next)
|
||
{
|
||
GTestCase *tc = iter->data;
|
||
|
||
if (strcmp (tc->name, "subprocess") != 0)
|
||
n++;
|
||
}
|
||
|
||
for (iter = suite->suites; iter; iter = iter->next)
|
||
{
|
||
GTestSuite *ts = iter->data;
|
||
|
||
if (strcmp (ts->name, "subprocess") != 0)
|
||
n += g_test_suite_count (ts);
|
||
}
|
||
|
||
return n;
|
||
}
|
||
|
||
/**
|
||
* g_test_run_suite:
|
||
* @suite: a #GTestSuite
|
||
*
|
||
* Execute the tests within @suite and all nested #GTestSuites.
|
||
* The test suites to be executed are filtered according to
|
||
* test path arguments (`-p testpath` and `-s testpath`) as parsed by
|
||
* g_test_init(). See the g_test_run() documentation for more
|
||
* information on the order that tests are run in.
|
||
*
|
||
* g_test_run_suite() or g_test_run() may only be called once
|
||
* in a program.
|
||
*
|
||
* Returns: 0 on success
|
||
*
|
||
* Since: 2.16
|
||
*/
|
||
int
|
||
g_test_run_suite (GTestSuite *suite)
|
||
{
|
||
int n_bad = 0;
|
||
|
||
g_return_val_if_fail (g_test_run_once == TRUE, -1);
|
||
|
||
g_test_run_once = FALSE;
|
||
test_count = g_test_suite_count (suite);
|
||
|
||
test_run_name = g_strdup_printf ("/%s", suite->name);
|
||
|
||
if (test_paths)
|
||
{
|
||
GSList *iter;
|
||
|
||
for (iter = test_paths; iter; iter = iter->next)
|
||
n_bad += g_test_run_suite_internal (suite, iter->data);
|
||
}
|
||
else
|
||
n_bad = g_test_run_suite_internal (suite, NULL);
|
||
|
||
g_free (test_run_name);
|
||
test_run_name = NULL;
|
||
|
||
return n_bad;
|
||
}
|
||
|
||
static void
|
||
gtest_default_log_handler (const gchar *log_domain,
|
||
GLogLevelFlags log_level,
|
||
const gchar *message,
|
||
gpointer unused_data)
|
||
{
|
||
const gchar *strv[16];
|
||
gboolean fatal = FALSE;
|
||
gchar *msg;
|
||
guint i = 0;
|
||
|
||
if (log_domain)
|
||
{
|
||
strv[i++] = log_domain;
|
||
strv[i++] = "-";
|
||
}
|
||
if (log_level & G_LOG_FLAG_FATAL)
|
||
{
|
||
strv[i++] = "FATAL-";
|
||
fatal = TRUE;
|
||
}
|
||
if (log_level & G_LOG_FLAG_RECURSION)
|
||
strv[i++] = "RECURSIVE-";
|
||
if (log_level & G_LOG_LEVEL_ERROR)
|
||
strv[i++] = "ERROR";
|
||
if (log_level & G_LOG_LEVEL_CRITICAL)
|
||
strv[i++] = "CRITICAL";
|
||
if (log_level & G_LOG_LEVEL_WARNING)
|
||
strv[i++] = "WARNING";
|
||
if (log_level & G_LOG_LEVEL_MESSAGE)
|
||
strv[i++] = "MESSAGE";
|
||
if (log_level & G_LOG_LEVEL_INFO)
|
||
strv[i++] = "INFO";
|
||
if (log_level & G_LOG_LEVEL_DEBUG)
|
||
strv[i++] = "DEBUG";
|
||
strv[i++] = ": ";
|
||
strv[i++] = message;
|
||
strv[i++] = NULL;
|
||
|
||
msg = g_strjoinv ("", (gchar**) strv);
|
||
g_test_log (fatal ? G_TEST_LOG_ERROR : G_TEST_LOG_MESSAGE, msg, NULL, 0, NULL);
|
||
g_log_default_handler (log_domain, log_level, message, unused_data);
|
||
|
||
g_free (msg);
|
||
}
|
||
|
||
void
|
||
g_assertion_message (const char *domain,
|
||
const char *file,
|
||
int line,
|
||
const char *func,
|
||
const char *message)
|
||
{
|
||
char lstr[32];
|
||
char *s;
|
||
|
||
if (!message)
|
||
message = "code should not be reached";
|
||
g_snprintf (lstr, 32, "%d", line);
|
||
s = g_strconcat (domain ? domain : "", domain && domain[0] ? ":" : "",
|
||
"ERROR:", file, ":", lstr, ":",
|
||
func, func[0] ? ":" : "",
|
||
" ", message, NULL);
|
||
g_printerr ("**\n%s\n", s);
|
||
|
||
/* Don't print a fatal error indication if assertions are non-fatal, or
|
||
* if we are a child process that might be sharing the parent's stdout. */
|
||
if (test_nonfatal_assertions || test_in_subprocess || test_in_forked_child)
|
||
g_test_log (G_TEST_LOG_MESSAGE, s, NULL, 0, NULL);
|
||
else
|
||
g_test_log (G_TEST_LOG_ERROR, s, NULL, 0, NULL);
|
||
|
||
if (test_nonfatal_assertions)
|
||
{
|
||
g_free (s);
|
||
g_test_fail ();
|
||
return;
|
||
}
|
||
|
||
/* store assertion message in global variable, so that it can be found in a
|
||
* core dump */
|
||
if (__glib_assert_msg != NULL)
|
||
/* free the old one */
|
||
free (__glib_assert_msg);
|
||
__glib_assert_msg = (char*) malloc (strlen (s) + 1);
|
||
strcpy (__glib_assert_msg, s);
|
||
|
||
g_free (s);
|
||
|
||
if (test_in_subprocess)
|
||
{
|
||
/* If this is a test case subprocess then it probably hit this
|
||
* assertion on purpose, so just exit() rather than abort()ing,
|
||
* to avoid triggering any system crash-reporting daemon.
|
||
*/
|
||
_exit (1);
|
||
}
|
||
else
|
||
g_abort ();
|
||
}
|
||
|
||
/**
|
||
* g_assertion_message_expr: (skip)
|
||
* @domain: (nullable): log domain
|
||
* @file: file containing the assertion
|
||
* @line: line number of the assertion
|
||
* @func: function containing the assertion
|
||
* @expr: (nullable): expression which failed
|
||
*
|
||
* Internal function used to print messages from the public g_assert() and
|
||
* g_assert_not_reached() macros.
|
||
*/
|
||
void
|
||
g_assertion_message_expr (const char *domain,
|
||
const char *file,
|
||
int line,
|
||
const char *func,
|
||
const char *expr)
|
||
{
|
||
char *s;
|
||
if (!expr)
|
||
s = g_strdup ("code should not be reached");
|
||
else
|
||
s = g_strconcat ("assertion failed: (", expr, ")", NULL);
|
||
g_assertion_message (domain, file, line, func, s);
|
||
g_free (s);
|
||
|
||
/* Normally g_assertion_message() won't return, but we need this for
|
||
* when test_nonfatal_assertions is set, since
|
||
* g_assertion_message_expr() is used for always-fatal assertions.
|
||
*/
|
||
if (test_in_subprocess)
|
||
_exit (1);
|
||
else
|
||
g_abort ();
|
||
}
|
||
|
||
void
|
||
g_assertion_message_cmpnum (const char *domain,
|
||
const char *file,
|
||
int line,
|
||
const char *func,
|
||
const char *expr,
|
||
long double arg1,
|
||
const char *cmp,
|
||
long double arg2,
|
||
char numtype)
|
||
{
|
||
char *s = NULL;
|
||
|
||
switch (numtype)
|
||
{
|
||
case 'i': s = g_strdup_printf ("assertion failed (%s): (%" G_GINT64_MODIFIER "i %s %" G_GINT64_MODIFIER "i)", expr, (gint64) arg1, cmp, (gint64) arg2); break;
|
||
case 'x': s = g_strdup_printf ("assertion failed (%s): (0x%08" G_GINT64_MODIFIER "x %s 0x%08" G_GINT64_MODIFIER "x)", expr, (guint64) arg1, cmp, (guint64) arg2); break;
|
||
case 'f': s = g_strdup_printf ("assertion failed (%s): (%.9g %s %.9g)", expr, (double) arg1, cmp, (double) arg2); break;
|
||
/* ideally use: floats=%.7g double=%.17g */
|
||
}
|
||
g_assertion_message (domain, file, line, func, s);
|
||
g_free (s);
|
||
}
|
||
|
||
void
|
||
g_assertion_message_cmpstr (const char *domain,
|
||
const char *file,
|
||
int line,
|
||
const char *func,
|
||
const char *expr,
|
||
const char *arg1,
|
||
const char *cmp,
|
||
const char *arg2)
|
||
{
|
||
char *a1, *a2, *s, *t1 = NULL, *t2 = NULL;
|
||
a1 = arg1 ? g_strconcat ("\"", t1 = g_strescape (arg1, NULL), "\"", NULL) : g_strdup ("NULL");
|
||
a2 = arg2 ? g_strconcat ("\"", t2 = g_strescape (arg2, NULL), "\"", NULL) : g_strdup ("NULL");
|
||
g_free (t1);
|
||
g_free (t2);
|
||
s = g_strdup_printf ("assertion failed (%s): (%s %s %s)", expr, a1, cmp, a2);
|
||
g_free (a1);
|
||
g_free (a2);
|
||
g_assertion_message (domain, file, line, func, s);
|
||
g_free (s);
|
||
}
|
||
|
||
void
|
||
g_assertion_message_error (const char *domain,
|
||
const char *file,
|
||
int line,
|
||
const char *func,
|
||
const char *expr,
|
||
const GError *error,
|
||
GQuark error_domain,
|
||
int error_code)
|
||
{
|
||
GString *gstring;
|
||
|
||
/* This is used by both g_assert_error() and g_assert_no_error(), so there
|
||
* are three cases: expected an error but got the wrong error, expected
|
||
* an error but got no error, and expected no error but got an error.
|
||
*/
|
||
|
||
gstring = g_string_new ("assertion failed ");
|
||
if (error_domain)
|
||
g_string_append_printf (gstring, "(%s == (%s, %d)): ", expr,
|
||
g_quark_to_string (error_domain), error_code);
|
||
else
|
||
g_string_append_printf (gstring, "(%s == NULL): ", expr);
|
||
|
||
if (error)
|
||
g_string_append_printf (gstring, "%s (%s, %d)", error->message,
|
||
g_quark_to_string (error->domain), error->code);
|
||
else
|
||
g_string_append_printf (gstring, "%s is NULL", expr);
|
||
|
||
g_assertion_message (domain, file, line, func, gstring->str);
|
||
g_string_free (gstring, TRUE);
|
||
}
|
||
|
||
/**
|
||
* g_strcmp0:
|
||
* @str1: (nullable): a C string or %NULL
|
||
* @str2: (nullable): another C string or %NULL
|
||
*
|
||
* Compares @str1 and @str2 like strcmp(). Handles %NULL
|
||
* gracefully by sorting it before non-%NULL strings.
|
||
* Comparing two %NULL pointers returns 0.
|
||
*
|
||
* Returns: an integer less than, equal to, or greater than zero, if @str1 is <, == or > than @str2.
|
||
*
|
||
* Since: 2.16
|
||
*/
|
||
int
|
||
g_strcmp0 (const char *str1,
|
||
const char *str2)
|
||
{
|
||
if (!str1)
|
||
return -(str1 != str2);
|
||
if (!str2)
|
||
return str1 != str2;
|
||
return strcmp (str1, str2);
|
||
}
|
||
|
||
static void
|
||
test_trap_clear (void)
|
||
{
|
||
test_trap_last_status = 0;
|
||
test_trap_last_pid = 0;
|
||
g_clear_pointer (&test_trap_last_subprocess, g_free);
|
||
g_clear_pointer (&test_trap_last_stdout, g_free);
|
||
g_clear_pointer (&test_trap_last_stderr, g_free);
|
||
}
|
||
|
||
#ifdef G_OS_UNIX
|
||
|
||
static int
|
||
safe_dup2 (int fd1,
|
||
int fd2)
|
||
{
|
||
int ret;
|
||
do
|
||
ret = dup2 (fd1, fd2);
|
||
while (ret < 0 && errno == EINTR);
|
||
return ret;
|
||
}
|
||
|
||
#endif
|
||
|
||
typedef struct {
|
||
GPid pid;
|
||
GMainLoop *loop;
|
||
int child_status; /* unmodified platform-specific status */
|
||
|
||
GIOChannel *stdout_io;
|
||
gboolean echo_stdout;
|
||
GString *stdout_str;
|
||
|
||
GIOChannel *stderr_io;
|
||
gboolean echo_stderr;
|
||
GString *stderr_str;
|
||
} WaitForChildData;
|
||
|
||
static void
|
||
check_complete (WaitForChildData *data)
|
||
{
|
||
if (data->child_status != -1 && data->stdout_io == NULL && data->stderr_io == NULL)
|
||
g_main_loop_quit (data->loop);
|
||
}
|
||
|
||
static void
|
||
child_exited (GPid pid,
|
||
gint status,
|
||
gpointer user_data)
|
||
{
|
||
WaitForChildData *data = user_data;
|
||
|
||
g_assert (status != -1);
|
||
data->child_status = status;
|
||
|
||
check_complete (data);
|
||
}
|
||
|
||
static gboolean
|
||
child_timeout (gpointer user_data)
|
||
{
|
||
WaitForChildData *data = user_data;
|
||
|
||
#ifdef G_OS_WIN32
|
||
TerminateProcess (data->pid, G_TEST_STATUS_TIMED_OUT);
|
||
#else
|
||
kill (data->pid, SIGALRM);
|
||
#endif
|
||
|
||
return FALSE;
|
||
}
|
||
|
||
static gboolean
|
||
child_read (GIOChannel *io, GIOCondition cond, gpointer user_data)
|
||
{
|
||
WaitForChildData *data = user_data;
|
||
GIOStatus status;
|
||
gsize nread, nwrote, total;
|
||
gchar buf[4096];
|
||
FILE *echo_file = NULL;
|
||
|
||
status = g_io_channel_read_chars (io, buf, sizeof (buf), &nread, NULL);
|
||
if (status == G_IO_STATUS_ERROR || status == G_IO_STATUS_EOF)
|
||
{
|
||
// FIXME data->error = (status == G_IO_STATUS_ERROR);
|
||
if (io == data->stdout_io)
|
||
g_clear_pointer (&data->stdout_io, g_io_channel_unref);
|
||
else
|
||
g_clear_pointer (&data->stderr_io, g_io_channel_unref);
|
||
|
||
check_complete (data);
|
||
return FALSE;
|
||
}
|
||
else if (status == G_IO_STATUS_AGAIN)
|
||
return TRUE;
|
||
|
||
if (io == data->stdout_io)
|
||
{
|
||
g_string_append_len (data->stdout_str, buf, nread);
|
||
if (data->echo_stdout)
|
||
echo_file = stdout;
|
||
}
|
||
else
|
||
{
|
||
g_string_append_len (data->stderr_str, buf, nread);
|
||
if (data->echo_stderr)
|
||
echo_file = stderr;
|
||
}
|
||
|
||
if (echo_file)
|
||
{
|
||
for (total = 0; total < nread; total += nwrote)
|
||
{
|
||
int errsv;
|
||
|
||
nwrote = fwrite (buf + total, 1, nread - total, echo_file);
|
||
errsv = errno;
|
||
if (nwrote == 0)
|
||
g_error ("write failed: %s", g_strerror (errsv));
|
||
}
|
||
}
|
||
|
||
return TRUE;
|
||
}
|
||
|
||
static void
|
||
wait_for_child (GPid pid,
|
||
int stdout_fd, gboolean echo_stdout,
|
||
int stderr_fd, gboolean echo_stderr,
|
||
guint64 timeout)
|
||
{
|
||
WaitForChildData data;
|
||
GMainContext *context;
|
||
GSource *source;
|
||
|
||
data.pid = pid;
|
||
data.child_status = -1;
|
||
|
||
context = g_main_context_new ();
|
||
data.loop = g_main_loop_new (context, FALSE);
|
||
|
||
source = g_child_watch_source_new (pid);
|
||
g_source_set_callback (source, (GSourceFunc) child_exited, &data, NULL);
|
||
g_source_attach (source, context);
|
||
g_source_unref (source);
|
||
|
||
data.echo_stdout = echo_stdout;
|
||
data.stdout_str = g_string_new (NULL);
|
||
data.stdout_io = g_io_channel_unix_new (stdout_fd);
|
||
g_io_channel_set_close_on_unref (data.stdout_io, TRUE);
|
||
g_io_channel_set_encoding (data.stdout_io, NULL, NULL);
|
||
g_io_channel_set_buffered (data.stdout_io, FALSE);
|
||
source = g_io_create_watch (data.stdout_io, G_IO_IN | G_IO_ERR | G_IO_HUP);
|
||
g_source_set_callback (source, (GSourceFunc) child_read, &data, NULL);
|
||
g_source_attach (source, context);
|
||
g_source_unref (source);
|
||
|
||
data.echo_stderr = echo_stderr;
|
||
data.stderr_str = g_string_new (NULL);
|
||
data.stderr_io = g_io_channel_unix_new (stderr_fd);
|
||
g_io_channel_set_close_on_unref (data.stderr_io, TRUE);
|
||
g_io_channel_set_encoding (data.stderr_io, NULL, NULL);
|
||
g_io_channel_set_buffered (data.stderr_io, FALSE);
|
||
source = g_io_create_watch (data.stderr_io, G_IO_IN | G_IO_ERR | G_IO_HUP);
|
||
g_source_set_callback (source, (GSourceFunc) child_read, &data, NULL);
|
||
g_source_attach (source, context);
|
||
g_source_unref (source);
|
||
|
||
if (timeout)
|
||
{
|
||
source = g_timeout_source_new (0);
|
||
g_source_set_ready_time (source, g_get_monotonic_time () + timeout);
|
||
g_source_set_callback (source, (GSourceFunc) child_timeout, &data, NULL);
|
||
g_source_attach (source, context);
|
||
g_source_unref (source);
|
||
}
|
||
|
||
g_main_loop_run (data.loop);
|
||
g_main_loop_unref (data.loop);
|
||
g_main_context_unref (context);
|
||
|
||
test_trap_last_pid = pid;
|
||
test_trap_last_status = data.child_status;
|
||
test_trap_last_stdout = g_string_free (data.stdout_str, FALSE);
|
||
test_trap_last_stderr = g_string_free (data.stderr_str, FALSE);
|
||
|
||
g_clear_pointer (&data.stdout_io, g_io_channel_unref);
|
||
g_clear_pointer (&data.stderr_io, g_io_channel_unref);
|
||
}
|
||
|
||
/**
|
||
* g_test_trap_fork:
|
||
* @usec_timeout: Timeout for the forked test in micro seconds.
|
||
* @test_trap_flags: Flags to modify forking behaviour.
|
||
*
|
||
* Fork the current test program to execute a test case that might
|
||
* not return or that might abort.
|
||
*
|
||
* If @usec_timeout is non-0, the forked test case is aborted and
|
||
* considered failing if its run time exceeds it.
|
||
*
|
||
* The forking behavior can be configured with the #GTestTrapFlags flags.
|
||
*
|
||
* In the following example, the test code forks, the forked child
|
||
* process produces some sample output and exits successfully.
|
||
* The forking parent process then asserts successful child program
|
||
* termination and validates child program outputs.
|
||
*
|
||
* |[<!-- language="C" -->
|
||
* static void
|
||
* test_fork_patterns (void)
|
||
* {
|
||
* if (g_test_trap_fork (0, G_TEST_TRAP_SILENCE_STDOUT | G_TEST_TRAP_SILENCE_STDERR))
|
||
* {
|
||
* g_print ("some stdout text: somagic17\n");
|
||
* g_printerr ("some stderr text: semagic43\n");
|
||
* exit (0); // successful test run
|
||
* }
|
||
* g_test_trap_assert_passed ();
|
||
* g_test_trap_assert_stdout ("*somagic17*");
|
||
* g_test_trap_assert_stderr ("*semagic43*");
|
||
* }
|
||
* ]|
|
||
*
|
||
* Returns: %TRUE for the forked child and %FALSE for the executing parent process.
|
||
*
|
||
* Since: 2.16
|
||
*
|
||
* Deprecated: This function is implemented only on Unix platforms,
|
||
* and is not always reliable due to problems inherent in
|
||
* fork-without-exec. Use g_test_trap_subprocess() instead.
|
||
*/
|
||
G_GNUC_BEGIN_IGNORE_DEPRECATIONS
|
||
gboolean
|
||
g_test_trap_fork (guint64 usec_timeout,
|
||
GTestTrapFlags test_trap_flags)
|
||
{
|
||
#ifdef G_OS_UNIX
|
||
int stdout_pipe[2] = { -1, -1 };
|
||
int stderr_pipe[2] = { -1, -1 };
|
||
int errsv;
|
||
|
||
test_trap_clear();
|
||
if (pipe (stdout_pipe) < 0 || pipe (stderr_pipe) < 0)
|
||
{
|
||
errsv = errno;
|
||
g_error ("failed to create pipes to fork test program: %s", g_strerror (errsv));
|
||
}
|
||
test_trap_last_pid = fork ();
|
||
errsv = errno;
|
||
if (test_trap_last_pid < 0)
|
||
g_error ("failed to fork test program: %s", g_strerror (errsv));
|
||
if (test_trap_last_pid == 0) /* child */
|
||
{
|
||
int fd0 = -1;
|
||
test_in_forked_child = TRUE;
|
||
close (stdout_pipe[0]);
|
||
close (stderr_pipe[0]);
|
||
if (!(test_trap_flags & G_TEST_TRAP_INHERIT_STDIN))
|
||
{
|
||
fd0 = g_open ("/dev/null", O_RDONLY, 0);
|
||
if (fd0 < 0)
|
||
g_error ("failed to open /dev/null for stdin redirection");
|
||
}
|
||
if (safe_dup2 (stdout_pipe[1], 1) < 0 || safe_dup2 (stderr_pipe[1], 2) < 0 || (fd0 >= 0 && safe_dup2 (fd0, 0) < 0))
|
||
{
|
||
errsv = errno;
|
||
g_error ("failed to dup2() in forked test program: %s", g_strerror (errsv));
|
||
}
|
||
if (fd0 >= 3)
|
||
close (fd0);
|
||
if (stdout_pipe[1] >= 3)
|
||
close (stdout_pipe[1]);
|
||
if (stderr_pipe[1] >= 3)
|
||
close (stderr_pipe[1]);
|
||
|
||
/* We typically expect these child processes to crash, and some
|
||
* tests spawn a *lot* of them. Avoid spamming system crash
|
||
* collection programs such as systemd-coredump and abrt.
|
||
*/
|
||
#ifdef HAVE_SYS_RESOURCE_H
|
||
{
|
||
struct rlimit limit = { 0, 0 };
|
||
(void) setrlimit (RLIMIT_CORE, &limit);
|
||
}
|
||
#endif
|
||
|
||
return TRUE;
|
||
}
|
||
else /* parent */
|
||
{
|
||
test_run_forks++;
|
||
close (stdout_pipe[1]);
|
||
close (stderr_pipe[1]);
|
||
|
||
wait_for_child (test_trap_last_pid,
|
||
stdout_pipe[0], !(test_trap_flags & G_TEST_TRAP_SILENCE_STDOUT),
|
||
stderr_pipe[0], !(test_trap_flags & G_TEST_TRAP_SILENCE_STDERR),
|
||
usec_timeout);
|
||
return FALSE;
|
||
}
|
||
#else
|
||
g_message ("Not implemented: g_test_trap_fork");
|
||
|
||
return FALSE;
|
||
#endif
|
||
}
|
||
G_GNUC_END_IGNORE_DEPRECATIONS
|
||
|
||
/**
|
||
* g_test_trap_subprocess:
|
||
* @test_path: (nullable): Test to run in a subprocess
|
||
* @usec_timeout: Timeout for the subprocess test in micro seconds.
|
||
* @test_flags: Flags to modify subprocess behaviour.
|
||
*
|
||
* Respawns the test program to run only @test_path in a subprocess.
|
||
* This can be used for a test case that might not return, or that
|
||
* might abort.
|
||
*
|
||
* If @test_path is %NULL then the same test is re-run in a subprocess.
|
||
* You can use g_test_subprocess() to determine whether the test is in
|
||
* a subprocess or not.
|
||
*
|
||
* @test_path can also be the name of the parent test, followed by
|
||
* "`/subprocess/`" and then a name for the specific subtest (or just
|
||
* ending with "`/subprocess`" if the test only has one child test);
|
||
* tests with names of this form will automatically be skipped in the
|
||
* parent process.
|
||
*
|
||
* If @usec_timeout is non-0, the test subprocess is aborted and
|
||
* considered failing if its run time exceeds it.
|
||
*
|
||
* The subprocess behavior can be configured with the
|
||
* #GTestSubprocessFlags flags.
|
||
*
|
||
* You can use methods such as g_test_trap_assert_passed(),
|
||
* g_test_trap_assert_failed(), and g_test_trap_assert_stderr() to
|
||
* check the results of the subprocess. (But note that
|
||
* g_test_trap_assert_stdout() and g_test_trap_assert_stderr()
|
||
* cannot be used if @test_flags specifies that the child should
|
||
* inherit the parent stdout/stderr.)
|
||
*
|
||
* If your `main ()` needs to behave differently in
|
||
* the subprocess, you can call g_test_subprocess() (after calling
|
||
* g_test_init()) to see whether you are in a subprocess.
|
||
*
|
||
* The following example tests that calling
|
||
* `my_object_new(1000000)` will abort with an error
|
||
* message.
|
||
*
|
||
* |[<!-- language="C" -->
|
||
* static void
|
||
* test_create_large_object (void)
|
||
* {
|
||
* if (g_test_subprocess ())
|
||
* {
|
||
* my_object_new (1000000);
|
||
* return;
|
||
* }
|
||
*
|
||
* // Reruns this same test in a subprocess
|
||
* g_test_trap_subprocess (NULL, 0, 0);
|
||
* g_test_trap_assert_failed ();
|
||
* g_test_trap_assert_stderr ("*ERROR*too large*");
|
||
* }
|
||
*
|
||
* int
|
||
* main (int argc, char **argv)
|
||
* {
|
||
* g_test_init (&argc, &argv, NULL);
|
||
*
|
||
* g_test_add_func ("/myobject/create_large_object",
|
||
* test_create_large_object);
|
||
* return g_test_run ();
|
||
* }
|
||
* ]|
|
||
*
|
||
* Since: 2.38
|
||
*/
|
||
void
|
||
g_test_trap_subprocess (const char *test_path,
|
||
guint64 usec_timeout,
|
||
GTestSubprocessFlags test_flags)
|
||
{
|
||
GError *error = NULL;
|
||
GPtrArray *argv;
|
||
GSpawnFlags flags;
|
||
int stdout_fd, stderr_fd;
|
||
GPid pid;
|
||
|
||
/* Sanity check that they used GTestSubprocessFlags, not GTestTrapFlags */
|
||
g_assert ((test_flags & (G_TEST_TRAP_INHERIT_STDIN | G_TEST_TRAP_SILENCE_STDOUT | G_TEST_TRAP_SILENCE_STDERR)) == 0);
|
||
|
||
if (test_path)
|
||
{
|
||
if (!g_test_suite_case_exists (g_test_get_root (), test_path))
|
||
g_error ("g_test_trap_subprocess: test does not exist: %s", test_path);
|
||
}
|
||
else
|
||
{
|
||
test_path = test_run_name;
|
||
}
|
||
|
||
if (g_test_verbose ())
|
||
g_print ("GTest: subprocess: %s\n", test_path);
|
||
|
||
test_trap_clear ();
|
||
test_trap_last_subprocess = g_strdup (test_path);
|
||
|
||
argv = g_ptr_array_new ();
|
||
g_ptr_array_add (argv, test_argv0);
|
||
g_ptr_array_add (argv, "-q");
|
||
g_ptr_array_add (argv, "-p");
|
||
g_ptr_array_add (argv, (char *)test_path);
|
||
g_ptr_array_add (argv, "--GTestSubprocess");
|
||
if (test_log_fd != -1)
|
||
{
|
||
char log_fd_buf[128];
|
||
|
||
g_ptr_array_add (argv, "--GTestLogFD");
|
||
g_snprintf (log_fd_buf, sizeof (log_fd_buf), "%d", test_log_fd);
|
||
g_ptr_array_add (argv, log_fd_buf);
|
||
}
|
||
g_ptr_array_add (argv, NULL);
|
||
|
||
flags = G_SPAWN_DO_NOT_REAP_CHILD;
|
||
if (test_log_fd != -1)
|
||
flags |= G_SPAWN_LEAVE_DESCRIPTORS_OPEN;
|
||
if (test_flags & G_TEST_TRAP_INHERIT_STDIN)
|
||
flags |= G_SPAWN_CHILD_INHERITS_STDIN;
|
||
|
||
if (!g_spawn_async_with_pipes (test_initial_cwd,
|
||
(char **)argv->pdata,
|
||
NULL, flags,
|
||
NULL, NULL,
|
||
&pid, NULL, &stdout_fd, &stderr_fd,
|
||
&error))
|
||
{
|
||
g_error ("g_test_trap_subprocess() failed: %s",
|
||
error->message);
|
||
}
|
||
g_ptr_array_free (argv, TRUE);
|
||
|
||
wait_for_child (pid,
|
||
stdout_fd, !!(test_flags & G_TEST_SUBPROCESS_INHERIT_STDOUT),
|
||
stderr_fd, !!(test_flags & G_TEST_SUBPROCESS_INHERIT_STDERR),
|
||
usec_timeout);
|
||
}
|
||
|
||
/**
|
||
* g_test_subprocess:
|
||
*
|
||
* Returns %TRUE (after g_test_init() has been called) if the test
|
||
* program is running under g_test_trap_subprocess().
|
||
*
|
||
* Returns: %TRUE if the test program is running under
|
||
* g_test_trap_subprocess().
|
||
*
|
||
* Since: 2.38
|
||
*/
|
||
gboolean
|
||
g_test_subprocess (void)
|
||
{
|
||
return test_in_subprocess;
|
||
}
|
||
|
||
/**
|
||
* g_test_trap_has_passed:
|
||
*
|
||
* Check the result of the last g_test_trap_subprocess() call.
|
||
*
|
||
* Returns: %TRUE if the last test subprocess terminated successfully.
|
||
*
|
||
* Since: 2.16
|
||
*/
|
||
gboolean
|
||
g_test_trap_has_passed (void)
|
||
{
|
||
#ifdef G_OS_UNIX
|
||
return (WIFEXITED (test_trap_last_status) &&
|
||
WEXITSTATUS (test_trap_last_status) == 0);
|
||
#else
|
||
return test_trap_last_status == 0;
|
||
#endif
|
||
}
|
||
|
||
/**
|
||
* g_test_trap_reached_timeout:
|
||
*
|
||
* Check the result of the last g_test_trap_subprocess() call.
|
||
*
|
||
* Returns: %TRUE if the last test subprocess got killed due to a timeout.
|
||
*
|
||
* Since: 2.16
|
||
*/
|
||
gboolean
|
||
g_test_trap_reached_timeout (void)
|
||
{
|
||
#ifdef G_OS_UNIX
|
||
return (WIFSIGNALED (test_trap_last_status) &&
|
||
WTERMSIG (test_trap_last_status) == SIGALRM);
|
||
#else
|
||
return test_trap_last_status == G_TEST_STATUS_TIMED_OUT;
|
||
#endif
|
||
}
|
||
|
||
static gboolean
|
||
log_child_output (const gchar *process_id)
|
||
{
|
||
gchar *escaped;
|
||
|
||
#ifdef G_OS_UNIX
|
||
if (WIFEXITED (test_trap_last_status)) /* normal exit */
|
||
{
|
||
if (WEXITSTATUS (test_trap_last_status) == 0)
|
||
g_test_message ("child process (%s) exit status: 0 (success)",
|
||
process_id);
|
||
else
|
||
g_test_message ("child process (%s) exit status: %d (error)",
|
||
process_id, WEXITSTATUS (test_trap_last_status));
|
||
}
|
||
else if (WIFSIGNALED (test_trap_last_status) &&
|
||
WTERMSIG (test_trap_last_status) == SIGALRM)
|
||
{
|
||
g_test_message ("child process (%s) timed out", process_id);
|
||
}
|
||
else if (WIFSIGNALED (test_trap_last_status))
|
||
{
|
||
const gchar *maybe_dumped_core = "";
|
||
|
||
#ifdef WCOREDUMP
|
||
if (WCOREDUMP (test_trap_last_status))
|
||
maybe_dumped_core = ", core dumped";
|
||
#endif
|
||
|
||
g_test_message ("child process (%s) killed by signal %d (%s)%s",
|
||
process_id, WTERMSIG (test_trap_last_status),
|
||
g_strsignal (WTERMSIG (test_trap_last_status)),
|
||
maybe_dumped_core);
|
||
}
|
||
else
|
||
{
|
||
g_test_message ("child process (%s) unknown wait status %d",
|
||
process_id, test_trap_last_status);
|
||
}
|
||
#else
|
||
if (test_trap_last_status == 0)
|
||
g_test_message ("child process (%s) exit status: 0 (success)",
|
||
process_id);
|
||
else
|
||
g_test_message ("child process (%s) exit status: %d (error)",
|
||
process_id, test_trap_last_status);
|
||
#endif
|
||
|
||
escaped = g_strescape (test_trap_last_stdout, NULL);
|
||
g_test_message ("child process (%s) stdout: \"%s\"", process_id, escaped);
|
||
g_free (escaped);
|
||
|
||
escaped = g_strescape (test_trap_last_stderr, NULL);
|
||
g_test_message ("child process (%s) stderr: \"%s\"", process_id, escaped);
|
||
g_free (escaped);
|
||
|
||
/* so we can use short-circuiting:
|
||
* logged_child_output = logged_child_output || log_child_output (...) */
|
||
return TRUE;
|
||
}
|
||
|
||
void
|
||
g_test_trap_assertions (const char *domain,
|
||
const char *file,
|
||
int line,
|
||
const char *func,
|
||
guint64 assertion_flags, /* 0-pass, 1-fail, 2-outpattern, 4-errpattern */
|
||
const char *pattern)
|
||
{
|
||
gboolean must_pass = assertion_flags == 0;
|
||
gboolean must_fail = assertion_flags == 1;
|
||
gboolean match_result = 0 == (assertion_flags & 1);
|
||
gboolean logged_child_output = FALSE;
|
||
const char *stdout_pattern = (assertion_flags & 2) ? pattern : NULL;
|
||
const char *stderr_pattern = (assertion_flags & 4) ? pattern : NULL;
|
||
const char *match_error = match_result ? "failed to match" : "contains invalid match";
|
||
char *process_id;
|
||
|
||
#ifdef G_OS_UNIX
|
||
if (test_trap_last_subprocess != NULL)
|
||
{
|
||
process_id = g_strdup_printf ("%s [%d]", test_trap_last_subprocess,
|
||
test_trap_last_pid);
|
||
}
|
||
else if (test_trap_last_pid != 0)
|
||
process_id = g_strdup_printf ("%d", test_trap_last_pid);
|
||
#else
|
||
if (test_trap_last_subprocess != NULL)
|
||
process_id = g_strdup (test_trap_last_subprocess);
|
||
#endif
|
||
else
|
||
g_error ("g_test_trap_ assertion with no trapped test");
|
||
|
||
if (must_pass && !g_test_trap_has_passed())
|
||
{
|
||
char *msg;
|
||
|
||
logged_child_output = logged_child_output || log_child_output (process_id);
|
||
|
||
msg = g_strdup_printf ("child process (%s) failed unexpectedly", process_id);
|
||
g_assertion_message (domain, file, line, func, msg);
|
||
g_free (msg);
|
||
}
|
||
if (must_fail && g_test_trap_has_passed())
|
||
{
|
||
char *msg;
|
||
|
||
logged_child_output = logged_child_output || log_child_output (process_id);
|
||
|
||
msg = g_strdup_printf ("child process (%s) did not fail as expected", process_id);
|
||
g_assertion_message (domain, file, line, func, msg);
|
||
g_free (msg);
|
||
}
|
||
if (stdout_pattern && match_result == !g_pattern_match_simple (stdout_pattern, test_trap_last_stdout))
|
||
{
|
||
char *msg;
|
||
|
||
logged_child_output = logged_child_output || log_child_output (process_id);
|
||
|
||
msg = g_strdup_printf ("stdout of child process (%s) %s: %s\nstdout was:\n%s",
|
||
process_id, match_error, stdout_pattern, test_trap_last_stdout);
|
||
g_assertion_message (domain, file, line, func, msg);
|
||
g_free (msg);
|
||
}
|
||
if (stderr_pattern && match_result == !g_pattern_match_simple (stderr_pattern, test_trap_last_stderr))
|
||
{
|
||
char *msg;
|
||
|
||
logged_child_output = logged_child_output || log_child_output (process_id);
|
||
|
||
msg = g_strdup_printf ("stderr of child process (%s) %s: %s\nstderr was:\n%s",
|
||
process_id, match_error, stderr_pattern, test_trap_last_stderr);
|
||
g_assertion_message (domain, file, line, func, msg);
|
||
g_free (msg);
|
||
}
|
||
|
||
(void) logged_child_output; /* shut up scan-build about the final unread assignment */
|
||
|
||
g_free (process_id);
|
||
}
|
||
|
||
static void
|
||
gstring_overwrite_int (GString *gstring,
|
||
guint pos,
|
||
guint32 vuint)
|
||
{
|
||
vuint = g_htonl (vuint);
|
||
g_string_overwrite_len (gstring, pos, (const gchar*) &vuint, 4);
|
||
}
|
||
|
||
static void
|
||
gstring_append_int (GString *gstring,
|
||
guint32 vuint)
|
||
{
|
||
vuint = g_htonl (vuint);
|
||
g_string_append_len (gstring, (const gchar*) &vuint, 4);
|
||
}
|
||
|
||
static void
|
||
gstring_append_double (GString *gstring,
|
||
double vdouble)
|
||
{
|
||
union { double vdouble; guint64 vuint64; } u;
|
||
u.vdouble = vdouble;
|
||
u.vuint64 = GUINT64_TO_BE (u.vuint64);
|
||
g_string_append_len (gstring, (const gchar*) &u.vuint64, 8);
|
||
}
|
||
|
||
static guint8*
|
||
g_test_log_dump (GTestLogMsg *msg,
|
||
guint *len)
|
||
{
|
||
GString *gstring = g_string_sized_new (1024);
|
||
guint ui;
|
||
gstring_append_int (gstring, 0); /* message length */
|
||
gstring_append_int (gstring, msg->log_type);
|
||
gstring_append_int (gstring, msg->n_strings);
|
||
gstring_append_int (gstring, msg->n_nums);
|
||
gstring_append_int (gstring, 0); /* reserved */
|
||
for (ui = 0; ui < msg->n_strings; ui++)
|
||
{
|
||
guint l = strlen (msg->strings[ui]);
|
||
gstring_append_int (gstring, l);
|
||
g_string_append_len (gstring, msg->strings[ui], l);
|
||
}
|
||
for (ui = 0; ui < msg->n_nums; ui++)
|
||
gstring_append_double (gstring, msg->nums[ui]);
|
||
*len = gstring->len;
|
||
gstring_overwrite_int (gstring, 0, *len); /* message length */
|
||
return (guint8*) g_string_free (gstring, FALSE);
|
||
}
|
||
|
||
static inline long double
|
||
net_double (const gchar **ipointer)
|
||
{
|
||
union { guint64 vuint64; double vdouble; } u;
|
||
guint64 aligned_int64;
|
||
memcpy (&aligned_int64, *ipointer, 8);
|
||
*ipointer += 8;
|
||
u.vuint64 = GUINT64_FROM_BE (aligned_int64);
|
||
return u.vdouble;
|
||
}
|
||
|
||
static inline guint32
|
||
net_int (const gchar **ipointer)
|
||
{
|
||
guint32 aligned_int;
|
||
memcpy (&aligned_int, *ipointer, 4);
|
||
*ipointer += 4;
|
||
return g_ntohl (aligned_int);
|
||
}
|
||
|
||
static gboolean
|
||
g_test_log_extract (GTestLogBuffer *tbuffer)
|
||
{
|
||
const gchar *p = tbuffer->data->str;
|
||
GTestLogMsg msg;
|
||
guint mlength;
|
||
if (tbuffer->data->len < 4 * 5)
|
||
return FALSE;
|
||
mlength = net_int (&p);
|
||
if (tbuffer->data->len < mlength)
|
||
return FALSE;
|
||
msg.log_type = net_int (&p);
|
||
msg.n_strings = net_int (&p);
|
||
msg.n_nums = net_int (&p);
|
||
if (net_int (&p) == 0)
|
||
{
|
||
guint ui;
|
||
msg.strings = g_new0 (gchar*, msg.n_strings + 1);
|
||
msg.nums = g_new0 (long double, msg.n_nums);
|
||
for (ui = 0; ui < msg.n_strings; ui++)
|
||
{
|
||
guint sl = net_int (&p);
|
||
msg.strings[ui] = g_strndup (p, sl);
|
||
p += sl;
|
||
}
|
||
for (ui = 0; ui < msg.n_nums; ui++)
|
||
msg.nums[ui] = net_double (&p);
|
||
if (p <= tbuffer->data->str + mlength)
|
||
{
|
||
g_string_erase (tbuffer->data, 0, mlength);
|
||
tbuffer->msgs = g_slist_prepend (tbuffer->msgs, g_memdup (&msg, sizeof (msg)));
|
||
return TRUE;
|
||
}
|
||
|
||
g_free (msg.nums);
|
||
g_strfreev (msg.strings);
|
||
}
|
||
|
||
g_error ("corrupt log stream from test program");
|
||
return FALSE;
|
||
}
|
||
|
||
/**
|
||
* g_test_log_buffer_new:
|
||
*
|
||
* Internal function for gtester to decode test log messages, no ABI guarantees provided.
|
||
*/
|
||
GTestLogBuffer*
|
||
g_test_log_buffer_new (void)
|
||
{
|
||
GTestLogBuffer *tb = g_new0 (GTestLogBuffer, 1);
|
||
tb->data = g_string_sized_new (1024);
|
||
return tb;
|
||
}
|
||
|
||
/**
|
||
* g_test_log_buffer_free:
|
||
*
|
||
* Internal function for gtester to free test log messages, no ABI guarantees provided.
|
||
*/
|
||
void
|
||
g_test_log_buffer_free (GTestLogBuffer *tbuffer)
|
||
{
|
||
g_return_if_fail (tbuffer != NULL);
|
||
while (tbuffer->msgs)
|
||
g_test_log_msg_free (g_test_log_buffer_pop (tbuffer));
|
||
g_string_free (tbuffer->data, TRUE);
|
||
g_free (tbuffer);
|
||
}
|
||
|
||
/**
|
||
* g_test_log_buffer_push:
|
||
*
|
||
* Internal function for gtester to decode test log messages, no ABI guarantees provided.
|
||
*/
|
||
void
|
||
g_test_log_buffer_push (GTestLogBuffer *tbuffer,
|
||
guint n_bytes,
|
||
const guint8 *bytes)
|
||
{
|
||
g_return_if_fail (tbuffer != NULL);
|
||
if (n_bytes)
|
||
{
|
||
gboolean more_messages;
|
||
g_return_if_fail (bytes != NULL);
|
||
g_string_append_len (tbuffer->data, (const gchar*) bytes, n_bytes);
|
||
do
|
||
more_messages = g_test_log_extract (tbuffer);
|
||
while (more_messages);
|
||
}
|
||
}
|
||
|
||
/**
|
||
* g_test_log_buffer_pop:
|
||
*
|
||
* Internal function for gtester to retrieve test log messages, no ABI guarantees provided.
|
||
*/
|
||
GTestLogMsg*
|
||
g_test_log_buffer_pop (GTestLogBuffer *tbuffer)
|
||
{
|
||
GTestLogMsg *msg = NULL;
|
||
g_return_val_if_fail (tbuffer != NULL, NULL);
|
||
if (tbuffer->msgs)
|
||
{
|
||
GSList *slist = g_slist_last (tbuffer->msgs);
|
||
msg = slist->data;
|
||
tbuffer->msgs = g_slist_delete_link (tbuffer->msgs, slist);
|
||
}
|
||
return msg;
|
||
}
|
||
|
||
/**
|
||
* g_test_log_msg_free:
|
||
*
|
||
* Internal function for gtester to free test log messages, no ABI guarantees provided.
|
||
*/
|
||
void
|
||
g_test_log_msg_free (GTestLogMsg *tmsg)
|
||
{
|
||
g_return_if_fail (tmsg != NULL);
|
||
g_strfreev (tmsg->strings);
|
||
g_free (tmsg->nums);
|
||
g_free (tmsg);
|
||
}
|
||
|
||
static gchar *
|
||
g_test_build_filename_va (GTestFileType file_type,
|
||
const gchar *first_path,
|
||
va_list ap)
|
||
{
|
||
const gchar *pathv[16];
|
||
gsize num_path_segments;
|
||
|
||
if (file_type == G_TEST_DIST)
|
||
pathv[0] = test_disted_files_dir;
|
||
else if (file_type == G_TEST_BUILT)
|
||
pathv[0] = test_built_files_dir;
|
||
else
|
||
g_assert_not_reached ();
|
||
|
||
pathv[1] = first_path;
|
||
|
||
for (num_path_segments = 2; num_path_segments < G_N_ELEMENTS (pathv); num_path_segments++)
|
||
{
|
||
pathv[num_path_segments] = va_arg (ap, const char *);
|
||
if (pathv[num_path_segments] == NULL)
|
||
break;
|
||
}
|
||
|
||
g_assert_cmpint (num_path_segments, <, G_N_ELEMENTS (pathv));
|
||
|
||
return g_build_filenamev ((gchar **) pathv);
|
||
}
|
||
|
||
/**
|
||
* g_test_build_filename:
|
||
* @file_type: the type of file (built vs. distributed)
|
||
* @first_path: the first segment of the pathname
|
||
* @...: %NULL-terminated additional path segments
|
||
*
|
||
* Creates the pathname to a data file that is required for a test.
|
||
*
|
||
* This function is conceptually similar to g_build_filename() except
|
||
* that the first argument has been replaced with a #GTestFileType
|
||
* argument.
|
||
*
|
||
* The data file should either have been distributed with the module
|
||
* containing the test (%G_TEST_DIST) or built as part of the build
|
||
* system of that module (%G_TEST_BUILT).
|
||
*
|
||
* In order for this function to work in srcdir != builddir situations,
|
||
* the G_TEST_SRCDIR and G_TEST_BUILDDIR environment variables need to
|
||
* have been defined. As of 2.38, this is done by the glib.mk
|
||
* included in GLib. Please ensure that your copy is up to date before
|
||
* using this function.
|
||
*
|
||
* In case neither variable is set, this function will fall back to
|
||
* using the dirname portion of argv[0], possibly removing ".libs".
|
||
* This allows for casual running of tests directly from the commandline
|
||
* in the srcdir == builddir case and should also support running of
|
||
* installed tests, assuming the data files have been installed in the
|
||
* same relative path as the test binary.
|
||
*
|
||
* Returns: the path of the file, to be freed using g_free()
|
||
*
|
||
* Since: 2.38
|
||
**/
|
||
/**
|
||
* GTestFileType:
|
||
* @G_TEST_DIST: a file that was included in the distribution tarball
|
||
* @G_TEST_BUILT: a file that was built on the compiling machine
|
||
*
|
||
* The type of file to return the filename for, when used with
|
||
* g_test_build_filename().
|
||
*
|
||
* These two options correspond rather directly to the 'dist' and
|
||
* 'built' terminology that automake uses and are explicitly used to
|
||
* distinguish between the 'srcdir' and 'builddir' being separate. All
|
||
* files in your project should either be dist (in the
|
||
* `EXTRA_DIST` or `dist_schema_DATA`
|
||
* sense, in which case they will always be in the srcdir) or built (in
|
||
* the `BUILT_SOURCES` sense, in which case they will
|
||
* always be in the builddir).
|
||
*
|
||
* Note: as a general rule of automake, files that are generated only as
|
||
* part of the build-from-git process (but then are distributed with the
|
||
* tarball) always go in srcdir (even if doing a srcdir != builddir
|
||
* build from git) and are considered as distributed files.
|
||
*
|
||
* Since: 2.38
|
||
**/
|
||
gchar *
|
||
g_test_build_filename (GTestFileType file_type,
|
||
const gchar *first_path,
|
||
...)
|
||
{
|
||
gchar *result;
|
||
va_list ap;
|
||
|
||
g_assert (g_test_initialized ());
|
||
|
||
va_start (ap, first_path);
|
||
result = g_test_build_filename_va (file_type, first_path, ap);
|
||
va_end (ap);
|
||
|
||
return result;
|
||
}
|
||
|
||
/**
|
||
* g_test_get_dir:
|
||
* @file_type: the type of file (built vs. distributed)
|
||
*
|
||
* Gets the pathname of the directory containing test files of the type
|
||
* specified by @file_type.
|
||
*
|
||
* This is approximately the same as calling g_test_build_filename("."),
|
||
* but you don't need to free the return value.
|
||
*
|
||
* Returns: (type filename): the path of the directory, owned by GLib
|
||
*
|
||
* Since: 2.38
|
||
**/
|
||
const gchar *
|
||
g_test_get_dir (GTestFileType file_type)
|
||
{
|
||
g_assert (g_test_initialized ());
|
||
|
||
if (file_type == G_TEST_DIST)
|
||
return test_disted_files_dir;
|
||
else if (file_type == G_TEST_BUILT)
|
||
return test_built_files_dir;
|
||
|
||
g_assert_not_reached ();
|
||
}
|
||
|
||
/**
|
||
* g_test_get_filename:
|
||
* @file_type: the type of file (built vs. distributed)
|
||
* @first_path: the first segment of the pathname
|
||
* @...: %NULL-terminated additional path segments
|
||
*
|
||
* Gets the pathname to a data file that is required for a test.
|
||
*
|
||
* This is the same as g_test_build_filename() with two differences.
|
||
* The first difference is that must only use this function from within
|
||
* a testcase function. The second difference is that you need not free
|
||
* the return value -- it will be automatically freed when the testcase
|
||
* finishes running.
|
||
*
|
||
* It is safe to use this function from a thread inside of a testcase
|
||
* but you must ensure that all such uses occur before the main testcase
|
||
* function returns (ie: it is best to ensure that all threads have been
|
||
* joined).
|
||
*
|
||
* Returns: the path, automatically freed at the end of the testcase
|
||
*
|
||
* Since: 2.38
|
||
**/
|
||
const gchar *
|
||
g_test_get_filename (GTestFileType file_type,
|
||
const gchar *first_path,
|
||
...)
|
||
{
|
||
gchar *result;
|
||
GSList *node;
|
||
va_list ap;
|
||
|
||
g_assert (g_test_initialized ());
|
||
if (test_filename_free_list == NULL)
|
||
g_error ("g_test_get_filename() can only be used within testcase functions");
|
||
|
||
va_start (ap, first_path);
|
||
result = g_test_build_filename_va (file_type, first_path, ap);
|
||
va_end (ap);
|
||
|
||
node = g_slist_prepend (NULL, result);
|
||
do
|
||
node->next = *test_filename_free_list;
|
||
while (!g_atomic_pointer_compare_and_exchange (test_filename_free_list, node->next, node));
|
||
|
||
return result;
|
||
}
|