mirror of
https://gitlab.gnome.org/GNOME/glib.git
synced 2025-02-15 14:57:15 +01:00
It’s fairly consistently timing out on macOS. Looking at the verbose test output, it’s still making progress right up until when it times out (i.e. it hasn’t hit a `GRecMutex` bug and hasn’t deadlocked), so it seems that the test runner is just hopelessly overloaded/underpowered for the number of threads and iterations we’re asking it to test. Tone those numbers down for CI test runs then. Signed-off-by: Philip Withnall <pwithnall@endlessos.org>
287 lines
6.8 KiB
C
287 lines
6.8 KiB
C
/* Unit tests for GRecMutex
|
|
* Copyright (C) 2011 Red Hat, Inc
|
|
* Author: Matthias Clasen
|
|
*
|
|
* SPDX-License-Identifier: LicenseRef-old-glib-tests
|
|
*
|
|
* This work is provided "as is"; redistribution and modification
|
|
* in whole or in part, in any medium, physical or electronic is
|
|
* permitted without restriction.
|
|
*
|
|
* This work is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.
|
|
*
|
|
* In no event shall the authors or contributors be liable for any
|
|
* direct, indirect, incidental, special, exemplary, or consequential
|
|
* damages (including, but not limited to, procurement of substitute
|
|
* goods or services; loss of use, data, or profits; or business
|
|
* interruption) however caused and on any theory of liability, whether
|
|
* in contract, strict liability, or tort (including negligence or
|
|
* otherwise) arising in any way out of the use of this software, even
|
|
* if advised of the possibility of such damage.
|
|
*/
|
|
|
|
/* We are testing some deprecated APIs here */
|
|
#ifndef GLIB_DISABLE_DEPRECATION_WARNINGS
|
|
#define GLIB_DISABLE_DEPRECATION_WARNINGS
|
|
#endif
|
|
|
|
#include <glib.h>
|
|
|
|
#include <stdio.h>
|
|
|
|
static void
|
|
test_rec_mutex1 (void)
|
|
{
|
|
GRecMutex mutex;
|
|
|
|
g_rec_mutex_init (&mutex);
|
|
g_rec_mutex_lock (&mutex);
|
|
g_rec_mutex_unlock (&mutex);
|
|
g_rec_mutex_lock (&mutex);
|
|
g_rec_mutex_unlock (&mutex);
|
|
g_rec_mutex_clear (&mutex);
|
|
}
|
|
|
|
static void
|
|
test_rec_mutex2 (void)
|
|
{
|
|
static GRecMutex mutex;
|
|
|
|
g_rec_mutex_lock (&mutex);
|
|
g_rec_mutex_unlock (&mutex);
|
|
g_rec_mutex_lock (&mutex);
|
|
g_rec_mutex_unlock (&mutex);
|
|
}
|
|
|
|
static void
|
|
test_rec_mutex3 (void)
|
|
{
|
|
static GRecMutex mutex;
|
|
gboolean ret;
|
|
|
|
ret = g_rec_mutex_trylock (&mutex);
|
|
g_assert_true (ret);
|
|
|
|
ret = g_rec_mutex_trylock (&mutex);
|
|
g_assert_true (ret);
|
|
|
|
g_rec_mutex_unlock (&mutex);
|
|
g_rec_mutex_unlock (&mutex);
|
|
}
|
|
|
|
typedef struct {
|
|
size_t n_locks;
|
|
unsigned int n_iterations;
|
|
size_t n_threads;
|
|
GThread **threads; /* (array length=n_threads) */
|
|
GThread **owners; /* (array length=n_locks), each element is locked by the corresponding mutex in @locks */
|
|
GRecMutex *locks; /* (array length=n_locks) */
|
|
} ThreadTestData;
|
|
|
|
static void
|
|
thread_test_data_clear (ThreadTestData *data)
|
|
{
|
|
g_free (data->locks);
|
|
g_free (data->owners);
|
|
g_free (data->threads);
|
|
}
|
|
|
|
static void
|
|
acquire (ThreadTestData *data,
|
|
unsigned int nr)
|
|
{
|
|
GThread *self;
|
|
|
|
self = g_thread_self ();
|
|
|
|
if (!g_rec_mutex_trylock (&data->locks[nr]))
|
|
{
|
|
if (g_test_verbose ())
|
|
g_printerr ("thread %p going to block on lock %d\n", self, nr);
|
|
|
|
g_rec_mutex_lock (&data->locks[nr]);
|
|
}
|
|
|
|
g_assert_null (data->owners[nr]); /* hopefully nobody else is here */
|
|
data->owners[nr] = self;
|
|
|
|
/* let some other threads try to ruin our day */
|
|
g_thread_yield ();
|
|
g_thread_yield ();
|
|
|
|
g_assert_true (data->owners[nr] == self); /* hopefully this is still us... */
|
|
|
|
if (g_test_verbose ())
|
|
g_printerr ("thread %p recursively taking lock %d\n", self, nr);
|
|
|
|
g_rec_mutex_lock (&data->locks[nr]); /* we're recursive, after all */
|
|
|
|
g_assert_true (data->owners[nr] == self); /* hopefully this is still us... */
|
|
|
|
g_rec_mutex_unlock (&data->locks[nr]);
|
|
|
|
g_thread_yield ();
|
|
g_thread_yield ();
|
|
|
|
g_assert_true (data->owners[nr] == self); /* hopefully this is still us... */
|
|
data->owners[nr] = NULL; /* make way for the next guy */
|
|
|
|
g_rec_mutex_unlock (&data->locks[nr]);
|
|
}
|
|
|
|
static gpointer
|
|
thread_func (gpointer user_data)
|
|
{
|
|
ThreadTestData *data = user_data;
|
|
GRand *rand;
|
|
|
|
rand = g_rand_new ();
|
|
|
|
for (unsigned int i = 0; i < data->n_iterations; i++)
|
|
acquire (data, g_rand_int_range (rand, 0, data->n_locks));
|
|
|
|
g_rand_free (rand);
|
|
|
|
return NULL;
|
|
}
|
|
|
|
static void
|
|
test_rec_mutex4 (void)
|
|
{
|
|
ThreadTestData data;
|
|
|
|
data.n_locks = 48;
|
|
data.n_iterations = 10000;
|
|
data.n_threads = 100;
|
|
|
|
/* Some CI runners have a hard time with this much contention, so tone it down
|
|
* a bit for CI. */
|
|
if (!g_test_perf ())
|
|
{
|
|
data.n_locks /= 10;
|
|
data.n_iterations /= 10;
|
|
data.n_threads /= 10;
|
|
}
|
|
|
|
data.threads = g_new0 (GThread*, data.n_threads);
|
|
data.owners = g_new0 (GThread*, data.n_locks);
|
|
data.locks = g_new0 (GRecMutex, data.n_locks);
|
|
|
|
for (size_t i = 0; i < data.n_locks; i++)
|
|
g_rec_mutex_init (&data.locks[i]);
|
|
|
|
for (size_t i = 0; i < data.n_threads; i++)
|
|
data.threads[i] = g_thread_new ("test", thread_func, &data);
|
|
|
|
for (size_t i = 0; i < data.n_threads; i++)
|
|
g_thread_join (data.threads[i]);
|
|
|
|
for (size_t i = 0; i < data.n_locks; i++)
|
|
g_rec_mutex_clear (&data.locks[i]);
|
|
|
|
for (size_t i = 0; i < data.n_locks; i++)
|
|
g_assert_null (data.owners[i]);
|
|
|
|
thread_test_data_clear (&data);
|
|
}
|
|
|
|
static gint count_to = 0;
|
|
static gint depth;
|
|
|
|
static gboolean
|
|
do_addition (gint *value)
|
|
{
|
|
static GRecMutex lock;
|
|
gboolean more;
|
|
gint i;
|
|
|
|
/* test performance of "good" cases (ie: short critical sections) */
|
|
for (i = 0; i < depth; i++)
|
|
g_rec_mutex_lock (&lock);
|
|
|
|
if ((more = *value != count_to))
|
|
if (*value != -1)
|
|
(*value)++;
|
|
|
|
for (i = 0; i < depth; i++)
|
|
g_rec_mutex_unlock (&lock);
|
|
|
|
return more;
|
|
}
|
|
|
|
static gpointer
|
|
addition_thread (gpointer value)
|
|
{
|
|
while (do_addition (value));
|
|
|
|
return NULL;
|
|
}
|
|
|
|
static void
|
|
test_mutex_perf (gconstpointer data)
|
|
{
|
|
gint c = GPOINTER_TO_INT (data);
|
|
GThread *threads[100];
|
|
gint64 start_time;
|
|
gint n_threads;
|
|
gdouble rate;
|
|
gint x = -1;
|
|
gint i;
|
|
|
|
n_threads = c / 256;
|
|
depth = c % 256;
|
|
count_to = g_test_perf () ? 100000000 : n_threads + 1;
|
|
|
|
for (i = 0; i < n_threads - 1; i++)
|
|
threads[i] = g_thread_new ("test", addition_thread, &x);
|
|
|
|
/* avoid measuring thread setup/teardown time */
|
|
start_time = g_get_monotonic_time ();
|
|
g_atomic_int_set (&x, 0);
|
|
addition_thread (&x);
|
|
g_assert_cmpint (g_atomic_int_get (&x), ==, count_to);
|
|
rate = g_get_monotonic_time () - start_time;
|
|
rate = x / rate;
|
|
|
|
for (i = 0; i < n_threads - 1; i++)
|
|
g_thread_join (threads[i]);
|
|
|
|
g_test_maximized_result (rate, "%f mips", rate);
|
|
}
|
|
|
|
|
|
int
|
|
main (int argc, char *argv[])
|
|
{
|
|
g_test_init (&argc, &argv, NULL);
|
|
|
|
g_test_add_func ("/thread/rec-mutex1", test_rec_mutex1);
|
|
g_test_add_func ("/thread/rec-mutex2", test_rec_mutex2);
|
|
g_test_add_func ("/thread/rec-mutex3", test_rec_mutex3);
|
|
g_test_add_func ("/thread/rec-mutex4", test_rec_mutex4);
|
|
|
|
{
|
|
gint i, j;
|
|
|
|
for (i = 0; i < 5; i++)
|
|
for (j = 1; j <= 5; j++)
|
|
{
|
|
gchar name[80];
|
|
guint c;
|
|
|
|
c = i * 256 + j;
|
|
|
|
if (i)
|
|
sprintf (name, "/thread/rec-mutex/perf/contended%d/depth%d", i, j);
|
|
else
|
|
sprintf (name, "/thread/rec-mutex/perf/uncontended/depth%d", j);
|
|
|
|
g_test_add_data_func (name, GINT_TO_POINTER (c), test_mutex_perf);
|
|
}
|
|
}
|
|
|
|
return g_test_run ();
|
|
}
|