aboutsummaryrefslogtreecommitdiff
path: root/tests
diff options
context:
space:
mode:
authorPeter Maydell <peter.maydell@linaro.org>2021-03-31 16:38:49 +0100
committerPeter Maydell <peter.maydell@linaro.org>2021-03-31 16:38:49 +0100
commit1bd16067b652cce41a9214d0c62c73d5b45ab4b1 (patch)
tree594b7ebcd1e1e348f5a25f7670af6e7d8d8072ac /tests
parent6ee55e1d10c25c2f6bf5ce2084ad2327e17affa5 (diff)
parentb6489ac06695e257ea0a9841364577e247fdee30 (diff)
downloadqemu-1bd16067b652cce41a9214d0c62c73d5b45ab4b1.zip
qemu-1bd16067b652cce41a9214d0c62c73d5b45ab4b1.tar.gz
qemu-1bd16067b652cce41a9214d0c62c73d5b45ab4b1.tar.bz2
Merge remote-tracking branch 'remotes/stefanha-gitlab/tags/block-pull-request' into staging
Pull request A fix for VDI image files and more generally for CoRwlock. # gpg: Signature made Wed 31 Mar 2021 10:50:39 BST # gpg: using RSA key 8695A8BFD3F97CDAAC35775A9CA4ABB381AB73C8 # gpg: Good signature from "Stefan Hajnoczi <stefanha@redhat.com>" [full] # gpg: aka "Stefan Hajnoczi <stefanha@gmail.com>" [full] # Primary key fingerprint: 8695 A8BF D3F9 7CDA AC35 775A 9CA4 ABB3 81AB 73C8 * remotes/stefanha-gitlab/tags/block-pull-request: test-coroutine: Add rwlock downgrade test test-coroutine: Add rwlock upgrade test coroutine-lock: Reimplement CoRwlock to fix downgrade bug coroutine-lock: Store the coroutine in the CoWaitRecord only once block/vdi: Don't assume that blocks are larger than VdiHeader block/vdi: When writing new bmap entry fails, don't leak the buffer Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
Diffstat (limited to 'tests')
-rw-r--r--tests/unit/test-coroutine.c161
1 files changed, 161 insertions, 0 deletions
diff --git a/tests/unit/test-coroutine.c b/tests/unit/test-coroutine.c
index e946d93..aa77a3b 100644
--- a/tests/unit/test-coroutine.c
+++ b/tests/unit/test-coroutine.c
@@ -264,6 +264,165 @@ static void test_co_mutex_lockable(void)
g_assert(QEMU_MAKE_LOCKABLE(null_pointer) == NULL);
}
+static CoRwlock rwlock;
+
+/* Test that readers are properly sent back to the queue when upgrading,
+ * even if they are the sole readers. The test scenario is as follows:
+ *
+ *
+ * | c1 | c2 |
+ * |--------------+------------+
+ * | rdlock | |
+ * | yield | |
+ * | | wrlock |
+ * | | <queued> |
+ * | upgrade | |
+ * | <queued> | <dequeued> |
+ * | | unlock |
+ * | <dequeued> | |
+ * | unlock | |
+ */
+
+static void coroutine_fn rwlock_yield_upgrade(void *opaque)
+{
+ qemu_co_rwlock_rdlock(&rwlock);
+ qemu_coroutine_yield();
+
+ qemu_co_rwlock_upgrade(&rwlock);
+ qemu_co_rwlock_unlock(&rwlock);
+
+ *(bool *)opaque = true;
+}
+
+static void coroutine_fn rwlock_wrlock_yield(void *opaque)
+{
+ qemu_co_rwlock_wrlock(&rwlock);
+ qemu_coroutine_yield();
+
+ qemu_co_rwlock_unlock(&rwlock);
+ *(bool *)opaque = true;
+}
+
+static void test_co_rwlock_upgrade(void)
+{
+ bool c1_done = false;
+ bool c2_done = false;
+ Coroutine *c1, *c2;
+
+ qemu_co_rwlock_init(&rwlock);
+ c1 = qemu_coroutine_create(rwlock_yield_upgrade, &c1_done);
+ c2 = qemu_coroutine_create(rwlock_wrlock_yield, &c2_done);
+
+ qemu_coroutine_enter(c1);
+ qemu_coroutine_enter(c2);
+
+ /* c1 now should go to sleep. */
+ qemu_coroutine_enter(c1);
+ g_assert(!c1_done);
+
+ qemu_coroutine_enter(c2);
+ g_assert(c1_done);
+ g_assert(c2_done);
+}
+
+static void coroutine_fn rwlock_rdlock_yield(void *opaque)
+{
+ qemu_co_rwlock_rdlock(&rwlock);
+ qemu_coroutine_yield();
+
+ qemu_co_rwlock_unlock(&rwlock);
+ qemu_coroutine_yield();
+
+ *(bool *)opaque = true;
+}
+
+static void coroutine_fn rwlock_wrlock_downgrade(void *opaque)
+{
+ qemu_co_rwlock_wrlock(&rwlock);
+
+ qemu_co_rwlock_downgrade(&rwlock);
+ qemu_co_rwlock_unlock(&rwlock);
+ *(bool *)opaque = true;
+}
+
+static void coroutine_fn rwlock_rdlock(void *opaque)
+{
+ qemu_co_rwlock_rdlock(&rwlock);
+
+ qemu_co_rwlock_unlock(&rwlock);
+ *(bool *)opaque = true;
+}
+
+static void coroutine_fn rwlock_wrlock(void *opaque)
+{
+ qemu_co_rwlock_wrlock(&rwlock);
+
+ qemu_co_rwlock_unlock(&rwlock);
+ *(bool *)opaque = true;
+}
+
+/*
+ * Check that downgrading a reader-writer lock does not cause a hang.
+ *
+ * Four coroutines are used to produce a situation where there are
+ * both reader and writer hopefuls waiting to acquire an rwlock that
+ * is held by a reader.
+ *
+ * The correct sequence of operations we aim to provoke can be
+ * represented as:
+ *
+ * | c1 | c2 | c3 | c4 |
+ * |--------+------------+------------+------------|
+ * | rdlock | | | |
+ * | yield | | | |
+ * | | wrlock | | |
+ * | | <queued> | | |
+ * | | | rdlock | |
+ * | | | <queued> | |
+ * | | | | wrlock |
+ * | | | | <queued> |
+ * | unlock | | | |
+ * | yield | | | |
+ * | | <dequeued> | | |
+ * | | downgrade | | |
+ * | | | <dequeued> | |
+ * | | | unlock | |
+ * | | ... | | |
+ * | | unlock | | |
+ * | | | | <dequeued> |
+ * | | | | unlock |
+ */
+static void test_co_rwlock_downgrade(void)
+{
+ bool c1_done = false;
+ bool c2_done = false;
+ bool c3_done = false;
+ bool c4_done = false;
+ Coroutine *c1, *c2, *c3, *c4;
+
+ qemu_co_rwlock_init(&rwlock);
+
+ c1 = qemu_coroutine_create(rwlock_rdlock_yield, &c1_done);
+ c2 = qemu_coroutine_create(rwlock_wrlock_downgrade, &c2_done);
+ c3 = qemu_coroutine_create(rwlock_rdlock, &c3_done);
+ c4 = qemu_coroutine_create(rwlock_wrlock, &c4_done);
+
+ qemu_coroutine_enter(c1);
+ qemu_coroutine_enter(c2);
+ qemu_coroutine_enter(c3);
+ qemu_coroutine_enter(c4);
+
+ qemu_coroutine_enter(c1);
+
+ g_assert(c2_done);
+ g_assert(c3_done);
+ g_assert(c4_done);
+
+ qemu_coroutine_enter(c1);
+
+ g_assert(c1_done);
+}
+
/*
* Check that creation, enter, and return work
*/
@@ -501,6 +660,8 @@ int main(int argc, char **argv)
g_test_add_func("/basic/order", test_order);
g_test_add_func("/locking/co-mutex", test_co_mutex);
g_test_add_func("/locking/co-mutex/lockable", test_co_mutex_lockable);
+ g_test_add_func("/locking/co-rwlock/upgrade", test_co_rwlock_upgrade);
+ g_test_add_func("/locking/co-rwlock/downgrade", test_co_rwlock_downgrade);
if (g_test_perf()) {
g_test_add_func("/perf/lifecycle", perf_lifecycle);
g_test_add_func("/perf/nesting", perf_nesting);