qemu-devel
[Top][All Lists]
Advanced

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

[PATCH] tests/tcg: Skip failing ppc64 multi-threaded tests in CI


From: Nicholas Piggin
Subject: [PATCH] tests/tcg: Skip failing ppc64 multi-threaded tests in CI
Date: Fri, 26 Jul 2024 00:59:03 +1000

In Gitlab CI, some ppc64 multi-threaded tcg tests crash when run in the
clang-user job with an assertion failure in glibc that seems to
indicate corruption:

  signals: allocatestack.c:223: allocate_stack:
    Assertion `powerof2 (pagesize_m1 + 1)' failed.

Disable these tests on CI for now.

Signed-off-by: Nicholas Piggin <npiggin@gmail.com>
---
gcc compiled host does work in CI, though I don't know if there is a
simple way to test for that in the makefile. clang compiled crash does
not reproduce on my local system / compiler, so progress in debugging
this has been slow.

Thanks,
Nick
---
 tests/tcg/ppc64/Makefile.target | 15 +++++++++++++++
 1 file changed, 15 insertions(+)

diff --git a/tests/tcg/ppc64/Makefile.target b/tests/tcg/ppc64/Makefile.target
index 8c3e4e4038..d57f9338b4 100644
--- a/tests/tcg/ppc64/Makefile.target
+++ b/tests/tcg/ppc64/Makefile.target
@@ -11,6 +11,21 @@ config-cc.mak: Makefile
 
 -include config-cc.mak
 
+ifneq ($(GITLAB_CI),)
+# When multi-thread tests are run under the clang-user in CI, it causes crashes
+# in glibc, so disable for now.
+# signals: allocatestack.c:223: allocate_stack: Assertion `powerof2 
(pagesize_m1 + 1)' failed.
+run-signals: signals
+       $(call skip-test, $<, "Broken on Gitlab CI for clang-user")
+run-plugin-signals-with-%:
+       $(call skip-test, $<, "Broken on Gitlab CI for clang-user")
+
+run-threadcount: threadcount
+       $(call skip-test, $<, "Broken on Gitlab CI for clang-user")
+run-plugin-threadcount-with-%:
+       $(call skip-test, $<, "Broken on Gitlab CI for clang-user")
+endif
+
 ifneq ($(CROSS_CC_HAS_POWER8_VECTOR),)
 PPC64_TESTS=bcdsub non_signalling_xscv
 endif
-- 
2.45.2




reply via email to

[Prev in Thread] Current Thread [Next in Thread]