bug-coreutils
[Top][All Lists]
Advanced

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

bug#21061: coreutils-8.24 - Partially reproducible failures of tests/mis


From: Pádraig Brady
Subject: bug#21061: coreutils-8.24 - Partially reproducible failures of tests/misc/timeout-parameters.sh
Date: Wed, 15 Jul 2015 11:30:59 +0100
User-agent: Mozilla/5.0 (X11; Linux x86_64; rv:31.0) Gecko/20100101 Thunderbird/31.6.0

On 15/07/15 10:22, Peter Bray wrote:
> Greetings,
> 
> N.B. This bug report is for reference only, and documents only a
>       partially reproducible check failure. No Action Requested.
> 
> On Solaris 10 (Update 8 and Update 11) and Solaris 11.2 X86 VMs, and
> one Solaris 10 Update 10 (Non-VM) system, I see random "gmake check"
> failures for "tests/misc/timeout-parameters.sh".
> 
> Running the test by itself (with the command line below) on the same
> VMs / real system will sometimes succeed and sometimes fail.
> 
>    gmake check TESTS=tests/misc/timeout-parameters.sh VERBOSE=yes SUBDIRS=.
> 
> Looking through the attached "failure.log" file, I extracted the
> following command line test, which may exhibit the failure without all
> the make(1) and test infrastructure code:
> 
>    failures=0
>    for i in `./src/seq 1 100`
>    do
>      ./src/timeout 2.34e+5d sleep 0 \
>        || { echo fail; failures=`expr ${failures} + 1` }
>    done
>    echo "Total Failures: ${failures}"
> 
> On a real hardware system (Xeon E3-1245v2) with a 64-bit kernel,
> failures are very rare (only 1 test harness failure seen, no failures
> of the sample code above even with 1..1000 runs).
> 
> On virtual machines (also using Xeon E3-1245v2 running VMware ESXi
> 5.5d (latest patches) - two identical ESXi systems running similarly
> configured VMs), test harness failures and failures in the above
> command line check are rare for the 64-bit Solaris kernels.
> 
> Failures on Solaris 10 32-bit kernels (on both of these ESXi servers),
> are easily reproduced and vary between 5% (common) and 45% (rare).

Interesting. I'm not reproducing that in 5000 loops in the above test script
on 32 bit baremetal solaris 10 update 10.

I presume the large timeout value is causing early timer firing
on your systems for some reason? What does this return?

  time src/timeout 2.34e+5d sleep inf

Note on 32 bit, the 234000 days will be truncated to a itimerspec of:
  { {0,0}, {2147483647,999999999} }

A wild guess is that perhaps ntp is adjusting the system time
which causes the above timer to be adjusted in the kernel
and roll over to 0, thus triggering early?

thanks,
Pádraig.





reply via email to

[Prev in Thread] Current Thread [Next in Thread]