bpo-35513: Replace time.time() with time.monotonic() in tests (GH-11182)

Replace time.time() with time.monotonic() in tests to measure time
delta.

test_zipfile64: display progress every minute (60 secs) rather than
every 5 minutes (5*60 seconds).
(cherry picked from commit 2cf4c202ff)

Co-authored-by: Victor Stinner <vstinner@redhat.com>
This commit is contained in:
Miss Islington (bot) 2018-12-17 01:03:04 -08:00 committed by GitHub
parent 2d91a1325f
commit be69ff232d
No known key found for this signature in database
GPG key ID: 4AEE18F83AFDEB23
14 changed files with 68 additions and 66 deletions

View file

@ -2213,14 +2213,14 @@ def _start_server(urlhandler, hostname, port):
Let the server do its thing. We just need to monitor its status.
Use time.sleep so the loop doesn't hog the CPU.
>>> starttime = time.time()
>>> starttime = time.monotonic()
>>> timeout = 1 #seconds
This is a short timeout for testing purposes.
>>> while serverthread.serving:
... time.sleep(.01)
... if serverthread.serving and time.time() - starttime > timeout:
... if serverthread.serving and time.monotonic() - starttime > timeout:
... serverthread.stop()
... break

View file

@ -157,11 +157,11 @@ def __init__(self, func):
self.elapsed = None
def __call__(self, *args, **kwds):
t = time.time()
t = time.monotonic()
try:
return self.func(*args, **kwds)
finally:
self.elapsed = time.time() - t
self.elapsed = time.monotonic() - t
#
# Base class for test cases
@ -1036,9 +1036,9 @@ def test_no_import_lock_contention(self):
def test_timeout(self):
q = multiprocessing.Queue()
start = time.time()
start = time.monotonic()
self.assertRaises(pyqueue.Empty, q.get, True, 0.200)
delta = time.time() - start
delta = time.monotonic() - start
# bpo-30317: Tolerate a delta of 100 ms because of the bad clock
# resolution on Windows (usually 15.6 ms). x86 Windows7 3.x once
# failed because the delta was only 135.8 ms.
@ -1434,9 +1434,9 @@ def _test_waitfor_timeout_f(cls, cond, state, success, sem):
sem.release()
with cond:
expected = 0.1
dt = time.time()
dt = time.monotonic()
result = cond.wait_for(lambda : state.value==4, timeout=expected)
dt = time.time() - dt
dt = time.monotonic() - dt
# borrow logic in assertTimeout() from test/lock_tests.py
if not result and expected * 0.6 < dt < expected * 10.0:
success.value = True
@ -2527,7 +2527,7 @@ def test_map_no_failfast(self):
# process would fill the result queue (after the result handler thread
# terminated, hence not draining it anymore).
t_start = time.time()
t_start = time.monotonic()
with self.assertRaises(ValueError):
with self.Pool(2) as p:
@ -2539,7 +2539,7 @@ def test_map_no_failfast(self):
p.join()
# check that we indeed waited for all jobs
self.assertGreater(time.time() - t_start, 0.9)
self.assertGreater(time.monotonic() - t_start, 0.9)
def test_release_task_refs(self):
# Issue #29861: task arguments and results should not be kept
@ -4051,9 +4051,9 @@ def test_wait_timeout(self):
expected = 5
a, b = multiprocessing.Pipe()
start = time.time()
start = time.monotonic()
res = wait([a, b], expected)
delta = time.time() - start
delta = time.monotonic() - start
self.assertEqual(res, [])
self.assertLess(delta, expected * 2)
@ -4061,9 +4061,9 @@ def test_wait_timeout(self):
b.send(None)
start = time.time()
start = time.monotonic()
res = wait([a, b], 20)
delta = time.time() - start
delta = time.monotonic() - start
self.assertEqual(res, [a])
self.assertLess(delta, 0.4)
@ -4087,9 +4087,9 @@ def test_wait_integer(self):
self.assertIsInstance(p.sentinel, int)
self.assertTrue(sem.acquire(timeout=20))
start = time.time()
start = time.monotonic()
res = wait([a, p.sentinel, b], expected + 20)
delta = time.time() - start
delta = time.monotonic() - start
self.assertEqual(res, [p.sentinel])
self.assertLess(delta, expected + 2)
@ -4097,18 +4097,18 @@ def test_wait_integer(self):
a.send(None)
start = time.time()
start = time.monotonic()
res = wait([a, p.sentinel, b], 20)
delta = time.time() - start
delta = time.monotonic() - start
self.assertEqual(sorted_(res), sorted_([p.sentinel, b]))
self.assertLess(delta, 0.4)
b.send(None)
start = time.time()
start = time.monotonic()
res = wait([a, p.sentinel, b], 20)
delta = time.time() - start
delta = time.monotonic() - start
self.assertEqual(sorted_(res), sorted_([a, p.sentinel, b]))
self.assertLess(delta, 0.4)
@ -4119,9 +4119,9 @@ def test_wait_integer(self):
def test_neg_timeout(self):
from multiprocessing.connection import wait
a, b = multiprocessing.Pipe()
t = time.time()
t = time.monotonic()
res = wait([a], timeout=-1)
t = time.time() - t
t = time.monotonic() - t
self.assertEqual(res, [])
self.assertLess(t, 1)
a.close()

View file

@ -74,7 +74,7 @@ def tearDown(self):
support.reap_children()
def assertTimeout(self, actual, expected):
# The waiting and/or time.time() can be imprecise, which
# The waiting and/or time.monotonic() can be imprecise, which
# is why comparing to the expected value would sometimes fail
# (especially under Windows).
self.assertGreaterEqual(actual, expected * 0.6)
@ -190,16 +190,16 @@ def test_timeout(self):
# TIMEOUT_MAX is ok
lock.acquire(timeout=TIMEOUT_MAX)
lock.release()
t1 = time.time()
t1 = time.monotonic()
self.assertTrue(lock.acquire(timeout=5))
t2 = time.time()
t2 = time.monotonic()
# Just a sanity test that it didn't actually wait for the timeout.
self.assertLess(t2 - t1, 5)
results = []
def f():
t1 = time.time()
t1 = time.monotonic()
results.append(lock.acquire(timeout=0.5))
t2 = time.time()
t2 = time.monotonic()
results.append(t2 - t1)
Bunch(f, 1).wait_for_finished()
self.assertFalse(results[0])
@ -382,9 +382,9 @@ def test_timeout(self):
N = 5
def f():
results1.append(evt.wait(0.0))
t1 = time.time()
t1 = time.monotonic()
r = evt.wait(0.5)
t2 = time.time()
t2 = time.monotonic()
results2.append((r, t2 - t1))
Bunch(f, N).wait_for_finished()
self.assertEqual(results1, [False] * N)
@ -545,9 +545,9 @@ def test_timeout(self):
N = 5
def f():
cond.acquire()
t1 = time.time()
t1 = time.monotonic()
result = cond.wait(0.5)
t2 = time.time()
t2 = time.monotonic()
cond.release()
results.append((t2 - t1, result))
Bunch(f, N).wait_for_finished()
@ -584,9 +584,9 @@ def test_waitfor_timeout(self):
success = []
def f():
with cond:
dt = time.time()
dt = time.monotonic()
result = cond.wait_for(lambda : state==4, timeout=0.1)
dt = time.time() - dt
dt = time.monotonic() - dt
self.assertFalse(result)
self.assertTimeout(dt, 0.1)
success.append(None)
@ -692,9 +692,9 @@ def test_acquire_timeout(self):
self.assertFalse(sem.acquire(timeout=0.005))
sem.release()
self.assertTrue(sem.acquire(timeout=0.005))
t = time.time()
t = time.monotonic()
self.assertFalse(sem.acquire(timeout=0.5))
dt = time.time() - t
dt = time.monotonic() - t
self.assertTimeout(dt, 0.5)
def test_default_value(self):

View file

@ -2219,11 +2219,11 @@ def start_threads(threads, unlock=None):
try:
if unlock:
unlock()
endtime = starttime = time.time()
endtime = starttime = time.monotonic()
for timeout in range(1, 16):
endtime += 60
for t in started:
t.join(max(endtime - time.time(), 0.01))
t.join(max(endtime - time.monotonic(), 0.01))
started = [t for t in started if t.isAlive()]
if not started:
break

View file

@ -107,10 +107,10 @@ async def once():
def run_until(loop, pred, timeout=30):
deadline = time.time() + timeout
deadline = time.monotonic() + timeout
while not pred():
if timeout is not None:
timeout = deadline - time.time()
timeout = deadline - time.monotonic()
if timeout <= 0:
raise futures.TimeoutError()
loop.run_until_complete(tasks.sleep(0.001, loop=loop))

View file

@ -70,8 +70,8 @@ def capture_server(evt, buf, serv):
pass
else:
n = 200
start = time.time()
while n > 0 and time.time() - start < 3.0:
start = time.monotonic()
while n > 0 and time.monotonic() - start < 3.0:
r, w, e = select.select([conn], [], [], 0.1)
if r:
n -= 1

View file

@ -70,14 +70,14 @@ def delay_unlock(to_unlock, delay):
to_unlock.release()
self.lock.acquire()
start_time = int(time.time())
start_time = int(time.monotonic())
_thread.start_new_thread(delay_unlock,(self.lock, DELAY))
if support.verbose:
print()
print("*** Waiting for thread to release the lock "\
"(approx. %s sec.) ***" % DELAY)
self.lock.acquire()
end_time = int(time.time())
end_time = int(time.monotonic())
if support.verbose:
print("done")
self.assertGreaterEqual(end_time - start_time, DELAY,

View file

@ -77,10 +77,10 @@ def play_sound_file(self, data, rate, ssize, nchannels):
# set parameters based on .au file headers
dsp.setparameters(AFMT_S16_NE, nchannels, rate)
self.assertTrue(abs(expected_time - 3.51) < 1e-2, expected_time)
t1 = time.time()
t1 = time.monotonic()
dsp.write(data)
dsp.close()
t2 = time.time()
t2 = time.monotonic()
elapsed_time = t2 - t1
percent_diff = (abs(elapsed_time - expected_time) / expected_time) * 100

View file

@ -1013,12 +1013,12 @@ def my_url_handler(url, content_type):
serverthread = pydoc._start_server(my_url_handler, hostname='0.0.0.0', port=0)
self.assertIn('0.0.0.0', serverthread.docserver.address)
starttime = time.time()
starttime = time.monotonic()
timeout = 1 #seconds
while serverthread.serving:
time.sleep(.01)
if serverthread.serving and time.time() - starttime > timeout:
if serverthread.serving and time.monotonic() - starttime > timeout:
serverthread.stop()
break

View file

@ -1122,18 +1122,18 @@ def second_handler(signum=None, frame=None):
self.setsig(signal.SIGALRM, second_handler) # for ITIMER_REAL
expected_sigs = 0
deadline = time.time() + 15.0
deadline = time.monotonic() + 15.0
while expected_sigs < N:
os.kill(os.getpid(), signal.SIGPROF)
expected_sigs += 1
# Wait for handlers to run to avoid signal coalescing
while len(sigs) < expected_sigs and time.time() < deadline:
while len(sigs) < expected_sigs and time.monotonic() < deadline:
time.sleep(1e-5)
os.kill(os.getpid(), signal.SIGUSR1)
expected_sigs += 1
while len(sigs) < expected_sigs and time.time() < deadline:
while len(sigs) < expected_sigs and time.monotonic() < deadline:
time.sleep(1e-5)
# All ITIMER_REAL signals should have been delivered to the
@ -1156,7 +1156,7 @@ def handler(signum, frame):
self.setsig(signal.SIGALRM, handler) # for ITIMER_REAL
expected_sigs = 0
deadline = time.time() + 15.0
deadline = time.monotonic() + 15.0
while expected_sigs < N:
# Hopefully the SIGALRM will be received somewhere during
@ -1166,7 +1166,7 @@ def handler(signum, frame):
expected_sigs += 2
# Wait for handlers to run to avoid signal coalescing
while len(sigs) < expected_sigs and time.time() < deadline:
while len(sigs) < expected_sigs and time.monotonic() < deadline:
time.sleep(1e-5)
# All ITIMER_REAL signals should have been delivered to the

View file

@ -95,9 +95,9 @@ def test_lock_acquire_interruption(self):
lock = thread.allocate_lock()
lock.acquire()
signal.alarm(1)
t1 = time.time()
t1 = time.monotonic()
self.assertRaises(KeyboardInterrupt, lock.acquire, timeout=5)
dt = time.time() - t1
dt = time.monotonic() - t1
# Checking that KeyboardInterrupt was raised is not sufficient.
# We want to assert that lock.acquire() was interrupted because
# of the signal, not that the signal handler was called immediately
@ -136,9 +136,9 @@ def other_thread():
rlock.release()
time.sleep(0.01)
signal.alarm(1)
t1 = time.time()
t1 = time.monotonic()
self.assertRaises(KeyboardInterrupt, rlock.acquire, timeout=5)
dt = time.time() - t1
dt = time.monotonic() - t1
# See rationale above in test_lock_acquire_interruption
self.assertLess(dt, 3.0)
finally:
@ -203,9 +203,9 @@ def my_handler(signum, frame):
old_handler = signal.signal(signal.SIGUSR1, my_handler)
try:
def timed_acquire():
self.start = time.time()
self.start = time.monotonic()
lock.acquire(timeout=0.5)
self.end = time.time()
self.end = time.monotonic()
def send_signals():
for _ in range(40):
time.sleep(0.02)

View file

@ -127,11 +127,11 @@ def _sock_operation(self, count, timeout, method, *args):
self.sock.settimeout(timeout)
method = getattr(self.sock, method)
for i in range(count):
t1 = time.time()
t1 = time.monotonic()
try:
method(*args)
except socket.timeout as e:
delta = time.time() - t1
delta = time.monotonic() - t1
break
else:
self.fail('socket.timeout was not raised')

View file

@ -22,7 +22,7 @@
TESTFN2 = TESTFN + "2"
# How much time in seconds can pass before we print a 'Still working' message.
_PRINT_WORKING_MSG_INTERVAL = 5 * 60
_PRINT_WORKING_MSG_INTERVAL = 60
class TestsWithSourceFile(unittest.TestCase):
def setUp(self):
@ -43,12 +43,12 @@ def zipTest(self, f, compression):
# raw data to store.
filecount = 6*1024**3 // len(self.data)
next_time = time.time() + _PRINT_WORKING_MSG_INTERVAL
next_time = time.monotonic() + _PRINT_WORKING_MSG_INTERVAL
for num in range(filecount):
zipfp.writestr("testfn%d" % num, self.data)
# Print still working message since this test can be really slow
if next_time <= time.time():
next_time = time.time() + _PRINT_WORKING_MSG_INTERVAL
if next_time <= time.monotonic():
next_time = time.monotonic() + _PRINT_WORKING_MSG_INTERVAL
print((
' zipTest still writing %d of %d, be patient...' %
(num, filecount)), file=sys.__stdout__)
@ -60,8 +60,8 @@ def zipTest(self, f, compression):
for num in range(filecount):
self.assertEqual(zipfp.read("testfn%d" % num), self.data)
# Print still working message since this test can be really slow
if next_time <= time.time():
next_time = time.time() + _PRINT_WORKING_MSG_INTERVAL
if next_time <= time.monotonic():
next_time = time.monotonic() + _PRINT_WORKING_MSG_INTERVAL
print((
' zipTest still reading %d of %d, be patient...' %
(num, filecount)), file=sys.__stdout__)

View file

@ -0,0 +1,2 @@
Replace :func:`time.time` with :func:`time.monotonic` in tests to measure
time delta.