| 
									
										
										
										
											2022-10-25 22:00:50 -07:00
										 |  |  | """Python implementations of some algorithms for use by longobject.c.
 | 
					
						
							|  |  |  | The goal is to provide asymptotically faster algorithms that can be | 
					
						
							|  |  |  | used for operations on integers with many digits.  In those cases, the | 
					
						
							|  |  |  | performance overhead of the Python implementation is not significant | 
					
						
							|  |  |  | since the asymptotic behavior is what dominates runtime. Functions | 
					
						
							|  |  |  | provided by this module should be considered private and not part of any | 
					
						
							|  |  |  | public API. | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | Note: for ease of maintainability, please prefer clear code and avoid | 
					
						
							|  |  |  | "micro-optimizations".  This module will only be imported and used for | 
					
						
							|  |  |  | integers with a huge number of digits.  Saving a few microseconds with | 
					
						
							|  |  |  | tricky or non-obvious code is not worth it.  For people looking for | 
					
						
							|  |  |  | maximum performance, they should use something like gmpy2."""
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | import re | 
					
						
							|  |  |  | import decimal | 
					
						
							| 
									
										
											  
											
												gh-118164: Break a loop between _pydecimal and _pylong and optimize int to str conversion (GH-118483)
For converting large ints to strings, CPython invokes a function in _pylong.py,
which uses the decimal module to implement an asymptotically waaaaay
sub-quadratic algorithm. But if the C decimal module isn't available, CPython
uses _pydecimal.py instead. Which in turn frequently does str(int). If the int
is very large, _pylong ends up doing the work, which in turn asks decimal to do
"big" arithmetic, which in turn calls str(big_int), which in turn ... it can
become infinite mutual recursion.
This change introduces a different int->str function that doesn't use decimal.
It's asymptotically worse, "Karatsuba time" instead of quadratic time, so
still a huge improvement. _pylong switches to that when the C decimal isn't
available. It is also used for not too large integers (less than 450_000 bits),
where it is faster (up to 2 times for 30_000 bits) than the asymptotically
better implementation that uses the C decimal.
Co-authored-by: Tim Peters <tim.peters@gmail.com>
											
										 
											2024-05-05 08:20:06 +03:00
										 |  |  | try: | 
					
						
							|  |  |  |     import _decimal | 
					
						
							|  |  |  | except ImportError: | 
					
						
							|  |  |  |     _decimal = None | 
					
						
							| 
									
										
										
										
											2022-10-25 22:00:50 -07:00
										 |  |  | 
 | 
					
						
							| 
									
										
										
										
											2024-05-07 19:09:09 -05:00
										 |  |  | # A number of functions have this form, where `w` is a desired number of | 
					
						
							|  |  |  | # digits in base `base`: | 
					
						
							|  |  |  | # | 
					
						
							|  |  |  | #    def inner(...w...): | 
					
						
							|  |  |  | #        if w <= LIMIT: | 
					
						
							|  |  |  | #            return something | 
					
						
							|  |  |  | #        lo = w >> 1 | 
					
						
							|  |  |  | #        hi = w - lo | 
					
						
							|  |  |  | #        something involving base**lo, inner(...lo...), j, and inner(...hi...) | 
					
						
							|  |  |  | #    figure out largest w needed | 
					
						
							|  |  |  | #    result = inner(w) | 
					
						
							|  |  |  | # | 
					
						
							|  |  |  | # They all had some on-the-fly scheme to cache `base**lo` results for reuse. | 
					
						
							|  |  |  | # Power is costly. | 
					
						
							|  |  |  | # | 
					
						
							|  |  |  | # This routine aims to compute all amd only the needed powers in advance, as | 
					
						
							|  |  |  | # efficiently as reasonably possible. This isn't trivial, and all the | 
					
						
							|  |  |  | # on-the-fly methods did needless work in many cases. The driving code above | 
					
						
							|  |  |  | # changes to: | 
					
						
							|  |  |  | # | 
					
						
							|  |  |  | #    figure out largest w needed | 
					
						
							|  |  |  | #    mycache = compute_powers(w, base, LIMIT) | 
					
						
							|  |  |  | #    result = inner(w) | 
					
						
							|  |  |  | # | 
					
						
							|  |  |  | # and `mycache[lo]` replaces `base**lo` in the inner function. | 
					
						
							|  |  |  | # | 
					
						
							| 
									
										
										
										
											2024-05-18 19:19:57 -05:00
										 |  |  | # If an algorithm wants the powers of ceiling(w/2) instead of the floor, | 
					
						
							|  |  |  | # pass keyword argument `need_hi=True`. | 
					
						
							|  |  |  | # | 
					
						
							|  |  |  | # While this does give minor speedups (a few percent at best), the | 
					
						
							|  |  |  | # primary intent is to simplify the functions using this, by eliminating | 
					
						
							|  |  |  | # the need for them to craft their own ad-hoc caching schemes. | 
					
						
							|  |  |  | # | 
					
						
							|  |  |  | # See code near end of file for a block of code that can be enabled to | 
					
						
							|  |  |  | # run millions of tests. | 
					
						
							|  |  |  | def compute_powers(w, base, more_than, *, need_hi=False, show=False): | 
					
						
							| 
									
										
										
										
											2024-05-07 19:09:09 -05:00
										 |  |  |     seen = set() | 
					
						
							|  |  |  |     need = set() | 
					
						
							|  |  |  |     ws = {w} | 
					
						
							|  |  |  |     while ws: | 
					
						
							|  |  |  |         w = ws.pop() # any element is fine to use next | 
					
						
							|  |  |  |         if w in seen or w <= more_than: | 
					
						
							|  |  |  |             continue | 
					
						
							|  |  |  |         seen.add(w) | 
					
						
							|  |  |  |         lo = w >> 1 | 
					
						
							| 
									
										
										
										
											2024-05-18 19:19:57 -05:00
										 |  |  |         hi = w - lo | 
					
						
							|  |  |  |         # only _need_ one here; the other may, or may not, be needed | 
					
						
							|  |  |  |         which = hi if need_hi else lo | 
					
						
							|  |  |  |         need.add(which) | 
					
						
							|  |  |  |         ws.add(which) | 
					
						
							|  |  |  |         if lo != hi: | 
					
						
							|  |  |  |             ws.add(w - which) | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     # `need` is the set of exponents needed. To compute them all | 
					
						
							|  |  |  |     # efficiently, possibly add other exponents to `extra`. The goal is | 
					
						
							|  |  |  |     # to ensure that each exponent can be gotten from a smaller one via | 
					
						
							|  |  |  |     # multiplying by the base, squaring it, or squaring and then | 
					
						
							|  |  |  |     # multiplying by the base. | 
					
						
							|  |  |  |     # | 
					
						
							|  |  |  |     # If need_hi is False, this is already the case (w can always be | 
					
						
							|  |  |  |     # gotten from w >> 1 via one of the squaring strategies). But we do | 
					
						
							|  |  |  |     # the work anyway, just in case ;-) | 
					
						
							|  |  |  |     # | 
					
						
							|  |  |  |     # Note that speed is irrelevant. These loops are working on little | 
					
						
							|  |  |  |     # ints (exponents) and go around O(log w) times. The total cost is | 
					
						
							|  |  |  |     # insignificant compared to just one of the bigint multiplies. | 
					
						
							|  |  |  |     cands = need.copy() | 
					
						
							|  |  |  |     extra = set() | 
					
						
							|  |  |  |     while cands: | 
					
						
							|  |  |  |         w = max(cands) | 
					
						
							|  |  |  |         cands.remove(w) | 
					
						
							|  |  |  |         lo = w >> 1 | 
					
						
							|  |  |  |         if lo > more_than and w-1 not in cands and lo not in cands: | 
					
						
							|  |  |  |             extra.add(lo) | 
					
						
							|  |  |  |             cands.add(lo) | 
					
						
							|  |  |  |     assert need_hi or not extra | 
					
						
							| 
									
										
										
										
											2024-05-07 19:09:09 -05:00
										 |  |  | 
 | 
					
						
							|  |  |  |     d = {} | 
					
						
							| 
									
										
										
										
											2024-05-18 19:19:57 -05:00
										 |  |  |     for n in sorted(need | extra): | 
					
						
							|  |  |  |         lo = n >> 1 | 
					
						
							|  |  |  |         hi = n - lo | 
					
						
							|  |  |  |         if n-1 in d: | 
					
						
							| 
									
										
										
										
											2024-05-07 19:09:09 -05:00
										 |  |  |             if show: | 
					
						
							| 
									
										
										
										
											2024-05-18 19:19:57 -05:00
										 |  |  |                 print("* base", end="") | 
					
						
							|  |  |  |             result = d[n-1] * base # cheap! | 
					
						
							|  |  |  |         elif lo in d: | 
					
						
							|  |  |  |             # Multiplying a bigint by itself is about twice as fast | 
					
						
							|  |  |  |             # in CPython provided it's the same object. | 
					
						
							| 
									
										
										
										
											2024-05-07 19:09:09 -05:00
										 |  |  |             if show: | 
					
						
							| 
									
										
										
										
											2024-05-18 19:19:57 -05:00
										 |  |  |                 print("square", end="") | 
					
						
							|  |  |  |             result = d[lo] * d[lo] # same object | 
					
						
							| 
									
										
										
										
											2024-05-07 19:09:09 -05:00
										 |  |  |             if hi != lo: | 
					
						
							|  |  |  |                 if show: | 
					
						
							| 
									
										
										
										
											2024-05-18 19:19:57 -05:00
										 |  |  |                     print(" * base", end="") | 
					
						
							|  |  |  |                 assert 2 * lo + 1 == n | 
					
						
							|  |  |  |                 result *= base | 
					
						
							|  |  |  |         else: # rare | 
					
						
							|  |  |  |             if show: | 
					
						
							|  |  |  |                 print("pow", end='') | 
					
						
							|  |  |  |             result = base ** n | 
					
						
							|  |  |  |         if show: | 
					
						
							|  |  |  |             print(" at", n, "needed" if n in need else "extra") | 
					
						
							|  |  |  |         d[n] = result | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     assert need <= d.keys() | 
					
						
							|  |  |  |     if excess := d.keys() - need: | 
					
						
							|  |  |  |         assert need_hi | 
					
						
							|  |  |  |         for n in excess: | 
					
						
							|  |  |  |             del d[n] | 
					
						
							| 
									
										
										
										
											2024-05-07 19:09:09 -05:00
										 |  |  |     return d | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | _unbounded_dec_context = decimal.getcontext().copy() | 
					
						
							|  |  |  | _unbounded_dec_context.prec = decimal.MAX_PREC | 
					
						
							|  |  |  | _unbounded_dec_context.Emax = decimal.MAX_EMAX | 
					
						
							|  |  |  | _unbounded_dec_context.Emin = decimal.MIN_EMIN | 
					
						
							|  |  |  | _unbounded_dec_context.traps[decimal.Inexact] = 1 # sanity check | 
					
						
							| 
									
										
										
										
											2022-10-25 22:00:50 -07:00
										 |  |  | 
 | 
					
						
							|  |  |  | def int_to_decimal(n): | 
					
						
							|  |  |  |     """Asymptotically fast conversion of an 'int' to Decimal.""" | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     # Function due to Tim Peters.  See GH issue #90716 for details. | 
					
						
							|  |  |  |     # https://github.com/python/cpython/issues/90716 | 
					
						
							|  |  |  |     # | 
					
						
							|  |  |  |     # The implementation in longobject.c of base conversion algorithms | 
					
						
							|  |  |  |     # between power-of-2 and non-power-of-2 bases are quadratic time. | 
					
						
							|  |  |  |     # This function implements a divide-and-conquer algorithm that is | 
					
						
							|  |  |  |     # faster for large numbers.  Builds an equal decimal.Decimal in a | 
					
						
							|  |  |  |     # "clever" recursive way.  If we want a string representation, we | 
					
						
							|  |  |  |     # apply str to _that_. | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2024-05-07 19:09:09 -05:00
										 |  |  |     from decimal import Decimal as D | 
					
						
							|  |  |  |     BITLIM = 200 | 
					
						
							| 
									
										
										
										
											2022-10-25 22:00:50 -07:00
										 |  |  | 
 | 
					
						
							| 
									
										
										
										
											2024-05-07 19:09:09 -05:00
										 |  |  |     # Don't bother caching the "lo" mask in this; the time to compute it is | 
					
						
							|  |  |  |     # tiny compared to the multiply. | 
					
						
							| 
									
										
										
										
											2022-10-25 22:00:50 -07:00
										 |  |  |     def inner(n, w): | 
					
						
							|  |  |  |         if w <= BITLIM: | 
					
						
							|  |  |  |             return D(n) | 
					
						
							|  |  |  |         w2 = w >> 1 | 
					
						
							|  |  |  |         hi = n >> w2 | 
					
						
							| 
									
										
										
										
											2024-05-07 19:09:09 -05:00
										 |  |  |         lo = n & ((1 << w2) - 1) | 
					
						
							|  |  |  |         return inner(lo, w2) + inner(hi, w - w2) * w2pow[w2] | 
					
						
							| 
									
										
										
										
											2022-10-25 22:00:50 -07:00
										 |  |  | 
 | 
					
						
							| 
									
										
										
										
											2024-05-07 19:09:09 -05:00
										 |  |  |     with decimal.localcontext(_unbounded_dec_context): | 
					
						
							|  |  |  |         nbits = n.bit_length() | 
					
						
							|  |  |  |         w2pow = compute_powers(nbits, D(2), BITLIM) | 
					
						
							| 
									
										
										
										
											2022-10-25 22:00:50 -07:00
										 |  |  |         if n < 0: | 
					
						
							|  |  |  |             negate = True | 
					
						
							|  |  |  |             n = -n | 
					
						
							|  |  |  |         else: | 
					
						
							|  |  |  |             negate = False | 
					
						
							| 
									
										
										
										
											2024-05-07 19:09:09 -05:00
										 |  |  |         result = inner(n, nbits) | 
					
						
							| 
									
										
										
										
											2022-10-25 22:00:50 -07:00
										 |  |  |         if negate: | 
					
						
							|  |  |  |             result = -result | 
					
						
							|  |  |  |     return result | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | def int_to_decimal_string(n): | 
					
						
							|  |  |  |     """Asymptotically fast conversion of an 'int' to a decimal string.""" | 
					
						
							| 
									
										
											  
											
												gh-118164: Break a loop between _pydecimal and _pylong and optimize int to str conversion (GH-118483)
For converting large ints to strings, CPython invokes a function in _pylong.py,
which uses the decimal module to implement an asymptotically waaaaay
sub-quadratic algorithm. But if the C decimal module isn't available, CPython
uses _pydecimal.py instead. Which in turn frequently does str(int). If the int
is very large, _pylong ends up doing the work, which in turn asks decimal to do
"big" arithmetic, which in turn calls str(big_int), which in turn ... it can
become infinite mutual recursion.
This change introduces a different int->str function that doesn't use decimal.
It's asymptotically worse, "Karatsuba time" instead of quadratic time, so
still a huge improvement. _pylong switches to that when the C decimal isn't
available. It is also used for not too large integers (less than 450_000 bits),
where it is faster (up to 2 times for 30_000 bits) than the asymptotically
better implementation that uses the C decimal.
Co-authored-by: Tim Peters <tim.peters@gmail.com>
											
										 
											2024-05-05 08:20:06 +03:00
										 |  |  |     w = n.bit_length() | 
					
						
							|  |  |  |     if w > 450_000 and _decimal is not None: | 
					
						
							|  |  |  |         # It is only usable with the C decimal implementation. | 
					
						
							|  |  |  |         # _pydecimal.py calls str() on very large integers, which in its | 
					
						
							|  |  |  |         # turn calls int_to_decimal_string(), causing very deep recursion. | 
					
						
							|  |  |  |         return str(int_to_decimal(n)) | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     # Fallback algorithm for the case when the C decimal module isn't | 
					
						
							|  |  |  |     # available.  This algorithm is asymptotically worse than the algorithm | 
					
						
							|  |  |  |     # using the decimal module, but better than the quadratic time | 
					
						
							|  |  |  |     # implementation in longobject.c. | 
					
						
							| 
									
										
										
										
											2024-05-07 19:09:09 -05:00
										 |  |  | 
 | 
					
						
							|  |  |  |     DIGLIM = 1000 | 
					
						
							| 
									
										
											  
											
												gh-118164: Break a loop between _pydecimal and _pylong and optimize int to str conversion (GH-118483)
For converting large ints to strings, CPython invokes a function in _pylong.py,
which uses the decimal module to implement an asymptotically waaaaay
sub-quadratic algorithm. But if the C decimal module isn't available, CPython
uses _pydecimal.py instead. Which in turn frequently does str(int). If the int
is very large, _pylong ends up doing the work, which in turn asks decimal to do
"big" arithmetic, which in turn calls str(big_int), which in turn ... it can
become infinite mutual recursion.
This change introduces a different int->str function that doesn't use decimal.
It's asymptotically worse, "Karatsuba time" instead of quadratic time, so
still a huge improvement. _pylong switches to that when the C decimal isn't
available. It is also used for not too large integers (less than 450_000 bits),
where it is faster (up to 2 times for 30_000 bits) than the asymptotically
better implementation that uses the C decimal.
Co-authored-by: Tim Peters <tim.peters@gmail.com>
											
										 
											2024-05-05 08:20:06 +03:00
										 |  |  |     def inner(n, w): | 
					
						
							| 
									
										
										
										
											2024-05-07 19:09:09 -05:00
										 |  |  |         if w <= DIGLIM: | 
					
						
							| 
									
										
											  
											
												gh-118164: Break a loop between _pydecimal and _pylong and optimize int to str conversion (GH-118483)
For converting large ints to strings, CPython invokes a function in _pylong.py,
which uses the decimal module to implement an asymptotically waaaaay
sub-quadratic algorithm. But if the C decimal module isn't available, CPython
uses _pydecimal.py instead. Which in turn frequently does str(int). If the int
is very large, _pylong ends up doing the work, which in turn asks decimal to do
"big" arithmetic, which in turn calls str(big_int), which in turn ... it can
become infinite mutual recursion.
This change introduces a different int->str function that doesn't use decimal.
It's asymptotically worse, "Karatsuba time" instead of quadratic time, so
still a huge improvement. _pylong switches to that when the C decimal isn't
available. It is also used for not too large integers (less than 450_000 bits),
where it is faster (up to 2 times for 30_000 bits) than the asymptotically
better implementation that uses the C decimal.
Co-authored-by: Tim Peters <tim.peters@gmail.com>
											
										 
											2024-05-05 08:20:06 +03:00
										 |  |  |             return str(n) | 
					
						
							|  |  |  |         w2 = w >> 1 | 
					
						
							| 
									
										
										
										
											2024-05-07 19:09:09 -05:00
										 |  |  |         hi, lo = divmod(n, pow10[w2]) | 
					
						
							| 
									
										
											  
											
												gh-118164: Break a loop between _pydecimal and _pylong and optimize int to str conversion (GH-118483)
For converting large ints to strings, CPython invokes a function in _pylong.py,
which uses the decimal module to implement an asymptotically waaaaay
sub-quadratic algorithm. But if the C decimal module isn't available, CPython
uses _pydecimal.py instead. Which in turn frequently does str(int). If the int
is very large, _pylong ends up doing the work, which in turn asks decimal to do
"big" arithmetic, which in turn calls str(big_int), which in turn ... it can
become infinite mutual recursion.
This change introduces a different int->str function that doesn't use decimal.
It's asymptotically worse, "Karatsuba time" instead of quadratic time, so
still a huge improvement. _pylong switches to that when the C decimal isn't
available. It is also used for not too large integers (less than 450_000 bits),
where it is faster (up to 2 times for 30_000 bits) than the asymptotically
better implementation that uses the C decimal.
Co-authored-by: Tim Peters <tim.peters@gmail.com>
											
										 
											2024-05-05 08:20:06 +03:00
										 |  |  |         return inner(hi, w - w2) + inner(lo, w2).zfill(w2) | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     # The estimation of the number of decimal digits. | 
					
						
							|  |  |  |     # There is no harm in small error.  If we guess too large, there may | 
					
						
							|  |  |  |     # be leading 0's that need to be stripped.  If we guess too small, we | 
					
						
							|  |  |  |     # may need to call str() recursively for the remaining highest digits, | 
					
						
							|  |  |  |     # which can still potentially be a large integer. This is manifested | 
					
						
							|  |  |  |     # only if the number has way more than 10**15 digits, that exceeds | 
					
						
							|  |  |  |     # the 52-bit physical address limit in both Intel64 and AMD64. | 
					
						
							|  |  |  |     w = int(w * 0.3010299956639812 + 1)  # log10(2) | 
					
						
							| 
									
										
										
										
											2024-05-07 19:09:09 -05:00
										 |  |  |     pow10 = compute_powers(w, 5, DIGLIM) | 
					
						
							|  |  |  |     for k, v in pow10.items(): | 
					
						
							|  |  |  |         pow10[k] = v << k # 5**k << k == 5**k * 2**k == 10**k | 
					
						
							| 
									
										
											  
											
												gh-118164: Break a loop between _pydecimal and _pylong and optimize int to str conversion (GH-118483)
For converting large ints to strings, CPython invokes a function in _pylong.py,
which uses the decimal module to implement an asymptotically waaaaay
sub-quadratic algorithm. But if the C decimal module isn't available, CPython
uses _pydecimal.py instead. Which in turn frequently does str(int). If the int
is very large, _pylong ends up doing the work, which in turn asks decimal to do
"big" arithmetic, which in turn calls str(big_int), which in turn ... it can
become infinite mutual recursion.
This change introduces a different int->str function that doesn't use decimal.
It's asymptotically worse, "Karatsuba time" instead of quadratic time, so
still a huge improvement. _pylong switches to that when the C decimal isn't
available. It is also used for not too large integers (less than 450_000 bits),
where it is faster (up to 2 times for 30_000 bits) than the asymptotically
better implementation that uses the C decimal.
Co-authored-by: Tim Peters <tim.peters@gmail.com>
											
										 
											2024-05-05 08:20:06 +03:00
										 |  |  |     if n < 0: | 
					
						
							|  |  |  |         n = -n | 
					
						
							|  |  |  |         sign = '-' | 
					
						
							|  |  |  |     else: | 
					
						
							|  |  |  |         sign = '' | 
					
						
							|  |  |  |     s = inner(n, w) | 
					
						
							|  |  |  |     if s[0] == '0' and n: | 
					
						
							|  |  |  |         # If our guess of w is too large, there may be leading 0's that | 
					
						
							|  |  |  |         # need to be stripped. | 
					
						
							|  |  |  |         s = s.lstrip('0') | 
					
						
							|  |  |  |     return sign + s | 
					
						
							| 
									
										
										
										
											2022-10-25 22:00:50 -07:00
										 |  |  | 
 | 
					
						
							|  |  |  | def _str_to_int_inner(s): | 
					
						
							|  |  |  |     """Asymptotically fast conversion of a 'str' to an 'int'.""" | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     # Function due to Bjorn Martinsson.  See GH issue #90716 for details. | 
					
						
							|  |  |  |     # https://github.com/python/cpython/issues/90716 | 
					
						
							|  |  |  |     # | 
					
						
							|  |  |  |     # The implementation in longobject.c of base conversion algorithms | 
					
						
							|  |  |  |     # between power-of-2 and non-power-of-2 bases are quadratic time. | 
					
						
							|  |  |  |     # This function implements a divide-and-conquer algorithm making use | 
					
						
							|  |  |  |     # of Python's built in big int multiplication. Since Python uses the | 
					
						
							|  |  |  |     # Karatsuba algorithm for multiplication, the time complexity | 
					
						
							|  |  |  |     # of this function is O(len(s)**1.58). | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     DIGLIM = 2048 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     def inner(a, b): | 
					
						
							|  |  |  |         if b - a <= DIGLIM: | 
					
						
							|  |  |  |             return int(s[a:b]) | 
					
						
							|  |  |  |         mid = (a + b + 1) >> 1 | 
					
						
							| 
									
										
										
										
											2024-05-07 19:09:09 -05:00
										 |  |  |         return (inner(mid, b) | 
					
						
							|  |  |  |                 + ((inner(a, mid) * w5pow[b - mid]) | 
					
						
							|  |  |  |                     << (b - mid))) | 
					
						
							| 
									
										
										
										
											2022-10-25 22:00:50 -07:00
										 |  |  | 
 | 
					
						
							| 
									
										
										
										
											2024-05-07 19:09:09 -05:00
										 |  |  |     w5pow = compute_powers(len(s), 5, DIGLIM) | 
					
						
							| 
									
										
										
										
											2022-10-25 22:00:50 -07:00
										 |  |  |     return inner(0, len(s)) | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2024-05-18 19:19:57 -05:00
										 |  |  | # Asymptotically faster version, using the C decimal module. See | 
					
						
							|  |  |  | # comments at the end of the file. This uses decimal arithmetic to | 
					
						
							|  |  |  | # convert from base 10 to base 256. The latter is just a string of | 
					
						
							|  |  |  | # bytes, which CPython can convert very efficiently to a Python int. | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | # log of 10 to base 256 with best-possible 53-bit precision. Obtained | 
					
						
							|  |  |  | # via: | 
					
						
							|  |  |  | #    from mpmath import mp | 
					
						
							|  |  |  | #    mp.prec = 1000 | 
					
						
							|  |  |  | #    print(float(mp.log(10, 256)).hex()) | 
					
						
							|  |  |  | _LOG_10_BASE_256 = float.fromhex('0x1.a934f0979a371p-2') # about 0.415 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | # _spread is for internal testing. It maps a key to the number of times | 
					
						
							|  |  |  | # that condition obtained in _dec_str_to_int_inner: | 
					
						
							|  |  |  | #     key 0 - quotient guess was right | 
					
						
							|  |  |  | #     key 1 - quotient had to be boosted by 1, one time | 
					
						
							|  |  |  | #     key 999 - one adjustment wasn't enough, so fell back to divmod | 
					
						
							|  |  |  | from collections import defaultdict | 
					
						
							|  |  |  | _spread = defaultdict(int) | 
					
						
							|  |  |  | del defaultdict | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | def _dec_str_to_int_inner(s, *, GUARD=8): | 
					
						
							|  |  |  |     # Yes, BYTELIM is "large". Large enough that CPython will usually | 
					
						
							|  |  |  |     # use the Karatsuba _str_to_int_inner to convert the string. This | 
					
						
							|  |  |  |     # allowed reducing the cutoff for calling _this_ function from 3.5M | 
					
						
							|  |  |  |     # to 2M digits. We could almost certainly do even better by | 
					
						
							|  |  |  |     # fine-tuning this and/or using a larger output base than 256. | 
					
						
							|  |  |  |     BYTELIM = 100_000 | 
					
						
							|  |  |  |     D = decimal.Decimal | 
					
						
							|  |  |  |     result = bytearray() | 
					
						
							|  |  |  |     # See notes at end of file for discussion of GUARD. | 
					
						
							|  |  |  |     assert GUARD > 0 # if 0, `decimal` can blow up - .prec 0 not allowed | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     def inner(n, w): | 
					
						
							|  |  |  |         #assert n < D256 ** w # required, but too expensive to check | 
					
						
							|  |  |  |         if w <= BYTELIM: | 
					
						
							|  |  |  |             # XXX Stefan Pochmann discovered that, for 1024-bit ints, | 
					
						
							|  |  |  |             # `int(Decimal)` took 2.5x longer than `int(str(Decimal))`. | 
					
						
							|  |  |  |             # Worse, `int(Decimal) is still quadratic-time for much | 
					
						
							|  |  |  |             # larger ints. So unless/until all that is repaired, the | 
					
						
							|  |  |  |             # seemingly redundant `str(Decimal)` is crucial to speed. | 
					
						
							|  |  |  |             result.extend(int(str(n)).to_bytes(w)) # big-endian default | 
					
						
							|  |  |  |             return | 
					
						
							|  |  |  |         w1 = w >> 1 | 
					
						
							|  |  |  |         w2 = w - w1 | 
					
						
							|  |  |  |         if 0: | 
					
						
							|  |  |  |             # This is maximally clear, but "too slow". `decimal` | 
					
						
							|  |  |  |             # division is asymptotically fast, but we have no way to | 
					
						
							|  |  |  |             # tell it to reuse the high-precision reciprocal it computes | 
					
						
							|  |  |  |             # for pow256[w2], so it has to recompute it over & over & | 
					
						
							|  |  |  |             # over again :-( | 
					
						
							|  |  |  |             hi, lo = divmod(n, pow256[w2][0]) | 
					
						
							|  |  |  |         else: | 
					
						
							|  |  |  |             p256, recip = pow256[w2] | 
					
						
							|  |  |  |             # The integer part will have a number of digits about equal | 
					
						
							|  |  |  |             # to the difference between the log10s of `n` and `pow256` | 
					
						
							|  |  |  |             # (which, since these are integers, is roughly approximated | 
					
						
							|  |  |  |             # by `.adjusted()`). That's the working precision we need, | 
					
						
							|  |  |  |             ctx.prec = max(n.adjusted() - p256.adjusted(), 0) + GUARD | 
					
						
							|  |  |  |             hi = +n * +recip # unary `+` chops back to ctx.prec digits | 
					
						
							|  |  |  |             ctx.prec = decimal.MAX_PREC | 
					
						
							|  |  |  |             hi = hi.to_integral_value() # lose the fractional digits | 
					
						
							|  |  |  |             lo = n - hi * p256 | 
					
						
							|  |  |  |             # Because we've been uniformly rounding down, `hi` is a | 
					
						
							|  |  |  |             # lower bound on the correct quotient. | 
					
						
							|  |  |  |             assert lo >= 0 | 
					
						
							|  |  |  |             # Adjust quotient up if needed. It usually isn't. In random | 
					
						
							|  |  |  |             # testing on inputs through 5 billion digit strings, the | 
					
						
							|  |  |  |             # test triggered once in about 200 thousand tries. | 
					
						
							|  |  |  |             count = 0 | 
					
						
							|  |  |  |             if lo >= p256: | 
					
						
							|  |  |  |                 count = 1 | 
					
						
							|  |  |  |                 lo -= p256 | 
					
						
							|  |  |  |                 hi += 1 | 
					
						
							|  |  |  |                 if lo >= p256: | 
					
						
							|  |  |  |                     # Complete correction via an exact computation. I | 
					
						
							|  |  |  |                     # believe it's not possible to get here provided | 
					
						
							|  |  |  |                     # GUARD >= 3. It's tested by reducing GUARD below | 
					
						
							|  |  |  |                     # that. | 
					
						
							|  |  |  |                     count = 999 | 
					
						
							|  |  |  |                     hi2, lo = divmod(lo, p256) | 
					
						
							|  |  |  |                     hi += hi2 | 
					
						
							|  |  |  |             _spread[count] += 1 | 
					
						
							|  |  |  |             # The assert should always succeed, but way too slow to keep | 
					
						
							|  |  |  |             # enabled. | 
					
						
							|  |  |  |             #assert hi, lo == divmod(n, pow256[w2][0]) | 
					
						
							|  |  |  |         inner(hi, w1) | 
					
						
							|  |  |  |         del hi # at top levels, can free a lot of RAM "early" | 
					
						
							|  |  |  |         inner(lo, w2) | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     # How many base 256 digits are needed?. Mathematically, exactly | 
					
						
							|  |  |  |     # floor(log256(int(s))) + 1. There is no cheap way to compute this. | 
					
						
							|  |  |  |     # But we can get an upper bound, and that's necessary for our error | 
					
						
							|  |  |  |     # analysis to make sense. int(s) < 10**len(s), so the log needed is | 
					
						
							|  |  |  |     # < log256(10**len(s)) = len(s) * log256(10). However, using | 
					
						
							|  |  |  |     # finite-precision floating point for this, it's possible that the | 
					
						
							|  |  |  |     # computed value is a little less than the true value. If the true | 
					
						
							|  |  |  |     # value is at - or a little higher than - an integer, we can get an | 
					
						
							|  |  |  |     # off-by-1 error too low. So we add 2 instead of 1 if chopping lost | 
					
						
							|  |  |  |     # a fraction > 0.9. | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2024-08-12 12:16:41 +08:00
										 |  |  |     # The "WASI" test platform can complain about `len(s)` if it's too | 
					
						
							| 
									
										
										
										
											2024-05-18 19:19:57 -05:00
										 |  |  |     # large to fit in its idea of "an index-sized integer". | 
					
						
							|  |  |  |     lenS = s.__len__() | 
					
						
							|  |  |  |     log_ub = lenS * _LOG_10_BASE_256 | 
					
						
							|  |  |  |     log_ub_as_int = int(log_ub) | 
					
						
							|  |  |  |     w = log_ub_as_int + 1 + (log_ub - log_ub_as_int > 0.9) | 
					
						
							|  |  |  |     # And what if we've plain exhausted the limits of HW floats? We | 
					
						
							|  |  |  |     # could compute the log to any desired precision using `decimal`, | 
					
						
							|  |  |  |     # but it's not plausible that anyone will pass a string requiring | 
					
						
							|  |  |  |     # trillions of bytes (unless they're just trying to "break things"). | 
					
						
							|  |  |  |     if w.bit_length() >= 46: | 
					
						
							|  |  |  |         # "Only" had < 53 - 46 = 7 bits to spare in IEEE-754 double. | 
					
						
							|  |  |  |         raise ValueError(f"cannot convert string of len {lenS} to int") | 
					
						
							|  |  |  |     with decimal.localcontext(_unbounded_dec_context) as ctx: | 
					
						
							|  |  |  |         D256 = D(256) | 
					
						
							|  |  |  |         pow256 = compute_powers(w, D256, BYTELIM, need_hi=True) | 
					
						
							|  |  |  |         rpow256 = compute_powers(w, 1 / D256, BYTELIM, need_hi=True) | 
					
						
							|  |  |  |         # We're going to do inexact, chopped arithmetic, multiplying by | 
					
						
							|  |  |  |         # an approximation to the reciprocal of 256**i. We chop to get a | 
					
						
							|  |  |  |         # lower bound on the true integer quotient. Our approximation is | 
					
						
							|  |  |  |         # a lower bound, the multiplication is chopped too, and | 
					
						
							|  |  |  |         # to_integral_value() is also chopped. | 
					
						
							|  |  |  |         ctx.traps[decimal.Inexact] = 0 | 
					
						
							|  |  |  |         ctx.rounding = decimal.ROUND_DOWN | 
					
						
							|  |  |  |         for k, v in pow256.items(): | 
					
						
							|  |  |  |             # No need to save much more precision in the reciprocal than | 
					
						
							|  |  |  |             # the power of 256 has, plus some guard digits to absorb | 
					
						
							|  |  |  |             # most relevant rounding errors. This is highly significant: | 
					
						
							|  |  |  |             # 1/2**i has the same number of significant decimal digits | 
					
						
							|  |  |  |             # as 5**i, generally over twice the number in 2**i, | 
					
						
							|  |  |  |             ctx.prec = v.adjusted() + GUARD + 1 | 
					
						
							|  |  |  |             # The unary "+" chops the reciprocal back to that precision. | 
					
						
							|  |  |  |             pow256[k] = v, +rpow256[k] | 
					
						
							|  |  |  |         del rpow256 # exact reciprocals no longer needed | 
					
						
							|  |  |  |         ctx.prec = decimal.MAX_PREC | 
					
						
							|  |  |  |         inner(D(s), w) | 
					
						
							|  |  |  |     return int.from_bytes(result) | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2022-10-25 22:00:50 -07:00
										 |  |  | def int_from_string(s): | 
					
						
							|  |  |  |     """Asymptotically fast version of PyLong_FromString(), conversion
 | 
					
						
							|  |  |  |     of a string of decimal digits into an 'int'."""
 | 
					
						
							|  |  |  |     # PyLong_FromString() has already removed leading +/-, checked for invalid | 
					
						
							|  |  |  |     # use of underscore characters, checked that string consists of only digits | 
					
						
							|  |  |  |     # and underscores, and stripped leading whitespace.  The input can still | 
					
						
							|  |  |  |     # contain underscores and have trailing whitespace. | 
					
						
							|  |  |  |     s = s.rstrip().replace('_', '') | 
					
						
							| 
									
										
										
										
											2024-05-18 19:19:57 -05:00
										 |  |  |     func = _str_to_int_inner | 
					
						
							|  |  |  |     if len(s) >= 2_000_000 and _decimal is not None: | 
					
						
							|  |  |  |         func = _dec_str_to_int_inner | 
					
						
							|  |  |  |     return func(s) | 
					
						
							| 
									
										
										
										
											2022-10-25 22:00:50 -07:00
										 |  |  | 
 | 
					
						
							|  |  |  | def str_to_int(s): | 
					
						
							|  |  |  |     """Asymptotically fast version of decimal string to 'int' conversion.""" | 
					
						
							|  |  |  |     # FIXME: this doesn't support the full syntax that int() supports. | 
					
						
							|  |  |  |     m = re.match(r'\s*([+-]?)([0-9_]+)\s*', s) | 
					
						
							|  |  |  |     if not m: | 
					
						
							|  |  |  |         raise ValueError('invalid literal for int() with base 10') | 
					
						
							|  |  |  |     v = int_from_string(m.group(2)) | 
					
						
							|  |  |  |     if m.group(1) == '-': | 
					
						
							|  |  |  |         v = -v | 
					
						
							|  |  |  |     return v | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | # Fast integer division, based on code from Mark Dickinson, fast_div.py | 
					
						
							|  |  |  | # GH-47701. Additional refinements and optimizations by Bjorn Martinsson.  The | 
					
						
							|  |  |  | # algorithm is due to Burnikel and Ziegler, in their paper "Fast Recursive | 
					
						
							|  |  |  | # Division". | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | _DIV_LIMIT = 4000 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | def _div2n1n(a, b, n): | 
					
						
							|  |  |  |     """Divide a 2n-bit nonnegative integer a by an n-bit positive integer
 | 
					
						
							|  |  |  |     b, using a recursive divide-and-conquer algorithm. | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     Inputs: | 
					
						
							|  |  |  |       n is a positive integer | 
					
						
							|  |  |  |       b is a positive integer with exactly n bits | 
					
						
							|  |  |  |       a is a nonnegative integer such that a < 2**n * b | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     Output: | 
					
						
							|  |  |  |       (q, r) such that a = b*q+r and 0 <= r < b. | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     """
 | 
					
						
							|  |  |  |     if a.bit_length() - n <= _DIV_LIMIT: | 
					
						
							|  |  |  |         return divmod(a, b) | 
					
						
							|  |  |  |     pad = n & 1 | 
					
						
							|  |  |  |     if pad: | 
					
						
							|  |  |  |         a <<= 1 | 
					
						
							|  |  |  |         b <<= 1 | 
					
						
							|  |  |  |         n += 1 | 
					
						
							|  |  |  |     half_n = n >> 1 | 
					
						
							|  |  |  |     mask = (1 << half_n) - 1 | 
					
						
							|  |  |  |     b1, b2 = b >> half_n, b & mask | 
					
						
							|  |  |  |     q1, r = _div3n2n(a >> n, (a >> half_n) & mask, b, b1, b2, half_n) | 
					
						
							|  |  |  |     q2, r = _div3n2n(r, a & mask, b, b1, b2, half_n) | 
					
						
							|  |  |  |     if pad: | 
					
						
							|  |  |  |         r >>= 1 | 
					
						
							|  |  |  |     return q1 << half_n | q2, r | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | def _div3n2n(a12, a3, b, b1, b2, n): | 
					
						
							|  |  |  |     """Helper function for _div2n1n; not intended to be called directly.""" | 
					
						
							|  |  |  |     if a12 >> n == b1: | 
					
						
							|  |  |  |         q, r = (1 << n) - 1, a12 - (b1 << n) + b1 | 
					
						
							|  |  |  |     else: | 
					
						
							|  |  |  |         q, r = _div2n1n(a12, b1, n) | 
					
						
							|  |  |  |     r = (r << n | a3) - q * b2 | 
					
						
							|  |  |  |     while r < 0: | 
					
						
							|  |  |  |         q -= 1 | 
					
						
							|  |  |  |         r += b | 
					
						
							|  |  |  |     return q, r | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | def _int2digits(a, n): | 
					
						
							|  |  |  |     """Decompose non-negative int a into base 2**n
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     Input: | 
					
						
							|  |  |  |       a is a non-negative integer | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     Output: | 
					
						
							|  |  |  |       List of the digits of a in base 2**n in little-endian order, | 
					
						
							|  |  |  |       meaning the most significant digit is last. The most | 
					
						
							|  |  |  |       significant digit is guaranteed to be non-zero. | 
					
						
							|  |  |  |       If a is 0 then the output is an empty list. | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     """
 | 
					
						
							|  |  |  |     a_digits = [0] * ((a.bit_length() + n - 1) // n) | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     def inner(x, L, R): | 
					
						
							|  |  |  |         if L + 1 == R: | 
					
						
							|  |  |  |             a_digits[L] = x | 
					
						
							|  |  |  |             return | 
					
						
							|  |  |  |         mid = (L + R) >> 1 | 
					
						
							|  |  |  |         shift = (mid - L) * n | 
					
						
							|  |  |  |         upper = x >> shift | 
					
						
							|  |  |  |         lower = x ^ (upper << shift) | 
					
						
							|  |  |  |         inner(lower, L, mid) | 
					
						
							|  |  |  |         inner(upper, mid, R) | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     if a: | 
					
						
							|  |  |  |         inner(a, 0, len(a_digits)) | 
					
						
							|  |  |  |     return a_digits | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | def _digits2int(digits, n): | 
					
						
							|  |  |  |     """Combine base-2**n digits into an int. This function is the
 | 
					
						
							|  |  |  |     inverse of `_int2digits`. For more details, see _int2digits. | 
					
						
							|  |  |  |     """
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     def inner(L, R): | 
					
						
							|  |  |  |         if L + 1 == R: | 
					
						
							|  |  |  |             return digits[L] | 
					
						
							|  |  |  |         mid = (L + R) >> 1 | 
					
						
							|  |  |  |         shift = (mid - L) * n | 
					
						
							|  |  |  |         return (inner(mid, R) << shift) + inner(L, mid) | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     return inner(0, len(digits)) if digits else 0 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | def _divmod_pos(a, b): | 
					
						
							|  |  |  |     """Divide a non-negative integer a by a positive integer b, giving
 | 
					
						
							|  |  |  |     quotient and remainder."""
 | 
					
						
							|  |  |  |     # Use grade-school algorithm in base 2**n, n = nbits(b) | 
					
						
							|  |  |  |     n = b.bit_length() | 
					
						
							|  |  |  |     a_digits = _int2digits(a, n) | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     r = 0 | 
					
						
							|  |  |  |     q_digits = [] | 
					
						
							|  |  |  |     for a_digit in reversed(a_digits): | 
					
						
							|  |  |  |         q_digit, r = _div2n1n((r << n) + a_digit, b, n) | 
					
						
							|  |  |  |         q_digits.append(q_digit) | 
					
						
							|  |  |  |     q_digits.reverse() | 
					
						
							|  |  |  |     q = _digits2int(q_digits, n) | 
					
						
							|  |  |  |     return q, r | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | def int_divmod(a, b): | 
					
						
							|  |  |  |     """Asymptotically fast replacement for divmod, for 'int'.
 | 
					
						
							|  |  |  |     Its time complexity is O(n**1.58), where n = #bits(a) + #bits(b). | 
					
						
							|  |  |  |     """
 | 
					
						
							|  |  |  |     if b == 0: | 
					
						
							| 
									
										
										
										
											2024-06-03 19:03:56 +03:00
										 |  |  |         raise ZeroDivisionError('division by zero') | 
					
						
							| 
									
										
										
										
											2022-10-25 22:00:50 -07:00
										 |  |  |     elif b < 0: | 
					
						
							|  |  |  |         q, r = int_divmod(-a, -b) | 
					
						
							|  |  |  |         return q, -r | 
					
						
							|  |  |  |     elif a < 0: | 
					
						
							|  |  |  |         q, r = int_divmod(~a, b) | 
					
						
							|  |  |  |         return ~q, b + ~r | 
					
						
							|  |  |  |     else: | 
					
						
							|  |  |  |         return _divmod_pos(a, b) | 
					
						
							| 
									
										
										
										
											2024-05-18 19:19:57 -05:00
										 |  |  | 
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | # Notes on _dec_str_to_int_inner: | 
					
						
							|  |  |  | # | 
					
						
							|  |  |  | # Stefan Pochmann worked up a str->int function that used the decimal | 
					
						
							|  |  |  | # module to, in effect, convert from base 10 to base 256. This is | 
					
						
							|  |  |  | # "unnatural", in that it requires multiplying and dividing by large | 
					
						
							|  |  |  | # powers of 2, which `decimal` isn't naturally suited to. But | 
					
						
							|  |  |  | # `decimal`'s `*` and `/` are asymptotically superior to CPython's, so | 
					
						
							|  |  |  | # at _some_ point it could be expected to win. | 
					
						
							|  |  |  | # | 
					
						
							|  |  |  | # Alas, the crossover point was too high to be of much real interest. I | 
					
						
							|  |  |  | # (Tim) then worked on ways to replace its division with multiplication | 
					
						
							|  |  |  | # by a cached reciprocal approximation instead, fixing up errors | 
					
						
							|  |  |  | # afterwards. This reduced the crossover point significantly, | 
					
						
							|  |  |  | # | 
					
						
							|  |  |  | # I revisited the code, and found ways to improve and simplify it. The | 
					
						
							|  |  |  | # crossover point is at about 3.4 million digits now. | 
					
						
							|  |  |  | # | 
					
						
							|  |  |  | # About .adjusted() | 
					
						
							|  |  |  | # ----------------- | 
					
						
							|  |  |  | # Restrict to Decimal values x > 0. We don't use negative numbers in the | 
					
						
							|  |  |  | # code, and I don't want to have to keep typing, e.g., "absolute value". | 
					
						
							|  |  |  | # | 
					
						
							|  |  |  | # For convenience, I'll use `x.a` to mean `x.adjusted()`. x.a doesn't | 
					
						
							|  |  |  | # look at the digits of x, but instead returns an integer giving x's | 
					
						
							|  |  |  | # order of magnitude. These are equivalent: | 
					
						
							|  |  |  | # | 
					
						
							|  |  |  | # - x.a is the power-of-10 exponent of x's most significant digit. | 
					
						
							|  |  |  | # - x.a = the infinitely precise floor(log10(x)) | 
					
						
							|  |  |  | # - x can be written in this form, where f is a real with 1 <= f < 10: | 
					
						
							|  |  |  | #    x = f * 10**x.a | 
					
						
							|  |  |  | # | 
					
						
							|  |  |  | # Observation; if x is an integer, len(str(x)) = x.a + 1. | 
					
						
							|  |  |  | # | 
					
						
							|  |  |  | # Lemma 1: (x * y).a = x.a + y.a, or one larger | 
					
						
							|  |  |  | # | 
					
						
							|  |  |  | # Proof: Write x = f * 10**x.a and y = g * 10**y.a, where f and g are in | 
					
						
							|  |  |  | # [1, 10). Then x*y = f*g * 10**(x.a + y.a), where 1 <= f*g < 100. If | 
					
						
							|  |  |  | # f*g < 10, (x*y).a is x.a+y.a. Else divide f*g by 10 to bring it back | 
					
						
							|  |  |  | # into [1, 10], and add 1 to the exponent to compensate. Then (x*y).a is | 
					
						
							|  |  |  | # x.a+y.a+1. | 
					
						
							|  |  |  | # | 
					
						
							|  |  |  | # Lemma 2: ceiling(log10(x/y)) <= x.a - y.a + 1 | 
					
						
							|  |  |  | # | 
					
						
							|  |  |  | # Proof: Express x and y as in Lemma 1. Then x/y = f/g * 10**(x.a - | 
					
						
							|  |  |  | # y.a), where 1/10 < f/g < 10. If 1 <= f/g, (x/y).a is x.a-y.a. Else | 
					
						
							|  |  |  | # multiply f/g by 10 to bring it back into [1, 10], and subtract 1 from | 
					
						
							|  |  |  | # the exponent to compensate. Then (x/y).a is x.a-y.a-1. So the largest | 
					
						
							|  |  |  | # (x/y).a can be is x.a-y.a. Since that's the floor of log10(x/y). the | 
					
						
							|  |  |  | # ceiling is at most 1 larger (with equality iff f/g = 1 exactly). | 
					
						
							|  |  |  | # | 
					
						
							|  |  |  | # GUARD digits | 
					
						
							|  |  |  | # ------------ | 
					
						
							|  |  |  | # We only want the integer part of divisions, so don't need to build | 
					
						
							|  |  |  | # the full multiplication tree. But using _just_ the number of | 
					
						
							|  |  |  | # digits expected in the integer part ignores too much. What's left | 
					
						
							|  |  |  | # out can have a very significant effect on the quotient. So we use | 
					
						
							|  |  |  | # GUARD additional digits. | 
					
						
							|  |  |  | # | 
					
						
							|  |  |  | # The default 8 is more than enough so no more than 1 correction step | 
					
						
							|  |  |  | # was ever needed for all inputs tried through 2.5 billion digits. In | 
					
						
							|  |  |  | # fact, I believe 3 guard digits are always enough - but the proof is | 
					
						
							|  |  |  | # very involved, so better safe than sorry. | 
					
						
							|  |  |  | # | 
					
						
							|  |  |  | # Short course: | 
					
						
							|  |  |  | # | 
					
						
							|  |  |  | # If prec is the decimal precision in effect, and we're rounding down, | 
					
						
							|  |  |  | # the result of an operation is exactly equal to the infinitely precise | 
					
						
							|  |  |  | # result times 1-e for some real e with 0 <= e < 10**(1-prec). In | 
					
						
							|  |  |  | # | 
					
						
							|  |  |  | #     ctx.prec = max(n.adjusted() - p256.adjusted(), 0) + GUARD | 
					
						
							|  |  |  | #     hi = +n * +recip # unary `+` chops to ctx.prec digits | 
					
						
							|  |  |  | # | 
					
						
							| 
									
										
										
										
											2024-08-12 12:16:41 +08:00
										 |  |  | # we have 3 visible chopped operations, but there's also a 4th: | 
					
						
							| 
									
										
										
										
											2024-05-18 19:19:57 -05:00
										 |  |  | # precomputing a truncated `recip` as part of setup. | 
					
						
							|  |  |  | # | 
					
						
							|  |  |  | # So the computed product is exactly equal to the true product times | 
					
						
							|  |  |  | # (1-e1)*(1-e2)*(1-e3)*(1-e4); since the e's are all very small, an | 
					
						
							|  |  |  | # excellent approximation to the second factor is 1-(e1+e2+e3+e4) (the | 
					
						
							|  |  |  | # 2nd and higher order terms in the expanded product are too tiny to | 
					
						
							|  |  |  | # matter). If they're all as large as possible, that's | 
					
						
							|  |  |  | # | 
					
						
							|  |  |  | # 1 - 4*10**(1-prec). This, BTW, is all bog-standard FP error analysis. | 
					
						
							|  |  |  | # | 
					
						
							|  |  |  | # That implies the computed product is within 1 of the true product | 
					
						
							|  |  |  | # provided prec >= log10(true_product) + 1.602. | 
					
						
							|  |  |  | # | 
					
						
							|  |  |  | # Here are telegraphic details, rephrasing the initial condition in | 
					
						
							|  |  |  | # equivalent ways, step by step: | 
					
						
							|  |  |  | # | 
					
						
							|  |  |  | # prod - prod * (1 - 4*10**(1-prec)) <= 1 | 
					
						
							|  |  |  | # prod - prod + prod * 4*10**(1-prec)) <= 1 | 
					
						
							|  |  |  | # prod * 4*10**(1-prec)) <= 1 | 
					
						
							|  |  |  | # 10**(log10(prod)) * 4*10**(1-prec)) <= 1 | 
					
						
							|  |  |  | # 4*10**(1-prec+log10(prod))) <= 1 | 
					
						
							|  |  |  | # 10**(1-prec+log10(prod))) <= 1/4 | 
					
						
							|  |  |  | # 1-prec+log10(prod) <= log10(1/4) = -0.602 | 
					
						
							|  |  |  | # -prec <= -1.602 - log10(prod) | 
					
						
							|  |  |  | # prec >= log10(prod) + 1.602 | 
					
						
							|  |  |  | # | 
					
						
							|  |  |  | # The true product is the same as the true ratio n/p256. By Lemma 2 | 
					
						
							|  |  |  | # above, n.a - p256.a + 1 is an upper bound on the ceiling of | 
					
						
							|  |  |  | # log10(prod). Then 2 is the ceiling of 1.602. so n.a - p256.a + 3 is an | 
					
						
							|  |  |  | # upper bound on the right hand side of the inequality. Any prec >= that | 
					
						
							|  |  |  | # will work. | 
					
						
							|  |  |  | # | 
					
						
							|  |  |  | # But since this is just a sketch of a proof ;-), the code uses the | 
					
						
							|  |  |  | # empirically tested 8 instead of 3. 5 digits more or less makes no | 
					
						
							|  |  |  | # practical difference to speed - these ints are huge. And while | 
					
						
							|  |  |  | # increasing GUARD above 3 may not be necessary, every increase cuts the | 
					
						
							|  |  |  | # percentage of cases that need a correction at all. | 
					
						
							|  |  |  | # | 
					
						
							|  |  |  | # On Computing Reciprocals | 
					
						
							|  |  |  | # ------------------------ | 
					
						
							|  |  |  | # In general, the exact reciprocals we compute have over twice as many | 
					
						
							|  |  |  | # significant digits as needed. 1/256**i has the same number of | 
					
						
							|  |  |  | # significant decimal digits as 5**i. It's a significant waste of RAM | 
					
						
							|  |  |  | # to store all those unneeded digits. | 
					
						
							|  |  |  | # | 
					
						
							|  |  |  | # So we cut exact reciprocals back to the least precision that can | 
					
						
							|  |  |  | # be needed so that the error analysis above is valid, | 
					
						
							|  |  |  | # | 
					
						
							|  |  |  | # [Note: turns out it's very significantly faster to do it this way than | 
					
						
							|  |  |  | # to compute  1 / 256**i  directly to the desired precision, because the | 
					
						
							|  |  |  | # power method doesn't require division. It's also faster than computing | 
					
						
							|  |  |  | # (1/256)**i directly to the desired precision - no material division | 
					
						
							|  |  |  | # there, but `compute_powers()` is much smarter about _how_ to compute | 
					
						
							|  |  |  | # all the powers needed than repeated applications of `**` - that | 
					
						
							|  |  |  | # function invokes `**` for at most the few smallest powers needed.] | 
					
						
							|  |  |  | # | 
					
						
							|  |  |  | # The hard part is that chopping back to a shorter width occurs | 
					
						
							|  |  |  | # _outside_ of `inner`. We can't know then what `prec` `inner()` will | 
					
						
							|  |  |  | # need. We have to pick, for each value of `w2`, the largest possible | 
					
						
							|  |  |  | # value `prec` can become when `inner()` is working on `w2`. | 
					
						
							|  |  |  | # | 
					
						
							|  |  |  | # This is the `prec` inner() uses: | 
					
						
							|  |  |  | #     max(n.a - p256.a, 0) + GUARD | 
					
						
							|  |  |  | # and what setup uses (renaming its `v` to `p256` - same thing): | 
					
						
							|  |  |  | #     p256.a + GUARD + 1 | 
					
						
							|  |  |  | # | 
					
						
							|  |  |  | # We need that the second is always at least as large as the first, | 
					
						
							|  |  |  | # which is the same as requiring | 
					
						
							|  |  |  | # | 
					
						
							|  |  |  | #     n.a - 2 * p256.a <= 1 | 
					
						
							|  |  |  | # | 
					
						
							|  |  |  | # What's the largest n can be? n < 255**w = 256**(w2 + (w - w2)). The | 
					
						
							|  |  |  | # worst case in this context is when w ix even. and then w = 2*w2, so | 
					
						
							|  |  |  | # n < 256**(2*w2) = (256**w2)**2 = p256**2. By Lemma 1, then, n.a | 
					
						
							|  |  |  | # is at most p256.a + p256.a + 1. | 
					
						
							|  |  |  | # | 
					
						
							|  |  |  | # So the most n.a - 2 * p256.a can be is | 
					
						
							|  |  |  | # p256.a + p256.a + 1 - 2 * p256.a = 1. QED | 
					
						
							|  |  |  | # | 
					
						
							|  |  |  | # Note: an earlier version of the code split on floor(e/2) instead of on | 
					
						
							|  |  |  | # the ceiling. The worst case then is odd `w`, and a more involved proof | 
					
						
							|  |  |  | # was needed to show that adding 4 (instead of 1) may be necessary. | 
					
						
							|  |  |  | # Basically because, in that case, n may be up to 256 times larger than | 
					
						
							|  |  |  | # p256**2. Curiously enough, by splitting on the ceiling instead, | 
					
						
							|  |  |  | # nothing in any proof here actually depends on the output base (256). | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | # Enable for brute-force testing of compute_powers(). This takes about a | 
					
						
							|  |  |  | # minute, because it tries millions of cases. | 
					
						
							|  |  |  | if 0: | 
					
						
							| 
									
										
										
										
											2024-08-12 12:16:41 +08:00
										 |  |  |     def consumer(w, limit, need_hi): | 
					
						
							| 
									
										
										
										
											2024-05-18 19:19:57 -05:00
										 |  |  |         seen = set() | 
					
						
							|  |  |  |         need = set() | 
					
						
							|  |  |  |         def inner(w): | 
					
						
							|  |  |  |             if w <= limit: | 
					
						
							|  |  |  |                 return | 
					
						
							|  |  |  |             if w in seen: | 
					
						
							|  |  |  |                 return | 
					
						
							|  |  |  |             seen.add(w) | 
					
						
							|  |  |  |             lo = w >> 1 | 
					
						
							|  |  |  |             hi = w - lo | 
					
						
							|  |  |  |             need.add(hi if need_hi else lo) | 
					
						
							|  |  |  |             inner(lo) | 
					
						
							|  |  |  |             inner(hi) | 
					
						
							|  |  |  |         inner(w) | 
					
						
							| 
									
										
										
										
											2024-08-12 12:16:41 +08:00
										 |  |  |         exp = compute_powers(w, 1, limit, need_hi=need_hi) | 
					
						
							| 
									
										
										
										
											2024-05-18 19:19:57 -05:00
										 |  |  |         assert exp.keys() == need | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     from itertools import chain | 
					
						
							|  |  |  |     for need_hi in (False, True): | 
					
						
							|  |  |  |         for limit in (0, 1, 10, 100, 1_000, 10_000, 100_000): | 
					
						
							|  |  |  |             for w in chain(range(1, 100_000), | 
					
						
							|  |  |  |                            (10**i for i in range(5, 30))): | 
					
						
							|  |  |  |                 consumer(w, limit, need_hi) |