-
-
Notifications
You must be signed in to change notification settings - Fork 31.9k
sum() several times slower on Python 3 64-bit #68264
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
I got a report that summing numbers is noticably slower on Python 3. This is easily reproducible: $ time python2.7 -c "print sum(xrange(3, 10**9, 3)) + sum(xrange(5, 10**9, 5)) - sum(xrange(15, 10**9, 15))"
233333333166666668 real 0m6.165s $ time python3.4 -c "print(sum(range(3, 10**9, 3)) + sum(range(5, 10**9, 5)) - sum(range(15, 10**9, 15)))"
233333333166666668 real 0m16.413s I can't tell from initial poking what's the core issue here. Both examples produce equivalent bytecode, the builtin_sum() function is only noticably different in the fact that it uses PyLong_* across the board, including PyLong_AsLongAndOverlow. We'll need to profile this, which I didn't have time for yet. |
Can't reproduce on 32-bit Linux. $ time python2.7 -c "print sum(xrange(3, 10**9, 3)) + sum(xrange(5, 10**9, 5)) - sum(xrange(15, 10**9, 15))"
233333333166666668 real 1m11.614s real 1m11.658s $ python2.7 -m timeit -n1 -r1 "sum(xrange(3, 10**9, 3)) + sum(xrange(5, 10**9, 5)) - sum(xrange(15, 10**9, 15))"
1 loops, best of 1: 72 sec per loop
$ python3.4 -m timeit -n1 -r1 "sum(range(3, 10**9, 3)) + sum(range(5, 10**9, 5)) - sum(range(15, 10**9, 15))"
1 loops, best of 1: 72.5 sec per loop
$ python2.7 -m timeit -s "a = list(range(10**6))" -- "sum(a)"
10 loops, best of 3: 114 msec per loop
$ python3.4 -m timeit -s "a = list(range(10**6))" -- "sum(a)"
10 loops, best of 3: 83.5 msec per loop What is sys.int_info on your build? |
I reproduce under 64-bit Linux. So this may be because the Python long digit (30 bits) is smaller than the C long (64 bits). Lukasz: is there a specific use case? Note you can use Numpy for such calculations. |
Serhiy, this is 64-bit specific. Antoine, as far as I can tell, the main use case is: "Don't make it look like migrating to Python 3 is a terrible performance downgrade." As we discussed on the language summit this year [1], we have to be at least not worse to look appealing. This might be a flawed benchmark but people will make them anyway. In this particular case, there's internal usage at Twitter that unearthed it. The example is just a simplified repro. Some perf degradations were expected, like switching text to Unicode. In this case, the end result computed by both 2.7 and 3.4 is the same so we should be able to address this. |
If that's due to the different representation of Python 2's int type and Python 3's int type then I don't see an easy solution to this. |
Łukasz: there are three ingredients here - sum, (x)range and the integer addition that sum will be performing at each iteration. Is there any chance you can separate the effects on your machine? On my machine (OS X, 64-bit), I'm seeing *some* speed difference in the integer arithmetic, but not enough to explain the whole of the timing mismatch. One thing we've lost in Python 3 is the fast path for small-int addition *inside* the ceval loop. It may be possible to restore something there. |
Throwing out sum, I'm seeing significant slowdown simply from xrange versus range: taniyama:Desktop mdickinson$ python2 -m timeit -s 'x = xrange(3, 10**9, 3)' 'for e in x: pass' |
... not to forget the interpreter startup time on his machine. :) I did a tiny bit of profiling and about 90% of the time seems to be spent creating and deallocating throw-away PyLong objects. My guess is that it simply lacks a free-list in _PyLong_New(). |
It seems we (like the benchmarks posted) are spending a whole lot of time on something that's probably not relevant to any real-world situation. If someone has actual code that suffers from this, it would be good to know about it. |
I don't think it's irrelevant. Throw-away integers are really not uncommon. For-loops use them quite often, non-trivial arithmetic expressions can create a lot of intermediate temporaries. Speeding up the create-delete cycle of PyLong sounds like a very obvious thing to do. Imagine some code that iterates over a list of integers, applies some calculation to them, and then stores them in a new list, maybe even using a list comprehension or so. If you could speed up the intermediate calculation by avoiding overhead in creating temporary PyLong objects, such code could benefit a lot. I suspect that adding a free-list for single-digit PyLong objects (the most common case) would provide some visible benefit. |
Le 01/05/2015 08:09, Stefan Behnel a écrit :
That may be a good thing indeed. I'm just saying that the benchmarks |
I tried implementing a freelist. Patch attached, mostly adapted from the one in dictobject.c, but certainly needs a bit of cleanup. The results are not bad, about 10-20% faster: Original: $ ./python -m timeit 'sum(range(1, 100000))'
1000 loops, best of 3: 1.86 msec per loop
$ ./python -m timeit -s 'l = list(range(1000, 10000))' '[(i*2+5) // 7 for i in l]'
1000 loops, best of 3: 1.05 msec per loop With freelist: $ ./python -m timeit 'sum(range(1, 100000))'
1000 loops, best of 3: 1.52 msec per loop
$ ./python -m timeit -s 'l = list(range(1000, 10000))' '[(i*2+5) // 7 for i in l]'
1000 loops, best of 3: 931 usec per loop |
Antoine asked:
You might have missed Łukasz' earlier comment: "In this particular case, there's internal usage at Twitter that unearthed it. The example is just a simplified repro." |
bpo-24165 was created to pursue the path of a free-list for PyLong objects. |
FWIW, a PGO build of Py3.7 is now about 20% *faster* here than my Ubuntu 16/04 system Python 2.7, and for some (probably unrelated) reason, the system Python 3.5 is another 2% faster on my side. IMHO, the only other thing that seems obvious to try would be to inline the unpacking of single digit PyLongs into sum(). I attached a simple patch that does that, in case someone wants to test it out. For non-PGO builds, it's about 17% faster for me. Didn't take the time to benchmark PGO builds with it. |
Does that mean we can close this issue? Or do I misunderstand what you are comparing? 32 vs. 64 bits? PGO vs. non-PGO? OTOH on my Mac I still find that 3.10 with PGO is still more than twice as slow than 2.7. Thinking about it that's a bit odd, since (presumably) the majority of the work in sum() involves a long int result (even though the values returned by range() all fit in 30 bits, the sum quickly exceeds that). (earlier)
If my theory is correct that wouldn't help this particular case, right? FWIW just "for i in [x]range(15, 10**9, 15): pass" is about the same speed in Python 2.7 as in 3.11. |
The actual accumulation of a long int result is still as fast as it ever was. The main difference from Py2.7 isn't the addition, it is that detecting and extracting a small int added has become expensive. -- Python 2 fastpath -------------------------------------- if (PyInt_CheckExact(item)) { // Very cheap
long b = PyInt_AS_LONG(item); // Very cheap
long x = i_result + b; // Very cheap
if ((x^i_result) >= 0 || (x^b) >= 0) { // Semi cheap
i_result = x; // Zero cost
Py_DECREF(item); // Most expensive step, but still cheap
continue;
}
} -- Python 3 fastpath -------------------------------------- if (PyLong_CheckExact(item) || PyBool_Check(item)) { // Cheap
long b = PyLong_AsLongAndOverflow(item, &overflow); // Super Expensive
if (overflow == 0 && // Branch predictable test
(i_result >= 0 ? (b <= LONG_MAX - i_result) // Slower but better test
: (b >= LONG_MIN - i_result)))
{
i_result += b; // Very cheap
Py_DECREF(item);
continue;
}
} -- Supporting function ------------------------------------ long
PyLong_AsLongAndOverflow(PyObject *vv, int *overflow) // OMG, this does a lot of work
{
/* This version by Tim Peters */
PyLongObject *v;
unsigned long x, prev;
long res;
Py_ssize_t i;
int sign;
int do_decref = 0; /* if PyNumber_Index was called */
*overflow = 0;
if (vv == NULL) {
PyErr_BadInternalCall();
return -1;
}
if (PyLong_Check(vv)) {
v = (PyLongObject *)vv;
}
else {
v = (PyLongObject *)_PyNumber_Index(vv);
if (v == NULL)
return -1;
do_decref = 1;
}
switch (i) {
case -1:
res = -(sdigit)v->ob_digit[0];
break;
case 0:
res = 0;
break;
case 1:
res = v->ob_digit[0];
break;
default:
sign = 1;
x = 0;
if (i < 0) {
sign = -1;
i = -(i);
}
while (--i >= 0) {
prev = x;
x = (x << PyLong_SHIFT) | v->ob_digit[i];
if ((x >> PyLong_SHIFT) != prev) {
*overflow = sign;
goto exit;
}
}
/* Haven't lost any bits, but casting to long requires extra
* care (see comment above).
*/
if (x <= (unsigned long)LONG_MAX) {
res = (long)x * sign;
}
else if (sign < 0 && x == PY_ABS_LONG_MIN) {
res = LONG_MIN;
}
else {
*overflow = sign;
/* res is already set to -1 */
}
}
exit:
if (do_decref) {
Py_DECREF(v);
}
return res;
} |
I created a PR from my last patch, inlining the unpacking of single digit integers. Since most integers should fit into a single digit these days, this is as fast a path as it gets. |
Thanks, that gets to the heart of the issue. I marked the PR as approved (though there is a small coding nit you may want to fix). |
The patch looks fine, but it looks a bit like benchmark chasing. Is the speed of builtin sum() of a sequence of integers important enough to do this bit of inlining? (It may break if we change the internals of Py_Long, as Mark Shannon has been wanting to do for a while -- see faster-cpython/ideas#42.) |
Given that we already accepted essentially separate loops for the int, float and everything else cases, I think the answer is that it doesn't add much to the triplication.
I would assume that such a structural change would come with suitable macros to unpack the special 0-2 digit integers. Those would then apply here, too. As it stands, there are already some modules distributed over the source tree that use direct digit access: ceval.c, _decimal.c, marshal.c. They are easy to find with grep and my PR just adds one more. |
Sounds good, you have my blessing. |
What are microbenchmark results for PR 28469 in comparison with the baseline? |
Old, with PGO: New, with PGO: The results are a bit more mixed with PGO optimisation (I tried a couple of times), not sure why. Might just be normal fluctuation, bad benchmark value selection, or accidental PGO tuning, can't say. In any case, the 1-digit case (10000, 2**29) is again about 28% faster and none of the other cases seems (visibly) slower. I think this is a very clear net-win. |
Thank you again Stefan. Now no doubts are left. BTW, pyperf gives more stable results. I use it if have any doubts (either the results of timeit are not stable or the difference is less than say 10%). |
Unfortunately commit debd804 introduced a reference leak: ❯ ./python -m test test_grammar -R : == Tests result: FAILURE == 1 test failed: Total duration: 1.1 sec debd804 is the first bad commit
.../Core and Builtins/2021-09-20-10-02-12.bpo-24076.ZFgFSj.rst | 1 + |
Opened bpo-28493 to fix the refleak |
Sorry, I meant PR 28493 |
Sorry for that, Pablo. I knew exactly where the problem was, the second I read your notification. Thank you for resolving it so quickly. |
Always happy to help :) |
I'm still seeing a large slowdown (comparing the Ubuntu 2.7 and 3.11) $ time python2.7 -c "print sum(xrange(3, 10**9, 3)) + sum(xrange(5, 10**9, 5)) - sum(xrange(15, 10**9, 15))"
233333333166666668
real 0m2.907s
user 0m2.877s
sys 0m0.025s
$ time python3.11 -c "print(sum(range(3, 10**9, 3)) + sum(range(5, 10**9, 5)) - sum(range(15, 10**9, 15)))"
233333333166666668
real 0m6.915s
user 0m6.898s
sys 0m0.009s |
I think the fix is faster-cpython/ideas#147, but we haven't started implementing that. |
I'm reopening this as |
3.11 D:\python311>python
Python 3.11.0 (main, Oct 24 2022, 18:26:48) [MSC v.1933 64 bit (AMD64)] on win32
>>> import time
>>> t=time.time();sum(range(1,pow(10,8)+1));print(time.time()-t)
5000000050000000
4.157237768173218 vs D:\python310>python
Python 3.10.6 (tags/v3.10.6:9c7b4bd, Aug 1 2022, 21:53:49) [MSC v.1932 64 bit (AMD64)] on win32
>>> import time
>>> t=time.time();sum(range(1,pow(10,8)+1));print(time.time()-t)
5000000050000000
4.183239221572876 |
pypy 7.3.9 D:\pypy3.8-v7.3.9-win64>pypy
Python 3.8.12 (0089b4a7ab2306925a251b35912885d52ead1aba, Mar 16 2022, 13:51:04)
[PyPy 7.3.9 with MSC v.1929 64 bit (AMD64)] on win32
Type "help", "copyright", "credits" or "license" for more information.
>>>> import time
>>>> t=time.time();sum(range(1,pow(10,8)+1));print(time.time()-t)
5000000050000000
0.1780109405517578 |
This can be closed. On 3.14:
Much faster than @ambv benchmarks.
Discussion of further optimization belongs elsewhere (faster-cpython). |
I'm not sure. Absolute numbers in benchmarks aren't relevant here. OP probably run py2 tests on a different system than you. At least, you should run test for py2 and py3 on same system. Here are my tests. Py2:
Py3.13:
Py3.14a7:
For me it looks like issue is valid and it's a regression from py2, not a feature request. @picnixz ? @ambv ? |
Honestly, considering 2.7 has been EOL for a long time, I don't think we need to keep this specific issue open. I don't think we can do anything now and I would indeed prefer if faster-cpython comes up with a solution. Now, for performance loss, we sometimes treat them as bug, sometimes not. I categorized it as a FR because we won't backport the change I think. |
Why not? If v2.7 is better somewhere - it's still a regression.
I'm not sure that closing issue is a right thing even if we can't do anything now. Work is ongoing, e.g.: https://discuss.python.org/t/87950. I think it's a good thing to keep eye on this issue for people involved. BTW, Py3.14 has impressive speedup, but I doubt it's related to integer arithmetic. |
I'm with @skirpichev. Can someone at least summarize an explanation of the difference between 2.7 and 3.x? |
I think it was analyzed by #68264 (comment) but AFAICT, it's the call to
I wasn't aware of this one so thanks. By the way, feel free to re-open issues if I close them wrongly! |
My 2c:
Here more tests from my zoo:
|
Note: these values reflect the state of the issue at the time it was migrated and might not reflect the current state.
Show more details
GitHub fields:
bugs.python.org fields:
The text was updated successfully, but these errors were encountered: