Re: Python Windows Extensions for Mac

2011-08-21 Thread Chris Angelico
On Sun, Aug 21, 2011 at 6:38 AM, Johnny Venter  wrote:
> Yes, I want to make my queries from a remote non-Windows computer. Here is 
> the scenario:
>
> From my mac, I want to use python to access and read objects from a remote  
> Windows computer joined to a Windows 2003 functional level domain. Given 
> this, what is the best way to accomplish this?
>

Then the "use Python" part is relatively immaterial; what you need to
know is: What network protocol are you using to "access and read
objects"? Start by researching that; once you know the details (is it
even TCP/IP-based?), you can look into whether Python has facilities
for speaking that protocol.

ChrisA
-- 
http://mail.python.org/mailman/listinfo/python-list


Installing numpy on 2.7 (OS X 10.7)

2011-08-21 Thread jefflovejapan
I'm following the instructions given http://www.scipy.org/
Installing_SciPy/Mac_OS_X">here, but it isn't working.
Specifically, I'm getting:

Library/Frameworks/Python.framework/Versions/2.7/Resources/Python.app/
Contents/MacOS/Python: can't open file 'setup.py': [Errno 2] No such
file or directory

even though I completed the two downloads from github successfully.
Any ideas why this wouldn't be working, or any ideas where I can look
for better instructions?
-- 
http://mail.python.org/mailman/listinfo/python-list


Re: Python Windows Extensions for Mac

2011-08-21 Thread Alec Taylor
Perhaps you'd be better off with something like RunDeck (Free,
Open-Source, Cross-Platform, CopyLeft) for this kind of problem.

On Sun, Aug 21, 2011 at 5:30 PM, Chris Angelico  wrote:
> On Sun, Aug 21, 2011 at 6:38 AM, Johnny Venter  wrote:
>> Yes, I want to make my queries from a remote non-Windows computer. Here is 
>> the scenario:
>>
>> From my mac, I want to use python to access and read objects from a remote  
>> Windows computer joined to a Windows 2003 functional level domain. Given 
>> this, what is the best way to accomplish this?
>>
>
> Then the "use Python" part is relatively immaterial; what you need to
> know is: What network protocol are you using to "access and read
> objects"? Start by researching that; once you know the details (is it
> even TCP/IP-based?), you can look into whether Python has facilities
> for speaking that protocol.
>
> ChrisA
> --
> http://mail.python.org/mailman/listinfo/python-list
>
-- 
http://mail.python.org/mailman/listinfo/python-list


Re: Installing numpy on 2.7 (OS X 10.7)

2011-08-21 Thread Benjamin Kaplan
On Sun, Aug 21, 2011 at 6:03 AM, jefflovejapan  wrote:
> I'm following the instructions given http://www.scipy.org/
> Installing_SciPy/Mac_OS_X">here, but it isn't working.
> Specifically, I'm getting:
>
> Library/Frameworks/Python.framework/Versions/2.7/Resources/Python.app/
> Contents/MacOS/Python: can't open file 'setup.py': [Errno 2] No such
> file or directory
>
> even though I completed the two downloads from github successfully.
> Any ideas why this wouldn't be working, or any ideas where I can look
> for better instructions?
> --

As the error says, it can't find the setup.py file. My guess would be
you forgot to cd to the newly created directory. You have to "cd
numpy" (or cd scipy if you're trying to install that) first to get
into the directory with the install script.
-- 
http://mail.python.org/mailman/listinfo/python-list


Re: Python Windows Extensions for Mac

2011-08-21 Thread Irmen de Jong
On 21-8-2011 1:51, Johnny Venter wrote:
> Thank you all for the replies. I would like to query various Windows' objects 
> and
> resources from Mac and/or Linux such as Active Directory users, network 
> shares, group
> members, etc... What module or methods can I use with python to accomplish 
> this?

A way to approach this problem is installing Python + windows extensions on the 
actual
windows machine(s) you want to query.

Then create some form of a server process that does the windows specific stuff 
locally,
and exposes an interface with the functions you want to provide to your remote
machine(s).  Talk with the server process using some form of platform 
independent IPC,
for instance Pyro: http://pypi.python.org/pypi/Pyro4/

Be careful what methods you expose this way though (security issues!)

Irmen
-- 
http://mail.python.org/mailman/listinfo/python-list


Re: decorator issue with modules dbus & gobject

2011-08-21 Thread Emile van Sebille

On 8/18/2011 5:02 AM Makiavelik said...

Hi,
Here is a sample code that reproduces the issue :


Not really 'sample' enough to allow others to investigate...

ImportError: No module named gobject
ImportError: No module named dbus
ImportError: No module named dbus.mainloop.glib

Try to eliminate the non-standard dependencies.  I wanted to check the 
default value set in timeout but there's too much crap to clean up.


Emile

--
http://mail.python.org/mailman/listinfo/python-list


relative speed of incremention syntaxes (or "i=i+1" VS "i+=1")

2011-08-21 Thread Laurent
Hi Folks,

I was arguing with a guy who was sure that incrementing a variable i with "i += 
1" is faster than "i = i + 1". I couldn't tell if he was right or wrong so I 
did a little benchmark with the very useful timeit module.
Here are the results on my little Linux Eeepc Netbook (using Python 3.2):


Computing, please wait...

Results for 100 times "i = i + 1":
0.37591004371643066
0.3827171325683594
0.37238597869873047
0.37305116653442383
0.3725881576538086
0.37294602394104004
0.3712761402130127
0.37357497215270996
0.371567964553833
0.37359118461608887
Total 3.7396 seconds.

Results for 100 times "i += 1":
0.3821070194244385
0.3802030086517334
0.3828878402709961
0.3823058605194092
0.3801591396331787
0.38340115547180176
0.3795340061187744
0.38153910636901855
0.3835160732269287
0.381864070892334
Total 3.8175 seconds.

==> "i = i + 1" is 2.08% faster than "i += 1".



I did many tests and "i = i + 1" always seems to be around 2% faster than "i += 
1". This is no surprise as the += notation seems to be a syntaxic sugar layer 
that has to be converted to i = i + 1 anyway. Am I wrong in my interpretation?

Btw here's the trivial Python 3.2 script I made for this benchmark:


import timeit

r = 10
n = 100

s1 = "i = i + 1"
s2 = "i += 1"

t1 = timeit.Timer(stmt=s1, setup="i = 0")
t2 = timeit.Timer(stmt=s2, setup="i = 0")

print("Computing, please wait...")

results1 = t1.repeat(repeat=r, number=n)
results2 = t2.repeat(repeat=r, number=n)

print('\nResults for {} times "{}":'.format(n, s1))
sum1 = 0
for result in results1:
print(result)
sum1 += result
print("Total {:.5} seconds.".format(sum1))

print('\nResults for {} times "{}":'.format(n, s2))
sum2 = 0
for result in results2:
print(result)
sum2 += result
print("Total {:.5} seconds.".format(sum2))

print('\n==> "{}" is {:.3}% faster than "{}".'.format(s1,(sum2 / sum1) * 100 - 
100, s2))



Comments are welcome...

-- 
http://mail.python.org/mailman/listinfo/python-list


Re: relative speed of incremention syntaxes (or "i=i+1" VS "i+=1")

2011-08-21 Thread woooee
as the += notation seems to be a syntaxic sugar layer that has to be
converted to i = i + 1 anyway.

That has always been my understanding.  The faster way is to append to
a list as concatenating usually, requires the original string,
accessing an intermediate block of memory, and the memory for the
final string.
x_list.append(value)
to_string = "".join(x_list)
-- 
http://mail.python.org/mailman/listinfo/python-list


Re: relative speed of incremention syntaxes (or "i=i+1" VS "i+=1")

2011-08-21 Thread Laurent
Well I agree with you about string concatenation, but here I'm talking about 
integers incrementation...
-- 
http://mail.python.org/mailman/listinfo/python-list


Re: relative speed of incremention syntaxes (or "i=i+1" VS "i+=1")

2011-08-21 Thread Irmen de Jong

On 21-08-11 19:03, Laurent wrote:

Well I agree with you about string concatenation, but here I'm talking about 
integers incrementation...


Seems the two forms are not 100% identical:

>>> import dis
>>> def f1(x):
...  x=x+1
...
>>> def f2(x):
...  x+=1
...
>>>
>>> dis.dis(f1)
  2   0 LOAD_FAST0 (x)
  3 LOAD_CONST   1 (1)
  6 BINARY_ADD
  7 STORE_FAST   0 (x)
 10 LOAD_CONST   0 (None)
 13 RETURN_VALUE
>>> dis.dis(f2)
  2   0 LOAD_FAST0 (x)
  3 LOAD_CONST   1 (1)
  6 INPLACE_ADD
  7 STORE_FAST   0 (x)
 10 LOAD_CONST   0 (None)
 13 RETURN_VALUE
>>>


What the precise difference (semantics and speed) is between the 
BINARY_ADD and INPLACE_ADD opcodes, I dunno. Look in the Python source 
code or maybe someone knows it from memory :-)


Irmen

--
http://mail.python.org/mailman/listinfo/python-list


Re: extended slicing and negative stop value problem

2011-08-21 Thread Max
On Aug 20, 1:40 pm, Steven D'Aprano  wrote:
> Pardon me for breaking threading, but I don't have Max's original post.

Not sure why; I also can't see it! I'll copy it at the end just in
case.

> On Sat, Aug 20, 2011 at 7:20 PM, Max Moroz  wrote:
> > Would it be a good idea to change Python definition so that a[10, -1, -1]
>
> I presume you mean slice notation a[10:-1:-1].
>
> > referred to the elements starting with position 10, going down to the
> > beginning?
>
> If this was going to be "fixed", the time was probably
> about three years ago, when Python3 was just starting. Now such a change
> will probably need to wait for the hypothetical Python 4000.

Yeah, I was surprised that it didn't bother anyone..

> The usual advice is to do your slicing twice, reversing it the second time:
>
> a[0:11][::-1]
> # Instead of a[10:-1:-1], which looks like it should work, but doesn't.

It works nicely, but it is 1.3 times slower in my code (I am surprised
the interpreter doesn't optimize this).

-- 
http://mail.python.org/mailman/listinfo/python-list


Re: relative speed of incremention syntaxes (or "i=i+1" VS "i+=1")

2011-08-21 Thread Andreas Löscher
> What the precise difference (semantics and speed) is between the 
> BINARY_ADD and INPLACE_ADD opcodes, I dunno. Look in the Python source 
> code or maybe someone knows it from memory :-)
> 
> Irmen
> 
from Python/ceval.c:

1316case BINARY_ADD:
1317w = POP();
1318v = TOP();
1319if (PyInt_CheckExact(v) && PyInt_CheckExact(w)) {
1320/* INLINE: int + int */
1321register long a, b, i;
1322a = PyInt_AS_LONG(v);
1323b = PyInt_AS_LONG(w);
1324/* cast to avoid undefined behaviour
1325   on overflow */
1326i = (long)((unsigned long)a + b);
1327if ((i^a) < 0 && (i^b) < 0)
1328goto slow_add;
1329x = PyInt_FromLong(i);
1330}
1331else if (PyString_CheckExact(v) &&
1332 PyString_CheckExact(w)) {
1333x = string_concatenate(v, w, f, next_instr);
1334/* string_concatenate consumed the ref to v */
1335goto skip_decref_vx;
1336}
1337else {
1338  slow_add:
1339x = PyNumber_Add(v, w);
1340}
1341Py_DECREF(v);
1342  skip_decref_vx:
1343Py_DECREF(w);
1344SET_TOP(x);
1345if (x != NULL) continue;
1346break;

1532case INPLACE_ADD:
1533w = POP();
1534v = TOP();
1535if (PyInt_CheckExact(v) && PyInt_CheckExact(w)) {
1536/* INLINE: int + int */
1537register long a, b, i;
1538a = PyInt_AS_LONG(v);
1539b = PyInt_AS_LONG(w);
1540i = a + b;
1541if ((i^a) < 0 && (i^b) < 0)
1542goto slow_iadd;
1543x = PyInt_FromLong(i);
1544}
1545else if (PyString_CheckExact(v) &&
1546 PyString_CheckExact(w)) {
1547x = string_concatenate(v, w, f, next_instr);
1548/* string_concatenate consumed the ref to v */
1549goto skip_decref_v;
1550}
1551else {
1552  slow_iadd:
1553x = PyNumber_InPlaceAdd(v, w);
1554}
1555Py_DECREF(v);
1556  skip_decref_v:
1557Py_DECREF(w);
1558SET_TOP(x);
1559if (x != NULL) continue;
1560break;

As for using Integers, the first case (line 1319 and 1535) are true and
there is no difference in Code. However, Python uses a huge switch-case
construct to execute it's opcodes and INPLACE_ADD cames after
BINARY_ADD, hence the difference in speed. 

To be clear, this is nothing you should consider when writing fast code.
Complexity wise they both are the same. 



-- 
http://mail.python.org/mailman/listinfo/python-list


Re: relative speed of incremention syntaxes (or "i=i+1" VS "i+=1")

2011-08-21 Thread Laurent
Thanks for these explanations. So 2% speed difference just between "B..." and 
"I..." entries in a huge alphabetically-ordered switch case? Wow. Maybe there 
is some material for speed enhancement here... 
-- 
http://mail.python.org/mailman/listinfo/python-list


Re: relative speed of incremention syntaxes (or "i=i+1" VS "i+=1")

2011-08-21 Thread Hans Mulder

On 21/08/11 19:14:19, Irmen de Jong wrote:


What the precise difference (semantics and speed) is between the
BINARY_ADD and INPLACE_ADD opcodes, I dunno. Look in the Python source
code or maybe someone knows it from memory :-)


There is a clear difference in semantics: BINARY_ADD always produces
a new object, INPLACE_ADD may modify its left-hand operand in situ
(if it's mutable).

Integers are immutable, so for integers the semantics are the same,
but for lists, for example, the two are different:

>>> x = [2, 3, 5, 7]
>>> y = [11, 13]
>>> x+y
[2, 3, 5, 7, 11, 13]
>>> x
[2, 3, 5, 7]# x still has its original value
>>> x += y
>>> x
[2, 3, 5, 7, 11, 13]# x is now modified
>>>

For integers, I would not expect a measurable difference in speed.

Hope this helps,

-- HansM

--
http://mail.python.org/mailman/listinfo/python-list


Re: relative speed of incremention syntaxes (or "i=i+1" VS "i+=1")

2011-08-21 Thread Laurent
Well 2% more time after 1 million iterations so you're right I won't consider 
it.
-- 
http://mail.python.org/mailman/listinfo/python-list


Re: relative speed of incremention syntaxes (or "i=i+1" VS "i+=1")

2011-08-21 Thread Christian Heimes
Am 21.08.2011 19:27, schrieb Andreas Löscher:
> As for using Integers, the first case (line 1319 and 1535) are true and
> there is no difference in Code. However, Python uses a huge switch-case
> construct to execute it's opcodes and INPLACE_ADD cames after
> BINARY_ADD, hence the difference in speed. 

I don't think that's the reason. Modern compiles turn a switch statement
into a jump or branch table rather than a linear search like chained
elif statements. Python 3.2 on Linux should also be compiled with
computed gotos.

Christian

-- 
http://mail.python.org/mailman/listinfo/python-list


Re: try... except with unknown error types

2011-08-21 Thread Paul Rubin
Steven D'Aprano  writes:
>> But there's no way to know what that minimum is.  Python libraries throw
>> all sorts of exceptions that their documentation doesn't mention.
>
> Yes, you're absolutely correct. But it's also irrelevant. Most of those
> exceptions should not be caught, even if you know what they are, because
> they represent either bugs that should be fixed, or bad data which should
> raise an exception. A bare except, or except Exception, is hardly ever the
> right approach.

I'm not sure what to do instead.  The exceptions I'm currently dealing
with happen when certain network operations go wrong (e.g. network or
remote host is down, connection fails, etc.)  The remedy in each case is
to catch the exception, log the error, and try the operation again
later.  But there's no guaranteed-to-be-complete list in the Python docs
of all the exceptions that can be thrown.  A new and surprising mode of
network failure can lead to an unhandled exception, unless you catch
everything.

The Erlang approach is tempting.  Don't catch the exception at all--just
let the process crash, and restart it.  But that's a more heavyweight
operation in Python.  

> After all, unless you're writing
> software for a nuclear reactor, or an aeroplane's autopilot, chances are
> that *bugs don't really matter*. That is to say, if you release software
> with a hidden bug, the consequences generally aren't very important.

It's a retail application that would cause some business disruption and
a pissed off customer if the program went down.  Also it's in an
embedded box on a customer site.  It's not in Antarctica or anything
like that, but it's a few towns over, and someone would have to drive
there (probably through heavy traffic) if something went wrong that
power cycling the box couldn't fix.

> It took me a long time to realise that the world won't end if I write a
> piece of software with a bug. 

It's not the end of the world if I get run over a truck, but such an
event would be of enough consequence to me personally that I find it
worth going to some trouble to avoid it.
-- 
http://mail.python.org/mailman/listinfo/python-list


Re: relative speed of incremention syntaxes (or "i=i+1" VS "i+=1")

2011-08-21 Thread Roy Smith
In article ,
 Christian Heimes  wrote:

> Am 21.08.2011 19:27, schrieb Andreas Löscher:
> > As for using Integers, the first case (line 1319 and 1535) are true and
> > there is no difference in Code. However, Python uses a huge switch-case
> > construct to execute it's opcodes and INPLACE_ADD cames after
> > BINARY_ADD, hence the difference in speed. 
> 
> I don't think that's the reason. Modern compiles turn a switch statement
> into a jump or branch table rather than a linear search like chained
> elif statements.

This is true even for very small values of "modern".  I remember the 
Unix v6 C compiler (circa 1977) was able to do this.
-- 
http://mail.python.org/mailman/listinfo/python-list


Re: try... except with unknown error types

2011-08-21 Thread Chris Angelico
On Sun, Aug 21, 2011 at 7:26 PM, Paul Rubin  wrote:
> I'm not sure what to do instead.  The exceptions I'm currently dealing
> with happen when certain network operations go wrong (e.g. network or
> remote host is down, connection fails, etc.)  The remedy in each case is
> to catch the exception, log the error, and try the operation again
> later.  But there's no guaranteed-to-be-complete list in the Python docs
> of all the exceptions that can be thrown.  A new and surprising mode of
> network failure can lead to an unhandled exception, unless you catch
> everything.
>

A new and surprising mode of network failure would be indicated by a
new subclass of IOError or EnvironmentError. If you catch one of
those, you should catch it. That's the benefit of hierarchical
exceptions.

ChrisA
-- 
http://mail.python.org/mailman/listinfo/python-list


Re: try... except with unknown error types

2011-08-21 Thread Ethan Furman

Paul Rubin wrote:

Steven D'Aprano  writes:

But there's no way to know what that minimum is.  Python libraries throw
all sorts of exceptions that their documentation doesn't mention.

Yes, you're absolutely correct. But it's also irrelevant. Most of those
exceptions should not be caught, even if you know what they are, because
they represent either bugs that should be fixed, or bad data which should
raise an exception. A bare except, or except Exception, is hardly ever the
right approach.


I'm not sure what to do instead.  The exceptions I'm currently dealing
with happen when certain network operations go wrong (e.g. network or
remote host is down, connection fails, etc.)  The remedy in each case is
to catch the exception, log the error, and try the operation again
later.  But there's no guaranteed-to-be-complete list in the Python docs
of all the exceptions that can be thrown.  A new and surprising mode of
network failure can lead to an unhandled exception, unless you catch
everything.


In a case like this I can see catching everything so long as (which you 
say you are doing) you log the error somehow -- what's really 
frustrating is when the error is simply tossed with no record 
whatsoever... what a pain to debug!


~Ethan~
--
http://mail.python.org/mailman/listinfo/python-list


Re: relative speed of incremention syntaxes (or "i=i+1" VS "i+=1")

2011-08-21 Thread Terry Reedy

On 8/21/2011 1:27 PM, Andreas Löscher wrote:

What the precise difference (semantics and speed) is between the
BINARY_ADD and INPLACE_ADD opcodes, I dunno. Look in the Python source
code or maybe someone knows it from memory :-)

Irmen


from Python/ceval.c:

1316case BINARY_ADD:
1317w = POP();
1318v = TOP();
1319if (PyInt_CheckExact(v)&&  PyInt_CheckExact(w)) {
1320/* INLINE: int + int */
1321register long a, b, i;
1322a = PyInt_AS_LONG(v);
1323b = PyInt_AS_LONG(w);
1324/* cast to avoid undefined behaviour
1325   on overflow */
1326i = (long)((unsigned long)a + b);
1327if ((i^a)<  0&&  (i^b)<  0)
1328goto slow_add;
1329x = PyInt_FromLong(i);
1330}
1331else if (PyString_CheckExact(v)&&
1332 PyString_CheckExact(w)) {
1333x = string_concatenate(v, w, f, next_instr);
1334/* string_concatenate consumed the ref to v */
1335goto skip_decref_vx;
1336}
1337else {
1338  slow_add:
1339x = PyNumber_Add(v, w);
1340}
1341Py_DECREF(v);
1342  skip_decref_vx:
1343Py_DECREF(w);
1344SET_TOP(x);
1345if (x != NULL) continue;
1346break;

1532case INPLACE_ADD:
1533w = POP();
1534v = TOP();
1535if (PyInt_CheckExact(v)&&  PyInt_CheckExact(w)) {
1536/* INLINE: int + int */
1537register long a, b, i;
1538a = PyInt_AS_LONG(v);
1539b = PyInt_AS_LONG(w);
1540i = a + b;
1541if ((i^a)<  0&&  (i^b)<  0)
1542goto slow_iadd;
1543x = PyInt_FromLong(i);
1544}
1545else if (PyString_CheckExact(v)&&
1546 PyString_CheckExact(w)) {
1547x = string_concatenate(v, w, f, next_instr);
1548/* string_concatenate consumed the ref to v */
1549goto skip_decref_v;
1550}
1551else {
1552  slow_iadd:
1553x = PyNumber_InPlaceAdd(v, w);
1554}
1555Py_DECREF(v);
1556  skip_decref_v:
1557Py_DECREF(w);
1558SET_TOP(x);
1559if (x != NULL) continue;
1560break;

As for using Integers, the first case (line 1319 and 1535) are true and
there is no difference in Code. However, Python uses a huge switch-case
construct to execute it's opcodes and INPLACE_ADD cames after
BINARY_ADD, hence the difference in speed.

To be clear, this is nothing you should consider when writing fast code.
Complexity wise they both are the same.


With 64 bit 3.2.2 on my Win 7 Pentium, the difference was 4% and with 
floats (0.0 and 1.0), 6%


--
Terry Jan Reedy


--
http://mail.python.org/mailman/listinfo/python-list


Re: try... except with unknown error types

2011-08-21 Thread Terry Reedy

On 8/21/2011 2:26 PM, Paul Rubin wrote:

Steven D'Aprano  writes:

But there's no way to know what that minimum is.  Python libraries throw
all sorts of exceptions that their documentation doesn't mention.


Yes, you're absolutely correct. But it's also irrelevant. Most of those
exceptions should not be caught, even if you know what they are, because
they represent either bugs that should be fixed, or bad data which should
raise an exception. A bare except, or except Exception, is hardly ever the
right approach.


I'm not sure what to do instead.  The exceptions I'm currently dealing
with happen when certain network operations go wrong (e.g. network or
remote host is down, connection fails, etc.)  The remedy in each case is
to catch the exception, log the error, and try the operation again
later.  But there's no guaranteed-to-be-complete list in the Python docs
of all the exceptions that can be thrown.  A new and surprising mode of
network failure can lead to an unhandled exception, unless you catch
everything.


I would expect that catching socket.error (or even IOError) should catch 
all of those.


"exception socket.error
A subclass of IOError, this exception is raised for socket-related 
errors. It is recommended that you inspect its errno attribute to 
discriminate between different kinds of errors."



It's a retail application that would cause some business disruption and
a pissed off customer if the program went down.  Also it's in an
embedded box on a customer site.  It's not in Antarctica or anything
like that, but it's a few towns over, and someone would have to drive
there (probably through heavy traffic) if something went wrong that
power cycling the box couldn't fix.



--
Terry Jan Reedy

--
http://mail.python.org/mailman/listinfo/python-list


Re: relative speed of incremention syntaxes (or "i=i+1" VS "i+=1")

2011-08-21 Thread Laurent
Actually 6% between float themselves is just as not-understandable.
-- 
http://mail.python.org/mailman/listinfo/python-list


Re: relative speed of incremention syntaxes (or "i=i+1" VS "i+=1")

2011-08-21 Thread Laurent

> With 64 bit 3.2.2 on my Win 7 Pentium, the difference was 4% and with 
> floats (0.0 and 1.0), 6%

For floats it is understandable. But for integers, seriously, 4% is a lot. I 
would never have thought an interpreter would have differences like this in 
syntax for something as fundamental as adding 1.
-- 
http://mail.python.org/mailman/listinfo/python-list


Re: relative speed of incremention syntaxes (or "i=i+1" VS "i+=1")

2011-08-21 Thread Laurent
I did the test several times with floats on my machine and the difference is 
not as big as for integers:


==> "i = i + 1.0" is 0.732% faster than "i += 1.0".

It seems normal as the float addition is supposed to be slower than integer 
addition, thus the syntaxic difference is comparatively less important. 
-- 
http://mail.python.org/mailman/listinfo/python-list


Re: relative speed of incremention syntaxes (or "i=i+1" VS "i+=1")

2011-08-21 Thread Nobody
On Sun, 21 Aug 2011 09:52:23 -0700, Laurent wrote:

> I did many tests and "i = i + 1" always seems to be around 2% faster
> than "i += 1". This is no surprise as the += notation seems to be a
> syntaxic sugar layer that has to be converted to i = i + 1 anyway. Am I
> wrong in my interpretation? 

It depends. If the value on the left has an __iadd__ method, that will be
called; the value will be updated in-place, so all references to that
object will be affected:

> import numpy as np
> a = np.zeros(3)
> b = a
> a
array([ 0.,  0.,  0.])
> b
array([ 0.,  0.,  0.])
> a += 1
> a
array([ 1.,  1.,  1.])
> b
array([ 1.,  1.,  1.])

If the value on the left doesn't have an __iadd__ method, then addition is
performed and the name is re-bound to the result:

> a = a + 1
> a
array([ 2.,  2.,  2.])
> b
array([ 1.,  1.,  1.])

If you're writing code which could reasonably be expected to work with
arbitrary "numeric" values, you should decide which to use according to
whether in-place modification is appropriate rather than trivial
performance differences. If a difference of a few percent is significant,
Python is probably the wrong language in the first place.

-- 
http://mail.python.org/mailman/listinfo/python-list


Re: try... except with unknown error types

2011-08-21 Thread Roy Smith
In article <7xty9ahb84@ruckus.brouhaha.com>,
 Paul Rubin  wrote:

> It's a retail application that would cause some business disruption and
> a pissed off customer if the program went down.  Also it's in an
> embedded box on a customer site.  It's not in Antarctica or anything
> like that, but it's a few towns over, and someone would have to drive
> there (probably through heavy traffic) if something went wrong that
> power cycling the box couldn't fix.

I would do something like this:

try:
   do_some_network_stuff()
except IOError:
   do_normal_recovery()
except Exception:
   call_home()
   do_some_other_recovery()

You are right, in an embedded/unattended application, you don't want to 
ever crash.  If you ever get an error that you don't understand, you 
have no choice but to do the best you can to recover.  I would also 
generate some kind of alert back to home base to tell somebody to take a 
look at it and make sure things are fine.

I would expect that the whole application is wrapped in some kind of 
watchdog timer which will do a hard reset of the entire system.
-- 
http://mail.python.org/mailman/listinfo/python-list


Re: extended slicing and negative stop value problem

2011-08-21 Thread Chris Rebert
On Sun, Aug 21, 2011 at 10:27 AM, Max  wrote:
> On Aug 20, 1:40 pm, Steven D'Aprano  +comp.lang.pyt...@pearwood.info> wrote:
>> On Sat, Aug 20, 2011 at 7:20 PM, Max Moroz  wrote:
>> > Would it be a good idea to change Python definition so that a[10, -1, -1]
>>
>> I presume you mean slice notation a[10:-1:-1].
>>
>> > referred to the elements starting with position 10, going down to the
>> > beginning?

>> The usual advice is to do your slicing twice, reversing it the second time:
>>
>> a[0:11][::-1]
>> # Instead of a[10:-1:-1], which looks like it should work, but doesn't.
>
> It works nicely, but it is 1.3 times slower in my code (I am surprised
> the interpreter doesn't optimize this).

That would require CPython to assume certain slicing semantics for all
types (which it can't) or to check for this very specific case at
runtime (which would slow down all other [list] slicing operations).

A smarter implementation such as PyPy could indeed theoretically
optimize this case though.

Cheers,
Chris
-- 
http://mail.python.org/mailman/listinfo/python-list


Re: relative speed of incremention syntaxes (or "i=i+1" VS "i+=1")

2011-08-21 Thread Andreas Löscher
Am Sonntag, den 21.08.2011, 14:52 -0400 schrieb Roy Smith:
> In article ,
>  Christian Heimes  wrote:
> 
> > Am 21.08.2011 19:27, schrieb Andreas Lscher:
> > > As for using Integers, the first case (line 1319 and 1535) are true and
> > > there is no difference in Code. However, Python uses a huge switch-case
> > > construct to execute it's opcodes and INPLACE_ADD cames after
> > > BINARY_ADD, hence the difference in speed. 
> > 
> > I don't think that's the reason. Modern compiles turn a switch statement
> > into a jump or branch table rather than a linear search like chained
> > elif statements.
> 
> This is true even for very small values of "modern".  I remember the 
> Unix v6 C compiler (circa 1977) was able to do this.

What is the difference in speed between a jump table that is searched
from top to bottom in comparison to an ordinary if-then-elif...? The
difference can only be in the search algorithm regarding the table.
Without optimization (linear search) both are the same. If the compiler
applies some magic the difference can be relevant (linear complexity for
if-then-elif... and O(1) if you would use a dictionary). 

Hence the executed code for integers is the same, there must be a slower
path to the code of BINARY_ADD than to INPLACE_ADD. 

How would such an jump table work to behave the same liek a
switch-case-statement? Beware, that things like

   case PRINT_NEWLINE_TO:
1802w = stream = POP();
1803/* fall through to PRINT_NEWLINE */
1804
1805case PRINT_NEWLINE:

must be supported.


Bye the way:
First line of ceval.c since at least Version 2.4
1   
2   /* Execute compiled code */
3   
4   /* XXX TO DO:
5  XXX speed up searching for keywords by using a dictionary
6  XXX document it!
7  */

:-)

-- 
http://mail.python.org/mailman/listinfo/python-list


Re: relative speed of incremention syntaxes (or "i=i+1" VS "i+=1")

2011-08-21 Thread Andreas Löscher
Am Sonntag, den 21.08.2011, 12:53 -0700 schrieb Laurent:
> > With 64 bit 3.2.2 on my Win 7 Pentium, the difference was 4% and with 
> > floats (0.0 and 1.0), 6%
> 
> For floats it is understandable. But for integers, seriously, 4% is a lot. I 
> would never have thought an interpreter would have differences like this in 
> syntax for something as fundamental as adding 1.

It's not as bad as you think. The addition of two integers is a cheap
task (in terms of computation power). If you have two way's to to it,
every little think (jumps in the code etc. ) will have a larger impact
on the execution time than on an expensive operation.

But every improvement on your algorithm will easily result in a
significant shorter execution time than replaceing a+=1 with a=a+1 will
ever do. :-)

-- 
http://mail.python.org/mailman/listinfo/python-list


Re: relative speed of incremention syntaxes (or "i=i+1" VS "i+=1")

2011-08-21 Thread Chris Angelico
2011/8/22 Andreas Löscher :
> How would such an jump table work to behave the same liek a
> switch-case-statement?

If your switch statement uses a simple integer enum with sequential
values, then it can be done quite easily. Take this as an example:

switch (argc)
{
case 0: printf("No args at all, this is weird\n"); break;
case 1: printf("No args\n"); break;
case 2: printf("Default for second arg\n");
case 3: printf("Two args\n"); break;
default: printf("Too many args\n"); break;
}

I compiled this using Open Watcom C, looked at the disassembly, and
hereby translate it into pseudocode (I'll email/post the full 80x86
disassembly if you like):

1) Check if argc > 3 (unsigned comparison), if so jump to default case.
2) Left shift argc two places, add a constant offset, fetch a pointer
from there, and jump to it - that's the jump table. One JMP statement.
3) Code follows for each case.

Incidentally, the Open Watcom compiler actually turned several of the
cases into offset-load of the appropriate string pointer, and then a
jump to the single call to printf. The fall-through from 'case 2' to
'case 3' works fine, although it means that 'case 2' has to be
de-optimized from that one simplification.

This type of optimization works best when the case values are
sequential. (If I remove the 'case 0', the compiler decrements argc
and proceeds to continue as above.) Otherwise, the jump table has to
have a lot of copies of the "default" pointer.

Chris Angelico
-- 
http://mail.python.org/mailman/listinfo/python-list


Re: relative speed of incremention syntaxes (or "i=i+1" VS "i+=1")

2011-08-21 Thread Terry Reedy

On 8/21/2011 7:17 PM, Andreas Löscher wrote:

Am Sonntag, den 21.08.2011, 14:52 -0400 schrieb Roy Smith:

In article,
  Christian Heimes  wrote:


Am 21.08.2011 19:27, schrieb Andreas Lscher:

As for using Integers, the first case (line 1319 and 1535) are true and
there is no difference in Code. However, Python uses a huge switch-case
construct to execute it's opcodes and INPLACE_ADD cames after
BINARY_ADD, hence the difference in speed.


I don't think that's the reason. Modern compiles turn a switch statement
into a jump or branch table rather than a linear search like chained
elif statements.


This is true even for very small values of "modern".  I remember the
Unix v6 C compiler (circa 1977) was able to do this.


What is the difference in speed between a jump table that is searched
from top to bottom in comparison to an ordinary if-then-elif...? The
difference can only be in the search algorithm regarding the table.
Without optimization (linear search) both are the same. If the compiler
applies some magic the difference can be relevant (linear complexity for
if-then-elif... and O(1) if you would use a dictionary).


A jump or branch table is applicable when the case value values are all 
small ints, like bytes or less. For C, the table is simply an array of 
pointers (addressess, with entries for unused byte codes would be a void 
pointer). Hence, O(1) access.

https://secure.wikimedia.org/wikipedia/en/wiki/Jump_table


Hence the executed code for integers is the same, there must be a slower
path to the code of BINARY_ADD than to INPLACE_ADD.

How would such an jump table work to behave the same liek a
switch-case-statement? Beware, that things like

case PRINT_NEWLINE_TO:
1802w = stream = POP();
1803/* fall through to PRINT_NEWLINE */


add jump to address of the code for PRINT_NEWLINE


1804
1805case PRINT_NEWLINE:

must be supported.


--
Terry Jan Reedy


--
http://mail.python.org/mailman/listinfo/python-list


Re: relative speed of incremention syntaxes (or "i=i+1" VS "i+=1")

2011-08-21 Thread Chris Angelico
2011/8/22 Andreas Löscher :
> But every improvement on your algorithm will easily result in a
> significant shorter execution time than replaceing a+=1 with a=a+1 will
> ever do. :-)
>

Agreed. If Python needed a faster alternative to "a=a+1", then I would
recommend an "a.inc()" call or something; some way to avoid looking up
the value of 1. (An equivalent to C's ++a operation, if you like.) But
I think you'd be hard-pressed to find any situation where improving
the speed of incrementing will be significant, that wouldn't be better
served by algorithmic improvements (even just using "for a in
range()") or dropping to C.

ChrisA
-- 
http://mail.python.org/mailman/listinfo/python-list


Re: relative speed of incremention syntaxes (or "i=i+1" VS "i+=1")

2011-08-21 Thread Laurent Payot
I made Python my language of choice because of its readability and simpleness, 
and not because of its speed. But it's always good to know what is the fastest 
sometimes when you don't want to write a module in C. So I was just wondering 
if there was a difference. There is, of a few percent. Anyway I will keep on 
using the 2% slower "i += 1" because for me that's less prone to errors because 
you write the variable only once, and that's more important than speed.
-- 
http://mail.python.org/mailman/listinfo/python-list


Re: relative speed of incremention syntaxes (or "i=i+1" VS "i+=1")

2011-08-21 Thread Andreas Löscher
Am Sonntag, den 21.08.2011, 19:38 -0400 schrieb Terry Reedy:
> On 8/21/2011 7:17 PM, Andreas Löscher wrote:
> > Am Sonntag, den 21.08.2011, 14:52 -0400 schrieb Roy Smith:
> >> In article,
> >>   Christian Heimes  wrote:
> >>
> >>> Am 21.08.2011 19:27, schrieb Andreas Lscher:
>  As for using Integers, the first case (line 1319 and 1535) are true and
>  there is no difference in Code. However, Python uses a huge switch-case
>  construct to execute it's opcodes and INPLACE_ADD cames after
>  BINARY_ADD, hence the difference in speed.
> >>>
> >>> I don't think that's the reason. Modern compiles turn a switch statement
> >>> into a jump or branch table rather than a linear search like chained
> >>> elif statements.
> >>
> >> This is true even for very small values of "modern".  I remember the
> >> Unix v6 C compiler (circa 1977) was able to do this.
> >
> > What is the difference in speed between a jump table that is searched
> > from top to bottom in comparison to an ordinary if-then-elif...? The
> > difference can only be in the search algorithm regarding the table.
> > Without optimization (linear search) both are the same. If the compiler
> > applies some magic the difference can be relevant (linear complexity for
> > if-then-elif... and O(1) if you would use a dictionary).
> 
> A jump or branch table is applicable when the case value values are all 
> small ints, like bytes or less. For C, the table is simply an array of 
> pointers (addressess, with entries for unused byte codes would be a void 
> pointer). Hence, O(1) access.
> https://secure.wikimedia.org/wikipedia/en/wiki/Jump_table
> 
> > Hence the executed code for integers is the same, there must be a slower
> > path to the code of BINARY_ADD than to INPLACE_ADD.
> >
> > How would such an jump table work to behave the same liek a
> > switch-case-statement? Beware, that things like
> >
> > case PRINT_NEWLINE_TO:
> > 1802w = stream = POP();
> > 1803/* fall through to PRINT_NEWLINE */
> 
> add jump to address of the code for PRINT_NEWLINE
> 
> > 1804
> > 1805case PRINT_NEWLINE:
> >
> > must be supported.
> 

:-) too easy or too late 
thanks

-- 
http://mail.python.org/mailman/listinfo/python-list


Re: try... except with unknown error types

2011-08-21 Thread Steven D'Aprano
Paul Rubin wrote:

> Steven D'Aprano  writes:
>>> But there's no way to know what that minimum is.  Python libraries throw
>>> all sorts of exceptions that their documentation doesn't mention.
>>
>> Yes, you're absolutely correct. But it's also irrelevant. Most of those
>> exceptions should not be caught, even if you know what they are, because
>> they represent either bugs that should be fixed, or bad data which should
>> raise an exception. A bare except, or except Exception, is hardly ever
>> the right approach.
> 
> I'm not sure what to do instead.  The exceptions I'm currently dealing
> with happen when certain network operations go wrong (e.g. network or
> remote host is down, connection fails, etc.)  The remedy in each case is
> to catch the exception, log the error, and try the operation again
> later.  But there's no guaranteed-to-be-complete list in the Python docs
> of all the exceptions that can be thrown.  A new and surprising mode of
> network failure can lead to an unhandled exception, unless you catch
> everything.

I was waiting for you to raise network errors :)

Network errors are a particularly annoying case, because although rare and
undocumented, they are legitimate errors that should be caught. I feel your
pain.


> The Erlang approach is tempting.  Don't catch the exception at all--just
> let the process crash, and restart it.  But that's a more heavyweight
> operation in Python.

The Erland approach sounds good, but as I've never used it, I don't know how
well it works in practice.


>> After all, unless you're writing
>> software for a nuclear reactor, or an aeroplane's autopilot, chances are
>> that *bugs don't really matter*. That is to say, if you release software
>> with a hidden bug, the consequences generally aren't very important.
> 
> It's a retail application that would cause some business disruption and
> a pissed off customer if the program went down.  Also it's in an
> embedded box on a customer site.  It's not in Antarctica or anything
> like that, but it's a few towns over, and someone would have to drive
> there (probably through heavy traffic) if something went wrong that
> power cycling the box couldn't fix.

Customers are always pissed off when something goes wrong, but ask them to
pay an extra $300 for a battery backup unit to the hardware RAID controller
*they* insisted on against your advice, and they say no. But I'm not
bitter...

Customer gets pissed off. What's the consequences? Do they sue you? Leave?
Are they locked into a five year contract and have to pay you even if they
do leave? Do you have to fix the incident at no charge, or give them two
hours free support? Are you competing with armour-plated mature software
that never goes down, or is the field yours alone? How price sensitive are
your customers? Will they pay an extra $100,000 for an extra 9 in the
expected uptime?

Without knowing the consequences to *you* of failure, I can't tell you where
you should be spending your time: trying to predict errors ahead of time,
or fixing them once they've been seen. This is a business problem, not a
software problem.

Don't misunderstand me, I'm not suggesting that you shouldn't try to avoid
things going wrong. But it's a hard problem. Solving it is why they pay you
the big bucks *cough*:

http://www.joelonsoftware.com/items/2007/12/06.html

and is the difference between version 0.4 of an application, and version
3.4. If you're trying to go straight to version 3.4 without putting the
software through real-world testing, then your testing better be mean and
tough. *Really* tough.

http://www.codinghorror.com/blog/2011/04/working-with-the-chaos-monkey.html


-- 
Steven

-- 
http://mail.python.org/mailman/listinfo/python-list


Re: try... except with unknown error types

2011-08-21 Thread Steven D'Aprano
Chris Angelico wrote:


> A new and surprising mode of network failure would be indicated by a
> new subclass of IOError or EnvironmentError. 

/s/would/should/

I don't see why you expect this, when *existing* network-related failures
aren't:

>>> import socket
>>> issubclass(socket.error, EnvironmentError)
False

(Fortunately that specific example is fixed in Python 3.)

Besides, there's a world of difference between "should be" and "are".


-- 
Steven

-- 
http://mail.python.org/mailman/listinfo/python-list


Re: try... except with unknown error types

2011-08-21 Thread Chris Angelico
On Mon, Aug 22, 2011 at 1:30 AM, Steven D'Aprano
 wrote:
> /s/would/should/
>
> I don't see why you expect this, when *existing* network-related failures
> aren't

Ehh, granted. Definitely a case of "should". But certainly, there
won't be an infinite number of new exceptions invented; most of the
runtime issues you'll have will fall into a fairly small number of
exception types (either by subclassing or by parameters eg errno).

ChrisA
-- 
http://mail.python.org/mailman/listinfo/python-list


Re: relative speed of incremention syntaxes (or "i=i+1" VS "i+=1")

2011-08-21 Thread Terry Reedy

On 8/21/2011 5:07 PM, Nobody wrote:


If the value on the left has an __iadd__ method, that will be called;


Correct


the value will be updated in-place,


Not necessarily correct. The target is rebound to the return from the 
__iadd__ method. Augmented *assignment* is always assignment. This trips 
up people who try


t = (1, [])
t[1] += [1,2] # which *does* extend the array, but raises
TypeError: 'tuple' object does not support item assignment
# instead of
t[1].extend([1,2]) # which extends without raising an error


*IF* (and only if) the target object is mutable, then the __iadd__ may 
optionally mutate the target object. But the rebinding takes place 
nonetheless. Numbers, the subject of this thread, are not mutable and 
are not 'updated in-place'


class test:
def __init__(self, n):
self.n = n
def __iadd__(self, other):
return test(self.n + other.n)
def __repr__(self):
return repr(self.n)

r = test(1)
t = r
t += r
print(r, t)
# 1,2

That said, there is normally no reason to write an __ixxx__ method 
unless instances are mutable and the operation can be and is done in 
place therein. So the class above is for illustrative purposes only. A 
saner example is the following, which treats test examples as mutable 
number containers rather than as immutable surrogates.


class test:
def __init__(self, n):
self.n = n
def __add__(self, other):
return test(self.n + other.n)
def __iadd__(self, other):
n = self.n + other.n
self.n = n
return n
def __repr__(self):
return repr(self.n)

r = test(1)
t = r
t += r
print(r, t)
# 2 2

The interpreter cannot enforce that 'x += a' have the same effect as 'x 
= x+a', but it would break normal expectations to make the two different.



> so all references to that object will be affected:

Only if the target object is mutable and is mutated by the optional 
augmented assignment __ixxx__ methods.



>  import numpy as np
>  a = np.zeros(3)


Numpy arrays meet the qualification above.


If the value on the left doesn't have an __iadd__ method, then addition is
performed and the name is re-bound to the result:


As is also done with the results of __ixxx__ methods.

--
Terry Jan Reedy

--
http://mail.python.org/mailman/listinfo/python-list


Re: relative speed of incremention syntaxes (or "i=i+1" VS "i+=1")

2011-08-21 Thread Steven D'Aprano
Laurent wrote:

> 
>> With 64 bit 3.2.2 on my Win 7 Pentium, the difference was 4% and with
>> floats (0.0 and 1.0), 6%
> 
> For floats it is understandable. But for integers, seriously, 4% is a lot.
> I would never have thought an interpreter would have differences like this
> in syntax for something as fundamental as adding 1.

Why? Python integers are rich objects, not native ints. Adding 1 is a
moderately heavyweight operation far more complicated than the same
operation in C.

n=n+1 and n+=1 call different methods and do different things, they are
*not* just different syntax. Your intuition about what should and shouldn't
take the same time should not be trusted.

But really, we're talking about tiny differences in speed. Such trivial
differences are at, or beyond, the limit of what can realistically be
measured on a noisy PC running multiple processes (pretty much all PCs
these days). Here are three runs of each on my computer:


[steve@sylar python]$ python2.5 -m timeit -s 'n=0' 'n = n+1'
100 loops, best of 3: 0.508 usec per loop
[steve@sylar python]$ python2.5 -m timeit -s 'n=0' 'n = n+1'
100 loops, best of 3: 0.587 usec per loop
[steve@sylar python]$ python2.5 -m timeit -s 'n=0' 'n = n+1'
100 loops, best of 3: 0.251 usec per loop


[steve@sylar python]$ python2.5 -m timeit -s 'n=0' 'n += 1'
100 loops, best of 3: 0.226 usec per loop
[steve@sylar python]$ python2.5 -m timeit -s 'n=0' 'n += 1'
100 loops, best of 3: 0.494 usec per loop
[steve@sylar python]$ python2.5 -m timeit -s 'n=0' 'n += 1'
100 loops, best of 3: 0.53 usec per loop

Look at the variation between runs! About 130% variation between the fastest
and slowest for each expression. And there's no reason to think that the
fastest results shown is as fast as it can get. The time is dominated by
noise, not the addition.


For what it's worth, if I try it with a more recent Python:

[steve@sylar python]$ python3.2 -m timeit -s 'n=0' 'n = n+1'
100 loops, best of 3: 0.221 usec per loop
[steve@sylar python]$ python3.2 -m timeit -s 'n=0' 'n = n+1'
100 loops, best of 3: 0.202 usec per loop
[steve@sylar python]$ python3.2 -m timeit -s 'n=0' 'n = n+1'
100 loops, best of 3: 0.244 usec per loop

[steve@sylar python]$ python3.2 -m timeit -s 'n=0' 'n += 1'
100 loops, best of 3: 0.49 usec per loop
[steve@sylar python]$ python3.2 -m timeit -s 'n=0' 'n += 1'
100 loops, best of 3: 0.176 usec per loop
[steve@sylar python]$ python3.2 -m timeit -s 'n=0' 'n += 1'
100 loops, best of 3: 0.49 usec per loop


I simply do not believe that we can justify making *any* claim about the
relative speeds of n=n+1 and n+=1 other than "they are about the same". Any
result you get, faster or slower, will depend more on chance than on any
real or significant difference in the code.


-- 
Steven

-- 
http://mail.python.org/mailman/listinfo/python-list


Re: relative speed of incremention syntaxes (or "i=i+1" VS "i+=1")

2011-08-21 Thread Steven D'Aprano
Chris Angelico wrote:

> 2011/8/22 Andreas Löscher :
>> But every improvement on your algorithm will easily result in a
>> significant shorter execution time than replaceing a+=1 with a=a+1 will
>> ever do. :-)
>>
> 
> Agreed. If Python needed a faster alternative to "a=a+1", then I would
> recommend an "a.inc()" call or something; some way to avoid looking up
> the value of 1.

Keep in mind that ints in Python are objects, not memory locations, and can
be shared. If you are hoping for an in-place increment, I invite you to
consider the following code and try to predict what it would do:

a = 42
b = a
b.inc()
print(a)



-- 
Steven

-- 
http://mail.python.org/mailman/listinfo/python-list


Re: Installing numpy on 2.7 (OS X 10.7)

2011-08-21 Thread jefflovejapan
On Aug 21, 10:39 pm, Benjamin Kaplan  wrote:
> On Sun, Aug 21, 2011 at 6:03 AM, jefflovejapan  
> wrote:
> > I'm following the instructions given http://www.scipy.org/
> > Installing_SciPy/Mac_OS_X">here, but it isn't working.
> > Specifically, I'm getting:
>
> > Library/Frameworks/Python.framework/Versions/2.7/Resources/Python.app/
> > Contents/MacOS/Python: can't open file 'setup.py': [Errno 2] No such
> > file or directory
>
> > even though I completed the two downloads from github successfully.
> > Any ideas why this wouldn't be working, or any ideas where I can look
> > for better instructions?
> > --
>
> As the error says, it can't find the setup.py file. My guess would be
> you forgot to cd to the newly created directory. You have to "cd
> numpy" (or cd scipy if you're trying to install that) first to get
> into the directory with the install script.

Of course! Thanks a lot.
-- 
http://mail.python.org/mailman/listinfo/python-list


Re: relative speed of incremention syntaxes (or "i=i+1" VS "i+=1")

2011-08-21 Thread Christian Heimes
Am 22.08.2011 01:25, schrieb Andreas Löscher:
> It's not as bad as you think. The addition of two integers is a cheap
> task (in terms of computation power). If you have two way's to to it,
> every little think (jumps in the code etc. ) will have a larger impact
> on the execution time than on an expensive operation.
> 
> But every improvement on your algorithm will easily result in a
> significant shorter execution time than replaceing a+=1 with a=a+1 will
> ever do. :-)

You can learn an important lesson here. Since Python is a high level
language without a JIT (yet) integer operations are much slower than in
C or other low level languages. In general it's not a problem for most
people. However if your algorithm is speed bound to integer operations
or has a tight inner for loop than you can gain a considerable amount of
speedup with C code. Reference counting and Python object creation need
several CPU cycles. Also a good algorithm and a modern C compiler can
make use of SIMD instructions, too.

If you ever feel the need to implement something fast e.g. an encryption
algorithm then you'd better off with a C implementation. Cython is my
favourite tool for the job.

Christian

-- 
http://mail.python.org/mailman/listinfo/python-list


Re: Wait for a keypress before continuing?

2011-08-21 Thread Tim Roberts
John Doe  wrote:
>
>Tim Roberts  wrote: 
>
>> That exact code works perfectly for me.  The function returns as
>> soon as I press the escape key.  You are running this from a
>> console process, and not a GUI process, right? 
>
>No. I am running this from within Windows, all sorts of Windows. 
>
>So... Does that mean I will need something complex like a keyboard
>hook? Or what? 

I see that this conversation took a nasty turn while I was on vacation.

msvcrt.getch works with consoles.  If you have an application where stdin
and stdout are connected to a real, live console window (which looks just
like a command line window), then msvcrt.getch will work.  If not, then you
have to use the Windows APIs.  GetKeyboardState and GetKeyState can tell
you if a specific key is currently being pressed.
-- 
Tim Roberts, t...@probo.com
Providenza & Boekelheide, Inc.
-- 
http://mail.python.org/mailman/listinfo/python-list


Re: relative speed of incremention syntaxes (or "i=i+1" VS "i+=1")

2011-08-21 Thread Stephen Hansen
On 8/21/11 12:53 PM, Laurent wrote:
> 
>> With 64 bit 3.2.2 on my Win 7 Pentium, the difference was 4% and with 
>> floats (0.0 and 1.0), 6%
> 
> For floats it is understandable. But for integers, seriously, 4% is a lot. I 
> would never have thought an interpreter would have differences like this in 
> syntax for something as fundamental as adding 1.

Its, seriously, not even kind of a lot at all. Percentages without
context are meaningless: its 4% slower, sure -- but that is 4% of an
incredibly small probably constant-time amount of time.

Picking "i += 1" over "i = i + 1" based on one being 4% slower is sorta
kinda crazy. The difference in speed is probably related to churn and
cache as much as anything else (its not as consistent on my machine, for
example): or the ceval loop doing a few more case-tests between them as
others have mentioned. All in all, if 4% of a nanomicrofraction of a
chunk of time is that meaningful, you're probably best served not using
Python. :)

That said: my advice is always to avoid += like a plague. It is magic
and impossible to predict without intimate knowledge of exactly what's
on the left-side.

   i += 1
   n += x

Those two things look very similar, but they may do -completely-
different things depending on just what "n" is.

It may or may not do something that is like:

n = n + x

Or, it may do something that's more akin to

n.extend(x)
n = n

Those aren't even kind of equivalent actions. And things get more
complicated if 'n' is say, n[0] (especially if something goes wrong
between the extend and the rebinding).

Python's usually all explicit and pretty well-defined in how its basic
syntax and behaviors operate, and you usually don't really have to know
details about how a data-type works to predict exactly what it's doing:
in fact, its often beneficial to not pay too much attention to such
details, and just assume the data type will work approximately as you'd
expect. That way people can slip something-something to you and wink and
say of /course/ its a dict, darling. Try it, you'll like it, okay? This
sorta thing is encouraged, but it kinda depends on trusting objects to
behave a certain way and for things to be predictable in both how they
work and how they fail.

With "i = i + 1", I know that generally speaking, my "i" is being
assigned a new object and that's that, no matter what type "i" is.
(Okay: I do know that you could modify __add__ to do something
underhanded here, tweaking internal state and then returning self.
People going out of their way to behave unpredictably is not my
objection: supposedly easy and straight-forward normal Python-fu being
inherently unpredictable is).

For example: I just /know/ that it doesn't matter who or what may have
their own binding to that object before I go and increment it, they
won't be affected and everything just will work fine. With augmented
assignment, I can't be sure of that. Now, while I admit, you generally
do have to keep track in your head of which of your data-types are
mutable vs immutable and take care with sharing mutables, the fact that
"n += x" is described and generally thought of as merely syntactical
sugar for:

   n = n + x

... lets one easily think that this should be entirely safe, even with
mutable objects, because if += were merely syntactical sugar, it would
be. But its not! Because += is wiggly. It can do more then one entirely
different kind of behavior.

Anyways.  I've been kinda annoyed at augmented assignment for
years now :P

-- 

   Stephen Hansen
   ... Also: Ixokai
   ... Mail: me+list/python (AT) ixokai (DOT) io
   ... Blog: http://meh.ixokai.io/



signature.asc
Description: OpenPGP digital signature
-- 
http://mail.python.org/mailman/listinfo/python-list


Re: relative speed of incremention syntaxes (or "i=i+1" VS "i+=1")

2011-08-21 Thread Terry Reedy

On 8/21/2011 7:41 PM, Chris Angelico wrote:


Agreed. If Python needed a faster alternative to "a=a+1", then I would
recommend an "a.inc()" call or something


But looking up the method name, creating a bound method wrapper, and 
making the call would probably be slower than the syntax;-).


--
Terry Jan Reedy

--
http://mail.python.org/mailman/listinfo/python-list


Re: try... except with unknown error types

2011-08-21 Thread Roy Smith
In article <4e51a205$0$29974$c3e8da3$54964...@news.astraweb.com>,
 Steven D'Aprano  wrote:

> http://www.codinghorror.com/blog/2011/04/working-with-the-chaos-monkey.html

I *love* being the Chaos Monkey!

A few jobs ago, I had already turned in my resignation and was a 
short-timer, counting down the days to when I was out of there.  A new 
server had just come in and I was helping the guy who was going to 
replace me set it up.

This was the first server we had gotten with RAID and redundant power 
supplies and I wanted to test them out.  My accomplice was horrified, 
but technically I was still his boss, so I got to do what I wanted.  I 
opened up the drive shelve and yanked out a drive.  The console printed 
a nice neat warning about loss of media and kept chugging along, just 
like it should.

Then I pulled out a second drive.  Sure enough, the whole array failed, 
just like expected.

Then, I decided to have a go at the redundant power supplies.  We paid a 
lot of money for that, and I damn well was going to test them now, when 
there was no real risk.  I grabbed the big handle on the front of one of 
the (hot swappable) power supplies and pulled.  The whole thing went 
dark.  Turns out there had been an configuration error and the second 
power supply had never been installed (just a faceplate that looked like 
a power supply).  My buddy was pissed, but I figure I'd just done him a 
big favor.  Better to find out about it now than when the supply failed 
at some critical juncture.
-- 
http://mail.python.org/mailman/listinfo/python-list


Re: relative speed of incremention syntaxes (or "i=i+1" VS "i+=1")

2011-08-21 Thread Terry Reedy

On 8/21/2011 7:49 PM, Laurent Payot wrote:

I made Python my language of choice because of its readability and
simpleness, and not because of its speed. But it's always good to
know what is the fastest sometimes when you don't want to write a
module in C. So I was just wondering if there was a difference. There
is, of a few percent. Anyway I will keep on using the 2% slower "i +=
1" because for me that's less prone to errors because you write the
variable only once, and that's more important than speed.


For longer variable names, it is also easier and faster to read once one 
gets used to the idiom.


number_of_chars += 1 # versus
number_of_chars = number_of_chars + 1

Not repeating was a major reason for the addition.

--
Terry Jan Reedy

--
http://mail.python.org/mailman/listinfo/python-list


Re: try... except with unknown error types

2011-08-21 Thread Steven D'Aprano
On Mon, 22 Aug 2011 10:41 am Chris Angelico wrote:

> On Mon, Aug 22, 2011 at 1:30 AM, Steven D'Aprano
>  wrote:
>> /s/would/should/
>>
>> I don't see why you expect this, when *existing* network-related failures
>> aren't
> 
> Ehh, granted. Definitely a case of "should". But certainly, there
> won't be an infinite number of new exceptions invented; most of the
> runtime issues you'll have will fall into a fairly small number of
> exception types (either by subclassing or by parameters eg errno).

Er, you can't know that either.

Except that all exceptions must be rooted at BaseException, there *can* be
an infinite number of new exception types invented. Or functions could
raise unexpected, but known, exceptions, whether built-in or not.

Of course, the set of hypothetical exceptions that could theoretically be
raised is much, much bigger than the set of exceptions which you can
realistically expect to see, especially if you limit yourself to those that
don't leave you wondering about the sanity of the developer ("WTF? Why is
x+1 raising HTTPBadGatewayError?"). But still, as a general rule you should
not be surprised to see any of:

AttributeError
TypeError
ValueError
(and likely others)

The more important question is, if you get an exception, any exception, is
that a bug in the function you are calling, a bug in your calling code (you
supplied a bad argument), or an expected result that you should catch and
work around?


-- 
Steven

-- 
http://mail.python.org/mailman/listinfo/python-list


Re: relative speed of incremention syntaxes (or "i=i+1" VS "i+=1")

2011-08-21 Thread Steven D'Aprano
On Mon, 22 Aug 2011 12:08 pm Stephen Hansen wrote:

> Picking "i += 1" over "i = i + 1" based on one being 4% slower is sorta
> kinda crazy. The difference in speed is probably related to churn and
> cache as much as anything else (its not as consistent on my machine, for
> example): or the ceval loop doing a few more case-tests between them as
> others have mentioned. All in all, if 4% of a nanomicrofraction of a
> chunk of time is that meaningful, you're probably best served not using
> Python. :)

Agreed.

But...

> That said: my advice is always to avoid += like a plague. It is magic
> and impossible to predict without intimate knowledge of exactly what's
> on the left-side.
> 
>i += 1
>n += x
> 
> Those two things look very similar, but they may do -completely-
> different things depending on just what "n" is.


Technically, the *exact same criticism* can be applied to:

n = n + x

since either n or x could override __add__ or __radd__ and do anything it
bloody well likes. Including in-place modification of *either* argument
(where possible).


[...]
> With "i = i + 1", I know that generally speaking, my "i" is being
> assigned a new object and that's that, no matter what type "i" is.
> (Okay: I do know that you could modify __add__ to do something
> underhanded here, tweaking internal state and then returning self.

What makes you think that's underhanded?

To my mind, __add__ modifying self as a side-effect is unusual, but apart
from breaking the general rule of thumb to avoid side-effects, not
particularly evil. (But I would accept this is a code smell.)


> People going out of their way to behave unpredictably is not my
> objection: supposedly easy and straight-forward normal Python-fu being
> inherently unpredictable is).
> 
> For example: I just /know/ that it doesn't matter who or what may have
> their own binding to that object before I go and increment it, they
> won't be affected and everything just will work fine. 

But you can't /know/ that at all, unless you know that the object
isn't "underhanded" (however you define that!). And given some arbitrary
object, how can you know that?



-- 
Steven

-- 
http://mail.python.org/mailman/listinfo/python-list


Re: relative speed of incremention syntaxes (or "i=i+1" VS "i+=1")

2011-08-21 Thread casevh
On Aug 21, 10:27 am, Andreas Löscher  wrote:
>
> from Python/ceval.c:
>
> 1316            case BINARY_ADD:
> 1317                w = POP();
> 1318                v = TOP();
> 1319                if (PyInt_CheckExact(v) && PyInt_CheckExact(w)) {
> 1320                    /* INLINE: int + int */
> 1321                    register long a, b, i;
> 1322                    a = PyInt_AS_LONG(v);
> 1323                    b = PyInt_AS_LONG(w);
> 1324                    /* cast to avoid undefined behaviour
> 1325                       on overflow */
> 1326                    i = (long)((unsigned long)a + b);
> 1327                    if ((i^a) < 0 && (i^b) < 0)
> 1328                        goto slow_add;
> 1329                    x = PyInt_FromLong(i);
> 1330                }
> 1331                else if (PyString_CheckExact(v) &&
> 1332                         PyString_CheckExact(w)) {
> 1333                    x = string_concatenate(v, w, f, next_instr);
> 1334                    /* string_concatenate consumed the ref to v */
> 1335                    goto skip_decref_vx;
> 1336                }
> 1337                else {
> 1338                  slow_add:
> 1339                    x = PyNumber_Add(v, w);
> 1340                }
> 1341                Py_DECREF(v);
> 1342              skip_decref_vx:
> 1343                Py_DECREF(w);
> 1344                SET_TOP(x);
> 1345                if (x != NULL) continue;
> 1346                break;
>
> 1532            case INPLACE_ADD:
> 1533                w = POP();
> 1534                v = TOP();
> 1535                if (PyInt_CheckExact(v) && PyInt_CheckExact(w)) {
> 1536                    /* INLINE: int + int */
> 1537                    register long a, b, i;
> 1538                    a = PyInt_AS_LONG(v);
> 1539                    b = PyInt_AS_LONG(w);
> 1540                    i = a + b;
> 1541                    if ((i^a) < 0 && (i^b) < 0)
> 1542                        goto slow_iadd;
> 1543                    x = PyInt_FromLong(i);
> 1544                }
> 1545                else if (PyString_CheckExact(v) &&
> 1546                         PyString_CheckExact(w)) {
> 1547                    x = string_concatenate(v, w, f, next_instr);
> 1548                    /* string_concatenate consumed the ref to v */
> 1549                    goto skip_decref_v;
> 1550                }
> 1551                else {
> 1552                  slow_iadd:
> 1553                    x = PyNumber_InPlaceAdd(v, w);
> 1554                }
> 1555                Py_DECREF(v);
> 1556              skip_decref_v:
> 1557                Py_DECREF(w);
> 1558                SET_TOP(x);
> 1559                if (x != NULL) continue;
> 1560                break;
>
> As for using Integers, the first case (line 1319 and 1535) are true and
> there is no difference in Code. However, Python uses a huge switch-case
> construct to execute it's opcodes and INPLACE_ADD cames after
> BINARY_ADD, hence the difference in speed.

That fragment of cevel.c is from a 2.x version. Python 2.x supports
both a PyInt and PyLong type and the cevel loop optimized the PyInt
case only. On my system, I could not measure a difference between
binary and inplace addition.

Python 3.x behaves differently:

TARGET(BINARY_ADD)
w = POP();
v = TOP();
if (PyUnicode_CheckExact(v) &&
 PyUnicode_CheckExact(w)) {
x = unicode_concatenate(v, w, f, next_instr);
/* unicode_concatenate consumed the ref to v */
goto skip_decref_vx;
}
else {
x = PyNumber_Add(v, w);
}
Py_DECREF(v);
  skip_decref_vx:
Py_DECREF(w);
SET_TOP(x);
if (x != NULL) DISPATCH();
break;

TARGET(INPLACE_ADD)
w = POP();
v = TOP();
if (PyUnicode_CheckExact(v) &&
 PyUnicode_CheckExact(w)) {
x = unicode_concatenate(v, w, f, next_instr);
/* unicode_concatenate consumed the ref to v */
goto skip_decref_v;
}
else {
x = PyNumber_InPlaceAdd(v, w);
}
Py_DECREF(v);
  skip_decref_v:
Py_DECREF(w);
SET_TOP(x);
if (x != NULL) DISPATCH();
break;

cevel just calls PyNumber_Add or PyNumber_InPlaceAdd. If you look at
the code for PyNumber_InPlaceAdd (in abstract.c), it calls an internal
function binary_iop1 with pointers to nb_inplace_add and nb_add.
binary_iop1 then checks if nb_inplace_add exists. The PyLong type does
not implement nb_inplace_add so the check fails and binary_iop1 used
nb_add.

In recent version of gmpy and gmpy2, I implemented the nb_inplace_add
function and performance (for the gmpy.mpz type) is much better for
the in-place addition.

For the adventuresome, gmpy2 implements a mutable integer type calle

Re: relative speed of incremention syntaxes (or "i=i+1" VS "i+=1")

2011-08-21 Thread Stephen Hansen
On 8/21/11 9:14 PM, Steven D'Aprano wrote:
>> That said: my advice is always to avoid += like a plague. It is magic
>> and impossible to predict without intimate knowledge of exactly what's
>> on the left-side.
>>
>>i += 1
>>n += x
>>
>> Those two things look very similar, but they may do -completely-
>> different things depending on just what "n" is.
> 
> Technically, the *exact same criticism* can be applied to:
> 
> n = n + x
> 
> since either n or x could override __add__ or __radd__ and do anything it
> bloody well likes. Including in-place modification of *either* argument
> (where possible).

I know: I addressed that. :P See, you know I addressed that, because:

> [...]
>> With "i = i + 1", I know that generally speaking, my "i" is being
>> assigned a new object and that's that, no matter what type "i" is.
>> (Okay: I do know that you could modify __add__ to do something
>> underhanded here, tweaking internal state and then returning self.
> 
> What makes you think that's underhanded?

You quoted me addressing that very fact, and responded! :)

Its underhanded outside of narrow, well-defined situations such as ORM's
and the like where they're doing interesting and novel things with the
object model.

Underhanded doesn't mean there's no reason one should ever do it, but
its very much an unusual thing and objects that do things like that
should NOT freely interact with general Python objects: very surprising
behavior will result.

> To my mind, __add__ modifying self as a side-effect is unusual, but apart
> from breaking the general rule of thumb to avoid side-effects, not
> particularly evil. (But I would accept this is a code smell.)

I didn't really say its evil, just that its underhanded: that's like,
sketchy, not evil. :)

There's good reasons to do it on occasion, but if an object does
something like that then that is a very special kind of object and when
using it, you need to be very aware of its unusual semantics.

Having objects like that is fine.

However, for the /language/ to have unusual (and unpredictable, from
something of a distance at least) semantics, rubs me very much the wrong
way.

>> People going out of their way to behave unpredictably is not my
>> objection: supposedly easy and straight-forward normal Python-fu being
>> inherently unpredictable is).
>>
>> For example: I just /know/ that it doesn't matter who or what may have
>> their own binding to that object before I go and increment it, they
>> won't be affected and everything just will work fine. 
> 
> But you can't /know/ that at all, unless you know that the object
> isn't "underhanded" (however you define that!). And given some arbitrary
> object, how can you know that?

I CAN know it with sufficient certainty that I can rely on it, and if I
get messed up on it-- its never happened yet-- then I can fire someone
or get a new library, after staring at an odd traceback and scratching
my head for a minute.

Python has very soft rules, I know that. Objects /can/ do all kinds of
crazy things. I'm okay with that.

But I can rely on certain behaviors being consistent: the basic behavior
of the language is very straight-forward. It has hooks where you can do
lots of Special stuff, and if objects are Special, they're described as
Special.

ORM's have a lot of Special objects, for example. That's fine: when
messing with ORM objects, I know I have to take care with the operators
I use against them, because I know I'm not using a /usual/ Python
object. SQLAlchemy for example. This is not a criticism of SQLAlchemy:
having magic objects lets it construct SQL in ways that are vastly
easier then if they stuck to more regularly behaving objects.

But, += is Python itself adding an unpredictable behavior into the core
language, with its own base types behaving

The *exact same criticism* can NOT be applied to

n = n + x

Because my criticism isn't about one choosing to do crazy stuff with the
object model. I've never advocated Python be strict about rules.

But for Python, all by itself, with nothing but built-in and basic
types, to have a situation where:

a = a + b
a += b

... does two very distinctly different actions, even if in many or
even most circumstances the end-result is probably the same and probably
fine, is my criticism.

-- 

   Stephen Hansen
   ... Also: Ixokai
   ... Mail: me+list/python (AT) ixokai (DOT) io
   ... Blog: http://meh.ixokai.io/



signature.asc
Description: OpenPGP digital signature
-- 
http://mail.python.org/mailman/listinfo/python-list


Re: relative speed of incremention syntaxes (or "i=i+1" VS "i+=1")

2011-08-21 Thread Stephen Hansen
On 8/21/11 9:37 PM, Stephen Hansen wrote:
> But, += is Python itself adding an unpredictable behavior into the core
> language, with its own base types behaving

... very differently to fundamental, basic appearing operators.

Editing fail on my part.

Similarly:

> But for Python, all by itself, with nothing but built-in and basic
> types, to have a situation where:
> 
> a = a + b
> a += b

... would be clearer if the second example were "x += y".

> ... does two very distinctly different actions, even if in many or
> even most circumstances the end-result is probably the same and probably
> fine, is my criticism.
> 

-- 

   Stephen Hansen
   ... Also: Ixokai
   ... Mail: me+list/python (AT) ixokai (DOT) io
   ... Blog: http://meh.ixokai.io/



signature.asc
Description: OpenPGP digital signature
-- 
http://mail.python.org/mailman/listinfo/python-list


Adding modified methods from another class without subclassing

2011-08-21 Thread John O'Hagan
I have a class like this:

class MySeq():
def __init__(self, *seq, c=12):
self.__c = c
self.__pc = sorted(set([i % __c for i in seq]))
self.order = ([[self.__pc.index(i % __c), i // __c] for i in seq])
#other calculated attributes

@property
def pitches(self):
return [self.__pc[i[0]] + i[1] * self.__c for i in self.order]

#other methods

The "pitches" attribute initially reconstructs the "seq" arguments but can be 
modified by writing to the "order" attribute. 

The "pitches" attribute represents the instances and as such I found myself 
adding a lot of methods like:


def __getitem__(self, index):
return self.pitches[index]

def __len__(self):
return len(self.pitches)

def __iter__(self):
return iter(self.pitches)   

def __repr__(self):
return str(self.pitches)

and so on, and calling a lot of list methods on the "pitches" attribute of 
MySeq instances elsewhere. I thought of making MySeq a subclass of list with 
"pitches" as its contents, but then I would have had to override a lot of 
methods case-by-case, for example to ensure that any alterations to "pitches" 
were reflected in the other calculated attributes.

So I wrote this function which takes a method, modifies it to apply to an 
instance attribute, and takes care of any quirks:

def listmeth_to_attribute(meth, attr):
def new_meth(inst, *args):
#ensure comparison operators work:
args = [getattr(i, attr)  if isinstance(i, inst.__class__)
else i for i in args]
reference = getattr(inst, attr)
test = reference[:] 
result = meth(test, *args) 
#ensure instance is reinitialised
#if attribute has been changed: 
if test != reference:
inst.__init__(*test)
#ensure slices are of same class
if isinstance(result, meth.__objclass__):
result = inst.__class__(*result)
return result
return new_meth

and this decorator to apply this function to all the list methods and add them 
to MySeq:

def add_mod_methods(source_cls, modfunc, modfunc_args, *overrides):
"""Overides = any methods in target to override from source"""
def decorator(target_cls):
for name, meth in vars(source_cls).items():
if name not in dir(target_cls) or name in overrides:
setattr(target_cls, name, modfunc(meth, *modfunc_args))
return target_cls
return decorator

a kind of DIY single inheritance, used like this:

@add_mod_methods(list, listmeth_to_attribute, ('pitches',), '__repr__')
class MySeq():
.

Now I can call list methods transparently on MySeq instances, like subclassing 
but without all the overriding. If this works it will simplify a lot of code in 
my project. 

But the fact that I haven't seen this approach before increases the likelihood 
it may not be a good idea. I can almost hear the screams of "No, don't do 
that!" or the sound of me slapping my forehead when someone says "Why don't you 
just...". So before I put it in, I'd appreciate any comments, warnings, 
criticisms, alternatives etc.. 

Regards,

John
-- 
http://mail.python.org/mailman/listinfo/python-list


Re: Adding modified methods from another class without subclassing

2011-08-21 Thread Steven D'Aprano
On Mon, 22 Aug 2011 03:04 pm John O'Hagan wrote:

> The "pitches" attribute represents the instances and as such I found
> myself adding a lot of methods like:
> 
> def __getitem__(self, index):
> return self.pitches[index]
> 
> def __len__(self):
> return len(self.pitches)


Looks like a call for (semi-)automatic delegation!

Try something like this:


# Untested
class MySeq(object):
methods_to_delegate = ('__getitem__', '__len__', ...)
pitches = ...  # make sure pitches is defined
def __getattr__(self, name):
if name in self.__class__.methods_to_delegate:
return getattr(self.pitches, name)
return super(MySeq, object).__getattr__(self, name)
# will likely raise AttributeError




> But the fact that I haven't seen this approach before increases the
> likelihood it may not be a good idea. I can almost hear the screams of
> "No, don't do that!" 

The general technique is called delegation and is perfectly legitimate.

http://c2.com/cgi/wiki?WhatIsDelegation
http://en.wikipedia.org/wiki/Delegation_(programming)



-- 
Steven

-- 
http://mail.python.org/mailman/listinfo/python-list


Re: relative speed of incremention syntaxes (or "i=i+1" VS "i+=1")

2011-08-21 Thread Seebs
On 2011-08-21, Andreas L?scher  wrote:
> Am Sonntag, den 21.08.2011, 14:52 -0400 schrieb Roy Smith:
>> In article ,
>>  Christian Heimes  wrote:
>> > I don't think that's the reason. Modern compiles turn a switch statement
>> > into a jump or branch table rather than a linear search like chained
>> > elif statements.

>> This is true even for very small values of "modern".  I remember the 
>> Unix v6 C compiler (circa 1977) was able to do this.

> What is the difference in speed between a jump table that is searched
> from top to bottom

A jump table isn't searched, it's jumped into.  You don't look through the
table for a matching element, you grab the Nth element of the table.

-s
-- 
Copyright 2011, all wrongs reversed.  Peter Seebach / usenet-nos...@seebs.net
http://www.seebs.net/log/ <-- lawsuits, religion, and funny pictures
http://en.wikipedia.org/wiki/Fair_Game_(Scientology) <-- get educated!
I am not speaking for my employer, although they do rent some of my opinions.
-- 
http://mail.python.org/mailman/listinfo/python-list


Order of addresses returned by socket.gethostbyname_ex()

2011-08-21 Thread Tomas Lidén
In what order are the addresses returned by socket.gethostbyname_ex()?

We know that gethostbyname() is indeterministic but hope that
gethostbyname_ex() has a specified order.

Best regards,
Tomas
-- 
http://mail.python.org/mailman/listinfo/python-list