Antoine Pitrou <pit...@free.fr> added the comment: > OK' I'm back to being 100% on the side of rejecting both of these > changes. ASCII is not unocode, it is bytes. You can decode it to > unicode but it is not unicode. Those transformations operate bytes to > bytes, not bytes to unicode.
ASCII is just a subset of the unicode character set. > We made the bytes unicode separation to avoid the problem where you > have a working program that unexpectedly gets non ASCII input and > blows up with a unicode error. How is blowing up with a unicode error worse than blowing up with a ValueError? Both indicate wrong input. At worse the code could catch UnicodeError and re-raise it as ValueError, but I don't see the point. > The programer should have to explicitly encode to ASCII if they are > inadvisedly workimg with it in a string as part of a wire protocol > (why else would they be using these transforms). Inadvisedly? There are many situations where you can have base64 data in some unicode strings. ---------- _______________________________________ Python tracker <rep...@bugs.python.org> <http://bugs.python.org/issue13641> _______________________________________ _______________________________________________ Python-bugs-list mailing list Unsubscribe: http://mail.python.org/mailman/options/python-bugs-list/archive%40mail-archive.com