This issue tracker has been migrated to GitHub ,
and is currently read-only.
For more information,
see the GitHub FAQs in the Python's Developer Guide.
Created on 2011年08月11日 19:18 by tchrist, last changed 2022年04月11日 14:57 by admin.
| Files | ||||
|---|---|---|---|---|
| File name | Uploaded | Description | Edit | |
| alnum.python | tchrist, 2011年08月11日 19:18 | test case showing conformance bugs in Python re lib when handling Unicode | ||
| Messages (14) | |||
|---|---|---|---|
| msg141920 - (view) | Author: Tom Christiansen (tchrist) | Date: 2011年08月11日 19:18 | |
You cannot use Python's lib re for handling Unicode regular expressions because it violates the standard set out for the same in UTS#18 on Unicode Regular Expressions in RL1.2a on compatibility properties. What \w is allowed to match is clearly explained there, but Python has its own idea. Because it is in clear violation of the standard, it is misleading and wrong for Python to claim that the re.UNICODE flag makes \w and friends match Unicode. Here are the failed test cases when the attached file is run under v3.2; there are further failures when run under v2.7. FAIL lib re found non alphanumeric string café FAIL lib re found non alphanumeric string K FAIL lib re found non alphanumeric string ͅ FAIL lib re found non alphanumeric string ְ FAIL lib re found non alphanumeric string 0 FAIL lib re found non alphanumeric string 𐍁 FAIL lib re found non alphanumeric string Unicode FAIL lib re found non alphanumeric string 𐐔𐐯𐑅𐐨𐑉𐐯𐐻 FAIL lib re found non alphanumeric string connector‿punctuation FAIL lib re found non alphanumeric string Ὰͅ_Στο_Διάολο FAIL lib re found non alphanumeric string 𐌰𐍄𐍄𐌰‿𐌿𐌽𐍃𐌰𐍂‿𐌸𐌿‿𐌹𐌽‿𐌷𐌹𐌼𐌹𐌽𐌰𐌼 FAIL lib re found all alphanumeric string 123 FAIL lib re found all alphanumeric string 123 FAIL lib re found all alphanumeric string 1⁄41⁄23⁄4 FAIL lib re found all alphanumeric string (3) Note that Matthew Barnett's regex lib for Python handles all of these cases in comformance with The Unicode Standard. |
|||
| msg141993 - (view) | Author: Terry J. Reedy (terry.reedy) * (Python committer) | Date: 2011年08月12日 22:46 | |
However desireable it would be, I do not believe there is any claim in the manual that the re module follows the evolving Unicode consortium r.e. standard. If I understand, you are saying that this statement in the doc, "Matches Unicode word characters;" is not now correct and should be revised. Was it once correct? Could we add "by an older definition of 'word' character"? There has been some discussion of adding regex to the stdlib, possibly as a replacement for re. You posts indicate that regex is more improved then some realized, and hence has more incompatibilities that we realized, and hence is less suitable as a strictly backwards-compatible replacement. So I think it needs to be looked at as a parallel addition. I do not know Mathew's current position on the subject. |
|||
| msg142001 - (view) | Author: Tom Christiansen (tchrist) | Date: 2011年08月13日 00:18 | |
> Terry J. Reedy <tjreedy@udel.edu> added the comment: > However desireable it would be, I do not believe there is any claim in the = > manual that the re module follows the evolving Unicode consortium r.e. stan= My from the hip thought is that if re cannot be fixed to follow the Unicode Standard, it should be deprecated in favor of code that can if such is available, because you cannot process Unicode text with regular expressions otherwise. > dard. If I understand, you are saying that this statement in the doc, "Matc= > hes Unicode word characters;" is not now correct and should be revised. Was= > it once correct? Could we add "by an older definition of 'word' character"= > ? Yes, your hunch is exactly correct. They once had a lesser definition that they have now. It is very very old. I had to track this down for Java once. There is some discussion of a "word_character class" at least as far back as tr18v3 from back in 1998. http://www.unicode.org/reports/tr18/tr18-3.html By the time tr18v5 rolled around just a year later in 1999, the overall document has changed substantially, and you can clearly see its current shape there. Word characters are supposed to include all code points with the Alphabetic property, for example. http://www.unicode.org/reports/tr18/tr18-5.html However, the word "alphabetic" has *never* been synonymous in Unicode with \p{gc=Lu} \p{gc=Ll} \p{gc=Lt} \p{gc=Lm} \p{gc=Lo} as many people incorrectly assume, nor certainly to \p{gc=Lu} \p{gc=Ll} \p{gc=Lt} let alone to \p{gc=Lu} \p{gc=Ll} Rather, it has since its creation included code points that are not letters, such as all GC=Nl and also certain GC=So code points. And, notoriously, U+0345. Indeed it is here I first noticed that that Python had already broken with the Standard, because U+0345 COMBINING GREEK YPOGEGRAMMENI is GC=Mn, but Alphabetic=True, yet I have shown that Python's title method is messing up there. I wouldn't spend too much in archaeological digs, though, because lots of stuff has changed since the less millennium. It was in tr18v7 from 2003-05 that we hit paydirt, because this is when the famous Annex C of RL1.2a fame first appeared: http://www.unicode.org/reports/tr18/tr18-7.html#Compatibility_Properties Notice how it defines \w to be nothing more than \p{alpha}, \p{digit}, and \p{gc=Pc}. It does not yet contain the requirement that all Marks be counted as part of the word, just the few that are alphas -- which the U+0345 counts for, since it has an uppercase map of a capital iota! That particular change did not occur until tr18v8 in 2003-08, barely a scant three months later. http://www.unicode.org/reports/tr18/tr18-8.html#Compatibility_Properties Now at last we see word characters defined in the modern way that we have become used to. They must match any of: \p{alpha} \p{gc=Mark} \p{digit} \p{gc=Connector_Punctuation} BTW, Python is matching all of \p{GC=N} meaning \p{GC=Nd} \p{GC=Nl} \p{GC=No} instead of the required \p{GC=Nd} which is a synonym for \p{digit}. I don't know had that happened, because \w has never included all number code points in Unicode, only the decimal number ones. That all goes to show why, when citing conformance to some aspect of The Unicode Standard, one must be exceedingly careful just how one does so! The Unicode Consortium recognizes this is an issue, and I am pretty sure I can hear it in your own subtext as well. Kindly bear with and forgive me for momentarily sounding like a standard lawyer. I do this because to show not just why it is important to get references to the Unicode Standard correct, but indeed, how to do so. After I have given the formal requirements, I will then produce illustrations of various purported claims, some of which meet the citation requirements, and others which do not. ======================================================================= To begin with, there is an entire technical report on conformance. It includes: http://unicode.org/reports/tr33/ The Unicode Standard [Unicode] is a very large and complex standard. Because of this complexity, and because of the nature and role of the standard, it is often rather difficult to determine, in any particular case, just exactly what conformance to the Unicode Standard means. ... Conformance claims must be specific to versions of the Unicode Standard, but the level of specificity needed for a claim may vary according to the nature of the particular conformance claim. Some standards developed by the Unicode Consortium require separate conformance to a specific version (or later), of the Unicode Standard. This version is sometimes called the base version. In such cases, the version of the standard and the version of the Unicode Standard to which the conformance claim refers must be compatible. However, you don't need to read tr33, really, because *the* most important thing bits about conformance are to be found on pp. 57-58 of Chapter 3 of the published Unicode Standard: http://www.unicode.org/versions/Unicode6.0.0/ch03.pdf References to the Unicode Standard The documents associated with the major, minor, and update versions are called the major reference, minor reference, and update reference, respectively. For example, consider Uni- code Version 3.1.1. The major reference for that version is The Unicode Standard, Version 3.0 (ISBN 0-201-61633-5). The minor reference is Unicode Standard Annex #27, "The Uni- code Standard, Version 3.1." The update reference is Unicode Version 3.1.1. The exact list of contributory files, Unicode Standard Annexes, and Unicode Character Database files can be found at Enumerated Version 3.1.1. The reference for this version, Version 6.0.0, of the Unicode Standard, is The Unicode Consortium. The Unicode Standard, Version 6.0.0, defined by: The Unicode Standard, Version 6.0 (Mountain View, CA: The Uni- code Consortium, 2011. ISBN 978-1-936213-01-6) References to an update (or minor version prior to Version 5.2.0) include a reference to both the major version and the documents modifying it. For the standard citation format for other versions of the Unicode Standard, see "Versions" in Section B.6, Other Unicode Online Resources. Precision in Version Citation Because Unicode has an open repertoire with relatively frequent updates, it is important not to over-specify the version number. Wherever the precise behavior of all Unicode char- acters needs to be cited, the full three-field version number should be used, as in the first example below. However, trailing zeros are often omitted, as in the second example. In such a case, writing 3.1 is in all respects equivalent to writing 3.1.0. 1. The Unicode Standard, Version 3.1.1 2. The Unicode Standard, Version 3.1 3. The Unicode Standard, Version 3.0 or later 4. The Unicode Standard Where some basic level of content is all that is important, phrasing such as in the third example can be used. Where the important information is simply the overall architecture and semantics of the Unicode Standard, the version can be omitted entirely, as in example 4. References to Unicode Character Properties Properties and property values have defined names and abbreviations, such as Property: General_Category (gc) Property Value: Uppercase_Letter (Lu) To reference a given property and property value, these aliases are used, as in this example: The property value Uppercase_Letter from the General_Category prop- erty, as specified in Version 6.0.0 of the Unicode Standard. Then cite that version of the standard, using the standard citation format that is provided for each version of the Unicode Standard. When referencing multi-word properties or property values, it is permissible to omit the underscores in these aliases or to replace them by spaces. When referencing a Unicode character property, it is customary to prepend the word "Uni- code" to the name of the property, unless it is clear from context that the Unicode Standard is the source of the specification. References to Unicode Algorithms A reference to a Unicode algorithm must specify the name of the algorithm or its abbrevia- tion, followed by the version of the Unicode Standard, as in this example: The Unicode Bidirectional Algorithm, as specified in Version 6.0.0 of the Unicode Standard. See Unicode Standard Annex #9, "Unicode Bidirectional Algorithm," (http://www.unicode.org/reports/tr9/tr9-23.html) ======================================================================= Now for some concrete citation examples, both correct and dubious. In the JDK7 documentation for on the Character class we find: Character information is based on the Unicode Standard, version 6.0.0. That one is a perfectly good conformance citation, even if there seems a bit of wiggle in "is based on", but no matter. It is short and does everything it needs to. However, in the JDK7 documentation for the Pattern class we somewhat problematically find: Unicode support This class is in conformance with Level 1 of Unicode Technical Standard #18: Unicode Regular Expression, plus RL2.1 Canonical Equivalents. And similarly, in the JDK7 documentation for the Normalizer class we find: This class provides the method normalize which transforms Unicode text into an equivalent composed or decomposed form, allowing for easier sorting and searching of text. The normalize method supports the standard normalization forms described in Unicode Standard Annex #15 — Unicode Normalization Forms. The problem with those second two Java refs is that they to my reading appear to be in technical violation, for they give neither a full version number nor a date of publication. You *have* to give one or the other, or both. Java got themselves into a heap of trouble (so to speak) over this once before because it turned out that the version of the document they were actually in conformance with was quite literally from the previous millennium!! That's why you need to give versions and publication dates. Here are some other citations. First, from the perldelta manpage that the Perl 5.14 release ships with: Perl comes with the Unicode 6.0 data base updated with Corrigendum #8 <http://www.unicode.org/versions/corrigendum8.html>, with one exception noted below. See <http://unicode.org/versions/Unicode6.0.0/> for details on the new release. Perl does not support any Unicode provisional properties, including the new ones for this release. That is quite complete, as it even includes the specific which corrigenda we follow and explains the matter of properties. Or this from the perlunicode manpage of that same release: Unicode Regular Expression Support Level The following list of Unicode supported features for regular expressions describes all features currently directly supported by core Perl. The references to "Level N" and the section numbers refer to the Unicode Technical Standard #18, "Unicode Regular Expressions", version 13, from August 2008. See all that? Notice how it gives the name of the document, its revision number, and its publication date. You don't have to do all that for the main Unicode release, but you really ought to when referring to individual technical reports BECAUSE THESE GET UPDATED ASYNCRONOUSLY. I would suggest you pick a version of tr18 that you conform to, and state which of its requirements you do and do not meet. However, I cannot find any version of tr18 that has existed during the present millennium that Python comes even close to meeting more than one or two requirements for. Given that, it may be better to no longer make any claims regarding Unicode at all. That seems like back-peddaling to me, not future-thinking. Matthew's regex module, however, does *almost* everything right that re does wrong. It may be that as with Java's io vs nio classes (and now heaven forbid nio2!), you actually can't fix the only module and must create a wholly new namespace. I cannot answer that. For RL1.2 proper, the first properties requirement, Java was only missing a few, so they went and added the missing properties. I strongly urge you to do so because you cannot handle Unicode without properties. Rl1.2 requires only 11 of them, so it isn't too hard. Matthew supports many many more. However, because the \w&c issues are bigger, Java addressed the tr18 RL1.2a issues differently, this time by creating a new compilation flag called UNICODE_CHARACTER_CLASSES (with corresponding embedded "(?U)" regex flag.) Truth be told, even Perl has secret pattern compilation flags to govern this sort of thing (ascii, locale, unicode), but we (well, I) hope you never have to use or even notice them. That too might be a route forward for Python, although I am not quite sure how much flexibility and control of your lexical scope you have. However, the "from __future_" imports suggest you may have enough to do something slick so that only people who ask for it get it, and also importantly that they get it all over the place so don't have to add an extra flag or u'...' or whatever every single time. This isn't something I've looked much into, however. Hope this clarifies things. --tom |
|||
| msg142030 - (view) | Author: Antoine Pitrou (pitrou) * (Python committer) | Date: 2011年08月13日 17:36 | |
> However, because the \w&c issues are bigger, Java addressed the tr18 RL1.2a > issues differently, this time by creating a new compilation flag called > UNICODE_CHARACTER_CLASSES (with corresponding embedded "(?U)" regex flag.) > > Truth be told, even Perl has secret pattern compilation flags to govern > this sort of thing (ascii, locale, unicode), but we (well, I) hope you > never have to use or even notice them. > > That too might be a route forward for Python, although I am not quite sure > how much flexibility and control of your lexical scope you have. However, > the "from __future_" imports suggest you may have enough to do something > slick so that only people who ask for it get it, and also importantly that > they get it all over the place so don't have to add an extra flag or u'...' > or whatever every single time. If the current behaviour is buggy or sub-optimal, I think we should simply fix it (which might be done by replacing "re" with "regex" if someone wants to shepherd its inclusion in the stdlib). By the way, thanks for the detailed explanations, Tom. |
|||
| msg142112 - (view) | Author: Ezio Melotti (ezio.melotti) * (Python committer) | Date: 2011年08月15日 10:30 | |
If the regex module works fine here, I think it's better to leave the re module alone and include the regex module in 3.3. |
|||
| msg143040 - (view) | Author: Guido van Rossum (gvanrossum) * (Python committer) | Date: 2011年08月26日 21:22 | |
Really? The re module cannot be salvaged and we should add regex but keep the (buggy) re? That does not make a lot of sense to me. I think it should just be fixed in the re module. Or the re module should be *replaced* by the code from the regex module (but renamed to re, and with certain backwards compatibilities restored, probably). But I really hope the re module (really: the _sre extension module) can be fixed. We should also make a habit in our docs of citing specific versions of the Unicode standard, and specific TR numbers and versions where they apply. (And hopefully we can supply URLs to the Unicode consortium's canonical copies of those documents.) |
|||
| msg143091 - (view) | Author: Ezio Melotti (ezio.melotti) * (Python committer) | Date: 2011年08月28日 06:26 | |
> Or the re module should be *replaced* by the code from the regex module > (but renamed to re, and with certain backwards compatibilities > restored, probably). This is what I meant. > But I really hope the re module (really: the _sre extension module) > can be fixed. Start fixing these issues from scratch doesn't make much sense IMHO. We could "extract" the fixes from regex and merge them in re, but then again it's probably easier to just replace the whole module. > We should also make a habit in our docs of citing specific versions > of the Unicode standard, and specific TR numbers and versions where > they apply. While this is a good thing it's not always doable. Usually someone reports a bug related to something specified in some standard and only that part gets fixed. Sometimes everything else is also updated to follow the whole standard, but often this happens incrementally, so we can't say, e.g., "the re module supports Unicode x.y" unless we go through the whole standard and fix/implements everything. |
|||
| msg143092 - (view) | Author: Ezio Melotti (ezio.melotti) * (Python committer) | Date: 2011年08月28日 06:37 | |
> But I really hope the re module (really: the _sre extension module) > can be fixed. If you mean on 2.7/3.2, then I guess we could extract the fixes from regex, but we have to see if it's doable and someone will have to do it. Also consider that the regex module is available for 2.7/3.2, so we could suggest the users to use it if they have problems with the re bugs (even if that means having an additional dependency). ISTM that current plan is: * replace re with regex (and rename it) on 3.3 and fix all these bugs; * leave 2.7 and 3.2 with the old re and its bugs; * let people use the external regex module on 2.7/3.2 if they need to. If this is not ok, maybe it should be discussed on python-dev. |
|||
| msg143109 - (view) | Author: Guido van Rossum (gvanrossum) * (Python committer) | Date: 2011年08月28日 17:22 | |
[me] >> But I really hope the re module (really: the _sre extension module) >> can be fixed. [Ezio] > Start fixing these issues from scratch doesn't make much sense IMHO. We could "extract" the fixes from regex and merge them in re, but then again it's probably easier to just replace the whole module. I have changed my mind at least half-way. I am open to having regex (with some changes, details TBD) replace re in 3.3. (I am not yet 100% convinced, but I'm not rejecting it as strongly as I was when I wrote that comment in this bug. See the ongoing python-dev discussion on this topic.) >> We should also make a habit in our docs of citing specific versions >> of the Unicode standard, and specific TR numbers and versions where >> they apply. > > While this is a good thing it's not always doable. Usually someone reports a bug related to something specified in some standard and only that part gets fixed. Sometimes everything else is also updated to follow the whole standard, but often this happens incrementally, so we can't say, e.g., "the re module supports Unicode x.y" unless we go through the whole standard and fix/implements everything. Hm. I think that for Unicode it may actually be important enough to be consistent in following the whole standard that we should attempt to be consistent and not just chase bug reports. Now, we may consciously decide not to implement a certain recommendation of the standard. E.g. I'm not going to require that IronPython or Jython have string objects that support O(1) indexing of code points, even (assuming PEP 393 gets accepted) CPython will have them. But these decisions should be made explicitly, and documented clearly. Ideally, we need a "Unicode czar" -- a core developer whose job it is to keep track of Python's compliance with various parts and versions of the Unicode standard and who can nudge other developers towards fixing bugs or implementing features, or update the documentation in case things don't get added. (I like Tom's approach to Java 1.7, where he submitted proposed doc fixes explaining the deviations from the standard. Perhaps a bit passive-aggressive, but it was effective. :-) |
|||
| msg143113 - (view) | Author: Ezio Melotti (ezio.melotti) * (Python committer) | Date: 2011年08月28日 17:58 | |
> Ideally, we need a "Unicode czar" -- a core developer whose job it is > to keep track of Python's compliance with various parts and versions > of the Unicode standard and who can nudge other developers towards > fixing bugs or implementing features, or update the documentation in > case things don't get added. We should first do a full review of the latest Unicode standard and see what's missing. I think there might be parts of older Unicode versions (even < Unicode 5) that are not yet implemented. Chapter 3 is a good place where to start, but I'm not sure that's enough -- there are a few TRs that should be considered as well. If we manage to catch up with Unicode 6, then it shouldn't be too difficult to review the changes that every new version will introduce and open an issue for each (or a single issue if the changes are limited). FWIW I'm planning to look at the conformance of the UTF codecs and fix them (if necessary) whenever I'll have time. |
|||
| msg144663 - (view) | Author: Ezio Melotti (ezio.melotti) * (Python committer) | Date: 2011年09月30日 04:05 | |
The failing re tests after PEP 393 are: FAIL lib re found non alphanumeric string 'cafe' FAIL lib re found non alphanumeric string 'K' FAIL lib re found non alphanumeric string '' FAIL lib re found non alphanumeric string '' FAIL lib re found non alphanumeric string 'connector‿punctuation' FAIL lib re found non alphanumeric string 'Ὰ_Στο_Διάολο' FAIL lib re found non alphanumeric string '𐌰𐍄𐍄𐌰‿𐌿𐌽𐍃𐌰𐍂‿𐌸𐌿‿𐌹𐌽‿𐌷𐌹𐌼𐌹𐌽𐌰𐌼' FAIL lib re found all alphanumeric string '123' FAIL lib re found all alphanumeric string '123' FAIL lib re found all alphanumeric string '1⁄41⁄23⁄4' FAIL lib re found all alphanumeric string '(3)' |
|||
| msg313850 - (view) | Author: Terry J. Reedy (terry.reedy) * (Python committer) | Date: 2018年03月15日 00:33 | |
Whatever I may have said before, I favor supporting the Unicode standard for \w, which is related to the standard for identifiers. This is one of 2 issues about \w being defined too narrowly. I am somewhat arbitrarily closing #1693050 as a duplicate of this (fewer digits ;-). There are 3 issues about tokenize.tokenize failing on valid identifiers, defined as \w sequences whose first char is an identifier itself (and therefore a start char). In msg313814 of #32987, Serhiy indicates which start and continue identifier characters are matched by \W for re and regex. I am leaving #24194 open as the tokenizer name issue. |
|||
| msg334508 - (view) | Author: Henry S. Thompson (HThompson) | Date: 2019年01月29日 10:10 | |
This issue is also implicated in a failure of isalpha and friends.
Easy way to see this is to compare
>>> isalpha('İ')
True
>>> isalpha('İ'.lower())
False
This results from the use of a combining character to encode lower-case Turkish dotted i:
>>> len('İ'.lower())
2
>>> unicodedata.category('İ'.lower()[1])
'Mn'
|
|||
| msg361105 - (view) | Author: Henry S. Thompson (HThompson) | Date: 2020年01月31日 13:31 | |
[One year and 2 days later... :-[ Is this fixed in 3.9? If not, the Versions list above should be updated. The failure of lower() to preserve 'alpha-ness' is a serious bug, it causes significant failures in e.g. Turkish NLP, and it's _not_ just a failure of the documentation! Please can we move this to category Unicode and get at least this aspect of the problem fixed? Should I raise a separate issue on isalpha() etc.? |
|||
| History | |||
|---|---|---|---|
| Date | User | Action | Args |
| 2022年04月11日 14:57:20 | admin | set | github: 56940 |
| 2021年05月26日 18:46:48 | pitrou | set | stage: test needed -> needs patch versions: + Python 3.11, - Python 3.6, Python 3.7, Python 3.8 |
| 2020年02月03日 08:34:10 | vstinner | set | nosy:
- vstinner |
| 2020年01月31日 14:05:49 | terry.reedy | set | assignee: docs@python -> components: + Unicode, - Documentation nosy: + lemburg, benjamin.peterson, serhiy.storchaka |
| 2020年01月31日 13:31:30 | HThompson | set | messages: + msg361105 |
| 2019年09月07日 17:07:19 | JustinTArthur | set | nosy:
+ JustinTArthur |
| 2019年01月29日 10:10:15 | HThompson | set | nosy:
+ HThompson messages: + msg334508 |
| 2018年03月15日 00:33:15 | terry.reedy | set | stage: needs patch -> test needed messages: + msg313850 versions: + Python 3.6, Python 3.7, Python 3.8, - Python 2.7, Python 3.3, Python 3.4 |
| 2016年04月25日 06:08:17 | serhiy.storchaka | link | issue24194 dependencies |
| 2013年07月10日 19:11:02 | terry.reedy | set | versions: + Python 3.4, - Python 3.2 |
| 2011年09月30日 04:05:12 | ezio.melotti | set | messages: + msg144663 |
| 2011年08月28日 17:58:12 | ezio.melotti | set | messages: + msg143113 |
| 2011年08月28日 17:22:44 | gvanrossum | set | messages: + msg143109 |
| 2011年08月28日 06:37:44 | ezio.melotti | set | messages: + msg143092 |
| 2011年08月28日 06:26:11 | ezio.melotti | set | messages: + msg143091 |
| 2011年08月26日 21:22:26 | gvanrossum | set | nosy:
+ gvanrossum messages: + msg143040 |
| 2011年08月15日 10:30:03 | ezio.melotti | set | messages: + msg142112 |
| 2011年08月13日 17:36:36 | pitrou | set | messages: + msg142030 |
| 2011年08月13日 09:40:58 | pitrou | set | nosy:
+ vstinner |
| 2011年08月13日 00:57:20 | mrabarnett | set | nosy:
+ mrabarnett |
| 2011年08月13日 00:18:22 | tchrist | set | messages: + msg142001 |
| 2011年08月12日 22:46:48 | terry.reedy | set | assignee: docs@python components: + Documentation versions: + Python 3.2, Python 3.3 nosy: + terry.reedy, pitrou, docs@python messages: + msg141993 stage: needs patch |
| 2011年08月12日 18:03:10 | Arfrever | set | nosy:
+ Arfrever |
| 2011年08月12日 00:20:02 | ezio.melotti | set | nosy:
+ ezio.melotti |
| 2011年08月11日 19:18:31 | tchrist | create | |