| Commit message (Collapse) | Author | Age | Files | Lines |
|
|
|
|
| |
emitted in the tokenize module (GH-104846). (#104850)
(cherry picked from commit c8cf9b42eb2bfbd4c3e708ec28d32430248a1d7a)
|
| |
|
|
|
|
|
|
|
|
|
|
|
| |
This commit replaces the Python implementation of the tokenize module with an implementation
that reuses the real C tokenizer via a private extension module. The tokenize module now implements
a compatibility layer that transforms tokens from the C tokenizer into Python tokenize tokens for backward
compatibility.
As the C tokenizer does not emit some tokens that the Python tokenizer provides (such as comments and non-semantic newlines), a new special mode has been added to the C tokenizer mode that currently is only used via
the extension module that exposes it to the Python layer. This new mode forces the C tokenizer to emit these new extra tokens and add the appropriate metadata that is needed to match the old Python implementation.
Co-authored-by: Pablo Galindo <pablogsal@gmail.com>
|
|
|
| |
Co-authored-by: Erlend E. Aasland <erlend.aasland@protonmail.com>
|
| |
|
| |
|
| |
|
| |
|
|
|
|
| |
Fix test_tabnanny on VxWorks: adjust ENOENT error message,
use os.strerror().
|
| |
|
|
Testing strategy: whitebox.
|