summaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
-rw-r--r--Lib/test/test_ast.py8
-rw-r--r--Lib/test/test_exceptions.py2
-rw-r--r--Misc/NEWS.d/next/Core and Builtins/2021-10-07-21-26-44.bpo-45408.qUqzcd.rst2
-rw-r--r--Parser/pegen.c5
4 files changed, 15 insertions, 2 deletions
diff --git a/Lib/test/test_ast.py b/Lib/test/test_ast.py
index 326f3ab..39fc7e9 100644
--- a/Lib/test/test_ast.py
+++ b/Lib/test/test_ast.py
@@ -1044,6 +1044,14 @@ Module(
with self.assertRaisesRegex(ValueError, msg):
ast.literal_eval(node)
+ def test_literal_eval_syntax_errors(self):
+ msg = "unexpected character after line continuation character"
+ with self.assertRaisesRegex(SyntaxError, msg):
+ ast.literal_eval(r'''
+ \
+ (\
+ \ ''')
+
def test_bad_integer(self):
# issue13436: Bad error message with invalid numeric values
body = [ast.ImportFrom(module='time',
diff --git a/Lib/test/test_exceptions.py b/Lib/test/test_exceptions.py
index d04e5f5..4213dab 100644
--- a/Lib/test/test_exceptions.py
+++ b/Lib/test/test_exceptions.py
@@ -223,7 +223,7 @@ class ExceptionTests(unittest.TestCase):
check('x = "a', 1, 5)
check('lambda x: x = 2', 1, 1)
check('f{a + b + c}', 1, 2)
- check('[file for str(file) in []\n])', 2, 2)
+ check('[file for str(file) in []\n])', 1, 11)
check('a = « hello » « world »', 1, 5)
check('[\nfile\nfor str(file)\nin\n[]\n]', 3, 5)
check('[file for\n str(file) in []]', 2, 2)
diff --git a/Misc/NEWS.d/next/Core and Builtins/2021-10-07-21-26-44.bpo-45408.qUqzcd.rst b/Misc/NEWS.d/next/Core and Builtins/2021-10-07-21-26-44.bpo-45408.qUqzcd.rst
new file mode 100644
index 0000000..e4d4db9
--- /dev/null
+++ b/Misc/NEWS.d/next/Core and Builtins/2021-10-07-21-26-44.bpo-45408.qUqzcd.rst
@@ -0,0 +1,2 @@
+Fix a crash in the parser when reporting tokenizer errors that occur at the
+same time unclosed parentheses are detected. Patch by Pablo Galindo.
diff --git a/Parser/pegen.c b/Parser/pegen.c
index e20e926..1bb975d 100644
--- a/Parser/pegen.c
+++ b/Parser/pegen.c
@@ -1321,13 +1321,16 @@ _PyPegen_run_parser(Parser *p)
{
void *res = _PyPegen_parse(p);
if (res == NULL) {
+ if (PyErr_Occurred() && !PyErr_ExceptionMatches(PyExc_SyntaxError)) {
+ return NULL;
+ }
Token *last_token = p->tokens[p->fill - 1];
reset_parser_state(p);
_PyPegen_parse(p);
if (PyErr_Occurred()) {
// Prioritize tokenizer errors to custom syntax errors raised
// on the second phase only if the errors come from the parser.
- if (p->tok->done != E_ERROR && PyErr_ExceptionMatches(PyExc_SyntaxError)) {
+ if (p->tok->done == E_DONE && PyErr_ExceptionMatches(PyExc_SyntaxError)) {
_PyPegen_check_tokenizer_errors(p);
}
return NULL;