summaryrefslogtreecommitdiffstats
path: root/Lib/HTMLParser.py
diff options
context:
space:
mode:
authorEzio Melotti <ezio.melotti@gmail.com>2012-02-13 14:28:54 (GMT)
committerEzio Melotti <ezio.melotti@gmail.com>2012-02-13 14:28:54 (GMT)
commitf117443cb8afa3b2d91b4fef861db17866d6b6df (patch)
tree4355ba11b5406084325b960dbd4dc464ad8fa202 /Lib/HTMLParser.py
parent4b92cc3f7924e455b7e41cf1a66034a44ede0cc0 (diff)
downloadcpython-f117443cb8afa3b2d91b4fef861db17866d6b6df.zip
cpython-f117443cb8afa3b2d91b4fef861db17866d6b6df.tar.gz
cpython-f117443cb8afa3b2d91b4fef861db17866d6b6df.tar.bz2
#13993: HTMLParser is now able to handle broken end tags.
Diffstat (limited to 'Lib/HTMLParser.py')
-rw-r--r--Lib/HTMLParser.py34
1 files changed, 26 insertions, 8 deletions
diff --git a/Lib/HTMLParser.py b/Lib/HTMLParser.py
index 516bc70..6cc9ff1 100644
--- a/Lib/HTMLParser.py
+++ b/Lib/HTMLParser.py
@@ -23,6 +23,9 @@ starttagopen = re.compile('<[a-zA-Z]')
piclose = re.compile('>')
commentclose = re.compile(r'--\s*>')
tagfind = re.compile('[a-zA-Z][-.a-zA-Z0-9:_]*')
+# see http://www.w3.org/TR/html5/tokenization.html#tag-open-state
+# and http://www.w3.org/TR/html5/tokenization.html#tag-name-state
+tagfind_tolerant = re.compile('[a-zA-Z][^\t\n\r\f />\x00]*')
attrfind = re.compile(
r'\s*((?<=[\'"\s])[^\s/>][^\s/=>]*)(\s*=+\s*'
@@ -243,7 +246,7 @@ class HTMLParser(markupbase.ParserBase):
# see http://www.w3.org/TR/html5/tokenization.html#bogus-comment-state
def parse_bogus_comment(self, i, report=1):
rawdata = self.rawdata
- if rawdata[i:i+2] != '<!':
+ if rawdata[i:i+2] not in ('<!', '</'):
self.error('unexpected call to parse_comment()')
pos = rawdata.find('>', i+2)
if pos == -1:
@@ -353,23 +356,38 @@ class HTMLParser(markupbase.ParserBase):
match = endendtag.search(rawdata, i+1) # >
if not match:
return -1
- j = match.end()
+ gtpos = match.end()
match = endtagfind.match(rawdata, i) # </ + tag + >
if not match:
if self.cdata_elem is not None:
- self.handle_data(rawdata[i:j])
- return j
- self.error("bad end tag: %r" % (rawdata[i:j],))
+ self.handle_data(rawdata[i:gtpos])
+ return gtpos
+ # find the name: w3.org/TR/html5/tokenization.html#tag-name-state
+ namematch = tagfind_tolerant.match(rawdata, i+2)
+ if not namematch:
+ # w3.org/TR/html5/tokenization.html#end-tag-open-state
+ if rawdata[i:i+3] == '</>':
+ return i+3
+ else:
+ return self.parse_bogus_comment(i)
+ tagname = namematch.group().lower()
+ # consume and ignore other stuff between the name and the >
+ # Note: this is not 100% correct, since we might have things like
+ # </tag attr=">">, but looking for > after tha name should cover
+ # most of the cases and is much simpler
+ gtpos = rawdata.find('>', namematch.end())
+ self.handle_endtag(tagname)
+ return gtpos+1
elem = match.group(1).lower() # script or style
if self.cdata_elem is not None:
if elem != self.cdata_elem:
- self.handle_data(rawdata[i:j])
- return j
+ self.handle_data(rawdata[i:gtpos])
+ return gtpos
self.handle_endtag(elem)
self.clear_cdata_mode()
- return j
+ return gtpos
# Overridable -- finish processing of start+end tag: <tag.../>
def handle_startendtag(self, tag, attrs):