98 lines
3.2 KiB
Python
98 lines
3.2 KiB
Python
# -*- coding: utf-8 -*-
|
|
"""
|
|
Pygments tests with example files
|
|
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
|
|
|
:copyright: Copyright 2006-2012 by the Pygments team, see AUTHORS.
|
|
:license: BSD, see LICENSE for details.
|
|
"""
|
|
|
|
import os
|
|
import pprint
|
|
import difflib
|
|
import cPickle as pickle
|
|
|
|
from pygments.lexers import get_lexer_for_filename, get_lexer_by_name
|
|
from pygments.token import Error
|
|
from pygments.util import ClassNotFound, b
|
|
|
|
STORE_OUTPUT = False
|
|
|
|
# generate methods
|
|
def test_example_files():
|
|
testdir = os.path.dirname(__file__)
|
|
outdir = os.path.join(testdir, 'examplefiles', 'output')
|
|
if STORE_OUTPUT and not os.path.isdir(outdir):
|
|
os.makedirs(outdir)
|
|
for fn in os.listdir(os.path.join(testdir, 'examplefiles')):
|
|
if fn.startswith('.') or fn.endswith('#'):
|
|
continue
|
|
|
|
absfn = os.path.join(testdir, 'examplefiles', fn)
|
|
if not os.path.isfile(absfn):
|
|
continue
|
|
outfn = os.path.join(outdir, fn)
|
|
|
|
try:
|
|
lx = get_lexer_for_filename(absfn)
|
|
except ClassNotFound:
|
|
if "_" not in fn:
|
|
raise AssertionError('file %r has no registered extension, '
|
|
'nor is of the form <lexer>_filename '
|
|
'for overriding, thus no lexer found.'
|
|
% fn)
|
|
try:
|
|
name, rest = fn.split("_", 1)
|
|
lx = get_lexer_by_name(name)
|
|
except ClassNotFound:
|
|
raise AssertionError('no lexer found for file %r' % fn)
|
|
yield check_lexer, lx, absfn, outfn
|
|
|
|
def check_lexer(lx, absfn, outfn):
|
|
fp = open(absfn, 'rb')
|
|
try:
|
|
text = fp.read()
|
|
finally:
|
|
fp.close()
|
|
text = text.replace(b('\r\n'), b('\n'))
|
|
text = text.strip(b('\n')) + b('\n')
|
|
try:
|
|
text = text.decode('utf-8')
|
|
except UnicodeError:
|
|
text = text.decode('latin1')
|
|
ntext = []
|
|
tokens = []
|
|
for type, val in lx.get_tokens(text):
|
|
ntext.append(val)
|
|
assert type != Error, \
|
|
'lexer %s generated error token for %s: %r at position %d' % \
|
|
(lx, absfn, val, len(u''.join(ntext)))
|
|
tokens.append((type, val))
|
|
if u''.join(ntext) != text:
|
|
print '\n'.join(difflib.unified_diff(u''.join(ntext).splitlines(),
|
|
text.splitlines()))
|
|
raise AssertionError('round trip failed for ' + absfn)
|
|
|
|
# check output against previous run if enabled
|
|
if STORE_OUTPUT:
|
|
# no previous output -- store it
|
|
if not os.path.isfile(outfn):
|
|
fp = open(outfn, 'wb')
|
|
try:
|
|
pickle.dump(tokens, fp)
|
|
finally:
|
|
fp.close()
|
|
return
|
|
# otherwise load it and compare
|
|
fp = open(outfn, 'rb')
|
|
try:
|
|
stored_tokens = pickle.load(fp)
|
|
finally:
|
|
fp.close()
|
|
if stored_tokens != tokens:
|
|
f1 = pprint.pformat(stored_tokens)
|
|
f2 = pprint.pformat(tokens)
|
|
print '\n'.join(difflib.unified_diff(f1.splitlines(),
|
|
f2.splitlines()))
|
|
assert False, absfn
|