mirror of
https://github.com/kovidgoyal/calibre.git
synced 2026-05-13 10:42:17 -04:00
Replace some uses of plugins dict
This commit is contained in:
@@ -186,6 +186,7 @@ def _init():
|
||||
for names in dispatch
|
||||
)
|
||||
|
||||
|
||||
_init()
|
||||
|
||||
|
||||
@@ -196,6 +197,7 @@ def _unicode_replace(match, int=int, unichr=unichr, maxunicode=sys.maxunicode):
|
||||
else:
|
||||
return '\N{REPLACEMENT CHARACTER}' # U+FFFD
|
||||
|
||||
|
||||
UNICODE_UNESCAPE = functools.partial(
|
||||
re.compile(COMPILED_MACROS['unicode'], re.I).sub,
|
||||
_unicode_replace)
|
||||
@@ -204,12 +206,15 @@ NEWLINE_UNESCAPE = functools.partial(
|
||||
re.compile(r'()\\' + COMPILED_MACROS['nl']).sub,
|
||||
'')
|
||||
|
||||
|
||||
SIMPLE_UNESCAPE = functools.partial(
|
||||
re.compile(r'\\(%s)' % COMPILED_MACROS['simple_escape'] , re.I).sub,
|
||||
re.compile(r'\\(%s)' % COMPILED_MACROS['simple_escape'], re.I).sub,
|
||||
# Same as r'\1', but faster on CPython
|
||||
operator.methodcaller('group', 1))
|
||||
|
||||
FIND_NEWLINES = lambda x : list(re.compile(COMPILED_MACROS['nl']).finditer(x))
|
||||
|
||||
def FIND_NEWLINES(x):
|
||||
return list(re.compile(COMPILED_MACROS['nl']).finditer(x))
|
||||
|
||||
|
||||
class Token(object):
|
||||
@@ -439,11 +444,12 @@ class TokenList(list):
|
||||
"""
|
||||
return ''.join(token.as_css() for token in self)
|
||||
|
||||
|
||||
def load_c_tokenizer():
|
||||
from calibre.constants import plugins
|
||||
tokenizer, err = plugins['tokenizer']
|
||||
if err:
|
||||
raise RuntimeError('Failed to load module tokenizer: %s' % err)
|
||||
from calibre_extensions import tokenizer
|
||||
tokens = list(':;(){}[]') + ['DELIM', 'INTEGER', 'STRING']
|
||||
tokenizer.init(COMPILED_TOKEN_REGEXPS, UNICODE_UNESCAPE, NEWLINE_UNESCAPE, SIMPLE_UNESCAPE, FIND_NEWLINES, TOKEN_DISPATCH, COMPILED_TOKEN_INDEXES, *tokens)
|
||||
tokenizer.init(
|
||||
COMPILED_TOKEN_REGEXPS, UNICODE_UNESCAPE, NEWLINE_UNESCAPE,
|
||||
SIMPLE_UNESCAPE, FIND_NEWLINES, TOKEN_DISPATCH, COMPILED_TOKEN_INDEXES,
|
||||
*tokens)
|
||||
return tokenizer
|
||||
|
||||
Reference in New Issue
Block a user