diff --git a/emmet/abbreviation/tokenizer/__init__.py b/emmet/abbreviation/tokenizer/__init__.py index 307e20c..f7b5ae3 100644 --- a/emmet/abbreviation/tokenizer/__init__.py +++ b/emmet/abbreviation/tokenizer/__init__.py @@ -3,15 +3,15 @@ from .utils import Chars, escaped from . import tokens -OPERATOR_TYPES = dict([ - (Chars.Child, 'child'), - (Chars.Sibling, 'sibling'), - (Chars.Climb, 'climb'), - (Chars.Dot, 'class'), - (Chars.Hash, 'id'), - (Chars.Slash, 'close'), - (Chars.Equals, 'equal') -]) +OPERATOR_TYPES = { + Chars.Child: 'child', + Chars.Sibling: 'sibling', + Chars.Climb: 'climb', + Chars.Dot: 'class', + Chars.Hash: 'id', + Chars.Slash: 'close', + Chars.Equals: 'equal' +} def tokenize(source: str):