diff options
author | Matthäus G. Chajdas <dev@anteru.net> | 2020-09-08 20:33:25 +0200 |
---|---|---|
committer | Matthäus G. Chajdas <dev@anteru.net> | 2020-09-08 20:33:25 +0200 |
commit | 203ef1eff6daebab6f95b0b49e6e6a58168073fb (patch) | |
tree | 7defa199f48a34787f980b6400d8bbaa9380039a /tests/test_grammar_notation.py | |
parent | e09d4e0cf23d7c6069ddc690942ceb4cd23fd556 (diff) | |
parent | b2c91c70ee536b0472100d1273818f8bb45529fe (diff) | |
download | pygments-git-bug/angular-html.tar.gz |
Merge branch 'master' into bug/angular-htmlbug/angular-html
# Conflicts:
# tests/test_shell.py
Diffstat (limited to 'tests/test_grammar_notation.py')
-rw-r--r-- | tests/test_grammar_notation.py | 94 |
1 files changed, 47 insertions, 47 deletions
diff --git a/tests/test_grammar_notation.py b/tests/test_grammar_notation.py index 0d7e6865..9db46e62 100644 --- a/tests/test_grammar_notation.py +++ b/tests/test_grammar_notation.py @@ -19,21 +19,21 @@ def lexer_peg(): def test_peg_basic(lexer_peg): - fragment = u'rule<-("terminal"/nonterminal/[cls])*\n' + fragment = 'rule<-("terminal"/nonterminal/[cls])*\n' tokens = [ - (Token.Name.Class, u'rule'), - (Token.Operator, u'<-'), - (Token.Punctuation, u'('), - (Token.String.Double, u'"terminal"'), - (Token.Operator, u'/'), - (Token.Name.Class, u'nonterminal'), - (Token.Operator, u'/'), - (Token.Punctuation, u'['), - (Token.String, u'cls'), - (Token.Punctuation, u']'), - (Token.Punctuation, u')'), - (Token.Operator, u'*'), - (Token.Text, u'\n'), + (Token.Name.Class, 'rule'), + (Token.Operator, '<-'), + (Token.Punctuation, '('), + (Token.String.Double, '"terminal"'), + (Token.Operator, '/'), + (Token.Name.Class, 'nonterminal'), + (Token.Operator, '/'), + (Token.Punctuation, '['), + (Token.String, 'cls'), + (Token.Punctuation, ']'), + (Token.Punctuation, ')'), + (Token.Operator, '*'), + (Token.Text, '\n'), ] assert list(lexer_peg.get_tokens(fragment)) == tokens @@ -42,31 +42,31 @@ def test_peg_operators(lexer_peg): # see for example: # - https://github.com/gvanrossum/pegen # - https://nim-lang.org/docs/pegs.html - fragment = u"rule = 'a' | 'b'\n" + fragment = "rule = 'a' | 'b'\n" tokens = [ - (Token.Name.Class, u'rule'), - (Token.Text, u' '), - (Token.Operator, u'='), - (Token.Text, u' '), - (Token.String.Single, u"'a'"), - (Token.Text, u' '), - (Token.Operator, u'|'), - (Token.Text, u' '), - (Token.String.Single, u"'b'"), - (Token.Text, u'\n'), + (Token.Name.Class, 'rule'), + (Token.Text, ' '), + (Token.Operator, '='), + (Token.Text, ' '), + (Token.String.Single, "'a'"), + (Token.Text, ' '), + (Token.Operator, '|'), + (Token.Text, ' '), + (Token.String.Single, "'b'"), + (Token.Text, '\n'), ] assert list(lexer_peg.get_tokens(fragment)) == tokens - fragment = u"rule: 'a' ~ 'b'\n" + fragment = "rule: 'a' ~ 'b'\n" tokens = [ - (Token.Name.Class, u'rule'), - (Token.Operator, u':'), - (Token.Text, u' '), - (Token.String.Single, u"'a'"), - (Token.Text, u' '), - (Token.Operator, u'~'), - (Token.Text, u' '), - (Token.String.Single, u"'b'"), - (Token.Text, u'\n'), + (Token.Name.Class, 'rule'), + (Token.Operator, ':'), + (Token.Text, ' '), + (Token.String.Single, "'a'"), + (Token.Text, ' '), + (Token.Operator, '~'), + (Token.Text, ' '), + (Token.String.Single, "'b'"), + (Token.Text, '\n'), ] assert list(lexer_peg.get_tokens(fragment)) == tokens @@ -76,19 +76,19 @@ def test_peg_modified_strings(lexer_peg): # - http://textx.github.io/Arpeggio/ # - https://nim-lang.org/docs/pegs.html # - https://github.com/erikrose/parsimonious - fragment = u'~"regex" i"insensitive" "multimod"ilx ("not modified")\n' + fragment = '~"regex" i"insensitive" "multimod"ilx ("not modified")\n' tokens = [ # can't handle parsimonious-style regex while ~ is a cut operator - (Token.Operator, u'~'), - (Token.String.Double, u'"regex"'), - (Token.Text, u' '), - (Token.String.Double, u'i"insensitive"'), - (Token.Text, u' '), - (Token.String.Double, u'"multimod"ilx'), - (Token.Text, u' '), - (Token.Punctuation, u'('), - (Token.String.Double, u'"not modified"'), - (Token.Punctuation, u')'), - (Token.Text, u'\n'), + (Token.Operator, '~'), + (Token.String.Double, '"regex"'), + (Token.Text, ' '), + (Token.String.Double, 'i"insensitive"'), + (Token.Text, ' '), + (Token.String.Double, '"multimod"ilx'), + (Token.Text, ' '), + (Token.Punctuation, '('), + (Token.String.Double, '"not modified"'), + (Token.Punctuation, ')'), + (Token.Text, '\n'), ] assert list(lexer_peg.get_tokens(fragment)) == tokens |