summaryrefslogtreecommitdiff
path: root/tests/test_objectiveclexer.py
diff options
context:
space:
mode:
Diffstat (limited to 'tests/test_objectiveclexer.py')
-rw-r--r--tests/test_objectiveclexer.py84
1 files changed, 42 insertions, 42 deletions
diff --git a/tests/test_objectiveclexer.py b/tests/test_objectiveclexer.py
index 31f833cf..7264bad8 100644
--- a/tests/test_objectiveclexer.py
+++ b/tests/test_objectiveclexer.py
@@ -19,78 +19,78 @@ def lexer():
def test_literal_number_int(lexer):
- fragment = u'@(1);\n'
+ fragment = '@(1);\n'
expected = [
- (Token.Literal, u'@('),
- (Token.Literal.Number.Integer, u'1'),
- (Token.Literal, u')'),
- (Token.Punctuation, u';'),
- (Token.Text, u'\n'),
+ (Token.Literal, '@('),
+ (Token.Literal.Number.Integer, '1'),
+ (Token.Literal, ')'),
+ (Token.Punctuation, ';'),
+ (Token.Text, '\n'),
]
assert list(lexer.get_tokens(fragment)) == expected
def test_literal_number_expression(lexer):
- fragment = u'@(1+2);\n'
+ fragment = '@(1+2);\n'
expected = [
- (Token.Literal, u'@('),
- (Token.Literal.Number.Integer, u'1'),
- (Token.Operator, u'+'),
- (Token.Literal.Number.Integer, u'2'),
- (Token.Literal, u')'),
- (Token.Punctuation, u';'),
- (Token.Text, u'\n'),
+ (Token.Literal, '@('),
+ (Token.Literal.Number.Integer, '1'),
+ (Token.Operator, '+'),
+ (Token.Literal.Number.Integer, '2'),
+ (Token.Literal, ')'),
+ (Token.Punctuation, ';'),
+ (Token.Text, '\n'),
]
assert list(lexer.get_tokens(fragment)) == expected
def test_literal_number_nested_expression(lexer):
- fragment = u'@(1+(2+3));\n'
+ fragment = '@(1+(2+3));\n'
expected = [
- (Token.Literal, u'@('),
- (Token.Literal.Number.Integer, u'1'),
- (Token.Operator, u'+'),
- (Token.Punctuation, u'('),
- (Token.Literal.Number.Integer, u'2'),
- (Token.Operator, u'+'),
- (Token.Literal.Number.Integer, u'3'),
- (Token.Punctuation, u')'),
- (Token.Literal, u')'),
- (Token.Punctuation, u';'),
- (Token.Text, u'\n'),
+ (Token.Literal, '@('),
+ (Token.Literal.Number.Integer, '1'),
+ (Token.Operator, '+'),
+ (Token.Punctuation, '('),
+ (Token.Literal.Number.Integer, '2'),
+ (Token.Operator, '+'),
+ (Token.Literal.Number.Integer, '3'),
+ (Token.Punctuation, ')'),
+ (Token.Literal, ')'),
+ (Token.Punctuation, ';'),
+ (Token.Text, '\n'),
]
assert list(lexer.get_tokens(fragment)) == expected
def test_literal_number_bool(lexer):
- fragment = u'@NO;\n'
+ fragment = '@NO;\n'
expected = [
- (Token.Literal.Number, u'@NO'),
- (Token.Punctuation, u';'),
- (Token.Text, u'\n'),
+ (Token.Literal.Number, '@NO'),
+ (Token.Punctuation, ';'),
+ (Token.Text, '\n'),
]
assert list(lexer.get_tokens(fragment)) == expected
def test_literal_number_bool_expression(lexer):
- fragment = u'@(YES);\n'
+ fragment = '@(YES);\n'
expected = [
- (Token.Literal, u'@('),
- (Token.Name.Builtin, u'YES'),
- (Token.Literal, u')'),
- (Token.Punctuation, u';'),
- (Token.Text, u'\n'),
+ (Token.Literal, '@('),
+ (Token.Name.Builtin, 'YES'),
+ (Token.Literal, ')'),
+ (Token.Punctuation, ';'),
+ (Token.Text, '\n'),
]
assert list(lexer.get_tokens(fragment)) == expected
def test_module_import(lexer):
- fragment = u'@import ModuleA;\n'
+ fragment = '@import ModuleA;\n'
expected = [
- (Token.Keyword, u'@import'),
- (Token.Text, u' '),
- (Token.Name, u'ModuleA'),
- (Token.Punctuation, u';'),
- (Token.Text, u'\n'),
+ (Token.Keyword, '@import'),
+ (Token.Text, ' '),
+ (Token.Name, 'ModuleA'),
+ (Token.Punctuation, ';'),
+ (Token.Text, '\n'),
]
assert list(lexer.get_tokens(fragment)) == expected