From f7e07b7b61be4befd5eaafce93aeb0238c884315 Mon Sep 17 00:00:00 2001 From: Andi Albrecht Date: Mon, 26 Oct 2015 20:03:58 +0100 Subject: Code cleanup in tests. --- tests/test_filters.py | 2 -- tests/test_format.py | 9 ++++++--- tests/test_functions.py | 1 - tests/test_grouping.py | 6 ++++-- tests/test_parse.py | 2 +- tests/test_tokenize.py | 54 ++++++++++++++++++++++++------------------------- 6 files changed, 38 insertions(+), 36 deletions(-) (limited to 'tests') diff --git a/tests/test_filters.py b/tests/test_filters.py index eb61604..925b0b6 100644 --- a/tests/test_filters.py +++ b/tests/test_filters.py @@ -5,7 +5,6 @@ Created on 24/03/2012 ''' import unittest -from sqlparse.compat import u from sqlparse.filters import StripWhitespace, Tokens2Unicode from sqlparse.lexer import tokenize @@ -75,5 +74,4 @@ LIMIT 1""" if __name__ == "__main__": - #import sys;sys.argv = ['', 'Test.testName'] unittest.main() diff --git a/tests/test_format.py b/tests/test_format.py index a105b1c..e8875dd 100644 --- a/tests/test_format.py +++ b/tests/test_format.py @@ -93,9 +93,12 @@ class TestFormat(TestCaseBase): # Because of the use of self.ndiffAssertEqual(f(s1), "SELECT some_column LIKE 'value\r'") - self.ndiffAssertEqual(f(s2), "SELECT some_column LIKE 'value\r'\nWHERE id = 1\n") - self.ndiffAssertEqual(f(s3), "SELECT some_column LIKE 'value\\'\r' WHERE id = 1\n") - self.ndiffAssertEqual(f(s4), "SELECT some_column LIKE 'value\\\\\\'\r' WHERE id = 1\n") + self.ndiffAssertEqual( + f(s2), "SELECT some_column LIKE 'value\r'\nWHERE id = 1\n") + self.ndiffAssertEqual( + f(s3), "SELECT some_column LIKE 'value\\'\r' WHERE id = 1\n") + self.ndiffAssertEqual( + f(s4), "SELECT some_column LIKE 'value\\\\\\'\r' WHERE id = 1\n") def test_outputformat(self): sql = 'select * from foo;' diff --git a/tests/test_functions.py b/tests/test_functions.py index 52e2ce7..425ab7f 100644 --- a/tests/test_functions.py +++ b/tests/test_functions.py @@ -160,5 +160,4 @@ class Test_IsType(Test_SQL): if __name__ == "__main__": - #import sys;sys.argv = ['', 'Test.testName'] main() diff --git a/tests/test_grouping.py b/tests/test_grouping.py index fa68ab2..e846176 100644 --- a/tests/test_grouping.py +++ b/tests/test_grouping.py @@ -21,7 +21,8 @@ class TestGrouping(TestCaseBase): self.assert_(isinstance(parsed.tokens[-1], sql.Identifier)) self.assertEqual(len(parsed.tokens[2].tokens), 5) self.assert_(isinstance(parsed.tokens[2].tokens[3], sql.Identifier)) - self.assert_(isinstance(parsed.tokens[2].tokens[3].tokens[0], sql.Parenthesis)) + self.assert_(isinstance(parsed.tokens[2].tokens[3].tokens[0], + sql.Parenthesis)) self.assertEqual(len(parsed.tokens[2].tokens[3].tokens), 3) def test_comments(self): @@ -147,7 +148,8 @@ class TestGrouping(TestCaseBase): s = 'select x from (select y from foo where bar = 1) z' p = sqlparse.parse(s)[0] self.ndiffAssertEqual(s, u(p)) - self.assertTrue(isinstance(p.tokens[-1].tokens[0].tokens[-2], sql.Where)) + self.assertTrue(isinstance(p.tokens[-1].tokens[0].tokens[-2], + sql.Where)) def test_typecast(self): s = 'select foo::integer from bar' diff --git a/tests/test_parse.py b/tests/test_parse.py index fb7b24b..2ea0f40 100644 --- a/tests/test_parse.py +++ b/tests/test_parse.py @@ -221,7 +221,7 @@ def test_sqlite_identifiers(): assert (len(p) == 1 and isinstance(p[0], sqlparse.sql.IdentifierList) and [id.get_name() for id in p[0].get_identifiers()] - == ['[col1]', '[col2]']) + == ['[col1]', '[col2]']) p = sqlparse.parse('[col1]+[col2]')[0] types = [tok.ttype for tok in p.flatten()] diff --git a/tests/test_tokenize.py b/tests/test_tokenize.py index 619aefd..02c3f25 100644 --- a/tests/test_tokenize.py +++ b/tests/test_tokenize.py @@ -1,6 +1,5 @@ # -*- coding: utf-8 -*- -import sys import types import unittest @@ -9,8 +8,8 @@ import pytest import sqlparse from sqlparse import lexer from sqlparse import sql +from sqlparse import tokens as T from sqlparse.compat import StringIO -from sqlparse.tokens import * class TestTokenize(unittest.TestCase): @@ -22,14 +21,14 @@ class TestTokenize(unittest.TestCase): tokens = list(stream) self.assertEqual(len(tokens), 8) self.assertEqual(len(tokens[0]), 2) - self.assertEqual(tokens[0], (Keyword.DML, u'select')) - self.assertEqual(tokens[-1], (Punctuation, u';')) + self.assertEqual(tokens[0], (T.Keyword.DML, u'select')) + self.assertEqual(tokens[-1], (T.Punctuation, u';')) def test_backticks(self): s = '`foo`.`bar`' tokens = list(lexer.tokenize(s)) self.assertEqual(len(tokens), 3) - self.assertEqual(tokens[0], (Name, u'`foo`')) + self.assertEqual(tokens[0], (T.Name, u'`foo`')) def test_linebreaks(self): # issue1 s = 'foo\nbar\n' @@ -49,28 +48,28 @@ class TestTokenize(unittest.TestCase): s = "create created_foo" tokens = list(lexer.tokenize(s)) self.assertEqual(len(tokens), 3) - self.assertEqual(tokens[0][0], Keyword.DDL) - self.assertEqual(tokens[2][0], Name) + self.assertEqual(tokens[0][0], T.Keyword.DDL) + self.assertEqual(tokens[2][0], T.Name) self.assertEqual(tokens[2][1], u'created_foo') s = "enddate" tokens = list(lexer.tokenize(s)) self.assertEqual(len(tokens), 1) - self.assertEqual(tokens[0][0], Name) + self.assertEqual(tokens[0][0], T.Name) s = "join_col" tokens = list(lexer.tokenize(s)) self.assertEqual(len(tokens), 1) - self.assertEqual(tokens[0][0], Name) + self.assertEqual(tokens[0][0], T.Name) s = "left join_col" tokens = list(lexer.tokenize(s)) self.assertEqual(len(tokens), 3) - self.assertEqual(tokens[2][0], Name) + self.assertEqual(tokens[2][0], T.Name) self.assertEqual(tokens[2][1], 'join_col') def test_negative_numbers(self): s = "values(-1)" tokens = list(lexer.tokenize(s)) self.assertEqual(len(tokens), 4) - self.assertEqual(tokens[2][0], Number.Integer) + self.assertEqual(tokens[2][0], T.Number.Integer) self.assertEqual(tokens[2][1], '-1') def test_tab_expansion(self): @@ -88,15 +87,15 @@ class TestToken(unittest.TestCase): self.assertEqual(str(token), 'FoO') def test_repr(self): - token = sql.Token(Keyword, 'foo') + token = sql.Token(T.Keyword, 'foo') tst = "