Пример #1
0
def test_split_by_type_with_no_such_type():
    tokenized = Tokenized()
    tokenized.add(Token(TokenType.SIMPLE, "1"))
    tokenized.add(Token(TokenType.ASSIGNMENT, "="))
    splitted = tokenized.split_by_type(TokenType.PIPE)
    assert len(splitted) == 1
    assert splitted[0].to_string_list() == ["1", "="]
Пример #2
0
def test_size():
    tokenized = Tokenized()

    assert tokenized.size() == 0

    tokenized.add(Token(TokenType.SIMPLE, "simple"))
    assert tokenized.size() == 1

    tokenized.add(Token(TokenType.DOUBLE_QUOTED, "\"double\""))
    assert tokenized.size() == 2
Пример #3
0
def test_first():
    tokenized = Tokenized()

    tokenized.add(Token(TokenType.SIMPLE, "simple"))
    first = tokenized.first()
    assert first.get_token_type() == TokenType.SIMPLE
    assert first.get_string_value() == "simple"

    tokenized.add(Token(TokenType.DOUBLE_QUOTED, "\"double\""))
    assert first.get_token_type() == TokenType.SIMPLE
    assert first.get_string_value() == "simple"
Пример #4
0
def test_to_list():
    tokenized = Tokenized()
    tokenized.add(Token(TokenType.PIPE, "|"))
    tokenized.add(Token(TokenType.ASSIGNMENT, "="))

    token_list = tokenized.to_list()
    assert token_list[0].get_token_type() == TokenType.PIPE
    assert token_list[0].get_string_value() == "|"
    assert token_list[1].get_token_type() == TokenType.ASSIGNMENT
    assert token_list[1].get_string_value() == "="

    assert len(token_list) == 2
Пример #5
0
 def process(self, tokenized, environment):
     """
     Returns a Tokenized object with substituted variables from the environment
     in the given tokenized string.
     Substitution is not performed inside single quotes.
     An exception is throwing if there is no variable in the environment.
     Substitution is performed only in one pass.
     Token types are not changed after the substitution.
     """
     result = Tokenized()
     for token in tokenized.to_list():
         string_value = token.get_string_value()
         token_type = token.get_token_type()
         if token_type != TokenType.SINGLE_QUOTED:
             string_value = self.__substitute(string_value, environment)
         result.add(Token(token_type, string_value))
     return result
Пример #6
0
def test_not_contains():
    tokenized = Tokenized()

    assert not tokenized.contains_token(TokenType.SIMPLE)
    tokenized.add(Token(TokenType.SIMPLE, "simple"))

    assert not tokenized.contains_token(TokenType.DOUBLE_QUOTED)
    tokenized.add(Token(TokenType.DOUBLE_QUOTED, "\"double\""))

    assert not tokenized.contains_token(TokenType.SINGLE_QUOTED)
    tokenized.add(Token(TokenType.SINGLE_QUOTED, "'single'"))

    assert not tokenized.contains_token(TokenType.PIPE)
    tokenized.add(Token(TokenType.PIPE, "|"))

    assert not tokenized.contains_token(TokenType.ASSIGNMENT)
    tokenized.add(Token(TokenType.ASSIGNMENT, "="))

    assert not tokenized.contains_token(TokenType.EMPTY)
Пример #7
0
def test_split_by_type():
    tokenized = Tokenized()
    tokenized.add(Token(TokenType.SIMPLE, "1"))
    tokenized.add(Token(TokenType.PIPE, "|"))
    tokenized.add(Token(TokenType.PIPE, "|"))
    tokenized.add(Token(TokenType.SIMPLE, "2"))
    tokenized.add(Token(TokenType.SIMPLE, "3"))

    splitted = tokenized.split_by_type(TokenType.PIPE)

    assert len(splitted) == 3
    assert splitted[0].to_string_list() == ["1"]
    assert splitted[1].size() == 0
    assert splitted[2].to_string_list() == ["2", "3"]
Пример #8
0
def test_to_string_list():
    tokenized = Tokenized()
    tokenized.add(Token(TokenType.SIMPLE, "simple"))
    tokenized.add(Token(TokenType.DOUBLE_QUOTED, "\"double\""))
    tokenized.add(Token(TokenType.SINGLE_QUOTED, "'single'"))
    tokenized.add(Token(TokenType.PIPE, "|"))
    tokenized.add(Token(TokenType.ASSIGNMENT, "="))
    tokenized.add(Token(TokenType.EMPTY, " "))

    expected = ["simple", "\"double\"", "'single'", "|", "=", " "]
    assert expected == tokenized.to_string_list()
Пример #9
0
def test_make_string():
    tokenized = Tokenized()
    tokenized.add(Token(TokenType.SIMPLE, "simple"))
    tokenized.add(Token(TokenType.DOUBLE_QUOTED, "\"double\""))
    tokenized.add(Token(TokenType.SINGLE_QUOTED, "'single'"))
    tokenized.add(Token(TokenType.PIPE, "|"))
    tokenized.add(Token(TokenType.ASSIGNMENT, "="))
    tokenized.add(Token(TokenType.EMPTY, " "))

    expected = "simple\"double\"'single'|= "
    assert expected == tokenized.make_string()