Пример #1
0
def test_lexer_ident_colon_colors():
    lexer = Lexer("abc: #11223311, #aabbcc, #abc1, #fff, #dd, #e", {})
    tokens = [token for token in lexer]
    from stilus.nodes.null import null

    assert tokens[0] == Token("ident", Ident("abc", null), lineno=1, column=1)
    assert tokens[1] == Token(":", ":", " ", lineno=1, column=4)
    assert tokens[2] == Token(
        "color", RGBA(17, 34, 51, 0.67), lineno=1, column=6
    )
    assert tokens[4] == Token(
        "color", RGBA(170, 187, 204, 1), lineno=1, column=17
    )
    assert tokens[6] == Token(
        "color", RGBA(170, 187, 204, 0.067), lineno=1, column=26
    )
    assert tokens[8] == Token(
        "color", RGBA(255, 255, 255, 1), lineno=1, column=33
    )
    assert tokens[10] == Token(
        "color", RGBA(221, 221, 221, 1), lineno=1, column=39
    )
    assert tokens[12] == Token(
        "color", RGBA(238, 238, 238, 1), lineno=1, column=44
    )
Пример #2
0
def test_lexer_boolean_unicode():
    lexer = Lexer("if true:\n  return U+abcdef;\n", {})
    tokens = [token for token in lexer]
    assert tokens[1] == Token("boolean", true, "", lineno=1, column=4)
    assert tokens[4] == Token(
        "literal", Literal("U+abcdef"), lineno=2, column=17
    )
Пример #3
0
def test_lexer_keyword_string():
    lexer = Lexer('if "fizz":\n  return foo;\n', {})
    tokens = [token for token in lexer]
    assert tokens[0] == Token("if", "if", lineno=1, column=1)
    assert tokens[1] == Token(
        "string", String("fizz", '"'), lineno=1, column=4
    )
    assert tokens[3] == Token("return", "return", lineno=1, column=11)
Пример #4
0
def test_lexer_urlchars_important():
    lexer = Lexer('url("/images/foo.png")\n' "!important foo", {})
    tokens = [token for token in lexer]
    assert tokens[1] == Token(
        "string", String("/images/foo.png", '"'), lineno=1, column=5
    )
    assert tokens[4] == Token(
        "ident", Literal("!important"), lineno=2, column=1
    )
Пример #5
0
def test_lexer_ident_space():
    lexer = Lexer("abc def klm:\n  xyz abc\n", {})
    tokens = [token for token in lexer]
    from stilus.nodes.null import null

    assert tokens[0] == Token("ident", Ident("abc", null), lineno=1, column=1)
    assert tokens[1] == Token("space", lineno=1, column=4)
    assert tokens[2] == Token("ident", Ident("def", null), lineno=1, column=5)
    assert tokens[3] == Token("space", lineno=1, column=8)
    assert tokens[4] == Token("ident", Ident("klm", null), lineno=1, column=9)
Пример #6
0
def test_lexer_is_part_of_selector():
    lexer = Lexer("^if.null,[bar],abc  color: black\n", {})
    assert lexer.next() == Token("selector", "^", lineno=1, column=1)
    assert lexer.next() == Token("if", "if", lineno=1, column=2)
    assert lexer.next() == Token(".", ".", "", lineno=1, column=4)
    assert lexer.next() == Token("ident", Ident("null"), lineno=1, column=5)
    lexer = Lexer("^#fif: black\n", {})
    assert lexer.next() == Token("selector", "^", lineno=1, column=1)
    assert lexer.next() == Token(
        "color", RGBA(255, 255, 255, 1), lineno=1, column=2
    )
    assert lexer.next() == Token("ident", Ident("if"), lineno=1, column=4)
Пример #7
0
def test_lexer_empty_string():
    lexer = Lexer("Hello there!", {})
    assert lexer.stash == deque([])
    assert lexer.indent_stack == deque([])
    assert lexer.indent_re is None
    assert lexer.lineno == 1
    assert lexer.column == 1
    assert lexer.options == {}
    assert lexer.prev is None
    assert lexer.is_url is False
    assert lexer.at_eos is False
Пример #8
0
def test_lexer_ident_colon_null_newline_eos():
    lexer = Lexer("abc:\n  color: null\n", {})
    tokens = [token for token in lexer]
    from stilus.nodes.null import null

    assert tokens[0] == Token("ident", Ident("abc", null), lineno=1, column=1)
    assert tokens[1] == Token(":", ":", "", lineno=1, column=4)
    assert tokens[2] == Token(
        "ident", Ident("color", null), lineno=1, column=5
    )
    assert tokens[3] == Token(":", ":", " ", lineno=2, column=11)
    assert tokens[4] == Token("null", value=null, lineno=2, column=13)
    assert tokens[5] == Token("newline", lineno=2, column=17)
    assert tokens[6] == Token("eos", lineno=3, column=1)
Пример #9
0
def test_lexer_peek():
    lexer = Lexer("abc:\n  color: #11223311\n", {})
    abc = Token("ident", Ident("abc", null), lineno=1, column=1)
    assert lexer.peek() == abc
    assert lexer.peek() == abc
    assert lexer.next() == abc
    colon = Token(":", ":", "", lineno=1, column=4)
    assert lexer.peek() == Token(":", ":", "", lineno=1, column=4)
    assert lexer.next() == colon
Пример #10
0
def test_lexer_functions():
    lexer = Lexer(
        "mixin(add) {\n"
        "  mul = @(c, d) {\n"
        "    c * d\n"
        "  }\n"
        "  width: add(2, 3) + mul(4, 5)\n"
        "}\n",
        {},
    )
    tokens = [token for token in lexer]
    assert tokens[0] == Token(
        "function", Ident("mixin", null), "", lineno=1, column=1
    )
    anon_fun_token = Token("function", Ident("anonymous"), lineno=2, column=9)
    anon_fun_token.anonymous = True
    assert tokens[8] == anon_fun_token
Пример #11
0
def test_lexer_atrules():
    lexer = Lexer(
        "@viewport {\n"
        "  color: blue\n"
        "}\n"
        "\n"
        '@namespace svg "http://www.w3.org/2000/svg"\n'
        "@-moz-viewport\n",
        {},
    )
    tokens = [token for token in lexer]
    assert tokens[0] == Token("atrule", "viewport", lineno=1, column=1)
    assert tokens[9] == Token("namespace", lineno=5, column=1)
    assert tokens[12] == Token(
        "string",
        String("http://www.w3.org/2000/svg", '"'),
        lineno=5,
        column=16,
    )
    assert tokens[14] == Token("atrule", "-moz-viewport", lineno=6, column=1)
Пример #12
0
def test_lexer_function_paren_braces_sep_unit():
    lexer = Lexer(
        "bg()\n"
        "  background: blue\n"
        "\n"
        "body {\n"
        "  bg(); width: 100px\n"
        "}\n",
        {},
    )
    tokens = [token for token in lexer]
    from stilus.nodes.null import null

    assert tokens[0] == Token(
        "function", Ident("bg", null), "", lineno=1, column=1
    )
    assert tokens[1] == Token(")", ")", "", lineno=1, column=4)
    assert tokens[9] == Token("{", "{", lineno=4, column=6)
    assert tokens[11] == Token(
        "function", Ident("bg", null), "", lineno=5, column=3
    )
    assert tokens[13] == Token(";", None, lineno=5, column=7)
    assert tokens[16] == Token("unit", Unit(100.0, "px"), lineno=5, column=16)
    assert tokens[18] == Token("}", "}", lineno=6, column=1)
Пример #13
0
def test_lexer_clean():
    lexer = Lexer("Hello", {})
    assert lexer.clean("hello") == "hello"
    assert lexer.clean("\uFEFFhello") == "hello"
    # empty
    lexer = Lexer("", {})
    assert lexer.s == ""
    assert lexer.advance() == Token("eos", lineno=1, column=1)
    # empty lines
    lexer = Lexer("\n\n\n", {})
    assert lexer.s == "\n"
    assert lexer.advance() == Token("newline", lineno=1, column=1)
    assert lexer.advance() == Token("eos", lineno=2, column=1)
    # our most basic example
    lexer = Lexer("abc:\n  color: black\n", {})
    assert lexer.s == "abc:\rcolor: black\n"
    # weird stuff at the beginning and the end
    lexer = Lexer("\t\t\nabc;;;\n\t\n", {})
    assert lexer.s == "\t\t\nabc;;;\n"
    # whitespaces (and newlines) at the end must be removed
    lexer = Lexer("abc\t\n\t\f", {})
    assert lexer.s == "abc\n"
    # \r\n to \n
    lexer = Lexer("abc\r\ndef\n\n", {})
    assert lexer.s == "abc\ndef\n"
    lexer = Lexer("abc\r\n\r\ndef\n\n", {})
    assert lexer.s == "abc\n\ndef\n"
    # \ * string to \r
    lexer = Lexer("abc\n   \\    \ndef\n", {})
    assert lexer.s == "abc\n   \rdef\n"
    lexer = Lexer("abc\n   \\    \n\\  \ndef\n", {})
    assert lexer.s == "abc\n   \r\rdef\n"
    # comments
    lexer = Lexer("abc: // some comment\n  color: #FFF\n", {})
    assert lexer.s == "abc:\rcolor: #FFF\n"
    lexer = Lexer("abc: /* some comment\n * longer one\n * ends now */\n", {})
    assert lexer.s == "abc: /* some comment\n * longer one\n * ends now */\n"
    lexer = Lexer("abc: /*! some comment\n * longer one\n * ends now */\n", {})
    assert lexer.s == "abc: /*! some comment\n * longer one\n * ends now */\n"
    # more comments
    lexer = Lexer("abc(// another comment\ndef ,// yet another comment\n", {})
    assert lexer.s == "abc(\rdef ,\r"
    # whitespace, \n and ) or ,
    lexer = Lexer("abc: \t\f\t\n  )def\n", {})
    assert lexer.s == "abc:)\rdef\n"
    lexer = Lexer("abc: \t\f\t\n  ,def\n", {})
    assert lexer.s == "abc:,\rdef\n"
Пример #14
0
def test_lexer_eq():
    left = Lexer("foobar", {"fizz": "fuzz"})
    right = Lexer("foobar", {"fizz": "fuzz"})
    assert left == right
    wrong = Lexer("fizz", {})
    assert left != wrong
Пример #15
0
def test_lexer_escaped():
    lexer = Lexer("bar: 1 \\+ 2\n", {})
    tokens = [token for token in lexer]
    assert tokens[3] == Token("ident", Literal("+"), lineno=1, column=8)
Пример #16
0
def test_lexer_indent_outdent():
    lexer = Lexer("abc, def:\n  color: #12345678\n    foo: null\n", {})
    tokens = [token for token in lexer]
    assert tokens[7] == Token("indent", lineno=2, column=27)
    assert tokens[11] == Token("outdent", lineno=3, column=14)
Пример #17
0
def test_lexer_namedop():
    lexer = Lexer("foo is a bar\nfizz isnt a fuzz\n", {})
    tokens = [token for token in lexer]
    assert tokens[2] == Token("is a", "is a", " ", lineno=1, column=5)
    assert tokens[7] == Token("!=", "!=", " ", lineno=2, column=6)