示例#1
0
def test_tokenize_repeated():
    assert tokenize_output_sorted("""
        +-------+-------+
        | one   | two   |
        | ----- | ----- |
        | three | 12345 |
        +-------+-------+
    """) == ['12345', 'one', 'three', 'two']
示例#2
0
def test_tokenize_one():
    assert tokenize_output_sorted('one') == ['one']
示例#3
0
def test_tokenize_complex():
    assert tokenize_output_sorted(
        'one "two" Three=four {"qwe":"hello world"}') == [
            'Three', 'four', 'hello', 'one', 'qwe', 'two', 'world'
        ]
示例#4
0
def test_tokenize_javascript():
    assert tokenize_output_sorted("{Hello: 'hello world', test:null}") == [
        'Hello', 'hello', 'hello world', 'test', 'world'
    ]
示例#5
0
def test_tokenize_json_partial():
    assert tokenize_output_sorted('"test": "1",') == ['test']
示例#6
0
def test_tokenize_json():
    assert tokenize_output_sorted(
        '{"Hello": "hello world", "test": None}') == [
            'Hello', 'hello', 'hello world', 'test', 'world'
        ]
示例#7
0
def test_tokenize_env_substrig():
    assert tokenize_output_sorted('SHELL=bash\nPATH=/a/b:/c/d',
                                  substring='/c') == ['/a/b:/c/d', '/c/d']
示例#8
0
def test_tokenize_env():
    assert tokenize_output_sorted('SHELL=bash\nPATH=/a/b:/c/d') == [
        '/a/b', '/a/b:/c/d', '/c/d', 'PATH', 'SHELL', 'bash'
    ]
示例#9
0
def test_tokenize_substring():
    assert tokenize_output_sorted('one two three four five six',
                                  substring='e') == ['five', 'one', 'three']
示例#10
0
def test_tokenize_empty():
    assert tokenize_output_sorted('') == []
示例#11
0
def test_tokenize_specials():
    assert tokenize_output_sorted(
        '\n\t\r one \n\t\r "two" \n\t\r three \n\t\r') == [
            'one', 'three', 'two'
        ]
示例#12
0
def test_tokenize_one_2_three_4():
    assert tokenize_output_sorted('one 2 three 4') == ['one', 'three']
示例#13
0
def test_tokenize_empty_prefix():
    assert tokenize_output_sorted('one two three', substring='none') == []