예제 #1
0
파일: test.py 프로젝트: bakkot/jsparagus
    def testNegativeLookahead(self):
        tokenize = lexer.LexicalGrammar('a b')
        rules = {
            'goal': [
                [LookaheadRule(frozenset({'a'}), False), 'abs'],
            ],
            'abs': [
                ['a'],
                ['b'],
                ['abs', 'a'],
                ['abs', 'b'],
            ],
        }

        self.compile(tokenize, Grammar(rules))
        self.assertNoParse("a b", message="expected 'b', got 'a'")
        self.assertParse(
            'b a',
            ('goal', ('abs_2', 'b', 'a')))

        # In simple cases like this, the lookahead restriction can even
        # disambiguate a grammar that would otherwise be ambiguous.
        rules['goal'].append(prod(['a'], 'goal_a'))
        self.compile(tokenize, Grammar(rules))
        self.assertParse('a', ('goal_a', 'a'))
예제 #2
0
파일: test.py 프로젝트: evilpie/jsparagus
    def testNegativeLookahead(self):
        tokenize = lexer.LexicalGrammar('a b')
        rules = {
            'goal': [
                [LookaheadRule(frozenset({'a'}), False), 'abs'],
            ],
            'abs': [
                ['a'],
                ['b'],
                ['abs', 'a'],
                ['abs', 'b'],
            ],
        }

        parse = gen.compile(Grammar(rules))
        self.assertRaisesRegex(SyntaxError,
                               r"expected 'b', got 'a'",
                               lambda: parse(tokenize, "a b"))
        self.assertEqual(
            parse(tokenize, 'b a'),
            ('goal', ('abs 2', 'b', 'a'))
        )

        # In simple cases like this, the lookahead restriction can even
        # disambiguate a grammar that would otherwise be ambiguous.
        rules['goal'].append(prod(['a'], 'goal_a'))
        parse = gen.compile(Grammar(rules))
        self.assertEqual(
            parse(tokenize, 'a'),
            ('goal_a', 'a')
        )
예제 #3
0
파일: test.py 프로젝트: nbp/jsparagus
 def testExpandOptional(self):
     grammar = Grammar({'goal': [[]]})
     empties = {}
     # Unit test for gen.expand_optional_symbols_in_rhs
     self.assertEqual(
         list(
             gen.expand_optional_symbols_in_rhs(['ONE', 'TWO', '3'],
                                                grammar, empties)),
         [(['ONE', 'TWO', '3'], {})])
     self.assertEqual(
         list(
             gen.expand_optional_symbols_in_rhs(
                 ['a', 'b', Optional('c')], grammar, empties)),
         [(['a', 'b'], {
             2: None
         }), (['a', 'b', 'c'], {})])
     self.assertEqual(
         list(
             gen.expand_optional_symbols_in_rhs(
                 [Optional('a'), Optional('b')], grammar, empties)),
         [([], {
             0: None,
             1: None
         }), (['a'], {
             1: None
         }), (['b'], {
             0: None
         }), (['a', 'b'], {})])
예제 #4
0
파일: test.py 프로젝트: nbp/jsparagus
    def testArithmetic(self):
        tokenize = lexer.LexicalGrammar("+ - * / ( )",
                                        NUM=r'[0-9]\w*',
                                        VAR=r'[A-Za-z]\w*')
        arith_grammar = Grammar({
            'expr': [
                ['term'],
                ['expr', '+', 'term'],
                ['expr', '-', 'term'],
            ],
            'term': [
                ['prim'],
                ['term', '*', 'prim'],
                ['term', '/', 'prim'],
            ],
            'prim': [
                ['NUM'],
                ['VAR'],
                ['(', 'expr', ')'],
            ],
        })
        self.compile(tokenize, arith_grammar)

        self.assertParse('2 * 3 + 4 * (5 + 7)',
                         ('expr 1', ('term 1', '2', '*', '3'), '+',
                          ('term 1', '4', '*',
                           ('prim 2', '(', ('expr 1', '5', '+', '7'), ')'))))

        self.assertNoParse("(", message="unexpected end of input")
        self.assertNoParse(
            ")", message="expected one of ['(', 'NUM', 'VAR'], got ')'")
예제 #5
0
파일: test.py 프로젝트: mconca/jsparagus
    def testLookaheadDisambiguation(self):
        """A lookahead restriction should be able to rule out certain nonterminals entirely."""

        grammar = Grammar({
            'Script': [
                ['Statement'],
                ['Statement', 'Statement'],
            ],
            'Statement': [
                [
                    LookaheadRule(frozenset({'function'}), False),
                    'Expression', ';'
                ],
                ['Function'],
            ],
            'Function': [
                ['function', 'x', '(', ')', '{', '}'],
            ],
            'Expression': [
                ['Primary'],
                ['++', 'Primary'],
                ['Primary', '++'],
            ],
            'Primary': [
                ['Function'],
                ['x'],
            ],
        })

        self.compile(lexer.LexicalGrammar("function x ( ) { } ++ ;"), grammar)
        self.assertParse("function x() {}")
        self.assertParse("++function x() {};")
        self.assertNoParse("++function x() {}", message="unexpected end")
        self.assertNoParse("function x() {}++;", message="got ';'")
        self.assertParse("function x() {} ++x;")
예제 #6
0
파일: test.py 프로젝트: nbp/jsparagus
 def testLookaheadBeforeOptional(self):
     self.compile(
         lexer.LexicalGrammar('= : _',
                              PUBLIC=r'public\b',
                              IDENT=r'[a-z]+\b',
                              NUM=r'[0-9]\b'),
         Grammar({
             'decl': [
                 [
                     LookaheadRule(frozenset({'IDENT'}), True),
                     Optional('attrs'), 'pat', '=', 'NUM'
                 ],
             ],
             'attrs': [
                 ['attr'],
                 ['attrs', 'attr'],
             ],
             'attr': [
                 ['PUBLIC', ':'],
                 ['IDENT', ':'],
             ],
             'pat': [
                 ['IDENT'],
                 ['_'],
             ],
         }))
     self.assertEqual(self.parse("x = 0"), ("decl", None, "x", "=", "0"))
     self.assertParse("thread: x = 0")
     self.assertNoParse("public: x = 0",
                        message="expected 'IDENT', got 'PUBLIC'")
     self.assertNoParse("_ = 0", message="expected 'IDENT', got '_'")
     self.assertParse("funny: public: x = 0")
     self.assertParse("funny: _ = 0")
예제 #7
0
파일: test.py 프로젝트: evilpie/jsparagus
    def testLeftHandSideExpression(self):
        """Example of a grammar that's in SLR(1) but hard to smoosh into an LL(1) form.

        This is taken from the ECMAScript grammar.

        ...Of course, it's not really possible to enforce the desired syntactic
        restrictions in LR(k) either; the ES grammar matches `(x + y) = z` and
        an additional attribute grammar (IsValidSimpleAssignmentTarget) is
        necessary to rule it out.
        """
        self.compile(
            lexer.LexicalGrammar("= +", VAR=r'[a-z]+\b'),
            Grammar({
                'AssignmentExpression': [
                    ['AdditiveExpression'],
                    ['LeftHandSideExpression', '=', 'AssignmentExpression'],
                ],
                'AdditiveExpression': [
                    ['LeftHandSideExpression'],
                    ['AdditiveExpression', '+', 'LeftHandSideExpression'],
                ],
                'LeftHandSideExpression': [
                    ['VAR'],
                ]
            })
        )
        self.assertParse("z = x + y")
        self.assertNoParse(
            "x + y = z",
            message="expected one of ['+', 'end of input'], got '='")
예제 #8
0
파일: test.py 프로젝트: j-sow/jsparagus
    def testSimple(self):
        grammar = Grammar({
            'expr': [
                ['SYMBOL'],
                ['(', 'tail'],
            ],
            'tail': [
                [')'],
                ['expr', 'tail'],
            ],
        })
        self.compile(LispTokenizer, grammar)

        self.assertParse(
            "(lambda (x) (* x x))",
            ('expr_1',
                '(',
                ('tail_1',
                    'lambda',
                    ('tail_1',
                        ('expr_1', '(', ('tail_1', 'x', ')')),
                        ('tail_1',
                            ('expr_1',
                                '(',
                                ('tail_1',
                                    '*',
                                    ('tail_1',
                                        'x',
                                        ('tail_1', 'x', ')')))),
                            ')')))))
예제 #9
0
파일: test.py 프로젝트: evilpie/jsparagus
    def testConvenienceMethodTypeInference(self):
        """A method can be called only in an intermediate reduce expression."""

        # The action `f(g($0))`.
        action = CallMethod("f", [CallMethod("g", [0])])

        # The grammar `goal ::= NAME => f(g($1))`.
        grammar = Grammar(
            {
                'goal': [Production(['NAME'], action)],
            },
            variable_terminals=['NAME'])

        # Since the return value of f() is used as the value of a `goal`,
        # we infer that f() returns a goal.
        self.assertEqual(
            grammar.methods['f'].return_type,
            jsparagus.types.NtType('goal'))

        # Since the return value of g() isn't used except as an argument, we
        # just give it the type `g`. I guess NtType is a bit of a misnomer for
        # this.
        self.assertEqual(
            grammar.methods['g'].return_type,
            jsparagus.types.NtType('g'))

        # Since g() is passed to f(), we infer this:
        self.assertEqual(
            grammar.methods['f'].argument_types,
            [jsparagus.types.NtType('g')])
예제 #10
0
파일: test.py 프로젝트: evilpie/jsparagus
    def disabledNegativeLookaheadDisambiguation(self):
        tokenize = lexer.LexicalGrammar(
            '( ) { } ; function =',
            IDENT=r'[A-Za-z_][A-Za-z_0-9]*')
        grammar = Grammar({
            'stmts': [
                ['stmt'],
                ['stmts', 'stmt'],
            ],
            'stmt': [
                [LookaheadRule(set=frozenset({'function'}), positive=False),
                 'expr', ';'],
                ['fndecl'],
            ],
            'fndecl': [
                ['function', 'IDENT', '(', ')', '{', Optional('stmt'), '}'],
            ],
            'expr': [
                ['term'],
                ['IDENT', '=', 'expr'],
            ],
            'term': [
                ['(', 'expr', ')'],
                ['fndecl'],
                ['term', '(', 'expr', ')'],
            ],
        })
        parse = gen.compile(grammar)

        # Test that without the lookahead restriction, we reject this grammar
        # (it's ambiguous):
        del grammar['stmt'][0][0]
        self.assertRaisesRegex(ValueError,
                               'banana',
                               lambda: gen.compile(grammar))

        self.assertEqual(
            parse(tokenize, 'function f() { x = function y() {}; }'),
            ('stmt', 1,
                ('fndecl',
                    'function', 'f', '(', ')', '{',
                    ('stmt', 0,
                        ('expr', 1,
                            'x',
                            '=',
                            ('expr', 0,
                                ('term', 1,
                                    ('fndecl',
                                        'function', 'y', '(', ')',
                                        '{', None, '}')))),
                        ';'))))

        self.assertEqual(
            parse(tokenize, '(function g(){});'),
            ('stmts', 0,
                ('stmt', 0,
                    ('term', 1,
                        ('fndecl',
                            'function', 'g', '(', ')', '{', None, '}')),
                    ';')))
예제 #11
0
파일: test.py 프로젝트: evilpie/jsparagus
 def testEmptyGrammar(self):
     tokenize = lexer.LexicalGrammar("X")
     self.compile(tokenize, Grammar({'goal': [[]]}))
     self.assertParse("", ('goal',))
     self.assertNoParse(
         "X",
         message="expected 'end of input', got 'X' (line 1)")
예제 #12
0
파일: test.py 프로젝트: evilpie/jsparagus
    def testSimple(self):
        grammar = Grammar({
            'expr': [
                ['SYMBOL'],
                ['(', 'tail'],
            ],
            'tail': [
                [')'],
                ['expr', 'tail'],
            ],
        })
        parse = gen.compile(grammar)

        parsed = parse(LispTokenizer, "(lambda (x) (* x x))")
        self.assertEqual(
            parsed,
            ('expr 1',
                '(',
                ('tail 1',
                    'lambda',
                    ('tail 1',
                        ('expr 1', '(', ('tail 1', 'x', ')')),
                        ('tail 1',
                            ('expr 1',
                                '(',
                                ('tail 1',
                                    '*',
                                    ('tail 1',
                                        'x',
                                        ('tail 1', 'x', ')')))),
                            ')')))))
예제 #13
0
파일: test.py 프로젝트: evilpie/jsparagus
    def testLookaheadWithCanonicalLR(self):
        """Only a lookahead assertion makes this grammar unambiguous."""
        tokenize = lexer.LexicalGrammar("async => { } ;", Identifier=r'\w+')
        grammar = Grammar({
            "script": [
                ["Expression", ";"],
            ],
            "Expression": [
                ["PrimaryExpression"],
                ["async", "Identifier", "=>", "AsyncConciseBody"],
            ],
            "AsyncConciseBody": [
                [LookaheadRule(set=frozenset(["{"]), positive=False),
                 "Expression"],
                ["{", "}"],
            ],
            "PrimaryExpression": [
                ["{", "}"],
            ],
        })

        self.compile(tokenize, grammar)
        self.assertParse("{};")
        self.assertParse("async x => {};")
        self.assertParse("async x => async y => {};")
예제 #14
0
파일: test.py 프로젝트: evilpie/jsparagus
    def testDeepRecursion(self):
        grammar = Grammar({
            'expr': [
                ['SYMBOL'],
                ['(', ')'],
                ['(', 'exprs', ')'],
            ],
            'exprs': [
                ['expr'],
                ['exprs', 'expr'],
            ],
        })
        parse = gen.compile(grammar)

        N = 3000
        s = "x"
        t = ('expr 0', 'x')
        for i in range(N):
            s = "(" + s + ")"
            t = ('expr 2', '(', t, ')')

        result = parse(LispTokenizer, s)

        # Python can't check that result == t; it causes a RecursionError.
        # Testing that repr(result) == repr(t), same deal. So:
        for i in range(N):
            self.assertIsInstance(result, tuple)
            self.assertEqual(len(result), 4)
            self.assertEqual(result[0], 'expr 2')
            self.assertEqual(result[1], '(')
            self.assertEqual(result[3], ')')
            result = result[2]
예제 #15
0
파일: test.py 프로젝트: evilpie/jsparagus
 def testMultiGoal(self):
     tokenize = lexer.LexicalGrammar("WHILE DEF FN { } ( ) -> ;", ID=r'\w+')
     grammar = Grammar({
         "stmt": [
             ["expr", ";"],
             ["{", "stmts", "}"],
             ["WHILE", "(", "expr", ")", "stmt"],
             ["DEF", "ID", "(", "ID", ")", "{", Optional("stmts"), "}"],
         ],
         "stmts": [
             ["stmt"],
             ["stmts", "stmt"],
         ],
         "expr": [
             ["FN", "ID", "->", "expr"],
             ["call_expr"],
         ],
         "call_expr": [
             ["ID"],
             ["call_expr", "(", "expr", ")"],
             ["(", "expr", ")"],
         ],
     }, goal_nts=["stmts", "expr"])
     self.compile_multi(tokenize, grammar)
     self.assertParse("WHILE ( x ) { decx ( x ) ; }", goal="stmts")
     self.assertNoParse(
         "WHILE ( x ) { decx ( x ) ; }", goal="expr",
         message="expected one of ['(', 'FN', 'ID'], got 'WHILE'")
     self.assertParse("f(x);", goal="stmts")
     self.assertNoParse("f(x);", goal="expr",
                        message="expected 'end of input', got ';'")
     self.assertParse("(FN x -> f ( x ))(x)", goal="expr")
     self.assertNoParse("(FN x -> f ( x ))(x)", goal="stmts",
                        message="unexpected end of input")
예제 #16
0
파일: test.py 프로젝트: evilpie/jsparagus
    def testLeftFactorMultiLevel(self):
        """Test left-factoring again on a nonterminal introduced by
        left-factoring."""
        tokenize = lexer.LexicalGrammar("FOR IN TO BY ( ) = ;",
                                        VAR=r'[A-Za-z]+')

        # The first left-factoring pass on `stmt` will left-factor `FOR ( VAR`.
        # A second pass is needed to left-factor `= expr TO expr`.
        grammar = Grammar({
            'stmt': [
                ['expr', ';'],
                ['FOR', '(', 'VAR', 'IN', 'expr', ')', 'stmt'],
                ['FOR', '(', 'VAR', '=', 'expr', 'TO', 'expr', ')', 'stmt'],
                ['FOR', '(', 'VAR', '=', 'expr', 'TO', 'expr',
                 'BY', 'expr', ')', 'stmt'],
                ['IF', '(', 'expr', ')', 'stmt'],
            ],
            'expr': [
                ['VAR'],
            ],
        })
        parse = gen.compile(grammar)
        self.assertEqual(
            parse(tokenize, "FOR (x IN y) z;"),
            ('stmt 1', 'FOR', '(', 'x', 'IN', 'y', ')',
             ('stmt 0', 'z', ';')))
        self.assertEqual(
            parse(tokenize, "FOR (x = y TO z) x;"),
            ('stmt 2', 'FOR', '(', 'x', '=', 'y', 'TO', 'z', ')',
             ('stmt 0', 'x', ';')))
        self.assertEqual(
            parse(tokenize, "FOR (x = y TO z BY w) x;"),
            ('stmt 3', 'FOR', '(', 'x', '=', 'y', 'TO', 'z', 'BY', 'w', ')',
             ('stmt 0', 'x', ';')))
예제 #17
0
파일: test.py 프로젝트: evilpie/jsparagus
 def check(rules):
     grammar = Grammar(rules, goal_nts=['goal'])
     out = io.StringIO()
     self.assertRaisesRegex(
         ValueError,
         r"ambiguous grammar|reduce-reduce conflict",
         lambda: gen.generate_parser(out, grammar))
예제 #18
0
파일: test.py 프로젝트: evilpie/jsparagus
 def testList(self):
     list_grammar = Grammar({
         'prelist': [
             ['word', 'list']
         ],
         'list': [
             ['word'],
             ['list', 'word'],
         ],
         'word': [
             ['SYMBOL']
         ],
     })
     parse = gen.compile(list_grammar)
     self.assertEqual(
         parse(LispTokenizer,
               "the quick brown fox jumped over the lazy dog"),
         ('prelist',
             'the',
             ('list 1',
                 ('list 1',
                     ('list 1',
                         ('list 1',
                             ('list 1',
                                 ('list 1',
                                     ('list 1',
                                         'quick',
                                         'brown'),
                                     'fox'),
                                 'jumped'),
                             'over'),
                         'the'),
                     'lazy'),
                 'dog')))
예제 #19
0
파일: test.py 프로젝트: evilpie/jsparagus
 def testStaggeredItems(self):
     """Items in a state can have different amounts of leading context."""
     # In this example grammar, after "A" "B", we're in a state that
     # contains these two items (ignoring lookahead):
     #       goal ::= "A" "B" · y
     #       x ::= "B" · stars "X"
     #
     # Likewise, after `"A" "B" stars`, we have:
     #       x ::= "B" stars · "X"
     #       y ::= stars · "Y"
     #       stars ::= stars · "*"
     tokenize = lexer.LexicalGrammar("A B * X Y")
     grammar = Grammar({
         "goal": [
             ["A", "x"],
             ["A", "B", "y"],
         ],
         "x": [
             ["B", "stars", "X"],
         ],
         "y": [
             ["stars", "Y"],
         ],
         "stars": [
             ["*"],
             ["stars", "*"],
         ],
     })
     self.compile(tokenize, grammar)
     self.assertParse("A B * * * X")
     self.assertParse("A B * * * Y")
예제 #20
0
파일: test.py 프로젝트: evilpie/jsparagus
    def testCanonicalLR(self):
        """Example 4.39 (grammar 4.20) from the book."""

        # Modified as marked below
        grammar = Grammar({
            "S": [
                ["L", "=", "R"],
                ["R"],
            ],
            "L": [
                ["*", "R"],
                ["id"],
            ],
            "R": [
                ["L"],
                # added so we can have a negative test, showing that
                # `R = R` is not an S:
                ["7"],
            ],
        })
        self.compile(lexer.LexicalGrammar("id = * 7"), grammar)
        self.assertParse("id = *id")
        self.assertParse("*id = id")
        self.assertParse("id = 7")
        self.assertNoParse("7 = id",
                           message="expected 'end of input', got '='")
예제 #21
0
파일: test.py 프로젝트: evilpie/jsparagus
    def testReduceActions(self):
        tokenize = lexer.LexicalGrammar("+ - * / ( )",
                                        NUM=r'[0-9]\w*',
                                        VAR=r'[A-Za-z]\w*')
        grammar = Grammar({
            "expr": [
                ["term"],
                prod(["expr", "+", "term"], "add"),
                prod(["expr", "-", "term"], "sub"),
            ],
            "term": [
                ["unary"],
                prod(["term", "*", "unary"], "mul"),
                prod(["term", "/", "unary"], "div"),
            ],
            "unary": [
                ["prim"],
                prod(["-", "prim"], "neg"),
            ],
            "prim": [
                prod(["(", "expr", ")"], "parens"),
                prod(["NUM"], "num"),
                prod(["VAR"], "var"),
            ],
        }, goal_nts=['expr'])

        self.compile(tokenize, grammar)
        self.assertParse("X", ('var', 'X'))
        self.assertParse("3 + 4", ('add', ('num', '3'), '+', ('num', '4')))
        self.assertParse(
            "2 * 3 + 4 * (5 + 7)",
            (
                'add',
                ('mul', ('num', '2'), '*', ('num', '3')),
                '+',
                (
                    'mul',
                    ('num', '4'),
                    '*',
                    ('parens', '(',
                        ('add', ('num', '5'), '+', ('num', '7')), ')'))))
        self.assertParse(
            "1 / (1 + 1 / (1 + 1 / (1 + 1)))",
            (
                'div', ('num', '1'), '/', (
                    'parens', '(', (
                        'add', ('num', '1'), '+', (
                            'div', ('num', '1'), '/', (
                                'parens', '(', (
                                    'add', ('num', '1'), '+', (
                                        'div', ('num', '1'), '/', (
                                            'parens', '(', (
                                                'add', ('num', '1'), '+',
                                                ('num', '1')),
                                            ')'))),
                                ')'))),
                    ')')))
예제 #22
0
파일: test.py 프로젝트: nbp/jsparagus
 def testParameterizedProductions(self):
     passthru = ('Yield', Var('Yield')),
     name = Nt("name", passthru)
     stmt = Nt("stmt", passthru)
     stmts = Nt("stmts", passthru)
     grammar = Grammar(
         {
             'script': [
                 ['def'],
                 ['script', 'def'],
             ],
             'def': [
                 [
                     'function', 'IDENT', '(', ')', '{',
                     Nt('stmts', (('Yield', False), )), '}'
                 ],
                 [
                     'function', '*', 'IDENT', '(', ')', '{',
                     Nt('stmts', (('Yield', True), )), '}'
                 ],
             ],
             'stmts':
             NtDef(['Yield'], [
                 [stmt],
                 [stmts, stmt],
             ], None),
             'stmt':
             NtDef(['Yield'], [
                 [name, "(", ")", ";"],
                 [name, "=", name, ";"],
                 Production(["yield", name, ";"],
                            reducer=CallMethod("yield_stmt", [1]),
                            condition=('Yield', True)),
             ], None),
             'name':
             NtDef(
                 ['Yield'],
                 [
                     ["IDENT"],
                     # Specifically ask for a method here, because otherwise we
                     # wouldn't get one and then type checking would fail.
                     Production(["yield"],
                                CallMethod("yield_as_name", []),
                                condition=('Yield', False)),
                 ],
                 None),
         },
         variable_terminals=["IDENT"])
     self.compile(
         lexer.LexicalGrammar("( ) { } ; * = function yield",
                              IDENT=r'[A-Za-z]\w*'), grammar)
     self.assertParse("function* farm() { cow = pig; yield cow; }")
     self.assertNoParse(
         "function city() { yield toOncomingTraffic; }",
         message="expected one of ['(', ';', '='], got 'IDENT'")
     self.assertNoParse("function* farm() { yield = corn; yield yield; }",
                        message="expected 'IDENT', got '='")
예제 #23
0
파일: test.py 프로젝트: evilpie/jsparagus
 def testEnd(self):
     self.compile(
         lexer.LexicalGrammar("ONE TWO"),
         Grammar({
             'goal': [
                 ['ONE', 'TWO']
             ]
         })
     )
     self.assertNoParse("ONE TWO TWO",
                        message="expected 'end of input', got 'TWO'")
예제 #24
0
파일: test.py 프로젝트: nbp/jsparagus
 def testLeftFactorMulti(self):
     """Test left-factoring with common prefix of length >1."""
     tokenize = lexer.LexicalGrammar("A B C D E")
     grammar = Grammar({
         'goal': [
             ['A', 'B', 'C', 'D'],
             ['A', 'B', 'C', 'E'],
         ],
     })
     self.compile(tokenize, grammar)
     self.assertParse("A B C D", ('goal 0', 'A', 'B', 'C', 'D'))
     self.assertParse("A B C E", ('goal 1', 'A', 'B', 'C', 'E'))
예제 #25
0
파일: test.py 프로젝트: evilpie/jsparagus
 def testEpsilonFreeTransform(self):
     tokenize = lexer.LexicalGrammar('{ } X')
     grammar = Grammar({
         'goal': [
             ['{', 'xlist', '}'],
         ],
         'xlist': [
             [],
             ['xlist', 'X'],
         ],
     })
     self.compile(tokenize, grammar)
     self.assertParse("{}", ('goal', '{', ('xlist 0',), '}'))
예제 #26
0
파일: test.py 프로젝트: evilpie/jsparagus
    def testLeftFactor(self):
        """Most basic left-factoring test."""
        tokenize = lexer.LexicalGrammar("A B")
        grammar = Grammar({
            'goal': [
                ['A'],
                ['A', 'B'],
            ],
        })

        self.compile(tokenize, grammar)
        self.assertParse("A", 'A')
        self.assertParse("A B", ('goal 1', 'A', 'B'))
예제 #27
0
파일: test.py 프로젝트: mconca/jsparagus
    def testMissingParameterError(self):
        grammar = {
            'Foo': [
                ['Bar'],
            ],
            'Bar':
            NtDef(('Arg', ), [
                ['NUM'],
                Production(['STR'], reducer=0, condition=('Arg', True)),
            ], None),
        }

        self.assertRaisesRegex(ValueError, "missing parameters for 'Bar'",
                               lambda: Grammar(grammar))
예제 #28
0
파일: test.py 프로젝트: evilpie/jsparagus
 def testCheckCycleFree(self):
     tokenize = lexer.LexicalGrammar("!")
     grammar = Grammar({
         "problem": [
             ["one", "two"],
         ],
         "one": [
             ["!"],
         ],
         "two": [
             [Optional("problem")],
         ],
     })
     self.compile(tokenize, grammar)
     self.assertParse("! ! ! ! !")
예제 #29
0
파일: test.py 프로젝트: evilpie/jsparagus
    def testArithmetic(self):
        tokenize = lexer.LexicalGrammar(
            "+ - * / ( )",
            NUM=r'[0-9]\w*',
            VAR=r'[A-Za-z]\w*')
        arith_grammar = Grammar({
            'expr': [
                ['term'],
                ['expr', '+', 'term'],
                ['expr', '-', 'term'],
            ],
            'term': [
                ['prim'],
                ['term', '*', 'prim'],
                ['term', '/', 'prim'],
            ],
            'prim': [
                ['NUM'],
                ['VAR'],
                ['(', 'expr', ')'],
            ],
        })
        parse = gen.compile(arith_grammar)

        self.assertEqual(
            parse(tokenize, '2 * 3 + 4 * (5 + 7)'),
            ('expr 1',
                ('term 1', '2', '*', '3'),
                '+',
                ('term 1',
                    '4',
                    '*',
                    ('prim 2',
                        '(',
                        ('expr 1', '5', '+', '7'),
                        ')'))))

        self.assertRaisesRegex(
            SyntaxError,
            r"unexpected end of input",
            lambda: parse(tokenize, "("))
        self.assertRaisesRegex(
            SyntaxError,
            r"expected one of \['\(', 'NUM', 'VAR'], got '\)'",
            lambda: parse(tokenize, ")"))
예제 #30
0
파일: test.py 프로젝트: evilpie/jsparagus
 def testForLookahead(self):
     grammar = Grammar({
         'Stmt': [
             [';'],
             ['ForStmt'],
         ],
         'ForStmt': [
             ["for", "(", LookaheadRule(frozenset({"let"}), False),
              "Expr", ";", ";", ")", "Stmt"],
         ],
         'Expr': [
             ["0"],
             ["let"],
         ],
     })
     self.compile(lexer.LexicalGrammar("for ( let ; ) 0"), grammar)
     self.assertParse("for (0;;) ;")
     self.assertNoParse("for (let;;) ;", message="expected '0', got 'let'")