mirror of
https://git.savannah.gnu.org/git/bison.git
synced 2026-03-13 06:13:02 +00:00
Have Bison grammars parsed by a Bison grammar.
* src/reader.c, src/reader.h (prologue_augment): New. * src/reader.c (copy_definition): Remove. * src/reader.h, src/reader.c (gram_start_symbol_set, prologue_augment) (grammar_symbol_append, grammar_rule_begin, grammar_midrule_action) (grammar_current_rule_prec_set, grammar_current_rule_check) (grammar_current_rule_symbol_append) (grammar_current_rule_action_append): Export. * src/parse-gram.y (symbol_list_new, symbol_list_symbol_append_ (symbol_list_action_append): Remove. Hook the routines from reader. * src/scan-gram.l: In INITIAL, characters and strings are tokens. * src/system.h (ATTRIBUTE_NORETURN, ATTRIBUTE_UNUSED): Now. * src/reader.c (read_declarations): Remove, unused. * src/parse-gram.y: Handle the epilogue. * src/reader.h, src/reader.c (gram_start_symbol_set): Rename as... (grammar_start_symbol_set): this. * src/scan-gram.l: Be sure to ``use'' yycontrol to keep GCC quiet. * src/reader.c (readgram): Remove, unused. (reader): Adjust to insert eoftoken and axiom where appropriate. * src/reader.c (copy_dollar): Replace with... * src/scan-gram.h (handle_dollar): this. * src/parse-gram.y: Remove `%thong'. * src/reader.c (copy_at): Replace with... * src/scan-gram.h (handle_at): this. * src/complain.h, src/complain.c (warn_at, complain_at, fatal_at): New. * src/scan-gram.l (YY_LINES): Keep lineno synchronized for the time being. * src/reader.h, src/reader.c (grammar_rule_end): New. * src/parse.y (current_type, current_class): New. Implement `%nterm', `%token' support. Merge `%term' into `%token'. (string_as_id): New. * src/symtab.h, src/symtab.c (symbol_make_alias): Don't pass the type name. * src/parse-gram.y: Be sure to handle properly the beginning of rules. * src/parse-gram.y: Handle %type. * src/reader.c (grammar_rule_end): Call grammar_current_rule_check. * src/parse-gram.y: More directives support. * src/options.c: No longer handle source directives. * src/parse-gram.y: Fix %output. * src/parse-gram.y: Handle %union. Use the prologue locations. * src/reader.c (parse_union_decl): Remove. * src/reader.h, src/reader.c (epilogue_set): New. * src/parse-gram.y: Use it. * data/bison.simple, data/bison.c++: b4_stype is now either not defined, then default to int, or to the contents of %union, without `union' itself. Adjust. * src/muscle_tab.c (muscle_init): Don't predefine `stype'. * src/output.c (actions_output): Don't output braces, as they are already handled by the scanner. * src/scan-gram.l (SC_CHARACTER): Set the user_token_number of characters to themselves. * tests/reduce.at (Reduced Automaton): End the grammars with %% so that the epilogue has a proper #line. * src/parse-gram.y: Handle precedence/associativity. * src/symtab.c (symbol_precedence_set): Requires the symbol to be a terminal. * src/scan-gram.l (SC_BRACED_CODE): Catch strings and characters. * tests/calc.at: Do not use `%token "foo"' as it makes not sense at all to define terminals that cannot be emitted. * src/scan-gram.l: Escape M4 characters. * src/scan-gram.l: Working properly with escapes in user strings/characters. * tests/torture.at (AT_DATA_TRIANGULAR_GRAMMAR) (AT_DATA_HORIZONTAL_GRAMMAR): Respect the `%token ID NUM STRING' grammar. Use more modest sizes, as for the time being the parser does not release memory, and therefore the process swallows a huge amount of memory. * tests/torture.at (AT_DATA_LOOKAHEADS_GRAMMAR): Adjust to the stricter %token grammar. * src/symtab.h (associativity): Add `undef_assoc'. (symbol_precedence_set): Do nothing when passed an undef_assoc. * src/symtab.c (symbol_check_alias_consistence): Adjust. * tests/regression.at (Invalid %directive): Remove, as it is now meaningless. (Invalid inputs): Adjust to the new error messages. (Token definitions): The new grammar doesn't allow too many eccentricities. * src/lex.h, src/lex.c: Remove. * src/reader.c (lastprec, skip_to_char, read_signed_integer) (copy_character, copy_string2, copy_string, copy_identifier) (copy_comment, parse_token_decl, parse_type_decl, parse_assoc_decl) (parse_muscle_decl, parse_dquoted_param, parse_skel_decl) (parse_action): Remove. * po/POTFILES.in: Adjust.
This commit is contained in:
@@ -89,7 +89,7 @@ char quote[] = "@:>@@:>@,";
|
||||
%type <ival> exp
|
||||
|
||||
/* Exercise quotes in strings. */
|
||||
%token "fake @>:@@>:@,"
|
||||
%token FAKE "fake @>:@@>:@,"
|
||||
|
||||
%nonassoc '=' /* comparison */
|
||||
%left '-' '+'
|
||||
|
||||
@@ -204,6 +204,7 @@ not_reachable: useful { /* A not reachable action. */ }
|
||||
non_productive: non_productive useless_token
|
||||
{ /* Another non productive action. */ }
|
||||
;
|
||||
%%
|
||||
]])
|
||||
|
||||
AT_CHECK([[bison not-reduced.y]], 0, [],
|
||||
@@ -244,6 +245,7 @@ exp: useful { /* A useful action. */ }
|
||||
//non_productive: non_productive useless_token
|
||||
// { /* Another non productive action. */ }
|
||||
// ;
|
||||
%%
|
||||
]])
|
||||
|
||||
AT_CHECK([[bison reduced.y]])
|
||||
|
||||
@@ -287,40 +287,23 @@ AT_DATA([input.y],
|
||||
[[%%
|
||||
?
|
||||
default: 'a' }
|
||||
%{
|
||||
%&
|
||||
%a
|
||||
%-
|
||||
%{
|
||||
]])
|
||||
|
||||
AT_CHECK([bison input.y], [1], [],
|
||||
[[input.y:2: invalid input: `?'
|
||||
input.y:3: invalid input: `}'
|
||||
input.y:4: invalid input: `%{'
|
||||
input.y:5: invalid input: `%&'
|
||||
input.y:6: invalid input: `%a'
|
||||
input.y:7: invalid input: `%-'
|
||||
]])
|
||||
|
||||
AT_CLEANUP
|
||||
|
||||
|
||||
|
||||
## -------------------- ##
|
||||
## Invalid %directive. ##
|
||||
## -------------------- ##
|
||||
|
||||
|
||||
AT_SETUP([Invalid %directive])
|
||||
|
||||
AT_DATA([input.y],
|
||||
[[%invalid
|
||||
]])
|
||||
|
||||
AT_CHECK([bison input.y], [1], [],
|
||||
[[input.y:1: unrecognized: %invalid
|
||||
input.y:1: Skipping to next %
|
||||
input.y:2: fatal error: no input grammar
|
||||
[[input.y:2.1: invalid character: `?'
|
||||
input.y:3.14: invalid character: `}'
|
||||
input.y:4.1: invalid character: `%'
|
||||
input.y:4.2: invalid character: `&'
|
||||
input.y:5.1: invalid character: `%'
|
||||
input.y:6.1: invalid character: `%'
|
||||
input.y:6.2: invalid character: `-'
|
||||
input.y:7.1-8.0: unexpected end of file in a prologue
|
||||
input.y:7.1-8.0: parse error, unexpected PROLOGUE, expecting ";" or "|"
|
||||
input.y:8: symbol a is used, but is not defined as a token and has no rules
|
||||
]])
|
||||
|
||||
AT_CLEANUP
|
||||
@@ -340,13 +323,11 @@ AT_DATA([input.y],
|
||||
void yyerror (const char *s);
|
||||
int yylex (void);
|
||||
%}
|
||||
[%token "end of file"
|
||||
[%token YYEOF 0 "end of file"
|
||||
%token 'a' "a"
|
||||
%token "b" 'b'
|
||||
%token "c" c
|
||||
%token d "d"
|
||||
%token e 'e'
|
||||
%token 'f' e
|
||||
%token b "b"
|
||||
%token c 'c'
|
||||
%token 'd' d
|
||||
%%
|
||||
exp: "a";
|
||||
]])
|
||||
|
||||
@@ -33,7 +33,7 @@ AT_DATA([[input.y]],
|
||||
[$2])
|
||||
|
||||
AT_CHECK([bison input.y -o input.c])
|
||||
AT_CHECK([$CC $CFLAGS $CPPFLAGS input.c -o input], 1, [], [stderr])
|
||||
AT_CHECK([$CC $CFLAGS $CPPFLAGS input.c -c], 1, [], [stderr])
|
||||
# In case GCC displays column information, strip it down.
|
||||
#
|
||||
# input.y:4:2: #error "4" or input.y:4.2: #error "4"
|
||||
|
||||
@@ -57,7 +57,7 @@ EOF
|
||||
|
||||
for my $size (1 .. $max)
|
||||
{
|
||||
print "%token \"$size\" ", $size, "\n";
|
||||
print "%token t$size $size \"$size\"\n";
|
||||
};
|
||||
|
||||
print <<EOF;
|
||||
@@ -128,7 +128,9 @@ AT_SETUP([Big triangle])
|
||||
|
||||
# I have been able to go up to 2000 on my machine.
|
||||
# I tried 3000, a 29Mb grammar file, but then my system killed bison.
|
||||
AT_DATA_TRIANGULAR_GRAMMAR([input.y], [500])
|
||||
# With 500 and the new parser, which consume far too much memory,
|
||||
# it gets killed too. Of course the parser is to be cleaned.
|
||||
AT_DATA_TRIANGULAR_GRAMMAR([input.y], [200])
|
||||
AT_CHECK([bison input.y -v -o input.c])
|
||||
AT_CHECK([$CC $CFLAGS $CPPFLAGS input.c -o input], 0, [], [ignore])
|
||||
AT_CHECK([./input])
|
||||
@@ -164,7 +166,7 @@ EOF
|
||||
|
||||
for my $size (1 .. $max)
|
||||
{
|
||||
print "%token \"$size\" ", $size, "\n";
|
||||
print "%token t$size $size \"$size\"\n";
|
||||
};
|
||||
|
||||
print <<EOF;
|
||||
@@ -272,12 +274,12 @@ EOF
|
||||
print
|
||||
wrap ("%type <val> ",
|
||||
" ",
|
||||
map { "token$_" } (1 .. $max)),
|
||||
map { "n$_" } (1 .. $max)),
|
||||
"\n";
|
||||
|
||||
for my $count (1 .. $max)
|
||||
{
|
||||
print "%token \"$count\" $count\n";
|
||||
print "%token t$count $count \"$count\"\n";
|
||||
};
|
||||
|
||||
print <<EOF;
|
||||
@@ -288,18 +290,18 @@ input:
|
||||
;
|
||||
|
||||
exp:
|
||||
token1 "1" { assert (\@S|@1 == 1); }
|
||||
n1 "1" { assert (\@S|@1 == 1); }
|
||||
EOF
|
||||
|
||||
for my $count (2 .. $max)
|
||||
{
|
||||
print "| token$count \"$count\" { assert (\@S|@1 == $count); }\n";
|
||||
print "| n$count \"$count\" { assert (\@S|@1 == $count); }\n";
|
||||
};
|
||||
print ";\n";
|
||||
|
||||
for my $count (1 .. $max)
|
||||
{
|
||||
print "token$count: token { \$\$ = $count; };\n";
|
||||
print "n$count: token { \$\$ = $count; };\n";
|
||||
};
|
||||
|
||||
print <<EOF;
|
||||
|
||||
Reference in New Issue
Block a user