mirror of
https://git.savannah.gnu.org/git/bison.git
synced 2026-03-09 12:23:04 +00:00
(print_transitions): this. (print_transitions, print_errs, print_reductions): Align the lookaheads columns. (print_core, print_transitions, print_errs, print_state, print_grammar): Output empty lines separator before, not after. (state_default_rule_compute): Rename as... (state_default_rule): this. * tests/conflicts.at (Defaulted Conflicted Reduction), (Unresolved SR Conflicts, Resolved SR Conflicts): Adjust. * tests/regression.at (Rule Line Numbers, Web2c Report): Adjust.
637 lines
12 KiB
Plaintext
637 lines
12 KiB
Plaintext
# Bison Regressions. -*- Autotest -*-
|
||
# Copyright (C) 2001, 2002 Free Software Foundation, Inc.
|
||
|
||
# This program is free software; you can redistribute it and/or modify
|
||
# it under the terms of the GNU General Public License as published by
|
||
# the Free Software Foundation; either version 2, or (at your option)
|
||
# any later version.
|
||
|
||
# This program is distributed in the hope that it will be useful,
|
||
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||
# GNU General Public License for more details.
|
||
|
||
# You should have received a copy of the GNU General Public License
|
||
# along with this program; if not, write to the Free Software
|
||
# Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA
|
||
# 02111-1307, USA.
|
||
|
||
AT_BANNER([[Regression tests.]])
|
||
|
||
|
||
## ------------------------- ##
|
||
## Early token definitions. ##
|
||
## ------------------------- ##
|
||
|
||
|
||
AT_SETUP([Early token definitions])
|
||
|
||
# Found in GCJ: they expect the tokens to be defined before the user
|
||
# prologue, so that they can use the token definitions in it.
|
||
|
||
AT_DATA([input.y],
|
||
[[%{
|
||
void yyerror (const char *s);
|
||
int yylex (void);
|
||
%}
|
||
|
||
%union
|
||
{
|
||
int val;
|
||
};
|
||
%{
|
||
#ifndef MY_TOKEN
|
||
# error "MY_TOKEN not defined."
|
||
#endif
|
||
%}
|
||
%token MY_TOKEN
|
||
%%
|
||
exp: MY_TOKEN;
|
||
%%
|
||
]])
|
||
|
||
AT_CHECK([bison input.y -o input.c])
|
||
AT_CHECK([$CC $CFLAGS $CPPFLAGS input.c -c])
|
||
|
||
AT_CLEANUP
|
||
|
||
|
||
|
||
## ---------------- ##
|
||
## Braces parsing. ##
|
||
## ---------------- ##
|
||
|
||
|
||
AT_SETUP([Braces parsing])
|
||
|
||
AT_DATA([input.y],
|
||
[[/* Bison used to swallow the character after `}'. */
|
||
|
||
%%
|
||
exp: { tests = {{{{{{{{{{}}}}}}}}}}; };
|
||
%%
|
||
]])
|
||
|
||
AT_CHECK([bison -v input.y -o input.c])
|
||
|
||
AT_CHECK([fgrep 'tests = {{{{{{{{{{}}}}}}}}}};' input.c], 0, [ignore])
|
||
|
||
AT_CLEANUP
|
||
|
||
|
||
## ------------------ ##
|
||
## Duplicate string. ##
|
||
## ------------------ ##
|
||
|
||
|
||
AT_SETUP([Duplicate string])
|
||
|
||
AT_DATA([input.y],
|
||
[[/* `Bison -v' used to dump core when two tokens are defined with the same
|
||
string, as LE and GE below. */
|
||
|
||
%token NUM
|
||
%token LE "<="
|
||
%token GE "<="
|
||
|
||
%%
|
||
exp: '(' exp ')' | NUM ;
|
||
%%
|
||
]])
|
||
|
||
AT_CHECK([bison -v input.y -o input.c], 0, [],
|
||
[[input.y:6: warning: symbol `"<="' used more than once as a literal string
|
||
]])
|
||
|
||
AT_CLEANUP
|
||
|
||
|
||
## ------------------- ##
|
||
## Rule Line Numbers. ##
|
||
## ------------------- ##
|
||
|
||
AT_SETUP([Rule Line Numbers])
|
||
|
||
AT_KEYWORDS([report])
|
||
|
||
AT_DATA([input.y],
|
||
[[%%
|
||
expr:
|
||
'a'
|
||
|
||
{
|
||
|
||
}
|
||
|
||
'b'
|
||
|
||
{
|
||
|
||
}
|
||
|
||
|
|
||
|
||
|
||
{
|
||
|
||
|
||
}
|
||
|
||
'c'
|
||
|
||
{
|
||
|
||
};
|
||
]])
|
||
|
||
AT_CHECK([bison input.y -o input.c -v])
|
||
|
||
# Check the contents of the report.
|
||
AT_CHECK([cat input.output], [],
|
||
[[Grammar
|
||
|
||
0 $axiom: expr $
|
||
|
||
1 @1: /* empty */
|
||
|
||
2 expr: 'a' @1 'b'
|
||
|
||
3 @2: /* empty */
|
||
|
||
4 expr: @2 'c'
|
||
|
||
|
||
Terminals, with rules where they appear
|
||
|
||
$ (0) 0
|
||
'a' (97) 2
|
||
'b' (98) 2
|
||
'c' (99) 4
|
||
error (256)
|
||
|
||
|
||
Nonterminals, with rules where they appear
|
||
|
||
$axiom (6)
|
||
on left: 0
|
||
expr (7)
|
||
on left: 2 4, on right: 0
|
||
@1 (8)
|
||
on left: 1, on right: 2
|
||
@2 (9)
|
||
on left: 3, on right: 4
|
||
|
||
|
||
state 0
|
||
|
||
0 $axiom: . expr $
|
||
|
||
'a' shift, and go to state 1
|
||
|
||
$default reduce using rule 3 (@2)
|
||
|
||
expr go to state 2
|
||
@2 go to state 3
|
||
|
||
|
||
state 1
|
||
|
||
2 expr: 'a' . @1 'b'
|
||
|
||
$default reduce using rule 1 (@1)
|
||
|
||
@1 go to state 4
|
||
|
||
|
||
state 2
|
||
|
||
0 $axiom: expr . $
|
||
|
||
$ shift, and go to state 5
|
||
|
||
|
||
state 3
|
||
|
||
4 expr: @2 . 'c'
|
||
|
||
'c' shift, and go to state 6
|
||
|
||
|
||
state 4
|
||
|
||
2 expr: 'a' @1 . 'b'
|
||
|
||
'b' shift, and go to state 7
|
||
|
||
|
||
state 5
|
||
|
||
0 $axiom: expr $ .
|
||
|
||
$default accept
|
||
|
||
|
||
state 6
|
||
|
||
4 expr: @2 'c' .
|
||
|
||
$default reduce using rule 4 (expr)
|
||
|
||
|
||
state 7
|
||
|
||
2 expr: 'a' @1 'b' .
|
||
|
||
$default reduce using rule 2 (expr)
|
||
]])
|
||
|
||
AT_CLEANUP
|
||
|
||
|
||
|
||
## ---------------------- ##
|
||
## Mixing %token styles. ##
|
||
## ---------------------- ##
|
||
|
||
|
||
AT_SETUP([Mixing %token styles])
|
||
|
||
# Taken from the documentation.
|
||
AT_DATA([input.y],
|
||
[[%token <operator> OR "||"
|
||
%token <operator> LE 134 "<="
|
||
%left OR "<="
|
||
%%
|
||
exp: ;
|
||
%%
|
||
]])
|
||
|
||
AT_CHECK([bison -v input.y -o input.c])
|
||
|
||
AT_CLEANUP
|
||
|
||
|
||
|
||
## ---------------- ##
|
||
## Invalid inputs. ##
|
||
## ---------------- ##
|
||
|
||
|
||
AT_SETUP([Invalid inputs])
|
||
|
||
AT_DATA([input.y],
|
||
[[%%
|
||
?
|
||
default: 'a' }
|
||
%&
|
||
%a
|
||
%-
|
||
%{
|
||
]])
|
||
|
||
AT_CHECK([bison input.y], [1], [],
|
||
[[input.y:2.1: invalid character: `?'
|
||
input.y:3.14: invalid character: `}'
|
||
input.y:4.1: invalid character: `%'
|
||
input.y:4.2: invalid character: `&'
|
||
input.y:5.1: invalid character: `%'
|
||
input.y:6.1: invalid character: `%'
|
||
input.y:6.2: invalid character: `-'
|
||
input.y:7.1-8.0: unexpected end of file in a prologue
|
||
input.y:7.1-8.0: parse error, unexpected PROLOGUE, expecting ";" or "|"
|
||
input.y:5.2: symbol a is used, but is not defined as a token and has no rules
|
||
]])
|
||
|
||
AT_CLEANUP
|
||
|
||
|
||
|
||
## ------------------- ##
|
||
## Token definitions. ##
|
||
## ------------------- ##
|
||
|
||
|
||
AT_SETUP([Token definitions])
|
||
|
||
# Bison managed, when fed with `%token 'f' "f"' to #define 'f'!
|
||
AT_DATA([input.y],
|
||
[%{
|
||
void yyerror (const char *s);
|
||
int yylex (void);
|
||
%}
|
||
[%token YYEOF 0 "end of file"
|
||
%token 'a' "a"
|
||
%token b "b"
|
||
%token c 'c'
|
||
%token 'd' d
|
||
%%
|
||
exp: "a";
|
||
]])
|
||
|
||
AT_CHECK([bison input.y -o input.c])
|
||
AT_CHECK([$CC $CFLAGS $CPPFLAGS input.c -c])
|
||
AT_CLEANUP
|
||
|
||
|
||
|
||
## -------------- ##
|
||
## Web2c Report. ##
|
||
## -------------- ##
|
||
|
||
# The generation of the reduction was once wrong in Bison, and made it
|
||
# miss some reductions. In the following test case, the reduction on
|
||
# `undef_id_tok' in state 1 was missing. This is stripped down from
|
||
# the actual web2c.y.
|
||
|
||
AT_SETUP([Web2c Report])
|
||
|
||
AT_KEYWORDS([report])
|
||
|
||
AT_DATA([input.y],
|
||
[[%token undef_id_tok const_id_tok
|
||
|
||
%start CONST_DEC_PART
|
||
|
||
%%
|
||
CONST_DEC_PART:
|
||
CONST_DEC_LIST
|
||
;
|
||
|
||
CONST_DEC_LIST:
|
||
CONST_DEC
|
||
| CONST_DEC_LIST CONST_DEC
|
||
;
|
||
|
||
CONST_DEC:
|
||
{ } undef_id_tok '=' const_id_tok ';'
|
||
;
|
||
%%
|
||
]])
|
||
|
||
AT_CHECK([bison -v input.y])
|
||
AT_CHECK([cat input.output], 0,
|
||
[[Grammar
|
||
|
||
0 $axiom: CONST_DEC_PART $
|
||
|
||
1 CONST_DEC_PART: CONST_DEC_LIST
|
||
|
||
2 CONST_DEC_LIST: CONST_DEC
|
||
3 | CONST_DEC_LIST CONST_DEC
|
||
|
||
4 @1: /* empty */
|
||
|
||
5 CONST_DEC: @1 undef_id_tok '=' const_id_tok ';'
|
||
|
||
|
||
Terminals, with rules where they appear
|
||
|
||
$ (0) 0
|
||
';' (59) 5
|
||
'=' (61) 5
|
||
error (256)
|
||
undef_id_tok (258) 5
|
||
const_id_tok (259) 5
|
||
|
||
|
||
Nonterminals, with rules where they appear
|
||
|
||
$axiom (7)
|
||
on left: 0
|
||
CONST_DEC_PART (8)
|
||
on left: 1, on right: 0
|
||
CONST_DEC_LIST (9)
|
||
on left: 2 3, on right: 1 3
|
||
CONST_DEC (10)
|
||
on left: 5, on right: 2 3
|
||
@1 (11)
|
||
on left: 4, on right: 5
|
||
|
||
|
||
state 0
|
||
|
||
0 $axiom: . CONST_DEC_PART $
|
||
|
||
$default reduce using rule 4 (@1)
|
||
|
||
CONST_DEC_PART go to state 1
|
||
CONST_DEC_LIST go to state 2
|
||
CONST_DEC go to state 3
|
||
@1 go to state 4
|
||
|
||
|
||
state 1
|
||
|
||
0 $axiom: CONST_DEC_PART . $
|
||
|
||
$ shift, and go to state 5
|
||
|
||
|
||
state 2
|
||
|
||
1 CONST_DEC_PART: CONST_DEC_LIST .
|
||
3 CONST_DEC_LIST: CONST_DEC_LIST . CONST_DEC
|
||
|
||
undef_id_tok reduce using rule 4 (@1)
|
||
$default reduce using rule 1 (CONST_DEC_PART)
|
||
|
||
CONST_DEC go to state 6
|
||
@1 go to state 4
|
||
|
||
|
||
state 3
|
||
|
||
2 CONST_DEC_LIST: CONST_DEC .
|
||
|
||
$default reduce using rule 2 (CONST_DEC_LIST)
|
||
|
||
|
||
state 4
|
||
|
||
5 CONST_DEC: @1 . undef_id_tok '=' const_id_tok ';'
|
||
|
||
undef_id_tok shift, and go to state 7
|
||
|
||
|
||
state 5
|
||
|
||
0 $axiom: CONST_DEC_PART $ .
|
||
|
||
$default accept
|
||
|
||
|
||
state 6
|
||
|
||
3 CONST_DEC_LIST: CONST_DEC_LIST CONST_DEC .
|
||
|
||
$default reduce using rule 3 (CONST_DEC_LIST)
|
||
|
||
|
||
state 7
|
||
|
||
5 CONST_DEC: @1 undef_id_tok . '=' const_id_tok ';'
|
||
|
||
'=' shift, and go to state 8
|
||
|
||
|
||
state 8
|
||
|
||
5 CONST_DEC: @1 undef_id_tok '=' . const_id_tok ';'
|
||
|
||
const_id_tok shift, and go to state 9
|
||
|
||
|
||
state 9
|
||
|
||
5 CONST_DEC: @1 undef_id_tok '=' const_id_tok . ';'
|
||
|
||
';' shift, and go to state 10
|
||
|
||
|
||
state 10
|
||
|
||
5 CONST_DEC: @1 undef_id_tok '=' const_id_tok ';' .
|
||
|
||
$default reduce using rule 5 (CONST_DEC)
|
||
]])
|
||
|
||
AT_CLEANUP
|
||
|
||
|
||
## --------------- ##
|
||
## Web2c Actions. ##
|
||
## --------------- ##
|
||
|
||
# The generation of the mapping `state -> action' was once wrong in
|
||
# extremely specific situations. web2c.y exhibits this situation.
|
||
# Below is a stripped version of the grammar. It looks like one can
|
||
# simplify it further, but just don't: it is tuned to exhibit a bug,
|
||
# which disapears when applying sane grammar transformations.
|
||
#
|
||
# It used to be wrong on yydefact only:
|
||
#
|
||
# static const short yydefact[] =
|
||
# {
|
||
# - 2, 0, 1, 0, 0, 2, 3, 2, 5, 4,
|
||
# + 2, 0, 1, 0, 0, 0, 3, 2, 5, 4,
|
||
# 0, 0
|
||
# };
|
||
#
|
||
# but let's check all the tables.
|
||
|
||
|
||
AT_SETUP([Web2c Actions])
|
||
|
||
AT_KEYWORDS([report])
|
||
|
||
AT_DATA([input.y],
|
||
[[%%
|
||
statement: struct_stat;
|
||
struct_stat: /* empty. */ | if else;
|
||
if: "if" "const" "then" statement;
|
||
else: "else" statement;
|
||
%%
|
||
]])
|
||
|
||
AT_CHECK([bison -v input.y -o input.c])
|
||
|
||
# Check only the tables. We don't use --no-parser, because it is
|
||
# still to be implemented in the experimental branch of Bison.
|
||
[sed -n 's/ *$//;/^static const.*\[\] =/,/^}/p' input.c >tables.c]
|
||
|
||
AT_CHECK([[cat tables.c]], 0,
|
||
[[static const unsigned char yytranslate[] =
|
||
{
|
||
0, 2, 2, 2, 2, 2, 2, 2, 2, 2,
|
||
2, 2, 2, 2, 2, 2, 2, 2, 2, 2,
|
||
2, 2, 2, 2, 2, 2, 2, 2, 2, 2,
|
||
2, 2, 2, 2, 2, 2, 2, 2, 2, 2,
|
||
2, 2, 2, 2, 2, 2, 2, 2, 2, 2,
|
||
2, 2, 2, 2, 2, 2, 2, 2, 2, 2,
|
||
2, 2, 2, 2, 2, 2, 2, 2, 2, 2,
|
||
2, 2, 2, 2, 2, 2, 2, 2, 2, 2,
|
||
2, 2, 2, 2, 2, 2, 2, 2, 2, 2,
|
||
2, 2, 2, 2, 2, 2, 2, 2, 2, 2,
|
||
2, 2, 2, 2, 2, 2, 2, 2, 2, 2,
|
||
2, 2, 2, 2, 2, 2, 2, 2, 2, 2,
|
||
2, 2, 2, 2, 2, 2, 2, 2, 2, 2,
|
||
2, 2, 2, 2, 2, 2, 2, 2, 2, 2,
|
||
2, 2, 2, 2, 2, 2, 2, 2, 2, 2,
|
||
2, 2, 2, 2, 2, 2, 2, 2, 2, 2,
|
||
2, 2, 2, 2, 2, 2, 2, 2, 2, 2,
|
||
2, 2, 2, 2, 2, 2, 2, 2, 2, 2,
|
||
2, 2, 2, 2, 2, 2, 2, 2, 2, 2,
|
||
2, 2, 2, 2, 2, 2, 2, 2, 2, 2,
|
||
2, 2, 2, 2, 2, 2, 2, 2, 2, 2,
|
||
2, 2, 2, 2, 2, 2, 2, 2, 2, 2,
|
||
2, 2, 2, 2, 2, 2, 2, 2, 2, 2,
|
||
2, 2, 2, 2, 2, 2, 2, 2, 2, 2,
|
||
2, 2, 2, 2, 2, 2, 2, 2, 2, 2,
|
||
2, 2, 2, 2, 2, 2, 1, 2, 3, 4,
|
||
5, 6
|
||
};
|
||
static const unsigned char yyprhs[] =
|
||
{
|
||
0, 0, 3, 5, 6, 9, 14
|
||
};
|
||
static const signed char yyrhs[] =
|
||
{
|
||
8, 0, -1, 9, -1, -1, 10, 11, -1, 3,
|
||
4, 5, 8, -1, 6, 8, -1
|
||
};
|
||
static const unsigned char yyrline[] =
|
||
{
|
||
0, 2, 2, 3, 3, 4, 5
|
||
};
|
||
static const char *const yytname[] =
|
||
{
|
||
"$", "error", "$undefined.", "\"if\"", "\"const\"", "\"then\"",
|
||
"\"else\"", "$axiom", "statement", "struct_stat", "if", "else", 0
|
||
};
|
||
static const short yytoknum[] =
|
||
{
|
||
0, 256, 257, 258, 259, 260, 261, -1
|
||
};
|
||
static const unsigned char yyr1[] =
|
||
{
|
||
0, 7, 8, 9, 9, 10, 11
|
||
};
|
||
static const unsigned char yyr2[] =
|
||
{
|
||
0, 2, 1, 0, 2, 4, 2
|
||
};
|
||
static const short yydefact[] =
|
||
{
|
||
3, 0, 0, 2, 0, 0, 0, 3, 4, 3,
|
||
6, 5
|
||
};
|
||
static const short yydefgoto[] =
|
||
{
|
||
-1, 2, 3, 4, 8
|
||
};
|
||
static const short yypact[] =
|
||
{
|
||
-2, -1, 4,-32768, 0, 2,-32768, -2,-32768, -2,
|
||
-32768,-32768
|
||
};
|
||
static const short yypgoto[] =
|
||
{
|
||
-32768, -7,-32768,-32768,-32768
|
||
};
|
||
static const short yytable[] =
|
||
{
|
||
10, 1, 11, 5, 6, 0, 7, 9
|
||
};
|
||
static const short yycheck[] =
|
||
{
|
||
7, 3, 9, 4, 0, -1, 6, 5
|
||
};
|
||
static const unsigned char yystos[] =
|
||
{
|
||
0, 3, 8, 9, 10, 4, 0, 6, 11, 5,
|
||
8, 8
|
||
};
|
||
]])
|
||
|
||
AT_CLEANUP
|