repositories
/
TP_AL_C.git
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
Simplify a bit a print helper function.
[TP_AL_C.git]
/
lexer
/
syntactic_analyzer.c
diff --git
a/lexer/syntactic_analyzer.c
b/lexer/syntactic_analyzer.c
index 29524f01f7b39d82d2328f3c773be5565941d30e..ddef90b12707dc5f09009e2d62f182a46899cd0c 100644
(file)
--- a/
lexer/syntactic_analyzer.c
+++ b/
lexer/syntactic_analyzer.c
@@
-25,7
+25,7
@@
static void analyze_TEXT() {
fputws(L" ", target);
}
#if PRINT_TOKEN
fputws(L" ", target);
}
#if PRINT_TOKEN
- wprint_token
_stdout(
);
+ wprint_token
(stdout
);
#endif
token[tokenFound].type = tokenTypestr[tokenType];
tokenFound++;
#endif
token[tokenFound].type = tokenTypestr[tokenType];
tokenFound++;
@@
-48,7
+48,7
@@
static void analyze_P() {
if (tokenType == NPARA) {
scanner();
#if PRINT_TOKEN
if (tokenType == NPARA) {
scanner();
#if PRINT_TOKEN
- wprint_token
_stdout(
);
+ wprint_token
(stdout
);
#endif
if (target != NULL) {
fputws(L"<p>\n", target);
#endif
if (target != NULL) {
fputws(L"<p>\n", target);
@@
-60,7
+60,7
@@
static void analyze_P() {
if (tokenType == MOT) {
scanner();
#if PRINT_TOKEN
if (tokenType == MOT) {
scanner();
#if PRINT_TOKEN
- wprint_token
_stdout(
);
+ wprint_token
(stdout
);
#endif
if (target != NULL) {
fputws((const wchar_t*)token[tokenFound].value, target);
#endif
if (target != NULL) {
fputws((const wchar_t*)token[tokenFound].value, target);
@@
-91,7
+91,7
@@
static void analyze_HEAD() {
if (tokenType == MOTCLE) {
scanner();
#if PRINT_TOKEN
if (tokenType == MOTCLE) {
scanner();
#if PRINT_TOKEN
- wprint_token
_stdout(
);
+ wprint_token
(stdout
);
#endif /* PRINT_TOKEN */
token[tokenFound].type = tokenTypestr[tokenType];
tokenFound++;
#endif /* PRINT_TOKEN */
token[tokenFound].type = tokenTypestr[tokenType];
tokenFound++;
@@
-108,7
+108,7
@@
static void analyze_HEAD() {
if (tokenType == MOTCLE) {
scanner();
#if PRINT_TOKEN
if (tokenType == MOTCLE) {
scanner();
#if PRINT_TOKEN
- wprint_token
_stdout(
);
+ wprint_token
(stdout
);
#endif /* PRINT_TOKEN */
token[tokenFound].type = tokenTypestr[tokenType];
tokenFound++;
#endif /* PRINT_TOKEN */
token[tokenFound].type = tokenTypestr[tokenType];
tokenFound++;
@@
-132,7
+132,7
@@
static void analyze_H1() {
if (tokenType == SECTION) {
scanner();
#if PRINT_TOKEN
if (tokenType == SECTION) {
scanner();
#if PRINT_TOKEN
- wprint_token
_stdout(
);
+ wprint_token
(stdout
);
#endif /* PRINT_TOKEN */
if (target != NULL) {
fputws(L"<h1>\n", target);
#endif /* PRINT_TOKEN */
if (target != NULL) {
fputws(L"<h1>\n", target);
@@
-158,7
+158,7
@@
static void analyze_H2() {
if (tokenType == SSECTION) {
scanner();
#if PRINT_TOKEN
if (tokenType == SSECTION) {
scanner();
#if PRINT_TOKEN
- wprint_token
_stdout(
);
+ wprint_token
(stdout
);
#endif /* PRINT_TOKEN */
if (target != NULL) {
fputws(L"<h2>\n", target);
#endif /* PRINT_TOKEN */
if (target != NULL) {
fputws(L"<h2>\n", target);
@@
-234,7
+234,7
@@
void analyze_AXIOME() {
scanner();
/* print the lexical analysis result */
#if PRINT_TOKEN
scanner();
/* print the lexical analysis result */
#if PRINT_TOKEN
- wprint_token
_stdout(
);
+ wprint_token
(stdout
);
#endif /* PRINT_TOKEN */
token[tokenFound].type = tokenTypestr[tokenType];
tokenFound++;
#endif /* PRINT_TOKEN */
token[tokenFound].type = tokenTypestr[tokenType];
tokenFound++;