X-Git-Url: https://git.piment-noir.org/?p=TP_AL_C.git;a=blobdiff_plain;f=lexer%2Fsyntactic_analyzer.c;h=ddef90b12707dc5f09009e2d62f182a46899cd0c;hp=29524f01f7b39d82d2328f3c773be5565941d30e;hb=f196dae589431452a3eb37a2aa985d882f695407;hpb=7cfc17fb01df603682946e58acc973f7d9e5f0bc diff --git a/lexer/syntactic_analyzer.c b/lexer/syntactic_analyzer.c index 29524f0..ddef90b 100644 --- a/lexer/syntactic_analyzer.c +++ b/lexer/syntactic_analyzer.c @@ -25,7 +25,7 @@ static void analyze_TEXT() { fputws(L" ", target); } #if PRINT_TOKEN - wprint_token_stdout(); + wprint_token(stdout); #endif token[tokenFound].type = tokenTypestr[tokenType]; tokenFound++; @@ -48,7 +48,7 @@ static void analyze_P() { if (tokenType == NPARA) { scanner(); #if PRINT_TOKEN - wprint_token_stdout(); + wprint_token(stdout); #endif if (target != NULL) { fputws(L"

\n", target); @@ -60,7 +60,7 @@ static void analyze_P() { if (tokenType == MOT) { scanner(); #if PRINT_TOKEN - wprint_token_stdout(); + wprint_token(stdout); #endif if (target != NULL) { fputws((const wchar_t*)token[tokenFound].value, target); @@ -91,7 +91,7 @@ static void analyze_HEAD() { if (tokenType == MOTCLE) { scanner(); #if PRINT_TOKEN - wprint_token_stdout(); + wprint_token(stdout); #endif /* PRINT_TOKEN */ token[tokenFound].type = tokenTypestr[tokenType]; tokenFound++; @@ -108,7 +108,7 @@ static void analyze_HEAD() { if (tokenType == MOTCLE) { scanner(); #if PRINT_TOKEN - wprint_token_stdout(); + wprint_token(stdout); #endif /* PRINT_TOKEN */ token[tokenFound].type = tokenTypestr[tokenType]; tokenFound++; @@ -132,7 +132,7 @@ static void analyze_H1() { if (tokenType == SECTION) { scanner(); #if PRINT_TOKEN - wprint_token_stdout(); + wprint_token(stdout); #endif /* PRINT_TOKEN */ if (target != NULL) { fputws(L"

\n", target); @@ -158,7 +158,7 @@ static void analyze_H2() { if (tokenType == SSECTION) { scanner(); #if PRINT_TOKEN - wprint_token_stdout(); + wprint_token(stdout); #endif /* PRINT_TOKEN */ if (target != NULL) { fputws(L"

\n", target); @@ -234,7 +234,7 @@ void analyze_AXIOME() { scanner(); /* print the lexical analysis result */ #if PRINT_TOKEN - wprint_token_stdout(); + wprint_token(stdout); #endif /* PRINT_TOKEN */ token[tokenFound].type = tokenTypestr[tokenType]; tokenFound++;