Modify tokenizer to store last token in a wcstring

This commit is contained in:
ridiculousfish 2012-11-21 22:09:35 -08:00
parent f545fb2491
commit a3a7c48799
4 changed files with 18 additions and 72 deletions

View file

@ -95,7 +95,7 @@ static int indent(wcstring &out, const wcstring &in, int flags)
for (; tok_has_next(&tok); tok_next(&tok)) for (; tok_has_next(&tok); tok_next(&tok))
{ {
int type = tok_last_type(&tok); int type = tok_last_type(&tok);
wchar_t *last = tok_last(&tok); const wchar_t *last = tok_last(&tok);
switch (type) switch (type)
{ {

View file

@ -968,7 +968,7 @@ static void tokenize(const wchar_t * const buff, std::vector<int> &color, const
{ {
/*Parameter */ /*Parameter */
wchar_t *param = tok_last(&tok); const wchar_t *param = tok_last(&tok);
if (param[0] == L'-') if (param[0] == L'-')
{ {
if (wcscmp(param, L"--") == 0) if (wcscmp(param, L"--") == 0)

View file

@ -51,11 +51,6 @@ segments.
*/ */
#define SEP L" \n|\t;#\r<>^&" #define SEP L" \n|\t;#\r<>^&"
/**
Maximum length of a string containing a file descriptor number
*/
#define FD_STR_MAX_LEN 16
/** /**
Descriptions of all tokenizer errors Descriptions of all tokenizer errors
*/ */
@ -74,31 +69,7 @@ static const wchar_t *tok_desc[] =
N_(L"Redirect output to file if file does not exist"), N_(L"Redirect output to file if file does not exist"),
N_(L"Run job in background"), N_(L"Run job in background"),
N_(L"Comment") N_(L"Comment")
} };
;
/**
Tests if the tokenizer buffer is large enough to hold contents of
the specified length, and if not, reallocates the tokenizer buffer.
\return 0 if the system could not provide the memory needed, and 1 otherwise.
*/
static int check_size(tokenizer_t *tok, size_t len)
{
if (tok->last_len <= len)
{
wchar_t *tmp;
tok->last_len = len +1;
tmp = (wchar_t *)realloc(tok->last, sizeof(wchar_t)*tok->last_len);
if (tmp == 0)
{
wperror(L"realloc");
return 0;
}
tok->last = tmp;
}
return 1;
}
/** /**
Set the latest tokens string to be the specified error message Set the latest tokens string to be the specified error message
@ -107,14 +78,7 @@ static void tok_call_error(tokenizer_t *tok, int error_type, const wchar_t *erro
{ {
tok->last_type = TOK_ERROR; tok->last_type = TOK_ERROR;
tok->error = error_type; tok->error = error_type;
if (!check_size(tok, wcslen(error_message)+1)) tok->last_token = error_message;
{
if (tok->last != 0)
*tok->last=0;
return;
}
wcscpy(tok->last, error_message);
} }
int tok_get_error(tokenizer_t *tok) int tok_get_error(tokenizer_t *tok)
@ -123,7 +87,7 @@ int tok_get_error(tokenizer_t *tok)
} }
tokenizer_t::tokenizer_t(const wchar_t *b, tok_flags_t flags) : buff(NULL), orig_buff(NULL), last(NULL), last_type(0), last_len(0), last_pos(0), has_next(false), accept_unfinished(false), show_comments(false), last_quote(0), error(0), squash_errors(false), cached_lineno_offset(0), cached_lineno_count(0) tokenizer_t::tokenizer_t(const wchar_t *b, tok_flags_t flags) : buff(NULL), orig_buff(NULL), last_type(0), last_pos(0), has_next(false), accept_unfinished(false), show_comments(false), last_quote(0), error(0), squash_errors(false), cached_lineno_offset(0), cached_lineno_count(0)
{ {
/* We can only generate error messages on the main thread due to wgettext() thread safety issues. */ /* We can only generate error messages on the main thread due to wgettext() thread safety issues. */
@ -149,8 +113,6 @@ tokenizer_t::tokenizer_t(const wchar_t *b, tok_flags_t flags) : buff(NULL), orig
void tok_destroy(tokenizer_t *tok) void tok_destroy(tokenizer_t *tok)
{ {
CHECK(tok,); CHECK(tok,);
free(tok->last);
} }
int tok_last_type(tokenizer_t *tok) int tok_last_type(tokenizer_t *tok)
@ -161,11 +123,11 @@ int tok_last_type(tokenizer_t *tok)
return tok->last_type; return tok->last_type;
} }
wchar_t *tok_last(tokenizer_t *tok) const wchar_t *tok_last(tokenizer_t *tok)
{ {
CHECK(tok, 0); CHECK(tok, 0);
return tok->last; return tok->last_token.c_str();
} }
int tok_has_next(tokenizer_t *tok) int tok_has_next(tokenizer_t *tok)
@ -447,11 +409,7 @@ static void read_string(tokenizer_t *tok)
len = tok->buff - start; len = tok->buff - start;
if (!check_size(tok, len)) tok->last_token.assign(start, len);
return;
memcpy(tok->last, start, sizeof(wchar_t)*len);
tok->last[len] = L'\0';
tok->last_type = TOK_STRING; tok->last_type = TOK_STRING;
} }
@ -465,13 +423,10 @@ static void read_comment(tokenizer_t *tok)
start = tok->buff; start = tok->buff;
while (*(tok->buff)!= L'\n' && *(tok->buff)!= L'\0') while (*(tok->buff)!= L'\n' && *(tok->buff)!= L'\0')
tok->buff++; tok->buff++;
size_t len = tok->buff - start; size_t len = tok->buff - start;
if (!check_size(tok, len)) tok->last_token.assign(start, len);
return;
memcpy(tok->last, start, sizeof(wchar_t)*len);
tok->last[len] = L'\0';
tok->last_type = TOK_COMMENT; tok->last_type = TOK_COMMENT;
} }
@ -503,9 +458,8 @@ static void read_redirect(tokenizer_t *tok, int fd)
TOK_CALL_ERROR(tok, TOK_OTHER, PIPE_ERROR); TOK_CALL_ERROR(tok, TOK_OTHER, PIPE_ERROR);
return; return;
} }
check_size(tok, FD_STR_MAX_LEN);
tok->buff++; tok->buff++;
swprintf(tok->last, FD_STR_MAX_LEN, L"%d", fd); tok->last_token = to_string<int>(fd);
tok->last_type = TOK_PIPE; tok->last_type = TOK_PIPE;
return; return;
} }
@ -520,12 +474,7 @@ static void read_redirect(tokenizer_t *tok, int fd)
TOK_CALL_ERROR(tok, TOK_OTHER, REDIRECT_ERROR); TOK_CALL_ERROR(tok, TOK_OTHER, REDIRECT_ERROR);
} }
if (!check_size(tok, 2)) tok->last_token = to_string(fd);
{
return;
}
swprintf(tok->last, tok->last_len, L"%d", fd);
if (*tok->buff == L'&') if (*tok->buff == L'&')
{ {
@ -652,10 +601,7 @@ void tok_next(tokenizer_t *tok)
break; break;
case L'|': case L'|':
check_size(tok, 2); tok->last_token = L"1";
tok->last[0]=L'1';
tok->last[1]=L'\0';
tok->last_type = TOK_PIPE; tok->last_type = TOK_PIPE;
tok->buff++; tok->buff++;
break; break;

View file

@ -10,6 +10,7 @@
#define FISH_TOKENIZER_H #define FISH_TOKENIZER_H
#include <wchar.h> #include <wchar.h>
#include "common.h"
/** /**
Token types Token types
@ -72,13 +73,12 @@ struct tokenizer_t
const wchar_t *buff; const wchar_t *buff;
/** A copy of the original string */ /** A copy of the original string */
const wchar_t *orig_buff; const wchar_t *orig_buff;
/** A pointer to the last token*/ /** The last token */
wchar_t *last; wcstring last_token;
/** Type of last token*/ /** Type of last token*/
int last_type; int last_type;
/** Length of last token*/
size_t last_len;
/** Offset of last token*/ /** Offset of last token*/
size_t last_pos; size_t last_pos;
/** Whether there are more tokens*/ /** Whether there are more tokens*/
@ -128,7 +128,7 @@ int tok_last_type(tokenizer_t *tok);
/** /**
Returns the last token string. The string should not be freed by the caller. Returns the last token string. The string should not be freed by the caller.
*/ */
wchar_t *tok_last(tokenizer_t *tok); const wchar_t *tok_last(tokenizer_t *tok);
/** /**
Returns the type of quote from the last TOK_QSTRING Returns the type of quote from the last TOK_QSTRING