Some reformatting in query.h
This commit is contained in:
parent
b06df8533b
commit
4bc830a2c5
1 changed files with 37 additions and 35 deletions
20
src/query.h
20
src/query.h
|
|
@ -96,14 +96,14 @@ public:
|
||||||
explicit token_t(kind_t _kind = UNKNOWN,
|
explicit token_t(kind_t _kind = UNKNOWN,
|
||||||
const optional<string>& _value = none)
|
const optional<string>& _value = none)
|
||||||
: kind(_kind), value(_value) {
|
: kind(_kind), value(_value) {
|
||||||
TRACE_CTOR(lexer_t::token_t, "");
|
TRACE_CTOR(query_t::lexer_t::token_t, "");
|
||||||
}
|
}
|
||||||
token_t(const token_t& tok)
|
token_t(const token_t& tok)
|
||||||
: kind(tok.kind), value(tok.value) {
|
: kind(tok.kind), value(tok.value) {
|
||||||
TRACE_CTOR(lexer_t::token_t, "copy");
|
TRACE_CTOR(query_t::lexer_t::token_t, "copy");
|
||||||
}
|
}
|
||||||
~token_t() throw() {
|
~token_t() throw() {
|
||||||
TRACE_DTOR(lexer_t::token_t);
|
TRACE_DTOR(query_t::lexer_t::token_t);
|
||||||
}
|
}
|
||||||
|
|
||||||
token_t& operator=(const token_t& tok) {
|
token_t& operator=(const token_t& tok) {
|
||||||
|
|
@ -137,6 +137,8 @@ public:
|
||||||
case TERM: return string("TERM(") + *value + ")";
|
case TERM: return string("TERM(") + *value + ")";
|
||||||
case END_REACHED: return "END_REACHED";
|
case END_REACHED: return "END_REACHED";
|
||||||
}
|
}
|
||||||
|
assert(false);
|
||||||
|
return empty_string;
|
||||||
}
|
}
|
||||||
|
|
||||||
string symbol() const {
|
string symbol() const {
|
||||||
|
|
@ -180,7 +182,7 @@ public:
|
||||||
consume_whitespace(false),
|
consume_whitespace(false),
|
||||||
consume_next_arg(false)
|
consume_next_arg(false)
|
||||||
{
|
{
|
||||||
TRACE_CTOR(lexer_t, "");
|
TRACE_CTOR(query_t::lexer_t, "");
|
||||||
assert(begin != end);
|
assert(begin != end);
|
||||||
arg_i = (*begin).as_string().begin();
|
arg_i = (*begin).as_string().begin();
|
||||||
arg_end = (*begin).as_string().end();
|
arg_end = (*begin).as_string().end();
|
||||||
|
|
@ -192,10 +194,10 @@ public:
|
||||||
consume_next_arg(lexer.consume_next_arg),
|
consume_next_arg(lexer.consume_next_arg),
|
||||||
token_cache(lexer.token_cache)
|
token_cache(lexer.token_cache)
|
||||||
{
|
{
|
||||||
TRACE_CTOR(lexer_t, "copy");
|
TRACE_CTOR(query_t::lexer_t, "copy");
|
||||||
}
|
}
|
||||||
~lexer_t() throw() {
|
~lexer_t() throw() {
|
||||||
TRACE_DTOR(lexer_t);
|
TRACE_DTOR(query_t::lexer_t);
|
||||||
}
|
}
|
||||||
|
|
||||||
token_t next_token();
|
token_t next_token();
|
||||||
|
|
@ -227,14 +229,14 @@ protected:
|
||||||
public:
|
public:
|
||||||
parser_t(const value_t& _args)
|
parser_t(const value_t& _args)
|
||||||
: args(_args), lexer(args.begin(), args.end()) {
|
: args(_args), lexer(args.begin(), args.end()) {
|
||||||
TRACE_CTOR(parser_t, "");
|
TRACE_CTOR(query_t::parser_t, "");
|
||||||
}
|
}
|
||||||
parser_t(const parser_t& parser)
|
parser_t(const parser_t& parser)
|
||||||
: args(parser.args), lexer(parser.lexer) {
|
: args(parser.args), lexer(parser.lexer) {
|
||||||
TRACE_CTOR(parser_t, "copy");
|
TRACE_CTOR(query_t::parser_t, "copy");
|
||||||
}
|
}
|
||||||
~parser_t() throw() {
|
~parser_t() throw() {
|
||||||
TRACE_DTOR(parser_t);
|
TRACE_DTOR(query_t::parser_t);
|
||||||
}
|
}
|
||||||
|
|
||||||
expr_t::ptr_op_t parse() {
|
expr_t::ptr_op_t parse() {
|
||||||
|
|
|
||||||
Loading…
Add table
Reference in a new issue