diff options
author | Michael Muré <batolettre@gmail.com> | 2020-08-20 12:00:34 +0200 |
---|---|---|
committer | Mike Goldin <mike.goldin@protonmail.ch> | 2020-11-17 08:09:51 -0500 |
commit | b285c57dc62caac2c1f09e74eeece406b0e7cc00 (patch) | |
tree | aa2f15c8e3a250c8bd8d8f4a7b11a5c9819453a8 /query | |
parent | 902997f53771babb2f9ea1bb6288c2ec295c4c9e (diff) | |
download | git-bug-b285c57dc62caac2c1f09e74eeece406b0e7cc00.tar.gz |
query: expand the tokenizer/parser to parse arbitrary search terms
Diffstat (limited to 'query')
-rw-r--r-- | query/lexer.go | 43 | ||||
-rw-r--r-- | query/lexer_test.go | 33 | ||||
-rw-r--r-- | query/parser.go | 73 | ||||
-rw-r--r-- | query/parser_test.go | 14 | ||||
-rw-r--r-- | query/query.go | 3 |
5 files changed, 116 insertions, 50 deletions
diff --git a/query/lexer.go b/query/lexer.go index ca67d641..5ca700c7 100644 --- a/query/lexer.go +++ b/query/lexer.go @@ -6,9 +6,38 @@ import ( "unicode" ) +type tokenKind int + +const ( + _ tokenKind = iota + tokenKindKV + tokenKindSearch +) + type token struct { + kind tokenKind + + // KV qualifier string value string + + // Search + term string +} + +func newTokenKV(qualifier, value string) token { + return token{ + kind: tokenKindKV, + qualifier: qualifier, + value: value, + } +} + +func newTokenSearch(term string) token { + return token{ + kind: tokenKindSearch, + term: term, + } } // tokenize parse and break a input into tokens ready to be @@ -22,6 +51,13 @@ func tokenize(query string) ([]token, error) { var tokens []token for _, field := range fields { split := strings.Split(field, ":") + + // full text search + if len(split) == 1 { + tokens = append(tokens, newTokenSearch(removeQuote(field))) + continue + } + if len(split) != 2 { return nil, fmt.Errorf("can't tokenize \"%s\"", field) } @@ -33,14 +69,13 @@ func tokenize(query string) ([]token, error) { return nil, fmt.Errorf("empty value for qualifier \"%s\"", split[0]) } - tokens = append(tokens, token{ - qualifier: split[0], - value: removeQuote(split[1]), - }) + tokens = append(tokens, newTokenKV(split[0], removeQuote(split[1]))) } return tokens, nil } +// split the query into chunks by splitting on whitespaces but respecting +// quotes func splitQuery(query string) ([]string, error) { lastQuote := rune(0) inQuote := false diff --git a/query/lexer_test.go b/query/lexer_test.go index 36b9ba10..59f17dec 100644 --- a/query/lexer_test.go +++ b/query/lexer_test.go @@ -11,32 +11,43 @@ func TestTokenize(t *testing.T) { input string tokens []token }{ - {"gibberish", nil}, {"status:", nil}, {":value", nil}, - {"status:open", []token{{"status", "open"}}}, - {"status:closed", []token{{"status", "closed"}}}, + {"status:open", []token{newTokenKV("status", "open")}}, + {"status:closed", []token{newTokenKV("status", "closed")}}, - {"author:rene", []token{{"author", "rene"}}}, - {`author:"René Descartes"`, []token{{"author", "René Descartes"}}}, + {"author:rene", []token{newTokenKV("author", "rene")}}, + {`author:"René Descartes"`, []token{newTokenKV("author", "René Descartes")}}, { `status:open status:closed author:rene author:"René Descartes"`, []token{ - {"status", "open"}, - {"status", "closed"}, - {"author", "rene"}, - {"author", "René Descartes"}, + newTokenKV("status", "open"), + newTokenKV("status", "closed"), + newTokenKV("author", "rene"), + newTokenKV("author", "René Descartes"), }, }, // quotes - {`key:"value value"`, []token{{"key", "value value"}}}, - {`key:'value value'`, []token{{"key", "value value"}}}, + {`key:"value value"`, []token{newTokenKV("key", "value value")}}, + {`key:'value value'`, []token{newTokenKV("key", "value value")}}, // unmatched quotes {`key:'value value`, nil}, {`key:value value'`, nil}, + + // full text search + {"search", []token{newTokenSearch("search")}}, + {"search more terms", []token{ + newTokenSearch("search"), + newTokenSearch("more"), + newTokenSearch("terms"), + }}, + {"search \"more terms\"", []token{ + newTokenSearch("search"), + newTokenSearch("more terms"), + }}, } for _, tc := range tests { diff --git a/query/parser.go b/query/parser.go index a379f750..762a47e5 100644 --- a/query/parser.go +++ b/query/parser.go @@ -24,42 +24,49 @@ func Parse(query string) (*Query, error) { sortingDone := false for _, t := range tokens { - switch t.qualifier { - case "status", "state": - status, err := bug.StatusFromString(t.value) - if err != nil { - return nil, err - } - q.Status = append(q.Status, status) - case "author": - q.Author = append(q.Author, t.value) - case "actor": - q.Actor = append(q.Actor, t.value) - case "participant": - q.Participant = append(q.Participant, t.value) - case "label": - q.Label = append(q.Label, t.value) - case "title": - q.Title = append(q.Title, t.value) - case "no": - switch t.value { + switch t.kind { + case tokenKindSearch: + q.Search = append(q.Search, t.term) + break + + case tokenKindKV: + switch t.qualifier { + case "status", "state": + status, err := bug.StatusFromString(t.value) + if err != nil { + return nil, err + } + q.Status = append(q.Status, status) + case "author": + q.Author = append(q.Author, t.value) + case "actor": + q.Actor = append(q.Actor, t.value) + case "participant": + q.Participant = append(q.Participant, t.value) case "label": - q.NoLabel = true + q.Label = append(q.Label, t.value) + case "title": + q.Title = append(q.Title, t.value) + case "no": + switch t.value { + case "label": + q.NoLabel = true + default: + return nil, fmt.Errorf("unknown \"no\" filter \"%s\"", t.value) + } + case "sort": + if sortingDone { + return nil, fmt.Errorf("multiple sorting") + } + err = parseSorting(q, t.value) + if err != nil { + return nil, err + } + sortingDone = true + default: - return nil, fmt.Errorf("unknown \"no\" filter \"%s\"", t.value) + return nil, fmt.Errorf("unknown qualifier \"%s\"", t.qualifier) } - case "sort": - if sortingDone { - return nil, fmt.Errorf("multiple sorting") - } - err = parseSorting(q, t.value) - if err != nil { - return nil, err - } - sortingDone = true - - default: - return nil, fmt.Errorf("unknown qualifier \"%s\"", t.qualifier) } } return q, nil diff --git a/query/parser_test.go b/query/parser_test.go index 6a509adb..87dd870a 100644 --- a/query/parser_test.go +++ b/query/parser_test.go @@ -13,7 +13,7 @@ func TestParse(t *testing.T) { input string output *Query }{ - {"gibberish", nil}, + // KV {"status:", nil}, {":value", nil}, @@ -62,8 +62,18 @@ func TestParse(t *testing.T) { }}, {"sort:unknown", nil}, - {`status:open author:"René Descartes" participant:leonhard label:hello label:"Good first issue" sort:edit-desc`, + // Search + {"search", &Query{ + Search: []string{"search"}, + }}, + {"search \"more terms\"", &Query{ + Search: []string{"search", "more terms"}, + }}, + + // Complex + {`status:open author:"René Descartes" search participant:leonhard label:hello label:"Good first issue" sort:edit-desc "more terms"`, &Query{ + Search: []string{"search", "more terms"}, Filters: Filters{ Status: []bug.Status{bug.OpenStatus}, Author: []string{"René Descartes"}, diff --git a/query/query.go b/query/query.go index a499ad38..816d6414 100644 --- a/query/query.go +++ b/query/query.go @@ -7,6 +7,7 @@ import "github.com/MichaelMure/git-bug/bug" // manually. This query doesn't do anything by itself and need to be interpreted // for the specific domain of application. type Query struct { + Search Filters OrderBy OrderDirection @@ -20,6 +21,8 @@ func NewQuery() *Query { } } +type Search []string + // Filters is a collection of Filter that implement a complex filter type Filters struct { Status []bug.Status |