Features: Filter Aggregations, Nested Aggregations, Multi Match Queries, Highlights (#12)
* Add support for multi_match queries * Add support for highlights * Add support for nested aggregations and filtered aggregations * Update README * Fix formatting
This commit is contained in:
parent
259a3cd818
commit
fc16427405
18
README.md
18
README.md
|
@ -124,6 +124,7 @@ The following queries are currently supported:
|
|||
| `"match_phrase_prefix"` | `MatchPhrasePrefix()` |
|
||||
| `"match_all"` | `MatchAll()` |
|
||||
| `"match_none"` | `MatchNone()` |
|
||||
| `"multi_match"` | `MultiMatch()` |
|
||||
| `"exists"` | `Exists()` |
|
||||
| `"fuzzy"` | `Fuzzy()` |
|
||||
| `"ids"` | `IDs()` |
|
||||
|
@ -158,6 +159,23 @@ The following aggregations are currently supported:
|
|||
| `"top_hits"` | `TopHits()` |
|
||||
| `"terms"` | `TermsAgg()` |
|
||||
|
||||
### Supported Top Level Options
|
||||
|
||||
The following top level options are currently supported:
|
||||
|
||||
| ElasticSearch DSL | `esquery.Search` Function |
|
||||
| ------------------------|--------------------------------------- |
|
||||
| `"highlight"` | `Highlight()` |
|
||||
| `"explain"` | `Explain()` |
|
||||
| `"from"` | `From()` |
|
||||
| `"postFilter"` | `PostFilter()` |
|
||||
| `"query"` | `Query()` |
|
||||
| `"aggs"` | `Aggs()` |
|
||||
| `"size"` | `Size()` |
|
||||
| `"sort"` | `Sort()` |
|
||||
| `"source"` | `SourceIncludes(), SourceExcludes()` |
|
||||
| `"timeout"` | `Timeout()` |
|
||||
|
||||
#### Custom Queries and Aggregations
|
||||
|
||||
To execute an arbitrary query or aggregation (including those not yet supported by the library), use the `CustomQuery()` or `CustomAgg()` functions, respectively. Both accept any `map[string]interface{}` value.
|
||||
|
|
|
@ -58,5 +58,39 @@ func TestAggregations(t *testing.T) {
|
|||
},
|
||||
},
|
||||
},
|
||||
{
|
||||
"a complex, multi-aggregation, nested",
|
||||
Aggregate(
|
||||
NestedAgg("categories","categories").
|
||||
Aggs(TermsAgg("type","outdoors")),
|
||||
FilterAgg("filtered",
|
||||
Term("type", "t-shirt")),
|
||||
),
|
||||
map[string]interface{}{
|
||||
"aggs": map[string]interface{}{
|
||||
"categories": map[string]interface{}{
|
||||
"nested": map[string]interface{}{
|
||||
"path": "categories",
|
||||
},
|
||||
"aggs": map[string]interface{} {
|
||||
"type": map[string]interface{} {
|
||||
"terms": map[string]interface{} {
|
||||
"field": "outdoors",
|
||||
},
|
||||
},
|
||||
},
|
||||
},
|
||||
"filtered": map[string]interface{}{
|
||||
"filter": map[string]interface{}{
|
||||
"term": map[string]interface{}{
|
||||
"type": map[string]interface{} {
|
||||
"value": "t-shirt",
|
||||
},
|
||||
},
|
||||
},
|
||||
},
|
||||
},
|
||||
},
|
||||
},
|
||||
})
|
||||
}
|
||||
|
|
|
@ -0,0 +1,49 @@
|
|||
package esquery
|
||||
|
||||
type FilterAggregation struct {
|
||||
name string
|
||||
filter Mappable
|
||||
aggs []Aggregation
|
||||
}
|
||||
|
||||
// Filter creates a new aggregation of type "filter". The method name includes
|
||||
// the "Agg" suffix to prevent conflict with the "filter" query.
|
||||
func FilterAgg(name string, filter Mappable) *FilterAggregation {
|
||||
return &FilterAggregation{
|
||||
name: name,
|
||||
filter: filter,
|
||||
}
|
||||
}
|
||||
|
||||
// Name returns the name of the aggregation.
|
||||
func (agg *FilterAggregation) Name() string {
|
||||
return agg.name
|
||||
}
|
||||
|
||||
// Filter sets the filter items
|
||||
func (agg *FilterAggregation) Filter(filter Mappable) *FilterAggregation {
|
||||
agg.filter = filter
|
||||
return agg
|
||||
}
|
||||
|
||||
// Aggs sets sub-aggregations for the aggregation.
|
||||
func (agg *FilterAggregation) Aggs(aggs ...Aggregation) *FilterAggregation {
|
||||
agg.aggs = aggs
|
||||
return agg
|
||||
}
|
||||
|
||||
func (agg *FilterAggregation) Map() map[string]interface{} {
|
||||
outerMap := map[string]interface{}{
|
||||
"filter": agg.filter.Map(),
|
||||
}
|
||||
|
||||
if len(agg.aggs) > 0 {
|
||||
subAggs := make(map[string]map[string]interface{})
|
||||
for _, sub := range agg.aggs {
|
||||
subAggs[sub.Name()] = sub.Map()
|
||||
}
|
||||
outerMap["aggs"] = subAggs
|
||||
}
|
||||
|
||||
return outerMap
|
||||
}
|
|
@ -0,0 +1,42 @@
|
|||
package esquery
|
||||
|
||||
import "testing"
|
||||
|
||||
func TestFilterAggs(t *testing.T) {
|
||||
runMapTests(t, []mapTest{
|
||||
{
|
||||
"filter agg: simple",
|
||||
FilterAgg("filtered", Term("type", "t-shirt")),
|
||||
map[string]interface{}{
|
||||
"filter": map[string]interface{}{
|
||||
"term": map[string]interface{}{
|
||||
"type": map[string]interface{}{
|
||||
"value": "t-shirt",
|
||||
},
|
||||
},
|
||||
},
|
||||
},
|
||||
},
|
||||
{
|
||||
"filter agg: with aggs",
|
||||
FilterAgg("filtered", Term("type", "t-shirt")).
|
||||
Aggs(Avg("avg_price", "price")),
|
||||
map[string]interface{}{
|
||||
"filter": map[string]interface{}{
|
||||
"term": map[string]interface{}{
|
||||
"type": map[string]interface{}{
|
||||
"value": "t-shirt",
|
||||
},
|
||||
},
|
||||
},
|
||||
"aggs": map[string]interface{}{
|
||||
"avg_price": map[string]interface{}{
|
||||
"avg": map[string]interface{}{
|
||||
"field": "price",
|
||||
},
|
||||
},
|
||||
},
|
||||
},
|
||||
},
|
||||
})
|
||||
}
|
|
@ -0,0 +1,53 @@
|
|||
package esquery
|
||||
|
||||
type NestedAggregation struct {
|
||||
name string
|
||||
path string
|
||||
aggs []Aggregation
|
||||
}
|
||||
|
||||
// NestedAgg creates a new aggregation of type "nested". The method name includes
|
||||
// the "Agg" suffix to prevent conflict with the "nested" query.
|
||||
func NestedAgg(name string, path string) *NestedAggregation {
|
||||
return &NestedAggregation{
|
||||
name: name,
|
||||
path: path,
|
||||
}
|
||||
}
|
||||
|
||||
// Name returns the name of the aggregation.
|
||||
func (agg *NestedAggregation) Name() string {
|
||||
return agg.name
|
||||
}
|
||||
|
||||
// NumberOfFragments sets the aggregations path
|
||||
func (agg *NestedAggregation) Path(p string) *NestedAggregation {
|
||||
agg.path = p
|
||||
return agg
|
||||
}
|
||||
|
||||
// Aggs sets sub-aggregations for the aggregation.
|
||||
func (agg *NestedAggregation) Aggs(aggs ...Aggregation) *NestedAggregation {
|
||||
agg.aggs = aggs
|
||||
return agg
|
||||
}
|
||||
|
||||
func (agg *NestedAggregation) Map() map[string]interface{} {
|
||||
innerMap := map[string]interface{}{
|
||||
"path": agg.path,
|
||||
}
|
||||
|
||||
outerMap := map[string]interface{}{
|
||||
"nested": innerMap,
|
||||
}
|
||||
|
||||
if len(agg.aggs) > 0 {
|
||||
subAggs := make(map[string]map[string]interface{})
|
||||
for _, sub := range agg.aggs {
|
||||
subAggs[sub.Name()] = sub.Map()
|
||||
}
|
||||
outerMap["aggs"] = subAggs
|
||||
}
|
||||
|
||||
return outerMap
|
||||
}
|
|
@ -0,0 +1,34 @@
|
|||
package esquery
|
||||
|
||||
import "testing"
|
||||
|
||||
func TestNestedAggs(t *testing.T) {
|
||||
runMapTests(t, []mapTest{
|
||||
{
|
||||
"nested agg: simple",
|
||||
NestedAgg("simple", "categories"),
|
||||
map[string]interface{}{
|
||||
"nested": map[string]interface{}{
|
||||
"path": "categories",
|
||||
},
|
||||
},
|
||||
},
|
||||
{
|
||||
"nested agg: with aggs",
|
||||
NestedAgg("more_nested", "authors").
|
||||
Aggs(TermsAgg("authors", "name")),
|
||||
map[string]interface{}{
|
||||
"nested": map[string]interface{}{
|
||||
"path": "authors",
|
||||
},
|
||||
"aggs": map[string]interface{}{
|
||||
"authors": map[string]interface{}{
|
||||
"terms": map[string]interface{}{
|
||||
"field": "name",
|
||||
},
|
||||
},
|
||||
},
|
||||
},
|
||||
},
|
||||
})
|
||||
}
|
|
@ -0,0 +1,341 @@
|
|||
package esquery
|
||||
|
||||
import (
|
||||
"github.com/fatih/structs"
|
||||
)
|
||||
|
||||
// Map returns a map representation of the highlight; implementing the
|
||||
// Mappable interface.
|
||||
func (q *QueryHighlight) Map() map[string]interface{} {
|
||||
results := structs.Map(q.params)
|
||||
if q.highlightQuery != nil {
|
||||
results["query"] = q.highlightQuery.Map()
|
||||
}
|
||||
if q.fields != nil && len(q.fields) > 0 {
|
||||
fields := make(map[string]interface{})
|
||||
for k, v := range q.fields {
|
||||
fields[k] = v.Map()
|
||||
}
|
||||
results["fields"] = fields
|
||||
}
|
||||
return results
|
||||
}
|
||||
|
||||
type QueryHighlight struct {
|
||||
highlightQuery Mappable `structs:"highlight_query,omitempty"`
|
||||
fields map[string]*QueryHighlight `structs:"fields"`
|
||||
params highlighParams
|
||||
}
|
||||
|
||||
type highlighParams struct {
|
||||
PreTags []string `structs:"pre_tags,omitempty"`
|
||||
PostTags []string `structs:"post_tags,omitempty"`
|
||||
|
||||
FragmentSize uint16 `structs:"fragment_size,omitempty"`
|
||||
NumberOfFragments uint16 `structs:"number_of_fragments,omitempty"`
|
||||
Type HighlightType `structs:"type,string,omitempty"`
|
||||
BoundaryChars string `structs:"boundary_chars,omitempty"`
|
||||
BoundaryMaxScan uint16 `structs:"boundary_max_scan,omitempty"`
|
||||
BoundaryScanner HighlightBoundaryScanner `structs:"boundary_scanner,string,omitempty"`
|
||||
BoundaryScannerLocale string `structs:"boundary_scanner_locale,omitempty"`
|
||||
Encoder HighlightEncoder `structs:"encoder,string,omitempty"`
|
||||
ForceSource *bool `structs:"force_source,omitempty"`
|
||||
Fragmenter HighlightFragmenter `structs:"fragmenter,string,omitempty"`
|
||||
FragmentOffset uint16 `structs:"fragment_offset,omitempty"`
|
||||
MatchedFields []string `structs:"matched_fields,omitempty"`
|
||||
NoMatchSize uint16 `structs:"no_match_size,omitempty"`
|
||||
Order HighlightOrder `structs:"order,string,omitempty"`
|
||||
PhraseLimit uint16 `structs:"phrase_limit,omitempty"`
|
||||
RequireFieldMatch *bool `structs:"require_field_match,omitempty"`
|
||||
TagsSchema HighlightTagsSchema `structs:"tags_schema,string,omitempty"`
|
||||
}
|
||||
|
||||
// Highlight creates a new "query" of type "highlight"
|
||||
func Highlight() *QueryHighlight {
|
||||
return newHighlight()
|
||||
}
|
||||
|
||||
func newHighlight() *QueryHighlight {
|
||||
return &QueryHighlight{
|
||||
fields: make(map[string]*QueryHighlight),
|
||||
params: highlighParams{},
|
||||
}
|
||||
}
|
||||
|
||||
// PreTags sets the highlight query's pre_tags ignore unmapped field
|
||||
func (q *QueryHighlight) PreTags(s ...string) *QueryHighlight {
|
||||
q.params.PreTags = append(q.params.PreTags, s...)
|
||||
return q
|
||||
}
|
||||
|
||||
// PostTags sets the highlight query's post_tags ignore unmapped field
|
||||
func (q *QueryHighlight) PostTags(s ...string) *QueryHighlight {
|
||||
q.params.PostTags = append(q.params.PostTags, s...)
|
||||
return q
|
||||
}
|
||||
|
||||
// Field sets an entry the highlight query's fields
|
||||
func (q *QueryHighlight) Field(name string, h ...*QueryHighlight) *QueryHighlight {
|
||||
var fld *QueryHighlight
|
||||
if len(h) > 0 {
|
||||
fld = h[len(h)-1]
|
||||
} else {
|
||||
fld = &QueryHighlight{}
|
||||
}
|
||||
q.fields[name] = fld
|
||||
return q
|
||||
}
|
||||
|
||||
// Fields sets all entries for the highlight query's fields
|
||||
func (q *QueryHighlight) Fields(h map[string]*QueryHighlight) *QueryHighlight {
|
||||
q.fields = h
|
||||
return q
|
||||
}
|
||||
|
||||
// FragmentSize sets the highlight query's fragment_size ignore unmapped field
|
||||
func (q *QueryHighlight) FragmentSize(i uint16) *QueryHighlight {
|
||||
q.params.FragmentSize = i
|
||||
return q
|
||||
}
|
||||
|
||||
// NumberOfFragments sets the highlight query's number_of_fragments ignore unmapped field
|
||||
func (q *QueryHighlight) NumberOfFragments(i uint16) *QueryHighlight {
|
||||
q.params.NumberOfFragments = i
|
||||
return q
|
||||
}
|
||||
|
||||
// Type sets the highlight query's type ignore unmapped field
|
||||
func (q *QueryHighlight) Type(t HighlightType) *QueryHighlight {
|
||||
q.params.Type = t
|
||||
return q
|
||||
}
|
||||
|
||||
// BoundaryChars sets the highlight query's boundary_chars ignore unmapped field
|
||||
func (q *QueryHighlight) BoundaryChars(s string) *QueryHighlight {
|
||||
q.params.BoundaryChars = s
|
||||
return q
|
||||
}
|
||||
|
||||
// BoundaryMaxScan sets the highlight query's boundary_max_scan ignore unmapped field
|
||||
func (q *QueryHighlight) BoundaryMaxScan(i uint16) *QueryHighlight {
|
||||
q.params.BoundaryMaxScan = i
|
||||
return q
|
||||
}
|
||||
|
||||
// BoundaryScanner sets the highlight query's boundary_scanner ignore unmapped field
|
||||
func (q *QueryHighlight) BoundaryScanner(t HighlightBoundaryScanner) *QueryHighlight {
|
||||
q.params.BoundaryScanner = t
|
||||
return q
|
||||
}
|
||||
|
||||
// BoundaryScannerLocale sets the highlight query's boundary_scanner_locale ignore unmapped field
|
||||
func (q *QueryHighlight) BoundaryScannerLocale(l string) *QueryHighlight {
|
||||
q.params.BoundaryScannerLocale = l
|
||||
return q
|
||||
}
|
||||
|
||||
// Encoder sets the highlight query's encoder ignore unmapped field
|
||||
func (q *QueryHighlight) Encoder(e HighlightEncoder) *QueryHighlight {
|
||||
q.params.Encoder = e
|
||||
return q
|
||||
}
|
||||
|
||||
// ForceSource sets the highlight query's force_source ignore unmapped field
|
||||
func (q *QueryHighlight) ForceSource(b bool) *QueryHighlight {
|
||||
q.params.ForceSource = &b
|
||||
return q
|
||||
}
|
||||
|
||||
// Fragmenter sets the highlight query's fragmenter ignore unmapped field
|
||||
func (q *QueryHighlight) Fragmenter(f HighlightFragmenter) *QueryHighlight {
|
||||
q.params.Fragmenter = f
|
||||
return q
|
||||
}
|
||||
|
||||
// FragmentOffset sets the highlight query's fragment_offset ignore unmapped field
|
||||
func (q *QueryHighlight) FragmentOffset(i uint16) *QueryHighlight {
|
||||
q.params.FragmentOffset = i
|
||||
return q
|
||||
}
|
||||
|
||||
// HighlightQuery sets the highlight query's highlight_query ignore unmapped field
|
||||
func (q *QueryHighlight) HighlightQuery(b Mappable) *QueryHighlight {
|
||||
q.highlightQuery = b
|
||||
return q
|
||||
}
|
||||
|
||||
// MatchedFields sets the highlight query's matched_fields ignore unmapped field
|
||||
func (q *QueryHighlight) MatchedFields(s ...string) *QueryHighlight {
|
||||
q.params.MatchedFields = append(q.params.MatchedFields, s...)
|
||||
return q
|
||||
}
|
||||
|
||||
// NoMatchSize sets the highlight query's no_match_size ignore unmapped field
|
||||
func (q *QueryHighlight) NoMatchSize(i uint16) *QueryHighlight {
|
||||
q.params.NoMatchSize = i
|
||||
return q
|
||||
}
|
||||
|
||||
// Order sets the nested highlight's score order unmapped field
|
||||
func (q *QueryHighlight) Order(o HighlightOrder) *QueryHighlight {
|
||||
q.params.Order = o
|
||||
return q
|
||||
}
|
||||
|
||||
// PhraseLimit sets the highlight query's phrase_limit ignore unmapped field
|
||||
func (q *QueryHighlight) PhraseLimit(i uint16) *QueryHighlight {
|
||||
q.params.PhraseLimit = i
|
||||
return q
|
||||
}
|
||||
|
||||
// RequireFieldMatch sets the highlight query's require_field_match ignore unmapped field
|
||||
func (q *QueryHighlight) RequireFieldMatch(b bool) *QueryHighlight {
|
||||
q.params.RequireFieldMatch = &b
|
||||
return q
|
||||
}
|
||||
|
||||
// TagsSchema sets the highlight query's tags_schema ignore unmapped field
|
||||
func (q *QueryHighlight) TagsSchema(s HighlightTagsSchema) *QueryHighlight {
|
||||
q.params.TagsSchema = s
|
||||
return q
|
||||
}
|
||||
|
||||
type HighlightType uint8
|
||||
|
||||
const (
|
||||
// HighlighterUnified is the "unified" value
|
||||
HighlighterUnified HighlightType = iota
|
||||
|
||||
// HighlighterPlain is the "plain" value
|
||||
HighlighterPlain
|
||||
|
||||
// HighlighterFvh is the "fvh" value
|
||||
HighlighterFvh
|
||||
)
|
||||
|
||||
// String returns a string representation of the type parameter, as
|
||||
// known to ElasticSearch.
|
||||
func (a HighlightType) String() string {
|
||||
switch a {
|
||||
case HighlighterUnified:
|
||||
return "unified"
|
||||
case HighlighterPlain:
|
||||
return "plain"
|
||||
case HighlighterFvh:
|
||||
return "fvh"
|
||||
}
|
||||
return ""
|
||||
}
|
||||
|
||||
type HighlightBoundaryScanner uint8
|
||||
|
||||
const (
|
||||
BoundaryScannerDefault HighlightBoundaryScanner = iota
|
||||
|
||||
// BoundaryScannerChars is the "chars" value
|
||||
BoundaryScannerChars
|
||||
|
||||
// BoundaryScannerSentence is the "sentence" value
|
||||
BoundaryScannerSentence
|
||||
|
||||
// BoundaryScannerWord is the "word" value
|
||||
BoundaryScannerWord
|
||||
)
|
||||
|
||||
// String returns a string representation of the boundary_scanner parameter, as
|
||||
// known to ElasticSearch.
|
||||
func (a HighlightBoundaryScanner) String() string {
|
||||
switch a {
|
||||
case BoundaryScannerChars:
|
||||
return "chars"
|
||||
case BoundaryScannerSentence:
|
||||
return "sentence"
|
||||
case BoundaryScannerWord:
|
||||
return "word"
|
||||
}
|
||||
return ""
|
||||
}
|
||||
|
||||
type HighlightEncoder uint8
|
||||
|
||||
const (
|
||||
// EncoderDefault is the "default" value
|
||||
EncoderDefault HighlightEncoder = iota
|
||||
|
||||
// EncoderHtml is the "html" value
|
||||
EncoderHtml
|
||||
)
|
||||
|
||||
// String returns a string representation of the encoder parameter, as
|
||||
// known to ElasticSearch.
|
||||
func (a HighlightEncoder) String() string {
|
||||
switch a {
|
||||
case EncoderDefault:
|
||||
return "default"
|
||||
case EncoderHtml:
|
||||
return "html"
|
||||
}
|
||||
return ""
|
||||
}
|
||||
|
||||
type HighlightFragmenter uint8
|
||||
|
||||
const (
|
||||
// FragmentSpan is the "span" value
|
||||
FragmenterSpan HighlightFragmenter = iota
|
||||
|
||||
// FragmenterSimple is the "simple" value
|
||||
FragmenterSimple
|
||||
)
|
||||
|
||||
// String returns a string representation of the fragmenter parameter, as
|
||||
// known to ElasticSearch.
|
||||
func (a HighlightFragmenter) String() string {
|
||||
switch a {
|
||||
case FragmenterSpan:
|
||||
return "span"
|
||||
case FragmenterSimple:
|
||||
return "simple"
|
||||
}
|
||||
return ""
|
||||
}
|
||||
|
||||
type HighlightOrder uint8
|
||||
|
||||
const (
|
||||
// OrderNone is the "none" value
|
||||
OrderNone HighlightOrder = iota
|
||||
|
||||
// OrderScore is the "score" value
|
||||
OrderScore
|
||||
)
|
||||
|
||||
// String returns a string representation of the order parameter, as
|
||||
// known to ElasticSearch.
|
||||
func (a HighlightOrder) String() string {
|
||||
switch a {
|
||||
case OrderNone:
|
||||
return "none"
|
||||
case OrderScore:
|
||||
return "score"
|
||||
}
|
||||
return ""
|
||||
}
|
||||
|
||||
type HighlightTagsSchema uint8
|
||||
|
||||
const (
|
||||
TagsSchemaDefault HighlightTagsSchema = iota
|
||||
// TagsSchemaStyled is the "styled" value
|
||||
TagsSchemaStyled
|
||||
)
|
||||
|
||||
// String returns a string representation of the tags_schema parameter, as
|
||||
// known to ElasticSearch.
|
||||
func (a HighlightTagsSchema) String() string {
|
||||
switch a {
|
||||
case TagsSchemaStyled:
|
||||
return "styled"
|
||||
}
|
||||
return ""
|
||||
}
|
|
@ -0,0 +1,93 @@
|
|||
package esquery
|
||||
|
||||
import (
|
||||
"testing"
|
||||
)
|
||||
|
||||
func TestHighlight(t *testing.T) {
|
||||
runMapTests(t, []mapTest{
|
||||
{
|
||||
"simple highlight",
|
||||
Highlight().Field("content"),
|
||||
map[string]interface{}{
|
||||
"fields": map[string]interface{}{
|
||||
"content": map[string]interface{}{},
|
||||
},
|
||||
},
|
||||
},
|
||||
{
|
||||
"highlight all params",
|
||||
Highlight().
|
||||
PreTags("<pre>", "<code>").
|
||||
PostTags("</code>", "</pre>").
|
||||
Field("content",
|
||||
Highlight().
|
||||
BoundaryChars(".;,")).
|
||||
FragmentSize(150).
|
||||
NumberOfFragments(4).
|
||||
Type(HighlighterPlain).
|
||||
BoundaryChars("()[]").
|
||||
BoundaryMaxScan(32).
|
||||
BoundaryScanner(BoundaryScannerChars).
|
||||
BoundaryScannerLocale("en-US").
|
||||
Encoder(EncoderHtml).
|
||||
ForceSource(true).
|
||||
Fragmenter(FragmenterSimple).
|
||||
FragmentOffset(6).
|
||||
HighlightQuery(
|
||||
Bool().
|
||||
Must(
|
||||
Match("author").
|
||||
Query("some guy").
|
||||
Analyzer("analyzer?").
|
||||
Fuzziness("fuzz"))).
|
||||
MatchedFields("title", "body").
|
||||
NoMatchSize(64).
|
||||
Order(OrderScore).
|
||||
PhraseLimit(512).
|
||||
RequireFieldMatch(false).
|
||||
TagsSchema(TagsSchemaStyled),
|
||||
map[string]interface{}{
|
||||
"pre_tags": []string{"<pre>", "<code>"},
|
||||
"post_tags": []string{"</code>", "</pre>"},
|
||||
"fragment_size": 150,
|
||||
"number_of_fragments": 4,
|
||||
"type": "plain",
|
||||
"boundary_chars": "()[]",
|
||||
"boundary_scanner": "chars",
|
||||
"boundary_max_scan": 32,
|
||||
"boundary_scanner_locale": "en-US",
|
||||
"encoder": "html",
|
||||
"force_source": true,
|
||||
"fragment_offset": 6,
|
||||
"fragmenter": "simple",
|
||||
"matched_fields": []string{"title", "body"},
|
||||
"no_match_size": 64,
|
||||
"order": "score",
|
||||
"phrase_limit": 512,
|
||||
"require_field_match": false,
|
||||
"tags_schema": "styled",
|
||||
"fields": map[string]interface{}{
|
||||
"content": map[string]interface{}{
|
||||
"boundary_chars": ".;,",
|
||||
},
|
||||
},
|
||||
"query": map[string]interface{}{
|
||||
"bool": map[string]interface{}{
|
||||
"must": []map[string]interface{}{
|
||||
{
|
||||
"match": map[string]interface{}{
|
||||
"author": map[string]interface{}{
|
||||
"analyzer": "analyzer?",
|
||||
"fuzziness": "fuzz",
|
||||
"query": "some guy",
|
||||
},
|
||||
},
|
||||
},
|
||||
},
|
||||
},
|
||||
},
|
||||
},
|
||||
},
|
||||
})
|
||||
}
|
|
@ -0,0 +1,210 @@
|
|||
package esquery
|
||||
|
||||
import (
|
||||
"github.com/fatih/structs"
|
||||
)
|
||||
|
||||
type MultiMatchQuery struct {
|
||||
params multiMatchParams
|
||||
}
|
||||
|
||||
// Map returns a map representation of the query; implementing the
|
||||
// Mappable interface.
|
||||
func (q *MultiMatchQuery) Map() map[string]interface{} {
|
||||
return map[string]interface{}{
|
||||
"multi_match": structs.Map(q.params),
|
||||
}
|
||||
}
|
||||
|
||||
type multiMatchParams struct {
|
||||
Qry interface{} `structs:"query"`
|
||||
Fields []string `structs:"fields"`
|
||||
Type MultiMatchType `structs:"type,string,omitempty"`
|
||||
TieBrk float32 `structs:"tie_breaker,omitempty"`
|
||||
Boost float32 `structs:"boost,omitempty"`
|
||||
Anl string `structs:"analyzer,omitempty"`
|
||||
AutoGenerate *bool `structs:"auto_generate_synonyms_phrase_query,omitempty"`
|
||||
Fuzz string `structs:"fuzziness,omitempty"`
|
||||
MaxExp uint16 `structs:"max_expansions,omitempty"`
|
||||
PrefLen uint16 `structs:"prefix_length,omitempty"`
|
||||
Trans *bool `structs:"transpositions,omitempty"`
|
||||
FuzzyRw string `structs:"fuzzy_rewrite,omitempty"`
|
||||
Lent *bool `structs:"lenient,omitempty"`
|
||||
Op MatchOperator `structs:"operator,string,omitempty"`
|
||||
MinMatch string `structs:"minimum_should_match,omitempty"`
|
||||
ZeroTerms ZeroTerms `structs:"zero_terms_query,string,omitempty"`
|
||||
Slp uint16 `structs:"slop,omitempty"`
|
||||
}
|
||||
|
||||
// MultiMatch creates a new query of type "multi_match"
|
||||
func MultiMatch(simpleQuery ...interface{}) *MultiMatchQuery {
|
||||
return newMultiMatch(simpleQuery...)
|
||||
}
|
||||
|
||||
func newMultiMatch(simpleQuery ...interface{}) *MultiMatchQuery {
|
||||
var qry interface{}
|
||||
if len(simpleQuery) > 0 {
|
||||
qry = simpleQuery[len(simpleQuery)-1]
|
||||
}
|
||||
|
||||
return &MultiMatchQuery{
|
||||
params: multiMatchParams{
|
||||
Qry: qry,
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
// Query sets the data to find in the query's field (it is the "query" component
|
||||
// of the query).
|
||||
func (q *MultiMatchQuery) Query(data interface{}) *MultiMatchQuery {
|
||||
q.params.Qry = data
|
||||
return q
|
||||
}
|
||||
|
||||
// Analyzer sets the analyzer used to convert the text in the "query" value into
|
||||
// tokens.
|
||||
func (q *MultiMatchQuery) Analyzer(a string) *MultiMatchQuery {
|
||||
q.params.Anl = a
|
||||
return q
|
||||
}
|
||||
|
||||
// Fields sets the fields used in the query
|
||||
func (q *MultiMatchQuery) Fields(a ...string) *MultiMatchQuery {
|
||||
q.params.Fields = append(q.params.Fields, a...)
|
||||
return q
|
||||
}
|
||||
|
||||
// AutoGenerateSynonymsPhraseQuery sets the "auto_generate_synonyms_phrase_query"
|
||||
// boolean.
|
||||
func (q *MultiMatchQuery) AutoGenerateSynonymsPhraseQuery(b bool) *MultiMatchQuery {
|
||||
q.params.AutoGenerate = &b
|
||||
return q
|
||||
}
|
||||
|
||||
// Fuzziness set the maximum edit distance allowed for matching.
|
||||
func (q *MultiMatchQuery) Fuzziness(f string) *MultiMatchQuery {
|
||||
q.params.Fuzz = f
|
||||
return q
|
||||
}
|
||||
|
||||
// MaxExpansions sets the maximum number of terms to which the query will expand.
|
||||
func (q *MultiMatchQuery) MaxExpansions(e uint16) *MultiMatchQuery {
|
||||
q.params.MaxExp = e
|
||||
return q
|
||||
}
|
||||
|
||||
// PrefixLength sets the number of beginning characters left unchanged for fuzzy
|
||||
// matching.
|
||||
func (q *MultiMatchQuery) PrefixLength(l uint16) *MultiMatchQuery {
|
||||
q.params.PrefLen = l
|
||||
return q
|
||||
}
|
||||
|
||||
// TieBreaker
|
||||
func (q *MultiMatchQuery) TieBreaker(l float32) *MultiMatchQuery {
|
||||
q.params.TieBrk = l
|
||||
return q
|
||||
}
|
||||
|
||||
// Boost
|
||||
func (q *MultiMatchQuery) Boost(l float32) *MultiMatchQuery {
|
||||
q.params.Boost = l
|
||||
return q
|
||||
}
|
||||
|
||||
// Transpositions sets whether edits for fuzzy matching include transpositions
|
||||
// of two adjacent characters.
|
||||
func (q *MultiMatchQuery) Transpositions(b bool) *MultiMatchQuery {
|
||||
q.params.Trans = &b
|
||||
return q
|
||||
}
|
||||
|
||||
// FuzzyRewrite sets the method used to rewrite the query.
|
||||
func (q *MultiMatchQuery) FuzzyRewrite(s string) *MultiMatchQuery {
|
||||
q.params.FuzzyRw = s
|
||||
return q
|
||||
}
|
||||
|
||||
// Lenient sets whether format-based errors should be ignored.
|
||||
func (q *MultiMatchQuery) Lenient(b bool) *MultiMatchQuery {
|
||||
q.params.Lent = &b
|
||||
return q
|
||||
}
|
||||
|
||||
// Operator sets the boolean logic used to interpret text in the query value.
|
||||
func (q *MultiMatchQuery) Operator(op MatchOperator) *MultiMatchQuery {
|
||||
q.params.Op = op
|
||||
return q
|
||||
}
|
||||
|
||||
// Type sets the query type
|
||||
func (q *MultiMatchQuery) Type(t MultiMatchType) *MultiMatchQuery {
|
||||
q.params.Type = t
|
||||
return q
|
||||
}
|
||||
|
||||
// MinimumShouldMatch sets the minimum number of clauses that must match for a
|
||||
// document to be returned.
|
||||
func (q *MultiMatchQuery) MinimumShouldMatch(s string) *MultiMatchQuery {
|
||||
q.params.MinMatch = s
|
||||
return q
|
||||
}
|
||||
|
||||
// Slop sets the maximum number of positions allowed between matching tokens.
|
||||
func (q *MultiMatchQuery) Slop(n uint16) *MultiMatchQuery {
|
||||
q.params.Slp = n
|
||||
return q
|
||||
}
|
||||
|
||||
// ZeroTermsQuery sets the "zero_terms_query" option to use. This indicates
|
||||
// whether no documents are returned if the analyzer removes all tokens, such as
|
||||
// when using a stop filter.
|
||||
func (q *MultiMatchQuery) ZeroTermsQuery(s ZeroTerms) *MultiMatchQuery {
|
||||
q.params.ZeroTerms = s
|
||||
return q
|
||||
}
|
||||
|
||||
// MatchType is an enumeration type representing supported values for a
|
||||
// multi match query's "type" parameter.
|
||||
type MultiMatchType uint8
|
||||
|
||||
const (
|
||||
// TypeBestFields is the "best_fields" type
|
||||
MatchTypeBestFields MultiMatchType = iota
|
||||
|
||||
// TypeMostFields is the "most_fields" type
|
||||
MatchTypeMostFields
|
||||
|
||||
// TypeMostFields is the "cross_fields" type
|
||||
MatchTypeCrossFields
|
||||
|
||||
// TypeMostFields is the "phrase" type
|
||||
MatchTypePhrase
|
||||
|
||||
// TypeMostFields is the "phrase_prefix" type
|
||||
MatchTypePhrasePrefix
|
||||
|
||||
// TypeMostFields is the "bool_prefix" type
|
||||
MatchTypeBoolPrefix
|
||||
)
|
||||
|
||||
// String returns a string representation of the match operator, as known to
|
||||
// ElasticSearch.
|
||||
func (a MultiMatchType) String() string {
|
||||
switch a {
|
||||
case MatchTypeBestFields:
|
||||
return "best_fields"
|
||||
case MatchTypeMostFields:
|
||||
return "most_fields"
|
||||
case MatchTypeCrossFields:
|
||||
return "cross_fields"
|
||||
case MatchTypePhrase:
|
||||
return "phrase"
|
||||
case MatchTypePhrasePrefix:
|
||||
return "phrase_prefix"
|
||||
case MatchTypeBoolPrefix:
|
||||
return "bool_prefix"
|
||||
default:
|
||||
return ""
|
||||
}
|
||||
}
|
|
@ -0,0 +1,62 @@
|
|||
package esquery
|
||||
|
||||
import (
|
||||
"testing"
|
||||
)
|
||||
|
||||
func TestMultiMatch(t *testing.T) {
|
||||
runMapTests(t, []mapTest{
|
||||
{
|
||||
"simple multi_match",
|
||||
MultiMatch("value1", "value2").Fields("title"),
|
||||
map[string]interface{}{
|
||||
"multi_match": map[string]interface{}{
|
||||
"fields": []string{"title"},
|
||||
"query": "value2",
|
||||
},
|
||||
},
|
||||
},
|
||||
{
|
||||
"multi_match all params",
|
||||
MultiMatch("original").
|
||||
Query("test").
|
||||
Analyzer("stop").
|
||||
Fields("title", "body").
|
||||
AutoGenerateSynonymsPhraseQuery(true).
|
||||
Fuzziness("AUTO").
|
||||
MaxExpansions(16).
|
||||
PrefixLength(12).
|
||||
TieBreaker(0.3).
|
||||
Boost(6.4).
|
||||
Transpositions(true).
|
||||
FuzzyRewrite("scoring_boolean").
|
||||
Lenient(true).
|
||||
Operator(OperatorAnd).
|
||||
Type(MatchTypePhrase).
|
||||
MinimumShouldMatch("3<90%").
|
||||
Slop(2).
|
||||
ZeroTermsQuery(ZeroTermsAll),
|
||||
map[string]interface{}{
|
||||
"multi_match": map[string]interface{}{
|
||||
"analyzer": "stop",
|
||||
"auto_generate_synonyms_phrase_query": true,
|
||||
"boost": 6.4,
|
||||
"fuzziness": "AUTO",
|
||||
"fuzzy_rewrite": "scoring_boolean",
|
||||
"lenient": true,
|
||||
"max_expansions": 16,
|
||||
"minimum_should_match": "3<90%",
|
||||
"prefix_length": 12,
|
||||
"transpositions": true,
|
||||
"type": "phrase",
|
||||
"tie_breaker": 0.3,
|
||||
"operator": "AND",
|
||||
"zero_terms_query": "all",
|
||||
"slop": 2,
|
||||
"query": "test",
|
||||
"fields": []string{"title", "body"},
|
||||
},
|
||||
},
|
||||
},
|
||||
})
|
||||
}
|
23
search.go
23
search.go
|
@ -15,15 +15,16 @@ import (
|
|||
// Not all features of the search API are currently supported, but a request can
|
||||
// currently include a query, aggregations, and more.
|
||||
type SearchRequest struct {
|
||||
query Mappable
|
||||
aggs []Aggregation
|
||||
postFilter Mappable
|
||||
from *uint64
|
||||
size *uint64
|
||||
explain *bool
|
||||
timeout *time.Duration
|
||||
source Source
|
||||
from *uint64
|
||||
highlight Mappable
|
||||
postFilter Mappable
|
||||
query Mappable
|
||||
size *uint64
|
||||
sort Sort
|
||||
source Source
|
||||
timeout *time.Duration
|
||||
}
|
||||
|
||||
// Search creates a new SearchRequest object, to be filled via method chaining.
|
||||
|
@ -98,6 +99,13 @@ func (req *SearchRequest) SourceExcludes(keys ...string) *SearchRequest {
|
|||
return req
|
||||
}
|
||||
|
||||
// Highlight sets a highlight for the request.
|
||||
func (req *SearchRequest) Highlight(highlight Mappable) *SearchRequest {
|
||||
req.highlight = highlight
|
||||
return req
|
||||
}
|
||||
|
||||
|
||||
// Map implements the Mappable interface. It converts the request to into a
|
||||
// nested map[string]interface{}, as expected by the go-elasticsearch library.
|
||||
func (req *SearchRequest) Map() map[string]interface{} {
|
||||
|
@ -131,6 +139,9 @@ func (req *SearchRequest) Map() map[string]interface{} {
|
|||
if req.timeout != nil {
|
||||
m["timeout"] = fmt.Sprintf("%.0fs", req.timeout.Seconds())
|
||||
}
|
||||
if req.highlight != nil {
|
||||
m["highlight"] = req.highlight.Map()
|
||||
}
|
||||
|
||||
source := req.source.Map()
|
||||
if len(source) > 0 {
|
||||
|
|
Loading…
Reference in New Issue