Compare commits
4 Commits
Author | SHA1 | Date | |
---|---|---|---|
af8bb64366 | |||
0aee33d553 | |||
059550898e | |||
24564489b8 |
@ -37,6 +37,7 @@ esgo2dump --input=http://127.0.0.1:9200/some_index --output=./data.json --query_
|
|||||||
f_limit int
|
f_limit int
|
||||||
f_type string
|
f_type string
|
||||||
f_source string
|
f_source string
|
||||||
|
f_sort string
|
||||||
f_query string
|
f_query string
|
||||||
|
|
||||||
f_query_file string
|
f_query_file string
|
||||||
@ -57,6 +58,7 @@ func init() {
|
|||||||
rootCommand.Flags().StringVar(&es_oversion, "o-version", "7", "output(es) version")
|
rootCommand.Flags().StringVar(&es_oversion, "o-version", "7", "output(es) version")
|
||||||
rootCommand.Flags().StringVarP(&f_type, "type", "t", "data", "data/mapping/setting")
|
rootCommand.Flags().StringVarP(&f_type, "type", "t", "data", "data/mapping/setting")
|
||||||
rootCommand.Flags().StringVarP(&f_source, "source", "s", "", "query source, use ';' to separate")
|
rootCommand.Flags().StringVarP(&f_source, "source", "s", "", "query source, use ';' to separate")
|
||||||
|
rootCommand.Flags().StringVar(&f_sort, "sort", "", "sort, <field>:<direction> format, for example: time:desc or name:asc")
|
||||||
rootCommand.Flags().StringVarP(&f_query, "query", "q", "", `query dsl, example: {"bool":{"must":[{"term":{"name":{"value":"some_name"}}}],"must_not":[{"range":{"age":{"gte":18,"lt":60}}}]}}`)
|
rootCommand.Flags().StringVarP(&f_query, "query", "q", "", `query dsl, example: {"bool":{"must":[{"term":{"name":{"value":"some_name"}}}],"must_not":[{"range":{"age":{"gte":18,"lt":60}}}]}}`)
|
||||||
rootCommand.Flags().StringVar(&f_query_file, "query_file", "", `query json file (will execute line by line)`)
|
rootCommand.Flags().StringVar(&f_query_file, "query_file", "", `query json file (will execute line by line)`)
|
||||||
rootCommand.Flags().IntVarP(&f_limit, "limit", "l", 100, "")
|
rootCommand.Flags().IntVarP(&f_limit, "limit", "l", 100, "")
|
||||||
|
@ -11,6 +11,7 @@ import (
|
|||||||
"net/url"
|
"net/url"
|
||||||
"os"
|
"os"
|
||||||
"strings"
|
"strings"
|
||||||
|
"sync"
|
||||||
|
|
||||||
"github.com/loveuer/esgo2dump/internal/interfaces"
|
"github.com/loveuer/esgo2dump/internal/interfaces"
|
||||||
"github.com/loveuer/esgo2dump/internal/opt"
|
"github.com/loveuer/esgo2dump/internal/opt"
|
||||||
@ -192,24 +193,27 @@ func executeData(ctx context.Context, input, output interfaces.DumpIO) error {
|
|||||||
|
|
||||||
e2ch = make(chan error)
|
e2ch = make(chan error)
|
||||||
wch = make(chan []*model.ESSource)
|
wch = make(chan []*model.ESSource)
|
||||||
|
wg = sync.WaitGroup{}
|
||||||
)
|
)
|
||||||
|
|
||||||
go func() {
|
go func() {
|
||||||
defer func() {
|
wg.Add(1)
|
||||||
close(wch)
|
|
||||||
close(e2ch)
|
|
||||||
}()
|
|
||||||
|
|
||||||
if err = output.WriteData(ctx, wch); err != nil {
|
if err = output.WriteData(ctx, wch); err != nil {
|
||||||
e2ch <- err
|
e2ch <- err
|
||||||
}
|
}
|
||||||
|
|
||||||
|
wg.Done()
|
||||||
}()
|
}()
|
||||||
|
|
||||||
log.Info("Query: got queries=%d", len(queries))
|
log.Info("Query: got queries=%d", len(queries))
|
||||||
|
|
||||||
Loop:
|
Loop:
|
||||||
for _, query := range queries {
|
for qi, query := range queries {
|
||||||
dch, ech = input.ReadData(ctx, f_limit, query, sources)
|
bs, _ := json.Marshal(query)
|
||||||
|
|
||||||
|
log.Debug("Query[%d]: %s", qi, string(bs))
|
||||||
|
|
||||||
|
dch, ech = input.ReadData(ctx, f_limit, query, sources, []string{f_sort})
|
||||||
|
|
||||||
for {
|
for {
|
||||||
select {
|
select {
|
||||||
@ -233,6 +237,10 @@ Loop:
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
close(wch)
|
||||||
|
|
||||||
|
wg.Wait()
|
||||||
|
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -6,7 +6,7 @@ import (
|
|||||||
)
|
)
|
||||||
|
|
||||||
type DumpIO interface {
|
type DumpIO interface {
|
||||||
ReadData(ctx context.Context, size int, query map[string]any, includeFields []string) (<-chan []*model.ESSource, <-chan error)
|
ReadData(ctx context.Context, size int, query map[string]any, includeFields []string, sort []string) (<-chan []*model.ESSource, <-chan error)
|
||||||
WriteData(ctx context.Context, docsCh <-chan []*model.ESSource) error
|
WriteData(ctx context.Context, docsCh <-chan []*model.ESSource) error
|
||||||
|
|
||||||
ReadMapping(context.Context) (map[string]any, error)
|
ReadMapping(context.Context) (map[string]any, error)
|
||||||
|
9
internal/util/min.go
Normal file
9
internal/util/min.go
Normal file
@ -0,0 +1,9 @@
|
|||||||
|
package util
|
||||||
|
|
||||||
|
func Min[T ~string | ~int | ~int64 | ~uint64 | ~float64 | ~float32 | ~int32 | ~uint32 | ~int16 | ~uint16 | ~int8 | ~uint8](a, b T) T {
|
||||||
|
if a <= b {
|
||||||
|
return a
|
||||||
|
}
|
||||||
|
|
||||||
|
return b
|
||||||
|
}
|
@ -135,8 +135,8 @@ func (c *clientv6) Close() error {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (c *clientv6) ReadData(ctx context.Context, size int, query map[string]any, source []string) (<-chan []*model.ESSource, <-chan error) {
|
func (c *clientv6) ReadData(ctx context.Context, size int, query map[string]any, source []string, sort []string) (<-chan []*model.ESSource, <-chan error) {
|
||||||
dch, ech := es6.ReadData(ctx, c.client, c.index, size, 0, query, source)
|
dch, ech := es6.ReadData(ctx, c.client, c.index, size, 0, query, source, sort)
|
||||||
|
|
||||||
return dch, ech
|
return dch, ech
|
||||||
}
|
}
|
||||||
|
@ -70,63 +70,8 @@ func (c *client) Close() error {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
//func (c *client) WriteData(ctx context.Context, docs []*model.ESSource) (int, error) {
|
func (c *client) ReadData(ctx context.Context, size int, query map[string]any, source []string, sort []string) (<-chan []*model.ESSource, <-chan error) {
|
||||||
// var (
|
dch, ech := es7.ReadDataV2(ctx, c.client, c.index, size, 0, query, source, sort)
|
||||||
// err error
|
|
||||||
// indexer esutil.BulkIndexer
|
|
||||||
// count int
|
|
||||||
// be error
|
|
||||||
// )
|
|
||||||
// if indexer, err = esutil.NewBulkIndexer(esutil.BulkIndexerConfig{
|
|
||||||
// Client: c.client,
|
|
||||||
// Index: c.index,
|
|
||||||
// ErrorTrace: true,
|
|
||||||
// OnError: func(ctx context.Context, err error) {
|
|
||||||
//
|
|
||||||
// },
|
|
||||||
// }); err != nil {
|
|
||||||
// return 0, err
|
|
||||||
// }
|
|
||||||
//
|
|
||||||
// for _, doc := range docs {
|
|
||||||
// var bs []byte
|
|
||||||
//
|
|
||||||
// if bs, err = json.Marshal(doc.Content); err != nil {
|
|
||||||
// return 0, err
|
|
||||||
// }
|
|
||||||
//
|
|
||||||
// if err = indexer.Add(context.Background(), esutil.BulkIndexerItem{
|
|
||||||
// Action: "index",
|
|
||||||
// Index: c.index,
|
|
||||||
// DocumentID: doc.DocId,
|
|
||||||
// Body: bytes.NewReader(bs),
|
|
||||||
// OnFailure: func(ctx context.Context, item esutil.BulkIndexerItem, item2 esutil.BulkIndexerResponseItem, bulkErr error) {
|
|
||||||
// be = bulkErr
|
|
||||||
// },
|
|
||||||
// }); err != nil {
|
|
||||||
// return 0, err
|
|
||||||
// }
|
|
||||||
// count++
|
|
||||||
// }
|
|
||||||
//
|
|
||||||
// if err = indexer.Close(util.TimeoutCtx(ctx, opt.Timeout)); err != nil {
|
|
||||||
// return 0, err
|
|
||||||
// }
|
|
||||||
//
|
|
||||||
// if be != nil {
|
|
||||||
// return 0, be
|
|
||||||
// }
|
|
||||||
//
|
|
||||||
// stats := indexer.Stats()
|
|
||||||
// if stats.NumFailed > 0 {
|
|
||||||
// return count, fmt.Errorf("write to xes failed_count=%d bulk_count=%d", stats.NumFailed, count)
|
|
||||||
// }
|
|
||||||
//
|
|
||||||
// return count, nil
|
|
||||||
//}
|
|
||||||
|
|
||||||
func (c *client) ReadData(ctx context.Context, size int, query map[string]any, source []string) (<-chan []*model.ESSource, <-chan error) {
|
|
||||||
dch, ech := es7.ReadData(ctx, c.client, c.index, size, 0, query, source)
|
|
||||||
|
|
||||||
return dch, ech
|
return dch, ech
|
||||||
}
|
}
|
||||||
|
@ -110,7 +110,7 @@ func (c *client) IsFile() bool {
|
|||||||
return true
|
return true
|
||||||
}
|
}
|
||||||
|
|
||||||
func (c *client) ReadData(ctx context.Context, size int, _ map[string]any, _ []string) (<-chan []*model.ESSource, <-chan error) {
|
func (c *client) ReadData(ctx context.Context, size int, _ map[string]any, _ []string, _ []string) (<-chan []*model.ESSource, <-chan error) {
|
||||||
var (
|
var (
|
||||||
err error
|
err error
|
||||||
count = 0
|
count = 0
|
||||||
|
@ -4,6 +4,7 @@ type ESSource struct {
|
|||||||
DocId string `json:"_id"`
|
DocId string `json:"_id"`
|
||||||
Index string `json:"_index"`
|
Index string `json:"_index"`
|
||||||
Content map[string]any `json:"_source"`
|
Content map[string]any `json:"_source"`
|
||||||
|
Sort []any `json:"sort"`
|
||||||
}
|
}
|
||||||
|
|
||||||
type ESResponse struct {
|
type ESResponse struct {
|
||||||
|
@ -10,10 +10,11 @@ import (
|
|||||||
"github.com/loveuer/esgo2dump/internal/util"
|
"github.com/loveuer/esgo2dump/internal/util"
|
||||||
"github.com/loveuer/esgo2dump/log"
|
"github.com/loveuer/esgo2dump/log"
|
||||||
"github.com/loveuer/esgo2dump/model"
|
"github.com/loveuer/esgo2dump/model"
|
||||||
|
"github.com/samber/lo"
|
||||||
"time"
|
"time"
|
||||||
)
|
)
|
||||||
|
|
||||||
func ReadData(ctx context.Context, client *elastic.Client, index string, size, max int, query map[string]any, source []string) (<-chan []*model.ESSource, <-chan error) {
|
func ReadData(ctx context.Context, client *elastic.Client, index string, size, max int, query map[string]any, source []string, sort []string) (<-chan []*model.ESSource, <-chan error) {
|
||||||
var (
|
var (
|
||||||
dataCh = make(chan []*model.ESSource)
|
dataCh = make(chan []*model.ESSource)
|
||||||
errCh = make(chan error)
|
errCh = make(chan error)
|
||||||
@ -71,6 +72,16 @@ func ReadData(ctx context.Context, client *elastic.Client, index string, size, m
|
|||||||
qs = append(qs, client.Search.WithSourceIncludes(source...))
|
qs = append(qs, client.Search.WithSourceIncludes(source...))
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if len(sort) > 0 {
|
||||||
|
sorts := lo.Filter(sort, func(item string, index int) bool {
|
||||||
|
return item != ""
|
||||||
|
})
|
||||||
|
|
||||||
|
if len(sorts) > 0 {
|
||||||
|
qs = append(qs, client.Search.WithSort(sorts...))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
if query != nil && len(query) > 0 {
|
if query != nil && len(query) > 0 {
|
||||||
queryBs, _ := json.Marshal(map[string]any{"query": query})
|
queryBs, _ := json.Marshal(map[string]any{"query": query})
|
||||||
qs = append(qs, client.Search.WithBody(bytes.NewReader(queryBs)))
|
qs = append(qs, client.Search.WithBody(bytes.NewReader(queryBs)))
|
||||||
|
127
xes/es7/read.go
127
xes/es7/read.go
@ -10,10 +10,15 @@ import (
|
|||||||
"github.com/loveuer/esgo2dump/internal/util"
|
"github.com/loveuer/esgo2dump/internal/util"
|
||||||
"github.com/loveuer/esgo2dump/log"
|
"github.com/loveuer/esgo2dump/log"
|
||||||
"github.com/loveuer/esgo2dump/model"
|
"github.com/loveuer/esgo2dump/model"
|
||||||
|
"github.com/samber/lo"
|
||||||
"time"
|
"time"
|
||||||
)
|
)
|
||||||
|
|
||||||
func ReadData(ctx context.Context, client *elastic.Client, index string, size, max int, query map[string]any, source []string) (<-chan []*model.ESSource, <-chan error) {
|
// ReadData
|
||||||
|
// Deprecated
|
||||||
|
// @param[source]: a list of include fields to extract and return from the _source field.
|
||||||
|
// @param[sort]: a list of <field>:<direction> pairs.
|
||||||
|
func ReadData(ctx context.Context, client *elastic.Client, index string, size, max int, query map[string]any, source []string, sort []string) (<-chan []*model.ESSource, <-chan error) {
|
||||||
var (
|
var (
|
||||||
dataCh = make(chan []*model.ESSource)
|
dataCh = make(chan []*model.ESSource)
|
||||||
errCh = make(chan error)
|
errCh = make(chan error)
|
||||||
@ -71,6 +76,16 @@ func ReadData(ctx context.Context, client *elastic.Client, index string, size, m
|
|||||||
qs = append(qs, client.Search.WithSourceIncludes(source...))
|
qs = append(qs, client.Search.WithSourceIncludes(source...))
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if len(sort) > 0 {
|
||||||
|
sorts := lo.Filter(sort, func(item string, index int) bool {
|
||||||
|
return item != ""
|
||||||
|
})
|
||||||
|
|
||||||
|
if len(sorts) > 0 {
|
||||||
|
qs = append(qs, client.Search.WithSort(sorts...))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
if query != nil && len(query) > 0 {
|
if query != nil && len(query) > 0 {
|
||||||
queryBs, _ := json.Marshal(map[string]any{"query": query})
|
queryBs, _ := json.Marshal(map[string]any{"query": query})
|
||||||
qs = append(qs, client.Search.WithBody(bytes.NewReader(queryBs)))
|
qs = append(qs, client.Search.WithBody(bytes.NewReader(queryBs)))
|
||||||
@ -134,3 +149,113 @@ func ReadData(ctx context.Context, client *elastic.Client, index string, size, m
|
|||||||
|
|
||||||
return dataCh, errCh
|
return dataCh, errCh
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// ReadDataV2 es7 read data
|
||||||
|
/*
|
||||||
|
- @param[source]: a list of include fields to extract and return from the _source field.
|
||||||
|
- @param[sort]: a list of <field>:<direction> pairs.
|
||||||
|
*/
|
||||||
|
func ReadDataV2(
|
||||||
|
ctx context.Context,
|
||||||
|
client *elastic.Client,
|
||||||
|
index string,
|
||||||
|
size, max int,
|
||||||
|
query map[string]any,
|
||||||
|
source []string,
|
||||||
|
sort []string,
|
||||||
|
) (<-chan []*model.ESSource, <-chan error) {
|
||||||
|
var (
|
||||||
|
dataCh = make(chan []*model.ESSource)
|
||||||
|
errCh = make(chan error)
|
||||||
|
)
|
||||||
|
|
||||||
|
go func() {
|
||||||
|
var (
|
||||||
|
err error
|
||||||
|
bs []byte
|
||||||
|
resp *esapi.Response
|
||||||
|
searchAfter = make([]any, 0)
|
||||||
|
total = 0
|
||||||
|
body = make(map[string]any)
|
||||||
|
qs []func(request *esapi.SearchRequest)
|
||||||
|
)
|
||||||
|
|
||||||
|
if sort == nil {
|
||||||
|
sort = []string{}
|
||||||
|
}
|
||||||
|
|
||||||
|
if query != nil && len(query) > 0 {
|
||||||
|
body["query"] = query
|
||||||
|
}
|
||||||
|
|
||||||
|
sort = append(sort, "_id:ASC")
|
||||||
|
|
||||||
|
sorts := lo.Filter(sort, func(item string, index int) bool {
|
||||||
|
return item != ""
|
||||||
|
})
|
||||||
|
|
||||||
|
defer func() {
|
||||||
|
close(dataCh)
|
||||||
|
close(errCh)
|
||||||
|
}()
|
||||||
|
|
||||||
|
for {
|
||||||
|
qs = []func(*esapi.SearchRequest){
|
||||||
|
client.Search.WithContext(util.TimeoutCtx(ctx, 30)),
|
||||||
|
client.Search.WithIndex(index),
|
||||||
|
client.Search.WithSize(util.Min(size, max-total)),
|
||||||
|
client.Search.WithSort(sorts...),
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(source) > 0 {
|
||||||
|
qs = append(qs, client.Search.WithSourceIncludes(source...))
|
||||||
|
}
|
||||||
|
|
||||||
|
delete(body, "search_after")
|
||||||
|
if len(searchAfter) > 0 {
|
||||||
|
body["search_after"] = searchAfter
|
||||||
|
}
|
||||||
|
|
||||||
|
if bs, err = json.Marshal(body); err != nil {
|
||||||
|
errCh <- err
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
qs = append(qs, client.Search.WithBody(bytes.NewReader(bs)))
|
||||||
|
if resp, err = client.Search(qs...); err != nil {
|
||||||
|
errCh <- err
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
if resp.StatusCode != 200 {
|
||||||
|
errCh <- fmt.Errorf("resp status=%d, resp=%s", resp.StatusCode, resp.String())
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
var result = new(model.ESResponse)
|
||||||
|
decoder := json.NewDecoder(resp.Body)
|
||||||
|
if err = decoder.Decode(result); err != nil {
|
||||||
|
errCh <- err
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
if resp.StatusCode != 200 {
|
||||||
|
errCh <- fmt.Errorf("resp status=%d, resp=%s", resp.StatusCode, resp.String())
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
dataCh <- result.Hits.Hits
|
||||||
|
total += len(result.Hits.Hits)
|
||||||
|
|
||||||
|
if len(result.Hits.Hits) < size || (max > 0 && total >= max) {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
|
||||||
|
searchAfter = result.Hits.Hits[len(result.Hits.Hits)-1].Sort
|
||||||
|
}
|
||||||
|
|
||||||
|
}()
|
||||||
|
|
||||||
|
return dataCh, errCh
|
||||||
|
|
||||||
|
}
|
||||||
|
Reference in New Issue
Block a user