Skip to content

Commit

Permalink
updated zapx version, code cleanup
Browse files Browse the repository at this point in the history
  • Loading branch information
Thejas-bhat committed Jul 13, 2022
1 parent 05c5ea1 commit 6284a48
Show file tree
Hide file tree
Showing 5 changed files with 23 additions and 24 deletions.
2 changes: 1 addition & 1 deletion go.mod
Original file line number Diff line number Diff line change
Expand Up @@ -21,7 +21,7 @@ require (
github.com/blevesearch/zapx/v12 v12.3.4
github.com/blevesearch/zapx/v13 v13.3.4
github.com/blevesearch/zapx/v14 v14.3.4
github.com/blevesearch/zapx/v15 v15.3.4
github.com/blevesearch/zapx/v15 v15.3.5-0.20220713163830-ae843e553177
github.com/couchbase/moss v0.2.0
github.com/golang/protobuf v1.3.2
github.com/spf13/cobra v0.0.5
Expand Down
5 changes: 2 additions & 3 deletions go.sum
Original file line number Diff line number Diff line change
Expand Up @@ -20,7 +20,6 @@ github.com/blevesearch/gtreap v0.1.1/go.mod h1:QaQyDRAT51sotthUWAH4Sj08awFSSWzgY
github.com/blevesearch/mmap-go v1.0.2/go.mod h1:ol2qBqYaOUsGdm7aRMRrYGgPvnwLe6Y+7LMvAB5IbSA=
github.com/blevesearch/mmap-go v1.0.4 h1:OVhDhT5B/M1HNPpYPBKIEJaD0F3Si+CrEKULGCDPWmc=
github.com/blevesearch/mmap-go v1.0.4/go.mod h1:EWmEAOmdAS9z/pi/+Toxu99DnsbhG1TIxUoRmJw/pSs=
github.com/blevesearch/scorch_segment_api/v2 v2.1.0 h1:NFwteOpZEvJk5Vg0H6gD0hxupsG3JYocE4DBvsA2GZI=
github.com/blevesearch/scorch_segment_api/v2 v2.1.0/go.mod h1:uch7xyyO/Alxkuxa+CGs79vw0QY8BENSBjg6Mw5L5DE=
github.com/blevesearch/scorch_segment_api/v2 v2.1.1 h1:J8UDudUpDJz21d/hCMIshCeRordwnDTftgXcSDMUx40=
github.com/blevesearch/scorch_segment_api/v2 v2.1.1/go.mod h1:uch7xyyO/Alxkuxa+CGs79vw0QY8BENSBjg6Mw5L5DE=
Expand All @@ -42,8 +41,8 @@ github.com/blevesearch/zapx/v13 v13.3.4 h1:f646k6300VGRIR7eJ6lLtF8UC95NIWmF899j4
github.com/blevesearch/zapx/v13 v13.3.4/go.mod h1:Wl7hO1gT+IDvJb7i06g2iW5Qvw0KzncJPsBx7WGWhLA=
github.com/blevesearch/zapx/v14 v14.3.4 h1:/FVzSGFG5rbVWfPEqlcaJd8lZSJMQpTdmFhz/l2QI7w=
github.com/blevesearch/zapx/v14 v14.3.4/go.mod h1:b1YhRXXhAj9i+9aOwhRKCHUmJyYieK/QbDvPJDLddUk=
github.com/blevesearch/zapx/v15 v15.3.4 h1:/y6AOxRuBiZPFAItqcrKcXPPtlAwuW/jMoOFO7tc7rs=
github.com/blevesearch/zapx/v15 v15.3.4/go.mod h1:TQ/qDC2q7TSSpeC6Vgr9fDN56Ra0u49lZJQ4v30WEx4=
github.com/blevesearch/zapx/v15 v15.3.5-0.20220713163830-ae843e553177 h1:0/WYF9nS1HBgDc3z7ePdDz15CwMuNYZ4WeD7Kravm7M=
github.com/blevesearch/zapx/v15 v15.3.5-0.20220713163830-ae843e553177/go.mod h1:ii4ohMQC0TCUjYfq8OtrbABgeI1zljjyXBFpUe/dPDw=
github.com/coreos/etcd v3.3.10+incompatible/go.mod h1:uF7uidLiAD3TWHmW31ZFd/JWoc32PjwdhPthX9715RE=
github.com/coreos/go-etcd v2.0.0+incompatible/go.mod h1:Jez6KQU2B/sWsbdaef3ED8NzMklzPG4d5KIOhIy30Tk=
github.com/coreos/go-semver v0.2.0/go.mod h1:nnelYz7RCh+5ahJtPPxZlU+153eP4D4r3EedlOD2RNk=
Expand Down
2 changes: 1 addition & 1 deletion index/scorch/merge.go
Original file line number Diff line number Diff line change
Expand Up @@ -438,7 +438,7 @@ type segmentMerge struct {
func cumulateBytesRead(sbs []segment.Segment) uint64 {
var rv uint64
for _, seg := range sbs {
if segI, ok := seg.(segment.DiskStatsReporter); ok {
if segI, diskStatsAvailable := seg.(segment.DiskStatsReporter); diskStatsAvailable {
rv += segI.BytesRead()
}
}
Expand Down
32 changes: 16 additions & 16 deletions index/scorch/snapshot_index.go
Original file line number Diff line number Diff line change
Expand Up @@ -149,15 +149,15 @@ func (i *IndexSnapshot) newIndexSnapshotFieldDict(field string,
for index, segment := range i.segment {
go func(index int, segment *SegmentSnapshot) {
var prevBytesRead uint64
seg, ok := segment.segment.(diskStatsReporter)
if ok {
seg, diskStatsAvailable := segment.segment.(diskStatsReporter)
if diskStatsAvailable {
prevBytesRead = seg.BytesRead()
}
dict, err := segment.segment.Dictionary(field)
if err != nil {
results <- &asynchSegmentResult{err: err}
} else {
if ok {
if diskStatsAvailable {
atomic.AddUint64(&i.parent.stats.TotBytesReadAtQueryTime,
seg.BytesRead()-prevBytesRead)
}
Expand Down Expand Up @@ -436,8 +436,8 @@ func (i *IndexSnapshot) Document(id string) (rv index.Document, err error) {

rvd := document.NewDocument(id)
var prevBytesRead uint64
seg, ok := i.segment[segmentIndex].segment.(segment.DiskStatsReporter)
if ok {
seg, diskStatsAvailable := i.segment[segmentIndex].segment.(segment.DiskStatsReporter)
if diskStatsAvailable {
prevBytesRead = seg.BytesRead()
}
err = i.segment[segmentIndex].VisitDocument(localDocNum, func(name string, typ byte, val []byte, pos []uint64) bool {
Expand Down Expand Up @@ -469,7 +469,7 @@ func (i *IndexSnapshot) Document(id string) (rv index.Document, err error) {
if err != nil {
return nil, err
}
if ok {
if diskStatsAvailable {
delta := seg.BytesRead() - prevBytesRead
atomic.AddUint64(&i.parent.stats.TotBytesReadAtQueryTime, delta)
}
Expand Down Expand Up @@ -548,15 +548,15 @@ func (is *IndexSnapshot) TermFieldReader(term []byte, field string, includeFreq,
rv.dicts = make([]segment.TermDictionary, len(is.segment))
for i, segment := range is.segment {
var prevBytesRead uint64
segP, ok := segment.segment.(diskStatsReporter)
if ok {
segP, diskStatsAvailable := segment.segment.(diskStatsReporter)
if diskStatsAvailable {
prevBytesRead = segP.BytesRead()
}
dict, err := segment.segment.Dictionary(field)
if err != nil {
return nil, err
}
if ok {
if diskStatsAvailable {
atomic.AddUint64(&is.parent.stats.TotBytesReadAtQueryTime, segP.BytesRead()-prevBytesRead)
}
rv.dicts[i] = dict
Expand All @@ -565,7 +565,7 @@ func (is *IndexSnapshot) TermFieldReader(term []byte, field string, includeFreq,

for i, segment := range is.segment {
var prevBytesReadPL uint64
if postings, ok := rv.postings[i].(diskStatsReporter); ok {
if postings, diskStatsAvailable := rv.postings[i].(diskStatsReporter); diskStatsAvailable {
prevBytesReadPL = postings.BytesRead()
}
pl, err := rv.dicts[i].PostingsList(term, segment.deleted, rv.postings[i])
Expand All @@ -575,18 +575,18 @@ func (is *IndexSnapshot) TermFieldReader(term []byte, field string, includeFreq,
rv.postings[i] = pl

var prevBytesReadItr uint64
if itr, ok := rv.iterators[i].(diskStatsReporter); ok {
if itr, diskStatsAvailable := rv.iterators[i].(diskStatsReporter); diskStatsAvailable {
prevBytesReadItr = itr.BytesRead()
}
rv.iterators[i] = pl.Iterator(includeFreq, includeNorm, includeTermVectors, rv.iterators[i])

if postings, ok := pl.(diskStatsReporter); ok &&
if postings, diskStatsAvailable := pl.(diskStatsReporter); diskStatsAvailable &&
prevBytesReadPL < postings.BytesRead() {
atomic.AddUint64(&is.parent.stats.TotBytesReadAtQueryTime,
postings.BytesRead()-prevBytesReadPL)
}

if itr, ok := rv.iterators[i].(diskStatsReporter); ok &&
if itr, diskStatsAvailable := rv.iterators[i].(diskStatsReporter); diskStatsAvailable &&
prevBytesReadItr < itr.BytesRead() {
atomic.AddUint64(&is.parent.stats.TotBytesReadAtQueryTime,
itr.BytesRead()-prevBytesReadItr)
Expand Down Expand Up @@ -710,15 +710,15 @@ func (i *IndexSnapshot) documentVisitFieldTermsOnSegment(

if ssvOk && ssv != nil && len(vFields) > 0 {
var prevBytesRead uint64
ssvp, ok := ssv.(segment.DiskStatsReporter)
if ok {
ssvp, diskStatsAvailable := ssv.(segment.DiskStatsReporter)
if diskStatsAvailable {
prevBytesRead = ssvp.BytesRead()
}
dvs, err = ssv.VisitDocValues(localDocNum, fields, visitor, dvs)
if err != nil {
return nil, nil, err
}
if ok {
if diskStatsAvailable {
atomic.AddUint64(&i.parent.stats.TotBytesReadAtQueryTime, ssvp.BytesRead()-prevBytesRead)
}
}
Expand Down
6 changes: 3 additions & 3 deletions index/scorch/snapshot_index_tfr.go
Original file line number Diff line number Diff line change
Expand Up @@ -77,8 +77,8 @@ func (i *IndexSnapshotTermFieldReader) Next(preAlloced *index.TermFieldDoc) (*in
// find the next hit
for i.segmentOffset < len(i.iterators) {
prevBytesRead := uint64(0)
itr, ok := i.iterators[i.segmentOffset].(segment.DiskStatsReporter)
if ok {
itr, diskStatsAvailable := i.iterators[i.segmentOffset].(segment.DiskStatsReporter)
if diskStatsAvailable {
prevBytesRead = itr.BytesRead()
}
next, err := i.iterators[i.segmentOffset].Next()
Expand All @@ -98,7 +98,7 @@ func (i *IndexSnapshotTermFieldReader) Next(preAlloced *index.TermFieldDoc) (*in
// this is because there are chances of having a series of loadChunk calls,
// and they have to be added together before sending the bytesRead at this point
// upstream.
if ok {
if diskStatsAvailable {
delta := itr.BytesRead() - prevBytesRead
atomic.AddUint64(&i.snapshot.parent.stats.TotBytesReadAtQueryTime, uint64(delta))
}
Expand Down

0 comments on commit 6284a48

Please sign in to comment.