Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

fix(tsi1/partition/test): fix data races in test code (#57) #25338

Merged
merged 3 commits into from
Sep 17, 2024
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
4 changes: 2 additions & 2 deletions tsdb/index/tsi1/index.go
Original file line number Diff line number Diff line change
Expand Up @@ -269,8 +269,8 @@ func (i *Index) Open() (rErr error) {
i.partitions = make([]*Partition, i.PartitionN)
for j := 0; j < len(i.partitions); j++ {
p := NewPartition(i.sfile, filepath.Join(i.path, fmt.Sprint(j)))
p.MaxLogFileSize = i.maxLogFileSize
p.MaxLogFileAge = i.maxLogFileAge
p.maxLogFileSize = i.maxLogFileSize
p.maxLogFileAge = i.maxLogFileAge
p.nosync = i.disableFsync
p.logbufferSize = i.logfileBufferSize
p.logger = i.logger.With(zap.String("tsi1_partition", fmt.Sprint(j+1)))
Expand Down
145 changes: 78 additions & 67 deletions tsdb/index/tsi1/partition.go
Original file line number Diff line number Diff line change
Expand Up @@ -39,8 +39,7 @@ const ManifestFileName = "MANIFEST"

// Partition represents a collection of layered index files and WAL.
type Partition struct {
// exported for tests
Mu sync.RWMutex
mu sync.RWMutex
opened bool

sfile *tsdb.SeriesFile // series lookup file
Expand Down Expand Up @@ -70,8 +69,9 @@ type Partition struct {
id string // id portion of path.

// Log file compaction thresholds.
MaxLogFileSize int64
MaxLogFileAge time.Duration
// Should be read/changed under the lock after a partition is opened.
maxLogFileSize int64
maxLogFileAge time.Duration
nosync bool // when true, flushing and syncing of LogFile will be disabled.
logbufferSize int // the LogFile's buffer is set to this value.

Expand All @@ -98,8 +98,7 @@ func NewPartition(sfile *tsdb.SeriesFile, path string) *Partition {
sfile: sfile,
seriesIDSet: tsdb.NewSeriesIDSet(),
fileSet: &FileSet{},
MaxLogFileSize: tsdb.DefaultMaxIndexLogFileSize,
MaxLogFileAge: tsdb.DefaultCompactFullWriteColdDuration,
maxLogFileSize: tsdb.DefaultMaxIndexLogFileSize,

// compactionEnabled: true,
compactionInterrupt: make(chan struct{}),
Expand All @@ -111,6 +110,16 @@ func NewPartition(sfile *tsdb.SeriesFile, path string) *Partition {
return p
}

// SetMaxLogFileSize provides a setter for the partition setting of maxLogFileSize
// that is otherwise only available at creation time. Returns the previous value.
// Only for tests!
func (p *Partition) SetMaxLogFileSize(new int64) (old int64) {
p.mu.Lock()
old, p.maxLogFileSize = p.maxLogFileSize, new
p.mu.Unlock()
return old
}

// bytes estimates the memory footprint of this Partition, in bytes.
func (p *Partition) bytes() int {
var b int
Expand All @@ -135,8 +144,8 @@ func (p *Partition) bytes() int {
b += int(unsafe.Sizeof(p.fieldset)) + p.fieldset.Bytes()
b += int(unsafe.Sizeof(p.path)) + len(p.path)
b += int(unsafe.Sizeof(p.id)) + len(p.id)
b += int(unsafe.Sizeof(p.MaxLogFileSize))
b += int(unsafe.Sizeof(p.MaxLogFileAge))
b += int(unsafe.Sizeof(p.maxLogFileSize))
b += int(unsafe.Sizeof(p.maxLogFileAge))
b += int(unsafe.Sizeof(p.compactionInterrupt))
b += int(unsafe.Sizeof(p.compactionsDisabled))
b += int(unsafe.Sizeof(p.logger))
Expand All @@ -151,8 +160,8 @@ var ErrIncompatibleVersion = errors.New("incompatible tsi1 index MANIFEST")

// Open opens the partition.
func (p *Partition) Open() (rErr error) {
p.Mu.Lock()
defer p.Mu.Unlock()
p.mu.Lock()
defer p.mu.Unlock()

p.closing = make(chan struct{})

Expand Down Expand Up @@ -214,7 +223,7 @@ func (p *Partition) Open() (rErr error) {

// Make first log file active, if within threshold.
sz, _ := f.Stat()
if p.activeLogFile == nil && sz < p.MaxLogFileSize {
if p.activeLogFile == nil && sz < p.maxLogFileSize {
p.activeLogFile = f
}

Expand Down Expand Up @@ -340,8 +349,8 @@ func (p *Partition) buildSeriesSet() error {

// CurrentCompactionN returns the number of compactions currently running.
func (p *Partition) CurrentCompactionN() int {
p.Mu.RLock()
defer p.Mu.RUnlock()
p.mu.RLock()
defer p.mu.RUnlock()
return p.currentCompactionN
}

Expand All @@ -368,20 +377,22 @@ func (p *Partition) Close() error {
p.Wait()

// Lock index and close remaining
p.Mu.Lock()
defer p.Mu.Unlock()
p.mu.Lock()
defer p.mu.Unlock()

var err error
if p.fileSet == nil {
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Nice

return nil
}

// Close log files.
var err []error
for _, f := range p.fileSet.files {
if localErr := f.Close(); localErr != nil {
err = localErr
}
localErr := f.Close()
err = append(err, localErr)
}
p.fileSet.files = nil

return err
return errors.Join(err...)
}

// closing returns true if the partition is currently closing. It does not require
Expand All @@ -403,8 +414,8 @@ func (p *Partition) SeriesFile() *tsdb.SeriesFile { return p.sfile }

// NextSequence returns the next file identifier.
func (p *Partition) NextSequence() int {
p.Mu.Lock()
defer p.Mu.Unlock()
p.mu.Lock()
defer p.mu.Unlock()
return p.nextSequence()
}

Expand Down Expand Up @@ -446,8 +457,8 @@ func (p *Partition) manifest(newFileSet *FileSet) *Manifest {

// SetManifestPathForTest is only to force a bad path in testing
func (p *Partition) SetManifestPathForTest(path string) {
p.Mu.Lock()
defer p.Mu.Unlock()
p.mu.Lock()
defer p.mu.Unlock()
p.manifestPathFn = func() string { return path }
}

Expand All @@ -458,16 +469,16 @@ func (p *Partition) WithLogger(logger *zap.Logger) {

// SetFieldSet sets a shared field set from the engine.
func (p *Partition) SetFieldSet(fs *tsdb.MeasurementFieldSet) {
p.Mu.Lock()
p.mu.Lock()
p.fieldset = fs
p.Mu.Unlock()
p.mu.Unlock()
}

// FieldSet returns the fieldset.
func (p *Partition) FieldSet() *tsdb.MeasurementFieldSet {
p.Mu.Lock()
p.mu.Lock()
fs := p.fieldset
p.Mu.Unlock()
p.mu.Unlock()
return fs
}

Expand All @@ -477,8 +488,8 @@ func (p *Partition) RetainFileSet() (*FileSet, error) {
case <-p.closing:
return nil, tsdb.ErrIndexClosing
default:
p.Mu.RLock()
defer p.Mu.RUnlock()
p.mu.RLock()
defer p.mu.RUnlock()
return p.retainFileSet(), nil
}
}
Expand All @@ -491,8 +502,8 @@ func (p *Partition) retainFileSet() *FileSet {

// FileN returns the active files in the file set.
func (p *Partition) FileN() int {
p.Mu.RLock()
defer p.Mu.RUnlock()
p.mu.RLock()
defer p.mu.RUnlock()
return len(p.fileSet.files)
}

Expand Down Expand Up @@ -672,12 +683,12 @@ func (p *Partition) DropMeasurement(name []byte) error {
// Mark measurement as deleted.
entries = append(entries, LogEntry{Flag: LogEntryMeasurementTombstoneFlag, Name: name})

p.Mu.RLock()
p.mu.RLock()
if err := p.activeLogFile.Writes(entries); err != nil {
p.Mu.RUnlock()
p.mu.RUnlock()
return err
}
p.Mu.RUnlock()
p.mu.RUnlock()

// Check if the log file needs to be swapped.
if err := p.CheckLogFile(); err != nil {
Expand Down Expand Up @@ -705,14 +716,14 @@ func (p *Partition) createSeriesListIfNotExists(names [][]byte, tagsSlice []mode
defer fs.Release()

// Ensure fileset cannot change during insert.
p.Mu.RLock()
p.mu.RLock()
// Insert series into log file.
ids, err := p.activeLogFile.AddSeriesList(p.seriesIDSet, names, tagsSlice, tracker)
if err != nil {
p.Mu.RUnlock()
p.mu.RUnlock()
return nil, err
}
p.Mu.RUnlock()
p.mu.RUnlock()

if err := p.CheckLogFile(); err != nil {
return nil, err
Expand All @@ -723,8 +734,8 @@ func (p *Partition) createSeriesListIfNotExists(names [][]byte, tagsSlice []mode
func (p *Partition) DropSeries(seriesID uint64) error {
// Delete series from index.
if err := func() error {
p.Mu.RLock()
defer p.Mu.RUnlock()
p.mu.RLock()
defer p.mu.RUnlock()
return p.activeLogFile.DeleteSeriesID(seriesID)
}(); err != nil {
return err
Expand All @@ -743,8 +754,8 @@ func (p *Partition) DropSeriesList(seriesIDs []uint64) error {

// Delete series from index.
if err := func() error {
p.Mu.RLock()
defer p.Mu.RUnlock()
p.mu.RLock()
defer p.mu.RUnlock()
return p.activeLogFile.DeleteSeriesIDList(seriesIDs)
}(); err != nil {
return err
Expand Down Expand Up @@ -911,14 +922,14 @@ func (p *Partition) AssignShard(k string, shardID uint64) {}

// Compact requests a compaction of log files.
func (p *Partition) Compact() {
p.Mu.Lock()
defer p.Mu.Unlock()
p.mu.Lock()
defer p.mu.Unlock()
p.compact()
}

func (p *Partition) DisableCompactions() {
p.Mu.Lock()
defer p.Mu.Unlock()
p.mu.Lock()
defer p.mu.Unlock()
p.compactionsDisabled++

select {
Expand All @@ -934,8 +945,8 @@ func (p *Partition) DisableCompactions() {
}

func (p *Partition) EnableCompactions() {
p.Mu.Lock()
defer p.Mu.Unlock()
p.mu.Lock()
defer p.mu.Unlock()

// Already enabled?
if p.compactionsEnabled() {
Expand All @@ -953,9 +964,9 @@ func (p *Partition) runPeriodicCompaction() {
p.Compact()

// Avoid a race when using Reopen in tests
p.Mu.RLock()
p.mu.RLock()
closing := p.closing
p.Mu.RUnlock()
p.mu.RUnlock()

// check for compactions once an hour (usually not necessary but a nice safety check)
t := time.NewTicker(1 * time.Hour)
Expand All @@ -978,8 +989,8 @@ func (p *Partition) runPeriodicCompaction() {
// If checkRunning = true, only count as needing a compaction if there is not a compaction already
// in progress for the level that would be compacted
func (p *Partition) NeedsCompaction(checkRunning bool) bool {
p.Mu.RLock()
defer p.Mu.RUnlock()
p.mu.RLock()
defer p.mu.RUnlock()
if p.needsLogCompaction() {
return true
}
Expand Down Expand Up @@ -1032,10 +1043,10 @@ func (p *Partition) compact() {
p.currentCompactionN++
go func() {
p.compactLogFile(logFile)
p.Mu.Lock()
p.mu.Lock()
p.currentCompactionN--
p.levelCompacting[0] = false
p.Mu.Unlock()
p.mu.Unlock()
p.Compact()
}()
}
Expand Down Expand Up @@ -1075,10 +1086,10 @@ func (p *Partition) compact() {
p.compactToLevel(files, level+1, interrupt)

// Ensure compaction lock for the level is released.
p.Mu.Lock()
p.mu.Lock()
p.levelCompacting[level] = false
p.currentCompactionN--
p.Mu.Unlock()
p.mu.Unlock()

// Check for new compactions
p.Compact()
Expand Down Expand Up @@ -1156,8 +1167,8 @@ func (p *Partition) compactToLevel(files []*IndexFile, level int, interrupt <-ch

// Obtain lock to swap in index file and write manifest.
if err := func() (rErr error) {
p.Mu.Lock()
defer p.Mu.Unlock()
p.mu.Lock()
defer p.mu.Unlock()

// Replace previous files with new index file.
newFileSet := p.fileSet.MustReplace(IndexFiles(files).Files(), file)
Expand Down Expand Up @@ -1215,23 +1226,23 @@ func (p *Partition) Rebuild() {}
func (p *Partition) needsLogCompaction() bool {
size := p.activeLogFile.Size()
modTime := p.activeLogFile.ModTime()
return size >= p.MaxLogFileSize || (size > 0 && modTime.Before(time.Now().Add(-p.MaxLogFileAge)))
return size >= p.maxLogFileSize || (size > 0 && modTime.Before(time.Now().Add(-p.maxLogFileAge)))
}

func (p *Partition) CheckLogFile() error {
// Check log file under read lock.
needsCompaction := func() bool {
p.Mu.RLock()
defer p.Mu.RUnlock()
p.mu.RLock()
defer p.mu.RUnlock()
return p.needsLogCompaction()
}()
if !needsCompaction {
return nil
}

// If file size exceeded then recheck under write lock and swap files.
p.Mu.Lock()
defer p.Mu.Unlock()
p.mu.Lock()
defer p.mu.Unlock()
return p.checkLogFile()
}

Expand Down Expand Up @@ -1261,9 +1272,9 @@ func (p *Partition) compactLogFile(logFile *LogFile) {
return
}

p.Mu.Lock()
p.mu.Lock()
interrupt := p.compactionInterrupt
p.Mu.Unlock()
p.mu.Unlock()

start := time.Now()

Expand Down Expand Up @@ -1313,8 +1324,8 @@ func (p *Partition) compactLogFile(logFile *LogFile) {

// Obtain lock to swap in index file and write manifest.
if err := func() (rErr error) {
p.Mu.Lock()
defer p.Mu.Unlock()
p.mu.Lock()
defer p.mu.Unlock()

// Replace previous log file with index file.
newFileSet := p.fileSet.MustReplace([]File{logFile}, file)
Expand Down
Loading