package tsm1 import ( "bufio" "compress/gzip" "encoding/binary" "errors" "fmt" "io" "math" "os" "path/filepath" "strings" "sync" "github.com/influxdata/influxdb/v2/pkg/file" "github.com/influxdata/influxdb/v2/tsdb" ) const TombstoneFileExtension = "tombstone" const ( headerSize = 4 v2header = 0x1502 v3header = 0x1503 v4header = 0x1504 ) var errIncompatibleVersion = errors.New("incompatible v4 version") // Tombstoner records tombstones when entries are deleted. type Tombstoner struct { mu sync.RWMutex // Path is the location of the file to record tombstone. This should be the // full path to a TSM file. Path string FilterFn func(k []byte) bool // Tombstones that have been written but not flushed to disk yet. tombstones []Tombstone // cache of the stats for this tombstone tombstoneStats TombstoneStat // indicates that the stats may be out of sync with what is on disk and they // should be refreshed. statsLoaded bool // These are references used for pending writes that have not been committed. If // these are nil, then no pending writes are in progress. gz *gzip.Writer bw *bufio.Writer pendingFile *os.File tmp [8]byte lastAppliedOffset int64 // Optional observer for when tombstone files are written. obs tsdb.FileStoreObserver } // NewTombstoner constructs a Tombstoner for the given path. FilterFn can be nil. func NewTombstoner(path string, filterFn func(k []byte) bool) *Tombstoner { return &Tombstoner{ Path: path, FilterFn: filterFn, obs: noFileStoreObserver{}, } } // Tombstone represents an individual deletion. type Tombstone struct { // Key is the tombstoned series key. Key []byte // Min and Max are the min and max unix nanosecond time ranges of Key that are deleted. If // the full range is deleted, both values are -1. Min, Max int64 } // WithObserver sets a FileStoreObserver for when the tombstone file is written. func (t *Tombstoner) WithObserver(obs tsdb.FileStoreObserver) { t.obs = obs } // Add adds the all keys, across all timestamps, to the tombstone. func (t *Tombstoner) Add(keys [][]byte) error { return t.AddRange(keys, math.MinInt64, math.MaxInt64) } // AddRange adds all keys to the tombstone specifying only the data between min and max to be removed. func (t *Tombstoner) AddRange(keys [][]byte, min, max int64) error { for t.FilterFn != nil && len(keys) > 0 && !t.FilterFn(keys[0]) { keys = keys[1:] } if len(keys) == 0 { return nil } t.mu.Lock() defer t.mu.Unlock() // If this TSMFile has not been written (mainly in tests), don't write a // tombstone because the keys will not be written when it's actually saved. if t.Path == "" { return nil } t.statsLoaded = false if err := t.prepareV4(); err == errIncompatibleVersion { if cap(t.tombstones) < len(t.tombstones)+len(keys) { ts := make([]Tombstone, len(t.tombstones), len(t.tombstones)+len(keys)) copy(ts, t.tombstones) t.tombstones = ts } for _, k := range keys { if t.FilterFn != nil && !t.FilterFn(k) { continue } t.tombstones = append(t.tombstones, Tombstone{ Key: k, Min: min, Max: max, }) } return t.writeTombstoneV3(t.tombstones) } else if err != nil { return err } for _, k := range keys { if t.FilterFn != nil && !t.FilterFn(k) { continue } if err := t.writeTombstone(t.gz, Tombstone{ Key: k, Min: min, Max: max, }); err != nil { return err } } return nil } func (t *Tombstoner) Flush() error { t.mu.Lock() defer t.mu.Unlock() if err := t.commit(); err != nil { // Reset our temp references and clean up. _ = t.rollback() return err } return nil } func (t *Tombstoner) Rollback() error { t.mu.Lock() defer t.mu.Unlock() return t.rollback() } // Delete removes all the tombstone files from disk. func (t *Tombstoner) Delete() error { t.mu.Lock() defer t.mu.Unlock() if err := os.RemoveAll(t.tombstonePath()); err != nil { return err } t.statsLoaded = false t.lastAppliedOffset = 0 return nil } // HasTombstones return true if there are any tombstone entries recorded. func (t *Tombstoner) HasTombstones() bool { stats := t.TombstoneStats() if !stats.TombstoneExists { return false } if stats.Size > 0 { return true } t.mu.RLock() n := len(t.tombstones) t.mu.RUnlock() return n > 0 } // TombstoneFiles returns any tombstone files associated with Tombstoner's TSM file. func (t *Tombstoner) TombstoneStats() TombstoneStat { t.mu.RLock() if t.statsLoaded { stats := t.tombstoneStats t.mu.RUnlock() return stats } t.mu.RUnlock() stat, err := os.Stat(t.tombstonePath()) if err != nil { t.mu.Lock() // The file doesn't exist so record that we tried to load it so // we don't continue to keep trying. This is the common case. t.statsLoaded = os.IsNotExist(err) t.tombstoneStats.TombstoneExists = false stats := t.tombstoneStats t.mu.Unlock() return stats } t.mu.Lock() t.tombstoneStats = TombstoneStat{ TombstoneExists: true, Path: t.tombstonePath(), LastModified: stat.ModTime().UnixNano(), Size: uint32(stat.Size()), } t.statsLoaded = true stats := t.tombstoneStats t.mu.Unlock() return stats } // Walk calls fn for every Tombstone under the Tombstoner. func (t *Tombstoner) Walk(fn func(t Tombstone) error) error { t.mu.Lock() defer t.mu.Unlock() f, err := os.Open(t.tombstonePath()) if os.IsNotExist(err) { return nil } else if err != nil { return err } defer f.Close() var b [4]byte if _, err := f.Read(b[:]); err != nil { // Might be a zero length file which should not exist, but // an old bug allowed them to occur. Treat it as an empty // v1 tombstone file so we don't abort loading the TSM file. return t.readTombstoneV1(f, fn) } if _, err := f.Seek(0, io.SeekStart); err != nil { return err } header := binary.BigEndian.Uint32(b[:]) if header == v4header { return t.readTombstoneV4(f, fn) } else if header == v3header { return t.readTombstoneV3(f, fn) } else if header == v2header { return t.readTombstoneV2(f, fn) } return t.readTombstoneV1(f, fn) } func (t *Tombstoner) writeTombstoneV3(tombstones []Tombstone) error { tmp, err := os.CreateTemp(filepath.Dir(t.Path), TombstoneFileExtension) if err != nil { return err } defer tmp.Close() var b [8]byte bw := bufio.NewWriterSize(tmp, 1024*1024) binary.BigEndian.PutUint32(b[:4], v3header) if _, err := bw.Write(b[:4]); err != nil { return err } gz := gzip.NewWriter(bw) for _, ts := range tombstones { if err := t.writeTombstone(gz, ts); err != nil { return err } } t.gz = gz t.bw = bw t.pendingFile = tmp t.tombstones = t.tombstones[:0] return t.commit() } func (t *Tombstoner) prepareV4() error { if t.pendingFile != nil { return nil } tmpPath := fmt.Sprintf("%s.%s", t.tombstonePath(), CompactionTempExtension) tmp, err := os.OpenFile(tmpPath, os.O_CREATE|os.O_RDWR|os.O_EXCL, 0666) if err != nil { return err } removeTmp := func() { tmp.Close() os.Remove(tmp.Name()) } // Copy the existing v4 file if it exists f, err := os.Open(t.tombstonePath()) if !os.IsNotExist(err) { defer f.Close() var b [4]byte if n, err := f.Read(b[:]); n == 4 && err == nil { header := binary.BigEndian.Uint32(b[:]) // There is an existing tombstone on disk and it's not a v3. Just rewrite it as a v3 // version again. if header != v4header { removeTmp() return errIncompatibleVersion } // Seek back to the beginning we copy the header if _, err := f.Seek(0, io.SeekStart); err != nil { removeTmp() return err } // Copy the while file if _, err := io.Copy(tmp, f); err != nil { f.Close() removeTmp() return err } } } else if err != nil && !os.IsNotExist(err) { removeTmp() return err } var b [8]byte bw := bufio.NewWriterSize(tmp, 64*1024) // Write the header only if the file is new if os.IsNotExist(err) { binary.BigEndian.PutUint32(b[:4], v4header) if _, err := bw.Write(b[:4]); err != nil { removeTmp() return err } } // Write the tombstones gz := gzip.NewWriter(bw) t.pendingFile = tmp t.gz = gz t.bw = bw return nil } func (t *Tombstoner) commit() error { // No pending writes if t.pendingFile == nil { return nil } if err := t.gz.Close(); err != nil { return err } if err := t.bw.Flush(); err != nil { return err } // fsync the file to flush the write if err := t.pendingFile.Sync(); err != nil { return err } tmpFilename := t.pendingFile.Name() t.pendingFile.Close() if err := t.obs.FileFinishing(tmpFilename); err != nil { return err } if err := file.RenameFile(tmpFilename, t.tombstonePath()); err != nil { return err } if err := file.SyncDir(filepath.Dir(t.tombstonePath())); err != nil { return err } t.pendingFile = nil t.bw = nil t.gz = nil return nil } func (t *Tombstoner) rollback() error { if t.pendingFile == nil { return nil } tmpFilename := t.pendingFile.Name() t.pendingFile.Close() t.gz = nil t.bw = nil t.pendingFile = nil return os.Remove(tmpFilename) } // readTombstoneV1 reads the first version of tombstone files that were not // capable of storing a min and max time for a key. This is used for backwards // compatibility with versions prior to 0.13. This format is a simple newline // separated text file. func (t *Tombstoner) readTombstoneV1(f *os.File, fn func(t Tombstone) error) error { r := bufio.NewScanner(f) for r.Scan() { line := r.Text() if line == "" { continue } if err := fn(Tombstone{ Key: []byte(line), Min: math.MinInt64, Max: math.MaxInt64, }); err != nil { return err } } if err := r.Err(); err != nil { return err } for _, t := range t.tombstones { if err := fn(t); err != nil { return err } } return nil } // readTombstoneV2 reads the second version of tombstone files that are capable // of storing keys and the range of time for the key that points were deleted. This // format is binary. func (t *Tombstoner) readTombstoneV2(f *os.File, fn func(t Tombstone) error) error { // Skip header, already checked earlier if _, err := f.Seek(headerSize, io.SeekStart); err != nil { return err } n := int64(4) fi, err := f.Stat() if err != nil { return err } size := fi.Size() var ( min, max int64 key []byte ) b := make([]byte, 4096) for { if n >= size { break } if _, err = f.Read(b[:4]); err != nil { return err } n += 4 keyLen := int(binary.BigEndian.Uint32(b[:4])) if keyLen > len(b) { b = make([]byte, keyLen) } if _, err := f.Read(b[:keyLen]); err != nil { return err } key = b[:keyLen] n += int64(keyLen) if _, err := f.Read(b[:8]); err != nil { return err } n += 8 min = int64(binary.BigEndian.Uint64(b[:8])) if _, err := f.Read(b[:8]); err != nil { return err } n += 8 max = int64(binary.BigEndian.Uint64(b[:8])) if err := fn(Tombstone{ Key: key, Min: min, Max: max, }); err != nil { return err } } for _, t := range t.tombstones { if err := fn(t); err != nil { return err } } return nil } // readTombstoneV3 reads the third version of tombstone files that are capable // of storing keys and the range of time for the key that points were deleted. This // format is a binary and compressed with gzip. func (t *Tombstoner) readTombstoneV3(f *os.File, fn func(t Tombstone) error) error { // Skip header, already checked earlier if _, err := f.Seek(headerSize, io.SeekStart); err != nil { return err } var ( min, max int64 key []byte ) gr, err := gzip.NewReader(bufio.NewReader(f)) if err != nil { return err } defer gr.Close() b := make([]byte, 4096) for { if _, err = io.ReadFull(gr, b[:4]); err == io.EOF || err == io.ErrUnexpectedEOF { break } else if err != nil { return err } keyLen := int(binary.BigEndian.Uint32(b[:4])) if keyLen > len(b) { b = make([]byte, keyLen) } if _, err := io.ReadFull(gr, b[:keyLen]); err != nil { return err } // Copy the key since b is re-used key = make([]byte, keyLen) copy(key, b[:keyLen]) if _, err := io.ReadFull(gr, b[:8]); err != nil { return err } min = int64(binary.BigEndian.Uint64(b[:8])) if _, err := io.ReadFull(gr, b[:8]); err != nil { return err } max = int64(binary.BigEndian.Uint64(b[:8])) if err := fn(Tombstone{ Key: key, Min: min, Max: max, }); err != nil { return err } } for _, t := range t.tombstones { if err := fn(t); err != nil { return err } } return nil } // readTombstoneV4 reads the fourth version of tombstone files that are capable // of storing multiple v3 files appended together. func (t *Tombstoner) readTombstoneV4(f *os.File, fn func(t Tombstone) error) error { // Skip header, already checked earlier if t.lastAppliedOffset != 0 { if _, err := f.Seek(t.lastAppliedOffset, io.SeekStart); err != nil { return err } } else { if _, err := f.Seek(headerSize, io.SeekStart); err != nil { return err } } var ( min, max int64 key []byte ) br := bufio.NewReaderSize(f, 64*1024) gr, err := gzip.NewReader(br) if err == io.EOF { return nil } else if err != nil { return err } defer gr.Close() b := make([]byte, 4096) for { gr.Multistream(false) if err := func() error { for { if _, err = io.ReadFull(gr, b[:4]); err == io.EOF || err == io.ErrUnexpectedEOF { return nil } else if err != nil { return err } keyLen := int(binary.BigEndian.Uint32(b[:4])) if keyLen+16 > len(b) { b = make([]byte, keyLen+16) } if _, err := io.ReadFull(gr, b[:keyLen]); err != nil { return err } // Copy the key since b is re-used key = b[:keyLen] minBuf := b[keyLen : keyLen+8] maxBuf := b[keyLen+8 : keyLen+16] if _, err := io.ReadFull(gr, minBuf); err != nil { return err } min = int64(binary.BigEndian.Uint64(minBuf)) if _, err := io.ReadFull(gr, maxBuf); err != nil { return err } max = int64(binary.BigEndian.Uint64(maxBuf)) if err := fn(Tombstone{ Key: key, Min: min, Max: max, }); err != nil { return err } } }(); err != nil { return err } for _, t := range t.tombstones { if err := fn(t); err != nil { return err } } err = gr.Reset(br) if err == io.EOF { break } } // Save the position of tombstone file so we don't re-apply the same set again if there are // more deletes. pos, err := f.Seek(0, io.SeekCurrent) if err != nil { return err } t.lastAppliedOffset = pos return nil } func (t *Tombstoner) tombstonePath() string { if strings.HasSuffix(t.Path, TombstoneFileExtension) { return t.Path } // Filename is 0000001.tsm1 filename := filepath.Base(t.Path) // Strip off the tsm1 ext := filepath.Ext(filename) if ext != "" { filename = strings.TrimSuffix(filename, ext) } // Append the "tombstone" suffix to create a 0000001.tombstone file return filepath.Join(filepath.Dir(t.Path), filename+"."+TombstoneFileExtension) } func (t *Tombstoner) writeTombstone(dst io.Writer, ts Tombstone) error { binary.BigEndian.PutUint32(t.tmp[:4], uint32(len(ts.Key))) if _, err := dst.Write(t.tmp[:4]); err != nil { return err } if _, err := dst.Write([]byte(ts.Key)); err != nil { return err } binary.BigEndian.PutUint64(t.tmp[:], uint64(ts.Min)) if _, err := dst.Write(t.tmp[:]); err != nil { return err } binary.BigEndian.PutUint64(t.tmp[:], uint64(ts.Max)) _, err := dst.Write(t.tmp[:]) return err }