package zstd import ( "fmt" "math/bits" "github.com/klauspost/compress/zstd/internal/xxhash" ) const ( dictShardBits = 6 ) type fastBase struct { // cur is the offset at the start of hist cur int32 // maximum offset. Should be at least 2x block size. maxMatchOff int32 hist []byte crc *xxhash.Digest tmp [8]byte blk *blockEnc lastDictID uint32 lowMem bool } // CRC returns the underlying CRC writer. func (e *fastBase) CRC() *xxhash.Digest { return e.crc } // AppendCRC will append the CRC to the destination slice and return it. func (e *fastBase) AppendCRC(dst []byte) []byte { crc := e.crc.Sum(e.tmp[:0]) dst = append(dst, crc[7], crc[6], crc[5], crc[4]) return dst } // WindowSize returns the window size of the encoder, // or a window size small enough to contain the input size, if > 0. func (e *fastBase) WindowSize(size int) int32 { if size > 0 && size < int(e.maxMatchOff) { b := int32(1) << uint(bits.Len(uint(size))) // Keep minimum window. if b < 1024 { b = 1024 } return b } return e.maxMatchOff } // Block returns the current block. func (e *fastBase) Block() *blockEnc { return e.blk } func (e *fastBase) addBlock(src []byte) int32 { if debugAsserts && e.cur > bufferReset { panic(fmt.Sprintf("ecur (%d) > buffer reset (%d)", e.cur, bufferReset)) } // check if we have space already if len(e.hist)+len(src) > cap(e.hist) { if cap(e.hist) == 0 { e.ensureHist(len(src)) } else { if cap(e.hist) < int(e.maxMatchOff+maxCompressedBlockSize) { panic(fmt.Errorf("unexpected buffer cap %d, want at least %d with window %d", cap(e.hist), e.maxMatchOff+maxCompressedBlockSize, e.maxMatchOff)) } // Move down offset := int32(len(e.hist)) - e.maxMatchOff copy(e.hist[0:e.maxMatchOff], e.hist[offset:]) e.cur += offset e.hist = e.hist[:e.maxMatchOff] } } s := int32(len(e.hist)) e.hist = append(e.hist, src...) return s } // ensureHist will ensure that history can keep at least this many bytes. func (e *fastBase) ensureHist(n int) { if cap(e.hist) >= n { return } l := e.maxMatchOff if (e.lowMem && e.maxMatchOff > maxCompressedBlockSize) || e.maxMatchOff <= maxCompressedBlockSize { l += maxCompressedBlockSize } else { l += e.maxMatchOff } // Make it at least 1MB. if l < 1<<20 && !e.lowMem { l = 1 << 20 } // Make it at least the requested size. if l < int32(n) { l = int32(n) } e.hist = make([]byte, 0, l) } // useBlock will replace the block with the provided one, // but transfer recent offsets from the previous. func (e *fastBase) UseBlock(enc *blockEnc) { enc.reset(e.blk) e.blk = enc } func (e *fastBase) matchlenNoHist(s, t int32, src []byte) int32 { // Extend the match to be as long as possible. return int32(matchLen(src[s:], src[t:])) } func (e *fastBase) matchlen(s, t int32, src []byte) int32 { if debugAsserts { if s < 0 { err := fmt.Sprintf("s (%d) < 0", s) panic(err) } if t < 0 { err := fmt.Sprintf("s (%d) < 0", s) panic(err) } if s-t > e.maxMatchOff { err := fmt.Sprintf("s (%d) - t (%d) > maxMatchOff (%d)", s, t, e.maxMatchOff) panic(err) } if len(src)-int(s) > maxCompressedBlockSize { panic(fmt.Sprintf("len(src)-s (%d) > maxCompressedBlockSize (%d)", len(src)-int(s), maxCompressedBlockSize)) } } // Extend the match to be as long as possible. return int32(matchLen(src[s:], src[t:])) } // Reset the encoding table. func (e *fastBase) resetBase(d *dict, singleBlock bool) { if e.blk == nil { e.blk = &blockEnc{lowMem: e.lowMem} e.blk.init() } else { e.blk.reset(nil) } e.blk.initNewEncode() if e.crc == nil { e.crc = xxhash.New() } else { e.crc.Reset() } if d != nil { low := e.lowMem if singleBlock { e.lowMem = true } e.ensureHist(d.DictContentSize() + maxCompressedBlockSize) e.lowMem = low } // We offset current position so everything will be out of reach. // If above reset line, history will be purged. if e.cur < bufferReset { e.cur += e.maxMatchOff + int32(len(e.hist)) } e.hist = e.hist[:0] if d != nil { // Set offsets (currently not used) for i, off := range d.offsets { e.blk.recentOffsets[i] = uint32(off) e.blk.prevRecentOffsets[i] = e.blk.recentOffsets[i] } // Transfer litenc. e.blk.dictLitEnc = d.litEnc e.hist = append(e.hist, d.content...) } }