forked from toolshed/abra
		
	We were running behind and there were quite some deprecations to update. This was mostly in the upstream copy/pasta package but seems quite minimal.
		
			
				
	
	
		
			713 lines
		
	
	
		
			18 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
			
		
		
	
	
			713 lines
		
	
	
		
			18 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
| // Copyright 2019+ Klaus Post. All rights reserved.
 | |
| // License information can be found in the LICENSE file.
 | |
| // Based on work by Yann Collet, released under BSD License.
 | |
| 
 | |
| package zstd
 | |
| 
 | |
| import (
 | |
| 	"errors"
 | |
| 	"fmt"
 | |
| 	"hash/crc32"
 | |
| 	"io"
 | |
| 	"sync"
 | |
| 
 | |
| 	"github.com/klauspost/compress/huff0"
 | |
| 	"github.com/klauspost/compress/zstd/internal/xxhash"
 | |
| )
 | |
| 
 | |
| type blockType uint8
 | |
| 
 | |
| //go:generate stringer -type=blockType,literalsBlockType,seqCompMode,tableIndex
 | |
| 
 | |
| const (
 | |
| 	blockTypeRaw blockType = iota
 | |
| 	blockTypeRLE
 | |
| 	blockTypeCompressed
 | |
| 	blockTypeReserved
 | |
| )
 | |
| 
 | |
| type literalsBlockType uint8
 | |
| 
 | |
| const (
 | |
| 	literalsBlockRaw literalsBlockType = iota
 | |
| 	literalsBlockRLE
 | |
| 	literalsBlockCompressed
 | |
| 	literalsBlockTreeless
 | |
| )
 | |
| 
 | |
| const (
 | |
| 	// maxCompressedBlockSize is the biggest allowed compressed block size (128KB)
 | |
| 	maxCompressedBlockSize = 128 << 10
 | |
| 
 | |
| 	compressedBlockOverAlloc    = 16
 | |
| 	maxCompressedBlockSizeAlloc = 128<<10 + compressedBlockOverAlloc
 | |
| 
 | |
| 	// Maximum possible block size (all Raw+Uncompressed).
 | |
| 	maxBlockSize = (1 << 21) - 1
 | |
| 
 | |
| 	maxMatchLen  = 131074
 | |
| 	maxSequences = 0x7f00 + 0xffff
 | |
| 
 | |
| 	// We support slightly less than the reference decoder to be able to
 | |
| 	// use ints on 32 bit archs.
 | |
| 	maxOffsetBits = 30
 | |
| )
 | |
| 
 | |
| var (
 | |
| 	huffDecoderPool = sync.Pool{New: func() interface{} {
 | |
| 		return &huff0.Scratch{}
 | |
| 	}}
 | |
| 
 | |
| 	fseDecoderPool = sync.Pool{New: func() interface{} {
 | |
| 		return &fseDecoder{}
 | |
| 	}}
 | |
| )
 | |
| 
 | |
| type blockDec struct {
 | |
| 	// Raw source data of the block.
 | |
| 	data        []byte
 | |
| 	dataStorage []byte
 | |
| 
 | |
| 	// Destination of the decoded data.
 | |
| 	dst []byte
 | |
| 
 | |
| 	// Buffer for literals data.
 | |
| 	literalBuf []byte
 | |
| 
 | |
| 	// Window size of the block.
 | |
| 	WindowSize uint64
 | |
| 
 | |
| 	err error
 | |
| 
 | |
| 	// Check against this crc, if hasCRC is true.
 | |
| 	checkCRC uint32
 | |
| 	hasCRC   bool
 | |
| 
 | |
| 	// Frame to use for singlethreaded decoding.
 | |
| 	// Should not be used by the decoder itself since parent may be another frame.
 | |
| 	localFrame *frameDec
 | |
| 
 | |
| 	sequence []seqVals
 | |
| 
 | |
| 	async struct {
 | |
| 		newHist  *history
 | |
| 		literals []byte
 | |
| 		seqData  []byte
 | |
| 		seqSize  int // Size of uncompressed sequences
 | |
| 		fcs      uint64
 | |
| 	}
 | |
| 
 | |
| 	// Block is RLE, this is the size.
 | |
| 	RLESize uint32
 | |
| 
 | |
| 	Type blockType
 | |
| 
 | |
| 	// Is this the last block of a frame?
 | |
| 	Last bool
 | |
| 
 | |
| 	// Use less memory
 | |
| 	lowMem bool
 | |
| }
 | |
| 
 | |
| func (b *blockDec) String() string {
 | |
| 	if b == nil {
 | |
| 		return "<nil>"
 | |
| 	}
 | |
| 	return fmt.Sprintf("Steam Size: %d, Type: %v, Last: %t, Window: %d", len(b.data), b.Type, b.Last, b.WindowSize)
 | |
| }
 | |
| 
 | |
| func newBlockDec(lowMem bool) *blockDec {
 | |
| 	b := blockDec{
 | |
| 		lowMem: lowMem,
 | |
| 	}
 | |
| 	return &b
 | |
| }
 | |
| 
 | |
| // reset will reset the block.
 | |
| // Input must be a start of a block and will be at the end of the block when returned.
 | |
| func (b *blockDec) reset(br byteBuffer, windowSize uint64) error {
 | |
| 	b.WindowSize = windowSize
 | |
| 	tmp, err := br.readSmall(3)
 | |
| 	if err != nil {
 | |
| 		println("Reading block header:", err)
 | |
| 		return err
 | |
| 	}
 | |
| 	bh := uint32(tmp[0]) | (uint32(tmp[1]) << 8) | (uint32(tmp[2]) << 16)
 | |
| 	b.Last = bh&1 != 0
 | |
| 	b.Type = blockType((bh >> 1) & 3)
 | |
| 	// find size.
 | |
| 	cSize := int(bh >> 3)
 | |
| 	maxSize := maxCompressedBlockSizeAlloc
 | |
| 	switch b.Type {
 | |
| 	case blockTypeReserved:
 | |
| 		return ErrReservedBlockType
 | |
| 	case blockTypeRLE:
 | |
| 		if cSize > maxCompressedBlockSize || cSize > int(b.WindowSize) {
 | |
| 			if debugDecoder {
 | |
| 				printf("rle block too big: csize:%d block: %+v\n", uint64(cSize), b)
 | |
| 			}
 | |
| 			return ErrWindowSizeExceeded
 | |
| 		}
 | |
| 		b.RLESize = uint32(cSize)
 | |
| 		if b.lowMem {
 | |
| 			maxSize = cSize
 | |
| 		}
 | |
| 		cSize = 1
 | |
| 	case blockTypeCompressed:
 | |
| 		if debugDecoder {
 | |
| 			println("Data size on stream:", cSize)
 | |
| 		}
 | |
| 		b.RLESize = 0
 | |
| 		maxSize = maxCompressedBlockSizeAlloc
 | |
| 		if windowSize < maxCompressedBlockSize && b.lowMem {
 | |
| 			maxSize = int(windowSize) + compressedBlockOverAlloc
 | |
| 		}
 | |
| 		if cSize > maxCompressedBlockSize || uint64(cSize) > b.WindowSize {
 | |
| 			if debugDecoder {
 | |
| 				printf("compressed block too big: csize:%d block: %+v\n", uint64(cSize), b)
 | |
| 			}
 | |
| 			return ErrCompressedSizeTooBig
 | |
| 		}
 | |
| 		// Empty compressed blocks must at least be 2 bytes
 | |
| 		// for Literals_Block_Type and one for Sequences_Section_Header.
 | |
| 		if cSize < 2 {
 | |
| 			return ErrBlockTooSmall
 | |
| 		}
 | |
| 	case blockTypeRaw:
 | |
| 		if cSize > maxCompressedBlockSize || cSize > int(b.WindowSize) {
 | |
| 			if debugDecoder {
 | |
| 				printf("rle block too big: csize:%d block: %+v\n", uint64(cSize), b)
 | |
| 			}
 | |
| 			return ErrWindowSizeExceeded
 | |
| 		}
 | |
| 
 | |
| 		b.RLESize = 0
 | |
| 		// We do not need a destination for raw blocks.
 | |
| 		maxSize = -1
 | |
| 	default:
 | |
| 		panic("Invalid block type")
 | |
| 	}
 | |
| 
 | |
| 	// Read block data.
 | |
| 	if _, ok := br.(*byteBuf); !ok && cap(b.dataStorage) < cSize {
 | |
| 		// byteBuf doesn't need a destination buffer.
 | |
| 		if b.lowMem || cSize > maxCompressedBlockSize {
 | |
| 			b.dataStorage = make([]byte, 0, cSize+compressedBlockOverAlloc)
 | |
| 		} else {
 | |
| 			b.dataStorage = make([]byte, 0, maxCompressedBlockSizeAlloc)
 | |
| 		}
 | |
| 	}
 | |
| 	b.data, err = br.readBig(cSize, b.dataStorage)
 | |
| 	if err != nil {
 | |
| 		if debugDecoder {
 | |
| 			println("Reading block:", err, "(", cSize, ")", len(b.data))
 | |
| 			printf("%T", br)
 | |
| 		}
 | |
| 		return err
 | |
| 	}
 | |
| 	if cap(b.dst) <= maxSize {
 | |
| 		b.dst = make([]byte, 0, maxSize+1)
 | |
| 	}
 | |
| 	return nil
 | |
| }
 | |
| 
 | |
| // sendEOF will make the decoder send EOF on this frame.
 | |
| func (b *blockDec) sendErr(err error) {
 | |
| 	b.Last = true
 | |
| 	b.Type = blockTypeReserved
 | |
| 	b.err = err
 | |
| }
 | |
| 
 | |
| // Close will release resources.
 | |
| // Closed blockDec cannot be reset.
 | |
| func (b *blockDec) Close() {
 | |
| }
 | |
| 
 | |
| // decodeBuf
 | |
| func (b *blockDec) decodeBuf(hist *history) error {
 | |
| 	switch b.Type {
 | |
| 	case blockTypeRLE:
 | |
| 		if cap(b.dst) < int(b.RLESize) {
 | |
| 			if b.lowMem {
 | |
| 				b.dst = make([]byte, b.RLESize)
 | |
| 			} else {
 | |
| 				b.dst = make([]byte, maxCompressedBlockSize)
 | |
| 			}
 | |
| 		}
 | |
| 		b.dst = b.dst[:b.RLESize]
 | |
| 		v := b.data[0]
 | |
| 		for i := range b.dst {
 | |
| 			b.dst[i] = v
 | |
| 		}
 | |
| 		hist.appendKeep(b.dst)
 | |
| 		return nil
 | |
| 	case blockTypeRaw:
 | |
| 		hist.appendKeep(b.data)
 | |
| 		return nil
 | |
| 	case blockTypeCompressed:
 | |
| 		saved := b.dst
 | |
| 		// Append directly to history
 | |
| 		if hist.ignoreBuffer == 0 {
 | |
| 			b.dst = hist.b
 | |
| 			hist.b = nil
 | |
| 		} else {
 | |
| 			b.dst = b.dst[:0]
 | |
| 		}
 | |
| 		err := b.decodeCompressed(hist)
 | |
| 		if debugDecoder {
 | |
| 			println("Decompressed to total", len(b.dst), "bytes, hash:", xxhash.Sum64(b.dst), "error:", err)
 | |
| 		}
 | |
| 		if hist.ignoreBuffer == 0 {
 | |
| 			hist.b = b.dst
 | |
| 			b.dst = saved
 | |
| 		} else {
 | |
| 			hist.appendKeep(b.dst)
 | |
| 		}
 | |
| 		return err
 | |
| 	case blockTypeReserved:
 | |
| 		// Used for returning errors.
 | |
| 		return b.err
 | |
| 	default:
 | |
| 		panic("Invalid block type")
 | |
| 	}
 | |
| }
 | |
| 
 | |
| func (b *blockDec) decodeLiterals(in []byte, hist *history) (remain []byte, err error) {
 | |
| 	// There must be at least one byte for Literals_Block_Type and one for Sequences_Section_Header
 | |
| 	if len(in) < 2 {
 | |
| 		return in, ErrBlockTooSmall
 | |
| 	}
 | |
| 
 | |
| 	litType := literalsBlockType(in[0] & 3)
 | |
| 	var litRegenSize int
 | |
| 	var litCompSize int
 | |
| 	sizeFormat := (in[0] >> 2) & 3
 | |
| 	var fourStreams bool
 | |
| 	var literals []byte
 | |
| 	switch litType {
 | |
| 	case literalsBlockRaw, literalsBlockRLE:
 | |
| 		switch sizeFormat {
 | |
| 		case 0, 2:
 | |
| 			// Regenerated_Size uses 5 bits (0-31). Literals_Section_Header uses 1 byte.
 | |
| 			litRegenSize = int(in[0] >> 3)
 | |
| 			in = in[1:]
 | |
| 		case 1:
 | |
| 			// Regenerated_Size uses 12 bits (0-4095). Literals_Section_Header uses 2 bytes.
 | |
| 			litRegenSize = int(in[0]>>4) + (int(in[1]) << 4)
 | |
| 			in = in[2:]
 | |
| 		case 3:
 | |
| 			//  Regenerated_Size uses 20 bits (0-1048575). Literals_Section_Header uses 3 bytes.
 | |
| 			if len(in) < 3 {
 | |
| 				println("too small: litType:", litType, " sizeFormat", sizeFormat, len(in))
 | |
| 				return in, ErrBlockTooSmall
 | |
| 			}
 | |
| 			litRegenSize = int(in[0]>>4) + (int(in[1]) << 4) + (int(in[2]) << 12)
 | |
| 			in = in[3:]
 | |
| 		}
 | |
| 	case literalsBlockCompressed, literalsBlockTreeless:
 | |
| 		switch sizeFormat {
 | |
| 		case 0, 1:
 | |
| 			// Both Regenerated_Size and Compressed_Size use 10 bits (0-1023).
 | |
| 			if len(in) < 3 {
 | |
| 				println("too small: litType:", litType, " sizeFormat", sizeFormat, len(in))
 | |
| 				return in, ErrBlockTooSmall
 | |
| 			}
 | |
| 			n := uint64(in[0]>>4) + (uint64(in[1]) << 4) + (uint64(in[2]) << 12)
 | |
| 			litRegenSize = int(n & 1023)
 | |
| 			litCompSize = int(n >> 10)
 | |
| 			fourStreams = sizeFormat == 1
 | |
| 			in = in[3:]
 | |
| 		case 2:
 | |
| 			fourStreams = true
 | |
| 			if len(in) < 4 {
 | |
| 				println("too small: litType:", litType, " sizeFormat", sizeFormat, len(in))
 | |
| 				return in, ErrBlockTooSmall
 | |
| 			}
 | |
| 			n := uint64(in[0]>>4) + (uint64(in[1]) << 4) + (uint64(in[2]) << 12) + (uint64(in[3]) << 20)
 | |
| 			litRegenSize = int(n & 16383)
 | |
| 			litCompSize = int(n >> 14)
 | |
| 			in = in[4:]
 | |
| 		case 3:
 | |
| 			fourStreams = true
 | |
| 			if len(in) < 5 {
 | |
| 				println("too small: litType:", litType, " sizeFormat", sizeFormat, len(in))
 | |
| 				return in, ErrBlockTooSmall
 | |
| 			}
 | |
| 			n := uint64(in[0]>>4) + (uint64(in[1]) << 4) + (uint64(in[2]) << 12) + (uint64(in[3]) << 20) + (uint64(in[4]) << 28)
 | |
| 			litRegenSize = int(n & 262143)
 | |
| 			litCompSize = int(n >> 18)
 | |
| 			in = in[5:]
 | |
| 		}
 | |
| 	}
 | |
| 	if debugDecoder {
 | |
| 		println("literals type:", litType, "litRegenSize:", litRegenSize, "litCompSize:", litCompSize, "sizeFormat:", sizeFormat, "4X:", fourStreams)
 | |
| 	}
 | |
| 	if litRegenSize > int(b.WindowSize) || litRegenSize > maxCompressedBlockSize {
 | |
| 		return in, ErrWindowSizeExceeded
 | |
| 	}
 | |
| 
 | |
| 	switch litType {
 | |
| 	case literalsBlockRaw:
 | |
| 		if len(in) < litRegenSize {
 | |
| 			println("too small: litType:", litType, " sizeFormat", sizeFormat, "remain:", len(in), "want:", litRegenSize)
 | |
| 			return in, ErrBlockTooSmall
 | |
| 		}
 | |
| 		literals = in[:litRegenSize]
 | |
| 		in = in[litRegenSize:]
 | |
| 		//printf("Found %d uncompressed literals\n", litRegenSize)
 | |
| 	case literalsBlockRLE:
 | |
| 		if len(in) < 1 {
 | |
| 			println("too small: litType:", litType, " sizeFormat", sizeFormat, "remain:", len(in), "want:", 1)
 | |
| 			return in, ErrBlockTooSmall
 | |
| 		}
 | |
| 		if cap(b.literalBuf) < litRegenSize {
 | |
| 			if b.lowMem {
 | |
| 				b.literalBuf = make([]byte, litRegenSize, litRegenSize+compressedBlockOverAlloc)
 | |
| 			} else {
 | |
| 				b.literalBuf = make([]byte, litRegenSize, maxCompressedBlockSize+compressedBlockOverAlloc)
 | |
| 			}
 | |
| 		}
 | |
| 		literals = b.literalBuf[:litRegenSize]
 | |
| 		v := in[0]
 | |
| 		for i := range literals {
 | |
| 			literals[i] = v
 | |
| 		}
 | |
| 		in = in[1:]
 | |
| 		if debugDecoder {
 | |
| 			printf("Found %d RLE compressed literals\n", litRegenSize)
 | |
| 		}
 | |
| 	case literalsBlockTreeless:
 | |
| 		if len(in) < litCompSize {
 | |
| 			println("too small: litType:", litType, " sizeFormat", sizeFormat, "remain:", len(in), "want:", litCompSize)
 | |
| 			return in, ErrBlockTooSmall
 | |
| 		}
 | |
| 		// Store compressed literals, so we defer decoding until we get history.
 | |
| 		literals = in[:litCompSize]
 | |
| 		in = in[litCompSize:]
 | |
| 		if debugDecoder {
 | |
| 			printf("Found %d compressed literals\n", litCompSize)
 | |
| 		}
 | |
| 		huff := hist.huffTree
 | |
| 		if huff == nil {
 | |
| 			return in, errors.New("literal block was treeless, but no history was defined")
 | |
| 		}
 | |
| 		// Ensure we have space to store it.
 | |
| 		if cap(b.literalBuf) < litRegenSize {
 | |
| 			if b.lowMem {
 | |
| 				b.literalBuf = make([]byte, 0, litRegenSize+compressedBlockOverAlloc)
 | |
| 			} else {
 | |
| 				b.literalBuf = make([]byte, 0, maxCompressedBlockSize+compressedBlockOverAlloc)
 | |
| 			}
 | |
| 		}
 | |
| 		var err error
 | |
| 		// Use our out buffer.
 | |
| 		huff.MaxDecodedSize = litRegenSize
 | |
| 		if fourStreams {
 | |
| 			literals, err = huff.Decoder().Decompress4X(b.literalBuf[:0:litRegenSize], literals)
 | |
| 		} else {
 | |
| 			literals, err = huff.Decoder().Decompress1X(b.literalBuf[:0:litRegenSize], literals)
 | |
| 		}
 | |
| 		// Make sure we don't leak our literals buffer
 | |
| 		if err != nil {
 | |
| 			println("decompressing literals:", err)
 | |
| 			return in, err
 | |
| 		}
 | |
| 		if len(literals) != litRegenSize {
 | |
| 			return in, fmt.Errorf("literal output size mismatch want %d, got %d", litRegenSize, len(literals))
 | |
| 		}
 | |
| 
 | |
| 	case literalsBlockCompressed:
 | |
| 		if len(in) < litCompSize {
 | |
| 			println("too small: litType:", litType, " sizeFormat", sizeFormat, "remain:", len(in), "want:", litCompSize)
 | |
| 			return in, ErrBlockTooSmall
 | |
| 		}
 | |
| 		literals = in[:litCompSize]
 | |
| 		in = in[litCompSize:]
 | |
| 		// Ensure we have space to store it.
 | |
| 		if cap(b.literalBuf) < litRegenSize {
 | |
| 			if b.lowMem {
 | |
| 				b.literalBuf = make([]byte, 0, litRegenSize+compressedBlockOverAlloc)
 | |
| 			} else {
 | |
| 				b.literalBuf = make([]byte, 0, maxCompressedBlockSize+compressedBlockOverAlloc)
 | |
| 			}
 | |
| 		}
 | |
| 		huff := hist.huffTree
 | |
| 		if huff == nil || (hist.dict != nil && huff == hist.dict.litEnc) {
 | |
| 			huff = huffDecoderPool.Get().(*huff0.Scratch)
 | |
| 			if huff == nil {
 | |
| 				huff = &huff0.Scratch{}
 | |
| 			}
 | |
| 		}
 | |
| 		var err error
 | |
| 		if debugDecoder {
 | |
| 			println("huff table input:", len(literals), "CRC:", crc32.ChecksumIEEE(literals))
 | |
| 		}
 | |
| 		huff, literals, err = huff0.ReadTable(literals, huff)
 | |
| 		if err != nil {
 | |
| 			println("reading huffman table:", err)
 | |
| 			return in, err
 | |
| 		}
 | |
| 		hist.huffTree = huff
 | |
| 		huff.MaxDecodedSize = litRegenSize
 | |
| 		// Use our out buffer.
 | |
| 		if fourStreams {
 | |
| 			literals, err = huff.Decoder().Decompress4X(b.literalBuf[:0:litRegenSize], literals)
 | |
| 		} else {
 | |
| 			literals, err = huff.Decoder().Decompress1X(b.literalBuf[:0:litRegenSize], literals)
 | |
| 		}
 | |
| 		if err != nil {
 | |
| 			println("decoding compressed literals:", err)
 | |
| 			return in, err
 | |
| 		}
 | |
| 		// Make sure we don't leak our literals buffer
 | |
| 		if len(literals) != litRegenSize {
 | |
| 			return in, fmt.Errorf("literal output size mismatch want %d, got %d", litRegenSize, len(literals))
 | |
| 		}
 | |
| 		// Re-cap to get extra size.
 | |
| 		literals = b.literalBuf[:len(literals)]
 | |
| 		if debugDecoder {
 | |
| 			printf("Decompressed %d literals into %d bytes\n", litCompSize, litRegenSize)
 | |
| 		}
 | |
| 	}
 | |
| 	hist.decoders.literals = literals
 | |
| 	return in, nil
 | |
| }
 | |
| 
 | |
| // decodeCompressed will start decompressing a block.
 | |
| func (b *blockDec) decodeCompressed(hist *history) error {
 | |
| 	in := b.data
 | |
| 	in, err := b.decodeLiterals(in, hist)
 | |
| 	if err != nil {
 | |
| 		return err
 | |
| 	}
 | |
| 	err = b.prepareSequences(in, hist)
 | |
| 	if err != nil {
 | |
| 		return err
 | |
| 	}
 | |
| 	if hist.decoders.nSeqs == 0 {
 | |
| 		b.dst = append(b.dst, hist.decoders.literals...)
 | |
| 		return nil
 | |
| 	}
 | |
| 	before := len(hist.decoders.out)
 | |
| 	err = hist.decoders.decodeSync(hist.b[hist.ignoreBuffer:])
 | |
| 	if err != nil {
 | |
| 		return err
 | |
| 	}
 | |
| 	if hist.decoders.maxSyncLen > 0 {
 | |
| 		hist.decoders.maxSyncLen += uint64(before)
 | |
| 		hist.decoders.maxSyncLen -= uint64(len(hist.decoders.out))
 | |
| 	}
 | |
| 	b.dst = hist.decoders.out
 | |
| 	hist.recentOffsets = hist.decoders.prevOffset
 | |
| 	return nil
 | |
| }
 | |
| 
 | |
| func (b *blockDec) prepareSequences(in []byte, hist *history) (err error) {
 | |
| 	if debugDecoder {
 | |
| 		printf("prepareSequences: %d byte(s) input\n", len(in))
 | |
| 	}
 | |
| 	// Decode Sequences
 | |
| 	// https://github.com/facebook/zstd/blob/dev/doc/zstd_compression_format.md#sequences-section
 | |
| 	if len(in) < 1 {
 | |
| 		return ErrBlockTooSmall
 | |
| 	}
 | |
| 	var nSeqs int
 | |
| 	seqHeader := in[0]
 | |
| 	switch {
 | |
| 	case seqHeader < 128:
 | |
| 		nSeqs = int(seqHeader)
 | |
| 		in = in[1:]
 | |
| 	case seqHeader < 255:
 | |
| 		if len(in) < 2 {
 | |
| 			return ErrBlockTooSmall
 | |
| 		}
 | |
| 		nSeqs = int(seqHeader-128)<<8 | int(in[1])
 | |
| 		in = in[2:]
 | |
| 	case seqHeader == 255:
 | |
| 		if len(in) < 3 {
 | |
| 			return ErrBlockTooSmall
 | |
| 		}
 | |
| 		nSeqs = 0x7f00 + int(in[1]) + (int(in[2]) << 8)
 | |
| 		in = in[3:]
 | |
| 	}
 | |
| 	if nSeqs == 0 && len(in) != 0 {
 | |
| 		// When no sequences, there should not be any more data...
 | |
| 		if debugDecoder {
 | |
| 			printf("prepareSequences: 0 sequences, but %d byte(s) left on stream\n", len(in))
 | |
| 		}
 | |
| 		return ErrUnexpectedBlockSize
 | |
| 	}
 | |
| 
 | |
| 	var seqs = &hist.decoders
 | |
| 	seqs.nSeqs = nSeqs
 | |
| 	if nSeqs > 0 {
 | |
| 		if len(in) < 1 {
 | |
| 			return ErrBlockTooSmall
 | |
| 		}
 | |
| 		br := byteReader{b: in, off: 0}
 | |
| 		compMode := br.Uint8()
 | |
| 		br.advance(1)
 | |
| 		if debugDecoder {
 | |
| 			printf("Compression modes: 0b%b", compMode)
 | |
| 		}
 | |
| 		if compMode&3 != 0 {
 | |
| 			return errors.New("corrupt block: reserved bits not zero")
 | |
| 		}
 | |
| 		for i := uint(0); i < 3; i++ {
 | |
| 			mode := seqCompMode((compMode >> (6 - i*2)) & 3)
 | |
| 			if debugDecoder {
 | |
| 				println("Table", tableIndex(i), "is", mode)
 | |
| 			}
 | |
| 			var seq *sequenceDec
 | |
| 			switch tableIndex(i) {
 | |
| 			case tableLiteralLengths:
 | |
| 				seq = &seqs.litLengths
 | |
| 			case tableOffsets:
 | |
| 				seq = &seqs.offsets
 | |
| 			case tableMatchLengths:
 | |
| 				seq = &seqs.matchLengths
 | |
| 			default:
 | |
| 				panic("unknown table")
 | |
| 			}
 | |
| 			switch mode {
 | |
| 			case compModePredefined:
 | |
| 				if seq.fse != nil && !seq.fse.preDefined {
 | |
| 					fseDecoderPool.Put(seq.fse)
 | |
| 				}
 | |
| 				seq.fse = &fsePredef[i]
 | |
| 			case compModeRLE:
 | |
| 				if br.remain() < 1 {
 | |
| 					return ErrBlockTooSmall
 | |
| 				}
 | |
| 				v := br.Uint8()
 | |
| 				br.advance(1)
 | |
| 				if seq.fse == nil || seq.fse.preDefined {
 | |
| 					seq.fse = fseDecoderPool.Get().(*fseDecoder)
 | |
| 				}
 | |
| 				symb, err := decSymbolValue(v, symbolTableX[i])
 | |
| 				if err != nil {
 | |
| 					printf("RLE Transform table (%v) error: %v", tableIndex(i), err)
 | |
| 					return err
 | |
| 				}
 | |
| 				seq.fse.setRLE(symb)
 | |
| 				if debugDecoder {
 | |
| 					printf("RLE set to 0x%x, code: %v", symb, v)
 | |
| 				}
 | |
| 			case compModeFSE:
 | |
| 				if debugDecoder {
 | |
| 					println("Reading table for", tableIndex(i))
 | |
| 				}
 | |
| 				if seq.fse == nil || seq.fse.preDefined {
 | |
| 					seq.fse = fseDecoderPool.Get().(*fseDecoder)
 | |
| 				}
 | |
| 				err := seq.fse.readNCount(&br, uint16(maxTableSymbol[i]))
 | |
| 				if err != nil {
 | |
| 					println("Read table error:", err)
 | |
| 					return err
 | |
| 				}
 | |
| 				err = seq.fse.transform(symbolTableX[i])
 | |
| 				if err != nil {
 | |
| 					println("Transform table error:", err)
 | |
| 					return err
 | |
| 				}
 | |
| 				if debugDecoder {
 | |
| 					println("Read table ok", "symbolLen:", seq.fse.symbolLen)
 | |
| 				}
 | |
| 			case compModeRepeat:
 | |
| 				seq.repeat = true
 | |
| 			}
 | |
| 			if br.overread() {
 | |
| 				return io.ErrUnexpectedEOF
 | |
| 			}
 | |
| 		}
 | |
| 		in = br.unread()
 | |
| 	}
 | |
| 	if debugDecoder {
 | |
| 		println("Literals:", len(seqs.literals), "hash:", xxhash.Sum64(seqs.literals), "and", seqs.nSeqs, "sequences.")
 | |
| 	}
 | |
| 
 | |
| 	if nSeqs == 0 {
 | |
| 		if len(b.sequence) > 0 {
 | |
| 			b.sequence = b.sequence[:0]
 | |
| 		}
 | |
| 		return nil
 | |
| 	}
 | |
| 	br := seqs.br
 | |
| 	if br == nil {
 | |
| 		br = &bitReader{}
 | |
| 	}
 | |
| 	if err := br.init(in); err != nil {
 | |
| 		return err
 | |
| 	}
 | |
| 
 | |
| 	if err := seqs.initialize(br, hist, b.dst); err != nil {
 | |
| 		println("initializing sequences:", err)
 | |
| 		return err
 | |
| 	}
 | |
| 
 | |
| 	return nil
 | |
| }
 | |
| 
 | |
| func (b *blockDec) decodeSequences(hist *history) error {
 | |
| 	if cap(b.sequence) < hist.decoders.nSeqs {
 | |
| 		if b.lowMem {
 | |
| 			b.sequence = make([]seqVals, 0, hist.decoders.nSeqs)
 | |
| 		} else {
 | |
| 			b.sequence = make([]seqVals, 0, 0x7F00+0xffff)
 | |
| 		}
 | |
| 	}
 | |
| 	b.sequence = b.sequence[:hist.decoders.nSeqs]
 | |
| 	if hist.decoders.nSeqs == 0 {
 | |
| 		hist.decoders.seqSize = len(hist.decoders.literals)
 | |
| 		return nil
 | |
| 	}
 | |
| 	hist.decoders.windowSize = hist.windowSize
 | |
| 	hist.decoders.prevOffset = hist.recentOffsets
 | |
| 
 | |
| 	err := hist.decoders.decode(b.sequence)
 | |
| 	hist.recentOffsets = hist.decoders.prevOffset
 | |
| 	return err
 | |
| }
 | |
| 
 | |
| func (b *blockDec) executeSequences(hist *history) error {
 | |
| 	hbytes := hist.b
 | |
| 	if len(hbytes) > hist.windowSize {
 | |
| 		hbytes = hbytes[len(hbytes)-hist.windowSize:]
 | |
| 		// We do not need history anymore.
 | |
| 		if hist.dict != nil {
 | |
| 			hist.dict.content = nil
 | |
| 		}
 | |
| 	}
 | |
| 	hist.decoders.windowSize = hist.windowSize
 | |
| 	hist.decoders.out = b.dst[:0]
 | |
| 	err := hist.decoders.execute(b.sequence, hbytes)
 | |
| 	if err != nil {
 | |
| 		return err
 | |
| 	}
 | |
| 	return b.updateHistory(hist)
 | |
| }
 | |
| 
 | |
| func (b *blockDec) updateHistory(hist *history) error {
 | |
| 	if len(b.data) > maxCompressedBlockSize {
 | |
| 		return fmt.Errorf("compressed block size too large (%d)", len(b.data))
 | |
| 	}
 | |
| 	// Set output and release references.
 | |
| 	b.dst = hist.decoders.out
 | |
| 	hist.recentOffsets = hist.decoders.prevOffset
 | |
| 
 | |
| 	if b.Last {
 | |
| 		// if last block we don't care about history.
 | |
| 		println("Last block, no history returned")
 | |
| 		hist.b = hist.b[:0]
 | |
| 		return nil
 | |
| 	} else {
 | |
| 		hist.append(b.dst)
 | |
| 		if debugDecoder {
 | |
| 			println("Finished block with ", len(b.sequence), "sequences. Added", len(b.dst), "to history, now length", len(hist.b))
 | |
| 		}
 | |
| 	}
 | |
| 	hist.decoders.out, hist.decoders.literals = nil, nil
 | |
| 
 | |
| 	return nil
 | |
| }
 |