mirror of
https://github.com/zr-hebo/sniffer-agent.git
synced 2025-08-10 04:13:40 +08:00
add vendor packages避免用户网络太慢无法下载编译
This commit is contained in:
105
vendor/github.com/pierrec/lz4/lz4.go
generated
vendored
Normal file
105
vendor/github.com/pierrec/lz4/lz4.go
generated
vendored
Normal file
@@ -0,0 +1,105 @@
|
||||
// Package lz4 implements reading and writing lz4 compressed data (a frame),
|
||||
// as specified in http://fastcompression.blogspot.fr/2013/04/lz4-streaming-format-final.html,
|
||||
// using an io.Reader (decompression) and io.Writer (compression).
|
||||
// It is designed to minimize memory usage while maximizing throughput by being able to
|
||||
// [de]compress data concurrently.
|
||||
//
|
||||
// The Reader and the Writer support concurrent processing provided the supplied buffers are
|
||||
// large enough (in multiples of BlockMaxSize) and there is no block dependency.
|
||||
// Reader.WriteTo and Writer.ReadFrom do leverage the concurrency transparently.
|
||||
// The runtime.GOMAXPROCS() value is used to apply concurrency or not.
|
||||
//
|
||||
// Although the block level compression and decompression functions are exposed and are fully compatible
|
||||
// with the lz4 block format definition, they are low level and should not be used directly.
|
||||
// For a complete description of an lz4 compressed block, see:
|
||||
// http://fastcompression.blogspot.fr/2011/05/lz4-explained.html
|
||||
//
|
||||
// See https://github.com/Cyan4973/lz4 for the reference C implementation.
|
||||
package lz4
|
||||
|
||||
import (
|
||||
"hash"
|
||||
"sync"
|
||||
|
||||
"github.com/pierrec/xxHash/xxHash32"
|
||||
)
|
||||
|
||||
const (
|
||||
// Extension is the LZ4 frame file name extension
|
||||
Extension = ".lz4"
|
||||
// Version is the LZ4 frame format version
|
||||
Version = 1
|
||||
|
||||
frameMagic = uint32(0x184D2204)
|
||||
frameSkipMagic = uint32(0x184D2A50)
|
||||
|
||||
// The following constants are used to setup the compression algorithm.
|
||||
minMatch = 4 // the minimum size of the match sequence size (4 bytes)
|
||||
winSizeLog = 16 // LZ4 64Kb window size limit
|
||||
winSize = 1 << winSizeLog
|
||||
winMask = winSize - 1 // 64Kb window of previous data for dependent blocks
|
||||
|
||||
// hashLog determines the size of the hash table used to quickly find a previous match position.
|
||||
// Its value influences the compression speed and memory usage, the lower the faster,
|
||||
// but at the expense of the compression ratio.
|
||||
// 16 seems to be the best compromise.
|
||||
hashLog = 16
|
||||
hashTableSize = 1 << hashLog
|
||||
hashShift = uint((minMatch * 8) - hashLog)
|
||||
|
||||
mfLimit = 8 + minMatch // The last match cannot start within the last 12 bytes.
|
||||
skipStrength = 6 // variable step for fast scan
|
||||
|
||||
hasher = uint32(2654435761) // prime number used to hash minMatch
|
||||
)
|
||||
|
||||
// map the block max size id with its value in bytes: 64Kb, 256Kb, 1Mb and 4Mb.
|
||||
var bsMapID = map[byte]int{4: 64 << 10, 5: 256 << 10, 6: 1 << 20, 7: 4 << 20}
|
||||
var bsMapValue = map[int]byte{}
|
||||
|
||||
// Reversed.
|
||||
func init() {
|
||||
for i, v := range bsMapID {
|
||||
bsMapValue[v] = i
|
||||
}
|
||||
}
|
||||
|
||||
// Header describes the various flags that can be set on a Writer or obtained from a Reader.
|
||||
// The default values match those of the LZ4 frame format definition (http://fastcompression.blogspot.com/2013/04/lz4-streaming-format-final.html).
|
||||
//
|
||||
// NB. in a Reader, in case of concatenated frames, the Header values may change between Read() calls.
|
||||
// It is the caller responsibility to check them if necessary (typically when using the Reader concurrency).
|
||||
type Header struct {
|
||||
BlockDependency bool // compressed blocks are dependent (one block depends on the last 64Kb of the previous one)
|
||||
BlockChecksum bool // compressed blocks are checksumed
|
||||
NoChecksum bool // frame checksum
|
||||
BlockMaxSize int // the size of the decompressed data block (one of [64KB, 256KB, 1MB, 4MB]). Default=4MB.
|
||||
Size uint64 // the frame total size. It is _not_ computed by the Writer.
|
||||
HighCompression bool // use high compression (only for the Writer)
|
||||
done bool // whether the descriptor was processed (Read or Write and checked)
|
||||
// Removed as not supported
|
||||
// Dict bool // a dictionary id is to be used
|
||||
// DictID uint32 // the dictionary id read from the frame, if any.
|
||||
}
|
||||
|
||||
// xxhPool wraps the standard pool for xxHash items.
|
||||
// Putting items back in the pool automatically resets them.
|
||||
type xxhPool struct {
|
||||
sync.Pool
|
||||
}
|
||||
|
||||
func (p *xxhPool) Get() hash.Hash32 {
|
||||
return p.Pool.Get().(hash.Hash32)
|
||||
}
|
||||
|
||||
func (p *xxhPool) Put(h hash.Hash32) {
|
||||
h.Reset()
|
||||
p.Pool.Put(h)
|
||||
}
|
||||
|
||||
// hashPool is used by readers and writers and contains xxHash items.
|
||||
var hashPool = xxhPool{
|
||||
Pool: sync.Pool{
|
||||
New: func() interface{} { return xxHash32.New(0) },
|
||||
},
|
||||
}
|
Reference in New Issue
Block a user