Documentation ¶
Overview ¶
Package lz4 implements reading and writing lz4 compressed data (a frame), as specified in http://fastcompression.blogspot.fr/2013/04/lz4-streaming-format-final.html.
Although the block level compression and decompression functions are exposed and are fully compatible with the lz4 block format definition, they are low level and should not be used directly. For a complete description of an lz4 compressed block, see: http://fastcompression.blogspot.fr/2011/05/lz4-explained.html
See https://github.com/Cyan4973/lz4 for the reference C implementation.
Example ¶
package main import ( "io" "os" "strings" "github.com/pierrec/lz4" ) func main() { // Compress and uncompress an input string. s := "hello world" r := strings.NewReader(s) // The pipe will uncompress the data from the writer. pr, pw := io.Pipe() zw := lz4.NewWriter(pw) zr := lz4.NewReader(pr) go func() { // Compress the input string. _, _ = io.Copy(zw, r) _ = zw.Close() // Make sure the writer is closed _ = pw.Close() // Terminate the pipe }() _, _ = io.Copy(os.Stdout, zr) }
Output: hello world
Index ¶
Examples ¶
Constants ¶
const ( // Extension is the LZ4 frame file name extension Extension = ".lz4" // Version is the LZ4 frame format version Version = 1 )
Variables ¶
var ( // ErrInvalidSourceShortBuffer is returned by UncompressBlock or CompressBLock when a compressed // block is corrupted or the destination buffer is not large enough for the uncompressed data. ErrInvalidSourceShortBuffer = errors.New("lz4: invalid source or destination buffer too short") // ErrInvalid is returned when reading an invalid LZ4 archive. ErrInvalid = errors.New("lz4: bad magic number") // ErrBlockDependency is returned when attempting to decompress an archive created with block dependency. ErrBlockDependency = errors.New("lz4: block dependency not supported") // ErrUnsupportedSeek is returned when attempting to Seek any way but forward from the current position. ErrUnsupportedSeek = errors.New("lz4: can only seek forward from io.SeekCurrent") )
Functions ¶
func CompressBlock ¶
CompressBlock compresses the source buffer into the destination one. This is the fast version of LZ4 compression and also the default one.
The argument hashTable is scratch space for a hash table used by the compressor. If provided, it should have length at least 1<<16. If it is shorter (or nil), CompressBlock allocates its own hash table.
The size of the compressed data is returned. If it is 0 and no error, then the data is incompressible.
An error is returned if the destination buffer is too small.
Example ¶
package main import ( "fmt" "strings" "github.com/pierrec/lz4" ) func main() { s := "hello world" data := []byte(strings.Repeat(s, 100)) buf := make([]byte, len(data)) ht := make([]int, 64<<10) // buffer for the compression table n, err := lz4.CompressBlock(data, buf, ht) if err != nil { fmt.Println(err) } if n >= len(data) { fmt.Printf("`%s` is not compressible", s) } buf = buf[:n] // compressed data // Allocated a very large buffer for decompression. out := make([]byte, 10*len(data)) n, err = lz4.UncompressBlock(buf, out) if err != nil { fmt.Println(err) } out = out[:n] // uncompressed data fmt.Println(string(out[:len(s)])) }
Output: hello world
func CompressBlockBound ¶
CompressBlockBound returns the maximum size of a given buffer of size n, when not compressible.
func CompressBlockHC ¶
CompressBlockHC compresses the source buffer src into the destination dst with max search depth (use 0 or negative value for no max).
CompressBlockHC compression ratio is better than CompressBlock but it is also slower.
The size of the compressed data is returned. If it is 0 and no error, then the data is not compressible.
An error is returned if the destination buffer is too small.
func UncompressBlock ¶
UncompressBlock uncompresses the source buffer into the destination one, and returns the uncompressed size.
The destination buffer must be sized appropriately.
An error is returned if the source data is invalid or the destination buffer is too small.
Types ¶
type Header ¶
type Header struct { BlockChecksum bool // Compressed blocks checksum flag. NoChecksum bool // Frame checksum flag. BlockMaxSize int // Size of the uncompressed data block (one of [64KB, 256KB, 1MB, 4MB]). Default=4MB. Size uint64 // Frame total size. It is _not_ computed by the Writer. CompressionLevel int // Compression level (higher is better, use 0 for fastest compression). // contains filtered or unexported fields }
Header describes the various flags that can be set on a Writer or obtained from a Reader. The default values match those of the LZ4 frame format definition (http://fastcompression.blogspot.com/2013/04/lz4-streaming-format-final.html).
NB. in a Reader, in case of concatenated frames, the Header values may change between Read() calls. It is the caller's responsibility to check them if necessary.
type Reader ¶
type Reader struct { Header // Handler called when a block has been successfully read. // It provides the number of bytes read. OnBlockDone func(size int) // contains filtered or unexported fields }
Reader implements the LZ4 frame decoder. The Header is set after the first call to Read(). The Header may change between Read() calls in case of concatenated frames.
func NewReader ¶
NewReader returns a new LZ4 frame decoder. No access to the underlying io.Reader is performed.
func (*Reader) Read ¶
Read decompresses data from the underlying source into the supplied buffer.
Since there can be multiple streams concatenated, Header values may change between calls to Read(). If that is the case, no data is actually read from the underlying io.Reader, to allow for potential input buffer resizing.
func (*Reader) Reset ¶
Reset discards the Reader's state and makes it equivalent to the result of its original state from NewReader, but reading from r instead. This permits reusing a Reader rather than allocating a new one.
func (*Reader) Seek ¶
Seek implements io.Seeker, but supports seeking forward from the current position only. Any other seek will return an error. Allows skipping output bytes which aren't needed, which in some scenarios is faster than reading and discarding them. Note this may cause future calls to Read() to read 0 bytes if all of the data they would have returned is skipped.
type Writer ¶
type Writer struct { Header // Handler called when a block has been successfully written out. // It provides the number of bytes written. OnBlockDone func(size int) // contains filtered or unexported fields }
Writer implements the LZ4 frame encoder.
func NewWriter ¶
NewWriter returns a new LZ4 frame encoder. No access to the underlying io.Writer is performed. The supplied Header is checked at the first Write. It is ok to change it before the first Write but then not until a Reset() is performed.
func (*Writer) Close ¶
Close closes the Writer, flushing any unwritten data to the underlying io.Writer, but does not close the underlying io.Writer.
func (*Writer) Flush ¶
Flush flushes any pending compressed data to the underlying writer. Flush does not return until the data has been written. If the underlying writer returns an error, Flush returns that error.
func (*Writer) Reset ¶
Reset clears the state of the Writer z such that it is equivalent to its initial state from NewWriter, but instead writing to w. No access to the underlying io.Writer is performed.
func (*Writer) WithConcurrency ¶
WithConcurrency sets the number of concurrent go routines used for compression. A negative value sets the concurrency to GOMAXPROCS.