forked from Shiloh/githaven
520 lines
14 KiB
Go
520 lines
14 KiB
Go
package unsnap
|
|
|
|
import (
|
|
"bytes"
|
|
"encoding/binary"
|
|
"fmt"
|
|
"io"
|
|
"io/ioutil"
|
|
"os"
|
|
"strings"
|
|
|
|
"hash/crc32"
|
|
|
|
snappy "github.com/golang/snappy"
|
|
// The C library can be used, but this makes the binary dependent
|
|
// lots of extraneous c-libraries; it is no longer stand-alone. Yuck.
|
|
//
|
|
// Therefore we comment out the "dgryski/go-csnappy" path and use the
|
|
// "github.com/golang/snappy/snappy" above instead. If you are
|
|
// performance limited and can deal with distributing more libraries,
|
|
// then this is easy to swap.
|
|
//
|
|
// If you swap, note that some of the tests won't pass
|
|
// because snappy-go produces slightly different (but still
|
|
// conformant) encodings on some data. Here are bindings
|
|
// to the C-snappy:
|
|
// snappy "github.com/dgryski/go-csnappy"
|
|
)
|
|
|
|
// SnappyFile: create a drop-in-replacement/wrapper for an *os.File that handles doing the unsnappification online as more is read from it
|
|
|
|
type SnappyFile struct {
|
|
Fname string
|
|
|
|
Reader io.Reader
|
|
Writer io.Writer
|
|
|
|
// allow clients to substitute us for an os.File and just switch
|
|
// off compression if they don't want it.
|
|
SnappyEncodeDecodeOff bool // if true, we bypass straight to Filep
|
|
|
|
EncBuf FixedSizeRingBuf // holds any extra that isn't yet returned, encoded
|
|
DecBuf FixedSizeRingBuf // holds any extra that isn't yet returned, decoded
|
|
|
|
// for writing to stream-framed snappy
|
|
HeaderChunkWritten bool
|
|
|
|
// Sanity check: we can only read, or only write, to one SnappyFile.
|
|
// EncBuf and DecBuf are used differently in each mode. Verify
|
|
// that we are consistent with this flag.
|
|
Writing bool
|
|
}
|
|
|
|
var total int
|
|
|
|
// for debugging, show state of buffers
|
|
func (f *SnappyFile) Dump() {
|
|
fmt.Printf("EncBuf has length %d and contents:\n%s\n", len(f.EncBuf.Bytes()), string(f.EncBuf.Bytes()))
|
|
fmt.Printf("DecBuf has length %d and contents:\n%s\n", len(f.DecBuf.Bytes()), string(f.DecBuf.Bytes()))
|
|
}
|
|
|
|
func (f *SnappyFile) Read(p []byte) (n int, err error) {
|
|
|
|
if f.SnappyEncodeDecodeOff {
|
|
return f.Reader.Read(p)
|
|
}
|
|
|
|
if f.Writing {
|
|
panic("Reading on a write-only SnappyFile")
|
|
}
|
|
|
|
// before we unencrypt more, try to drain the DecBuf first
|
|
n, _ = f.DecBuf.Read(p)
|
|
if n > 0 {
|
|
total += n
|
|
return n, nil
|
|
}
|
|
|
|
//nEncRead, nDecAdded, err := UnsnapOneFrame(f.Filep, &f.EncBuf, &f.DecBuf, f.Fname)
|
|
_, _, err = UnsnapOneFrame(f.Reader, &f.EncBuf, &f.DecBuf, f.Fname)
|
|
if err != nil && err != io.EOF {
|
|
panic(err)
|
|
}
|
|
|
|
n, _ = f.DecBuf.Read(p)
|
|
|
|
if n > 0 {
|
|
total += n
|
|
return n, nil
|
|
}
|
|
if f.DecBuf.Readable == 0 {
|
|
if f.DecBuf.Readable == 0 && f.EncBuf.Readable == 0 {
|
|
// only now (when EncBuf is empty) can we give io.EOF.
|
|
// Any earlier, and we leave stuff un-decoded!
|
|
return 0, io.EOF
|
|
}
|
|
}
|
|
return 0, nil
|
|
}
|
|
|
|
func Open(name string) (file *SnappyFile, err error) {
|
|
fp, err := os.Open(name)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
// encoding in snappy can apparently go beyond the original size, so
|
|
// we make our buffers big enough, 2*max snappy chunk => 2 * CHUNK_MAX(65536)
|
|
|
|
snap := NewReader(fp)
|
|
snap.Fname = name
|
|
return snap, nil
|
|
}
|
|
|
|
func NewReader(r io.Reader) *SnappyFile {
|
|
return &SnappyFile{
|
|
Reader: r,
|
|
EncBuf: *NewFixedSizeRingBuf(CHUNK_MAX * 2), // buffer of snappy encoded bytes
|
|
DecBuf: *NewFixedSizeRingBuf(CHUNK_MAX * 2), // buffer of snapppy decoded bytes
|
|
Writing: false,
|
|
}
|
|
}
|
|
|
|
func NewWriter(w io.Writer) *SnappyFile {
|
|
return &SnappyFile{
|
|
Writer: w,
|
|
EncBuf: *NewFixedSizeRingBuf(65536), // on writing: temp for testing compression
|
|
DecBuf: *NewFixedSizeRingBuf(65536 * 2), // on writing: final buffer of snappy framed and encoded bytes
|
|
Writing: true,
|
|
}
|
|
}
|
|
|
|
func Create(name string) (file *SnappyFile, err error) {
|
|
fp, err := os.Create(name)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
snap := NewWriter(fp)
|
|
snap.Fname = name
|
|
return snap, nil
|
|
}
|
|
|
|
func (f *SnappyFile) Close() error {
|
|
if f.Writing {
|
|
wc, ok := f.Writer.(io.WriteCloser)
|
|
if ok {
|
|
return wc.Close()
|
|
}
|
|
return nil
|
|
}
|
|
rc, ok := f.Reader.(io.ReadCloser)
|
|
if ok {
|
|
return rc.Close()
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (f *SnappyFile) Sync() error {
|
|
file, ok := f.Writer.(*os.File)
|
|
if ok {
|
|
return file.Sync()
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// for an increment of a frame at a time:
|
|
// read from r into encBuf (encBuf is still encoded, thus the name), and write unsnappified frames into outDecodedBuf
|
|
// the returned n: number of bytes read from the encrypted encBuf
|
|
func UnsnapOneFrame(r io.Reader, encBuf *FixedSizeRingBuf, outDecodedBuf *FixedSizeRingBuf, fname string) (nEnc int64, nDec int64, err error) {
|
|
// b, err := ioutil.ReadAll(r)
|
|
// if err != nil {
|
|
// panic(err)
|
|
// }
|
|
|
|
nEnc = 0
|
|
nDec = 0
|
|
|
|
// read up to 65536 bytes from r into encBuf, at least a snappy frame
|
|
nread, err := io.CopyN(encBuf, r, 65536) // returns nwrotebytes, err
|
|
nEnc += nread
|
|
if err != nil {
|
|
if err == io.EOF {
|
|
if nread == 0 {
|
|
if encBuf.Readable == 0 {
|
|
return nEnc, nDec, io.EOF
|
|
}
|
|
// else we have bytes in encBuf, so decode them!
|
|
err = nil
|
|
} else {
|
|
// continue below, processing the nread bytes
|
|
err = nil
|
|
}
|
|
} else {
|
|
// may be an odd already closed... don't panic on that
|
|
if strings.Contains(err.Error(), "file already closed") {
|
|
err = nil
|
|
} else {
|
|
panic(err)
|
|
}
|
|
}
|
|
}
|
|
|
|
// flag for printing chunk size alignment messages
|
|
verbose := false
|
|
|
|
const snappyStreamHeaderSz = 10
|
|
const headerSz = 4
|
|
const crc32Sz = 4
|
|
// the magic 18 bytes accounts for the snappy streaming header and the first chunks size and checksum
|
|
// http://code.google.com/p/snappy/source/browse/trunk/framing_format.txt
|
|
|
|
chunk := (*encBuf).Bytes()
|
|
|
|
// however we exit, advance as
|
|
// defer func() { (*encBuf).Next(N) }()
|
|
|
|
// 65536 is the max size of a snappy framed chunk. See
|
|
// http://code.google.com/p/snappy/source/browse/trunk/framing_format.txt:91
|
|
// buf := make([]byte, 65536)
|
|
|
|
// fmt.Printf("read from file, b is len:%d with value: %#v\n", len(b), b)
|
|
// fmt.Printf("read from file, bcut is len:%d with value: %#v\n", len(bcut), bcut)
|
|
|
|
//fmt.Printf("raw bytes of chunksz are: %v\n", b[11:14])
|
|
|
|
fourbytes := make([]byte, 4)
|
|
chunkCount := 0
|
|
|
|
for nDec < 65536 {
|
|
if len(chunk) == 0 {
|
|
break
|
|
}
|
|
chunkCount++
|
|
fourbytes[3] = 0
|
|
copy(fourbytes, chunk[1:4])
|
|
chunksz := binary.LittleEndian.Uint32(fourbytes)
|
|
chunk_type := chunk[0]
|
|
|
|
switch true {
|
|
case chunk_type == 0xff:
|
|
{ // stream identifier
|
|
|
|
streamHeader := chunk[:snappyStreamHeaderSz]
|
|
if 0 != bytes.Compare(streamHeader, []byte{0xff, 0x06, 0x00, 0x00, 0x73, 0x4e, 0x61, 0x50, 0x70, 0x59}) {
|
|
panic("file had chunk starting with 0xff but then no magic snappy streaming protocol bytes, aborting.")
|
|
} else {
|
|
//fmt.Printf("got streaming snappy magic header just fine.\n")
|
|
}
|
|
chunk = chunk[snappyStreamHeaderSz:]
|
|
(*encBuf).Advance(snappyStreamHeaderSz)
|
|
nEnc += snappyStreamHeaderSz
|
|
continue
|
|
}
|
|
case chunk_type == 0x00:
|
|
{ // compressed data
|
|
if verbose {
|
|
fmt.Fprintf(os.Stderr, "chunksz is %d while total bytes avail are: %d\n", int(chunksz), len(chunk)-4)
|
|
}
|
|
|
|
crc := binary.LittleEndian.Uint32(chunk[headerSz:(headerSz + crc32Sz)])
|
|
section := chunk[(headerSz + crc32Sz):(headerSz + chunksz)]
|
|
|
|
dec, ok := snappy.Decode(nil, section)
|
|
if ok != nil {
|
|
// we've probably truncated a snappy frame at this point
|
|
// ok=snappy: corrupt input
|
|
// len(dec) == 0
|
|
//
|
|
panic(fmt.Sprintf("could not decode snappy stream: '%s' and len dec=%d and ok=%v\n", fname, len(dec), ok))
|
|
|
|
// get back to caller with what we've got so far
|
|
return nEnc, nDec, nil
|
|
}
|
|
// fmt.Printf("ok, b is %#v , %#v\n", ok, dec)
|
|
|
|
// spit out decoded text
|
|
// n, err := w.Write(dec)
|
|
//fmt.Printf("len(dec) = %d, outDecodedBuf.Readable=%d\n", len(dec), outDecodedBuf.Readable)
|
|
bnb := bytes.NewBuffer(dec)
|
|
n, err := io.Copy(outDecodedBuf, bnb)
|
|
if err != nil {
|
|
//fmt.Printf("got n=%d, err= %s ; when trying to io.Copy(outDecodedBuf: N=%d, Readable=%d)\n", n, err, outDecodedBuf.N, outDecodedBuf.Readable)
|
|
panic(err)
|
|
}
|
|
if n != int64(len(dec)) {
|
|
panic("could not write all bytes to outDecodedBuf")
|
|
}
|
|
nDec += n
|
|
|
|
// verify the crc32 rotated checksum
|
|
m32 := masked_crc32c(dec)
|
|
if m32 != crc {
|
|
panic(fmt.Sprintf("crc32 masked failiure. expected: %v but got: %v", crc, m32))
|
|
} else {
|
|
//fmt.Printf("\nchecksums match: %v == %v\n", crc, m32)
|
|
}
|
|
|
|
// move to next header
|
|
inc := (headerSz + int(chunksz))
|
|
chunk = chunk[inc:]
|
|
(*encBuf).Advance(inc)
|
|
nEnc += int64(inc)
|
|
continue
|
|
}
|
|
case chunk_type == 0x01:
|
|
{ // uncompressed data
|
|
|
|
//n, err := w.Write(chunk[(headerSz+crc32Sz):(headerSz + int(chunksz))])
|
|
n, err := io.Copy(outDecodedBuf, bytes.NewBuffer(chunk[(headerSz+crc32Sz):(headerSz+int(chunksz))]))
|
|
if verbose {
|
|
//fmt.Printf("debug: n=%d err=%v chunksz=%d outDecodedBuf='%v'\n", n, err, chunksz, outDecodedBuf)
|
|
}
|
|
if err != nil {
|
|
panic(err)
|
|
}
|
|
if n != int64(chunksz-crc32Sz) {
|
|
panic("could not write all bytes to stdout")
|
|
}
|
|
nDec += n
|
|
|
|
inc := (headerSz + int(chunksz))
|
|
chunk = chunk[inc:]
|
|
(*encBuf).Advance(inc)
|
|
nEnc += int64(inc)
|
|
continue
|
|
}
|
|
case chunk_type == 0xfe:
|
|
fallthrough // padding, just skip it
|
|
case chunk_type >= 0x80 && chunk_type <= 0xfd:
|
|
{ // Reserved skippable chunks
|
|
//fmt.Printf("\nin reserved skippable chunks, at nEnc=%v\n", nEnc)
|
|
inc := (headerSz + int(chunksz))
|
|
chunk = chunk[inc:]
|
|
nEnc += int64(inc)
|
|
(*encBuf).Advance(inc)
|
|
continue
|
|
}
|
|
|
|
default:
|
|
panic(fmt.Sprintf("unrecognized/unsupported chunk type %#v", chunk_type))
|
|
}
|
|
|
|
} // end for{}
|
|
|
|
return nEnc, nDec, err
|
|
//return int64(N), nil
|
|
}
|
|
|
|
// for whole file at once:
|
|
//
|
|
// receive on stdin a stream of bytes in the snappy-streaming framed
|
|
// format, defined here: http://code.google.com/p/snappy/source/browse/trunk/framing_format.txt
|
|
// Grab each frame, run it through the snappy decoder, and spit out
|
|
// each frame all joined back-to-back on stdout.
|
|
//
|
|
func Unsnappy(r io.Reader, w io.Writer) (err error) {
|
|
b, err := ioutil.ReadAll(r)
|
|
if err != nil {
|
|
panic(err)
|
|
}
|
|
|
|
// flag for printing chunk size alignment messages
|
|
verbose := false
|
|
|
|
const snappyStreamHeaderSz = 10
|
|
const headerSz = 4
|
|
const crc32Sz = 4
|
|
// the magic 18 bytes accounts for the snappy streaming header and the first chunks size and checksum
|
|
// http://code.google.com/p/snappy/source/browse/trunk/framing_format.txt
|
|
|
|
chunk := b[:]
|
|
|
|
// 65536 is the max size of a snappy framed chunk. See
|
|
// http://code.google.com/p/snappy/source/browse/trunk/framing_format.txt:91
|
|
//buf := make([]byte, 65536)
|
|
|
|
// fmt.Printf("read from file, b is len:%d with value: %#v\n", len(b), b)
|
|
// fmt.Printf("read from file, bcut is len:%d with value: %#v\n", len(bcut), bcut)
|
|
|
|
//fmt.Printf("raw bytes of chunksz are: %v\n", b[11:14])
|
|
|
|
fourbytes := make([]byte, 4)
|
|
chunkCount := 0
|
|
|
|
for {
|
|
if len(chunk) == 0 {
|
|
break
|
|
}
|
|
chunkCount++
|
|
fourbytes[3] = 0
|
|
copy(fourbytes, chunk[1:4])
|
|
chunksz := binary.LittleEndian.Uint32(fourbytes)
|
|
chunk_type := chunk[0]
|
|
|
|
switch true {
|
|
case chunk_type == 0xff:
|
|
{ // stream identifier
|
|
|
|
streamHeader := chunk[:snappyStreamHeaderSz]
|
|
if 0 != bytes.Compare(streamHeader, []byte{0xff, 0x06, 0x00, 0x00, 0x73, 0x4e, 0x61, 0x50, 0x70, 0x59}) {
|
|
panic("file had chunk starting with 0xff but then no magic snappy streaming protocol bytes, aborting.")
|
|
} else {
|
|
//fmt.Printf("got streaming snappy magic header just fine.\n")
|
|
}
|
|
chunk = chunk[snappyStreamHeaderSz:]
|
|
continue
|
|
}
|
|
case chunk_type == 0x00:
|
|
{ // compressed data
|
|
if verbose {
|
|
fmt.Fprintf(os.Stderr, "chunksz is %d while total bytes avail are: %d\n", int(chunksz), len(chunk)-4)
|
|
}
|
|
|
|
//crc := binary.LittleEndian.Uint32(chunk[headerSz:(headerSz + crc32Sz)])
|
|
section := chunk[(headerSz + crc32Sz):(headerSz + chunksz)]
|
|
|
|
dec, ok := snappy.Decode(nil, section)
|
|
if ok != nil {
|
|
panic("could not decode snappy stream")
|
|
}
|
|
// fmt.Printf("ok, b is %#v , %#v\n", ok, dec)
|
|
|
|
// spit out decoded text
|
|
n, err := w.Write(dec)
|
|
if err != nil {
|
|
panic(err)
|
|
}
|
|
if n != len(dec) {
|
|
panic("could not write all bytes to stdout")
|
|
}
|
|
|
|
// TODO: verify the crc32 rotated checksum?
|
|
|
|
// move to next header
|
|
chunk = chunk[(headerSz + int(chunksz)):]
|
|
continue
|
|
}
|
|
case chunk_type == 0x01:
|
|
{ // uncompressed data
|
|
|
|
//crc := binary.LittleEndian.Uint32(chunk[headerSz:(headerSz + crc32Sz)])
|
|
section := chunk[(headerSz + crc32Sz):(headerSz + chunksz)]
|
|
|
|
n, err := w.Write(section)
|
|
if err != nil {
|
|
panic(err)
|
|
}
|
|
if n != int(chunksz-crc32Sz) {
|
|
panic("could not write all bytes to stdout")
|
|
}
|
|
|
|
chunk = chunk[(headerSz + int(chunksz)):]
|
|
continue
|
|
}
|
|
case chunk_type == 0xfe:
|
|
fallthrough // padding, just skip it
|
|
case chunk_type >= 0x80 && chunk_type <= 0xfd:
|
|
{ // Reserved skippable chunks
|
|
chunk = chunk[(headerSz + int(chunksz)):]
|
|
continue
|
|
}
|
|
|
|
default:
|
|
panic(fmt.Sprintf("unrecognized/unsupported chunk type %#v", chunk_type))
|
|
}
|
|
|
|
} // end for{}
|
|
|
|
return nil
|
|
}
|
|
|
|
// 0xff 0x06 0x00 0x00 sNaPpY
|
|
var SnappyStreamHeaderMagic = []byte{0xff, 0x06, 0x00, 0x00, 0x73, 0x4e, 0x61, 0x50, 0x70, 0x59}
|
|
|
|
const CHUNK_MAX = 65536
|
|
const _STREAM_TO_STREAM_BLOCK_SIZE = CHUNK_MAX
|
|
const _STREAM_IDENTIFIER = `sNaPpY`
|
|
const _COMPRESSED_CHUNK = 0x00
|
|
const _UNCOMPRESSED_CHUNK = 0x01
|
|
const _IDENTIFIER_CHUNK = 0xff
|
|
const _RESERVED_UNSKIPPABLE0 = 0x02 // chunk ranges are [inclusive, exclusive)
|
|
const _RESERVED_UNSKIPPABLE1 = 0x80
|
|
const _RESERVED_SKIPPABLE0 = 0x80
|
|
const _RESERVED_SKIPPABLE1 = 0xff
|
|
|
|
// the minimum percent of bytes compression must save to be enabled in automatic
|
|
// mode
|
|
const _COMPRESSION_THRESHOLD = .125
|
|
|
|
var crctab *crc32.Table
|
|
|
|
func init() {
|
|
crctab = crc32.MakeTable(crc32.Castagnoli) // this is correct table, matches the crc32c.c code used by python
|
|
}
|
|
|
|
func masked_crc32c(data []byte) uint32 {
|
|
|
|
// see the framing format specification, http://code.google.com/p/snappy/source/browse/trunk/framing_format.txt
|
|
var crc uint32 = crc32.Checksum(data, crctab)
|
|
return (uint32((crc>>15)|(crc<<17)) + 0xa282ead8)
|
|
}
|
|
|
|
func ReadSnappyStreamCompressedFile(filename string) ([]byte, error) {
|
|
|
|
snappyFile, err := Open(filename)
|
|
if err != nil {
|
|
return []byte{}, err
|
|
}
|
|
|
|
var bb bytes.Buffer
|
|
_, err = bb.ReadFrom(snappyFile)
|
|
if err == io.EOF {
|
|
err = nil
|
|
}
|
|
if err != nil {
|
|
panic(err)
|
|
}
|
|
|
|
return bb.Bytes(), err
|
|
}
|