package flate
import (
"errors"
"fmt"
"io"
"math"
)
const (
NoCompression = 0
BestSpeed = 1
BestCompression = 9
DefaultCompression = -1
HuffmanOnly = -2
)
const (
logWindowSize = 15
windowSize = 1 << logWindowSize
windowMask = windowSize - 1
baseMatchLength = 3
minMatchLength = 4
maxMatchLength = 258
baseMatchOffset = 1
maxMatchOffset = 1 << 15
maxFlateBlockTokens = 1 << 14
maxStoreBlockSize = 65535
hashBits = 17
hashSize = 1 << hashBits
hashMask = (1 << hashBits ) - 1
maxHashOffset = 1 << 24
skipNever = math .MaxInt32
)
type compressionLevel struct {
level, good, lazy, nice, chain, fastSkipHashing int
}
var levels = []compressionLevel {
{0 , 0 , 0 , 0 , 0 , 0 },
{1 , 0 , 0 , 0 , 0 , 0 },
{2 , 4 , 0 , 16 , 8 , 5 },
{3 , 4 , 0 , 32 , 32 , 6 },
{4 , 4 , 4 , 16 , 16 , skipNever },
{5 , 8 , 16 , 32 , 32 , skipNever },
{6 , 8 , 16 , 128 , 128 , skipNever },
{7 , 8 , 32 , 128 , 256 , skipNever },
{8 , 32 , 128 , 258 , 1024 , skipNever },
{9 , 32 , 258 , 258 , 4096 , skipNever },
}
type compressor struct {
compressionLevel
w *huffmanBitWriter
bulkHasher func ([]byte , []uint32 )
fill func (*compressor , []byte ) int
step func (*compressor )
bestSpeed *deflateFast
chainHead int
hashHead [hashSize ]uint32
hashPrev [windowSize ]uint32
hashOffset int
index int
window []byte
windowEnd int
blockStart int
byteAvailable bool
sync bool
tokens []token
length int
offset int
maxInsertIndex int
err error
hashMatch [maxMatchLength - 1 ]uint32
}
func (d *compressor ) fillDeflate (b []byte ) int {
if d .index >= 2 *windowSize -(minMatchLength +maxMatchLength ) {
copy (d .window , d .window [windowSize :2 *windowSize ])
d .index -= windowSize
d .windowEnd -= windowSize
if d .blockStart >= windowSize {
d .blockStart -= windowSize
} else {
d .blockStart = math .MaxInt32
}
d .hashOffset += windowSize
if d .hashOffset > maxHashOffset {
delta := d .hashOffset - 1
d .hashOffset -= delta
d .chainHead -= delta
for i , v := range d .hashPrev [:] {
if int (v ) > delta {
d .hashPrev [i ] = uint32 (int (v ) - delta )
} else {
d .hashPrev [i ] = 0
}
}
for i , v := range d .hashHead [:] {
if int (v ) > delta {
d .hashHead [i ] = uint32 (int (v ) - delta )
} else {
d .hashHead [i ] = 0
}
}
}
}
n := copy (d .window [d .windowEnd :], b )
d .windowEnd += n
return n
}
func (d *compressor ) writeBlock (tokens []token , index int ) error {
if index > 0 {
var window []byte
if d .blockStart <= index {
window = d .window [d .blockStart :index ]
}
d .blockStart = index
d .w .writeBlock (tokens , false , window )
return d .w .err
}
return nil
}
func (d *compressor ) fillWindow (b []byte ) {
if d .compressionLevel .level < 2 {
return
}
if d .index != 0 || d .windowEnd != 0 {
panic ("internal error: fillWindow called with stale data" )
}
if len (b ) > windowSize {
b = b [len (b )-windowSize :]
}
n := copy (d .window , b )
loops := (n + 256 - minMatchLength ) / 256
for j := 0 ; j < loops ; j ++ {
index := j * 256
end := index + 256 + minMatchLength - 1
if end > n {
end = n
}
toCheck := d .window [index :end ]
dstSize := len (toCheck ) - minMatchLength + 1
if dstSize <= 0 {
continue
}
dst := d .hashMatch [:dstSize ]
d .bulkHasher (toCheck , dst )
for i , val := range dst {
di := i + index
hh := &d .hashHead [val &hashMask ]
d .hashPrev [di &windowMask ] = *hh
*hh = uint32 (di + d .hashOffset )
}
}
d .windowEnd = n
d .index = n
}
func (d *compressor ) findMatch (pos int , prevHead int , prevLength int , lookahead int ) (length , offset int , ok bool ) {
minMatchLook := maxMatchLength
if lookahead < minMatchLook {
minMatchLook = lookahead
}
win := d .window [0 : pos +minMatchLook ]
nice := len (win ) - pos
if d .nice < nice {
nice = d .nice
}
tries := d .chain
length = prevLength
if length >= d .good {
tries >>= 2
}
wEnd := win [pos +length ]
wPos := win [pos :]
minIndex := pos - windowSize
for i := prevHead ; tries > 0 ; tries -- {
if wEnd == win [i +length ] {
n := matchLen (win [i :], wPos , minMatchLook )
if n > length && (n > minMatchLength || pos -i <= 4096 ) {
length = n
offset = pos - i
ok = true
if n >= nice {
break
}
wEnd = win [pos +n ]
}
}
if i == minIndex {
break
}
i = int (d .hashPrev [i &windowMask ]) - d .hashOffset
if i < minIndex || i < 0 {
break
}
}
return
}
func (d *compressor ) writeStoredBlock (buf []byte ) error {
if d .w .writeStoredHeader (len (buf ), false ); d .w .err != nil {
return d .w .err
}
d .w .writeBytes (buf )
return d .w .err
}
const hashmul = 0x1e35a7bd
func hash4(b []byte ) uint32 {
return ((uint32 (b [3 ]) | uint32 (b [2 ])<<8 | uint32 (b [1 ])<<16 | uint32 (b [0 ])<<24 ) * hashmul ) >> (32 - hashBits )
}
func bulkHash4(b []byte , dst []uint32 ) {
if len (b ) < minMatchLength {
return
}
hb := uint32 (b [3 ]) | uint32 (b [2 ])<<8 | uint32 (b [1 ])<<16 | uint32 (b [0 ])<<24
dst [0 ] = (hb * hashmul ) >> (32 - hashBits )
end := len (b ) - minMatchLength + 1
for i := 1 ; i < end ; i ++ {
hb = (hb << 8 ) | uint32 (b [i +3 ])
dst [i ] = (hb * hashmul ) >> (32 - hashBits )
}
}
func matchLen(a , b []byte , max int ) int {
a = a [:max ]
b = b [:len (a )]
for i , av := range a {
if b [i ] != av {
return i
}
}
return max
}
func (d *compressor ) encSpeed () {
if d .windowEnd < maxStoreBlockSize {
if !d .sync {
return
}
if d .windowEnd < 128 {
switch {
case d .windowEnd == 0 :
return
case d .windowEnd <= 16 :
d .err = d .writeStoredBlock (d .window [:d .windowEnd ])
default :
d .w .writeBlockHuff (false , d .window [:d .windowEnd ])
d .err = d .w .err
}
d .windowEnd = 0
d .bestSpeed .reset ()
return
}
}
d .tokens = d .bestSpeed .encode (d .tokens [:0 ], d .window [:d .windowEnd ])
if len (d .tokens ) > d .windowEnd -(d .windowEnd >>4 ) {
d .w .writeBlockHuff (false , d .window [:d .windowEnd ])
} else {
d .w .writeBlockDynamic (d .tokens , false , d .window [:d .windowEnd ])
}
d .err = d .w .err
d .windowEnd = 0
}
func (d *compressor ) initDeflate () {
d .window = make ([]byte , 2 *windowSize )
d .hashOffset = 1
d .tokens = make ([]token , 0 , maxFlateBlockTokens +1 )
d .length = minMatchLength - 1
d .offset = 0
d .byteAvailable = false
d .index = 0
d .chainHead = -1
d .bulkHasher = bulkHash4
}
func (d *compressor ) deflate () {
if d .windowEnd -d .index < minMatchLength +maxMatchLength && !d .sync {
return
}
d .maxInsertIndex = d .windowEnd - (minMatchLength - 1 )
Loop :
for {
if d .index > d .windowEnd {
panic ("index > windowEnd" )
}
lookahead := d .windowEnd - d .index
if lookahead < minMatchLength +maxMatchLength {
if !d .sync {
break Loop
}
if d .index > d .windowEnd {
panic ("index > windowEnd" )
}
if lookahead == 0 {
if d .byteAvailable {
d .tokens = append (d .tokens , literalToken (uint32 (d .window [d .index -1 ])))
d .byteAvailable = false
}
if len (d .tokens ) > 0 {
if d .err = d .writeBlock (d .tokens , d .index ); d .err != nil {
return
}
d .tokens = d .tokens [:0 ]
}
break Loop
}
}
if d .index < d .maxInsertIndex {
hash := hash4 (d .window [d .index : d .index +minMatchLength ])
hh := &d .hashHead [hash &hashMask ]
d .chainHead = int (*hh )
d .hashPrev [d .index &windowMask ] = uint32 (d .chainHead )
*hh = uint32 (d .index + d .hashOffset )
}
prevLength := d .length
prevOffset := d .offset
d .length = minMatchLength - 1
d .offset = 0
minIndex := d .index - windowSize
if minIndex < 0 {
minIndex = 0
}
if d .chainHead -d .hashOffset >= minIndex &&
(d .fastSkipHashing != skipNever && lookahead > minMatchLength -1 ||
d .fastSkipHashing == skipNever && lookahead > prevLength && prevLength < d .lazy ) {
if newLength , newOffset , ok := d .findMatch (d .index , d .chainHead -d .hashOffset , minMatchLength -1 , lookahead ); ok {
d .length = newLength
d .offset = newOffset
}
}
if d .fastSkipHashing != skipNever && d .length >= minMatchLength ||
d .fastSkipHashing == skipNever && prevLength >= minMatchLength && d .length <= prevLength {
if d .fastSkipHashing != skipNever {
d .tokens = append (d .tokens , matchToken (uint32 (d .length -baseMatchLength ), uint32 (d .offset -baseMatchOffset )))
} else {
d .tokens = append (d .tokens , matchToken (uint32 (prevLength -baseMatchLength ), uint32 (prevOffset -baseMatchOffset )))
}
if d .length <= d .fastSkipHashing {
var newIndex int
if d .fastSkipHashing != skipNever {
newIndex = d .index + d .length
} else {
newIndex = d .index + prevLength - 1
}
index := d .index
for index ++; index < newIndex ; index ++ {
if index < d .maxInsertIndex {
hash := hash4 (d .window [index : index +minMatchLength ])
hh := &d .hashHead [hash &hashMask ]
d .hashPrev [index &windowMask ] = *hh
*hh = uint32 (index + d .hashOffset )
}
}
d .index = index
if d .fastSkipHashing == skipNever {
d .byteAvailable = false
d .length = minMatchLength - 1
}
} else {
d .index += d .length
}
if len (d .tokens ) == maxFlateBlockTokens {
if d .err = d .writeBlock (d .tokens , d .index ); d .err != nil {
return
}
d .tokens = d .tokens [:0 ]
}
} else {
if d .fastSkipHashing != skipNever || d .byteAvailable {
i := d .index - 1
if d .fastSkipHashing != skipNever {
i = d .index
}
d .tokens = append (d .tokens , literalToken (uint32 (d .window [i ])))
if len (d .tokens ) == maxFlateBlockTokens {
if d .err = d .writeBlock (d .tokens , i +1 ); d .err != nil {
return
}
d .tokens = d .tokens [:0 ]
}
}
d .index ++
if d .fastSkipHashing == skipNever {
d .byteAvailable = true
}
}
}
}
func (d *compressor ) fillStore (b []byte ) int {
n := copy (d .window [d .windowEnd :], b )
d .windowEnd += n
return n
}
func (d *compressor ) store () {
if d .windowEnd > 0 && (d .windowEnd == maxStoreBlockSize || d .sync ) {
d .err = d .writeStoredBlock (d .window [:d .windowEnd ])
d .windowEnd = 0
}
}
func (d *compressor ) storeHuff () {
if d .windowEnd < len (d .window ) && !d .sync || d .windowEnd == 0 {
return
}
d .w .writeBlockHuff (false , d .window [:d .windowEnd ])
d .err = d .w .err
d .windowEnd = 0
}
func (d *compressor ) write (b []byte ) (n int , err error ) {
if d .err != nil {
return 0 , d .err
}
n = len (b )
for len (b ) > 0 {
d .step (d )
b = b [d .fill (d , b ):]
if d .err != nil {
return 0 , d .err
}
}
return n , nil
}
func (d *compressor ) syncFlush () error {
if d .err != nil {
return d .err
}
d .sync = true
d .step (d )
if d .err == nil {
d .w .writeStoredHeader (0 , false )
d .w .flush ()
d .err = d .w .err
}
d .sync = false
return d .err
}
func (d *compressor ) init (w io .Writer , level int ) (err error ) {
d .w = newHuffmanBitWriter (w )
switch {
case level == NoCompression :
d .window = make ([]byte , maxStoreBlockSize )
d .fill = (*compressor ).fillStore
d .step = (*compressor ).store
case level == HuffmanOnly :
d .window = make ([]byte , maxStoreBlockSize )
d .fill = (*compressor ).fillStore
d .step = (*compressor ).storeHuff
case level == BestSpeed :
d .compressionLevel = levels [level ]
d .window = make ([]byte , maxStoreBlockSize )
d .fill = (*compressor ).fillStore
d .step = (*compressor ).encSpeed
d .bestSpeed = newDeflateFast ()
d .tokens = make ([]token , maxStoreBlockSize )
case level == DefaultCompression :
level = 6
fallthrough
case 2 <= level && level <= 9 :
d .compressionLevel = levels [level ]
d .initDeflate ()
d .fill = (*compressor ).fillDeflate
d .step = (*compressor ).deflate
default :
return fmt .Errorf ("flate: invalid compression level %d: want value in range [-2, 9]" , level )
}
return nil
}
func (d *compressor ) reset (w io .Writer ) {
d .w .reset (w )
d .sync = false
d .err = nil
switch d .compressionLevel .level {
case NoCompression :
d .windowEnd = 0
case BestSpeed :
d .windowEnd = 0
d .tokens = d .tokens [:0 ]
d .bestSpeed .reset ()
default :
d .chainHead = -1
for i := range d .hashHead {
d .hashHead [i ] = 0
}
for i := range d .hashPrev {
d .hashPrev [i ] = 0
}
d .hashOffset = 1
d .index , d .windowEnd = 0 , 0
d .blockStart , d .byteAvailable = 0 , false
d .tokens = d .tokens [:0 ]
d .length = minMatchLength - 1
d .offset = 0
d .maxInsertIndex = 0
}
}
func (d *compressor ) close () error {
if d .err == errWriterClosed {
return nil
}
if d .err != nil {
return d .err
}
d .sync = true
d .step (d )
if d .err != nil {
return d .err
}
if d .w .writeStoredHeader (0 , true ); d .w .err != nil {
return d .w .err
}
d .w .flush ()
if d .w .err != nil {
return d .w .err
}
d .err = errWriterClosed
return nil
}
func NewWriter (w io .Writer , level int ) (*Writer , error ) {
var dw Writer
if err := dw .d .init (w , level ); err != nil {
return nil , err
}
return &dw , nil
}
func NewWriterDict (w io .Writer , level int , dict []byte ) (*Writer , error ) {
dw := &dictWriter {w }
zw , err := NewWriter (dw , level )
if err != nil {
return nil , err
}
zw .d .fillWindow (dict )
zw .dict = append (zw .dict , dict ...)
return zw , err
}
type dictWriter struct {
w io .Writer
}
func (w *dictWriter ) Write (b []byte ) (n int , err error ) {
return w .w .Write (b )
}
var errWriterClosed = errors .New ("flate: closed writer" )
type Writer struct {
d compressor
dict []byte
}
func (w *Writer ) Write (data []byte ) (n int , err error ) {
return w .d .write (data )
}
func (w *Writer ) Flush () error {
return w .d .syncFlush ()
}
func (w *Writer ) Close () error {
return w .d .close ()
}
func (w *Writer ) Reset (dst io .Writer ) {
if dw , ok := w .d .w .writer .(*dictWriter ); ok {
dw .w = dst
w .d .reset (dw )
w .d .fillWindow (w .dict )
} else {
w .d .reset (dst )
}
}
The pages are generated with Golds v0.7.0-preview . (GOOS=linux GOARCH=amd64)
Golds is a Go 101 project developed by Tapir Liu .
PR and bug reports are welcome and can be submitted to the issue list .
Please follow @zigo_101 (reachable from the left QR code) to get the latest news of Golds .