id
int32 0
167k
| repo
stringlengths 5
54
| path
stringlengths 4
155
| func_name
stringlengths 1
118
| original_string
stringlengths 52
85.5k
| language
stringclasses 1
value | code
stringlengths 52
85.5k
| code_tokens
listlengths 21
1.41k
| docstring
stringlengths 6
2.61k
| docstring_tokens
listlengths 3
215
| sha
stringlengths 40
40
| url
stringlengths 85
252
|
---|---|---|---|---|---|---|---|---|---|---|---|
8,300 | klauspost/compress | huff0/compress.go | huffSort | func (s *Scratch) huffSort() {
type rankPos struct {
base uint32
current uint32
}
// Clear nodes
nodes := s.nodes[:huffNodesLen+1]
s.nodes = nodes
nodes = nodes[1 : huffNodesLen+1]
// Sort into buckets based on length of symbol count.
var rank [32]rankPos
for _, v := range s.count[:s.symbolLen] {
r := highBit32(v+1) & 31
rank[r].base++
}
for n := 30; n > 0; n-- {
rank[n-1].base += rank[n].base
}
for n := range rank[:] {
rank[n].current = rank[n].base
}
for n, c := range s.count[:s.symbolLen] {
r := (highBit32(c+1) + 1) & 31
pos := rank[r].current
rank[r].current++
prev := nodes[(pos-1)&huffNodesMask]
for pos > rank[r].base && c > prev.count {
nodes[pos&huffNodesMask] = prev
pos--
prev = nodes[(pos-1)&huffNodesMask]
}
nodes[pos&huffNodesMask] = nodeElt{count: c, symbol: byte(n)}
}
return
} | go | func (s *Scratch) huffSort() {
type rankPos struct {
base uint32
current uint32
}
// Clear nodes
nodes := s.nodes[:huffNodesLen+1]
s.nodes = nodes
nodes = nodes[1 : huffNodesLen+1]
// Sort into buckets based on length of symbol count.
var rank [32]rankPos
for _, v := range s.count[:s.symbolLen] {
r := highBit32(v+1) & 31
rank[r].base++
}
for n := 30; n > 0; n-- {
rank[n-1].base += rank[n].base
}
for n := range rank[:] {
rank[n].current = rank[n].base
}
for n, c := range s.count[:s.symbolLen] {
r := (highBit32(c+1) + 1) & 31
pos := rank[r].current
rank[r].current++
prev := nodes[(pos-1)&huffNodesMask]
for pos > rank[r].base && c > prev.count {
nodes[pos&huffNodesMask] = prev
pos--
prev = nodes[(pos-1)&huffNodesMask]
}
nodes[pos&huffNodesMask] = nodeElt{count: c, symbol: byte(n)}
}
return
} | [
"func",
"(",
"s",
"*",
"Scratch",
")",
"huffSort",
"(",
")",
"{",
"type",
"rankPos",
"struct",
"{",
"base",
"uint32",
"\n",
"current",
"uint32",
"\n",
"}",
"\n\n",
"// Clear nodes",
"nodes",
":=",
"s",
".",
"nodes",
"[",
":",
"huffNodesLen",
"+",
"1",
"]",
"\n",
"s",
".",
"nodes",
"=",
"nodes",
"\n",
"nodes",
"=",
"nodes",
"[",
"1",
":",
"huffNodesLen",
"+",
"1",
"]",
"\n\n",
"// Sort into buckets based on length of symbol count.",
"var",
"rank",
"[",
"32",
"]",
"rankPos",
"\n",
"for",
"_",
",",
"v",
":=",
"range",
"s",
".",
"count",
"[",
":",
"s",
".",
"symbolLen",
"]",
"{",
"r",
":=",
"highBit32",
"(",
"v",
"+",
"1",
")",
"&",
"31",
"\n",
"rank",
"[",
"r",
"]",
".",
"base",
"++",
"\n",
"}",
"\n",
"for",
"n",
":=",
"30",
";",
"n",
">",
"0",
";",
"n",
"--",
"{",
"rank",
"[",
"n",
"-",
"1",
"]",
".",
"base",
"+=",
"rank",
"[",
"n",
"]",
".",
"base",
"\n",
"}",
"\n",
"for",
"n",
":=",
"range",
"rank",
"[",
":",
"]",
"{",
"rank",
"[",
"n",
"]",
".",
"current",
"=",
"rank",
"[",
"n",
"]",
".",
"base",
"\n",
"}",
"\n",
"for",
"n",
",",
"c",
":=",
"range",
"s",
".",
"count",
"[",
":",
"s",
".",
"symbolLen",
"]",
"{",
"r",
":=",
"(",
"highBit32",
"(",
"c",
"+",
"1",
")",
"+",
"1",
")",
"&",
"31",
"\n",
"pos",
":=",
"rank",
"[",
"r",
"]",
".",
"current",
"\n",
"rank",
"[",
"r",
"]",
".",
"current",
"++",
"\n",
"prev",
":=",
"nodes",
"[",
"(",
"pos",
"-",
"1",
")",
"&",
"huffNodesMask",
"]",
"\n",
"for",
"pos",
">",
"rank",
"[",
"r",
"]",
".",
"base",
"&&",
"c",
">",
"prev",
".",
"count",
"{",
"nodes",
"[",
"pos",
"&",
"huffNodesMask",
"]",
"=",
"prev",
"\n",
"pos",
"--",
"\n",
"prev",
"=",
"nodes",
"[",
"(",
"pos",
"-",
"1",
")",
"&",
"huffNodesMask",
"]",
"\n",
"}",
"\n",
"nodes",
"[",
"pos",
"&",
"huffNodesMask",
"]",
"=",
"nodeElt",
"{",
"count",
":",
"c",
",",
"symbol",
":",
"byte",
"(",
"n",
")",
"}",
"\n",
"}",
"\n",
"return",
"\n",
"}"
]
| // huffSort will sort symbols, decreasing order. | [
"huffSort",
"will",
"sort",
"symbols",
"decreasing",
"order",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/huff0/compress.go#L455-L491 |
8,301 | klauspost/compress | zlib/writer.go | NewWriter | func NewWriter(w io.Writer) *Writer {
z, _ := NewWriterLevelDict(w, DefaultCompression, nil)
return z
} | go | func NewWriter(w io.Writer) *Writer {
z, _ := NewWriterLevelDict(w, DefaultCompression, nil)
return z
} | [
"func",
"NewWriter",
"(",
"w",
"io",
".",
"Writer",
")",
"*",
"Writer",
"{",
"z",
",",
"_",
":=",
"NewWriterLevelDict",
"(",
"w",
",",
"DefaultCompression",
",",
"nil",
")",
"\n",
"return",
"z",
"\n",
"}"
]
| // NewWriter creates a new Writer.
// Writes to the returned Writer are compressed and written to w.
//
// It is the caller's responsibility to call Close on the WriteCloser when done.
// Writes may be buffered and not flushed until Close. | [
"NewWriter",
"creates",
"a",
"new",
"Writer",
".",
"Writes",
"to",
"the",
"returned",
"Writer",
"are",
"compressed",
"and",
"written",
"to",
"w",
".",
"It",
"is",
"the",
"caller",
"s",
"responsibility",
"to",
"call",
"Close",
"on",
"the",
"WriteCloser",
"when",
"done",
".",
"Writes",
"may",
"be",
"buffered",
"and",
"not",
"flushed",
"until",
"Close",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zlib/writer.go#L45-L48 |
8,302 | klauspost/compress | zlib/writer.go | NewWriterLevel | func NewWriterLevel(w io.Writer, level int) (*Writer, error) {
return NewWriterLevelDict(w, level, nil)
} | go | func NewWriterLevel(w io.Writer, level int) (*Writer, error) {
return NewWriterLevelDict(w, level, nil)
} | [
"func",
"NewWriterLevel",
"(",
"w",
"io",
".",
"Writer",
",",
"level",
"int",
")",
"(",
"*",
"Writer",
",",
"error",
")",
"{",
"return",
"NewWriterLevelDict",
"(",
"w",
",",
"level",
",",
"nil",
")",
"\n",
"}"
]
| // NewWriterLevel is like NewWriter but specifies the compression level instead
// of assuming DefaultCompression.
//
// The compression level can be DefaultCompression, NoCompression, HuffmanOnly
// or any integer value between BestSpeed and BestCompression inclusive.
// The error returned will be nil if the level is valid. | [
"NewWriterLevel",
"is",
"like",
"NewWriter",
"but",
"specifies",
"the",
"compression",
"level",
"instead",
"of",
"assuming",
"DefaultCompression",
".",
"The",
"compression",
"level",
"can",
"be",
"DefaultCompression",
"NoCompression",
"HuffmanOnly",
"or",
"any",
"integer",
"value",
"between",
"BestSpeed",
"and",
"BestCompression",
"inclusive",
".",
"The",
"error",
"returned",
"will",
"be",
"nil",
"if",
"the",
"level",
"is",
"valid",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zlib/writer.go#L56-L58 |
8,303 | klauspost/compress | zlib/writer.go | NewWriterLevelDict | func NewWriterLevelDict(w io.Writer, level int, dict []byte) (*Writer, error) {
if level < HuffmanOnly || level > BestCompression {
return nil, fmt.Errorf("zlib: invalid compression level: %d", level)
}
return &Writer{
w: w,
level: level,
dict: dict,
}, nil
} | go | func NewWriterLevelDict(w io.Writer, level int, dict []byte) (*Writer, error) {
if level < HuffmanOnly || level > BestCompression {
return nil, fmt.Errorf("zlib: invalid compression level: %d", level)
}
return &Writer{
w: w,
level: level,
dict: dict,
}, nil
} | [
"func",
"NewWriterLevelDict",
"(",
"w",
"io",
".",
"Writer",
",",
"level",
"int",
",",
"dict",
"[",
"]",
"byte",
")",
"(",
"*",
"Writer",
",",
"error",
")",
"{",
"if",
"level",
"<",
"HuffmanOnly",
"||",
"level",
">",
"BestCompression",
"{",
"return",
"nil",
",",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"level",
")",
"\n",
"}",
"\n",
"return",
"&",
"Writer",
"{",
"w",
":",
"w",
",",
"level",
":",
"level",
",",
"dict",
":",
"dict",
",",
"}",
",",
"nil",
"\n",
"}"
]
| // NewWriterLevelDict is like NewWriterLevel but specifies a dictionary to
// compress with.
//
// The dictionary may be nil. If not, its contents should not be modified until
// the Writer is closed. | [
"NewWriterLevelDict",
"is",
"like",
"NewWriterLevel",
"but",
"specifies",
"a",
"dictionary",
"to",
"compress",
"with",
".",
"The",
"dictionary",
"may",
"be",
"nil",
".",
"If",
"not",
"its",
"contents",
"should",
"not",
"be",
"modified",
"until",
"the",
"Writer",
"is",
"closed",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zlib/writer.go#L65-L74 |
8,304 | klauspost/compress | zlib/writer.go | Reset | func (z *Writer) Reset(w io.Writer) {
z.w = w
// z.level and z.dict left unchanged.
if z.compressor != nil {
z.compressor.Reset(w)
}
if z.digest != nil {
z.digest.Reset()
}
z.err = nil
z.scratch = [4]byte{}
z.wroteHeader = false
} | go | func (z *Writer) Reset(w io.Writer) {
z.w = w
// z.level and z.dict left unchanged.
if z.compressor != nil {
z.compressor.Reset(w)
}
if z.digest != nil {
z.digest.Reset()
}
z.err = nil
z.scratch = [4]byte{}
z.wroteHeader = false
} | [
"func",
"(",
"z",
"*",
"Writer",
")",
"Reset",
"(",
"w",
"io",
".",
"Writer",
")",
"{",
"z",
".",
"w",
"=",
"w",
"\n",
"// z.level and z.dict left unchanged.",
"if",
"z",
".",
"compressor",
"!=",
"nil",
"{",
"z",
".",
"compressor",
".",
"Reset",
"(",
"w",
")",
"\n",
"}",
"\n",
"if",
"z",
".",
"digest",
"!=",
"nil",
"{",
"z",
".",
"digest",
".",
"Reset",
"(",
")",
"\n",
"}",
"\n",
"z",
".",
"err",
"=",
"nil",
"\n",
"z",
".",
"scratch",
"=",
"[",
"4",
"]",
"byte",
"{",
"}",
"\n",
"z",
".",
"wroteHeader",
"=",
"false",
"\n",
"}"
]
| // Reset clears the state of the Writer z such that it is equivalent to its
// initial state from NewWriterLevel or NewWriterLevelDict, but instead writing
// to w. | [
"Reset",
"clears",
"the",
"state",
"of",
"the",
"Writer",
"z",
"such",
"that",
"it",
"is",
"equivalent",
"to",
"its",
"initial",
"state",
"from",
"NewWriterLevel",
"or",
"NewWriterLevelDict",
"but",
"instead",
"writing",
"to",
"w",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zlib/writer.go#L79-L91 |
8,305 | klauspost/compress | zlib/writer.go | writeHeader | func (z *Writer) writeHeader() (err error) {
z.wroteHeader = true
// ZLIB has a two-byte header (as documented in RFC 1950).
// The first four bits is the CINFO (compression info), which is 7 for the default deflate window size.
// The next four bits is the CM (compression method), which is 8 for deflate.
z.scratch[0] = 0x78
// The next two bits is the FLEVEL (compression level). The four values are:
// 0=fastest, 1=fast, 2=default, 3=best.
// The next bit, FDICT, is set if a dictionary is given.
// The final five FCHECK bits form a mod-31 checksum.
switch z.level {
case -2, 0, 1:
z.scratch[1] = 0 << 6
case 2, 3, 4, 5:
z.scratch[1] = 1 << 6
case 6, -1:
z.scratch[1] = 2 << 6
case 7, 8, 9:
z.scratch[1] = 3 << 6
default:
panic("unreachable")
}
if z.dict != nil {
z.scratch[1] |= 1 << 5
}
z.scratch[1] += uint8(31 - (uint16(z.scratch[0])<<8+uint16(z.scratch[1]))%31)
if _, err = z.w.Write(z.scratch[0:2]); err != nil {
return err
}
if z.dict != nil {
// The next four bytes are the Adler-32 checksum of the dictionary.
checksum := adler32.Checksum(z.dict)
z.scratch[0] = uint8(checksum >> 24)
z.scratch[1] = uint8(checksum >> 16)
z.scratch[2] = uint8(checksum >> 8)
z.scratch[3] = uint8(checksum >> 0)
if _, err = z.w.Write(z.scratch[0:4]); err != nil {
return err
}
}
if z.compressor == nil {
// Initialize deflater unless the Writer is being reused
// after a Reset call.
z.compressor, err = flate.NewWriterDict(z.w, z.level, z.dict)
if err != nil {
return err
}
z.digest = adler32.New()
}
return nil
} | go | func (z *Writer) writeHeader() (err error) {
z.wroteHeader = true
// ZLIB has a two-byte header (as documented in RFC 1950).
// The first four bits is the CINFO (compression info), which is 7 for the default deflate window size.
// The next four bits is the CM (compression method), which is 8 for deflate.
z.scratch[0] = 0x78
// The next two bits is the FLEVEL (compression level). The four values are:
// 0=fastest, 1=fast, 2=default, 3=best.
// The next bit, FDICT, is set if a dictionary is given.
// The final five FCHECK bits form a mod-31 checksum.
switch z.level {
case -2, 0, 1:
z.scratch[1] = 0 << 6
case 2, 3, 4, 5:
z.scratch[1] = 1 << 6
case 6, -1:
z.scratch[1] = 2 << 6
case 7, 8, 9:
z.scratch[1] = 3 << 6
default:
panic("unreachable")
}
if z.dict != nil {
z.scratch[1] |= 1 << 5
}
z.scratch[1] += uint8(31 - (uint16(z.scratch[0])<<8+uint16(z.scratch[1]))%31)
if _, err = z.w.Write(z.scratch[0:2]); err != nil {
return err
}
if z.dict != nil {
// The next four bytes are the Adler-32 checksum of the dictionary.
checksum := adler32.Checksum(z.dict)
z.scratch[0] = uint8(checksum >> 24)
z.scratch[1] = uint8(checksum >> 16)
z.scratch[2] = uint8(checksum >> 8)
z.scratch[3] = uint8(checksum >> 0)
if _, err = z.w.Write(z.scratch[0:4]); err != nil {
return err
}
}
if z.compressor == nil {
// Initialize deflater unless the Writer is being reused
// after a Reset call.
z.compressor, err = flate.NewWriterDict(z.w, z.level, z.dict)
if err != nil {
return err
}
z.digest = adler32.New()
}
return nil
} | [
"func",
"(",
"z",
"*",
"Writer",
")",
"writeHeader",
"(",
")",
"(",
"err",
"error",
")",
"{",
"z",
".",
"wroteHeader",
"=",
"true",
"\n",
"// ZLIB has a two-byte header (as documented in RFC 1950).",
"// The first four bits is the CINFO (compression info), which is 7 for the default deflate window size.",
"// The next four bits is the CM (compression method), which is 8 for deflate.",
"z",
".",
"scratch",
"[",
"0",
"]",
"=",
"0x78",
"\n",
"// The next two bits is the FLEVEL (compression level). The four values are:",
"// 0=fastest, 1=fast, 2=default, 3=best.",
"// The next bit, FDICT, is set if a dictionary is given.",
"// The final five FCHECK bits form a mod-31 checksum.",
"switch",
"z",
".",
"level",
"{",
"case",
"-",
"2",
",",
"0",
",",
"1",
":",
"z",
".",
"scratch",
"[",
"1",
"]",
"=",
"0",
"<<",
"6",
"\n",
"case",
"2",
",",
"3",
",",
"4",
",",
"5",
":",
"z",
".",
"scratch",
"[",
"1",
"]",
"=",
"1",
"<<",
"6",
"\n",
"case",
"6",
",",
"-",
"1",
":",
"z",
".",
"scratch",
"[",
"1",
"]",
"=",
"2",
"<<",
"6",
"\n",
"case",
"7",
",",
"8",
",",
"9",
":",
"z",
".",
"scratch",
"[",
"1",
"]",
"=",
"3",
"<<",
"6",
"\n",
"default",
":",
"panic",
"(",
"\"",
"\"",
")",
"\n",
"}",
"\n",
"if",
"z",
".",
"dict",
"!=",
"nil",
"{",
"z",
".",
"scratch",
"[",
"1",
"]",
"|=",
"1",
"<<",
"5",
"\n",
"}",
"\n",
"z",
".",
"scratch",
"[",
"1",
"]",
"+=",
"uint8",
"(",
"31",
"-",
"(",
"uint16",
"(",
"z",
".",
"scratch",
"[",
"0",
"]",
")",
"<<",
"8",
"+",
"uint16",
"(",
"z",
".",
"scratch",
"[",
"1",
"]",
")",
")",
"%",
"31",
")",
"\n",
"if",
"_",
",",
"err",
"=",
"z",
".",
"w",
".",
"Write",
"(",
"z",
".",
"scratch",
"[",
"0",
":",
"2",
"]",
")",
";",
"err",
"!=",
"nil",
"{",
"return",
"err",
"\n",
"}",
"\n",
"if",
"z",
".",
"dict",
"!=",
"nil",
"{",
"// The next four bytes are the Adler-32 checksum of the dictionary.",
"checksum",
":=",
"adler32",
".",
"Checksum",
"(",
"z",
".",
"dict",
")",
"\n",
"z",
".",
"scratch",
"[",
"0",
"]",
"=",
"uint8",
"(",
"checksum",
">>",
"24",
")",
"\n",
"z",
".",
"scratch",
"[",
"1",
"]",
"=",
"uint8",
"(",
"checksum",
">>",
"16",
")",
"\n",
"z",
".",
"scratch",
"[",
"2",
"]",
"=",
"uint8",
"(",
"checksum",
">>",
"8",
")",
"\n",
"z",
".",
"scratch",
"[",
"3",
"]",
"=",
"uint8",
"(",
"checksum",
">>",
"0",
")",
"\n",
"if",
"_",
",",
"err",
"=",
"z",
".",
"w",
".",
"Write",
"(",
"z",
".",
"scratch",
"[",
"0",
":",
"4",
"]",
")",
";",
"err",
"!=",
"nil",
"{",
"return",
"err",
"\n",
"}",
"\n",
"}",
"\n",
"if",
"z",
".",
"compressor",
"==",
"nil",
"{",
"// Initialize deflater unless the Writer is being reused",
"// after a Reset call.",
"z",
".",
"compressor",
",",
"err",
"=",
"flate",
".",
"NewWriterDict",
"(",
"z",
".",
"w",
",",
"z",
".",
"level",
",",
"z",
".",
"dict",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"return",
"err",
"\n",
"}",
"\n",
"z",
".",
"digest",
"=",
"adler32",
".",
"New",
"(",
")",
"\n",
"}",
"\n",
"return",
"nil",
"\n",
"}"
]
| // writeHeader writes the ZLIB header. | [
"writeHeader",
"writes",
"the",
"ZLIB",
"header",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zlib/writer.go#L94-L144 |
8,306 | klauspost/compress | zlib/writer.go | Write | func (z *Writer) Write(p []byte) (n int, err error) {
if !z.wroteHeader {
z.err = z.writeHeader()
}
if z.err != nil {
return 0, z.err
}
if len(p) == 0 {
return 0, nil
}
n, err = z.compressor.Write(p)
if err != nil {
z.err = err
return
}
z.digest.Write(p)
return
} | go | func (z *Writer) Write(p []byte) (n int, err error) {
if !z.wroteHeader {
z.err = z.writeHeader()
}
if z.err != nil {
return 0, z.err
}
if len(p) == 0 {
return 0, nil
}
n, err = z.compressor.Write(p)
if err != nil {
z.err = err
return
}
z.digest.Write(p)
return
} | [
"func",
"(",
"z",
"*",
"Writer",
")",
"Write",
"(",
"p",
"[",
"]",
"byte",
")",
"(",
"n",
"int",
",",
"err",
"error",
")",
"{",
"if",
"!",
"z",
".",
"wroteHeader",
"{",
"z",
".",
"err",
"=",
"z",
".",
"writeHeader",
"(",
")",
"\n",
"}",
"\n",
"if",
"z",
".",
"err",
"!=",
"nil",
"{",
"return",
"0",
",",
"z",
".",
"err",
"\n",
"}",
"\n",
"if",
"len",
"(",
"p",
")",
"==",
"0",
"{",
"return",
"0",
",",
"nil",
"\n",
"}",
"\n",
"n",
",",
"err",
"=",
"z",
".",
"compressor",
".",
"Write",
"(",
"p",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"z",
".",
"err",
"=",
"err",
"\n",
"return",
"\n",
"}",
"\n",
"z",
".",
"digest",
".",
"Write",
"(",
"p",
")",
"\n",
"return",
"\n",
"}"
]
| // Write writes a compressed form of p to the underlying io.Writer. The
// compressed bytes are not necessarily flushed until the Writer is closed or
// explicitly flushed. | [
"Write",
"writes",
"a",
"compressed",
"form",
"of",
"p",
"to",
"the",
"underlying",
"io",
".",
"Writer",
".",
"The",
"compressed",
"bytes",
"are",
"not",
"necessarily",
"flushed",
"until",
"the",
"Writer",
"is",
"closed",
"or",
"explicitly",
"flushed",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zlib/writer.go#L149-L166 |
8,307 | klauspost/compress | zip/writer.go | NewWriter | func NewWriter(w io.Writer) *Writer {
return &Writer{cw: &countWriter{w: bufio.NewWriter(w)}}
} | go | func NewWriter(w io.Writer) *Writer {
return &Writer{cw: &countWriter{w: bufio.NewWriter(w)}}
} | [
"func",
"NewWriter",
"(",
"w",
"io",
".",
"Writer",
")",
"*",
"Writer",
"{",
"return",
"&",
"Writer",
"{",
"cw",
":",
"&",
"countWriter",
"{",
"w",
":",
"bufio",
".",
"NewWriter",
"(",
"w",
")",
"}",
"}",
"\n",
"}"
]
| // NewWriter returns a new Writer writing a zip file to w. | [
"NewWriter",
"returns",
"a",
"new",
"Writer",
"writing",
"a",
"zip",
"file",
"to",
"w",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zip/writer.go#L33-L35 |
8,308 | klauspost/compress | zip/writer.go | SetOffset | func (w *Writer) SetOffset(n int64) {
if w.cw.count != 0 {
panic("zip: SetOffset called after data was written")
}
w.cw.count = n
} | go | func (w *Writer) SetOffset(n int64) {
if w.cw.count != 0 {
panic("zip: SetOffset called after data was written")
}
w.cw.count = n
} | [
"func",
"(",
"w",
"*",
"Writer",
")",
"SetOffset",
"(",
"n",
"int64",
")",
"{",
"if",
"w",
".",
"cw",
".",
"count",
"!=",
"0",
"{",
"panic",
"(",
"\"",
"\"",
")",
"\n",
"}",
"\n",
"w",
".",
"cw",
".",
"count",
"=",
"n",
"\n",
"}"
]
| // SetOffset sets the offset of the beginning of the zip data within the
// underlying writer. It should be used when the zip data is appended to an
// existing file, such as a binary executable.
// It must be called before any data is written. | [
"SetOffset",
"sets",
"the",
"offset",
"of",
"the",
"beginning",
"of",
"the",
"zip",
"data",
"within",
"the",
"underlying",
"writer",
".",
"It",
"should",
"be",
"used",
"when",
"the",
"zip",
"data",
"is",
"appended",
"to",
"an",
"existing",
"file",
"such",
"as",
"a",
"binary",
"executable",
".",
"It",
"must",
"be",
"called",
"before",
"any",
"data",
"is",
"written",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zip/writer.go#L41-L46 |
8,309 | klauspost/compress | zip/writer.go | Flush | func (w *Writer) Flush() error {
return w.cw.w.(*bufio.Writer).Flush()
} | go | func (w *Writer) Flush() error {
return w.cw.w.(*bufio.Writer).Flush()
} | [
"func",
"(",
"w",
"*",
"Writer",
")",
"Flush",
"(",
")",
"error",
"{",
"return",
"w",
".",
"cw",
".",
"w",
".",
"(",
"*",
"bufio",
".",
"Writer",
")",
".",
"Flush",
"(",
")",
"\n",
"}"
]
| // Flush flushes any buffered data to the underlying writer.
// Calling Flush is not normally necessary; calling Close is sufficient. | [
"Flush",
"flushes",
"any",
"buffered",
"data",
"to",
"the",
"underlying",
"writer",
".",
"Calling",
"Flush",
"is",
"not",
"normally",
"necessary",
";",
"calling",
"Close",
"is",
"sufficient",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zip/writer.go#L50-L52 |
8,310 | klauspost/compress | zip/writer.go | RegisterCompressor | func (w *Writer) RegisterCompressor(method uint16, comp Compressor) {
if w.compressors == nil {
w.compressors = make(map[uint16]Compressor)
}
w.compressors[method] = comp
} | go | func (w *Writer) RegisterCompressor(method uint16, comp Compressor) {
if w.compressors == nil {
w.compressors = make(map[uint16]Compressor)
}
w.compressors[method] = comp
} | [
"func",
"(",
"w",
"*",
"Writer",
")",
"RegisterCompressor",
"(",
"method",
"uint16",
",",
"comp",
"Compressor",
")",
"{",
"if",
"w",
".",
"compressors",
"==",
"nil",
"{",
"w",
".",
"compressors",
"=",
"make",
"(",
"map",
"[",
"uint16",
"]",
"Compressor",
")",
"\n",
"}",
"\n",
"w",
".",
"compressors",
"[",
"method",
"]",
"=",
"comp",
"\n",
"}"
]
| // RegisterCompressor registers or overrides a custom compressor for a specific
// method ID. If a compressor for a given method is not found, Writer will
// default to looking up the compressor at the package level. | [
"RegisterCompressor",
"registers",
"or",
"overrides",
"a",
"custom",
"compressor",
"for",
"a",
"specific",
"method",
"ID",
".",
"If",
"a",
"compressor",
"for",
"a",
"given",
"method",
"is",
"not",
"found",
"Writer",
"will",
"default",
"to",
"looking",
"up",
"the",
"compressor",
"at",
"the",
"package",
"level",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zip/writer.go#L276-L281 |
8,311 | klauspost/compress | huff0/huff0.go | estimateSize | func (c cTable) estimateSize(hist []uint32) int {
nbBits := uint32(7)
for i, v := range c[:len(hist)] {
nbBits += uint32(v.nBits) * hist[i]
}
return int(nbBits >> 3)
} | go | func (c cTable) estimateSize(hist []uint32) int {
nbBits := uint32(7)
for i, v := range c[:len(hist)] {
nbBits += uint32(v.nBits) * hist[i]
}
return int(nbBits >> 3)
} | [
"func",
"(",
"c",
"cTable",
")",
"estimateSize",
"(",
"hist",
"[",
"]",
"uint32",
")",
"int",
"{",
"nbBits",
":=",
"uint32",
"(",
"7",
")",
"\n",
"for",
"i",
",",
"v",
":=",
"range",
"c",
"[",
":",
"len",
"(",
"hist",
")",
"]",
"{",
"nbBits",
"+=",
"uint32",
"(",
"v",
".",
"nBits",
")",
"*",
"hist",
"[",
"i",
"]",
"\n",
"}",
"\n",
"return",
"int",
"(",
"nbBits",
">>",
"3",
")",
"\n",
"}"
]
| // estimateSize returns the estimated size in bytes of the input represented in the
// histogram supplied. | [
"estimateSize",
"returns",
"the",
"estimated",
"size",
"in",
"bytes",
"of",
"the",
"input",
"represented",
"in",
"the",
"histogram",
"supplied",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/huff0/huff0.go#L217-L223 |
8,312 | klauspost/compress | huff0/huff0.go | minSize | func (s *Scratch) minSize(total int) int {
nbBits := float64(7)
fTotal := float64(total)
for _, v := range s.count[:s.symbolLen] {
n := float64(v)
if n > 0 {
nbBits += math.Log2(fTotal/n) * n
}
}
return int(nbBits) >> 3
} | go | func (s *Scratch) minSize(total int) int {
nbBits := float64(7)
fTotal := float64(total)
for _, v := range s.count[:s.symbolLen] {
n := float64(v)
if n > 0 {
nbBits += math.Log2(fTotal/n) * n
}
}
return int(nbBits) >> 3
} | [
"func",
"(",
"s",
"*",
"Scratch",
")",
"minSize",
"(",
"total",
"int",
")",
"int",
"{",
"nbBits",
":=",
"float64",
"(",
"7",
")",
"\n",
"fTotal",
":=",
"float64",
"(",
"total",
")",
"\n",
"for",
"_",
",",
"v",
":=",
"range",
"s",
".",
"count",
"[",
":",
"s",
".",
"symbolLen",
"]",
"{",
"n",
":=",
"float64",
"(",
"v",
")",
"\n",
"if",
"n",
">",
"0",
"{",
"nbBits",
"+=",
"math",
".",
"Log2",
"(",
"fTotal",
"/",
"n",
")",
"*",
"n",
"\n",
"}",
"\n",
"}",
"\n",
"return",
"int",
"(",
"nbBits",
")",
">>",
"3",
"\n",
"}"
]
| // minSize returns the minimum possible size considering the shannon limit. | [
"minSize",
"returns",
"the",
"minimum",
"possible",
"size",
"considering",
"the",
"shannon",
"limit",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/huff0/huff0.go#L226-L236 |
8,313 | klauspost/compress | zstd/fse_decoder.go | decSymbolValue | func decSymbolValue(symb uint8, t []baseOffset) (decSymbol, error) {
if int(symb) >= len(t) {
return decSymbol{}, fmt.Errorf("rle symbol %d >= max %d", symb, len(t))
}
lu := t[symb]
return decSymbol{
addBits: lu.addBits,
baseline: lu.baseLine,
}, nil
} | go | func decSymbolValue(symb uint8, t []baseOffset) (decSymbol, error) {
if int(symb) >= len(t) {
return decSymbol{}, fmt.Errorf("rle symbol %d >= max %d", symb, len(t))
}
lu := t[symb]
return decSymbol{
addBits: lu.addBits,
baseline: lu.baseLine,
}, nil
} | [
"func",
"decSymbolValue",
"(",
"symb",
"uint8",
",",
"t",
"[",
"]",
"baseOffset",
")",
"(",
"decSymbol",
",",
"error",
")",
"{",
"if",
"int",
"(",
"symb",
")",
">=",
"len",
"(",
"t",
")",
"{",
"return",
"decSymbol",
"{",
"}",
",",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"symb",
",",
"len",
"(",
"t",
")",
")",
"\n",
"}",
"\n",
"lu",
":=",
"t",
"[",
"symb",
"]",
"\n",
"return",
"decSymbol",
"{",
"addBits",
":",
"lu",
".",
"addBits",
",",
"baseline",
":",
"lu",
".",
"baseLine",
",",
"}",
",",
"nil",
"\n",
"}"
]
| // decSymbolValue returns the transformed decSymbol for the given symbol. | [
"decSymbolValue",
"returns",
"the",
"transformed",
"decSymbol",
"for",
"the",
"given",
"symbol",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zstd/fse_decoder.go#L186-L195 |
8,314 | klauspost/compress | zstd/fse_decoder.go | setRLE | func (s *fseDecoder) setRLE(symbol decSymbol) {
s.actualTableLog = 0
s.maxBits = 0
s.dt[0] = symbol
} | go | func (s *fseDecoder) setRLE(symbol decSymbol) {
s.actualTableLog = 0
s.maxBits = 0
s.dt[0] = symbol
} | [
"func",
"(",
"s",
"*",
"fseDecoder",
")",
"setRLE",
"(",
"symbol",
"decSymbol",
")",
"{",
"s",
".",
"actualTableLog",
"=",
"0",
"\n",
"s",
".",
"maxBits",
"=",
"0",
"\n",
"s",
".",
"dt",
"[",
"0",
"]",
"=",
"symbol",
"\n",
"}"
]
| // setRLE will set the decoder til RLE mode. | [
"setRLE",
"will",
"set",
"the",
"decoder",
"til",
"RLE",
"mode",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zstd/fse_decoder.go#L198-L202 |
8,315 | klauspost/compress | zstd/fse_decoder.go | transform | func (s *fseDecoder) transform(t []baseOffset) error {
tableSize := uint16(1 << s.actualTableLog)
s.maxBits = 0
for i, v := range s.dt[:tableSize] {
if int(v.addBits) >= len(t) {
return fmt.Errorf("invalid decoding table entry %d, symbol %d >= max (%d)", i, v.addBits, len(t))
}
lu := t[v.addBits]
if lu.addBits > s.maxBits {
s.maxBits = lu.addBits
}
s.dt[i&maxTableMask] = decSymbol{
newState: v.newState,
nbBits: v.nbBits,
addBits: lu.addBits,
baseline: lu.baseLine,
}
}
return nil
} | go | func (s *fseDecoder) transform(t []baseOffset) error {
tableSize := uint16(1 << s.actualTableLog)
s.maxBits = 0
for i, v := range s.dt[:tableSize] {
if int(v.addBits) >= len(t) {
return fmt.Errorf("invalid decoding table entry %d, symbol %d >= max (%d)", i, v.addBits, len(t))
}
lu := t[v.addBits]
if lu.addBits > s.maxBits {
s.maxBits = lu.addBits
}
s.dt[i&maxTableMask] = decSymbol{
newState: v.newState,
nbBits: v.nbBits,
addBits: lu.addBits,
baseline: lu.baseLine,
}
}
return nil
} | [
"func",
"(",
"s",
"*",
"fseDecoder",
")",
"transform",
"(",
"t",
"[",
"]",
"baseOffset",
")",
"error",
"{",
"tableSize",
":=",
"uint16",
"(",
"1",
"<<",
"s",
".",
"actualTableLog",
")",
"\n",
"s",
".",
"maxBits",
"=",
"0",
"\n",
"for",
"i",
",",
"v",
":=",
"range",
"s",
".",
"dt",
"[",
":",
"tableSize",
"]",
"{",
"if",
"int",
"(",
"v",
".",
"addBits",
")",
">=",
"len",
"(",
"t",
")",
"{",
"return",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"i",
",",
"v",
".",
"addBits",
",",
"len",
"(",
"t",
")",
")",
"\n",
"}",
"\n",
"lu",
":=",
"t",
"[",
"v",
".",
"addBits",
"]",
"\n",
"if",
"lu",
".",
"addBits",
">",
"s",
".",
"maxBits",
"{",
"s",
".",
"maxBits",
"=",
"lu",
".",
"addBits",
"\n",
"}",
"\n",
"s",
".",
"dt",
"[",
"i",
"&",
"maxTableMask",
"]",
"=",
"decSymbol",
"{",
"newState",
":",
"v",
".",
"newState",
",",
"nbBits",
":",
"v",
".",
"nbBits",
",",
"addBits",
":",
"lu",
".",
"addBits",
",",
"baseline",
":",
"lu",
".",
"baseLine",
",",
"}",
"\n",
"}",
"\n",
"return",
"nil",
"\n",
"}"
]
| // transform will transform the decoder table into a table usable for
// decoding without having to apply the transformation while decoding.
// The state will contain the base value and the number of bits to read. | [
"transform",
"will",
"transform",
"the",
"decoder",
"table",
"into",
"a",
"table",
"usable",
"for",
"decoding",
"without",
"having",
"to",
"apply",
"the",
"transformation",
"while",
"decoding",
".",
"The",
"state",
"will",
"contain",
"the",
"base",
"value",
"and",
"the",
"number",
"of",
"bits",
"to",
"read",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zstd/fse_decoder.go#L269-L288 |
8,316 | klauspost/compress | zstd/fse_decoder.go | init | func (s *fseState) init(br *bitReader, tableLog uint8, dt []decSymbol) {
s.dt = dt
br.fill()
s.state = dt[br.getBits(tableLog)]
} | go | func (s *fseState) init(br *bitReader, tableLog uint8, dt []decSymbol) {
s.dt = dt
br.fill()
s.state = dt[br.getBits(tableLog)]
} | [
"func",
"(",
"s",
"*",
"fseState",
")",
"init",
"(",
"br",
"*",
"bitReader",
",",
"tableLog",
"uint8",
",",
"dt",
"[",
"]",
"decSymbol",
")",
"{",
"s",
".",
"dt",
"=",
"dt",
"\n",
"br",
".",
"fill",
"(",
")",
"\n",
"s",
".",
"state",
"=",
"dt",
"[",
"br",
".",
"getBits",
"(",
"tableLog",
")",
"]",
"\n",
"}"
]
| // Initialize and decodeAsync first state and symbol. | [
"Initialize",
"and",
"decodeAsync",
"first",
"state",
"and",
"symbol",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zstd/fse_decoder.go#L297-L301 |
8,317 | klauspost/compress | zstd/fse_decoder.go | next | func (s *fseState) next(br *bitReader) {
lowBits := uint16(br.getBits(s.state.nbBits))
s.state = s.dt[s.state.newState+lowBits]
} | go | func (s *fseState) next(br *bitReader) {
lowBits := uint16(br.getBits(s.state.nbBits))
s.state = s.dt[s.state.newState+lowBits]
} | [
"func",
"(",
"s",
"*",
"fseState",
")",
"next",
"(",
"br",
"*",
"bitReader",
")",
"{",
"lowBits",
":=",
"uint16",
"(",
"br",
".",
"getBits",
"(",
"s",
".",
"state",
".",
"nbBits",
")",
")",
"\n",
"s",
".",
"state",
"=",
"s",
".",
"dt",
"[",
"s",
".",
"state",
".",
"newState",
"+",
"lowBits",
"]",
"\n",
"}"
]
| // next returns the current symbol and sets the next state.
// At least tablelog bits must be available in the bit reader. | [
"next",
"returns",
"the",
"current",
"symbol",
"and",
"sets",
"the",
"next",
"state",
".",
"At",
"least",
"tablelog",
"bits",
"must",
"be",
"available",
"in",
"the",
"bit",
"reader",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zstd/fse_decoder.go#L305-L308 |
8,318 | klauspost/compress | zstd/fse_decoder.go | final | func (s *fseState) final() (int, uint8) {
return int(s.state.baseline), s.state.addBits
} | go | func (s *fseState) final() (int, uint8) {
return int(s.state.baseline), s.state.addBits
} | [
"func",
"(",
"s",
"*",
"fseState",
")",
"final",
"(",
")",
"(",
"int",
",",
"uint8",
")",
"{",
"return",
"int",
"(",
"s",
".",
"state",
".",
"baseline",
")",
",",
"s",
".",
"state",
".",
"addBits",
"\n",
"}"
]
| // final returns the current state symbol without decoding the next. | [
"final",
"returns",
"the",
"current",
"state",
"symbol",
"without",
"decoding",
"the",
"next",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zstd/fse_decoder.go#L317-L319 |
8,319 | klauspost/compress | zlib/reader.go | NewReader | func NewReader(r io.Reader) (io.ReadCloser, error) {
return NewReaderDict(r, nil)
} | go | func NewReader(r io.Reader) (io.ReadCloser, error) {
return NewReaderDict(r, nil)
} | [
"func",
"NewReader",
"(",
"r",
"io",
".",
"Reader",
")",
"(",
"io",
".",
"ReadCloser",
",",
"error",
")",
"{",
"return",
"NewReaderDict",
"(",
"r",
",",
"nil",
")",
"\n",
"}"
]
| // NewReader creates a new ReadCloser.
// Reads from the returned ReadCloser read and decompress data from r.
// If r does not implement io.ByteReader, the decompressor may read more
// data than necessary from r.
// It is the caller's responsibility to call Close on the ReadCloser when done.
//
// The ReadCloser returned by NewReader also implements Resetter. | [
"NewReader",
"creates",
"a",
"new",
"ReadCloser",
".",
"Reads",
"from",
"the",
"returned",
"ReadCloser",
"read",
"and",
"decompress",
"data",
"from",
"r",
".",
"If",
"r",
"does",
"not",
"implement",
"io",
".",
"ByteReader",
"the",
"decompressor",
"may",
"read",
"more",
"data",
"than",
"necessary",
"from",
"r",
".",
"It",
"is",
"the",
"caller",
"s",
"responsibility",
"to",
"call",
"Close",
"on",
"the",
"ReadCloser",
"when",
"done",
".",
"The",
"ReadCloser",
"returned",
"by",
"NewReader",
"also",
"implements",
"Resetter",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zlib/reader.go#L71-L73 |
8,320 | klauspost/compress | zlib/reader.go | NewReaderDict | func NewReaderDict(r io.Reader, dict []byte) (io.ReadCloser, error) {
z := new(reader)
err := z.Reset(r, dict)
if err != nil {
return nil, err
}
return z, nil
} | go | func NewReaderDict(r io.Reader, dict []byte) (io.ReadCloser, error) {
z := new(reader)
err := z.Reset(r, dict)
if err != nil {
return nil, err
}
return z, nil
} | [
"func",
"NewReaderDict",
"(",
"r",
"io",
".",
"Reader",
",",
"dict",
"[",
"]",
"byte",
")",
"(",
"io",
".",
"ReadCloser",
",",
"error",
")",
"{",
"z",
":=",
"new",
"(",
"reader",
")",
"\n",
"err",
":=",
"z",
".",
"Reset",
"(",
"r",
",",
"dict",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"return",
"nil",
",",
"err",
"\n",
"}",
"\n",
"return",
"z",
",",
"nil",
"\n",
"}"
]
| // NewReaderDict is like NewReader but uses a preset dictionary.
// NewReaderDict ignores the dictionary if the compressed data does not refer to it.
// If the compressed data refers to a different dictionary, NewReaderDict returns ErrDictionary.
//
// The ReadCloser returned by NewReaderDict also implements Resetter. | [
"NewReaderDict",
"is",
"like",
"NewReader",
"but",
"uses",
"a",
"preset",
"dictionary",
".",
"NewReaderDict",
"ignores",
"the",
"dictionary",
"if",
"the",
"compressed",
"data",
"does",
"not",
"refer",
"to",
"it",
".",
"If",
"the",
"compressed",
"data",
"refers",
"to",
"a",
"different",
"dictionary",
"NewReaderDict",
"returns",
"ErrDictionary",
".",
"The",
"ReadCloser",
"returned",
"by",
"NewReaderDict",
"also",
"implements",
"Resetter",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zlib/reader.go#L80-L87 |
8,321 | klauspost/compress | zlib/reader.go | Close | func (z *reader) Close() error {
if z.err != nil && z.err != io.EOF {
return z.err
}
z.err = z.decompressor.Close()
return z.err
} | go | func (z *reader) Close() error {
if z.err != nil && z.err != io.EOF {
return z.err
}
z.err = z.decompressor.Close()
return z.err
} | [
"func",
"(",
"z",
"*",
"reader",
")",
"Close",
"(",
")",
"error",
"{",
"if",
"z",
".",
"err",
"!=",
"nil",
"&&",
"z",
".",
"err",
"!=",
"io",
".",
"EOF",
"{",
"return",
"z",
".",
"err",
"\n",
"}",
"\n",
"z",
".",
"err",
"=",
"z",
".",
"decompressor",
".",
"Close",
"(",
")",
"\n",
"return",
"z",
".",
"err",
"\n",
"}"
]
| // Calling Close does not close the wrapped io.Reader originally passed to NewReader.
// In order for the ZLIB checksum to be verified, the reader must be
// fully consumed until the io.EOF. | [
"Calling",
"Close",
"does",
"not",
"close",
"the",
"wrapped",
"io",
".",
"Reader",
"originally",
"passed",
"to",
"NewReader",
".",
"In",
"order",
"for",
"the",
"ZLIB",
"checksum",
"to",
"be",
"verified",
"the",
"reader",
"must",
"be",
"fully",
"consumed",
"until",
"the",
"io",
".",
"EOF",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zlib/reader.go#L122-L128 |
8,322 | klauspost/compress | zstd/history.go | reset | func (h *history) reset() {
h.b = h.b[:0]
h.error = false
h.recentOffsets = [3]int{1, 4, 8}
if f := h.decoders.litLengths.fse; f != nil && !f.preDefined {
fseDecoderPool.Put(f)
}
if f := h.decoders.offsets.fse; f != nil && !f.preDefined {
fseDecoderPool.Put(f)
}
if f := h.decoders.matchLengths.fse; f != nil && !f.preDefined {
fseDecoderPool.Put(f)
}
h.decoders = sequenceDecs{}
if h.huffTree != nil {
huffDecoderPool.Put(h.huffTree)
}
h.huffTree = nil
//printf("history created: %+v (l: %d, c: %d)", *h, len(h.b), cap(h.b))
} | go | func (h *history) reset() {
h.b = h.b[:0]
h.error = false
h.recentOffsets = [3]int{1, 4, 8}
if f := h.decoders.litLengths.fse; f != nil && !f.preDefined {
fseDecoderPool.Put(f)
}
if f := h.decoders.offsets.fse; f != nil && !f.preDefined {
fseDecoderPool.Put(f)
}
if f := h.decoders.matchLengths.fse; f != nil && !f.preDefined {
fseDecoderPool.Put(f)
}
h.decoders = sequenceDecs{}
if h.huffTree != nil {
huffDecoderPool.Put(h.huffTree)
}
h.huffTree = nil
//printf("history created: %+v (l: %d, c: %d)", *h, len(h.b), cap(h.b))
} | [
"func",
"(",
"h",
"*",
"history",
")",
"reset",
"(",
")",
"{",
"h",
".",
"b",
"=",
"h",
".",
"b",
"[",
":",
"0",
"]",
"\n",
"h",
".",
"error",
"=",
"false",
"\n",
"h",
".",
"recentOffsets",
"=",
"[",
"3",
"]",
"int",
"{",
"1",
",",
"4",
",",
"8",
"}",
"\n",
"if",
"f",
":=",
"h",
".",
"decoders",
".",
"litLengths",
".",
"fse",
";",
"f",
"!=",
"nil",
"&&",
"!",
"f",
".",
"preDefined",
"{",
"fseDecoderPool",
".",
"Put",
"(",
"f",
")",
"\n",
"}",
"\n",
"if",
"f",
":=",
"h",
".",
"decoders",
".",
"offsets",
".",
"fse",
";",
"f",
"!=",
"nil",
"&&",
"!",
"f",
".",
"preDefined",
"{",
"fseDecoderPool",
".",
"Put",
"(",
"f",
")",
"\n",
"}",
"\n",
"if",
"f",
":=",
"h",
".",
"decoders",
".",
"matchLengths",
".",
"fse",
";",
"f",
"!=",
"nil",
"&&",
"!",
"f",
".",
"preDefined",
"{",
"fseDecoderPool",
".",
"Put",
"(",
"f",
")",
"\n",
"}",
"\n",
"h",
".",
"decoders",
"=",
"sequenceDecs",
"{",
"}",
"\n",
"if",
"h",
".",
"huffTree",
"!=",
"nil",
"{",
"huffDecoderPool",
".",
"Put",
"(",
"h",
".",
"huffTree",
")",
"\n",
"}",
"\n",
"h",
".",
"huffTree",
"=",
"nil",
"\n",
"//printf(\"history created: %+v (l: %d, c: %d)\", *h, len(h.b), cap(h.b))",
"}"
]
| // reset will reset the history to initial state of a frame.
// The history must already have been initialized to the desired size. | [
"reset",
"will",
"reset",
"the",
"history",
"to",
"initial",
"state",
"of",
"a",
"frame",
".",
"The",
"history",
"must",
"already",
"have",
"been",
"initialized",
"to",
"the",
"desired",
"size",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zstd/history.go#L24-L43 |
8,323 | klauspost/compress | zstd/history.go | append | func (h *history) append(b []byte) {
if len(b) >= h.windowSize {
// Discard all history by simply overwriting
h.b = h.b[:h.windowSize]
copy(h.b, b[len(b)-h.windowSize:])
return
}
// If there is space, append it.
if len(b) < cap(h.b)-len(h.b) {
h.b = append(h.b, b...)
return
}
// Move data down so we only have window size left.
// We know we have less than window size in b at this point.
discard := len(b) + len(h.b) - h.windowSize
copy(h.b, h.b[discard:])
h.b = h.b[:h.windowSize]
copy(h.b[h.windowSize-len(b):], b)
} | go | func (h *history) append(b []byte) {
if len(b) >= h.windowSize {
// Discard all history by simply overwriting
h.b = h.b[:h.windowSize]
copy(h.b, b[len(b)-h.windowSize:])
return
}
// If there is space, append it.
if len(b) < cap(h.b)-len(h.b) {
h.b = append(h.b, b...)
return
}
// Move data down so we only have window size left.
// We know we have less than window size in b at this point.
discard := len(b) + len(h.b) - h.windowSize
copy(h.b, h.b[discard:])
h.b = h.b[:h.windowSize]
copy(h.b[h.windowSize-len(b):], b)
} | [
"func",
"(",
"h",
"*",
"history",
")",
"append",
"(",
"b",
"[",
"]",
"byte",
")",
"{",
"if",
"len",
"(",
"b",
")",
">=",
"h",
".",
"windowSize",
"{",
"// Discard all history by simply overwriting",
"h",
".",
"b",
"=",
"h",
".",
"b",
"[",
":",
"h",
".",
"windowSize",
"]",
"\n",
"copy",
"(",
"h",
".",
"b",
",",
"b",
"[",
"len",
"(",
"b",
")",
"-",
"h",
".",
"windowSize",
":",
"]",
")",
"\n",
"return",
"\n",
"}",
"\n\n",
"// If there is space, append it.",
"if",
"len",
"(",
"b",
")",
"<",
"cap",
"(",
"h",
".",
"b",
")",
"-",
"len",
"(",
"h",
".",
"b",
")",
"{",
"h",
".",
"b",
"=",
"append",
"(",
"h",
".",
"b",
",",
"b",
"...",
")",
"\n",
"return",
"\n",
"}",
"\n\n",
"// Move data down so we only have window size left.",
"// We know we have less than window size in b at this point.",
"discard",
":=",
"len",
"(",
"b",
")",
"+",
"len",
"(",
"h",
".",
"b",
")",
"-",
"h",
".",
"windowSize",
"\n",
"copy",
"(",
"h",
".",
"b",
",",
"h",
".",
"b",
"[",
"discard",
":",
"]",
")",
"\n",
"h",
".",
"b",
"=",
"h",
".",
"b",
"[",
":",
"h",
".",
"windowSize",
"]",
"\n",
"copy",
"(",
"h",
".",
"b",
"[",
"h",
".",
"windowSize",
"-",
"len",
"(",
"b",
")",
":",
"]",
",",
"b",
")",
"\n",
"}"
]
| // append bytes to history.
// This function will make sure there is space for it,
// if the buffer has been allocated with enough extra space. | [
"append",
"bytes",
"to",
"history",
".",
"This",
"function",
"will",
"make",
"sure",
"there",
"is",
"space",
"for",
"it",
"if",
"the",
"buffer",
"has",
"been",
"allocated",
"with",
"enough",
"extra",
"space",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zstd/history.go#L48-L68 |
8,324 | klauspost/compress | zstd/history.go | appendKeep | func (h *history) appendKeep(b []byte) {
h.b = append(h.b, b...)
} | go | func (h *history) appendKeep(b []byte) {
h.b = append(h.b, b...)
} | [
"func",
"(",
"h",
"*",
"history",
")",
"appendKeep",
"(",
"b",
"[",
"]",
"byte",
")",
"{",
"h",
".",
"b",
"=",
"append",
"(",
"h",
".",
"b",
",",
"b",
"...",
")",
"\n",
"}"
]
| // append bytes to history without ever discarding anything. | [
"append",
"bytes",
"to",
"history",
"without",
"ever",
"discarding",
"anything",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zstd/history.go#L71-L73 |
8,325 | klauspost/compress | huff0/decompress.go | Decompress1X | func (s *Scratch) Decompress1X(in []byte) (out []byte, err error) {
if len(s.dt.single) == 0 {
return nil, errors.New("no table loaded")
}
var br bitReader
err = br.init(in)
if err != nil {
return nil, err
}
s.Out = s.Out[:0]
decode := func() byte {
val := br.peekBitsFast(s.actualTableLog) /* note : actualTableLog >= 1 */
v := s.dt.single[val]
br.bitsRead += v.nBits
return v.byte
}
hasDec := func(v dEntrySingle) byte {
br.bitsRead += v.nBits
return v.byte
}
// Avoid bounds check by always having full sized table.
const tlSize = 1 << tableLogMax
const tlMask = tlSize - 1
dt := s.dt.single[:tlSize]
// Use temp table to avoid bound checks/append penalty.
var tmp = s.huffWeight[:256]
var off uint8
for br.off >= 8 {
br.fillFast()
tmp[off+0] = hasDec(dt[br.peekBitsFast(s.actualTableLog)&tlMask])
tmp[off+1] = hasDec(dt[br.peekBitsFast(s.actualTableLog)&tlMask])
br.fillFast()
tmp[off+2] = hasDec(dt[br.peekBitsFast(s.actualTableLog)&tlMask])
tmp[off+3] = hasDec(dt[br.peekBitsFast(s.actualTableLog)&tlMask])
off += 4
if off == 0 {
s.Out = append(s.Out, tmp...)
}
}
s.Out = append(s.Out, tmp[:off]...)
for !br.finished() {
br.fill()
s.Out = append(s.Out, decode())
}
return s.Out, br.close()
} | go | func (s *Scratch) Decompress1X(in []byte) (out []byte, err error) {
if len(s.dt.single) == 0 {
return nil, errors.New("no table loaded")
}
var br bitReader
err = br.init(in)
if err != nil {
return nil, err
}
s.Out = s.Out[:0]
decode := func() byte {
val := br.peekBitsFast(s.actualTableLog) /* note : actualTableLog >= 1 */
v := s.dt.single[val]
br.bitsRead += v.nBits
return v.byte
}
hasDec := func(v dEntrySingle) byte {
br.bitsRead += v.nBits
return v.byte
}
// Avoid bounds check by always having full sized table.
const tlSize = 1 << tableLogMax
const tlMask = tlSize - 1
dt := s.dt.single[:tlSize]
// Use temp table to avoid bound checks/append penalty.
var tmp = s.huffWeight[:256]
var off uint8
for br.off >= 8 {
br.fillFast()
tmp[off+0] = hasDec(dt[br.peekBitsFast(s.actualTableLog)&tlMask])
tmp[off+1] = hasDec(dt[br.peekBitsFast(s.actualTableLog)&tlMask])
br.fillFast()
tmp[off+2] = hasDec(dt[br.peekBitsFast(s.actualTableLog)&tlMask])
tmp[off+3] = hasDec(dt[br.peekBitsFast(s.actualTableLog)&tlMask])
off += 4
if off == 0 {
s.Out = append(s.Out, tmp...)
}
}
s.Out = append(s.Out, tmp[:off]...)
for !br.finished() {
br.fill()
s.Out = append(s.Out, decode())
}
return s.Out, br.close()
} | [
"func",
"(",
"s",
"*",
"Scratch",
")",
"Decompress1X",
"(",
"in",
"[",
"]",
"byte",
")",
"(",
"out",
"[",
"]",
"byte",
",",
"err",
"error",
")",
"{",
"if",
"len",
"(",
"s",
".",
"dt",
".",
"single",
")",
"==",
"0",
"{",
"return",
"nil",
",",
"errors",
".",
"New",
"(",
"\"",
"\"",
")",
"\n",
"}",
"\n",
"var",
"br",
"bitReader",
"\n",
"err",
"=",
"br",
".",
"init",
"(",
"in",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"return",
"nil",
",",
"err",
"\n",
"}",
"\n",
"s",
".",
"Out",
"=",
"s",
".",
"Out",
"[",
":",
"0",
"]",
"\n\n",
"decode",
":=",
"func",
"(",
")",
"byte",
"{",
"val",
":=",
"br",
".",
"peekBitsFast",
"(",
"s",
".",
"actualTableLog",
")",
"/* note : actualTableLog >= 1 */",
"\n",
"v",
":=",
"s",
".",
"dt",
".",
"single",
"[",
"val",
"]",
"\n",
"br",
".",
"bitsRead",
"+=",
"v",
".",
"nBits",
"\n",
"return",
"v",
".",
"byte",
"\n",
"}",
"\n",
"hasDec",
":=",
"func",
"(",
"v",
"dEntrySingle",
")",
"byte",
"{",
"br",
".",
"bitsRead",
"+=",
"v",
".",
"nBits",
"\n",
"return",
"v",
".",
"byte",
"\n",
"}",
"\n\n",
"// Avoid bounds check by always having full sized table.",
"const",
"tlSize",
"=",
"1",
"<<",
"tableLogMax",
"\n",
"const",
"tlMask",
"=",
"tlSize",
"-",
"1",
"\n",
"dt",
":=",
"s",
".",
"dt",
".",
"single",
"[",
":",
"tlSize",
"]",
"\n\n",
"// Use temp table to avoid bound checks/append penalty.",
"var",
"tmp",
"=",
"s",
".",
"huffWeight",
"[",
":",
"256",
"]",
"\n",
"var",
"off",
"uint8",
"\n\n",
"for",
"br",
".",
"off",
">=",
"8",
"{",
"br",
".",
"fillFast",
"(",
")",
"\n",
"tmp",
"[",
"off",
"+",
"0",
"]",
"=",
"hasDec",
"(",
"dt",
"[",
"br",
".",
"peekBitsFast",
"(",
"s",
".",
"actualTableLog",
")",
"&",
"tlMask",
"]",
")",
"\n",
"tmp",
"[",
"off",
"+",
"1",
"]",
"=",
"hasDec",
"(",
"dt",
"[",
"br",
".",
"peekBitsFast",
"(",
"s",
".",
"actualTableLog",
")",
"&",
"tlMask",
"]",
")",
"\n",
"br",
".",
"fillFast",
"(",
")",
"\n",
"tmp",
"[",
"off",
"+",
"2",
"]",
"=",
"hasDec",
"(",
"dt",
"[",
"br",
".",
"peekBitsFast",
"(",
"s",
".",
"actualTableLog",
")",
"&",
"tlMask",
"]",
")",
"\n",
"tmp",
"[",
"off",
"+",
"3",
"]",
"=",
"hasDec",
"(",
"dt",
"[",
"br",
".",
"peekBitsFast",
"(",
"s",
".",
"actualTableLog",
")",
"&",
"tlMask",
"]",
")",
"\n",
"off",
"+=",
"4",
"\n",
"if",
"off",
"==",
"0",
"{",
"s",
".",
"Out",
"=",
"append",
"(",
"s",
".",
"Out",
",",
"tmp",
"...",
")",
"\n",
"}",
"\n",
"}",
"\n\n",
"s",
".",
"Out",
"=",
"append",
"(",
"s",
".",
"Out",
",",
"tmp",
"[",
":",
"off",
"]",
"...",
")",
"\n\n",
"for",
"!",
"br",
".",
"finished",
"(",
")",
"{",
"br",
".",
"fill",
"(",
")",
"\n",
"s",
".",
"Out",
"=",
"append",
"(",
"s",
".",
"Out",
",",
"decode",
"(",
")",
")",
"\n",
"}",
"\n",
"return",
"s",
".",
"Out",
",",
"br",
".",
"close",
"(",
")",
"\n",
"}"
]
| // Decompress1X will decompress a 1X encoded stream.
// The length of the supplied input must match the end of a block exactly.
// Before this is called, the table must be initialized with ReadTable unless
// the encoder re-used the table. | [
"Decompress1X",
"will",
"decompress",
"a",
"1X",
"encoded",
"stream",
".",
"The",
"length",
"of",
"the",
"supplied",
"input",
"must",
"match",
"the",
"end",
"of",
"a",
"block",
"exactly",
".",
"Before",
"this",
"is",
"called",
"the",
"table",
"must",
"be",
"initialized",
"with",
"ReadTable",
"unless",
"the",
"encoder",
"re",
"-",
"used",
"the",
"table",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/huff0/decompress.go#L156-L207 |
8,326 | klauspost/compress | huff0/decompress.go | matches | func (s *Scratch) matches(ct cTable, w io.Writer) {
if s == nil || len(s.dt.single) == 0 {
return
}
dt := s.dt.single[:1<<s.actualTableLog]
tablelog := s.actualTableLog
ok := 0
broken := 0
for sym, enc := range ct {
errs := 0
broken++
if enc.nBits == 0 {
for _, dec := range dt {
if dec.byte == byte(sym) {
fmt.Fprintf(w, "symbol %x has decoder, but no encoder\n", sym)
errs++
break
}
}
if errs == 0 {
broken--
}
continue
}
// Unused bits in input
ub := tablelog - enc.nBits
top := enc.val << ub
// decoder looks at top bits.
dec := dt[top]
if dec.nBits != enc.nBits {
fmt.Fprintf(w, "symbol 0x%x bit size mismatch (enc: %d, dec:%d).\n", sym, enc.nBits, dec.nBits)
errs++
}
if dec.byte != uint8(sym) {
fmt.Fprintf(w, "symbol 0x%x decoder output mismatch (enc: %d, dec:%d).\n", sym, sym, dec.byte)
errs++
}
if errs > 0 {
fmt.Fprintf(w, "%d errros in base, stopping\n", errs)
continue
}
// Ensure that all combinations are covered.
for i := uint16(0); i < (1 << ub); i++ {
vval := top | i
dec := dt[vval]
if dec.nBits != enc.nBits {
fmt.Fprintf(w, "symbol 0x%x bit size mismatch (enc: %d, dec:%d).\n", vval, enc.nBits, dec.nBits)
errs++
}
if dec.byte != uint8(sym) {
fmt.Fprintf(w, "symbol 0x%x decoder output mismatch (enc: %d, dec:%d).\n", vval, sym, dec.byte)
errs++
}
if errs > 20 {
fmt.Fprintf(w, "%d errros, stopping\n", errs)
break
}
}
if errs == 0 {
ok++
broken--
}
}
if broken > 0 {
fmt.Fprintf(w, "%d broken, %d ok\n", broken, ok)
}
} | go | func (s *Scratch) matches(ct cTable, w io.Writer) {
if s == nil || len(s.dt.single) == 0 {
return
}
dt := s.dt.single[:1<<s.actualTableLog]
tablelog := s.actualTableLog
ok := 0
broken := 0
for sym, enc := range ct {
errs := 0
broken++
if enc.nBits == 0 {
for _, dec := range dt {
if dec.byte == byte(sym) {
fmt.Fprintf(w, "symbol %x has decoder, but no encoder\n", sym)
errs++
break
}
}
if errs == 0 {
broken--
}
continue
}
// Unused bits in input
ub := tablelog - enc.nBits
top := enc.val << ub
// decoder looks at top bits.
dec := dt[top]
if dec.nBits != enc.nBits {
fmt.Fprintf(w, "symbol 0x%x bit size mismatch (enc: %d, dec:%d).\n", sym, enc.nBits, dec.nBits)
errs++
}
if dec.byte != uint8(sym) {
fmt.Fprintf(w, "symbol 0x%x decoder output mismatch (enc: %d, dec:%d).\n", sym, sym, dec.byte)
errs++
}
if errs > 0 {
fmt.Fprintf(w, "%d errros in base, stopping\n", errs)
continue
}
// Ensure that all combinations are covered.
for i := uint16(0); i < (1 << ub); i++ {
vval := top | i
dec := dt[vval]
if dec.nBits != enc.nBits {
fmt.Fprintf(w, "symbol 0x%x bit size mismatch (enc: %d, dec:%d).\n", vval, enc.nBits, dec.nBits)
errs++
}
if dec.byte != uint8(sym) {
fmt.Fprintf(w, "symbol 0x%x decoder output mismatch (enc: %d, dec:%d).\n", vval, sym, dec.byte)
errs++
}
if errs > 20 {
fmt.Fprintf(w, "%d errros, stopping\n", errs)
break
}
}
if errs == 0 {
ok++
broken--
}
}
if broken > 0 {
fmt.Fprintf(w, "%d broken, %d ok\n", broken, ok)
}
} | [
"func",
"(",
"s",
"*",
"Scratch",
")",
"matches",
"(",
"ct",
"cTable",
",",
"w",
"io",
".",
"Writer",
")",
"{",
"if",
"s",
"==",
"nil",
"||",
"len",
"(",
"s",
".",
"dt",
".",
"single",
")",
"==",
"0",
"{",
"return",
"\n",
"}",
"\n",
"dt",
":=",
"s",
".",
"dt",
".",
"single",
"[",
":",
"1",
"<<",
"s",
".",
"actualTableLog",
"]",
"\n",
"tablelog",
":=",
"s",
".",
"actualTableLog",
"\n",
"ok",
":=",
"0",
"\n",
"broken",
":=",
"0",
"\n",
"for",
"sym",
",",
"enc",
":=",
"range",
"ct",
"{",
"errs",
":=",
"0",
"\n",
"broken",
"++",
"\n",
"if",
"enc",
".",
"nBits",
"==",
"0",
"{",
"for",
"_",
",",
"dec",
":=",
"range",
"dt",
"{",
"if",
"dec",
".",
"byte",
"==",
"byte",
"(",
"sym",
")",
"{",
"fmt",
".",
"Fprintf",
"(",
"w",
",",
"\"",
"\\n",
"\"",
",",
"sym",
")",
"\n",
"errs",
"++",
"\n",
"break",
"\n",
"}",
"\n",
"}",
"\n",
"if",
"errs",
"==",
"0",
"{",
"broken",
"--",
"\n",
"}",
"\n",
"continue",
"\n",
"}",
"\n",
"// Unused bits in input",
"ub",
":=",
"tablelog",
"-",
"enc",
".",
"nBits",
"\n",
"top",
":=",
"enc",
".",
"val",
"<<",
"ub",
"\n",
"// decoder looks at top bits.",
"dec",
":=",
"dt",
"[",
"top",
"]",
"\n",
"if",
"dec",
".",
"nBits",
"!=",
"enc",
".",
"nBits",
"{",
"fmt",
".",
"Fprintf",
"(",
"w",
",",
"\"",
"\\n",
"\"",
",",
"sym",
",",
"enc",
".",
"nBits",
",",
"dec",
".",
"nBits",
")",
"\n",
"errs",
"++",
"\n",
"}",
"\n",
"if",
"dec",
".",
"byte",
"!=",
"uint8",
"(",
"sym",
")",
"{",
"fmt",
".",
"Fprintf",
"(",
"w",
",",
"\"",
"\\n",
"\"",
",",
"sym",
",",
"sym",
",",
"dec",
".",
"byte",
")",
"\n",
"errs",
"++",
"\n",
"}",
"\n",
"if",
"errs",
">",
"0",
"{",
"fmt",
".",
"Fprintf",
"(",
"w",
",",
"\"",
"\\n",
"\"",
",",
"errs",
")",
"\n",
"continue",
"\n",
"}",
"\n",
"// Ensure that all combinations are covered.",
"for",
"i",
":=",
"uint16",
"(",
"0",
")",
";",
"i",
"<",
"(",
"1",
"<<",
"ub",
")",
";",
"i",
"++",
"{",
"vval",
":=",
"top",
"|",
"i",
"\n",
"dec",
":=",
"dt",
"[",
"vval",
"]",
"\n",
"if",
"dec",
".",
"nBits",
"!=",
"enc",
".",
"nBits",
"{",
"fmt",
".",
"Fprintf",
"(",
"w",
",",
"\"",
"\\n",
"\"",
",",
"vval",
",",
"enc",
".",
"nBits",
",",
"dec",
".",
"nBits",
")",
"\n",
"errs",
"++",
"\n",
"}",
"\n",
"if",
"dec",
".",
"byte",
"!=",
"uint8",
"(",
"sym",
")",
"{",
"fmt",
".",
"Fprintf",
"(",
"w",
",",
"\"",
"\\n",
"\"",
",",
"vval",
",",
"sym",
",",
"dec",
".",
"byte",
")",
"\n",
"errs",
"++",
"\n",
"}",
"\n",
"if",
"errs",
">",
"20",
"{",
"fmt",
".",
"Fprintf",
"(",
"w",
",",
"\"",
"\\n",
"\"",
",",
"errs",
")",
"\n",
"break",
"\n",
"}",
"\n",
"}",
"\n",
"if",
"errs",
"==",
"0",
"{",
"ok",
"++",
"\n",
"broken",
"--",
"\n",
"}",
"\n",
"}",
"\n",
"if",
"broken",
">",
"0",
"{",
"fmt",
".",
"Fprintf",
"(",
"w",
",",
"\"",
"\\n",
"\"",
",",
"broken",
",",
"ok",
")",
"\n",
"}",
"\n",
"}"
]
| // matches will compare a decoding table to a coding table.
// Errors are written to the writer.
// Nothing will be written if table is ok. | [
"matches",
"will",
"compare",
"a",
"decoding",
"table",
"to",
"a",
"coding",
"table",
".",
"Errors",
"are",
"written",
"to",
"the",
"writer",
".",
"Nothing",
"will",
"be",
"written",
"if",
"table",
"is",
"ok",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/huff0/decompress.go#L331-L397 |
8,327 | klauspost/compress | fse/decompress.go | Decompress | func Decompress(b []byte, s *Scratch) ([]byte, error) {
s, err := s.prepare(b)
if err != nil {
return nil, err
}
s.Out = s.Out[:0]
err = s.readNCount()
if err != nil {
return nil, err
}
err = s.buildDtable()
if err != nil {
return nil, err
}
err = s.decompress()
if err != nil {
return nil, err
}
return s.Out, nil
} | go | func Decompress(b []byte, s *Scratch) ([]byte, error) {
s, err := s.prepare(b)
if err != nil {
return nil, err
}
s.Out = s.Out[:0]
err = s.readNCount()
if err != nil {
return nil, err
}
err = s.buildDtable()
if err != nil {
return nil, err
}
err = s.decompress()
if err != nil {
return nil, err
}
return s.Out, nil
} | [
"func",
"Decompress",
"(",
"b",
"[",
"]",
"byte",
",",
"s",
"*",
"Scratch",
")",
"(",
"[",
"]",
"byte",
",",
"error",
")",
"{",
"s",
",",
"err",
":=",
"s",
".",
"prepare",
"(",
"b",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"return",
"nil",
",",
"err",
"\n",
"}",
"\n",
"s",
".",
"Out",
"=",
"s",
".",
"Out",
"[",
":",
"0",
"]",
"\n",
"err",
"=",
"s",
".",
"readNCount",
"(",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"return",
"nil",
",",
"err",
"\n",
"}",
"\n",
"err",
"=",
"s",
".",
"buildDtable",
"(",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"return",
"nil",
",",
"err",
"\n",
"}",
"\n",
"err",
"=",
"s",
".",
"decompress",
"(",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"return",
"nil",
",",
"err",
"\n",
"}",
"\n\n",
"return",
"s",
".",
"Out",
",",
"nil",
"\n",
"}"
]
| // Decompress a block of data.
// You can provide a scratch buffer to avoid allocations.
// If nil is provided a temporary one will be allocated.
// It is possible, but by no way guaranteed that corrupt data will
// return an error.
// It is up to the caller to verify integrity of the returned data.
// Use a predefined Scrach to set maximum acceptable output size. | [
"Decompress",
"a",
"block",
"of",
"data",
".",
"You",
"can",
"provide",
"a",
"scratch",
"buffer",
"to",
"avoid",
"allocations",
".",
"If",
"nil",
"is",
"provided",
"a",
"temporary",
"one",
"will",
"be",
"allocated",
".",
"It",
"is",
"possible",
"but",
"by",
"no",
"way",
"guaranteed",
"that",
"corrupt",
"data",
"will",
"return",
"an",
"error",
".",
"It",
"is",
"up",
"to",
"the",
"caller",
"to",
"verify",
"integrity",
"of",
"the",
"returned",
"data",
".",
"Use",
"a",
"predefined",
"Scrach",
"to",
"set",
"maximum",
"acceptable",
"output",
"size",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/fse/decompress.go#L19-L39 |
8,328 | klauspost/compress | fse/decompress.go | allocDtable | func (s *Scratch) allocDtable() {
tableSize := 1 << s.actualTableLog
if cap(s.decTable) < int(tableSize) {
s.decTable = make([]decSymbol, tableSize)
}
s.decTable = s.decTable[:tableSize]
if cap(s.ct.tableSymbol) < 256 {
s.ct.tableSymbol = make([]byte, 256)
}
s.ct.tableSymbol = s.ct.tableSymbol[:256]
if cap(s.ct.stateTable) < 256 {
s.ct.stateTable = make([]uint16, 256)
}
s.ct.stateTable = s.ct.stateTable[:256]
} | go | func (s *Scratch) allocDtable() {
tableSize := 1 << s.actualTableLog
if cap(s.decTable) < int(tableSize) {
s.decTable = make([]decSymbol, tableSize)
}
s.decTable = s.decTable[:tableSize]
if cap(s.ct.tableSymbol) < 256 {
s.ct.tableSymbol = make([]byte, 256)
}
s.ct.tableSymbol = s.ct.tableSymbol[:256]
if cap(s.ct.stateTable) < 256 {
s.ct.stateTable = make([]uint16, 256)
}
s.ct.stateTable = s.ct.stateTable[:256]
} | [
"func",
"(",
"s",
"*",
"Scratch",
")",
"allocDtable",
"(",
")",
"{",
"tableSize",
":=",
"1",
"<<",
"s",
".",
"actualTableLog",
"\n",
"if",
"cap",
"(",
"s",
".",
"decTable",
")",
"<",
"int",
"(",
"tableSize",
")",
"{",
"s",
".",
"decTable",
"=",
"make",
"(",
"[",
"]",
"decSymbol",
",",
"tableSize",
")",
"\n",
"}",
"\n",
"s",
".",
"decTable",
"=",
"s",
".",
"decTable",
"[",
":",
"tableSize",
"]",
"\n\n",
"if",
"cap",
"(",
"s",
".",
"ct",
".",
"tableSymbol",
")",
"<",
"256",
"{",
"s",
".",
"ct",
".",
"tableSymbol",
"=",
"make",
"(",
"[",
"]",
"byte",
",",
"256",
")",
"\n",
"}",
"\n",
"s",
".",
"ct",
".",
"tableSymbol",
"=",
"s",
".",
"ct",
".",
"tableSymbol",
"[",
":",
"256",
"]",
"\n\n",
"if",
"cap",
"(",
"s",
".",
"ct",
".",
"stateTable",
")",
"<",
"256",
"{",
"s",
".",
"ct",
".",
"stateTable",
"=",
"make",
"(",
"[",
"]",
"uint16",
",",
"256",
")",
"\n",
"}",
"\n",
"s",
".",
"ct",
".",
"stateTable",
"=",
"s",
".",
"ct",
".",
"stateTable",
"[",
":",
"256",
"]",
"\n",
"}"
]
| // allocDtable will allocate decoding tables if they are not big enough. | [
"allocDtable",
"will",
"allocate",
"decoding",
"tables",
"if",
"they",
"are",
"not",
"big",
"enough",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/fse/decompress.go#L173-L189 |
8,329 | klauspost/compress | fse/decompress.go | decompress | func (s *Scratch) decompress() error {
br := &s.bits
br.init(s.br.unread())
var s1, s2 decoder
// Initialize and decode first state and symbol.
s1.init(br, s.decTable, s.actualTableLog)
s2.init(br, s.decTable, s.actualTableLog)
// Use temp table to avoid bound checks/append penalty.
var tmp = s.ct.tableSymbol[:256]
var off uint8
// Main part
if !s.zeroBits {
for br.off >= 8 {
br.fillFast()
tmp[off+0] = s1.nextFast()
tmp[off+1] = s2.nextFast()
br.fillFast()
tmp[off+2] = s1.nextFast()
tmp[off+3] = s2.nextFast()
off += 4
if off == 0 {
s.Out = append(s.Out, tmp...)
}
}
} else {
for br.off >= 8 {
br.fillFast()
tmp[off+0] = s1.next()
tmp[off+1] = s2.next()
br.fillFast()
tmp[off+2] = s1.next()
tmp[off+3] = s2.next()
off += 4
if off == 0 {
s.Out = append(s.Out, tmp...)
off = 0
if len(s.Out) >= s.DecompressLimit {
return fmt.Errorf("output size (%d) > DecompressLimit (%d)", len(s.Out), s.DecompressLimit)
}
}
}
}
s.Out = append(s.Out, tmp[:off]...)
// Final bits, a bit more expensive check
for {
if s1.finished() {
s.Out = append(s.Out, s1.final(), s2.final())
break
}
br.fill()
s.Out = append(s.Out, s1.next())
if s2.finished() {
s.Out = append(s.Out, s2.final(), s1.final())
break
}
s.Out = append(s.Out, s2.next())
if len(s.Out) >= s.DecompressLimit {
return fmt.Errorf("output size (%d) > DecompressLimit (%d)", len(s.Out), s.DecompressLimit)
}
}
return br.close()
} | go | func (s *Scratch) decompress() error {
br := &s.bits
br.init(s.br.unread())
var s1, s2 decoder
// Initialize and decode first state and symbol.
s1.init(br, s.decTable, s.actualTableLog)
s2.init(br, s.decTable, s.actualTableLog)
// Use temp table to avoid bound checks/append penalty.
var tmp = s.ct.tableSymbol[:256]
var off uint8
// Main part
if !s.zeroBits {
for br.off >= 8 {
br.fillFast()
tmp[off+0] = s1.nextFast()
tmp[off+1] = s2.nextFast()
br.fillFast()
tmp[off+2] = s1.nextFast()
tmp[off+3] = s2.nextFast()
off += 4
if off == 0 {
s.Out = append(s.Out, tmp...)
}
}
} else {
for br.off >= 8 {
br.fillFast()
tmp[off+0] = s1.next()
tmp[off+1] = s2.next()
br.fillFast()
tmp[off+2] = s1.next()
tmp[off+3] = s2.next()
off += 4
if off == 0 {
s.Out = append(s.Out, tmp...)
off = 0
if len(s.Out) >= s.DecompressLimit {
return fmt.Errorf("output size (%d) > DecompressLimit (%d)", len(s.Out), s.DecompressLimit)
}
}
}
}
s.Out = append(s.Out, tmp[:off]...)
// Final bits, a bit more expensive check
for {
if s1.finished() {
s.Out = append(s.Out, s1.final(), s2.final())
break
}
br.fill()
s.Out = append(s.Out, s1.next())
if s2.finished() {
s.Out = append(s.Out, s2.final(), s1.final())
break
}
s.Out = append(s.Out, s2.next())
if len(s.Out) >= s.DecompressLimit {
return fmt.Errorf("output size (%d) > DecompressLimit (%d)", len(s.Out), s.DecompressLimit)
}
}
return br.close()
} | [
"func",
"(",
"s",
"*",
"Scratch",
")",
"decompress",
"(",
")",
"error",
"{",
"br",
":=",
"&",
"s",
".",
"bits",
"\n",
"br",
".",
"init",
"(",
"s",
".",
"br",
".",
"unread",
"(",
")",
")",
"\n\n",
"var",
"s1",
",",
"s2",
"decoder",
"\n",
"// Initialize and decode first state and symbol.",
"s1",
".",
"init",
"(",
"br",
",",
"s",
".",
"decTable",
",",
"s",
".",
"actualTableLog",
")",
"\n",
"s2",
".",
"init",
"(",
"br",
",",
"s",
".",
"decTable",
",",
"s",
".",
"actualTableLog",
")",
"\n\n",
"// Use temp table to avoid bound checks/append penalty.",
"var",
"tmp",
"=",
"s",
".",
"ct",
".",
"tableSymbol",
"[",
":",
"256",
"]",
"\n",
"var",
"off",
"uint8",
"\n\n",
"// Main part",
"if",
"!",
"s",
".",
"zeroBits",
"{",
"for",
"br",
".",
"off",
">=",
"8",
"{",
"br",
".",
"fillFast",
"(",
")",
"\n",
"tmp",
"[",
"off",
"+",
"0",
"]",
"=",
"s1",
".",
"nextFast",
"(",
")",
"\n",
"tmp",
"[",
"off",
"+",
"1",
"]",
"=",
"s2",
".",
"nextFast",
"(",
")",
"\n",
"br",
".",
"fillFast",
"(",
")",
"\n",
"tmp",
"[",
"off",
"+",
"2",
"]",
"=",
"s1",
".",
"nextFast",
"(",
")",
"\n",
"tmp",
"[",
"off",
"+",
"3",
"]",
"=",
"s2",
".",
"nextFast",
"(",
")",
"\n",
"off",
"+=",
"4",
"\n",
"if",
"off",
"==",
"0",
"{",
"s",
".",
"Out",
"=",
"append",
"(",
"s",
".",
"Out",
",",
"tmp",
"...",
")",
"\n",
"}",
"\n",
"}",
"\n",
"}",
"else",
"{",
"for",
"br",
".",
"off",
">=",
"8",
"{",
"br",
".",
"fillFast",
"(",
")",
"\n",
"tmp",
"[",
"off",
"+",
"0",
"]",
"=",
"s1",
".",
"next",
"(",
")",
"\n",
"tmp",
"[",
"off",
"+",
"1",
"]",
"=",
"s2",
".",
"next",
"(",
")",
"\n",
"br",
".",
"fillFast",
"(",
")",
"\n",
"tmp",
"[",
"off",
"+",
"2",
"]",
"=",
"s1",
".",
"next",
"(",
")",
"\n",
"tmp",
"[",
"off",
"+",
"3",
"]",
"=",
"s2",
".",
"next",
"(",
")",
"\n",
"off",
"+=",
"4",
"\n",
"if",
"off",
"==",
"0",
"{",
"s",
".",
"Out",
"=",
"append",
"(",
"s",
".",
"Out",
",",
"tmp",
"...",
")",
"\n",
"off",
"=",
"0",
"\n",
"if",
"len",
"(",
"s",
".",
"Out",
")",
">=",
"s",
".",
"DecompressLimit",
"{",
"return",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"len",
"(",
"s",
".",
"Out",
")",
",",
"s",
".",
"DecompressLimit",
")",
"\n",
"}",
"\n",
"}",
"\n",
"}",
"\n",
"}",
"\n",
"s",
".",
"Out",
"=",
"append",
"(",
"s",
".",
"Out",
",",
"tmp",
"[",
":",
"off",
"]",
"...",
")",
"\n\n",
"// Final bits, a bit more expensive check",
"for",
"{",
"if",
"s1",
".",
"finished",
"(",
")",
"{",
"s",
".",
"Out",
"=",
"append",
"(",
"s",
".",
"Out",
",",
"s1",
".",
"final",
"(",
")",
",",
"s2",
".",
"final",
"(",
")",
")",
"\n",
"break",
"\n",
"}",
"\n",
"br",
".",
"fill",
"(",
")",
"\n",
"s",
".",
"Out",
"=",
"append",
"(",
"s",
".",
"Out",
",",
"s1",
".",
"next",
"(",
")",
")",
"\n",
"if",
"s2",
".",
"finished",
"(",
")",
"{",
"s",
".",
"Out",
"=",
"append",
"(",
"s",
".",
"Out",
",",
"s2",
".",
"final",
"(",
")",
",",
"s1",
".",
"final",
"(",
")",
")",
"\n",
"break",
"\n",
"}",
"\n",
"s",
".",
"Out",
"=",
"append",
"(",
"s",
".",
"Out",
",",
"s2",
".",
"next",
"(",
")",
")",
"\n",
"if",
"len",
"(",
"s",
".",
"Out",
")",
">=",
"s",
".",
"DecompressLimit",
"{",
"return",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"len",
"(",
"s",
".",
"Out",
")",
",",
"s",
".",
"DecompressLimit",
")",
"\n",
"}",
"\n",
"}",
"\n",
"return",
"br",
".",
"close",
"(",
")",
"\n",
"}"
]
| // decompress will decompress the bitstream.
// If the buffer is over-read an error is returned. | [
"decompress",
"will",
"decompress",
"the",
"bitstream",
".",
"If",
"the",
"buffer",
"is",
"over",
"-",
"read",
"an",
"error",
"is",
"returned",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/fse/decompress.go#L261-L326 |
8,330 | klauspost/compress | fse/decompress.go | init | func (d *decoder) init(in *bitReader, dt []decSymbol, tableLog uint8) {
d.dt = dt
d.br = in
d.state = uint16(in.getBits(tableLog))
} | go | func (d *decoder) init(in *bitReader, dt []decSymbol, tableLog uint8) {
d.dt = dt
d.br = in
d.state = uint16(in.getBits(tableLog))
} | [
"func",
"(",
"d",
"*",
"decoder",
")",
"init",
"(",
"in",
"*",
"bitReader",
",",
"dt",
"[",
"]",
"decSymbol",
",",
"tableLog",
"uint8",
")",
"{",
"d",
".",
"dt",
"=",
"dt",
"\n",
"d",
".",
"br",
"=",
"in",
"\n",
"d",
".",
"state",
"=",
"uint16",
"(",
"in",
".",
"getBits",
"(",
"tableLog",
")",
")",
"\n",
"}"
]
| // init will initialize the decoder and read the first state from the stream. | [
"init",
"will",
"initialize",
"the",
"decoder",
"and",
"read",
"the",
"first",
"state",
"from",
"the",
"stream",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/fse/decompress.go#L336-L340 |
8,331 | klauspost/compress | fse/decompress.go | next | func (d *decoder) next() uint8 {
n := &d.dt[d.state]
lowBits := d.br.getBits(n.nbBits)
d.state = n.newState + lowBits
return n.symbol
} | go | func (d *decoder) next() uint8 {
n := &d.dt[d.state]
lowBits := d.br.getBits(n.nbBits)
d.state = n.newState + lowBits
return n.symbol
} | [
"func",
"(",
"d",
"*",
"decoder",
")",
"next",
"(",
")",
"uint8",
"{",
"n",
":=",
"&",
"d",
".",
"dt",
"[",
"d",
".",
"state",
"]",
"\n",
"lowBits",
":=",
"d",
".",
"br",
".",
"getBits",
"(",
"n",
".",
"nbBits",
")",
"\n",
"d",
".",
"state",
"=",
"n",
".",
"newState",
"+",
"lowBits",
"\n",
"return",
"n",
".",
"symbol",
"\n",
"}"
]
| // next returns the next symbol and sets the next state.
// At least tablelog bits must be available in the bit reader. | [
"next",
"returns",
"the",
"next",
"symbol",
"and",
"sets",
"the",
"next",
"state",
".",
"At",
"least",
"tablelog",
"bits",
"must",
"be",
"available",
"in",
"the",
"bit",
"reader",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/fse/decompress.go#L344-L349 |
8,332 | klauspost/compress | zstd/framedec.go | next | func (d *frameDec) next(block *blockDec) error {
println("decoding new block")
err := block.reset(d.rawInput, d.WindowSize)
if err != nil {
println("block error:", err)
// Signal the frame decoder we have a problem.
d.sendErr(block, err)
return err
}
block.input <- struct{}{}
if debug {
println("next block:", block)
}
d.asyncRunningMu.Lock()
defer d.asyncRunningMu.Unlock()
if !d.asyncRunning {
return nil
}
if block.Last {
// We indicate the frame is done by sending io.EOF
d.decoding <- block
return io.EOF
}
d.decoding <- block
return nil
} | go | func (d *frameDec) next(block *blockDec) error {
println("decoding new block")
err := block.reset(d.rawInput, d.WindowSize)
if err != nil {
println("block error:", err)
// Signal the frame decoder we have a problem.
d.sendErr(block, err)
return err
}
block.input <- struct{}{}
if debug {
println("next block:", block)
}
d.asyncRunningMu.Lock()
defer d.asyncRunningMu.Unlock()
if !d.asyncRunning {
return nil
}
if block.Last {
// We indicate the frame is done by sending io.EOF
d.decoding <- block
return io.EOF
}
d.decoding <- block
return nil
} | [
"func",
"(",
"d",
"*",
"frameDec",
")",
"next",
"(",
"block",
"*",
"blockDec",
")",
"error",
"{",
"println",
"(",
"\"",
"\"",
")",
"\n",
"err",
":=",
"block",
".",
"reset",
"(",
"d",
".",
"rawInput",
",",
"d",
".",
"WindowSize",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"println",
"(",
"\"",
"\"",
",",
"err",
")",
"\n",
"// Signal the frame decoder we have a problem.",
"d",
".",
"sendErr",
"(",
"block",
",",
"err",
")",
"\n",
"return",
"err",
"\n",
"}",
"\n",
"block",
".",
"input",
"<-",
"struct",
"{",
"}",
"{",
"}",
"\n",
"if",
"debug",
"{",
"println",
"(",
"\"",
"\"",
",",
"block",
")",
"\n",
"}",
"\n",
"d",
".",
"asyncRunningMu",
".",
"Lock",
"(",
")",
"\n",
"defer",
"d",
".",
"asyncRunningMu",
".",
"Unlock",
"(",
")",
"\n",
"if",
"!",
"d",
".",
"asyncRunning",
"{",
"return",
"nil",
"\n",
"}",
"\n",
"if",
"block",
".",
"Last",
"{",
"// We indicate the frame is done by sending io.EOF",
"d",
".",
"decoding",
"<-",
"block",
"\n",
"return",
"io",
".",
"EOF",
"\n",
"}",
"\n",
"d",
".",
"decoding",
"<-",
"block",
"\n",
"return",
"nil",
"\n",
"}"
]
| // next will start decoding the next block from stream. | [
"next",
"will",
"start",
"decoding",
"the",
"next",
"block",
"from",
"stream",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zstd/framedec.go#L233-L258 |
8,333 | klauspost/compress | zstd/framedec.go | sendErr | func (d *frameDec) sendErr(block *blockDec, err error) bool {
d.asyncRunningMu.Lock()
defer d.asyncRunningMu.Unlock()
if !d.asyncRunning {
return false
}
println("sending error", err.Error())
block.sendErr(err)
d.decoding <- block
return true
} | go | func (d *frameDec) sendErr(block *blockDec, err error) bool {
d.asyncRunningMu.Lock()
defer d.asyncRunningMu.Unlock()
if !d.asyncRunning {
return false
}
println("sending error", err.Error())
block.sendErr(err)
d.decoding <- block
return true
} | [
"func",
"(",
"d",
"*",
"frameDec",
")",
"sendErr",
"(",
"block",
"*",
"blockDec",
",",
"err",
"error",
")",
"bool",
"{",
"d",
".",
"asyncRunningMu",
".",
"Lock",
"(",
")",
"\n",
"defer",
"d",
".",
"asyncRunningMu",
".",
"Unlock",
"(",
")",
"\n",
"if",
"!",
"d",
".",
"asyncRunning",
"{",
"return",
"false",
"\n",
"}",
"\n\n",
"println",
"(",
"\"",
"\"",
",",
"err",
".",
"Error",
"(",
")",
")",
"\n",
"block",
".",
"sendErr",
"(",
"err",
")",
"\n",
"d",
".",
"decoding",
"<-",
"block",
"\n",
"return",
"true",
"\n",
"}"
]
| // sendEOF will queue an error block on the frame.
// This will cause the frame decoder to return when it encounters the block.
// Returns true if the decoder was added. | [
"sendEOF",
"will",
"queue",
"an",
"error",
"block",
"on",
"the",
"frame",
".",
"This",
"will",
"cause",
"the",
"frame",
"decoder",
"to",
"return",
"when",
"it",
"encounters",
"the",
"block",
".",
"Returns",
"true",
"if",
"the",
"decoder",
"was",
"added",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zstd/framedec.go#L263-L274 |
8,334 | klauspost/compress | zstd/framedec.go | checkCRC | func (d *frameDec) checkCRC() error {
if !d.HasCheckSum {
return nil
}
var tmp [8]byte
gotB := d.crc.Sum(tmp[:0])
// Flip to match file order.
gotB[0] = gotB[7]
gotB[1] = gotB[6]
gotB[2] = gotB[5]
gotB[3] = gotB[4]
// We can overwrite upper tmp now
want := d.rawInput.readSmall(4)
if want == nil {
println("CRC missing?")
return io.ErrUnexpectedEOF
}
if !bytes.Equal(gotB[:4], want) {
println("CRC Check Failed:", gotB[:4], "!=", want)
return ErrCRCMismatch
}
println("CRC ok")
return nil
} | go | func (d *frameDec) checkCRC() error {
if !d.HasCheckSum {
return nil
}
var tmp [8]byte
gotB := d.crc.Sum(tmp[:0])
// Flip to match file order.
gotB[0] = gotB[7]
gotB[1] = gotB[6]
gotB[2] = gotB[5]
gotB[3] = gotB[4]
// We can overwrite upper tmp now
want := d.rawInput.readSmall(4)
if want == nil {
println("CRC missing?")
return io.ErrUnexpectedEOF
}
if !bytes.Equal(gotB[:4], want) {
println("CRC Check Failed:", gotB[:4], "!=", want)
return ErrCRCMismatch
}
println("CRC ok")
return nil
} | [
"func",
"(",
"d",
"*",
"frameDec",
")",
"checkCRC",
"(",
")",
"error",
"{",
"if",
"!",
"d",
".",
"HasCheckSum",
"{",
"return",
"nil",
"\n",
"}",
"\n",
"var",
"tmp",
"[",
"8",
"]",
"byte",
"\n",
"gotB",
":=",
"d",
".",
"crc",
".",
"Sum",
"(",
"tmp",
"[",
":",
"0",
"]",
")",
"\n",
"// Flip to match file order.",
"gotB",
"[",
"0",
"]",
"=",
"gotB",
"[",
"7",
"]",
"\n",
"gotB",
"[",
"1",
"]",
"=",
"gotB",
"[",
"6",
"]",
"\n",
"gotB",
"[",
"2",
"]",
"=",
"gotB",
"[",
"5",
"]",
"\n",
"gotB",
"[",
"3",
"]",
"=",
"gotB",
"[",
"4",
"]",
"\n\n",
"// We can overwrite upper tmp now",
"want",
":=",
"d",
".",
"rawInput",
".",
"readSmall",
"(",
"4",
")",
"\n",
"if",
"want",
"==",
"nil",
"{",
"println",
"(",
"\"",
"\"",
")",
"\n",
"return",
"io",
".",
"ErrUnexpectedEOF",
"\n",
"}",
"\n\n",
"if",
"!",
"bytes",
".",
"Equal",
"(",
"gotB",
"[",
":",
"4",
"]",
",",
"want",
")",
"{",
"println",
"(",
"\"",
"\"",
",",
"gotB",
"[",
":",
"4",
"]",
",",
"\"",
"\"",
",",
"want",
")",
"\n",
"return",
"ErrCRCMismatch",
"\n",
"}",
"\n",
"println",
"(",
"\"",
"\"",
")",
"\n",
"return",
"nil",
"\n",
"}"
]
| // checkCRC will check the checksum if the frame has one.
// Will return ErrCRCMismatch if crc check failed, otherwise nil. | [
"checkCRC",
"will",
"check",
"the",
"checksum",
"if",
"the",
"frame",
"has",
"one",
".",
"Will",
"return",
"ErrCRCMismatch",
"if",
"crc",
"check",
"failed",
"otherwise",
"nil",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zstd/framedec.go#L278-L303 |
8,335 | klauspost/compress | zstd/framedec.go | runDecoder | func (d *frameDec) runDecoder(dst []byte, dec *blockDec) ([]byte, error) {
// TODO: Init to dictionary
d.history.reset()
saved := d.history.b
// We use the history for output to avoid copying it.
d.history.b = dst
// Store input length, so we only check new data.
crcStart := len(dst)
var err error
for {
err = dec.reset(d.rawInput, d.WindowSize)
if err != nil {
break
}
if debug {
println("next block:", dec)
}
err = dec.decodeBuf(&d.history)
if err != nil || dec.Last {
break
}
if uint64(len(d.history.b)) > d.o.maxDecodedSize {
err = ErrDecoderSizeExceeded
break
}
}
dst = d.history.b
if err == nil {
if d.HasCheckSum {
var n int
n, err = d.crc.Write(dst[crcStart:])
if err == nil {
if n != len(dst)-crcStart {
err = io.ErrShortWrite
}
}
err = d.checkCRC()
}
}
d.history.b = saved
return dst, err
} | go | func (d *frameDec) runDecoder(dst []byte, dec *blockDec) ([]byte, error) {
// TODO: Init to dictionary
d.history.reset()
saved := d.history.b
// We use the history for output to avoid copying it.
d.history.b = dst
// Store input length, so we only check new data.
crcStart := len(dst)
var err error
for {
err = dec.reset(d.rawInput, d.WindowSize)
if err != nil {
break
}
if debug {
println("next block:", dec)
}
err = dec.decodeBuf(&d.history)
if err != nil || dec.Last {
break
}
if uint64(len(d.history.b)) > d.o.maxDecodedSize {
err = ErrDecoderSizeExceeded
break
}
}
dst = d.history.b
if err == nil {
if d.HasCheckSum {
var n int
n, err = d.crc.Write(dst[crcStart:])
if err == nil {
if n != len(dst)-crcStart {
err = io.ErrShortWrite
}
}
err = d.checkCRC()
}
}
d.history.b = saved
return dst, err
} | [
"func",
"(",
"d",
"*",
"frameDec",
")",
"runDecoder",
"(",
"dst",
"[",
"]",
"byte",
",",
"dec",
"*",
"blockDec",
")",
"(",
"[",
"]",
"byte",
",",
"error",
")",
"{",
"// TODO: Init to dictionary",
"d",
".",
"history",
".",
"reset",
"(",
")",
"\n",
"saved",
":=",
"d",
".",
"history",
".",
"b",
"\n\n",
"// We use the history for output to avoid copying it.",
"d",
".",
"history",
".",
"b",
"=",
"dst",
"\n",
"// Store input length, so we only check new data.",
"crcStart",
":=",
"len",
"(",
"dst",
")",
"\n",
"var",
"err",
"error",
"\n",
"for",
"{",
"err",
"=",
"dec",
".",
"reset",
"(",
"d",
".",
"rawInput",
",",
"d",
".",
"WindowSize",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"break",
"\n",
"}",
"\n",
"if",
"debug",
"{",
"println",
"(",
"\"",
"\"",
",",
"dec",
")",
"\n",
"}",
"\n",
"err",
"=",
"dec",
".",
"decodeBuf",
"(",
"&",
"d",
".",
"history",
")",
"\n",
"if",
"err",
"!=",
"nil",
"||",
"dec",
".",
"Last",
"{",
"break",
"\n",
"}",
"\n",
"if",
"uint64",
"(",
"len",
"(",
"d",
".",
"history",
".",
"b",
")",
")",
">",
"d",
".",
"o",
".",
"maxDecodedSize",
"{",
"err",
"=",
"ErrDecoderSizeExceeded",
"\n",
"break",
"\n",
"}",
"\n",
"}",
"\n",
"dst",
"=",
"d",
".",
"history",
".",
"b",
"\n",
"if",
"err",
"==",
"nil",
"{",
"if",
"d",
".",
"HasCheckSum",
"{",
"var",
"n",
"int",
"\n",
"n",
",",
"err",
"=",
"d",
".",
"crc",
".",
"Write",
"(",
"dst",
"[",
"crcStart",
":",
"]",
")",
"\n",
"if",
"err",
"==",
"nil",
"{",
"if",
"n",
"!=",
"len",
"(",
"dst",
")",
"-",
"crcStart",
"{",
"err",
"=",
"io",
".",
"ErrShortWrite",
"\n",
"}",
"\n",
"}",
"\n",
"err",
"=",
"d",
".",
"checkCRC",
"(",
")",
"\n",
"}",
"\n",
"}",
"\n",
"d",
".",
"history",
".",
"b",
"=",
"saved",
"\n",
"return",
"dst",
",",
"err",
"\n",
"}"
]
| // runDecoder will create a sync decoder that will decodeAsync a block of data. | [
"runDecoder",
"will",
"create",
"a",
"sync",
"decoder",
"that",
"will",
"decodeAsync",
"a",
"block",
"of",
"data",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zstd/framedec.go#L418-L460 |
8,336 | klauspost/compress | flate/huffman_bit_writer.go | dynamicSize | func (w *huffmanBitWriter) dynamicSize(litEnc, offEnc *huffmanEncoder, extraBits int) (size, numCodegens int) {
numCodegens = len(w.codegenFreq)
for numCodegens > 4 && w.codegenFreq[codegenOrder[numCodegens-1]] == 0 {
numCodegens--
}
header := 3 + 5 + 5 + 4 + (3 * numCodegens) +
w.codegenEncoding.bitLength(w.codegenFreq[:]) +
int(w.codegenFreq[16])*2 +
int(w.codegenFreq[17])*3 +
int(w.codegenFreq[18])*7
size = header +
litEnc.bitLength(w.literalFreq) +
offEnc.bitLength(w.offsetFreq) +
extraBits
return size, numCodegens
} | go | func (w *huffmanBitWriter) dynamicSize(litEnc, offEnc *huffmanEncoder, extraBits int) (size, numCodegens int) {
numCodegens = len(w.codegenFreq)
for numCodegens > 4 && w.codegenFreq[codegenOrder[numCodegens-1]] == 0 {
numCodegens--
}
header := 3 + 5 + 5 + 4 + (3 * numCodegens) +
w.codegenEncoding.bitLength(w.codegenFreq[:]) +
int(w.codegenFreq[16])*2 +
int(w.codegenFreq[17])*3 +
int(w.codegenFreq[18])*7
size = header +
litEnc.bitLength(w.literalFreq) +
offEnc.bitLength(w.offsetFreq) +
extraBits
return size, numCodegens
} | [
"func",
"(",
"w",
"*",
"huffmanBitWriter",
")",
"dynamicSize",
"(",
"litEnc",
",",
"offEnc",
"*",
"huffmanEncoder",
",",
"extraBits",
"int",
")",
"(",
"size",
",",
"numCodegens",
"int",
")",
"{",
"numCodegens",
"=",
"len",
"(",
"w",
".",
"codegenFreq",
")",
"\n",
"for",
"numCodegens",
">",
"4",
"&&",
"w",
".",
"codegenFreq",
"[",
"codegenOrder",
"[",
"numCodegens",
"-",
"1",
"]",
"]",
"==",
"0",
"{",
"numCodegens",
"--",
"\n",
"}",
"\n",
"header",
":=",
"3",
"+",
"5",
"+",
"5",
"+",
"4",
"+",
"(",
"3",
"*",
"numCodegens",
")",
"+",
"w",
".",
"codegenEncoding",
".",
"bitLength",
"(",
"w",
".",
"codegenFreq",
"[",
":",
"]",
")",
"+",
"int",
"(",
"w",
".",
"codegenFreq",
"[",
"16",
"]",
")",
"*",
"2",
"+",
"int",
"(",
"w",
".",
"codegenFreq",
"[",
"17",
"]",
")",
"*",
"3",
"+",
"int",
"(",
"w",
".",
"codegenFreq",
"[",
"18",
"]",
")",
"*",
"7",
"\n",
"size",
"=",
"header",
"+",
"litEnc",
".",
"bitLength",
"(",
"w",
".",
"literalFreq",
")",
"+",
"offEnc",
".",
"bitLength",
"(",
"w",
".",
"offsetFreq",
")",
"+",
"extraBits",
"\n\n",
"return",
"size",
",",
"numCodegens",
"\n",
"}"
]
| // dynamicSize returns the size of dynamically encoded data in bits. | [
"dynamicSize",
"returns",
"the",
"size",
"of",
"dynamically",
"encoded",
"data",
"in",
"bits",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/flate/huffman_bit_writer.go#L296-L312 |
8,337 | klauspost/compress | flate/huffman_bit_writer.go | fixedSize | func (w *huffmanBitWriter) fixedSize(extraBits int) int {
return 3 +
fixedLiteralEncoding.bitLength(w.literalFreq) +
fixedOffsetEncoding.bitLength(w.offsetFreq) +
extraBits
} | go | func (w *huffmanBitWriter) fixedSize(extraBits int) int {
return 3 +
fixedLiteralEncoding.bitLength(w.literalFreq) +
fixedOffsetEncoding.bitLength(w.offsetFreq) +
extraBits
} | [
"func",
"(",
"w",
"*",
"huffmanBitWriter",
")",
"fixedSize",
"(",
"extraBits",
"int",
")",
"int",
"{",
"return",
"3",
"+",
"fixedLiteralEncoding",
".",
"bitLength",
"(",
"w",
".",
"literalFreq",
")",
"+",
"fixedOffsetEncoding",
".",
"bitLength",
"(",
"w",
".",
"offsetFreq",
")",
"+",
"extraBits",
"\n",
"}"
]
| // fixedSize returns the size of dynamically encoded data in bits. | [
"fixedSize",
"returns",
"the",
"size",
"of",
"dynamically",
"encoded",
"data",
"in",
"bits",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/flate/huffman_bit_writer.go#L315-L320 |
8,338 | klauspost/compress | flate/huffman_bit_writer.go | storedSize | func (w *huffmanBitWriter) storedSize(in []byte) (int, bool) {
if in == nil {
return 0, false
}
if len(in) <= maxStoreBlockSize {
return (len(in) + 5) * 8, true
}
return 0, false
} | go | func (w *huffmanBitWriter) storedSize(in []byte) (int, bool) {
if in == nil {
return 0, false
}
if len(in) <= maxStoreBlockSize {
return (len(in) + 5) * 8, true
}
return 0, false
} | [
"func",
"(",
"w",
"*",
"huffmanBitWriter",
")",
"storedSize",
"(",
"in",
"[",
"]",
"byte",
")",
"(",
"int",
",",
"bool",
")",
"{",
"if",
"in",
"==",
"nil",
"{",
"return",
"0",
",",
"false",
"\n",
"}",
"\n",
"if",
"len",
"(",
"in",
")",
"<=",
"maxStoreBlockSize",
"{",
"return",
"(",
"len",
"(",
"in",
")",
"+",
"5",
")",
"*",
"8",
",",
"true",
"\n",
"}",
"\n",
"return",
"0",
",",
"false",
"\n",
"}"
]
| // storedSize calculates the stored size, including header.
// The function returns the size in bits and whether the block
// fits inside a single block. | [
"storedSize",
"calculates",
"the",
"stored",
"size",
"including",
"header",
".",
"The",
"function",
"returns",
"the",
"size",
"in",
"bits",
"and",
"whether",
"the",
"block",
"fits",
"inside",
"a",
"single",
"block",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/flate/huffman_bit_writer.go#L325-L333 |
8,339 | klauspost/compress | flate/huffman_bit_writer.go | writeDynamicHeader | func (w *huffmanBitWriter) writeDynamicHeader(numLiterals int, numOffsets int, numCodegens int, isEof bool) {
if w.err != nil {
return
}
var firstBits int32 = 4
if isEof {
firstBits = 5
}
w.writeBits(firstBits, 3)
w.writeBits(int32(numLiterals-257), 5)
w.writeBits(int32(numOffsets-1), 5)
w.writeBits(int32(numCodegens-4), 4)
for i := 0; i < numCodegens; i++ {
value := uint(w.codegenEncoding.codes[codegenOrder[i]].len)
w.writeBits(int32(value), 3)
}
i := 0
for {
var codeWord int = int(w.codegen[i])
i++
if codeWord == badCode {
break
}
w.writeCode(w.codegenEncoding.codes[uint32(codeWord)])
switch codeWord {
case 16:
w.writeBits(int32(w.codegen[i]), 2)
i++
break
case 17:
w.writeBits(int32(w.codegen[i]), 3)
i++
break
case 18:
w.writeBits(int32(w.codegen[i]), 7)
i++
break
}
}
} | go | func (w *huffmanBitWriter) writeDynamicHeader(numLiterals int, numOffsets int, numCodegens int, isEof bool) {
if w.err != nil {
return
}
var firstBits int32 = 4
if isEof {
firstBits = 5
}
w.writeBits(firstBits, 3)
w.writeBits(int32(numLiterals-257), 5)
w.writeBits(int32(numOffsets-1), 5)
w.writeBits(int32(numCodegens-4), 4)
for i := 0; i < numCodegens; i++ {
value := uint(w.codegenEncoding.codes[codegenOrder[i]].len)
w.writeBits(int32(value), 3)
}
i := 0
for {
var codeWord int = int(w.codegen[i])
i++
if codeWord == badCode {
break
}
w.writeCode(w.codegenEncoding.codes[uint32(codeWord)])
switch codeWord {
case 16:
w.writeBits(int32(w.codegen[i]), 2)
i++
break
case 17:
w.writeBits(int32(w.codegen[i]), 3)
i++
break
case 18:
w.writeBits(int32(w.codegen[i]), 7)
i++
break
}
}
} | [
"func",
"(",
"w",
"*",
"huffmanBitWriter",
")",
"writeDynamicHeader",
"(",
"numLiterals",
"int",
",",
"numOffsets",
"int",
",",
"numCodegens",
"int",
",",
"isEof",
"bool",
")",
"{",
"if",
"w",
".",
"err",
"!=",
"nil",
"{",
"return",
"\n",
"}",
"\n",
"var",
"firstBits",
"int32",
"=",
"4",
"\n",
"if",
"isEof",
"{",
"firstBits",
"=",
"5",
"\n",
"}",
"\n",
"w",
".",
"writeBits",
"(",
"firstBits",
",",
"3",
")",
"\n",
"w",
".",
"writeBits",
"(",
"int32",
"(",
"numLiterals",
"-",
"257",
")",
",",
"5",
")",
"\n",
"w",
".",
"writeBits",
"(",
"int32",
"(",
"numOffsets",
"-",
"1",
")",
",",
"5",
")",
"\n",
"w",
".",
"writeBits",
"(",
"int32",
"(",
"numCodegens",
"-",
"4",
")",
",",
"4",
")",
"\n\n",
"for",
"i",
":=",
"0",
";",
"i",
"<",
"numCodegens",
";",
"i",
"++",
"{",
"value",
":=",
"uint",
"(",
"w",
".",
"codegenEncoding",
".",
"codes",
"[",
"codegenOrder",
"[",
"i",
"]",
"]",
".",
"len",
")",
"\n",
"w",
".",
"writeBits",
"(",
"int32",
"(",
"value",
")",
",",
"3",
")",
"\n",
"}",
"\n\n",
"i",
":=",
"0",
"\n",
"for",
"{",
"var",
"codeWord",
"int",
"=",
"int",
"(",
"w",
".",
"codegen",
"[",
"i",
"]",
")",
"\n",
"i",
"++",
"\n",
"if",
"codeWord",
"==",
"badCode",
"{",
"break",
"\n",
"}",
"\n",
"w",
".",
"writeCode",
"(",
"w",
".",
"codegenEncoding",
".",
"codes",
"[",
"uint32",
"(",
"codeWord",
")",
"]",
")",
"\n\n",
"switch",
"codeWord",
"{",
"case",
"16",
":",
"w",
".",
"writeBits",
"(",
"int32",
"(",
"w",
".",
"codegen",
"[",
"i",
"]",
")",
",",
"2",
")",
"\n",
"i",
"++",
"\n",
"break",
"\n",
"case",
"17",
":",
"w",
".",
"writeBits",
"(",
"int32",
"(",
"w",
".",
"codegen",
"[",
"i",
"]",
")",
",",
"3",
")",
"\n",
"i",
"++",
"\n",
"break",
"\n",
"case",
"18",
":",
"w",
".",
"writeBits",
"(",
"int32",
"(",
"w",
".",
"codegen",
"[",
"i",
"]",
")",
",",
"7",
")",
"\n",
"i",
"++",
"\n",
"break",
"\n",
"}",
"\n",
"}",
"\n",
"}"
]
| // Write the header of a dynamic Huffman block to the output stream.
//
// numLiterals The number of literals specified in codegen
// numOffsets The number of offsets specified in codegen
// numCodegens The number of codegens used in codegen | [
"Write",
"the",
"header",
"of",
"a",
"dynamic",
"Huffman",
"block",
"to",
"the",
"output",
"stream",
".",
"numLiterals",
"The",
"number",
"of",
"literals",
"specified",
"in",
"codegen",
"numOffsets",
"The",
"number",
"of",
"offsets",
"specified",
"in",
"codegen",
"numCodegens",
"The",
"number",
"of",
"codegens",
"used",
"in",
"codegen"
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/flate/huffman_bit_writer.go#L367-L409 |
8,340 | klauspost/compress | flate/huffman_bit_writer.go | writeBlock | func (w *huffmanBitWriter) writeBlock(tokens []token, eof bool, input []byte) {
if w.err != nil {
return
}
tokens = append(tokens, endBlockMarker)
numLiterals, numOffsets := w.indexTokens(tokens)
var extraBits int
storedSize, storable := w.storedSize(input)
if storable {
// We only bother calculating the costs of the extra bits required by
// the length of offset fields (which will be the same for both fixed
// and dynamic encoding), if we need to compare those two encodings
// against stored encoding.
for lengthCode := lengthCodesStart + 8; lengthCode < numLiterals; lengthCode++ {
// First eight length codes have extra size = 0.
extraBits += int(w.literalFreq[lengthCode]) * int(lengthExtraBits[lengthCode-lengthCodesStart])
}
for offsetCode := 4; offsetCode < numOffsets; offsetCode++ {
// First four offset codes have extra size = 0.
extraBits += int(w.offsetFreq[offsetCode]) * int(offsetExtraBits[offsetCode])
}
}
// Figure out smallest code.
// Fixed Huffman baseline.
var literalEncoding = fixedLiteralEncoding
var offsetEncoding = fixedOffsetEncoding
var size = w.fixedSize(extraBits)
// Dynamic Huffman?
var numCodegens int
// Generate codegen and codegenFrequencies, which indicates how to encode
// the literalEncoding and the offsetEncoding.
w.generateCodegen(numLiterals, numOffsets, w.literalEncoding, w.offsetEncoding)
w.codegenEncoding.generate(w.codegenFreq[:], 7)
dynamicSize, numCodegens := w.dynamicSize(w.literalEncoding, w.offsetEncoding, extraBits)
if dynamicSize < size {
size = dynamicSize
literalEncoding = w.literalEncoding
offsetEncoding = w.offsetEncoding
}
// Stored bytes?
if storable && storedSize < size {
w.writeStoredHeader(len(input), eof)
w.writeBytes(input)
return
}
// Huffman.
if literalEncoding == fixedLiteralEncoding {
w.writeFixedHeader(eof)
} else {
w.writeDynamicHeader(numLiterals, numOffsets, numCodegens, eof)
}
// Write the tokens.
w.writeTokens(tokens, literalEncoding.codes, offsetEncoding.codes)
} | go | func (w *huffmanBitWriter) writeBlock(tokens []token, eof bool, input []byte) {
if w.err != nil {
return
}
tokens = append(tokens, endBlockMarker)
numLiterals, numOffsets := w.indexTokens(tokens)
var extraBits int
storedSize, storable := w.storedSize(input)
if storable {
// We only bother calculating the costs of the extra bits required by
// the length of offset fields (which will be the same for both fixed
// and dynamic encoding), if we need to compare those two encodings
// against stored encoding.
for lengthCode := lengthCodesStart + 8; lengthCode < numLiterals; lengthCode++ {
// First eight length codes have extra size = 0.
extraBits += int(w.literalFreq[lengthCode]) * int(lengthExtraBits[lengthCode-lengthCodesStart])
}
for offsetCode := 4; offsetCode < numOffsets; offsetCode++ {
// First four offset codes have extra size = 0.
extraBits += int(w.offsetFreq[offsetCode]) * int(offsetExtraBits[offsetCode])
}
}
// Figure out smallest code.
// Fixed Huffman baseline.
var literalEncoding = fixedLiteralEncoding
var offsetEncoding = fixedOffsetEncoding
var size = w.fixedSize(extraBits)
// Dynamic Huffman?
var numCodegens int
// Generate codegen and codegenFrequencies, which indicates how to encode
// the literalEncoding and the offsetEncoding.
w.generateCodegen(numLiterals, numOffsets, w.literalEncoding, w.offsetEncoding)
w.codegenEncoding.generate(w.codegenFreq[:], 7)
dynamicSize, numCodegens := w.dynamicSize(w.literalEncoding, w.offsetEncoding, extraBits)
if dynamicSize < size {
size = dynamicSize
literalEncoding = w.literalEncoding
offsetEncoding = w.offsetEncoding
}
// Stored bytes?
if storable && storedSize < size {
w.writeStoredHeader(len(input), eof)
w.writeBytes(input)
return
}
// Huffman.
if literalEncoding == fixedLiteralEncoding {
w.writeFixedHeader(eof)
} else {
w.writeDynamicHeader(numLiterals, numOffsets, numCodegens, eof)
}
// Write the tokens.
w.writeTokens(tokens, literalEncoding.codes, offsetEncoding.codes)
} | [
"func",
"(",
"w",
"*",
"huffmanBitWriter",
")",
"writeBlock",
"(",
"tokens",
"[",
"]",
"token",
",",
"eof",
"bool",
",",
"input",
"[",
"]",
"byte",
")",
"{",
"if",
"w",
".",
"err",
"!=",
"nil",
"{",
"return",
"\n",
"}",
"\n\n",
"tokens",
"=",
"append",
"(",
"tokens",
",",
"endBlockMarker",
")",
"\n",
"numLiterals",
",",
"numOffsets",
":=",
"w",
".",
"indexTokens",
"(",
"tokens",
")",
"\n\n",
"var",
"extraBits",
"int",
"\n",
"storedSize",
",",
"storable",
":=",
"w",
".",
"storedSize",
"(",
"input",
")",
"\n",
"if",
"storable",
"{",
"// We only bother calculating the costs of the extra bits required by",
"// the length of offset fields (which will be the same for both fixed",
"// and dynamic encoding), if we need to compare those two encodings",
"// against stored encoding.",
"for",
"lengthCode",
":=",
"lengthCodesStart",
"+",
"8",
";",
"lengthCode",
"<",
"numLiterals",
";",
"lengthCode",
"++",
"{",
"// First eight length codes have extra size = 0.",
"extraBits",
"+=",
"int",
"(",
"w",
".",
"literalFreq",
"[",
"lengthCode",
"]",
")",
"*",
"int",
"(",
"lengthExtraBits",
"[",
"lengthCode",
"-",
"lengthCodesStart",
"]",
")",
"\n",
"}",
"\n",
"for",
"offsetCode",
":=",
"4",
";",
"offsetCode",
"<",
"numOffsets",
";",
"offsetCode",
"++",
"{",
"// First four offset codes have extra size = 0.",
"extraBits",
"+=",
"int",
"(",
"w",
".",
"offsetFreq",
"[",
"offsetCode",
"]",
")",
"*",
"int",
"(",
"offsetExtraBits",
"[",
"offsetCode",
"]",
")",
"\n",
"}",
"\n",
"}",
"\n\n",
"// Figure out smallest code.",
"// Fixed Huffman baseline.",
"var",
"literalEncoding",
"=",
"fixedLiteralEncoding",
"\n",
"var",
"offsetEncoding",
"=",
"fixedOffsetEncoding",
"\n",
"var",
"size",
"=",
"w",
".",
"fixedSize",
"(",
"extraBits",
")",
"\n\n",
"// Dynamic Huffman?",
"var",
"numCodegens",
"int",
"\n\n",
"// Generate codegen and codegenFrequencies, which indicates how to encode",
"// the literalEncoding and the offsetEncoding.",
"w",
".",
"generateCodegen",
"(",
"numLiterals",
",",
"numOffsets",
",",
"w",
".",
"literalEncoding",
",",
"w",
".",
"offsetEncoding",
")",
"\n",
"w",
".",
"codegenEncoding",
".",
"generate",
"(",
"w",
".",
"codegenFreq",
"[",
":",
"]",
",",
"7",
")",
"\n",
"dynamicSize",
",",
"numCodegens",
":=",
"w",
".",
"dynamicSize",
"(",
"w",
".",
"literalEncoding",
",",
"w",
".",
"offsetEncoding",
",",
"extraBits",
")",
"\n\n",
"if",
"dynamicSize",
"<",
"size",
"{",
"size",
"=",
"dynamicSize",
"\n",
"literalEncoding",
"=",
"w",
".",
"literalEncoding",
"\n",
"offsetEncoding",
"=",
"w",
".",
"offsetEncoding",
"\n",
"}",
"\n\n",
"// Stored bytes?",
"if",
"storable",
"&&",
"storedSize",
"<",
"size",
"{",
"w",
".",
"writeStoredHeader",
"(",
"len",
"(",
"input",
")",
",",
"eof",
")",
"\n",
"w",
".",
"writeBytes",
"(",
"input",
")",
"\n",
"return",
"\n",
"}",
"\n\n",
"// Huffman.",
"if",
"literalEncoding",
"==",
"fixedLiteralEncoding",
"{",
"w",
".",
"writeFixedHeader",
"(",
"eof",
")",
"\n",
"}",
"else",
"{",
"w",
".",
"writeDynamicHeader",
"(",
"numLiterals",
",",
"numOffsets",
",",
"numCodegens",
",",
"eof",
")",
"\n",
"}",
"\n\n",
"// Write the tokens.",
"w",
".",
"writeTokens",
"(",
"tokens",
",",
"literalEncoding",
".",
"codes",
",",
"offsetEncoding",
".",
"codes",
")",
"\n",
"}"
]
| // writeBlock will write a block of tokens with the smallest encoding.
// The original input can be supplied, and if the huffman encoded data
// is larger than the original bytes, the data will be written as a
// stored block.
// If the input is nil, the tokens will always be Huffman encoded. | [
"writeBlock",
"will",
"write",
"a",
"block",
"of",
"tokens",
"with",
"the",
"smallest",
"encoding",
".",
"The",
"original",
"input",
"can",
"be",
"supplied",
"and",
"if",
"the",
"huffman",
"encoded",
"data",
"is",
"larger",
"than",
"the",
"original",
"bytes",
"the",
"data",
"will",
"be",
"written",
"as",
"a",
"stored",
"block",
".",
"If",
"the",
"input",
"is",
"nil",
"the",
"tokens",
"will",
"always",
"be",
"Huffman",
"encoded",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/flate/huffman_bit_writer.go#L442-L504 |
8,341 | klauspost/compress | flate/huffman_bit_writer.go | indexTokens | func (w *huffmanBitWriter) indexTokens(tokens []token) (numLiterals, numOffsets int) {
for i := range w.literalFreq {
w.literalFreq[i] = 0
}
for i := range w.offsetFreq {
w.offsetFreq[i] = 0
}
for _, t := range tokens {
if t < matchType {
w.literalFreq[t.literal()]++
continue
}
length := t.length()
offset := t.offset()
w.literalFreq[lengthCodesStart+lengthCode(length)]++
w.offsetFreq[offsetCode(offset)]++
}
// get the number of literals
numLiterals = len(w.literalFreq)
for w.literalFreq[numLiterals-1] == 0 {
numLiterals--
}
// get the number of offsets
numOffsets = len(w.offsetFreq)
for numOffsets > 0 && w.offsetFreq[numOffsets-1] == 0 {
numOffsets--
}
if numOffsets == 0 {
// We haven't found a single match. If we want to go with the dynamic encoding,
// we should count at least one offset to be sure that the offset huffman tree could be encoded.
w.offsetFreq[0] = 1
numOffsets = 1
}
w.literalEncoding.generate(w.literalFreq, 15)
w.offsetEncoding.generate(w.offsetFreq, 15)
return
} | go | func (w *huffmanBitWriter) indexTokens(tokens []token) (numLiterals, numOffsets int) {
for i := range w.literalFreq {
w.literalFreq[i] = 0
}
for i := range w.offsetFreq {
w.offsetFreq[i] = 0
}
for _, t := range tokens {
if t < matchType {
w.literalFreq[t.literal()]++
continue
}
length := t.length()
offset := t.offset()
w.literalFreq[lengthCodesStart+lengthCode(length)]++
w.offsetFreq[offsetCode(offset)]++
}
// get the number of literals
numLiterals = len(w.literalFreq)
for w.literalFreq[numLiterals-1] == 0 {
numLiterals--
}
// get the number of offsets
numOffsets = len(w.offsetFreq)
for numOffsets > 0 && w.offsetFreq[numOffsets-1] == 0 {
numOffsets--
}
if numOffsets == 0 {
// We haven't found a single match. If we want to go with the dynamic encoding,
// we should count at least one offset to be sure that the offset huffman tree could be encoded.
w.offsetFreq[0] = 1
numOffsets = 1
}
w.literalEncoding.generate(w.literalFreq, 15)
w.offsetEncoding.generate(w.offsetFreq, 15)
return
} | [
"func",
"(",
"w",
"*",
"huffmanBitWriter",
")",
"indexTokens",
"(",
"tokens",
"[",
"]",
"token",
")",
"(",
"numLiterals",
",",
"numOffsets",
"int",
")",
"{",
"for",
"i",
":=",
"range",
"w",
".",
"literalFreq",
"{",
"w",
".",
"literalFreq",
"[",
"i",
"]",
"=",
"0",
"\n",
"}",
"\n",
"for",
"i",
":=",
"range",
"w",
".",
"offsetFreq",
"{",
"w",
".",
"offsetFreq",
"[",
"i",
"]",
"=",
"0",
"\n",
"}",
"\n\n",
"for",
"_",
",",
"t",
":=",
"range",
"tokens",
"{",
"if",
"t",
"<",
"matchType",
"{",
"w",
".",
"literalFreq",
"[",
"t",
".",
"literal",
"(",
")",
"]",
"++",
"\n",
"continue",
"\n",
"}",
"\n",
"length",
":=",
"t",
".",
"length",
"(",
")",
"\n",
"offset",
":=",
"t",
".",
"offset",
"(",
")",
"\n",
"w",
".",
"literalFreq",
"[",
"lengthCodesStart",
"+",
"lengthCode",
"(",
"length",
")",
"]",
"++",
"\n",
"w",
".",
"offsetFreq",
"[",
"offsetCode",
"(",
"offset",
")",
"]",
"++",
"\n",
"}",
"\n\n",
"// get the number of literals",
"numLiterals",
"=",
"len",
"(",
"w",
".",
"literalFreq",
")",
"\n",
"for",
"w",
".",
"literalFreq",
"[",
"numLiterals",
"-",
"1",
"]",
"==",
"0",
"{",
"numLiterals",
"--",
"\n",
"}",
"\n",
"// get the number of offsets",
"numOffsets",
"=",
"len",
"(",
"w",
".",
"offsetFreq",
")",
"\n",
"for",
"numOffsets",
">",
"0",
"&&",
"w",
".",
"offsetFreq",
"[",
"numOffsets",
"-",
"1",
"]",
"==",
"0",
"{",
"numOffsets",
"--",
"\n",
"}",
"\n",
"if",
"numOffsets",
"==",
"0",
"{",
"// We haven't found a single match. If we want to go with the dynamic encoding,",
"// we should count at least one offset to be sure that the offset huffman tree could be encoded.",
"w",
".",
"offsetFreq",
"[",
"0",
"]",
"=",
"1",
"\n",
"numOffsets",
"=",
"1",
"\n",
"}",
"\n",
"w",
".",
"literalEncoding",
".",
"generate",
"(",
"w",
".",
"literalFreq",
",",
"15",
")",
"\n",
"w",
".",
"offsetEncoding",
".",
"generate",
"(",
"w",
".",
"offsetFreq",
",",
"15",
")",
"\n",
"return",
"\n",
"}"
]
| // indexTokens indexes a slice of tokens, and updates
// literalFreq and offsetFreq, and generates literalEncoding
// and offsetEncoding.
// The number of literal and offset tokens is returned. | [
"indexTokens",
"indexes",
"a",
"slice",
"of",
"tokens",
"and",
"updates",
"literalFreq",
"and",
"offsetFreq",
"and",
"generates",
"literalEncoding",
"and",
"offsetEncoding",
".",
"The",
"number",
"of",
"literal",
"and",
"offset",
"tokens",
"is",
"returned",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/flate/huffman_bit_writer.go#L543-L581 |
8,342 | klauspost/compress | flate/huffman_bit_writer.go | writeTokens | func (w *huffmanBitWriter) writeTokens(tokens []token, leCodes, oeCodes []hcode) {
if w.err != nil {
return
}
for _, t := range tokens {
if t < matchType {
w.writeCode(leCodes[t.literal()])
continue
}
// Write the length
length := t.length()
lengthCode := lengthCode(length)
w.writeCode(leCodes[lengthCode+lengthCodesStart])
extraLengthBits := uint(lengthExtraBits[lengthCode])
if extraLengthBits > 0 {
extraLength := int32(length - lengthBase[lengthCode])
w.writeBits(extraLength, extraLengthBits)
}
// Write the offset
offset := t.offset()
offsetCode := offsetCode(offset)
w.writeCode(oeCodes[offsetCode])
extraOffsetBits := uint(offsetExtraBits[offsetCode])
if extraOffsetBits > 0 {
extraOffset := int32(offset - offsetBase[offsetCode])
w.writeBits(extraOffset, extraOffsetBits)
}
}
} | go | func (w *huffmanBitWriter) writeTokens(tokens []token, leCodes, oeCodes []hcode) {
if w.err != nil {
return
}
for _, t := range tokens {
if t < matchType {
w.writeCode(leCodes[t.literal()])
continue
}
// Write the length
length := t.length()
lengthCode := lengthCode(length)
w.writeCode(leCodes[lengthCode+lengthCodesStart])
extraLengthBits := uint(lengthExtraBits[lengthCode])
if extraLengthBits > 0 {
extraLength := int32(length - lengthBase[lengthCode])
w.writeBits(extraLength, extraLengthBits)
}
// Write the offset
offset := t.offset()
offsetCode := offsetCode(offset)
w.writeCode(oeCodes[offsetCode])
extraOffsetBits := uint(offsetExtraBits[offsetCode])
if extraOffsetBits > 0 {
extraOffset := int32(offset - offsetBase[offsetCode])
w.writeBits(extraOffset, extraOffsetBits)
}
}
} | [
"func",
"(",
"w",
"*",
"huffmanBitWriter",
")",
"writeTokens",
"(",
"tokens",
"[",
"]",
"token",
",",
"leCodes",
",",
"oeCodes",
"[",
"]",
"hcode",
")",
"{",
"if",
"w",
".",
"err",
"!=",
"nil",
"{",
"return",
"\n",
"}",
"\n",
"for",
"_",
",",
"t",
":=",
"range",
"tokens",
"{",
"if",
"t",
"<",
"matchType",
"{",
"w",
".",
"writeCode",
"(",
"leCodes",
"[",
"t",
".",
"literal",
"(",
")",
"]",
")",
"\n",
"continue",
"\n",
"}",
"\n",
"// Write the length",
"length",
":=",
"t",
".",
"length",
"(",
")",
"\n",
"lengthCode",
":=",
"lengthCode",
"(",
"length",
")",
"\n",
"w",
".",
"writeCode",
"(",
"leCodes",
"[",
"lengthCode",
"+",
"lengthCodesStart",
"]",
")",
"\n",
"extraLengthBits",
":=",
"uint",
"(",
"lengthExtraBits",
"[",
"lengthCode",
"]",
")",
"\n",
"if",
"extraLengthBits",
">",
"0",
"{",
"extraLength",
":=",
"int32",
"(",
"length",
"-",
"lengthBase",
"[",
"lengthCode",
"]",
")",
"\n",
"w",
".",
"writeBits",
"(",
"extraLength",
",",
"extraLengthBits",
")",
"\n",
"}",
"\n",
"// Write the offset",
"offset",
":=",
"t",
".",
"offset",
"(",
")",
"\n",
"offsetCode",
":=",
"offsetCode",
"(",
"offset",
")",
"\n",
"w",
".",
"writeCode",
"(",
"oeCodes",
"[",
"offsetCode",
"]",
")",
"\n",
"extraOffsetBits",
":=",
"uint",
"(",
"offsetExtraBits",
"[",
"offsetCode",
"]",
")",
"\n",
"if",
"extraOffsetBits",
">",
"0",
"{",
"extraOffset",
":=",
"int32",
"(",
"offset",
"-",
"offsetBase",
"[",
"offsetCode",
"]",
")",
"\n",
"w",
".",
"writeBits",
"(",
"extraOffset",
",",
"extraOffsetBits",
")",
"\n",
"}",
"\n",
"}",
"\n",
"}"
]
| // writeTokens writes a slice of tokens to the output.
// codes for literal and offset encoding must be supplied. | [
"writeTokens",
"writes",
"a",
"slice",
"of",
"tokens",
"to",
"the",
"output",
".",
"codes",
"for",
"literal",
"and",
"offset",
"encoding",
"must",
"be",
"supplied",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/flate/huffman_bit_writer.go#L585-L613 |
8,343 | klauspost/compress | flate/huffman_bit_writer.go | writeBlockHuff | func (w *huffmanBitWriter) writeBlockHuff(eof bool, input []byte) {
if w.err != nil {
return
}
// Clear histogram
for i := range w.literalFreq {
w.literalFreq[i] = 0
}
// Add everything as literals
histogram(input, w.literalFreq)
w.literalFreq[endBlockMarker] = 1
const numLiterals = endBlockMarker + 1
const numOffsets = 1
w.literalEncoding.generate(w.literalFreq, 15)
// Figure out smallest code.
// Always use dynamic Huffman or Store
var numCodegens int
// Generate codegen and codegenFrequencies, which indicates how to encode
// the literalEncoding and the offsetEncoding.
w.generateCodegen(numLiterals, numOffsets, w.literalEncoding, huffOffset)
w.codegenEncoding.generate(w.codegenFreq[:], 7)
size, numCodegens := w.dynamicSize(w.literalEncoding, huffOffset, 0)
// Store bytes, if we don't get a reasonable improvement.
if ssize, storable := w.storedSize(input); storable && ssize < (size+size>>4) {
w.writeStoredHeader(len(input), eof)
w.writeBytes(input)
return
}
// Huffman.
w.writeDynamicHeader(numLiterals, numOffsets, numCodegens, eof)
encoding := w.literalEncoding.codes[:257]
n := w.nbytes
for _, t := range input {
// Bitwriting inlined, ~30% speedup
c := encoding[t]
w.bits |= uint64(c.code) << w.nbits
w.nbits += uint(c.len)
if w.nbits < 48 {
continue
}
// Store 6 bytes
bits := w.bits
w.bits >>= 48
w.nbits -= 48
bytes := w.bytes[n : n+6]
bytes[0] = byte(bits)
bytes[1] = byte(bits >> 8)
bytes[2] = byte(bits >> 16)
bytes[3] = byte(bits >> 24)
bytes[4] = byte(bits >> 32)
bytes[5] = byte(bits >> 40)
n += 6
if n < bufferFlushSize {
continue
}
w.write(w.bytes[:n])
if w.err != nil {
return // Return early in the event of write failures
}
n = 0
}
w.nbytes = n
w.writeCode(encoding[endBlockMarker])
} | go | func (w *huffmanBitWriter) writeBlockHuff(eof bool, input []byte) {
if w.err != nil {
return
}
// Clear histogram
for i := range w.literalFreq {
w.literalFreq[i] = 0
}
// Add everything as literals
histogram(input, w.literalFreq)
w.literalFreq[endBlockMarker] = 1
const numLiterals = endBlockMarker + 1
const numOffsets = 1
w.literalEncoding.generate(w.literalFreq, 15)
// Figure out smallest code.
// Always use dynamic Huffman or Store
var numCodegens int
// Generate codegen and codegenFrequencies, which indicates how to encode
// the literalEncoding and the offsetEncoding.
w.generateCodegen(numLiterals, numOffsets, w.literalEncoding, huffOffset)
w.codegenEncoding.generate(w.codegenFreq[:], 7)
size, numCodegens := w.dynamicSize(w.literalEncoding, huffOffset, 0)
// Store bytes, if we don't get a reasonable improvement.
if ssize, storable := w.storedSize(input); storable && ssize < (size+size>>4) {
w.writeStoredHeader(len(input), eof)
w.writeBytes(input)
return
}
// Huffman.
w.writeDynamicHeader(numLiterals, numOffsets, numCodegens, eof)
encoding := w.literalEncoding.codes[:257]
n := w.nbytes
for _, t := range input {
// Bitwriting inlined, ~30% speedup
c := encoding[t]
w.bits |= uint64(c.code) << w.nbits
w.nbits += uint(c.len)
if w.nbits < 48 {
continue
}
// Store 6 bytes
bits := w.bits
w.bits >>= 48
w.nbits -= 48
bytes := w.bytes[n : n+6]
bytes[0] = byte(bits)
bytes[1] = byte(bits >> 8)
bytes[2] = byte(bits >> 16)
bytes[3] = byte(bits >> 24)
bytes[4] = byte(bits >> 32)
bytes[5] = byte(bits >> 40)
n += 6
if n < bufferFlushSize {
continue
}
w.write(w.bytes[:n])
if w.err != nil {
return // Return early in the event of write failures
}
n = 0
}
w.nbytes = n
w.writeCode(encoding[endBlockMarker])
} | [
"func",
"(",
"w",
"*",
"huffmanBitWriter",
")",
"writeBlockHuff",
"(",
"eof",
"bool",
",",
"input",
"[",
"]",
"byte",
")",
"{",
"if",
"w",
".",
"err",
"!=",
"nil",
"{",
"return",
"\n",
"}",
"\n\n",
"// Clear histogram",
"for",
"i",
":=",
"range",
"w",
".",
"literalFreq",
"{",
"w",
".",
"literalFreq",
"[",
"i",
"]",
"=",
"0",
"\n",
"}",
"\n\n",
"// Add everything as literals",
"histogram",
"(",
"input",
",",
"w",
".",
"literalFreq",
")",
"\n\n",
"w",
".",
"literalFreq",
"[",
"endBlockMarker",
"]",
"=",
"1",
"\n\n",
"const",
"numLiterals",
"=",
"endBlockMarker",
"+",
"1",
"\n",
"const",
"numOffsets",
"=",
"1",
"\n\n",
"w",
".",
"literalEncoding",
".",
"generate",
"(",
"w",
".",
"literalFreq",
",",
"15",
")",
"\n\n",
"// Figure out smallest code.",
"// Always use dynamic Huffman or Store",
"var",
"numCodegens",
"int",
"\n\n",
"// Generate codegen and codegenFrequencies, which indicates how to encode",
"// the literalEncoding and the offsetEncoding.",
"w",
".",
"generateCodegen",
"(",
"numLiterals",
",",
"numOffsets",
",",
"w",
".",
"literalEncoding",
",",
"huffOffset",
")",
"\n",
"w",
".",
"codegenEncoding",
".",
"generate",
"(",
"w",
".",
"codegenFreq",
"[",
":",
"]",
",",
"7",
")",
"\n",
"size",
",",
"numCodegens",
":=",
"w",
".",
"dynamicSize",
"(",
"w",
".",
"literalEncoding",
",",
"huffOffset",
",",
"0",
")",
"\n\n",
"// Store bytes, if we don't get a reasonable improvement.",
"if",
"ssize",
",",
"storable",
":=",
"w",
".",
"storedSize",
"(",
"input",
")",
";",
"storable",
"&&",
"ssize",
"<",
"(",
"size",
"+",
"size",
">>",
"4",
")",
"{",
"w",
".",
"writeStoredHeader",
"(",
"len",
"(",
"input",
")",
",",
"eof",
")",
"\n",
"w",
".",
"writeBytes",
"(",
"input",
")",
"\n",
"return",
"\n",
"}",
"\n\n",
"// Huffman.",
"w",
".",
"writeDynamicHeader",
"(",
"numLiterals",
",",
"numOffsets",
",",
"numCodegens",
",",
"eof",
")",
"\n",
"encoding",
":=",
"w",
".",
"literalEncoding",
".",
"codes",
"[",
":",
"257",
"]",
"\n",
"n",
":=",
"w",
".",
"nbytes",
"\n",
"for",
"_",
",",
"t",
":=",
"range",
"input",
"{",
"// Bitwriting inlined, ~30% speedup",
"c",
":=",
"encoding",
"[",
"t",
"]",
"\n",
"w",
".",
"bits",
"|=",
"uint64",
"(",
"c",
".",
"code",
")",
"<<",
"w",
".",
"nbits",
"\n",
"w",
".",
"nbits",
"+=",
"uint",
"(",
"c",
".",
"len",
")",
"\n",
"if",
"w",
".",
"nbits",
"<",
"48",
"{",
"continue",
"\n",
"}",
"\n",
"// Store 6 bytes",
"bits",
":=",
"w",
".",
"bits",
"\n",
"w",
".",
"bits",
">>=",
"48",
"\n",
"w",
".",
"nbits",
"-=",
"48",
"\n",
"bytes",
":=",
"w",
".",
"bytes",
"[",
"n",
":",
"n",
"+",
"6",
"]",
"\n",
"bytes",
"[",
"0",
"]",
"=",
"byte",
"(",
"bits",
")",
"\n",
"bytes",
"[",
"1",
"]",
"=",
"byte",
"(",
"bits",
">>",
"8",
")",
"\n",
"bytes",
"[",
"2",
"]",
"=",
"byte",
"(",
"bits",
">>",
"16",
")",
"\n",
"bytes",
"[",
"3",
"]",
"=",
"byte",
"(",
"bits",
">>",
"24",
")",
"\n",
"bytes",
"[",
"4",
"]",
"=",
"byte",
"(",
"bits",
">>",
"32",
")",
"\n",
"bytes",
"[",
"5",
"]",
"=",
"byte",
"(",
"bits",
">>",
"40",
")",
"\n",
"n",
"+=",
"6",
"\n",
"if",
"n",
"<",
"bufferFlushSize",
"{",
"continue",
"\n",
"}",
"\n",
"w",
".",
"write",
"(",
"w",
".",
"bytes",
"[",
":",
"n",
"]",
")",
"\n",
"if",
"w",
".",
"err",
"!=",
"nil",
"{",
"return",
"// Return early in the event of write failures",
"\n",
"}",
"\n",
"n",
"=",
"0",
"\n",
"}",
"\n",
"w",
".",
"nbytes",
"=",
"n",
"\n",
"w",
".",
"writeCode",
"(",
"encoding",
"[",
"endBlockMarker",
"]",
")",
"\n",
"}"
]
| // writeBlockHuff encodes a block of bytes as either
// Huffman encoded literals or uncompressed bytes if the
// results only gains very little from compression. | [
"writeBlockHuff",
"encodes",
"a",
"block",
"of",
"bytes",
"as",
"either",
"Huffman",
"encoded",
"literals",
"or",
"uncompressed",
"bytes",
"if",
"the",
"results",
"only",
"gains",
"very",
"little",
"from",
"compression",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/flate/huffman_bit_writer.go#L629-L701 |
8,344 | klauspost/compress | zip/reader.go | OpenReader | func OpenReader(name string) (*ReadCloser, error) {
f, err := os.Open(name)
if err != nil {
return nil, err
}
fi, err := f.Stat()
if err != nil {
f.Close()
return nil, err
}
r := new(ReadCloser)
if err := r.init(f, fi.Size()); err != nil {
f.Close()
return nil, err
}
r.f = f
return r, nil
} | go | func OpenReader(name string) (*ReadCloser, error) {
f, err := os.Open(name)
if err != nil {
return nil, err
}
fi, err := f.Stat()
if err != nil {
f.Close()
return nil, err
}
r := new(ReadCloser)
if err := r.init(f, fi.Size()); err != nil {
f.Close()
return nil, err
}
r.f = f
return r, nil
} | [
"func",
"OpenReader",
"(",
"name",
"string",
")",
"(",
"*",
"ReadCloser",
",",
"error",
")",
"{",
"f",
",",
"err",
":=",
"os",
".",
"Open",
"(",
"name",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"return",
"nil",
",",
"err",
"\n",
"}",
"\n",
"fi",
",",
"err",
":=",
"f",
".",
"Stat",
"(",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"f",
".",
"Close",
"(",
")",
"\n",
"return",
"nil",
",",
"err",
"\n",
"}",
"\n",
"r",
":=",
"new",
"(",
"ReadCloser",
")",
"\n",
"if",
"err",
":=",
"r",
".",
"init",
"(",
"f",
",",
"fi",
".",
"Size",
"(",
")",
")",
";",
"err",
"!=",
"nil",
"{",
"f",
".",
"Close",
"(",
")",
"\n",
"return",
"nil",
",",
"err",
"\n",
"}",
"\n",
"r",
".",
"f",
"=",
"f",
"\n",
"return",
"r",
",",
"nil",
"\n",
"}"
]
| // OpenReader will open the Zip file specified by name and return a ReadCloser. | [
"OpenReader",
"will",
"open",
"the",
"Zip",
"file",
"specified",
"by",
"name",
"and",
"return",
"a",
"ReadCloser",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zip/reader.go#L49-L66 |
8,345 | klauspost/compress | zip/reader.go | NewReader | func NewReader(r io.ReaderAt, size int64) (*Reader, error) {
zr := new(Reader)
if err := zr.init(r, size); err != nil {
return nil, err
}
return zr, nil
} | go | func NewReader(r io.ReaderAt, size int64) (*Reader, error) {
zr := new(Reader)
if err := zr.init(r, size); err != nil {
return nil, err
}
return zr, nil
} | [
"func",
"NewReader",
"(",
"r",
"io",
".",
"ReaderAt",
",",
"size",
"int64",
")",
"(",
"*",
"Reader",
",",
"error",
")",
"{",
"zr",
":=",
"new",
"(",
"Reader",
")",
"\n",
"if",
"err",
":=",
"zr",
".",
"init",
"(",
"r",
",",
"size",
")",
";",
"err",
"!=",
"nil",
"{",
"return",
"nil",
",",
"err",
"\n",
"}",
"\n",
"return",
"zr",
",",
"nil",
"\n",
"}"
]
| // NewReader returns a new Reader reading from r, which is assumed to
// have the given size in bytes. | [
"NewReader",
"returns",
"a",
"new",
"Reader",
"reading",
"from",
"r",
"which",
"is",
"assumed",
"to",
"have",
"the",
"given",
"size",
"in",
"bytes",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zip/reader.go#L70-L76 |
8,346 | klauspost/compress | zip/reader.go | RegisterDecompressor | func (z *Reader) RegisterDecompressor(method uint16, dcomp Decompressor) {
if z.decompressors == nil {
z.decompressors = make(map[uint16]Decompressor)
}
z.decompressors[method] = dcomp
} | go | func (z *Reader) RegisterDecompressor(method uint16, dcomp Decompressor) {
if z.decompressors == nil {
z.decompressors = make(map[uint16]Decompressor)
}
z.decompressors[method] = dcomp
} | [
"func",
"(",
"z",
"*",
"Reader",
")",
"RegisterDecompressor",
"(",
"method",
"uint16",
",",
"dcomp",
"Decompressor",
")",
"{",
"if",
"z",
".",
"decompressors",
"==",
"nil",
"{",
"z",
".",
"decompressors",
"=",
"make",
"(",
"map",
"[",
"uint16",
"]",
"Decompressor",
")",
"\n",
"}",
"\n",
"z",
".",
"decompressors",
"[",
"method",
"]",
"=",
"dcomp",
"\n",
"}"
]
| // RegisterDecompressor registers or overrides a custom decompressor for a
// specific method ID. If a decompressor for a given method is not found,
// Reader will default to looking up the decompressor at the package level.
//
// Must not be called concurrently with Open on any Files in the Reader. | [
"RegisterDecompressor",
"registers",
"or",
"overrides",
"a",
"custom",
"decompressor",
"for",
"a",
"specific",
"method",
"ID",
".",
"If",
"a",
"decompressor",
"for",
"a",
"given",
"method",
"is",
"not",
"found",
"Reader",
"will",
"default",
"to",
"looking",
"up",
"the",
"decompressor",
"at",
"the",
"package",
"level",
".",
"Must",
"not",
"be",
"called",
"concurrently",
"with",
"Open",
"on",
"any",
"Files",
"in",
"the",
"Reader",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zip/reader.go#L123-L128 |
8,347 | klauspost/compress | zip/reader.go | DataOffset | func (f *File) DataOffset() (offset int64, err error) {
bodyOffset, err := f.findBodyOffset()
if err != nil {
return
}
return f.headerOffset + bodyOffset, nil
} | go | func (f *File) DataOffset() (offset int64, err error) {
bodyOffset, err := f.findBodyOffset()
if err != nil {
return
}
return f.headerOffset + bodyOffset, nil
} | [
"func",
"(",
"f",
"*",
"File",
")",
"DataOffset",
"(",
")",
"(",
"offset",
"int64",
",",
"err",
"error",
")",
"{",
"bodyOffset",
",",
"err",
":=",
"f",
".",
"findBodyOffset",
"(",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"return",
"\n",
"}",
"\n",
"return",
"f",
".",
"headerOffset",
"+",
"bodyOffset",
",",
"nil",
"\n",
"}"
]
| // DataOffset returns the offset of the file's possibly-compressed
// data, relative to the beginning of the zip file.
//
// Most callers should instead use Open, which transparently
// decompresses data and verifies checksums. | [
"DataOffset",
"returns",
"the",
"offset",
"of",
"the",
"file",
"s",
"possibly",
"-",
"compressed",
"data",
"relative",
"to",
"the",
"beginning",
"of",
"the",
"zip",
"file",
".",
"Most",
"callers",
"should",
"instead",
"use",
"Open",
"which",
"transparently",
"decompresses",
"data",
"and",
"verifies",
"checksums",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zip/reader.go#L148-L154 |
8,348 | klauspost/compress | zip/reader.go | Open | func (f *File) Open() (rc io.ReadCloser, err error) {
bodyOffset, err := f.findBodyOffset()
if err != nil {
return
}
size := int64(f.CompressedSize64)
r := io.NewSectionReader(f.zipr, f.headerOffset+bodyOffset, size)
dcomp := f.zip.decompressor(f.Method)
if dcomp == nil {
err = ErrAlgorithm
return
}
rc = dcomp(r)
var desr io.Reader
if f.hasDataDescriptor() {
desr = io.NewSectionReader(f.zipr, f.headerOffset+bodyOffset+size, dataDescriptorLen)
}
rc = &checksumReader{
rc: rc,
hash: crc32.NewIEEE(),
f: f,
desr: desr,
}
return
} | go | func (f *File) Open() (rc io.ReadCloser, err error) {
bodyOffset, err := f.findBodyOffset()
if err != nil {
return
}
size := int64(f.CompressedSize64)
r := io.NewSectionReader(f.zipr, f.headerOffset+bodyOffset, size)
dcomp := f.zip.decompressor(f.Method)
if dcomp == nil {
err = ErrAlgorithm
return
}
rc = dcomp(r)
var desr io.Reader
if f.hasDataDescriptor() {
desr = io.NewSectionReader(f.zipr, f.headerOffset+bodyOffset+size, dataDescriptorLen)
}
rc = &checksumReader{
rc: rc,
hash: crc32.NewIEEE(),
f: f,
desr: desr,
}
return
} | [
"func",
"(",
"f",
"*",
"File",
")",
"Open",
"(",
")",
"(",
"rc",
"io",
".",
"ReadCloser",
",",
"err",
"error",
")",
"{",
"bodyOffset",
",",
"err",
":=",
"f",
".",
"findBodyOffset",
"(",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"return",
"\n",
"}",
"\n",
"size",
":=",
"int64",
"(",
"f",
".",
"CompressedSize64",
")",
"\n",
"r",
":=",
"io",
".",
"NewSectionReader",
"(",
"f",
".",
"zipr",
",",
"f",
".",
"headerOffset",
"+",
"bodyOffset",
",",
"size",
")",
"\n",
"dcomp",
":=",
"f",
".",
"zip",
".",
"decompressor",
"(",
"f",
".",
"Method",
")",
"\n",
"if",
"dcomp",
"==",
"nil",
"{",
"err",
"=",
"ErrAlgorithm",
"\n",
"return",
"\n",
"}",
"\n",
"rc",
"=",
"dcomp",
"(",
"r",
")",
"\n",
"var",
"desr",
"io",
".",
"Reader",
"\n",
"if",
"f",
".",
"hasDataDescriptor",
"(",
")",
"{",
"desr",
"=",
"io",
".",
"NewSectionReader",
"(",
"f",
".",
"zipr",
",",
"f",
".",
"headerOffset",
"+",
"bodyOffset",
"+",
"size",
",",
"dataDescriptorLen",
")",
"\n",
"}",
"\n",
"rc",
"=",
"&",
"checksumReader",
"{",
"rc",
":",
"rc",
",",
"hash",
":",
"crc32",
".",
"NewIEEE",
"(",
")",
",",
"f",
":",
"f",
",",
"desr",
":",
"desr",
",",
"}",
"\n",
"return",
"\n",
"}"
]
| // Open returns a ReadCloser that provides access to the File's contents.
// Multiple files may be read concurrently. | [
"Open",
"returns",
"a",
"ReadCloser",
"that",
"provides",
"access",
"to",
"the",
"File",
"s",
"contents",
".",
"Multiple",
"files",
"may",
"be",
"read",
"concurrently",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zip/reader.go#L158-L182 |
8,349 | klauspost/compress | zip/reader.go | findBodyOffset | func (f *File) findBodyOffset() (int64, error) {
var buf [fileHeaderLen]byte
if _, err := f.zipr.ReadAt(buf[:], f.headerOffset); err != nil {
return 0, err
}
b := readBuf(buf[:])
if sig := b.uint32(); sig != fileHeaderSignature {
return 0, ErrFormat
}
b = b[22:] // skip over most of the header
filenameLen := int(b.uint16())
extraLen := int(b.uint16())
return int64(fileHeaderLen + filenameLen + extraLen), nil
} | go | func (f *File) findBodyOffset() (int64, error) {
var buf [fileHeaderLen]byte
if _, err := f.zipr.ReadAt(buf[:], f.headerOffset); err != nil {
return 0, err
}
b := readBuf(buf[:])
if sig := b.uint32(); sig != fileHeaderSignature {
return 0, ErrFormat
}
b = b[22:] // skip over most of the header
filenameLen := int(b.uint16())
extraLen := int(b.uint16())
return int64(fileHeaderLen + filenameLen + extraLen), nil
} | [
"func",
"(",
"f",
"*",
"File",
")",
"findBodyOffset",
"(",
")",
"(",
"int64",
",",
"error",
")",
"{",
"var",
"buf",
"[",
"fileHeaderLen",
"]",
"byte",
"\n",
"if",
"_",
",",
"err",
":=",
"f",
".",
"zipr",
".",
"ReadAt",
"(",
"buf",
"[",
":",
"]",
",",
"f",
".",
"headerOffset",
")",
";",
"err",
"!=",
"nil",
"{",
"return",
"0",
",",
"err",
"\n",
"}",
"\n",
"b",
":=",
"readBuf",
"(",
"buf",
"[",
":",
"]",
")",
"\n",
"if",
"sig",
":=",
"b",
".",
"uint32",
"(",
")",
";",
"sig",
"!=",
"fileHeaderSignature",
"{",
"return",
"0",
",",
"ErrFormat",
"\n",
"}",
"\n",
"b",
"=",
"b",
"[",
"22",
":",
"]",
"// skip over most of the header",
"\n",
"filenameLen",
":=",
"int",
"(",
"b",
".",
"uint16",
"(",
")",
")",
"\n",
"extraLen",
":=",
"int",
"(",
"b",
".",
"uint16",
"(",
")",
")",
"\n",
"return",
"int64",
"(",
"fileHeaderLen",
"+",
"filenameLen",
"+",
"extraLen",
")",
",",
"nil",
"\n",
"}"
]
| // findBodyOffset does the minimum work to verify the file has a header
// and returns the file body offset. | [
"findBodyOffset",
"does",
"the",
"minimum",
"work",
"to",
"verify",
"the",
"file",
"has",
"a",
"header",
"and",
"returns",
"the",
"file",
"body",
"offset",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zip/reader.go#L234-L247 |
8,350 | klauspost/compress | zip/reader.go | readDirectory64End | func readDirectory64End(r io.ReaderAt, offset int64, d *directoryEnd) (err error) {
buf := make([]byte, directory64EndLen)
if _, err := r.ReadAt(buf, offset); err != nil {
return err
}
b := readBuf(buf)
if sig := b.uint32(); sig != directory64EndSignature {
return ErrFormat
}
b = b[12:] // skip dir size, version and version needed (uint64 + 2x uint16)
d.diskNbr = b.uint32() // number of this disk
d.dirDiskNbr = b.uint32() // number of the disk with the start of the central directory
d.dirRecordsThisDisk = b.uint64() // total number of entries in the central directory on this disk
d.directoryRecords = b.uint64() // total number of entries in the central directory
d.directorySize = b.uint64() // size of the central directory
d.directoryOffset = b.uint64() // offset of start of central directory with respect to the starting disk number
return nil
} | go | func readDirectory64End(r io.ReaderAt, offset int64, d *directoryEnd) (err error) {
buf := make([]byte, directory64EndLen)
if _, err := r.ReadAt(buf, offset); err != nil {
return err
}
b := readBuf(buf)
if sig := b.uint32(); sig != directory64EndSignature {
return ErrFormat
}
b = b[12:] // skip dir size, version and version needed (uint64 + 2x uint16)
d.diskNbr = b.uint32() // number of this disk
d.dirDiskNbr = b.uint32() // number of the disk with the start of the central directory
d.dirRecordsThisDisk = b.uint64() // total number of entries in the central directory on this disk
d.directoryRecords = b.uint64() // total number of entries in the central directory
d.directorySize = b.uint64() // size of the central directory
d.directoryOffset = b.uint64() // offset of start of central directory with respect to the starting disk number
return nil
} | [
"func",
"readDirectory64End",
"(",
"r",
"io",
".",
"ReaderAt",
",",
"offset",
"int64",
",",
"d",
"*",
"directoryEnd",
")",
"(",
"err",
"error",
")",
"{",
"buf",
":=",
"make",
"(",
"[",
"]",
"byte",
",",
"directory64EndLen",
")",
"\n",
"if",
"_",
",",
"err",
":=",
"r",
".",
"ReadAt",
"(",
"buf",
",",
"offset",
")",
";",
"err",
"!=",
"nil",
"{",
"return",
"err",
"\n",
"}",
"\n\n",
"b",
":=",
"readBuf",
"(",
"buf",
")",
"\n",
"if",
"sig",
":=",
"b",
".",
"uint32",
"(",
")",
";",
"sig",
"!=",
"directory64EndSignature",
"{",
"return",
"ErrFormat",
"\n",
"}",
"\n\n",
"b",
"=",
"b",
"[",
"12",
":",
"]",
"// skip dir size, version and version needed (uint64 + 2x uint16)",
"\n",
"d",
".",
"diskNbr",
"=",
"b",
".",
"uint32",
"(",
")",
"// number of this disk",
"\n",
"d",
".",
"dirDiskNbr",
"=",
"b",
".",
"uint32",
"(",
")",
"// number of the disk with the start of the central directory",
"\n",
"d",
".",
"dirRecordsThisDisk",
"=",
"b",
".",
"uint64",
"(",
")",
"// total number of entries in the central directory on this disk",
"\n",
"d",
".",
"directoryRecords",
"=",
"b",
".",
"uint64",
"(",
")",
"// total number of entries in the central directory",
"\n",
"d",
".",
"directorySize",
"=",
"b",
".",
"uint64",
"(",
")",
"// size of the central directory",
"\n",
"d",
".",
"directoryOffset",
"=",
"b",
".",
"uint64",
"(",
")",
"// offset of start of central directory with respect to the starting disk number",
"\n\n",
"return",
"nil",
"\n",
"}"
]
| // readDirectory64End reads the zip64 directory end and updates the
// directory end with the zip64 directory end values. | [
"readDirectory64End",
"reads",
"the",
"zip64",
"directory",
"end",
"and",
"updates",
"the",
"directory",
"end",
"with",
"the",
"zip64",
"directory",
"end",
"values",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zip/reader.go#L466-L486 |
8,351 | klauspost/compress | zip/struct.go | FileInfoHeader | func FileInfoHeader(fi os.FileInfo) (*FileHeader, error) {
size := fi.Size()
fh := &FileHeader{
Name: fi.Name(),
UncompressedSize64: uint64(size),
}
fh.SetModTime(fi.ModTime())
fh.SetMode(fi.Mode())
if fh.UncompressedSize64 > uint32max {
fh.UncompressedSize = uint32max
} else {
fh.UncompressedSize = uint32(fh.UncompressedSize64)
}
return fh, nil
} | go | func FileInfoHeader(fi os.FileInfo) (*FileHeader, error) {
size := fi.Size()
fh := &FileHeader{
Name: fi.Name(),
UncompressedSize64: uint64(size),
}
fh.SetModTime(fi.ModTime())
fh.SetMode(fi.Mode())
if fh.UncompressedSize64 > uint32max {
fh.UncompressedSize = uint32max
} else {
fh.UncompressedSize = uint32(fh.UncompressedSize64)
}
return fh, nil
} | [
"func",
"FileInfoHeader",
"(",
"fi",
"os",
".",
"FileInfo",
")",
"(",
"*",
"FileHeader",
",",
"error",
")",
"{",
"size",
":=",
"fi",
".",
"Size",
"(",
")",
"\n",
"fh",
":=",
"&",
"FileHeader",
"{",
"Name",
":",
"fi",
".",
"Name",
"(",
")",
",",
"UncompressedSize64",
":",
"uint64",
"(",
"size",
")",
",",
"}",
"\n",
"fh",
".",
"SetModTime",
"(",
"fi",
".",
"ModTime",
"(",
")",
")",
"\n",
"fh",
".",
"SetMode",
"(",
"fi",
".",
"Mode",
"(",
")",
")",
"\n",
"if",
"fh",
".",
"UncompressedSize64",
">",
"uint32max",
"{",
"fh",
".",
"UncompressedSize",
"=",
"uint32max",
"\n",
"}",
"else",
"{",
"fh",
".",
"UncompressedSize",
"=",
"uint32",
"(",
"fh",
".",
"UncompressedSize64",
")",
"\n",
"}",
"\n",
"return",
"fh",
",",
"nil",
"\n",
"}"
]
| // FileInfoHeader creates a partially-populated FileHeader from an
// os.FileInfo.
// Because os.FileInfo's Name method returns only the base name of
// the file it describes, it may be necessary to modify the Name field
// of the returned header to provide the full path name of the file. | [
"FileInfoHeader",
"creates",
"a",
"partially",
"-",
"populated",
"FileHeader",
"from",
"an",
"os",
".",
"FileInfo",
".",
"Because",
"os",
".",
"FileInfo",
"s",
"Name",
"method",
"returns",
"only",
"the",
"base",
"name",
"of",
"the",
"file",
"it",
"describes",
"it",
"may",
"be",
"necessary",
"to",
"modify",
"the",
"Name",
"field",
"of",
"the",
"returned",
"header",
"to",
"provide",
"the",
"full",
"path",
"name",
"of",
"the",
"file",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zip/struct.go#L120-L134 |
8,352 | klauspost/compress | zip/struct.go | ModTime | func (h *FileHeader) ModTime() time.Time {
return msDosTimeToTime(h.ModifiedDate, h.ModifiedTime)
} | go | func (h *FileHeader) ModTime() time.Time {
return msDosTimeToTime(h.ModifiedDate, h.ModifiedTime)
} | [
"func",
"(",
"h",
"*",
"FileHeader",
")",
"ModTime",
"(",
")",
"time",
".",
"Time",
"{",
"return",
"msDosTimeToTime",
"(",
"h",
".",
"ModifiedDate",
",",
"h",
".",
"ModifiedTime",
")",
"\n",
"}"
]
| // ModTime returns the modification time in UTC.
// The resolution is 2s. | [
"ModTime",
"returns",
"the",
"modification",
"time",
"in",
"UTC",
".",
"The",
"resolution",
"is",
"2s",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zip/struct.go#L179-L181 |
8,353 | klauspost/compress | zip/struct.go | SetModTime | func (h *FileHeader) SetModTime(t time.Time) {
h.ModifiedDate, h.ModifiedTime = timeToMsDosTime(t)
} | go | func (h *FileHeader) SetModTime(t time.Time) {
h.ModifiedDate, h.ModifiedTime = timeToMsDosTime(t)
} | [
"func",
"(",
"h",
"*",
"FileHeader",
")",
"SetModTime",
"(",
"t",
"time",
".",
"Time",
")",
"{",
"h",
".",
"ModifiedDate",
",",
"h",
".",
"ModifiedTime",
"=",
"timeToMsDosTime",
"(",
"t",
")",
"\n",
"}"
]
| // SetModTime sets the ModifiedTime and ModifiedDate fields to the given time in UTC.
// The resolution is 2s. | [
"SetModTime",
"sets",
"the",
"ModifiedTime",
"and",
"ModifiedDate",
"fields",
"to",
"the",
"given",
"time",
"in",
"UTC",
".",
"The",
"resolution",
"is",
"2s",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zip/struct.go#L185-L187 |
8,354 | klauspost/compress | zip/struct.go | Mode | func (h *FileHeader) Mode() (mode os.FileMode) {
switch h.CreatorVersion >> 8 {
case creatorUnix, creatorMacOSX:
mode = unixModeToFileMode(h.ExternalAttrs >> 16)
case creatorNTFS, creatorVFAT, creatorFAT:
mode = msdosModeToFileMode(h.ExternalAttrs)
}
if len(h.Name) > 0 && h.Name[len(h.Name)-1] == '/' {
mode |= os.ModeDir
}
return mode
} | go | func (h *FileHeader) Mode() (mode os.FileMode) {
switch h.CreatorVersion >> 8 {
case creatorUnix, creatorMacOSX:
mode = unixModeToFileMode(h.ExternalAttrs >> 16)
case creatorNTFS, creatorVFAT, creatorFAT:
mode = msdosModeToFileMode(h.ExternalAttrs)
}
if len(h.Name) > 0 && h.Name[len(h.Name)-1] == '/' {
mode |= os.ModeDir
}
return mode
} | [
"func",
"(",
"h",
"*",
"FileHeader",
")",
"Mode",
"(",
")",
"(",
"mode",
"os",
".",
"FileMode",
")",
"{",
"switch",
"h",
".",
"CreatorVersion",
">>",
"8",
"{",
"case",
"creatorUnix",
",",
"creatorMacOSX",
":",
"mode",
"=",
"unixModeToFileMode",
"(",
"h",
".",
"ExternalAttrs",
">>",
"16",
")",
"\n",
"case",
"creatorNTFS",
",",
"creatorVFAT",
",",
"creatorFAT",
":",
"mode",
"=",
"msdosModeToFileMode",
"(",
"h",
".",
"ExternalAttrs",
")",
"\n",
"}",
"\n",
"if",
"len",
"(",
"h",
".",
"Name",
")",
">",
"0",
"&&",
"h",
".",
"Name",
"[",
"len",
"(",
"h",
".",
"Name",
")",
"-",
"1",
"]",
"==",
"'/'",
"{",
"mode",
"|=",
"os",
".",
"ModeDir",
"\n",
"}",
"\n",
"return",
"mode",
"\n",
"}"
]
| // Mode returns the permission and mode bits for the FileHeader. | [
"Mode",
"returns",
"the",
"permission",
"and",
"mode",
"bits",
"for",
"the",
"FileHeader",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zip/struct.go#L209-L220 |
8,355 | klauspost/compress | zip/struct.go | SetMode | func (h *FileHeader) SetMode(mode os.FileMode) {
h.CreatorVersion = h.CreatorVersion&0xff | creatorUnix<<8
h.ExternalAttrs = fileModeToUnixMode(mode) << 16
// set MSDOS attributes too, as the original zip does.
if mode&os.ModeDir != 0 {
h.ExternalAttrs |= msdosDir
}
if mode&0200 == 0 {
h.ExternalAttrs |= msdosReadOnly
}
} | go | func (h *FileHeader) SetMode(mode os.FileMode) {
h.CreatorVersion = h.CreatorVersion&0xff | creatorUnix<<8
h.ExternalAttrs = fileModeToUnixMode(mode) << 16
// set MSDOS attributes too, as the original zip does.
if mode&os.ModeDir != 0 {
h.ExternalAttrs |= msdosDir
}
if mode&0200 == 0 {
h.ExternalAttrs |= msdosReadOnly
}
} | [
"func",
"(",
"h",
"*",
"FileHeader",
")",
"SetMode",
"(",
"mode",
"os",
".",
"FileMode",
")",
"{",
"h",
".",
"CreatorVersion",
"=",
"h",
".",
"CreatorVersion",
"&",
"0xff",
"|",
"creatorUnix",
"<<",
"8",
"\n",
"h",
".",
"ExternalAttrs",
"=",
"fileModeToUnixMode",
"(",
"mode",
")",
"<<",
"16",
"\n\n",
"// set MSDOS attributes too, as the original zip does.",
"if",
"mode",
"&",
"os",
".",
"ModeDir",
"!=",
"0",
"{",
"h",
".",
"ExternalAttrs",
"|=",
"msdosDir",
"\n",
"}",
"\n",
"if",
"mode",
"&",
"0200",
"==",
"0",
"{",
"h",
".",
"ExternalAttrs",
"|=",
"msdosReadOnly",
"\n",
"}",
"\n",
"}"
]
| // SetMode changes the permission and mode bits for the FileHeader. | [
"SetMode",
"changes",
"the",
"permission",
"and",
"mode",
"bits",
"for",
"the",
"FileHeader",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zip/struct.go#L223-L234 |
8,356 | klauspost/compress | zip/struct.go | isZip64 | func (fh *FileHeader) isZip64() bool {
return fh.CompressedSize64 >= uint32max || fh.UncompressedSize64 >= uint32max
} | go | func (fh *FileHeader) isZip64() bool {
return fh.CompressedSize64 >= uint32max || fh.UncompressedSize64 >= uint32max
} | [
"func",
"(",
"fh",
"*",
"FileHeader",
")",
"isZip64",
"(",
")",
"bool",
"{",
"return",
"fh",
".",
"CompressedSize64",
">=",
"uint32max",
"||",
"fh",
".",
"UncompressedSize64",
">=",
"uint32max",
"\n",
"}"
]
| // isZip64 reports whether the file size exceeds the 32 bit limit | [
"isZip64",
"reports",
"whether",
"the",
"file",
"size",
"exceeds",
"the",
"32",
"bit",
"limit"
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zip/struct.go#L237-L239 |
8,357 | klauspost/compress | flate/huffman_code.go | set | func (h *hcode) set(code uint16, length uint16) {
h.len = length
h.code = code
} | go | func (h *hcode) set(code uint16, length uint16) {
h.len = length
h.code = code
} | [
"func",
"(",
"h",
"*",
"hcode",
")",
"set",
"(",
"code",
"uint16",
",",
"length",
"uint16",
")",
"{",
"h",
".",
"len",
"=",
"length",
"\n",
"h",
".",
"code",
"=",
"code",
"\n",
"}"
]
| // set sets the code and length of an hcode. | [
"set",
"sets",
"the",
"code",
"and",
"length",
"of",
"an",
"hcode",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/flate/huffman_code.go#L51-L54 |
8,358 | klauspost/compress | flate/huffman_code.go | generateFixedLiteralEncoding | func generateFixedLiteralEncoding() *huffmanEncoder {
h := newHuffmanEncoder(maxNumLit)
codes := h.codes
var ch uint16
for ch = 0; ch < maxNumLit; ch++ {
var bits uint16
var size uint16
switch {
case ch < 144:
// size 8, 000110000 .. 10111111
bits = ch + 48
size = 8
break
case ch < 256:
// size 9, 110010000 .. 111111111
bits = ch + 400 - 144
size = 9
break
case ch < 280:
// size 7, 0000000 .. 0010111
bits = ch - 256
size = 7
break
default:
// size 8, 11000000 .. 11000111
bits = ch + 192 - 280
size = 8
}
codes[ch] = hcode{code: reverseBits(bits, byte(size)), len: size}
}
return h
} | go | func generateFixedLiteralEncoding() *huffmanEncoder {
h := newHuffmanEncoder(maxNumLit)
codes := h.codes
var ch uint16
for ch = 0; ch < maxNumLit; ch++ {
var bits uint16
var size uint16
switch {
case ch < 144:
// size 8, 000110000 .. 10111111
bits = ch + 48
size = 8
break
case ch < 256:
// size 9, 110010000 .. 111111111
bits = ch + 400 - 144
size = 9
break
case ch < 280:
// size 7, 0000000 .. 0010111
bits = ch - 256
size = 7
break
default:
// size 8, 11000000 .. 11000111
bits = ch + 192 - 280
size = 8
}
codes[ch] = hcode{code: reverseBits(bits, byte(size)), len: size}
}
return h
} | [
"func",
"generateFixedLiteralEncoding",
"(",
")",
"*",
"huffmanEncoder",
"{",
"h",
":=",
"newHuffmanEncoder",
"(",
"maxNumLit",
")",
"\n",
"codes",
":=",
"h",
".",
"codes",
"\n",
"var",
"ch",
"uint16",
"\n",
"for",
"ch",
"=",
"0",
";",
"ch",
"<",
"maxNumLit",
";",
"ch",
"++",
"{",
"var",
"bits",
"uint16",
"\n",
"var",
"size",
"uint16",
"\n",
"switch",
"{",
"case",
"ch",
"<",
"144",
":",
"// size 8, 000110000 .. 10111111",
"bits",
"=",
"ch",
"+",
"48",
"\n",
"size",
"=",
"8",
"\n",
"break",
"\n",
"case",
"ch",
"<",
"256",
":",
"// size 9, 110010000 .. 111111111",
"bits",
"=",
"ch",
"+",
"400",
"-",
"144",
"\n",
"size",
"=",
"9",
"\n",
"break",
"\n",
"case",
"ch",
"<",
"280",
":",
"// size 7, 0000000 .. 0010111",
"bits",
"=",
"ch",
"-",
"256",
"\n",
"size",
"=",
"7",
"\n",
"break",
"\n",
"default",
":",
"// size 8, 11000000 .. 11000111",
"bits",
"=",
"ch",
"+",
"192",
"-",
"280",
"\n",
"size",
"=",
"8",
"\n",
"}",
"\n",
"codes",
"[",
"ch",
"]",
"=",
"hcode",
"{",
"code",
":",
"reverseBits",
"(",
"bits",
",",
"byte",
"(",
"size",
")",
")",
",",
"len",
":",
"size",
"}",
"\n",
"}",
"\n",
"return",
"h",
"\n",
"}"
]
| // Generates a HuffmanCode corresponding to the fixed literal table | [
"Generates",
"a",
"HuffmanCode",
"corresponding",
"to",
"the",
"fixed",
"literal",
"table"
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/flate/huffman_code.go#L63-L94 |
8,359 | klauspost/compress | flate/huffman_code.go | assignEncodingAndSize | func (h *huffmanEncoder) assignEncodingAndSize(bitCount []int32, list []literalNode) {
code := uint16(0)
for n, bits := range bitCount {
code <<= 1
if n == 0 || bits == 0 {
continue
}
// The literals list[len(list)-bits] .. list[len(list)-bits]
// are encoded using "bits" bits, and get the values
// code, code + 1, .... The code values are
// assigned in literal order (not frequency order).
chunk := list[len(list)-int(bits):]
h.lns.sort(chunk)
for _, node := range chunk {
h.codes[node.literal] = hcode{code: reverseBits(code, uint8(n)), len: uint16(n)}
code++
}
list = list[0 : len(list)-int(bits)]
}
} | go | func (h *huffmanEncoder) assignEncodingAndSize(bitCount []int32, list []literalNode) {
code := uint16(0)
for n, bits := range bitCount {
code <<= 1
if n == 0 || bits == 0 {
continue
}
// The literals list[len(list)-bits] .. list[len(list)-bits]
// are encoded using "bits" bits, and get the values
// code, code + 1, .... The code values are
// assigned in literal order (not frequency order).
chunk := list[len(list)-int(bits):]
h.lns.sort(chunk)
for _, node := range chunk {
h.codes[node.literal] = hcode{code: reverseBits(code, uint8(n)), len: uint16(n)}
code++
}
list = list[0 : len(list)-int(bits)]
}
} | [
"func",
"(",
"h",
"*",
"huffmanEncoder",
")",
"assignEncodingAndSize",
"(",
"bitCount",
"[",
"]",
"int32",
",",
"list",
"[",
"]",
"literalNode",
")",
"{",
"code",
":=",
"uint16",
"(",
"0",
")",
"\n",
"for",
"n",
",",
"bits",
":=",
"range",
"bitCount",
"{",
"code",
"<<=",
"1",
"\n",
"if",
"n",
"==",
"0",
"||",
"bits",
"==",
"0",
"{",
"continue",
"\n",
"}",
"\n",
"// The literals list[len(list)-bits] .. list[len(list)-bits]",
"// are encoded using \"bits\" bits, and get the values",
"// code, code + 1, .... The code values are",
"// assigned in literal order (not frequency order).",
"chunk",
":=",
"list",
"[",
"len",
"(",
"list",
")",
"-",
"int",
"(",
"bits",
")",
":",
"]",
"\n\n",
"h",
".",
"lns",
".",
"sort",
"(",
"chunk",
")",
"\n",
"for",
"_",
",",
"node",
":=",
"range",
"chunk",
"{",
"h",
".",
"codes",
"[",
"node",
".",
"literal",
"]",
"=",
"hcode",
"{",
"code",
":",
"reverseBits",
"(",
"code",
",",
"uint8",
"(",
"n",
")",
")",
",",
"len",
":",
"uint16",
"(",
"n",
")",
"}",
"\n",
"code",
"++",
"\n",
"}",
"\n",
"list",
"=",
"list",
"[",
"0",
":",
"len",
"(",
"list",
")",
"-",
"int",
"(",
"bits",
")",
"]",
"\n",
"}",
"\n",
"}"
]
| // Look at the leaves and assign them a bit count and an encoding as specified
// in RFC 1951 3.2.2 | [
"Look",
"at",
"the",
"leaves",
"and",
"assign",
"them",
"a",
"bit",
"count",
"and",
"an",
"encoding",
"as",
"specified",
"in",
"RFC",
"1951",
"3",
".",
"2",
".",
"2"
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/flate/huffman_code.go#L247-L267 |
8,360 | klauspost/compress | flate/token.go | offsetCode | func offsetCode(off uint32) uint32 {
if off < uint32(len(offsetCodes)) {
return offsetCodes[off]
} else if off>>7 < uint32(len(offsetCodes)) {
return offsetCodes[off>>7] + 14
} else {
return offsetCodes[off>>14] + 28
}
} | go | func offsetCode(off uint32) uint32 {
if off < uint32(len(offsetCodes)) {
return offsetCodes[off]
} else if off>>7 < uint32(len(offsetCodes)) {
return offsetCodes[off>>7] + 14
} else {
return offsetCodes[off>>14] + 28
}
} | [
"func",
"offsetCode",
"(",
"off",
"uint32",
")",
"uint32",
"{",
"if",
"off",
"<",
"uint32",
"(",
"len",
"(",
"offsetCodes",
")",
")",
"{",
"return",
"offsetCodes",
"[",
"off",
"]",
"\n",
"}",
"else",
"if",
"off",
">>",
"7",
"<",
"uint32",
"(",
"len",
"(",
"offsetCodes",
")",
")",
"{",
"return",
"offsetCodes",
"[",
"off",
">>",
"7",
"]",
"+",
"14",
"\n",
"}",
"else",
"{",
"return",
"offsetCodes",
"[",
"off",
">>",
"14",
"]",
"+",
"28",
"\n",
"}",
"\n",
"}"
]
| // Returns the offset code corresponding to a specific offset | [
"Returns",
"the",
"offset",
"code",
"corresponding",
"to",
"a",
"specific",
"offset"
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/flate/token.go#L107-L115 |
8,361 | klauspost/compress | flate/dict_decoder.go | init | func (dd *dictDecoder) init(size int, dict []byte) {
*dd = dictDecoder{hist: dd.hist}
if cap(dd.hist) < size {
dd.hist = make([]byte, size)
}
dd.hist = dd.hist[:size]
if len(dict) > len(dd.hist) {
dict = dict[len(dict)-len(dd.hist):]
}
dd.wrPos = copy(dd.hist, dict)
if dd.wrPos == len(dd.hist) {
dd.wrPos = 0
dd.full = true
}
dd.rdPos = dd.wrPos
} | go | func (dd *dictDecoder) init(size int, dict []byte) {
*dd = dictDecoder{hist: dd.hist}
if cap(dd.hist) < size {
dd.hist = make([]byte, size)
}
dd.hist = dd.hist[:size]
if len(dict) > len(dd.hist) {
dict = dict[len(dict)-len(dd.hist):]
}
dd.wrPos = copy(dd.hist, dict)
if dd.wrPos == len(dd.hist) {
dd.wrPos = 0
dd.full = true
}
dd.rdPos = dd.wrPos
} | [
"func",
"(",
"dd",
"*",
"dictDecoder",
")",
"init",
"(",
"size",
"int",
",",
"dict",
"[",
"]",
"byte",
")",
"{",
"*",
"dd",
"=",
"dictDecoder",
"{",
"hist",
":",
"dd",
".",
"hist",
"}",
"\n\n",
"if",
"cap",
"(",
"dd",
".",
"hist",
")",
"<",
"size",
"{",
"dd",
".",
"hist",
"=",
"make",
"(",
"[",
"]",
"byte",
",",
"size",
")",
"\n",
"}",
"\n",
"dd",
".",
"hist",
"=",
"dd",
".",
"hist",
"[",
":",
"size",
"]",
"\n\n",
"if",
"len",
"(",
"dict",
")",
">",
"len",
"(",
"dd",
".",
"hist",
")",
"{",
"dict",
"=",
"dict",
"[",
"len",
"(",
"dict",
")",
"-",
"len",
"(",
"dd",
".",
"hist",
")",
":",
"]",
"\n",
"}",
"\n",
"dd",
".",
"wrPos",
"=",
"copy",
"(",
"dd",
".",
"hist",
",",
"dict",
")",
"\n",
"if",
"dd",
".",
"wrPos",
"==",
"len",
"(",
"dd",
".",
"hist",
")",
"{",
"dd",
".",
"wrPos",
"=",
"0",
"\n",
"dd",
".",
"full",
"=",
"true",
"\n",
"}",
"\n",
"dd",
".",
"rdPos",
"=",
"dd",
".",
"wrPos",
"\n",
"}"
]
| // init initializes dictDecoder to have a sliding window dictionary of the given
// size. If a preset dict is provided, it will initialize the dictionary with
// the contents of dict. | [
"init",
"initializes",
"dictDecoder",
"to",
"have",
"a",
"sliding",
"window",
"dictionary",
"of",
"the",
"given",
"size",
".",
"If",
"a",
"preset",
"dict",
"is",
"provided",
"it",
"will",
"initialize",
"the",
"dictionary",
"with",
"the",
"contents",
"of",
"dict",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/flate/dict_decoder.go#L39-L56 |
8,362 | klauspost/compress | flate/dict_decoder.go | histSize | func (dd *dictDecoder) histSize() int {
if dd.full {
return len(dd.hist)
}
return dd.wrPos
} | go | func (dd *dictDecoder) histSize() int {
if dd.full {
return len(dd.hist)
}
return dd.wrPos
} | [
"func",
"(",
"dd",
"*",
"dictDecoder",
")",
"histSize",
"(",
")",
"int",
"{",
"if",
"dd",
".",
"full",
"{",
"return",
"len",
"(",
"dd",
".",
"hist",
")",
"\n",
"}",
"\n",
"return",
"dd",
".",
"wrPos",
"\n",
"}"
]
| // histSize reports the total amount of historical data in the dictionary. | [
"histSize",
"reports",
"the",
"total",
"amount",
"of",
"historical",
"data",
"in",
"the",
"dictionary",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/flate/dict_decoder.go#L59-L64 |
8,363 | klauspost/compress | flate/dict_decoder.go | readFlush | func (dd *dictDecoder) readFlush() []byte {
toRead := dd.hist[dd.rdPos:dd.wrPos]
dd.rdPos = dd.wrPos
if dd.wrPos == len(dd.hist) {
dd.wrPos, dd.rdPos = 0, 0
dd.full = true
}
return toRead
} | go | func (dd *dictDecoder) readFlush() []byte {
toRead := dd.hist[dd.rdPos:dd.wrPos]
dd.rdPos = dd.wrPos
if dd.wrPos == len(dd.hist) {
dd.wrPos, dd.rdPos = 0, 0
dd.full = true
}
return toRead
} | [
"func",
"(",
"dd",
"*",
"dictDecoder",
")",
"readFlush",
"(",
")",
"[",
"]",
"byte",
"{",
"toRead",
":=",
"dd",
".",
"hist",
"[",
"dd",
".",
"rdPos",
":",
"dd",
".",
"wrPos",
"]",
"\n",
"dd",
".",
"rdPos",
"=",
"dd",
".",
"wrPos",
"\n",
"if",
"dd",
".",
"wrPos",
"==",
"len",
"(",
"dd",
".",
"hist",
")",
"{",
"dd",
".",
"wrPos",
",",
"dd",
".",
"rdPos",
"=",
"0",
",",
"0",
"\n",
"dd",
".",
"full",
"=",
"true",
"\n",
"}",
"\n",
"return",
"toRead",
"\n",
"}"
]
| // readFlush returns a slice of the historical buffer that is ready to be
// emitted to the user. The data returned by readFlush must be fully consumed
// before calling any other dictDecoder methods. | [
"readFlush",
"returns",
"a",
"slice",
"of",
"the",
"historical",
"buffer",
"that",
"is",
"ready",
"to",
"be",
"emitted",
"to",
"the",
"user",
".",
"The",
"data",
"returned",
"by",
"readFlush",
"must",
"be",
"fully",
"consumed",
"before",
"calling",
"any",
"other",
"dictDecoder",
"methods",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/flate/dict_decoder.go#L176-L184 |
8,364 | klauspost/compress | flate/snappy.go | Reset | func (e *snappyGen) Reset() {
e.prev = e.prev[:0]
e.cur += maxMatchOffset
} | go | func (e *snappyGen) Reset() {
e.prev = e.prev[:0]
e.cur += maxMatchOffset
} | [
"func",
"(",
"e",
"*",
"snappyGen",
")",
"Reset",
"(",
")",
"{",
"e",
".",
"prev",
"=",
"e",
".",
"prev",
"[",
":",
"0",
"]",
"\n",
"e",
".",
"cur",
"+=",
"maxMatchOffset",
"\n",
"}"
]
| // Reset the encoding table. | [
"Reset",
"the",
"encoding",
"table",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/flate/snappy.go#L897-L900 |
8,365 | klauspost/compress | zstd/decoder_options.go | WithDecoderLowmem | func WithDecoderLowmem(b bool) DOption {
return func(o *decoderOptions) error { o.lowMem = b; return nil }
} | go | func WithDecoderLowmem(b bool) DOption {
return func(o *decoderOptions) error { o.lowMem = b; return nil }
} | [
"func",
"WithDecoderLowmem",
"(",
"b",
"bool",
")",
"DOption",
"{",
"return",
"func",
"(",
"o",
"*",
"decoderOptions",
")",
"error",
"{",
"o",
".",
"lowMem",
"=",
"b",
";",
"return",
"nil",
"}",
"\n",
"}"
]
| // WithDecoderLowmem will set whether to use a lower amount of memory,
// but possibly have to allocate more while running. | [
"WithDecoderLowmem",
"will",
"set",
"whether",
"to",
"use",
"a",
"lower",
"amount",
"of",
"memory",
"but",
"possibly",
"have",
"to",
"allocate",
"more",
"while",
"running",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zstd/decoder_options.go#L34-L36 |
8,366 | klauspost/compress | zstd/decoder_options.go | WithDecoderConcurrency | func WithDecoderConcurrency(n int) DOption {
return func(o *decoderOptions) error {
if n <= 0 {
return fmt.Errorf("Concurrency must be at least 1")
}
o.concurrent = n
return nil
}
} | go | func WithDecoderConcurrency(n int) DOption {
return func(o *decoderOptions) error {
if n <= 0 {
return fmt.Errorf("Concurrency must be at least 1")
}
o.concurrent = n
return nil
}
} | [
"func",
"WithDecoderConcurrency",
"(",
"n",
"int",
")",
"DOption",
"{",
"return",
"func",
"(",
"o",
"*",
"decoderOptions",
")",
"error",
"{",
"if",
"n",
"<=",
"0",
"{",
"return",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
")",
"\n",
"}",
"\n",
"o",
".",
"concurrent",
"=",
"n",
"\n",
"return",
"nil",
"\n",
"}",
"\n",
"}"
]
| // WithDecoderConcurrency will set the concurrency,
// meaning the maximum number of decoders to run concurrently.
// The value supplied must be at least 1.
// By default this will be set to GOMAXPROCS. | [
"WithDecoderConcurrency",
"will",
"set",
"the",
"concurrency",
"meaning",
"the",
"maximum",
"number",
"of",
"decoders",
"to",
"run",
"concurrently",
".",
"The",
"value",
"supplied",
"must",
"be",
"at",
"least",
"1",
".",
"By",
"default",
"this",
"will",
"be",
"set",
"to",
"GOMAXPROCS",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zstd/decoder_options.go#L42-L50 |
8,367 | klauspost/compress | fse/fse.go | HistogramFinished | func (s *Scratch) HistogramFinished(maxSymbol uint8, maxCount int) {
s.maxCount = maxCount
s.symbolLen = uint16(maxSymbol) + 1
s.clearCount = maxCount != 0
} | go | func (s *Scratch) HistogramFinished(maxSymbol uint8, maxCount int) {
s.maxCount = maxCount
s.symbolLen = uint16(maxSymbol) + 1
s.clearCount = maxCount != 0
} | [
"func",
"(",
"s",
"*",
"Scratch",
")",
"HistogramFinished",
"(",
"maxSymbol",
"uint8",
",",
"maxCount",
"int",
")",
"{",
"s",
".",
"maxCount",
"=",
"maxCount",
"\n",
"s",
".",
"symbolLen",
"=",
"uint16",
"(",
"maxSymbol",
")",
"+",
"1",
"\n",
"s",
".",
"clearCount",
"=",
"maxCount",
"!=",
"0",
"\n",
"}"
]
| // HistogramFinished can be called to indicate that the histogram has been populated.
// maxSymbol is the index of the highest set symbol of the next data segment.
// maxCount is the number of entries in the most populated entry.
// These are accepted at face value. | [
"HistogramFinished",
"can",
"be",
"called",
"to",
"indicate",
"that",
"the",
"histogram",
"has",
"been",
"populated",
".",
"maxSymbol",
"is",
"the",
"index",
"of",
"the",
"highest",
"set",
"symbol",
"of",
"the",
"next",
"data",
"segment",
".",
"maxCount",
"is",
"the",
"number",
"of",
"entries",
"in",
"the",
"most",
"populated",
"entry",
".",
"These",
"are",
"accepted",
"at",
"face",
"value",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/fse/fse.go#L98-L102 |
8,368 | klauspost/compress | fse/fse.go | prepare | func (s *Scratch) prepare(in []byte) (*Scratch, error) {
if s == nil {
s = &Scratch{}
}
if s.MaxSymbolValue == 0 {
s.MaxSymbolValue = 255
}
if s.TableLog == 0 {
s.TableLog = defaultTablelog
}
if s.TableLog > maxTableLog {
return nil, fmt.Errorf("tableLog (%d) > maxTableLog (%d)", s.TableLog, maxTableLog)
}
if cap(s.Out) == 0 {
s.Out = make([]byte, 0, len(in))
}
if s.clearCount && s.maxCount == 0 {
for i := range s.count {
s.count[i] = 0
}
s.clearCount = false
}
s.br.init(in)
if s.DecompressLimit == 0 {
// Max size 2GB.
s.DecompressLimit = 2 << 30
}
return s, nil
} | go | func (s *Scratch) prepare(in []byte) (*Scratch, error) {
if s == nil {
s = &Scratch{}
}
if s.MaxSymbolValue == 0 {
s.MaxSymbolValue = 255
}
if s.TableLog == 0 {
s.TableLog = defaultTablelog
}
if s.TableLog > maxTableLog {
return nil, fmt.Errorf("tableLog (%d) > maxTableLog (%d)", s.TableLog, maxTableLog)
}
if cap(s.Out) == 0 {
s.Out = make([]byte, 0, len(in))
}
if s.clearCount && s.maxCount == 0 {
for i := range s.count {
s.count[i] = 0
}
s.clearCount = false
}
s.br.init(in)
if s.DecompressLimit == 0 {
// Max size 2GB.
s.DecompressLimit = 2 << 30
}
return s, nil
} | [
"func",
"(",
"s",
"*",
"Scratch",
")",
"prepare",
"(",
"in",
"[",
"]",
"byte",
")",
"(",
"*",
"Scratch",
",",
"error",
")",
"{",
"if",
"s",
"==",
"nil",
"{",
"s",
"=",
"&",
"Scratch",
"{",
"}",
"\n",
"}",
"\n",
"if",
"s",
".",
"MaxSymbolValue",
"==",
"0",
"{",
"s",
".",
"MaxSymbolValue",
"=",
"255",
"\n",
"}",
"\n",
"if",
"s",
".",
"TableLog",
"==",
"0",
"{",
"s",
".",
"TableLog",
"=",
"defaultTablelog",
"\n",
"}",
"\n",
"if",
"s",
".",
"TableLog",
">",
"maxTableLog",
"{",
"return",
"nil",
",",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"s",
".",
"TableLog",
",",
"maxTableLog",
")",
"\n",
"}",
"\n",
"if",
"cap",
"(",
"s",
".",
"Out",
")",
"==",
"0",
"{",
"s",
".",
"Out",
"=",
"make",
"(",
"[",
"]",
"byte",
",",
"0",
",",
"len",
"(",
"in",
")",
")",
"\n",
"}",
"\n",
"if",
"s",
".",
"clearCount",
"&&",
"s",
".",
"maxCount",
"==",
"0",
"{",
"for",
"i",
":=",
"range",
"s",
".",
"count",
"{",
"s",
".",
"count",
"[",
"i",
"]",
"=",
"0",
"\n",
"}",
"\n",
"s",
".",
"clearCount",
"=",
"false",
"\n",
"}",
"\n",
"s",
".",
"br",
".",
"init",
"(",
"in",
")",
"\n",
"if",
"s",
".",
"DecompressLimit",
"==",
"0",
"{",
"// Max size 2GB.",
"s",
".",
"DecompressLimit",
"=",
"2",
"<<",
"30",
"\n",
"}",
"\n\n",
"return",
"s",
",",
"nil",
"\n",
"}"
]
| // prepare will prepare and allocate scratch tables used for both compression and decompression. | [
"prepare",
"will",
"prepare",
"and",
"allocate",
"scratch",
"tables",
"used",
"for",
"both",
"compression",
"and",
"decompression",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/fse/fse.go#L105-L134 |
8,369 | klauspost/compress | compressible.go | Estimate | func Estimate(b []byte) float64 {
if len(b) < 16 {
return 0
}
// Correctly predicted order 1
hits := 0
lastMatch := false
var o1 [256]byte
var hist [256]int
c1 := byte(0)
for _, c := range b {
if c == o1[c1] {
// We only count a hit if there was two correct predictions in a row.
if lastMatch {
hits++
}
lastMatch = true
} else {
lastMatch = false
}
o1[c1] = c
c1 = c
hist[c]++
}
// Use x^0.6 to give better spread
prediction := math.Pow(float64(hits)/float64(len(b)), 0.6)
// Calculate histogram distribution
variance := float64(0)
avg := float64(len(b)) / 256
for _, v := range hist {
Δ := float64(v) - avg
variance += Δ * Δ
}
stddev := math.Sqrt(float64(variance)) / float64(len(b))
exp := math.Sqrt(1 / float64(len(b)))
// Subtract expected stddev
stddev -= exp
if stddev < 0 {
stddev = 0
}
stddev *= 1 + exp
// Use x^0.4 to give better spread
entropy := math.Pow(stddev, 0.4)
// 50/50 weight between prediction and histogram distribution
return math.Pow((prediction+entropy)/2, 0.9)
} | go | func Estimate(b []byte) float64 {
if len(b) < 16 {
return 0
}
// Correctly predicted order 1
hits := 0
lastMatch := false
var o1 [256]byte
var hist [256]int
c1 := byte(0)
for _, c := range b {
if c == o1[c1] {
// We only count a hit if there was two correct predictions in a row.
if lastMatch {
hits++
}
lastMatch = true
} else {
lastMatch = false
}
o1[c1] = c
c1 = c
hist[c]++
}
// Use x^0.6 to give better spread
prediction := math.Pow(float64(hits)/float64(len(b)), 0.6)
// Calculate histogram distribution
variance := float64(0)
avg := float64(len(b)) / 256
for _, v := range hist {
Δ := float64(v) - avg
variance += Δ * Δ
}
stddev := math.Sqrt(float64(variance)) / float64(len(b))
exp := math.Sqrt(1 / float64(len(b)))
// Subtract expected stddev
stddev -= exp
if stddev < 0 {
stddev = 0
}
stddev *= 1 + exp
// Use x^0.4 to give better spread
entropy := math.Pow(stddev, 0.4)
// 50/50 weight between prediction and histogram distribution
return math.Pow((prediction+entropy)/2, 0.9)
} | [
"func",
"Estimate",
"(",
"b",
"[",
"]",
"byte",
")",
"float64",
"{",
"if",
"len",
"(",
"b",
")",
"<",
"16",
"{",
"return",
"0",
"\n",
"}",
"\n\n",
"// Correctly predicted order 1",
"hits",
":=",
"0",
"\n",
"lastMatch",
":=",
"false",
"\n",
"var",
"o1",
"[",
"256",
"]",
"byte",
"\n",
"var",
"hist",
"[",
"256",
"]",
"int",
"\n",
"c1",
":=",
"byte",
"(",
"0",
")",
"\n",
"for",
"_",
",",
"c",
":=",
"range",
"b",
"{",
"if",
"c",
"==",
"o1",
"[",
"c1",
"]",
"{",
"// We only count a hit if there was two correct predictions in a row.",
"if",
"lastMatch",
"{",
"hits",
"++",
"\n",
"}",
"\n",
"lastMatch",
"=",
"true",
"\n",
"}",
"else",
"{",
"lastMatch",
"=",
"false",
"\n",
"}",
"\n",
"o1",
"[",
"c1",
"]",
"=",
"c",
"\n",
"c1",
"=",
"c",
"\n",
"hist",
"[",
"c",
"]",
"++",
"\n",
"}",
"\n\n",
"// Use x^0.6 to give better spread",
"prediction",
":=",
"math",
".",
"Pow",
"(",
"float64",
"(",
"hits",
")",
"/",
"float64",
"(",
"len",
"(",
"b",
")",
")",
",",
"0.6",
")",
"\n\n",
"// Calculate histogram distribution",
"variance",
":=",
"float64",
"(",
"0",
")",
"\n",
"avg",
":=",
"float64",
"(",
"len",
"(",
"b",
")",
")",
"/",
"256",
"\n\n",
"for",
"_",
",",
"v",
":=",
"range",
"hist",
"{",
"Δ ",
"= ",
"loat64(",
"v",
")",
" ",
" ",
"vg",
"\n",
"variance",
"+=",
"Δ ",
" ",
"",
"\n",
"}",
"\n\n",
"stddev",
":=",
"math",
".",
"Sqrt",
"(",
"float64",
"(",
"variance",
")",
")",
"/",
"float64",
"(",
"len",
"(",
"b",
")",
")",
"\n",
"exp",
":=",
"math",
".",
"Sqrt",
"(",
"1",
"/",
"float64",
"(",
"len",
"(",
"b",
")",
")",
")",
"\n\n",
"// Subtract expected stddev",
"stddev",
"-=",
"exp",
"\n",
"if",
"stddev",
"<",
"0",
"{",
"stddev",
"=",
"0",
"\n",
"}",
"\n",
"stddev",
"*=",
"1",
"+",
"exp",
"\n\n",
"// Use x^0.4 to give better spread",
"entropy",
":=",
"math",
".",
"Pow",
"(",
"stddev",
",",
"0.4",
")",
"\n\n",
"// 50/50 weight between prediction and histogram distribution",
"return",
"math",
".",
"Pow",
"(",
"(",
"prediction",
"+",
"entropy",
")",
"/",
"2",
",",
"0.9",
")",
"\n",
"}"
]
| // Estimate returns a normalized compressibility estimate of block b.
// Values close to zero are likely uncompressible.
// Values above 0.1 are likely to be compressible.
// Values above 0.5 are very compressible.
// Very small lengths will return 0. | [
"Estimate",
"returns",
"a",
"normalized",
"compressibility",
"estimate",
"of",
"block",
"b",
".",
"Values",
"close",
"to",
"zero",
"are",
"likely",
"uncompressible",
".",
"Values",
"above",
"0",
".",
"1",
"are",
"likely",
"to",
"be",
"compressible",
".",
"Values",
"above",
"0",
".",
"5",
"are",
"very",
"compressible",
".",
"Very",
"small",
"lengths",
"will",
"return",
"0",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/compressible.go#L10-L63 |
8,370 | klauspost/compress | fse/bytereader.go | init | func (b *byteReader) init(in []byte) {
b.b = in
b.off = 0
} | go | func (b *byteReader) init(in []byte) {
b.b = in
b.off = 0
} | [
"func",
"(",
"b",
"*",
"byteReader",
")",
"init",
"(",
"in",
"[",
"]",
"byte",
")",
"{",
"b",
".",
"b",
"=",
"in",
"\n",
"b",
".",
"off",
"=",
"0",
"\n",
"}"
]
| // init will initialize the reader and set the input. | [
"init",
"will",
"initialize",
"the",
"reader",
"and",
"set",
"the",
"input",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/fse/bytereader.go#L18-L21 |
8,371 | klauspost/compress | zstd/bytereader.go | Uint8 | func (b *byteReader) Uint8() uint8 {
v := b.b[b.off]
return v
} | go | func (b *byteReader) Uint8() uint8 {
v := b.b[b.off]
return v
} | [
"func",
"(",
"b",
"*",
"byteReader",
")",
"Uint8",
"(",
")",
"uint8",
"{",
"v",
":=",
"b",
".",
"b",
"[",
"b",
".",
"off",
"]",
"\n",
"return",
"v",
"\n",
"}"
]
| // Uint8 returns the next byte | [
"Uint8",
"returns",
"the",
"next",
"byte"
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zstd/bytereader.go#L43-L46 |
8,372 | klauspost/compress | flate/inflate.go | dataBlock | func (f *decompressor) dataBlock() {
// Uncompressed.
// Discard current half-byte.
f.nb = 0
f.b = 0
// Length then ones-complement of length.
nr, err := io.ReadFull(f.r, f.buf[0:4])
f.roffset += int64(nr)
if err != nil {
f.err = noEOF(err)
return
}
n := int(f.buf[0]) | int(f.buf[1])<<8
nn := int(f.buf[2]) | int(f.buf[3])<<8
if uint16(nn) != uint16(^n) {
f.err = CorruptInputError(f.roffset)
return
}
if n == 0 {
f.toRead = f.dict.readFlush()
f.finishBlock()
return
}
f.copyLen = n
f.copyData()
} | go | func (f *decompressor) dataBlock() {
// Uncompressed.
// Discard current half-byte.
f.nb = 0
f.b = 0
// Length then ones-complement of length.
nr, err := io.ReadFull(f.r, f.buf[0:4])
f.roffset += int64(nr)
if err != nil {
f.err = noEOF(err)
return
}
n := int(f.buf[0]) | int(f.buf[1])<<8
nn := int(f.buf[2]) | int(f.buf[3])<<8
if uint16(nn) != uint16(^n) {
f.err = CorruptInputError(f.roffset)
return
}
if n == 0 {
f.toRead = f.dict.readFlush()
f.finishBlock()
return
}
f.copyLen = n
f.copyData()
} | [
"func",
"(",
"f",
"*",
"decompressor",
")",
"dataBlock",
"(",
")",
"{",
"// Uncompressed.",
"// Discard current half-byte.",
"f",
".",
"nb",
"=",
"0",
"\n",
"f",
".",
"b",
"=",
"0",
"\n\n",
"// Length then ones-complement of length.",
"nr",
",",
"err",
":=",
"io",
".",
"ReadFull",
"(",
"f",
".",
"r",
",",
"f",
".",
"buf",
"[",
"0",
":",
"4",
"]",
")",
"\n",
"f",
".",
"roffset",
"+=",
"int64",
"(",
"nr",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"f",
".",
"err",
"=",
"noEOF",
"(",
"err",
")",
"\n",
"return",
"\n",
"}",
"\n",
"n",
":=",
"int",
"(",
"f",
".",
"buf",
"[",
"0",
"]",
")",
"|",
"int",
"(",
"f",
".",
"buf",
"[",
"1",
"]",
")",
"<<",
"8",
"\n",
"nn",
":=",
"int",
"(",
"f",
".",
"buf",
"[",
"2",
"]",
")",
"|",
"int",
"(",
"f",
".",
"buf",
"[",
"3",
"]",
")",
"<<",
"8",
"\n",
"if",
"uint16",
"(",
"nn",
")",
"!=",
"uint16",
"(",
"^",
"n",
")",
"{",
"f",
".",
"err",
"=",
"CorruptInputError",
"(",
"f",
".",
"roffset",
")",
"\n",
"return",
"\n",
"}",
"\n\n",
"if",
"n",
"==",
"0",
"{",
"f",
".",
"toRead",
"=",
"f",
".",
"dict",
".",
"readFlush",
"(",
")",
"\n",
"f",
".",
"finishBlock",
"(",
")",
"\n",
"return",
"\n",
"}",
"\n\n",
"f",
".",
"copyLen",
"=",
"n",
"\n",
"f",
".",
"copyData",
"(",
")",
"\n",
"}"
]
| // Copy a single uncompressed data block from input to output. | [
"Copy",
"a",
"single",
"uncompressed",
"data",
"block",
"from",
"input",
"to",
"output",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/flate/inflate.go#L675-L703 |
8,373 | klauspost/compress | flate/inflate.go | copyData | func (f *decompressor) copyData() {
buf := f.dict.writeSlice()
if len(buf) > f.copyLen {
buf = buf[:f.copyLen]
}
cnt, err := io.ReadFull(f.r, buf)
f.roffset += int64(cnt)
f.copyLen -= cnt
f.dict.writeMark(cnt)
if err != nil {
f.err = noEOF(err)
return
}
if f.dict.availWrite() == 0 || f.copyLen > 0 {
f.toRead = f.dict.readFlush()
f.step = (*decompressor).copyData
return
}
f.finishBlock()
} | go | func (f *decompressor) copyData() {
buf := f.dict.writeSlice()
if len(buf) > f.copyLen {
buf = buf[:f.copyLen]
}
cnt, err := io.ReadFull(f.r, buf)
f.roffset += int64(cnt)
f.copyLen -= cnt
f.dict.writeMark(cnt)
if err != nil {
f.err = noEOF(err)
return
}
if f.dict.availWrite() == 0 || f.copyLen > 0 {
f.toRead = f.dict.readFlush()
f.step = (*decompressor).copyData
return
}
f.finishBlock()
} | [
"func",
"(",
"f",
"*",
"decompressor",
")",
"copyData",
"(",
")",
"{",
"buf",
":=",
"f",
".",
"dict",
".",
"writeSlice",
"(",
")",
"\n",
"if",
"len",
"(",
"buf",
")",
">",
"f",
".",
"copyLen",
"{",
"buf",
"=",
"buf",
"[",
":",
"f",
".",
"copyLen",
"]",
"\n",
"}",
"\n\n",
"cnt",
",",
"err",
":=",
"io",
".",
"ReadFull",
"(",
"f",
".",
"r",
",",
"buf",
")",
"\n",
"f",
".",
"roffset",
"+=",
"int64",
"(",
"cnt",
")",
"\n",
"f",
".",
"copyLen",
"-=",
"cnt",
"\n",
"f",
".",
"dict",
".",
"writeMark",
"(",
"cnt",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"f",
".",
"err",
"=",
"noEOF",
"(",
"err",
")",
"\n",
"return",
"\n",
"}",
"\n\n",
"if",
"f",
".",
"dict",
".",
"availWrite",
"(",
")",
"==",
"0",
"||",
"f",
".",
"copyLen",
">",
"0",
"{",
"f",
".",
"toRead",
"=",
"f",
".",
"dict",
".",
"readFlush",
"(",
")",
"\n",
"f",
".",
"step",
"=",
"(",
"*",
"decompressor",
")",
".",
"copyData",
"\n",
"return",
"\n",
"}",
"\n",
"f",
".",
"finishBlock",
"(",
")",
"\n",
"}"
]
| // copyData copies f.copyLen bytes from the underlying reader into f.hist.
// It pauses for reads when f.hist is full. | [
"copyData",
"copies",
"f",
".",
"copyLen",
"bytes",
"from",
"the",
"underlying",
"reader",
"into",
"f",
".",
"hist",
".",
"It",
"pauses",
"for",
"reads",
"when",
"f",
".",
"hist",
"is",
"full",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/flate/inflate.go#L707-L728 |
8,374 | klauspost/compress | flate/inflate.go | noEOF | func noEOF(e error) error {
if e == io.EOF {
return io.ErrUnexpectedEOF
}
return e
} | go | func noEOF(e error) error {
if e == io.EOF {
return io.ErrUnexpectedEOF
}
return e
} | [
"func",
"noEOF",
"(",
"e",
"error",
")",
"error",
"{",
"if",
"e",
"==",
"io",
".",
"EOF",
"{",
"return",
"io",
".",
"ErrUnexpectedEOF",
"\n",
"}",
"\n",
"return",
"e",
"\n",
"}"
]
| // noEOF returns err, unless err == io.EOF, in which case it returns io.ErrUnexpectedEOF. | [
"noEOF",
"returns",
"err",
"unless",
"err",
"==",
"io",
".",
"EOF",
"in",
"which",
"case",
"it",
"returns",
"io",
".",
"ErrUnexpectedEOF",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/flate/inflate.go#L741-L746 |
8,375 | klauspost/compress | flate/inflate.go | huffSym | func (f *decompressor) huffSym(h *huffmanDecoder) (int, error) {
// Since a huffmanDecoder can be empty or be composed of a degenerate tree
// with single element, huffSym must error on these two edge cases. In both
// cases, the chunks slice will be 0 for the invalid sequence, leading it
// satisfy the n == 0 check below.
n := uint(h.min)
// Optimization. Compiler isn't smart enough to keep f.b,f.nb in registers,
// but is smart enough to keep local variables in registers, so use nb and b,
// inline call to moreBits and reassign b,nb back to f on return.
nb, b := f.nb, f.b
for {
for nb < n {
c, err := f.r.ReadByte()
if err != nil {
f.b = b
f.nb = nb
return 0, noEOF(err)
}
f.roffset++
b |= uint32(c) << (nb & 31)
nb += 8
}
chunk := h.chunks[b&(huffmanNumChunks-1)]
n = uint(chunk & huffmanCountMask)
if n > huffmanChunkBits {
chunk = h.links[chunk>>huffmanValueShift][(b>>huffmanChunkBits)&h.linkMask]
n = uint(chunk & huffmanCountMask)
}
if n <= nb {
if n == 0 {
f.b = b
f.nb = nb
f.err = CorruptInputError(f.roffset)
return 0, f.err
}
f.b = b >> (n & 31)
f.nb = nb - n
return int(chunk >> huffmanValueShift), nil
}
}
} | go | func (f *decompressor) huffSym(h *huffmanDecoder) (int, error) {
// Since a huffmanDecoder can be empty or be composed of a degenerate tree
// with single element, huffSym must error on these two edge cases. In both
// cases, the chunks slice will be 0 for the invalid sequence, leading it
// satisfy the n == 0 check below.
n := uint(h.min)
// Optimization. Compiler isn't smart enough to keep f.b,f.nb in registers,
// but is smart enough to keep local variables in registers, so use nb and b,
// inline call to moreBits and reassign b,nb back to f on return.
nb, b := f.nb, f.b
for {
for nb < n {
c, err := f.r.ReadByte()
if err != nil {
f.b = b
f.nb = nb
return 0, noEOF(err)
}
f.roffset++
b |= uint32(c) << (nb & 31)
nb += 8
}
chunk := h.chunks[b&(huffmanNumChunks-1)]
n = uint(chunk & huffmanCountMask)
if n > huffmanChunkBits {
chunk = h.links[chunk>>huffmanValueShift][(b>>huffmanChunkBits)&h.linkMask]
n = uint(chunk & huffmanCountMask)
}
if n <= nb {
if n == 0 {
f.b = b
f.nb = nb
f.err = CorruptInputError(f.roffset)
return 0, f.err
}
f.b = b >> (n & 31)
f.nb = nb - n
return int(chunk >> huffmanValueShift), nil
}
}
} | [
"func",
"(",
"f",
"*",
"decompressor",
")",
"huffSym",
"(",
"h",
"*",
"huffmanDecoder",
")",
"(",
"int",
",",
"error",
")",
"{",
"// Since a huffmanDecoder can be empty or be composed of a degenerate tree",
"// with single element, huffSym must error on these two edge cases. In both",
"// cases, the chunks slice will be 0 for the invalid sequence, leading it",
"// satisfy the n == 0 check below.",
"n",
":=",
"uint",
"(",
"h",
".",
"min",
")",
"\n",
"// Optimization. Compiler isn't smart enough to keep f.b,f.nb in registers,",
"// but is smart enough to keep local variables in registers, so use nb and b,",
"// inline call to moreBits and reassign b,nb back to f on return.",
"nb",
",",
"b",
":=",
"f",
".",
"nb",
",",
"f",
".",
"b",
"\n",
"for",
"{",
"for",
"nb",
"<",
"n",
"{",
"c",
",",
"err",
":=",
"f",
".",
"r",
".",
"ReadByte",
"(",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"f",
".",
"b",
"=",
"b",
"\n",
"f",
".",
"nb",
"=",
"nb",
"\n",
"return",
"0",
",",
"noEOF",
"(",
"err",
")",
"\n",
"}",
"\n",
"f",
".",
"roffset",
"++",
"\n",
"b",
"|=",
"uint32",
"(",
"c",
")",
"<<",
"(",
"nb",
"&",
"31",
")",
"\n",
"nb",
"+=",
"8",
"\n",
"}",
"\n",
"chunk",
":=",
"h",
".",
"chunks",
"[",
"b",
"&",
"(",
"huffmanNumChunks",
"-",
"1",
")",
"]",
"\n",
"n",
"=",
"uint",
"(",
"chunk",
"&",
"huffmanCountMask",
")",
"\n",
"if",
"n",
">",
"huffmanChunkBits",
"{",
"chunk",
"=",
"h",
".",
"links",
"[",
"chunk",
">>",
"huffmanValueShift",
"]",
"[",
"(",
"b",
">>",
"huffmanChunkBits",
")",
"&",
"h",
".",
"linkMask",
"]",
"\n",
"n",
"=",
"uint",
"(",
"chunk",
"&",
"huffmanCountMask",
")",
"\n",
"}",
"\n",
"if",
"n",
"<=",
"nb",
"{",
"if",
"n",
"==",
"0",
"{",
"f",
".",
"b",
"=",
"b",
"\n",
"f",
".",
"nb",
"=",
"nb",
"\n",
"f",
".",
"err",
"=",
"CorruptInputError",
"(",
"f",
".",
"roffset",
")",
"\n",
"return",
"0",
",",
"f",
".",
"err",
"\n",
"}",
"\n",
"f",
".",
"b",
"=",
"b",
">>",
"(",
"n",
"&",
"31",
")",
"\n",
"f",
".",
"nb",
"=",
"nb",
"-",
"n",
"\n",
"return",
"int",
"(",
"chunk",
">>",
"huffmanValueShift",
")",
",",
"nil",
"\n",
"}",
"\n",
"}",
"\n",
"}"
]
| // Read the next Huffman-encoded symbol from f according to h. | [
"Read",
"the",
"next",
"Huffman",
"-",
"encoded",
"symbol",
"from",
"f",
"according",
"to",
"h",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/flate/inflate.go#L760-L800 |
8,376 | klauspost/compress | flate/inflate.go | NewReader | func NewReader(r io.Reader) io.ReadCloser {
fixedHuffmanDecoderInit()
var f decompressor
f.r = makeReader(r)
f.bits = new([maxNumLit + maxNumDist]int)
f.codebits = new([numCodes]int)
f.step = (*decompressor).nextBlock
f.dict.init(maxMatchOffset, nil)
return &f
} | go | func NewReader(r io.Reader) io.ReadCloser {
fixedHuffmanDecoderInit()
var f decompressor
f.r = makeReader(r)
f.bits = new([maxNumLit + maxNumDist]int)
f.codebits = new([numCodes]int)
f.step = (*decompressor).nextBlock
f.dict.init(maxMatchOffset, nil)
return &f
} | [
"func",
"NewReader",
"(",
"r",
"io",
".",
"Reader",
")",
"io",
".",
"ReadCloser",
"{",
"fixedHuffmanDecoderInit",
"(",
")",
"\n\n",
"var",
"f",
"decompressor",
"\n",
"f",
".",
"r",
"=",
"makeReader",
"(",
"r",
")",
"\n",
"f",
".",
"bits",
"=",
"new",
"(",
"[",
"maxNumLit",
"+",
"maxNumDist",
"]",
"int",
")",
"\n",
"f",
".",
"codebits",
"=",
"new",
"(",
"[",
"numCodes",
"]",
"int",
")",
"\n",
"f",
".",
"step",
"=",
"(",
"*",
"decompressor",
")",
".",
"nextBlock",
"\n",
"f",
".",
"dict",
".",
"init",
"(",
"maxMatchOffset",
",",
"nil",
")",
"\n",
"return",
"&",
"f",
"\n",
"}"
]
| // NewReader returns a new ReadCloser that can be used
// to read the uncompressed version of r.
// If r does not also implement io.ByteReader,
// the decompressor may read more data than necessary from r.
// It is the caller's responsibility to call Close on the ReadCloser
// when finished reading.
//
// The ReadCloser returned by NewReader also implements Resetter. | [
"NewReader",
"returns",
"a",
"new",
"ReadCloser",
"that",
"can",
"be",
"used",
"to",
"read",
"the",
"uncompressed",
"version",
"of",
"r",
".",
"If",
"r",
"does",
"not",
"also",
"implement",
"io",
".",
"ByteReader",
"the",
"decompressor",
"may",
"read",
"more",
"data",
"than",
"necessary",
"from",
"r",
".",
"It",
"is",
"the",
"caller",
"s",
"responsibility",
"to",
"call",
"Close",
"on",
"the",
"ReadCloser",
"when",
"finished",
"reading",
".",
"The",
"ReadCloser",
"returned",
"by",
"NewReader",
"also",
"implements",
"Resetter",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/flate/inflate.go#L851-L861 |
8,377 | klauspost/compress | zstd/fse_predefined.go | fillBase | func fillBase(dst []baseOffset, base uint32, bits ...uint8) {
if len(bits) != len(dst) {
panic(fmt.Sprintf("len(dst) (%d) != len(bits) (%d)", len(dst), len(bits)))
}
for i, bit := range bits {
if base > math.MaxInt32 {
panic(fmt.Sprintf("invalid decoding table, base overflows int32"))
}
dst[i] = baseOffset{
baseLine: base,
addBits: bit,
}
base += 1 << bit
}
} | go | func fillBase(dst []baseOffset, base uint32, bits ...uint8) {
if len(bits) != len(dst) {
panic(fmt.Sprintf("len(dst) (%d) != len(bits) (%d)", len(dst), len(bits)))
}
for i, bit := range bits {
if base > math.MaxInt32 {
panic(fmt.Sprintf("invalid decoding table, base overflows int32"))
}
dst[i] = baseOffset{
baseLine: base,
addBits: bit,
}
base += 1 << bit
}
} | [
"func",
"fillBase",
"(",
"dst",
"[",
"]",
"baseOffset",
",",
"base",
"uint32",
",",
"bits",
"...",
"uint8",
")",
"{",
"if",
"len",
"(",
"bits",
")",
"!=",
"len",
"(",
"dst",
")",
"{",
"panic",
"(",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
",",
"len",
"(",
"dst",
")",
",",
"len",
"(",
"bits",
")",
")",
")",
"\n",
"}",
"\n",
"for",
"i",
",",
"bit",
":=",
"range",
"bits",
"{",
"if",
"base",
">",
"math",
".",
"MaxInt32",
"{",
"panic",
"(",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
")",
")",
"\n",
"}",
"\n\n",
"dst",
"[",
"i",
"]",
"=",
"baseOffset",
"{",
"baseLine",
":",
"base",
",",
"addBits",
":",
"bit",
",",
"}",
"\n",
"base",
"+=",
"1",
"<<",
"bit",
"\n",
"}",
"\n",
"}"
]
| // fillBase will precalculate base offsets with the given bit distributions. | [
"fillBase",
"will",
"precalculate",
"base",
"offsets",
"with",
"the",
"given",
"bit",
"distributions",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zstd/fse_predefined.go#L43-L58 |
8,378 | klauspost/compress | fse/compress.go | Compress | func Compress(in []byte, s *Scratch) ([]byte, error) {
if len(in) <= 1 {
return nil, ErrIncompressible
}
if len(in) >= 2<<30 {
return nil, errors.New("input too big, must be < 2GB")
}
s, err := s.prepare(in)
if err != nil {
return nil, err
}
// Create histogram, if none was provided.
maxCount := s.maxCount
if maxCount == 0 {
maxCount = s.countSimple(in)
}
// Reset for next run.
s.clearCount = true
s.maxCount = 0
if maxCount == len(in) {
// One symbol, use RLE
return nil, ErrUseRLE
}
if maxCount == 1 || maxCount < (len(in)>>7) {
// Each symbol present maximum once or too well distributed.
return nil, ErrIncompressible
}
s.optimalTableLog()
err = s.normalizeCount()
if err != nil {
return nil, err
}
err = s.writeCount()
if err != nil {
return nil, err
}
if false {
err = s.validateNorm()
if err != nil {
return nil, err
}
}
err = s.buildCTable()
if err != nil {
return nil, err
}
err = s.compress(in)
if err != nil {
return nil, err
}
s.Out = s.bw.out
// Check if we compressed.
if len(s.Out) >= len(in) {
return nil, ErrIncompressible
}
return s.Out, nil
} | go | func Compress(in []byte, s *Scratch) ([]byte, error) {
if len(in) <= 1 {
return nil, ErrIncompressible
}
if len(in) >= 2<<30 {
return nil, errors.New("input too big, must be < 2GB")
}
s, err := s.prepare(in)
if err != nil {
return nil, err
}
// Create histogram, if none was provided.
maxCount := s.maxCount
if maxCount == 0 {
maxCount = s.countSimple(in)
}
// Reset for next run.
s.clearCount = true
s.maxCount = 0
if maxCount == len(in) {
// One symbol, use RLE
return nil, ErrUseRLE
}
if maxCount == 1 || maxCount < (len(in)>>7) {
// Each symbol present maximum once or too well distributed.
return nil, ErrIncompressible
}
s.optimalTableLog()
err = s.normalizeCount()
if err != nil {
return nil, err
}
err = s.writeCount()
if err != nil {
return nil, err
}
if false {
err = s.validateNorm()
if err != nil {
return nil, err
}
}
err = s.buildCTable()
if err != nil {
return nil, err
}
err = s.compress(in)
if err != nil {
return nil, err
}
s.Out = s.bw.out
// Check if we compressed.
if len(s.Out) >= len(in) {
return nil, ErrIncompressible
}
return s.Out, nil
} | [
"func",
"Compress",
"(",
"in",
"[",
"]",
"byte",
",",
"s",
"*",
"Scratch",
")",
"(",
"[",
"]",
"byte",
",",
"error",
")",
"{",
"if",
"len",
"(",
"in",
")",
"<=",
"1",
"{",
"return",
"nil",
",",
"ErrIncompressible",
"\n",
"}",
"\n",
"if",
"len",
"(",
"in",
")",
">=",
"2",
"<<",
"30",
"{",
"return",
"nil",
",",
"errors",
".",
"New",
"(",
"\"",
"\"",
")",
"\n",
"}",
"\n",
"s",
",",
"err",
":=",
"s",
".",
"prepare",
"(",
"in",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"return",
"nil",
",",
"err",
"\n",
"}",
"\n\n",
"// Create histogram, if none was provided.",
"maxCount",
":=",
"s",
".",
"maxCount",
"\n",
"if",
"maxCount",
"==",
"0",
"{",
"maxCount",
"=",
"s",
".",
"countSimple",
"(",
"in",
")",
"\n",
"}",
"\n",
"// Reset for next run.",
"s",
".",
"clearCount",
"=",
"true",
"\n",
"s",
".",
"maxCount",
"=",
"0",
"\n",
"if",
"maxCount",
"==",
"len",
"(",
"in",
")",
"{",
"// One symbol, use RLE",
"return",
"nil",
",",
"ErrUseRLE",
"\n",
"}",
"\n",
"if",
"maxCount",
"==",
"1",
"||",
"maxCount",
"<",
"(",
"len",
"(",
"in",
")",
">>",
"7",
")",
"{",
"// Each symbol present maximum once or too well distributed.",
"return",
"nil",
",",
"ErrIncompressible",
"\n",
"}",
"\n",
"s",
".",
"optimalTableLog",
"(",
")",
"\n",
"err",
"=",
"s",
".",
"normalizeCount",
"(",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"return",
"nil",
",",
"err",
"\n",
"}",
"\n",
"err",
"=",
"s",
".",
"writeCount",
"(",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"return",
"nil",
",",
"err",
"\n",
"}",
"\n\n",
"if",
"false",
"{",
"err",
"=",
"s",
".",
"validateNorm",
"(",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"return",
"nil",
",",
"err",
"\n",
"}",
"\n",
"}",
"\n\n",
"err",
"=",
"s",
".",
"buildCTable",
"(",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"return",
"nil",
",",
"err",
"\n",
"}",
"\n",
"err",
"=",
"s",
".",
"compress",
"(",
"in",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"return",
"nil",
",",
"err",
"\n",
"}",
"\n",
"s",
".",
"Out",
"=",
"s",
".",
"bw",
".",
"out",
"\n",
"// Check if we compressed.",
"if",
"len",
"(",
"s",
".",
"Out",
")",
">=",
"len",
"(",
"in",
")",
"{",
"return",
"nil",
",",
"ErrIncompressible",
"\n",
"}",
"\n",
"return",
"s",
".",
"Out",
",",
"nil",
"\n",
"}"
]
| // Compress the input bytes. Input must be < 2GB.
// Provide a Scratch buffer to avoid memory allocations.
// Note that the output is also kept in the scratch buffer.
// If input is too hard to compress, ErrIncompressible is returned.
// If input is a single byte value repeated ErrUseRLE is returned. | [
"Compress",
"the",
"input",
"bytes",
".",
"Input",
"must",
"be",
"<",
"2GB",
".",
"Provide",
"a",
"Scratch",
"buffer",
"to",
"avoid",
"memory",
"allocations",
".",
"Note",
"that",
"the",
"output",
"is",
"also",
"kept",
"in",
"the",
"scratch",
"buffer",
".",
"If",
"input",
"is",
"too",
"hard",
"to",
"compress",
"ErrIncompressible",
"is",
"returned",
".",
"If",
"input",
"is",
"a",
"single",
"byte",
"value",
"repeated",
"ErrUseRLE",
"is",
"returned",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/fse/compress.go#L18-L77 |
8,379 | klauspost/compress | fse/compress.go | init | func (c *cState) init(bw *bitWriter, ct *cTable, tableLog uint8, first symbolTransform) {
c.bw = bw
c.stateTable = ct.stateTable
nbBitsOut := (first.deltaNbBits + (1 << 15)) >> 16
im := int32((nbBitsOut << 16) - first.deltaNbBits)
lu := (im >> nbBitsOut) + first.deltaFindState
c.state = c.stateTable[lu]
return
} | go | func (c *cState) init(bw *bitWriter, ct *cTable, tableLog uint8, first symbolTransform) {
c.bw = bw
c.stateTable = ct.stateTable
nbBitsOut := (first.deltaNbBits + (1 << 15)) >> 16
im := int32((nbBitsOut << 16) - first.deltaNbBits)
lu := (im >> nbBitsOut) + first.deltaFindState
c.state = c.stateTable[lu]
return
} | [
"func",
"(",
"c",
"*",
"cState",
")",
"init",
"(",
"bw",
"*",
"bitWriter",
",",
"ct",
"*",
"cTable",
",",
"tableLog",
"uint8",
",",
"first",
"symbolTransform",
")",
"{",
"c",
".",
"bw",
"=",
"bw",
"\n",
"c",
".",
"stateTable",
"=",
"ct",
".",
"stateTable",
"\n\n",
"nbBitsOut",
":=",
"(",
"first",
".",
"deltaNbBits",
"+",
"(",
"1",
"<<",
"15",
")",
")",
">>",
"16",
"\n",
"im",
":=",
"int32",
"(",
"(",
"nbBitsOut",
"<<",
"16",
")",
"-",
"first",
".",
"deltaNbBits",
")",
"\n",
"lu",
":=",
"(",
"im",
">>",
"nbBitsOut",
")",
"+",
"first",
".",
"deltaFindState",
"\n",
"c",
".",
"state",
"=",
"c",
".",
"stateTable",
"[",
"lu",
"]",
"\n",
"return",
"\n",
"}"
]
| // init will initialize the compression state to the first symbol of the stream. | [
"init",
"will",
"initialize",
"the",
"compression",
"state",
"to",
"the",
"first",
"symbol",
"of",
"the",
"stream",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/fse/compress.go#L87-L96 |
8,380 | klauspost/compress | fse/compress.go | flush | func (c *cState) flush(tableLog uint8) {
c.bw.flush32()
c.bw.addBits16NC(c.state, tableLog)
c.bw.flush()
} | go | func (c *cState) flush(tableLog uint8) {
c.bw.flush32()
c.bw.addBits16NC(c.state, tableLog)
c.bw.flush()
} | [
"func",
"(",
"c",
"*",
"cState",
")",
"flush",
"(",
"tableLog",
"uint8",
")",
"{",
"c",
".",
"bw",
".",
"flush32",
"(",
")",
"\n",
"c",
".",
"bw",
".",
"addBits16NC",
"(",
"c",
".",
"state",
",",
"tableLog",
")",
"\n",
"c",
".",
"bw",
".",
"flush",
"(",
")",
"\n",
"}"
]
| // flush will write the tablelog to the output and flush the remaining full bytes. | [
"flush",
"will",
"write",
"the",
"tablelog",
"to",
"the",
"output",
"and",
"flush",
"the",
"remaining",
"full",
"bytes",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/fse/compress.go#L115-L119 |
8,381 | klauspost/compress | fse/compress.go | String | func (s symbolTransform) String() string {
return fmt.Sprintf("dnbits: %08x, fs:%d", s.deltaNbBits, s.deltaFindState)
} | go | func (s symbolTransform) String() string {
return fmt.Sprintf("dnbits: %08x, fs:%d", s.deltaNbBits, s.deltaFindState)
} | [
"func",
"(",
"s",
"symbolTransform",
")",
"String",
"(",
")",
"string",
"{",
"return",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
",",
"s",
".",
"deltaNbBits",
",",
"s",
".",
"deltaFindState",
")",
"\n",
"}"
]
| // String prints values as a human readable string. | [
"String",
"prints",
"values",
"as",
"a",
"human",
"readable",
"string",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/fse/compress.go#L318-L320 |
8,382 | klauspost/compress | fse/compress.go | allocCtable | func (s *Scratch) allocCtable() {
tableSize := 1 << s.actualTableLog
// get tableSymbol that is big enough.
if cap(s.ct.tableSymbol) < int(tableSize) {
s.ct.tableSymbol = make([]byte, tableSize)
}
s.ct.tableSymbol = s.ct.tableSymbol[:tableSize]
ctSize := tableSize
if cap(s.ct.stateTable) < ctSize {
s.ct.stateTable = make([]uint16, ctSize)
}
s.ct.stateTable = s.ct.stateTable[:ctSize]
if cap(s.ct.symbolTT) < int(s.symbolLen) {
s.ct.symbolTT = make([]symbolTransform, 256)
}
s.ct.symbolTT = s.ct.symbolTT[:256]
} | go | func (s *Scratch) allocCtable() {
tableSize := 1 << s.actualTableLog
// get tableSymbol that is big enough.
if cap(s.ct.tableSymbol) < int(tableSize) {
s.ct.tableSymbol = make([]byte, tableSize)
}
s.ct.tableSymbol = s.ct.tableSymbol[:tableSize]
ctSize := tableSize
if cap(s.ct.stateTable) < ctSize {
s.ct.stateTable = make([]uint16, ctSize)
}
s.ct.stateTable = s.ct.stateTable[:ctSize]
if cap(s.ct.symbolTT) < int(s.symbolLen) {
s.ct.symbolTT = make([]symbolTransform, 256)
}
s.ct.symbolTT = s.ct.symbolTT[:256]
} | [
"func",
"(",
"s",
"*",
"Scratch",
")",
"allocCtable",
"(",
")",
"{",
"tableSize",
":=",
"1",
"<<",
"s",
".",
"actualTableLog",
"\n",
"// get tableSymbol that is big enough.",
"if",
"cap",
"(",
"s",
".",
"ct",
".",
"tableSymbol",
")",
"<",
"int",
"(",
"tableSize",
")",
"{",
"s",
".",
"ct",
".",
"tableSymbol",
"=",
"make",
"(",
"[",
"]",
"byte",
",",
"tableSize",
")",
"\n",
"}",
"\n",
"s",
".",
"ct",
".",
"tableSymbol",
"=",
"s",
".",
"ct",
".",
"tableSymbol",
"[",
":",
"tableSize",
"]",
"\n\n",
"ctSize",
":=",
"tableSize",
"\n",
"if",
"cap",
"(",
"s",
".",
"ct",
".",
"stateTable",
")",
"<",
"ctSize",
"{",
"s",
".",
"ct",
".",
"stateTable",
"=",
"make",
"(",
"[",
"]",
"uint16",
",",
"ctSize",
")",
"\n",
"}",
"\n",
"s",
".",
"ct",
".",
"stateTable",
"=",
"s",
".",
"ct",
".",
"stateTable",
"[",
":",
"ctSize",
"]",
"\n\n",
"if",
"cap",
"(",
"s",
".",
"ct",
".",
"symbolTT",
")",
"<",
"int",
"(",
"s",
".",
"symbolLen",
")",
"{",
"s",
".",
"ct",
".",
"symbolTT",
"=",
"make",
"(",
"[",
"]",
"symbolTransform",
",",
"256",
")",
"\n",
"}",
"\n",
"s",
".",
"ct",
".",
"symbolTT",
"=",
"s",
".",
"ct",
".",
"symbolTT",
"[",
":",
"256",
"]",
"\n",
"}"
]
| // allocCtable will allocate tables needed for compression.
// If existing tables a re big enough, they are simply re-used. | [
"allocCtable",
"will",
"allocate",
"tables",
"needed",
"for",
"compression",
".",
"If",
"existing",
"tables",
"a",
"re",
"big",
"enough",
"they",
"are",
"simply",
"re",
"-",
"used",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/fse/compress.go#L331-L349 |
8,383 | klauspost/compress | fse/compress.go | normalizeCount | func (s *Scratch) normalizeCount() error {
var (
tableLog = s.actualTableLog
scale = 62 - uint64(tableLog)
step = (1 << 62) / uint64(s.br.remain())
vStep = uint64(1) << (scale - 20)
stillToDistribute = int16(1 << tableLog)
largest int
largestP int16
lowThreshold = (uint32)(s.br.remain() >> tableLog)
)
for i, cnt := range s.count[:s.symbolLen] {
// already handled
// if (count[s] == s.length) return 0; /* rle special case */
if cnt == 0 {
s.norm[i] = 0
continue
}
if cnt <= lowThreshold {
s.norm[i] = -1
stillToDistribute--
} else {
proba := (int16)((uint64(cnt) * step) >> scale)
if proba < 8 {
restToBeat := vStep * uint64(rtbTable[proba])
v := uint64(cnt)*step - (uint64(proba) << scale)
if v > restToBeat {
proba++
}
}
if proba > largestP {
largestP = proba
largest = i
}
s.norm[i] = proba
stillToDistribute -= proba
}
}
if -stillToDistribute >= (s.norm[largest] >> 1) {
// corner case, need another normalization method
return s.normalizeCount2()
}
s.norm[largest] += stillToDistribute
return nil
} | go | func (s *Scratch) normalizeCount() error {
var (
tableLog = s.actualTableLog
scale = 62 - uint64(tableLog)
step = (1 << 62) / uint64(s.br.remain())
vStep = uint64(1) << (scale - 20)
stillToDistribute = int16(1 << tableLog)
largest int
largestP int16
lowThreshold = (uint32)(s.br.remain() >> tableLog)
)
for i, cnt := range s.count[:s.symbolLen] {
// already handled
// if (count[s] == s.length) return 0; /* rle special case */
if cnt == 0 {
s.norm[i] = 0
continue
}
if cnt <= lowThreshold {
s.norm[i] = -1
stillToDistribute--
} else {
proba := (int16)((uint64(cnt) * step) >> scale)
if proba < 8 {
restToBeat := vStep * uint64(rtbTable[proba])
v := uint64(cnt)*step - (uint64(proba) << scale)
if v > restToBeat {
proba++
}
}
if proba > largestP {
largestP = proba
largest = i
}
s.norm[i] = proba
stillToDistribute -= proba
}
}
if -stillToDistribute >= (s.norm[largest] >> 1) {
// corner case, need another normalization method
return s.normalizeCount2()
}
s.norm[largest] += stillToDistribute
return nil
} | [
"func",
"(",
"s",
"*",
"Scratch",
")",
"normalizeCount",
"(",
")",
"error",
"{",
"var",
"(",
"tableLog",
"=",
"s",
".",
"actualTableLog",
"\n",
"scale",
"=",
"62",
"-",
"uint64",
"(",
"tableLog",
")",
"\n",
"step",
"=",
"(",
"1",
"<<",
"62",
")",
"/",
"uint64",
"(",
"s",
".",
"br",
".",
"remain",
"(",
")",
")",
"\n",
"vStep",
"=",
"uint64",
"(",
"1",
")",
"<<",
"(",
"scale",
"-",
"20",
")",
"\n",
"stillToDistribute",
"=",
"int16",
"(",
"1",
"<<",
"tableLog",
")",
"\n",
"largest",
"int",
"\n",
"largestP",
"int16",
"\n",
"lowThreshold",
"=",
"(",
"uint32",
")",
"(",
"s",
".",
"br",
".",
"remain",
"(",
")",
">>",
"tableLog",
")",
"\n",
")",
"\n\n",
"for",
"i",
",",
"cnt",
":=",
"range",
"s",
".",
"count",
"[",
":",
"s",
".",
"symbolLen",
"]",
"{",
"// already handled",
"// if (count[s] == s.length) return 0; /* rle special case */",
"if",
"cnt",
"==",
"0",
"{",
"s",
".",
"norm",
"[",
"i",
"]",
"=",
"0",
"\n",
"continue",
"\n",
"}",
"\n",
"if",
"cnt",
"<=",
"lowThreshold",
"{",
"s",
".",
"norm",
"[",
"i",
"]",
"=",
"-",
"1",
"\n",
"stillToDistribute",
"--",
"\n",
"}",
"else",
"{",
"proba",
":=",
"(",
"int16",
")",
"(",
"(",
"uint64",
"(",
"cnt",
")",
"*",
"step",
")",
">>",
"scale",
")",
"\n",
"if",
"proba",
"<",
"8",
"{",
"restToBeat",
":=",
"vStep",
"*",
"uint64",
"(",
"rtbTable",
"[",
"proba",
"]",
")",
"\n",
"v",
":=",
"uint64",
"(",
"cnt",
")",
"*",
"step",
"-",
"(",
"uint64",
"(",
"proba",
")",
"<<",
"scale",
")",
"\n",
"if",
"v",
">",
"restToBeat",
"{",
"proba",
"++",
"\n",
"}",
"\n",
"}",
"\n",
"if",
"proba",
">",
"largestP",
"{",
"largestP",
"=",
"proba",
"\n",
"largest",
"=",
"i",
"\n",
"}",
"\n",
"s",
".",
"norm",
"[",
"i",
"]",
"=",
"proba",
"\n",
"stillToDistribute",
"-=",
"proba",
"\n",
"}",
"\n",
"}",
"\n\n",
"if",
"-",
"stillToDistribute",
">=",
"(",
"s",
".",
"norm",
"[",
"largest",
"]",
">>",
"1",
")",
"{",
"// corner case, need another normalization method",
"return",
"s",
".",
"normalizeCount2",
"(",
")",
"\n",
"}",
"\n",
"s",
".",
"norm",
"[",
"largest",
"]",
"+=",
"stillToDistribute",
"\n",
"return",
"nil",
"\n",
"}"
]
| // normalizeCount will normalize the count of the symbols so
// the total is equal to the table size. | [
"normalizeCount",
"will",
"normalize",
"the",
"count",
"of",
"the",
"symbols",
"so",
"the",
"total",
"is",
"equal",
"to",
"the",
"table",
"size",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/fse/compress.go#L511-L558 |
8,384 | klauspost/compress | fse/compress.go | validateNorm | func (s *Scratch) validateNorm() (err error) {
var total int
for _, v := range s.norm[:s.symbolLen] {
if v >= 0 {
total += int(v)
} else {
total -= int(v)
}
}
defer func() {
if err == nil {
return
}
fmt.Printf("selected TableLog: %d, Symbol length: %d\n", s.actualTableLog, s.symbolLen)
for i, v := range s.norm[:s.symbolLen] {
fmt.Printf("%3d: %5d -> %4d \n", i, s.count[i], v)
}
}()
if total != (1 << s.actualTableLog) {
return fmt.Errorf("warning: Total == %d != %d", total, 1<<s.actualTableLog)
}
for i, v := range s.count[s.symbolLen:] {
if v != 0 {
return fmt.Errorf("warning: Found symbol out of range, %d after cut", i)
}
}
return nil
} | go | func (s *Scratch) validateNorm() (err error) {
var total int
for _, v := range s.norm[:s.symbolLen] {
if v >= 0 {
total += int(v)
} else {
total -= int(v)
}
}
defer func() {
if err == nil {
return
}
fmt.Printf("selected TableLog: %d, Symbol length: %d\n", s.actualTableLog, s.symbolLen)
for i, v := range s.norm[:s.symbolLen] {
fmt.Printf("%3d: %5d -> %4d \n", i, s.count[i], v)
}
}()
if total != (1 << s.actualTableLog) {
return fmt.Errorf("warning: Total == %d != %d", total, 1<<s.actualTableLog)
}
for i, v := range s.count[s.symbolLen:] {
if v != 0 {
return fmt.Errorf("warning: Found symbol out of range, %d after cut", i)
}
}
return nil
} | [
"func",
"(",
"s",
"*",
"Scratch",
")",
"validateNorm",
"(",
")",
"(",
"err",
"error",
")",
"{",
"var",
"total",
"int",
"\n",
"for",
"_",
",",
"v",
":=",
"range",
"s",
".",
"norm",
"[",
":",
"s",
".",
"symbolLen",
"]",
"{",
"if",
"v",
">=",
"0",
"{",
"total",
"+=",
"int",
"(",
"v",
")",
"\n",
"}",
"else",
"{",
"total",
"-=",
"int",
"(",
"v",
")",
"\n",
"}",
"\n",
"}",
"\n",
"defer",
"func",
"(",
")",
"{",
"if",
"err",
"==",
"nil",
"{",
"return",
"\n",
"}",
"\n",
"fmt",
".",
"Printf",
"(",
"\"",
"\\n",
"\"",
",",
"s",
".",
"actualTableLog",
",",
"s",
".",
"symbolLen",
")",
"\n",
"for",
"i",
",",
"v",
":=",
"range",
"s",
".",
"norm",
"[",
":",
"s",
".",
"symbolLen",
"]",
"{",
"fmt",
".",
"Printf",
"(",
"\"",
"\\n",
"\"",
",",
"i",
",",
"s",
".",
"count",
"[",
"i",
"]",
",",
"v",
")",
"\n",
"}",
"\n",
"}",
"(",
")",
"\n",
"if",
"total",
"!=",
"(",
"1",
"<<",
"s",
".",
"actualTableLog",
")",
"{",
"return",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"total",
",",
"1",
"<<",
"s",
".",
"actualTableLog",
")",
"\n",
"}",
"\n",
"for",
"i",
",",
"v",
":=",
"range",
"s",
".",
"count",
"[",
"s",
".",
"symbolLen",
":",
"]",
"{",
"if",
"v",
"!=",
"0",
"{",
"return",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"i",
")",
"\n",
"}",
"\n",
"}",
"\n",
"return",
"nil",
"\n",
"}"
]
| // validateNorm validates the normalized histogram table. | [
"validateNorm",
"validates",
"the",
"normalized",
"histogram",
"table",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/fse/compress.go#L657-L684 |
8,385 | klauspost/compress | zstd/decoder.go | Read | func (d *Decoder) Read(p []byte) (int, error) {
if d.stream == nil {
return 0, errors.New("no input has been initialized")
}
var n int
for {
if len(d.current.b) > 0 {
filled := copy(p, d.current.b)
p = p[filled:]
d.current.b = d.current.b[filled:]
n += filled
}
if len(p) == 0 {
break
}
if len(d.current.b) == 0 {
// We have an error and no more data
if d.current.err != nil {
break
}
d.nextBlock()
}
}
if len(d.current.b) > 0 {
// Only return error at end of block
return n, nil
}
if d.current.err != nil {
d.drainOutput()
}
if debug {
println("returning", n, d.current.err, len(d.decoders))
}
return n, d.current.err
} | go | func (d *Decoder) Read(p []byte) (int, error) {
if d.stream == nil {
return 0, errors.New("no input has been initialized")
}
var n int
for {
if len(d.current.b) > 0 {
filled := copy(p, d.current.b)
p = p[filled:]
d.current.b = d.current.b[filled:]
n += filled
}
if len(p) == 0 {
break
}
if len(d.current.b) == 0 {
// We have an error and no more data
if d.current.err != nil {
break
}
d.nextBlock()
}
}
if len(d.current.b) > 0 {
// Only return error at end of block
return n, nil
}
if d.current.err != nil {
d.drainOutput()
}
if debug {
println("returning", n, d.current.err, len(d.decoders))
}
return n, d.current.err
} | [
"func",
"(",
"d",
"*",
"Decoder",
")",
"Read",
"(",
"p",
"[",
"]",
"byte",
")",
"(",
"int",
",",
"error",
")",
"{",
"if",
"d",
".",
"stream",
"==",
"nil",
"{",
"return",
"0",
",",
"errors",
".",
"New",
"(",
"\"",
"\"",
")",
"\n",
"}",
"\n",
"var",
"n",
"int",
"\n",
"for",
"{",
"if",
"len",
"(",
"d",
".",
"current",
".",
"b",
")",
">",
"0",
"{",
"filled",
":=",
"copy",
"(",
"p",
",",
"d",
".",
"current",
".",
"b",
")",
"\n",
"p",
"=",
"p",
"[",
"filled",
":",
"]",
"\n",
"d",
".",
"current",
".",
"b",
"=",
"d",
".",
"current",
".",
"b",
"[",
"filled",
":",
"]",
"\n",
"n",
"+=",
"filled",
"\n",
"}",
"\n",
"if",
"len",
"(",
"p",
")",
"==",
"0",
"{",
"break",
"\n",
"}",
"\n",
"if",
"len",
"(",
"d",
".",
"current",
".",
"b",
")",
"==",
"0",
"{",
"// We have an error and no more data",
"if",
"d",
".",
"current",
".",
"err",
"!=",
"nil",
"{",
"break",
"\n",
"}",
"\n",
"d",
".",
"nextBlock",
"(",
")",
"\n",
"}",
"\n",
"}",
"\n",
"if",
"len",
"(",
"d",
".",
"current",
".",
"b",
")",
">",
"0",
"{",
"// Only return error at end of block",
"return",
"n",
",",
"nil",
"\n",
"}",
"\n",
"if",
"d",
".",
"current",
".",
"err",
"!=",
"nil",
"{",
"d",
".",
"drainOutput",
"(",
")",
"\n",
"}",
"\n",
"if",
"debug",
"{",
"println",
"(",
"\"",
"\"",
",",
"n",
",",
"d",
".",
"current",
".",
"err",
",",
"len",
"(",
"d",
".",
"decoders",
")",
")",
"\n",
"}",
"\n",
"return",
"n",
",",
"d",
".",
"current",
".",
"err",
"\n",
"}"
]
| // Read bytes from the decompressed stream into p.
// Returns the number of bytes written and any error that occurred.
// When the stream is done, io.EOF will be returned. | [
"Read",
"bytes",
"from",
"the",
"decompressed",
"stream",
"into",
"p",
".",
"Returns",
"the",
"number",
"of",
"bytes",
"written",
"and",
"any",
"error",
"that",
"occurred",
".",
"When",
"the",
"stream",
"is",
"done",
"io",
".",
"EOF",
"will",
"be",
"returned",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zstd/decoder.go#L105-L139 |
8,386 | klauspost/compress | zstd/decoder.go | Reset | func (d *Decoder) Reset(r io.Reader) error {
if d.current.err == ErrDecoderClosed {
return d.current.err
}
if r == nil {
return errors.New("nil Reader sent as input")
}
// TODO: If r is a *bytes.Buffer, we could automatically switch to sync operation.
if d.stream == nil {
d.stream = make(chan decodeStream, 1)
go d.startStreamDecoder(d.stream)
}
d.drainOutput()
// Remove current block.
d.current.decodeOutput = decodeOutput{}
d.current.err = nil
d.current.cancel = make(chan struct{})
d.current.flushed = false
d.current.d = nil
d.stream <- decodeStream{
r: r,
output: d.current.output,
cancel: d.current.cancel,
}
return nil
} | go | func (d *Decoder) Reset(r io.Reader) error {
if d.current.err == ErrDecoderClosed {
return d.current.err
}
if r == nil {
return errors.New("nil Reader sent as input")
}
// TODO: If r is a *bytes.Buffer, we could automatically switch to sync operation.
if d.stream == nil {
d.stream = make(chan decodeStream, 1)
go d.startStreamDecoder(d.stream)
}
d.drainOutput()
// Remove current block.
d.current.decodeOutput = decodeOutput{}
d.current.err = nil
d.current.cancel = make(chan struct{})
d.current.flushed = false
d.current.d = nil
d.stream <- decodeStream{
r: r,
output: d.current.output,
cancel: d.current.cancel,
}
return nil
} | [
"func",
"(",
"d",
"*",
"Decoder",
")",
"Reset",
"(",
"r",
"io",
".",
"Reader",
")",
"error",
"{",
"if",
"d",
".",
"current",
".",
"err",
"==",
"ErrDecoderClosed",
"{",
"return",
"d",
".",
"current",
".",
"err",
"\n",
"}",
"\n",
"if",
"r",
"==",
"nil",
"{",
"return",
"errors",
".",
"New",
"(",
"\"",
"\"",
")",
"\n",
"}",
"\n\n",
"// TODO: If r is a *bytes.Buffer, we could automatically switch to sync operation.",
"if",
"d",
".",
"stream",
"==",
"nil",
"{",
"d",
".",
"stream",
"=",
"make",
"(",
"chan",
"decodeStream",
",",
"1",
")",
"\n",
"go",
"d",
".",
"startStreamDecoder",
"(",
"d",
".",
"stream",
")",
"\n",
"}",
"\n\n",
"d",
".",
"drainOutput",
"(",
")",
"\n\n",
"// Remove current block.",
"d",
".",
"current",
".",
"decodeOutput",
"=",
"decodeOutput",
"{",
"}",
"\n",
"d",
".",
"current",
".",
"err",
"=",
"nil",
"\n",
"d",
".",
"current",
".",
"cancel",
"=",
"make",
"(",
"chan",
"struct",
"{",
"}",
")",
"\n",
"d",
".",
"current",
".",
"flushed",
"=",
"false",
"\n",
"d",
".",
"current",
".",
"d",
"=",
"nil",
"\n\n",
"d",
".",
"stream",
"<-",
"decodeStream",
"{",
"r",
":",
"r",
",",
"output",
":",
"d",
".",
"current",
".",
"output",
",",
"cancel",
":",
"d",
".",
"current",
".",
"cancel",
",",
"}",
"\n",
"return",
"nil",
"\n",
"}"
]
| // Reset will reset the decoder the supplied stream after the current has finished processing.
// Note that this functionality cannot be used after Close has been called. | [
"Reset",
"will",
"reset",
"the",
"decoder",
"the",
"supplied",
"stream",
"after",
"the",
"current",
"has",
"finished",
"processing",
".",
"Note",
"that",
"this",
"functionality",
"cannot",
"be",
"used",
"after",
"Close",
"has",
"been",
"called",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zstd/decoder.go#L143-L173 |
8,387 | klauspost/compress | zstd/decoder.go | drainOutput | func (d *Decoder) drainOutput() {
if d.current.cancel != nil {
println("cancelling current")
close(d.current.cancel)
d.current.cancel = nil
}
if d.current.d != nil {
println("re-adding current decoder", d.current.d, len(d.decoders))
d.decoders <- d.current.d
d.current.d = nil
d.current.b = nil
}
if d.current.output == nil || d.current.flushed {
println("current already flushed")
return
}
for {
select {
case v := <-d.current.output:
if v.d != nil {
println("got decoder", v.d)
d.decoders <- v.d
}
if v.err == errEndOfStream {
println("current flushed")
d.current.flushed = true
return
}
}
}
} | go | func (d *Decoder) drainOutput() {
if d.current.cancel != nil {
println("cancelling current")
close(d.current.cancel)
d.current.cancel = nil
}
if d.current.d != nil {
println("re-adding current decoder", d.current.d, len(d.decoders))
d.decoders <- d.current.d
d.current.d = nil
d.current.b = nil
}
if d.current.output == nil || d.current.flushed {
println("current already flushed")
return
}
for {
select {
case v := <-d.current.output:
if v.d != nil {
println("got decoder", v.d)
d.decoders <- v.d
}
if v.err == errEndOfStream {
println("current flushed")
d.current.flushed = true
return
}
}
}
} | [
"func",
"(",
"d",
"*",
"Decoder",
")",
"drainOutput",
"(",
")",
"{",
"if",
"d",
".",
"current",
".",
"cancel",
"!=",
"nil",
"{",
"println",
"(",
"\"",
"\"",
")",
"\n",
"close",
"(",
"d",
".",
"current",
".",
"cancel",
")",
"\n",
"d",
".",
"current",
".",
"cancel",
"=",
"nil",
"\n",
"}",
"\n",
"if",
"d",
".",
"current",
".",
"d",
"!=",
"nil",
"{",
"println",
"(",
"\"",
"\"",
",",
"d",
".",
"current",
".",
"d",
",",
"len",
"(",
"d",
".",
"decoders",
")",
")",
"\n",
"d",
".",
"decoders",
"<-",
"d",
".",
"current",
".",
"d",
"\n",
"d",
".",
"current",
".",
"d",
"=",
"nil",
"\n",
"d",
".",
"current",
".",
"b",
"=",
"nil",
"\n",
"}",
"\n",
"if",
"d",
".",
"current",
".",
"output",
"==",
"nil",
"||",
"d",
".",
"current",
".",
"flushed",
"{",
"println",
"(",
"\"",
"\"",
")",
"\n",
"return",
"\n",
"}",
"\n",
"for",
"{",
"select",
"{",
"case",
"v",
":=",
"<-",
"d",
".",
"current",
".",
"output",
":",
"if",
"v",
".",
"d",
"!=",
"nil",
"{",
"println",
"(",
"\"",
"\"",
",",
"v",
".",
"d",
")",
"\n",
"d",
".",
"decoders",
"<-",
"v",
".",
"d",
"\n",
"}",
"\n",
"if",
"v",
".",
"err",
"==",
"errEndOfStream",
"{",
"println",
"(",
"\"",
"\"",
")",
"\n",
"d",
".",
"current",
".",
"flushed",
"=",
"true",
"\n",
"return",
"\n",
"}",
"\n",
"}",
"\n",
"}",
"\n",
"}"
]
| // drainOutput will drain the output until errEndOfStream is sent. | [
"drainOutput",
"will",
"drain",
"the",
"output",
"until",
"errEndOfStream",
"is",
"sent",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zstd/decoder.go#L176-L206 |
8,388 | klauspost/compress | zstd/decoder.go | WriteTo | func (d *Decoder) WriteTo(w io.Writer) (int64, error) {
if d.stream == nil {
return 0, errors.New("no input has been initialized")
}
var n int64
for d.current.err == nil {
if len(d.current.b) > 0 {
n2, err2 := w.Write(d.current.b)
n += int64(n2)
if err2 != nil && d.current.err == nil {
d.current.err = err2
break
}
}
d.nextBlock()
}
err := d.current.err
if err != nil {
d.drainOutput()
}
if err == io.EOF {
err = nil
}
return n, err
} | go | func (d *Decoder) WriteTo(w io.Writer) (int64, error) {
if d.stream == nil {
return 0, errors.New("no input has been initialized")
}
var n int64
for d.current.err == nil {
if len(d.current.b) > 0 {
n2, err2 := w.Write(d.current.b)
n += int64(n2)
if err2 != nil && d.current.err == nil {
d.current.err = err2
break
}
}
d.nextBlock()
}
err := d.current.err
if err != nil {
d.drainOutput()
}
if err == io.EOF {
err = nil
}
return n, err
} | [
"func",
"(",
"d",
"*",
"Decoder",
")",
"WriteTo",
"(",
"w",
"io",
".",
"Writer",
")",
"(",
"int64",
",",
"error",
")",
"{",
"if",
"d",
".",
"stream",
"==",
"nil",
"{",
"return",
"0",
",",
"errors",
".",
"New",
"(",
"\"",
"\"",
")",
"\n",
"}",
"\n",
"var",
"n",
"int64",
"\n",
"for",
"d",
".",
"current",
".",
"err",
"==",
"nil",
"{",
"if",
"len",
"(",
"d",
".",
"current",
".",
"b",
")",
">",
"0",
"{",
"n2",
",",
"err2",
":=",
"w",
".",
"Write",
"(",
"d",
".",
"current",
".",
"b",
")",
"\n",
"n",
"+=",
"int64",
"(",
"n2",
")",
"\n",
"if",
"err2",
"!=",
"nil",
"&&",
"d",
".",
"current",
".",
"err",
"==",
"nil",
"{",
"d",
".",
"current",
".",
"err",
"=",
"err2",
"\n",
"break",
"\n",
"}",
"\n",
"}",
"\n",
"d",
".",
"nextBlock",
"(",
")",
"\n",
"}",
"\n",
"err",
":=",
"d",
".",
"current",
".",
"err",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"d",
".",
"drainOutput",
"(",
")",
"\n",
"}",
"\n",
"if",
"err",
"==",
"io",
".",
"EOF",
"{",
"err",
"=",
"nil",
"\n",
"}",
"\n",
"return",
"n",
",",
"err",
"\n",
"}"
]
| // WriteTo writes data to w until there's no more data to write or when an error occurs.
// The return value n is the number of bytes written.
// Any error encountered during the write is also returned. | [
"WriteTo",
"writes",
"data",
"to",
"w",
"until",
"there",
"s",
"no",
"more",
"data",
"to",
"write",
"or",
"when",
"an",
"error",
"occurs",
".",
"The",
"return",
"value",
"n",
"is",
"the",
"number",
"of",
"bytes",
"written",
".",
"Any",
"error",
"encountered",
"during",
"the",
"write",
"is",
"also",
"returned",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zstd/decoder.go#L211-L235 |
8,389 | klauspost/compress | zstd/decoder.go | DecodeAll | func (d *Decoder) DecodeAll(input, dst []byte) ([]byte, error) {
if d.current.err == ErrDecoderClosed {
return dst, ErrDecoderClosed
}
//println(len(d.frames), len(d.decoders), d.current)
block, frame := <-d.decoders, <-d.frames
defer func() {
d.decoders <- block
frame.rawInput = nil
d.frames <- frame
}()
if cap(dst) == 0 {
// Allocate 1MB by default.
dst = make([]byte, 0, 1<<20)
}
br := byteBuf(input)
for {
err := frame.reset(&br)
if err == io.EOF {
return dst, nil
}
if err != nil {
return dst, err
}
if frame.FrameContentSize > d.o.maxDecodedSize-uint64(len(dst)) {
return dst, ErrDecoderSizeExceeded
}
if frame.FrameContentSize > 0 && frame.FrameContentSize < 1<<30 {
// Never preallocate moe than 1 GB up front.
if uint64(cap(dst)) < frame.FrameContentSize {
dst2 := make([]byte, len(dst), len(dst)+int(frame.FrameContentSize))
copy(dst2, dst)
dst = dst2
}
}
dst, err = frame.runDecoder(dst, block)
if err != nil {
return dst, err
}
if len(br) == 0 {
break
}
}
return dst, nil
} | go | func (d *Decoder) DecodeAll(input, dst []byte) ([]byte, error) {
if d.current.err == ErrDecoderClosed {
return dst, ErrDecoderClosed
}
//println(len(d.frames), len(d.decoders), d.current)
block, frame := <-d.decoders, <-d.frames
defer func() {
d.decoders <- block
frame.rawInput = nil
d.frames <- frame
}()
if cap(dst) == 0 {
// Allocate 1MB by default.
dst = make([]byte, 0, 1<<20)
}
br := byteBuf(input)
for {
err := frame.reset(&br)
if err == io.EOF {
return dst, nil
}
if err != nil {
return dst, err
}
if frame.FrameContentSize > d.o.maxDecodedSize-uint64(len(dst)) {
return dst, ErrDecoderSizeExceeded
}
if frame.FrameContentSize > 0 && frame.FrameContentSize < 1<<30 {
// Never preallocate moe than 1 GB up front.
if uint64(cap(dst)) < frame.FrameContentSize {
dst2 := make([]byte, len(dst), len(dst)+int(frame.FrameContentSize))
copy(dst2, dst)
dst = dst2
}
}
dst, err = frame.runDecoder(dst, block)
if err != nil {
return dst, err
}
if len(br) == 0 {
break
}
}
return dst, nil
} | [
"func",
"(",
"d",
"*",
"Decoder",
")",
"DecodeAll",
"(",
"input",
",",
"dst",
"[",
"]",
"byte",
")",
"(",
"[",
"]",
"byte",
",",
"error",
")",
"{",
"if",
"d",
".",
"current",
".",
"err",
"==",
"ErrDecoderClosed",
"{",
"return",
"dst",
",",
"ErrDecoderClosed",
"\n",
"}",
"\n",
"//println(len(d.frames), len(d.decoders), d.current)",
"block",
",",
"frame",
":=",
"<-",
"d",
".",
"decoders",
",",
"<-",
"d",
".",
"frames",
"\n",
"defer",
"func",
"(",
")",
"{",
"d",
".",
"decoders",
"<-",
"block",
"\n",
"frame",
".",
"rawInput",
"=",
"nil",
"\n",
"d",
".",
"frames",
"<-",
"frame",
"\n",
"}",
"(",
")",
"\n",
"if",
"cap",
"(",
"dst",
")",
"==",
"0",
"{",
"// Allocate 1MB by default.",
"dst",
"=",
"make",
"(",
"[",
"]",
"byte",
",",
"0",
",",
"1",
"<<",
"20",
")",
"\n",
"}",
"\n",
"br",
":=",
"byteBuf",
"(",
"input",
")",
"\n",
"for",
"{",
"err",
":=",
"frame",
".",
"reset",
"(",
"&",
"br",
")",
"\n",
"if",
"err",
"==",
"io",
".",
"EOF",
"{",
"return",
"dst",
",",
"nil",
"\n",
"}",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"return",
"dst",
",",
"err",
"\n",
"}",
"\n",
"if",
"frame",
".",
"FrameContentSize",
">",
"d",
".",
"o",
".",
"maxDecodedSize",
"-",
"uint64",
"(",
"len",
"(",
"dst",
")",
")",
"{",
"return",
"dst",
",",
"ErrDecoderSizeExceeded",
"\n",
"}",
"\n",
"if",
"frame",
".",
"FrameContentSize",
">",
"0",
"&&",
"frame",
".",
"FrameContentSize",
"<",
"1",
"<<",
"30",
"{",
"// Never preallocate moe than 1 GB up front.",
"if",
"uint64",
"(",
"cap",
"(",
"dst",
")",
")",
"<",
"frame",
".",
"FrameContentSize",
"{",
"dst2",
":=",
"make",
"(",
"[",
"]",
"byte",
",",
"len",
"(",
"dst",
")",
",",
"len",
"(",
"dst",
")",
"+",
"int",
"(",
"frame",
".",
"FrameContentSize",
")",
")",
"\n",
"copy",
"(",
"dst2",
",",
"dst",
")",
"\n",
"dst",
"=",
"dst2",
"\n",
"}",
"\n",
"}",
"\n",
"dst",
",",
"err",
"=",
"frame",
".",
"runDecoder",
"(",
"dst",
",",
"block",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"return",
"dst",
",",
"err",
"\n",
"}",
"\n",
"if",
"len",
"(",
"br",
")",
"==",
"0",
"{",
"break",
"\n",
"}",
"\n",
"}",
"\n",
"return",
"dst",
",",
"nil",
"\n",
"}"
]
| // DecodeAll allows stateless decoding of a blob of bytes.
// Output will be appended to dst, so if the destination size is known
// you can pre-allocate the destination slice to avoid allocations.
// DecodeAll can be used concurrently.
// The Decoder concurrency limits will be respected. | [
"DecodeAll",
"allows",
"stateless",
"decoding",
"of",
"a",
"blob",
"of",
"bytes",
".",
"Output",
"will",
"be",
"appended",
"to",
"dst",
"so",
"if",
"the",
"destination",
"size",
"is",
"known",
"you",
"can",
"pre",
"-",
"allocate",
"the",
"destination",
"slice",
"to",
"avoid",
"allocations",
".",
"DecodeAll",
"can",
"be",
"used",
"concurrently",
".",
"The",
"Decoder",
"concurrency",
"limits",
"will",
"be",
"respected",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zstd/decoder.go#L242-L286 |
8,390 | klauspost/compress | zstd/decoder.go | nextBlock | func (d *Decoder) nextBlock() {
if d.current.d != nil {
d.decoders <- d.current.d
d.current.d = nil
}
if d.current.err != nil {
// Keep error state.
return
}
d.current.decodeOutput = <-d.current.output
if debug {
println("got", len(d.current.b), "bytes, error:", d.current.err)
}
} | go | func (d *Decoder) nextBlock() {
if d.current.d != nil {
d.decoders <- d.current.d
d.current.d = nil
}
if d.current.err != nil {
// Keep error state.
return
}
d.current.decodeOutput = <-d.current.output
if debug {
println("got", len(d.current.b), "bytes, error:", d.current.err)
}
} | [
"func",
"(",
"d",
"*",
"Decoder",
")",
"nextBlock",
"(",
")",
"{",
"if",
"d",
".",
"current",
".",
"d",
"!=",
"nil",
"{",
"d",
".",
"decoders",
"<-",
"d",
".",
"current",
".",
"d",
"\n",
"d",
".",
"current",
".",
"d",
"=",
"nil",
"\n",
"}",
"\n",
"if",
"d",
".",
"current",
".",
"err",
"!=",
"nil",
"{",
"// Keep error state.",
"return",
"\n",
"}",
"\n",
"d",
".",
"current",
".",
"decodeOutput",
"=",
"<-",
"d",
".",
"current",
".",
"output",
"\n",
"if",
"debug",
"{",
"println",
"(",
"\"",
"\"",
",",
"len",
"(",
"d",
".",
"current",
".",
"b",
")",
",",
"\"",
"\"",
",",
"d",
".",
"current",
".",
"err",
")",
"\n",
"}",
"\n",
"}"
]
| // nextBlock returns the next block.
// If an error occurs d.err will be set. | [
"nextBlock",
"returns",
"the",
"next",
"block",
".",
"If",
"an",
"error",
"occurs",
"d",
".",
"err",
"will",
"be",
"set",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zstd/decoder.go#L290-L303 |
8,391 | klauspost/compress | zstd/decoder.go | Close | func (d *Decoder) Close() {
if d.current.err == ErrDecoderClosed {
return
}
d.drainOutput()
if d.stream != nil {
close(d.stream)
d.streamWg.Wait()
d.stream = nil
}
if d.decoders != nil {
close(d.decoders)
for dec := range d.decoders {
dec.Close()
}
d.decoders = nil
}
if d.current.d != nil {
d.current.d.Close()
d.current.d = nil
}
d.current.err = ErrDecoderClosed
} | go | func (d *Decoder) Close() {
if d.current.err == ErrDecoderClosed {
return
}
d.drainOutput()
if d.stream != nil {
close(d.stream)
d.streamWg.Wait()
d.stream = nil
}
if d.decoders != nil {
close(d.decoders)
for dec := range d.decoders {
dec.Close()
}
d.decoders = nil
}
if d.current.d != nil {
d.current.d.Close()
d.current.d = nil
}
d.current.err = ErrDecoderClosed
} | [
"func",
"(",
"d",
"*",
"Decoder",
")",
"Close",
"(",
")",
"{",
"if",
"d",
".",
"current",
".",
"err",
"==",
"ErrDecoderClosed",
"{",
"return",
"\n",
"}",
"\n",
"d",
".",
"drainOutput",
"(",
")",
"\n",
"if",
"d",
".",
"stream",
"!=",
"nil",
"{",
"close",
"(",
"d",
".",
"stream",
")",
"\n",
"d",
".",
"streamWg",
".",
"Wait",
"(",
")",
"\n",
"d",
".",
"stream",
"=",
"nil",
"\n",
"}",
"\n",
"if",
"d",
".",
"decoders",
"!=",
"nil",
"{",
"close",
"(",
"d",
".",
"decoders",
")",
"\n",
"for",
"dec",
":=",
"range",
"d",
".",
"decoders",
"{",
"dec",
".",
"Close",
"(",
")",
"\n",
"}",
"\n",
"d",
".",
"decoders",
"=",
"nil",
"\n",
"}",
"\n",
"if",
"d",
".",
"current",
".",
"d",
"!=",
"nil",
"{",
"d",
".",
"current",
".",
"d",
".",
"Close",
"(",
")",
"\n",
"d",
".",
"current",
".",
"d",
"=",
"nil",
"\n",
"}",
"\n",
"d",
".",
"current",
".",
"err",
"=",
"ErrDecoderClosed",
"\n",
"}"
]
| // Close will release all resources.
// It is NOT possible to reuse the decoder after this. | [
"Close",
"will",
"release",
"all",
"resources",
".",
"It",
"is",
"NOT",
"possible",
"to",
"reuse",
"the",
"decoder",
"after",
"this",
"."
]
| ae52aff18558bd92cbe681549bfe9e8cbffd5903 | https://github.com/klauspost/compress/blob/ae52aff18558bd92cbe681549bfe9e8cbffd5903/zstd/decoder.go#L307-L329 |
8,392 | go-gl/gl | v4.3-compatibility/gl/package.go | BeginConditionalRender | func BeginConditionalRender(id uint32, mode uint32) {
C.glowBeginConditionalRender(gpBeginConditionalRender, (C.GLuint)(id), (C.GLenum)(mode))
} | go | func BeginConditionalRender(id uint32, mode uint32) {
C.glowBeginConditionalRender(gpBeginConditionalRender, (C.GLuint)(id), (C.GLenum)(mode))
} | [
"func",
"BeginConditionalRender",
"(",
"id",
"uint32",
",",
"mode",
"uint32",
")",
"{",
"C",
".",
"glowBeginConditionalRender",
"(",
"gpBeginConditionalRender",
",",
"(",
"C",
".",
"GLuint",
")",
"(",
"id",
")",
",",
"(",
"C",
".",
"GLenum",
")",
"(",
"mode",
")",
")",
"\n",
"}"
]
| // start conditional rendering | [
"start",
"conditional",
"rendering"
]
| bf2b1f2f34d7f6a60a99a830f65dcd6afb0c6587 | https://github.com/go-gl/gl/blob/bf2b1f2f34d7f6a60a99a830f65dcd6afb0c6587/v4.3-compatibility/gl/package.go#L19819-L19821 |
8,393 | go-gl/gl | v4.3-compatibility/gl/package.go | BindFragDataLocation | func BindFragDataLocation(program uint32, color uint32, name *uint8) {
C.glowBindFragDataLocation(gpBindFragDataLocation, (C.GLuint)(program), (C.GLuint)(color), (*C.GLchar)(unsafe.Pointer(name)))
} | go | func BindFragDataLocation(program uint32, color uint32, name *uint8) {
C.glowBindFragDataLocation(gpBindFragDataLocation, (C.GLuint)(program), (C.GLuint)(color), (*C.GLchar)(unsafe.Pointer(name)))
} | [
"func",
"BindFragDataLocation",
"(",
"program",
"uint32",
",",
"color",
"uint32",
",",
"name",
"*",
"uint8",
")",
"{",
"C",
".",
"glowBindFragDataLocation",
"(",
"gpBindFragDataLocation",
",",
"(",
"C",
".",
"GLuint",
")",
"(",
"program",
")",
",",
"(",
"C",
".",
"GLuint",
")",
"(",
"color",
")",
",",
"(",
"*",
"C",
".",
"GLchar",
")",
"(",
"unsafe",
".",
"Pointer",
"(",
"name",
")",
")",
")",
"\n",
"}"
]
| // bind a user-defined varying out variable to a fragment shader color number | [
"bind",
"a",
"user",
"-",
"defined",
"varying",
"out",
"variable",
"to",
"a",
"fragment",
"shader",
"color",
"number"
]
| bf2b1f2f34d7f6a60a99a830f65dcd6afb0c6587 | https://github.com/go-gl/gl/blob/bf2b1f2f34d7f6a60a99a830f65dcd6afb0c6587/v4.3-compatibility/gl/package.go#L19924-L19926 |
8,394 | go-gl/gl | v4.3-compatibility/gl/package.go | ClampColor | func ClampColor(target uint32, clamp uint32) {
C.glowClampColor(gpClampColor, (C.GLenum)(target), (C.GLenum)(clamp))
} | go | func ClampColor(target uint32, clamp uint32) {
C.glowClampColor(gpClampColor, (C.GLenum)(target), (C.GLenum)(clamp))
} | [
"func",
"ClampColor",
"(",
"target",
"uint32",
",",
"clamp",
"uint32",
")",
"{",
"C",
".",
"glowClampColor",
"(",
"gpClampColor",
",",
"(",
"C",
".",
"GLenum",
")",
"(",
"target",
")",
",",
"(",
"C",
".",
"GLenum",
")",
"(",
"clamp",
")",
")",
"\n",
"}"
]
| // specify whether data read via should be clamped | [
"specify",
"whether",
"data",
"read",
"via",
"should",
"be",
"clamped"
]
| bf2b1f2f34d7f6a60a99a830f65dcd6afb0c6587 | https://github.com/go-gl/gl/blob/bf2b1f2f34d7f6a60a99a830f65dcd6afb0c6587/v4.3-compatibility/gl/package.go#L20282-L20284 |
8,395 | go-gl/gl | v4.3-compatibility/gl/package.go | ColorTable | func ColorTable(target uint32, internalformat uint32, width int32, format uint32, xtype uint32, table unsafe.Pointer) {
C.glowColorTable(gpColorTable, (C.GLenum)(target), (C.GLenum)(internalformat), (C.GLsizei)(width), (C.GLenum)(format), (C.GLenum)(xtype), table)
} | go | func ColorTable(target uint32, internalformat uint32, width int32, format uint32, xtype uint32, table unsafe.Pointer) {
C.glowColorTable(gpColorTable, (C.GLenum)(target), (C.GLenum)(internalformat), (C.GLsizei)(width), (C.GLenum)(format), (C.GLenum)(xtype), table)
} | [
"func",
"ColorTable",
"(",
"target",
"uint32",
",",
"internalformat",
"uint32",
",",
"width",
"int32",
",",
"format",
"uint32",
",",
"xtype",
"uint32",
",",
"table",
"unsafe",
".",
"Pointer",
")",
"{",
"C",
".",
"glowColorTable",
"(",
"gpColorTable",
",",
"(",
"C",
".",
"GLenum",
")",
"(",
"target",
")",
",",
"(",
"C",
".",
"GLenum",
")",
"(",
"internalformat",
")",
",",
"(",
"C",
".",
"GLsizei",
")",
"(",
"width",
")",
",",
"(",
"C",
".",
"GLenum",
")",
"(",
"format",
")",
",",
"(",
"C",
".",
"GLenum",
")",
"(",
"xtype",
")",
",",
"table",
")",
"\n",
"}"
]
| // define a color lookup table | [
"define",
"a",
"color",
"lookup",
"table"
]
| bf2b1f2f34d7f6a60a99a830f65dcd6afb0c6587 | https://github.com/go-gl/gl/blob/bf2b1f2f34d7f6a60a99a830f65dcd6afb0c6587/v4.3-compatibility/gl/package.go#L20646-L20648 |
8,396 | go-gl/gl | v4.3-compatibility/gl/package.go | ConvolutionFilter1D | func ConvolutionFilter1D(target uint32, internalformat uint32, width int32, format uint32, xtype uint32, image unsafe.Pointer) {
C.glowConvolutionFilter1D(gpConvolutionFilter1D, (C.GLenum)(target), (C.GLenum)(internalformat), (C.GLsizei)(width), (C.GLenum)(format), (C.GLenum)(xtype), image)
} | go | func ConvolutionFilter1D(target uint32, internalformat uint32, width int32, format uint32, xtype uint32, image unsafe.Pointer) {
C.glowConvolutionFilter1D(gpConvolutionFilter1D, (C.GLenum)(target), (C.GLenum)(internalformat), (C.GLsizei)(width), (C.GLenum)(format), (C.GLenum)(xtype), image)
} | [
"func",
"ConvolutionFilter1D",
"(",
"target",
"uint32",
",",
"internalformat",
"uint32",
",",
"width",
"int32",
",",
"format",
"uint32",
",",
"xtype",
"uint32",
",",
"image",
"unsafe",
".",
"Pointer",
")",
"{",
"C",
".",
"glowConvolutionFilter1D",
"(",
"gpConvolutionFilter1D",
",",
"(",
"C",
".",
"GLenum",
")",
"(",
"target",
")",
",",
"(",
"C",
".",
"GLenum",
")",
"(",
"internalformat",
")",
",",
"(",
"C",
".",
"GLsizei",
")",
"(",
"width",
")",
",",
"(",
"C",
".",
"GLenum",
")",
"(",
"format",
")",
",",
"(",
"C",
".",
"GLenum",
")",
"(",
"xtype",
")",
",",
"image",
")",
"\n",
"}"
]
| // define a one-dimensional convolution filter | [
"define",
"a",
"one",
"-",
"dimensional",
"convolution",
"filter"
]
| bf2b1f2f34d7f6a60a99a830f65dcd6afb0c6587 | https://github.com/go-gl/gl/blob/bf2b1f2f34d7f6a60a99a830f65dcd6afb0c6587/v4.3-compatibility/gl/package.go#L20812-L20814 |
8,397 | go-gl/gl | v4.3-compatibility/gl/package.go | ConvolutionFilter2D | func ConvolutionFilter2D(target uint32, internalformat uint32, width int32, height int32, format uint32, xtype uint32, image unsafe.Pointer) {
C.glowConvolutionFilter2D(gpConvolutionFilter2D, (C.GLenum)(target), (C.GLenum)(internalformat), (C.GLsizei)(width), (C.GLsizei)(height), (C.GLenum)(format), (C.GLenum)(xtype), image)
} | go | func ConvolutionFilter2D(target uint32, internalformat uint32, width int32, height int32, format uint32, xtype uint32, image unsafe.Pointer) {
C.glowConvolutionFilter2D(gpConvolutionFilter2D, (C.GLenum)(target), (C.GLenum)(internalformat), (C.GLsizei)(width), (C.GLsizei)(height), (C.GLenum)(format), (C.GLenum)(xtype), image)
} | [
"func",
"ConvolutionFilter2D",
"(",
"target",
"uint32",
",",
"internalformat",
"uint32",
",",
"width",
"int32",
",",
"height",
"int32",
",",
"format",
"uint32",
",",
"xtype",
"uint32",
",",
"image",
"unsafe",
".",
"Pointer",
")",
"{",
"C",
".",
"glowConvolutionFilter2D",
"(",
"gpConvolutionFilter2D",
",",
"(",
"C",
".",
"GLenum",
")",
"(",
"target",
")",
",",
"(",
"C",
".",
"GLenum",
")",
"(",
"internalformat",
")",
",",
"(",
"C",
".",
"GLsizei",
")",
"(",
"width",
")",
",",
"(",
"C",
".",
"GLsizei",
")",
"(",
"height",
")",
",",
"(",
"C",
".",
"GLenum",
")",
"(",
"format",
")",
",",
"(",
"C",
".",
"GLenum",
")",
"(",
"xtype",
")",
",",
"image",
")",
"\n",
"}"
]
| // define a two-dimensional convolution filter | [
"define",
"a",
"two",
"-",
"dimensional",
"convolution",
"filter"
]
| bf2b1f2f34d7f6a60a99a830f65dcd6afb0c6587 | https://github.com/go-gl/gl/blob/bf2b1f2f34d7f6a60a99a830f65dcd6afb0c6587/v4.3-compatibility/gl/package.go#L20820-L20822 |
8,398 | go-gl/gl | v4.3-compatibility/gl/package.go | CopyColorTable | func CopyColorTable(target uint32, internalformat uint32, x int32, y int32, width int32) {
C.glowCopyColorTable(gpCopyColorTable, (C.GLenum)(target), (C.GLenum)(internalformat), (C.GLint)(x), (C.GLint)(y), (C.GLsizei)(width))
} | go | func CopyColorTable(target uint32, internalformat uint32, x int32, y int32, width int32) {
C.glowCopyColorTable(gpCopyColorTable, (C.GLenum)(target), (C.GLenum)(internalformat), (C.GLint)(x), (C.GLint)(y), (C.GLsizei)(width))
} | [
"func",
"CopyColorTable",
"(",
"target",
"uint32",
",",
"internalformat",
"uint32",
",",
"x",
"int32",
",",
"y",
"int32",
",",
"width",
"int32",
")",
"{",
"C",
".",
"glowCopyColorTable",
"(",
"gpCopyColorTable",
",",
"(",
"C",
".",
"GLenum",
")",
"(",
"target",
")",
",",
"(",
"C",
".",
"GLenum",
")",
"(",
"internalformat",
")",
",",
"(",
"C",
".",
"GLint",
")",
"(",
"x",
")",
",",
"(",
"C",
".",
"GLint",
")",
"(",
"y",
")",
",",
"(",
"C",
".",
"GLsizei",
")",
"(",
"width",
")",
")",
"\n",
"}"
]
| // copy pixels into a color table | [
"copy",
"pixels",
"into",
"a",
"color",
"table"
]
| bf2b1f2f34d7f6a60a99a830f65dcd6afb0c6587 | https://github.com/go-gl/gl/blob/bf2b1f2f34d7f6a60a99a830f65dcd6afb0c6587/v4.3-compatibility/gl/package.go#L20871-L20873 |
8,399 | go-gl/gl | v4.3-compatibility/gl/package.go | CopyConvolutionFilter1D | func CopyConvolutionFilter1D(target uint32, internalformat uint32, x int32, y int32, width int32) {
C.glowCopyConvolutionFilter1D(gpCopyConvolutionFilter1D, (C.GLenum)(target), (C.GLenum)(internalformat), (C.GLint)(x), (C.GLint)(y), (C.GLsizei)(width))
} | go | func CopyConvolutionFilter1D(target uint32, internalformat uint32, x int32, y int32, width int32) {
C.glowCopyConvolutionFilter1D(gpCopyConvolutionFilter1D, (C.GLenum)(target), (C.GLenum)(internalformat), (C.GLint)(x), (C.GLint)(y), (C.GLsizei)(width))
} | [
"func",
"CopyConvolutionFilter1D",
"(",
"target",
"uint32",
",",
"internalformat",
"uint32",
",",
"x",
"int32",
",",
"y",
"int32",
",",
"width",
"int32",
")",
"{",
"C",
".",
"glowCopyConvolutionFilter1D",
"(",
"gpCopyConvolutionFilter1D",
",",
"(",
"C",
".",
"GLenum",
")",
"(",
"target",
")",
",",
"(",
"C",
".",
"GLenum",
")",
"(",
"internalformat",
")",
",",
"(",
"C",
".",
"GLint",
")",
"(",
"x",
")",
",",
"(",
"C",
".",
"GLint",
")",
"(",
"y",
")",
",",
"(",
"C",
".",
"GLsizei",
")",
"(",
"width",
")",
")",
"\n",
"}"
]
| // copy pixels into a one-dimensional convolution filter | [
"copy",
"pixels",
"into",
"a",
"one",
"-",
"dimensional",
"convolution",
"filter"
]
| bf2b1f2f34d7f6a60a99a830f65dcd6afb0c6587 | https://github.com/go-gl/gl/blob/bf2b1f2f34d7f6a60a99a830f65dcd6afb0c6587/v4.3-compatibility/gl/package.go#L20879-L20881 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.