id
int32
0
167k
repo
stringlengths
5
54
path
stringlengths
4
155
func_name
stringlengths
1
118
original_string
stringlengths
52
85.5k
language
stringclasses
1 value
code
stringlengths
52
85.5k
code_tokens
listlengths
21
1.41k
docstring
stringlengths
6
2.61k
docstring_tokens
listlengths
3
215
sha
stringlengths
40
40
url
stringlengths
85
252
9,600
couchbase/gocb
error.go
IsHTTPError
func IsHTTPError(err error) bool { cause := errors.Cause(err) if nErr, ok := cause.(HTTPError); ok { return nErr.HTTPError() } return false }
go
func IsHTTPError(err error) bool { cause := errors.Cause(err) if nErr, ok := cause.(HTTPError); ok { return nErr.HTTPError() } return false }
[ "func", "IsHTTPError", "(", "err", "error", ")", "bool", "{", "cause", ":=", "errors", ".", "Cause", "(", "err", ")", "\n", "if", "nErr", ",", "ok", ":=", "cause", ".", "(", "HTTPError", ")", ";", "ok", "{", "return", "nErr", ".", "HTTPError", "(", ")", "\n", "}", "\n\n", "return", "false", "\n", "}" ]
// IsHTTPError indicates whether the passed error is a // HTTP error.
[ "IsHTTPError", "indicates", "whether", "the", "passed", "error", "is", "a", "HTTP", "error", "." ]
bca37798dd8f7c60a98041ed7089d39290c71718
https://github.com/couchbase/gocb/blob/bca37798dd8f7c60a98041ed7089d39290c71718/error.go#L182-L189
9,601
couchbase/gocb
error.go
IsNetworkError
func IsNetworkError(err error) bool { cause := errors.Cause(err) if nErr, ok := cause.(NetworkError); ok { return nErr.NetworkError() } return false }
go
func IsNetworkError(err error) bool { cause := errors.Cause(err) if nErr, ok := cause.(NetworkError); ok { return nErr.NetworkError() } return false }
[ "func", "IsNetworkError", "(", "err", "error", ")", "bool", "{", "cause", ":=", "errors", ".", "Cause", "(", "err", ")", "\n", "if", "nErr", ",", "ok", ":=", "cause", ".", "(", "NetworkError", ")", ";", "ok", "{", "return", "nErr", ".", "NetworkError", "(", ")", "\n", "}", "\n\n", "return", "false", "\n", "}" ]
// IsNetworkError indicates whether the passed error is a // network error.
[ "IsNetworkError", "indicates", "whether", "the", "passed", "error", "is", "a", "network", "error", "." ]
bca37798dd8f7c60a98041ed7089d39290c71718
https://github.com/couchbase/gocb/blob/bca37798dd8f7c60a98041ed7089d39290c71718/error.go#L193-L200
9,602
couchbase/gocb
error.go
IsServiceNotFoundError
func IsServiceNotFoundError(err error) bool { cause := errors.Cause(err) if nErr, ok := cause.(ServiceNotFoundError); ok { return nErr.ServiceNotFoundError() } return false }
go
func IsServiceNotFoundError(err error) bool { cause := errors.Cause(err) if nErr, ok := cause.(ServiceNotFoundError); ok { return nErr.ServiceNotFoundError() } return false }
[ "func", "IsServiceNotFoundError", "(", "err", "error", ")", "bool", "{", "cause", ":=", "errors", ".", "Cause", "(", "err", ")", "\n", "if", "nErr", ",", "ok", ":=", "cause", ".", "(", "ServiceNotFoundError", ")", ";", "ok", "{", "return", "nErr", ".", "ServiceNotFoundError", "(", ")", "\n", "}", "\n\n", "return", "false", "\n", "}" ]
// IsServiceNotFoundError indicates whether the passed error occurred due to // the requested service not being found.
[ "IsServiceNotFoundError", "indicates", "whether", "the", "passed", "error", "occurred", "due", "to", "the", "requested", "service", "not", "being", "found", "." ]
bca37798dd8f7c60a98041ed7089d39290c71718
https://github.com/couchbase/gocb/blob/bca37798dd8f7c60a98041ed7089d39290c71718/error.go#L204-L211
9,603
couchbase/gocb
error.go
IsTimeoutError
func IsTimeoutError(err error) bool { switch errType := errors.Cause(err).(type) { case TimeoutError: return errType.Timeout() default: return false } }
go
func IsTimeoutError(err error) bool { switch errType := errors.Cause(err).(type) { case TimeoutError: return errType.Timeout() default: return false } }
[ "func", "IsTimeoutError", "(", "err", "error", ")", "bool", "{", "switch", "errType", ":=", "errors", ".", "Cause", "(", "err", ")", ".", "(", "type", ")", "{", "case", "TimeoutError", ":", "return", "errType", ".", "Timeout", "(", ")", "\n", "default", ":", "return", "false", "\n", "}", "\n", "}" ]
// IsTimeoutError verifies whether or not the cause for an error is a timeout.
[ "IsTimeoutError", "verifies", "whether", "or", "not", "the", "cause", "for", "an", "error", "is", "a", "timeout", "." ]
bca37798dd8f7c60a98041ed7089d39290c71718
https://github.com/couchbase/gocb/blob/bca37798dd8f7c60a98041ed7089d39290c71718/error.go#L214-L221
9,604
couchbase/gocb
error.go
IsAuthenticationError
func IsAuthenticationError(err error) bool { cause := errors.Cause(err) if kvErr, ok := cause.(KeyValueError); ok && kvErr.KVError() { return kvErr.StatusCode() == int(gocbcore.StatusAuthError) } return false }
go
func IsAuthenticationError(err error) bool { cause := errors.Cause(err) if kvErr, ok := cause.(KeyValueError); ok && kvErr.KVError() { return kvErr.StatusCode() == int(gocbcore.StatusAuthError) } return false }
[ "func", "IsAuthenticationError", "(", "err", "error", ")", "bool", "{", "cause", ":=", "errors", ".", "Cause", "(", "err", ")", "\n", "if", "kvErr", ",", "ok", ":=", "cause", ".", "(", "KeyValueError", ")", ";", "ok", "&&", "kvErr", ".", "KVError", "(", ")", "{", "return", "kvErr", ".", "StatusCode", "(", ")", "==", "int", "(", "gocbcore", ".", "StatusAuthError", ")", "\n", "}", "\n\n", "return", "false", "\n", "}" ]
// IsAuthenticationError verifies whether or not the cause for an error is an authentication error.
[ "IsAuthenticationError", "verifies", "whether", "or", "not", "the", "cause", "for", "an", "error", "is", "an", "authentication", "error", "." ]
bca37798dd8f7c60a98041ed7089d39290c71718
https://github.com/couchbase/gocb/blob/bca37798dd8f7c60a98041ed7089d39290c71718/error.go#L224-L231
9,605
couchbase/gocb
error.go
IsBucketMissingError
func IsBucketMissingError(err error) bool { cause := errors.Cause(err) if kvErr, ok := cause.(KeyValueError); ok && kvErr.KVError() { return kvErr.StatusCode() == int(gocbcore.StatusNoBucket) } return false }
go
func IsBucketMissingError(err error) bool { cause := errors.Cause(err) if kvErr, ok := cause.(KeyValueError); ok && kvErr.KVError() { return kvErr.StatusCode() == int(gocbcore.StatusNoBucket) } return false }
[ "func", "IsBucketMissingError", "(", "err", "error", ")", "bool", "{", "cause", ":=", "errors", ".", "Cause", "(", "err", ")", "\n", "if", "kvErr", ",", "ok", ":=", "cause", ".", "(", "KeyValueError", ")", ";", "ok", "&&", "kvErr", ".", "KVError", "(", ")", "{", "return", "kvErr", ".", "StatusCode", "(", ")", "==", "int", "(", "gocbcore", ".", "StatusNoBucket", ")", "\n", "}", "\n\n", "return", "false", "\n", "}" ]
// IsBucketMissingError verifies whether or not the cause for an error is a bucket missing error.
[ "IsBucketMissingError", "verifies", "whether", "or", "not", "the", "cause", "for", "an", "error", "is", "a", "bucket", "missing", "error", "." ]
bca37798dd8f7c60a98041ed7089d39290c71718
https://github.com/couchbase/gocb/blob/bca37798dd8f7c60a98041ed7089d39290c71718/error.go#L234-L241
9,606
couchbase/gocb
error.go
IsAccessError
func IsAccessError(err error) bool { cause := errors.Cause(err) if kvErr, ok := cause.(KeyValueError); ok && kvErr.KVError() { return kvErr.StatusCode() == int(gocbcore.StatusAccessError) } return false }
go
func IsAccessError(err error) bool { cause := errors.Cause(err) if kvErr, ok := cause.(KeyValueError); ok && kvErr.KVError() { return kvErr.StatusCode() == int(gocbcore.StatusAccessError) } return false }
[ "func", "IsAccessError", "(", "err", "error", ")", "bool", "{", "cause", ":=", "errors", ".", "Cause", "(", "err", ")", "\n", "if", "kvErr", ",", "ok", ":=", "cause", ".", "(", "KeyValueError", ")", ";", "ok", "&&", "kvErr", ".", "KVError", "(", ")", "{", "return", "kvErr", ".", "StatusCode", "(", ")", "==", "int", "(", "gocbcore", ".", "StatusAccessError", ")", "\n", "}", "\n\n", "return", "false", "\n", "}" ]
// IsAccessError verifies whether or not the cause for an error is an access error.
[ "IsAccessError", "verifies", "whether", "or", "not", "the", "cause", "for", "an", "error", "is", "an", "access", "error", "." ]
bca37798dd8f7c60a98041ed7089d39290c71718
https://github.com/couchbase/gocb/blob/bca37798dd8f7c60a98041ed7089d39290c71718/error.go#L244-L251
9,607
couchbase/gocb
error.go
IsConfigurationError
func IsConfigurationError(err error) bool { switch errType := errors.Cause(err).(type) { case ConfigurationError: return errType.ConfigurationError() default: return false } }
go
func IsConfigurationError(err error) bool { switch errType := errors.Cause(err).(type) { case ConfigurationError: return errType.ConfigurationError() default: return false } }
[ "func", "IsConfigurationError", "(", "err", "error", ")", "bool", "{", "switch", "errType", ":=", "errors", ".", "Cause", "(", "err", ")", ".", "(", "type", ")", "{", "case", "ConfigurationError", ":", "return", "errType", ".", "ConfigurationError", "(", ")", "\n", "default", ":", "return", "false", "\n", "}", "\n", "}" ]
// IsConfigurationError verifies whether or not the cause for an error is a configuration error.
[ "IsConfigurationError", "verifies", "whether", "or", "not", "the", "cause", "for", "an", "error", "is", "a", "configuration", "error", "." ]
bca37798dd8f7c60a98041ed7089d39290c71718
https://github.com/couchbase/gocb/blob/bca37798dd8f7c60a98041ed7089d39290c71718/error.go#L254-L261
9,608
couchbase/gocb
error.go
IsSubdocPathNotFoundError
func IsSubdocPathNotFoundError(err error) bool { cause := errors.Cause(err) if kvErr, ok := cause.(KeyValueError); ok && kvErr.KVError() { return kvErr.StatusCode() == int(gocbcore.StatusSubDocPathNotFound) } return false }
go
func IsSubdocPathNotFoundError(err error) bool { cause := errors.Cause(err) if kvErr, ok := cause.(KeyValueError); ok && kvErr.KVError() { return kvErr.StatusCode() == int(gocbcore.StatusSubDocPathNotFound) } return false }
[ "func", "IsSubdocPathNotFoundError", "(", "err", "error", ")", "bool", "{", "cause", ":=", "errors", ".", "Cause", "(", "err", ")", "\n", "if", "kvErr", ",", "ok", ":=", "cause", ".", "(", "KeyValueError", ")", ";", "ok", "&&", "kvErr", ".", "KVError", "(", ")", "{", "return", "kvErr", ".", "StatusCode", "(", ")", "==", "int", "(", "gocbcore", ".", "StatusSubDocPathNotFound", ")", "\n", "}", "\n\n", "return", "false", "\n", "}" ]
// IsSubdocPathNotFoundError verifies whether or not the cause for an error is due to a subdoc operation path not found.
[ "IsSubdocPathNotFoundError", "verifies", "whether", "or", "not", "the", "cause", "for", "an", "error", "is", "due", "to", "a", "subdoc", "operation", "path", "not", "found", "." ]
bca37798dd8f7c60a98041ed7089d39290c71718
https://github.com/couchbase/gocb/blob/bca37798dd8f7c60a98041ed7089d39290c71718/error.go#L264-L271
9,609
couchbase/gocb
error.go
IsDurabilityError
func IsDurabilityError(err error) bool { switch errType := errors.Cause(err).(type) { case DurabilityError: return errType.DurabilityError() default: return false } }
go
func IsDurabilityError(err error) bool { switch errType := errors.Cause(err).(type) { case DurabilityError: return errType.DurabilityError() default: return false } }
[ "func", "IsDurabilityError", "(", "err", "error", ")", "bool", "{", "switch", "errType", ":=", "errors", ".", "Cause", "(", "err", ")", ".", "(", "type", ")", "{", "case", "DurabilityError", ":", "return", "errType", ".", "DurabilityError", "(", ")", "\n", "default", ":", "return", "false", "\n", "}", "\n", "}" ]
// IsDurabilityError verifies whether or not the cause for an error is due to a durability error.
[ "IsDurabilityError", "verifies", "whether", "or", "not", "the", "cause", "for", "an", "error", "is", "due", "to", "a", "durability", "error", "." ]
bca37798dd8f7c60a98041ed7089d39290c71718
https://github.com/couchbase/gocb/blob/bca37798dd8f7c60a98041ed7089d39290c71718/error.go#L284-L291
9,610
couchbase/gocb
error.go
IsNoResultsError
func IsNoResultsError(err error) bool { switch errType := errors.Cause(err).(type) { case NoResultsError: return errType.NoResultsError() default: return false } }
go
func IsNoResultsError(err error) bool { switch errType := errors.Cause(err).(type) { case NoResultsError: return errType.NoResultsError() default: return false } }
[ "func", "IsNoResultsError", "(", "err", "error", ")", "bool", "{", "switch", "errType", ":=", "errors", ".", "Cause", "(", "err", ")", ".", "(", "type", ")", "{", "case", "NoResultsError", ":", "return", "errType", ".", "NoResultsError", "(", ")", "\n", "default", ":", "return", "false", "\n", "}", "\n", "}" ]
// IsNoResultsError verifies whether or not the cause for an error is due no results being available to a query.
[ "IsNoResultsError", "verifies", "whether", "or", "not", "the", "cause", "for", "an", "error", "is", "due", "no", "results", "being", "available", "to", "a", "query", "." ]
bca37798dd8f7c60a98041ed7089d39290c71718
https://github.com/couchbase/gocb/blob/bca37798dd8f7c60a98041ed7089d39290c71718/error.go#L294-L301
9,611
couchbase/gocb
cluster_analyticsquery.go
NextBytes
func (r *AnalyticsResults) NextBytes() []byte { if r.streamResult.Closed() { return nil } raw, err := r.streamResult.NextBytes() if err != nil { r.err = err return nil } return raw }
go
func (r *AnalyticsResults) NextBytes() []byte { if r.streamResult.Closed() { return nil } raw, err := r.streamResult.NextBytes() if err != nil { r.err = err return nil } return raw }
[ "func", "(", "r", "*", "AnalyticsResults", ")", "NextBytes", "(", ")", "[", "]", "byte", "{", "if", "r", ".", "streamResult", ".", "Closed", "(", ")", "{", "return", "nil", "\n", "}", "\n\n", "raw", ",", "err", ":=", "r", ".", "streamResult", ".", "NextBytes", "(", ")", "\n", "if", "err", "!=", "nil", "{", "r", ".", "err", "=", "err", "\n", "return", "nil", "\n", "}", "\n\n", "return", "raw", "\n", "}" ]
// NextBytes returns the next result from the results as a byte array.
[ "NextBytes", "returns", "the", "next", "result", "from", "the", "results", "as", "a", "byte", "array", "." ]
bca37798dd8f7c60a98041ed7089d39290c71718
https://github.com/couchbase/gocb/blob/bca37798dd8f7c60a98041ed7089d39290c71718/cluster_analyticsquery.go#L103-L115
9,612
couchbase/gocb
cluster_analyticsquery.go
AnalyticsQuery
func (c *Cluster) AnalyticsQuery(statement string, opts *AnalyticsQueryOptions) (*AnalyticsResults, error) { if opts == nil { opts = &AnalyticsQueryOptions{} } ctx := opts.Context if ctx == nil { ctx = context.Background() } var span opentracing.Span if opts.ParentSpanContext == nil { span = opentracing.GlobalTracer().StartSpan("ExecuteAnalyticsQuery", opentracing.Tag{Key: "couchbase.service", Value: "cbas"}) } else { span = opentracing.GlobalTracer().StartSpan("ExecuteAnalyticsQuery", opentracing.Tag{Key: "couchbase.service", Value: "cbas"}, opentracing.ChildOf(opts.ParentSpanContext)) } defer span.Finish() provider, err := c.getHTTPProvider() if err != nil { return nil, err } return c.analyticsQuery(ctx, span.Context(), statement, opts, provider) }
go
func (c *Cluster) AnalyticsQuery(statement string, opts *AnalyticsQueryOptions) (*AnalyticsResults, error) { if opts == nil { opts = &AnalyticsQueryOptions{} } ctx := opts.Context if ctx == nil { ctx = context.Background() } var span opentracing.Span if opts.ParentSpanContext == nil { span = opentracing.GlobalTracer().StartSpan("ExecuteAnalyticsQuery", opentracing.Tag{Key: "couchbase.service", Value: "cbas"}) } else { span = opentracing.GlobalTracer().StartSpan("ExecuteAnalyticsQuery", opentracing.Tag{Key: "couchbase.service", Value: "cbas"}, opentracing.ChildOf(opts.ParentSpanContext)) } defer span.Finish() provider, err := c.getHTTPProvider() if err != nil { return nil, err } return c.analyticsQuery(ctx, span.Context(), statement, opts, provider) }
[ "func", "(", "c", "*", "Cluster", ")", "AnalyticsQuery", "(", "statement", "string", ",", "opts", "*", "AnalyticsQueryOptions", ")", "(", "*", "AnalyticsResults", ",", "error", ")", "{", "if", "opts", "==", "nil", "{", "opts", "=", "&", "AnalyticsQueryOptions", "{", "}", "\n", "}", "\n", "ctx", ":=", "opts", ".", "Context", "\n", "if", "ctx", "==", "nil", "{", "ctx", "=", "context", ".", "Background", "(", ")", "\n", "}", "\n\n", "var", "span", "opentracing", ".", "Span", "\n", "if", "opts", ".", "ParentSpanContext", "==", "nil", "{", "span", "=", "opentracing", ".", "GlobalTracer", "(", ")", ".", "StartSpan", "(", "\"", "\"", ",", "opentracing", ".", "Tag", "{", "Key", ":", "\"", "\"", ",", "Value", ":", "\"", "\"", "}", ")", "\n", "}", "else", "{", "span", "=", "opentracing", ".", "GlobalTracer", "(", ")", ".", "StartSpan", "(", "\"", "\"", ",", "opentracing", ".", "Tag", "{", "Key", ":", "\"", "\"", ",", "Value", ":", "\"", "\"", "}", ",", "opentracing", ".", "ChildOf", "(", "opts", ".", "ParentSpanContext", ")", ")", "\n", "}", "\n", "defer", "span", ".", "Finish", "(", ")", "\n\n", "provider", ",", "err", ":=", "c", ".", "getHTTPProvider", "(", ")", "\n", "if", "err", "!=", "nil", "{", "return", "nil", ",", "err", "\n", "}", "\n\n", "return", "c", ".", "analyticsQuery", "(", "ctx", ",", "span", ".", "Context", "(", ")", ",", "statement", ",", "opts", ",", "provider", ")", "\n", "}" ]
// AnalyticsQuery performs an analytics query and returns a list of rows or an error.
[ "AnalyticsQuery", "performs", "an", "analytics", "query", "and", "returns", "a", "list", "of", "rows", "or", "an", "error", "." ]
bca37798dd8f7c60a98041ed7089d39290c71718
https://github.com/couchbase/gocb/blob/bca37798dd8f7c60a98041ed7089d39290c71718/cluster_analyticsquery.go#L319-L344
9,613
apex/gateway
gateway.go
ListenAndServe
func ListenAndServe(addr string, h http.Handler) error { if h == nil { h = http.DefaultServeMux } lambda.Start(func(ctx context.Context, e events.APIGatewayProxyRequest) (events.APIGatewayProxyResponse, error) { r, err := NewRequest(ctx, e) if err != nil { return events.APIGatewayProxyResponse{}, err } w := NewResponse() h.ServeHTTP(w, r) return w.End(), nil }) return nil }
go
func ListenAndServe(addr string, h http.Handler) error { if h == nil { h = http.DefaultServeMux } lambda.Start(func(ctx context.Context, e events.APIGatewayProxyRequest) (events.APIGatewayProxyResponse, error) { r, err := NewRequest(ctx, e) if err != nil { return events.APIGatewayProxyResponse{}, err } w := NewResponse() h.ServeHTTP(w, r) return w.End(), nil }) return nil }
[ "func", "ListenAndServe", "(", "addr", "string", ",", "h", "http", ".", "Handler", ")", "error", "{", "if", "h", "==", "nil", "{", "h", "=", "http", ".", "DefaultServeMux", "\n", "}", "\n\n", "lambda", ".", "Start", "(", "func", "(", "ctx", "context", ".", "Context", ",", "e", "events", ".", "APIGatewayProxyRequest", ")", "(", "events", ".", "APIGatewayProxyResponse", ",", "error", ")", "{", "r", ",", "err", ":=", "NewRequest", "(", "ctx", ",", "e", ")", "\n", "if", "err", "!=", "nil", "{", "return", "events", ".", "APIGatewayProxyResponse", "{", "}", ",", "err", "\n", "}", "\n\n", "w", ":=", "NewResponse", "(", ")", "\n", "h", ".", "ServeHTTP", "(", "w", ",", "r", ")", "\n", "return", "w", ".", "End", "(", ")", ",", "nil", "\n", "}", ")", "\n\n", "return", "nil", "\n", "}" ]
// ListenAndServe is a drop-in replacement for // http.ListenAndServe for use within AWS Lambda. // // ListenAndServe always returns a non-nil error.
[ "ListenAndServe", "is", "a", "drop", "-", "in", "replacement", "for", "http", ".", "ListenAndServe", "for", "use", "within", "AWS", "Lambda", ".", "ListenAndServe", "always", "returns", "a", "non", "-", "nil", "error", "." ]
21c9971861be4e8be9352e7a2540710b0c46610f
https://github.com/apex/gateway/blob/21c9971861be4e8be9352e7a2540710b0c46610f/gateway.go#L16-L33
9,614
apex/gateway
context.go
newContext
func newContext(ctx context.Context, e events.APIGatewayProxyRequest) context.Context { return context.WithValue(ctx, requestContextKey, e.RequestContext) }
go
func newContext(ctx context.Context, e events.APIGatewayProxyRequest) context.Context { return context.WithValue(ctx, requestContextKey, e.RequestContext) }
[ "func", "newContext", "(", "ctx", "context", ".", "Context", ",", "e", "events", ".", "APIGatewayProxyRequest", ")", "context", ".", "Context", "{", "return", "context", ".", "WithValue", "(", "ctx", ",", "requestContextKey", ",", "e", ".", "RequestContext", ")", "\n", "}" ]
// newContext returns a new Context with specific api gateway proxy values.
[ "newContext", "returns", "a", "new", "Context", "with", "specific", "api", "gateway", "proxy", "values", "." ]
21c9971861be4e8be9352e7a2540710b0c46610f
https://github.com/apex/gateway/blob/21c9971861be4e8be9352e7a2540710b0c46610f/context.go#L16-L18
9,615
apex/gateway
context.go
RequestContext
func RequestContext(ctx context.Context) (events.APIGatewayProxyRequestContext, bool) { c, ok := ctx.Value(requestContextKey).(events.APIGatewayProxyRequestContext) return c, ok }
go
func RequestContext(ctx context.Context) (events.APIGatewayProxyRequestContext, bool) { c, ok := ctx.Value(requestContextKey).(events.APIGatewayProxyRequestContext) return c, ok }
[ "func", "RequestContext", "(", "ctx", "context", ".", "Context", ")", "(", "events", ".", "APIGatewayProxyRequestContext", ",", "bool", ")", "{", "c", ",", "ok", ":=", "ctx", ".", "Value", "(", "requestContextKey", ")", ".", "(", "events", ".", "APIGatewayProxyRequestContext", ")", "\n", "return", "c", ",", "ok", "\n", "}" ]
// RequestContext returns the APIGatewayProxyRequestContext value stored in ctx.
[ "RequestContext", "returns", "the", "APIGatewayProxyRequestContext", "value", "stored", "in", "ctx", "." ]
21c9971861be4e8be9352e7a2540710b0c46610f
https://github.com/apex/gateway/blob/21c9971861be4e8be9352e7a2540710b0c46610f/context.go#L21-L24
9,616
apex/gateway
response.go
Header
func (w *ResponseWriter) Header() http.Header { if w.header == nil { w.header = make(http.Header) } return w.header }
go
func (w *ResponseWriter) Header() http.Header { if w.header == nil { w.header = make(http.Header) } return w.header }
[ "func", "(", "w", "*", "ResponseWriter", ")", "Header", "(", ")", "http", ".", "Header", "{", "if", "w", ".", "header", "==", "nil", "{", "w", ".", "header", "=", "make", "(", "http", ".", "Header", ")", "\n", "}", "\n\n", "return", "w", ".", "header", "\n", "}" ]
// Header implementation.
[ "Header", "implementation", "." ]
21c9971861be4e8be9352e7a2540710b0c46610f
https://github.com/apex/gateway/blob/21c9971861be4e8be9352e7a2540710b0c46610f/response.go#L31-L37
9,617
apex/gateway
response.go
Write
func (w *ResponseWriter) Write(b []byte) (int, error) { if !w.wroteHeader { w.WriteHeader(http.StatusOK) } return w.buf.Write(b) }
go
func (w *ResponseWriter) Write(b []byte) (int, error) { if !w.wroteHeader { w.WriteHeader(http.StatusOK) } return w.buf.Write(b) }
[ "func", "(", "w", "*", "ResponseWriter", ")", "Write", "(", "b", "[", "]", "byte", ")", "(", "int", ",", "error", ")", "{", "if", "!", "w", ".", "wroteHeader", "{", "w", ".", "WriteHeader", "(", "http", ".", "StatusOK", ")", "\n", "}", "\n\n", "return", "w", ".", "buf", ".", "Write", "(", "b", ")", "\n", "}" ]
// Write implementation.
[ "Write", "implementation", "." ]
21c9971861be4e8be9352e7a2540710b0c46610f
https://github.com/apex/gateway/blob/21c9971861be4e8be9352e7a2540710b0c46610f/response.go#L40-L46
9,618
apex/gateway
response.go
WriteHeader
func (w *ResponseWriter) WriteHeader(status int) { if w.wroteHeader { return } if w.Header().Get("Content-Type") == "" { w.Header().Set("Content-Type", "text/plain; charset=utf8") } w.out.StatusCode = status h := make(map[string]string) for k, v := range w.Header() { if len(v) > 0 { h[k] = v[len(v)-1] } } w.out.Headers = h w.wroteHeader = true }
go
func (w *ResponseWriter) WriteHeader(status int) { if w.wroteHeader { return } if w.Header().Get("Content-Type") == "" { w.Header().Set("Content-Type", "text/plain; charset=utf8") } w.out.StatusCode = status h := make(map[string]string) for k, v := range w.Header() { if len(v) > 0 { h[k] = v[len(v)-1] } } w.out.Headers = h w.wroteHeader = true }
[ "func", "(", "w", "*", "ResponseWriter", ")", "WriteHeader", "(", "status", "int", ")", "{", "if", "w", ".", "wroteHeader", "{", "return", "\n", "}", "\n\n", "if", "w", ".", "Header", "(", ")", ".", "Get", "(", "\"", "\"", ")", "==", "\"", "\"", "{", "w", ".", "Header", "(", ")", ".", "Set", "(", "\"", "\"", ",", "\"", "\"", ")", "\n", "}", "\n\n", "w", ".", "out", ".", "StatusCode", "=", "status", "\n\n", "h", ":=", "make", "(", "map", "[", "string", "]", "string", ")", "\n\n", "for", "k", ",", "v", ":=", "range", "w", ".", "Header", "(", ")", "{", "if", "len", "(", "v", ")", ">", "0", "{", "h", "[", "k", "]", "=", "v", "[", "len", "(", "v", ")", "-", "1", "]", "\n", "}", "\n", "}", "\n\n", "w", ".", "out", ".", "Headers", "=", "h", "\n", "w", ".", "wroteHeader", "=", "true", "\n", "}" ]
// WriteHeader implementation.
[ "WriteHeader", "implementation", "." ]
21c9971861be4e8be9352e7a2540710b0c46610f
https://github.com/apex/gateway/blob/21c9971861be4e8be9352e7a2540710b0c46610f/response.go#L49-L70
9,619
apex/gateway
response.go
End
func (w *ResponseWriter) End() events.APIGatewayProxyResponse { w.out.IsBase64Encoded = isBinary(w.header) if w.out.IsBase64Encoded { w.out.Body = base64.StdEncoding.EncodeToString(w.buf.Bytes()) } else { w.out.Body = w.buf.String() } // notify end w.closeNotifyCh <- true return w.out }
go
func (w *ResponseWriter) End() events.APIGatewayProxyResponse { w.out.IsBase64Encoded = isBinary(w.header) if w.out.IsBase64Encoded { w.out.Body = base64.StdEncoding.EncodeToString(w.buf.Bytes()) } else { w.out.Body = w.buf.String() } // notify end w.closeNotifyCh <- true return w.out }
[ "func", "(", "w", "*", "ResponseWriter", ")", "End", "(", ")", "events", ".", "APIGatewayProxyResponse", "{", "w", ".", "out", ".", "IsBase64Encoded", "=", "isBinary", "(", "w", ".", "header", ")", "\n\n", "if", "w", ".", "out", ".", "IsBase64Encoded", "{", "w", ".", "out", ".", "Body", "=", "base64", ".", "StdEncoding", ".", "EncodeToString", "(", "w", ".", "buf", ".", "Bytes", "(", ")", ")", "\n", "}", "else", "{", "w", ".", "out", ".", "Body", "=", "w", ".", "buf", ".", "String", "(", ")", "\n", "}", "\n\n", "// notify end", "w", ".", "closeNotifyCh", "<-", "true", "\n\n", "return", "w", ".", "out", "\n", "}" ]
// End the request.
[ "End", "the", "request", "." ]
21c9971861be4e8be9352e7a2540710b0c46610f
https://github.com/apex/gateway/blob/21c9971861be4e8be9352e7a2540710b0c46610f/response.go#L78-L91
9,620
apex/gateway
response.go
isBinary
func isBinary(h http.Header) bool { switch { case !isTextMime(h.Get("Content-Type")): return true case h.Get("Content-Encoding") == "gzip": return true default: return false } }
go
func isBinary(h http.Header) bool { switch { case !isTextMime(h.Get("Content-Type")): return true case h.Get("Content-Encoding") == "gzip": return true default: return false } }
[ "func", "isBinary", "(", "h", "http", ".", "Header", ")", "bool", "{", "switch", "{", "case", "!", "isTextMime", "(", "h", ".", "Get", "(", "\"", "\"", ")", ")", ":", "return", "true", "\n", "case", "h", ".", "Get", "(", "\"", "\"", ")", "==", "\"", "\"", ":", "return", "true", "\n", "default", ":", "return", "false", "\n", "}", "\n", "}" ]
// isBinary returns true if the response reprensents binary.
[ "isBinary", "returns", "true", "if", "the", "response", "reprensents", "binary", "." ]
21c9971861be4e8be9352e7a2540710b0c46610f
https://github.com/apex/gateway/blob/21c9971861be4e8be9352e7a2540710b0c46610f/response.go#L94-L103
9,621
apex/gateway
response.go
isTextMime
func isTextMime(kind string) bool { mt, _, err := mime.ParseMediaType(kind) if err != nil { return false } if strings.HasPrefix(mt, "text/") { return true } switch mt { case "image/svg+xml": return true case "application/json": return true case "application/xml": return true default: return false } }
go
func isTextMime(kind string) bool { mt, _, err := mime.ParseMediaType(kind) if err != nil { return false } if strings.HasPrefix(mt, "text/") { return true } switch mt { case "image/svg+xml": return true case "application/json": return true case "application/xml": return true default: return false } }
[ "func", "isTextMime", "(", "kind", "string", ")", "bool", "{", "mt", ",", "_", ",", "err", ":=", "mime", ".", "ParseMediaType", "(", "kind", ")", "\n", "if", "err", "!=", "nil", "{", "return", "false", "\n", "}", "\n\n", "if", "strings", ".", "HasPrefix", "(", "mt", ",", "\"", "\"", ")", "{", "return", "true", "\n", "}", "\n\n", "switch", "mt", "{", "case", "\"", "\"", ":", "return", "true", "\n", "case", "\"", "\"", ":", "return", "true", "\n", "case", "\"", "\"", ":", "return", "true", "\n", "default", ":", "return", "false", "\n", "}", "\n", "}" ]
// isTextMime returns true if the content type represents textual data.
[ "isTextMime", "returns", "true", "if", "the", "content", "type", "represents", "textual", "data", "." ]
21c9971861be4e8be9352e7a2540710b0c46610f
https://github.com/apex/gateway/blob/21c9971861be4e8be9352e7a2540710b0c46610f/response.go#L106-L126
9,622
apex/gateway
request.go
NewRequest
func NewRequest(ctx context.Context, e events.APIGatewayProxyRequest) (*http.Request, error) { // path u, err := url.Parse(e.Path) if err != nil { return nil, errors.Wrap(err, "parsing path") } // querystring q := u.Query() for k, v := range e.QueryStringParameters { q.Set(k, v) } u.RawQuery = q.Encode() // base64 encoded body body := e.Body if e.IsBase64Encoded { b, err := base64.StdEncoding.DecodeString(body) if err != nil { return nil, errors.Wrap(err, "decoding base64 body") } body = string(b) } // new request req, err := http.NewRequest(e.HTTPMethod, u.String(), strings.NewReader(body)) if err != nil { return nil, errors.Wrap(err, "creating request") } // remote addr req.RemoteAddr = e.RequestContext.Identity.SourceIP // header fields for k, v := range e.Headers { req.Header.Set(k, v) } // content-length if req.Header.Get("Content-Length") == "" && body != "" { req.Header.Set("Content-Length", strconv.Itoa(len(body))) } // custom fields req.Header.Set("X-Request-Id", e.RequestContext.RequestID) req.Header.Set("X-Stage", e.RequestContext.Stage) // custom context values req = req.WithContext(newContext(ctx, e)) // xray support if traceID := ctx.Value("x-amzn-trace-id"); traceID != nil { req.Header.Set("X-Amzn-Trace-Id", fmt.Sprintf("%v", traceID)) } // host req.URL.Host = req.Header.Get("Host") req.Host = req.URL.Host return req, nil }
go
func NewRequest(ctx context.Context, e events.APIGatewayProxyRequest) (*http.Request, error) { // path u, err := url.Parse(e.Path) if err != nil { return nil, errors.Wrap(err, "parsing path") } // querystring q := u.Query() for k, v := range e.QueryStringParameters { q.Set(k, v) } u.RawQuery = q.Encode() // base64 encoded body body := e.Body if e.IsBase64Encoded { b, err := base64.StdEncoding.DecodeString(body) if err != nil { return nil, errors.Wrap(err, "decoding base64 body") } body = string(b) } // new request req, err := http.NewRequest(e.HTTPMethod, u.String(), strings.NewReader(body)) if err != nil { return nil, errors.Wrap(err, "creating request") } // remote addr req.RemoteAddr = e.RequestContext.Identity.SourceIP // header fields for k, v := range e.Headers { req.Header.Set(k, v) } // content-length if req.Header.Get("Content-Length") == "" && body != "" { req.Header.Set("Content-Length", strconv.Itoa(len(body))) } // custom fields req.Header.Set("X-Request-Id", e.RequestContext.RequestID) req.Header.Set("X-Stage", e.RequestContext.Stage) // custom context values req = req.WithContext(newContext(ctx, e)) // xray support if traceID := ctx.Value("x-amzn-trace-id"); traceID != nil { req.Header.Set("X-Amzn-Trace-Id", fmt.Sprintf("%v", traceID)) } // host req.URL.Host = req.Header.Get("Host") req.Host = req.URL.Host return req, nil }
[ "func", "NewRequest", "(", "ctx", "context", ".", "Context", ",", "e", "events", ".", "APIGatewayProxyRequest", ")", "(", "*", "http", ".", "Request", ",", "error", ")", "{", "// path", "u", ",", "err", ":=", "url", ".", "Parse", "(", "e", ".", "Path", ")", "\n", "if", "err", "!=", "nil", "{", "return", "nil", ",", "errors", ".", "Wrap", "(", "err", ",", "\"", "\"", ")", "\n", "}", "\n\n", "// querystring", "q", ":=", "u", ".", "Query", "(", ")", "\n", "for", "k", ",", "v", ":=", "range", "e", ".", "QueryStringParameters", "{", "q", ".", "Set", "(", "k", ",", "v", ")", "\n", "}", "\n", "u", ".", "RawQuery", "=", "q", ".", "Encode", "(", ")", "\n\n", "// base64 encoded body", "body", ":=", "e", ".", "Body", "\n", "if", "e", ".", "IsBase64Encoded", "{", "b", ",", "err", ":=", "base64", ".", "StdEncoding", ".", "DecodeString", "(", "body", ")", "\n", "if", "err", "!=", "nil", "{", "return", "nil", ",", "errors", ".", "Wrap", "(", "err", ",", "\"", "\"", ")", "\n", "}", "\n", "body", "=", "string", "(", "b", ")", "\n", "}", "\n\n", "// new request", "req", ",", "err", ":=", "http", ".", "NewRequest", "(", "e", ".", "HTTPMethod", ",", "u", ".", "String", "(", ")", ",", "strings", ".", "NewReader", "(", "body", ")", ")", "\n", "if", "err", "!=", "nil", "{", "return", "nil", ",", "errors", ".", "Wrap", "(", "err", ",", "\"", "\"", ")", "\n", "}", "\n\n", "// remote addr", "req", ".", "RemoteAddr", "=", "e", ".", "RequestContext", ".", "Identity", ".", "SourceIP", "\n\n", "// header fields", "for", "k", ",", "v", ":=", "range", "e", ".", "Headers", "{", "req", ".", "Header", ".", "Set", "(", "k", ",", "v", ")", "\n", "}", "\n\n", "// content-length", "if", "req", ".", "Header", ".", "Get", "(", "\"", "\"", ")", "==", "\"", "\"", "&&", "body", "!=", "\"", "\"", "{", "req", ".", "Header", ".", "Set", "(", "\"", "\"", ",", "strconv", ".", "Itoa", "(", "len", "(", "body", ")", ")", ")", "\n", "}", "\n\n", "// custom fields", "req", ".", "Header", ".", "Set", "(", "\"", "\"", ",", "e", ".", "RequestContext", ".", "RequestID", ")", "\n", "req", ".", "Header", ".", "Set", "(", "\"", "\"", ",", "e", ".", "RequestContext", ".", "Stage", ")", "\n\n", "// custom context values", "req", "=", "req", ".", "WithContext", "(", "newContext", "(", "ctx", ",", "e", ")", ")", "\n\n", "// xray support", "if", "traceID", ":=", "ctx", ".", "Value", "(", "\"", "\"", ")", ";", "traceID", "!=", "nil", "{", "req", ".", "Header", ".", "Set", "(", "\"", "\"", ",", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "traceID", ")", ")", "\n", "}", "\n\n", "// host", "req", ".", "URL", ".", "Host", "=", "req", ".", "Header", ".", "Get", "(", "\"", "\"", ")", "\n", "req", ".", "Host", "=", "req", ".", "URL", ".", "Host", "\n\n", "return", "req", ",", "nil", "\n", "}" ]
// NewRequest returns a new http.Request from the given Lambda event.
[ "NewRequest", "returns", "a", "new", "http", ".", "Request", "from", "the", "given", "Lambda", "event", "." ]
21c9971861be4e8be9352e7a2540710b0c46610f
https://github.com/apex/gateway/blob/21c9971861be4e8be9352e7a2540710b0c46610f/request.go#L17-L77
9,623
dsnet/compress
internal/prefix/writer.go
BitsWritten
func (pw *Writer) BitsWritten() int64 { return 8*pw.Offset + 8*int64(pw.cntBuf) + int64(pw.numBits) }
go
func (pw *Writer) BitsWritten() int64 { return 8*pw.Offset + 8*int64(pw.cntBuf) + int64(pw.numBits) }
[ "func", "(", "pw", "*", "Writer", ")", "BitsWritten", "(", ")", "int64", "{", "return", "8", "*", "pw", ".", "Offset", "+", "8", "*", "int64", "(", "pw", ".", "cntBuf", ")", "+", "int64", "(", "pw", ".", "numBits", ")", "\n", "}" ]
// BitsWritten reports the total number of bits issued to any Write method.
[ "BitsWritten", "reports", "the", "total", "number", "of", "bits", "issued", "to", "any", "Write", "method", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/internal/prefix/writer.go#L38-L40
9,624
dsnet/compress
internal/prefix/writer.go
WritePads
func (pw *Writer) WritePads(v uint) { nb := -pw.numBits & 7 pw.bufBits |= uint64(v) << pw.numBits pw.numBits += nb }
go
func (pw *Writer) WritePads(v uint) { nb := -pw.numBits & 7 pw.bufBits |= uint64(v) << pw.numBits pw.numBits += nb }
[ "func", "(", "pw", "*", "Writer", ")", "WritePads", "(", "v", "uint", ")", "{", "nb", ":=", "-", "pw", ".", "numBits", "&", "7", "\n", "pw", ".", "bufBits", "|=", "uint64", "(", "v", ")", "<<", "pw", ".", "numBits", "\n", "pw", ".", "numBits", "+=", "nb", "\n", "}" ]
// WritePads writes 0-7 bits to the bit buffer to achieve byte-alignment.
[ "WritePads", "writes", "0", "-", "7", "bits", "to", "the", "bit", "buffer", "to", "achieve", "byte", "-", "alignment", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/internal/prefix/writer.go#L43-L47
9,625
dsnet/compress
internal/prefix/writer.go
Write
func (pw *Writer) Write(buf []byte) (cnt int, err error) { if pw.numBits > 0 || pw.cntBuf > 0 { if pw.numBits%8 != 0 { return 0, errorf(errors.Invalid, "non-aligned bit buffer") } if _, err := pw.Flush(); err != nil { return 0, err } } cnt, err = pw.wr.Write(buf) pw.Offset += int64(cnt) return cnt, err }
go
func (pw *Writer) Write(buf []byte) (cnt int, err error) { if pw.numBits > 0 || pw.cntBuf > 0 { if pw.numBits%8 != 0 { return 0, errorf(errors.Invalid, "non-aligned bit buffer") } if _, err := pw.Flush(); err != nil { return 0, err } } cnt, err = pw.wr.Write(buf) pw.Offset += int64(cnt) return cnt, err }
[ "func", "(", "pw", "*", "Writer", ")", "Write", "(", "buf", "[", "]", "byte", ")", "(", "cnt", "int", ",", "err", "error", ")", "{", "if", "pw", ".", "numBits", ">", "0", "||", "pw", ".", "cntBuf", ">", "0", "{", "if", "pw", ".", "numBits", "%", "8", "!=", "0", "{", "return", "0", ",", "errorf", "(", "errors", ".", "Invalid", ",", "\"", "\"", ")", "\n", "}", "\n", "if", "_", ",", "err", ":=", "pw", ".", "Flush", "(", ")", ";", "err", "!=", "nil", "{", "return", "0", ",", "err", "\n", "}", "\n", "}", "\n", "cnt", ",", "err", "=", "pw", ".", "wr", ".", "Write", "(", "buf", ")", "\n", "pw", ".", "Offset", "+=", "int64", "(", "cnt", ")", "\n", "return", "cnt", ",", "err", "\n", "}" ]
// Write writes bytes from buf. // The bit-ordering mode does not affect this method.
[ "Write", "writes", "bytes", "from", "buf", ".", "The", "bit", "-", "ordering", "mode", "does", "not", "affect", "this", "method", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/internal/prefix/writer.go#L51-L63
9,626
dsnet/compress
internal/prefix/writer.go
TryWriteBits
func (pw *Writer) TryWriteBits(v, nb uint) bool { if 64-pw.numBits < nb { return false } pw.bufBits |= uint64(v) << pw.numBits pw.numBits += nb return true }
go
func (pw *Writer) TryWriteBits(v, nb uint) bool { if 64-pw.numBits < nb { return false } pw.bufBits |= uint64(v) << pw.numBits pw.numBits += nb return true }
[ "func", "(", "pw", "*", "Writer", ")", "TryWriteBits", "(", "v", ",", "nb", "uint", ")", "bool", "{", "if", "64", "-", "pw", ".", "numBits", "<", "nb", "{", "return", "false", "\n", "}", "\n", "pw", ".", "bufBits", "|=", "uint64", "(", "v", ")", "<<", "pw", ".", "numBits", "\n", "pw", ".", "numBits", "+=", "nb", "\n", "return", "true", "\n", "}" ]
// TryWriteBits attempts to write nb bits using the contents of the bit buffer // alone. It reports whether it succeeded. // // This method is designed to be inlined for performance reasons.
[ "TryWriteBits", "attempts", "to", "write", "nb", "bits", "using", "the", "contents", "of", "the", "bit", "buffer", "alone", ".", "It", "reports", "whether", "it", "succeeded", ".", "This", "method", "is", "designed", "to", "be", "inlined", "for", "performance", "reasons", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/internal/prefix/writer.go#L78-L85
9,627
dsnet/compress
internal/prefix/writer.go
WriteBits
func (pw *Writer) WriteBits(v, nb uint) { if _, err := pw.PushBits(); err != nil { errors.Panic(err) } pw.bufBits |= uint64(v) << pw.numBits pw.numBits += nb }
go
func (pw *Writer) WriteBits(v, nb uint) { if _, err := pw.PushBits(); err != nil { errors.Panic(err) } pw.bufBits |= uint64(v) << pw.numBits pw.numBits += nb }
[ "func", "(", "pw", "*", "Writer", ")", "WriteBits", "(", "v", ",", "nb", "uint", ")", "{", "if", "_", ",", "err", ":=", "pw", ".", "PushBits", "(", ")", ";", "err", "!=", "nil", "{", "errors", ".", "Panic", "(", "err", ")", "\n", "}", "\n", "pw", ".", "bufBits", "|=", "uint64", "(", "v", ")", "<<", "pw", ".", "numBits", "\n", "pw", ".", "numBits", "+=", "nb", "\n", "}" ]
// WriteBits writes nb bits of v to the underlying writer.
[ "WriteBits", "writes", "nb", "bits", "of", "v", "to", "the", "underlying", "writer", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/internal/prefix/writer.go#L88-L94
9,628
dsnet/compress
internal/prefix/writer.go
TryWriteSymbol
func (pw *Writer) TryWriteSymbol(sym uint, pe *Encoder) bool { chunk := pe.chunks[uint32(sym)&pe.chunkMask] nb := uint(chunk & countMask) if 64-pw.numBits < nb { return false } pw.bufBits |= uint64(chunk>>countBits) << pw.numBits pw.numBits += nb return true }
go
func (pw *Writer) TryWriteSymbol(sym uint, pe *Encoder) bool { chunk := pe.chunks[uint32(sym)&pe.chunkMask] nb := uint(chunk & countMask) if 64-pw.numBits < nb { return false } pw.bufBits |= uint64(chunk>>countBits) << pw.numBits pw.numBits += nb return true }
[ "func", "(", "pw", "*", "Writer", ")", "TryWriteSymbol", "(", "sym", "uint", ",", "pe", "*", "Encoder", ")", "bool", "{", "chunk", ":=", "pe", ".", "chunks", "[", "uint32", "(", "sym", ")", "&", "pe", ".", "chunkMask", "]", "\n", "nb", ":=", "uint", "(", "chunk", "&", "countMask", ")", "\n", "if", "64", "-", "pw", ".", "numBits", "<", "nb", "{", "return", "false", "\n", "}", "\n", "pw", ".", "bufBits", "|=", "uint64", "(", "chunk", ">>", "countBits", ")", "<<", "pw", ".", "numBits", "\n", "pw", ".", "numBits", "+=", "nb", "\n", "return", "true", "\n", "}" ]
// TryWriteSymbol attempts to encode the next symbol using the contents of the // bit buffer alone. It reports whether it succeeded. // // This method is designed to be inlined for performance reasons.
[ "TryWriteSymbol", "attempts", "to", "encode", "the", "next", "symbol", "using", "the", "contents", "of", "the", "bit", "buffer", "alone", ".", "It", "reports", "whether", "it", "succeeded", ".", "This", "method", "is", "designed", "to", "be", "inlined", "for", "performance", "reasons", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/internal/prefix/writer.go#L100-L109
9,629
dsnet/compress
internal/prefix/writer.go
WriteSymbol
func (pw *Writer) WriteSymbol(sym uint, pe *Encoder) { if _, err := pw.PushBits(); err != nil { errors.Panic(err) } chunk := pe.chunks[uint32(sym)&pe.chunkMask] nb := uint(chunk & countMask) pw.bufBits |= uint64(chunk>>countBits) << pw.numBits pw.numBits += nb }
go
func (pw *Writer) WriteSymbol(sym uint, pe *Encoder) { if _, err := pw.PushBits(); err != nil { errors.Panic(err) } chunk := pe.chunks[uint32(sym)&pe.chunkMask] nb := uint(chunk & countMask) pw.bufBits |= uint64(chunk>>countBits) << pw.numBits pw.numBits += nb }
[ "func", "(", "pw", "*", "Writer", ")", "WriteSymbol", "(", "sym", "uint", ",", "pe", "*", "Encoder", ")", "{", "if", "_", ",", "err", ":=", "pw", ".", "PushBits", "(", ")", ";", "err", "!=", "nil", "{", "errors", ".", "Panic", "(", "err", ")", "\n", "}", "\n", "chunk", ":=", "pe", ".", "chunks", "[", "uint32", "(", "sym", ")", "&", "pe", ".", "chunkMask", "]", "\n", "nb", ":=", "uint", "(", "chunk", "&", "countMask", ")", "\n", "pw", ".", "bufBits", "|=", "uint64", "(", "chunk", ">>", "countBits", ")", "<<", "pw", ".", "numBits", "\n", "pw", ".", "numBits", "+=", "nb", "\n", "}" ]
// WriteSymbol writes the symbol using the provided prefix Encoder.
[ "WriteSymbol", "writes", "the", "symbol", "using", "the", "provided", "prefix", "Encoder", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/internal/prefix/writer.go#L112-L120
9,630
dsnet/compress
internal/prefix/writer.go
Flush
func (pw *Writer) Flush() (int64, error) { if pw.numBits < 8 && pw.cntBuf == 0 { return pw.Offset, nil } if _, err := pw.PushBits(); err != nil { return pw.Offset, err } cnt, err := pw.wr.Write(pw.buf[:pw.cntBuf]) pw.cntBuf -= cnt pw.Offset += int64(cnt) return pw.Offset, err }
go
func (pw *Writer) Flush() (int64, error) { if pw.numBits < 8 && pw.cntBuf == 0 { return pw.Offset, nil } if _, err := pw.PushBits(); err != nil { return pw.Offset, err } cnt, err := pw.wr.Write(pw.buf[:pw.cntBuf]) pw.cntBuf -= cnt pw.Offset += int64(cnt) return pw.Offset, err }
[ "func", "(", "pw", "*", "Writer", ")", "Flush", "(", ")", "(", "int64", ",", "error", ")", "{", "if", "pw", ".", "numBits", "<", "8", "&&", "pw", ".", "cntBuf", "==", "0", "{", "return", "pw", ".", "Offset", ",", "nil", "\n", "}", "\n", "if", "_", ",", "err", ":=", "pw", ".", "PushBits", "(", ")", ";", "err", "!=", "nil", "{", "return", "pw", ".", "Offset", ",", "err", "\n", "}", "\n", "cnt", ",", "err", ":=", "pw", ".", "wr", ".", "Write", "(", "pw", ".", "buf", "[", ":", "pw", ".", "cntBuf", "]", ")", "\n", "pw", ".", "cntBuf", "-=", "cnt", "\n", "pw", ".", "Offset", "+=", "int64", "(", "cnt", ")", "\n", "return", "pw", ".", "Offset", ",", "err", "\n", "}" ]
// Flush flushes all complete bytes from the bit buffer to the byte buffer, and // then flushes all bytes in the byte buffer to the underlying writer. // After this call, the bit Writer is will only withhold 7 bits at most.
[ "Flush", "flushes", "all", "complete", "bytes", "from", "the", "bit", "buffer", "to", "the", "byte", "buffer", "and", "then", "flushes", "all", "bytes", "in", "the", "byte", "buffer", "to", "the", "underlying", "writer", ".", "After", "this", "call", "the", "bit", "Writer", "is", "will", "only", "withhold", "7", "bits", "at", "most", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/internal/prefix/writer.go#L125-L136
9,631
dsnet/compress
internal/prefix/writer.go
PushBits
func (pw *Writer) PushBits() (uint, error) { if pw.cntBuf >= len(pw.buf)-8 { cnt, err := pw.wr.Write(pw.buf[:pw.cntBuf]) pw.cntBuf -= cnt pw.Offset += int64(cnt) if err != nil { return 0, err } } u := pw.bufBits if pw.bigEndian { // Swap all the bits within each byte. u = (u&0xaaaaaaaaaaaaaaaa)>>1 | (u&0x5555555555555555)<<1 u = (u&0xcccccccccccccccc)>>2 | (u&0x3333333333333333)<<2 u = (u&0xf0f0f0f0f0f0f0f0)>>4 | (u&0x0f0f0f0f0f0f0f0f)<<4 } // Starting with Go 1.7, the compiler should use a wide integer // store here if the architecture supports it. binary.LittleEndian.PutUint64(pw.buf[pw.cntBuf:], u) nb := pw.numBits / 8 // Number of bytes to copy from bit buffer pw.cntBuf += int(nb) pw.bufBits >>= 8 * nb pw.numBits -= 8 * nb return 8 * nb, nil }
go
func (pw *Writer) PushBits() (uint, error) { if pw.cntBuf >= len(pw.buf)-8 { cnt, err := pw.wr.Write(pw.buf[:pw.cntBuf]) pw.cntBuf -= cnt pw.Offset += int64(cnt) if err != nil { return 0, err } } u := pw.bufBits if pw.bigEndian { // Swap all the bits within each byte. u = (u&0xaaaaaaaaaaaaaaaa)>>1 | (u&0x5555555555555555)<<1 u = (u&0xcccccccccccccccc)>>2 | (u&0x3333333333333333)<<2 u = (u&0xf0f0f0f0f0f0f0f0)>>4 | (u&0x0f0f0f0f0f0f0f0f)<<4 } // Starting with Go 1.7, the compiler should use a wide integer // store here if the architecture supports it. binary.LittleEndian.PutUint64(pw.buf[pw.cntBuf:], u) nb := pw.numBits / 8 // Number of bytes to copy from bit buffer pw.cntBuf += int(nb) pw.bufBits >>= 8 * nb pw.numBits -= 8 * nb return 8 * nb, nil }
[ "func", "(", "pw", "*", "Writer", ")", "PushBits", "(", ")", "(", "uint", ",", "error", ")", "{", "if", "pw", ".", "cntBuf", ">=", "len", "(", "pw", ".", "buf", ")", "-", "8", "{", "cnt", ",", "err", ":=", "pw", ".", "wr", ".", "Write", "(", "pw", ".", "buf", "[", ":", "pw", ".", "cntBuf", "]", ")", "\n", "pw", ".", "cntBuf", "-=", "cnt", "\n", "pw", ".", "Offset", "+=", "int64", "(", "cnt", ")", "\n", "if", "err", "!=", "nil", "{", "return", "0", ",", "err", "\n", "}", "\n", "}", "\n\n", "u", ":=", "pw", ".", "bufBits", "\n", "if", "pw", ".", "bigEndian", "{", "// Swap all the bits within each byte.", "u", "=", "(", "u", "&", "0xaaaaaaaaaaaaaaaa", ")", ">>", "1", "|", "(", "u", "&", "0x5555555555555555", ")", "<<", "1", "\n", "u", "=", "(", "u", "&", "0xcccccccccccccccc", ")", ">>", "2", "|", "(", "u", "&", "0x3333333333333333", ")", "<<", "2", "\n", "u", "=", "(", "u", "&", "0xf0f0f0f0f0f0f0f0", ")", ">>", "4", "|", "(", "u", "&", "0x0f0f0f0f0f0f0f0f", ")", "<<", "4", "\n", "}", "\n", "// Starting with Go 1.7, the compiler should use a wide integer", "// store here if the architecture supports it.", "binary", ".", "LittleEndian", ".", "PutUint64", "(", "pw", ".", "buf", "[", "pw", ".", "cntBuf", ":", "]", ",", "u", ")", "\n\n", "nb", ":=", "pw", ".", "numBits", "/", "8", "// Number of bytes to copy from bit buffer", "\n", "pw", ".", "cntBuf", "+=", "int", "(", "nb", ")", "\n", "pw", ".", "bufBits", ">>=", "8", "*", "nb", "\n", "pw", ".", "numBits", "-=", "8", "*", "nb", "\n", "return", "8", "*", "nb", ",", "nil", "\n", "}" ]
// PushBits pushes as many bytes as possible from the bit buffer to the byte // buffer, reporting the number of bits pushed.
[ "PushBits", "pushes", "as", "many", "bytes", "as", "possible", "from", "the", "bit", "buffer", "to", "the", "byte", "buffer", "reporting", "the", "number", "of", "bits", "pushed", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/internal/prefix/writer.go#L140-L166
9,632
dsnet/compress
internal/tool/bench/main.go
getFiles
func getFiles(paths []string, globs []string) []file { var fs []file for _, p := range paths { for _, g := range globs { ms, _ := filepath.Glob(filepath.Join(p, g)) for _, m := range ms { r, err1 := filepath.Rel(p, m) fi, err2 := os.Stat(m) if err1 == nil && err2 == nil && !fi.IsDir() { fs = append(fs, file{Abs: m, Rel: r}) } } } } return fs }
go
func getFiles(paths []string, globs []string) []file { var fs []file for _, p := range paths { for _, g := range globs { ms, _ := filepath.Glob(filepath.Join(p, g)) for _, m := range ms { r, err1 := filepath.Rel(p, m) fi, err2 := os.Stat(m) if err1 == nil && err2 == nil && !fi.IsDir() { fs = append(fs, file{Abs: m, Rel: r}) } } } } return fs }
[ "func", "getFiles", "(", "paths", "[", "]", "string", ",", "globs", "[", "]", "string", ")", "[", "]", "file", "{", "var", "fs", "[", "]", "file", "\n", "for", "_", ",", "p", ":=", "range", "paths", "{", "for", "_", ",", "g", ":=", "range", "globs", "{", "ms", ",", "_", ":=", "filepath", ".", "Glob", "(", "filepath", ".", "Join", "(", "p", ",", "g", ")", ")", "\n", "for", "_", ",", "m", ":=", "range", "ms", "{", "r", ",", "err1", ":=", "filepath", ".", "Rel", "(", "p", ",", "m", ")", "\n", "fi", ",", "err2", ":=", "os", ".", "Stat", "(", "m", ")", "\n", "if", "err1", "==", "nil", "&&", "err2", "==", "nil", "&&", "!", "fi", ".", "IsDir", "(", ")", "{", "fs", "=", "append", "(", "fs", ",", "file", "{", "Abs", ":", "m", ",", "Rel", ":", "r", "}", ")", "\n", "}", "\n", "}", "\n", "}", "\n", "}", "\n", "return", "fs", "\n", "}" ]
// getFiles returns a list of files found by applying the glob matching on // all the specified paths. This function ignores any errors.
[ "getFiles", "returns", "a", "list", "of", "files", "found", "by", "applying", "the", "glob", "matching", "on", "all", "the", "specified", "paths", ".", "This", "function", "ignores", "any", "errors", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/internal/tool/bench/main.go#L93-L108
9,633
dsnet/compress
brotli/common.go
errWrap
func errWrap(err error, replaceCode int) error { if cerr, ok := err.(errors.Error); ok { if errors.IsInvalid(cerr) { cerr.Code = replaceCode } err = errorf(cerr.Code, "%s", cerr.Msg) } return err }
go
func errWrap(err error, replaceCode int) error { if cerr, ok := err.(errors.Error); ok { if errors.IsInvalid(cerr) { cerr.Code = replaceCode } err = errorf(cerr.Code, "%s", cerr.Msg) } return err }
[ "func", "errWrap", "(", "err", "error", ",", "replaceCode", "int", ")", "error", "{", "if", "cerr", ",", "ok", ":=", "err", ".", "(", "errors", ".", "Error", ")", ";", "ok", "{", "if", "errors", ".", "IsInvalid", "(", "cerr", ")", "{", "cerr", ".", "Code", "=", "replaceCode", "\n", "}", "\n", "err", "=", "errorf", "(", "cerr", ".", "Code", ",", "\"", "\"", ",", "cerr", ".", "Msg", ")", "\n", "}", "\n", "return", "err", "\n", "}" ]
// errWrap converts a lower-level errors.Error to be one from this package. // The replaceCode passed in will be used to replace the code for any errors // with the errors.Invalid code. // // For the Reader, set this to errors.Corrupted. // For the Writer, set this to errors.Internal.
[ "errWrap", "converts", "a", "lower", "-", "level", "errors", ".", "Error", "to", "be", "one", "from", "this", "package", ".", "The", "replaceCode", "passed", "in", "will", "be", "used", "to", "replace", "the", "code", "for", "any", "errors", "with", "the", "errors", ".", "Invalid", "code", ".", "For", "the", "Reader", "set", "this", "to", "errors", ".", "Corrupted", ".", "For", "the", "Writer", "set", "this", "to", "errors", ".", "Internal", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/brotli/common.go#L25-L33
9,634
dsnet/compress
brotli/common.go
reverseUint32
func reverseUint32(v uint32) (x uint32) { x |= uint32(reverseLUT[byte(v>>0)]) << 24 x |= uint32(reverseLUT[byte(v>>8)]) << 16 x |= uint32(reverseLUT[byte(v>>16)]) << 8 x |= uint32(reverseLUT[byte(v>>24)]) << 0 return x }
go
func reverseUint32(v uint32) (x uint32) { x |= uint32(reverseLUT[byte(v>>0)]) << 24 x |= uint32(reverseLUT[byte(v>>8)]) << 16 x |= uint32(reverseLUT[byte(v>>16)]) << 8 x |= uint32(reverseLUT[byte(v>>24)]) << 0 return x }
[ "func", "reverseUint32", "(", "v", "uint32", ")", "(", "x", "uint32", ")", "{", "x", "|=", "uint32", "(", "reverseLUT", "[", "byte", "(", "v", ">>", "0", ")", "]", ")", "<<", "24", "\n", "x", "|=", "uint32", "(", "reverseLUT", "[", "byte", "(", "v", ">>", "8", ")", "]", ")", "<<", "16", "\n", "x", "|=", "uint32", "(", "reverseLUT", "[", "byte", "(", "v", ">>", "16", ")", "]", ")", "<<", "8", "\n", "x", "|=", "uint32", "(", "reverseLUT", "[", "byte", "(", "v", ">>", "24", ")", "]", ")", "<<", "0", "\n", "return", "x", "\n", "}" ]
// reverseUint32 reverses all bits of v.
[ "reverseUint32", "reverses", "all", "bits", "of", "v", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/brotli/common.go#L76-L82
9,635
dsnet/compress
xflate/internal/meta/reader.go
NewReader
func NewReader(rd io.Reader) *Reader { mr := new(Reader) mr.Reset(rd) return mr }
go
func NewReader(rd io.Reader) *Reader { mr := new(Reader) mr.Reset(rd) return mr }
[ "func", "NewReader", "(", "rd", "io", ".", "Reader", ")", "*", "Reader", "{", "mr", ":=", "new", "(", "Reader", ")", "\n", "mr", ".", "Reset", "(", "rd", ")", "\n", "return", "mr", "\n", "}" ]
// NewReader creates a new Reader reading from the given reader. // If rd does not also implement compress.ByteReader or compress.BufferedReader, // then the decoder may read more data than necessary from rd.
[ "NewReader", "creates", "a", "new", "Reader", "reading", "from", "the", "given", "reader", ".", "If", "rd", "does", "not", "also", "implement", "compress", ".", "ByteReader", "or", "compress", ".", "BufferedReader", "then", "the", "decoder", "may", "read", "more", "data", "than", "necessary", "from", "rd", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/xflate/internal/meta/reader.go#L39-L43
9,636
dsnet/compress
xflate/internal/meta/reader.go
Reset
func (mr *Reader) Reset(rd io.Reader) { *mr = Reader{ br: mr.br, bw: mr.bw, bb: mr.bb, } if br, ok := rd.(*prefix.Reader); ok { // Use input Reader directly as a prefix.Reader. mr.rd = br } else { // Use pre-allocated prefix.Reader to wrap input Reader. mr.rd = &mr.br mr.rd.Init(rd, false) } return }
go
func (mr *Reader) Reset(rd io.Reader) { *mr = Reader{ br: mr.br, bw: mr.bw, bb: mr.bb, } if br, ok := rd.(*prefix.Reader); ok { // Use input Reader directly as a prefix.Reader. mr.rd = br } else { // Use pre-allocated prefix.Reader to wrap input Reader. mr.rd = &mr.br mr.rd.Init(rd, false) } return }
[ "func", "(", "mr", "*", "Reader", ")", "Reset", "(", "rd", "io", ".", "Reader", ")", "{", "*", "mr", "=", "Reader", "{", "br", ":", "mr", ".", "br", ",", "bw", ":", "mr", ".", "bw", ",", "bb", ":", "mr", ".", "bb", ",", "}", "\n", "if", "br", ",", "ok", ":=", "rd", ".", "(", "*", "prefix", ".", "Reader", ")", ";", "ok", "{", "// Use input Reader directly as a prefix.Reader.", "mr", ".", "rd", "=", "br", "\n", "}", "else", "{", "// Use pre-allocated prefix.Reader to wrap input Reader.", "mr", ".", "rd", "=", "&", "mr", ".", "br", "\n", "mr", ".", "rd", ".", "Init", "(", "rd", ",", "false", ")", "\n", "}", "\n", "return", "\n", "}" ]
// Reset discards the Reader's state and makes it equivalent to the result // of a call to NewReader, but reading from rd instead. // // This is used to reduce memory allocations.
[ "Reset", "discards", "the", "Reader", "s", "state", "and", "makes", "it", "equivalent", "to", "the", "result", "of", "a", "call", "to", "NewReader", "but", "reading", "from", "rd", "instead", ".", "This", "is", "used", "to", "reduce", "memory", "allocations", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/xflate/internal/meta/reader.go#L49-L64
9,637
dsnet/compress
xflate/internal/meta/reader.go
Read
func (mr *Reader) Read(buf []byte) (int, error) { if mr.err != nil { return 0, mr.err } var rdCnt int for len(buf) > 0 { if len(mr.buf) > 0 { cpCnt := copy(buf, mr.buf) mr.buf = mr.buf[cpCnt:] rdCnt += cpCnt break } if mr.final != FinalNil { mr.FinalMode = mr.final mr.err = io.EOF break } mr.err = mr.decodeBlock() if mr.err != nil { break } } mr.OutputOffset += int64(rdCnt) return rdCnt, mr.err }
go
func (mr *Reader) Read(buf []byte) (int, error) { if mr.err != nil { return 0, mr.err } var rdCnt int for len(buf) > 0 { if len(mr.buf) > 0 { cpCnt := copy(buf, mr.buf) mr.buf = mr.buf[cpCnt:] rdCnt += cpCnt break } if mr.final != FinalNil { mr.FinalMode = mr.final mr.err = io.EOF break } mr.err = mr.decodeBlock() if mr.err != nil { break } } mr.OutputOffset += int64(rdCnt) return rdCnt, mr.err }
[ "func", "(", "mr", "*", "Reader", ")", "Read", "(", "buf", "[", "]", "byte", ")", "(", "int", ",", "error", ")", "{", "if", "mr", ".", "err", "!=", "nil", "{", "return", "0", ",", "mr", ".", "err", "\n", "}", "\n\n", "var", "rdCnt", "int", "\n", "for", "len", "(", "buf", ")", ">", "0", "{", "if", "len", "(", "mr", ".", "buf", ")", ">", "0", "{", "cpCnt", ":=", "copy", "(", "buf", ",", "mr", ".", "buf", ")", "\n", "mr", ".", "buf", "=", "mr", ".", "buf", "[", "cpCnt", ":", "]", "\n", "rdCnt", "+=", "cpCnt", "\n", "break", "\n", "}", "\n\n", "if", "mr", ".", "final", "!=", "FinalNil", "{", "mr", ".", "FinalMode", "=", "mr", ".", "final", "\n", "mr", ".", "err", "=", "io", ".", "EOF", "\n", "break", "\n", "}", "\n\n", "mr", ".", "err", "=", "mr", ".", "decodeBlock", "(", ")", "\n", "if", "mr", ".", "err", "!=", "nil", "{", "break", "\n", "}", "\n", "}", "\n\n", "mr", ".", "OutputOffset", "+=", "int64", "(", "rdCnt", ")", "\n", "return", "rdCnt", ",", "mr", ".", "err", "\n", "}" ]
// Read reads the decoded meta data from the underlying io.Reader. // This returns io.EOF either when a meta block with final bits set is found or // when io.EOF is hit in the underlying reader.
[ "Read", "reads", "the", "decoded", "meta", "data", "from", "the", "underlying", "io", ".", "Reader", ".", "This", "returns", "io", ".", "EOF", "either", "when", "a", "meta", "block", "with", "final", "bits", "set", "is", "found", "or", "when", "io", ".", "EOF", "is", "hit", "in", "the", "underlying", "reader", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/xflate/internal/meta/reader.go#L69-L97
9,638
dsnet/compress
xflate/internal/meta/reader.go
Close
func (mr *Reader) Close() error { if mr.err == errClosed { return nil } if mr.err != nil && mr.err != io.EOF { return mr.err } mr.FinalMode = mr.final mr.err = errClosed mr.rd = nil // Release reference to underlying Reader return nil }
go
func (mr *Reader) Close() error { if mr.err == errClosed { return nil } if mr.err != nil && mr.err != io.EOF { return mr.err } mr.FinalMode = mr.final mr.err = errClosed mr.rd = nil // Release reference to underlying Reader return nil }
[ "func", "(", "mr", "*", "Reader", ")", "Close", "(", ")", "error", "{", "if", "mr", ".", "err", "==", "errClosed", "{", "return", "nil", "\n", "}", "\n", "if", "mr", ".", "err", "!=", "nil", "&&", "mr", ".", "err", "!=", "io", ".", "EOF", "{", "return", "mr", ".", "err", "\n", "}", "\n\n", "mr", ".", "FinalMode", "=", "mr", ".", "final", "\n", "mr", ".", "err", "=", "errClosed", "\n", "mr", ".", "rd", "=", "nil", "// Release reference to underlying Reader", "\n", "return", "nil", "\n", "}" ]
// Close ends the meta stream. // The FinalMode encountered becomes valid after calling Close.
[ "Close", "ends", "the", "meta", "stream", ".", "The", "FinalMode", "encountered", "becomes", "valid", "after", "calling", "Close", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/xflate/internal/meta/reader.go#L101-L113
9,639
dsnet/compress
brotli/context.go
initContextLUTs
func initContextLUTs() { for i := 0; i < 256; i++ { for m := 0; m < numContextModes; m++ { base := m << 8 // Operations performed here are specified in RFC section 7.1. switch m { case contextLSB6: contextP1LUT[base+i] = byte(i) & 0x3f contextP2LUT[base+i] = 0 case contextMSB6: contextP1LUT[base+i] = byte(i) >> 2 contextP2LUT[base+i] = 0 case contextUTF8: contextP1LUT[base+i] = contextLUT0[byte(i)] contextP2LUT[base+i] = contextLUT1[byte(i)] case contextSigned: contextP1LUT[base+i] = contextLUT2[byte(i)] << 3 contextP2LUT[base+i] = contextLUT2[byte(i)] default: panic("unknown context mode") } } } }
go
func initContextLUTs() { for i := 0; i < 256; i++ { for m := 0; m < numContextModes; m++ { base := m << 8 // Operations performed here are specified in RFC section 7.1. switch m { case contextLSB6: contextP1LUT[base+i] = byte(i) & 0x3f contextP2LUT[base+i] = 0 case contextMSB6: contextP1LUT[base+i] = byte(i) >> 2 contextP2LUT[base+i] = 0 case contextUTF8: contextP1LUT[base+i] = contextLUT0[byte(i)] contextP2LUT[base+i] = contextLUT1[byte(i)] case contextSigned: contextP1LUT[base+i] = contextLUT2[byte(i)] << 3 contextP2LUT[base+i] = contextLUT2[byte(i)] default: panic("unknown context mode") } } } }
[ "func", "initContextLUTs", "(", ")", "{", "for", "i", ":=", "0", ";", "i", "<", "256", ";", "i", "++", "{", "for", "m", ":=", "0", ";", "m", "<", "numContextModes", ";", "m", "++", "{", "base", ":=", "m", "<<", "8", "\n\n", "// Operations performed here are specified in RFC section 7.1.", "switch", "m", "{", "case", "contextLSB6", ":", "contextP1LUT", "[", "base", "+", "i", "]", "=", "byte", "(", "i", ")", "&", "0x3f", "\n", "contextP2LUT", "[", "base", "+", "i", "]", "=", "0", "\n", "case", "contextMSB6", ":", "contextP1LUT", "[", "base", "+", "i", "]", "=", "byte", "(", "i", ")", ">>", "2", "\n", "contextP2LUT", "[", "base", "+", "i", "]", "=", "0", "\n", "case", "contextUTF8", ":", "contextP1LUT", "[", "base", "+", "i", "]", "=", "contextLUT0", "[", "byte", "(", "i", ")", "]", "\n", "contextP2LUT", "[", "base", "+", "i", "]", "=", "contextLUT1", "[", "byte", "(", "i", ")", "]", "\n", "case", "contextSigned", ":", "contextP1LUT", "[", "base", "+", "i", "]", "=", "contextLUT2", "[", "byte", "(", "i", ")", "]", "<<", "3", "\n", "contextP2LUT", "[", "base", "+", "i", "]", "=", "contextLUT2", "[", "byte", "(", "i", ")", "]", "\n", "default", ":", "panic", "(", "\"", "\"", ")", "\n", "}", "\n", "}", "\n", "}", "\n", "}" ]
// initContextLUTs computes LUTs so that context ID computation can be // efficiently without any branches.
[ "initContextLUTs", "computes", "LUTs", "so", "that", "context", "ID", "computation", "can", "be", "efficiently", "without", "any", "branches", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/brotli/context.go#L91-L115
9,640
dsnet/compress
brotli/context.go
getLitContextID
func getLitContextID(p1, p2 byte, mode uint8) uint8 { base := uint(mode) << 8 return contextP1LUT[base+uint(p1)] | contextP2LUT[base+uint(p2)] }
go
func getLitContextID(p1, p2 byte, mode uint8) uint8 { base := uint(mode) << 8 return contextP1LUT[base+uint(p1)] | contextP2LUT[base+uint(p2)] }
[ "func", "getLitContextID", "(", "p1", ",", "p2", "byte", ",", "mode", "uint8", ")", "uint8", "{", "base", ":=", "uint", "(", "mode", ")", "<<", "8", "\n", "return", "contextP1LUT", "[", "base", "+", "uint", "(", "p1", ")", "]", "|", "contextP2LUT", "[", "base", "+", "uint", "(", "p2", ")", "]", "\n", "}" ]
// getLitContextID computes the context ID for literals from RFC section 7.1. // Bytes p1 and p2 are the last and second-to-last byte, respectively.
[ "getLitContextID", "computes", "the", "context", "ID", "for", "literals", "from", "RFC", "section", "7", ".", "1", ".", "Bytes", "p1", "and", "p2", "are", "the", "last", "and", "second", "-", "to", "-", "last", "byte", "respectively", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/brotli/context.go#L119-L122
9,641
dsnet/compress
brotli/dict_decoder.go
HistSize
func (dd *dictDecoder) HistSize() int { if dd.full { return dd.size } return dd.wrPos }
go
func (dd *dictDecoder) HistSize() int { if dd.full { return dd.size } return dd.wrPos }
[ "func", "(", "dd", "*", "dictDecoder", ")", "HistSize", "(", ")", "int", "{", "if", "dd", ".", "full", "{", "return", "dd", ".", "size", "\n", "}", "\n", "return", "dd", ".", "wrPos", "\n", "}" ]
// HistSize reports the total amount of historical data in the dictionary.
[ "HistSize", "reports", "the", "total", "amount", "of", "historical", "data", "in", "the", "dictionary", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/brotli/dict_decoder.go#L49-L54
9,642
dsnet/compress
brotli/dict_decoder.go
ReadFlush
func (dd *dictDecoder) ReadFlush() []byte { toRead := dd.hist[dd.rdPos:dd.wrPos] dd.rdPos = dd.wrPos if dd.wrPos == len(dd.hist) { if len(dd.hist) == dd.size { dd.wrPos, dd.rdPos = 0, 0 dd.full = true } else { // Allocate a larger history buffer. size := cap(dd.hist) * growFactor if size > dd.size { size = dd.size } hist := make([]byte, size) copy(hist, dd.hist) dd.hist = hist } } return toRead }
go
func (dd *dictDecoder) ReadFlush() []byte { toRead := dd.hist[dd.rdPos:dd.wrPos] dd.rdPos = dd.wrPos if dd.wrPos == len(dd.hist) { if len(dd.hist) == dd.size { dd.wrPos, dd.rdPos = 0, 0 dd.full = true } else { // Allocate a larger history buffer. size := cap(dd.hist) * growFactor if size > dd.size { size = dd.size } hist := make([]byte, size) copy(hist, dd.hist) dd.hist = hist } } return toRead }
[ "func", "(", "dd", "*", "dictDecoder", ")", "ReadFlush", "(", ")", "[", "]", "byte", "{", "toRead", ":=", "dd", ".", "hist", "[", "dd", ".", "rdPos", ":", "dd", ".", "wrPos", "]", "\n", "dd", ".", "rdPos", "=", "dd", ".", "wrPos", "\n", "if", "dd", ".", "wrPos", "==", "len", "(", "dd", ".", "hist", ")", "{", "if", "len", "(", "dd", ".", "hist", ")", "==", "dd", ".", "size", "{", "dd", ".", "wrPos", ",", "dd", ".", "rdPos", "=", "0", ",", "0", "\n", "dd", ".", "full", "=", "true", "\n", "}", "else", "{", "// Allocate a larger history buffer.", "size", ":=", "cap", "(", "dd", ".", "hist", ")", "*", "growFactor", "\n", "if", "size", ">", "dd", ".", "size", "{", "size", "=", "dd", ".", "size", "\n", "}", "\n", "hist", ":=", "make", "(", "[", "]", "byte", ",", "size", ")", "\n", "copy", "(", "hist", ",", "dd", ".", "hist", ")", "\n", "dd", ".", "hist", "=", "hist", "\n", "}", "\n", "}", "\n", "return", "toRead", "\n", "}" ]
// ReadFlush returns a slice of the historical buffer that is ready to be // emitted to the user. A call to ReadFlush is only valid after all of the data // from a previous call to ReadFlush has been consumed.
[ "ReadFlush", "returns", "a", "slice", "of", "the", "historical", "buffer", "that", "is", "ready", "to", "be", "emitted", "to", "the", "user", ".", "A", "call", "to", "ReadFlush", "is", "only", "valid", "after", "all", "of", "the", "data", "from", "a", "previous", "call", "to", "ReadFlush", "has", "been", "consumed", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/brotli/dict_decoder.go#L105-L124
9,643
dsnet/compress
brotli/dict_decoder.go
LastBytes
func (dd *dictDecoder) LastBytes() (p1, p2 byte) { if dd.wrPos > 1 { return dd.hist[dd.wrPos-1], dd.hist[dd.wrPos-2] } else if dd.wrPos > 0 { return dd.hist[dd.wrPos-1], dd.hist[len(dd.hist)-1] } else { return dd.hist[len(dd.hist)-1], dd.hist[len(dd.hist)-2] } }
go
func (dd *dictDecoder) LastBytes() (p1, p2 byte) { if dd.wrPos > 1 { return dd.hist[dd.wrPos-1], dd.hist[dd.wrPos-2] } else if dd.wrPos > 0 { return dd.hist[dd.wrPos-1], dd.hist[len(dd.hist)-1] } else { return dd.hist[len(dd.hist)-1], dd.hist[len(dd.hist)-2] } }
[ "func", "(", "dd", "*", "dictDecoder", ")", "LastBytes", "(", ")", "(", "p1", ",", "p2", "byte", ")", "{", "if", "dd", ".", "wrPos", ">", "1", "{", "return", "dd", ".", "hist", "[", "dd", ".", "wrPos", "-", "1", "]", ",", "dd", ".", "hist", "[", "dd", ".", "wrPos", "-", "2", "]", "\n", "}", "else", "if", "dd", ".", "wrPos", ">", "0", "{", "return", "dd", ".", "hist", "[", "dd", ".", "wrPos", "-", "1", "]", ",", "dd", ".", "hist", "[", "len", "(", "dd", ".", "hist", ")", "-", "1", "]", "\n", "}", "else", "{", "return", "dd", ".", "hist", "[", "len", "(", "dd", ".", "hist", ")", "-", "1", "]", ",", "dd", ".", "hist", "[", "len", "(", "dd", ".", "hist", ")", "-", "2", "]", "\n", "}", "\n", "}" ]
// LastBytes reports the last 2 bytes in the dictionary. If they do not exist, // then zero values are returned.
[ "LastBytes", "reports", "the", "last", "2", "bytes", "in", "the", "dictionary", ".", "If", "they", "do", "not", "exist", "then", "zero", "values", "are", "returned", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/brotli/dict_decoder.go#L128-L136
9,644
dsnet/compress
internal/tool/bench/config.go
setDefaults
func setDefaults() { formats = defaultFormats() tests = defaultTests() codecs = defaultCodecs() paths = defaultPaths() globs = []string{"*.txt", "*.bin"} levels = []int{1, 6, 9} sizes = []int{1e4, 1e5, 1e6} }
go
func setDefaults() { formats = defaultFormats() tests = defaultTests() codecs = defaultCodecs() paths = defaultPaths() globs = []string{"*.txt", "*.bin"} levels = []int{1, 6, 9} sizes = []int{1e4, 1e5, 1e6} }
[ "func", "setDefaults", "(", ")", "{", "formats", "=", "defaultFormats", "(", ")", "\n", "tests", "=", "defaultTests", "(", ")", "\n", "codecs", "=", "defaultCodecs", "(", ")", "\n", "paths", "=", "defaultPaths", "(", ")", "\n", "globs", "=", "[", "]", "string", "{", "\"", "\"", ",", "\"", "\"", "}", "\n", "levels", "=", "[", "]", "int", "{", "1", ",", "6", ",", "9", "}", "\n", "sizes", "=", "[", "]", "int", "{", "1e4", ",", "1e5", ",", "1e6", "}", "\n", "}" ]
// setDefaults configures the top-level parameters with default values. // This function method must be called after all init functions have executed // since they register the various codecs.
[ "setDefaults", "configures", "the", "top", "-", "level", "parameters", "with", "default", "values", ".", "This", "function", "method", "must", "be", "called", "after", "all", "init", "functions", "have", "executed", "since", "they", "register", "the", "various", "codecs", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/internal/tool/bench/config.go#L120-L128
9,645
dsnet/compress
internal/tool/bench/config.go
intName
func intName(n int64) string { switch n { case 1e3, 1e4, 1e5, 1e6, 1e7, 1e8, 1e9, 1e10, 1e11, 1e12: s := fmt.Sprintf("%e", float64(n)) re := regexp.MustCompile("\\.0*e\\+0*") return re.ReplaceAllString(s, "e") default: s := unitconv.FormatPrefix(float64(n), unitconv.Base1024, 2) return strings.Replace(s, ".00", "", -1) } }
go
func intName(n int64) string { switch n { case 1e3, 1e4, 1e5, 1e6, 1e7, 1e8, 1e9, 1e10, 1e11, 1e12: s := fmt.Sprintf("%e", float64(n)) re := regexp.MustCompile("\\.0*e\\+0*") return re.ReplaceAllString(s, "e") default: s := unitconv.FormatPrefix(float64(n), unitconv.Base1024, 2) return strings.Replace(s, ".00", "", -1) } }
[ "func", "intName", "(", "n", "int64", ")", "string", "{", "switch", "n", "{", "case", "1e3", ",", "1e4", ",", "1e5", ",", "1e6", ",", "1e7", ",", "1e8", ",", "1e9", ",", "1e10", ",", "1e11", ",", "1e12", ":", "s", ":=", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "float64", "(", "n", ")", ")", "\n", "re", ":=", "regexp", ".", "MustCompile", "(", "\"", "\\\\", "\\\\", "\"", ")", "\n", "return", "re", ".", "ReplaceAllString", "(", "s", ",", "\"", "\"", ")", "\n", "default", ":", "s", ":=", "unitconv", ".", "FormatPrefix", "(", "float64", "(", "n", ")", ",", "unitconv", ".", "Base1024", ",", "2", ")", "\n", "return", "strings", ".", "Replace", "(", "s", ",", "\"", "\"", ",", "\"", "\"", ",", "-", "1", ")", "\n", "}", "\n", "}" ]
// intName returns a shorter representation of the input integer. // It uses scientific notation for exact powers of 10. // It uses SI suffixes for powers of 1024. // If the number is small enough, it will be printed as is.
[ "intName", "returns", "a", "shorter", "representation", "of", "the", "input", "integer", ".", "It", "uses", "scientific", "notation", "for", "exact", "powers", "of", "10", ".", "It", "uses", "SI", "suffixes", "for", "powers", "of", "1024", ".", "If", "the", "number", "is", "small", "enough", "it", "will", "be", "printed", "as", "is", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/internal/tool/bench/config.go#L275-L285
9,646
dsnet/compress
internal/prefix/reader.go
BitsRead
func (pr *Reader) BitsRead() int64 { offset := 8*pr.Offset - int64(pr.numBits) if pr.bufRd != nil { discardBits := pr.discardBits + int(pr.fedBits-pr.numBits) offset = 8*pr.Offset + int64(discardBits) } return offset }
go
func (pr *Reader) BitsRead() int64 { offset := 8*pr.Offset - int64(pr.numBits) if pr.bufRd != nil { discardBits := pr.discardBits + int(pr.fedBits-pr.numBits) offset = 8*pr.Offset + int64(discardBits) } return offset }
[ "func", "(", "pr", "*", "Reader", ")", "BitsRead", "(", ")", "int64", "{", "offset", ":=", "8", "*", "pr", ".", "Offset", "-", "int64", "(", "pr", ".", "numBits", ")", "\n", "if", "pr", ".", "bufRd", "!=", "nil", "{", "discardBits", ":=", "pr", ".", "discardBits", "+", "int", "(", "pr", ".", "fedBits", "-", "pr", ".", "numBits", ")", "\n", "offset", "=", "8", "*", "pr", ".", "Offset", "+", "int64", "(", "discardBits", ")", "\n", "}", "\n", "return", "offset", "\n", "}" ]
// BitsRead reports the total number of bits emitted from any Read method.
[ "BitsRead", "reports", "the", "total", "number", "of", "bits", "emitted", "from", "any", "Read", "method", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/internal/prefix/reader.go#L97-L104
9,647
dsnet/compress
internal/prefix/reader.go
Read
func (pr *Reader) Read(buf []byte) (cnt int, err error) { if pr.numBits > 0 { if pr.numBits%8 != 0 { return 0, errorf(errors.Invalid, "non-aligned bit buffer") } for cnt = 0; len(buf) > cnt && pr.numBits > 0; cnt++ { if pr.bigEndian { buf[cnt] = internal.ReverseLUT[byte(pr.bufBits)] } else { buf[cnt] = byte(pr.bufBits) } pr.bufBits >>= 8 pr.numBits -= 8 } return cnt, nil } if _, err := pr.Flush(); err != nil { return 0, err } cnt, err = pr.rd.Read(buf) pr.Offset += int64(cnt) return cnt, err }
go
func (pr *Reader) Read(buf []byte) (cnt int, err error) { if pr.numBits > 0 { if pr.numBits%8 != 0 { return 0, errorf(errors.Invalid, "non-aligned bit buffer") } for cnt = 0; len(buf) > cnt && pr.numBits > 0; cnt++ { if pr.bigEndian { buf[cnt] = internal.ReverseLUT[byte(pr.bufBits)] } else { buf[cnt] = byte(pr.bufBits) } pr.bufBits >>= 8 pr.numBits -= 8 } return cnt, nil } if _, err := pr.Flush(); err != nil { return 0, err } cnt, err = pr.rd.Read(buf) pr.Offset += int64(cnt) return cnt, err }
[ "func", "(", "pr", "*", "Reader", ")", "Read", "(", "buf", "[", "]", "byte", ")", "(", "cnt", "int", ",", "err", "error", ")", "{", "if", "pr", ".", "numBits", ">", "0", "{", "if", "pr", ".", "numBits", "%", "8", "!=", "0", "{", "return", "0", ",", "errorf", "(", "errors", ".", "Invalid", ",", "\"", "\"", ")", "\n", "}", "\n", "for", "cnt", "=", "0", ";", "len", "(", "buf", ")", ">", "cnt", "&&", "pr", ".", "numBits", ">", "0", ";", "cnt", "++", "{", "if", "pr", ".", "bigEndian", "{", "buf", "[", "cnt", "]", "=", "internal", ".", "ReverseLUT", "[", "byte", "(", "pr", ".", "bufBits", ")", "]", "\n", "}", "else", "{", "buf", "[", "cnt", "]", "=", "byte", "(", "pr", ".", "bufBits", ")", "\n", "}", "\n", "pr", ".", "bufBits", ">>=", "8", "\n", "pr", ".", "numBits", "-=", "8", "\n", "}", "\n", "return", "cnt", ",", "nil", "\n", "}", "\n", "if", "_", ",", "err", ":=", "pr", ".", "Flush", "(", ")", ";", "err", "!=", "nil", "{", "return", "0", ",", "err", "\n", "}", "\n", "cnt", ",", "err", "=", "pr", ".", "rd", ".", "Read", "(", "buf", ")", "\n", "pr", ".", "Offset", "+=", "int64", "(", "cnt", ")", "\n", "return", "cnt", ",", "err", "\n", "}" ]
// Read reads bytes into buf. // The bit-ordering mode does not affect this method.
[ "Read", "reads", "bytes", "into", "buf", ".", "The", "bit", "-", "ordering", "mode", "does", "not", "affect", "this", "method", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/internal/prefix/reader.go#L123-L145
9,648
dsnet/compress
internal/prefix/reader.go
ReadOffset
func (pr *Reader) ReadOffset(pd *Decoder, rcs RangeCodes) uint { rc := rcs[pr.ReadSymbol(pd)] return uint(rc.Base) + pr.ReadBits(uint(rc.Len)) }
go
func (pr *Reader) ReadOffset(pd *Decoder, rcs RangeCodes) uint { rc := rcs[pr.ReadSymbol(pd)] return uint(rc.Base) + pr.ReadBits(uint(rc.Len)) }
[ "func", "(", "pr", "*", "Reader", ")", "ReadOffset", "(", "pd", "*", "Decoder", ",", "rcs", "RangeCodes", ")", "uint", "{", "rc", ":=", "rcs", "[", "pr", ".", "ReadSymbol", "(", "pd", ")", "]", "\n", "return", "uint", "(", "rc", ".", "Base", ")", "+", "pr", ".", "ReadBits", "(", "uint", "(", "rc", ".", "Len", ")", ")", "\n", "}" ]
// ReadOffset reads an offset value using the provided RangeCodes indexed by // the symbol read.
[ "ReadOffset", "reads", "an", "offset", "value", "using", "the", "provided", "RangeCodes", "indexed", "by", "the", "symbol", "read", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/internal/prefix/reader.go#L149-L152
9,649
dsnet/compress
internal/prefix/reader.go
ReadBits
func (pr *Reader) ReadBits(nb uint) uint { if err := pr.PullBits(nb); err != nil { errors.Panic(err) } val := uint(pr.bufBits & uint64(1<<nb-1)) pr.bufBits >>= nb pr.numBits -= nb return val }
go
func (pr *Reader) ReadBits(nb uint) uint { if err := pr.PullBits(nb); err != nil { errors.Panic(err) } val := uint(pr.bufBits & uint64(1<<nb-1)) pr.bufBits >>= nb pr.numBits -= nb return val }
[ "func", "(", "pr", "*", "Reader", ")", "ReadBits", "(", "nb", "uint", ")", "uint", "{", "if", "err", ":=", "pr", ".", "PullBits", "(", "nb", ")", ";", "err", "!=", "nil", "{", "errors", ".", "Panic", "(", "err", ")", "\n", "}", "\n", "val", ":=", "uint", "(", "pr", ".", "bufBits", "&", "uint64", "(", "1", "<<", "nb", "-", "1", ")", ")", "\n", "pr", ".", "bufBits", ">>=", "nb", "\n", "pr", ".", "numBits", "-=", "nb", "\n", "return", "val", "\n", "}" ]
// ReadBits reads nb bits in from the underlying reader.
[ "ReadBits", "reads", "nb", "bits", "in", "from", "the", "underlying", "reader", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/internal/prefix/reader.go#L169-L177
9,650
dsnet/compress
internal/prefix/reader.go
ReadSymbol
func (pr *Reader) ReadSymbol(pd *Decoder) uint { if len(pd.chunks) == 0 { panicf(errors.Invalid, "decode with empty prefix tree") } nb := uint(pd.MinBits) for { if err := pr.PullBits(nb); err != nil { errors.Panic(err) } chunk := pd.chunks[uint32(pr.bufBits)&pd.chunkMask] nb = uint(chunk & countMask) if nb > uint(pd.chunkBits) { linkIdx := chunk >> countBits chunk = pd.links[linkIdx][uint32(pr.bufBits>>pd.chunkBits)&pd.linkMask] nb = uint(chunk & countMask) } if nb <= pr.numBits { pr.bufBits >>= nb pr.numBits -= nb return uint(chunk >> countBits) } } }
go
func (pr *Reader) ReadSymbol(pd *Decoder) uint { if len(pd.chunks) == 0 { panicf(errors.Invalid, "decode with empty prefix tree") } nb := uint(pd.MinBits) for { if err := pr.PullBits(nb); err != nil { errors.Panic(err) } chunk := pd.chunks[uint32(pr.bufBits)&pd.chunkMask] nb = uint(chunk & countMask) if nb > uint(pd.chunkBits) { linkIdx := chunk >> countBits chunk = pd.links[linkIdx][uint32(pr.bufBits>>pd.chunkBits)&pd.linkMask] nb = uint(chunk & countMask) } if nb <= pr.numBits { pr.bufBits >>= nb pr.numBits -= nb return uint(chunk >> countBits) } } }
[ "func", "(", "pr", "*", "Reader", ")", "ReadSymbol", "(", "pd", "*", "Decoder", ")", "uint", "{", "if", "len", "(", "pd", ".", "chunks", ")", "==", "0", "{", "panicf", "(", "errors", ".", "Invalid", ",", "\"", "\"", ")", "\n", "}", "\n\n", "nb", ":=", "uint", "(", "pd", ".", "MinBits", ")", "\n", "for", "{", "if", "err", ":=", "pr", ".", "PullBits", "(", "nb", ")", ";", "err", "!=", "nil", "{", "errors", ".", "Panic", "(", "err", ")", "\n", "}", "\n", "chunk", ":=", "pd", ".", "chunks", "[", "uint32", "(", "pr", ".", "bufBits", ")", "&", "pd", ".", "chunkMask", "]", "\n", "nb", "=", "uint", "(", "chunk", "&", "countMask", ")", "\n", "if", "nb", ">", "uint", "(", "pd", ".", "chunkBits", ")", "{", "linkIdx", ":=", "chunk", ">>", "countBits", "\n", "chunk", "=", "pd", ".", "links", "[", "linkIdx", "]", "[", "uint32", "(", "pr", ".", "bufBits", ">>", "pd", ".", "chunkBits", ")", "&", "pd", ".", "linkMask", "]", "\n", "nb", "=", "uint", "(", "chunk", "&", "countMask", ")", "\n", "}", "\n", "if", "nb", "<=", "pr", ".", "numBits", "{", "pr", ".", "bufBits", ">>=", "nb", "\n", "pr", ".", "numBits", "-=", "nb", "\n", "return", "uint", "(", "chunk", ">>", "countBits", ")", "\n", "}", "\n", "}", "\n", "}" ]
// ReadSymbol reads the next symbol using the provided prefix Decoder.
[ "ReadSymbol", "reads", "the", "next", "symbol", "using", "the", "provided", "prefix", "Decoder", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/internal/prefix/reader.go#L198-L221
9,651
dsnet/compress
internal/prefix/reader.go
Flush
func (pr *Reader) Flush() (int64, error) { if pr.bufRd == nil { return pr.Offset, nil } // Update the number of total bits to discard. pr.discardBits += int(pr.fedBits - pr.numBits) pr.fedBits = pr.numBits // Discard some bytes to update read offset. var err error nd := (pr.discardBits + 7) / 8 // Round up to nearest byte nd, err = pr.bufRd.Discard(nd) pr.discardBits -= nd * 8 // -7..0 pr.Offset += int64(nd) // These are invalid after Discard. pr.bufPeek = nil return pr.Offset, err }
go
func (pr *Reader) Flush() (int64, error) { if pr.bufRd == nil { return pr.Offset, nil } // Update the number of total bits to discard. pr.discardBits += int(pr.fedBits - pr.numBits) pr.fedBits = pr.numBits // Discard some bytes to update read offset. var err error nd := (pr.discardBits + 7) / 8 // Round up to nearest byte nd, err = pr.bufRd.Discard(nd) pr.discardBits -= nd * 8 // -7..0 pr.Offset += int64(nd) // These are invalid after Discard. pr.bufPeek = nil return pr.Offset, err }
[ "func", "(", "pr", "*", "Reader", ")", "Flush", "(", ")", "(", "int64", ",", "error", ")", "{", "if", "pr", ".", "bufRd", "==", "nil", "{", "return", "pr", ".", "Offset", ",", "nil", "\n", "}", "\n\n", "// Update the number of total bits to discard.", "pr", ".", "discardBits", "+=", "int", "(", "pr", ".", "fedBits", "-", "pr", ".", "numBits", ")", "\n", "pr", ".", "fedBits", "=", "pr", ".", "numBits", "\n\n", "// Discard some bytes to update read offset.", "var", "err", "error", "\n", "nd", ":=", "(", "pr", ".", "discardBits", "+", "7", ")", "/", "8", "// Round up to nearest byte", "\n", "nd", ",", "err", "=", "pr", ".", "bufRd", ".", "Discard", "(", "nd", ")", "\n", "pr", ".", "discardBits", "-=", "nd", "*", "8", "// -7..0", "\n", "pr", ".", "Offset", "+=", "int64", "(", "nd", ")", "\n\n", "// These are invalid after Discard.", "pr", ".", "bufPeek", "=", "nil", "\n", "return", "pr", ".", "Offset", ",", "err", "\n", "}" ]
// Flush updates the read offset of the underlying ByteReader. // If reader is a compress.BufferedReader, then this calls Discard to update // the read offset.
[ "Flush", "updates", "the", "read", "offset", "of", "the", "underlying", "ByteReader", ".", "If", "reader", "is", "a", "compress", ".", "BufferedReader", "then", "this", "calls", "Discard", "to", "update", "the", "read", "offset", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/internal/prefix/reader.go#L226-L245
9,652
dsnet/compress
brotli/transform.go
transformUppercase
func transformUppercase(word []byte, once bool) { for i := 0; i < len(word); { c := word[i] if c < 192 { if c >= 97 && c <= 122 { word[i] ^= 32 } i += 1 } else if c < 224 { if i+1 < len(word) { word[i+1] ^= 32 } i += 2 } else { if i+2 < len(word) { word[i+2] ^= 5 } i += 3 } if once { return } } }
go
func transformUppercase(word []byte, once bool) { for i := 0; i < len(word); { c := word[i] if c < 192 { if c >= 97 && c <= 122 { word[i] ^= 32 } i += 1 } else if c < 224 { if i+1 < len(word) { word[i+1] ^= 32 } i += 2 } else { if i+2 < len(word) { word[i+2] ^= 5 } i += 3 } if once { return } } }
[ "func", "transformUppercase", "(", "word", "[", "]", "byte", ",", "once", "bool", ")", "{", "for", "i", ":=", "0", ";", "i", "<", "len", "(", "word", ")", ";", "{", "c", ":=", "word", "[", "i", "]", "\n", "if", "c", "<", "192", "{", "if", "c", ">=", "97", "&&", "c", "<=", "122", "{", "word", "[", "i", "]", "^=", "32", "\n", "}", "\n", "i", "+=", "1", "\n", "}", "else", "if", "c", "<", "224", "{", "if", "i", "+", "1", "<", "len", "(", "word", ")", "{", "word", "[", "i", "+", "1", "]", "^=", "32", "\n", "}", "\n", "i", "+=", "2", "\n", "}", "else", "{", "if", "i", "+", "2", "<", "len", "(", "word", ")", "{", "word", "[", "i", "+", "2", "]", "^=", "5", "\n", "}", "\n", "i", "+=", "3", "\n", "}", "\n", "if", "once", "{", "return", "\n", "}", "\n", "}", "\n", "}" ]
// transformUppercase transform the word to be in uppercase using the algorithm // presented in RFC section 8. If once is set, then loop only executes once.
[ "transformUppercase", "transform", "the", "word", "to", "be", "in", "uppercase", "using", "the", "algorithm", "presented", "in", "RFC", "section", "8", ".", "If", "once", "is", "set", "then", "loop", "only", "executes", "once", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/brotli/transform.go#L204-L227
9,653
dsnet/compress
internal/tool/bench/benchmarks.go
BenchmarkEncoder
func BenchmarkEncoder(input []byte, enc Encoder, lvl int) testing.BenchmarkResult { return testing.Benchmark(func(b *testing.B) { b.StopTimer() if enc == nil { b.Fatalf("unexpected error: nil Encoder") } runtime.GC() b.StartTimer() for i := 0; i < b.N; i++ { wr := enc(ioutil.Discard, lvl) _, err := io.Copy(wr, bytes.NewBuffer(input)) if err := wr.Close(); err != nil { b.Fatalf("unexpected error: %v", err) } if err != nil { b.Fatalf("unexpected error: %v", err) } b.SetBytes(int64(len(input))) } }) }
go
func BenchmarkEncoder(input []byte, enc Encoder, lvl int) testing.BenchmarkResult { return testing.Benchmark(func(b *testing.B) { b.StopTimer() if enc == nil { b.Fatalf("unexpected error: nil Encoder") } runtime.GC() b.StartTimer() for i := 0; i < b.N; i++ { wr := enc(ioutil.Discard, lvl) _, err := io.Copy(wr, bytes.NewBuffer(input)) if err := wr.Close(); err != nil { b.Fatalf("unexpected error: %v", err) } if err != nil { b.Fatalf("unexpected error: %v", err) } b.SetBytes(int64(len(input))) } }) }
[ "func", "BenchmarkEncoder", "(", "input", "[", "]", "byte", ",", "enc", "Encoder", ",", "lvl", "int", ")", "testing", ".", "BenchmarkResult", "{", "return", "testing", ".", "Benchmark", "(", "func", "(", "b", "*", "testing", ".", "B", ")", "{", "b", ".", "StopTimer", "(", ")", "\n", "if", "enc", "==", "nil", "{", "b", ".", "Fatalf", "(", "\"", "\"", ")", "\n", "}", "\n", "runtime", ".", "GC", "(", ")", "\n", "b", ".", "StartTimer", "(", ")", "\n", "for", "i", ":=", "0", ";", "i", "<", "b", ".", "N", ";", "i", "++", "{", "wr", ":=", "enc", "(", "ioutil", ".", "Discard", ",", "lvl", ")", "\n", "_", ",", "err", ":=", "io", ".", "Copy", "(", "wr", ",", "bytes", ".", "NewBuffer", "(", "input", ")", ")", "\n", "if", "err", ":=", "wr", ".", "Close", "(", ")", ";", "err", "!=", "nil", "{", "b", ".", "Fatalf", "(", "\"", "\"", ",", "err", ")", "\n", "}", "\n", "if", "err", "!=", "nil", "{", "b", ".", "Fatalf", "(", "\"", "\"", ",", "err", ")", "\n", "}", "\n", "b", ".", "SetBytes", "(", "int64", "(", "len", "(", "input", ")", ")", ")", "\n", "}", "\n", "}", ")", "\n", "}" ]
// BenchmarkEncoder benchmarks a single encoder on the given input data using // the selected compression level and reports the result.
[ "BenchmarkEncoder", "benchmarks", "a", "single", "encoder", "on", "the", "given", "input", "data", "using", "the", "selected", "compression", "level", "and", "reports", "the", "result", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/internal/tool/bench/benchmarks.go#L28-L48
9,654
dsnet/compress
internal/tool/bench/benchmarks.go
BenchmarkDecoder
func BenchmarkDecoder(input []byte, dec Decoder) testing.BenchmarkResult { return testing.Benchmark(func(b *testing.B) { b.StopTimer() if dec == nil { b.Fatalf("unexpected error: nil Decoder") } runtime.GC() b.StartTimer() for i := 0; i < b.N; i++ { rd := dec(bufio.NewReader(bytes.NewBuffer(input))) cnt, err := io.Copy(ioutil.Discard, rd) if err := rd.Close(); err != nil { b.Fatalf("unexpected error: %v", err) } if err != nil { b.Fatalf("unexpected error: %v", err) } b.SetBytes(cnt) } }) }
go
func BenchmarkDecoder(input []byte, dec Decoder) testing.BenchmarkResult { return testing.Benchmark(func(b *testing.B) { b.StopTimer() if dec == nil { b.Fatalf("unexpected error: nil Decoder") } runtime.GC() b.StartTimer() for i := 0; i < b.N; i++ { rd := dec(bufio.NewReader(bytes.NewBuffer(input))) cnt, err := io.Copy(ioutil.Discard, rd) if err := rd.Close(); err != nil { b.Fatalf("unexpected error: %v", err) } if err != nil { b.Fatalf("unexpected error: %v", err) } b.SetBytes(cnt) } }) }
[ "func", "BenchmarkDecoder", "(", "input", "[", "]", "byte", ",", "dec", "Decoder", ")", "testing", ".", "BenchmarkResult", "{", "return", "testing", ".", "Benchmark", "(", "func", "(", "b", "*", "testing", ".", "B", ")", "{", "b", ".", "StopTimer", "(", ")", "\n", "if", "dec", "==", "nil", "{", "b", ".", "Fatalf", "(", "\"", "\"", ")", "\n", "}", "\n", "runtime", ".", "GC", "(", ")", "\n", "b", ".", "StartTimer", "(", ")", "\n", "for", "i", ":=", "0", ";", "i", "<", "b", ".", "N", ";", "i", "++", "{", "rd", ":=", "dec", "(", "bufio", ".", "NewReader", "(", "bytes", ".", "NewBuffer", "(", "input", ")", ")", ")", "\n", "cnt", ",", "err", ":=", "io", ".", "Copy", "(", "ioutil", ".", "Discard", ",", "rd", ")", "\n", "if", "err", ":=", "rd", ".", "Close", "(", ")", ";", "err", "!=", "nil", "{", "b", ".", "Fatalf", "(", "\"", "\"", ",", "err", ")", "\n", "}", "\n", "if", "err", "!=", "nil", "{", "b", ".", "Fatalf", "(", "\"", "\"", ",", "err", ")", "\n", "}", "\n", "b", ".", "SetBytes", "(", "cnt", ")", "\n", "}", "\n", "}", ")", "\n", "}" ]
// BenchmarkDecoder benchmarks a single decoder on the given pre-compressed // input data and reports the result.
[ "BenchmarkDecoder", "benchmarks", "a", "single", "decoder", "on", "the", "given", "pre", "-", "compressed", "input", "data", "and", "reports", "the", "result", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/internal/tool/bench/benchmarks.go#L71-L91
9,655
dsnet/compress
internal/prefix/encoder.go
Init
func (pe *Encoder) Init(codes PrefixCodes) { // Handle special case trees. if len(codes) <= 1 { switch { case len(codes) == 0: // Empty tree (should error if used later) *pe = Encoder{chunks: pe.chunks[:0], NumSyms: 0} case len(codes) == 1 && codes[0].Len == 0: // Single code tree (bit-length of zero) pe.chunks = append(pe.chunks[:0], codes[0].Val<<countBits|0) *pe = Encoder{chunks: pe.chunks[:1], NumSyms: 1} default: panic("invalid codes") } return } if internal.Debug && !sort.IsSorted(prefixCodesBySymbol(codes)) { panic("input codes is not sorted") } if internal.Debug && !(codes.checkLengths() && codes.checkPrefixes()) { panic("detected incomplete or overlapping codes") } // Enough chunks to contain all the symbols. numChunks := 1 for n := len(codes) - 1; n > 0; n >>= 1 { numChunks <<= 1 } pe.NumSyms = uint32(len(codes)) retry: // Allocate and reset chunks. pe.chunks = allocUint32s(pe.chunks, numChunks) pe.chunkMask = uint32(numChunks - 1) for i := range pe.chunks { pe.chunks[i] = 0 // Logic below relies on zero value as uninitialized } // Insert each symbol, checking that there are no conflicts. for _, c := range codes { if pe.chunks[c.Sym&pe.chunkMask] > 0 { // Collision found our "hash" table, so grow and try again. numChunks <<= 1 goto retry } pe.chunks[c.Sym&pe.chunkMask] = c.Val<<countBits | c.Len } }
go
func (pe *Encoder) Init(codes PrefixCodes) { // Handle special case trees. if len(codes) <= 1 { switch { case len(codes) == 0: // Empty tree (should error if used later) *pe = Encoder{chunks: pe.chunks[:0], NumSyms: 0} case len(codes) == 1 && codes[0].Len == 0: // Single code tree (bit-length of zero) pe.chunks = append(pe.chunks[:0], codes[0].Val<<countBits|0) *pe = Encoder{chunks: pe.chunks[:1], NumSyms: 1} default: panic("invalid codes") } return } if internal.Debug && !sort.IsSorted(prefixCodesBySymbol(codes)) { panic("input codes is not sorted") } if internal.Debug && !(codes.checkLengths() && codes.checkPrefixes()) { panic("detected incomplete or overlapping codes") } // Enough chunks to contain all the symbols. numChunks := 1 for n := len(codes) - 1; n > 0; n >>= 1 { numChunks <<= 1 } pe.NumSyms = uint32(len(codes)) retry: // Allocate and reset chunks. pe.chunks = allocUint32s(pe.chunks, numChunks) pe.chunkMask = uint32(numChunks - 1) for i := range pe.chunks { pe.chunks[i] = 0 // Logic below relies on zero value as uninitialized } // Insert each symbol, checking that there are no conflicts. for _, c := range codes { if pe.chunks[c.Sym&pe.chunkMask] > 0 { // Collision found our "hash" table, so grow and try again. numChunks <<= 1 goto retry } pe.chunks[c.Sym&pe.chunkMask] = c.Val<<countBits | c.Len } }
[ "func", "(", "pe", "*", "Encoder", ")", "Init", "(", "codes", "PrefixCodes", ")", "{", "// Handle special case trees.", "if", "len", "(", "codes", ")", "<=", "1", "{", "switch", "{", "case", "len", "(", "codes", ")", "==", "0", ":", "// Empty tree (should error if used later)", "*", "pe", "=", "Encoder", "{", "chunks", ":", "pe", ".", "chunks", "[", ":", "0", "]", ",", "NumSyms", ":", "0", "}", "\n", "case", "len", "(", "codes", ")", "==", "1", "&&", "codes", "[", "0", "]", ".", "Len", "==", "0", ":", "// Single code tree (bit-length of zero)", "pe", ".", "chunks", "=", "append", "(", "pe", ".", "chunks", "[", ":", "0", "]", ",", "codes", "[", "0", "]", ".", "Val", "<<", "countBits", "|", "0", ")", "\n", "*", "pe", "=", "Encoder", "{", "chunks", ":", "pe", ".", "chunks", "[", ":", "1", "]", ",", "NumSyms", ":", "1", "}", "\n", "default", ":", "panic", "(", "\"", "\"", ")", "\n", "}", "\n", "return", "\n", "}", "\n", "if", "internal", ".", "Debug", "&&", "!", "sort", ".", "IsSorted", "(", "prefixCodesBySymbol", "(", "codes", ")", ")", "{", "panic", "(", "\"", "\"", ")", "\n", "}", "\n", "if", "internal", ".", "Debug", "&&", "!", "(", "codes", ".", "checkLengths", "(", ")", "&&", "codes", ".", "checkPrefixes", "(", ")", ")", "{", "panic", "(", "\"", "\"", ")", "\n", "}", "\n\n", "// Enough chunks to contain all the symbols.", "numChunks", ":=", "1", "\n", "for", "n", ":=", "len", "(", "codes", ")", "-", "1", ";", "n", ">", "0", ";", "n", ">>=", "1", "{", "numChunks", "<<=", "1", "\n", "}", "\n", "pe", ".", "NumSyms", "=", "uint32", "(", "len", "(", "codes", ")", ")", "\n\n", "retry", ":", "// Allocate and reset chunks.", "pe", ".", "chunks", "=", "allocUint32s", "(", "pe", ".", "chunks", ",", "numChunks", ")", "\n", "pe", ".", "chunkMask", "=", "uint32", "(", "numChunks", "-", "1", ")", "\n", "for", "i", ":=", "range", "pe", ".", "chunks", "{", "pe", ".", "chunks", "[", "i", "]", "=", "0", "// Logic below relies on zero value as uninitialized", "\n", "}", "\n\n", "// Insert each symbol, checking that there are no conflicts.", "for", "_", ",", "c", ":=", "range", "codes", "{", "if", "pe", ".", "chunks", "[", "c", ".", "Sym", "&", "pe", ".", "chunkMask", "]", ">", "0", "{", "// Collision found our \"hash\" table, so grow and try again.", "numChunks", "<<=", "1", "\n", "goto", "retry", "\n", "}", "\n", "pe", ".", "chunks", "[", "c", ".", "Sym", "&", "pe", ".", "chunkMask", "]", "=", "c", ".", "Val", "<<", "countBits", "|", "c", ".", "Len", "\n", "}", "\n", "}" ]
// Init initializes Encoder according to the codes provided.
[ "Init", "initializes", "Encoder", "according", "to", "the", "codes", "provided", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/internal/prefix/encoder.go#L21-L66
9,656
dsnet/compress
internal/tool/fuzz/xflate_meta/xflate_meta.go
decodeMeta
func decodeMeta(data []byte) ([]byte, bool) { r := bytes.NewReader(data) mr := xflate.NewMetaReader(r) b, err := ioutil.ReadAll(mr) if err != nil { return nil, false } pos := int(r.Size()) - r.Len() decompressMeta(data[:pos]) return b, true }
go
func decodeMeta(data []byte) ([]byte, bool) { r := bytes.NewReader(data) mr := xflate.NewMetaReader(r) b, err := ioutil.ReadAll(mr) if err != nil { return nil, false } pos := int(r.Size()) - r.Len() decompressMeta(data[:pos]) return b, true }
[ "func", "decodeMeta", "(", "data", "[", "]", "byte", ")", "(", "[", "]", "byte", ",", "bool", ")", "{", "r", ":=", "bytes", ".", "NewReader", "(", "data", ")", "\n", "mr", ":=", "xflate", ".", "NewMetaReader", "(", "r", ")", "\n", "b", ",", "err", ":=", "ioutil", ".", "ReadAll", "(", "mr", ")", "\n", "if", "err", "!=", "nil", "{", "return", "nil", ",", "false", "\n", "}", "\n", "pos", ":=", "int", "(", "r", ".", "Size", "(", ")", ")", "-", "r", ".", "Len", "(", ")", "\n", "decompressMeta", "(", "data", "[", ":", "pos", "]", ")", "\n", "return", "b", ",", "true", "\n", "}" ]
// decodeMeta attempts to decode the metadata. // If successful, it verifies that meta-encoded blocks are DEFLATE blocks.
[ "decodeMeta", "attempts", "to", "decode", "the", "metadata", ".", "If", "successful", "it", "verifies", "that", "meta", "-", "encoded", "blocks", "are", "DEFLATE", "blocks", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/internal/tool/fuzz/xflate_meta/xflate_meta.go#L30-L40
9,657
dsnet/compress
internal/tool/fuzz/xflate_meta/xflate_meta.go
decompressMeta
func decompressMeta(data []byte) { // Make a copy and append DEFLATE terminator block. data = append([]byte(nil), data...) data = append(data, []byte{0x01, 0x00, 0x00, 0xff, 0xff}...) r := bytes.NewReader(data) for r.Len() > 0 { zr := flate.NewReader(r) b, err := ioutil.ReadAll(zr) if err != nil { panic(err) } if len(b) > 0 { panic("non-zero meta-encoded block") } if err := zr.Close(); err != nil { panic(err) } } }
go
func decompressMeta(data []byte) { // Make a copy and append DEFLATE terminator block. data = append([]byte(nil), data...) data = append(data, []byte{0x01, 0x00, 0x00, 0xff, 0xff}...) r := bytes.NewReader(data) for r.Len() > 0 { zr := flate.NewReader(r) b, err := ioutil.ReadAll(zr) if err != nil { panic(err) } if len(b) > 0 { panic("non-zero meta-encoded block") } if err := zr.Close(); err != nil { panic(err) } } }
[ "func", "decompressMeta", "(", "data", "[", "]", "byte", ")", "{", "// Make a copy and append DEFLATE terminator block.", "data", "=", "append", "(", "[", "]", "byte", "(", "nil", ")", ",", "data", "...", ")", "\n", "data", "=", "append", "(", "data", ",", "[", "]", "byte", "{", "0x01", ",", "0x00", ",", "0x00", ",", "0xff", ",", "0xff", "}", "...", ")", "\n\n", "r", ":=", "bytes", ".", "NewReader", "(", "data", ")", "\n", "for", "r", ".", "Len", "(", ")", ">", "0", "{", "zr", ":=", "flate", ".", "NewReader", "(", "r", ")", "\n", "b", ",", "err", ":=", "ioutil", ".", "ReadAll", "(", "zr", ")", "\n", "if", "err", "!=", "nil", "{", "panic", "(", "err", ")", "\n", "}", "\n", "if", "len", "(", "b", ")", ">", "0", "{", "panic", "(", "\"", "\"", ")", "\n", "}", "\n", "if", "err", ":=", "zr", ".", "Close", "(", ")", ";", "err", "!=", "nil", "{", "panic", "(", "err", ")", "\n", "}", "\n", "}", "\n", "}" ]
// decompressMeta attempts to decompress the meta-encoded blocks. // It expects decompression to succeed and to output nothing.
[ "decompressMeta", "attempts", "to", "decompress", "the", "meta", "-", "encoded", "blocks", ".", "It", "expects", "decompression", "to", "succeed", "and", "to", "output", "nothing", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/internal/tool/fuzz/xflate_meta/xflate_meta.go#L44-L63
9,658
dsnet/compress
brotli/dict.go
initDictLUTs
func initDictLUTs() { // Sweep from minDictLen to maxDictLen, inclusive. for i := minDictLen; i <= maxDictLen; i++ { dictSizes[i] = 1 << uint(dictBitSizes[i]) dictOffsets[i] = dictOffsets[i-1] + (i-1)*dictSizes[i-1] } }
go
func initDictLUTs() { // Sweep from minDictLen to maxDictLen, inclusive. for i := minDictLen; i <= maxDictLen; i++ { dictSizes[i] = 1 << uint(dictBitSizes[i]) dictOffsets[i] = dictOffsets[i-1] + (i-1)*dictSizes[i-1] } }
[ "func", "initDictLUTs", "(", ")", "{", "// Sweep from minDictLen to maxDictLen, inclusive.", "for", "i", ":=", "minDictLen", ";", "i", "<=", "maxDictLen", ";", "i", "++", "{", "dictSizes", "[", "i", "]", "=", "1", "<<", "uint", "(", "dictBitSizes", "[", "i", "]", ")", "\n", "dictOffsets", "[", "i", "]", "=", "dictOffsets", "[", "i", "-", "1", "]", "+", "(", "i", "-", "1", ")", "*", "dictSizes", "[", "i", "-", "1", "]", "\n", "}", "\n", "}" ]
// We dynamically compute NWORDS and DOFFSET according to the algorithm // provided in RFC section 8.
[ "We", "dynamically", "compute", "NWORDS", "and", "DOFFSET", "according", "to", "the", "algorithm", "provided", "in", "RFC", "section", "8", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/brotli/dict.go#L29-L35
9,659
dsnet/compress
internal/prefix/range.go
MakeRangeCodes
func MakeRangeCodes(minBase uint, bits []uint) (rc RangeCodes) { for _, nb := range bits { rc = append(rc, RangeCode{Base: uint32(minBase), Len: uint32(nb)}) minBase += 1 << nb } return rc }
go
func MakeRangeCodes(minBase uint, bits []uint) (rc RangeCodes) { for _, nb := range bits { rc = append(rc, RangeCode{Base: uint32(minBase), Len: uint32(nb)}) minBase += 1 << nb } return rc }
[ "func", "MakeRangeCodes", "(", "minBase", "uint", ",", "bits", "[", "]", "uint", ")", "(", "rc", "RangeCodes", ")", "{", "for", "_", ",", "nb", ":=", "range", "bits", "{", "rc", "=", "append", "(", "rc", ",", "RangeCode", "{", "Base", ":", "uint32", "(", "minBase", ")", ",", "Len", ":", "uint32", "(", "nb", ")", "}", ")", "\n", "minBase", "+=", "1", "<<", "nb", "\n", "}", "\n", "return", "rc", "\n", "}" ]
// MakeRangeCodes creates a RangeCodes, where each region is assumed to be // contiguously stacked, without any gaps, with bit-lengths taken from bits.
[ "MakeRangeCodes", "creates", "a", "RangeCodes", "where", "each", "region", "is", "assumed", "to", "be", "contiguously", "stacked", "without", "any", "gaps", "with", "bit", "-", "lengths", "taken", "from", "bits", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/internal/prefix/range.go#L24-L30
9,660
dsnet/compress
flate/prefix.go
handleDegenerateCodes
func handleDegenerateCodes(codes prefix.PrefixCodes, maxSyms uint) prefix.PrefixCodes { if len(codes) != 1 { return codes } return append(codes, prefix.PrefixCode{Sym: uint32(maxSyms), Len: 1}) }
go
func handleDegenerateCodes(codes prefix.PrefixCodes, maxSyms uint) prefix.PrefixCodes { if len(codes) != 1 { return codes } return append(codes, prefix.PrefixCode{Sym: uint32(maxSyms), Len: 1}) }
[ "func", "handleDegenerateCodes", "(", "codes", "prefix", ".", "PrefixCodes", ",", "maxSyms", "uint", ")", "prefix", ".", "PrefixCodes", "{", "if", "len", "(", "codes", ")", "!=", "1", "{", "return", "codes", "\n", "}", "\n", "return", "append", "(", "codes", ",", "prefix", ".", "PrefixCode", "{", "Sym", ":", "uint32", "(", "maxSyms", ")", ",", "Len", ":", "1", "}", ")", "\n", "}" ]
// RFC section 3.2.7 allows degenerate prefix trees with only node, but requires // a single bit for that node. This causes an unbalanced tree where the "1" code // is unused. The canonical prefix code generation algorithm breaks with this. // // To handle this case, we artificially insert another node for the "1" code // that uses a symbol larger than the alphabet to force an error later if // the code ends up getting used.
[ "RFC", "section", "3", ".", "2", ".", "7", "allows", "degenerate", "prefix", "trees", "with", "only", "node", "but", "requires", "a", "single", "bit", "for", "that", "node", ".", "This", "causes", "an", "unbalanced", "tree", "where", "the", "1", "code", "is", "unused", ".", "The", "canonical", "prefix", "code", "generation", "algorithm", "breaks", "with", "this", ".", "To", "handle", "this", "case", "we", "artificially", "insert", "another", "node", "for", "the", "1", "code", "that", "uses", "a", "symbol", "larger", "than", "the", "alphabet", "to", "force", "an", "error", "later", "if", "the", "code", "ends", "up", "getting", "used", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/flate/prefix.go#L203-L208
9,661
dsnet/compress
internal/common.go
ReverseUint32
func ReverseUint32(v uint32) (x uint32) { x |= uint32(ReverseLUT[byte(v>>0)]) << 24 x |= uint32(ReverseLUT[byte(v>>8)]) << 16 x |= uint32(ReverseLUT[byte(v>>16)]) << 8 x |= uint32(ReverseLUT[byte(v>>24)]) << 0 return x }
go
func ReverseUint32(v uint32) (x uint32) { x |= uint32(ReverseLUT[byte(v>>0)]) << 24 x |= uint32(ReverseLUT[byte(v>>8)]) << 16 x |= uint32(ReverseLUT[byte(v>>16)]) << 8 x |= uint32(ReverseLUT[byte(v>>24)]) << 0 return x }
[ "func", "ReverseUint32", "(", "v", "uint32", ")", "(", "x", "uint32", ")", "{", "x", "|=", "uint32", "(", "ReverseLUT", "[", "byte", "(", "v", ">>", "0", ")", "]", ")", "<<", "24", "\n", "x", "|=", "uint32", "(", "ReverseLUT", "[", "byte", "(", "v", ">>", "8", ")", "]", ")", "<<", "16", "\n", "x", "|=", "uint32", "(", "ReverseLUT", "[", "byte", "(", "v", ">>", "16", ")", "]", ")", "<<", "8", "\n", "x", "|=", "uint32", "(", "ReverseLUT", "[", "byte", "(", "v", ">>", "24", ")", "]", ")", "<<", "0", "\n", "return", "x", "\n", "}" ]
// ReverseUint32 reverses all bits of v.
[ "ReverseUint32", "reverses", "all", "bits", "of", "v", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/internal/common.go#L34-L40
9,662
dsnet/compress
internal/common.go
ReverseUint32N
func ReverseUint32N(v uint32, n uint) (x uint32) { return ReverseUint32(v << (32 - n)) }
go
func ReverseUint32N(v uint32, n uint) (x uint32) { return ReverseUint32(v << (32 - n)) }
[ "func", "ReverseUint32N", "(", "v", "uint32", ",", "n", "uint", ")", "(", "x", "uint32", ")", "{", "return", "ReverseUint32", "(", "v", "<<", "(", "32", "-", "n", ")", ")", "\n", "}" ]
// ReverseUint32N reverses the lower n bits of v.
[ "ReverseUint32N", "reverses", "the", "lower", "n", "bits", "of", "v", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/internal/common.go#L43-L45
9,663
dsnet/compress
internal/common.go
ReverseUint64
func ReverseUint64(v uint64) (x uint64) { x |= uint64(ReverseLUT[byte(v>>0)]) << 56 x |= uint64(ReverseLUT[byte(v>>8)]) << 48 x |= uint64(ReverseLUT[byte(v>>16)]) << 40 x |= uint64(ReverseLUT[byte(v>>24)]) << 32 x |= uint64(ReverseLUT[byte(v>>32)]) << 24 x |= uint64(ReverseLUT[byte(v>>40)]) << 16 x |= uint64(ReverseLUT[byte(v>>48)]) << 8 x |= uint64(ReverseLUT[byte(v>>56)]) << 0 return x }
go
func ReverseUint64(v uint64) (x uint64) { x |= uint64(ReverseLUT[byte(v>>0)]) << 56 x |= uint64(ReverseLUT[byte(v>>8)]) << 48 x |= uint64(ReverseLUT[byte(v>>16)]) << 40 x |= uint64(ReverseLUT[byte(v>>24)]) << 32 x |= uint64(ReverseLUT[byte(v>>32)]) << 24 x |= uint64(ReverseLUT[byte(v>>40)]) << 16 x |= uint64(ReverseLUT[byte(v>>48)]) << 8 x |= uint64(ReverseLUT[byte(v>>56)]) << 0 return x }
[ "func", "ReverseUint64", "(", "v", "uint64", ")", "(", "x", "uint64", ")", "{", "x", "|=", "uint64", "(", "ReverseLUT", "[", "byte", "(", "v", ">>", "0", ")", "]", ")", "<<", "56", "\n", "x", "|=", "uint64", "(", "ReverseLUT", "[", "byte", "(", "v", ">>", "8", ")", "]", ")", "<<", "48", "\n", "x", "|=", "uint64", "(", "ReverseLUT", "[", "byte", "(", "v", ">>", "16", ")", "]", ")", "<<", "40", "\n", "x", "|=", "uint64", "(", "ReverseLUT", "[", "byte", "(", "v", ">>", "24", ")", "]", ")", "<<", "32", "\n", "x", "|=", "uint64", "(", "ReverseLUT", "[", "byte", "(", "v", ">>", "32", ")", "]", ")", "<<", "24", "\n", "x", "|=", "uint64", "(", "ReverseLUT", "[", "byte", "(", "v", ">>", "40", ")", "]", ")", "<<", "16", "\n", "x", "|=", "uint64", "(", "ReverseLUT", "[", "byte", "(", "v", ">>", "48", ")", "]", ")", "<<", "8", "\n", "x", "|=", "uint64", "(", "ReverseLUT", "[", "byte", "(", "v", ">>", "56", ")", "]", ")", "<<", "0", "\n", "return", "x", "\n", "}" ]
// ReverseUint64 reverses all bits of v.
[ "ReverseUint64", "reverses", "all", "bits", "of", "v", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/internal/common.go#L48-L58
9,664
dsnet/compress
internal/common.go
ReverseUint64N
func ReverseUint64N(v uint64, n uint) (x uint64) { return ReverseUint64(v << (64 - n)) }
go
func ReverseUint64N(v uint64, n uint) (x uint64) { return ReverseUint64(v << (64 - n)) }
[ "func", "ReverseUint64N", "(", "v", "uint64", ",", "n", "uint", ")", "(", "x", "uint64", ")", "{", "return", "ReverseUint64", "(", "v", "<<", "(", "64", "-", "n", ")", ")", "\n", "}" ]
// ReverseUint64N reverses the lower n bits of v.
[ "ReverseUint64N", "reverses", "the", "lower", "n", "bits", "of", "v", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/internal/common.go#L61-L63
9,665
dsnet/compress
xflate/internal/meta/meta.go
ReverseSearch
func ReverseSearch(data []byte) int { var magic uint32 for i := len(data) - 1; i >= 0; i-- { magic = (magic << 8) | uint32(data[i]) if magic&magicMask == magicVals { return i } } return -1 }
go
func ReverseSearch(data []byte) int { var magic uint32 for i := len(data) - 1; i >= 0; i-- { magic = (magic << 8) | uint32(data[i]) if magic&magicMask == magicVals { return i } } return -1 }
[ "func", "ReverseSearch", "(", "data", "[", "]", "byte", ")", "int", "{", "var", "magic", "uint32", "\n", "for", "i", ":=", "len", "(", "data", ")", "-", "1", ";", "i", ">=", "0", ";", "i", "--", "{", "magic", "=", "(", "magic", "<<", "8", ")", "|", "uint32", "(", "data", "[", "i", "]", ")", "\n", "if", "magic", "&", "magicMask", "==", "magicVals", "{", "return", "i", "\n", "}", "\n", "}", "\n", "return", "-", "1", "\n", "}" ]
// ReverseSearch searches for a meta header in reverse. This returns the last // index where the header was found. If not found, it returns -1.
[ "ReverseSearch", "searches", "for", "a", "meta", "header", "in", "reverse", ".", "This", "returns", "the", "last", "index", "where", "the", "header", "was", "found", ".", "If", "not", "found", "it", "returns", "-", "1", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/xflate/internal/meta/meta.go#L35-L44
9,666
dsnet/compress
xflate/internal/meta/meta.go
numBits
func numBits(b byte) (zeros, ones int) { ones = int(oneBitsLUT[b]) zeros = 8 - ones return }
go
func numBits(b byte) (zeros, ones int) { ones = int(oneBitsLUT[b]) zeros = 8 - ones return }
[ "func", "numBits", "(", "b", "byte", ")", "(", "zeros", ",", "ones", "int", ")", "{", "ones", "=", "int", "(", "oneBitsLUT", "[", "b", "]", ")", "\n", "zeros", "=", "8", "-", "ones", "\n", "return", "\n", "}" ]
// numBits counts the number of zero and one bits in the byte.
[ "numBits", "counts", "the", "number", "of", "zero", "and", "one", "bits", "in", "the", "byte", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/xflate/internal/meta/meta.go#L156-L160
9,667
dsnet/compress
xflate/reader.go
NewReader
func NewReader(rs io.ReadSeeker, conf *ReaderConfig) (*Reader, error) { xr := new(Reader) err := xr.Reset(rs) return xr, err }
go
func NewReader(rs io.ReadSeeker, conf *ReaderConfig) (*Reader, error) { xr := new(Reader) err := xr.Reset(rs) return xr, err }
[ "func", "NewReader", "(", "rs", "io", ".", "ReadSeeker", ",", "conf", "*", "ReaderConfig", ")", "(", "*", "Reader", ",", "error", ")", "{", "xr", ":=", "new", "(", "Reader", ")", "\n", "err", ":=", "xr", ".", "Reset", "(", "rs", ")", "\n", "return", "xr", ",", "err", "\n", "}" ]
// NewReader creates a new Reader reading the given reader rs. This reader can // only decompress files in the XFLATE format. If the underlying stream is // regular DEFLATE and not XFLATE, then this returns error. // // Regardless of the current offset in rs, this function Seeks to the end of rs // in order to determine the total compressed size. The Reader returned has its // offset set to the start of the stream. // // If conf is nil, then default configuration values are used. Reader copies // all configuration values as necessary and does not store conf.
[ "NewReader", "creates", "a", "new", "Reader", "reading", "the", "given", "reader", "rs", ".", "This", "reader", "can", "only", "decompress", "files", "in", "the", "XFLATE", "format", ".", "If", "the", "underlying", "stream", "is", "regular", "DEFLATE", "and", "not", "XFLATE", "then", "this", "returns", "error", ".", "Regardless", "of", "the", "current", "offset", "in", "rs", "this", "function", "Seeks", "to", "the", "end", "of", "rs", "in", "order", "to", "determine", "the", "total", "compressed", "size", ".", "The", "Reader", "returned", "has", "its", "offset", "set", "to", "the", "start", "of", "the", "stream", ".", "If", "conf", "is", "nil", "then", "default", "configuration", "values", "are", "used", ".", "Reader", "copies", "all", "configuration", "values", "as", "necessary", "and", "does", "not", "store", "conf", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/xflate/reader.go#L113-L117
9,668
dsnet/compress
xflate/reader.go
Reset
func (xr *Reader) Reset(rs io.ReadSeeker) error { *xr = Reader{ rd: rs, mr: xr.mr, zr: xr.zr, idx: xr.idx, br: xr.br, bw: xr.bw, idxs: xr.idxs, chunks: xr.chunks, } if xr.zr == nil { xr.zr, _ = newFlateReader(nil) } xr.idx.Reset() // Read entire index. var backSize, footSize int64 if backSize, footSize, xr.err = xr.decodeFooter(); xr.err != nil { return xr.err } if xr.err = xr.decodeIndexes(backSize); xr.err != nil { return xr.err } if !xr.idx.AppendRecord(footSize, 0, footerType) { xr.err = errCorrupted return xr.err } // Setup initial chunk reader. _, xr.err = xr.Seek(0, io.SeekStart) return xr.err }
go
func (xr *Reader) Reset(rs io.ReadSeeker) error { *xr = Reader{ rd: rs, mr: xr.mr, zr: xr.zr, idx: xr.idx, br: xr.br, bw: xr.bw, idxs: xr.idxs, chunks: xr.chunks, } if xr.zr == nil { xr.zr, _ = newFlateReader(nil) } xr.idx.Reset() // Read entire index. var backSize, footSize int64 if backSize, footSize, xr.err = xr.decodeFooter(); xr.err != nil { return xr.err } if xr.err = xr.decodeIndexes(backSize); xr.err != nil { return xr.err } if !xr.idx.AppendRecord(footSize, 0, footerType) { xr.err = errCorrupted return xr.err } // Setup initial chunk reader. _, xr.err = xr.Seek(0, io.SeekStart) return xr.err }
[ "func", "(", "xr", "*", "Reader", ")", "Reset", "(", "rs", "io", ".", "ReadSeeker", ")", "error", "{", "*", "xr", "=", "Reader", "{", "rd", ":", "rs", ",", "mr", ":", "xr", ".", "mr", ",", "zr", ":", "xr", ".", "zr", ",", "idx", ":", "xr", ".", "idx", ",", "br", ":", "xr", ".", "br", ",", "bw", ":", "xr", ".", "bw", ",", "idxs", ":", "xr", ".", "idxs", ",", "chunks", ":", "xr", ".", "chunks", ",", "}", "\n", "if", "xr", ".", "zr", "==", "nil", "{", "xr", ".", "zr", ",", "_", "=", "newFlateReader", "(", "nil", ")", "\n", "}", "\n", "xr", ".", "idx", ".", "Reset", "(", ")", "\n\n", "// Read entire index.", "var", "backSize", ",", "footSize", "int64", "\n", "if", "backSize", ",", "footSize", ",", "xr", ".", "err", "=", "xr", ".", "decodeFooter", "(", ")", ";", "xr", ".", "err", "!=", "nil", "{", "return", "xr", ".", "err", "\n", "}", "\n", "if", "xr", ".", "err", "=", "xr", ".", "decodeIndexes", "(", "backSize", ")", ";", "xr", ".", "err", "!=", "nil", "{", "return", "xr", ".", "err", "\n", "}", "\n", "if", "!", "xr", ".", "idx", ".", "AppendRecord", "(", "footSize", ",", "0", ",", "footerType", ")", "{", "xr", ".", "err", "=", "errCorrupted", "\n", "return", "xr", ".", "err", "\n", "}", "\n\n", "// Setup initial chunk reader.", "_", ",", "xr", ".", "err", "=", "xr", ".", "Seek", "(", "0", ",", "io", ".", "SeekStart", ")", "\n", "return", "xr", ".", "err", "\n", "}" ]
// Reset discards the Reader's state and makes it equivalent to the result // of a call to NewReader, but reading from rd instead. This method may return // an error if it is unable to parse the index. // // This is used to reduce memory allocations.
[ "Reset", "discards", "the", "Reader", "s", "state", "and", "makes", "it", "equivalent", "to", "the", "result", "of", "a", "call", "to", "NewReader", "but", "reading", "from", "rd", "instead", ".", "This", "method", "may", "return", "an", "error", "if", "it", "is", "unable", "to", "parse", "the", "index", ".", "This", "is", "used", "to", "reduce", "memory", "allocations", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/xflate/reader.go#L124-L157
9,669
dsnet/compress
xflate/reader.go
Read
func (xr *Reader) Read(buf []byte) (int, error) { if xr.err != nil { return 0, xr.err } // Discard some data to reach the expected raw offset. if xr.discard > 0 { var n int64 xr.lr = io.LimitedReader{R: xr.zr, N: xr.discard} n, xr.err = io.Copy(ioutil.Discard, &xr.lr) if xr.err != nil { return 0, xr.err } if n != xr.discard { xr.err = errCorrupted return 0, xr.err } xr.discard = 0 } var cnt int for cnt == 0 && xr.err == nil { cnt, xr.err = xr.zr.Read(buf) xr.offset += int64(cnt) if xr.err == io.EOF { xr.err = nil // Clear io.EOF temporarily // Verify that the compressed section ends with an empty raw block. if xr.chk.typ == deflateType && xr.cr.sync != 0x0000ffff { xr.err = errCorrupted break } // Verify that the compressed and raw sizes match. if xr.chk.typ != footerType { xr.chk.csize += int64(len(endBlock)) // Side of effect of using chunkReader } if xr.chk.csize != xr.zr.InputOffset || xr.chk.rsize != xr.zr.OutputOffset { xr.err = errCorrupted break } // Seek to next chunk. if _, xr.err = xr.Seek(xr.offset, io.SeekStart); xr.err != nil { break } if xr.chk.typ == unknownType { xr.err = io.EOF } } } return cnt, xr.err }
go
func (xr *Reader) Read(buf []byte) (int, error) { if xr.err != nil { return 0, xr.err } // Discard some data to reach the expected raw offset. if xr.discard > 0 { var n int64 xr.lr = io.LimitedReader{R: xr.zr, N: xr.discard} n, xr.err = io.Copy(ioutil.Discard, &xr.lr) if xr.err != nil { return 0, xr.err } if n != xr.discard { xr.err = errCorrupted return 0, xr.err } xr.discard = 0 } var cnt int for cnt == 0 && xr.err == nil { cnt, xr.err = xr.zr.Read(buf) xr.offset += int64(cnt) if xr.err == io.EOF { xr.err = nil // Clear io.EOF temporarily // Verify that the compressed section ends with an empty raw block. if xr.chk.typ == deflateType && xr.cr.sync != 0x0000ffff { xr.err = errCorrupted break } // Verify that the compressed and raw sizes match. if xr.chk.typ != footerType { xr.chk.csize += int64(len(endBlock)) // Side of effect of using chunkReader } if xr.chk.csize != xr.zr.InputOffset || xr.chk.rsize != xr.zr.OutputOffset { xr.err = errCorrupted break } // Seek to next chunk. if _, xr.err = xr.Seek(xr.offset, io.SeekStart); xr.err != nil { break } if xr.chk.typ == unknownType { xr.err = io.EOF } } } return cnt, xr.err }
[ "func", "(", "xr", "*", "Reader", ")", "Read", "(", "buf", "[", "]", "byte", ")", "(", "int", ",", "error", ")", "{", "if", "xr", ".", "err", "!=", "nil", "{", "return", "0", ",", "xr", ".", "err", "\n", "}", "\n\n", "// Discard some data to reach the expected raw offset.", "if", "xr", ".", "discard", ">", "0", "{", "var", "n", "int64", "\n", "xr", ".", "lr", "=", "io", ".", "LimitedReader", "{", "R", ":", "xr", ".", "zr", ",", "N", ":", "xr", ".", "discard", "}", "\n", "n", ",", "xr", ".", "err", "=", "io", ".", "Copy", "(", "ioutil", ".", "Discard", ",", "&", "xr", ".", "lr", ")", "\n", "if", "xr", ".", "err", "!=", "nil", "{", "return", "0", ",", "xr", ".", "err", "\n", "}", "\n", "if", "n", "!=", "xr", ".", "discard", "{", "xr", ".", "err", "=", "errCorrupted", "\n", "return", "0", ",", "xr", ".", "err", "\n", "}", "\n", "xr", ".", "discard", "=", "0", "\n", "}", "\n\n", "var", "cnt", "int", "\n", "for", "cnt", "==", "0", "&&", "xr", ".", "err", "==", "nil", "{", "cnt", ",", "xr", ".", "err", "=", "xr", ".", "zr", ".", "Read", "(", "buf", ")", "\n", "xr", ".", "offset", "+=", "int64", "(", "cnt", ")", "\n", "if", "xr", ".", "err", "==", "io", ".", "EOF", "{", "xr", ".", "err", "=", "nil", "// Clear io.EOF temporarily", "\n\n", "// Verify that the compressed section ends with an empty raw block.", "if", "xr", ".", "chk", ".", "typ", "==", "deflateType", "&&", "xr", ".", "cr", ".", "sync", "!=", "0x0000ffff", "{", "xr", ".", "err", "=", "errCorrupted", "\n", "break", "\n", "}", "\n\n", "// Verify that the compressed and raw sizes match.", "if", "xr", ".", "chk", ".", "typ", "!=", "footerType", "{", "xr", ".", "chk", ".", "csize", "+=", "int64", "(", "len", "(", "endBlock", ")", ")", "// Side of effect of using chunkReader", "\n", "}", "\n", "if", "xr", ".", "chk", ".", "csize", "!=", "xr", ".", "zr", ".", "InputOffset", "||", "xr", ".", "chk", ".", "rsize", "!=", "xr", ".", "zr", ".", "OutputOffset", "{", "xr", ".", "err", "=", "errCorrupted", "\n", "break", "\n", "}", "\n\n", "// Seek to next chunk.", "if", "_", ",", "xr", ".", "err", "=", "xr", ".", "Seek", "(", "xr", ".", "offset", ",", "io", ".", "SeekStart", ")", ";", "xr", ".", "err", "!=", "nil", "{", "break", "\n", "}", "\n", "if", "xr", ".", "chk", ".", "typ", "==", "unknownType", "{", "xr", ".", "err", "=", "io", ".", "EOF", "\n", "}", "\n", "}", "\n", "}", "\n\n", "return", "cnt", ",", "xr", ".", "err", "\n", "}" ]
// Read reads decompressed data from the underlying io.Reader. // This method automatically proceeds to the next chunk when the current one // has been fully read.
[ "Read", "reads", "decompressed", "data", "from", "the", "underlying", "io", ".", "Reader", ".", "This", "method", "automatically", "proceeds", "to", "the", "next", "chunk", "when", "the", "current", "one", "has", "been", "fully", "read", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/xflate/reader.go#L162-L215
9,670
dsnet/compress
xflate/reader.go
Seek
func (xr *Reader) Seek(offset int64, whence int) (int64, error) { if xr.err != nil && xr.err != io.EOF { return 0, xr.err } // Determine which position to seek to. var pos int64 end := xr.idx.LastRecord().RawOffset switch whence { case io.SeekStart: pos = offset case io.SeekCurrent: pos = xr.offset + offset case io.SeekEnd: pos = end + offset default: return 0, errorf(errors.Invalid, "invalid whence: %d", whence) } if pos < 0 { return 0, errorf(errors.Invalid, "negative position: %d", pos) } // As an optimization if the new position is within the current chunk, // then just adjust the discard value. discard := pos - xr.offset remain := xr.chk.rsize - xr.zr.OutputOffset if discard > 0 && remain > 0 && discard < remain { xr.offset, xr.discard = pos, discard return pos, nil } // Query the index for the chunk to start decoding from. // Attempt to use the subsequent record before resorting to binary search. prev, curr := xr.idx.GetRecords(xr.ri) if !(prev.RawOffset <= pos && pos <= curr.RawOffset) { xr.ri = xr.idx.Search(pos) prev, curr = xr.idx.GetRecords(xr.ri) } xr.ri++ if xr.ri > len(xr.idx.Records) { xr.ri = len(xr.idx.Records) } // Setup a chunk reader at the given position. xr.chk = chunk{ csize: curr.CompOffset - prev.CompOffset, rsize: curr.RawOffset - prev.RawOffset, typ: curr.Type, } xr.offset, xr.discard = pos, pos-prev.RawOffset if pos > end { // In case pos is really large, only discard data that actually exists. xr.discard = end - prev.RawOffset } _, xr.err = xr.rd.Seek(prev.CompOffset, io.SeekStart) xr.cr.Reset(xr.rd, xr.chk.csize) xr.zr.Reset(&xr.cr) return pos, xr.err }
go
func (xr *Reader) Seek(offset int64, whence int) (int64, error) { if xr.err != nil && xr.err != io.EOF { return 0, xr.err } // Determine which position to seek to. var pos int64 end := xr.idx.LastRecord().RawOffset switch whence { case io.SeekStart: pos = offset case io.SeekCurrent: pos = xr.offset + offset case io.SeekEnd: pos = end + offset default: return 0, errorf(errors.Invalid, "invalid whence: %d", whence) } if pos < 0 { return 0, errorf(errors.Invalid, "negative position: %d", pos) } // As an optimization if the new position is within the current chunk, // then just adjust the discard value. discard := pos - xr.offset remain := xr.chk.rsize - xr.zr.OutputOffset if discard > 0 && remain > 0 && discard < remain { xr.offset, xr.discard = pos, discard return pos, nil } // Query the index for the chunk to start decoding from. // Attempt to use the subsequent record before resorting to binary search. prev, curr := xr.idx.GetRecords(xr.ri) if !(prev.RawOffset <= pos && pos <= curr.RawOffset) { xr.ri = xr.idx.Search(pos) prev, curr = xr.idx.GetRecords(xr.ri) } xr.ri++ if xr.ri > len(xr.idx.Records) { xr.ri = len(xr.idx.Records) } // Setup a chunk reader at the given position. xr.chk = chunk{ csize: curr.CompOffset - prev.CompOffset, rsize: curr.RawOffset - prev.RawOffset, typ: curr.Type, } xr.offset, xr.discard = pos, pos-prev.RawOffset if pos > end { // In case pos is really large, only discard data that actually exists. xr.discard = end - prev.RawOffset } _, xr.err = xr.rd.Seek(prev.CompOffset, io.SeekStart) xr.cr.Reset(xr.rd, xr.chk.csize) xr.zr.Reset(&xr.cr) return pos, xr.err }
[ "func", "(", "xr", "*", "Reader", ")", "Seek", "(", "offset", "int64", ",", "whence", "int", ")", "(", "int64", ",", "error", ")", "{", "if", "xr", ".", "err", "!=", "nil", "&&", "xr", ".", "err", "!=", "io", ".", "EOF", "{", "return", "0", ",", "xr", ".", "err", "\n", "}", "\n\n", "// Determine which position to seek to.", "var", "pos", "int64", "\n", "end", ":=", "xr", ".", "idx", ".", "LastRecord", "(", ")", ".", "RawOffset", "\n", "switch", "whence", "{", "case", "io", ".", "SeekStart", ":", "pos", "=", "offset", "\n", "case", "io", ".", "SeekCurrent", ":", "pos", "=", "xr", ".", "offset", "+", "offset", "\n", "case", "io", ".", "SeekEnd", ":", "pos", "=", "end", "+", "offset", "\n", "default", ":", "return", "0", ",", "errorf", "(", "errors", ".", "Invalid", ",", "\"", "\"", ",", "whence", ")", "\n", "}", "\n", "if", "pos", "<", "0", "{", "return", "0", ",", "errorf", "(", "errors", ".", "Invalid", ",", "\"", "\"", ",", "pos", ")", "\n", "}", "\n\n", "// As an optimization if the new position is within the current chunk,", "// then just adjust the discard value.", "discard", ":=", "pos", "-", "xr", ".", "offset", "\n", "remain", ":=", "xr", ".", "chk", ".", "rsize", "-", "xr", ".", "zr", ".", "OutputOffset", "\n", "if", "discard", ">", "0", "&&", "remain", ">", "0", "&&", "discard", "<", "remain", "{", "xr", ".", "offset", ",", "xr", ".", "discard", "=", "pos", ",", "discard", "\n", "return", "pos", ",", "nil", "\n", "}", "\n\n", "// Query the index for the chunk to start decoding from.", "// Attempt to use the subsequent record before resorting to binary search.", "prev", ",", "curr", ":=", "xr", ".", "idx", ".", "GetRecords", "(", "xr", ".", "ri", ")", "\n", "if", "!", "(", "prev", ".", "RawOffset", "<=", "pos", "&&", "pos", "<=", "curr", ".", "RawOffset", ")", "{", "xr", ".", "ri", "=", "xr", ".", "idx", ".", "Search", "(", "pos", ")", "\n", "prev", ",", "curr", "=", "xr", ".", "idx", ".", "GetRecords", "(", "xr", ".", "ri", ")", "\n", "}", "\n", "xr", ".", "ri", "++", "\n", "if", "xr", ".", "ri", ">", "len", "(", "xr", ".", "idx", ".", "Records", ")", "{", "xr", ".", "ri", "=", "len", "(", "xr", ".", "idx", ".", "Records", ")", "\n", "}", "\n\n", "// Setup a chunk reader at the given position.", "xr", ".", "chk", "=", "chunk", "{", "csize", ":", "curr", ".", "CompOffset", "-", "prev", ".", "CompOffset", ",", "rsize", ":", "curr", ".", "RawOffset", "-", "prev", ".", "RawOffset", ",", "typ", ":", "curr", ".", "Type", ",", "}", "\n", "xr", ".", "offset", ",", "xr", ".", "discard", "=", "pos", ",", "pos", "-", "prev", ".", "RawOffset", "\n", "if", "pos", ">", "end", "{", "// In case pos is really large, only discard data that actually exists.", "xr", ".", "discard", "=", "end", "-", "prev", ".", "RawOffset", "\n", "}", "\n", "_", ",", "xr", ".", "err", "=", "xr", ".", "rd", ".", "Seek", "(", "prev", ".", "CompOffset", ",", "io", ".", "SeekStart", ")", "\n", "xr", ".", "cr", ".", "Reset", "(", "xr", ".", "rd", ",", "xr", ".", "chk", ".", "csize", ")", "\n", "xr", ".", "zr", ".", "Reset", "(", "&", "xr", ".", "cr", ")", "\n", "return", "pos", ",", "xr", ".", "err", "\n", "}" ]
// Seek sets the offset for the next Read operation, interpreted according to // the whence value provided. It is permitted to seek to offsets in the middle // of a compressed chunk. The next call to Read will automatically discard some // number of bytes before returning the requested data.
[ "Seek", "sets", "the", "offset", "for", "the", "next", "Read", "operation", "interpreted", "according", "to", "the", "whence", "value", "provided", ".", "It", "is", "permitted", "to", "seek", "to", "offsets", "in", "the", "middle", "of", "a", "compressed", "chunk", ".", "The", "next", "call", "to", "Read", "will", "automatically", "discard", "some", "number", "of", "bytes", "before", "returning", "the", "requested", "data", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/xflate/reader.go#L221-L279
9,671
dsnet/compress
xflate/reader.go
Close
func (xr *Reader) Close() error { if xr.err == errClosed { return nil } if xr.err != nil && xr.err != io.EOF { return xr.err } xr.err = errClosed return nil }
go
func (xr *Reader) Close() error { if xr.err == errClosed { return nil } if xr.err != nil && xr.err != io.EOF { return xr.err } xr.err = errClosed return nil }
[ "func", "(", "xr", "*", "Reader", ")", "Close", "(", ")", "error", "{", "if", "xr", ".", "err", "==", "errClosed", "{", "return", "nil", "\n", "}", "\n", "if", "xr", ".", "err", "!=", "nil", "&&", "xr", ".", "err", "!=", "io", ".", "EOF", "{", "return", "xr", ".", "err", "\n", "}", "\n", "xr", ".", "err", "=", "errClosed", "\n", "return", "nil", "\n", "}" ]
// Close ends the XFLATE stream.
[ "Close", "ends", "the", "XFLATE", "stream", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/xflate/reader.go#L282-L291
9,672
dsnet/compress
xflate/reader.go
decodeIndexes
func (xr *Reader) decodeIndexes(backSize int64) error { pos, err := xr.rd.Seek(0, io.SeekCurrent) if err != nil { return err } // Read all indexes. var compSize int64 xr.idxs = xr.idxs[:0] for { // Seek backwards past index and compressed blocks. newPos := pos - (backSize + compSize) if newPos < 0 || newPos > pos { return errCorrupted // Integer overflow for new seek position } if pos, err = xr.rd.Seek(newPos, io.SeekStart); err != nil { return err } if backSize == 0 { break } // Read the index. if cap(xr.idxs) > len(xr.idxs) { xr.idxs = xr.idxs[:len(xr.idxs)+1] } else { xr.idxs = append(xr.idxs, index{}) } idx := &xr.idxs[len(xr.idxs)-1] idx.Reset() idx.IndexSize = backSize if err = xr.decodeIndex(idx); err != nil { return err } backSize, compSize = idx.BackSize, idx.LastRecord().CompOffset } if pos != 0 { return errCorrupted } // Compact all indexes into one. for i := len(xr.idxs) - 1; i >= 0; i-- { idx := xr.idxs[i] if !xr.idx.AppendIndex(&idx) { return errCorrupted } if !xr.idx.AppendRecord(idx.IndexSize, 0, indexType) { return errCorrupted } } return nil }
go
func (xr *Reader) decodeIndexes(backSize int64) error { pos, err := xr.rd.Seek(0, io.SeekCurrent) if err != nil { return err } // Read all indexes. var compSize int64 xr.idxs = xr.idxs[:0] for { // Seek backwards past index and compressed blocks. newPos := pos - (backSize + compSize) if newPos < 0 || newPos > pos { return errCorrupted // Integer overflow for new seek position } if pos, err = xr.rd.Seek(newPos, io.SeekStart); err != nil { return err } if backSize == 0 { break } // Read the index. if cap(xr.idxs) > len(xr.idxs) { xr.idxs = xr.idxs[:len(xr.idxs)+1] } else { xr.idxs = append(xr.idxs, index{}) } idx := &xr.idxs[len(xr.idxs)-1] idx.Reset() idx.IndexSize = backSize if err = xr.decodeIndex(idx); err != nil { return err } backSize, compSize = idx.BackSize, idx.LastRecord().CompOffset } if pos != 0 { return errCorrupted } // Compact all indexes into one. for i := len(xr.idxs) - 1; i >= 0; i-- { idx := xr.idxs[i] if !xr.idx.AppendIndex(&idx) { return errCorrupted } if !xr.idx.AppendRecord(idx.IndexSize, 0, indexType) { return errCorrupted } } return nil }
[ "func", "(", "xr", "*", "Reader", ")", "decodeIndexes", "(", "backSize", "int64", ")", "error", "{", "pos", ",", "err", ":=", "xr", ".", "rd", ".", "Seek", "(", "0", ",", "io", ".", "SeekCurrent", ")", "\n", "if", "err", "!=", "nil", "{", "return", "err", "\n", "}", "\n\n", "// Read all indexes.", "var", "compSize", "int64", "\n", "xr", ".", "idxs", "=", "xr", ".", "idxs", "[", ":", "0", "]", "\n", "for", "{", "// Seek backwards past index and compressed blocks.", "newPos", ":=", "pos", "-", "(", "backSize", "+", "compSize", ")", "\n", "if", "newPos", "<", "0", "||", "newPos", ">", "pos", "{", "return", "errCorrupted", "// Integer overflow for new seek position", "\n", "}", "\n", "if", "pos", ",", "err", "=", "xr", ".", "rd", ".", "Seek", "(", "newPos", ",", "io", ".", "SeekStart", ")", ";", "err", "!=", "nil", "{", "return", "err", "\n", "}", "\n", "if", "backSize", "==", "0", "{", "break", "\n", "}", "\n\n", "// Read the index.", "if", "cap", "(", "xr", ".", "idxs", ")", ">", "len", "(", "xr", ".", "idxs", ")", "{", "xr", ".", "idxs", "=", "xr", ".", "idxs", "[", ":", "len", "(", "xr", ".", "idxs", ")", "+", "1", "]", "\n", "}", "else", "{", "xr", ".", "idxs", "=", "append", "(", "xr", ".", "idxs", ",", "index", "{", "}", ")", "\n", "}", "\n", "idx", ":=", "&", "xr", ".", "idxs", "[", "len", "(", "xr", ".", "idxs", ")", "-", "1", "]", "\n", "idx", ".", "Reset", "(", ")", "\n", "idx", ".", "IndexSize", "=", "backSize", "\n", "if", "err", "=", "xr", ".", "decodeIndex", "(", "idx", ")", ";", "err", "!=", "nil", "{", "return", "err", "\n", "}", "\n", "backSize", ",", "compSize", "=", "idx", ".", "BackSize", ",", "idx", ".", "LastRecord", "(", ")", ".", "CompOffset", "\n", "}", "\n", "if", "pos", "!=", "0", "{", "return", "errCorrupted", "\n", "}", "\n\n", "// Compact all indexes into one.", "for", "i", ":=", "len", "(", "xr", ".", "idxs", ")", "-", "1", ";", "i", ">=", "0", ";", "i", "--", "{", "idx", ":=", "xr", ".", "idxs", "[", "i", "]", "\n", "if", "!", "xr", ".", "idx", ".", "AppendIndex", "(", "&", "idx", ")", "{", "return", "errCorrupted", "\n", "}", "\n", "if", "!", "xr", ".", "idx", ".", "AppendRecord", "(", "idx", ".", "IndexSize", ",", "0", ",", "indexType", ")", "{", "return", "errCorrupted", "\n", "}", "\n", "}", "\n", "return", "nil", "\n", "}" ]
// decodeIndexes iteratively decodes all of the indexes in the XFLATE stream. // Even if the index is fragmented in the source stream, this method will merge // all of the index fragments into a single index table.
[ "decodeIndexes", "iteratively", "decodes", "all", "of", "the", "indexes", "in", "the", "XFLATE", "stream", ".", "Even", "if", "the", "index", "is", "fragmented", "in", "the", "source", "stream", "this", "method", "will", "merge", "all", "of", "the", "index", "fragments", "into", "a", "single", "index", "table", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/xflate/reader.go#L296-L347
9,673
dsnet/compress
xflate/reader.go
decodeIndex
func (xr *Reader) decodeIndex(idx *index) error { // Helper function to read VLIs. var errVLI error readVLI := func() int64 { x, n := binary.Uvarint(xr.bw.Bytes()) if n <= 0 || x > math.MaxInt64 { errVLI = errCorrupted return 0 } xr.bw.Next(n) return int64(x) } // Read the index and restore the underlying reader offset. xr.br.Reset() xr.lr = io.LimitedReader{R: xr.rd, N: idx.IndexSize} n, err := io.Copy(&xr.br, &xr.lr) if err != nil { return err } if _, err := xr.rd.Seek(-n, io.SeekCurrent); err != nil { return err } // Parse the index. var crc uint32 xr.chunks = xr.chunks[:0] xr.bw.Reset() xr.mr.Reset(&xr.br) if _, err := io.Copy(&xr.bw, &xr.mr); err != nil { return errWrap(err) } if xr.bw.Len() > 4 { crc = crc32.ChecksumIEEE(xr.bw.Bytes()[:xr.bw.Len()-4]) } idx.BackSize = readVLI() numRecs := readVLI() totalCompSize := readVLI() totalRawSize := readVLI() if errVLI != nil { return errVLI } for i := int64(0); i < numRecs; i++ { xr.chunks = append(xr.chunks, chunk{readVLI(), readVLI(), 0}) } if xr.bw.Len() != 4 || binary.LittleEndian.Uint32(xr.bw.Bytes()) != crc { return errCorrupted } if xr.mr.FinalMode != meta.FinalMeta { return errCorrupted } if xr.mr.InputOffset != idx.IndexSize { return errCorrupted } // Convert individual index sizes to be absolute offsets. for _, chk := range xr.chunks { if chk.csize <= 4 { return errCorrupted // Every chunk has a sync marker } if !idx.AppendRecord(chk.csize, chk.rsize, deflateType) { return errCorrupted } } lastRec := idx.LastRecord() if lastRec.CompOffset != totalCompSize || lastRec.RawOffset != totalRawSize { return errCorrupted } return nil }
go
func (xr *Reader) decodeIndex(idx *index) error { // Helper function to read VLIs. var errVLI error readVLI := func() int64 { x, n := binary.Uvarint(xr.bw.Bytes()) if n <= 0 || x > math.MaxInt64 { errVLI = errCorrupted return 0 } xr.bw.Next(n) return int64(x) } // Read the index and restore the underlying reader offset. xr.br.Reset() xr.lr = io.LimitedReader{R: xr.rd, N: idx.IndexSize} n, err := io.Copy(&xr.br, &xr.lr) if err != nil { return err } if _, err := xr.rd.Seek(-n, io.SeekCurrent); err != nil { return err } // Parse the index. var crc uint32 xr.chunks = xr.chunks[:0] xr.bw.Reset() xr.mr.Reset(&xr.br) if _, err := io.Copy(&xr.bw, &xr.mr); err != nil { return errWrap(err) } if xr.bw.Len() > 4 { crc = crc32.ChecksumIEEE(xr.bw.Bytes()[:xr.bw.Len()-4]) } idx.BackSize = readVLI() numRecs := readVLI() totalCompSize := readVLI() totalRawSize := readVLI() if errVLI != nil { return errVLI } for i := int64(0); i < numRecs; i++ { xr.chunks = append(xr.chunks, chunk{readVLI(), readVLI(), 0}) } if xr.bw.Len() != 4 || binary.LittleEndian.Uint32(xr.bw.Bytes()) != crc { return errCorrupted } if xr.mr.FinalMode != meta.FinalMeta { return errCorrupted } if xr.mr.InputOffset != idx.IndexSize { return errCorrupted } // Convert individual index sizes to be absolute offsets. for _, chk := range xr.chunks { if chk.csize <= 4 { return errCorrupted // Every chunk has a sync marker } if !idx.AppendRecord(chk.csize, chk.rsize, deflateType) { return errCorrupted } } lastRec := idx.LastRecord() if lastRec.CompOffset != totalCompSize || lastRec.RawOffset != totalRawSize { return errCorrupted } return nil }
[ "func", "(", "xr", "*", "Reader", ")", "decodeIndex", "(", "idx", "*", "index", ")", "error", "{", "// Helper function to read VLIs.", "var", "errVLI", "error", "\n", "readVLI", ":=", "func", "(", ")", "int64", "{", "x", ",", "n", ":=", "binary", ".", "Uvarint", "(", "xr", ".", "bw", ".", "Bytes", "(", ")", ")", "\n", "if", "n", "<=", "0", "||", "x", ">", "math", ".", "MaxInt64", "{", "errVLI", "=", "errCorrupted", "\n", "return", "0", "\n", "}", "\n", "xr", ".", "bw", ".", "Next", "(", "n", ")", "\n", "return", "int64", "(", "x", ")", "\n", "}", "\n\n", "// Read the index and restore the underlying reader offset.", "xr", ".", "br", ".", "Reset", "(", ")", "\n", "xr", ".", "lr", "=", "io", ".", "LimitedReader", "{", "R", ":", "xr", ".", "rd", ",", "N", ":", "idx", ".", "IndexSize", "}", "\n", "n", ",", "err", ":=", "io", ".", "Copy", "(", "&", "xr", ".", "br", ",", "&", "xr", ".", "lr", ")", "\n", "if", "err", "!=", "nil", "{", "return", "err", "\n", "}", "\n", "if", "_", ",", "err", ":=", "xr", ".", "rd", ".", "Seek", "(", "-", "n", ",", "io", ".", "SeekCurrent", ")", ";", "err", "!=", "nil", "{", "return", "err", "\n", "}", "\n\n", "// Parse the index.", "var", "crc", "uint32", "\n", "xr", ".", "chunks", "=", "xr", ".", "chunks", "[", ":", "0", "]", "\n", "xr", ".", "bw", ".", "Reset", "(", ")", "\n", "xr", ".", "mr", ".", "Reset", "(", "&", "xr", ".", "br", ")", "\n", "if", "_", ",", "err", ":=", "io", ".", "Copy", "(", "&", "xr", ".", "bw", ",", "&", "xr", ".", "mr", ")", ";", "err", "!=", "nil", "{", "return", "errWrap", "(", "err", ")", "\n", "}", "\n", "if", "xr", ".", "bw", ".", "Len", "(", ")", ">", "4", "{", "crc", "=", "crc32", ".", "ChecksumIEEE", "(", "xr", ".", "bw", ".", "Bytes", "(", ")", "[", ":", "xr", ".", "bw", ".", "Len", "(", ")", "-", "4", "]", ")", "\n", "}", "\n", "idx", ".", "BackSize", "=", "readVLI", "(", ")", "\n", "numRecs", ":=", "readVLI", "(", ")", "\n", "totalCompSize", ":=", "readVLI", "(", ")", "\n", "totalRawSize", ":=", "readVLI", "(", ")", "\n", "if", "errVLI", "!=", "nil", "{", "return", "errVLI", "\n", "}", "\n\n", "for", "i", ":=", "int64", "(", "0", ")", ";", "i", "<", "numRecs", ";", "i", "++", "{", "xr", ".", "chunks", "=", "append", "(", "xr", ".", "chunks", ",", "chunk", "{", "readVLI", "(", ")", ",", "readVLI", "(", ")", ",", "0", "}", ")", "\n", "}", "\n", "if", "xr", ".", "bw", ".", "Len", "(", ")", "!=", "4", "||", "binary", ".", "LittleEndian", ".", "Uint32", "(", "xr", ".", "bw", ".", "Bytes", "(", ")", ")", "!=", "crc", "{", "return", "errCorrupted", "\n", "}", "\n", "if", "xr", ".", "mr", ".", "FinalMode", "!=", "meta", ".", "FinalMeta", "{", "return", "errCorrupted", "\n", "}", "\n", "if", "xr", ".", "mr", ".", "InputOffset", "!=", "idx", ".", "IndexSize", "{", "return", "errCorrupted", "\n", "}", "\n\n", "// Convert individual index sizes to be absolute offsets.", "for", "_", ",", "chk", ":=", "range", "xr", ".", "chunks", "{", "if", "chk", ".", "csize", "<=", "4", "{", "return", "errCorrupted", "// Every chunk has a sync marker", "\n", "}", "\n", "if", "!", "idx", ".", "AppendRecord", "(", "chk", ".", "csize", ",", "chk", ".", "rsize", ",", "deflateType", ")", "{", "return", "errCorrupted", "\n", "}", "\n", "}", "\n", "lastRec", ":=", "idx", ".", "LastRecord", "(", ")", "\n", "if", "lastRec", ".", "CompOffset", "!=", "totalCompSize", "||", "lastRec", ".", "RawOffset", "!=", "totalRawSize", "{", "return", "errCorrupted", "\n", "}", "\n", "return", "nil", "\n", "}" ]
// decodeIndex decodes the index from a meta encoded stream. // The current offset must be set to the start of the encoded index and // index.IndexSize must be populated. If successful, the index.Records and // index.BackSize fields will be populated. This method will attempt to reset // the read offset to the start of the index.
[ "decodeIndex", "decodes", "the", "index", "from", "a", "meta", "encoded", "stream", ".", "The", "current", "offset", "must", "be", "set", "to", "the", "start", "of", "the", "encoded", "index", "and", "index", ".", "IndexSize", "must", "be", "populated", ".", "If", "successful", "the", "index", ".", "Records", "and", "index", ".", "BackSize", "fields", "will", "be", "populated", ".", "This", "method", "will", "attempt", "to", "reset", "the", "read", "offset", "to", "the", "start", "of", "the", "index", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/xflate/reader.go#L354-L424
9,674
dsnet/compress
xflate/reader.go
decodeFooter
func (xr *Reader) decodeFooter() (backSize, footSize int64, err error) { // Read the last few bytes of the stream. end, err := xr.rd.Seek(0, io.SeekEnd) if err != nil { return 0, 0, err } if end > meta.MaxEncBytes { end = meta.MaxEncBytes } if _, err := xr.rd.Seek(-end, io.SeekEnd); err != nil { return 0, 0, err } xr.br.Reset() if _, err := io.Copy(&xr.br, xr.rd); err != nil { return 0, 0, err } // Search for and read the meta block. idx := meta.ReverseSearch(xr.br.Bytes()) if idx < 0 { return 0, 0, errCorrupted } xr.br.Next(idx) // Skip data until magic marker xr.bw.Reset() xr.mr.Reset(&xr.br) if _, err := io.Copy(&xr.bw, &xr.mr); err != nil { return 0, 0, errWrap(err) } if xr.br.Len() != 0 || xr.mr.NumBlocks != 1 { return 0, 0, errCorrupted } if xr.mr.FinalMode != meta.FinalStream { return 0, 0, errCorrupted } if _, err := xr.rd.Seek(-xr.mr.InputOffset, io.SeekCurrent); err != nil { return 0, 0, err } // Parse the footer. bufRaw := xr.bw.Bytes() if len(bufRaw) < 3 || !bytes.Equal(bufRaw[:3], magic[:]) { return 0, 0, errCorrupted // Magic value mismatch } backSizeU64, cnt := binary.Uvarint(bufRaw[3:]) if cnt <= 0 { return 0, 0, errCorrupted // Integer overflow for VLI } if len(bufRaw[3+cnt:]) > 0 { return 0, 0, errCorrupted // Trailing unread bytes } return int64(backSizeU64), xr.mr.InputOffset, nil }
go
func (xr *Reader) decodeFooter() (backSize, footSize int64, err error) { // Read the last few bytes of the stream. end, err := xr.rd.Seek(0, io.SeekEnd) if err != nil { return 0, 0, err } if end > meta.MaxEncBytes { end = meta.MaxEncBytes } if _, err := xr.rd.Seek(-end, io.SeekEnd); err != nil { return 0, 0, err } xr.br.Reset() if _, err := io.Copy(&xr.br, xr.rd); err != nil { return 0, 0, err } // Search for and read the meta block. idx := meta.ReverseSearch(xr.br.Bytes()) if idx < 0 { return 0, 0, errCorrupted } xr.br.Next(idx) // Skip data until magic marker xr.bw.Reset() xr.mr.Reset(&xr.br) if _, err := io.Copy(&xr.bw, &xr.mr); err != nil { return 0, 0, errWrap(err) } if xr.br.Len() != 0 || xr.mr.NumBlocks != 1 { return 0, 0, errCorrupted } if xr.mr.FinalMode != meta.FinalStream { return 0, 0, errCorrupted } if _, err := xr.rd.Seek(-xr.mr.InputOffset, io.SeekCurrent); err != nil { return 0, 0, err } // Parse the footer. bufRaw := xr.bw.Bytes() if len(bufRaw) < 3 || !bytes.Equal(bufRaw[:3], magic[:]) { return 0, 0, errCorrupted // Magic value mismatch } backSizeU64, cnt := binary.Uvarint(bufRaw[3:]) if cnt <= 0 { return 0, 0, errCorrupted // Integer overflow for VLI } if len(bufRaw[3+cnt:]) > 0 { return 0, 0, errCorrupted // Trailing unread bytes } return int64(backSizeU64), xr.mr.InputOffset, nil }
[ "func", "(", "xr", "*", "Reader", ")", "decodeFooter", "(", ")", "(", "backSize", ",", "footSize", "int64", ",", "err", "error", ")", "{", "// Read the last few bytes of the stream.", "end", ",", "err", ":=", "xr", ".", "rd", ".", "Seek", "(", "0", ",", "io", ".", "SeekEnd", ")", "\n", "if", "err", "!=", "nil", "{", "return", "0", ",", "0", ",", "err", "\n", "}", "\n", "if", "end", ">", "meta", ".", "MaxEncBytes", "{", "end", "=", "meta", ".", "MaxEncBytes", "\n", "}", "\n", "if", "_", ",", "err", ":=", "xr", ".", "rd", ".", "Seek", "(", "-", "end", ",", "io", ".", "SeekEnd", ")", ";", "err", "!=", "nil", "{", "return", "0", ",", "0", ",", "err", "\n", "}", "\n\n", "xr", ".", "br", ".", "Reset", "(", ")", "\n", "if", "_", ",", "err", ":=", "io", ".", "Copy", "(", "&", "xr", ".", "br", ",", "xr", ".", "rd", ")", ";", "err", "!=", "nil", "{", "return", "0", ",", "0", ",", "err", "\n", "}", "\n\n", "// Search for and read the meta block.", "idx", ":=", "meta", ".", "ReverseSearch", "(", "xr", ".", "br", ".", "Bytes", "(", ")", ")", "\n", "if", "idx", "<", "0", "{", "return", "0", ",", "0", ",", "errCorrupted", "\n", "}", "\n", "xr", ".", "br", ".", "Next", "(", "idx", ")", "// Skip data until magic marker", "\n\n", "xr", ".", "bw", ".", "Reset", "(", ")", "\n", "xr", ".", "mr", ".", "Reset", "(", "&", "xr", ".", "br", ")", "\n", "if", "_", ",", "err", ":=", "io", ".", "Copy", "(", "&", "xr", ".", "bw", ",", "&", "xr", ".", "mr", ")", ";", "err", "!=", "nil", "{", "return", "0", ",", "0", ",", "errWrap", "(", "err", ")", "\n", "}", "\n", "if", "xr", ".", "br", ".", "Len", "(", ")", "!=", "0", "||", "xr", ".", "mr", ".", "NumBlocks", "!=", "1", "{", "return", "0", ",", "0", ",", "errCorrupted", "\n", "}", "\n", "if", "xr", ".", "mr", ".", "FinalMode", "!=", "meta", ".", "FinalStream", "{", "return", "0", ",", "0", ",", "errCorrupted", "\n", "}", "\n", "if", "_", ",", "err", ":=", "xr", ".", "rd", ".", "Seek", "(", "-", "xr", ".", "mr", ".", "InputOffset", ",", "io", ".", "SeekCurrent", ")", ";", "err", "!=", "nil", "{", "return", "0", ",", "0", ",", "err", "\n", "}", "\n\n", "// Parse the footer.", "bufRaw", ":=", "xr", ".", "bw", ".", "Bytes", "(", ")", "\n", "if", "len", "(", "bufRaw", ")", "<", "3", "||", "!", "bytes", ".", "Equal", "(", "bufRaw", "[", ":", "3", "]", ",", "magic", "[", ":", "]", ")", "{", "return", "0", ",", "0", ",", "errCorrupted", "// Magic value mismatch", "\n", "}", "\n", "backSizeU64", ",", "cnt", ":=", "binary", ".", "Uvarint", "(", "bufRaw", "[", "3", ":", "]", ")", "\n", "if", "cnt", "<=", "0", "{", "return", "0", ",", "0", ",", "errCorrupted", "// Integer overflow for VLI", "\n", "}", "\n", "if", "len", "(", "bufRaw", "[", "3", "+", "cnt", ":", "]", ")", ">", "0", "{", "return", "0", ",", "0", ",", "errCorrupted", "// Trailing unread bytes", "\n", "}", "\n", "return", "int64", "(", "backSizeU64", ")", ",", "xr", ".", "mr", ".", "InputOffset", ",", "nil", "\n", "}" ]
// decodeFooter seeks to the end of the stream, searches for the footer // and decodes it. If successful, it will return the backSize for the preceding // index and the size of the footer itself. This method will attempt to reset // the read offset to the start of the footer.
[ "decodeFooter", "seeks", "to", "the", "end", "of", "the", "stream", "searches", "for", "the", "footer", "and", "decodes", "it", ".", "If", "successful", "it", "will", "return", "the", "backSize", "for", "the", "preceding", "index", "and", "the", "size", "of", "the", "footer", "itself", ".", "This", "method", "will", "attempt", "to", "reset", "the", "read", "offset", "to", "the", "start", "of", "the", "footer", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/xflate/reader.go#L430-L483
9,675
dsnet/compress
xflate/index.go
AppendRecord
func (idx *index) AppendRecord(compSize, rawSize int64, typ int) bool { if rawSize < 0 || compSize < 0 { return false // Invalid size } lastRec := idx.LastRecord() rec := record{ CompOffset: lastRec.CompOffset + compSize, RawOffset: lastRec.RawOffset + rawSize, Type: typ, } if rec.CompOffset < lastRec.CompOffset || rec.RawOffset < lastRec.RawOffset { return false // Overflow detected } idx.Records = append(idx.Records, rec) return true }
go
func (idx *index) AppendRecord(compSize, rawSize int64, typ int) bool { if rawSize < 0 || compSize < 0 { return false // Invalid size } lastRec := idx.LastRecord() rec := record{ CompOffset: lastRec.CompOffset + compSize, RawOffset: lastRec.RawOffset + rawSize, Type: typ, } if rec.CompOffset < lastRec.CompOffset || rec.RawOffset < lastRec.RawOffset { return false // Overflow detected } idx.Records = append(idx.Records, rec) return true }
[ "func", "(", "idx", "*", "index", ")", "AppendRecord", "(", "compSize", ",", "rawSize", "int64", ",", "typ", "int", ")", "bool", "{", "if", "rawSize", "<", "0", "||", "compSize", "<", "0", "{", "return", "false", "// Invalid size", "\n", "}", "\n\n", "lastRec", ":=", "idx", ".", "LastRecord", "(", ")", "\n", "rec", ":=", "record", "{", "CompOffset", ":", "lastRec", ".", "CompOffset", "+", "compSize", ",", "RawOffset", ":", "lastRec", ".", "RawOffset", "+", "rawSize", ",", "Type", ":", "typ", ",", "}", "\n", "if", "rec", ".", "CompOffset", "<", "lastRec", ".", "CompOffset", "||", "rec", ".", "RawOffset", "<", "lastRec", ".", "RawOffset", "{", "return", "false", "// Overflow detected", "\n", "}", "\n", "idx", ".", "Records", "=", "append", "(", "idx", ".", "Records", ",", "rec", ")", "\n", "return", "true", "\n", "}" ]
// AppendRecord appends a new record to the end of the index and reports whether // the operation was successful or not.
[ "AppendRecord", "appends", "a", "new", "record", "to", "the", "end", "of", "the", "index", "and", "reports", "whether", "the", "operation", "was", "successful", "or", "not", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/xflate/index.go#L40-L56
9,676
dsnet/compress
xflate/index.go
AppendIndex
func (idx *index) AppendIndex(other *index) bool { var preRec record for i, rec := range other.Records { csize, rsize := rec.CompOffset-preRec.CompOffset, rec.RawOffset-preRec.RawOffset if !idx.AppendRecord(csize, rsize, rec.Type) { idx.Records = idx.Records[:len(idx.Records)-i] // Ensure atomic append return false } preRec = rec } return true }
go
func (idx *index) AppendIndex(other *index) bool { var preRec record for i, rec := range other.Records { csize, rsize := rec.CompOffset-preRec.CompOffset, rec.RawOffset-preRec.RawOffset if !idx.AppendRecord(csize, rsize, rec.Type) { idx.Records = idx.Records[:len(idx.Records)-i] // Ensure atomic append return false } preRec = rec } return true }
[ "func", "(", "idx", "*", "index", ")", "AppendIndex", "(", "other", "*", "index", ")", "bool", "{", "var", "preRec", "record", "\n", "for", "i", ",", "rec", ":=", "range", "other", ".", "Records", "{", "csize", ",", "rsize", ":=", "rec", ".", "CompOffset", "-", "preRec", ".", "CompOffset", ",", "rec", ".", "RawOffset", "-", "preRec", ".", "RawOffset", "\n", "if", "!", "idx", ".", "AppendRecord", "(", "csize", ",", "rsize", ",", "rec", ".", "Type", ")", "{", "idx", ".", "Records", "=", "idx", ".", "Records", "[", ":", "len", "(", "idx", ".", "Records", ")", "-", "i", "]", "// Ensure atomic append", "\n", "return", "false", "\n", "}", "\n", "preRec", "=", "rec", "\n", "}", "\n", "return", "true", "\n", "}" ]
// AppendIndex appends the contents of another index onto the current receiver // and reports whether the operation was successful or not.
[ "AppendIndex", "appends", "the", "contents", "of", "another", "index", "onto", "the", "current", "receiver", "and", "reports", "whether", "the", "operation", "was", "successful", "or", "not", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/xflate/index.go#L60-L71
9,677
dsnet/compress
xflate/index.go
GetRecords
func (idx *index) GetRecords(i int) (prev, curr record) { recs := idx.Records if i > len(recs) { i = len(recs) } if i-1 >= 0 && i-1 < len(recs) { prev = recs[i-1] } if i >= 0 && i < len(recs) { curr = recs[i] } else { curr = prev curr.Type = unknownType } return prev, curr }
go
func (idx *index) GetRecords(i int) (prev, curr record) { recs := idx.Records if i > len(recs) { i = len(recs) } if i-1 >= 0 && i-1 < len(recs) { prev = recs[i-1] } if i >= 0 && i < len(recs) { curr = recs[i] } else { curr = prev curr.Type = unknownType } return prev, curr }
[ "func", "(", "idx", "*", "index", ")", "GetRecords", "(", "i", "int", ")", "(", "prev", ",", "curr", "record", ")", "{", "recs", ":=", "idx", ".", "Records", "\n", "if", "i", ">", "len", "(", "recs", ")", "{", "i", "=", "len", "(", "recs", ")", "\n", "}", "\n", "if", "i", "-", "1", ">=", "0", "&&", "i", "-", "1", "<", "len", "(", "recs", ")", "{", "prev", "=", "recs", "[", "i", "-", "1", "]", "\n", "}", "\n", "if", "i", ">=", "0", "&&", "i", "<", "len", "(", "recs", ")", "{", "curr", "=", "recs", "[", "i", "]", "\n", "}", "else", "{", "curr", "=", "prev", "\n", "curr", ".", "Type", "=", "unknownType", "\n", "}", "\n", "return", "prev", ",", "curr", "\n", "}" ]
// GetRecords returns the previous and current records at the given position. // This method will automatically bind the search position within the bounds // of the index. Thus, this will return zero value records if the position is // too low, and the last record if the value is too high.
[ "GetRecords", "returns", "the", "previous", "and", "current", "records", "at", "the", "given", "position", ".", "This", "method", "will", "automatically", "bind", "the", "search", "position", "within", "the", "bounds", "of", "the", "index", ".", "Thus", "this", "will", "return", "zero", "value", "records", "if", "the", "position", "is", "too", "low", "and", "the", "last", "record", "if", "the", "value", "is", "too", "high", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/xflate/index.go#L116-L131
9,678
dsnet/compress
xflate/index.go
LastRecord
func (idx *index) LastRecord() record { var rec record if len(idx.Records) > 0 { rec = idx.Records[len(idx.Records)-1] } return rec }
go
func (idx *index) LastRecord() record { var rec record if len(idx.Records) > 0 { rec = idx.Records[len(idx.Records)-1] } return rec }
[ "func", "(", "idx", "*", "index", ")", "LastRecord", "(", ")", "record", "{", "var", "rec", "record", "\n", "if", "len", "(", "idx", ".", "Records", ")", ">", "0", "{", "rec", "=", "idx", ".", "Records", "[", "len", "(", "idx", ".", "Records", ")", "-", "1", "]", "\n", "}", "\n", "return", "rec", "\n", "}" ]
// LastRecord returns the last record if it exists, otherwise the zero value.
[ "LastRecord", "returns", "the", "last", "record", "if", "it", "exists", "otherwise", "the", "zero", "value", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/xflate/index.go#L134-L140
9,679
dsnet/compress
internal/prefix/wrap.go
update
func (r *stringReader) update() { pos, _ := r.Seek(0, io.SeekCurrent) if off := pos - r.pos; off >= 0 && off < int64(len(r.buf)) { r.buf, r.pos = r.buf[off:], pos } else { r.buf, r.pos = nil, pos } }
go
func (r *stringReader) update() { pos, _ := r.Seek(0, io.SeekCurrent) if off := pos - r.pos; off >= 0 && off < int64(len(r.buf)) { r.buf, r.pos = r.buf[off:], pos } else { r.buf, r.pos = nil, pos } }
[ "func", "(", "r", "*", "stringReader", ")", "update", "(", ")", "{", "pos", ",", "_", ":=", "r", ".", "Seek", "(", "0", ",", "io", ".", "SeekCurrent", ")", "\n", "if", "off", ":=", "pos", "-", "r", ".", "pos", ";", "off", ">=", "0", "&&", "off", "<", "int64", "(", "len", "(", "r", ".", "buf", ")", ")", "{", "r", ".", "buf", ",", "r", ".", "pos", "=", "r", ".", "buf", "[", "off", ":", "]", ",", "pos", "\n", "}", "else", "{", "r", ".", "buf", ",", "r", ".", "pos", "=", "nil", ",", "pos", "\n", "}", "\n", "}" ]
// update reslices the internal buffer to be consistent with the read offset.
[ "update", "reslices", "the", "internal", "buffer", "to", "be", "consistent", "with", "the", "read", "offset", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/internal/prefix/wrap.go#L139-L146
9,680
dsnet/compress
xflate/xflate_stats.go
computeRecords
func computeRecords(r io.Reader, lvl, chnkSize int) (strmRec indexRecord, chnkRecs []indexRecord) { var cw1, cw2 countWriter zw1, _ := flate.NewWriter(&cw1, lvl) // Streamed compressor zw2, _ := flate.NewWriter(&cw2, lvl) // Chunked compressor buf := make([]byte, chnkSize) for { // Read a full chunks worth of data. cnt, err := io.ReadFull(r, buf) strmRec.rawSize += int64(cnt) if err == io.EOF { break } // Write chunk to both compressors. if _, err := zw1.Write(buf[:cnt]); err != nil { log.Fatal(err) } if _, err := zw2.Write(buf[:cnt]); err != nil { log.Fatal(err) } // Flush the chunked compressor, append the record, and reset. if err := zw2.Flush(); err != nil { log.Fatal(err) } chnkRecs = append(chnkRecs, indexRecord{rawSize: int64(cnt), compSize: int64(cw2)}) cw2 = 0 zw2.Reset(&cw2) if err == io.ErrUnexpectedEOF { break } if err != nil { log.Fatal(err) } } // Flush the streamed compressor and record the compressed size. if err := zw1.Flush(); err != nil { log.Fatal(err) } strmRec.compSize = int64(cw1) return strmRec, chnkRecs }
go
func computeRecords(r io.Reader, lvl, chnkSize int) (strmRec indexRecord, chnkRecs []indexRecord) { var cw1, cw2 countWriter zw1, _ := flate.NewWriter(&cw1, lvl) // Streamed compressor zw2, _ := flate.NewWriter(&cw2, lvl) // Chunked compressor buf := make([]byte, chnkSize) for { // Read a full chunks worth of data. cnt, err := io.ReadFull(r, buf) strmRec.rawSize += int64(cnt) if err == io.EOF { break } // Write chunk to both compressors. if _, err := zw1.Write(buf[:cnt]); err != nil { log.Fatal(err) } if _, err := zw2.Write(buf[:cnt]); err != nil { log.Fatal(err) } // Flush the chunked compressor, append the record, and reset. if err := zw2.Flush(); err != nil { log.Fatal(err) } chnkRecs = append(chnkRecs, indexRecord{rawSize: int64(cnt), compSize: int64(cw2)}) cw2 = 0 zw2.Reset(&cw2) if err == io.ErrUnexpectedEOF { break } if err != nil { log.Fatal(err) } } // Flush the streamed compressor and record the compressed size. if err := zw1.Flush(); err != nil { log.Fatal(err) } strmRec.compSize = int64(cw1) return strmRec, chnkRecs }
[ "func", "computeRecords", "(", "r", "io", ".", "Reader", ",", "lvl", ",", "chnkSize", "int", ")", "(", "strmRec", "indexRecord", ",", "chnkRecs", "[", "]", "indexRecord", ")", "{", "var", "cw1", ",", "cw2", "countWriter", "\n", "zw1", ",", "_", ":=", "flate", ".", "NewWriter", "(", "&", "cw1", ",", "lvl", ")", "// Streamed compressor", "\n", "zw2", ",", "_", ":=", "flate", ".", "NewWriter", "(", "&", "cw2", ",", "lvl", ")", "// Chunked compressor", "\n", "buf", ":=", "make", "(", "[", "]", "byte", ",", "chnkSize", ")", "\n", "for", "{", "// Read a full chunks worth of data.", "cnt", ",", "err", ":=", "io", ".", "ReadFull", "(", "r", ",", "buf", ")", "\n", "strmRec", ".", "rawSize", "+=", "int64", "(", "cnt", ")", "\n", "if", "err", "==", "io", ".", "EOF", "{", "break", "\n", "}", "\n\n", "// Write chunk to both compressors.", "if", "_", ",", "err", ":=", "zw1", ".", "Write", "(", "buf", "[", ":", "cnt", "]", ")", ";", "err", "!=", "nil", "{", "log", ".", "Fatal", "(", "err", ")", "\n", "}", "\n", "if", "_", ",", "err", ":=", "zw2", ".", "Write", "(", "buf", "[", ":", "cnt", "]", ")", ";", "err", "!=", "nil", "{", "log", ".", "Fatal", "(", "err", ")", "\n", "}", "\n\n", "// Flush the chunked compressor, append the record, and reset.", "if", "err", ":=", "zw2", ".", "Flush", "(", ")", ";", "err", "!=", "nil", "{", "log", ".", "Fatal", "(", "err", ")", "\n", "}", "\n", "chnkRecs", "=", "append", "(", "chnkRecs", ",", "indexRecord", "{", "rawSize", ":", "int64", "(", "cnt", ")", ",", "compSize", ":", "int64", "(", "cw2", ")", "}", ")", "\n", "cw2", "=", "0", "\n", "zw2", ".", "Reset", "(", "&", "cw2", ")", "\n\n", "if", "err", "==", "io", ".", "ErrUnexpectedEOF", "{", "break", "\n", "}", "\n", "if", "err", "!=", "nil", "{", "log", ".", "Fatal", "(", "err", ")", "\n", "}", "\n", "}", "\n\n", "// Flush the streamed compressor and record the compressed size.", "if", "err", ":=", "zw1", ".", "Flush", "(", ")", ";", "err", "!=", "nil", "{", "log", ".", "Fatal", "(", "err", ")", "\n", "}", "\n", "strmRec", ".", "compSize", "=", "int64", "(", "cw1", ")", "\n", "return", "strmRec", ",", "chnkRecs", "\n", "}" ]
// computeRecords computes the records the raw input size and the compressed // output size. strmRec is a single record for when the input is compressed as // a single stream. chnkRecs is a list of records for when the input is // compressed individually as chunks.
[ "computeRecords", "computes", "the", "records", "the", "raw", "input", "size", "and", "the", "compressed", "output", "size", ".", "strmRec", "is", "a", "single", "record", "for", "when", "the", "input", "is", "compressed", "as", "a", "single", "stream", ".", "chnkRecs", "is", "a", "list", "of", "records", "for", "when", "the", "input", "is", "compressed", "individually", "as", "chunks", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/xflate/xflate_stats.go#L236-L279
9,681
dsnet/compress
xflate/xflate_stats.go
compress
func compress(b []byte, lvl int) []byte { var buf bytes.Buffer w, err := flate.NewWriter(&buf, lvl) if err != nil { log.Fatal(err) } if _, err := io.Copy(w, bytes.NewReader(b)); err != nil { log.Fatal(err) } if err := w.Close(); err != nil { log.Fatal(err) } return buf.Bytes() }
go
func compress(b []byte, lvl int) []byte { var buf bytes.Buffer w, err := flate.NewWriter(&buf, lvl) if err != nil { log.Fatal(err) } if _, err := io.Copy(w, bytes.NewReader(b)); err != nil { log.Fatal(err) } if err := w.Close(); err != nil { log.Fatal(err) } return buf.Bytes() }
[ "func", "compress", "(", "b", "[", "]", "byte", ",", "lvl", "int", ")", "[", "]", "byte", "{", "var", "buf", "bytes", ".", "Buffer", "\n", "w", ",", "err", ":=", "flate", ".", "NewWriter", "(", "&", "buf", ",", "lvl", ")", "\n", "if", "err", "!=", "nil", "{", "log", ".", "Fatal", "(", "err", ")", "\n", "}", "\n", "if", "_", ",", "err", ":=", "io", ".", "Copy", "(", "w", ",", "bytes", ".", "NewReader", "(", "b", ")", ")", ";", "err", "!=", "nil", "{", "log", ".", "Fatal", "(", "err", ")", "\n", "}", "\n", "if", "err", ":=", "w", ".", "Close", "(", ")", ";", "err", "!=", "nil", "{", "log", ".", "Fatal", "(", "err", ")", "\n", "}", "\n", "return", "buf", ".", "Bytes", "(", ")", "\n", "}" ]
// compress compresses the input buffer at the given level.
[ "compress", "compresses", "the", "input", "buffer", "at", "the", "given", "level", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/xflate/xflate_stats.go#L282-L295
9,682
dsnet/compress
xflate/xflate_stats.go
encode
func encode(b []byte) []byte { var buf bytes.Buffer mw := meta.NewWriter(&buf) mw.FinalMode = meta.FinalMeta if _, err := io.Copy(mw, bytes.NewReader(b)); err != nil { log.Fatal(err) } if err := mw.Close(); err != nil { log.Fatal(err) } return buf.Bytes() }
go
func encode(b []byte) []byte { var buf bytes.Buffer mw := meta.NewWriter(&buf) mw.FinalMode = meta.FinalMeta if _, err := io.Copy(mw, bytes.NewReader(b)); err != nil { log.Fatal(err) } if err := mw.Close(); err != nil { log.Fatal(err) } return buf.Bytes() }
[ "func", "encode", "(", "b", "[", "]", "byte", ")", "[", "]", "byte", "{", "var", "buf", "bytes", ".", "Buffer", "\n", "mw", ":=", "meta", ".", "NewWriter", "(", "&", "buf", ")", "\n", "mw", ".", "FinalMode", "=", "meta", ".", "FinalMeta", "\n", "if", "_", ",", "err", ":=", "io", ".", "Copy", "(", "mw", ",", "bytes", ".", "NewReader", "(", "b", ")", ")", ";", "err", "!=", "nil", "{", "log", ".", "Fatal", "(", "err", ")", "\n", "}", "\n", "if", "err", ":=", "mw", ".", "Close", "(", ")", ";", "err", "!=", "nil", "{", "log", ".", "Fatal", "(", "err", ")", "\n", "}", "\n", "return", "buf", ".", "Bytes", "(", ")", "\n", "}" ]
// encode encodes the input using XFLATE's meta encoding.
[ "encode", "encodes", "the", "input", "using", "XFLATE", "s", "meta", "encoding", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/xflate/xflate_stats.go#L298-L309
9,683
dsnet/compress
xflate/writer.go
NewWriter
func NewWriter(wr io.Writer, conf *WriterConfig) (*Writer, error) { var lvl int var nchk, nidx int64 if conf != nil { lvl = conf.Level switch { case conf.ChunkSize < 0: return nil, errorf(errors.Invalid, "invalid chunk size: %d", conf.ChunkSize) case conf.ChunkSize > 0: nchk = conf.ChunkSize } switch { case conf.IndexSize < 0: nidx = -1 case conf.IndexSize > 0: nidx = conf.IndexSize } } zw, err := newFlateWriter(wr, lvl) if err != nil { return nil, err } xw := &Writer{wr: wr, zw: zw, nchk: nchk, nidx: nidx} xw.Reset(wr) return xw, nil }
go
func NewWriter(wr io.Writer, conf *WriterConfig) (*Writer, error) { var lvl int var nchk, nidx int64 if conf != nil { lvl = conf.Level switch { case conf.ChunkSize < 0: return nil, errorf(errors.Invalid, "invalid chunk size: %d", conf.ChunkSize) case conf.ChunkSize > 0: nchk = conf.ChunkSize } switch { case conf.IndexSize < 0: nidx = -1 case conf.IndexSize > 0: nidx = conf.IndexSize } } zw, err := newFlateWriter(wr, lvl) if err != nil { return nil, err } xw := &Writer{wr: wr, zw: zw, nchk: nchk, nidx: nidx} xw.Reset(wr) return xw, nil }
[ "func", "NewWriter", "(", "wr", "io", ".", "Writer", ",", "conf", "*", "WriterConfig", ")", "(", "*", "Writer", ",", "error", ")", "{", "var", "lvl", "int", "\n", "var", "nchk", ",", "nidx", "int64", "\n", "if", "conf", "!=", "nil", "{", "lvl", "=", "conf", ".", "Level", "\n", "switch", "{", "case", "conf", ".", "ChunkSize", "<", "0", ":", "return", "nil", ",", "errorf", "(", "errors", ".", "Invalid", ",", "\"", "\"", ",", "conf", ".", "ChunkSize", ")", "\n", "case", "conf", ".", "ChunkSize", ">", "0", ":", "nchk", "=", "conf", ".", "ChunkSize", "\n", "}", "\n", "switch", "{", "case", "conf", ".", "IndexSize", "<", "0", ":", "nidx", "=", "-", "1", "\n", "case", "conf", ".", "IndexSize", ">", "0", ":", "nidx", "=", "conf", ".", "IndexSize", "\n", "}", "\n", "}", "\n\n", "zw", ",", "err", ":=", "newFlateWriter", "(", "wr", ",", "lvl", ")", "\n", "if", "err", "!=", "nil", "{", "return", "nil", ",", "err", "\n", "}", "\n", "xw", ":=", "&", "Writer", "{", "wr", ":", "wr", ",", "zw", ":", "zw", ",", "nchk", ":", "nchk", ",", "nidx", ":", "nidx", "}", "\n", "xw", ".", "Reset", "(", "wr", ")", "\n", "return", "xw", ",", "nil", "\n", "}" ]
// NewWriter creates a new Writer writing to the given writer. // It is the caller's responsibility to call Close to complete the stream. // // If conf is nil, then default configuration values are used. Writer copies // all configuration values as necessary and does not store conf.
[ "NewWriter", "creates", "a", "new", "Writer", "writing", "to", "the", "given", "writer", ".", "It", "is", "the", "caller", "s", "responsibility", "to", "call", "Close", "to", "complete", "the", "stream", ".", "If", "conf", "is", "nil", "then", "default", "configuration", "values", "are", "used", ".", "Writer", "copies", "all", "configuration", "values", "as", "necessary", "and", "does", "not", "store", "conf", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/xflate/writer.go#L78-L104
9,684
dsnet/compress
xflate/writer.go
Reset
func (xw *Writer) Reset(wr io.Writer) error { *xw = Writer{ wr: wr, mw: xw.mw, zw: xw.zw, nchk: xw.nchk, nidx: xw.nidx, idx: xw.idx, } if xw.zw == nil { xw.zw, _ = newFlateWriter(wr, DefaultCompression) } else { xw.zw.Reset(wr) } if xw.nchk == 0 { xw.nchk = DefaultChunkSize } if xw.nidx == 0 { xw.nidx = DefaultIndexSize } xw.idx.Reset() return nil }
go
func (xw *Writer) Reset(wr io.Writer) error { *xw = Writer{ wr: wr, mw: xw.mw, zw: xw.zw, nchk: xw.nchk, nidx: xw.nidx, idx: xw.idx, } if xw.zw == nil { xw.zw, _ = newFlateWriter(wr, DefaultCompression) } else { xw.zw.Reset(wr) } if xw.nchk == 0 { xw.nchk = DefaultChunkSize } if xw.nidx == 0 { xw.nidx = DefaultIndexSize } xw.idx.Reset() return nil }
[ "func", "(", "xw", "*", "Writer", ")", "Reset", "(", "wr", "io", ".", "Writer", ")", "error", "{", "*", "xw", "=", "Writer", "{", "wr", ":", "wr", ",", "mw", ":", "xw", ".", "mw", ",", "zw", ":", "xw", ".", "zw", ",", "nchk", ":", "xw", ".", "nchk", ",", "nidx", ":", "xw", ".", "nidx", ",", "idx", ":", "xw", ".", "idx", ",", "}", "\n", "if", "xw", ".", "zw", "==", "nil", "{", "xw", ".", "zw", ",", "_", "=", "newFlateWriter", "(", "wr", ",", "DefaultCompression", ")", "\n", "}", "else", "{", "xw", ".", "zw", ".", "Reset", "(", "wr", ")", "\n", "}", "\n", "if", "xw", ".", "nchk", "==", "0", "{", "xw", ".", "nchk", "=", "DefaultChunkSize", "\n", "}", "\n", "if", "xw", ".", "nidx", "==", "0", "{", "xw", ".", "nidx", "=", "DefaultIndexSize", "\n", "}", "\n", "xw", ".", "idx", ".", "Reset", "(", ")", "\n", "return", "nil", "\n", "}" ]
// Reset discards the Writer's state and makes it equivalent to the result // of a call to NewWriter, but writes to wr instead. Any configurations from // a prior call to NewWriter will be preserved. // // This is used to reduce memory allocations.
[ "Reset", "discards", "the", "Writer", "s", "state", "and", "makes", "it", "equivalent", "to", "the", "result", "of", "a", "call", "to", "NewWriter", "but", "writes", "to", "wr", "instead", ".", "Any", "configurations", "from", "a", "prior", "call", "to", "NewWriter", "will", "be", "preserved", ".", "This", "is", "used", "to", "reduce", "memory", "allocations", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/xflate/writer.go#L111-L133
9,685
dsnet/compress
xflate/writer.go
Write
func (xw *Writer) Write(buf []byte) (int, error) { if xw.err != nil { return 0, xw.err } var n, cnt int for len(buf) > 0 && xw.err == nil { // Flush chunk if necessary. remain := xw.nchk - xw.zw.InputOffset if remain <= 0 { xw.err = xw.Flush(FlushFull) continue } if remain > int64(len(buf)) { remain = int64(len(buf)) } // Compress data for current chunk. offset := xw.zw.OutputOffset n, xw.err = xw.zw.Write(buf[:remain]) xw.OutputOffset += xw.zw.OutputOffset - offset buf = buf[n:] cnt += n } xw.InputOffset += int64(cnt) return cnt, xw.err }
go
func (xw *Writer) Write(buf []byte) (int, error) { if xw.err != nil { return 0, xw.err } var n, cnt int for len(buf) > 0 && xw.err == nil { // Flush chunk if necessary. remain := xw.nchk - xw.zw.InputOffset if remain <= 0 { xw.err = xw.Flush(FlushFull) continue } if remain > int64(len(buf)) { remain = int64(len(buf)) } // Compress data for current chunk. offset := xw.zw.OutputOffset n, xw.err = xw.zw.Write(buf[:remain]) xw.OutputOffset += xw.zw.OutputOffset - offset buf = buf[n:] cnt += n } xw.InputOffset += int64(cnt) return cnt, xw.err }
[ "func", "(", "xw", "*", "Writer", ")", "Write", "(", "buf", "[", "]", "byte", ")", "(", "int", ",", "error", ")", "{", "if", "xw", ".", "err", "!=", "nil", "{", "return", "0", ",", "xw", ".", "err", "\n", "}", "\n\n", "var", "n", ",", "cnt", "int", "\n", "for", "len", "(", "buf", ")", ">", "0", "&&", "xw", ".", "err", "==", "nil", "{", "// Flush chunk if necessary.", "remain", ":=", "xw", ".", "nchk", "-", "xw", ".", "zw", ".", "InputOffset", "\n", "if", "remain", "<=", "0", "{", "xw", ".", "err", "=", "xw", ".", "Flush", "(", "FlushFull", ")", "\n", "continue", "\n", "}", "\n", "if", "remain", ">", "int64", "(", "len", "(", "buf", ")", ")", "{", "remain", "=", "int64", "(", "len", "(", "buf", ")", ")", "\n", "}", "\n\n", "// Compress data for current chunk.", "offset", ":=", "xw", ".", "zw", ".", "OutputOffset", "\n", "n", ",", "xw", ".", "err", "=", "xw", ".", "zw", ".", "Write", "(", "buf", "[", ":", "remain", "]", ")", "\n", "xw", ".", "OutputOffset", "+=", "xw", ".", "zw", ".", "OutputOffset", "-", "offset", "\n", "buf", "=", "buf", "[", "n", ":", "]", "\n", "cnt", "+=", "n", "\n", "}", "\n\n", "xw", ".", "InputOffset", "+=", "int64", "(", "cnt", ")", "\n", "return", "cnt", ",", "xw", ".", "err", "\n", "}" ]
// Write writes the compressed form of buf to the underlying io.Writer. // This automatically breaks the input into multiple chunks, writes them out, // and records the sizes of each chunk in the index table.
[ "Write", "writes", "the", "compressed", "form", "of", "buf", "to", "the", "underlying", "io", ".", "Writer", ".", "This", "automatically", "breaks", "the", "input", "into", "multiple", "chunks", "writes", "them", "out", "and", "records", "the", "sizes", "of", "each", "chunk", "in", "the", "index", "table", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/xflate/writer.go#L138-L165
9,686
dsnet/compress
xflate/writer.go
Flush
func (xw *Writer) Flush(mode FlushMode) error { if xw.err != nil { return xw.err } switch mode { case FlushSync: offset := xw.zw.OutputOffset xw.err = xw.zw.Flush() xw.OutputOffset += xw.zw.OutputOffset - offset return xw.err case FlushFull: if xw.err = xw.Flush(FlushSync); xw.err != nil { return xw.err } xw.idx.AppendRecord(xw.zw.OutputOffset, xw.zw.InputOffset, deflateType) xw.zw.Reset(xw.wr) if int64(len(xw.idx.Records)) == xw.nidx { xw.err = xw.Flush(FlushIndex) } return xw.err case FlushIndex: if xw.zw.InputOffset+xw.zw.OutputOffset > 0 { if err := xw.Flush(FlushFull); err != nil { return err } } xw.err = xw.encodeIndex(&xw.idx) backSize := xw.idx.IndexSize xw.idx.Reset() xw.idx.BackSize = backSize return xw.err default: return errorf(errors.Invalid, "invalid flush mode: %d", mode) } }
go
func (xw *Writer) Flush(mode FlushMode) error { if xw.err != nil { return xw.err } switch mode { case FlushSync: offset := xw.zw.OutputOffset xw.err = xw.zw.Flush() xw.OutputOffset += xw.zw.OutputOffset - offset return xw.err case FlushFull: if xw.err = xw.Flush(FlushSync); xw.err != nil { return xw.err } xw.idx.AppendRecord(xw.zw.OutputOffset, xw.zw.InputOffset, deflateType) xw.zw.Reset(xw.wr) if int64(len(xw.idx.Records)) == xw.nidx { xw.err = xw.Flush(FlushIndex) } return xw.err case FlushIndex: if xw.zw.InputOffset+xw.zw.OutputOffset > 0 { if err := xw.Flush(FlushFull); err != nil { return err } } xw.err = xw.encodeIndex(&xw.idx) backSize := xw.idx.IndexSize xw.idx.Reset() xw.idx.BackSize = backSize return xw.err default: return errorf(errors.Invalid, "invalid flush mode: %d", mode) } }
[ "func", "(", "xw", "*", "Writer", ")", "Flush", "(", "mode", "FlushMode", ")", "error", "{", "if", "xw", ".", "err", "!=", "nil", "{", "return", "xw", ".", "err", "\n", "}", "\n\n", "switch", "mode", "{", "case", "FlushSync", ":", "offset", ":=", "xw", ".", "zw", ".", "OutputOffset", "\n", "xw", ".", "err", "=", "xw", ".", "zw", ".", "Flush", "(", ")", "\n", "xw", ".", "OutputOffset", "+=", "xw", ".", "zw", ".", "OutputOffset", "-", "offset", "\n", "return", "xw", ".", "err", "\n", "case", "FlushFull", ":", "if", "xw", ".", "err", "=", "xw", ".", "Flush", "(", "FlushSync", ")", ";", "xw", ".", "err", "!=", "nil", "{", "return", "xw", ".", "err", "\n", "}", "\n", "xw", ".", "idx", ".", "AppendRecord", "(", "xw", ".", "zw", ".", "OutputOffset", ",", "xw", ".", "zw", ".", "InputOffset", ",", "deflateType", ")", "\n", "xw", ".", "zw", ".", "Reset", "(", "xw", ".", "wr", ")", "\n", "if", "int64", "(", "len", "(", "xw", ".", "idx", ".", "Records", ")", ")", "==", "xw", ".", "nidx", "{", "xw", ".", "err", "=", "xw", ".", "Flush", "(", "FlushIndex", ")", "\n", "}", "\n", "return", "xw", ".", "err", "\n", "case", "FlushIndex", ":", "if", "xw", ".", "zw", ".", "InputOffset", "+", "xw", ".", "zw", ".", "OutputOffset", ">", "0", "{", "if", "err", ":=", "xw", ".", "Flush", "(", "FlushFull", ")", ";", "err", "!=", "nil", "{", "return", "err", "\n", "}", "\n", "}", "\n", "xw", ".", "err", "=", "xw", ".", "encodeIndex", "(", "&", "xw", ".", "idx", ")", "\n", "backSize", ":=", "xw", ".", "idx", ".", "IndexSize", "\n", "xw", ".", "idx", ".", "Reset", "(", ")", "\n", "xw", ".", "idx", ".", "BackSize", "=", "backSize", "\n", "return", "xw", ".", "err", "\n", "default", ":", "return", "errorf", "(", "errors", ".", "Invalid", ",", "\"", "\"", ",", "mode", ")", "\n", "}", "\n", "}" ]
// Flush flushes the current write buffer to the underlying writer. // Flushing is entirely optional and should be used sparingly.
[ "Flush", "flushes", "the", "current", "write", "buffer", "to", "the", "underlying", "writer", ".", "Flushing", "is", "entirely", "optional", "and", "should", "be", "used", "sparingly", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/xflate/writer.go#L169-L204
9,687
dsnet/compress
xflate/writer.go
Close
func (xw *Writer) Close() error { if xw.err == errClosed { return nil } if xw.err != nil { return xw.err } // Flush final index. if xw.zw.OutputOffset+xw.zw.InputOffset > 0 || len(xw.idx.Records) > 0 { xw.err = xw.Flush(FlushIndex) if xw.err != nil { return xw.err } } // Encode the footer. err := xw.encodeFooter(xw.idx.BackSize) if err != nil { xw.err = err } else { xw.err = errClosed } return err }
go
func (xw *Writer) Close() error { if xw.err == errClosed { return nil } if xw.err != nil { return xw.err } // Flush final index. if xw.zw.OutputOffset+xw.zw.InputOffset > 0 || len(xw.idx.Records) > 0 { xw.err = xw.Flush(FlushIndex) if xw.err != nil { return xw.err } } // Encode the footer. err := xw.encodeFooter(xw.idx.BackSize) if err != nil { xw.err = err } else { xw.err = errClosed } return err }
[ "func", "(", "xw", "*", "Writer", ")", "Close", "(", ")", "error", "{", "if", "xw", ".", "err", "==", "errClosed", "{", "return", "nil", "\n", "}", "\n", "if", "xw", ".", "err", "!=", "nil", "{", "return", "xw", ".", "err", "\n", "}", "\n\n", "// Flush final index.", "if", "xw", ".", "zw", ".", "OutputOffset", "+", "xw", ".", "zw", ".", "InputOffset", ">", "0", "||", "len", "(", "xw", ".", "idx", ".", "Records", ")", ">", "0", "{", "xw", ".", "err", "=", "xw", ".", "Flush", "(", "FlushIndex", ")", "\n", "if", "xw", ".", "err", "!=", "nil", "{", "return", "xw", ".", "err", "\n", "}", "\n", "}", "\n\n", "// Encode the footer.", "err", ":=", "xw", ".", "encodeFooter", "(", "xw", ".", "idx", ".", "BackSize", ")", "\n", "if", "err", "!=", "nil", "{", "xw", ".", "err", "=", "err", "\n", "}", "else", "{", "xw", ".", "err", "=", "errClosed", "\n", "}", "\n", "return", "err", "\n", "}" ]
// Close ends the XFLATE stream and flushes all buffered data. // This method automatically writes an index if any chunks have been written // since the last FlushIndex.
[ "Close", "ends", "the", "XFLATE", "stream", "and", "flushes", "all", "buffered", "data", ".", "This", "method", "automatically", "writes", "an", "index", "if", "any", "chunks", "have", "been", "written", "since", "the", "last", "FlushIndex", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/xflate/writer.go#L209-L233
9,688
dsnet/compress
xflate/writer.go
encodeIndex
func (xw *Writer) encodeIndex(index *index) error { // Helper function to write VLIs. var crc uint32 var errVLI error writeVLI := func(x int64) { b := xw.scratch[:binary.PutUvarint(xw.scratch[:], uint64(x))] crc = crc32.Update(crc, crc32.MakeTable(crc32.IEEE), b) if _, err := xw.mw.Write(b); err != nil { errVLI = errWrap(err) } } // Write the index. xw.mw.Reset(xw.wr) defer func() { xw.OutputOffset += xw.mw.OutputOffset }() xw.mw.FinalMode = meta.FinalMeta writeVLI(index.BackSize) writeVLI(int64(len(index.Records))) writeVLI(index.LastRecord().CompOffset) writeVLI(index.LastRecord().RawOffset) var preRec record for _, rec := range index.Records { writeVLI(rec.CompOffset - preRec.CompOffset) writeVLI(rec.RawOffset - preRec.RawOffset) preRec = rec } if errVLI != nil { return errWrap(errVLI) } binary.LittleEndian.PutUint32(xw.scratch[:], crc) if _, err := xw.mw.Write(xw.scratch[:4]); err != nil { return errWrap(err) } if err := xw.mw.Close(); err != nil { return errWrap(err) } index.IndexSize = xw.mw.OutputOffset // Record the encoded size return nil }
go
func (xw *Writer) encodeIndex(index *index) error { // Helper function to write VLIs. var crc uint32 var errVLI error writeVLI := func(x int64) { b := xw.scratch[:binary.PutUvarint(xw.scratch[:], uint64(x))] crc = crc32.Update(crc, crc32.MakeTable(crc32.IEEE), b) if _, err := xw.mw.Write(b); err != nil { errVLI = errWrap(err) } } // Write the index. xw.mw.Reset(xw.wr) defer func() { xw.OutputOffset += xw.mw.OutputOffset }() xw.mw.FinalMode = meta.FinalMeta writeVLI(index.BackSize) writeVLI(int64(len(index.Records))) writeVLI(index.LastRecord().CompOffset) writeVLI(index.LastRecord().RawOffset) var preRec record for _, rec := range index.Records { writeVLI(rec.CompOffset - preRec.CompOffset) writeVLI(rec.RawOffset - preRec.RawOffset) preRec = rec } if errVLI != nil { return errWrap(errVLI) } binary.LittleEndian.PutUint32(xw.scratch[:], crc) if _, err := xw.mw.Write(xw.scratch[:4]); err != nil { return errWrap(err) } if err := xw.mw.Close(); err != nil { return errWrap(err) } index.IndexSize = xw.mw.OutputOffset // Record the encoded size return nil }
[ "func", "(", "xw", "*", "Writer", ")", "encodeIndex", "(", "index", "*", "index", ")", "error", "{", "// Helper function to write VLIs.", "var", "crc", "uint32", "\n", "var", "errVLI", "error", "\n", "writeVLI", ":=", "func", "(", "x", "int64", ")", "{", "b", ":=", "xw", ".", "scratch", "[", ":", "binary", ".", "PutUvarint", "(", "xw", ".", "scratch", "[", ":", "]", ",", "uint64", "(", "x", ")", ")", "]", "\n", "crc", "=", "crc32", ".", "Update", "(", "crc", ",", "crc32", ".", "MakeTable", "(", "crc32", ".", "IEEE", ")", ",", "b", ")", "\n", "if", "_", ",", "err", ":=", "xw", ".", "mw", ".", "Write", "(", "b", ")", ";", "err", "!=", "nil", "{", "errVLI", "=", "errWrap", "(", "err", ")", "\n", "}", "\n", "}", "\n\n", "// Write the index.", "xw", ".", "mw", ".", "Reset", "(", "xw", ".", "wr", ")", "\n", "defer", "func", "(", ")", "{", "xw", ".", "OutputOffset", "+=", "xw", ".", "mw", ".", "OutputOffset", "}", "(", ")", "\n", "xw", ".", "mw", ".", "FinalMode", "=", "meta", ".", "FinalMeta", "\n", "writeVLI", "(", "index", ".", "BackSize", ")", "\n", "writeVLI", "(", "int64", "(", "len", "(", "index", ".", "Records", ")", ")", ")", "\n", "writeVLI", "(", "index", ".", "LastRecord", "(", ")", ".", "CompOffset", ")", "\n", "writeVLI", "(", "index", ".", "LastRecord", "(", ")", ".", "RawOffset", ")", "\n", "var", "preRec", "record", "\n", "for", "_", ",", "rec", ":=", "range", "index", ".", "Records", "{", "writeVLI", "(", "rec", ".", "CompOffset", "-", "preRec", ".", "CompOffset", ")", "\n", "writeVLI", "(", "rec", ".", "RawOffset", "-", "preRec", ".", "RawOffset", ")", "\n", "preRec", "=", "rec", "\n", "}", "\n", "if", "errVLI", "!=", "nil", "{", "return", "errWrap", "(", "errVLI", ")", "\n", "}", "\n\n", "binary", ".", "LittleEndian", ".", "PutUint32", "(", "xw", ".", "scratch", "[", ":", "]", ",", "crc", ")", "\n", "if", "_", ",", "err", ":=", "xw", ".", "mw", ".", "Write", "(", "xw", ".", "scratch", "[", ":", "4", "]", ")", ";", "err", "!=", "nil", "{", "return", "errWrap", "(", "err", ")", "\n", "}", "\n", "if", "err", ":=", "xw", ".", "mw", ".", "Close", "(", ")", ";", "err", "!=", "nil", "{", "return", "errWrap", "(", "err", ")", "\n", "}", "\n", "index", ".", "IndexSize", "=", "xw", ".", "mw", ".", "OutputOffset", "// Record the encoded size", "\n", "return", "nil", "\n", "}" ]
// encodeIndex encodes the index into a meta encoded stream. // The index.Records and index.BackSize fields must be populated. // The index.IndexSize field will be populated upon successful write.
[ "encodeIndex", "encodes", "the", "index", "into", "a", "meta", "encoded", "stream", ".", "The", "index", ".", "Records", "and", "index", ".", "BackSize", "fields", "must", "be", "populated", ".", "The", "index", ".", "IndexSize", "field", "will", "be", "populated", "upon", "successful", "write", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/xflate/writer.go#L238-L277
9,689
dsnet/compress
xflate/writer.go
encodeFooter
func (xw *Writer) encodeFooter(backSize int64) error { var n int n += copy(xw.scratch[n:], magic[:]) n += binary.PutUvarint(xw.scratch[n:], uint64(backSize)) xw.mw.Reset(xw.wr) defer func() { xw.OutputOffset += xw.mw.OutputOffset }() xw.mw.FinalMode = meta.FinalStream if _, err := xw.mw.Write(xw.scratch[:n]); err != nil { return errWrap(err) } if err := xw.mw.Close(); err != nil { return errWrap(err) } if xw.mw.NumBlocks != 1 { return errorf(errors.Internal, "footer was not a single block") } return nil }
go
func (xw *Writer) encodeFooter(backSize int64) error { var n int n += copy(xw.scratch[n:], magic[:]) n += binary.PutUvarint(xw.scratch[n:], uint64(backSize)) xw.mw.Reset(xw.wr) defer func() { xw.OutputOffset += xw.mw.OutputOffset }() xw.mw.FinalMode = meta.FinalStream if _, err := xw.mw.Write(xw.scratch[:n]); err != nil { return errWrap(err) } if err := xw.mw.Close(); err != nil { return errWrap(err) } if xw.mw.NumBlocks != 1 { return errorf(errors.Internal, "footer was not a single block") } return nil }
[ "func", "(", "xw", "*", "Writer", ")", "encodeFooter", "(", "backSize", "int64", ")", "error", "{", "var", "n", "int", "\n", "n", "+=", "copy", "(", "xw", ".", "scratch", "[", "n", ":", "]", ",", "magic", "[", ":", "]", ")", "\n", "n", "+=", "binary", ".", "PutUvarint", "(", "xw", ".", "scratch", "[", "n", ":", "]", ",", "uint64", "(", "backSize", ")", ")", "\n\n", "xw", ".", "mw", ".", "Reset", "(", "xw", ".", "wr", ")", "\n", "defer", "func", "(", ")", "{", "xw", ".", "OutputOffset", "+=", "xw", ".", "mw", ".", "OutputOffset", "}", "(", ")", "\n", "xw", ".", "mw", ".", "FinalMode", "=", "meta", ".", "FinalStream", "\n", "if", "_", ",", "err", ":=", "xw", ".", "mw", ".", "Write", "(", "xw", ".", "scratch", "[", ":", "n", "]", ")", ";", "err", "!=", "nil", "{", "return", "errWrap", "(", "err", ")", "\n", "}", "\n", "if", "err", ":=", "xw", ".", "mw", ".", "Close", "(", ")", ";", "err", "!=", "nil", "{", "return", "errWrap", "(", "err", ")", "\n", "}", "\n", "if", "xw", ".", "mw", ".", "NumBlocks", "!=", "1", "{", "return", "errorf", "(", "errors", ".", "Internal", ",", "\"", "\"", ")", "\n", "}", "\n", "return", "nil", "\n", "}" ]
// encodeFooter writes the final footer, encoding the provided backSize into it.
[ "encodeFooter", "writes", "the", "final", "footer", "encoding", "the", "provided", "backSize", "into", "it", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/xflate/writer.go#L280-L298
9,690
dsnet/compress
xflate/internal/meta/writer.go
NewWriter
func NewWriter(wr io.Writer) *Writer { mw := new(Writer) mw.Reset(wr) return mw }
go
func NewWriter(wr io.Writer) *Writer { mw := new(Writer) mw.Reset(wr) return mw }
[ "func", "NewWriter", "(", "wr", "io", ".", "Writer", ")", "*", "Writer", "{", "mw", ":=", "new", "(", "Writer", ")", "\n", "mw", ".", "Reset", "(", "wr", ")", "\n", "return", "mw", "\n", "}" ]
// NewWriter creates a new Writer writing to the given writer. // It is the caller's responsibility to call Close to complete the meta stream.
[ "NewWriter", "creates", "a", "new", "Writer", "writing", "to", "the", "given", "writer", ".", "It", "is", "the", "caller", "s", "responsibility", "to", "call", "Close", "to", "complete", "the", "meta", "stream", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/xflate/internal/meta/writer.go#L40-L44
9,691
dsnet/compress
xflate/internal/meta/writer.go
Reset
func (mw *Writer) Reset(wr io.Writer) { *mw = Writer{ wr: wr, bw: mw.bw, bb: mw.bb, cnts: mw.cnts, } return }
go
func (mw *Writer) Reset(wr io.Writer) { *mw = Writer{ wr: wr, bw: mw.bw, bb: mw.bb, cnts: mw.cnts, } return }
[ "func", "(", "mw", "*", "Writer", ")", "Reset", "(", "wr", "io", ".", "Writer", ")", "{", "*", "mw", "=", "Writer", "{", "wr", ":", "wr", ",", "bw", ":", "mw", ".", "bw", ",", "bb", ":", "mw", ".", "bb", ",", "cnts", ":", "mw", ".", "cnts", ",", "}", "\n", "return", "\n", "}" ]
// Reset discards the Writer's state and makes it equivalent to the result // of a call to NewWriter, but writes to wr instead. // // This is used to reduce memory allocations.
[ "Reset", "discards", "the", "Writer", "s", "state", "and", "makes", "it", "equivalent", "to", "the", "result", "of", "a", "call", "to", "NewWriter", "but", "writes", "to", "wr", "instead", ".", "This", "is", "used", "to", "reduce", "memory", "allocations", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/xflate/internal/meta/writer.go#L50-L58
9,692
dsnet/compress
xflate/internal/meta/writer.go
Write
func (mw *Writer) Write(buf []byte) (int, error) { if mw.err != nil { return 0, mw.err } var wrCnt int for _, b := range buf { zeros, ones := numBits(b) // If possible, avoid flushing to maintain high efficiency. if ensured := mw.bufCnt < EnsureRawBytes; ensured { goto skipEncode } if huffLen, _ := mw.computeHuffLen(mw.buf0s+zeros, mw.buf1s+ones); huffLen > 0 { goto skipEncode } mw.err = mw.encodeBlock(FinalNil) if mw.err != nil { break } skipEncode: mw.buf0s += zeros mw.buf1s += ones mw.buf[mw.bufCnt] = b mw.bufCnt++ wrCnt++ } mw.InputOffset += int64(wrCnt) return wrCnt, mw.err }
go
func (mw *Writer) Write(buf []byte) (int, error) { if mw.err != nil { return 0, mw.err } var wrCnt int for _, b := range buf { zeros, ones := numBits(b) // If possible, avoid flushing to maintain high efficiency. if ensured := mw.bufCnt < EnsureRawBytes; ensured { goto skipEncode } if huffLen, _ := mw.computeHuffLen(mw.buf0s+zeros, mw.buf1s+ones); huffLen > 0 { goto skipEncode } mw.err = mw.encodeBlock(FinalNil) if mw.err != nil { break } skipEncode: mw.buf0s += zeros mw.buf1s += ones mw.buf[mw.bufCnt] = b mw.bufCnt++ wrCnt++ } mw.InputOffset += int64(wrCnt) return wrCnt, mw.err }
[ "func", "(", "mw", "*", "Writer", ")", "Write", "(", "buf", "[", "]", "byte", ")", "(", "int", ",", "error", ")", "{", "if", "mw", ".", "err", "!=", "nil", "{", "return", "0", ",", "mw", ".", "err", "\n", "}", "\n\n", "var", "wrCnt", "int", "\n", "for", "_", ",", "b", ":=", "range", "buf", "{", "zeros", ",", "ones", ":=", "numBits", "(", "b", ")", "\n\n", "// If possible, avoid flushing to maintain high efficiency.", "if", "ensured", ":=", "mw", ".", "bufCnt", "<", "EnsureRawBytes", ";", "ensured", "{", "goto", "skipEncode", "\n", "}", "\n", "if", "huffLen", ",", "_", ":=", "mw", ".", "computeHuffLen", "(", "mw", ".", "buf0s", "+", "zeros", ",", "mw", ".", "buf1s", "+", "ones", ")", ";", "huffLen", ">", "0", "{", "goto", "skipEncode", "\n", "}", "\n\n", "mw", ".", "err", "=", "mw", ".", "encodeBlock", "(", "FinalNil", ")", "\n", "if", "mw", ".", "err", "!=", "nil", "{", "break", "\n", "}", "\n\n", "skipEncode", ":", "mw", ".", "buf0s", "+=", "zeros", "\n", "mw", ".", "buf1s", "+=", "ones", "\n", "mw", ".", "buf", "[", "mw", ".", "bufCnt", "]", "=", "b", "\n", "mw", ".", "bufCnt", "++", "\n", "wrCnt", "++", "\n", "}", "\n\n", "mw", ".", "InputOffset", "+=", "int64", "(", "wrCnt", ")", "\n", "return", "wrCnt", ",", "mw", ".", "err", "\n", "}" ]
// Write writes the encoded form of buf to the underlying io.Writer. // The Writer may buffer the input in order to produce larger meta blocks.
[ "Write", "writes", "the", "encoded", "form", "of", "buf", "to", "the", "underlying", "io", ".", "Writer", ".", "The", "Writer", "may", "buffer", "the", "input", "in", "order", "to", "produce", "larger", "meta", "blocks", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/xflate/internal/meta/writer.go#L62-L94
9,693
dsnet/compress
xflate/internal/meta/writer.go
Close
func (mw *Writer) Close() error { if mw.err == errClosed { return nil } if mw.err != nil { return mw.err } err := mw.encodeBlock(mw.FinalMode) if err != nil { mw.err = err } else { mw.err = errClosed } mw.wr = nil // Release reference to underlying Writer return err }
go
func (mw *Writer) Close() error { if mw.err == errClosed { return nil } if mw.err != nil { return mw.err } err := mw.encodeBlock(mw.FinalMode) if err != nil { mw.err = err } else { mw.err = errClosed } mw.wr = nil // Release reference to underlying Writer return err }
[ "func", "(", "mw", "*", "Writer", ")", "Close", "(", ")", "error", "{", "if", "mw", ".", "err", "==", "errClosed", "{", "return", "nil", "\n", "}", "\n", "if", "mw", ".", "err", "!=", "nil", "{", "return", "mw", ".", "err", "\n", "}", "\n\n", "err", ":=", "mw", ".", "encodeBlock", "(", "mw", ".", "FinalMode", ")", "\n", "if", "err", "!=", "nil", "{", "mw", ".", "err", "=", "err", "\n", "}", "else", "{", "mw", ".", "err", "=", "errClosed", "\n", "}", "\n", "mw", ".", "wr", "=", "nil", "// Release reference to underlying Writer", "\n", "return", "err", "\n", "}" ]
// Close ends the meta stream and flushes all buffered data. // The desired FinalMode must be set prior to calling Close.
[ "Close", "ends", "the", "meta", "stream", "and", "flushes", "all", "buffered", "data", ".", "The", "desired", "FinalMode", "must", "be", "set", "prior", "to", "calling", "Close", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/xflate/internal/meta/writer.go#L98-L114
9,694
dsnet/compress
bzip2/common.go
update
func (c *crc) update(buf []byte) { cval := internal.ReverseUint32(c.val) for len(buf) > 0 { n := len(buf) if n > len(c.buf) { n = len(c.buf) } for i, b := range buf[:n] { c.buf[i] = internal.ReverseLUT[b] } cval = crc32.Update(cval, crc32.IEEETable, c.buf[:n]) buf = buf[n:] } c.val = internal.ReverseUint32(cval) }
go
func (c *crc) update(buf []byte) { cval := internal.ReverseUint32(c.val) for len(buf) > 0 { n := len(buf) if n > len(c.buf) { n = len(c.buf) } for i, b := range buf[:n] { c.buf[i] = internal.ReverseLUT[b] } cval = crc32.Update(cval, crc32.IEEETable, c.buf[:n]) buf = buf[n:] } c.val = internal.ReverseUint32(cval) }
[ "func", "(", "c", "*", "crc", ")", "update", "(", "buf", "[", "]", "byte", ")", "{", "cval", ":=", "internal", ".", "ReverseUint32", "(", "c", ".", "val", ")", "\n", "for", "len", "(", "buf", ")", ">", "0", "{", "n", ":=", "len", "(", "buf", ")", "\n", "if", "n", ">", "len", "(", "c", ".", "buf", ")", "{", "n", "=", "len", "(", "c", ".", "buf", ")", "\n", "}", "\n", "for", "i", ",", "b", ":=", "range", "buf", "[", ":", "n", "]", "{", "c", ".", "buf", "[", "i", "]", "=", "internal", ".", "ReverseLUT", "[", "b", "]", "\n", "}", "\n", "cval", "=", "crc32", ".", "Update", "(", "cval", ",", "crc32", ".", "IEEETable", ",", "c", ".", "buf", "[", ":", "n", "]", ")", "\n", "buf", "=", "buf", "[", "n", ":", "]", "\n", "}", "\n", "c", ".", "val", "=", "internal", ".", "ReverseUint32", "(", "cval", ")", "\n", "}" ]
// update computes the CRC-32 of appending buf to c.
[ "update", "computes", "the", "CRC", "-", "32", "of", "appending", "buf", "to", "c", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/bzip2/common.go#L96-L110
9,695
dsnet/compress
flate/reader.go
readBlockHeader
func (zr *Reader) readBlockHeader() { zr.last = zr.rd.ReadBits(1) == 1 switch zr.rd.ReadBits(2) { case 0: // Raw block (RFC section 3.2.4). zr.rd.ReadPads() n := uint16(zr.rd.ReadBits(16)) nn := uint16(zr.rd.ReadBits(16)) if n^nn != 0xffff { panicf(errors.Corrupted, "raw block size mismatch") } zr.blkLen = int(n) // By convention, an empty block flushes the read buffer. if zr.blkLen == 0 { zr.toRead = zr.dict.ReadFlush() zr.finishBlock() return } zr.step = (*Reader).readRawData case 1: // Fixed prefix block (RFC section 3.2.6). zr.litTree, zr.distTree = &decLit, &decDist zr.step = (*Reader).readBlock case 2: // Dynamic prefix block (RFC section 3.2.7). zr.litTree, zr.distTree = &zr.pd1, &zr.pd2 zr.rd.ReadPrefixCodes(zr.litTree, zr.distTree) zr.step = (*Reader).readBlock default: // Reserved block (RFC section 3.2.3). panicf(errors.Corrupted, "encountered reserved block") } }
go
func (zr *Reader) readBlockHeader() { zr.last = zr.rd.ReadBits(1) == 1 switch zr.rd.ReadBits(2) { case 0: // Raw block (RFC section 3.2.4). zr.rd.ReadPads() n := uint16(zr.rd.ReadBits(16)) nn := uint16(zr.rd.ReadBits(16)) if n^nn != 0xffff { panicf(errors.Corrupted, "raw block size mismatch") } zr.blkLen = int(n) // By convention, an empty block flushes the read buffer. if zr.blkLen == 0 { zr.toRead = zr.dict.ReadFlush() zr.finishBlock() return } zr.step = (*Reader).readRawData case 1: // Fixed prefix block (RFC section 3.2.6). zr.litTree, zr.distTree = &decLit, &decDist zr.step = (*Reader).readBlock case 2: // Dynamic prefix block (RFC section 3.2.7). zr.litTree, zr.distTree = &zr.pd1, &zr.pd2 zr.rd.ReadPrefixCodes(zr.litTree, zr.distTree) zr.step = (*Reader).readBlock default: // Reserved block (RFC section 3.2.3). panicf(errors.Corrupted, "encountered reserved block") } }
[ "func", "(", "zr", "*", "Reader", ")", "readBlockHeader", "(", ")", "{", "zr", ".", "last", "=", "zr", ".", "rd", ".", "ReadBits", "(", "1", ")", "==", "1", "\n", "switch", "zr", ".", "rd", ".", "ReadBits", "(", "2", ")", "{", "case", "0", ":", "// Raw block (RFC section 3.2.4).", "zr", ".", "rd", ".", "ReadPads", "(", ")", "\n\n", "n", ":=", "uint16", "(", "zr", ".", "rd", ".", "ReadBits", "(", "16", ")", ")", "\n", "nn", ":=", "uint16", "(", "zr", ".", "rd", ".", "ReadBits", "(", "16", ")", ")", "\n", "if", "n", "^", "nn", "!=", "0xffff", "{", "panicf", "(", "errors", ".", "Corrupted", ",", "\"", "\"", ")", "\n", "}", "\n", "zr", ".", "blkLen", "=", "int", "(", "n", ")", "\n\n", "// By convention, an empty block flushes the read buffer.", "if", "zr", ".", "blkLen", "==", "0", "{", "zr", ".", "toRead", "=", "zr", ".", "dict", ".", "ReadFlush", "(", ")", "\n", "zr", ".", "finishBlock", "(", ")", "\n", "return", "\n", "}", "\n", "zr", ".", "step", "=", "(", "*", "Reader", ")", ".", "readRawData", "\n", "case", "1", ":", "// Fixed prefix block (RFC section 3.2.6).", "zr", ".", "litTree", ",", "zr", ".", "distTree", "=", "&", "decLit", ",", "&", "decDist", "\n", "zr", ".", "step", "=", "(", "*", "Reader", ")", ".", "readBlock", "\n", "case", "2", ":", "// Dynamic prefix block (RFC section 3.2.7).", "zr", ".", "litTree", ",", "zr", ".", "distTree", "=", "&", "zr", ".", "pd1", ",", "&", "zr", ".", "pd2", "\n", "zr", ".", "rd", ".", "ReadPrefixCodes", "(", "zr", ".", "litTree", ",", "zr", ".", "distTree", ")", "\n", "zr", ".", "step", "=", "(", "*", "Reader", ")", ".", "readBlock", "\n", "default", ":", "// Reserved block (RFC section 3.2.3).", "panicf", "(", "errors", ".", "Corrupted", ",", "\"", "\"", ")", "\n", "}", "\n", "}" ]
// readBlockHeader reads the block header according to RFC section 3.2.3.
[ "readBlockHeader", "reads", "the", "block", "header", "according", "to", "RFC", "section", "3", ".", "2", ".", "3", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/flate/reader.go#L100-L134
9,696
dsnet/compress
flate/reader.go
readRawData
func (zr *Reader) readRawData() { buf := zr.dict.WriteSlice() if len(buf) > zr.blkLen { buf = buf[:zr.blkLen] } cnt, err := zr.rd.Read(buf) zr.blkLen -= cnt zr.dict.WriteMark(cnt) if err != nil { if err == io.EOF { err = io.ErrUnexpectedEOF } errors.Panic(err) } if zr.blkLen > 0 { zr.toRead = zr.dict.ReadFlush() zr.step = (*Reader).readRawData // We need to continue this work return } zr.finishBlock() }
go
func (zr *Reader) readRawData() { buf := zr.dict.WriteSlice() if len(buf) > zr.blkLen { buf = buf[:zr.blkLen] } cnt, err := zr.rd.Read(buf) zr.blkLen -= cnt zr.dict.WriteMark(cnt) if err != nil { if err == io.EOF { err = io.ErrUnexpectedEOF } errors.Panic(err) } if zr.blkLen > 0 { zr.toRead = zr.dict.ReadFlush() zr.step = (*Reader).readRawData // We need to continue this work return } zr.finishBlock() }
[ "func", "(", "zr", "*", "Reader", ")", "readRawData", "(", ")", "{", "buf", ":=", "zr", ".", "dict", ".", "WriteSlice", "(", ")", "\n", "if", "len", "(", "buf", ")", ">", "zr", ".", "blkLen", "{", "buf", "=", "buf", "[", ":", "zr", ".", "blkLen", "]", "\n", "}", "\n\n", "cnt", ",", "err", ":=", "zr", ".", "rd", ".", "Read", "(", "buf", ")", "\n", "zr", ".", "blkLen", "-=", "cnt", "\n", "zr", ".", "dict", ".", "WriteMark", "(", "cnt", ")", "\n", "if", "err", "!=", "nil", "{", "if", "err", "==", "io", ".", "EOF", "{", "err", "=", "io", ".", "ErrUnexpectedEOF", "\n", "}", "\n", "errors", ".", "Panic", "(", "err", ")", "\n", "}", "\n\n", "if", "zr", ".", "blkLen", ">", "0", "{", "zr", ".", "toRead", "=", "zr", ".", "dict", ".", "ReadFlush", "(", ")", "\n", "zr", ".", "step", "=", "(", "*", "Reader", ")", ".", "readRawData", "// We need to continue this work", "\n", "return", "\n", "}", "\n", "zr", ".", "finishBlock", "(", ")", "\n", "}" ]
// readRawData reads raw data according to RFC section 3.2.4.
[ "readRawData", "reads", "raw", "data", "according", "to", "RFC", "section", "3", ".", "2", ".", "4", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/flate/reader.go#L137-L159
9,697
dsnet/compress
flate/reader.go
readBlock
func (zr *Reader) readBlock() { const ( stateInit = iota // Zero value must be stateInit stateDict ) switch zr.stepState { case stateInit: goto readLiteral case stateDict: goto copyDistance } readLiteral: // Read literal and/or (length, distance) according to RFC section 3.2.3. { if zr.dict.AvailSize() == 0 { zr.toRead = zr.dict.ReadFlush() zr.step = (*Reader).readBlock zr.stepState = stateInit // Need to continue work here return } // Read the literal symbol. litSym, ok := zr.rd.TryReadSymbol(zr.litTree) if !ok { litSym = zr.rd.ReadSymbol(zr.litTree) } switch { case litSym < endBlockSym: zr.dict.WriteByte(byte(litSym)) goto readLiteral case litSym == endBlockSym: zr.finishBlock() zr.stepState = stateInit // Next call to readBlock must start here return case litSym < maxNumLitSyms: // Decode the copy length. rec := lenRanges[litSym-257] extra, ok := zr.rd.TryReadBits(uint(rec.Len)) if !ok { extra = zr.rd.ReadBits(uint(rec.Len)) } zr.cpyLen = int(rec.Base) + int(extra) // Read the distance symbol. distSym, ok := zr.rd.TryReadSymbol(zr.distTree) if !ok { distSym = zr.rd.ReadSymbol(zr.distTree) } if distSym >= maxNumDistSyms { panicf(errors.Corrupted, "invalid distance symbol: %d", distSym) } // Decode the copy distance. rec = distRanges[distSym] extra, ok = zr.rd.TryReadBits(uint(rec.Len)) if !ok { extra = zr.rd.ReadBits(uint(rec.Len)) } zr.dist = int(rec.Base) + int(extra) if zr.dist > zr.dict.HistSize() { panicf(errors.Corrupted, "copy distance exceeds window history") } goto copyDistance default: panicf(errors.Corrupted, "invalid literal symbol: %d", litSym) } } copyDistance: // Perform a backwards copy according to RFC section 3.2.3. { cnt := zr.dict.TryWriteCopy(zr.dist, zr.cpyLen) if cnt == 0 { cnt = zr.dict.WriteCopy(zr.dist, zr.cpyLen) } zr.cpyLen -= cnt if zr.cpyLen > 0 { zr.toRead = zr.dict.ReadFlush() zr.step = (*Reader).readBlock zr.stepState = stateDict // Need to continue work here return } goto readLiteral } }
go
func (zr *Reader) readBlock() { const ( stateInit = iota // Zero value must be stateInit stateDict ) switch zr.stepState { case stateInit: goto readLiteral case stateDict: goto copyDistance } readLiteral: // Read literal and/or (length, distance) according to RFC section 3.2.3. { if zr.dict.AvailSize() == 0 { zr.toRead = zr.dict.ReadFlush() zr.step = (*Reader).readBlock zr.stepState = stateInit // Need to continue work here return } // Read the literal symbol. litSym, ok := zr.rd.TryReadSymbol(zr.litTree) if !ok { litSym = zr.rd.ReadSymbol(zr.litTree) } switch { case litSym < endBlockSym: zr.dict.WriteByte(byte(litSym)) goto readLiteral case litSym == endBlockSym: zr.finishBlock() zr.stepState = stateInit // Next call to readBlock must start here return case litSym < maxNumLitSyms: // Decode the copy length. rec := lenRanges[litSym-257] extra, ok := zr.rd.TryReadBits(uint(rec.Len)) if !ok { extra = zr.rd.ReadBits(uint(rec.Len)) } zr.cpyLen = int(rec.Base) + int(extra) // Read the distance symbol. distSym, ok := zr.rd.TryReadSymbol(zr.distTree) if !ok { distSym = zr.rd.ReadSymbol(zr.distTree) } if distSym >= maxNumDistSyms { panicf(errors.Corrupted, "invalid distance symbol: %d", distSym) } // Decode the copy distance. rec = distRanges[distSym] extra, ok = zr.rd.TryReadBits(uint(rec.Len)) if !ok { extra = zr.rd.ReadBits(uint(rec.Len)) } zr.dist = int(rec.Base) + int(extra) if zr.dist > zr.dict.HistSize() { panicf(errors.Corrupted, "copy distance exceeds window history") } goto copyDistance default: panicf(errors.Corrupted, "invalid literal symbol: %d", litSym) } } copyDistance: // Perform a backwards copy according to RFC section 3.2.3. { cnt := zr.dict.TryWriteCopy(zr.dist, zr.cpyLen) if cnt == 0 { cnt = zr.dict.WriteCopy(zr.dist, zr.cpyLen) } zr.cpyLen -= cnt if zr.cpyLen > 0 { zr.toRead = zr.dict.ReadFlush() zr.step = (*Reader).readBlock zr.stepState = stateDict // Need to continue work here return } goto readLiteral } }
[ "func", "(", "zr", "*", "Reader", ")", "readBlock", "(", ")", "{", "const", "(", "stateInit", "=", "iota", "// Zero value must be stateInit", "\n", "stateDict", "\n", ")", "\n\n", "switch", "zr", ".", "stepState", "{", "case", "stateInit", ":", "goto", "readLiteral", "\n", "case", "stateDict", ":", "goto", "copyDistance", "\n", "}", "\n\n", "readLiteral", ":", "// Read literal and/or (length, distance) according to RFC section 3.2.3.", "{", "if", "zr", ".", "dict", ".", "AvailSize", "(", ")", "==", "0", "{", "zr", ".", "toRead", "=", "zr", ".", "dict", ".", "ReadFlush", "(", ")", "\n", "zr", ".", "step", "=", "(", "*", "Reader", ")", ".", "readBlock", "\n", "zr", ".", "stepState", "=", "stateInit", "// Need to continue work here", "\n", "return", "\n", "}", "\n\n", "// Read the literal symbol.", "litSym", ",", "ok", ":=", "zr", ".", "rd", ".", "TryReadSymbol", "(", "zr", ".", "litTree", ")", "\n", "if", "!", "ok", "{", "litSym", "=", "zr", ".", "rd", ".", "ReadSymbol", "(", "zr", ".", "litTree", ")", "\n", "}", "\n", "switch", "{", "case", "litSym", "<", "endBlockSym", ":", "zr", ".", "dict", ".", "WriteByte", "(", "byte", "(", "litSym", ")", ")", "\n", "goto", "readLiteral", "\n", "case", "litSym", "==", "endBlockSym", ":", "zr", ".", "finishBlock", "(", ")", "\n", "zr", ".", "stepState", "=", "stateInit", "// Next call to readBlock must start here", "\n", "return", "\n", "case", "litSym", "<", "maxNumLitSyms", ":", "// Decode the copy length.", "rec", ":=", "lenRanges", "[", "litSym", "-", "257", "]", "\n", "extra", ",", "ok", ":=", "zr", ".", "rd", ".", "TryReadBits", "(", "uint", "(", "rec", ".", "Len", ")", ")", "\n", "if", "!", "ok", "{", "extra", "=", "zr", ".", "rd", ".", "ReadBits", "(", "uint", "(", "rec", ".", "Len", ")", ")", "\n", "}", "\n", "zr", ".", "cpyLen", "=", "int", "(", "rec", ".", "Base", ")", "+", "int", "(", "extra", ")", "\n\n", "// Read the distance symbol.", "distSym", ",", "ok", ":=", "zr", ".", "rd", ".", "TryReadSymbol", "(", "zr", ".", "distTree", ")", "\n", "if", "!", "ok", "{", "distSym", "=", "zr", ".", "rd", ".", "ReadSymbol", "(", "zr", ".", "distTree", ")", "\n", "}", "\n", "if", "distSym", ">=", "maxNumDistSyms", "{", "panicf", "(", "errors", ".", "Corrupted", ",", "\"", "\"", ",", "distSym", ")", "\n", "}", "\n\n", "// Decode the copy distance.", "rec", "=", "distRanges", "[", "distSym", "]", "\n", "extra", ",", "ok", "=", "zr", ".", "rd", ".", "TryReadBits", "(", "uint", "(", "rec", ".", "Len", ")", ")", "\n", "if", "!", "ok", "{", "extra", "=", "zr", ".", "rd", ".", "ReadBits", "(", "uint", "(", "rec", ".", "Len", ")", ")", "\n", "}", "\n", "zr", ".", "dist", "=", "int", "(", "rec", ".", "Base", ")", "+", "int", "(", "extra", ")", "\n", "if", "zr", ".", "dist", ">", "zr", ".", "dict", ".", "HistSize", "(", ")", "{", "panicf", "(", "errors", ".", "Corrupted", ",", "\"", "\"", ")", "\n", "}", "\n\n", "goto", "copyDistance", "\n", "default", ":", "panicf", "(", "errors", ".", "Corrupted", ",", "\"", "\"", ",", "litSym", ")", "\n", "}", "\n", "}", "\n\n", "copyDistance", ":", "// Perform a backwards copy according to RFC section 3.2.3.", "{", "cnt", ":=", "zr", ".", "dict", ".", "TryWriteCopy", "(", "zr", ".", "dist", ",", "zr", ".", "cpyLen", ")", "\n", "if", "cnt", "==", "0", "{", "cnt", "=", "zr", ".", "dict", ".", "WriteCopy", "(", "zr", ".", "dist", ",", "zr", ".", "cpyLen", ")", "\n", "}", "\n", "zr", ".", "cpyLen", "-=", "cnt", "\n\n", "if", "zr", ".", "cpyLen", ">", "0", "{", "zr", ".", "toRead", "=", "zr", ".", "dict", ".", "ReadFlush", "(", ")", "\n", "zr", ".", "step", "=", "(", "*", "Reader", ")", ".", "readBlock", "\n", "zr", ".", "stepState", "=", "stateDict", "// Need to continue work here", "\n", "return", "\n", "}", "\n", "goto", "readLiteral", "\n", "}", "\n", "}" ]
// readCommands reads block commands according to RFC section 3.2.3.
[ "readCommands", "reads", "block", "commands", "according", "to", "RFC", "section", "3", ".", "2", ".", "3", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/flate/reader.go#L162-L250
9,698
dsnet/compress
flate/reader.go
finishBlock
func (zr *Reader) finishBlock() { if zr.last { zr.rd.ReadPads() zr.err = io.EOF } zr.step = (*Reader).readBlockHeader }
go
func (zr *Reader) finishBlock() { if zr.last { zr.rd.ReadPads() zr.err = io.EOF } zr.step = (*Reader).readBlockHeader }
[ "func", "(", "zr", "*", "Reader", ")", "finishBlock", "(", ")", "{", "if", "zr", ".", "last", "{", "zr", ".", "rd", ".", "ReadPads", "(", ")", "\n", "zr", ".", "err", "=", "io", ".", "EOF", "\n", "}", "\n", "zr", ".", "step", "=", "(", "*", "Reader", ")", ".", "readBlockHeader", "\n", "}" ]
// finishBlock checks if we have hit io.EOF.
[ "finishBlock", "checks", "if", "we", "have", "hit", "io", ".", "EOF", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/flate/reader.go#L253-L259
9,699
dsnet/compress
brotli/bit_reader.go
FlushOffset
func (br *bitReader) FlushOffset() int64 { if br.bufRd == nil { return br.offset } // Update the number of total bits to discard. br.discardBits += int(br.fedBits - br.numBits) br.fedBits = br.numBits // Discard some bytes to update read offset. nd := (br.discardBits + 7) / 8 // Round up to nearest byte nd, _ = br.bufRd.Discard(nd) br.discardBits -= nd * 8 // -7..0 br.offset += int64(nd) // These are invalid after Discard. br.bufPeek = nil return br.offset }
go
func (br *bitReader) FlushOffset() int64 { if br.bufRd == nil { return br.offset } // Update the number of total bits to discard. br.discardBits += int(br.fedBits - br.numBits) br.fedBits = br.numBits // Discard some bytes to update read offset. nd := (br.discardBits + 7) / 8 // Round up to nearest byte nd, _ = br.bufRd.Discard(nd) br.discardBits -= nd * 8 // -7..0 br.offset += int64(nd) // These are invalid after Discard. br.bufPeek = nil return br.offset }
[ "func", "(", "br", "*", "bitReader", ")", "FlushOffset", "(", ")", "int64", "{", "if", "br", ".", "bufRd", "==", "nil", "{", "return", "br", ".", "offset", "\n", "}", "\n\n", "// Update the number of total bits to discard.", "br", ".", "discardBits", "+=", "int", "(", "br", ".", "fedBits", "-", "br", ".", "numBits", ")", "\n", "br", ".", "fedBits", "=", "br", ".", "numBits", "\n\n", "// Discard some bytes to update read offset.", "nd", ":=", "(", "br", ".", "discardBits", "+", "7", ")", "/", "8", "// Round up to nearest byte", "\n", "nd", ",", "_", "=", "br", ".", "bufRd", ".", "Discard", "(", "nd", ")", "\n", "br", ".", "discardBits", "-=", "nd", "*", "8", "// -7..0", "\n", "br", ".", "offset", "+=", "int64", "(", "nd", ")", "\n\n", "// These are invalid after Discard.", "br", ".", "bufPeek", "=", "nil", "\n", "return", "br", ".", "offset", "\n", "}" ]
// FlushOffset updates the read offset of the underlying byteReader. // If the byteReader is a bufio.Reader, then this calls Discard to update the // read offset.
[ "FlushOffset", "updates", "the", "read", "offset", "of", "the", "underlying", "byteReader", ".", "If", "the", "byteReader", "is", "a", "bufio", ".", "Reader", "then", "this", "calls", "Discard", "to", "update", "the", "read", "offset", "." ]
da652975a8eea9fa0735aba8056747a751db0bd3
https://github.com/dsnet/compress/blob/da652975a8eea9fa0735aba8056747a751db0bd3/brotli/bit_reader.go#L61-L79