Bump github.com/aws/aws-sdk-go from 1.43.36 to 1.44.24 (#2111)

Bumps [github.com/aws/aws-sdk-go](https://github.com/aws/aws-sdk-go) from 1.43.36 to 1.44.24.
- [Release notes](https://github.com/aws/aws-sdk-go/releases)
- [Changelog](https://github.com/aws/aws-sdk-go/blob/main/CHANGELOG.md)
- [Commits](https://github.com/aws/aws-sdk-go/compare/v1.43.36...v1.44.24)

---
updated-dependencies:
- dependency-name: github.com/aws/aws-sdk-go
  dependency-type: direct:production
  update-type: version-update:semver-minor
...

Signed-off-by: dependabot[bot] <support@github.com>

Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
This commit is contained in:
dependabot[bot] 2022-06-01 13:14:12 -04:00 committed by GitHub
parent c7841d7653
commit 3d2c088655
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
28 changed files with 911 additions and 59 deletions

2
go.mod
View File

@ -13,7 +13,7 @@ replace (
require ( require (
cloud.google.com/go/storage v1.22.0 cloud.google.com/go/storage v1.22.0
github.com/Azure/azure-storage-blob-go v0.14.0 github.com/Azure/azure-storage-blob-go v0.14.0
github.com/aws/aws-sdk-go v1.43.36 github.com/aws/aws-sdk-go v1.44.24
github.com/awslabs/amazon-ecr-credential-helper/ecr-login v0.0.0-20220228164355-396b2034c795 github.com/awslabs/amazon-ecr-credential-helper/ecr-login v0.0.0-20220228164355-396b2034c795
github.com/chrismellard/docker-credential-acr-env v0.0.0-20220119192733-fe33c00cee21 github.com/chrismellard/docker-credential-acr-env v0.0.0-20220119192733-fe33c00cee21
github.com/containerd/cgroups v1.0.3 // indirect github.com/containerd/cgroups v1.0.3 // indirect

4
go.sum
View File

@ -276,8 +276,8 @@ github.com/aws/aws-sdk-go v1.25.37/go.mod h1:KmX6BPdI08NWTb3/sm4ZGu5ShLoqVDhKgpi
github.com/aws/aws-sdk-go v1.27.1/go.mod h1:KmX6BPdI08NWTb3/sm4ZGu5ShLoqVDhKgpiN924inxo= github.com/aws/aws-sdk-go v1.27.1/go.mod h1:KmX6BPdI08NWTb3/sm4ZGu5ShLoqVDhKgpiN924inxo=
github.com/aws/aws-sdk-go v1.31.6/go.mod h1:5zCpMtNQVjRREroY7sYe8lOMRSxkhG6MZveU8YkpAk0= github.com/aws/aws-sdk-go v1.31.6/go.mod h1:5zCpMtNQVjRREroY7sYe8lOMRSxkhG6MZveU8YkpAk0=
github.com/aws/aws-sdk-go v1.36.30/go.mod h1:hcU610XS61/+aQV88ixoOzUoG7v3b31pl2zKMmprdro= github.com/aws/aws-sdk-go v1.36.30/go.mod h1:hcU610XS61/+aQV88ixoOzUoG7v3b31pl2zKMmprdro=
github.com/aws/aws-sdk-go v1.43.36 h1:8a+pYKNT7wSxUy3fi5dSqKQdfmit7SYGg5fv4zf+WuA= github.com/aws/aws-sdk-go v1.44.24 h1:3nOkwJBJLiGBmJKWp3z0utyXuBkxyGkRRwWjrTItJaY=
github.com/aws/aws-sdk-go v1.43.36/go.mod h1:y4AeaBuwd2Lk+GepC1E9v0qOiTws0MIWAX4oIKwKHZo= github.com/aws/aws-sdk-go v1.44.24/go.mod h1:y4AeaBuwd2Lk+GepC1E9v0qOiTws0MIWAX4oIKwKHZo=
github.com/aws/aws-sdk-go-v2 v1.7.1/go.mod h1:L5LuPC1ZgDr2xQS7AmIec/Jlc7O/Y1u2KxJyNVab250= github.com/aws/aws-sdk-go-v2 v1.7.1/go.mod h1:L5LuPC1ZgDr2xQS7AmIec/Jlc7O/Y1u2KxJyNVab250=
github.com/aws/aws-sdk-go-v2 v1.14.0 h1:IzSYBJHu0ZdUi27kIW6xVrs0eSxI4AzwbenzfXhhVs4= github.com/aws/aws-sdk-go-v2 v1.14.0 h1:IzSYBJHu0ZdUi27kIW6xVrs0eSxI4AzwbenzfXhhVs4=
github.com/aws/aws-sdk-go-v2 v1.14.0/go.mod h1:ZA3Y8V0LrlWj63MQAnRHgKf/5QB//LSZCPNWlWrNGLU= github.com/aws/aws-sdk-go-v2 v1.14.0/go.mod h1:ZA3Y8V0LrlWj63MQAnRHgKf/5QB//LSZCPNWlWrNGLU=

File diff suppressed because it is too large Load Diff

View File

@ -330,6 +330,9 @@ func MakeAddToUserAgentFreeFormHandler(s string) func(*Request) {
// WithSetRequestHeaders updates the operation request's HTTP header to contain // WithSetRequestHeaders updates the operation request's HTTP header to contain
// the header key value pairs provided. If the header key already exists in the // the header key value pairs provided. If the header key already exists in the
// request's HTTP header set, the existing value(s) will be replaced. // request's HTTP header set, the existing value(s) will be replaced.
//
// Header keys added will be added as canonical format with title casing
// applied via http.Header.Set method.
func WithSetRequestHeaders(h map[string]string) Option { func WithSetRequestHeaders(h map[string]string) Option {
return withRequestHeader(h).SetRequestHeaders return withRequestHeader(h).SetRequestHeaders
} }
@ -338,6 +341,6 @@ type withRequestHeader map[string]string
func (h withRequestHeader) SetRequestHeaders(r *Request) { func (h withRequestHeader) SetRequestHeaders(r *Request) {
for k, v := range h { for k, v := range h {
r.HTTPRequest.Header[k] = []string{v} r.HTTPRequest.Header.Set(k, v)
} }
} }

View File

@ -5,4 +5,4 @@ package aws
const SDKName = "aws-sdk-go" const SDKName = "aws-sdk-go"
// SDKVersion is the version of this SDK // SDKVersion is the version of this SDK
const SDKVersion = "1.43.36" const SDKVersion = "1.44.24"

View File

@ -8113,8 +8113,9 @@ func (c *S3) PutBucketLifecycleConfigurationRequest(input *PutBucketLifecycleCon
// Rules // Rules
// //
// You specify the lifecycle configuration in your request body. The lifecycle // You specify the lifecycle configuration in your request body. The lifecycle
// configuration is specified as XML consisting of one or more rules. Each rule // configuration is specified as XML consisting of one or more rules. An Amazon
// consists of the following: // S3 Lifecycle configuration can have up to 1,000 rules. This limit is not
// adjustable. Each rule consists of the following:
// //
// * Filter identifying a subset of objects to which the rule applies. The // * Filter identifying a subset of objects to which the rule applies. The
// filter can be based on a key name prefix, object tags, or a combination // filter can be based on a key name prefix, object tags, or a combination
@ -10918,9 +10919,11 @@ func (c *S3) UploadPartRequest(input *UploadPartInput) (req *request.Request, ou
// Part numbers can be any number from 1 to 10,000, inclusive. A part number // Part numbers can be any number from 1 to 10,000, inclusive. A part number
// uniquely identifies a part and also defines its position within the object // uniquely identifies a part and also defines its position within the object
// being created. If you upload a new part using the same part number that was // being created. If you upload a new part using the same part number that was
// used with a previous part, the previously uploaded part is overwritten. Each // used with a previous part, the previously uploaded part is overwritten.
// part must be at least 5 MB in size, except the last part. There is no size //
// limit on the last part of your multipart upload. // For information about maximum and minimum part sizes and other multipart
// upload specifications, see Multipart upload limits (https://docs.aws.amazon.com/AmazonS3/latest/userguide/qfacts.html)
// in the Amazon S3 User Guide.
// //
// To ensure that data is not corrupted when traversing the network, specify // To ensure that data is not corrupted when traversing the network, specify
// the Content-MD5 header in the upload part request. Amazon S3 checks the part // the Content-MD5 header in the upload part request. Amazon S3 checks the part
@ -11068,8 +11071,8 @@ func (c *S3) UploadPartCopyRequest(input *UploadPartCopyInput) (req *request.Req
// your request and a byte range by adding the request header x-amz-copy-source-range // your request and a byte range by adding the request header x-amz-copy-source-range
// in your request. // in your request.
// //
// The minimum allowable part size for a multipart upload is 5 MB. For more // For information about maximum and minimum part sizes and other multipart
// information about multipart upload limits, go to Quick Facts (https://docs.aws.amazon.com/AmazonS3/latest/dev/qfacts.html) // upload specifications, see Multipart upload limits (https://docs.aws.amazon.com/AmazonS3/latest/userguide/qfacts.html)
// in the Amazon S3 User Guide. // in the Amazon S3 User Guide.
// //
// Instead of using an existing object as part data, you might use the UploadPart // Instead of using an existing object as part data, you might use the UploadPart
@ -29347,9 +29350,9 @@ type NoncurrentVersionExpiration struct {
NewerNoncurrentVersions *int64 `type:"integer"` NewerNoncurrentVersions *int64 `type:"integer"`
// Specifies the number of days an object is noncurrent before Amazon S3 can // Specifies the number of days an object is noncurrent before Amazon S3 can
// perform the associated action. For information about the noncurrent days // perform the associated action. The value must be a non-zero positive integer.
// calculations, see How Amazon S3 Calculates When an Object Became Noncurrent // For information about the noncurrent days calculations, see How Amazon S3
// (https://docs.aws.amazon.com/AmazonS3/latest/dev/intro-lifecycle-rules.html#non-current-days-calculations) // Calculates When an Object Became Noncurrent (https://docs.aws.amazon.com/AmazonS3/latest/dev/intro-lifecycle-rules.html#non-current-days-calculations)
// in the Amazon S3 User Guide. // in the Amazon S3 User Guide.
NoncurrentDays *int64 `type:"integer"` NoncurrentDays *int64 `type:"integer"`
} }
@ -29662,7 +29665,9 @@ type Object struct {
// //
// * If an object is created by either the Multipart Upload or Part Copy // * If an object is created by either the Multipart Upload or Part Copy
// operation, the ETag is not an MD5 digest, regardless of the method of // operation, the ETag is not an MD5 digest, regardless of the method of
// encryption. // encryption. If an object is larger than 16 MB, the Amazon Web Services
// Management Console will upload or copy that object as a Multipart Upload,
// and therefore the ETag will not be an MD5 digest.
ETag *string `type:"string"` ETag *string `type:"string"`
// The name that you assign to an object. You use the object key to retrieve // The name that you assign to an object. You use the object key to retrieve

View File

@ -1279,6 +1279,12 @@ func (c *STS) GetSessionTokenRequest(input *GetSessionTokenInput) (req *request.
// and Comparing the Amazon Web Services STS API operations (https://docs.aws.amazon.com/IAM/latest/UserGuide/id_credentials_temp_request.html#stsapi_comparison) // and Comparing the Amazon Web Services STS API operations (https://docs.aws.amazon.com/IAM/latest/UserGuide/id_credentials_temp_request.html#stsapi_comparison)
// in the IAM User Guide. // in the IAM User Guide.
// //
// No permissions are required for users to perform this operation. The purpose
// of the sts:GetSessionToken operation is to authenticate the user using MFA.
// You cannot use policies to control authentication operations. For more information,
// see Permissions for GetSessionToken (https://docs.aws.amazon.com/IAM/latest/UserGuide/id_credentials_temp_control-access_getsessiontoken.html)
// in the IAM User Guide.
//
// Session Duration // Session Duration
// //
// The GetSessionToken operation must be called by using the long-term Amazon // The GetSessionToken operation must be called by using the long-term Amazon

View File

@ -253,7 +253,7 @@ func getStatFileContentUint64(filePath string) uint64 {
res, err := parseUint(trimmed, 10, 64) res, err := parseUint(trimmed, 10, 64)
if err != nil { if err != nil {
logrus.Errorf("Unable to parse %q as a uint from Cgroup file %q", string(contents), filePath) logrus.Errorf("unable to parse %q as a uint from Cgroup file %q", string(contents), filePath)
return res return res
} }

View File

@ -132,7 +132,7 @@ func removeDockerfile(c modifiableContext, filesToRemove ...string) error {
for _, fileToRemove := range filesToRemove { for _, fileToRemove := range filesToRemove {
if rm, _ := fileutils.Matches(fileToRemove, excludes); rm { if rm, _ := fileutils.Matches(fileToRemove, excludes); rm {
if err := c.Remove(fileToRemove); err != nil { if err := c.Remove(fileToRemove); err != nil {
logrus.Errorf("Failed to remove %s: %v", fileToRemove, err) logrus.Errorf("failed to remove %s: %v", fileToRemove, err)
} }
} }
} }

View File

@ -681,7 +681,7 @@ func (container *Container) InitializeStdio(iop *cio.DirectIO) (cio.IO, error) {
if container.StreamConfig.Stdin() == nil && !container.Config.Tty { if container.StreamConfig.Stdin() == nil && !container.Config.Tty {
if iop.Stdin != nil { if iop.Stdin != nil {
if err := iop.Stdin.Close(); err != nil { if err := iop.Stdin.Close(); err != nil {
logrus.Warnf("Error closing stdin: %+v", err) logrus.Warnf("error closing stdin: %+v", err)
} }
} }
} }

View File

@ -1,4 +1,3 @@
//go:build !windows
// +build !windows // +build !windows
package container // import "github.com/docker/docker/container" package container // import "github.com/docker/docker/container"
@ -145,7 +144,7 @@ func (container *Container) CopyImagePathContent(v volume.Volume, destination st
defer func() { defer func() {
if err := v.Unmount(id); err != nil { if err := v.Unmount(id); err != nil {
logrus.Warnf("Error while unmounting volume %s: %v", v.Name(), err) logrus.Warnf("error while unmounting volume %s: %v", v.Name(), err)
} }
}() }()
if err := label.Relabel(path, container.MountLabel, true); err != nil && !errors.Is(err, syscall.ENOTSUP) { if err := label.Relabel(path, container.MountLabel, true); err != nil && !errors.Is(err, syscall.ENOTSUP) {

View File

@ -63,8 +63,8 @@ func (c *Config) CopyStreams(ctx context.Context, cfg *AttachConfig) <-chan erro
// Connect stdin of container to the attach stdin stream. // Connect stdin of container to the attach stdin stream.
if cfg.Stdin != nil { if cfg.Stdin != nil {
group.Go(func() error { group.Go(func() error {
logrus.Debug("Attach: stdin: begin") logrus.Debug("attach: stdin: begin")
defer logrus.Debug("Attach: stdin: end") defer logrus.Debug("attach: stdin: end")
defer func() { defer func() {
if cfg.CloseStdin && !cfg.TTY { if cfg.CloseStdin && !cfg.TTY {
@ -98,8 +98,8 @@ func (c *Config) CopyStreams(ctx context.Context, cfg *AttachConfig) <-chan erro
} }
attachStream := func(name string, stream io.Writer, streamPipe io.ReadCloser) error { attachStream := func(name string, stream io.Writer, streamPipe io.ReadCloser) error {
logrus.Debugf("Attach: %s: begin", name) logrus.Debugf("attach: %s: begin", name)
defer logrus.Debugf("Attach: %s: end", name) defer logrus.Debugf("attach: %s: end", name)
defer func() { defer func() {
// Make sure stdin gets closed // Make sure stdin gets closed
if cfg.Stdin != nil { if cfg.Stdin != nil {
@ -132,7 +132,7 @@ func (c *Config) CopyStreams(ctx context.Context, cfg *AttachConfig) <-chan erro
errs := make(chan error, 1) errs := make(chan error, 1)
go func() { go func() {
defer logrus.Debug("Attach done") defer logrus.Debug("attach done")
groupErr := make(chan error, 1) groupErr := make(chan error, 1)
go func() { go func() {
groupErr <- group.Wait() groupErr <- group.Wait()

View File

@ -122,7 +122,7 @@ func (c *Config) CopyToPipe(iop *cio.DirectIO) {
c.wg.Add(1) c.wg.Add(1)
go func() { go func() {
if _, err := pools.Copy(w, r); err != nil { if _, err := pools.Copy(w, r); err != nil {
logrus.Errorf("Stream copy error: %v", err) logrus.Errorf("stream copy error: %v", err)
} }
r.Close() r.Close()
c.wg.Done() c.wg.Done()
@ -141,7 +141,7 @@ func (c *Config) CopyToPipe(iop *cio.DirectIO) {
go func() { go func() {
pools.Copy(iop.Stdin, stdin) pools.Copy(iop.Stdin, stdin)
if err := iop.Stdin.Close(); err != nil { if err := iop.Stdin.Close(); err != nil {
logrus.Warnf("Failed to close stdin: %v", err) logrus.Warnf("failed to close stdin: %v", err)
} }
}() }()
} }

View File

@ -382,7 +382,7 @@ func (v *memdbView) transform(container *Container) *Snapshot {
for port, bindings := range container.NetworkSettings.Ports { for port, bindings := range container.NetworkSettings.Ports {
p, err := nat.ParsePort(port.Port()) p, err := nat.ParsePort(port.Port())
if err != nil { if err != nil {
logrus.Warnf("Invalid port map %+v", err) logrus.Warnf("invalid port map %+v", err)
continue continue
} }
if len(bindings) == 0 { if len(bindings) == 0 {
@ -395,7 +395,7 @@ func (v *memdbView) transform(container *Container) *Snapshot {
for _, binding := range bindings { for _, binding := range bindings {
h, err := nat.ParsePort(binding.HostPort) h, err := nat.ParsePort(binding.HostPort)
if err != nil { if err != nil {
logrus.Warnf("Invalid host port map %+v", err) logrus.Warnf("invalid host port map %+v", err)
continue continue
} }
snapshot.Ports = append(snapshot.Ports, types.Port{ snapshot.Ports = append(snapshot.Ports, types.Port{

View File

@ -71,7 +71,7 @@ func (c *Config) InitializeStdio(iop *cio.DirectIO) (cio.IO, error) {
if c.StreamConfig.Stdin() == nil && !c.Tty && runtime.GOOS == "windows" { if c.StreamConfig.Stdin() == nil && !c.Tty && runtime.GOOS == "windows" {
if iop.Stdin != nil { if iop.Stdin != nil {
if err := iop.Stdin.Close(); err != nil { if err := iop.Stdin.Close(); err != nil {
logrus.Errorf("Error closing exec stdin: %+v", err) logrus.Errorf("error closing exec stdin: %+v", err)
} }
} }
} }

View File

@ -293,7 +293,7 @@ func IsInitialized(driverHome string) bool {
return false return false
} }
if err != nil { if err != nil {
logrus.Warnf("Graphdriver.IsInitialized: stat failed: %v", err) logrus.Warnf("graphdriver.IsInitialized: stat failed: %v", err)
} }
return !isEmptyDir(driverHome) return !isEmptyDir(driverHome)
} }

View File

@ -85,7 +85,7 @@ func (fl *follow) waitRead() error {
} }
return errRetry return errRetry
case err := <-fl.fileWatcher.Errors(): case err := <-fl.fileWatcher.Errors():
logrus.Debugf("Logger got error watching file: %v", err) logrus.Debugf("logger got error watching file: %v", err)
// Something happened, let's try and stay alive and create a new watcher // Something happened, let's try and stay alive and create a new watcher
if fl.retries <= 5 { if fl.retries <= 5 {
fl.fileWatcher.Close() fl.fileWatcher.Close()

View File

@ -76,7 +76,7 @@ func (s *fs) Walk(f DigestWalkFunc) error {
for _, v := range dir { for _, v := range dir {
dgst := digest.NewDigestFromHex(string(digest.Canonical), v.Name()) dgst := digest.NewDigestFromHex(string(digest.Canonical), v.Name())
if err := dgst.Validate(); err != nil { if err := dgst.Validate(); err != nil {
logrus.Debugf("Skipping invalid digest %s: %s", dgst, err) logrus.Debugf("skipping invalid digest %s: %s", dgst, err)
continue continue
} }
if err := f(dgst); err != nil { if err := f(dgst); err != nil {

View File

@ -69,19 +69,19 @@ func (is *store) restore() error {
err := is.fs.Walk(func(dgst digest.Digest) error { err := is.fs.Walk(func(dgst digest.Digest) error {
img, err := is.Get(IDFromDigest(dgst)) img, err := is.Get(IDFromDigest(dgst))
if err != nil { if err != nil {
logrus.Errorf("Invalid image %v, %v", dgst, err) logrus.Errorf("invalid image %v, %v", dgst, err)
return nil return nil
} }
var l layer.Layer var l layer.Layer
if chainID := img.RootFS.ChainID(); chainID != "" { if chainID := img.RootFS.ChainID(); chainID != "" {
if !system.IsOSSupported(img.OperatingSystem()) { if !system.IsOSSupported(img.OperatingSystem()) {
logrus.Errorf("Not restoring image with unsupported operating system %v, %v, %s", dgst, chainID, img.OperatingSystem()) logrus.Errorf("not restoring image with unsupported operating system %v, %v, %s", dgst, chainID, img.OperatingSystem())
return nil return nil
} }
l, err = is.lss[img.OperatingSystem()].Get(chainID) l, err = is.lss[img.OperatingSystem()].Get(chainID)
if err != nil { if err != nil {
if err == layer.ErrLayerDoesNotExist { if err == layer.ErrLayerDoesNotExist {
logrus.Errorf("Layer does not exist, not restoring image %v, %v, %s", dgst, chainID, img.OperatingSystem()) logrus.Errorf("layer does not exist, not restoring image %v, %v, %s", dgst, chainID, img.OperatingSystem())
return nil return nil
} }
return err return err
@ -244,7 +244,7 @@ func (is *store) Delete(id ID) ([]layer.Metadata, error) {
} }
if err := is.digestSet.Remove(id.Digest()); err != nil { if err := is.digestSet.Remove(id.Digest()); err != nil {
logrus.Errorf("Error removing %s from digest set: %q", id, err) logrus.Errorf("error removing %s from digest set: %q", id, err)
} }
delete(is.images, id) delete(is.images, id)
is.fs.Delete(id.Digest()) is.fs.Delete(id.Digest())
@ -330,7 +330,7 @@ func (is *store) imagesMap(all bool) map[ID]*Image {
} }
img, err := is.Get(id) img, err := is.Get(id)
if err != nil { if err != nil {
logrus.Errorf("Invalid image access: %q, error: %q", id, err) logrus.Errorf("invalid image access: %q, error: %q", id, err)
continue continue
} }
images[id] = img images[id] = img

View File

@ -339,7 +339,7 @@ func (fms *fileMetadataStore) getOrphan() ([]roLayer, error) {
} }
cacheID := strings.TrimSpace(string(contentBytes)) cacheID := strings.TrimSpace(string(contentBytes))
if cacheID == "" { if cacheID == "" {
logrus.Error("Invalid cache ID") logrus.Error("invalid cache ID")
continue continue
} }

View File

@ -764,9 +764,9 @@ func (ls *layerStore) Cleanup() error {
if err != nil { if err != nil {
logrus.Errorf("Cannot get orphan layers: %v", err) logrus.Errorf("Cannot get orphan layers: %v", err)
} }
logrus.Debugf("Found %v orphan layers", len(orphanLayers)) logrus.Debugf("found %v orphan layers", len(orphanLayers))
for _, orphan := range orphanLayers { for _, orphan := range orphanLayers {
logrus.Debugf("Removing orphan layer, chain ID: %v , cache ID: %v", orphan.chainID, orphan.cacheID) logrus.Debugf("removing orphan layer, chain ID: %v , cache ID: %v", orphan.chainID, orphan.cacheID)
err = ls.driver.Remove(orphan.cacheID) err = ls.driver.Remove(orphan.cacheID)
if err != nil && !os.IsNotExist(err) { if err != nil && !os.IsNotExist(err) {
logrus.WithError(err).WithField("cache-id", orphan.cacheID).Error("cannot remove orphan layer") logrus.WithError(err).WithField("cache-id", orphan.cacheID).Error("cannot remove orphan layer")

View File

@ -165,7 +165,7 @@ func gzDecompress(ctx context.Context, buf io.Reader) (io.ReadCloser, error) {
unpigzPath, err := exec.LookPath("unpigz") unpigzPath, err := exec.LookPath("unpigz")
if err != nil { if err != nil {
logrus.Debugf("Unpigz binary not found, falling back to go gzip library") logrus.Debugf("unpigz binary not found, falling back to go gzip library")
return gzip.NewReader(buf) return gzip.NewReader(buf)
} }

View File

@ -438,7 +438,7 @@ func ExportChanges(dir string, changes []Change, uidMaps, gidMaps []idtools.IDMa
logrus.Debugf("Can't close layer: %s", err) logrus.Debugf("Can't close layer: %s", err)
} }
if err := writer.Close(); err != nil { if err := writer.Close(); err != nil {
logrus.Debugf("Failed close Changes writer: %s", err) logrus.Debugf("failed close Changes writer: %s", err)
} }
}() }()
return reader, nil return reader, nil

View File

@ -108,7 +108,7 @@ func TarResourceRebase(sourcePath, rebaseName string) (content io.ReadCloser, er
sourceDir, sourceBase := SplitPathDirEntry(sourcePath) sourceDir, sourceBase := SplitPathDirEntry(sourcePath)
opts := TarResourceRebaseOpts(sourceBase, rebaseName) opts := TarResourceRebaseOpts(sourceBase, rebaseName)
logrus.Debugf("Copying %q from %q", sourceBase, sourceDir) logrus.Debugf("copying %q from %q", sourceBase, sourceDir)
return TarWithOptions(sourceDir, opts) return TarWithOptions(sourceDir, opts)
} }

View File

@ -159,7 +159,7 @@ func (w *filePoller) watch(f *os.File, lastFi os.FileInfo, chClose chan struct{}
select { select {
case <-timer.C: case <-timer.C:
case <-chClose: case <-chClose:
logrus.Debugf("Watch for %s closed", f.Name()) logrus.Debugf("watch for %s closed", f.Name())
return return
} }

View File

@ -1,4 +1,3 @@
//go:build linux || freebsd || openbsd
// +build linux freebsd openbsd // +build linux freebsd openbsd
// Package kernel provides helper function to get, parse and compare kernel // Package kernel provides helper function to get, parse and compare kernel
@ -26,7 +25,7 @@ func GetKernelVersion() (*VersionInfo, error) {
// the given version. // the given version.
func CheckKernelVersion(k, major, minor int) bool { func CheckKernelVersion(k, major, minor int) bool {
if v, err := GetKernelVersion(); err != nil { if v, err := GetKernelVersion(); err != nil {
logrus.Warnf("Error getting kernel version: %s", err) logrus.Warnf("error getting kernel version: %s", err)
} else { } else {
if CompareKernelVersion(*v, VersionInfo{Kernel: k, Major: major, Minor: minor}) < 0 { if CompareKernelVersion(*v, VersionInfo{Kernel: k, Major: major, Minor: minor}) < 0 {
return false return false

View File

@ -291,7 +291,7 @@ func (hooks *Hooks) MarshalJSON() ([]byte, error) {
case CommandHook: case CommandHook:
serializableHooks = append(serializableHooks, chook) serializableHooks = append(serializableHooks, chook)
default: default:
logrus.Warnf("Cannot serialize hook of type %T, skipping", hook) logrus.Warnf("cannot serialize hook of type %T, skipping", hook)
} }
} }

2
vendor/modules.txt vendored
View File

@ -104,7 +104,7 @@ github.com/ProtonMail/go-crypto/openpgp/s2k
# github.com/acomagu/bufpipe v1.0.3 # github.com/acomagu/bufpipe v1.0.3
## explicit; go 1.12 ## explicit; go 1.12
github.com/acomagu/bufpipe github.com/acomagu/bufpipe
# github.com/aws/aws-sdk-go v1.43.36 # github.com/aws/aws-sdk-go v1.44.24
## explicit; go 1.11 ## explicit; go 1.11
github.com/aws/aws-sdk-go/aws github.com/aws/aws-sdk-go/aws
github.com/aws/aws-sdk-go/aws/arn github.com/aws/aws-sdk-go/aws/arn