diff options
Diffstat (limited to 'vendor/github.com/minio/minio-go')
94 files changed, 0 insertions, 34611 deletions
diff --git a/vendor/github.com/minio/minio-go/v7/.gitignore b/vendor/github.com/minio/minio-go/v7/.gitignore deleted file mode 100644 index 8ae0384..0000000 --- a/vendor/github.com/minio/minio-go/v7/.gitignore +++ /dev/null | |||
@@ -1,6 +0,0 @@ | |||
1 | *~ | ||
2 | *.test | ||
3 | validator | ||
4 | golangci-lint | ||
5 | functional_tests | ||
6 | .idea \ No newline at end of file | ||
diff --git a/vendor/github.com/minio/minio-go/v7/.golangci.yml b/vendor/github.com/minio/minio-go/v7/.golangci.yml deleted file mode 100644 index 875b949..0000000 --- a/vendor/github.com/minio/minio-go/v7/.golangci.yml +++ /dev/null | |||
@@ -1,27 +0,0 @@ | |||
1 | linters-settings: | ||
2 | misspell: | ||
3 | locale: US | ||
4 | |||
5 | linters: | ||
6 | disable-all: true | ||
7 | enable: | ||
8 | - typecheck | ||
9 | - goimports | ||
10 | - misspell | ||
11 | - revive | ||
12 | - govet | ||
13 | - ineffassign | ||
14 | - gosimple | ||
15 | - unused | ||
16 | - gocritic | ||
17 | |||
18 | issues: | ||
19 | exclude-use-default: false | ||
20 | exclude: | ||
21 | # todo fix these when we get enough time. | ||
22 | - "singleCaseSwitch: should rewrite switch statement to if statement" | ||
23 | - "unlambda: replace" | ||
24 | - "captLocal:" | ||
25 | - "ifElseChain:" | ||
26 | - "elseif:" | ||
27 | - "should have a package comment" | ||
diff --git a/vendor/github.com/minio/minio-go/v7/CNAME b/vendor/github.com/minio/minio-go/v7/CNAME deleted file mode 100644 index d365a7b..0000000 --- a/vendor/github.com/minio/minio-go/v7/CNAME +++ /dev/null | |||
@@ -1 +0,0 @@ | |||
1 | minio-go.min.io \ No newline at end of file | ||
diff --git a/vendor/github.com/minio/minio-go/v7/CONTRIBUTING.md b/vendor/github.com/minio/minio-go/v7/CONTRIBUTING.md deleted file mode 100644 index 24522ef..0000000 --- a/vendor/github.com/minio/minio-go/v7/CONTRIBUTING.md +++ /dev/null | |||
@@ -1,22 +0,0 @@ | |||
1 | ### Developer Guidelines | ||
2 | |||
3 | ``minio-go`` welcomes your contribution. To make the process as seamless as possible, we ask for the following: | ||
4 | |||
5 | * Go ahead and fork the project and make your changes. We encourage pull requests to discuss code changes. | ||
6 | - Fork it | ||
7 | - Create your feature branch (git checkout -b my-new-feature) | ||
8 | - Commit your changes (git commit -am 'Add some feature') | ||
9 | - Push to the branch (git push origin my-new-feature) | ||
10 | - Create new Pull Request | ||
11 | |||
12 | * When you're ready to create a pull request, be sure to: | ||
13 | - Have test cases for the new code. If you have questions about how to do it, please ask in your pull request. | ||
14 | - Run `go fmt` | ||
15 | - Squash your commits into a single commit. `git rebase -i`. It's okay to force update your pull request. | ||
16 | - Make sure `go test -race ./...` and `go build` completes. | ||
17 | NOTE: go test runs functional tests and requires you to have a AWS S3 account. Set them as environment variables | ||
18 | ``ACCESS_KEY`` and ``SECRET_KEY``. To run shorter version of the tests please use ``go test -short -race ./...`` | ||
19 | |||
20 | * Read [Effective Go](https://github.com/golang/go/wiki/CodeReviewComments) article from Golang project | ||
21 | - `minio-go` project is strictly conformant with Golang style | ||
22 | - if you happen to observe offending code, please feel free to send a pull request | ||
diff --git a/vendor/github.com/minio/minio-go/v7/LICENSE b/vendor/github.com/minio/minio-go/v7/LICENSE deleted file mode 100644 index d645695..0000000 --- a/vendor/github.com/minio/minio-go/v7/LICENSE +++ /dev/null | |||
@@ -1,202 +0,0 @@ | |||
1 | |||
2 | Apache License | ||
3 | Version 2.0, January 2004 | ||
4 | http://www.apache.org/licenses/ | ||
5 | |||
6 | TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION | ||
7 | |||
8 | 1. Definitions. | ||
9 | |||
10 | "License" shall mean the terms and conditions for use, reproduction, | ||
11 | and distribution as defined by Sections 1 through 9 of this document. | ||
12 | |||
13 | "Licensor" shall mean the copyright owner or entity authorized by | ||
14 | the copyright owner that is granting the License. | ||
15 | |||
16 | "Legal Entity" shall mean the union of the acting entity and all | ||
17 | other entities that control, are controlled by, or are under common | ||
18 | control with that entity. For the purposes of this definition, | ||
19 | "control" means (i) the power, direct or indirect, to cause the | ||
20 | direction or management of such entity, whether by contract or | ||
21 | otherwise, or (ii) ownership of fifty percent (50%) or more of the | ||
22 | outstanding shares, or (iii) beneficial ownership of such entity. | ||
23 | |||
24 | "You" (or "Your") shall mean an individual or Legal Entity | ||
25 | exercising permissions granted by this License. | ||
26 | |||
27 | "Source" form shall mean the preferred form for making modifications, | ||
28 | including but not limited to software source code, documentation | ||
29 | source, and configuration files. | ||
30 | |||
31 | "Object" form shall mean any form resulting from mechanical | ||
32 | transformation or translation of a Source form, including but | ||
33 | not limited to compiled object code, generated documentation, | ||
34 | and conversions to other media types. | ||
35 | |||
36 | "Work" shall mean the work of authorship, whether in Source or | ||
37 | Object form, made available under the License, as indicated by a | ||
38 | copyright notice that is included in or attached to the work | ||
39 | (an example is provided in the Appendix below). | ||
40 | |||
41 | "Derivative Works" shall mean any work, whether in Source or Object | ||
42 | form, that is based on (or derived from) the Work and for which the | ||
43 | editorial revisions, annotations, elaborations, or other modifications | ||
44 | represent, as a whole, an original work of authorship. For the purposes | ||
45 | of this License, Derivative Works shall not include works that remain | ||
46 | separable from, or merely link (or bind by name) to the interfaces of, | ||
47 | the Work and Derivative Works thereof. | ||
48 | |||
49 | "Contribution" shall mean any work of authorship, including | ||
50 | the original version of the Work and any modifications or additions | ||
51 | to that Work or Derivative Works thereof, that is intentionally | ||
52 | submitted to Licensor for inclusion in the Work by the copyright owner | ||
53 | or by an individual or Legal Entity authorized to submit on behalf of | ||
54 | the copyright owner. For the purposes of this definition, "submitted" | ||
55 | means any form of electronic, verbal, or written communication sent | ||
56 | to the Licensor or its representatives, including but not limited to | ||
57 | communication on electronic mailing lists, source code control systems, | ||
58 | and issue tracking systems that are managed by, or on behalf of, the | ||
59 | Licensor for the purpose of discussing and improving the Work, but | ||
60 | excluding communication that is conspicuously marked or otherwise | ||
61 | designated in writing by the copyright owner as "Not a Contribution." | ||
62 | |||
63 | "Contributor" shall mean Licensor and any individual or Legal Entity | ||
64 | on behalf of whom a Contribution has been received by Licensor and | ||
65 | subsequently incorporated within the Work. | ||
66 | |||
67 | 2. Grant of Copyright License. Subject to the terms and conditions of | ||
68 | this License, each Contributor hereby grants to You a perpetual, | ||
69 | worldwide, non-exclusive, no-charge, royalty-free, irrevocable | ||
70 | copyright license to reproduce, prepare Derivative Works of, | ||
71 | publicly display, publicly perform, sublicense, and distribute the | ||
72 | Work and such Derivative Works in Source or Object form. | ||
73 | |||
74 | 3. Grant of Patent License. Subject to the terms and conditions of | ||
75 | this License, each Contributor hereby grants to You a perpetual, | ||
76 | worldwide, non-exclusive, no-charge, royalty-free, irrevocable | ||
77 | (except as stated in this section) patent license to make, have made, | ||
78 | use, offer to sell, sell, import, and otherwise transfer the Work, | ||
79 | where such license applies only to those patent claims licensable | ||
80 | by such Contributor that are necessarily infringed by their | ||
81 | Contribution(s) alone or by combination of their Contribution(s) | ||
82 | with the Work to which such Contribution(s) was submitted. If You | ||
83 | institute patent litigation against any entity (including a | ||
84 | cross-claim or counterclaim in a lawsuit) alleging that the Work | ||
85 | or a Contribution incorporated within the Work constitutes direct | ||
86 | or contributory patent infringement, then any patent licenses | ||
87 | granted to You under this License for that Work shall terminate | ||
88 | as of the date such litigation is filed. | ||
89 | |||
90 | 4. Redistribution. You may reproduce and distribute copies of the | ||
91 | Work or Derivative Works thereof in any medium, with or without | ||
92 | modifications, and in Source or Object form, provided that You | ||
93 | meet the following conditions: | ||
94 | |||
95 | (a) You must give any other recipients of the Work or | ||
96 | Derivative Works a copy of this License; and | ||
97 | |||
98 | (b) You must cause any modified files to carry prominent notices | ||
99 | stating that You changed the files; and | ||
100 | |||
101 | (c) You must retain, in the Source form of any Derivative Works | ||
102 | that You distribute, all copyright, patent, trademark, and | ||
103 | attribution notices from the Source form of the Work, | ||
104 | excluding those notices that do not pertain to any part of | ||
105 | the Derivative Works; and | ||
106 | |||
107 | (d) If the Work includes a "NOTICE" text file as part of its | ||
108 | distribution, then any Derivative Works that You distribute must | ||
109 | include a readable copy of the attribution notices contained | ||
110 | within such NOTICE file, excluding those notices that do not | ||
111 | pertain to any part of the Derivative Works, in at least one | ||
112 | of the following places: within a NOTICE text file distributed | ||
113 | as part of the Derivative Works; within the Source form or | ||
114 | documentation, if provided along with the Derivative Works; or, | ||
115 | within a display generated by the Derivative Works, if and | ||
116 | wherever such third-party notices normally appear. The contents | ||
117 | of the NOTICE file are for informational purposes only and | ||
118 | do not modify the License. You may add Your own attribution | ||
119 | notices within Derivative Works that You distribute, alongside | ||
120 | or as an addendum to the NOTICE text from the Work, provided | ||
121 | that such additional attribution notices cannot be construed | ||
122 | as modifying the License. | ||
123 | |||
124 | You may add Your own copyright statement to Your modifications and | ||
125 | may provide additional or different license terms and conditions | ||
126 | for use, reproduction, or distribution of Your modifications, or | ||
127 | for any such Derivative Works as a whole, provided Your use, | ||
128 | reproduction, and distribution of the Work otherwise complies with | ||
129 | the conditions stated in this License. | ||
130 | |||
131 | 5. Submission of Contributions. Unless You explicitly state otherwise, | ||
132 | any Contribution intentionally submitted for inclusion in the Work | ||
133 | by You to the Licensor shall be under the terms and conditions of | ||
134 | this License, without any additional terms or conditions. | ||
135 | Notwithstanding the above, nothing herein shall supersede or modify | ||
136 | the terms of any separate license agreement you may have executed | ||
137 | with Licensor regarding such Contributions. | ||
138 | |||
139 | 6. Trademarks. This License does not grant permission to use the trade | ||
140 | names, trademarks, service marks, or product names of the Licensor, | ||
141 | except as required for reasonable and customary use in describing the | ||
142 | origin of the Work and reproducing the content of the NOTICE file. | ||
143 | |||
144 | 7. Disclaimer of Warranty. Unless required by applicable law or | ||
145 | agreed to in writing, Licensor provides the Work (and each | ||
146 | Contributor provides its Contributions) on an "AS IS" BASIS, | ||
147 | WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or | ||
148 | implied, including, without limitation, any warranties or conditions | ||
149 | of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A | ||
150 | PARTICULAR PURPOSE. You are solely responsible for determining the | ||
151 | appropriateness of using or redistributing the Work and assume any | ||
152 | risks associated with Your exercise of permissions under this License. | ||
153 | |||
154 | 8. Limitation of Liability. In no event and under no legal theory, | ||
155 | whether in tort (including negligence), contract, or otherwise, | ||
156 | unless required by applicable law (such as deliberate and grossly | ||
157 | negligent acts) or agreed to in writing, shall any Contributor be | ||
158 | liable to You for damages, including any direct, indirect, special, | ||
159 | incidental, or consequential damages of any character arising as a | ||
160 | result of this License or out of the use or inability to use the | ||
161 | Work (including but not limited to damages for loss of goodwill, | ||
162 | work stoppage, computer failure or malfunction, or any and all | ||
163 | other commercial damages or losses), even if such Contributor | ||
164 | has been advised of the possibility of such damages. | ||
165 | |||
166 | 9. Accepting Warranty or Additional Liability. While redistributing | ||
167 | the Work or Derivative Works thereof, You may choose to offer, | ||
168 | and charge a fee for, acceptance of support, warranty, indemnity, | ||
169 | or other liability obligations and/or rights consistent with this | ||
170 | License. However, in accepting such obligations, You may act only | ||
171 | on Your own behalf and on Your sole responsibility, not on behalf | ||
172 | of any other Contributor, and only if You agree to indemnify, | ||
173 | defend, and hold each Contributor harmless for any liability | ||
174 | incurred by, or claims asserted against, such Contributor by reason | ||
175 | of your accepting any such warranty or additional liability. | ||
176 | |||
177 | END OF TERMS AND CONDITIONS | ||
178 | |||
179 | APPENDIX: How to apply the Apache License to your work. | ||
180 | |||
181 | To apply the Apache License to your work, attach the following | ||
182 | boilerplate notice, with the fields enclosed by brackets "[]" | ||
183 | replaced with your own identifying information. (Don't include | ||
184 | the brackets!) The text should be enclosed in the appropriate | ||
185 | comment syntax for the file format. We also recommend that a | ||
186 | file or class name and description of purpose be included on the | ||
187 | same "printed page" as the copyright notice for easier | ||
188 | identification within third-party archives. | ||
189 | |||
190 | Copyright [yyyy] [name of copyright owner] | ||
191 | |||
192 | Licensed under the Apache License, Version 2.0 (the "License"); | ||
193 | you may not use this file except in compliance with the License. | ||
194 | You may obtain a copy of the License at | ||
195 | |||
196 | http://www.apache.org/licenses/LICENSE-2.0 | ||
197 | |||
198 | Unless required by applicable law or agreed to in writing, software | ||
199 | distributed under the License is distributed on an "AS IS" BASIS, | ||
200 | WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
201 | See the License for the specific language governing permissions and | ||
202 | limitations under the License. | ||
diff --git a/vendor/github.com/minio/minio-go/v7/MAINTAINERS.md b/vendor/github.com/minio/minio-go/v7/MAINTAINERS.md deleted file mode 100644 index f640dfb..0000000 --- a/vendor/github.com/minio/minio-go/v7/MAINTAINERS.md +++ /dev/null | |||
@@ -1,35 +0,0 @@ | |||
1 | # For maintainers only | ||
2 | |||
3 | ## Responsibilities | ||
4 | |||
5 | Please go through this link [Maintainer Responsibility](https://gist.github.com/abperiasamy/f4d9b31d3186bbd26522) | ||
6 | |||
7 | ### Making new releases | ||
8 | Tag and sign your release commit, additionally this step requires you to have access to MinIO's trusted private key. | ||
9 | ```sh | ||
10 | $ export GNUPGHOME=/media/${USER}/minio/trusted | ||
11 | $ git tag -s 4.0.0 | ||
12 | $ git push | ||
13 | $ git push --tags | ||
14 | ``` | ||
15 | |||
16 | ### Update version | ||
17 | Once release has been made update `libraryVersion` constant in `api.go` to next to be released version. | ||
18 | |||
19 | ```sh | ||
20 | $ grep libraryVersion api.go | ||
21 | libraryVersion = "4.0.1" | ||
22 | ``` | ||
23 | |||
24 | Commit your changes | ||
25 | ``` | ||
26 | $ git commit -a -m "Update version for next release" --author "MinIO Trusted <[email protected]>" | ||
27 | ``` | ||
28 | |||
29 | ### Announce | ||
30 | Announce new release by adding release notes at https://github.com/minio/minio-go/releases from `[email protected]` account. Release notes requires two sections `highlights` and `changelog`. Highlights is a bulleted list of salient features in this release and Changelog contains list of all commits since the last release. | ||
31 | |||
32 | To generate `changelog` | ||
33 | ```sh | ||
34 | $ git log --no-color --pretty=format:'-%d %s (%cr) <%an>' <last_release_tag>..<latest_release_tag> | ||
35 | ``` | ||
diff --git a/vendor/github.com/minio/minio-go/v7/Makefile b/vendor/github.com/minio/minio-go/v7/Makefile deleted file mode 100644 index 68444aa..0000000 --- a/vendor/github.com/minio/minio-go/v7/Makefile +++ /dev/null | |||
@@ -1,38 +0,0 @@ | |||
1 | GOPATH := $(shell go env GOPATH) | ||
2 | TMPDIR := $(shell mktemp -d) | ||
3 | |||
4 | all: checks | ||
5 | |||
6 | .PHONY: examples docs | ||
7 | |||
8 | checks: lint vet test examples functional-test | ||
9 | |||
10 | lint: | ||
11 | @mkdir -p ${GOPATH}/bin | ||
12 | @echo "Installing golangci-lint" && curl -sSfL https://raw.githubusercontent.com/golangci/golangci-lint/master/install.sh | sh -s -- -b $(GOPATH)/bin | ||
13 | @echo "Running $@ check" | ||
14 | @GO111MODULE=on ${GOPATH}/bin/golangci-lint cache clean | ||
15 | @GO111MODULE=on ${GOPATH}/bin/golangci-lint run --timeout=5m --config ./.golangci.yml | ||
16 | |||
17 | vet: | ||
18 | @GO111MODULE=on go vet ./... | ||
19 | @echo "Installing staticcheck" && go install honnef.co/go/tools/cmd/staticcheck@latest | ||
20 | ${GOPATH}/bin/staticcheck -tests=false -checks="all,-ST1000,-ST1003,-ST1016,-ST1020,-ST1021,-ST1022,-ST1023,-ST1005" | ||
21 | |||
22 | test: | ||
23 | @GO111MODULE=on SERVER_ENDPOINT=localhost:9000 ACCESS_KEY=minioadmin SECRET_KEY=minioadmin ENABLE_HTTPS=1 MINT_MODE=full go test -race -v ./... | ||
24 | |||
25 | examples: | ||
26 | @echo "Building s3 examples" | ||
27 | @cd ./examples/s3 && $(foreach v,$(wildcard examples/s3/*.go),go build -mod=mod -o ${TMPDIR}/$(basename $(v)) $(notdir $(v)) || exit 1;) | ||
28 | @echo "Building minio examples" | ||
29 | @cd ./examples/minio && $(foreach v,$(wildcard examples/minio/*.go),go build -mod=mod -o ${TMPDIR}/$(basename $(v)) $(notdir $(v)) || exit 1;) | ||
30 | |||
31 | functional-test: | ||
32 | @GO111MODULE=on go build -race functional_tests.go | ||
33 | @SERVER_ENDPOINT=localhost:9000 ACCESS_KEY=minioadmin SECRET_KEY=minioadmin ENABLE_HTTPS=1 MINT_MODE=full ./functional_tests | ||
34 | |||
35 | clean: | ||
36 | @echo "Cleaning up all the generated files" | ||
37 | @find . -name '*.test' | xargs rm -fv | ||
38 | @find . -name '*~' | xargs rm -fv | ||
diff --git a/vendor/github.com/minio/minio-go/v7/NOTICE b/vendor/github.com/minio/minio-go/v7/NOTICE deleted file mode 100644 index 1e8fd3b..0000000 --- a/vendor/github.com/minio/minio-go/v7/NOTICE +++ /dev/null | |||
@@ -1,9 +0,0 @@ | |||
1 | MinIO Cloud Storage, (C) 2014-2020 MinIO, Inc. | ||
2 | |||
3 | This product includes software developed at MinIO, Inc. | ||
4 | (https://min.io/). | ||
5 | |||
6 | The MinIO project contains unmodified/modified subcomponents too with | ||
7 | separate copyright notices and license terms. Your use of the source | ||
8 | code for these subcomponents is subject to the terms and conditions | ||
9 | of Apache License Version 2.0 | ||
diff --git a/vendor/github.com/minio/minio-go/v7/README.md b/vendor/github.com/minio/minio-go/v7/README.md deleted file mode 100644 index 82f70a1..0000000 --- a/vendor/github.com/minio/minio-go/v7/README.md +++ /dev/null | |||
@@ -1,312 +0,0 @@ | |||
1 | # MinIO Go Client SDK for Amazon S3 Compatible Cloud Storage [![Slack](https://slack.min.io/slack?type=svg)](https://slack.min.io) [![Sourcegraph](https://sourcegraph.com/github.com/minio/minio-go/-/badge.svg)](https://sourcegraph.com/github.com/minio/minio-go?badge) [![Apache V2 License](https://img.shields.io/badge/license-Apache%20V2-blue.svg)](https://github.com/minio/minio-go/blob/master/LICENSE) | ||
2 | |||
3 | The MinIO Go Client SDK provides straightforward APIs to access any Amazon S3 compatible object storage. | ||
4 | |||
5 | This Quickstart Guide covers how to install the MinIO client SDK, connect to MinIO, and create a sample file uploader. | ||
6 | For a complete list of APIs and examples, see the [godoc documentation](https://pkg.go.dev/github.com/minio/minio-go/v7) or [Go Client API Reference](https://min.io/docs/minio/linux/developers/go/API.html). | ||
7 | |||
8 | These examples presume a working [Go development environment](https://golang.org/doc/install) and the [MinIO `mc` command line tool](https://min.io/docs/minio/linux/reference/minio-mc.html). | ||
9 | |||
10 | ## Download from Github | ||
11 | |||
12 | From your project directory: | ||
13 | |||
14 | ```sh | ||
15 | go get github.com/minio/minio-go/v7 | ||
16 | ``` | ||
17 | |||
18 | ## Initialize a MinIO Client Object | ||
19 | |||
20 | The MinIO client requires the following parameters to connect to an Amazon S3 compatible object storage: | ||
21 | |||
22 | | Parameter | Description | | ||
23 | | ----------------- | ---------------------------------------------------------- | | ||
24 | | `endpoint` | URL to object storage service. | | ||
25 | | `_minio.Options_` | All the options such as credentials, custom transport etc. | | ||
26 | |||
27 | ```go | ||
28 | package main | ||
29 | |||
30 | import ( | ||
31 | "log" | ||
32 | |||
33 | "github.com/minio/minio-go/v7" | ||
34 | "github.com/minio/minio-go/v7/pkg/credentials" | ||
35 | ) | ||
36 | |||
37 | func main() { | ||
38 | endpoint := "play.min.io" | ||
39 | accessKeyID := "Q3AM3UQ867SPQQA43P2F" | ||
40 | secretAccessKey := "zuf+tfteSlswRu7BJ86wekitnifILbZam1KYY3TG" | ||
41 | useSSL := true | ||
42 | |||
43 | // Initialize minio client object. | ||
44 | minioClient, err := minio.New(endpoint, &minio.Options{ | ||
45 | Creds: credentials.NewStaticV4(accessKeyID, secretAccessKey, ""), | ||
46 | Secure: useSSL, | ||
47 | }) | ||
48 | if err != nil { | ||
49 | log.Fatalln(err) | ||
50 | } | ||
51 | |||
52 | log.Printf("%#v\n", minioClient) // minioClient is now set up | ||
53 | } | ||
54 | ``` | ||
55 | |||
56 | ## Example - File Uploader | ||
57 | |||
58 | This sample code connects to an object storage server, creates a bucket, and uploads a file to the bucket. | ||
59 | It uses the MinIO `play` server, a public MinIO cluster located at [https://play.min.io](https://play.min.io). | ||
60 | |||
61 | The `play` server runs the latest stable version of MinIO and may be used for testing and development. | ||
62 | The access credentials shown in this example are open to the public and all data uploaded to `play` should be considered public and non-protected. | ||
63 | |||
64 | ### FileUploader.go | ||
65 | |||
66 | This example does the following: | ||
67 | |||
68 | - Connects to the MinIO `play` server using the provided credentials. | ||
69 | - Creates a bucket named `testbucket`. | ||
70 | - Uploads a file named `testdata` from `/tmp`. | ||
71 | - Verifies the file was created using `mc ls`. | ||
72 | |||
73 | ```go | ||
74 | // FileUploader.go MinIO example | ||
75 | package main | ||
76 | |||
77 | import ( | ||
78 | "context" | ||
79 | "log" | ||
80 | |||
81 | "github.com/minio/minio-go/v7" | ||
82 | "github.com/minio/minio-go/v7/pkg/credentials" | ||
83 | ) | ||
84 | |||
85 | func main() { | ||
86 | ctx := context.Background() | ||
87 | endpoint := "play.min.io" | ||
88 | accessKeyID := "Q3AM3UQ867SPQQA43P2F" | ||
89 | secretAccessKey := "zuf+tfteSlswRu7BJ86wekitnifILbZam1KYY3TG" | ||
90 | useSSL := true | ||
91 | |||
92 | // Initialize minio client object. | ||
93 | minioClient, err := minio.New(endpoint, &minio.Options{ | ||
94 | Creds: credentials.NewStaticV4(accessKeyID, secretAccessKey, ""), | ||
95 | Secure: useSSL, | ||
96 | }) | ||
97 | if err != nil { | ||
98 | log.Fatalln(err) | ||
99 | } | ||
100 | |||
101 | // Make a new bucket called testbucket. | ||
102 | bucketName := "testbucket" | ||
103 | location := "us-east-1" | ||
104 | |||
105 | err = minioClient.MakeBucket(ctx, bucketName, minio.MakeBucketOptions{Region: location}) | ||
106 | if err != nil { | ||
107 | // Check to see if we already own this bucket (which happens if you run this twice) | ||
108 | exists, errBucketExists := minioClient.BucketExists(ctx, bucketName) | ||
109 | if errBucketExists == nil && exists { | ||
110 | log.Printf("We already own %s\n", bucketName) | ||
111 | } else { | ||
112 | log.Fatalln(err) | ||
113 | } | ||
114 | } else { | ||
115 | log.Printf("Successfully created %s\n", bucketName) | ||
116 | } | ||
117 | |||
118 | // Upload the test file | ||
119 | // Change the value of filePath if the file is in another location | ||
120 | objectName := "testdata" | ||
121 | filePath := "/tmp/testdata" | ||
122 | contentType := "application/octet-stream" | ||
123 | |||
124 | // Upload the test file with FPutObject | ||
125 | info, err := minioClient.FPutObject(ctx, bucketName, objectName, filePath, minio.PutObjectOptions{ContentType: contentType}) | ||
126 | if err != nil { | ||
127 | log.Fatalln(err) | ||
128 | } | ||
129 | |||
130 | log.Printf("Successfully uploaded %s of size %d\n", objectName, info.Size) | ||
131 | } | ||
132 | ``` | ||
133 | |||
134 | **1. Create a test file containing data:** | ||
135 | |||
136 | You can do this with `dd` on Linux or macOS systems: | ||
137 | |||
138 | ```sh | ||
139 | dd if=/dev/urandom of=/tmp/testdata bs=2048 count=10 | ||
140 | ``` | ||
141 | |||
142 | or `fsutil` on Windows: | ||
143 | |||
144 | ```sh | ||
145 | fsutil file createnew "C:\Users\<username>\Desktop\sample.txt" 20480 | ||
146 | ``` | ||
147 | |||
148 | **2. Run FileUploader with the following commands:** | ||
149 | |||
150 | ```sh | ||
151 | go mod init example/FileUploader | ||
152 | go get github.com/minio/minio-go/v7 | ||
153 | go get github.com/minio/minio-go/v7/pkg/credentials | ||
154 | go run FileUploader.go | ||
155 | ``` | ||
156 | |||
157 | The output resembles the following: | ||
158 | |||
159 | ```sh | ||
160 | 2023/11/01 14:27:55 Successfully created testbucket | ||
161 | 2023/11/01 14:27:55 Successfully uploaded testdata of size 20480 | ||
162 | ``` | ||
163 | |||
164 | **3. Verify the Uploaded File With `mc ls`:** | ||
165 | |||
166 | ```sh | ||
167 | mc ls play/testbucket | ||
168 | [2023-11-01 14:27:55 UTC] 20KiB STANDARD TestDataFile | ||
169 | ``` | ||
170 | |||
171 | ## API Reference | ||
172 | |||
173 | The full API Reference is available here. | ||
174 | |||
175 | * [Complete API Reference](https://min.io/docs/minio/linux/developers/go/API.html) | ||
176 | |||
177 | ### API Reference : Bucket Operations | ||
178 | |||
179 | * [`MakeBucket`](https://min.io/docs/minio/linux/developers/go/API.html#MakeBucket) | ||
180 | * [`ListBuckets`](https://min.io/docs/minio/linux/developers/go/API.html#ListBuckets) | ||
181 | * [`BucketExists`](https://min.io/docs/minio/linux/developers/go/API.html#BucketExists) | ||
182 | * [`RemoveBucket`](https://min.io/docs/minio/linux/developers/go/API.html#RemoveBucket) | ||
183 | * [`ListObjects`](https://min.io/docs/minio/linux/developers/go/API.html#ListObjects) | ||
184 | * [`ListIncompleteUploads`](https://min.io/docs/minio/linux/developers/go/API.html#ListIncompleteUploads) | ||
185 | |||
186 | ### API Reference : Bucket policy Operations | ||
187 | |||
188 | * [`SetBucketPolicy`](https://min.io/docs/minio/linux/developers/go/API.html#SetBucketPolicy) | ||
189 | * [`GetBucketPolicy`](https://min.io/docs/minio/linux/developers/go/API.html#GetBucketPolicy) | ||
190 | |||
191 | ### API Reference : Bucket notification Operations | ||
192 | |||
193 | * [`SetBucketNotification`](https://min.io/docs/minio/linux/developers/go/API.html#SetBucketNotification) | ||
194 | * [`GetBucketNotification`](https://min.io/docs/minio/linux/developers/go/API.html#GetBucketNotification) | ||
195 | * [`RemoveAllBucketNotification`](https://min.io/docs/minio/linux/developers/go/API.html#RemoveAllBucketNotification) | ||
196 | * [`ListenBucketNotification`](https://min.io/docs/minio/linux/developers/go/API.html#ListenBucketNotification) (MinIO Extension) | ||
197 | * [`ListenNotification`](https://min.io/docs/minio/linux/developers/go/API.html#ListenNotification) (MinIO Extension) | ||
198 | |||
199 | ### API Reference : File Object Operations | ||
200 | |||
201 | * [`FPutObject`](https://min.io/docs/minio/linux/developers/go/API.html#FPutObject) | ||
202 | * [`FGetObject`](https://min.io/docs/minio/linux/developers/go/API.html#FGetObject) | ||
203 | |||
204 | ### API Reference : Object Operations | ||
205 | |||
206 | * [`GetObject`](https://min.io/docs/minio/linux/developers/go/API.html#GetObject) | ||
207 | * [`PutObject`](https://min.io/docs/minio/linux/developers/go/API.html#PutObject) | ||
208 | * [`PutObjectStreaming`](https://min.io/docs/minio/linux/developers/go/API.html#PutObjectStreaming) | ||
209 | * [`StatObject`](https://min.io/docs/minio/linux/developers/go/API.html#StatObject) | ||
210 | * [`CopyObject`](https://min.io/docs/minio/linux/developers/go/API.html#CopyObject) | ||
211 | * [`RemoveObject`](https://min.io/docs/minio/linux/developers/go/API.html#RemoveObject) | ||
212 | * [`RemoveObjects`](https://min.io/docs/minio/linux/developers/go/API.html#RemoveObjects) | ||
213 | * [`RemoveIncompleteUpload`](https://min.io/docs/minio/linux/developers/go/API.html#RemoveIncompleteUpload) | ||
214 | * [`SelectObjectContent`](https://min.io/docs/minio/linux/developers/go/API.html#SelectObjectContent) | ||
215 | |||
216 | ### API Reference : Presigned Operations | ||
217 | |||
218 | * [`PresignedGetObject`](https://min.io/docs/minio/linux/developers/go/API.html#PresignedGetObject) | ||
219 | * [`PresignedPutObject`](https://min.io/docs/minio/linux/developers/go/API.html#PresignedPutObject) | ||
220 | * [`PresignedHeadObject`](https://min.io/docs/minio/linux/developers/go/API.html#PresignedHeadObject) | ||
221 | * [`PresignedPostPolicy`](https://min.io/docs/minio/linux/developers/go/API.html#PresignedPostPolicy) | ||
222 | |||
223 | ### API Reference : Client custom settings | ||
224 | |||
225 | * [`SetAppInfo`](https://min.io/docs/minio/linux/developers/go/API.html#SetAppInfo) | ||
226 | * [`TraceOn`](https://min.io/docs/minio/linux/developers/go/API.html#TraceOn) | ||
227 | * [`TraceOff`](https://min.io/docs/minio/linux/developers/go/API.html#TraceOff) | ||
228 | |||
229 | ## Full Examples | ||
230 | |||
231 | ### Full Examples : Bucket Operations | ||
232 | |||
233 | * [makebucket.go](https://github.com/minio/minio-go/blob/master/examples/s3/makebucket.go) | ||
234 | * [listbuckets.go](https://github.com/minio/minio-go/blob/master/examples/s3/listbuckets.go) | ||
235 | * [bucketexists.go](https://github.com/minio/minio-go/blob/master/examples/s3/bucketexists.go) | ||
236 | * [removebucket.go](https://github.com/minio/minio-go/blob/master/examples/s3/removebucket.go) | ||
237 | * [listobjects.go](https://github.com/minio/minio-go/blob/master/examples/s3/listobjects.go) | ||
238 | * [listobjectsV2.go](https://github.com/minio/minio-go/blob/master/examples/s3/listobjectsV2.go) | ||
239 | * [listincompleteuploads.go](https://github.com/minio/minio-go/blob/master/examples/s3/listincompleteuploads.go) | ||
240 | |||
241 | ### Full Examples : Bucket policy Operations | ||
242 | |||
243 | * [setbucketpolicy.go](https://github.com/minio/minio-go/blob/master/examples/s3/setbucketpolicy.go) | ||
244 | * [getbucketpolicy.go](https://github.com/minio/minio-go/blob/master/examples/s3/getbucketpolicy.go) | ||
245 | * [listbucketpolicies.go](https://github.com/minio/minio-go/blob/master/examples/s3/listbucketpolicies.go) | ||
246 | |||
247 | ### Full Examples : Bucket lifecycle Operations | ||
248 | |||
249 | * [setbucketlifecycle.go](https://github.com/minio/minio-go/blob/master/examples/s3/setbucketlifecycle.go) | ||
250 | * [getbucketlifecycle.go](https://github.com/minio/minio-go/blob/master/examples/s3/getbucketlifecycle.go) | ||
251 | |||
252 | ### Full Examples : Bucket encryption Operations | ||
253 | |||
254 | * [setbucketencryption.go](https://github.com/minio/minio-go/blob/master/examples/s3/setbucketencryption.go) | ||
255 | * [getbucketencryption.go](https://github.com/minio/minio-go/blob/master/examples/s3/getbucketencryption.go) | ||
256 | * [deletebucketencryption.go](https://github.com/minio/minio-go/blob/master/examples/s3/deletebucketencryption.go) | ||
257 | |||
258 | ### Full Examples : Bucket replication Operations | ||
259 | |||
260 | * [setbucketreplication.go](https://github.com/minio/minio-go/blob/master/examples/s3/setbucketreplication.go) | ||
261 | * [getbucketreplication.go](https://github.com/minio/minio-go/blob/master/examples/s3/getbucketreplication.go) | ||
262 | * [removebucketreplication.go](https://github.com/minio/minio-go/blob/master/examples/s3/removebucketreplication.go) | ||
263 | |||
264 | ### Full Examples : Bucket notification Operations | ||
265 | |||
266 | * [setbucketnotification.go](https://github.com/minio/minio-go/blob/master/examples/s3/setbucketnotification.go) | ||
267 | * [getbucketnotification.go](https://github.com/minio/minio-go/blob/master/examples/s3/getbucketnotification.go) | ||
268 | * [removeallbucketnotification.go](https://github.com/minio/minio-go/blob/master/examples/s3/removeallbucketnotification.go) | ||
269 | * [listenbucketnotification.go](https://github.com/minio/minio-go/blob/master/examples/minio/listenbucketnotification.go) (MinIO Extension) | ||
270 | * [listennotification.go](https://github.com/minio/minio-go/blob/master/examples/minio/listen-notification.go) (MinIO Extension) | ||
271 | |||
272 | ### Full Examples : File Object Operations | ||
273 | |||
274 | * [fputobject.go](https://github.com/minio/minio-go/blob/master/examples/s3/fputobject.go) | ||
275 | * [fgetobject.go](https://github.com/minio/minio-go/blob/master/examples/s3/fgetobject.go) | ||
276 | |||
277 | ### Full Examples : Object Operations | ||
278 | |||
279 | * [putobject.go](https://github.com/minio/minio-go/blob/master/examples/s3/putobject.go) | ||
280 | * [getobject.go](https://github.com/minio/minio-go/blob/master/examples/s3/getobject.go) | ||
281 | * [statobject.go](https://github.com/minio/minio-go/blob/master/examples/s3/statobject.go) | ||
282 | * [copyobject.go](https://github.com/minio/minio-go/blob/master/examples/s3/copyobject.go) | ||
283 | * [removeobject.go](https://github.com/minio/minio-go/blob/master/examples/s3/removeobject.go) | ||
284 | * [removeincompleteupload.go](https://github.com/minio/minio-go/blob/master/examples/s3/removeincompleteupload.go) | ||
285 | * [removeobjects.go](https://github.com/minio/minio-go/blob/master/examples/s3/removeobjects.go) | ||
286 | |||
287 | ### Full Examples : Encrypted Object Operations | ||
288 | |||
289 | * [put-encrypted-object.go](https://github.com/minio/minio-go/blob/master/examples/s3/put-encrypted-object.go) | ||
290 | * [get-encrypted-object.go](https://github.com/minio/minio-go/blob/master/examples/s3/get-encrypted-object.go) | ||
291 | * [fput-encrypted-object.go](https://github.com/minio/minio-go/blob/master/examples/s3/fputencrypted-object.go) | ||
292 | |||
293 | ### Full Examples : Presigned Operations | ||
294 | |||
295 | * [presignedgetobject.go](https://github.com/minio/minio-go/blob/master/examples/s3/presignedgetobject.go) | ||
296 | * [presignedputobject.go](https://github.com/minio/minio-go/blob/master/examples/s3/presignedputobject.go) | ||
297 | * [presignedheadobject.go](https://github.com/minio/minio-go/blob/master/examples/s3/presignedheadobject.go) | ||
298 | * [presignedpostpolicy.go](https://github.com/minio/minio-go/blob/master/examples/s3/presignedpostpolicy.go) | ||
299 | |||
300 | ## Explore Further | ||
301 | |||
302 | * [Godoc Documentation](https://pkg.go.dev/github.com/minio/minio-go/v7) | ||
303 | * [Complete Documentation](https://min.io/docs/minio/kubernetes/upstream/index.html) | ||
304 | * [MinIO Go Client SDK API Reference](https://min.io/docs/minio/linux/developers/go/API.html) | ||
305 | |||
306 | ## Contribute | ||
307 | |||
308 | [Contributors Guide](https://github.com/minio/minio-go/blob/master/CONTRIBUTING.md) | ||
309 | |||
310 | ## License | ||
311 | |||
312 | This SDK is distributed under the [Apache License, Version 2.0](https://www.apache.org/licenses/LICENSE-2.0), see [LICENSE](https://github.com/minio/minio-go/blob/master/LICENSE) and [NOTICE](https://github.com/minio/minio-go/blob/master/NOTICE) for more information. | ||
diff --git a/vendor/github.com/minio/minio-go/v7/api-bucket-encryption.go b/vendor/github.com/minio/minio-go/v7/api-bucket-encryption.go deleted file mode 100644 index 24f94e0..0000000 --- a/vendor/github.com/minio/minio-go/v7/api-bucket-encryption.go +++ /dev/null | |||
@@ -1,134 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2020 MinIO, Inc. | ||
4 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
5 | * you may not use this file except in compliance with the License. | ||
6 | * You may obtain a copy of the License at | ||
7 | * | ||
8 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
9 | * | ||
10 | * Unless required by applicable law or agreed to in writing, software | ||
11 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
12 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
13 | * See the License for the specific language governing permissions and | ||
14 | * limitations under the License. | ||
15 | */ | ||
16 | |||
17 | package minio | ||
18 | |||
19 | import ( | ||
20 | "bytes" | ||
21 | "context" | ||
22 | "encoding/xml" | ||
23 | "net/http" | ||
24 | "net/url" | ||
25 | |||
26 | "github.com/minio/minio-go/v7/pkg/s3utils" | ||
27 | "github.com/minio/minio-go/v7/pkg/sse" | ||
28 | ) | ||
29 | |||
30 | // SetBucketEncryption sets the default encryption configuration on an existing bucket. | ||
31 | func (c *Client) SetBucketEncryption(ctx context.Context, bucketName string, config *sse.Configuration) error { | ||
32 | // Input validation. | ||
33 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
34 | return err | ||
35 | } | ||
36 | |||
37 | if config == nil { | ||
38 | return errInvalidArgument("configuration cannot be empty") | ||
39 | } | ||
40 | |||
41 | buf, err := xml.Marshal(config) | ||
42 | if err != nil { | ||
43 | return err | ||
44 | } | ||
45 | |||
46 | // Get resources properly escaped and lined up before | ||
47 | // using them in http request. | ||
48 | urlValues := make(url.Values) | ||
49 | urlValues.Set("encryption", "") | ||
50 | |||
51 | // Content-length is mandatory to set a default encryption configuration | ||
52 | reqMetadata := requestMetadata{ | ||
53 | bucketName: bucketName, | ||
54 | queryValues: urlValues, | ||
55 | contentBody: bytes.NewReader(buf), | ||
56 | contentLength: int64(len(buf)), | ||
57 | contentMD5Base64: sumMD5Base64(buf), | ||
58 | } | ||
59 | |||
60 | // Execute PUT to upload a new bucket default encryption configuration. | ||
61 | resp, err := c.executeMethod(ctx, http.MethodPut, reqMetadata) | ||
62 | defer closeResponse(resp) | ||
63 | if err != nil { | ||
64 | return err | ||
65 | } | ||
66 | if resp.StatusCode != http.StatusOK { | ||
67 | return httpRespToErrorResponse(resp, bucketName, "") | ||
68 | } | ||
69 | return nil | ||
70 | } | ||
71 | |||
72 | // RemoveBucketEncryption removes the default encryption configuration on a bucket with a context to control cancellations and timeouts. | ||
73 | func (c *Client) RemoveBucketEncryption(ctx context.Context, bucketName string) error { | ||
74 | // Input validation. | ||
75 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
76 | return err | ||
77 | } | ||
78 | |||
79 | // Get resources properly escaped and lined up before | ||
80 | // using them in http request. | ||
81 | urlValues := make(url.Values) | ||
82 | urlValues.Set("encryption", "") | ||
83 | |||
84 | // DELETE default encryption configuration on a bucket. | ||
85 | resp, err := c.executeMethod(ctx, http.MethodDelete, requestMetadata{ | ||
86 | bucketName: bucketName, | ||
87 | queryValues: urlValues, | ||
88 | contentSHA256Hex: emptySHA256Hex, | ||
89 | }) | ||
90 | defer closeResponse(resp) | ||
91 | if err != nil { | ||
92 | return err | ||
93 | } | ||
94 | if resp.StatusCode != http.StatusOK && resp.StatusCode != http.StatusNoContent { | ||
95 | return httpRespToErrorResponse(resp, bucketName, "") | ||
96 | } | ||
97 | return nil | ||
98 | } | ||
99 | |||
100 | // GetBucketEncryption gets the default encryption configuration | ||
101 | // on an existing bucket with a context to control cancellations and timeouts. | ||
102 | func (c *Client) GetBucketEncryption(ctx context.Context, bucketName string) (*sse.Configuration, error) { | ||
103 | // Input validation. | ||
104 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
105 | return nil, err | ||
106 | } | ||
107 | |||
108 | // Get resources properly escaped and lined up before | ||
109 | // using them in http request. | ||
110 | urlValues := make(url.Values) | ||
111 | urlValues.Set("encryption", "") | ||
112 | |||
113 | // Execute GET on bucket to get the default encryption configuration. | ||
114 | resp, err := c.executeMethod(ctx, http.MethodGet, requestMetadata{ | ||
115 | bucketName: bucketName, | ||
116 | queryValues: urlValues, | ||
117 | }) | ||
118 | |||
119 | defer closeResponse(resp) | ||
120 | if err != nil { | ||
121 | return nil, err | ||
122 | } | ||
123 | |||
124 | if resp.StatusCode != http.StatusOK { | ||
125 | return nil, httpRespToErrorResponse(resp, bucketName, "") | ||
126 | } | ||
127 | |||
128 | encryptionConfig := &sse.Configuration{} | ||
129 | if err = xmlDecoder(resp.Body, encryptionConfig); err != nil { | ||
130 | return nil, err | ||
131 | } | ||
132 | |||
133 | return encryptionConfig, nil | ||
134 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/api-bucket-lifecycle.go b/vendor/github.com/minio/minio-go/v7/api-bucket-lifecycle.go deleted file mode 100644 index fec5cec..0000000 --- a/vendor/github.com/minio/minio-go/v7/api-bucket-lifecycle.go +++ /dev/null | |||
@@ -1,169 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2020 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package minio | ||
19 | |||
20 | import ( | ||
21 | "bytes" | ||
22 | "context" | ||
23 | "encoding/xml" | ||
24 | "io" | ||
25 | "net/http" | ||
26 | "net/url" | ||
27 | "time" | ||
28 | |||
29 | "github.com/minio/minio-go/v7/pkg/lifecycle" | ||
30 | "github.com/minio/minio-go/v7/pkg/s3utils" | ||
31 | ) | ||
32 | |||
33 | // SetBucketLifecycle set the lifecycle on an existing bucket. | ||
34 | func (c *Client) SetBucketLifecycle(ctx context.Context, bucketName string, config *lifecycle.Configuration) error { | ||
35 | // Input validation. | ||
36 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
37 | return err | ||
38 | } | ||
39 | |||
40 | // If lifecycle is empty then delete it. | ||
41 | if config.Empty() { | ||
42 | return c.removeBucketLifecycle(ctx, bucketName) | ||
43 | } | ||
44 | |||
45 | buf, err := xml.Marshal(config) | ||
46 | if err != nil { | ||
47 | return err | ||
48 | } | ||
49 | |||
50 | // Save the updated lifecycle. | ||
51 | return c.putBucketLifecycle(ctx, bucketName, buf) | ||
52 | } | ||
53 | |||
54 | // Saves a new bucket lifecycle. | ||
55 | func (c *Client) putBucketLifecycle(ctx context.Context, bucketName string, buf []byte) error { | ||
56 | // Get resources properly escaped and lined up before | ||
57 | // using them in http request. | ||
58 | urlValues := make(url.Values) | ||
59 | urlValues.Set("lifecycle", "") | ||
60 | |||
61 | // Content-length is mandatory for put lifecycle request | ||
62 | reqMetadata := requestMetadata{ | ||
63 | bucketName: bucketName, | ||
64 | queryValues: urlValues, | ||
65 | contentBody: bytes.NewReader(buf), | ||
66 | contentLength: int64(len(buf)), | ||
67 | contentMD5Base64: sumMD5Base64(buf), | ||
68 | } | ||
69 | |||
70 | // Execute PUT to upload a new bucket lifecycle. | ||
71 | resp, err := c.executeMethod(ctx, http.MethodPut, reqMetadata) | ||
72 | defer closeResponse(resp) | ||
73 | if err != nil { | ||
74 | return err | ||
75 | } | ||
76 | if resp != nil { | ||
77 | if resp.StatusCode != http.StatusOK { | ||
78 | return httpRespToErrorResponse(resp, bucketName, "") | ||
79 | } | ||
80 | } | ||
81 | return nil | ||
82 | } | ||
83 | |||
84 | // Remove lifecycle from a bucket. | ||
85 | func (c *Client) removeBucketLifecycle(ctx context.Context, bucketName string) error { | ||
86 | // Get resources properly escaped and lined up before | ||
87 | // using them in http request. | ||
88 | urlValues := make(url.Values) | ||
89 | urlValues.Set("lifecycle", "") | ||
90 | |||
91 | // Execute DELETE on objectName. | ||
92 | resp, err := c.executeMethod(ctx, http.MethodDelete, requestMetadata{ | ||
93 | bucketName: bucketName, | ||
94 | queryValues: urlValues, | ||
95 | contentSHA256Hex: emptySHA256Hex, | ||
96 | }) | ||
97 | defer closeResponse(resp) | ||
98 | if err != nil { | ||
99 | return err | ||
100 | } | ||
101 | return nil | ||
102 | } | ||
103 | |||
104 | // GetBucketLifecycle fetch bucket lifecycle configuration | ||
105 | func (c *Client) GetBucketLifecycle(ctx context.Context, bucketName string) (*lifecycle.Configuration, error) { | ||
106 | lc, _, err := c.GetBucketLifecycleWithInfo(ctx, bucketName) | ||
107 | return lc, err | ||
108 | } | ||
109 | |||
110 | // GetBucketLifecycleWithInfo fetch bucket lifecycle configuration along with when it was last updated | ||
111 | func (c *Client) GetBucketLifecycleWithInfo(ctx context.Context, bucketName string) (*lifecycle.Configuration, time.Time, error) { | ||
112 | // Input validation. | ||
113 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
114 | return nil, time.Time{}, err | ||
115 | } | ||
116 | |||
117 | bucketLifecycle, updatedAt, err := c.getBucketLifecycle(ctx, bucketName) | ||
118 | if err != nil { | ||
119 | return nil, time.Time{}, err | ||
120 | } | ||
121 | |||
122 | config := lifecycle.NewConfiguration() | ||
123 | if err = xml.Unmarshal(bucketLifecycle, config); err != nil { | ||
124 | return nil, time.Time{}, err | ||
125 | } | ||
126 | return config, updatedAt, nil | ||
127 | } | ||
128 | |||
129 | // Request server for current bucket lifecycle. | ||
130 | func (c *Client) getBucketLifecycle(ctx context.Context, bucketName string) ([]byte, time.Time, error) { | ||
131 | // Get resources properly escaped and lined up before | ||
132 | // using them in http request. | ||
133 | urlValues := make(url.Values) | ||
134 | urlValues.Set("lifecycle", "") | ||
135 | urlValues.Set("withUpdatedAt", "true") | ||
136 | |||
137 | // Execute GET on bucket to get lifecycle. | ||
138 | resp, err := c.executeMethod(ctx, http.MethodGet, requestMetadata{ | ||
139 | bucketName: bucketName, | ||
140 | queryValues: urlValues, | ||
141 | }) | ||
142 | |||
143 | defer closeResponse(resp) | ||
144 | if err != nil { | ||
145 | return nil, time.Time{}, err | ||
146 | } | ||
147 | |||
148 | if resp != nil { | ||
149 | if resp.StatusCode != http.StatusOK { | ||
150 | return nil, time.Time{}, httpRespToErrorResponse(resp, bucketName, "") | ||
151 | } | ||
152 | } | ||
153 | |||
154 | lcBytes, err := io.ReadAll(resp.Body) | ||
155 | if err != nil { | ||
156 | return nil, time.Time{}, err | ||
157 | } | ||
158 | |||
159 | const minIOLifecycleCfgUpdatedAt = "X-Minio-LifecycleConfig-UpdatedAt" | ||
160 | var updatedAt time.Time | ||
161 | if timeStr := resp.Header.Get(minIOLifecycleCfgUpdatedAt); timeStr != "" { | ||
162 | updatedAt, err = time.Parse(iso8601DateFormat, timeStr) | ||
163 | if err != nil { | ||
164 | return nil, time.Time{}, err | ||
165 | } | ||
166 | } | ||
167 | |||
168 | return lcBytes, updatedAt, nil | ||
169 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/api-bucket-notification.go b/vendor/github.com/minio/minio-go/v7/api-bucket-notification.go deleted file mode 100644 index 8de5c01..0000000 --- a/vendor/github.com/minio/minio-go/v7/api-bucket-notification.go +++ /dev/null | |||
@@ -1,261 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2017-2020 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package minio | ||
19 | |||
20 | import ( | ||
21 | "bufio" | ||
22 | "bytes" | ||
23 | "context" | ||
24 | "encoding/xml" | ||
25 | "net/http" | ||
26 | "net/url" | ||
27 | "time" | ||
28 | |||
29 | jsoniter "github.com/json-iterator/go" | ||
30 | "github.com/minio/minio-go/v7/pkg/notification" | ||
31 | "github.com/minio/minio-go/v7/pkg/s3utils" | ||
32 | ) | ||
33 | |||
34 | // SetBucketNotification saves a new bucket notification with a context to control cancellations and timeouts. | ||
35 | func (c *Client) SetBucketNotification(ctx context.Context, bucketName string, config notification.Configuration) error { | ||
36 | // Input validation. | ||
37 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
38 | return err | ||
39 | } | ||
40 | |||
41 | // Get resources properly escaped and lined up before | ||
42 | // using them in http request. | ||
43 | urlValues := make(url.Values) | ||
44 | urlValues.Set("notification", "") | ||
45 | |||
46 | notifBytes, err := xml.Marshal(&config) | ||
47 | if err != nil { | ||
48 | return err | ||
49 | } | ||
50 | |||
51 | notifBuffer := bytes.NewReader(notifBytes) | ||
52 | reqMetadata := requestMetadata{ | ||
53 | bucketName: bucketName, | ||
54 | queryValues: urlValues, | ||
55 | contentBody: notifBuffer, | ||
56 | contentLength: int64(len(notifBytes)), | ||
57 | contentMD5Base64: sumMD5Base64(notifBytes), | ||
58 | contentSHA256Hex: sum256Hex(notifBytes), | ||
59 | } | ||
60 | |||
61 | // Execute PUT to upload a new bucket notification. | ||
62 | resp, err := c.executeMethod(ctx, http.MethodPut, reqMetadata) | ||
63 | defer closeResponse(resp) | ||
64 | if err != nil { | ||
65 | return err | ||
66 | } | ||
67 | if resp != nil { | ||
68 | if resp.StatusCode != http.StatusOK { | ||
69 | return httpRespToErrorResponse(resp, bucketName, "") | ||
70 | } | ||
71 | } | ||
72 | return nil | ||
73 | } | ||
74 | |||
75 | // RemoveAllBucketNotification - Remove bucket notification clears all previously specified config | ||
76 | func (c *Client) RemoveAllBucketNotification(ctx context.Context, bucketName string) error { | ||
77 | return c.SetBucketNotification(ctx, bucketName, notification.Configuration{}) | ||
78 | } | ||
79 | |||
80 | // GetBucketNotification returns current bucket notification configuration | ||
81 | func (c *Client) GetBucketNotification(ctx context.Context, bucketName string) (bucketNotification notification.Configuration, err error) { | ||
82 | // Input validation. | ||
83 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
84 | return notification.Configuration{}, err | ||
85 | } | ||
86 | return c.getBucketNotification(ctx, bucketName) | ||
87 | } | ||
88 | |||
89 | // Request server for notification rules. | ||
90 | func (c *Client) getBucketNotification(ctx context.Context, bucketName string) (notification.Configuration, error) { | ||
91 | urlValues := make(url.Values) | ||
92 | urlValues.Set("notification", "") | ||
93 | |||
94 | // Execute GET on bucket to list objects. | ||
95 | resp, err := c.executeMethod(ctx, http.MethodGet, requestMetadata{ | ||
96 | bucketName: bucketName, | ||
97 | queryValues: urlValues, | ||
98 | contentSHA256Hex: emptySHA256Hex, | ||
99 | }) | ||
100 | |||
101 | defer closeResponse(resp) | ||
102 | if err != nil { | ||
103 | return notification.Configuration{}, err | ||
104 | } | ||
105 | return processBucketNotificationResponse(bucketName, resp) | ||
106 | } | ||
107 | |||
108 | // processes the GetNotification http response from the server. | ||
109 | func processBucketNotificationResponse(bucketName string, resp *http.Response) (notification.Configuration, error) { | ||
110 | if resp.StatusCode != http.StatusOK { | ||
111 | errResponse := httpRespToErrorResponse(resp, bucketName, "") | ||
112 | return notification.Configuration{}, errResponse | ||
113 | } | ||
114 | var bucketNotification notification.Configuration | ||
115 | err := xmlDecoder(resp.Body, &bucketNotification) | ||
116 | if err != nil { | ||
117 | return notification.Configuration{}, err | ||
118 | } | ||
119 | return bucketNotification, nil | ||
120 | } | ||
121 | |||
122 | // ListenNotification listen for all events, this is a MinIO specific API | ||
123 | func (c *Client) ListenNotification(ctx context.Context, prefix, suffix string, events []string) <-chan notification.Info { | ||
124 | return c.ListenBucketNotification(ctx, "", prefix, suffix, events) | ||
125 | } | ||
126 | |||
127 | // ListenBucketNotification listen for bucket events, this is a MinIO specific API | ||
128 | func (c *Client) ListenBucketNotification(ctx context.Context, bucketName, prefix, suffix string, events []string) <-chan notification.Info { | ||
129 | notificationInfoCh := make(chan notification.Info, 1) | ||
130 | const notificationCapacity = 4 * 1024 * 1024 | ||
131 | notificationEventBuffer := make([]byte, notificationCapacity) | ||
132 | // Only success, start a routine to start reading line by line. | ||
133 | go func(notificationInfoCh chan<- notification.Info) { | ||
134 | defer close(notificationInfoCh) | ||
135 | |||
136 | // Validate the bucket name. | ||
137 | if bucketName != "" { | ||
138 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
139 | select { | ||
140 | case notificationInfoCh <- notification.Info{ | ||
141 | Err: err, | ||
142 | }: | ||
143 | case <-ctx.Done(): | ||
144 | } | ||
145 | return | ||
146 | } | ||
147 | } | ||
148 | |||
149 | // Check ARN partition to verify if listening bucket is supported | ||
150 | if s3utils.IsAmazonEndpoint(*c.endpointURL) || s3utils.IsGoogleEndpoint(*c.endpointURL) { | ||
151 | select { | ||
152 | case notificationInfoCh <- notification.Info{ | ||
153 | Err: errAPINotSupported("Listening for bucket notification is specific only to `minio` server endpoints"), | ||
154 | }: | ||
155 | case <-ctx.Done(): | ||
156 | } | ||
157 | return | ||
158 | } | ||
159 | |||
160 | // Continuously run and listen on bucket notification. | ||
161 | // Create a done channel to control 'ListObjects' go routine. | ||
162 | retryDoneCh := make(chan struct{}, 1) | ||
163 | |||
164 | // Indicate to our routine to exit cleanly upon return. | ||
165 | defer close(retryDoneCh) | ||
166 | |||
167 | // Prepare urlValues to pass into the request on every loop | ||
168 | urlValues := make(url.Values) | ||
169 | urlValues.Set("ping", "10") | ||
170 | urlValues.Set("prefix", prefix) | ||
171 | urlValues.Set("suffix", suffix) | ||
172 | urlValues["events"] = events | ||
173 | |||
174 | // Wait on the jitter retry loop. | ||
175 | for range c.newRetryTimerContinous(time.Second, time.Second*30, MaxJitter, retryDoneCh) { | ||
176 | // Execute GET on bucket to list objects. | ||
177 | resp, err := c.executeMethod(ctx, http.MethodGet, requestMetadata{ | ||
178 | bucketName: bucketName, | ||
179 | queryValues: urlValues, | ||
180 | contentSHA256Hex: emptySHA256Hex, | ||
181 | }) | ||
182 | if err != nil { | ||
183 | select { | ||
184 | case notificationInfoCh <- notification.Info{ | ||
185 | Err: err, | ||
186 | }: | ||
187 | case <-ctx.Done(): | ||
188 | } | ||
189 | return | ||
190 | } | ||
191 | |||
192 | // Validate http response, upon error return quickly. | ||
193 | if resp.StatusCode != http.StatusOK { | ||
194 | errResponse := httpRespToErrorResponse(resp, bucketName, "") | ||
195 | select { | ||
196 | case notificationInfoCh <- notification.Info{ | ||
197 | Err: errResponse, | ||
198 | }: | ||
199 | case <-ctx.Done(): | ||
200 | } | ||
201 | return | ||
202 | } | ||
203 | |||
204 | // Initialize a new bufio scanner, to read line by line. | ||
205 | bio := bufio.NewScanner(resp.Body) | ||
206 | |||
207 | // Use a higher buffer to support unexpected | ||
208 | // caching done by proxies | ||
209 | bio.Buffer(notificationEventBuffer, notificationCapacity) | ||
210 | json := jsoniter.ConfigCompatibleWithStandardLibrary | ||
211 | |||
212 | // Unmarshal each line, returns marshaled values. | ||
213 | for bio.Scan() { | ||
214 | var notificationInfo notification.Info | ||
215 | if err = json.Unmarshal(bio.Bytes(), ¬ificationInfo); err != nil { | ||
216 | // Unexpected error during json unmarshal, send | ||
217 | // the error to caller for actionable as needed. | ||
218 | select { | ||
219 | case notificationInfoCh <- notification.Info{ | ||
220 | Err: err, | ||
221 | }: | ||
222 | case <-ctx.Done(): | ||
223 | return | ||
224 | } | ||
225 | closeResponse(resp) | ||
226 | continue | ||
227 | } | ||
228 | |||
229 | // Empty events pinged from the server | ||
230 | if len(notificationInfo.Records) == 0 && notificationInfo.Err == nil { | ||
231 | continue | ||
232 | } | ||
233 | |||
234 | // Send notificationInfo | ||
235 | select { | ||
236 | case notificationInfoCh <- notificationInfo: | ||
237 | case <-ctx.Done(): | ||
238 | closeResponse(resp) | ||
239 | return | ||
240 | } | ||
241 | } | ||
242 | |||
243 | if err = bio.Err(); err != nil { | ||
244 | select { | ||
245 | case notificationInfoCh <- notification.Info{ | ||
246 | Err: err, | ||
247 | }: | ||
248 | case <-ctx.Done(): | ||
249 | return | ||
250 | } | ||
251 | } | ||
252 | |||
253 | // Close current connection before looping further. | ||
254 | closeResponse(resp) | ||
255 | |||
256 | } | ||
257 | }(notificationInfoCh) | ||
258 | |||
259 | // Returns the notification info channel, for caller to start reading from. | ||
260 | return notificationInfoCh | ||
261 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/api-bucket-policy.go b/vendor/github.com/minio/minio-go/v7/api-bucket-policy.go deleted file mode 100644 index dbb5259..0000000 --- a/vendor/github.com/minio/minio-go/v7/api-bucket-policy.go +++ /dev/null | |||
@@ -1,147 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2020 MinIO, Inc. | ||
4 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
5 | * you may not use this file except in compliance with the License. | ||
6 | * You may obtain a copy of the License at | ||
7 | * | ||
8 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
9 | * | ||
10 | * Unless required by applicable law or agreed to in writing, software | ||
11 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
12 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
13 | * See the License for the specific language governing permissions and | ||
14 | * limitations under the License. | ||
15 | */ | ||
16 | |||
17 | package minio | ||
18 | |||
19 | import ( | ||
20 | "context" | ||
21 | "io" | ||
22 | "net/http" | ||
23 | "net/url" | ||
24 | "strings" | ||
25 | |||
26 | "github.com/minio/minio-go/v7/pkg/s3utils" | ||
27 | ) | ||
28 | |||
29 | // SetBucketPolicy sets the access permissions on an existing bucket. | ||
30 | func (c *Client) SetBucketPolicy(ctx context.Context, bucketName, policy string) error { | ||
31 | // Input validation. | ||
32 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
33 | return err | ||
34 | } | ||
35 | |||
36 | // If policy is empty then delete the bucket policy. | ||
37 | if policy == "" { | ||
38 | return c.removeBucketPolicy(ctx, bucketName) | ||
39 | } | ||
40 | |||
41 | // Save the updated policies. | ||
42 | return c.putBucketPolicy(ctx, bucketName, policy) | ||
43 | } | ||
44 | |||
45 | // Saves a new bucket policy. | ||
46 | func (c *Client) putBucketPolicy(ctx context.Context, bucketName, policy string) error { | ||
47 | // Get resources properly escaped and lined up before | ||
48 | // using them in http request. | ||
49 | urlValues := make(url.Values) | ||
50 | urlValues.Set("policy", "") | ||
51 | |||
52 | reqMetadata := requestMetadata{ | ||
53 | bucketName: bucketName, | ||
54 | queryValues: urlValues, | ||
55 | contentBody: strings.NewReader(policy), | ||
56 | contentLength: int64(len(policy)), | ||
57 | } | ||
58 | |||
59 | // Execute PUT to upload a new bucket policy. | ||
60 | resp, err := c.executeMethod(ctx, http.MethodPut, reqMetadata) | ||
61 | defer closeResponse(resp) | ||
62 | if err != nil { | ||
63 | return err | ||
64 | } | ||
65 | if resp != nil { | ||
66 | if resp.StatusCode != http.StatusNoContent && resp.StatusCode != http.StatusOK { | ||
67 | return httpRespToErrorResponse(resp, bucketName, "") | ||
68 | } | ||
69 | } | ||
70 | return nil | ||
71 | } | ||
72 | |||
73 | // Removes all policies on a bucket. | ||
74 | func (c *Client) removeBucketPolicy(ctx context.Context, bucketName string) error { | ||
75 | // Get resources properly escaped and lined up before | ||
76 | // using them in http request. | ||
77 | urlValues := make(url.Values) | ||
78 | urlValues.Set("policy", "") | ||
79 | |||
80 | // Execute DELETE on objectName. | ||
81 | resp, err := c.executeMethod(ctx, http.MethodDelete, requestMetadata{ | ||
82 | bucketName: bucketName, | ||
83 | queryValues: urlValues, | ||
84 | contentSHA256Hex: emptySHA256Hex, | ||
85 | }) | ||
86 | defer closeResponse(resp) | ||
87 | if err != nil { | ||
88 | return err | ||
89 | } | ||
90 | |||
91 | if resp.StatusCode != http.StatusNoContent { | ||
92 | return httpRespToErrorResponse(resp, bucketName, "") | ||
93 | } | ||
94 | |||
95 | return nil | ||
96 | } | ||
97 | |||
98 | // GetBucketPolicy returns the current policy | ||
99 | func (c *Client) GetBucketPolicy(ctx context.Context, bucketName string) (string, error) { | ||
100 | // Input validation. | ||
101 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
102 | return "", err | ||
103 | } | ||
104 | bucketPolicy, err := c.getBucketPolicy(ctx, bucketName) | ||
105 | if err != nil { | ||
106 | errResponse := ToErrorResponse(err) | ||
107 | if errResponse.Code == "NoSuchBucketPolicy" { | ||
108 | return "", nil | ||
109 | } | ||
110 | return "", err | ||
111 | } | ||
112 | return bucketPolicy, nil | ||
113 | } | ||
114 | |||
115 | // Request server for current bucket policy. | ||
116 | func (c *Client) getBucketPolicy(ctx context.Context, bucketName string) (string, error) { | ||
117 | // Get resources properly escaped and lined up before | ||
118 | // using them in http request. | ||
119 | urlValues := make(url.Values) | ||
120 | urlValues.Set("policy", "") | ||
121 | |||
122 | // Execute GET on bucket to list objects. | ||
123 | resp, err := c.executeMethod(ctx, http.MethodGet, requestMetadata{ | ||
124 | bucketName: bucketName, | ||
125 | queryValues: urlValues, | ||
126 | contentSHA256Hex: emptySHA256Hex, | ||
127 | }) | ||
128 | |||
129 | defer closeResponse(resp) | ||
130 | if err != nil { | ||
131 | return "", err | ||
132 | } | ||
133 | |||
134 | if resp != nil { | ||
135 | if resp.StatusCode != http.StatusOK { | ||
136 | return "", httpRespToErrorResponse(resp, bucketName, "") | ||
137 | } | ||
138 | } | ||
139 | |||
140 | bucketPolicyBuf, err := io.ReadAll(resp.Body) | ||
141 | if err != nil { | ||
142 | return "", err | ||
143 | } | ||
144 | |||
145 | policy := string(bucketPolicyBuf) | ||
146 | return policy, err | ||
147 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/api-bucket-replication.go b/vendor/github.com/minio/minio-go/v7/api-bucket-replication.go deleted file mode 100644 index b12bb13..0000000 --- a/vendor/github.com/minio/minio-go/v7/api-bucket-replication.go +++ /dev/null | |||
@@ -1,355 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2020 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package minio | ||
19 | |||
20 | import ( | ||
21 | "bytes" | ||
22 | "context" | ||
23 | "encoding/json" | ||
24 | "encoding/xml" | ||
25 | "io" | ||
26 | "net/http" | ||
27 | "net/url" | ||
28 | "time" | ||
29 | |||
30 | "github.com/google/uuid" | ||
31 | "github.com/minio/minio-go/v7/pkg/replication" | ||
32 | "github.com/minio/minio-go/v7/pkg/s3utils" | ||
33 | ) | ||
34 | |||
35 | // RemoveBucketReplication removes a replication config on an existing bucket. | ||
36 | func (c *Client) RemoveBucketReplication(ctx context.Context, bucketName string) error { | ||
37 | return c.removeBucketReplication(ctx, bucketName) | ||
38 | } | ||
39 | |||
40 | // SetBucketReplication sets a replication config on an existing bucket. | ||
41 | func (c *Client) SetBucketReplication(ctx context.Context, bucketName string, cfg replication.Config) error { | ||
42 | // Input validation. | ||
43 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
44 | return err | ||
45 | } | ||
46 | |||
47 | // If replication is empty then delete it. | ||
48 | if cfg.Empty() { | ||
49 | return c.removeBucketReplication(ctx, bucketName) | ||
50 | } | ||
51 | // Save the updated replication. | ||
52 | return c.putBucketReplication(ctx, bucketName, cfg) | ||
53 | } | ||
54 | |||
55 | // Saves a new bucket replication. | ||
56 | func (c *Client) putBucketReplication(ctx context.Context, bucketName string, cfg replication.Config) error { | ||
57 | // Get resources properly escaped and lined up before | ||
58 | // using them in http request. | ||
59 | urlValues := make(url.Values) | ||
60 | urlValues.Set("replication", "") | ||
61 | replication, err := xml.Marshal(cfg) | ||
62 | if err != nil { | ||
63 | return err | ||
64 | } | ||
65 | |||
66 | reqMetadata := requestMetadata{ | ||
67 | bucketName: bucketName, | ||
68 | queryValues: urlValues, | ||
69 | contentBody: bytes.NewReader(replication), | ||
70 | contentLength: int64(len(replication)), | ||
71 | contentMD5Base64: sumMD5Base64(replication), | ||
72 | } | ||
73 | |||
74 | // Execute PUT to upload a new bucket replication config. | ||
75 | resp, err := c.executeMethod(ctx, http.MethodPut, reqMetadata) | ||
76 | defer closeResponse(resp) | ||
77 | if err != nil { | ||
78 | return err | ||
79 | } | ||
80 | |||
81 | if resp.StatusCode != http.StatusOK { | ||
82 | return httpRespToErrorResponse(resp, bucketName, "") | ||
83 | } | ||
84 | |||
85 | return nil | ||
86 | } | ||
87 | |||
88 | // Remove replication from a bucket. | ||
89 | func (c *Client) removeBucketReplication(ctx context.Context, bucketName string) error { | ||
90 | // Get resources properly escaped and lined up before | ||
91 | // using them in http request. | ||
92 | urlValues := make(url.Values) | ||
93 | urlValues.Set("replication", "") | ||
94 | |||
95 | // Execute DELETE on objectName. | ||
96 | resp, err := c.executeMethod(ctx, http.MethodDelete, requestMetadata{ | ||
97 | bucketName: bucketName, | ||
98 | queryValues: urlValues, | ||
99 | contentSHA256Hex: emptySHA256Hex, | ||
100 | }) | ||
101 | defer closeResponse(resp) | ||
102 | if err != nil { | ||
103 | return err | ||
104 | } | ||
105 | if resp.StatusCode != http.StatusOK { | ||
106 | return httpRespToErrorResponse(resp, bucketName, "") | ||
107 | } | ||
108 | return nil | ||
109 | } | ||
110 | |||
111 | // GetBucketReplication fetches bucket replication configuration.If config is not | ||
112 | // found, returns empty config with nil error. | ||
113 | func (c *Client) GetBucketReplication(ctx context.Context, bucketName string) (cfg replication.Config, err error) { | ||
114 | // Input validation. | ||
115 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
116 | return cfg, err | ||
117 | } | ||
118 | bucketReplicationCfg, err := c.getBucketReplication(ctx, bucketName) | ||
119 | if err != nil { | ||
120 | errResponse := ToErrorResponse(err) | ||
121 | if errResponse.Code == "ReplicationConfigurationNotFoundError" { | ||
122 | return cfg, nil | ||
123 | } | ||
124 | return cfg, err | ||
125 | } | ||
126 | return bucketReplicationCfg, nil | ||
127 | } | ||
128 | |||
129 | // Request server for current bucket replication config. | ||
130 | func (c *Client) getBucketReplication(ctx context.Context, bucketName string) (cfg replication.Config, err error) { | ||
131 | // Get resources properly escaped and lined up before | ||
132 | // using them in http request. | ||
133 | urlValues := make(url.Values) | ||
134 | urlValues.Set("replication", "") | ||
135 | |||
136 | // Execute GET on bucket to get replication config. | ||
137 | resp, err := c.executeMethod(ctx, http.MethodGet, requestMetadata{ | ||
138 | bucketName: bucketName, | ||
139 | queryValues: urlValues, | ||
140 | }) | ||
141 | |||
142 | defer closeResponse(resp) | ||
143 | if err != nil { | ||
144 | return cfg, err | ||
145 | } | ||
146 | |||
147 | if resp.StatusCode != http.StatusOK { | ||
148 | return cfg, httpRespToErrorResponse(resp, bucketName, "") | ||
149 | } | ||
150 | |||
151 | if err = xmlDecoder(resp.Body, &cfg); err != nil { | ||
152 | return cfg, err | ||
153 | } | ||
154 | |||
155 | return cfg, nil | ||
156 | } | ||
157 | |||
158 | // GetBucketReplicationMetrics fetches bucket replication status metrics | ||
159 | func (c *Client) GetBucketReplicationMetrics(ctx context.Context, bucketName string) (s replication.Metrics, err error) { | ||
160 | // Input validation. | ||
161 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
162 | return s, err | ||
163 | } | ||
164 | // Get resources properly escaped and lined up before | ||
165 | // using them in http request. | ||
166 | urlValues := make(url.Values) | ||
167 | urlValues.Set("replication-metrics", "") | ||
168 | |||
169 | // Execute GET on bucket to get replication config. | ||
170 | resp, err := c.executeMethod(ctx, http.MethodGet, requestMetadata{ | ||
171 | bucketName: bucketName, | ||
172 | queryValues: urlValues, | ||
173 | }) | ||
174 | |||
175 | defer closeResponse(resp) | ||
176 | if err != nil { | ||
177 | return s, err | ||
178 | } | ||
179 | |||
180 | if resp.StatusCode != http.StatusOK { | ||
181 | return s, httpRespToErrorResponse(resp, bucketName, "") | ||
182 | } | ||
183 | respBytes, err := io.ReadAll(resp.Body) | ||
184 | if err != nil { | ||
185 | return s, err | ||
186 | } | ||
187 | |||
188 | if err := json.Unmarshal(respBytes, &s); err != nil { | ||
189 | return s, err | ||
190 | } | ||
191 | return s, nil | ||
192 | } | ||
193 | |||
194 | // mustGetUUID - get a random UUID. | ||
195 | func mustGetUUID() string { | ||
196 | u, err := uuid.NewRandom() | ||
197 | if err != nil { | ||
198 | return "" | ||
199 | } | ||
200 | return u.String() | ||
201 | } | ||
202 | |||
203 | // ResetBucketReplication kicks off replication of previously replicated objects if ExistingObjectReplication | ||
204 | // is enabled in the replication config | ||
205 | func (c *Client) ResetBucketReplication(ctx context.Context, bucketName string, olderThan time.Duration) (rID string, err error) { | ||
206 | rID = mustGetUUID() | ||
207 | _, err = c.resetBucketReplicationOnTarget(ctx, bucketName, olderThan, "", rID) | ||
208 | if err != nil { | ||
209 | return rID, err | ||
210 | } | ||
211 | return rID, nil | ||
212 | } | ||
213 | |||
214 | // ResetBucketReplicationOnTarget kicks off replication of previously replicated objects if | ||
215 | // ExistingObjectReplication is enabled in the replication config | ||
216 | func (c *Client) ResetBucketReplicationOnTarget(ctx context.Context, bucketName string, olderThan time.Duration, tgtArn string) (replication.ResyncTargetsInfo, error) { | ||
217 | return c.resetBucketReplicationOnTarget(ctx, bucketName, olderThan, tgtArn, mustGetUUID()) | ||
218 | } | ||
219 | |||
220 | // ResetBucketReplication kicks off replication of previously replicated objects if ExistingObjectReplication | ||
221 | // is enabled in the replication config | ||
222 | func (c *Client) resetBucketReplicationOnTarget(ctx context.Context, bucketName string, olderThan time.Duration, tgtArn, resetID string) (rinfo replication.ResyncTargetsInfo, err error) { | ||
223 | // Input validation. | ||
224 | if err = s3utils.CheckValidBucketName(bucketName); err != nil { | ||
225 | return | ||
226 | } | ||
227 | // Get resources properly escaped and lined up before | ||
228 | // using them in http request. | ||
229 | urlValues := make(url.Values) | ||
230 | urlValues.Set("replication-reset", "") | ||
231 | if olderThan > 0 { | ||
232 | urlValues.Set("older-than", olderThan.String()) | ||
233 | } | ||
234 | if tgtArn != "" { | ||
235 | urlValues.Set("arn", tgtArn) | ||
236 | } | ||
237 | urlValues.Set("reset-id", resetID) | ||
238 | // Execute GET on bucket to get replication config. | ||
239 | resp, err := c.executeMethod(ctx, http.MethodPut, requestMetadata{ | ||
240 | bucketName: bucketName, | ||
241 | queryValues: urlValues, | ||
242 | }) | ||
243 | |||
244 | defer closeResponse(resp) | ||
245 | if err != nil { | ||
246 | return rinfo, err | ||
247 | } | ||
248 | |||
249 | if resp.StatusCode != http.StatusOK { | ||
250 | return rinfo, httpRespToErrorResponse(resp, bucketName, "") | ||
251 | } | ||
252 | |||
253 | if err = json.NewDecoder(resp.Body).Decode(&rinfo); err != nil { | ||
254 | return rinfo, err | ||
255 | } | ||
256 | return rinfo, nil | ||
257 | } | ||
258 | |||
259 | // GetBucketReplicationResyncStatus gets the status of replication resync | ||
260 | func (c *Client) GetBucketReplicationResyncStatus(ctx context.Context, bucketName, arn string) (rinfo replication.ResyncTargetsInfo, err error) { | ||
261 | // Input validation. | ||
262 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
263 | return rinfo, err | ||
264 | } | ||
265 | // Get resources properly escaped and lined up before | ||
266 | // using them in http request. | ||
267 | urlValues := make(url.Values) | ||
268 | urlValues.Set("replication-reset-status", "") | ||
269 | if arn != "" { | ||
270 | urlValues.Set("arn", arn) | ||
271 | } | ||
272 | // Execute GET on bucket to get replication config. | ||
273 | resp, err := c.executeMethod(ctx, http.MethodGet, requestMetadata{ | ||
274 | bucketName: bucketName, | ||
275 | queryValues: urlValues, | ||
276 | }) | ||
277 | |||
278 | defer closeResponse(resp) | ||
279 | if err != nil { | ||
280 | return rinfo, err | ||
281 | } | ||
282 | |||
283 | if resp.StatusCode != http.StatusOK { | ||
284 | return rinfo, httpRespToErrorResponse(resp, bucketName, "") | ||
285 | } | ||
286 | |||
287 | if err = json.NewDecoder(resp.Body).Decode(&rinfo); err != nil { | ||
288 | return rinfo, err | ||
289 | } | ||
290 | return rinfo, nil | ||
291 | } | ||
292 | |||
293 | // GetBucketReplicationMetricsV2 fetches bucket replication status metrics | ||
294 | func (c *Client) GetBucketReplicationMetricsV2(ctx context.Context, bucketName string) (s replication.MetricsV2, err error) { | ||
295 | // Input validation. | ||
296 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
297 | return s, err | ||
298 | } | ||
299 | // Get resources properly escaped and lined up before | ||
300 | // using them in http request. | ||
301 | urlValues := make(url.Values) | ||
302 | urlValues.Set("replication-metrics", "2") | ||
303 | |||
304 | // Execute GET on bucket to get replication metrics. | ||
305 | resp, err := c.executeMethod(ctx, http.MethodGet, requestMetadata{ | ||
306 | bucketName: bucketName, | ||
307 | queryValues: urlValues, | ||
308 | }) | ||
309 | |||
310 | defer closeResponse(resp) | ||
311 | if err != nil { | ||
312 | return s, err | ||
313 | } | ||
314 | |||
315 | if resp.StatusCode != http.StatusOK { | ||
316 | return s, httpRespToErrorResponse(resp, bucketName, "") | ||
317 | } | ||
318 | respBytes, err := io.ReadAll(resp.Body) | ||
319 | if err != nil { | ||
320 | return s, err | ||
321 | } | ||
322 | |||
323 | if err := json.Unmarshal(respBytes, &s); err != nil { | ||
324 | return s, err | ||
325 | } | ||
326 | return s, nil | ||
327 | } | ||
328 | |||
329 | // CheckBucketReplication validates if replication is set up properly for a bucket | ||
330 | func (c *Client) CheckBucketReplication(ctx context.Context, bucketName string) (err error) { | ||
331 | // Input validation. | ||
332 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
333 | return err | ||
334 | } | ||
335 | // Get resources properly escaped and lined up before | ||
336 | // using them in http request. | ||
337 | urlValues := make(url.Values) | ||
338 | urlValues.Set("replication-check", "") | ||
339 | |||
340 | // Execute GET on bucket to get replication config. | ||
341 | resp, err := c.executeMethod(ctx, http.MethodGet, requestMetadata{ | ||
342 | bucketName: bucketName, | ||
343 | queryValues: urlValues, | ||
344 | }) | ||
345 | |||
346 | defer closeResponse(resp) | ||
347 | if err != nil { | ||
348 | return err | ||
349 | } | ||
350 | |||
351 | if resp.StatusCode != http.StatusOK { | ||
352 | return httpRespToErrorResponse(resp, bucketName, "") | ||
353 | } | ||
354 | return nil | ||
355 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/api-bucket-tagging.go b/vendor/github.com/minio/minio-go/v7/api-bucket-tagging.go deleted file mode 100644 index 86d7429..0000000 --- a/vendor/github.com/minio/minio-go/v7/api-bucket-tagging.go +++ /dev/null | |||
@@ -1,134 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2020 MinIO, Inc. | ||
4 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
5 | * you may not use this file except in compliance with the License. | ||
6 | * You may obtain a copy of the License at | ||
7 | * | ||
8 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
9 | * | ||
10 | * Unless required by applicable law or agreed to in writing, software | ||
11 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
12 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
13 | * See the License for the specific language governing permissions and | ||
14 | * limitations under the License. | ||
15 | */ | ||
16 | |||
17 | package minio | ||
18 | |||
19 | import ( | ||
20 | "bytes" | ||
21 | "context" | ||
22 | "encoding/xml" | ||
23 | "errors" | ||
24 | "io" | ||
25 | "net/http" | ||
26 | "net/url" | ||
27 | |||
28 | "github.com/minio/minio-go/v7/pkg/s3utils" | ||
29 | "github.com/minio/minio-go/v7/pkg/tags" | ||
30 | ) | ||
31 | |||
32 | // GetBucketTagging fetch tagging configuration for a bucket with a | ||
33 | // context to control cancellations and timeouts. | ||
34 | func (c *Client) GetBucketTagging(ctx context.Context, bucketName string) (*tags.Tags, error) { | ||
35 | // Input validation. | ||
36 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
37 | return nil, err | ||
38 | } | ||
39 | |||
40 | // Get resources properly escaped and lined up before | ||
41 | // using them in http request. | ||
42 | urlValues := make(url.Values) | ||
43 | urlValues.Set("tagging", "") | ||
44 | |||
45 | // Execute GET on bucket to get tagging configuration. | ||
46 | resp, err := c.executeMethod(ctx, http.MethodGet, requestMetadata{ | ||
47 | bucketName: bucketName, | ||
48 | queryValues: urlValues, | ||
49 | }) | ||
50 | |||
51 | defer closeResponse(resp) | ||
52 | if err != nil { | ||
53 | return nil, err | ||
54 | } | ||
55 | |||
56 | if resp.StatusCode != http.StatusOK { | ||
57 | return nil, httpRespToErrorResponse(resp, bucketName, "") | ||
58 | } | ||
59 | |||
60 | defer io.Copy(io.Discard, resp.Body) | ||
61 | return tags.ParseBucketXML(resp.Body) | ||
62 | } | ||
63 | |||
64 | // SetBucketTagging sets tagging configuration for a bucket | ||
65 | // with a context to control cancellations and timeouts. | ||
66 | func (c *Client) SetBucketTagging(ctx context.Context, bucketName string, tags *tags.Tags) error { | ||
67 | // Input validation. | ||
68 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
69 | return err | ||
70 | } | ||
71 | |||
72 | if tags == nil { | ||
73 | return errors.New("nil tags passed") | ||
74 | } | ||
75 | |||
76 | buf, err := xml.Marshal(tags) | ||
77 | if err != nil { | ||
78 | return err | ||
79 | } | ||
80 | |||
81 | // Get resources properly escaped and lined up before | ||
82 | // using them in http request. | ||
83 | urlValues := make(url.Values) | ||
84 | urlValues.Set("tagging", "") | ||
85 | |||
86 | // Content-length is mandatory to set a default encryption configuration | ||
87 | reqMetadata := requestMetadata{ | ||
88 | bucketName: bucketName, | ||
89 | queryValues: urlValues, | ||
90 | contentBody: bytes.NewReader(buf), | ||
91 | contentLength: int64(len(buf)), | ||
92 | contentMD5Base64: sumMD5Base64(buf), | ||
93 | } | ||
94 | |||
95 | // Execute PUT on bucket to put tagging configuration. | ||
96 | resp, err := c.executeMethod(ctx, http.MethodPut, reqMetadata) | ||
97 | defer closeResponse(resp) | ||
98 | if err != nil { | ||
99 | return err | ||
100 | } | ||
101 | if resp.StatusCode != http.StatusOK && resp.StatusCode != http.StatusNoContent { | ||
102 | return httpRespToErrorResponse(resp, bucketName, "") | ||
103 | } | ||
104 | return nil | ||
105 | } | ||
106 | |||
107 | // RemoveBucketTagging removes tagging configuration for a | ||
108 | // bucket with a context to control cancellations and timeouts. | ||
109 | func (c *Client) RemoveBucketTagging(ctx context.Context, bucketName string) error { | ||
110 | // Input validation. | ||
111 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
112 | return err | ||
113 | } | ||
114 | |||
115 | // Get resources properly escaped and lined up before | ||
116 | // using them in http request. | ||
117 | urlValues := make(url.Values) | ||
118 | urlValues.Set("tagging", "") | ||
119 | |||
120 | // Execute DELETE on bucket to remove tagging configuration. | ||
121 | resp, err := c.executeMethod(ctx, http.MethodDelete, requestMetadata{ | ||
122 | bucketName: bucketName, | ||
123 | queryValues: urlValues, | ||
124 | contentSHA256Hex: emptySHA256Hex, | ||
125 | }) | ||
126 | defer closeResponse(resp) | ||
127 | if err != nil { | ||
128 | return err | ||
129 | } | ||
130 | if resp.StatusCode != http.StatusOK && resp.StatusCode != http.StatusNoContent { | ||
131 | return httpRespToErrorResponse(resp, bucketName, "") | ||
132 | } | ||
133 | return nil | ||
134 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/api-bucket-versioning.go b/vendor/github.com/minio/minio-go/v7/api-bucket-versioning.go deleted file mode 100644 index 8c84e4f..0000000 --- a/vendor/github.com/minio/minio-go/v7/api-bucket-versioning.go +++ /dev/null | |||
@@ -1,146 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2020 MinIO, Inc. | ||
4 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
5 | * you may not use this file except in compliance with the License. | ||
6 | * You may obtain a copy of the License at | ||
7 | * | ||
8 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
9 | * | ||
10 | * Unless required by applicable law or agreed to in writing, software | ||
11 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
12 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
13 | * See the License for the specific language governing permissions and | ||
14 | * limitations under the License. | ||
15 | */ | ||
16 | |||
17 | package minio | ||
18 | |||
19 | import ( | ||
20 | "bytes" | ||
21 | "context" | ||
22 | "encoding/xml" | ||
23 | "net/http" | ||
24 | "net/url" | ||
25 | |||
26 | "github.com/minio/minio-go/v7/pkg/s3utils" | ||
27 | ) | ||
28 | |||
29 | // SetBucketVersioning sets a bucket versioning configuration | ||
30 | func (c *Client) SetBucketVersioning(ctx context.Context, bucketName string, config BucketVersioningConfiguration) error { | ||
31 | // Input validation. | ||
32 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
33 | return err | ||
34 | } | ||
35 | |||
36 | buf, err := xml.Marshal(config) | ||
37 | if err != nil { | ||
38 | return err | ||
39 | } | ||
40 | |||
41 | // Get resources properly escaped and lined up before | ||
42 | // using them in http request. | ||
43 | urlValues := make(url.Values) | ||
44 | urlValues.Set("versioning", "") | ||
45 | |||
46 | reqMetadata := requestMetadata{ | ||
47 | bucketName: bucketName, | ||
48 | queryValues: urlValues, | ||
49 | contentBody: bytes.NewReader(buf), | ||
50 | contentLength: int64(len(buf)), | ||
51 | contentMD5Base64: sumMD5Base64(buf), | ||
52 | contentSHA256Hex: sum256Hex(buf), | ||
53 | } | ||
54 | |||
55 | // Execute PUT to set a bucket versioning. | ||
56 | resp, err := c.executeMethod(ctx, http.MethodPut, reqMetadata) | ||
57 | defer closeResponse(resp) | ||
58 | if err != nil { | ||
59 | return err | ||
60 | } | ||
61 | if resp != nil { | ||
62 | if resp.StatusCode != http.StatusOK { | ||
63 | return httpRespToErrorResponse(resp, bucketName, "") | ||
64 | } | ||
65 | } | ||
66 | return nil | ||
67 | } | ||
68 | |||
69 | // EnableVersioning - enable object versioning in given bucket. | ||
70 | func (c *Client) EnableVersioning(ctx context.Context, bucketName string) error { | ||
71 | return c.SetBucketVersioning(ctx, bucketName, BucketVersioningConfiguration{Status: "Enabled"}) | ||
72 | } | ||
73 | |||
74 | // SuspendVersioning - suspend object versioning in given bucket. | ||
75 | func (c *Client) SuspendVersioning(ctx context.Context, bucketName string) error { | ||
76 | return c.SetBucketVersioning(ctx, bucketName, BucketVersioningConfiguration{Status: "Suspended"}) | ||
77 | } | ||
78 | |||
79 | // ExcludedPrefix - holds individual prefixes excluded from being versioned. | ||
80 | type ExcludedPrefix struct { | ||
81 | Prefix string | ||
82 | } | ||
83 | |||
84 | // BucketVersioningConfiguration is the versioning configuration structure | ||
85 | type BucketVersioningConfiguration struct { | ||
86 | XMLName xml.Name `xml:"VersioningConfiguration"` | ||
87 | Status string `xml:"Status"` | ||
88 | MFADelete string `xml:"MfaDelete,omitempty"` | ||
89 | // MinIO extension - allows selective, prefix-level versioning exclusion. | ||
90 | // Requires versioning to be enabled | ||
91 | ExcludedPrefixes []ExcludedPrefix `xml:",omitempty"` | ||
92 | ExcludeFolders bool `xml:",omitempty"` | ||
93 | } | ||
94 | |||
95 | // Various supported states | ||
96 | const ( | ||
97 | Enabled = "Enabled" | ||
98 | // Disabled State = "Disabled" only used by MFA Delete not supported yet. | ||
99 | Suspended = "Suspended" | ||
100 | ) | ||
101 | |||
102 | // Enabled returns true if bucket versioning is enabled | ||
103 | func (b BucketVersioningConfiguration) Enabled() bool { | ||
104 | return b.Status == Enabled | ||
105 | } | ||
106 | |||
107 | // Suspended returns true if bucket versioning is suspended | ||
108 | func (b BucketVersioningConfiguration) Suspended() bool { | ||
109 | return b.Status == Suspended | ||
110 | } | ||
111 | |||
112 | // GetBucketVersioning gets the versioning configuration on | ||
113 | // an existing bucket with a context to control cancellations and timeouts. | ||
114 | func (c *Client) GetBucketVersioning(ctx context.Context, bucketName string) (BucketVersioningConfiguration, error) { | ||
115 | // Input validation. | ||
116 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
117 | return BucketVersioningConfiguration{}, err | ||
118 | } | ||
119 | |||
120 | // Get resources properly escaped and lined up before | ||
121 | // using them in http request. | ||
122 | urlValues := make(url.Values) | ||
123 | urlValues.Set("versioning", "") | ||
124 | |||
125 | // Execute GET on bucket to get the versioning configuration. | ||
126 | resp, err := c.executeMethod(ctx, http.MethodGet, requestMetadata{ | ||
127 | bucketName: bucketName, | ||
128 | queryValues: urlValues, | ||
129 | }) | ||
130 | |||
131 | defer closeResponse(resp) | ||
132 | if err != nil { | ||
133 | return BucketVersioningConfiguration{}, err | ||
134 | } | ||
135 | |||
136 | if resp.StatusCode != http.StatusOK { | ||
137 | return BucketVersioningConfiguration{}, httpRespToErrorResponse(resp, bucketName, "") | ||
138 | } | ||
139 | |||
140 | versioningConfig := BucketVersioningConfiguration{} | ||
141 | if err = xmlDecoder(resp.Body, &versioningConfig); err != nil { | ||
142 | return versioningConfig, err | ||
143 | } | ||
144 | |||
145 | return versioningConfig, nil | ||
146 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/api-compose-object.go b/vendor/github.com/minio/minio-go/v7/api-compose-object.go deleted file mode 100644 index e64a244..0000000 --- a/vendor/github.com/minio/minio-go/v7/api-compose-object.go +++ /dev/null | |||
@@ -1,594 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2017, 2018 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package minio | ||
19 | |||
20 | import ( | ||
21 | "context" | ||
22 | "fmt" | ||
23 | "io" | ||
24 | "net/http" | ||
25 | "net/url" | ||
26 | "strconv" | ||
27 | "strings" | ||
28 | "time" | ||
29 | |||
30 | "github.com/google/uuid" | ||
31 | "github.com/minio/minio-go/v7/pkg/encrypt" | ||
32 | "github.com/minio/minio-go/v7/pkg/s3utils" | ||
33 | ) | ||
34 | |||
35 | // CopyDestOptions represents options specified by user for CopyObject/ComposeObject APIs | ||
36 | type CopyDestOptions struct { | ||
37 | Bucket string // points to destination bucket | ||
38 | Object string // points to destination object | ||
39 | |||
40 | // `Encryption` is the key info for server-side-encryption with customer | ||
41 | // provided key. If it is nil, no encryption is performed. | ||
42 | Encryption encrypt.ServerSide | ||
43 | |||
44 | // `userMeta` is the user-metadata key-value pairs to be set on the | ||
45 | // destination. The keys are automatically prefixed with `x-amz-meta-` | ||
46 | // if needed. If nil is passed, and if only a single source (of any | ||
47 | // size) is provided in the ComposeObject call, then metadata from the | ||
48 | // source is copied to the destination. | ||
49 | // if no user-metadata is provided, it is copied from source | ||
50 | // (when there is only once source object in the compose | ||
51 | // request) | ||
52 | UserMetadata map[string]string | ||
53 | // UserMetadata is only set to destination if ReplaceMetadata is true | ||
54 | // other value is UserMetadata is ignored and we preserve src.UserMetadata | ||
55 | // NOTE: if you set this value to true and now metadata is present | ||
56 | // in UserMetadata your destination object will not have any metadata | ||
57 | // set. | ||
58 | ReplaceMetadata bool | ||
59 | |||
60 | // `userTags` is the user defined object tags to be set on destination. | ||
61 | // This will be set only if the `replaceTags` field is set to true. | ||
62 | // Otherwise this field is ignored | ||
63 | UserTags map[string]string | ||
64 | ReplaceTags bool | ||
65 | |||
66 | // Specifies whether you want to apply a Legal Hold to the copied object. | ||
67 | LegalHold LegalHoldStatus | ||
68 | |||
69 | // Object Retention related fields | ||
70 | Mode RetentionMode | ||
71 | RetainUntilDate time.Time | ||
72 | |||
73 | Size int64 // Needs to be specified if progress bar is specified. | ||
74 | // Progress of the entire copy operation will be sent here. | ||
75 | Progress io.Reader | ||
76 | } | ||
77 | |||
78 | // Process custom-metadata to remove a `x-amz-meta-` prefix if | ||
79 | // present and validate that keys are distinct (after this | ||
80 | // prefix removal). | ||
81 | func filterCustomMeta(userMeta map[string]string) map[string]string { | ||
82 | m := make(map[string]string) | ||
83 | for k, v := range userMeta { | ||
84 | if strings.HasPrefix(strings.ToLower(k), "x-amz-meta-") { | ||
85 | k = k[len("x-amz-meta-"):] | ||
86 | } | ||
87 | if _, ok := m[k]; ok { | ||
88 | continue | ||
89 | } | ||
90 | m[k] = v | ||
91 | } | ||
92 | return m | ||
93 | } | ||
94 | |||
95 | // Marshal converts all the CopyDestOptions into their | ||
96 | // equivalent HTTP header representation | ||
97 | func (opts CopyDestOptions) Marshal(header http.Header) { | ||
98 | const replaceDirective = "REPLACE" | ||
99 | if opts.ReplaceTags { | ||
100 | header.Set(amzTaggingHeaderDirective, replaceDirective) | ||
101 | if tags := s3utils.TagEncode(opts.UserTags); tags != "" { | ||
102 | header.Set(amzTaggingHeader, tags) | ||
103 | } | ||
104 | } | ||
105 | |||
106 | if opts.LegalHold != LegalHoldStatus("") { | ||
107 | header.Set(amzLegalHoldHeader, opts.LegalHold.String()) | ||
108 | } | ||
109 | |||
110 | if opts.Mode != RetentionMode("") && !opts.RetainUntilDate.IsZero() { | ||
111 | header.Set(amzLockMode, opts.Mode.String()) | ||
112 | header.Set(amzLockRetainUntil, opts.RetainUntilDate.Format(time.RFC3339)) | ||
113 | } | ||
114 | |||
115 | if opts.Encryption != nil { | ||
116 | opts.Encryption.Marshal(header) | ||
117 | } | ||
118 | |||
119 | if opts.ReplaceMetadata { | ||
120 | header.Set("x-amz-metadata-directive", replaceDirective) | ||
121 | for k, v := range filterCustomMeta(opts.UserMetadata) { | ||
122 | if isAmzHeader(k) || isStandardHeader(k) || isStorageClassHeader(k) { | ||
123 | header.Set(k, v) | ||
124 | } else { | ||
125 | header.Set("x-amz-meta-"+k, v) | ||
126 | } | ||
127 | } | ||
128 | } | ||
129 | } | ||
130 | |||
131 | // toDestinationInfo returns a validated copyOptions object. | ||
132 | func (opts CopyDestOptions) validate() (err error) { | ||
133 | // Input validation. | ||
134 | if err = s3utils.CheckValidBucketName(opts.Bucket); err != nil { | ||
135 | return err | ||
136 | } | ||
137 | if err = s3utils.CheckValidObjectName(opts.Object); err != nil { | ||
138 | return err | ||
139 | } | ||
140 | if opts.Progress != nil && opts.Size < 0 { | ||
141 | return errInvalidArgument("For progress bar effective size needs to be specified") | ||
142 | } | ||
143 | return nil | ||
144 | } | ||
145 | |||
146 | // CopySrcOptions represents a source object to be copied, using | ||
147 | // server-side copying APIs. | ||
148 | type CopySrcOptions struct { | ||
149 | Bucket, Object string | ||
150 | VersionID string | ||
151 | MatchETag string | ||
152 | NoMatchETag string | ||
153 | MatchModifiedSince time.Time | ||
154 | MatchUnmodifiedSince time.Time | ||
155 | MatchRange bool | ||
156 | Start, End int64 | ||
157 | Encryption encrypt.ServerSide | ||
158 | } | ||
159 | |||
160 | // Marshal converts all the CopySrcOptions into their | ||
161 | // equivalent HTTP header representation | ||
162 | func (opts CopySrcOptions) Marshal(header http.Header) { | ||
163 | // Set the source header | ||
164 | header.Set("x-amz-copy-source", s3utils.EncodePath(opts.Bucket+"/"+opts.Object)) | ||
165 | if opts.VersionID != "" { | ||
166 | header.Set("x-amz-copy-source", s3utils.EncodePath(opts.Bucket+"/"+opts.Object)+"?versionId="+opts.VersionID) | ||
167 | } | ||
168 | |||
169 | if opts.MatchETag != "" { | ||
170 | header.Set("x-amz-copy-source-if-match", opts.MatchETag) | ||
171 | } | ||
172 | if opts.NoMatchETag != "" { | ||
173 | header.Set("x-amz-copy-source-if-none-match", opts.NoMatchETag) | ||
174 | } | ||
175 | |||
176 | if !opts.MatchModifiedSince.IsZero() { | ||
177 | header.Set("x-amz-copy-source-if-modified-since", opts.MatchModifiedSince.Format(http.TimeFormat)) | ||
178 | } | ||
179 | if !opts.MatchUnmodifiedSince.IsZero() { | ||
180 | header.Set("x-amz-copy-source-if-unmodified-since", opts.MatchUnmodifiedSince.Format(http.TimeFormat)) | ||
181 | } | ||
182 | |||
183 | if opts.Encryption != nil { | ||
184 | encrypt.SSECopy(opts.Encryption).Marshal(header) | ||
185 | } | ||
186 | } | ||
187 | |||
188 | func (opts CopySrcOptions) validate() (err error) { | ||
189 | // Input validation. | ||
190 | if err = s3utils.CheckValidBucketName(opts.Bucket); err != nil { | ||
191 | return err | ||
192 | } | ||
193 | if err = s3utils.CheckValidObjectName(opts.Object); err != nil { | ||
194 | return err | ||
195 | } | ||
196 | if opts.Start > opts.End || opts.Start < 0 { | ||
197 | return errInvalidArgument("start must be non-negative, and start must be at most end.") | ||
198 | } | ||
199 | return nil | ||
200 | } | ||
201 | |||
202 | // Low level implementation of CopyObject API, supports only upto 5GiB worth of copy. | ||
203 | func (c *Client) copyObjectDo(ctx context.Context, srcBucket, srcObject, destBucket, destObject string, | ||
204 | metadata map[string]string, srcOpts CopySrcOptions, dstOpts PutObjectOptions, | ||
205 | ) (ObjectInfo, error) { | ||
206 | // Build headers. | ||
207 | headers := make(http.Header) | ||
208 | |||
209 | // Set all the metadata headers. | ||
210 | for k, v := range metadata { | ||
211 | headers.Set(k, v) | ||
212 | } | ||
213 | if !dstOpts.Internal.ReplicationStatus.Empty() { | ||
214 | headers.Set(amzBucketReplicationStatus, string(dstOpts.Internal.ReplicationStatus)) | ||
215 | } | ||
216 | if !dstOpts.Internal.SourceMTime.IsZero() { | ||
217 | headers.Set(minIOBucketSourceMTime, dstOpts.Internal.SourceMTime.Format(time.RFC3339Nano)) | ||
218 | } | ||
219 | if dstOpts.Internal.SourceETag != "" { | ||
220 | headers.Set(minIOBucketSourceETag, dstOpts.Internal.SourceETag) | ||
221 | } | ||
222 | if dstOpts.Internal.ReplicationRequest { | ||
223 | headers.Set(minIOBucketReplicationRequest, "true") | ||
224 | } | ||
225 | if dstOpts.Internal.ReplicationValidityCheck { | ||
226 | headers.Set(minIOBucketReplicationCheck, "true") | ||
227 | } | ||
228 | if !dstOpts.Internal.LegalholdTimestamp.IsZero() { | ||
229 | headers.Set(minIOBucketReplicationObjectLegalHoldTimestamp, dstOpts.Internal.LegalholdTimestamp.Format(time.RFC3339Nano)) | ||
230 | } | ||
231 | if !dstOpts.Internal.RetentionTimestamp.IsZero() { | ||
232 | headers.Set(minIOBucketReplicationObjectRetentionTimestamp, dstOpts.Internal.RetentionTimestamp.Format(time.RFC3339Nano)) | ||
233 | } | ||
234 | if !dstOpts.Internal.TaggingTimestamp.IsZero() { | ||
235 | headers.Set(minIOBucketReplicationTaggingTimestamp, dstOpts.Internal.TaggingTimestamp.Format(time.RFC3339Nano)) | ||
236 | } | ||
237 | |||
238 | if len(dstOpts.UserTags) != 0 { | ||
239 | headers.Set(amzTaggingHeader, s3utils.TagEncode(dstOpts.UserTags)) | ||
240 | } | ||
241 | |||
242 | reqMetadata := requestMetadata{ | ||
243 | bucketName: destBucket, | ||
244 | objectName: destObject, | ||
245 | customHeader: headers, | ||
246 | } | ||
247 | if dstOpts.Internal.SourceVersionID != "" { | ||
248 | if dstOpts.Internal.SourceVersionID != nullVersionID { | ||
249 | if _, err := uuid.Parse(dstOpts.Internal.SourceVersionID); err != nil { | ||
250 | return ObjectInfo{}, errInvalidArgument(err.Error()) | ||
251 | } | ||
252 | } | ||
253 | urlValues := make(url.Values) | ||
254 | urlValues.Set("versionId", dstOpts.Internal.SourceVersionID) | ||
255 | reqMetadata.queryValues = urlValues | ||
256 | } | ||
257 | |||
258 | // Set the source header | ||
259 | headers.Set("x-amz-copy-source", s3utils.EncodePath(srcBucket+"/"+srcObject)) | ||
260 | if srcOpts.VersionID != "" { | ||
261 | headers.Set("x-amz-copy-source", s3utils.EncodePath(srcBucket+"/"+srcObject)+"?versionId="+srcOpts.VersionID) | ||
262 | } | ||
263 | // Send upload-part-copy request | ||
264 | resp, err := c.executeMethod(ctx, http.MethodPut, reqMetadata) | ||
265 | defer closeResponse(resp) | ||
266 | if err != nil { | ||
267 | return ObjectInfo{}, err | ||
268 | } | ||
269 | |||
270 | // Check if we got an error response. | ||
271 | if resp.StatusCode != http.StatusOK { | ||
272 | return ObjectInfo{}, httpRespToErrorResponse(resp, srcBucket, srcObject) | ||
273 | } | ||
274 | |||
275 | cpObjRes := copyObjectResult{} | ||
276 | err = xmlDecoder(resp.Body, &cpObjRes) | ||
277 | if err != nil { | ||
278 | return ObjectInfo{}, err | ||
279 | } | ||
280 | |||
281 | objInfo := ObjectInfo{ | ||
282 | Key: destObject, | ||
283 | ETag: strings.Trim(cpObjRes.ETag, "\""), | ||
284 | LastModified: cpObjRes.LastModified, | ||
285 | } | ||
286 | return objInfo, nil | ||
287 | } | ||
288 | |||
289 | func (c *Client) copyObjectPartDo(ctx context.Context, srcBucket, srcObject, destBucket, destObject, uploadID string, | ||
290 | partID int, startOffset, length int64, metadata map[string]string, | ||
291 | ) (p CompletePart, err error) { | ||
292 | headers := make(http.Header) | ||
293 | |||
294 | // Set source | ||
295 | headers.Set("x-amz-copy-source", s3utils.EncodePath(srcBucket+"/"+srcObject)) | ||
296 | |||
297 | if startOffset < 0 { | ||
298 | return p, errInvalidArgument("startOffset must be non-negative") | ||
299 | } | ||
300 | |||
301 | if length >= 0 { | ||
302 | headers.Set("x-amz-copy-source-range", fmt.Sprintf("bytes=%d-%d", startOffset, startOffset+length-1)) | ||
303 | } | ||
304 | |||
305 | for k, v := range metadata { | ||
306 | headers.Set(k, v) | ||
307 | } | ||
308 | |||
309 | queryValues := make(url.Values) | ||
310 | queryValues.Set("partNumber", strconv.Itoa(partID)) | ||
311 | queryValues.Set("uploadId", uploadID) | ||
312 | |||
313 | resp, err := c.executeMethod(ctx, http.MethodPut, requestMetadata{ | ||
314 | bucketName: destBucket, | ||
315 | objectName: destObject, | ||
316 | customHeader: headers, | ||
317 | queryValues: queryValues, | ||
318 | }) | ||
319 | defer closeResponse(resp) | ||
320 | if err != nil { | ||
321 | return | ||
322 | } | ||
323 | |||
324 | // Check if we got an error response. | ||
325 | if resp.StatusCode != http.StatusOK { | ||
326 | return p, httpRespToErrorResponse(resp, destBucket, destObject) | ||
327 | } | ||
328 | |||
329 | // Decode copy-part response on success. | ||
330 | cpObjRes := copyObjectResult{} | ||
331 | err = xmlDecoder(resp.Body, &cpObjRes) | ||
332 | if err != nil { | ||
333 | return p, err | ||
334 | } | ||
335 | p.PartNumber, p.ETag = partID, cpObjRes.ETag | ||
336 | return p, nil | ||
337 | } | ||
338 | |||
339 | // uploadPartCopy - helper function to create a part in a multipart | ||
340 | // upload via an upload-part-copy request | ||
341 | // https://docs.aws.amazon.com/AmazonS3/latest/API/mpUploadUploadPartCopy.html | ||
342 | func (c *Client) uploadPartCopy(ctx context.Context, bucket, object, uploadID string, partNumber int, | ||
343 | headers http.Header, | ||
344 | ) (p CompletePart, err error) { | ||
345 | // Build query parameters | ||
346 | urlValues := make(url.Values) | ||
347 | urlValues.Set("partNumber", strconv.Itoa(partNumber)) | ||
348 | urlValues.Set("uploadId", uploadID) | ||
349 | |||
350 | // Send upload-part-copy request | ||
351 | resp, err := c.executeMethod(ctx, http.MethodPut, requestMetadata{ | ||
352 | bucketName: bucket, | ||
353 | objectName: object, | ||
354 | customHeader: headers, | ||
355 | queryValues: urlValues, | ||
356 | }) | ||
357 | defer closeResponse(resp) | ||
358 | if err != nil { | ||
359 | return p, err | ||
360 | } | ||
361 | |||
362 | // Check if we got an error response. | ||
363 | if resp.StatusCode != http.StatusOK { | ||
364 | return p, httpRespToErrorResponse(resp, bucket, object) | ||
365 | } | ||
366 | |||
367 | // Decode copy-part response on success. | ||
368 | cpObjRes := copyObjectResult{} | ||
369 | err = xmlDecoder(resp.Body, &cpObjRes) | ||
370 | if err != nil { | ||
371 | return p, err | ||
372 | } | ||
373 | p.PartNumber, p.ETag = partNumber, cpObjRes.ETag | ||
374 | return p, nil | ||
375 | } | ||
376 | |||
377 | // ComposeObject - creates an object using server-side copying | ||
378 | // of existing objects. It takes a list of source objects (with optional offsets) | ||
379 | // and concatenates them into a new object using only server-side copying | ||
380 | // operations. Optionally takes progress reader hook for applications to | ||
381 | // look at current progress. | ||
382 | func (c *Client) ComposeObject(ctx context.Context, dst CopyDestOptions, srcs ...CopySrcOptions) (UploadInfo, error) { | ||
383 | if len(srcs) < 1 || len(srcs) > maxPartsCount { | ||
384 | return UploadInfo{}, errInvalidArgument("There must be as least one and up to 10000 source objects.") | ||
385 | } | ||
386 | |||
387 | for _, src := range srcs { | ||
388 | if err := src.validate(); err != nil { | ||
389 | return UploadInfo{}, err | ||
390 | } | ||
391 | } | ||
392 | |||
393 | if err := dst.validate(); err != nil { | ||
394 | return UploadInfo{}, err | ||
395 | } | ||
396 | |||
397 | srcObjectInfos := make([]ObjectInfo, len(srcs)) | ||
398 | srcObjectSizes := make([]int64, len(srcs)) | ||
399 | var totalSize, totalParts int64 | ||
400 | var err error | ||
401 | for i, src := range srcs { | ||
402 | opts := StatObjectOptions{ServerSideEncryption: encrypt.SSE(src.Encryption), VersionID: src.VersionID} | ||
403 | srcObjectInfos[i], err = c.StatObject(context.Background(), src.Bucket, src.Object, opts) | ||
404 | if err != nil { | ||
405 | return UploadInfo{}, err | ||
406 | } | ||
407 | |||
408 | srcCopySize := srcObjectInfos[i].Size | ||
409 | // Check if a segment is specified, and if so, is the | ||
410 | // segment within object bounds? | ||
411 | if src.MatchRange { | ||
412 | // Since range is specified, | ||
413 | // 0 <= src.start <= src.end | ||
414 | // so only invalid case to check is: | ||
415 | if src.End >= srcCopySize || src.Start < 0 { | ||
416 | return UploadInfo{}, errInvalidArgument( | ||
417 | fmt.Sprintf("CopySrcOptions %d has invalid segment-to-copy [%d, %d] (size is %d)", | ||
418 | i, src.Start, src.End, srcCopySize)) | ||
419 | } | ||
420 | srcCopySize = src.End - src.Start + 1 | ||
421 | } | ||
422 | |||
423 | // Only the last source may be less than `absMinPartSize` | ||
424 | if srcCopySize < absMinPartSize && i < len(srcs)-1 { | ||
425 | return UploadInfo{}, errInvalidArgument( | ||
426 | fmt.Sprintf("CopySrcOptions %d is too small (%d) and it is not the last part", i, srcCopySize)) | ||
427 | } | ||
428 | |||
429 | // Is data to copy too large? | ||
430 | totalSize += srcCopySize | ||
431 | if totalSize > maxMultipartPutObjectSize { | ||
432 | return UploadInfo{}, errInvalidArgument(fmt.Sprintf("Cannot compose an object of size %d (> 5TiB)", totalSize)) | ||
433 | } | ||
434 | |||
435 | // record source size | ||
436 | srcObjectSizes[i] = srcCopySize | ||
437 | |||
438 | // calculate parts needed for current source | ||
439 | totalParts += partsRequired(srcCopySize) | ||
440 | // Do we need more parts than we are allowed? | ||
441 | if totalParts > maxPartsCount { | ||
442 | return UploadInfo{}, errInvalidArgument(fmt.Sprintf( | ||
443 | "Your proposed compose object requires more than %d parts", maxPartsCount)) | ||
444 | } | ||
445 | } | ||
446 | |||
447 | // Single source object case (i.e. when only one source is | ||
448 | // involved, it is being copied wholly and at most 5GiB in | ||
449 | // size, emptyfiles are also supported). | ||
450 | if (totalParts == 1 && srcs[0].Start == -1 && totalSize <= maxPartSize) || (totalSize == 0) { | ||
451 | return c.CopyObject(ctx, dst, srcs[0]) | ||
452 | } | ||
453 | |||
454 | // Now, handle multipart-copy cases. | ||
455 | |||
456 | // 1. Ensure that the object has not been changed while | ||
457 | // we are copying data. | ||
458 | for i, src := range srcs { | ||
459 | src.MatchETag = srcObjectInfos[i].ETag | ||
460 | } | ||
461 | |||
462 | // 2. Initiate a new multipart upload. | ||
463 | |||
464 | // Set user-metadata on the destination object. If no | ||
465 | // user-metadata is specified, and there is only one source, | ||
466 | // (only) then metadata from source is copied. | ||
467 | var userMeta map[string]string | ||
468 | if dst.ReplaceMetadata { | ||
469 | userMeta = dst.UserMetadata | ||
470 | } else { | ||
471 | userMeta = srcObjectInfos[0].UserMetadata | ||
472 | } | ||
473 | |||
474 | var userTags map[string]string | ||
475 | if dst.ReplaceTags { | ||
476 | userTags = dst.UserTags | ||
477 | } else { | ||
478 | userTags = srcObjectInfos[0].UserTags | ||
479 | } | ||
480 | |||
481 | uploadID, err := c.newUploadID(ctx, dst.Bucket, dst.Object, PutObjectOptions{ | ||
482 | ServerSideEncryption: dst.Encryption, | ||
483 | UserMetadata: userMeta, | ||
484 | UserTags: userTags, | ||
485 | Mode: dst.Mode, | ||
486 | RetainUntilDate: dst.RetainUntilDate, | ||
487 | LegalHold: dst.LegalHold, | ||
488 | }) | ||
489 | if err != nil { | ||
490 | return UploadInfo{}, err | ||
491 | } | ||
492 | |||
493 | // 3. Perform copy part uploads | ||
494 | objParts := []CompletePart{} | ||
495 | partIndex := 1 | ||
496 | for i, src := range srcs { | ||
497 | h := make(http.Header) | ||
498 | src.Marshal(h) | ||
499 | if dst.Encryption != nil && dst.Encryption.Type() == encrypt.SSEC { | ||
500 | dst.Encryption.Marshal(h) | ||
501 | } | ||
502 | |||
503 | // calculate start/end indices of parts after | ||
504 | // splitting. | ||
505 | startIdx, endIdx := calculateEvenSplits(srcObjectSizes[i], src) | ||
506 | for j, start := range startIdx { | ||
507 | end := endIdx[j] | ||
508 | |||
509 | // Add (or reset) source range header for | ||
510 | // upload part copy request. | ||
511 | h.Set("x-amz-copy-source-range", | ||
512 | fmt.Sprintf("bytes=%d-%d", start, end)) | ||
513 | |||
514 | // make upload-part-copy request | ||
515 | complPart, err := c.uploadPartCopy(ctx, dst.Bucket, | ||
516 | dst.Object, uploadID, partIndex, h) | ||
517 | if err != nil { | ||
518 | return UploadInfo{}, err | ||
519 | } | ||
520 | if dst.Progress != nil { | ||
521 | io.CopyN(io.Discard, dst.Progress, end-start+1) | ||
522 | } | ||
523 | objParts = append(objParts, complPart) | ||
524 | partIndex++ | ||
525 | } | ||
526 | } | ||
527 | |||
528 | // 4. Make final complete-multipart request. | ||
529 | uploadInfo, err := c.completeMultipartUpload(ctx, dst.Bucket, dst.Object, uploadID, | ||
530 | completeMultipartUpload{Parts: objParts}, PutObjectOptions{ServerSideEncryption: dst.Encryption}) | ||
531 | if err != nil { | ||
532 | return UploadInfo{}, err | ||
533 | } | ||
534 | |||
535 | uploadInfo.Size = totalSize | ||
536 | return uploadInfo, nil | ||
537 | } | ||
538 | |||
539 | // partsRequired is maximum parts possible with | ||
540 | // max part size of ceiling(maxMultipartPutObjectSize / (maxPartsCount - 1)) | ||
541 | func partsRequired(size int64) int64 { | ||
542 | maxPartSize := maxMultipartPutObjectSize / (maxPartsCount - 1) | ||
543 | r := size / int64(maxPartSize) | ||
544 | if size%int64(maxPartSize) > 0 { | ||
545 | r++ | ||
546 | } | ||
547 | return r | ||
548 | } | ||
549 | |||
550 | // calculateEvenSplits - computes splits for a source and returns | ||
551 | // start and end index slices. Splits happen evenly to be sure that no | ||
552 | // part is less than 5MiB, as that could fail the multipart request if | ||
553 | // it is not the last part. | ||
554 | func calculateEvenSplits(size int64, src CopySrcOptions) (startIndex, endIndex []int64) { | ||
555 | if size == 0 { | ||
556 | return | ||
557 | } | ||
558 | |||
559 | reqParts := partsRequired(size) | ||
560 | startIndex = make([]int64, reqParts) | ||
561 | endIndex = make([]int64, reqParts) | ||
562 | // Compute number of required parts `k`, as: | ||
563 | // | ||
564 | // k = ceiling(size / copyPartSize) | ||
565 | // | ||
566 | // Now, distribute the `size` bytes in the source into | ||
567 | // k parts as evenly as possible: | ||
568 | // | ||
569 | // r parts sized (q+1) bytes, and | ||
570 | // (k - r) parts sized q bytes, where | ||
571 | // | ||
572 | // size = q * k + r (by simple division of size by k, | ||
573 | // so that 0 <= r < k) | ||
574 | // | ||
575 | start := src.Start | ||
576 | if start == -1 { | ||
577 | start = 0 | ||
578 | } | ||
579 | quot, rem := size/reqParts, size%reqParts | ||
580 | nextStart := start | ||
581 | for j := int64(0); j < reqParts; j++ { | ||
582 | curPartSize := quot | ||
583 | if j < rem { | ||
584 | curPartSize++ | ||
585 | } | ||
586 | |||
587 | cStart := nextStart | ||
588 | cEnd := cStart + curPartSize - 1 | ||
589 | nextStart = cEnd + 1 | ||
590 | |||
591 | startIndex[j], endIndex[j] = cStart, cEnd | ||
592 | } | ||
593 | return | ||
594 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/api-copy-object.go b/vendor/github.com/minio/minio-go/v7/api-copy-object.go deleted file mode 100644 index 0c95d91..0000000 --- a/vendor/github.com/minio/minio-go/v7/api-copy-object.go +++ /dev/null | |||
@@ -1,76 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2017, 2018 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package minio | ||
19 | |||
20 | import ( | ||
21 | "context" | ||
22 | "io" | ||
23 | "net/http" | ||
24 | ) | ||
25 | |||
26 | // CopyObject - copy a source object into a new object | ||
27 | func (c *Client) CopyObject(ctx context.Context, dst CopyDestOptions, src CopySrcOptions) (UploadInfo, error) { | ||
28 | if err := src.validate(); err != nil { | ||
29 | return UploadInfo{}, err | ||
30 | } | ||
31 | |||
32 | if err := dst.validate(); err != nil { | ||
33 | return UploadInfo{}, err | ||
34 | } | ||
35 | |||
36 | header := make(http.Header) | ||
37 | dst.Marshal(header) | ||
38 | src.Marshal(header) | ||
39 | |||
40 | resp, err := c.executeMethod(ctx, http.MethodPut, requestMetadata{ | ||
41 | bucketName: dst.Bucket, | ||
42 | objectName: dst.Object, | ||
43 | customHeader: header, | ||
44 | }) | ||
45 | if err != nil { | ||
46 | return UploadInfo{}, err | ||
47 | } | ||
48 | defer closeResponse(resp) | ||
49 | |||
50 | if resp.StatusCode != http.StatusOK { | ||
51 | return UploadInfo{}, httpRespToErrorResponse(resp, dst.Bucket, dst.Object) | ||
52 | } | ||
53 | |||
54 | // Update the progress properly after successful copy. | ||
55 | if dst.Progress != nil { | ||
56 | io.Copy(io.Discard, io.LimitReader(dst.Progress, dst.Size)) | ||
57 | } | ||
58 | |||
59 | cpObjRes := copyObjectResult{} | ||
60 | if err = xmlDecoder(resp.Body, &cpObjRes); err != nil { | ||
61 | return UploadInfo{}, err | ||
62 | } | ||
63 | |||
64 | // extract lifecycle expiry date and rule ID | ||
65 | expTime, ruleID := amzExpirationToExpiryDateRuleID(resp.Header.Get(amzExpiration)) | ||
66 | |||
67 | return UploadInfo{ | ||
68 | Bucket: dst.Bucket, | ||
69 | Key: dst.Object, | ||
70 | LastModified: cpObjRes.LastModified, | ||
71 | ETag: trimEtag(resp.Header.Get("ETag")), | ||
72 | VersionID: resp.Header.Get(amzVersionID), | ||
73 | Expiration: expTime, | ||
74 | ExpirationRuleID: ruleID, | ||
75 | }, nil | ||
76 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/api-datatypes.go b/vendor/github.com/minio/minio-go/v7/api-datatypes.go deleted file mode 100644 index 97a6f80..0000000 --- a/vendor/github.com/minio/minio-go/v7/api-datatypes.go +++ /dev/null | |||
@@ -1,254 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2015-2020 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package minio | ||
19 | |||
20 | import ( | ||
21 | "encoding/xml" | ||
22 | "io" | ||
23 | "net/http" | ||
24 | "net/url" | ||
25 | "strings" | ||
26 | "time" | ||
27 | ) | ||
28 | |||
29 | // BucketInfo container for bucket metadata. | ||
30 | type BucketInfo struct { | ||
31 | // The name of the bucket. | ||
32 | Name string `json:"name"` | ||
33 | // Date the bucket was created. | ||
34 | CreationDate time.Time `json:"creationDate"` | ||
35 | } | ||
36 | |||
37 | // StringMap represents map with custom UnmarshalXML | ||
38 | type StringMap map[string]string | ||
39 | |||
40 | // UnmarshalXML unmarshals the XML into a map of string to strings, | ||
41 | // creating a key in the map for each tag and setting it's value to the | ||
42 | // tags contents. | ||
43 | // | ||
44 | // The fact this function is on the pointer of Map is important, so that | ||
45 | // if m is nil it can be initialized, which is often the case if m is | ||
46 | // nested in another xml structural. This is also why the first thing done | ||
47 | // on the first line is initialize it. | ||
48 | func (m *StringMap) UnmarshalXML(d *xml.Decoder, _ xml.StartElement) error { | ||
49 | *m = StringMap{} | ||
50 | for { | ||
51 | // Format is <key>value</key> | ||
52 | var e struct { | ||
53 | XMLName xml.Name | ||
54 | Value string `xml:",chardata"` | ||
55 | } | ||
56 | err := d.Decode(&e) | ||
57 | if err == io.EOF { | ||
58 | break | ||
59 | } | ||
60 | if err != nil { | ||
61 | return err | ||
62 | } | ||
63 | (*m)[e.XMLName.Local] = e.Value | ||
64 | } | ||
65 | return nil | ||
66 | } | ||
67 | |||
68 | // URLMap represents map with custom UnmarshalXML | ||
69 | type URLMap map[string]string | ||
70 | |||
71 | // UnmarshalXML unmarshals the XML into a map of string to strings, | ||
72 | // creating a key in the map for each tag and setting it's value to the | ||
73 | // tags contents. | ||
74 | // | ||
75 | // The fact this function is on the pointer of Map is important, so that | ||
76 | // if m is nil it can be initialized, which is often the case if m is | ||
77 | // nested in another xml structural. This is also why the first thing done | ||
78 | // on the first line is initialize it. | ||
79 | func (m *URLMap) UnmarshalXML(d *xml.Decoder, se xml.StartElement) error { | ||
80 | *m = URLMap{} | ||
81 | var tgs string | ||
82 | if err := d.DecodeElement(&tgs, &se); err != nil { | ||
83 | if err == io.EOF { | ||
84 | return nil | ||
85 | } | ||
86 | return err | ||
87 | } | ||
88 | for tgs != "" { | ||
89 | var key string | ||
90 | key, tgs, _ = stringsCut(tgs, "&") | ||
91 | if key == "" { | ||
92 | continue | ||
93 | } | ||
94 | key, value, _ := stringsCut(key, "=") | ||
95 | key, err := url.QueryUnescape(key) | ||
96 | if err != nil { | ||
97 | return err | ||
98 | } | ||
99 | |||
100 | value, err = url.QueryUnescape(value) | ||
101 | if err != nil { | ||
102 | return err | ||
103 | } | ||
104 | (*m)[key] = value | ||
105 | } | ||
106 | return nil | ||
107 | } | ||
108 | |||
109 | // stringsCut slices s around the first instance of sep, | ||
110 | // returning the text before and after sep. | ||
111 | // The found result reports whether sep appears in s. | ||
112 | // If sep does not appear in s, cut returns s, "", false. | ||
113 | func stringsCut(s, sep string) (before, after string, found bool) { | ||
114 | if i := strings.Index(s, sep); i >= 0 { | ||
115 | return s[:i], s[i+len(sep):], true | ||
116 | } | ||
117 | return s, "", false | ||
118 | } | ||
119 | |||
120 | // Owner name. | ||
121 | type Owner struct { | ||
122 | XMLName xml.Name `xml:"Owner" json:"owner"` | ||
123 | DisplayName string `xml:"ID" json:"name"` | ||
124 | ID string `xml:"DisplayName" json:"id"` | ||
125 | } | ||
126 | |||
127 | // UploadInfo contains information about the | ||
128 | // newly uploaded or copied object. | ||
129 | type UploadInfo struct { | ||
130 | Bucket string | ||
131 | Key string | ||
132 | ETag string | ||
133 | Size int64 | ||
134 | LastModified time.Time | ||
135 | Location string | ||
136 | VersionID string | ||
137 | |||
138 | // Lifecycle expiry-date and ruleID associated with the expiry | ||
139 | // not to be confused with `Expires` HTTP header. | ||
140 | Expiration time.Time | ||
141 | ExpirationRuleID string | ||
142 | |||
143 | // Verified checksum values, if any. | ||
144 | // Values are base64 (standard) encoded. | ||
145 | // For multipart objects this is a checksum of the checksum of each part. | ||
146 | ChecksumCRC32 string | ||
147 | ChecksumCRC32C string | ||
148 | ChecksumSHA1 string | ||
149 | ChecksumSHA256 string | ||
150 | } | ||
151 | |||
152 | // RestoreInfo contains information of the restore operation of an archived object | ||
153 | type RestoreInfo struct { | ||
154 | // Is the restoring operation is still ongoing | ||
155 | OngoingRestore bool | ||
156 | // When the restored copy of the archived object will be removed | ||
157 | ExpiryTime time.Time | ||
158 | } | ||
159 | |||
160 | // ObjectInfo container for object metadata. | ||
161 | type ObjectInfo struct { | ||
162 | // An ETag is optionally set to md5sum of an object. In case of multipart objects, | ||
163 | // ETag is of the form MD5SUM-N where MD5SUM is md5sum of all individual md5sums of | ||
164 | // each parts concatenated into one string. | ||
165 | ETag string `json:"etag"` | ||
166 | |||
167 | Key string `json:"name"` // Name of the object | ||
168 | LastModified time.Time `json:"lastModified"` // Date and time the object was last modified. | ||
169 | Size int64 `json:"size"` // Size in bytes of the object. | ||
170 | ContentType string `json:"contentType"` // A standard MIME type describing the format of the object data. | ||
171 | Expires time.Time `json:"expires"` // The date and time at which the object is no longer able to be cached. | ||
172 | |||
173 | // Collection of additional metadata on the object. | ||
174 | // eg: x-amz-meta-*, content-encoding etc. | ||
175 | Metadata http.Header `json:"metadata" xml:"-"` | ||
176 | |||
177 | // x-amz-meta-* headers stripped "x-amz-meta-" prefix containing the first value. | ||
178 | // Only returned by MinIO servers. | ||
179 | UserMetadata StringMap `json:"userMetadata,omitempty"` | ||
180 | |||
181 | // x-amz-tagging values in their k/v values. | ||
182 | // Only returned by MinIO servers. | ||
183 | UserTags URLMap `json:"userTags,omitempty" xml:"UserTags"` | ||
184 | |||
185 | // x-amz-tagging-count value | ||
186 | UserTagCount int | ||
187 | |||
188 | // Owner name. | ||
189 | Owner Owner | ||
190 | |||
191 | // ACL grant. | ||
192 | Grant []Grant | ||
193 | |||
194 | // The class of storage used to store the object. | ||
195 | StorageClass string `json:"storageClass"` | ||
196 | |||
197 | // Versioning related information | ||
198 | IsLatest bool | ||
199 | IsDeleteMarker bool | ||
200 | VersionID string `xml:"VersionId"` | ||
201 | |||
202 | // x-amz-replication-status value is either in one of the following states | ||
203 | // - COMPLETED | ||
204 | // - PENDING | ||
205 | // - FAILED | ||
206 | // - REPLICA (on the destination) | ||
207 | ReplicationStatus string `xml:"ReplicationStatus"` | ||
208 | // set to true if delete marker has backing object version on target, and eligible to replicate | ||
209 | ReplicationReady bool | ||
210 | // Lifecycle expiry-date and ruleID associated with the expiry | ||
211 | // not to be confused with `Expires` HTTP header. | ||
212 | Expiration time.Time | ||
213 | ExpirationRuleID string | ||
214 | |||
215 | Restore *RestoreInfo | ||
216 | |||
217 | // Checksum values | ||
218 | ChecksumCRC32 string | ||
219 | ChecksumCRC32C string | ||
220 | ChecksumSHA1 string | ||
221 | ChecksumSHA256 string | ||
222 | |||
223 | Internal *struct { | ||
224 | K int // Data blocks | ||
225 | M int // Parity blocks | ||
226 | } `xml:"Internal"` | ||
227 | |||
228 | // Error | ||
229 | Err error `json:"-"` | ||
230 | } | ||
231 | |||
232 | // ObjectMultipartInfo container for multipart object metadata. | ||
233 | type ObjectMultipartInfo struct { | ||
234 | // Date and time at which the multipart upload was initiated. | ||
235 | Initiated time.Time `type:"timestamp" timestampFormat:"iso8601"` | ||
236 | |||
237 | Initiator initiator | ||
238 | Owner owner | ||
239 | |||
240 | // The type of storage to use for the object. Defaults to 'STANDARD'. | ||
241 | StorageClass string | ||
242 | |||
243 | // Key of the object for which the multipart upload was initiated. | ||
244 | Key string | ||
245 | |||
246 | // Size in bytes of the object. | ||
247 | Size int64 | ||
248 | |||
249 | // Upload ID that identifies the multipart upload. | ||
250 | UploadID string `xml:"UploadId"` | ||
251 | |||
252 | // Error | ||
253 | Err error | ||
254 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/api-error-response.go b/vendor/github.com/minio/minio-go/v7/api-error-response.go deleted file mode 100644 index 7df211f..0000000 --- a/vendor/github.com/minio/minio-go/v7/api-error-response.go +++ /dev/null | |||
@@ -1,284 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2015-2020 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package minio | ||
19 | |||
20 | import ( | ||
21 | "bytes" | ||
22 | "encoding/xml" | ||
23 | "fmt" | ||
24 | "io" | ||
25 | "net/http" | ||
26 | "strings" | ||
27 | ) | ||
28 | |||
29 | /* **** SAMPLE ERROR RESPONSE **** | ||
30 | <?xml version="1.0" encoding="UTF-8"?> | ||
31 | <Error> | ||
32 | <Code>AccessDenied</Code> | ||
33 | <Message>Access Denied</Message> | ||
34 | <BucketName>bucketName</BucketName> | ||
35 | <Key>objectName</Key> | ||
36 | <RequestId>F19772218238A85A</RequestId> | ||
37 | <HostId>GuWkjyviSiGHizehqpmsD1ndz5NClSP19DOT+s2mv7gXGQ8/X1lhbDGiIJEXpGFD</HostId> | ||
38 | </Error> | ||
39 | */ | ||
40 | |||
41 | // ErrorResponse - Is the typed error returned by all API operations. | ||
42 | // ErrorResponse struct should be comparable since it is compared inside | ||
43 | // golang http API (https://github.com/golang/go/issues/29768) | ||
44 | type ErrorResponse struct { | ||
45 | XMLName xml.Name `xml:"Error" json:"-"` | ||
46 | Code string | ||
47 | Message string | ||
48 | BucketName string | ||
49 | Key string | ||
50 | Resource string | ||
51 | RequestID string `xml:"RequestId"` | ||
52 | HostID string `xml:"HostId"` | ||
53 | |||
54 | // Region where the bucket is located. This header is returned | ||
55 | // only in HEAD bucket and ListObjects response. | ||
56 | Region string | ||
57 | |||
58 | // Captures the server string returned in response header. | ||
59 | Server string | ||
60 | |||
61 | // Underlying HTTP status code for the returned error | ||
62 | StatusCode int `xml:"-" json:"-"` | ||
63 | } | ||
64 | |||
65 | // ToErrorResponse - Returns parsed ErrorResponse struct from body and | ||
66 | // http headers. | ||
67 | // | ||
68 | // For example: | ||
69 | // | ||
70 | // import s3 "github.com/minio/minio-go/v7" | ||
71 | // ... | ||
72 | // ... | ||
73 | // reader, stat, err := s3.GetObject(...) | ||
74 | // if err != nil { | ||
75 | // resp := s3.ToErrorResponse(err) | ||
76 | // } | ||
77 | // ... | ||
78 | func ToErrorResponse(err error) ErrorResponse { | ||
79 | switch err := err.(type) { | ||
80 | case ErrorResponse: | ||
81 | return err | ||
82 | default: | ||
83 | return ErrorResponse{} | ||
84 | } | ||
85 | } | ||
86 | |||
87 | // Error - Returns S3 error string. | ||
88 | func (e ErrorResponse) Error() string { | ||
89 | if e.Message == "" { | ||
90 | msg, ok := s3ErrorResponseMap[e.Code] | ||
91 | if !ok { | ||
92 | msg = fmt.Sprintf("Error response code %s.", e.Code) | ||
93 | } | ||
94 | return msg | ||
95 | } | ||
96 | return e.Message | ||
97 | } | ||
98 | |||
99 | // Common string for errors to report issue location in unexpected | ||
100 | // cases. | ||
101 | const ( | ||
102 | reportIssue = "Please report this issue at https://github.com/minio/minio-go/issues." | ||
103 | ) | ||
104 | |||
105 | // xmlDecodeAndBody reads the whole body up to 1MB and | ||
106 | // tries to XML decode it into v. | ||
107 | // The body that was read and any error from reading or decoding is returned. | ||
108 | func xmlDecodeAndBody(bodyReader io.Reader, v interface{}) ([]byte, error) { | ||
109 | // read the whole body (up to 1MB) | ||
110 | const maxBodyLength = 1 << 20 | ||
111 | body, err := io.ReadAll(io.LimitReader(bodyReader, maxBodyLength)) | ||
112 | if err != nil { | ||
113 | return nil, err | ||
114 | } | ||
115 | return bytes.TrimSpace(body), xmlDecoder(bytes.NewReader(body), v) | ||
116 | } | ||
117 | |||
118 | // httpRespToErrorResponse returns a new encoded ErrorResponse | ||
119 | // structure as error. | ||
120 | func httpRespToErrorResponse(resp *http.Response, bucketName, objectName string) error { | ||
121 | if resp == nil { | ||
122 | msg := "Empty http response. " + reportIssue | ||
123 | return errInvalidArgument(msg) | ||
124 | } | ||
125 | |||
126 | errResp := ErrorResponse{ | ||
127 | StatusCode: resp.StatusCode, | ||
128 | Server: resp.Header.Get("Server"), | ||
129 | } | ||
130 | |||
131 | errBody, err := xmlDecodeAndBody(resp.Body, &errResp) | ||
132 | // Xml decoding failed with no body, fall back to HTTP headers. | ||
133 | if err != nil { | ||
134 | switch resp.StatusCode { | ||
135 | case http.StatusNotFound: | ||
136 | if objectName == "" { | ||
137 | errResp = ErrorResponse{ | ||
138 | StatusCode: resp.StatusCode, | ||
139 | Code: "NoSuchBucket", | ||
140 | Message: "The specified bucket does not exist.", | ||
141 | BucketName: bucketName, | ||
142 | } | ||
143 | } else { | ||
144 | errResp = ErrorResponse{ | ||
145 | StatusCode: resp.StatusCode, | ||
146 | Code: "NoSuchKey", | ||
147 | Message: "The specified key does not exist.", | ||
148 | BucketName: bucketName, | ||
149 | Key: objectName, | ||
150 | } | ||
151 | } | ||
152 | case http.StatusForbidden: | ||
153 | errResp = ErrorResponse{ | ||
154 | StatusCode: resp.StatusCode, | ||
155 | Code: "AccessDenied", | ||
156 | Message: "Access Denied.", | ||
157 | BucketName: bucketName, | ||
158 | Key: objectName, | ||
159 | } | ||
160 | case http.StatusConflict: | ||
161 | errResp = ErrorResponse{ | ||
162 | StatusCode: resp.StatusCode, | ||
163 | Code: "Conflict", | ||
164 | Message: "Bucket not empty.", | ||
165 | BucketName: bucketName, | ||
166 | } | ||
167 | case http.StatusPreconditionFailed: | ||
168 | errResp = ErrorResponse{ | ||
169 | StatusCode: resp.StatusCode, | ||
170 | Code: "PreconditionFailed", | ||
171 | Message: s3ErrorResponseMap["PreconditionFailed"], | ||
172 | BucketName: bucketName, | ||
173 | Key: objectName, | ||
174 | } | ||
175 | default: | ||
176 | msg := resp.Status | ||
177 | if len(errBody) > 0 { | ||
178 | msg = string(errBody) | ||
179 | if len(msg) > 1024 { | ||
180 | msg = msg[:1024] + "..." | ||
181 | } | ||
182 | } | ||
183 | errResp = ErrorResponse{ | ||
184 | StatusCode: resp.StatusCode, | ||
185 | Code: resp.Status, | ||
186 | Message: msg, | ||
187 | BucketName: bucketName, | ||
188 | } | ||
189 | } | ||
190 | } | ||
191 | |||
192 | code := resp.Header.Get("x-minio-error-code") | ||
193 | if code != "" { | ||
194 | errResp.Code = code | ||
195 | } | ||
196 | desc := resp.Header.Get("x-minio-error-desc") | ||
197 | if desc != "" { | ||
198 | errResp.Message = strings.Trim(desc, `"`) | ||
199 | } | ||
200 | |||
201 | // Save hostID, requestID and region information | ||
202 | // from headers if not available through error XML. | ||
203 | if errResp.RequestID == "" { | ||
204 | errResp.RequestID = resp.Header.Get("x-amz-request-id") | ||
205 | } | ||
206 | if errResp.HostID == "" { | ||
207 | errResp.HostID = resp.Header.Get("x-amz-id-2") | ||
208 | } | ||
209 | if errResp.Region == "" { | ||
210 | errResp.Region = resp.Header.Get("x-amz-bucket-region") | ||
211 | } | ||
212 | if errResp.Code == "InvalidRegion" && errResp.Region != "" { | ||
213 | errResp.Message = fmt.Sprintf("Region does not match, expecting region ‘%s’.", errResp.Region) | ||
214 | } | ||
215 | |||
216 | return errResp | ||
217 | } | ||
218 | |||
219 | // errTransferAccelerationBucket - bucket name is invalid to be used with transfer acceleration. | ||
220 | func errTransferAccelerationBucket(bucketName string) error { | ||
221 | return ErrorResponse{ | ||
222 | StatusCode: http.StatusBadRequest, | ||
223 | Code: "InvalidArgument", | ||
224 | Message: "The name of the bucket used for Transfer Acceleration must be DNS-compliant and must not contain periods ‘.’.", | ||
225 | BucketName: bucketName, | ||
226 | } | ||
227 | } | ||
228 | |||
229 | // errEntityTooLarge - Input size is larger than supported maximum. | ||
230 | func errEntityTooLarge(totalSize, maxObjectSize int64, bucketName, objectName string) error { | ||
231 | msg := fmt.Sprintf("Your proposed upload size ‘%d’ exceeds the maximum allowed object size ‘%d’ for single PUT operation.", totalSize, maxObjectSize) | ||
232 | return ErrorResponse{ | ||
233 | StatusCode: http.StatusBadRequest, | ||
234 | Code: "EntityTooLarge", | ||
235 | Message: msg, | ||
236 | BucketName: bucketName, | ||
237 | Key: objectName, | ||
238 | } | ||
239 | } | ||
240 | |||
241 | // errEntityTooSmall - Input size is smaller than supported minimum. | ||
242 | func errEntityTooSmall(totalSize int64, bucketName, objectName string) error { | ||
243 | msg := fmt.Sprintf("Your proposed upload size ‘%d’ is below the minimum allowed object size ‘0B’ for single PUT operation.", totalSize) | ||
244 | return ErrorResponse{ | ||
245 | StatusCode: http.StatusBadRequest, | ||
246 | Code: "EntityTooSmall", | ||
247 | Message: msg, | ||
248 | BucketName: bucketName, | ||
249 | Key: objectName, | ||
250 | } | ||
251 | } | ||
252 | |||
253 | // errUnexpectedEOF - Unexpected end of file reached. | ||
254 | func errUnexpectedEOF(totalRead, totalSize int64, bucketName, objectName string) error { | ||
255 | msg := fmt.Sprintf("Data read ‘%d’ is not equal to the size ‘%d’ of the input Reader.", totalRead, totalSize) | ||
256 | return ErrorResponse{ | ||
257 | StatusCode: http.StatusBadRequest, | ||
258 | Code: "UnexpectedEOF", | ||
259 | Message: msg, | ||
260 | BucketName: bucketName, | ||
261 | Key: objectName, | ||
262 | } | ||
263 | } | ||
264 | |||
265 | // errInvalidArgument - Invalid argument response. | ||
266 | func errInvalidArgument(message string) error { | ||
267 | return ErrorResponse{ | ||
268 | StatusCode: http.StatusBadRequest, | ||
269 | Code: "InvalidArgument", | ||
270 | Message: message, | ||
271 | RequestID: "minio", | ||
272 | } | ||
273 | } | ||
274 | |||
275 | // errAPINotSupported - API not supported response | ||
276 | // The specified API call is not supported | ||
277 | func errAPINotSupported(message string) error { | ||
278 | return ErrorResponse{ | ||
279 | StatusCode: http.StatusNotImplemented, | ||
280 | Code: "APINotSupported", | ||
281 | Message: message, | ||
282 | RequestID: "minio", | ||
283 | } | ||
284 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/api-get-object-acl.go b/vendor/github.com/minio/minio-go/v7/api-get-object-acl.go deleted file mode 100644 index 9041d99..0000000 --- a/vendor/github.com/minio/minio-go/v7/api-get-object-acl.go +++ /dev/null | |||
@@ -1,152 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2018 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package minio | ||
19 | |||
20 | import ( | ||
21 | "context" | ||
22 | "encoding/xml" | ||
23 | "net/http" | ||
24 | "net/url" | ||
25 | ) | ||
26 | |||
27 | // Grantee represents the person being granted permissions. | ||
28 | type Grantee struct { | ||
29 | XMLName xml.Name `xml:"Grantee"` | ||
30 | ID string `xml:"ID"` | ||
31 | DisplayName string `xml:"DisplayName"` | ||
32 | URI string `xml:"URI"` | ||
33 | } | ||
34 | |||
35 | // Grant holds grant information | ||
36 | type Grant struct { | ||
37 | XMLName xml.Name `xml:"Grant"` | ||
38 | Grantee Grantee | ||
39 | Permission string `xml:"Permission"` | ||
40 | } | ||
41 | |||
42 | // AccessControlList contains the set of grantees and the permissions assigned to each grantee. | ||
43 | type AccessControlList struct { | ||
44 | XMLName xml.Name `xml:"AccessControlList"` | ||
45 | Grant []Grant | ||
46 | Permission string `xml:"Permission"` | ||
47 | } | ||
48 | |||
49 | type accessControlPolicy struct { | ||
50 | XMLName xml.Name `xml:"AccessControlPolicy"` | ||
51 | Owner Owner | ||
52 | AccessControlList AccessControlList | ||
53 | } | ||
54 | |||
55 | // GetObjectACL get object ACLs | ||
56 | func (c *Client) GetObjectACL(ctx context.Context, bucketName, objectName string) (*ObjectInfo, error) { | ||
57 | resp, err := c.executeMethod(ctx, http.MethodGet, requestMetadata{ | ||
58 | bucketName: bucketName, | ||
59 | objectName: objectName, | ||
60 | queryValues: url.Values{ | ||
61 | "acl": []string{""}, | ||
62 | }, | ||
63 | }) | ||
64 | if err != nil { | ||
65 | return nil, err | ||
66 | } | ||
67 | defer closeResponse(resp) | ||
68 | |||
69 | if resp.StatusCode != http.StatusOK { | ||
70 | return nil, httpRespToErrorResponse(resp, bucketName, objectName) | ||
71 | } | ||
72 | |||
73 | res := &accessControlPolicy{} | ||
74 | |||
75 | if err := xmlDecoder(resp.Body, res); err != nil { | ||
76 | return nil, err | ||
77 | } | ||
78 | |||
79 | objInfo, err := c.StatObject(ctx, bucketName, objectName, StatObjectOptions{}) | ||
80 | if err != nil { | ||
81 | return nil, err | ||
82 | } | ||
83 | |||
84 | objInfo.Owner.DisplayName = res.Owner.DisplayName | ||
85 | objInfo.Owner.ID = res.Owner.ID | ||
86 | |||
87 | objInfo.Grant = append(objInfo.Grant, res.AccessControlList.Grant...) | ||
88 | |||
89 | cannedACL := getCannedACL(res) | ||
90 | if cannedACL != "" { | ||
91 | objInfo.Metadata.Add("X-Amz-Acl", cannedACL) | ||
92 | return &objInfo, nil | ||
93 | } | ||
94 | |||
95 | grantACL := getAmzGrantACL(res) | ||
96 | for k, v := range grantACL { | ||
97 | objInfo.Metadata[k] = v | ||
98 | } | ||
99 | |||
100 | return &objInfo, nil | ||
101 | } | ||
102 | |||
103 | func getCannedACL(aCPolicy *accessControlPolicy) string { | ||
104 | grants := aCPolicy.AccessControlList.Grant | ||
105 | |||
106 | switch { | ||
107 | case len(grants) == 1: | ||
108 | if grants[0].Grantee.URI == "" && grants[0].Permission == "FULL_CONTROL" { | ||
109 | return "private" | ||
110 | } | ||
111 | case len(grants) == 2: | ||
112 | for _, g := range grants { | ||
113 | if g.Grantee.URI == "http://acs.amazonaws.com/groups/global/AuthenticatedUsers" && g.Permission == "READ" { | ||
114 | return "authenticated-read" | ||
115 | } | ||
116 | if g.Grantee.URI == "http://acs.amazonaws.com/groups/global/AllUsers" && g.Permission == "READ" { | ||
117 | return "public-read" | ||
118 | } | ||
119 | if g.Permission == "READ" && g.Grantee.ID == aCPolicy.Owner.ID { | ||
120 | return "bucket-owner-read" | ||
121 | } | ||
122 | } | ||
123 | case len(grants) == 3: | ||
124 | for _, g := range grants { | ||
125 | if g.Grantee.URI == "http://acs.amazonaws.com/groups/global/AllUsers" && g.Permission == "WRITE" { | ||
126 | return "public-read-write" | ||
127 | } | ||
128 | } | ||
129 | } | ||
130 | return "" | ||
131 | } | ||
132 | |||
133 | func getAmzGrantACL(aCPolicy *accessControlPolicy) map[string][]string { | ||
134 | grants := aCPolicy.AccessControlList.Grant | ||
135 | res := map[string][]string{} | ||
136 | |||
137 | for _, g := range grants { | ||
138 | switch { | ||
139 | case g.Permission == "READ": | ||
140 | res["X-Amz-Grant-Read"] = append(res["X-Amz-Grant-Read"], "id="+g.Grantee.ID) | ||
141 | case g.Permission == "WRITE": | ||
142 | res["X-Amz-Grant-Write"] = append(res["X-Amz-Grant-Write"], "id="+g.Grantee.ID) | ||
143 | case g.Permission == "READ_ACP": | ||
144 | res["X-Amz-Grant-Read-Acp"] = append(res["X-Amz-Grant-Read-Acp"], "id="+g.Grantee.ID) | ||
145 | case g.Permission == "WRITE_ACP": | ||
146 | res["X-Amz-Grant-Write-Acp"] = append(res["X-Amz-Grant-Write-Acp"], "id="+g.Grantee.ID) | ||
147 | case g.Permission == "FULL_CONTROL": | ||
148 | res["X-Amz-Grant-Full-Control"] = append(res["X-Amz-Grant-Full-Control"], "id="+g.Grantee.ID) | ||
149 | } | ||
150 | } | ||
151 | return res | ||
152 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/api-get-object-file.go b/vendor/github.com/minio/minio-go/v7/api-get-object-file.go deleted file mode 100644 index 2332dbf..0000000 --- a/vendor/github.com/minio/minio-go/v7/api-get-object-file.go +++ /dev/null | |||
@@ -1,127 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2015-2017 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package minio | ||
19 | |||
20 | import ( | ||
21 | "context" | ||
22 | "io" | ||
23 | "os" | ||
24 | "path/filepath" | ||
25 | |||
26 | "github.com/minio/minio-go/v7/pkg/s3utils" | ||
27 | ) | ||
28 | |||
29 | // FGetObject - download contents of an object to a local file. | ||
30 | // The options can be used to specify the GET request further. | ||
31 | func (c *Client) FGetObject(ctx context.Context, bucketName, objectName, filePath string, opts GetObjectOptions) error { | ||
32 | // Input validation. | ||
33 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
34 | return err | ||
35 | } | ||
36 | if err := s3utils.CheckValidObjectName(objectName); err != nil { | ||
37 | return err | ||
38 | } | ||
39 | |||
40 | // Verify if destination already exists. | ||
41 | st, err := os.Stat(filePath) | ||
42 | if err == nil { | ||
43 | // If the destination exists and is a directory. | ||
44 | if st.IsDir() { | ||
45 | return errInvalidArgument("fileName is a directory.") | ||
46 | } | ||
47 | } | ||
48 | |||
49 | // Proceed if file does not exist. return for all other errors. | ||
50 | if err != nil { | ||
51 | if !os.IsNotExist(err) { | ||
52 | return err | ||
53 | } | ||
54 | } | ||
55 | |||
56 | // Extract top level directory. | ||
57 | objectDir, _ := filepath.Split(filePath) | ||
58 | if objectDir != "" { | ||
59 | // Create any missing top level directories. | ||
60 | if err := os.MkdirAll(objectDir, 0o700); err != nil { | ||
61 | return err | ||
62 | } | ||
63 | } | ||
64 | |||
65 | // Gather md5sum. | ||
66 | objectStat, err := c.StatObject(ctx, bucketName, objectName, StatObjectOptions(opts)) | ||
67 | if err != nil { | ||
68 | return err | ||
69 | } | ||
70 | |||
71 | // Write to a temporary file "fileName.part.minio" before saving. | ||
72 | filePartPath := filePath + objectStat.ETag + ".part.minio" | ||
73 | |||
74 | // If exists, open in append mode. If not create it as a part file. | ||
75 | filePart, err := os.OpenFile(filePartPath, os.O_CREATE|os.O_APPEND|os.O_WRONLY, 0o600) | ||
76 | if err != nil { | ||
77 | return err | ||
78 | } | ||
79 | |||
80 | // If we return early with an error, be sure to close and delete | ||
81 | // filePart. If we have an error along the way there is a chance | ||
82 | // that filePart is somehow damaged, and we should discard it. | ||
83 | closeAndRemove := true | ||
84 | defer func() { | ||
85 | if closeAndRemove { | ||
86 | _ = filePart.Close() | ||
87 | _ = os.Remove(filePartPath) | ||
88 | } | ||
89 | }() | ||
90 | |||
91 | // Issue Stat to get the current offset. | ||
92 | st, err = filePart.Stat() | ||
93 | if err != nil { | ||
94 | return err | ||
95 | } | ||
96 | |||
97 | // Initialize get object request headers to set the | ||
98 | // appropriate range offsets to read from. | ||
99 | if st.Size() > 0 { | ||
100 | opts.SetRange(st.Size(), 0) | ||
101 | } | ||
102 | |||
103 | // Seek to current position for incoming reader. | ||
104 | objectReader, objectStat, _, err := c.getObject(ctx, bucketName, objectName, opts) | ||
105 | if err != nil { | ||
106 | return err | ||
107 | } | ||
108 | |||
109 | // Write to the part file. | ||
110 | if _, err = io.CopyN(filePart, objectReader, objectStat.Size); err != nil { | ||
111 | return err | ||
112 | } | ||
113 | |||
114 | // Close the file before rename, this is specifically needed for Windows users. | ||
115 | closeAndRemove = false | ||
116 | if err = filePart.Close(); err != nil { | ||
117 | return err | ||
118 | } | ||
119 | |||
120 | // Safely completed. Now commit by renaming to actual filename. | ||
121 | if err = os.Rename(filePartPath, filePath); err != nil { | ||
122 | return err | ||
123 | } | ||
124 | |||
125 | // Return. | ||
126 | return nil | ||
127 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/api-get-object.go b/vendor/github.com/minio/minio-go/v7/api-get-object.go deleted file mode 100644 index 9e6b154..0000000 --- a/vendor/github.com/minio/minio-go/v7/api-get-object.go +++ /dev/null | |||
@@ -1,683 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2015-2020 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package minio | ||
19 | |||
20 | import ( | ||
21 | "context" | ||
22 | "errors" | ||
23 | "fmt" | ||
24 | "io" | ||
25 | "net/http" | ||
26 | "sync" | ||
27 | |||
28 | "github.com/minio/minio-go/v7/pkg/s3utils" | ||
29 | ) | ||
30 | |||
31 | // GetObject wrapper function that accepts a request context | ||
32 | func (c *Client) GetObject(ctx context.Context, bucketName, objectName string, opts GetObjectOptions) (*Object, error) { | ||
33 | // Input validation. | ||
34 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
35 | return nil, err | ||
36 | } | ||
37 | if err := s3utils.CheckValidObjectName(objectName); err != nil { | ||
38 | return nil, err | ||
39 | } | ||
40 | |||
41 | gctx, cancel := context.WithCancel(ctx) | ||
42 | |||
43 | // Detect if snowball is server location we are talking to. | ||
44 | var snowball bool | ||
45 | if location, ok := c.bucketLocCache.Get(bucketName); ok { | ||
46 | snowball = location == "snowball" | ||
47 | } | ||
48 | |||
49 | var ( | ||
50 | err error | ||
51 | httpReader io.ReadCloser | ||
52 | objectInfo ObjectInfo | ||
53 | totalRead int | ||
54 | ) | ||
55 | |||
56 | // Create request channel. | ||
57 | reqCh := make(chan getRequest) | ||
58 | // Create response channel. | ||
59 | resCh := make(chan getResponse) | ||
60 | |||
61 | // This routine feeds partial object data as and when the caller reads. | ||
62 | go func() { | ||
63 | defer close(resCh) | ||
64 | defer func() { | ||
65 | // Close the http response body before returning. | ||
66 | // This ends the connection with the server. | ||
67 | if httpReader != nil { | ||
68 | httpReader.Close() | ||
69 | } | ||
70 | }() | ||
71 | defer cancel() | ||
72 | |||
73 | // Used to verify if etag of object has changed since last read. | ||
74 | var etag string | ||
75 | |||
76 | for req := range reqCh { | ||
77 | // If this is the first request we may not need to do a getObject request yet. | ||
78 | if req.isFirstReq { | ||
79 | // First request is a Read/ReadAt. | ||
80 | if req.isReadOp { | ||
81 | // Differentiate between wanting the whole object and just a range. | ||
82 | if req.isReadAt { | ||
83 | // If this is a ReadAt request only get the specified range. | ||
84 | // Range is set with respect to the offset and length of the buffer requested. | ||
85 | // Do not set objectInfo from the first readAt request because it will not get | ||
86 | // the whole object. | ||
87 | opts.SetRange(req.Offset, req.Offset+int64(len(req.Buffer))-1) | ||
88 | } else if req.Offset > 0 { | ||
89 | opts.SetRange(req.Offset, 0) | ||
90 | } | ||
91 | httpReader, objectInfo, _, err = c.getObject(gctx, bucketName, objectName, opts) | ||
92 | if err != nil { | ||
93 | resCh <- getResponse{Error: err} | ||
94 | return | ||
95 | } | ||
96 | etag = objectInfo.ETag | ||
97 | // Read at least firstReq.Buffer bytes, if not we have | ||
98 | // reached our EOF. | ||
99 | size, err := readFull(httpReader, req.Buffer) | ||
100 | totalRead += size | ||
101 | if size > 0 && err == io.ErrUnexpectedEOF { | ||
102 | if int64(size) < objectInfo.Size { | ||
103 | // In situations when returned size | ||
104 | // is less than the expected content | ||
105 | // length set by the server, make sure | ||
106 | // we return io.ErrUnexpectedEOF | ||
107 | err = io.ErrUnexpectedEOF | ||
108 | } else { | ||
109 | // If an EOF happens after reading some but not | ||
110 | // all the bytes ReadFull returns ErrUnexpectedEOF | ||
111 | err = io.EOF | ||
112 | } | ||
113 | } else if size == 0 && err == io.EOF && objectInfo.Size > 0 { | ||
114 | // Special cases when server writes more data | ||
115 | // than the content-length, net/http response | ||
116 | // body returns an error, instead of converting | ||
117 | // it to io.EOF - return unexpected EOF. | ||
118 | err = io.ErrUnexpectedEOF | ||
119 | } | ||
120 | // Send back the first response. | ||
121 | resCh <- getResponse{ | ||
122 | objectInfo: objectInfo, | ||
123 | Size: size, | ||
124 | Error: err, | ||
125 | didRead: true, | ||
126 | } | ||
127 | } else { | ||
128 | // First request is a Stat or Seek call. | ||
129 | // Only need to run a StatObject until an actual Read or ReadAt request comes through. | ||
130 | |||
131 | // Remove range header if already set, for stat Operations to get original file size. | ||
132 | delete(opts.headers, "Range") | ||
133 | objectInfo, err = c.StatObject(gctx, bucketName, objectName, StatObjectOptions(opts)) | ||
134 | if err != nil { | ||
135 | resCh <- getResponse{ | ||
136 | Error: err, | ||
137 | } | ||
138 | // Exit the go-routine. | ||
139 | return | ||
140 | } | ||
141 | etag = objectInfo.ETag | ||
142 | // Send back the first response. | ||
143 | resCh <- getResponse{ | ||
144 | objectInfo: objectInfo, | ||
145 | } | ||
146 | } | ||
147 | } else if req.settingObjectInfo { // Request is just to get objectInfo. | ||
148 | // Remove range header if already set, for stat Operations to get original file size. | ||
149 | delete(opts.headers, "Range") | ||
150 | // Check whether this is snowball | ||
151 | // if yes do not use If-Match feature | ||
152 | // it doesn't work. | ||
153 | if etag != "" && !snowball { | ||
154 | opts.SetMatchETag(etag) | ||
155 | } | ||
156 | objectInfo, err := c.StatObject(gctx, bucketName, objectName, StatObjectOptions(opts)) | ||
157 | if err != nil { | ||
158 | resCh <- getResponse{ | ||
159 | Error: err, | ||
160 | } | ||
161 | // Exit the goroutine. | ||
162 | return | ||
163 | } | ||
164 | // Send back the objectInfo. | ||
165 | resCh <- getResponse{ | ||
166 | objectInfo: objectInfo, | ||
167 | } | ||
168 | } else { | ||
169 | // Offset changes fetch the new object at an Offset. | ||
170 | // Because the httpReader may not be set by the first | ||
171 | // request if it was a stat or seek it must be checked | ||
172 | // if the object has been read or not to only initialize | ||
173 | // new ones when they haven't been already. | ||
174 | // All readAt requests are new requests. | ||
175 | if req.DidOffsetChange || !req.beenRead { | ||
176 | // Check whether this is snowball | ||
177 | // if yes do not use If-Match feature | ||
178 | // it doesn't work. | ||
179 | if etag != "" && !snowball { | ||
180 | opts.SetMatchETag(etag) | ||
181 | } | ||
182 | if httpReader != nil { | ||
183 | // Close previously opened http reader. | ||
184 | httpReader.Close() | ||
185 | } | ||
186 | // If this request is a readAt only get the specified range. | ||
187 | if req.isReadAt { | ||
188 | // Range is set with respect to the offset and length of the buffer requested. | ||
189 | opts.SetRange(req.Offset, req.Offset+int64(len(req.Buffer))-1) | ||
190 | } else if req.Offset > 0 { // Range is set with respect to the offset. | ||
191 | opts.SetRange(req.Offset, 0) | ||
192 | } else { | ||
193 | // Remove range header if already set | ||
194 | delete(opts.headers, "Range") | ||
195 | } | ||
196 | httpReader, objectInfo, _, err = c.getObject(gctx, bucketName, objectName, opts) | ||
197 | if err != nil { | ||
198 | resCh <- getResponse{ | ||
199 | Error: err, | ||
200 | } | ||
201 | return | ||
202 | } | ||
203 | totalRead = 0 | ||
204 | } | ||
205 | |||
206 | // Read at least req.Buffer bytes, if not we have | ||
207 | // reached our EOF. | ||
208 | size, err := readFull(httpReader, req.Buffer) | ||
209 | totalRead += size | ||
210 | if size > 0 && err == io.ErrUnexpectedEOF { | ||
211 | if int64(totalRead) < objectInfo.Size { | ||
212 | // In situations when returned size | ||
213 | // is less than the expected content | ||
214 | // length set by the server, make sure | ||
215 | // we return io.ErrUnexpectedEOF | ||
216 | err = io.ErrUnexpectedEOF | ||
217 | } else { | ||
218 | // If an EOF happens after reading some but not | ||
219 | // all the bytes ReadFull returns ErrUnexpectedEOF | ||
220 | err = io.EOF | ||
221 | } | ||
222 | } else if size == 0 && err == io.EOF && objectInfo.Size > 0 { | ||
223 | // Special cases when server writes more data | ||
224 | // than the content-length, net/http response | ||
225 | // body returns an error, instead of converting | ||
226 | // it to io.EOF - return unexpected EOF. | ||
227 | err = io.ErrUnexpectedEOF | ||
228 | } | ||
229 | |||
230 | // Reply back how much was read. | ||
231 | resCh <- getResponse{ | ||
232 | Size: size, | ||
233 | Error: err, | ||
234 | didRead: true, | ||
235 | objectInfo: objectInfo, | ||
236 | } | ||
237 | } | ||
238 | } | ||
239 | }() | ||
240 | |||
241 | // Create a newObject through the information sent back by reqCh. | ||
242 | return newObject(gctx, cancel, reqCh, resCh), nil | ||
243 | } | ||
244 | |||
245 | // get request message container to communicate with internal | ||
246 | // go-routine. | ||
247 | type getRequest struct { | ||
248 | Buffer []byte | ||
249 | Offset int64 // readAt offset. | ||
250 | DidOffsetChange bool // Tracks the offset changes for Seek requests. | ||
251 | beenRead bool // Determines if this is the first time an object is being read. | ||
252 | isReadAt bool // Determines if this request is a request to a specific range | ||
253 | isReadOp bool // Determines if this request is a Read or Read/At request. | ||
254 | isFirstReq bool // Determines if this request is the first time an object is being accessed. | ||
255 | settingObjectInfo bool // Determines if this request is to set the objectInfo of an object. | ||
256 | } | ||
257 | |||
258 | // get response message container to reply back for the request. | ||
259 | type getResponse struct { | ||
260 | Size int | ||
261 | Error error | ||
262 | didRead bool // Lets subsequent calls know whether or not httpReader has been initiated. | ||
263 | objectInfo ObjectInfo // Used for the first request. | ||
264 | } | ||
265 | |||
266 | // Object represents an open object. It implements | ||
267 | // Reader, ReaderAt, Seeker, Closer for a HTTP stream. | ||
268 | type Object struct { | ||
269 | // Mutex. | ||
270 | mutex *sync.Mutex | ||
271 | |||
272 | // User allocated and defined. | ||
273 | reqCh chan<- getRequest | ||
274 | resCh <-chan getResponse | ||
275 | ctx context.Context | ||
276 | cancel context.CancelFunc | ||
277 | currOffset int64 | ||
278 | objectInfo ObjectInfo | ||
279 | |||
280 | // Ask lower level to initiate data fetching based on currOffset | ||
281 | seekData bool | ||
282 | |||
283 | // Keeps track of closed call. | ||
284 | isClosed bool | ||
285 | |||
286 | // Keeps track of if this is the first call. | ||
287 | isStarted bool | ||
288 | |||
289 | // Previous error saved for future calls. | ||
290 | prevErr error | ||
291 | |||
292 | // Keeps track of if this object has been read yet. | ||
293 | beenRead bool | ||
294 | |||
295 | // Keeps track of if objectInfo has been set yet. | ||
296 | objectInfoSet bool | ||
297 | } | ||
298 | |||
299 | // doGetRequest - sends and blocks on the firstReqCh and reqCh of an object. | ||
300 | // Returns back the size of the buffer read, if anything was read, as well | ||
301 | // as any error encountered. For all first requests sent on the object | ||
302 | // it is also responsible for sending back the objectInfo. | ||
303 | func (o *Object) doGetRequest(request getRequest) (getResponse, error) { | ||
304 | select { | ||
305 | case <-o.ctx.Done(): | ||
306 | return getResponse{}, o.ctx.Err() | ||
307 | case o.reqCh <- request: | ||
308 | } | ||
309 | |||
310 | response := <-o.resCh | ||
311 | |||
312 | // Return any error to the top level. | ||
313 | if response.Error != nil { | ||
314 | return response, response.Error | ||
315 | } | ||
316 | |||
317 | // This was the first request. | ||
318 | if !o.isStarted { | ||
319 | // The object has been operated on. | ||
320 | o.isStarted = true | ||
321 | } | ||
322 | // Set the objectInfo if the request was not readAt | ||
323 | // and it hasn't been set before. | ||
324 | if !o.objectInfoSet && !request.isReadAt { | ||
325 | o.objectInfo = response.objectInfo | ||
326 | o.objectInfoSet = true | ||
327 | } | ||
328 | // Set beenRead only if it has not been set before. | ||
329 | if !o.beenRead { | ||
330 | o.beenRead = response.didRead | ||
331 | } | ||
332 | // Data are ready on the wire, no need to reinitiate connection in lower level | ||
333 | o.seekData = false | ||
334 | |||
335 | return response, nil | ||
336 | } | ||
337 | |||
338 | // setOffset - handles the setting of offsets for | ||
339 | // Read/ReadAt/Seek requests. | ||
340 | func (o *Object) setOffset(bytesRead int64) error { | ||
341 | // Update the currentOffset. | ||
342 | o.currOffset += bytesRead | ||
343 | |||
344 | if o.objectInfo.Size > -1 && o.currOffset >= o.objectInfo.Size { | ||
345 | return io.EOF | ||
346 | } | ||
347 | return nil | ||
348 | } | ||
349 | |||
350 | // Read reads up to len(b) bytes into b. It returns the number of | ||
351 | // bytes read (0 <= n <= len(b)) and any error encountered. Returns | ||
352 | // io.EOF upon end of file. | ||
353 | func (o *Object) Read(b []byte) (n int, err error) { | ||
354 | if o == nil { | ||
355 | return 0, errInvalidArgument("Object is nil") | ||
356 | } | ||
357 | |||
358 | // Locking. | ||
359 | o.mutex.Lock() | ||
360 | defer o.mutex.Unlock() | ||
361 | |||
362 | // prevErr is previous error saved from previous operation. | ||
363 | if o.prevErr != nil || o.isClosed { | ||
364 | return 0, o.prevErr | ||
365 | } | ||
366 | |||
367 | // Create a new request. | ||
368 | readReq := getRequest{ | ||
369 | isReadOp: true, | ||
370 | beenRead: o.beenRead, | ||
371 | Buffer: b, | ||
372 | } | ||
373 | |||
374 | // Alert that this is the first request. | ||
375 | if !o.isStarted { | ||
376 | readReq.isFirstReq = true | ||
377 | } | ||
378 | |||
379 | // Ask to establish a new data fetch routine based on seekData flag | ||
380 | readReq.DidOffsetChange = o.seekData | ||
381 | readReq.Offset = o.currOffset | ||
382 | |||
383 | // Send and receive from the first request. | ||
384 | response, err := o.doGetRequest(readReq) | ||
385 | if err != nil && err != io.EOF { | ||
386 | // Save the error for future calls. | ||
387 | o.prevErr = err | ||
388 | return response.Size, err | ||
389 | } | ||
390 | |||
391 | // Bytes read. | ||
392 | bytesRead := int64(response.Size) | ||
393 | |||
394 | // Set the new offset. | ||
395 | oerr := o.setOffset(bytesRead) | ||
396 | if oerr != nil { | ||
397 | // Save the error for future calls. | ||
398 | o.prevErr = oerr | ||
399 | return response.Size, oerr | ||
400 | } | ||
401 | |||
402 | // Return the response. | ||
403 | return response.Size, err | ||
404 | } | ||
405 | |||
406 | // Stat returns the ObjectInfo structure describing Object. | ||
407 | func (o *Object) Stat() (ObjectInfo, error) { | ||
408 | if o == nil { | ||
409 | return ObjectInfo{}, errInvalidArgument("Object is nil") | ||
410 | } | ||
411 | // Locking. | ||
412 | o.mutex.Lock() | ||
413 | defer o.mutex.Unlock() | ||
414 | |||
415 | if o.prevErr != nil && o.prevErr != io.EOF || o.isClosed { | ||
416 | return ObjectInfo{}, o.prevErr | ||
417 | } | ||
418 | |||
419 | // This is the first request. | ||
420 | if !o.isStarted || !o.objectInfoSet { | ||
421 | // Send the request and get the response. | ||
422 | _, err := o.doGetRequest(getRequest{ | ||
423 | isFirstReq: !o.isStarted, | ||
424 | settingObjectInfo: !o.objectInfoSet, | ||
425 | }) | ||
426 | if err != nil { | ||
427 | o.prevErr = err | ||
428 | return ObjectInfo{}, err | ||
429 | } | ||
430 | } | ||
431 | |||
432 | return o.objectInfo, nil | ||
433 | } | ||
434 | |||
435 | // ReadAt reads len(b) bytes from the File starting at byte offset | ||
436 | // off. It returns the number of bytes read and the error, if any. | ||
437 | // ReadAt always returns a non-nil error when n < len(b). At end of | ||
438 | // file, that error is io.EOF. | ||
439 | func (o *Object) ReadAt(b []byte, offset int64) (n int, err error) { | ||
440 | if o == nil { | ||
441 | return 0, errInvalidArgument("Object is nil") | ||
442 | } | ||
443 | |||
444 | // Locking. | ||
445 | o.mutex.Lock() | ||
446 | defer o.mutex.Unlock() | ||
447 | |||
448 | // prevErr is error which was saved in previous operation. | ||
449 | if o.prevErr != nil && o.prevErr != io.EOF || o.isClosed { | ||
450 | return 0, o.prevErr | ||
451 | } | ||
452 | |||
453 | // Set the current offset to ReadAt offset, because the current offset will be shifted at the end of this method. | ||
454 | o.currOffset = offset | ||
455 | |||
456 | // Can only compare offsets to size when size has been set. | ||
457 | if o.objectInfoSet { | ||
458 | // If offset is negative than we return io.EOF. | ||
459 | // If offset is greater than or equal to object size we return io.EOF. | ||
460 | if (o.objectInfo.Size > -1 && offset >= o.objectInfo.Size) || offset < 0 { | ||
461 | return 0, io.EOF | ||
462 | } | ||
463 | } | ||
464 | |||
465 | // Create the new readAt request. | ||
466 | readAtReq := getRequest{ | ||
467 | isReadOp: true, | ||
468 | isReadAt: true, | ||
469 | DidOffsetChange: true, // Offset always changes. | ||
470 | beenRead: o.beenRead, // Set if this is the first request to try and read. | ||
471 | Offset: offset, // Set the offset. | ||
472 | Buffer: b, | ||
473 | } | ||
474 | |||
475 | // Alert that this is the first request. | ||
476 | if !o.isStarted { | ||
477 | readAtReq.isFirstReq = true | ||
478 | } | ||
479 | |||
480 | // Send and receive from the first request. | ||
481 | response, err := o.doGetRequest(readAtReq) | ||
482 | if err != nil && err != io.EOF { | ||
483 | // Save the error. | ||
484 | o.prevErr = err | ||
485 | return response.Size, err | ||
486 | } | ||
487 | // Bytes read. | ||
488 | bytesRead := int64(response.Size) | ||
489 | // There is no valid objectInfo yet | ||
490 | // to compare against for EOF. | ||
491 | if !o.objectInfoSet { | ||
492 | // Update the currentOffset. | ||
493 | o.currOffset += bytesRead | ||
494 | } else { | ||
495 | // If this was not the first request update | ||
496 | // the offsets and compare against objectInfo | ||
497 | // for EOF. | ||
498 | oerr := o.setOffset(bytesRead) | ||
499 | if oerr != nil { | ||
500 | o.prevErr = oerr | ||
501 | return response.Size, oerr | ||
502 | } | ||
503 | } | ||
504 | return response.Size, err | ||
505 | } | ||
506 | |||
507 | // Seek sets the offset for the next Read or Write to offset, | ||
508 | // interpreted according to whence: 0 means relative to the | ||
509 | // origin of the file, 1 means relative to the current offset, | ||
510 | // and 2 means relative to the end. | ||
511 | // Seek returns the new offset and an error, if any. | ||
512 | // | ||
513 | // Seeking to a negative offset is an error. Seeking to any positive | ||
514 | // offset is legal, subsequent io operations succeed until the | ||
515 | // underlying object is not closed. | ||
516 | func (o *Object) Seek(offset int64, whence int) (n int64, err error) { | ||
517 | if o == nil { | ||
518 | return 0, errInvalidArgument("Object is nil") | ||
519 | } | ||
520 | |||
521 | // Locking. | ||
522 | o.mutex.Lock() | ||
523 | defer o.mutex.Unlock() | ||
524 | |||
525 | // At EOF seeking is legal allow only io.EOF, for any other errors we return. | ||
526 | if o.prevErr != nil && o.prevErr != io.EOF { | ||
527 | return 0, o.prevErr | ||
528 | } | ||
529 | |||
530 | // Negative offset is valid for whence of '2'. | ||
531 | if offset < 0 && whence != 2 { | ||
532 | return 0, errInvalidArgument(fmt.Sprintf("Negative position not allowed for %d", whence)) | ||
533 | } | ||
534 | |||
535 | // This is the first request. So before anything else | ||
536 | // get the ObjectInfo. | ||
537 | if !o.isStarted || !o.objectInfoSet { | ||
538 | // Create the new Seek request. | ||
539 | seekReq := getRequest{ | ||
540 | isReadOp: false, | ||
541 | Offset: offset, | ||
542 | isFirstReq: true, | ||
543 | } | ||
544 | // Send and receive from the seek request. | ||
545 | _, err := o.doGetRequest(seekReq) | ||
546 | if err != nil { | ||
547 | // Save the error. | ||
548 | o.prevErr = err | ||
549 | return 0, err | ||
550 | } | ||
551 | } | ||
552 | |||
553 | newOffset := o.currOffset | ||
554 | |||
555 | // Switch through whence. | ||
556 | switch whence { | ||
557 | default: | ||
558 | return 0, errInvalidArgument(fmt.Sprintf("Invalid whence %d", whence)) | ||
559 | case 0: | ||
560 | if o.objectInfo.Size > -1 && offset > o.objectInfo.Size { | ||
561 | return 0, io.EOF | ||
562 | } | ||
563 | newOffset = offset | ||
564 | case 1: | ||
565 | if o.objectInfo.Size > -1 && o.currOffset+offset > o.objectInfo.Size { | ||
566 | return 0, io.EOF | ||
567 | } | ||
568 | newOffset += offset | ||
569 | case 2: | ||
570 | // If we don't know the object size return an error for io.SeekEnd | ||
571 | if o.objectInfo.Size < 0 { | ||
572 | return 0, errInvalidArgument("Whence END is not supported when the object size is unknown") | ||
573 | } | ||
574 | // Seeking to positive offset is valid for whence '2', but | ||
575 | // since we are backing a Reader we have reached 'EOF' if | ||
576 | // offset is positive. | ||
577 | if offset > 0 { | ||
578 | return 0, io.EOF | ||
579 | } | ||
580 | // Seeking to negative position not allowed for whence. | ||
581 | if o.objectInfo.Size+offset < 0 { | ||
582 | return 0, errInvalidArgument(fmt.Sprintf("Seeking at negative offset not allowed for %d", whence)) | ||
583 | } | ||
584 | newOffset = o.objectInfo.Size + offset | ||
585 | } | ||
586 | // Reset the saved error since we successfully seeked, let the Read | ||
587 | // and ReadAt decide. | ||
588 | if o.prevErr == io.EOF { | ||
589 | o.prevErr = nil | ||
590 | } | ||
591 | |||
592 | // Ask lower level to fetch again from source when necessary | ||
593 | o.seekData = (newOffset != o.currOffset) || o.seekData | ||
594 | o.currOffset = newOffset | ||
595 | |||
596 | // Return the effective offset. | ||
597 | return o.currOffset, nil | ||
598 | } | ||
599 | |||
600 | // Close - The behavior of Close after the first call returns error | ||
601 | // for subsequent Close() calls. | ||
602 | func (o *Object) Close() (err error) { | ||
603 | if o == nil { | ||
604 | return errInvalidArgument("Object is nil") | ||
605 | } | ||
606 | |||
607 | // Locking. | ||
608 | o.mutex.Lock() | ||
609 | defer o.mutex.Unlock() | ||
610 | |||
611 | // if already closed return an error. | ||
612 | if o.isClosed { | ||
613 | return o.prevErr | ||
614 | } | ||
615 | |||
616 | // Close successfully. | ||
617 | o.cancel() | ||
618 | |||
619 | // Close the request channel to indicate the internal go-routine to exit. | ||
620 | close(o.reqCh) | ||
621 | |||
622 | // Save for future operations. | ||
623 | errMsg := "Object is already closed. Bad file descriptor." | ||
624 | o.prevErr = errors.New(errMsg) | ||
625 | // Save here that we closed done channel successfully. | ||
626 | o.isClosed = true | ||
627 | return nil | ||
628 | } | ||
629 | |||
630 | // newObject instantiates a new *minio.Object* | ||
631 | // ObjectInfo will be set by setObjectInfo | ||
632 | func newObject(ctx context.Context, cancel context.CancelFunc, reqCh chan<- getRequest, resCh <-chan getResponse) *Object { | ||
633 | return &Object{ | ||
634 | ctx: ctx, | ||
635 | cancel: cancel, | ||
636 | mutex: &sync.Mutex{}, | ||
637 | reqCh: reqCh, | ||
638 | resCh: resCh, | ||
639 | } | ||
640 | } | ||
641 | |||
642 | // getObject - retrieve object from Object Storage. | ||
643 | // | ||
644 | // Additionally this function also takes range arguments to download the specified | ||
645 | // range bytes of an object. Setting offset and length = 0 will download the full object. | ||
646 | // | ||
647 | // For more information about the HTTP Range header. | ||
648 | // go to http://www.w3.org/Protocols/rfc2616/rfc2616-sec14.html#sec14.35. | ||
649 | func (c *Client) getObject(ctx context.Context, bucketName, objectName string, opts GetObjectOptions) (io.ReadCloser, ObjectInfo, http.Header, error) { | ||
650 | // Validate input arguments. | ||
651 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
652 | return nil, ObjectInfo{}, nil, err | ||
653 | } | ||
654 | if err := s3utils.CheckValidObjectName(objectName); err != nil { | ||
655 | return nil, ObjectInfo{}, nil, err | ||
656 | } | ||
657 | |||
658 | // Execute GET on objectName. | ||
659 | resp, err := c.executeMethod(ctx, http.MethodGet, requestMetadata{ | ||
660 | bucketName: bucketName, | ||
661 | objectName: objectName, | ||
662 | queryValues: opts.toQueryValues(), | ||
663 | customHeader: opts.Header(), | ||
664 | contentSHA256Hex: emptySHA256Hex, | ||
665 | }) | ||
666 | if err != nil { | ||
667 | return nil, ObjectInfo{}, nil, err | ||
668 | } | ||
669 | if resp != nil { | ||
670 | if resp.StatusCode != http.StatusOK && resp.StatusCode != http.StatusPartialContent { | ||
671 | return nil, ObjectInfo{}, nil, httpRespToErrorResponse(resp, bucketName, objectName) | ||
672 | } | ||
673 | } | ||
674 | |||
675 | objectStat, err := ToObjectInfo(bucketName, objectName, resp.Header) | ||
676 | if err != nil { | ||
677 | closeResponse(resp) | ||
678 | return nil, ObjectInfo{}, nil, err | ||
679 | } | ||
680 | |||
681 | // do not close body here, caller will close | ||
682 | return resp.Body, objectStat, resp.Header, nil | ||
683 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/api-get-options.go b/vendor/github.com/minio/minio-go/v7/api-get-options.go deleted file mode 100644 index a0216e2..0000000 --- a/vendor/github.com/minio/minio-go/v7/api-get-options.go +++ /dev/null | |||
@@ -1,203 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2015-2020 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package minio | ||
19 | |||
20 | import ( | ||
21 | "fmt" | ||
22 | "net/http" | ||
23 | "net/url" | ||
24 | "strconv" | ||
25 | "time" | ||
26 | |||
27 | "github.com/minio/minio-go/v7/pkg/encrypt" | ||
28 | ) | ||
29 | |||
30 | // AdvancedGetOptions for internal use by MinIO server - not intended for client use. | ||
31 | type AdvancedGetOptions struct { | ||
32 | ReplicationDeleteMarker bool | ||
33 | IsReplicationReadyForDeleteMarker bool | ||
34 | ReplicationProxyRequest string | ||
35 | } | ||
36 | |||
37 | // GetObjectOptions are used to specify additional headers or options | ||
38 | // during GET requests. | ||
39 | type GetObjectOptions struct { | ||
40 | headers map[string]string | ||
41 | reqParams url.Values | ||
42 | ServerSideEncryption encrypt.ServerSide | ||
43 | VersionID string | ||
44 | PartNumber int | ||
45 | |||
46 | // Include any checksums, if object was uploaded with checksum. | ||
47 | // For multipart objects this is a checksum of part checksums. | ||
48 | // https://docs.aws.amazon.com/AmazonS3/latest/userguide/checking-object-integrity.html | ||
49 | Checksum bool | ||
50 | |||
51 | // To be not used by external applications | ||
52 | Internal AdvancedGetOptions | ||
53 | } | ||
54 | |||
55 | // StatObjectOptions are used to specify additional headers or options | ||
56 | // during GET info/stat requests. | ||
57 | type StatObjectOptions = GetObjectOptions | ||
58 | |||
59 | // Header returns the http.Header representation of the GET options. | ||
60 | func (o GetObjectOptions) Header() http.Header { | ||
61 | headers := make(http.Header, len(o.headers)) | ||
62 | for k, v := range o.headers { | ||
63 | headers.Set(k, v) | ||
64 | } | ||
65 | if o.ServerSideEncryption != nil && o.ServerSideEncryption.Type() == encrypt.SSEC { | ||
66 | o.ServerSideEncryption.Marshal(headers) | ||
67 | } | ||
68 | // this header is set for active-active replication scenario where GET/HEAD | ||
69 | // to site A is proxy'd to site B if object/version missing on site A. | ||
70 | if o.Internal.ReplicationProxyRequest != "" { | ||
71 | headers.Set(minIOBucketReplicationProxyRequest, o.Internal.ReplicationProxyRequest) | ||
72 | } | ||
73 | if o.Checksum { | ||
74 | headers.Set("x-amz-checksum-mode", "ENABLED") | ||
75 | } | ||
76 | return headers | ||
77 | } | ||
78 | |||
79 | // Set adds a key value pair to the options. The | ||
80 | // key-value pair will be part of the HTTP GET request | ||
81 | // headers. | ||
82 | func (o *GetObjectOptions) Set(key, value string) { | ||
83 | if o.headers == nil { | ||
84 | o.headers = make(map[string]string) | ||
85 | } | ||
86 | o.headers[http.CanonicalHeaderKey(key)] = value | ||
87 | } | ||
88 | |||
89 | // SetReqParam - set request query string parameter | ||
90 | // supported key: see supportedQueryValues and allowedCustomQueryPrefix. | ||
91 | // If an unsupported key is passed in, it will be ignored and nothing will be done. | ||
92 | func (o *GetObjectOptions) SetReqParam(key, value string) { | ||
93 | if !isCustomQueryValue(key) && !isStandardQueryValue(key) { | ||
94 | // do nothing | ||
95 | return | ||
96 | } | ||
97 | if o.reqParams == nil { | ||
98 | o.reqParams = make(url.Values) | ||
99 | } | ||
100 | o.reqParams.Set(key, value) | ||
101 | } | ||
102 | |||
103 | // AddReqParam - add request query string parameter | ||
104 | // supported key: see supportedQueryValues and allowedCustomQueryPrefix. | ||
105 | // If an unsupported key is passed in, it will be ignored and nothing will be done. | ||
106 | func (o *GetObjectOptions) AddReqParam(key, value string) { | ||
107 | if !isCustomQueryValue(key) && !isStandardQueryValue(key) { | ||
108 | // do nothing | ||
109 | return | ||
110 | } | ||
111 | if o.reqParams == nil { | ||
112 | o.reqParams = make(url.Values) | ||
113 | } | ||
114 | o.reqParams.Add(key, value) | ||
115 | } | ||
116 | |||
117 | // SetMatchETag - set match etag. | ||
118 | func (o *GetObjectOptions) SetMatchETag(etag string) error { | ||
119 | if etag == "" { | ||
120 | return errInvalidArgument("ETag cannot be empty.") | ||
121 | } | ||
122 | o.Set("If-Match", "\""+etag+"\"") | ||
123 | return nil | ||
124 | } | ||
125 | |||
126 | // SetMatchETagExcept - set match etag except. | ||
127 | func (o *GetObjectOptions) SetMatchETagExcept(etag string) error { | ||
128 | if etag == "" { | ||
129 | return errInvalidArgument("ETag cannot be empty.") | ||
130 | } | ||
131 | o.Set("If-None-Match", "\""+etag+"\"") | ||
132 | return nil | ||
133 | } | ||
134 | |||
135 | // SetUnmodified - set unmodified time since. | ||
136 | func (o *GetObjectOptions) SetUnmodified(modTime time.Time) error { | ||
137 | if modTime.IsZero() { | ||
138 | return errInvalidArgument("Modified since cannot be empty.") | ||
139 | } | ||
140 | o.Set("If-Unmodified-Since", modTime.Format(http.TimeFormat)) | ||
141 | return nil | ||
142 | } | ||
143 | |||
144 | // SetModified - set modified time since. | ||
145 | func (o *GetObjectOptions) SetModified(modTime time.Time) error { | ||
146 | if modTime.IsZero() { | ||
147 | return errInvalidArgument("Modified since cannot be empty.") | ||
148 | } | ||
149 | o.Set("If-Modified-Since", modTime.Format(http.TimeFormat)) | ||
150 | return nil | ||
151 | } | ||
152 | |||
153 | // SetRange - set the start and end offset of the object to be read. | ||
154 | // See https://tools.ietf.org/html/rfc7233#section-3.1 for reference. | ||
155 | func (o *GetObjectOptions) SetRange(start, end int64) error { | ||
156 | switch { | ||
157 | case start == 0 && end < 0: | ||
158 | // Read last '-end' bytes. `bytes=-N`. | ||
159 | o.Set("Range", fmt.Sprintf("bytes=%d", end)) | ||
160 | case 0 < start && end == 0: | ||
161 | // Read everything starting from offset | ||
162 | // 'start'. `bytes=N-`. | ||
163 | o.Set("Range", fmt.Sprintf("bytes=%d-", start)) | ||
164 | case 0 <= start && start <= end: | ||
165 | // Read everything starting at 'start' till the | ||
166 | // 'end'. `bytes=N-M` | ||
167 | o.Set("Range", fmt.Sprintf("bytes=%d-%d", start, end)) | ||
168 | default: | ||
169 | // All other cases such as | ||
170 | // bytes=-3- | ||
171 | // bytes=5-3 | ||
172 | // bytes=-2-4 | ||
173 | // bytes=-3-0 | ||
174 | // bytes=-3--2 | ||
175 | // are invalid. | ||
176 | return errInvalidArgument( | ||
177 | fmt.Sprintf( | ||
178 | "Invalid range specified: start=%d end=%d", | ||
179 | start, end)) | ||
180 | } | ||
181 | return nil | ||
182 | } | ||
183 | |||
184 | // toQueryValues - Convert the versionId, partNumber, and reqParams in Options to query string parameters. | ||
185 | func (o *GetObjectOptions) toQueryValues() url.Values { | ||
186 | urlValues := make(url.Values) | ||
187 | if o.VersionID != "" { | ||
188 | urlValues.Set("versionId", o.VersionID) | ||
189 | } | ||
190 | if o.PartNumber > 0 { | ||
191 | urlValues.Set("partNumber", strconv.Itoa(o.PartNumber)) | ||
192 | } | ||
193 | |||
194 | if o.reqParams != nil { | ||
195 | for key, values := range o.reqParams { | ||
196 | for _, value := range values { | ||
197 | urlValues.Add(key, value) | ||
198 | } | ||
199 | } | ||
200 | } | ||
201 | |||
202 | return urlValues | ||
203 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/api-list.go b/vendor/github.com/minio/minio-go/v7/api-list.go deleted file mode 100644 index 31b6edf..0000000 --- a/vendor/github.com/minio/minio-go/v7/api-list.go +++ /dev/null | |||
@@ -1,1057 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2015-2020 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package minio | ||
19 | |||
20 | import ( | ||
21 | "context" | ||
22 | "fmt" | ||
23 | "net/http" | ||
24 | "net/url" | ||
25 | "time" | ||
26 | |||
27 | "github.com/minio/minio-go/v7/pkg/s3utils" | ||
28 | ) | ||
29 | |||
30 | // ListBuckets list all buckets owned by this authenticated user. | ||
31 | // | ||
32 | // This call requires explicit authentication, no anonymous requests are | ||
33 | // allowed for listing buckets. | ||
34 | // | ||
35 | // api := client.New(....) | ||
36 | // for message := range api.ListBuckets(context.Background()) { | ||
37 | // fmt.Println(message) | ||
38 | // } | ||
39 | func (c *Client) ListBuckets(ctx context.Context) ([]BucketInfo, error) { | ||
40 | // Execute GET on service. | ||
41 | resp, err := c.executeMethod(ctx, http.MethodGet, requestMetadata{contentSHA256Hex: emptySHA256Hex}) | ||
42 | defer closeResponse(resp) | ||
43 | if err != nil { | ||
44 | return nil, err | ||
45 | } | ||
46 | if resp != nil { | ||
47 | if resp.StatusCode != http.StatusOK { | ||
48 | return nil, httpRespToErrorResponse(resp, "", "") | ||
49 | } | ||
50 | } | ||
51 | listAllMyBucketsResult := listAllMyBucketsResult{} | ||
52 | err = xmlDecoder(resp.Body, &listAllMyBucketsResult) | ||
53 | if err != nil { | ||
54 | return nil, err | ||
55 | } | ||
56 | return listAllMyBucketsResult.Buckets.Bucket, nil | ||
57 | } | ||
58 | |||
59 | // Bucket List Operations. | ||
60 | func (c *Client) listObjectsV2(ctx context.Context, bucketName string, opts ListObjectsOptions) <-chan ObjectInfo { | ||
61 | // Allocate new list objects channel. | ||
62 | objectStatCh := make(chan ObjectInfo, 1) | ||
63 | // Default listing is delimited at "/" | ||
64 | delimiter := "/" | ||
65 | if opts.Recursive { | ||
66 | // If recursive we do not delimit. | ||
67 | delimiter = "" | ||
68 | } | ||
69 | |||
70 | // Return object owner information by default | ||
71 | fetchOwner := true | ||
72 | |||
73 | sendObjectInfo := func(info ObjectInfo) { | ||
74 | select { | ||
75 | case objectStatCh <- info: | ||
76 | case <-ctx.Done(): | ||
77 | } | ||
78 | } | ||
79 | |||
80 | // Validate bucket name. | ||
81 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
82 | defer close(objectStatCh) | ||
83 | sendObjectInfo(ObjectInfo{ | ||
84 | Err: err, | ||
85 | }) | ||
86 | return objectStatCh | ||
87 | } | ||
88 | |||
89 | // Validate incoming object prefix. | ||
90 | if err := s3utils.CheckValidObjectNamePrefix(opts.Prefix); err != nil { | ||
91 | defer close(objectStatCh) | ||
92 | sendObjectInfo(ObjectInfo{ | ||
93 | Err: err, | ||
94 | }) | ||
95 | return objectStatCh | ||
96 | } | ||
97 | |||
98 | // Initiate list objects goroutine here. | ||
99 | go func(objectStatCh chan<- ObjectInfo) { | ||
100 | defer func() { | ||
101 | if contextCanceled(ctx) { | ||
102 | objectStatCh <- ObjectInfo{ | ||
103 | Err: ctx.Err(), | ||
104 | } | ||
105 | } | ||
106 | close(objectStatCh) | ||
107 | }() | ||
108 | |||
109 | // Save continuationToken for next request. | ||
110 | var continuationToken string | ||
111 | for { | ||
112 | // Get list of objects a maximum of 1000 per request. | ||
113 | result, err := c.listObjectsV2Query(ctx, bucketName, opts.Prefix, continuationToken, | ||
114 | fetchOwner, opts.WithMetadata, delimiter, opts.StartAfter, opts.MaxKeys, opts.headers) | ||
115 | if err != nil { | ||
116 | sendObjectInfo(ObjectInfo{ | ||
117 | Err: err, | ||
118 | }) | ||
119 | return | ||
120 | } | ||
121 | |||
122 | // If contents are available loop through and send over channel. | ||
123 | for _, object := range result.Contents { | ||
124 | object.ETag = trimEtag(object.ETag) | ||
125 | select { | ||
126 | // Send object content. | ||
127 | case objectStatCh <- object: | ||
128 | // If receives done from the caller, return here. | ||
129 | case <-ctx.Done(): | ||
130 | return | ||
131 | } | ||
132 | } | ||
133 | |||
134 | // Send all common prefixes if any. | ||
135 | // NOTE: prefixes are only present if the request is delimited. | ||
136 | for _, obj := range result.CommonPrefixes { | ||
137 | select { | ||
138 | // Send object prefixes. | ||
139 | case objectStatCh <- ObjectInfo{Key: obj.Prefix}: | ||
140 | // If receives done from the caller, return here. | ||
141 | case <-ctx.Done(): | ||
142 | return | ||
143 | } | ||
144 | } | ||
145 | |||
146 | // If continuation token present, save it for next request. | ||
147 | if result.NextContinuationToken != "" { | ||
148 | continuationToken = result.NextContinuationToken | ||
149 | } | ||
150 | |||
151 | // Listing ends result is not truncated, return right here. | ||
152 | if !result.IsTruncated { | ||
153 | return | ||
154 | } | ||
155 | |||
156 | // Add this to catch broken S3 API implementations. | ||
157 | if continuationToken == "" { | ||
158 | sendObjectInfo(ObjectInfo{ | ||
159 | Err: fmt.Errorf("listObjectsV2 is truncated without continuationToken, %s S3 server is incompatible with S3 API", c.endpointURL), | ||
160 | }) | ||
161 | return | ||
162 | } | ||
163 | } | ||
164 | }(objectStatCh) | ||
165 | return objectStatCh | ||
166 | } | ||
167 | |||
168 | // listObjectsV2Query - (List Objects V2) - List some or all (up to 1000) of the objects in a bucket. | ||
169 | // | ||
170 | // You can use the request parameters as selection criteria to return a subset of the objects in a bucket. | ||
171 | // request parameters :- | ||
172 | // --------- | ||
173 | // ?prefix - Limits the response to keys that begin with the specified prefix. | ||
174 | // ?continuation-token - Used to continue iterating over a set of objects | ||
175 | // ?metadata - Specifies if we want metadata for the objects as part of list operation. | ||
176 | // ?delimiter - A delimiter is a character you use to group keys. | ||
177 | // ?start-after - Sets a marker to start listing lexically at this key onwards. | ||
178 | // ?max-keys - Sets the maximum number of keys returned in the response body. | ||
179 | func (c *Client) listObjectsV2Query(ctx context.Context, bucketName, objectPrefix, continuationToken string, fetchOwner, metadata bool, delimiter, startAfter string, maxkeys int, headers http.Header) (ListBucketV2Result, error) { | ||
180 | // Validate bucket name. | ||
181 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
182 | return ListBucketV2Result{}, err | ||
183 | } | ||
184 | // Validate object prefix. | ||
185 | if err := s3utils.CheckValidObjectNamePrefix(objectPrefix); err != nil { | ||
186 | return ListBucketV2Result{}, err | ||
187 | } | ||
188 | // Get resources properly escaped and lined up before | ||
189 | // using them in http request. | ||
190 | urlValues := make(url.Values) | ||
191 | |||
192 | // Always set list-type in ListObjects V2 | ||
193 | urlValues.Set("list-type", "2") | ||
194 | |||
195 | if metadata { | ||
196 | urlValues.Set("metadata", "true") | ||
197 | } | ||
198 | |||
199 | // Set this conditionally if asked | ||
200 | if startAfter != "" { | ||
201 | urlValues.Set("start-after", startAfter) | ||
202 | } | ||
203 | |||
204 | // Always set encoding-type in ListObjects V2 | ||
205 | urlValues.Set("encoding-type", "url") | ||
206 | |||
207 | // Set object prefix, prefix value to be set to empty is okay. | ||
208 | urlValues.Set("prefix", objectPrefix) | ||
209 | |||
210 | // Set delimiter, delimiter value to be set to empty is okay. | ||
211 | urlValues.Set("delimiter", delimiter) | ||
212 | |||
213 | // Set continuation token | ||
214 | if continuationToken != "" { | ||
215 | urlValues.Set("continuation-token", continuationToken) | ||
216 | } | ||
217 | |||
218 | // Fetch owner when listing | ||
219 | if fetchOwner { | ||
220 | urlValues.Set("fetch-owner", "true") | ||
221 | } | ||
222 | |||
223 | // Set max keys. | ||
224 | if maxkeys > 0 { | ||
225 | urlValues.Set("max-keys", fmt.Sprintf("%d", maxkeys)) | ||
226 | } | ||
227 | |||
228 | // Execute GET on bucket to list objects. | ||
229 | resp, err := c.executeMethod(ctx, http.MethodGet, requestMetadata{ | ||
230 | bucketName: bucketName, | ||
231 | queryValues: urlValues, | ||
232 | contentSHA256Hex: emptySHA256Hex, | ||
233 | customHeader: headers, | ||
234 | }) | ||
235 | defer closeResponse(resp) | ||
236 | if err != nil { | ||
237 | return ListBucketV2Result{}, err | ||
238 | } | ||
239 | if resp != nil { | ||
240 | if resp.StatusCode != http.StatusOK { | ||
241 | return ListBucketV2Result{}, httpRespToErrorResponse(resp, bucketName, "") | ||
242 | } | ||
243 | } | ||
244 | |||
245 | // Decode listBuckets XML. | ||
246 | listBucketResult := ListBucketV2Result{} | ||
247 | if err = xmlDecoder(resp.Body, &listBucketResult); err != nil { | ||
248 | return listBucketResult, err | ||
249 | } | ||
250 | |||
251 | // This is an additional verification check to make | ||
252 | // sure proper responses are received. | ||
253 | if listBucketResult.IsTruncated && listBucketResult.NextContinuationToken == "" { | ||
254 | return listBucketResult, ErrorResponse{ | ||
255 | Code: "NotImplemented", | ||
256 | Message: "Truncated response should have continuation token set", | ||
257 | } | ||
258 | } | ||
259 | |||
260 | for i, obj := range listBucketResult.Contents { | ||
261 | listBucketResult.Contents[i].Key, err = decodeS3Name(obj.Key, listBucketResult.EncodingType) | ||
262 | if err != nil { | ||
263 | return listBucketResult, err | ||
264 | } | ||
265 | listBucketResult.Contents[i].LastModified = listBucketResult.Contents[i].LastModified.Truncate(time.Millisecond) | ||
266 | } | ||
267 | |||
268 | for i, obj := range listBucketResult.CommonPrefixes { | ||
269 | listBucketResult.CommonPrefixes[i].Prefix, err = decodeS3Name(obj.Prefix, listBucketResult.EncodingType) | ||
270 | if err != nil { | ||
271 | return listBucketResult, err | ||
272 | } | ||
273 | } | ||
274 | |||
275 | // Success. | ||
276 | return listBucketResult, nil | ||
277 | } | ||
278 | |||
279 | func (c *Client) listObjects(ctx context.Context, bucketName string, opts ListObjectsOptions) <-chan ObjectInfo { | ||
280 | // Allocate new list objects channel. | ||
281 | objectStatCh := make(chan ObjectInfo, 1) | ||
282 | // Default listing is delimited at "/" | ||
283 | delimiter := "/" | ||
284 | if opts.Recursive { | ||
285 | // If recursive we do not delimit. | ||
286 | delimiter = "" | ||
287 | } | ||
288 | |||
289 | sendObjectInfo := func(info ObjectInfo) { | ||
290 | select { | ||
291 | case objectStatCh <- info: | ||
292 | case <-ctx.Done(): | ||
293 | } | ||
294 | } | ||
295 | |||
296 | // Validate bucket name. | ||
297 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
298 | defer close(objectStatCh) | ||
299 | sendObjectInfo(ObjectInfo{ | ||
300 | Err: err, | ||
301 | }) | ||
302 | return objectStatCh | ||
303 | } | ||
304 | // Validate incoming object prefix. | ||
305 | if err := s3utils.CheckValidObjectNamePrefix(opts.Prefix); err != nil { | ||
306 | defer close(objectStatCh) | ||
307 | sendObjectInfo(ObjectInfo{ | ||
308 | Err: err, | ||
309 | }) | ||
310 | return objectStatCh | ||
311 | } | ||
312 | |||
313 | // Initiate list objects goroutine here. | ||
314 | go func(objectStatCh chan<- ObjectInfo) { | ||
315 | defer func() { | ||
316 | if contextCanceled(ctx) { | ||
317 | objectStatCh <- ObjectInfo{ | ||
318 | Err: ctx.Err(), | ||
319 | } | ||
320 | } | ||
321 | close(objectStatCh) | ||
322 | }() | ||
323 | |||
324 | marker := opts.StartAfter | ||
325 | for { | ||
326 | // Get list of objects a maximum of 1000 per request. | ||
327 | result, err := c.listObjectsQuery(ctx, bucketName, opts.Prefix, marker, delimiter, opts.MaxKeys, opts.headers) | ||
328 | if err != nil { | ||
329 | sendObjectInfo(ObjectInfo{ | ||
330 | Err: err, | ||
331 | }) | ||
332 | return | ||
333 | } | ||
334 | |||
335 | // If contents are available loop through and send over channel. | ||
336 | for _, object := range result.Contents { | ||
337 | // Save the marker. | ||
338 | marker = object.Key | ||
339 | object.ETag = trimEtag(object.ETag) | ||
340 | select { | ||
341 | // Send object content. | ||
342 | case objectStatCh <- object: | ||
343 | // If receives done from the caller, return here. | ||
344 | case <-ctx.Done(): | ||
345 | return | ||
346 | } | ||
347 | } | ||
348 | |||
349 | // Send all common prefixes if any. | ||
350 | // NOTE: prefixes are only present if the request is delimited. | ||
351 | for _, obj := range result.CommonPrefixes { | ||
352 | select { | ||
353 | // Send object prefixes. | ||
354 | case objectStatCh <- ObjectInfo{Key: obj.Prefix}: | ||
355 | // If receives done from the caller, return here. | ||
356 | case <-ctx.Done(): | ||
357 | return | ||
358 | } | ||
359 | } | ||
360 | |||
361 | // If next marker present, save it for next request. | ||
362 | if result.NextMarker != "" { | ||
363 | marker = result.NextMarker | ||
364 | } | ||
365 | |||
366 | // Listing ends result is not truncated, return right here. | ||
367 | if !result.IsTruncated { | ||
368 | return | ||
369 | } | ||
370 | } | ||
371 | }(objectStatCh) | ||
372 | return objectStatCh | ||
373 | } | ||
374 | |||
375 | func (c *Client) listObjectVersions(ctx context.Context, bucketName string, opts ListObjectsOptions) <-chan ObjectInfo { | ||
376 | // Allocate new list objects channel. | ||
377 | resultCh := make(chan ObjectInfo, 1) | ||
378 | // Default listing is delimited at "/" | ||
379 | delimiter := "/" | ||
380 | if opts.Recursive { | ||
381 | // If recursive we do not delimit. | ||
382 | delimiter = "" | ||
383 | } | ||
384 | |||
385 | sendObjectInfo := func(info ObjectInfo) { | ||
386 | select { | ||
387 | case resultCh <- info: | ||
388 | case <-ctx.Done(): | ||
389 | } | ||
390 | } | ||
391 | |||
392 | // Validate bucket name. | ||
393 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
394 | defer close(resultCh) | ||
395 | sendObjectInfo(ObjectInfo{ | ||
396 | Err: err, | ||
397 | }) | ||
398 | return resultCh | ||
399 | } | ||
400 | |||
401 | // Validate incoming object prefix. | ||
402 | if err := s3utils.CheckValidObjectNamePrefix(opts.Prefix); err != nil { | ||
403 | defer close(resultCh) | ||
404 | sendObjectInfo(ObjectInfo{ | ||
405 | Err: err, | ||
406 | }) | ||
407 | return resultCh | ||
408 | } | ||
409 | |||
410 | // Initiate list objects goroutine here. | ||
411 | go func(resultCh chan<- ObjectInfo) { | ||
412 | defer func() { | ||
413 | if contextCanceled(ctx) { | ||
414 | resultCh <- ObjectInfo{ | ||
415 | Err: ctx.Err(), | ||
416 | } | ||
417 | } | ||
418 | close(resultCh) | ||
419 | }() | ||
420 | |||
421 | var ( | ||
422 | keyMarker = "" | ||
423 | versionIDMarker = "" | ||
424 | ) | ||
425 | |||
426 | for { | ||
427 | // Get list of objects a maximum of 1000 per request. | ||
428 | result, err := c.listObjectVersionsQuery(ctx, bucketName, opts, keyMarker, versionIDMarker, delimiter) | ||
429 | if err != nil { | ||
430 | sendObjectInfo(ObjectInfo{ | ||
431 | Err: err, | ||
432 | }) | ||
433 | return | ||
434 | } | ||
435 | |||
436 | // If contents are available loop through and send over channel. | ||
437 | for _, version := range result.Versions { | ||
438 | info := ObjectInfo{ | ||
439 | ETag: trimEtag(version.ETag), | ||
440 | Key: version.Key, | ||
441 | LastModified: version.LastModified.Truncate(time.Millisecond), | ||
442 | Size: version.Size, | ||
443 | Owner: version.Owner, | ||
444 | StorageClass: version.StorageClass, | ||
445 | IsLatest: version.IsLatest, | ||
446 | VersionID: version.VersionID, | ||
447 | IsDeleteMarker: version.isDeleteMarker, | ||
448 | UserTags: version.UserTags, | ||
449 | UserMetadata: version.UserMetadata, | ||
450 | Internal: version.Internal, | ||
451 | } | ||
452 | select { | ||
453 | // Send object version info. | ||
454 | case resultCh <- info: | ||
455 | // If receives done from the caller, return here. | ||
456 | case <-ctx.Done(): | ||
457 | return | ||
458 | } | ||
459 | } | ||
460 | |||
461 | // Send all common prefixes if any. | ||
462 | // NOTE: prefixes are only present if the request is delimited. | ||
463 | for _, obj := range result.CommonPrefixes { | ||
464 | select { | ||
465 | // Send object prefixes. | ||
466 | case resultCh <- ObjectInfo{Key: obj.Prefix}: | ||
467 | // If receives done from the caller, return here. | ||
468 | case <-ctx.Done(): | ||
469 | return | ||
470 | } | ||
471 | } | ||
472 | |||
473 | // If next key marker is present, save it for next request. | ||
474 | if result.NextKeyMarker != "" { | ||
475 | keyMarker = result.NextKeyMarker | ||
476 | } | ||
477 | |||
478 | // If next version id marker is present, save it for next request. | ||
479 | if result.NextVersionIDMarker != "" { | ||
480 | versionIDMarker = result.NextVersionIDMarker | ||
481 | } | ||
482 | |||
483 | // Listing ends result is not truncated, return right here. | ||
484 | if !result.IsTruncated { | ||
485 | return | ||
486 | } | ||
487 | } | ||
488 | }(resultCh) | ||
489 | return resultCh | ||
490 | } | ||
491 | |||
492 | // listObjectVersions - (List Object Versions) - List some or all (up to 1000) of the existing objects | ||
493 | // and their versions in a bucket. | ||
494 | // | ||
495 | // You can use the request parameters as selection criteria to return a subset of the objects in a bucket. | ||
496 | // request parameters :- | ||
497 | // --------- | ||
498 | // ?key-marker - Specifies the key to start with when listing objects in a bucket. | ||
499 | // ?version-id-marker - Specifies the version id marker to start with when listing objects with versions in a bucket. | ||
500 | // ?delimiter - A delimiter is a character you use to group keys. | ||
501 | // ?prefix - Limits the response to keys that begin with the specified prefix. | ||
502 | // ?max-keys - Sets the maximum number of keys returned in the response body. | ||
503 | func (c *Client) listObjectVersionsQuery(ctx context.Context, bucketName string, opts ListObjectsOptions, keyMarker, versionIDMarker, delimiter string) (ListVersionsResult, error) { | ||
504 | // Validate bucket name. | ||
505 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
506 | return ListVersionsResult{}, err | ||
507 | } | ||
508 | // Validate object prefix. | ||
509 | if err := s3utils.CheckValidObjectNamePrefix(opts.Prefix); err != nil { | ||
510 | return ListVersionsResult{}, err | ||
511 | } | ||
512 | // Get resources properly escaped and lined up before | ||
513 | // using them in http request. | ||
514 | urlValues := make(url.Values) | ||
515 | |||
516 | // Set versions to trigger versioning API | ||
517 | urlValues.Set("versions", "") | ||
518 | |||
519 | // Set object prefix, prefix value to be set to empty is okay. | ||
520 | urlValues.Set("prefix", opts.Prefix) | ||
521 | |||
522 | // Set delimiter, delimiter value to be set to empty is okay. | ||
523 | urlValues.Set("delimiter", delimiter) | ||
524 | |||
525 | // Set object marker. | ||
526 | if keyMarker != "" { | ||
527 | urlValues.Set("key-marker", keyMarker) | ||
528 | } | ||
529 | |||
530 | // Set max keys. | ||
531 | if opts.MaxKeys > 0 { | ||
532 | urlValues.Set("max-keys", fmt.Sprintf("%d", opts.MaxKeys)) | ||
533 | } | ||
534 | |||
535 | // Set version ID marker | ||
536 | if versionIDMarker != "" { | ||
537 | urlValues.Set("version-id-marker", versionIDMarker) | ||
538 | } | ||
539 | |||
540 | if opts.WithMetadata { | ||
541 | urlValues.Set("metadata", "true") | ||
542 | } | ||
543 | |||
544 | // Always set encoding-type | ||
545 | urlValues.Set("encoding-type", "url") | ||
546 | |||
547 | // Execute GET on bucket to list objects. | ||
548 | resp, err := c.executeMethod(ctx, http.MethodGet, requestMetadata{ | ||
549 | bucketName: bucketName, | ||
550 | queryValues: urlValues, | ||
551 | contentSHA256Hex: emptySHA256Hex, | ||
552 | customHeader: opts.headers, | ||
553 | }) | ||
554 | defer closeResponse(resp) | ||
555 | if err != nil { | ||
556 | return ListVersionsResult{}, err | ||
557 | } | ||
558 | if resp != nil { | ||
559 | if resp.StatusCode != http.StatusOK { | ||
560 | return ListVersionsResult{}, httpRespToErrorResponse(resp, bucketName, "") | ||
561 | } | ||
562 | } | ||
563 | |||
564 | // Decode ListVersionsResult XML. | ||
565 | listObjectVersionsOutput := ListVersionsResult{} | ||
566 | err = xmlDecoder(resp.Body, &listObjectVersionsOutput) | ||
567 | if err != nil { | ||
568 | return ListVersionsResult{}, err | ||
569 | } | ||
570 | |||
571 | for i, obj := range listObjectVersionsOutput.Versions { | ||
572 | listObjectVersionsOutput.Versions[i].Key, err = decodeS3Name(obj.Key, listObjectVersionsOutput.EncodingType) | ||
573 | if err != nil { | ||
574 | return listObjectVersionsOutput, err | ||
575 | } | ||
576 | } | ||
577 | |||
578 | for i, obj := range listObjectVersionsOutput.CommonPrefixes { | ||
579 | listObjectVersionsOutput.CommonPrefixes[i].Prefix, err = decodeS3Name(obj.Prefix, listObjectVersionsOutput.EncodingType) | ||
580 | if err != nil { | ||
581 | return listObjectVersionsOutput, err | ||
582 | } | ||
583 | } | ||
584 | |||
585 | if listObjectVersionsOutput.NextKeyMarker != "" { | ||
586 | listObjectVersionsOutput.NextKeyMarker, err = decodeS3Name(listObjectVersionsOutput.NextKeyMarker, listObjectVersionsOutput.EncodingType) | ||
587 | if err != nil { | ||
588 | return listObjectVersionsOutput, err | ||
589 | } | ||
590 | } | ||
591 | |||
592 | return listObjectVersionsOutput, nil | ||
593 | } | ||
594 | |||
595 | // listObjects - (List Objects) - List some or all (up to 1000) of the objects in a bucket. | ||
596 | // | ||
597 | // You can use the request parameters as selection criteria to return a subset of the objects in a bucket. | ||
598 | // request parameters :- | ||
599 | // --------- | ||
600 | // ?marker - Specifies the key to start with when listing objects in a bucket. | ||
601 | // ?delimiter - A delimiter is a character you use to group keys. | ||
602 | // ?prefix - Limits the response to keys that begin with the specified prefix. | ||
603 | // ?max-keys - Sets the maximum number of keys returned in the response body. | ||
604 | func (c *Client) listObjectsQuery(ctx context.Context, bucketName, objectPrefix, objectMarker, delimiter string, maxkeys int, headers http.Header) (ListBucketResult, error) { | ||
605 | // Validate bucket name. | ||
606 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
607 | return ListBucketResult{}, err | ||
608 | } | ||
609 | // Validate object prefix. | ||
610 | if err := s3utils.CheckValidObjectNamePrefix(objectPrefix); err != nil { | ||
611 | return ListBucketResult{}, err | ||
612 | } | ||
613 | // Get resources properly escaped and lined up before | ||
614 | // using them in http request. | ||
615 | urlValues := make(url.Values) | ||
616 | |||
617 | // Set object prefix, prefix value to be set to empty is okay. | ||
618 | urlValues.Set("prefix", objectPrefix) | ||
619 | |||
620 | // Set delimiter, delimiter value to be set to empty is okay. | ||
621 | urlValues.Set("delimiter", delimiter) | ||
622 | |||
623 | // Set object marker. | ||
624 | if objectMarker != "" { | ||
625 | urlValues.Set("marker", objectMarker) | ||
626 | } | ||
627 | |||
628 | // Set max keys. | ||
629 | if maxkeys > 0 { | ||
630 | urlValues.Set("max-keys", fmt.Sprintf("%d", maxkeys)) | ||
631 | } | ||
632 | |||
633 | // Always set encoding-type | ||
634 | urlValues.Set("encoding-type", "url") | ||
635 | |||
636 | // Execute GET on bucket to list objects. | ||
637 | resp, err := c.executeMethod(ctx, http.MethodGet, requestMetadata{ | ||
638 | bucketName: bucketName, | ||
639 | queryValues: urlValues, | ||
640 | contentSHA256Hex: emptySHA256Hex, | ||
641 | customHeader: headers, | ||
642 | }) | ||
643 | defer closeResponse(resp) | ||
644 | if err != nil { | ||
645 | return ListBucketResult{}, err | ||
646 | } | ||
647 | if resp != nil { | ||
648 | if resp.StatusCode != http.StatusOK { | ||
649 | return ListBucketResult{}, httpRespToErrorResponse(resp, bucketName, "") | ||
650 | } | ||
651 | } | ||
652 | // Decode listBuckets XML. | ||
653 | listBucketResult := ListBucketResult{} | ||
654 | err = xmlDecoder(resp.Body, &listBucketResult) | ||
655 | if err != nil { | ||
656 | return listBucketResult, err | ||
657 | } | ||
658 | |||
659 | for i, obj := range listBucketResult.Contents { | ||
660 | listBucketResult.Contents[i].Key, err = decodeS3Name(obj.Key, listBucketResult.EncodingType) | ||
661 | if err != nil { | ||
662 | return listBucketResult, err | ||
663 | } | ||
664 | listBucketResult.Contents[i].LastModified = listBucketResult.Contents[i].LastModified.Truncate(time.Millisecond) | ||
665 | } | ||
666 | |||
667 | for i, obj := range listBucketResult.CommonPrefixes { | ||
668 | listBucketResult.CommonPrefixes[i].Prefix, err = decodeS3Name(obj.Prefix, listBucketResult.EncodingType) | ||
669 | if err != nil { | ||
670 | return listBucketResult, err | ||
671 | } | ||
672 | } | ||
673 | |||
674 | if listBucketResult.NextMarker != "" { | ||
675 | listBucketResult.NextMarker, err = decodeS3Name(listBucketResult.NextMarker, listBucketResult.EncodingType) | ||
676 | if err != nil { | ||
677 | return listBucketResult, err | ||
678 | } | ||
679 | } | ||
680 | |||
681 | return listBucketResult, nil | ||
682 | } | ||
683 | |||
684 | // ListObjectsOptions holds all options of a list object request | ||
685 | type ListObjectsOptions struct { | ||
686 | // Include objects versions in the listing | ||
687 | WithVersions bool | ||
688 | // Include objects metadata in the listing | ||
689 | WithMetadata bool | ||
690 | // Only list objects with the prefix | ||
691 | Prefix string | ||
692 | // Ignore '/' delimiter | ||
693 | Recursive bool | ||
694 | // The maximum number of objects requested per | ||
695 | // batch, advanced use-case not useful for most | ||
696 | // applications | ||
697 | MaxKeys int | ||
698 | // StartAfter start listing lexically at this | ||
699 | // object onwards, this value can also be set | ||
700 | // for Marker when `UseV1` is set to true. | ||
701 | StartAfter string | ||
702 | |||
703 | // Use the deprecated list objects V1 API | ||
704 | UseV1 bool | ||
705 | |||
706 | headers http.Header | ||
707 | } | ||
708 | |||
709 | // Set adds a key value pair to the options. The | ||
710 | // key-value pair will be part of the HTTP GET request | ||
711 | // headers. | ||
712 | func (o *ListObjectsOptions) Set(key, value string) { | ||
713 | if o.headers == nil { | ||
714 | o.headers = make(http.Header) | ||
715 | } | ||
716 | o.headers.Set(key, value) | ||
717 | } | ||
718 | |||
719 | // ListObjects returns objects list after evaluating the passed options. | ||
720 | // | ||
721 | // api := client.New(....) | ||
722 | // for object := range api.ListObjects(ctx, "mytestbucket", minio.ListObjectsOptions{Prefix: "starthere", Recursive:true}) { | ||
723 | // fmt.Println(object) | ||
724 | // } | ||
725 | // | ||
726 | // If caller cancels the context, then the last entry on the 'chan ObjectInfo' will be the context.Error() | ||
727 | // caller must drain the channel entirely and wait until channel is closed before proceeding, without | ||
728 | // waiting on the channel to be closed completely you might leak goroutines. | ||
729 | func (c *Client) ListObjects(ctx context.Context, bucketName string, opts ListObjectsOptions) <-chan ObjectInfo { | ||
730 | if opts.WithVersions { | ||
731 | return c.listObjectVersions(ctx, bucketName, opts) | ||
732 | } | ||
733 | |||
734 | // Use legacy list objects v1 API | ||
735 | if opts.UseV1 { | ||
736 | return c.listObjects(ctx, bucketName, opts) | ||
737 | } | ||
738 | |||
739 | // Check whether this is snowball region, if yes ListObjectsV2 doesn't work, fallback to listObjectsV1. | ||
740 | if location, ok := c.bucketLocCache.Get(bucketName); ok { | ||
741 | if location == "snowball" { | ||
742 | return c.listObjects(ctx, bucketName, opts) | ||
743 | } | ||
744 | } | ||
745 | |||
746 | return c.listObjectsV2(ctx, bucketName, opts) | ||
747 | } | ||
748 | |||
749 | // ListIncompleteUploads - List incompletely uploaded multipart objects. | ||
750 | // | ||
751 | // ListIncompleteUploads lists all incompleted objects matching the | ||
752 | // objectPrefix from the specified bucket. If recursion is enabled | ||
753 | // it would list all subdirectories and all its contents. | ||
754 | // | ||
755 | // Your input parameters are just bucketName, objectPrefix, recursive. | ||
756 | // If you enable recursive as 'true' this function will return back all | ||
757 | // the multipart objects in a given bucket name. | ||
758 | // | ||
759 | // api := client.New(....) | ||
760 | // // Recurively list all objects in 'mytestbucket' | ||
761 | // recursive := true | ||
762 | // for message := range api.ListIncompleteUploads(context.Background(), "mytestbucket", "starthere", recursive) { | ||
763 | // fmt.Println(message) | ||
764 | // } | ||
765 | func (c *Client) ListIncompleteUploads(ctx context.Context, bucketName, objectPrefix string, recursive bool) <-chan ObjectMultipartInfo { | ||
766 | return c.listIncompleteUploads(ctx, bucketName, objectPrefix, recursive) | ||
767 | } | ||
768 | |||
769 | // contextCanceled returns whether a context is canceled. | ||
770 | func contextCanceled(ctx context.Context) bool { | ||
771 | select { | ||
772 | case <-ctx.Done(): | ||
773 | return true | ||
774 | default: | ||
775 | return false | ||
776 | } | ||
777 | } | ||
778 | |||
779 | // listIncompleteUploads lists all incomplete uploads. | ||
780 | func (c *Client) listIncompleteUploads(ctx context.Context, bucketName, objectPrefix string, recursive bool) <-chan ObjectMultipartInfo { | ||
781 | // Allocate channel for multipart uploads. | ||
782 | objectMultipartStatCh := make(chan ObjectMultipartInfo, 1) | ||
783 | // Delimiter is set to "/" by default. | ||
784 | delimiter := "/" | ||
785 | if recursive { | ||
786 | // If recursive do not delimit. | ||
787 | delimiter = "" | ||
788 | } | ||
789 | // Validate bucket name. | ||
790 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
791 | defer close(objectMultipartStatCh) | ||
792 | objectMultipartStatCh <- ObjectMultipartInfo{ | ||
793 | Err: err, | ||
794 | } | ||
795 | return objectMultipartStatCh | ||
796 | } | ||
797 | // Validate incoming object prefix. | ||
798 | if err := s3utils.CheckValidObjectNamePrefix(objectPrefix); err != nil { | ||
799 | defer close(objectMultipartStatCh) | ||
800 | objectMultipartStatCh <- ObjectMultipartInfo{ | ||
801 | Err: err, | ||
802 | } | ||
803 | return objectMultipartStatCh | ||
804 | } | ||
805 | go func(objectMultipartStatCh chan<- ObjectMultipartInfo) { | ||
806 | defer func() { | ||
807 | if contextCanceled(ctx) { | ||
808 | objectMultipartStatCh <- ObjectMultipartInfo{ | ||
809 | Err: ctx.Err(), | ||
810 | } | ||
811 | } | ||
812 | close(objectMultipartStatCh) | ||
813 | }() | ||
814 | |||
815 | // object and upload ID marker for future requests. | ||
816 | var objectMarker string | ||
817 | var uploadIDMarker string | ||
818 | for { | ||
819 | // list all multipart uploads. | ||
820 | result, err := c.listMultipartUploadsQuery(ctx, bucketName, objectMarker, uploadIDMarker, objectPrefix, delimiter, 0) | ||
821 | if err != nil { | ||
822 | objectMultipartStatCh <- ObjectMultipartInfo{ | ||
823 | Err: err, | ||
824 | } | ||
825 | return | ||
826 | } | ||
827 | objectMarker = result.NextKeyMarker | ||
828 | uploadIDMarker = result.NextUploadIDMarker | ||
829 | |||
830 | // Send all multipart uploads. | ||
831 | for _, obj := range result.Uploads { | ||
832 | // Calculate total size of the uploaded parts if 'aggregateSize' is enabled. | ||
833 | select { | ||
834 | // Send individual uploads here. | ||
835 | case objectMultipartStatCh <- obj: | ||
836 | // If the context is canceled | ||
837 | case <-ctx.Done(): | ||
838 | return | ||
839 | } | ||
840 | } | ||
841 | // Send all common prefixes if any. | ||
842 | // NOTE: prefixes are only present if the request is delimited. | ||
843 | for _, obj := range result.CommonPrefixes { | ||
844 | select { | ||
845 | // Send delimited prefixes here. | ||
846 | case objectMultipartStatCh <- ObjectMultipartInfo{Key: obj.Prefix, Size: 0}: | ||
847 | // If context is canceled. | ||
848 | case <-ctx.Done(): | ||
849 | return | ||
850 | } | ||
851 | } | ||
852 | // Listing ends if result not truncated, return right here. | ||
853 | if !result.IsTruncated { | ||
854 | return | ||
855 | } | ||
856 | } | ||
857 | }(objectMultipartStatCh) | ||
858 | // return. | ||
859 | return objectMultipartStatCh | ||
860 | } | ||
861 | |||
862 | // listMultipartUploadsQuery - (List Multipart Uploads). | ||
863 | // - Lists some or all (up to 1000) in-progress multipart uploads in a bucket. | ||
864 | // | ||
865 | // You can use the request parameters as selection criteria to return a subset of the uploads in a bucket. | ||
866 | // request parameters. :- | ||
867 | // --------- | ||
868 | // ?key-marker - Specifies the multipart upload after which listing should begin. | ||
869 | // ?upload-id-marker - Together with key-marker specifies the multipart upload after which listing should begin. | ||
870 | // ?delimiter - A delimiter is a character you use to group keys. | ||
871 | // ?prefix - Limits the response to keys that begin with the specified prefix. | ||
872 | // ?max-uploads - Sets the maximum number of multipart uploads returned in the response body. | ||
873 | func (c *Client) listMultipartUploadsQuery(ctx context.Context, bucketName, keyMarker, uploadIDMarker, prefix, delimiter string, maxUploads int) (ListMultipartUploadsResult, error) { | ||
874 | // Get resources properly escaped and lined up before using them in http request. | ||
875 | urlValues := make(url.Values) | ||
876 | // Set uploads. | ||
877 | urlValues.Set("uploads", "") | ||
878 | // Set object key marker. | ||
879 | if keyMarker != "" { | ||
880 | urlValues.Set("key-marker", keyMarker) | ||
881 | } | ||
882 | // Set upload id marker. | ||
883 | if uploadIDMarker != "" { | ||
884 | urlValues.Set("upload-id-marker", uploadIDMarker) | ||
885 | } | ||
886 | |||
887 | // Set object prefix, prefix value to be set to empty is okay. | ||
888 | urlValues.Set("prefix", prefix) | ||
889 | |||
890 | // Set delimiter, delimiter value to be set to empty is okay. | ||
891 | urlValues.Set("delimiter", delimiter) | ||
892 | |||
893 | // Always set encoding-type | ||
894 | urlValues.Set("encoding-type", "url") | ||
895 | |||
896 | // maxUploads should be 1000 or less. | ||
897 | if maxUploads > 0 { | ||
898 | // Set max-uploads. | ||
899 | urlValues.Set("max-uploads", fmt.Sprintf("%d", maxUploads)) | ||
900 | } | ||
901 | |||
902 | // Execute GET on bucketName to list multipart uploads. | ||
903 | resp, err := c.executeMethod(ctx, http.MethodGet, requestMetadata{ | ||
904 | bucketName: bucketName, | ||
905 | queryValues: urlValues, | ||
906 | contentSHA256Hex: emptySHA256Hex, | ||
907 | }) | ||
908 | defer closeResponse(resp) | ||
909 | if err != nil { | ||
910 | return ListMultipartUploadsResult{}, err | ||
911 | } | ||
912 | if resp != nil { | ||
913 | if resp.StatusCode != http.StatusOK { | ||
914 | return ListMultipartUploadsResult{}, httpRespToErrorResponse(resp, bucketName, "") | ||
915 | } | ||
916 | } | ||
917 | // Decode response body. | ||
918 | listMultipartUploadsResult := ListMultipartUploadsResult{} | ||
919 | err = xmlDecoder(resp.Body, &listMultipartUploadsResult) | ||
920 | if err != nil { | ||
921 | return listMultipartUploadsResult, err | ||
922 | } | ||
923 | |||
924 | listMultipartUploadsResult.NextKeyMarker, err = decodeS3Name(listMultipartUploadsResult.NextKeyMarker, listMultipartUploadsResult.EncodingType) | ||
925 | if err != nil { | ||
926 | return listMultipartUploadsResult, err | ||
927 | } | ||
928 | |||
929 | listMultipartUploadsResult.NextUploadIDMarker, err = decodeS3Name(listMultipartUploadsResult.NextUploadIDMarker, listMultipartUploadsResult.EncodingType) | ||
930 | if err != nil { | ||
931 | return listMultipartUploadsResult, err | ||
932 | } | ||
933 | |||
934 | for i, obj := range listMultipartUploadsResult.Uploads { | ||
935 | listMultipartUploadsResult.Uploads[i].Key, err = decodeS3Name(obj.Key, listMultipartUploadsResult.EncodingType) | ||
936 | if err != nil { | ||
937 | return listMultipartUploadsResult, err | ||
938 | } | ||
939 | } | ||
940 | |||
941 | for i, obj := range listMultipartUploadsResult.CommonPrefixes { | ||
942 | listMultipartUploadsResult.CommonPrefixes[i].Prefix, err = decodeS3Name(obj.Prefix, listMultipartUploadsResult.EncodingType) | ||
943 | if err != nil { | ||
944 | return listMultipartUploadsResult, err | ||
945 | } | ||
946 | } | ||
947 | |||
948 | return listMultipartUploadsResult, nil | ||
949 | } | ||
950 | |||
951 | // listObjectParts list all object parts recursively. | ||
952 | // | ||
953 | //lint:ignore U1000 Keep this around | ||
954 | func (c *Client) listObjectParts(ctx context.Context, bucketName, objectName, uploadID string) (partsInfo map[int]ObjectPart, err error) { | ||
955 | // Part number marker for the next batch of request. | ||
956 | var nextPartNumberMarker int | ||
957 | partsInfo = make(map[int]ObjectPart) | ||
958 | for { | ||
959 | // Get list of uploaded parts a maximum of 1000 per request. | ||
960 | listObjPartsResult, err := c.listObjectPartsQuery(ctx, bucketName, objectName, uploadID, nextPartNumberMarker, 1000) | ||
961 | if err != nil { | ||
962 | return nil, err | ||
963 | } | ||
964 | // Append to parts info. | ||
965 | for _, part := range listObjPartsResult.ObjectParts { | ||
966 | // Trim off the odd double quotes from ETag in the beginning and end. | ||
967 | part.ETag = trimEtag(part.ETag) | ||
968 | partsInfo[part.PartNumber] = part | ||
969 | } | ||
970 | // Keep part number marker, for the next iteration. | ||
971 | nextPartNumberMarker = listObjPartsResult.NextPartNumberMarker | ||
972 | // Listing ends result is not truncated, return right here. | ||
973 | if !listObjPartsResult.IsTruncated { | ||
974 | break | ||
975 | } | ||
976 | } | ||
977 | |||
978 | // Return all the parts. | ||
979 | return partsInfo, nil | ||
980 | } | ||
981 | |||
982 | // findUploadIDs lists all incomplete uploads and find the uploadIDs of the matching object name. | ||
983 | func (c *Client) findUploadIDs(ctx context.Context, bucketName, objectName string) ([]string, error) { | ||
984 | var uploadIDs []string | ||
985 | // Make list incomplete uploads recursive. | ||
986 | isRecursive := true | ||
987 | // List all incomplete uploads. | ||
988 | for mpUpload := range c.listIncompleteUploads(ctx, bucketName, objectName, isRecursive) { | ||
989 | if mpUpload.Err != nil { | ||
990 | return nil, mpUpload.Err | ||
991 | } | ||
992 | if objectName == mpUpload.Key { | ||
993 | uploadIDs = append(uploadIDs, mpUpload.UploadID) | ||
994 | } | ||
995 | } | ||
996 | // Return the latest upload id. | ||
997 | return uploadIDs, nil | ||
998 | } | ||
999 | |||
1000 | // listObjectPartsQuery (List Parts query) | ||
1001 | // - lists some or all (up to 1000) parts that have been uploaded | ||
1002 | // for a specific multipart upload | ||
1003 | // | ||
1004 | // You can use the request parameters as selection criteria to return | ||
1005 | // a subset of the uploads in a bucket, request parameters :- | ||
1006 | // --------- | ||
1007 | // ?part-number-marker - Specifies the part after which listing should | ||
1008 | // begin. | ||
1009 | // ?max-parts - Maximum parts to be listed per request. | ||
1010 | func (c *Client) listObjectPartsQuery(ctx context.Context, bucketName, objectName, uploadID string, partNumberMarker, maxParts int) (ListObjectPartsResult, error) { | ||
1011 | // Get resources properly escaped and lined up before using them in http request. | ||
1012 | urlValues := make(url.Values) | ||
1013 | // Set part number marker. | ||
1014 | urlValues.Set("part-number-marker", fmt.Sprintf("%d", partNumberMarker)) | ||
1015 | // Set upload id. | ||
1016 | urlValues.Set("uploadId", uploadID) | ||
1017 | |||
1018 | // maxParts should be 1000 or less. | ||
1019 | if maxParts > 0 { | ||
1020 | // Set max parts. | ||
1021 | urlValues.Set("max-parts", fmt.Sprintf("%d", maxParts)) | ||
1022 | } | ||
1023 | |||
1024 | // Execute GET on objectName to get list of parts. | ||
1025 | resp, err := c.executeMethod(ctx, http.MethodGet, requestMetadata{ | ||
1026 | bucketName: bucketName, | ||
1027 | objectName: objectName, | ||
1028 | queryValues: urlValues, | ||
1029 | contentSHA256Hex: emptySHA256Hex, | ||
1030 | }) | ||
1031 | defer closeResponse(resp) | ||
1032 | if err != nil { | ||
1033 | return ListObjectPartsResult{}, err | ||
1034 | } | ||
1035 | if resp != nil { | ||
1036 | if resp.StatusCode != http.StatusOK { | ||
1037 | return ListObjectPartsResult{}, httpRespToErrorResponse(resp, bucketName, objectName) | ||
1038 | } | ||
1039 | } | ||
1040 | // Decode list object parts XML. | ||
1041 | listObjectPartsResult := ListObjectPartsResult{} | ||
1042 | err = xmlDecoder(resp.Body, &listObjectPartsResult) | ||
1043 | if err != nil { | ||
1044 | return listObjectPartsResult, err | ||
1045 | } | ||
1046 | return listObjectPartsResult, nil | ||
1047 | } | ||
1048 | |||
1049 | // Decode an S3 object name according to the encoding type | ||
1050 | func decodeS3Name(name, encodingType string) (string, error) { | ||
1051 | switch encodingType { | ||
1052 | case "url": | ||
1053 | return url.QueryUnescape(name) | ||
1054 | default: | ||
1055 | return name, nil | ||
1056 | } | ||
1057 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/api-object-legal-hold.go b/vendor/github.com/minio/minio-go/v7/api-object-legal-hold.go deleted file mode 100644 index 0c027d5..0000000 --- a/vendor/github.com/minio/minio-go/v7/api-object-legal-hold.go +++ /dev/null | |||
@@ -1,176 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2020 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package minio | ||
19 | |||
20 | import ( | ||
21 | "bytes" | ||
22 | "context" | ||
23 | "encoding/xml" | ||
24 | "fmt" | ||
25 | "net/http" | ||
26 | "net/url" | ||
27 | |||
28 | "github.com/minio/minio-go/v7/pkg/s3utils" | ||
29 | ) | ||
30 | |||
31 | // objectLegalHold - object legal hold specified in | ||
32 | // https://docs.aws.amazon.com/AmazonS3/latest/API/archive-RESTObjectPUTLegalHold.html | ||
33 | type objectLegalHold struct { | ||
34 | XMLNS string `xml:"xmlns,attr,omitempty"` | ||
35 | XMLName xml.Name `xml:"LegalHold"` | ||
36 | Status LegalHoldStatus `xml:"Status,omitempty"` | ||
37 | } | ||
38 | |||
39 | // PutObjectLegalHoldOptions represents options specified by user for PutObjectLegalHold call | ||
40 | type PutObjectLegalHoldOptions struct { | ||
41 | VersionID string | ||
42 | Status *LegalHoldStatus | ||
43 | } | ||
44 | |||
45 | // GetObjectLegalHoldOptions represents options specified by user for GetObjectLegalHold call | ||
46 | type GetObjectLegalHoldOptions struct { | ||
47 | VersionID string | ||
48 | } | ||
49 | |||
50 | // LegalHoldStatus - object legal hold status. | ||
51 | type LegalHoldStatus string | ||
52 | |||
53 | const ( | ||
54 | // LegalHoldEnabled indicates legal hold is enabled | ||
55 | LegalHoldEnabled LegalHoldStatus = "ON" | ||
56 | |||
57 | // LegalHoldDisabled indicates legal hold is disabled | ||
58 | LegalHoldDisabled LegalHoldStatus = "OFF" | ||
59 | ) | ||
60 | |||
61 | func (r LegalHoldStatus) String() string { | ||
62 | return string(r) | ||
63 | } | ||
64 | |||
65 | // IsValid - check whether this legal hold status is valid or not. | ||
66 | func (r LegalHoldStatus) IsValid() bool { | ||
67 | return r == LegalHoldEnabled || r == LegalHoldDisabled | ||
68 | } | ||
69 | |||
70 | func newObjectLegalHold(status *LegalHoldStatus) (*objectLegalHold, error) { | ||
71 | if status == nil { | ||
72 | return nil, fmt.Errorf("Status not set") | ||
73 | } | ||
74 | if !status.IsValid() { | ||
75 | return nil, fmt.Errorf("invalid legal hold status `%v`", status) | ||
76 | } | ||
77 | legalHold := &objectLegalHold{ | ||
78 | Status: *status, | ||
79 | } | ||
80 | return legalHold, nil | ||
81 | } | ||
82 | |||
83 | // PutObjectLegalHold : sets object legal hold for a given object and versionID. | ||
84 | func (c *Client) PutObjectLegalHold(ctx context.Context, bucketName, objectName string, opts PutObjectLegalHoldOptions) error { | ||
85 | // Input validation. | ||
86 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
87 | return err | ||
88 | } | ||
89 | |||
90 | if err := s3utils.CheckValidObjectName(objectName); err != nil { | ||
91 | return err | ||
92 | } | ||
93 | |||
94 | // Get resources properly escaped and lined up before | ||
95 | // using them in http request. | ||
96 | urlValues := make(url.Values) | ||
97 | urlValues.Set("legal-hold", "") | ||
98 | |||
99 | if opts.VersionID != "" { | ||
100 | urlValues.Set("versionId", opts.VersionID) | ||
101 | } | ||
102 | |||
103 | lh, err := newObjectLegalHold(opts.Status) | ||
104 | if err != nil { | ||
105 | return err | ||
106 | } | ||
107 | |||
108 | lhData, err := xml.Marshal(lh) | ||
109 | if err != nil { | ||
110 | return err | ||
111 | } | ||
112 | |||
113 | reqMetadata := requestMetadata{ | ||
114 | bucketName: bucketName, | ||
115 | objectName: objectName, | ||
116 | queryValues: urlValues, | ||
117 | contentBody: bytes.NewReader(lhData), | ||
118 | contentLength: int64(len(lhData)), | ||
119 | contentMD5Base64: sumMD5Base64(lhData), | ||
120 | contentSHA256Hex: sum256Hex(lhData), | ||
121 | } | ||
122 | |||
123 | // Execute PUT Object Legal Hold. | ||
124 | resp, err := c.executeMethod(ctx, http.MethodPut, reqMetadata) | ||
125 | defer closeResponse(resp) | ||
126 | if err != nil { | ||
127 | return err | ||
128 | } | ||
129 | if resp != nil { | ||
130 | if resp.StatusCode != http.StatusOK && resp.StatusCode != http.StatusNoContent { | ||
131 | return httpRespToErrorResponse(resp, bucketName, objectName) | ||
132 | } | ||
133 | } | ||
134 | return nil | ||
135 | } | ||
136 | |||
137 | // GetObjectLegalHold gets legal-hold status of given object. | ||
138 | func (c *Client) GetObjectLegalHold(ctx context.Context, bucketName, objectName string, opts GetObjectLegalHoldOptions) (status *LegalHoldStatus, err error) { | ||
139 | // Input validation. | ||
140 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
141 | return nil, err | ||
142 | } | ||
143 | |||
144 | if err := s3utils.CheckValidObjectName(objectName); err != nil { | ||
145 | return nil, err | ||
146 | } | ||
147 | urlValues := make(url.Values) | ||
148 | urlValues.Set("legal-hold", "") | ||
149 | |||
150 | if opts.VersionID != "" { | ||
151 | urlValues.Set("versionId", opts.VersionID) | ||
152 | } | ||
153 | |||
154 | // Execute GET on bucket to list objects. | ||
155 | resp, err := c.executeMethod(ctx, http.MethodGet, requestMetadata{ | ||
156 | bucketName: bucketName, | ||
157 | objectName: objectName, | ||
158 | queryValues: urlValues, | ||
159 | contentSHA256Hex: emptySHA256Hex, | ||
160 | }) | ||
161 | defer closeResponse(resp) | ||
162 | if err != nil { | ||
163 | return nil, err | ||
164 | } | ||
165 | if resp != nil { | ||
166 | if resp.StatusCode != http.StatusOK { | ||
167 | return nil, httpRespToErrorResponse(resp, bucketName, objectName) | ||
168 | } | ||
169 | } | ||
170 | lh := &objectLegalHold{} | ||
171 | if err = xml.NewDecoder(resp.Body).Decode(lh); err != nil { | ||
172 | return nil, err | ||
173 | } | ||
174 | |||
175 | return &lh.Status, nil | ||
176 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/api-object-lock.go b/vendor/github.com/minio/minio-go/v7/api-object-lock.go deleted file mode 100644 index f0a4398..0000000 --- a/vendor/github.com/minio/minio-go/v7/api-object-lock.go +++ /dev/null | |||
@@ -1,241 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2019 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package minio | ||
19 | |||
20 | import ( | ||
21 | "bytes" | ||
22 | "context" | ||
23 | "encoding/xml" | ||
24 | "fmt" | ||
25 | "net/http" | ||
26 | "net/url" | ||
27 | "time" | ||
28 | |||
29 | "github.com/minio/minio-go/v7/pkg/s3utils" | ||
30 | ) | ||
31 | |||
32 | // RetentionMode - object retention mode. | ||
33 | type RetentionMode string | ||
34 | |||
35 | const ( | ||
36 | // Governance - governance mode. | ||
37 | Governance RetentionMode = "GOVERNANCE" | ||
38 | |||
39 | // Compliance - compliance mode. | ||
40 | Compliance RetentionMode = "COMPLIANCE" | ||
41 | ) | ||
42 | |||
43 | func (r RetentionMode) String() string { | ||
44 | return string(r) | ||
45 | } | ||
46 | |||
47 | // IsValid - check whether this retention mode is valid or not. | ||
48 | func (r RetentionMode) IsValid() bool { | ||
49 | return r == Governance || r == Compliance | ||
50 | } | ||
51 | |||
52 | // ValidityUnit - retention validity unit. | ||
53 | type ValidityUnit string | ||
54 | |||
55 | const ( | ||
56 | // Days - denotes no. of days. | ||
57 | Days ValidityUnit = "DAYS" | ||
58 | |||
59 | // Years - denotes no. of years. | ||
60 | Years ValidityUnit = "YEARS" | ||
61 | ) | ||
62 | |||
63 | func (unit ValidityUnit) String() string { | ||
64 | return string(unit) | ||
65 | } | ||
66 | |||
67 | // IsValid - check whether this validity unit is valid or not. | ||
68 | func (unit ValidityUnit) isValid() bool { | ||
69 | return unit == Days || unit == Years | ||
70 | } | ||
71 | |||
72 | // Retention - bucket level retention configuration. | ||
73 | type Retention struct { | ||
74 | Mode RetentionMode | ||
75 | Validity time.Duration | ||
76 | } | ||
77 | |||
78 | func (r Retention) String() string { | ||
79 | return fmt.Sprintf("{Mode:%v, Validity:%v}", r.Mode, r.Validity) | ||
80 | } | ||
81 | |||
82 | // IsEmpty - returns whether retention is empty or not. | ||
83 | func (r Retention) IsEmpty() bool { | ||
84 | return r.Mode == "" || r.Validity == 0 | ||
85 | } | ||
86 | |||
87 | // objectLockConfig - object lock configuration specified in | ||
88 | // https://docs.aws.amazon.com/AmazonS3/latest/API/Type_API_ObjectLockConfiguration.html | ||
89 | type objectLockConfig struct { | ||
90 | XMLNS string `xml:"xmlns,attr,omitempty"` | ||
91 | XMLName xml.Name `xml:"ObjectLockConfiguration"` | ||
92 | ObjectLockEnabled string `xml:"ObjectLockEnabled"` | ||
93 | Rule *struct { | ||
94 | DefaultRetention struct { | ||
95 | Mode RetentionMode `xml:"Mode"` | ||
96 | Days *uint `xml:"Days"` | ||
97 | Years *uint `xml:"Years"` | ||
98 | } `xml:"DefaultRetention"` | ||
99 | } `xml:"Rule,omitempty"` | ||
100 | } | ||
101 | |||
102 | func newObjectLockConfig(mode *RetentionMode, validity *uint, unit *ValidityUnit) (*objectLockConfig, error) { | ||
103 | config := &objectLockConfig{ | ||
104 | ObjectLockEnabled: "Enabled", | ||
105 | } | ||
106 | |||
107 | if mode != nil && validity != nil && unit != nil { | ||
108 | if !mode.IsValid() { | ||
109 | return nil, fmt.Errorf("invalid retention mode `%v`", mode) | ||
110 | } | ||
111 | |||
112 | if !unit.isValid() { | ||
113 | return nil, fmt.Errorf("invalid validity unit `%v`", unit) | ||
114 | } | ||
115 | |||
116 | config.Rule = &struct { | ||
117 | DefaultRetention struct { | ||
118 | Mode RetentionMode `xml:"Mode"` | ||
119 | Days *uint `xml:"Days"` | ||
120 | Years *uint `xml:"Years"` | ||
121 | } `xml:"DefaultRetention"` | ||
122 | }{} | ||
123 | |||
124 | config.Rule.DefaultRetention.Mode = *mode | ||
125 | if *unit == Days { | ||
126 | config.Rule.DefaultRetention.Days = validity | ||
127 | } else { | ||
128 | config.Rule.DefaultRetention.Years = validity | ||
129 | } | ||
130 | |||
131 | return config, nil | ||
132 | } | ||
133 | |||
134 | if mode == nil && validity == nil && unit == nil { | ||
135 | return config, nil | ||
136 | } | ||
137 | |||
138 | return nil, fmt.Errorf("all of retention mode, validity and validity unit must be passed") | ||
139 | } | ||
140 | |||
141 | // SetBucketObjectLockConfig sets object lock configuration in given bucket. mode, validity and unit are either all set or all nil. | ||
142 | func (c *Client) SetBucketObjectLockConfig(ctx context.Context, bucketName string, mode *RetentionMode, validity *uint, unit *ValidityUnit) error { | ||
143 | // Input validation. | ||
144 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
145 | return err | ||
146 | } | ||
147 | |||
148 | // Get resources properly escaped and lined up before | ||
149 | // using them in http request. | ||
150 | urlValues := make(url.Values) | ||
151 | urlValues.Set("object-lock", "") | ||
152 | |||
153 | config, err := newObjectLockConfig(mode, validity, unit) | ||
154 | if err != nil { | ||
155 | return err | ||
156 | } | ||
157 | |||
158 | configData, err := xml.Marshal(config) | ||
159 | if err != nil { | ||
160 | return err | ||
161 | } | ||
162 | |||
163 | reqMetadata := requestMetadata{ | ||
164 | bucketName: bucketName, | ||
165 | queryValues: urlValues, | ||
166 | contentBody: bytes.NewReader(configData), | ||
167 | contentLength: int64(len(configData)), | ||
168 | contentMD5Base64: sumMD5Base64(configData), | ||
169 | contentSHA256Hex: sum256Hex(configData), | ||
170 | } | ||
171 | |||
172 | // Execute PUT bucket object lock configuration. | ||
173 | resp, err := c.executeMethod(ctx, http.MethodPut, reqMetadata) | ||
174 | defer closeResponse(resp) | ||
175 | if err != nil { | ||
176 | return err | ||
177 | } | ||
178 | if resp != nil { | ||
179 | if resp.StatusCode != http.StatusOK { | ||
180 | return httpRespToErrorResponse(resp, bucketName, "") | ||
181 | } | ||
182 | } | ||
183 | return nil | ||
184 | } | ||
185 | |||
186 | // GetObjectLockConfig gets object lock configuration of given bucket. | ||
187 | func (c *Client) GetObjectLockConfig(ctx context.Context, bucketName string) (objectLock string, mode *RetentionMode, validity *uint, unit *ValidityUnit, err error) { | ||
188 | // Input validation. | ||
189 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
190 | return "", nil, nil, nil, err | ||
191 | } | ||
192 | |||
193 | urlValues := make(url.Values) | ||
194 | urlValues.Set("object-lock", "") | ||
195 | |||
196 | // Execute GET on bucket to list objects. | ||
197 | resp, err := c.executeMethod(ctx, http.MethodGet, requestMetadata{ | ||
198 | bucketName: bucketName, | ||
199 | queryValues: urlValues, | ||
200 | contentSHA256Hex: emptySHA256Hex, | ||
201 | }) | ||
202 | defer closeResponse(resp) | ||
203 | if err != nil { | ||
204 | return "", nil, nil, nil, err | ||
205 | } | ||
206 | if resp != nil { | ||
207 | if resp.StatusCode != http.StatusOK { | ||
208 | return "", nil, nil, nil, httpRespToErrorResponse(resp, bucketName, "") | ||
209 | } | ||
210 | } | ||
211 | config := &objectLockConfig{} | ||
212 | if err = xml.NewDecoder(resp.Body).Decode(config); err != nil { | ||
213 | return "", nil, nil, nil, err | ||
214 | } | ||
215 | |||
216 | if config.Rule != nil { | ||
217 | mode = &config.Rule.DefaultRetention.Mode | ||
218 | if config.Rule.DefaultRetention.Days != nil { | ||
219 | validity = config.Rule.DefaultRetention.Days | ||
220 | days := Days | ||
221 | unit = &days | ||
222 | } else { | ||
223 | validity = config.Rule.DefaultRetention.Years | ||
224 | years := Years | ||
225 | unit = &years | ||
226 | } | ||
227 | return config.ObjectLockEnabled, mode, validity, unit, nil | ||
228 | } | ||
229 | return config.ObjectLockEnabled, nil, nil, nil, nil | ||
230 | } | ||
231 | |||
232 | // GetBucketObjectLockConfig gets object lock configuration of given bucket. | ||
233 | func (c *Client) GetBucketObjectLockConfig(ctx context.Context, bucketName string) (mode *RetentionMode, validity *uint, unit *ValidityUnit, err error) { | ||
234 | _, mode, validity, unit, err = c.GetObjectLockConfig(ctx, bucketName) | ||
235 | return mode, validity, unit, err | ||
236 | } | ||
237 | |||
238 | // SetObjectLockConfig sets object lock configuration in given bucket. mode, validity and unit are either all set or all nil. | ||
239 | func (c *Client) SetObjectLockConfig(ctx context.Context, bucketName string, mode *RetentionMode, validity *uint, unit *ValidityUnit) error { | ||
240 | return c.SetBucketObjectLockConfig(ctx, bucketName, mode, validity, unit) | ||
241 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/api-object-retention.go b/vendor/github.com/minio/minio-go/v7/api-object-retention.go deleted file mode 100644 index b29cb1f..0000000 --- a/vendor/github.com/minio/minio-go/v7/api-object-retention.go +++ /dev/null | |||
@@ -1,165 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2019-2020 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package minio | ||
19 | |||
20 | import ( | ||
21 | "bytes" | ||
22 | "context" | ||
23 | "encoding/xml" | ||
24 | "fmt" | ||
25 | "net/http" | ||
26 | "net/url" | ||
27 | "time" | ||
28 | |||
29 | "github.com/minio/minio-go/v7/pkg/s3utils" | ||
30 | ) | ||
31 | |||
32 | // objectRetention - object retention specified in | ||
33 | // https://docs.aws.amazon.com/AmazonS3/latest/API/Type_API_ObjectLockConfiguration.html | ||
34 | type objectRetention struct { | ||
35 | XMLNS string `xml:"xmlns,attr,omitempty"` | ||
36 | XMLName xml.Name `xml:"Retention"` | ||
37 | Mode RetentionMode `xml:"Mode,omitempty"` | ||
38 | RetainUntilDate *time.Time `type:"timestamp" timestampFormat:"iso8601" xml:"RetainUntilDate,omitempty"` | ||
39 | } | ||
40 | |||
41 | func newObjectRetention(mode *RetentionMode, date *time.Time) (*objectRetention, error) { | ||
42 | objectRetention := &objectRetention{} | ||
43 | |||
44 | if date != nil && !date.IsZero() { | ||
45 | objectRetention.RetainUntilDate = date | ||
46 | } | ||
47 | if mode != nil { | ||
48 | if !mode.IsValid() { | ||
49 | return nil, fmt.Errorf("invalid retention mode `%v`", mode) | ||
50 | } | ||
51 | objectRetention.Mode = *mode | ||
52 | } | ||
53 | |||
54 | return objectRetention, nil | ||
55 | } | ||
56 | |||
57 | // PutObjectRetentionOptions represents options specified by user for PutObject call | ||
58 | type PutObjectRetentionOptions struct { | ||
59 | GovernanceBypass bool | ||
60 | Mode *RetentionMode | ||
61 | RetainUntilDate *time.Time | ||
62 | VersionID string | ||
63 | } | ||
64 | |||
65 | // PutObjectRetention sets object retention for a given object and versionID. | ||
66 | func (c *Client) PutObjectRetention(ctx context.Context, bucketName, objectName string, opts PutObjectRetentionOptions) error { | ||
67 | // Input validation. | ||
68 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
69 | return err | ||
70 | } | ||
71 | |||
72 | if err := s3utils.CheckValidObjectName(objectName); err != nil { | ||
73 | return err | ||
74 | } | ||
75 | |||
76 | // Get resources properly escaped and lined up before | ||
77 | // using them in http request. | ||
78 | urlValues := make(url.Values) | ||
79 | urlValues.Set("retention", "") | ||
80 | |||
81 | if opts.VersionID != "" { | ||
82 | urlValues.Set("versionId", opts.VersionID) | ||
83 | } | ||
84 | |||
85 | retention, err := newObjectRetention(opts.Mode, opts.RetainUntilDate) | ||
86 | if err != nil { | ||
87 | return err | ||
88 | } | ||
89 | |||
90 | retentionData, err := xml.Marshal(retention) | ||
91 | if err != nil { | ||
92 | return err | ||
93 | } | ||
94 | |||
95 | // Build headers. | ||
96 | headers := make(http.Header) | ||
97 | |||
98 | if opts.GovernanceBypass { | ||
99 | // Set the bypass goverenance retention header | ||
100 | headers.Set(amzBypassGovernance, "true") | ||
101 | } | ||
102 | |||
103 | reqMetadata := requestMetadata{ | ||
104 | bucketName: bucketName, | ||
105 | objectName: objectName, | ||
106 | queryValues: urlValues, | ||
107 | contentBody: bytes.NewReader(retentionData), | ||
108 | contentLength: int64(len(retentionData)), | ||
109 | contentMD5Base64: sumMD5Base64(retentionData), | ||
110 | contentSHA256Hex: sum256Hex(retentionData), | ||
111 | customHeader: headers, | ||
112 | } | ||
113 | |||
114 | // Execute PUT Object Retention. | ||
115 | resp, err := c.executeMethod(ctx, http.MethodPut, reqMetadata) | ||
116 | defer closeResponse(resp) | ||
117 | if err != nil { | ||
118 | return err | ||
119 | } | ||
120 | if resp != nil { | ||
121 | if resp.StatusCode != http.StatusOK && resp.StatusCode != http.StatusNoContent { | ||
122 | return httpRespToErrorResponse(resp, bucketName, objectName) | ||
123 | } | ||
124 | } | ||
125 | return nil | ||
126 | } | ||
127 | |||
128 | // GetObjectRetention gets retention of given object. | ||
129 | func (c *Client) GetObjectRetention(ctx context.Context, bucketName, objectName, versionID string) (mode *RetentionMode, retainUntilDate *time.Time, err error) { | ||
130 | // Input validation. | ||
131 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
132 | return nil, nil, err | ||
133 | } | ||
134 | |||
135 | if err := s3utils.CheckValidObjectName(objectName); err != nil { | ||
136 | return nil, nil, err | ||
137 | } | ||
138 | urlValues := make(url.Values) | ||
139 | urlValues.Set("retention", "") | ||
140 | if versionID != "" { | ||
141 | urlValues.Set("versionId", versionID) | ||
142 | } | ||
143 | // Execute GET on bucket to list objects. | ||
144 | resp, err := c.executeMethod(ctx, http.MethodGet, requestMetadata{ | ||
145 | bucketName: bucketName, | ||
146 | objectName: objectName, | ||
147 | queryValues: urlValues, | ||
148 | contentSHA256Hex: emptySHA256Hex, | ||
149 | }) | ||
150 | defer closeResponse(resp) | ||
151 | if err != nil { | ||
152 | return nil, nil, err | ||
153 | } | ||
154 | if resp != nil { | ||
155 | if resp.StatusCode != http.StatusOK { | ||
156 | return nil, nil, httpRespToErrorResponse(resp, bucketName, objectName) | ||
157 | } | ||
158 | } | ||
159 | retention := &objectRetention{} | ||
160 | if err = xml.NewDecoder(resp.Body).Decode(retention); err != nil { | ||
161 | return nil, nil, err | ||
162 | } | ||
163 | |||
164 | return &retention.Mode, retention.RetainUntilDate, nil | ||
165 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/api-object-tagging.go b/vendor/github.com/minio/minio-go/v7/api-object-tagging.go deleted file mode 100644 index 6623e26..0000000 --- a/vendor/github.com/minio/minio-go/v7/api-object-tagging.go +++ /dev/null | |||
@@ -1,177 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2020 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package minio | ||
19 | |||
20 | import ( | ||
21 | "bytes" | ||
22 | "context" | ||
23 | "encoding/xml" | ||
24 | "net/http" | ||
25 | "net/url" | ||
26 | |||
27 | "github.com/minio/minio-go/v7/pkg/s3utils" | ||
28 | "github.com/minio/minio-go/v7/pkg/tags" | ||
29 | ) | ||
30 | |||
31 | // PutObjectTaggingOptions holds an object version id | ||
32 | // to update tag(s) of a specific object version | ||
33 | type PutObjectTaggingOptions struct { | ||
34 | VersionID string | ||
35 | Internal AdvancedObjectTaggingOptions | ||
36 | } | ||
37 | |||
38 | // AdvancedObjectTaggingOptions for internal use by MinIO server - not intended for client use. | ||
39 | type AdvancedObjectTaggingOptions struct { | ||
40 | ReplicationProxyRequest string | ||
41 | } | ||
42 | |||
43 | // PutObjectTagging replaces or creates object tag(s) and can target | ||
44 | // a specific object version in a versioned bucket. | ||
45 | func (c *Client) PutObjectTagging(ctx context.Context, bucketName, objectName string, otags *tags.Tags, opts PutObjectTaggingOptions) error { | ||
46 | // Input validation. | ||
47 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
48 | return err | ||
49 | } | ||
50 | |||
51 | // Get resources properly escaped and lined up before | ||
52 | // using them in http request. | ||
53 | urlValues := make(url.Values) | ||
54 | urlValues.Set("tagging", "") | ||
55 | |||
56 | if opts.VersionID != "" { | ||
57 | urlValues.Set("versionId", opts.VersionID) | ||
58 | } | ||
59 | headers := make(http.Header, 0) | ||
60 | if opts.Internal.ReplicationProxyRequest != "" { | ||
61 | headers.Set(minIOBucketReplicationProxyRequest, opts.Internal.ReplicationProxyRequest) | ||
62 | } | ||
63 | reqBytes, err := xml.Marshal(otags) | ||
64 | if err != nil { | ||
65 | return err | ||
66 | } | ||
67 | |||
68 | reqMetadata := requestMetadata{ | ||
69 | bucketName: bucketName, | ||
70 | objectName: objectName, | ||
71 | queryValues: urlValues, | ||
72 | contentBody: bytes.NewReader(reqBytes), | ||
73 | contentLength: int64(len(reqBytes)), | ||
74 | contentMD5Base64: sumMD5Base64(reqBytes), | ||
75 | customHeader: headers, | ||
76 | } | ||
77 | |||
78 | // Execute PUT to set a object tagging. | ||
79 | resp, err := c.executeMethod(ctx, http.MethodPut, reqMetadata) | ||
80 | defer closeResponse(resp) | ||
81 | if err != nil { | ||
82 | return err | ||
83 | } | ||
84 | if resp != nil { | ||
85 | if resp.StatusCode != http.StatusOK { | ||
86 | return httpRespToErrorResponse(resp, bucketName, objectName) | ||
87 | } | ||
88 | } | ||
89 | return nil | ||
90 | } | ||
91 | |||
92 | // GetObjectTaggingOptions holds the object version ID | ||
93 | // to fetch the tagging key/value pairs | ||
94 | type GetObjectTaggingOptions struct { | ||
95 | VersionID string | ||
96 | Internal AdvancedObjectTaggingOptions | ||
97 | } | ||
98 | |||
99 | // GetObjectTagging fetches object tag(s) with options to target | ||
100 | // a specific object version in a versioned bucket. | ||
101 | func (c *Client) GetObjectTagging(ctx context.Context, bucketName, objectName string, opts GetObjectTaggingOptions) (*tags.Tags, error) { | ||
102 | // Get resources properly escaped and lined up before | ||
103 | // using them in http request. | ||
104 | urlValues := make(url.Values) | ||
105 | urlValues.Set("tagging", "") | ||
106 | |||
107 | if opts.VersionID != "" { | ||
108 | urlValues.Set("versionId", opts.VersionID) | ||
109 | } | ||
110 | headers := make(http.Header, 0) | ||
111 | if opts.Internal.ReplicationProxyRequest != "" { | ||
112 | headers.Set(minIOBucketReplicationProxyRequest, opts.Internal.ReplicationProxyRequest) | ||
113 | } | ||
114 | // Execute GET on object to get object tag(s) | ||
115 | resp, err := c.executeMethod(ctx, http.MethodGet, requestMetadata{ | ||
116 | bucketName: bucketName, | ||
117 | objectName: objectName, | ||
118 | queryValues: urlValues, | ||
119 | customHeader: headers, | ||
120 | }) | ||
121 | |||
122 | defer closeResponse(resp) | ||
123 | if err != nil { | ||
124 | return nil, err | ||
125 | } | ||
126 | |||
127 | if resp != nil { | ||
128 | if resp.StatusCode != http.StatusOK { | ||
129 | return nil, httpRespToErrorResponse(resp, bucketName, objectName) | ||
130 | } | ||
131 | } | ||
132 | |||
133 | return tags.ParseObjectXML(resp.Body) | ||
134 | } | ||
135 | |||
136 | // RemoveObjectTaggingOptions holds the version id of the object to remove | ||
137 | type RemoveObjectTaggingOptions struct { | ||
138 | VersionID string | ||
139 | Internal AdvancedObjectTaggingOptions | ||
140 | } | ||
141 | |||
142 | // RemoveObjectTagging removes object tag(s) with options to control a specific object | ||
143 | // version in a versioned bucket | ||
144 | func (c *Client) RemoveObjectTagging(ctx context.Context, bucketName, objectName string, opts RemoveObjectTaggingOptions) error { | ||
145 | // Get resources properly escaped and lined up before | ||
146 | // using them in http request. | ||
147 | urlValues := make(url.Values) | ||
148 | urlValues.Set("tagging", "") | ||
149 | |||
150 | if opts.VersionID != "" { | ||
151 | urlValues.Set("versionId", opts.VersionID) | ||
152 | } | ||
153 | headers := make(http.Header, 0) | ||
154 | if opts.Internal.ReplicationProxyRequest != "" { | ||
155 | headers.Set(minIOBucketReplicationProxyRequest, opts.Internal.ReplicationProxyRequest) | ||
156 | } | ||
157 | // Execute DELETE on object to remove object tag(s) | ||
158 | resp, err := c.executeMethod(ctx, http.MethodDelete, requestMetadata{ | ||
159 | bucketName: bucketName, | ||
160 | objectName: objectName, | ||
161 | queryValues: urlValues, | ||
162 | customHeader: headers, | ||
163 | }) | ||
164 | |||
165 | defer closeResponse(resp) | ||
166 | if err != nil { | ||
167 | return err | ||
168 | } | ||
169 | |||
170 | if resp != nil { | ||
171 | // S3 returns "204 No content" after Object tag deletion. | ||
172 | if resp.StatusCode != http.StatusNoContent { | ||
173 | return httpRespToErrorResponse(resp, bucketName, objectName) | ||
174 | } | ||
175 | } | ||
176 | return err | ||
177 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/api-presigned.go b/vendor/github.com/minio/minio-go/v7/api-presigned.go deleted file mode 100644 index 9e85f81..0000000 --- a/vendor/github.com/minio/minio-go/v7/api-presigned.go +++ /dev/null | |||
@@ -1,228 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2015-2017 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package minio | ||
19 | |||
20 | import ( | ||
21 | "context" | ||
22 | "errors" | ||
23 | "net/http" | ||
24 | "net/url" | ||
25 | "time" | ||
26 | |||
27 | "github.com/minio/minio-go/v7/pkg/s3utils" | ||
28 | "github.com/minio/minio-go/v7/pkg/signer" | ||
29 | ) | ||
30 | |||
31 | // presignURL - Returns a presigned URL for an input 'method'. | ||
32 | // Expires maximum is 7days - ie. 604800 and minimum is 1. | ||
33 | func (c *Client) presignURL(ctx context.Context, method, bucketName, objectName string, expires time.Duration, reqParams url.Values, extraHeaders http.Header) (u *url.URL, err error) { | ||
34 | // Input validation. | ||
35 | if method == "" { | ||
36 | return nil, errInvalidArgument("method cannot be empty.") | ||
37 | } | ||
38 | if err = s3utils.CheckValidBucketName(bucketName); err != nil { | ||
39 | return nil, err | ||
40 | } | ||
41 | if err = isValidExpiry(expires); err != nil { | ||
42 | return nil, err | ||
43 | } | ||
44 | |||
45 | // Convert expires into seconds. | ||
46 | expireSeconds := int64(expires / time.Second) | ||
47 | reqMetadata := requestMetadata{ | ||
48 | presignURL: true, | ||
49 | bucketName: bucketName, | ||
50 | objectName: objectName, | ||
51 | expires: expireSeconds, | ||
52 | queryValues: reqParams, | ||
53 | extraPresignHeader: extraHeaders, | ||
54 | } | ||
55 | |||
56 | // Instantiate a new request. | ||
57 | // Since expires is set newRequest will presign the request. | ||
58 | var req *http.Request | ||
59 | if req, err = c.newRequest(ctx, method, reqMetadata); err != nil { | ||
60 | return nil, err | ||
61 | } | ||
62 | return req.URL, nil | ||
63 | } | ||
64 | |||
65 | // PresignedGetObject - Returns a presigned URL to access an object | ||
66 | // data without credentials. URL can have a maximum expiry of | ||
67 | // upto 7days or a minimum of 1sec. Additionally you can override | ||
68 | // a set of response headers using the query parameters. | ||
69 | func (c *Client) PresignedGetObject(ctx context.Context, bucketName, objectName string, expires time.Duration, reqParams url.Values) (u *url.URL, err error) { | ||
70 | if err = s3utils.CheckValidObjectName(objectName); err != nil { | ||
71 | return nil, err | ||
72 | } | ||
73 | return c.presignURL(ctx, http.MethodGet, bucketName, objectName, expires, reqParams, nil) | ||
74 | } | ||
75 | |||
76 | // PresignedHeadObject - Returns a presigned URL to access | ||
77 | // object metadata without credentials. URL can have a maximum expiry | ||
78 | // of upto 7days or a minimum of 1sec. Additionally you can override | ||
79 | // a set of response headers using the query parameters. | ||
80 | func (c *Client) PresignedHeadObject(ctx context.Context, bucketName, objectName string, expires time.Duration, reqParams url.Values) (u *url.URL, err error) { | ||
81 | if err = s3utils.CheckValidObjectName(objectName); err != nil { | ||
82 | return nil, err | ||
83 | } | ||
84 | return c.presignURL(ctx, http.MethodHead, bucketName, objectName, expires, reqParams, nil) | ||
85 | } | ||
86 | |||
87 | // PresignedPutObject - Returns a presigned URL to upload an object | ||
88 | // without credentials. URL can have a maximum expiry of upto 7days | ||
89 | // or a minimum of 1sec. | ||
90 | func (c *Client) PresignedPutObject(ctx context.Context, bucketName, objectName string, expires time.Duration) (u *url.URL, err error) { | ||
91 | if err = s3utils.CheckValidObjectName(objectName); err != nil { | ||
92 | return nil, err | ||
93 | } | ||
94 | return c.presignURL(ctx, http.MethodPut, bucketName, objectName, expires, nil, nil) | ||
95 | } | ||
96 | |||
97 | // PresignHeader - similar to Presign() but allows including HTTP headers that | ||
98 | // will be used to build the signature. The request using the resulting URL will | ||
99 | // need to have the exact same headers to be added for signature validation to | ||
100 | // pass. | ||
101 | // | ||
102 | // FIXME: The extra header parameter should be included in Presign() in the next | ||
103 | // major version bump, and this function should then be deprecated. | ||
104 | func (c *Client) PresignHeader(ctx context.Context, method, bucketName, objectName string, expires time.Duration, reqParams url.Values, extraHeaders http.Header) (u *url.URL, err error) { | ||
105 | return c.presignURL(ctx, method, bucketName, objectName, expires, reqParams, extraHeaders) | ||
106 | } | ||
107 | |||
108 | // Presign - returns a presigned URL for any http method of your choice along | ||
109 | // with custom request params and extra signed headers. URL can have a maximum | ||
110 | // expiry of upto 7days or a minimum of 1sec. | ||
111 | func (c *Client) Presign(ctx context.Context, method, bucketName, objectName string, expires time.Duration, reqParams url.Values) (u *url.URL, err error) { | ||
112 | return c.presignURL(ctx, method, bucketName, objectName, expires, reqParams, nil) | ||
113 | } | ||
114 | |||
115 | // PresignedPostPolicy - Returns POST urlString, form data to upload an object. | ||
116 | func (c *Client) PresignedPostPolicy(ctx context.Context, p *PostPolicy) (u *url.URL, formData map[string]string, err error) { | ||
117 | // Validate input arguments. | ||
118 | if p.expiration.IsZero() { | ||
119 | return nil, nil, errors.New("Expiration time must be specified") | ||
120 | } | ||
121 | if _, ok := p.formData["key"]; !ok { | ||
122 | return nil, nil, errors.New("object key must be specified") | ||
123 | } | ||
124 | if _, ok := p.formData["bucket"]; !ok { | ||
125 | return nil, nil, errors.New("bucket name must be specified") | ||
126 | } | ||
127 | |||
128 | bucketName := p.formData["bucket"] | ||
129 | // Fetch the bucket location. | ||
130 | location, err := c.getBucketLocation(ctx, bucketName) | ||
131 | if err != nil { | ||
132 | return nil, nil, err | ||
133 | } | ||
134 | |||
135 | isVirtualHost := c.isVirtualHostStyleRequest(*c.endpointURL, bucketName) | ||
136 | |||
137 | u, err = c.makeTargetURL(bucketName, "", location, isVirtualHost, nil) | ||
138 | if err != nil { | ||
139 | return nil, nil, err | ||
140 | } | ||
141 | |||
142 | // Get credentials from the configured credentials provider. | ||
143 | credValues, err := c.credsProvider.Get() | ||
144 | if err != nil { | ||
145 | return nil, nil, err | ||
146 | } | ||
147 | |||
148 | var ( | ||
149 | signerType = credValues.SignerType | ||
150 | sessionToken = credValues.SessionToken | ||
151 | accessKeyID = credValues.AccessKeyID | ||
152 | secretAccessKey = credValues.SecretAccessKey | ||
153 | ) | ||
154 | |||
155 | if signerType.IsAnonymous() { | ||
156 | return nil, nil, errInvalidArgument("Presigned operations are not supported for anonymous credentials") | ||
157 | } | ||
158 | |||
159 | // Keep time. | ||
160 | t := time.Now().UTC() | ||
161 | // For signature version '2' handle here. | ||
162 | if signerType.IsV2() { | ||
163 | policyBase64 := p.base64() | ||
164 | p.formData["policy"] = policyBase64 | ||
165 | // For Google endpoint set this value to be 'GoogleAccessId'. | ||
166 | if s3utils.IsGoogleEndpoint(*c.endpointURL) { | ||
167 | p.formData["GoogleAccessId"] = accessKeyID | ||
168 | } else { | ||
169 | // For all other endpoints set this value to be 'AWSAccessKeyId'. | ||
170 | p.formData["AWSAccessKeyId"] = accessKeyID | ||
171 | } | ||
172 | // Sign the policy. | ||
173 | p.formData["signature"] = signer.PostPresignSignatureV2(policyBase64, secretAccessKey) | ||
174 | return u, p.formData, nil | ||
175 | } | ||
176 | |||
177 | // Add date policy. | ||
178 | if err = p.addNewPolicy(policyCondition{ | ||
179 | matchType: "eq", | ||
180 | condition: "$x-amz-date", | ||
181 | value: t.Format(iso8601DateFormat), | ||
182 | }); err != nil { | ||
183 | return nil, nil, err | ||
184 | } | ||
185 | |||
186 | // Add algorithm policy. | ||
187 | if err = p.addNewPolicy(policyCondition{ | ||
188 | matchType: "eq", | ||
189 | condition: "$x-amz-algorithm", | ||
190 | value: signV4Algorithm, | ||
191 | }); err != nil { | ||
192 | return nil, nil, err | ||
193 | } | ||
194 | |||
195 | // Add a credential policy. | ||
196 | credential := signer.GetCredential(accessKeyID, location, t, signer.ServiceTypeS3) | ||
197 | if err = p.addNewPolicy(policyCondition{ | ||
198 | matchType: "eq", | ||
199 | condition: "$x-amz-credential", | ||
200 | value: credential, | ||
201 | }); err != nil { | ||
202 | return nil, nil, err | ||
203 | } | ||
204 | |||
205 | if sessionToken != "" { | ||
206 | if err = p.addNewPolicy(policyCondition{ | ||
207 | matchType: "eq", | ||
208 | condition: "$x-amz-security-token", | ||
209 | value: sessionToken, | ||
210 | }); err != nil { | ||
211 | return nil, nil, err | ||
212 | } | ||
213 | } | ||
214 | |||
215 | // Get base64 encoded policy. | ||
216 | policyBase64 := p.base64() | ||
217 | |||
218 | // Fill in the form data. | ||
219 | p.formData["policy"] = policyBase64 | ||
220 | p.formData["x-amz-algorithm"] = signV4Algorithm | ||
221 | p.formData["x-amz-credential"] = credential | ||
222 | p.formData["x-amz-date"] = t.Format(iso8601DateFormat) | ||
223 | if sessionToken != "" { | ||
224 | p.formData["x-amz-security-token"] = sessionToken | ||
225 | } | ||
226 | p.formData["x-amz-signature"] = signer.PostPresignSignatureV4(policyBase64, t, secretAccessKey, location) | ||
227 | return u, p.formData, nil | ||
228 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/api-put-bucket.go b/vendor/github.com/minio/minio-go/v7/api-put-bucket.go deleted file mode 100644 index 7376669..0000000 --- a/vendor/github.com/minio/minio-go/v7/api-put-bucket.go +++ /dev/null | |||
@@ -1,123 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2015-2020 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package minio | ||
19 | |||
20 | import ( | ||
21 | "bytes" | ||
22 | "context" | ||
23 | "encoding/xml" | ||
24 | "net/http" | ||
25 | |||
26 | "github.com/minio/minio-go/v7/pkg/s3utils" | ||
27 | ) | ||
28 | |||
29 | // Bucket operations | ||
30 | func (c *Client) makeBucket(ctx context.Context, bucketName string, opts MakeBucketOptions) (err error) { | ||
31 | // Validate the input arguments. | ||
32 | if err := s3utils.CheckValidBucketNameStrict(bucketName); err != nil { | ||
33 | return err | ||
34 | } | ||
35 | |||
36 | err = c.doMakeBucket(ctx, bucketName, opts.Region, opts.ObjectLocking) | ||
37 | if err != nil && (opts.Region == "" || opts.Region == "us-east-1") { | ||
38 | if resp, ok := err.(ErrorResponse); ok && resp.Code == "AuthorizationHeaderMalformed" && resp.Region != "" { | ||
39 | err = c.doMakeBucket(ctx, bucketName, resp.Region, opts.ObjectLocking) | ||
40 | } | ||
41 | } | ||
42 | return err | ||
43 | } | ||
44 | |||
45 | func (c *Client) doMakeBucket(ctx context.Context, bucketName, location string, objectLockEnabled bool) (err error) { | ||
46 | defer func() { | ||
47 | // Save the location into cache on a successful makeBucket response. | ||
48 | if err == nil { | ||
49 | c.bucketLocCache.Set(bucketName, location) | ||
50 | } | ||
51 | }() | ||
52 | |||
53 | // If location is empty, treat is a default region 'us-east-1'. | ||
54 | if location == "" { | ||
55 | location = "us-east-1" | ||
56 | // For custom region clients, default | ||
57 | // to custom region instead not 'us-east-1'. | ||
58 | if c.region != "" { | ||
59 | location = c.region | ||
60 | } | ||
61 | } | ||
62 | // PUT bucket request metadata. | ||
63 | reqMetadata := requestMetadata{ | ||
64 | bucketName: bucketName, | ||
65 | bucketLocation: location, | ||
66 | } | ||
67 | |||
68 | if objectLockEnabled { | ||
69 | headers := make(http.Header) | ||
70 | headers.Add("x-amz-bucket-object-lock-enabled", "true") | ||
71 | reqMetadata.customHeader = headers | ||
72 | } | ||
73 | |||
74 | // If location is not 'us-east-1' create bucket location config. | ||
75 | if location != "us-east-1" && location != "" { | ||
76 | createBucketConfig := createBucketConfiguration{} | ||
77 | createBucketConfig.Location = location | ||
78 | var createBucketConfigBytes []byte | ||
79 | createBucketConfigBytes, err = xml.Marshal(createBucketConfig) | ||
80 | if err != nil { | ||
81 | return err | ||
82 | } | ||
83 | reqMetadata.contentMD5Base64 = sumMD5Base64(createBucketConfigBytes) | ||
84 | reqMetadata.contentSHA256Hex = sum256Hex(createBucketConfigBytes) | ||
85 | reqMetadata.contentBody = bytes.NewReader(createBucketConfigBytes) | ||
86 | reqMetadata.contentLength = int64(len(createBucketConfigBytes)) | ||
87 | } | ||
88 | |||
89 | // Execute PUT to create a new bucket. | ||
90 | resp, err := c.executeMethod(ctx, http.MethodPut, reqMetadata) | ||
91 | defer closeResponse(resp) | ||
92 | if err != nil { | ||
93 | return err | ||
94 | } | ||
95 | |||
96 | if resp != nil { | ||
97 | if resp.StatusCode != http.StatusOK { | ||
98 | return httpRespToErrorResponse(resp, bucketName, "") | ||
99 | } | ||
100 | } | ||
101 | |||
102 | // Success. | ||
103 | return nil | ||
104 | } | ||
105 | |||
106 | // MakeBucketOptions holds all options to tweak bucket creation | ||
107 | type MakeBucketOptions struct { | ||
108 | // Bucket location | ||
109 | Region string | ||
110 | // Enable object locking | ||
111 | ObjectLocking bool | ||
112 | } | ||
113 | |||
114 | // MakeBucket creates a new bucket with bucketName with a context to control cancellations and timeouts. | ||
115 | // | ||
116 | // Location is an optional argument, by default all buckets are | ||
117 | // created in US Standard Region. | ||
118 | // | ||
119 | // For Amazon S3 for more supported regions - http://docs.aws.amazon.com/general/latest/gr/rande.html | ||
120 | // For Google Cloud Storage for more supported regions - https://cloud.google.com/storage/docs/bucket-locations | ||
121 | func (c *Client) MakeBucket(ctx context.Context, bucketName string, opts MakeBucketOptions) (err error) { | ||
122 | return c.makeBucket(ctx, bucketName, opts) | ||
123 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/api-put-object-common.go b/vendor/github.com/minio/minio-go/v7/api-put-object-common.go deleted file mode 100644 index 9ccb97c..0000000 --- a/vendor/github.com/minio/minio-go/v7/api-put-object-common.go +++ /dev/null | |||
@@ -1,149 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2015-2017 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package minio | ||
19 | |||
20 | import ( | ||
21 | "context" | ||
22 | "io" | ||
23 | "math" | ||
24 | "os" | ||
25 | |||
26 | "github.com/minio/minio-go/v7/pkg/s3utils" | ||
27 | ) | ||
28 | |||
29 | const nullVersionID = "null" | ||
30 | |||
31 | // Verify if reader is *minio.Object | ||
32 | func isObject(reader io.Reader) (ok bool) { | ||
33 | _, ok = reader.(*Object) | ||
34 | return | ||
35 | } | ||
36 | |||
37 | // Verify if reader is a generic ReaderAt | ||
38 | func isReadAt(reader io.Reader) (ok bool) { | ||
39 | var v *os.File | ||
40 | v, ok = reader.(*os.File) | ||
41 | if ok { | ||
42 | // Stdin, Stdout and Stderr all have *os.File type | ||
43 | // which happen to also be io.ReaderAt compatible | ||
44 | // we need to add special conditions for them to | ||
45 | // be ignored by this function. | ||
46 | for _, f := range []string{ | ||
47 | "/dev/stdin", | ||
48 | "/dev/stdout", | ||
49 | "/dev/stderr", | ||
50 | } { | ||
51 | if f == v.Name() { | ||
52 | ok = false | ||
53 | break | ||
54 | } | ||
55 | } | ||
56 | } else { | ||
57 | _, ok = reader.(io.ReaderAt) | ||
58 | } | ||
59 | return | ||
60 | } | ||
61 | |||
62 | // OptimalPartInfo - calculate the optimal part info for a given | ||
63 | // object size. | ||
64 | // | ||
65 | // NOTE: Assumption here is that for any object to be uploaded to any S3 compatible | ||
66 | // object storage it will have the following parameters as constants. | ||
67 | // | ||
68 | // maxPartsCount - 10000 | ||
69 | // minPartSize - 16MiB | ||
70 | // maxMultipartPutObjectSize - 5TiB | ||
71 | func OptimalPartInfo(objectSize int64, configuredPartSize uint64) (totalPartsCount int, partSize, lastPartSize int64, err error) { | ||
72 | // object size is '-1' set it to 5TiB. | ||
73 | var unknownSize bool | ||
74 | if objectSize == -1 { | ||
75 | unknownSize = true | ||
76 | objectSize = maxMultipartPutObjectSize | ||
77 | } | ||
78 | |||
79 | // object size is larger than supported maximum. | ||
80 | if objectSize > maxMultipartPutObjectSize { | ||
81 | err = errEntityTooLarge(objectSize, maxMultipartPutObjectSize, "", "") | ||
82 | return | ||
83 | } | ||
84 | |||
85 | var partSizeFlt float64 | ||
86 | if configuredPartSize > 0 { | ||
87 | if int64(configuredPartSize) > objectSize { | ||
88 | err = errEntityTooLarge(int64(configuredPartSize), objectSize, "", "") | ||
89 | return | ||
90 | } | ||
91 | |||
92 | if !unknownSize { | ||
93 | if objectSize > (int64(configuredPartSize) * maxPartsCount) { | ||
94 | err = errInvalidArgument("Part size * max_parts(10000) is lesser than input objectSize.") | ||
95 | return | ||
96 | } | ||
97 | } | ||
98 | |||
99 | if configuredPartSize < absMinPartSize { | ||
100 | err = errInvalidArgument("Input part size is smaller than allowed minimum of 5MiB.") | ||
101 | return | ||
102 | } | ||
103 | |||
104 | if configuredPartSize > maxPartSize { | ||
105 | err = errInvalidArgument("Input part size is bigger than allowed maximum of 5GiB.") | ||
106 | return | ||
107 | } | ||
108 | |||
109 | partSizeFlt = float64(configuredPartSize) | ||
110 | if unknownSize { | ||
111 | // If input has unknown size and part size is configured | ||
112 | // keep it to maximum allowed as per 10000 parts. | ||
113 | objectSize = int64(configuredPartSize) * maxPartsCount | ||
114 | } | ||
115 | } else { | ||
116 | configuredPartSize = minPartSize | ||
117 | // Use floats for part size for all calculations to avoid | ||
118 | // overflows during float64 to int64 conversions. | ||
119 | partSizeFlt = float64(objectSize / maxPartsCount) | ||
120 | partSizeFlt = math.Ceil(partSizeFlt/float64(configuredPartSize)) * float64(configuredPartSize) | ||
121 | } | ||
122 | |||
123 | // Total parts count. | ||
124 | totalPartsCount = int(math.Ceil(float64(objectSize) / partSizeFlt)) | ||
125 | // Part size. | ||
126 | partSize = int64(partSizeFlt) | ||
127 | // Last part size. | ||
128 | lastPartSize = objectSize - int64(totalPartsCount-1)*partSize | ||
129 | return totalPartsCount, partSize, lastPartSize, nil | ||
130 | } | ||
131 | |||
132 | // getUploadID - fetch upload id if already present for an object name | ||
133 | // or initiate a new request to fetch a new upload id. | ||
134 | func (c *Client) newUploadID(ctx context.Context, bucketName, objectName string, opts PutObjectOptions) (uploadID string, err error) { | ||
135 | // Input validation. | ||
136 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
137 | return "", err | ||
138 | } | ||
139 | if err := s3utils.CheckValidObjectName(objectName); err != nil { | ||
140 | return "", err | ||
141 | } | ||
142 | |||
143 | // Initiate multipart upload for an object. | ||
144 | initMultipartUploadResult, err := c.initiateMultipartUpload(ctx, bucketName, objectName, opts) | ||
145 | if err != nil { | ||
146 | return "", err | ||
147 | } | ||
148 | return initMultipartUploadResult.UploadID, nil | ||
149 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/api-put-object-fan-out.go b/vendor/github.com/minio/minio-go/v7/api-put-object-fan-out.go deleted file mode 100644 index 0ae9142..0000000 --- a/vendor/github.com/minio/minio-go/v7/api-put-object-fan-out.go +++ /dev/null | |||
@@ -1,164 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2023 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package minio | ||
19 | |||
20 | import ( | ||
21 | "context" | ||
22 | "encoding/json" | ||
23 | "errors" | ||
24 | "io" | ||
25 | "mime/multipart" | ||
26 | "net/http" | ||
27 | "strconv" | ||
28 | "strings" | ||
29 | "time" | ||
30 | |||
31 | "github.com/minio/minio-go/v7/pkg/encrypt" | ||
32 | ) | ||
33 | |||
34 | // PutObjectFanOutEntry is per object entry fan-out metadata | ||
35 | type PutObjectFanOutEntry struct { | ||
36 | Key string `json:"key"` | ||
37 | UserMetadata map[string]string `json:"metadata,omitempty"` | ||
38 | UserTags map[string]string `json:"tags,omitempty"` | ||
39 | ContentType string `json:"contentType,omitempty"` | ||
40 | ContentEncoding string `json:"contentEncoding,omitempty"` | ||
41 | ContentDisposition string `json:"contentDisposition,omitempty"` | ||
42 | ContentLanguage string `json:"contentLanguage,omitempty"` | ||
43 | CacheControl string `json:"cacheControl,omitempty"` | ||
44 | Retention RetentionMode `json:"retention,omitempty"` | ||
45 | RetainUntilDate *time.Time `json:"retainUntil,omitempty"` | ||
46 | } | ||
47 | |||
48 | // PutObjectFanOutRequest this is the request structure sent | ||
49 | // to the server to fan-out the stream to multiple objects. | ||
50 | type PutObjectFanOutRequest struct { | ||
51 | Entries []PutObjectFanOutEntry | ||
52 | Checksum Checksum | ||
53 | SSE encrypt.ServerSide | ||
54 | } | ||
55 | |||
56 | // PutObjectFanOutResponse this is the response structure sent | ||
57 | // by the server upon success or failure for each object | ||
58 | // fan-out keys. Additionally, this response carries ETag, | ||
59 | // VersionID and LastModified for each object fan-out. | ||
60 | type PutObjectFanOutResponse struct { | ||
61 | Key string `json:"key"` | ||
62 | ETag string `json:"etag,omitempty"` | ||
63 | VersionID string `json:"versionId,omitempty"` | ||
64 | LastModified *time.Time `json:"lastModified,omitempty"` | ||
65 | Error string `json:"error,omitempty"` | ||
66 | } | ||
67 | |||
68 | // PutObjectFanOut - is a variant of PutObject instead of writing a single object from a single | ||
69 | // stream multiple objects are written, defined via a list of PutObjectFanOutRequests. Each entry | ||
70 | // in PutObjectFanOutRequest carries an object keyname and its relevant metadata if any. `Key` is | ||
71 | // mandatory, rest of the other options in PutObjectFanOutRequest are optional. | ||
72 | func (c *Client) PutObjectFanOut(ctx context.Context, bucket string, fanOutData io.Reader, fanOutReq PutObjectFanOutRequest) ([]PutObjectFanOutResponse, error) { | ||
73 | if len(fanOutReq.Entries) == 0 { | ||
74 | return nil, errInvalidArgument("fan out requests cannot be empty") | ||
75 | } | ||
76 | |||
77 | policy := NewPostPolicy() | ||
78 | policy.SetBucket(bucket) | ||
79 | policy.SetKey(strconv.FormatInt(time.Now().UnixNano(), 16)) | ||
80 | |||
81 | // Expires in 15 minutes. | ||
82 | policy.SetExpires(time.Now().UTC().Add(15 * time.Minute)) | ||
83 | |||
84 | // Set encryption headers if any. | ||
85 | policy.SetEncryption(fanOutReq.SSE) | ||
86 | |||
87 | // Set checksum headers if any. | ||
88 | policy.SetChecksum(fanOutReq.Checksum) | ||
89 | |||
90 | url, formData, err := c.PresignedPostPolicy(ctx, policy) | ||
91 | if err != nil { | ||
92 | return nil, err | ||
93 | } | ||
94 | |||
95 | r, w := io.Pipe() | ||
96 | |||
97 | req, err := http.NewRequest(http.MethodPost, url.String(), r) | ||
98 | if err != nil { | ||
99 | w.Close() | ||
100 | return nil, err | ||
101 | } | ||
102 | |||
103 | var b strings.Builder | ||
104 | enc := json.NewEncoder(&b) | ||
105 | for _, req := range fanOutReq.Entries { | ||
106 | if req.Key == "" { | ||
107 | w.Close() | ||
108 | return nil, errors.New("PutObjectFanOutRequest.Key is mandatory and cannot be empty") | ||
109 | } | ||
110 | if err = enc.Encode(&req); err != nil { | ||
111 | w.Close() | ||
112 | return nil, err | ||
113 | } | ||
114 | } | ||
115 | |||
116 | mwriter := multipart.NewWriter(w) | ||
117 | req.Header.Add("Content-Type", mwriter.FormDataContentType()) | ||
118 | |||
119 | go func() { | ||
120 | defer w.Close() | ||
121 | defer mwriter.Close() | ||
122 | |||
123 | for k, v := range formData { | ||
124 | if err := mwriter.WriteField(k, v); err != nil { | ||
125 | return | ||
126 | } | ||
127 | } | ||
128 | |||
129 | if err := mwriter.WriteField("x-minio-fanout-list", b.String()); err != nil { | ||
130 | return | ||
131 | } | ||
132 | |||
133 | mw, err := mwriter.CreateFormFile("file", "fanout-content") | ||
134 | if err != nil { | ||
135 | return | ||
136 | } | ||
137 | |||
138 | if _, err = io.Copy(mw, fanOutData); err != nil { | ||
139 | return | ||
140 | } | ||
141 | }() | ||
142 | |||
143 | resp, err := c.do(req) | ||
144 | if err != nil { | ||
145 | return nil, err | ||
146 | } | ||
147 | defer closeResponse(resp) | ||
148 | |||
149 | if resp.StatusCode != http.StatusOK { | ||
150 | return nil, httpRespToErrorResponse(resp, bucket, "fanout-content") | ||
151 | } | ||
152 | |||
153 | dec := json.NewDecoder(resp.Body) | ||
154 | fanOutResp := make([]PutObjectFanOutResponse, 0, len(fanOutReq.Entries)) | ||
155 | for dec.More() { | ||
156 | var m PutObjectFanOutResponse | ||
157 | if err = dec.Decode(&m); err != nil { | ||
158 | return nil, err | ||
159 | } | ||
160 | fanOutResp = append(fanOutResp, m) | ||
161 | } | ||
162 | |||
163 | return fanOutResp, nil | ||
164 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/api-put-object-file-context.go b/vendor/github.com/minio/minio-go/v7/api-put-object-file-context.go deleted file mode 100644 index 4d29dfc..0000000 --- a/vendor/github.com/minio/minio-go/v7/api-put-object-file-context.go +++ /dev/null | |||
@@ -1,64 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2017 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package minio | ||
19 | |||
20 | import ( | ||
21 | "context" | ||
22 | "mime" | ||
23 | "os" | ||
24 | "path/filepath" | ||
25 | |||
26 | "github.com/minio/minio-go/v7/pkg/s3utils" | ||
27 | ) | ||
28 | |||
29 | // FPutObject - Create an object in a bucket, with contents from file at filePath. Allows request cancellation. | ||
30 | func (c *Client) FPutObject(ctx context.Context, bucketName, objectName, filePath string, opts PutObjectOptions) (info UploadInfo, err error) { | ||
31 | // Input validation. | ||
32 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
33 | return UploadInfo{}, err | ||
34 | } | ||
35 | if err := s3utils.CheckValidObjectName(objectName); err != nil { | ||
36 | return UploadInfo{}, err | ||
37 | } | ||
38 | |||
39 | // Open the referenced file. | ||
40 | fileReader, err := os.Open(filePath) | ||
41 | // If any error fail quickly here. | ||
42 | if err != nil { | ||
43 | return UploadInfo{}, err | ||
44 | } | ||
45 | defer fileReader.Close() | ||
46 | |||
47 | // Save the file stat. | ||
48 | fileStat, err := fileReader.Stat() | ||
49 | if err != nil { | ||
50 | return UploadInfo{}, err | ||
51 | } | ||
52 | |||
53 | // Save the file size. | ||
54 | fileSize := fileStat.Size() | ||
55 | |||
56 | // Set contentType based on filepath extension if not given or default | ||
57 | // value of "application/octet-stream" if the extension has no associated type. | ||
58 | if opts.ContentType == "" { | ||
59 | if opts.ContentType = mime.TypeByExtension(filepath.Ext(filePath)); opts.ContentType == "" { | ||
60 | opts.ContentType = "application/octet-stream" | ||
61 | } | ||
62 | } | ||
63 | return c.PutObject(ctx, bucketName, objectName, fileReader, fileSize, opts) | ||
64 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/api-put-object-multipart.go b/vendor/github.com/minio/minio-go/v7/api-put-object-multipart.go deleted file mode 100644 index 5f117af..0000000 --- a/vendor/github.com/minio/minio-go/v7/api-put-object-multipart.go +++ /dev/null | |||
@@ -1,465 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2015-2017 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package minio | ||
19 | |||
20 | import ( | ||
21 | "bytes" | ||
22 | "context" | ||
23 | "encoding/base64" | ||
24 | "encoding/hex" | ||
25 | "encoding/xml" | ||
26 | "fmt" | ||
27 | "hash/crc32" | ||
28 | "io" | ||
29 | "net/http" | ||
30 | "net/url" | ||
31 | "sort" | ||
32 | "strconv" | ||
33 | "strings" | ||
34 | |||
35 | "github.com/google/uuid" | ||
36 | "github.com/minio/minio-go/v7/pkg/encrypt" | ||
37 | "github.com/minio/minio-go/v7/pkg/s3utils" | ||
38 | ) | ||
39 | |||
40 | func (c *Client) putObjectMultipart(ctx context.Context, bucketName, objectName string, reader io.Reader, size int64, | ||
41 | opts PutObjectOptions, | ||
42 | ) (info UploadInfo, err error) { | ||
43 | info, err = c.putObjectMultipartNoStream(ctx, bucketName, objectName, reader, opts) | ||
44 | if err != nil { | ||
45 | errResp := ToErrorResponse(err) | ||
46 | // Verify if multipart functionality is not available, if not | ||
47 | // fall back to single PutObject operation. | ||
48 | if errResp.Code == "AccessDenied" && strings.Contains(errResp.Message, "Access Denied") { | ||
49 | // Verify if size of reader is greater than '5GiB'. | ||
50 | if size > maxSinglePutObjectSize { | ||
51 | return UploadInfo{}, errEntityTooLarge(size, maxSinglePutObjectSize, bucketName, objectName) | ||
52 | } | ||
53 | // Fall back to uploading as single PutObject operation. | ||
54 | return c.putObject(ctx, bucketName, objectName, reader, size, opts) | ||
55 | } | ||
56 | } | ||
57 | return info, err | ||
58 | } | ||
59 | |||
60 | func (c *Client) putObjectMultipartNoStream(ctx context.Context, bucketName, objectName string, reader io.Reader, opts PutObjectOptions) (info UploadInfo, err error) { | ||
61 | // Input validation. | ||
62 | if err = s3utils.CheckValidBucketName(bucketName); err != nil { | ||
63 | return UploadInfo{}, err | ||
64 | } | ||
65 | if err = s3utils.CheckValidObjectName(objectName); err != nil { | ||
66 | return UploadInfo{}, err | ||
67 | } | ||
68 | |||
69 | // Total data read and written to server. should be equal to | ||
70 | // 'size' at the end of the call. | ||
71 | var totalUploadedSize int64 | ||
72 | |||
73 | // Complete multipart upload. | ||
74 | var complMultipartUpload completeMultipartUpload | ||
75 | |||
76 | // Calculate the optimal parts info for a given size. | ||
77 | totalPartsCount, partSize, _, err := OptimalPartInfo(-1, opts.PartSize) | ||
78 | if err != nil { | ||
79 | return UploadInfo{}, err | ||
80 | } | ||
81 | |||
82 | // Choose hash algorithms to be calculated by hashCopyN, | ||
83 | // avoid sha256 with non-v4 signature request or | ||
84 | // HTTPS connection. | ||
85 | hashAlgos, hashSums := c.hashMaterials(opts.SendContentMd5, !opts.DisableContentSha256) | ||
86 | if len(hashSums) == 0 { | ||
87 | if opts.UserMetadata == nil { | ||
88 | opts.UserMetadata = make(map[string]string, 1) | ||
89 | } | ||
90 | opts.UserMetadata["X-Amz-Checksum-Algorithm"] = "CRC32C" | ||
91 | } | ||
92 | |||
93 | // Initiate a new multipart upload. | ||
94 | uploadID, err := c.newUploadID(ctx, bucketName, objectName, opts) | ||
95 | if err != nil { | ||
96 | return UploadInfo{}, err | ||
97 | } | ||
98 | delete(opts.UserMetadata, "X-Amz-Checksum-Algorithm") | ||
99 | |||
100 | defer func() { | ||
101 | if err != nil { | ||
102 | c.abortMultipartUpload(ctx, bucketName, objectName, uploadID) | ||
103 | } | ||
104 | }() | ||
105 | |||
106 | // Part number always starts with '1'. | ||
107 | partNumber := 1 | ||
108 | |||
109 | // Initialize parts uploaded map. | ||
110 | partsInfo := make(map[int]ObjectPart) | ||
111 | |||
112 | // Create a buffer. | ||
113 | buf := make([]byte, partSize) | ||
114 | |||
115 | // Create checksums | ||
116 | // CRC32C is ~50% faster on AMD64 @ 30GB/s | ||
117 | var crcBytes []byte | ||
118 | customHeader := make(http.Header) | ||
119 | crc := crc32.New(crc32.MakeTable(crc32.Castagnoli)) | ||
120 | for partNumber <= totalPartsCount { | ||
121 | length, rErr := readFull(reader, buf) | ||
122 | if rErr == io.EOF && partNumber > 1 { | ||
123 | break | ||
124 | } | ||
125 | |||
126 | if rErr != nil && rErr != io.ErrUnexpectedEOF && rErr != io.EOF { | ||
127 | return UploadInfo{}, rErr | ||
128 | } | ||
129 | |||
130 | // Calculates hash sums while copying partSize bytes into cw. | ||
131 | for k, v := range hashAlgos { | ||
132 | v.Write(buf[:length]) | ||
133 | hashSums[k] = v.Sum(nil) | ||
134 | v.Close() | ||
135 | } | ||
136 | |||
137 | // Update progress reader appropriately to the latest offset | ||
138 | // as we read from the source. | ||
139 | rd := newHook(bytes.NewReader(buf[:length]), opts.Progress) | ||
140 | |||
141 | // Checksums.. | ||
142 | var ( | ||
143 | md5Base64 string | ||
144 | sha256Hex string | ||
145 | ) | ||
146 | |||
147 | if hashSums["md5"] != nil { | ||
148 | md5Base64 = base64.StdEncoding.EncodeToString(hashSums["md5"]) | ||
149 | } | ||
150 | if hashSums["sha256"] != nil { | ||
151 | sha256Hex = hex.EncodeToString(hashSums["sha256"]) | ||
152 | } | ||
153 | if len(hashSums) == 0 { | ||
154 | crc.Reset() | ||
155 | crc.Write(buf[:length]) | ||
156 | cSum := crc.Sum(nil) | ||
157 | customHeader.Set("x-amz-checksum-crc32c", base64.StdEncoding.EncodeToString(cSum)) | ||
158 | crcBytes = append(crcBytes, cSum...) | ||
159 | } | ||
160 | |||
161 | p := uploadPartParams{bucketName: bucketName, objectName: objectName, uploadID: uploadID, reader: rd, partNumber: partNumber, md5Base64: md5Base64, sha256Hex: sha256Hex, size: int64(length), sse: opts.ServerSideEncryption, streamSha256: !opts.DisableContentSha256, customHeader: customHeader} | ||
162 | // Proceed to upload the part. | ||
163 | objPart, uerr := c.uploadPart(ctx, p) | ||
164 | if uerr != nil { | ||
165 | return UploadInfo{}, uerr | ||
166 | } | ||
167 | |||
168 | // Save successfully uploaded part metadata. | ||
169 | partsInfo[partNumber] = objPart | ||
170 | |||
171 | // Save successfully uploaded size. | ||
172 | totalUploadedSize += int64(length) | ||
173 | |||
174 | // Increment part number. | ||
175 | partNumber++ | ||
176 | |||
177 | // For unknown size, Read EOF we break away. | ||
178 | // We do not have to upload till totalPartsCount. | ||
179 | if rErr == io.EOF { | ||
180 | break | ||
181 | } | ||
182 | } | ||
183 | |||
184 | // Loop over total uploaded parts to save them in | ||
185 | // Parts array before completing the multipart request. | ||
186 | for i := 1; i < partNumber; i++ { | ||
187 | part, ok := partsInfo[i] | ||
188 | if !ok { | ||
189 | return UploadInfo{}, errInvalidArgument(fmt.Sprintf("Missing part number %d", i)) | ||
190 | } | ||
191 | complMultipartUpload.Parts = append(complMultipartUpload.Parts, CompletePart{ | ||
192 | ETag: part.ETag, | ||
193 | PartNumber: part.PartNumber, | ||
194 | ChecksumCRC32: part.ChecksumCRC32, | ||
195 | ChecksumCRC32C: part.ChecksumCRC32C, | ||
196 | ChecksumSHA1: part.ChecksumSHA1, | ||
197 | ChecksumSHA256: part.ChecksumSHA256, | ||
198 | }) | ||
199 | } | ||
200 | |||
201 | // Sort all completed parts. | ||
202 | sort.Sort(completedParts(complMultipartUpload.Parts)) | ||
203 | opts = PutObjectOptions{ | ||
204 | ServerSideEncryption: opts.ServerSideEncryption, | ||
205 | } | ||
206 | if len(crcBytes) > 0 { | ||
207 | // Add hash of hashes. | ||
208 | crc.Reset() | ||
209 | crc.Write(crcBytes) | ||
210 | opts.UserMetadata = map[string]string{"X-Amz-Checksum-Crc32c": base64.StdEncoding.EncodeToString(crc.Sum(nil))} | ||
211 | } | ||
212 | uploadInfo, err := c.completeMultipartUpload(ctx, bucketName, objectName, uploadID, complMultipartUpload, opts) | ||
213 | if err != nil { | ||
214 | return UploadInfo{}, err | ||
215 | } | ||
216 | |||
217 | uploadInfo.Size = totalUploadedSize | ||
218 | return uploadInfo, nil | ||
219 | } | ||
220 | |||
221 | // initiateMultipartUpload - Initiates a multipart upload and returns an upload ID. | ||
222 | func (c *Client) initiateMultipartUpload(ctx context.Context, bucketName, objectName string, opts PutObjectOptions) (initiateMultipartUploadResult, error) { | ||
223 | // Input validation. | ||
224 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
225 | return initiateMultipartUploadResult{}, err | ||
226 | } | ||
227 | if err := s3utils.CheckValidObjectName(objectName); err != nil { | ||
228 | return initiateMultipartUploadResult{}, err | ||
229 | } | ||
230 | |||
231 | // Initialize url queries. | ||
232 | urlValues := make(url.Values) | ||
233 | urlValues.Set("uploads", "") | ||
234 | |||
235 | if opts.Internal.SourceVersionID != "" { | ||
236 | if opts.Internal.SourceVersionID != nullVersionID { | ||
237 | if _, err := uuid.Parse(opts.Internal.SourceVersionID); err != nil { | ||
238 | return initiateMultipartUploadResult{}, errInvalidArgument(err.Error()) | ||
239 | } | ||
240 | } | ||
241 | urlValues.Set("versionId", opts.Internal.SourceVersionID) | ||
242 | } | ||
243 | |||
244 | // Set ContentType header. | ||
245 | customHeader := opts.Header() | ||
246 | |||
247 | reqMetadata := requestMetadata{ | ||
248 | bucketName: bucketName, | ||
249 | objectName: objectName, | ||
250 | queryValues: urlValues, | ||
251 | customHeader: customHeader, | ||
252 | } | ||
253 | |||
254 | // Execute POST on an objectName to initiate multipart upload. | ||
255 | resp, err := c.executeMethod(ctx, http.MethodPost, reqMetadata) | ||
256 | defer closeResponse(resp) | ||
257 | if err != nil { | ||
258 | return initiateMultipartUploadResult{}, err | ||
259 | } | ||
260 | if resp != nil { | ||
261 | if resp.StatusCode != http.StatusOK { | ||
262 | return initiateMultipartUploadResult{}, httpRespToErrorResponse(resp, bucketName, objectName) | ||
263 | } | ||
264 | } | ||
265 | // Decode xml for new multipart upload. | ||
266 | initiateMultipartUploadResult := initiateMultipartUploadResult{} | ||
267 | err = xmlDecoder(resp.Body, &initiateMultipartUploadResult) | ||
268 | if err != nil { | ||
269 | return initiateMultipartUploadResult, err | ||
270 | } | ||
271 | return initiateMultipartUploadResult, nil | ||
272 | } | ||
273 | |||
274 | type uploadPartParams struct { | ||
275 | bucketName string | ||
276 | objectName string | ||
277 | uploadID string | ||
278 | reader io.Reader | ||
279 | partNumber int | ||
280 | md5Base64 string | ||
281 | sha256Hex string | ||
282 | size int64 | ||
283 | sse encrypt.ServerSide | ||
284 | streamSha256 bool | ||
285 | customHeader http.Header | ||
286 | trailer http.Header | ||
287 | } | ||
288 | |||
289 | // uploadPart - Uploads a part in a multipart upload. | ||
290 | func (c *Client) uploadPart(ctx context.Context, p uploadPartParams) (ObjectPart, error) { | ||
291 | // Input validation. | ||
292 | if err := s3utils.CheckValidBucketName(p.bucketName); err != nil { | ||
293 | return ObjectPart{}, err | ||
294 | } | ||
295 | if err := s3utils.CheckValidObjectName(p.objectName); err != nil { | ||
296 | return ObjectPart{}, err | ||
297 | } | ||
298 | if p.size > maxPartSize { | ||
299 | return ObjectPart{}, errEntityTooLarge(p.size, maxPartSize, p.bucketName, p.objectName) | ||
300 | } | ||
301 | if p.size <= -1 { | ||
302 | return ObjectPart{}, errEntityTooSmall(p.size, p.bucketName, p.objectName) | ||
303 | } | ||
304 | if p.partNumber <= 0 { | ||
305 | return ObjectPart{}, errInvalidArgument("Part number cannot be negative or equal to zero.") | ||
306 | } | ||
307 | if p.uploadID == "" { | ||
308 | return ObjectPart{}, errInvalidArgument("UploadID cannot be empty.") | ||
309 | } | ||
310 | |||
311 | // Get resources properly escaped and lined up before using them in http request. | ||
312 | urlValues := make(url.Values) | ||
313 | // Set part number. | ||
314 | urlValues.Set("partNumber", strconv.Itoa(p.partNumber)) | ||
315 | // Set upload id. | ||
316 | urlValues.Set("uploadId", p.uploadID) | ||
317 | |||
318 | // Set encryption headers, if any. | ||
319 | if p.customHeader == nil { | ||
320 | p.customHeader = make(http.Header) | ||
321 | } | ||
322 | // https://docs.aws.amazon.com/AmazonS3/latest/API/mpUploadUploadPart.html | ||
323 | // Server-side encryption is supported by the S3 Multipart Upload actions. | ||
324 | // Unless you are using a customer-provided encryption key, you don't need | ||
325 | // to specify the encryption parameters in each UploadPart request. | ||
326 | if p.sse != nil && p.sse.Type() == encrypt.SSEC { | ||
327 | p.sse.Marshal(p.customHeader) | ||
328 | } | ||
329 | |||
330 | reqMetadata := requestMetadata{ | ||
331 | bucketName: p.bucketName, | ||
332 | objectName: p.objectName, | ||
333 | queryValues: urlValues, | ||
334 | customHeader: p.customHeader, | ||
335 | contentBody: p.reader, | ||
336 | contentLength: p.size, | ||
337 | contentMD5Base64: p.md5Base64, | ||
338 | contentSHA256Hex: p.sha256Hex, | ||
339 | streamSha256: p.streamSha256, | ||
340 | trailer: p.trailer, | ||
341 | } | ||
342 | |||
343 | // Execute PUT on each part. | ||
344 | resp, err := c.executeMethod(ctx, http.MethodPut, reqMetadata) | ||
345 | defer closeResponse(resp) | ||
346 | if err != nil { | ||
347 | return ObjectPart{}, err | ||
348 | } | ||
349 | if resp != nil { | ||
350 | if resp.StatusCode != http.StatusOK { | ||
351 | return ObjectPart{}, httpRespToErrorResponse(resp, p.bucketName, p.objectName) | ||
352 | } | ||
353 | } | ||
354 | // Once successfully uploaded, return completed part. | ||
355 | h := resp.Header | ||
356 | objPart := ObjectPart{ | ||
357 | ChecksumCRC32: h.Get("x-amz-checksum-crc32"), | ||
358 | ChecksumCRC32C: h.Get("x-amz-checksum-crc32c"), | ||
359 | ChecksumSHA1: h.Get("x-amz-checksum-sha1"), | ||
360 | ChecksumSHA256: h.Get("x-amz-checksum-sha256"), | ||
361 | } | ||
362 | objPart.Size = p.size | ||
363 | objPart.PartNumber = p.partNumber | ||
364 | // Trim off the odd double quotes from ETag in the beginning and end. | ||
365 | objPart.ETag = trimEtag(h.Get("ETag")) | ||
366 | return objPart, nil | ||
367 | } | ||
368 | |||
369 | // completeMultipartUpload - Completes a multipart upload by assembling previously uploaded parts. | ||
370 | func (c *Client) completeMultipartUpload(ctx context.Context, bucketName, objectName, uploadID string, | ||
371 | complete completeMultipartUpload, opts PutObjectOptions, | ||
372 | ) (UploadInfo, error) { | ||
373 | // Input validation. | ||
374 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
375 | return UploadInfo{}, err | ||
376 | } | ||
377 | if err := s3utils.CheckValidObjectName(objectName); err != nil { | ||
378 | return UploadInfo{}, err | ||
379 | } | ||
380 | |||
381 | // Initialize url queries. | ||
382 | urlValues := make(url.Values) | ||
383 | urlValues.Set("uploadId", uploadID) | ||
384 | // Marshal complete multipart body. | ||
385 | completeMultipartUploadBytes, err := xml.Marshal(complete) | ||
386 | if err != nil { | ||
387 | return UploadInfo{}, err | ||
388 | } | ||
389 | |||
390 | headers := opts.Header() | ||
391 | if s3utils.IsAmazonEndpoint(*c.endpointURL) { | ||
392 | headers.Del(encrypt.SseKmsKeyID) // Remove X-Amz-Server-Side-Encryption-Aws-Kms-Key-Id not supported in CompleteMultipartUpload | ||
393 | headers.Del(encrypt.SseGenericHeader) // Remove X-Amz-Server-Side-Encryption not supported in CompleteMultipartUpload | ||
394 | headers.Del(encrypt.SseEncryptionContext) // Remove X-Amz-Server-Side-Encryption-Context not supported in CompleteMultipartUpload | ||
395 | } | ||
396 | |||
397 | // Instantiate all the complete multipart buffer. | ||
398 | completeMultipartUploadBuffer := bytes.NewReader(completeMultipartUploadBytes) | ||
399 | reqMetadata := requestMetadata{ | ||
400 | bucketName: bucketName, | ||
401 | objectName: objectName, | ||
402 | queryValues: urlValues, | ||
403 | contentBody: completeMultipartUploadBuffer, | ||
404 | contentLength: int64(len(completeMultipartUploadBytes)), | ||
405 | contentSHA256Hex: sum256Hex(completeMultipartUploadBytes), | ||
406 | customHeader: headers, | ||
407 | } | ||
408 | |||
409 | // Execute POST to complete multipart upload for an objectName. | ||
410 | resp, err := c.executeMethod(ctx, http.MethodPost, reqMetadata) | ||
411 | defer closeResponse(resp) | ||
412 | if err != nil { | ||
413 | return UploadInfo{}, err | ||
414 | } | ||
415 | if resp != nil { | ||
416 | if resp.StatusCode != http.StatusOK { | ||
417 | return UploadInfo{}, httpRespToErrorResponse(resp, bucketName, objectName) | ||
418 | } | ||
419 | } | ||
420 | |||
421 | // Read resp.Body into a []bytes to parse for Error response inside the body | ||
422 | var b []byte | ||
423 | b, err = io.ReadAll(resp.Body) | ||
424 | if err != nil { | ||
425 | return UploadInfo{}, err | ||
426 | } | ||
427 | // Decode completed multipart upload response on success. | ||
428 | completeMultipartUploadResult := completeMultipartUploadResult{} | ||
429 | err = xmlDecoder(bytes.NewReader(b), &completeMultipartUploadResult) | ||
430 | if err != nil { | ||
431 | // xml parsing failure due to presence an ill-formed xml fragment | ||
432 | return UploadInfo{}, err | ||
433 | } else if completeMultipartUploadResult.Bucket == "" { | ||
434 | // xml's Decode method ignores well-formed xml that don't apply to the type of value supplied. | ||
435 | // In this case, it would leave completeMultipartUploadResult with the corresponding zero-values | ||
436 | // of the members. | ||
437 | |||
438 | // Decode completed multipart upload response on failure | ||
439 | completeMultipartUploadErr := ErrorResponse{} | ||
440 | err = xmlDecoder(bytes.NewReader(b), &completeMultipartUploadErr) | ||
441 | if err != nil { | ||
442 | // xml parsing failure due to presence an ill-formed xml fragment | ||
443 | return UploadInfo{}, err | ||
444 | } | ||
445 | return UploadInfo{}, completeMultipartUploadErr | ||
446 | } | ||
447 | |||
448 | // extract lifecycle expiry date and rule ID | ||
449 | expTime, ruleID := amzExpirationToExpiryDateRuleID(resp.Header.Get(amzExpiration)) | ||
450 | |||
451 | return UploadInfo{ | ||
452 | Bucket: completeMultipartUploadResult.Bucket, | ||
453 | Key: completeMultipartUploadResult.Key, | ||
454 | ETag: trimEtag(completeMultipartUploadResult.ETag), | ||
455 | VersionID: resp.Header.Get(amzVersionID), | ||
456 | Location: completeMultipartUploadResult.Location, | ||
457 | Expiration: expTime, | ||
458 | ExpirationRuleID: ruleID, | ||
459 | |||
460 | ChecksumSHA256: completeMultipartUploadResult.ChecksumSHA256, | ||
461 | ChecksumSHA1: completeMultipartUploadResult.ChecksumSHA1, | ||
462 | ChecksumCRC32: completeMultipartUploadResult.ChecksumCRC32, | ||
463 | ChecksumCRC32C: completeMultipartUploadResult.ChecksumCRC32C, | ||
464 | }, nil | ||
465 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/api-put-object-streaming.go b/vendor/github.com/minio/minio-go/v7/api-put-object-streaming.go deleted file mode 100644 index 9182d4e..0000000 --- a/vendor/github.com/minio/minio-go/v7/api-put-object-streaming.go +++ /dev/null | |||
@@ -1,809 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2017 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package minio | ||
19 | |||
20 | import ( | ||
21 | "bytes" | ||
22 | "context" | ||
23 | "encoding/base64" | ||
24 | "fmt" | ||
25 | "hash/crc32" | ||
26 | "io" | ||
27 | "net/http" | ||
28 | "net/url" | ||
29 | "sort" | ||
30 | "strings" | ||
31 | "sync" | ||
32 | |||
33 | "github.com/google/uuid" | ||
34 | "github.com/minio/minio-go/v7/pkg/s3utils" | ||
35 | ) | ||
36 | |||
37 | // putObjectMultipartStream - upload a large object using | ||
38 | // multipart upload and streaming signature for signing payload. | ||
39 | // Comprehensive put object operation involving multipart uploads. | ||
40 | // | ||
41 | // Following code handles these types of readers. | ||
42 | // | ||
43 | // - *minio.Object | ||
44 | // - Any reader which has a method 'ReadAt()' | ||
45 | func (c *Client) putObjectMultipartStream(ctx context.Context, bucketName, objectName string, | ||
46 | reader io.Reader, size int64, opts PutObjectOptions, | ||
47 | ) (info UploadInfo, err error) { | ||
48 | if opts.ConcurrentStreamParts && opts.NumThreads > 1 { | ||
49 | info, err = c.putObjectMultipartStreamParallel(ctx, bucketName, objectName, reader, opts) | ||
50 | } else if !isObject(reader) && isReadAt(reader) && !opts.SendContentMd5 { | ||
51 | // Verify if the reader implements ReadAt and it is not a *minio.Object then we will use parallel uploader. | ||
52 | info, err = c.putObjectMultipartStreamFromReadAt(ctx, bucketName, objectName, reader.(io.ReaderAt), size, opts) | ||
53 | } else { | ||
54 | info, err = c.putObjectMultipartStreamOptionalChecksum(ctx, bucketName, objectName, reader, size, opts) | ||
55 | } | ||
56 | if err != nil { | ||
57 | errResp := ToErrorResponse(err) | ||
58 | // Verify if multipart functionality is not available, if not | ||
59 | // fall back to single PutObject operation. | ||
60 | if errResp.Code == "AccessDenied" && strings.Contains(errResp.Message, "Access Denied") { | ||
61 | // Verify if size of reader is greater than '5GiB'. | ||
62 | if size > maxSinglePutObjectSize { | ||
63 | return UploadInfo{}, errEntityTooLarge(size, maxSinglePutObjectSize, bucketName, objectName) | ||
64 | } | ||
65 | // Fall back to uploading as single PutObject operation. | ||
66 | return c.putObject(ctx, bucketName, objectName, reader, size, opts) | ||
67 | } | ||
68 | } | ||
69 | return info, err | ||
70 | } | ||
71 | |||
72 | // uploadedPartRes - the response received from a part upload. | ||
73 | type uploadedPartRes struct { | ||
74 | Error error // Any error encountered while uploading the part. | ||
75 | PartNum int // Number of the part uploaded. | ||
76 | Size int64 // Size of the part uploaded. | ||
77 | Part ObjectPart | ||
78 | } | ||
79 | |||
80 | type uploadPartReq struct { | ||
81 | PartNum int // Number of the part uploaded. | ||
82 | Part ObjectPart // Size of the part uploaded. | ||
83 | } | ||
84 | |||
85 | // putObjectMultipartFromReadAt - Uploads files bigger than 128MiB. | ||
86 | // Supports all readers which implements io.ReaderAt interface | ||
87 | // (ReadAt method). | ||
88 | // | ||
89 | // NOTE: This function is meant to be used for all readers which | ||
90 | // implement io.ReaderAt which allows us for resuming multipart | ||
91 | // uploads but reading at an offset, which would avoid re-read the | ||
92 | // data which was already uploaded. Internally this function uses | ||
93 | // temporary files for staging all the data, these temporary files are | ||
94 | // cleaned automatically when the caller i.e http client closes the | ||
95 | // stream after uploading all the contents successfully. | ||
96 | func (c *Client) putObjectMultipartStreamFromReadAt(ctx context.Context, bucketName, objectName string, | ||
97 | reader io.ReaderAt, size int64, opts PutObjectOptions, | ||
98 | ) (info UploadInfo, err error) { | ||
99 | // Input validation. | ||
100 | if err = s3utils.CheckValidBucketName(bucketName); err != nil { | ||
101 | return UploadInfo{}, err | ||
102 | } | ||
103 | if err = s3utils.CheckValidObjectName(objectName); err != nil { | ||
104 | return UploadInfo{}, err | ||
105 | } | ||
106 | |||
107 | // Calculate the optimal parts info for a given size. | ||
108 | totalPartsCount, partSize, lastPartSize, err := OptimalPartInfo(size, opts.PartSize) | ||
109 | if err != nil { | ||
110 | return UploadInfo{}, err | ||
111 | } | ||
112 | |||
113 | withChecksum := c.trailingHeaderSupport | ||
114 | if withChecksum { | ||
115 | if opts.UserMetadata == nil { | ||
116 | opts.UserMetadata = make(map[string]string, 1) | ||
117 | } | ||
118 | opts.UserMetadata["X-Amz-Checksum-Algorithm"] = "CRC32C" | ||
119 | } | ||
120 | // Initiate a new multipart upload. | ||
121 | uploadID, err := c.newUploadID(ctx, bucketName, objectName, opts) | ||
122 | if err != nil { | ||
123 | return UploadInfo{}, err | ||
124 | } | ||
125 | delete(opts.UserMetadata, "X-Amz-Checksum-Algorithm") | ||
126 | |||
127 | // Aborts the multipart upload in progress, if the | ||
128 | // function returns any error, since we do not resume | ||
129 | // we should purge the parts which have been uploaded | ||
130 | // to relinquish storage space. | ||
131 | defer func() { | ||
132 | if err != nil { | ||
133 | c.abortMultipartUpload(ctx, bucketName, objectName, uploadID) | ||
134 | } | ||
135 | }() | ||
136 | |||
137 | // Total data read and written to server. should be equal to 'size' at the end of the call. | ||
138 | var totalUploadedSize int64 | ||
139 | |||
140 | // Complete multipart upload. | ||
141 | var complMultipartUpload completeMultipartUpload | ||
142 | |||
143 | // Declare a channel that sends the next part number to be uploaded. | ||
144 | uploadPartsCh := make(chan uploadPartReq) | ||
145 | |||
146 | // Declare a channel that sends back the response of a part upload. | ||
147 | uploadedPartsCh := make(chan uploadedPartRes) | ||
148 | |||
149 | // Used for readability, lastPartNumber is always totalPartsCount. | ||
150 | lastPartNumber := totalPartsCount | ||
151 | |||
152 | partitionCtx, partitionCancel := context.WithCancel(ctx) | ||
153 | defer partitionCancel() | ||
154 | // Send each part number to the channel to be processed. | ||
155 | go func() { | ||
156 | defer close(uploadPartsCh) | ||
157 | |||
158 | for p := 1; p <= totalPartsCount; p++ { | ||
159 | select { | ||
160 | case <-partitionCtx.Done(): | ||
161 | return | ||
162 | case uploadPartsCh <- uploadPartReq{PartNum: p}: | ||
163 | } | ||
164 | } | ||
165 | }() | ||
166 | |||
167 | // Receive each part number from the channel allowing three parallel uploads. | ||
168 | for w := 1; w <= opts.getNumThreads(); w++ { | ||
169 | go func(partSize int64) { | ||
170 | for { | ||
171 | var uploadReq uploadPartReq | ||
172 | var ok bool | ||
173 | select { | ||
174 | case <-ctx.Done(): | ||
175 | return | ||
176 | case uploadReq, ok = <-uploadPartsCh: | ||
177 | if !ok { | ||
178 | return | ||
179 | } | ||
180 | // Each worker will draw from the part channel and upload in parallel. | ||
181 | } | ||
182 | |||
183 | // If partNumber was not uploaded we calculate the missing | ||
184 | // part offset and size. For all other part numbers we | ||
185 | // calculate offset based on multiples of partSize. | ||
186 | readOffset := int64(uploadReq.PartNum-1) * partSize | ||
187 | |||
188 | // As a special case if partNumber is lastPartNumber, we | ||
189 | // calculate the offset based on the last part size. | ||
190 | if uploadReq.PartNum == lastPartNumber { | ||
191 | readOffset = size - lastPartSize | ||
192 | partSize = lastPartSize | ||
193 | } | ||
194 | |||
195 | sectionReader := newHook(io.NewSectionReader(reader, readOffset, partSize), opts.Progress) | ||
196 | trailer := make(http.Header, 1) | ||
197 | if withChecksum { | ||
198 | crc := crc32.New(crc32.MakeTable(crc32.Castagnoli)) | ||
199 | trailer.Set("x-amz-checksum-crc32c", base64.StdEncoding.EncodeToString(crc.Sum(nil))) | ||
200 | sectionReader = newHashReaderWrapper(sectionReader, crc, func(hash []byte) { | ||
201 | trailer.Set("x-amz-checksum-crc32c", base64.StdEncoding.EncodeToString(hash)) | ||
202 | }) | ||
203 | } | ||
204 | |||
205 | // Proceed to upload the part. | ||
206 | p := uploadPartParams{ | ||
207 | bucketName: bucketName, | ||
208 | objectName: objectName, | ||
209 | uploadID: uploadID, | ||
210 | reader: sectionReader, | ||
211 | partNumber: uploadReq.PartNum, | ||
212 | size: partSize, | ||
213 | sse: opts.ServerSideEncryption, | ||
214 | streamSha256: !opts.DisableContentSha256, | ||
215 | sha256Hex: "", | ||
216 | trailer: trailer, | ||
217 | } | ||
218 | objPart, err := c.uploadPart(ctx, p) | ||
219 | if err != nil { | ||
220 | uploadedPartsCh <- uploadedPartRes{ | ||
221 | Error: err, | ||
222 | } | ||
223 | // Exit the goroutine. | ||
224 | return | ||
225 | } | ||
226 | |||
227 | // Save successfully uploaded part metadata. | ||
228 | uploadReq.Part = objPart | ||
229 | |||
230 | // Send successful part info through the channel. | ||
231 | uploadedPartsCh <- uploadedPartRes{ | ||
232 | Size: objPart.Size, | ||
233 | PartNum: uploadReq.PartNum, | ||
234 | Part: uploadReq.Part, | ||
235 | } | ||
236 | } | ||
237 | }(partSize) | ||
238 | } | ||
239 | |||
240 | // Gather the responses as they occur and update any | ||
241 | // progress bar. | ||
242 | for u := 1; u <= totalPartsCount; u++ { | ||
243 | select { | ||
244 | case <-ctx.Done(): | ||
245 | return UploadInfo{}, ctx.Err() | ||
246 | case uploadRes := <-uploadedPartsCh: | ||
247 | if uploadRes.Error != nil { | ||
248 | return UploadInfo{}, uploadRes.Error | ||
249 | } | ||
250 | |||
251 | // Update the totalUploadedSize. | ||
252 | totalUploadedSize += uploadRes.Size | ||
253 | complMultipartUpload.Parts = append(complMultipartUpload.Parts, CompletePart{ | ||
254 | ETag: uploadRes.Part.ETag, | ||
255 | PartNumber: uploadRes.Part.PartNumber, | ||
256 | ChecksumCRC32: uploadRes.Part.ChecksumCRC32, | ||
257 | ChecksumCRC32C: uploadRes.Part.ChecksumCRC32C, | ||
258 | ChecksumSHA1: uploadRes.Part.ChecksumSHA1, | ||
259 | ChecksumSHA256: uploadRes.Part.ChecksumSHA256, | ||
260 | }) | ||
261 | } | ||
262 | } | ||
263 | |||
264 | // Verify if we uploaded all the data. | ||
265 | if totalUploadedSize != size { | ||
266 | return UploadInfo{}, errUnexpectedEOF(totalUploadedSize, size, bucketName, objectName) | ||
267 | } | ||
268 | |||
269 | // Sort all completed parts. | ||
270 | sort.Sort(completedParts(complMultipartUpload.Parts)) | ||
271 | |||
272 | opts = PutObjectOptions{ | ||
273 | ServerSideEncryption: opts.ServerSideEncryption, | ||
274 | } | ||
275 | if withChecksum { | ||
276 | // Add hash of hashes. | ||
277 | crc := crc32.New(crc32.MakeTable(crc32.Castagnoli)) | ||
278 | for _, part := range complMultipartUpload.Parts { | ||
279 | cs, err := base64.StdEncoding.DecodeString(part.ChecksumCRC32C) | ||
280 | if err == nil { | ||
281 | crc.Write(cs) | ||
282 | } | ||
283 | } | ||
284 | opts.UserMetadata = map[string]string{"X-Amz-Checksum-Crc32c": base64.StdEncoding.EncodeToString(crc.Sum(nil))} | ||
285 | } | ||
286 | |||
287 | uploadInfo, err := c.completeMultipartUpload(ctx, bucketName, objectName, uploadID, complMultipartUpload, opts) | ||
288 | if err != nil { | ||
289 | return UploadInfo{}, err | ||
290 | } | ||
291 | |||
292 | uploadInfo.Size = totalUploadedSize | ||
293 | return uploadInfo, nil | ||
294 | } | ||
295 | |||
296 | func (c *Client) putObjectMultipartStreamOptionalChecksum(ctx context.Context, bucketName, objectName string, | ||
297 | reader io.Reader, size int64, opts PutObjectOptions, | ||
298 | ) (info UploadInfo, err error) { | ||
299 | // Input validation. | ||
300 | if err = s3utils.CheckValidBucketName(bucketName); err != nil { | ||
301 | return UploadInfo{}, err | ||
302 | } | ||
303 | if err = s3utils.CheckValidObjectName(objectName); err != nil { | ||
304 | return UploadInfo{}, err | ||
305 | } | ||
306 | |||
307 | if !opts.SendContentMd5 { | ||
308 | if opts.UserMetadata == nil { | ||
309 | opts.UserMetadata = make(map[string]string, 1) | ||
310 | } | ||
311 | opts.UserMetadata["X-Amz-Checksum-Algorithm"] = "CRC32C" | ||
312 | } | ||
313 | |||
314 | // Calculate the optimal parts info for a given size. | ||
315 | totalPartsCount, partSize, lastPartSize, err := OptimalPartInfo(size, opts.PartSize) | ||
316 | if err != nil { | ||
317 | return UploadInfo{}, err | ||
318 | } | ||
319 | // Initiates a new multipart request | ||
320 | uploadID, err := c.newUploadID(ctx, bucketName, objectName, opts) | ||
321 | if err != nil { | ||
322 | return UploadInfo{}, err | ||
323 | } | ||
324 | delete(opts.UserMetadata, "X-Amz-Checksum-Algorithm") | ||
325 | |||
326 | // Aborts the multipart upload if the function returns | ||
327 | // any error, since we do not resume we should purge | ||
328 | // the parts which have been uploaded to relinquish | ||
329 | // storage space. | ||
330 | defer func() { | ||
331 | if err != nil { | ||
332 | c.abortMultipartUpload(ctx, bucketName, objectName, uploadID) | ||
333 | } | ||
334 | }() | ||
335 | |||
336 | // Create checksums | ||
337 | // CRC32C is ~50% faster on AMD64 @ 30GB/s | ||
338 | var crcBytes []byte | ||
339 | customHeader := make(http.Header) | ||
340 | crc := crc32.New(crc32.MakeTable(crc32.Castagnoli)) | ||
341 | md5Hash := c.md5Hasher() | ||
342 | defer md5Hash.Close() | ||
343 | |||
344 | // Total data read and written to server. should be equal to 'size' at the end of the call. | ||
345 | var totalUploadedSize int64 | ||
346 | |||
347 | // Initialize parts uploaded map. | ||
348 | partsInfo := make(map[int]ObjectPart) | ||
349 | |||
350 | // Create a buffer. | ||
351 | buf := make([]byte, partSize) | ||
352 | |||
353 | // Avoid declaring variables in the for loop | ||
354 | var md5Base64 string | ||
355 | |||
356 | // Part number always starts with '1'. | ||
357 | var partNumber int | ||
358 | for partNumber = 1; partNumber <= totalPartsCount; partNumber++ { | ||
359 | |||
360 | // Proceed to upload the part. | ||
361 | if partNumber == totalPartsCount { | ||
362 | partSize = lastPartSize | ||
363 | } | ||
364 | |||
365 | length, rerr := readFull(reader, buf) | ||
366 | if rerr == io.EOF && partNumber > 1 { | ||
367 | break | ||
368 | } | ||
369 | |||
370 | if rerr != nil && rerr != io.ErrUnexpectedEOF && err != io.EOF { | ||
371 | return UploadInfo{}, rerr | ||
372 | } | ||
373 | |||
374 | // Calculate md5sum. | ||
375 | if opts.SendContentMd5 { | ||
376 | md5Hash.Reset() | ||
377 | md5Hash.Write(buf[:length]) | ||
378 | md5Base64 = base64.StdEncoding.EncodeToString(md5Hash.Sum(nil)) | ||
379 | } else { | ||
380 | // Add CRC32C instead. | ||
381 | crc.Reset() | ||
382 | crc.Write(buf[:length]) | ||
383 | cSum := crc.Sum(nil) | ||
384 | customHeader.Set("x-amz-checksum-crc32c", base64.StdEncoding.EncodeToString(cSum)) | ||
385 | crcBytes = append(crcBytes, cSum...) | ||
386 | } | ||
387 | |||
388 | // Update progress reader appropriately to the latest offset | ||
389 | // as we read from the source. | ||
390 | hooked := newHook(bytes.NewReader(buf[:length]), opts.Progress) | ||
391 | p := uploadPartParams{bucketName: bucketName, objectName: objectName, uploadID: uploadID, reader: hooked, partNumber: partNumber, md5Base64: md5Base64, size: partSize, sse: opts.ServerSideEncryption, streamSha256: !opts.DisableContentSha256, customHeader: customHeader} | ||
392 | objPart, uerr := c.uploadPart(ctx, p) | ||
393 | if uerr != nil { | ||
394 | return UploadInfo{}, uerr | ||
395 | } | ||
396 | |||
397 | // Save successfully uploaded part metadata. | ||
398 | partsInfo[partNumber] = objPart | ||
399 | |||
400 | // Save successfully uploaded size. | ||
401 | totalUploadedSize += partSize | ||
402 | } | ||
403 | |||
404 | // Verify if we uploaded all the data. | ||
405 | if size > 0 { | ||
406 | if totalUploadedSize != size { | ||
407 | return UploadInfo{}, errUnexpectedEOF(totalUploadedSize, size, bucketName, objectName) | ||
408 | } | ||
409 | } | ||
410 | |||
411 | // Complete multipart upload. | ||
412 | var complMultipartUpload completeMultipartUpload | ||
413 | |||
414 | // Loop over total uploaded parts to save them in | ||
415 | // Parts array before completing the multipart request. | ||
416 | for i := 1; i < partNumber; i++ { | ||
417 | part, ok := partsInfo[i] | ||
418 | if !ok { | ||
419 | return UploadInfo{}, errInvalidArgument(fmt.Sprintf("Missing part number %d", i)) | ||
420 | } | ||
421 | complMultipartUpload.Parts = append(complMultipartUpload.Parts, CompletePart{ | ||
422 | ETag: part.ETag, | ||
423 | PartNumber: part.PartNumber, | ||
424 | ChecksumCRC32: part.ChecksumCRC32, | ||
425 | ChecksumCRC32C: part.ChecksumCRC32C, | ||
426 | ChecksumSHA1: part.ChecksumSHA1, | ||
427 | ChecksumSHA256: part.ChecksumSHA256, | ||
428 | }) | ||
429 | } | ||
430 | |||
431 | // Sort all completed parts. | ||
432 | sort.Sort(completedParts(complMultipartUpload.Parts)) | ||
433 | |||
434 | opts = PutObjectOptions{ | ||
435 | ServerSideEncryption: opts.ServerSideEncryption, | ||
436 | } | ||
437 | if len(crcBytes) > 0 { | ||
438 | // Add hash of hashes. | ||
439 | crc.Reset() | ||
440 | crc.Write(crcBytes) | ||
441 | opts.UserMetadata = map[string]string{"X-Amz-Checksum-Crc32c": base64.StdEncoding.EncodeToString(crc.Sum(nil))} | ||
442 | } | ||
443 | uploadInfo, err := c.completeMultipartUpload(ctx, bucketName, objectName, uploadID, complMultipartUpload, opts) | ||
444 | if err != nil { | ||
445 | return UploadInfo{}, err | ||
446 | } | ||
447 | |||
448 | uploadInfo.Size = totalUploadedSize | ||
449 | return uploadInfo, nil | ||
450 | } | ||
451 | |||
452 | // putObjectMultipartStreamParallel uploads opts.NumThreads parts in parallel. | ||
453 | // This is expected to take opts.PartSize * opts.NumThreads * (GOGC / 100) bytes of buffer. | ||
454 | func (c *Client) putObjectMultipartStreamParallel(ctx context.Context, bucketName, objectName string, | ||
455 | reader io.Reader, opts PutObjectOptions, | ||
456 | ) (info UploadInfo, err error) { | ||
457 | // Input validation. | ||
458 | if err = s3utils.CheckValidBucketName(bucketName); err != nil { | ||
459 | return UploadInfo{}, err | ||
460 | } | ||
461 | |||
462 | if err = s3utils.CheckValidObjectName(objectName); err != nil { | ||
463 | return UploadInfo{}, err | ||
464 | } | ||
465 | |||
466 | if !opts.SendContentMd5 { | ||
467 | if opts.UserMetadata == nil { | ||
468 | opts.UserMetadata = make(map[string]string, 1) | ||
469 | } | ||
470 | opts.UserMetadata["X-Amz-Checksum-Algorithm"] = "CRC32C" | ||
471 | } | ||
472 | |||
473 | // Cancel all when an error occurs. | ||
474 | ctx, cancel := context.WithCancel(ctx) | ||
475 | defer cancel() | ||
476 | |||
477 | // Calculate the optimal parts info for a given size. | ||
478 | totalPartsCount, partSize, _, err := OptimalPartInfo(-1, opts.PartSize) | ||
479 | if err != nil { | ||
480 | return UploadInfo{}, err | ||
481 | } | ||
482 | |||
483 | // Initiates a new multipart request | ||
484 | uploadID, err := c.newUploadID(ctx, bucketName, objectName, opts) | ||
485 | if err != nil { | ||
486 | return UploadInfo{}, err | ||
487 | } | ||
488 | delete(opts.UserMetadata, "X-Amz-Checksum-Algorithm") | ||
489 | |||
490 | // Aborts the multipart upload if the function returns | ||
491 | // any error, since we do not resume we should purge | ||
492 | // the parts which have been uploaded to relinquish | ||
493 | // storage space. | ||
494 | defer func() { | ||
495 | if err != nil { | ||
496 | c.abortMultipartUpload(ctx, bucketName, objectName, uploadID) | ||
497 | } | ||
498 | }() | ||
499 | |||
500 | // Create checksums | ||
501 | // CRC32C is ~50% faster on AMD64 @ 30GB/s | ||
502 | var crcBytes []byte | ||
503 | crc := crc32.New(crc32.MakeTable(crc32.Castagnoli)) | ||
504 | |||
505 | // Total data read and written to server. should be equal to 'size' at the end of the call. | ||
506 | var totalUploadedSize int64 | ||
507 | |||
508 | // Initialize parts uploaded map. | ||
509 | partsInfo := make(map[int]ObjectPart) | ||
510 | |||
511 | // Create a buffer. | ||
512 | nBuffers := int64(opts.NumThreads) | ||
513 | bufs := make(chan []byte, nBuffers) | ||
514 | all := make([]byte, nBuffers*partSize) | ||
515 | for i := int64(0); i < nBuffers; i++ { | ||
516 | bufs <- all[i*partSize : i*partSize+partSize] | ||
517 | } | ||
518 | |||
519 | var wg sync.WaitGroup | ||
520 | var mu sync.Mutex | ||
521 | errCh := make(chan error, opts.NumThreads) | ||
522 | |||
523 | reader = newHook(reader, opts.Progress) | ||
524 | |||
525 | // Part number always starts with '1'. | ||
526 | var partNumber int | ||
527 | for partNumber = 1; partNumber <= totalPartsCount; partNumber++ { | ||
528 | // Proceed to upload the part. | ||
529 | var buf []byte | ||
530 | select { | ||
531 | case buf = <-bufs: | ||
532 | case err = <-errCh: | ||
533 | cancel() | ||
534 | wg.Wait() | ||
535 | return UploadInfo{}, err | ||
536 | } | ||
537 | |||
538 | if int64(len(buf)) != partSize { | ||
539 | return UploadInfo{}, fmt.Errorf("read buffer < %d than expected partSize: %d", len(buf), partSize) | ||
540 | } | ||
541 | |||
542 | length, rerr := readFull(reader, buf) | ||
543 | if rerr == io.EOF && partNumber > 1 { | ||
544 | // Done | ||
545 | break | ||
546 | } | ||
547 | |||
548 | if rerr != nil && rerr != io.ErrUnexpectedEOF && err != io.EOF { | ||
549 | cancel() | ||
550 | wg.Wait() | ||
551 | return UploadInfo{}, rerr | ||
552 | } | ||
553 | |||
554 | // Calculate md5sum. | ||
555 | customHeader := make(http.Header) | ||
556 | if !opts.SendContentMd5 { | ||
557 | // Add CRC32C instead. | ||
558 | crc.Reset() | ||
559 | crc.Write(buf[:length]) | ||
560 | cSum := crc.Sum(nil) | ||
561 | customHeader.Set("x-amz-checksum-crc32c", base64.StdEncoding.EncodeToString(cSum)) | ||
562 | crcBytes = append(crcBytes, cSum...) | ||
563 | } | ||
564 | |||
565 | wg.Add(1) | ||
566 | go func(partNumber int) { | ||
567 | // Avoid declaring variables in the for loop | ||
568 | var md5Base64 string | ||
569 | |||
570 | if opts.SendContentMd5 { | ||
571 | md5Hash := c.md5Hasher() | ||
572 | md5Hash.Write(buf[:length]) | ||
573 | md5Base64 = base64.StdEncoding.EncodeToString(md5Hash.Sum(nil)) | ||
574 | md5Hash.Close() | ||
575 | } | ||
576 | |||
577 | defer wg.Done() | ||
578 | p := uploadPartParams{ | ||
579 | bucketName: bucketName, | ||
580 | objectName: objectName, | ||
581 | uploadID: uploadID, | ||
582 | reader: bytes.NewReader(buf[:length]), | ||
583 | partNumber: partNumber, | ||
584 | md5Base64: md5Base64, | ||
585 | size: int64(length), | ||
586 | sse: opts.ServerSideEncryption, | ||
587 | streamSha256: !opts.DisableContentSha256, | ||
588 | customHeader: customHeader, | ||
589 | } | ||
590 | objPart, uerr := c.uploadPart(ctx, p) | ||
591 | if uerr != nil { | ||
592 | errCh <- uerr | ||
593 | return | ||
594 | } | ||
595 | |||
596 | // Save successfully uploaded part metadata. | ||
597 | mu.Lock() | ||
598 | partsInfo[partNumber] = objPart | ||
599 | mu.Unlock() | ||
600 | |||
601 | // Send buffer back so it can be reused. | ||
602 | bufs <- buf | ||
603 | }(partNumber) | ||
604 | |||
605 | // Save successfully uploaded size. | ||
606 | totalUploadedSize += int64(length) | ||
607 | } | ||
608 | wg.Wait() | ||
609 | |||
610 | // Collect any error | ||
611 | select { | ||
612 | case err = <-errCh: | ||
613 | return UploadInfo{}, err | ||
614 | default: | ||
615 | } | ||
616 | |||
617 | // Complete multipart upload. | ||
618 | var complMultipartUpload completeMultipartUpload | ||
619 | |||
620 | // Loop over total uploaded parts to save them in | ||
621 | // Parts array before completing the multipart request. | ||
622 | for i := 1; i < partNumber; i++ { | ||
623 | part, ok := partsInfo[i] | ||
624 | if !ok { | ||
625 | return UploadInfo{}, errInvalidArgument(fmt.Sprintf("Missing part number %d", i)) | ||
626 | } | ||
627 | complMultipartUpload.Parts = append(complMultipartUpload.Parts, CompletePart{ | ||
628 | ETag: part.ETag, | ||
629 | PartNumber: part.PartNumber, | ||
630 | ChecksumCRC32: part.ChecksumCRC32, | ||
631 | ChecksumCRC32C: part.ChecksumCRC32C, | ||
632 | ChecksumSHA1: part.ChecksumSHA1, | ||
633 | ChecksumSHA256: part.ChecksumSHA256, | ||
634 | }) | ||
635 | } | ||
636 | |||
637 | // Sort all completed parts. | ||
638 | sort.Sort(completedParts(complMultipartUpload.Parts)) | ||
639 | |||
640 | opts = PutObjectOptions{} | ||
641 | if len(crcBytes) > 0 { | ||
642 | // Add hash of hashes. | ||
643 | crc.Reset() | ||
644 | crc.Write(crcBytes) | ||
645 | opts.UserMetadata = map[string]string{"X-Amz-Checksum-Crc32c": base64.StdEncoding.EncodeToString(crc.Sum(nil))} | ||
646 | } | ||
647 | uploadInfo, err := c.completeMultipartUpload(ctx, bucketName, objectName, uploadID, complMultipartUpload, opts) | ||
648 | if err != nil { | ||
649 | return UploadInfo{}, err | ||
650 | } | ||
651 | |||
652 | uploadInfo.Size = totalUploadedSize | ||
653 | return uploadInfo, nil | ||
654 | } | ||
655 | |||
656 | // putObject special function used Google Cloud Storage. This special function | ||
657 | // is used for Google Cloud Storage since Google's multipart API is not S3 compatible. | ||
658 | func (c *Client) putObject(ctx context.Context, bucketName, objectName string, reader io.Reader, size int64, opts PutObjectOptions) (info UploadInfo, err error) { | ||
659 | // Input validation. | ||
660 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
661 | return UploadInfo{}, err | ||
662 | } | ||
663 | if err := s3utils.CheckValidObjectName(objectName); err != nil { | ||
664 | return UploadInfo{}, err | ||
665 | } | ||
666 | |||
667 | // Size -1 is only supported on Google Cloud Storage, we error | ||
668 | // out in all other situations. | ||
669 | if size < 0 && !s3utils.IsGoogleEndpoint(*c.endpointURL) { | ||
670 | return UploadInfo{}, errEntityTooSmall(size, bucketName, objectName) | ||
671 | } | ||
672 | |||
673 | if opts.SendContentMd5 && s3utils.IsGoogleEndpoint(*c.endpointURL) && size < 0 { | ||
674 | return UploadInfo{}, errInvalidArgument("MD5Sum cannot be calculated with size '-1'") | ||
675 | } | ||
676 | |||
677 | var readSeeker io.Seeker | ||
678 | if size > 0 { | ||
679 | if isReadAt(reader) && !isObject(reader) { | ||
680 | seeker, ok := reader.(io.Seeker) | ||
681 | if ok { | ||
682 | offset, err := seeker.Seek(0, io.SeekCurrent) | ||
683 | if err != nil { | ||
684 | return UploadInfo{}, errInvalidArgument(err.Error()) | ||
685 | } | ||
686 | reader = io.NewSectionReader(reader.(io.ReaderAt), offset, size) | ||
687 | readSeeker = reader.(io.Seeker) | ||
688 | } | ||
689 | } | ||
690 | } | ||
691 | |||
692 | var md5Base64 string | ||
693 | if opts.SendContentMd5 { | ||
694 | // Calculate md5sum. | ||
695 | hash := c.md5Hasher() | ||
696 | |||
697 | if readSeeker != nil { | ||
698 | if _, err := io.Copy(hash, reader); err != nil { | ||
699 | return UploadInfo{}, err | ||
700 | } | ||
701 | // Seek back to beginning of io.NewSectionReader's offset. | ||
702 | _, err = readSeeker.Seek(0, io.SeekStart) | ||
703 | if err != nil { | ||
704 | return UploadInfo{}, errInvalidArgument(err.Error()) | ||
705 | } | ||
706 | } else { | ||
707 | // Create a buffer. | ||
708 | buf := make([]byte, size) | ||
709 | |||
710 | length, err := readFull(reader, buf) | ||
711 | if err != nil && err != io.ErrUnexpectedEOF && err != io.EOF { | ||
712 | return UploadInfo{}, err | ||
713 | } | ||
714 | |||
715 | hash.Write(buf[:length]) | ||
716 | reader = bytes.NewReader(buf[:length]) | ||
717 | } | ||
718 | |||
719 | md5Base64 = base64.StdEncoding.EncodeToString(hash.Sum(nil)) | ||
720 | hash.Close() | ||
721 | } | ||
722 | |||
723 | // Update progress reader appropriately to the latest offset as we | ||
724 | // read from the source. | ||
725 | progressReader := newHook(reader, opts.Progress) | ||
726 | |||
727 | // This function does not calculate sha256 and md5sum for payload. | ||
728 | // Execute put object. | ||
729 | return c.putObjectDo(ctx, bucketName, objectName, progressReader, md5Base64, "", size, opts) | ||
730 | } | ||
731 | |||
732 | // putObjectDo - executes the put object http operation. | ||
733 | // NOTE: You must have WRITE permissions on a bucket to add an object to it. | ||
734 | func (c *Client) putObjectDo(ctx context.Context, bucketName, objectName string, reader io.Reader, md5Base64, sha256Hex string, size int64, opts PutObjectOptions) (UploadInfo, error) { | ||
735 | // Input validation. | ||
736 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
737 | return UploadInfo{}, err | ||
738 | } | ||
739 | if err := s3utils.CheckValidObjectName(objectName); err != nil { | ||
740 | return UploadInfo{}, err | ||
741 | } | ||
742 | // Set headers. | ||
743 | customHeader := opts.Header() | ||
744 | |||
745 | // Add CRC when client supports it, MD5 is not set, not Google and we don't add SHA256 to chunks. | ||
746 | addCrc := c.trailingHeaderSupport && md5Base64 == "" && !s3utils.IsGoogleEndpoint(*c.endpointURL) && (opts.DisableContentSha256 || c.secure) | ||
747 | |||
748 | if addCrc { | ||
749 | // If user has added checksums, don't add them ourselves. | ||
750 | for k := range opts.UserMetadata { | ||
751 | if strings.HasPrefix(strings.ToLower(k), "x-amz-checksum-") { | ||
752 | addCrc = false | ||
753 | } | ||
754 | } | ||
755 | } | ||
756 | // Populate request metadata. | ||
757 | reqMetadata := requestMetadata{ | ||
758 | bucketName: bucketName, | ||
759 | objectName: objectName, | ||
760 | customHeader: customHeader, | ||
761 | contentBody: reader, | ||
762 | contentLength: size, | ||
763 | contentMD5Base64: md5Base64, | ||
764 | contentSHA256Hex: sha256Hex, | ||
765 | streamSha256: !opts.DisableContentSha256, | ||
766 | addCrc: addCrc, | ||
767 | } | ||
768 | if opts.Internal.SourceVersionID != "" { | ||
769 | if opts.Internal.SourceVersionID != nullVersionID { | ||
770 | if _, err := uuid.Parse(opts.Internal.SourceVersionID); err != nil { | ||
771 | return UploadInfo{}, errInvalidArgument(err.Error()) | ||
772 | } | ||
773 | } | ||
774 | urlValues := make(url.Values) | ||
775 | urlValues.Set("versionId", opts.Internal.SourceVersionID) | ||
776 | reqMetadata.queryValues = urlValues | ||
777 | } | ||
778 | |||
779 | // Execute PUT an objectName. | ||
780 | resp, err := c.executeMethod(ctx, http.MethodPut, reqMetadata) | ||
781 | defer closeResponse(resp) | ||
782 | if err != nil { | ||
783 | return UploadInfo{}, err | ||
784 | } | ||
785 | if resp != nil { | ||
786 | if resp.StatusCode != http.StatusOK { | ||
787 | return UploadInfo{}, httpRespToErrorResponse(resp, bucketName, objectName) | ||
788 | } | ||
789 | } | ||
790 | |||
791 | // extract lifecycle expiry date and rule ID | ||
792 | expTime, ruleID := amzExpirationToExpiryDateRuleID(resp.Header.Get(amzExpiration)) | ||
793 | h := resp.Header | ||
794 | return UploadInfo{ | ||
795 | Bucket: bucketName, | ||
796 | Key: objectName, | ||
797 | ETag: trimEtag(h.Get("ETag")), | ||
798 | VersionID: h.Get(amzVersionID), | ||
799 | Size: size, | ||
800 | Expiration: expTime, | ||
801 | ExpirationRuleID: ruleID, | ||
802 | |||
803 | // Checksum values | ||
804 | ChecksumCRC32: h.Get("x-amz-checksum-crc32"), | ||
805 | ChecksumCRC32C: h.Get("x-amz-checksum-crc32c"), | ||
806 | ChecksumSHA1: h.Get("x-amz-checksum-sha1"), | ||
807 | ChecksumSHA256: h.Get("x-amz-checksum-sha256"), | ||
808 | }, nil | ||
809 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/api-put-object.go b/vendor/github.com/minio/minio-go/v7/api-put-object.go deleted file mode 100644 index bbd8924..0000000 --- a/vendor/github.com/minio/minio-go/v7/api-put-object.go +++ /dev/null | |||
@@ -1,473 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2015-2017 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package minio | ||
19 | |||
20 | import ( | ||
21 | "bytes" | ||
22 | "context" | ||
23 | "encoding/base64" | ||
24 | "errors" | ||
25 | "fmt" | ||
26 | "hash/crc32" | ||
27 | "io" | ||
28 | "net/http" | ||
29 | "sort" | ||
30 | "time" | ||
31 | |||
32 | "github.com/minio/minio-go/v7/pkg/encrypt" | ||
33 | "github.com/minio/minio-go/v7/pkg/s3utils" | ||
34 | "golang.org/x/net/http/httpguts" | ||
35 | ) | ||
36 | |||
37 | // ReplicationStatus represents replication status of object | ||
38 | type ReplicationStatus string | ||
39 | |||
40 | const ( | ||
41 | // ReplicationStatusPending indicates replication is pending | ||
42 | ReplicationStatusPending ReplicationStatus = "PENDING" | ||
43 | // ReplicationStatusComplete indicates replication completed ok | ||
44 | ReplicationStatusComplete ReplicationStatus = "COMPLETED" | ||
45 | // ReplicationStatusFailed indicates replication failed | ||
46 | ReplicationStatusFailed ReplicationStatus = "FAILED" | ||
47 | // ReplicationStatusReplica indicates object is a replica of a source | ||
48 | ReplicationStatusReplica ReplicationStatus = "REPLICA" | ||
49 | ) | ||
50 | |||
51 | // Empty returns true if no replication status set. | ||
52 | func (r ReplicationStatus) Empty() bool { | ||
53 | return r == "" | ||
54 | } | ||
55 | |||
56 | // AdvancedPutOptions for internal use - to be utilized by replication, ILM transition | ||
57 | // implementation on MinIO server | ||
58 | type AdvancedPutOptions struct { | ||
59 | SourceVersionID string | ||
60 | SourceETag string | ||
61 | ReplicationStatus ReplicationStatus | ||
62 | SourceMTime time.Time | ||
63 | ReplicationRequest bool | ||
64 | RetentionTimestamp time.Time | ||
65 | TaggingTimestamp time.Time | ||
66 | LegalholdTimestamp time.Time | ||
67 | ReplicationValidityCheck bool | ||
68 | } | ||
69 | |||
70 | // PutObjectOptions represents options specified by user for PutObject call | ||
71 | type PutObjectOptions struct { | ||
72 | UserMetadata map[string]string | ||
73 | UserTags map[string]string | ||
74 | Progress io.Reader | ||
75 | ContentType string | ||
76 | ContentEncoding string | ||
77 | ContentDisposition string | ||
78 | ContentLanguage string | ||
79 | CacheControl string | ||
80 | Expires time.Time | ||
81 | Mode RetentionMode | ||
82 | RetainUntilDate time.Time | ||
83 | ServerSideEncryption encrypt.ServerSide | ||
84 | NumThreads uint | ||
85 | StorageClass string | ||
86 | WebsiteRedirectLocation string | ||
87 | PartSize uint64 | ||
88 | LegalHold LegalHoldStatus | ||
89 | SendContentMd5 bool | ||
90 | DisableContentSha256 bool | ||
91 | DisableMultipart bool | ||
92 | |||
93 | // ConcurrentStreamParts will create NumThreads buffers of PartSize bytes, | ||
94 | // fill them serially and upload them in parallel. | ||
95 | // This can be used for faster uploads on non-seekable or slow-to-seek input. | ||
96 | ConcurrentStreamParts bool | ||
97 | Internal AdvancedPutOptions | ||
98 | |||
99 | customHeaders http.Header | ||
100 | } | ||
101 | |||
102 | // SetMatchETag if etag matches while PUT MinIO returns an error | ||
103 | // this is a MinIO specific extension to support optimistic locking | ||
104 | // semantics. | ||
105 | func (opts *PutObjectOptions) SetMatchETag(etag string) { | ||
106 | if opts.customHeaders == nil { | ||
107 | opts.customHeaders = http.Header{} | ||
108 | } | ||
109 | opts.customHeaders.Set("If-Match", "\""+etag+"\"") | ||
110 | } | ||
111 | |||
112 | // SetMatchETagExcept if etag does not match while PUT MinIO returns an | ||
113 | // error this is a MinIO specific extension to support optimistic locking | ||
114 | // semantics. | ||
115 | func (opts *PutObjectOptions) SetMatchETagExcept(etag string) { | ||
116 | if opts.customHeaders == nil { | ||
117 | opts.customHeaders = http.Header{} | ||
118 | } | ||
119 | opts.customHeaders.Set("If-None-Match", "\""+etag+"\"") | ||
120 | } | ||
121 | |||
122 | // getNumThreads - gets the number of threads to be used in the multipart | ||
123 | // put object operation | ||
124 | func (opts PutObjectOptions) getNumThreads() (numThreads int) { | ||
125 | if opts.NumThreads > 0 { | ||
126 | numThreads = int(opts.NumThreads) | ||
127 | } else { | ||
128 | numThreads = totalWorkers | ||
129 | } | ||
130 | return | ||
131 | } | ||
132 | |||
133 | // Header - constructs the headers from metadata entered by user in | ||
134 | // PutObjectOptions struct | ||
135 | func (opts PutObjectOptions) Header() (header http.Header) { | ||
136 | header = make(http.Header) | ||
137 | |||
138 | contentType := opts.ContentType | ||
139 | if contentType == "" { | ||
140 | contentType = "application/octet-stream" | ||
141 | } | ||
142 | header.Set("Content-Type", contentType) | ||
143 | |||
144 | if opts.ContentEncoding != "" { | ||
145 | header.Set("Content-Encoding", opts.ContentEncoding) | ||
146 | } | ||
147 | if opts.ContentDisposition != "" { | ||
148 | header.Set("Content-Disposition", opts.ContentDisposition) | ||
149 | } | ||
150 | if opts.ContentLanguage != "" { | ||
151 | header.Set("Content-Language", opts.ContentLanguage) | ||
152 | } | ||
153 | if opts.CacheControl != "" { | ||
154 | header.Set("Cache-Control", opts.CacheControl) | ||
155 | } | ||
156 | |||
157 | if !opts.Expires.IsZero() { | ||
158 | header.Set("Expires", opts.Expires.UTC().Format(http.TimeFormat)) | ||
159 | } | ||
160 | |||
161 | if opts.Mode != "" { | ||
162 | header.Set(amzLockMode, opts.Mode.String()) | ||
163 | } | ||
164 | |||
165 | if !opts.RetainUntilDate.IsZero() { | ||
166 | header.Set("X-Amz-Object-Lock-Retain-Until-Date", opts.RetainUntilDate.Format(time.RFC3339)) | ||
167 | } | ||
168 | |||
169 | if opts.LegalHold != "" { | ||
170 | header.Set(amzLegalHoldHeader, opts.LegalHold.String()) | ||
171 | } | ||
172 | |||
173 | if opts.ServerSideEncryption != nil { | ||
174 | opts.ServerSideEncryption.Marshal(header) | ||
175 | } | ||
176 | |||
177 | if opts.StorageClass != "" { | ||
178 | header.Set(amzStorageClass, opts.StorageClass) | ||
179 | } | ||
180 | |||
181 | if opts.WebsiteRedirectLocation != "" { | ||
182 | header.Set(amzWebsiteRedirectLocation, opts.WebsiteRedirectLocation) | ||
183 | } | ||
184 | |||
185 | if !opts.Internal.ReplicationStatus.Empty() { | ||
186 | header.Set(amzBucketReplicationStatus, string(opts.Internal.ReplicationStatus)) | ||
187 | } | ||
188 | if !opts.Internal.SourceMTime.IsZero() { | ||
189 | header.Set(minIOBucketSourceMTime, opts.Internal.SourceMTime.Format(time.RFC3339Nano)) | ||
190 | } | ||
191 | if opts.Internal.SourceETag != "" { | ||
192 | header.Set(minIOBucketSourceETag, opts.Internal.SourceETag) | ||
193 | } | ||
194 | if opts.Internal.ReplicationRequest { | ||
195 | header.Set(minIOBucketReplicationRequest, "true") | ||
196 | } | ||
197 | if opts.Internal.ReplicationValidityCheck { | ||
198 | header.Set(minIOBucketReplicationCheck, "true") | ||
199 | } | ||
200 | if !opts.Internal.LegalholdTimestamp.IsZero() { | ||
201 | header.Set(minIOBucketReplicationObjectLegalHoldTimestamp, opts.Internal.LegalholdTimestamp.Format(time.RFC3339Nano)) | ||
202 | } | ||
203 | if !opts.Internal.RetentionTimestamp.IsZero() { | ||
204 | header.Set(minIOBucketReplicationObjectRetentionTimestamp, opts.Internal.RetentionTimestamp.Format(time.RFC3339Nano)) | ||
205 | } | ||
206 | if !opts.Internal.TaggingTimestamp.IsZero() { | ||
207 | header.Set(minIOBucketReplicationTaggingTimestamp, opts.Internal.TaggingTimestamp.Format(time.RFC3339Nano)) | ||
208 | } | ||
209 | |||
210 | if len(opts.UserTags) != 0 { | ||
211 | header.Set(amzTaggingHeader, s3utils.TagEncode(opts.UserTags)) | ||
212 | } | ||
213 | |||
214 | for k, v := range opts.UserMetadata { | ||
215 | if isAmzHeader(k) || isStandardHeader(k) || isStorageClassHeader(k) { | ||
216 | header.Set(k, v) | ||
217 | } else { | ||
218 | header.Set("x-amz-meta-"+k, v) | ||
219 | } | ||
220 | } | ||
221 | |||
222 | // set any other additional custom headers. | ||
223 | for k, v := range opts.customHeaders { | ||
224 | header[k] = v | ||
225 | } | ||
226 | |||
227 | return | ||
228 | } | ||
229 | |||
230 | // validate() checks if the UserMetadata map has standard headers or and raises an error if so. | ||
231 | func (opts PutObjectOptions) validate() (err error) { | ||
232 | for k, v := range opts.UserMetadata { | ||
233 | if !httpguts.ValidHeaderFieldName(k) || isStandardHeader(k) || isSSEHeader(k) || isStorageClassHeader(k) { | ||
234 | return errInvalidArgument(k + " unsupported user defined metadata name") | ||
235 | } | ||
236 | if !httpguts.ValidHeaderFieldValue(v) { | ||
237 | return errInvalidArgument(v + " unsupported user defined metadata value") | ||
238 | } | ||
239 | } | ||
240 | if opts.Mode != "" && !opts.Mode.IsValid() { | ||
241 | return errInvalidArgument(opts.Mode.String() + " unsupported retention mode") | ||
242 | } | ||
243 | if opts.LegalHold != "" && !opts.LegalHold.IsValid() { | ||
244 | return errInvalidArgument(opts.LegalHold.String() + " unsupported legal-hold status") | ||
245 | } | ||
246 | return nil | ||
247 | } | ||
248 | |||
249 | // completedParts is a collection of parts sortable by their part numbers. | ||
250 | // used for sorting the uploaded parts before completing the multipart request. | ||
251 | type completedParts []CompletePart | ||
252 | |||
253 | func (a completedParts) Len() int { return len(a) } | ||
254 | func (a completedParts) Swap(i, j int) { a[i], a[j] = a[j], a[i] } | ||
255 | func (a completedParts) Less(i, j int) bool { return a[i].PartNumber < a[j].PartNumber } | ||
256 | |||
257 | // PutObject creates an object in a bucket. | ||
258 | // | ||
259 | // You must have WRITE permissions on a bucket to create an object. | ||
260 | // | ||
261 | // - For size smaller than 16MiB PutObject automatically does a | ||
262 | // single atomic PUT operation. | ||
263 | // | ||
264 | // - For size larger than 16MiB PutObject automatically does a | ||
265 | // multipart upload operation. | ||
266 | // | ||
267 | // - For size input as -1 PutObject does a multipart Put operation | ||
268 | // until input stream reaches EOF. Maximum object size that can | ||
269 | // be uploaded through this operation will be 5TiB. | ||
270 | // | ||
271 | // WARNING: Passing down '-1' will use memory and these cannot | ||
272 | // be reused for best outcomes for PutObject(), pass the size always. | ||
273 | // | ||
274 | // NOTE: Upon errors during upload multipart operation is entirely aborted. | ||
275 | func (c *Client) PutObject(ctx context.Context, bucketName, objectName string, reader io.Reader, objectSize int64, | ||
276 | opts PutObjectOptions, | ||
277 | ) (info UploadInfo, err error) { | ||
278 | if objectSize < 0 && opts.DisableMultipart { | ||
279 | return UploadInfo{}, errors.New("object size must be provided with disable multipart upload") | ||
280 | } | ||
281 | |||
282 | err = opts.validate() | ||
283 | if err != nil { | ||
284 | return UploadInfo{}, err | ||
285 | } | ||
286 | |||
287 | return c.putObjectCommon(ctx, bucketName, objectName, reader, objectSize, opts) | ||
288 | } | ||
289 | |||
290 | func (c *Client) putObjectCommon(ctx context.Context, bucketName, objectName string, reader io.Reader, size int64, opts PutObjectOptions) (info UploadInfo, err error) { | ||
291 | // Check for largest object size allowed. | ||
292 | if size > int64(maxMultipartPutObjectSize) { | ||
293 | return UploadInfo{}, errEntityTooLarge(size, maxMultipartPutObjectSize, bucketName, objectName) | ||
294 | } | ||
295 | |||
296 | // NOTE: Streaming signature is not supported by GCS. | ||
297 | if s3utils.IsGoogleEndpoint(*c.endpointURL) { | ||
298 | return c.putObject(ctx, bucketName, objectName, reader, size, opts) | ||
299 | } | ||
300 | |||
301 | partSize := opts.PartSize | ||
302 | if opts.PartSize == 0 { | ||
303 | partSize = minPartSize | ||
304 | } | ||
305 | |||
306 | if c.overrideSignerType.IsV2() { | ||
307 | if size >= 0 && size < int64(partSize) || opts.DisableMultipart { | ||
308 | return c.putObject(ctx, bucketName, objectName, reader, size, opts) | ||
309 | } | ||
310 | return c.putObjectMultipart(ctx, bucketName, objectName, reader, size, opts) | ||
311 | } | ||
312 | |||
313 | if size < 0 { | ||
314 | if opts.DisableMultipart { | ||
315 | return UploadInfo{}, errors.New("no length provided and multipart disabled") | ||
316 | } | ||
317 | if opts.ConcurrentStreamParts && opts.NumThreads > 1 { | ||
318 | return c.putObjectMultipartStreamParallel(ctx, bucketName, objectName, reader, opts) | ||
319 | } | ||
320 | return c.putObjectMultipartStreamNoLength(ctx, bucketName, objectName, reader, opts) | ||
321 | } | ||
322 | |||
323 | if size < int64(partSize) || opts.DisableMultipart { | ||
324 | return c.putObject(ctx, bucketName, objectName, reader, size, opts) | ||
325 | } | ||
326 | |||
327 | return c.putObjectMultipartStream(ctx, bucketName, objectName, reader, size, opts) | ||
328 | } | ||
329 | |||
330 | func (c *Client) putObjectMultipartStreamNoLength(ctx context.Context, bucketName, objectName string, reader io.Reader, opts PutObjectOptions) (info UploadInfo, err error) { | ||
331 | // Input validation. | ||
332 | if err = s3utils.CheckValidBucketName(bucketName); err != nil { | ||
333 | return UploadInfo{}, err | ||
334 | } | ||
335 | if err = s3utils.CheckValidObjectName(objectName); err != nil { | ||
336 | return UploadInfo{}, err | ||
337 | } | ||
338 | |||
339 | // Total data read and written to server. should be equal to | ||
340 | // 'size' at the end of the call. | ||
341 | var totalUploadedSize int64 | ||
342 | |||
343 | // Complete multipart upload. | ||
344 | var complMultipartUpload completeMultipartUpload | ||
345 | |||
346 | // Calculate the optimal parts info for a given size. | ||
347 | totalPartsCount, partSize, _, err := OptimalPartInfo(-1, opts.PartSize) | ||
348 | if err != nil { | ||
349 | return UploadInfo{}, err | ||
350 | } | ||
351 | |||
352 | if !opts.SendContentMd5 { | ||
353 | if opts.UserMetadata == nil { | ||
354 | opts.UserMetadata = make(map[string]string, 1) | ||
355 | } | ||
356 | opts.UserMetadata["X-Amz-Checksum-Algorithm"] = "CRC32C" | ||
357 | } | ||
358 | |||
359 | // Initiate a new multipart upload. | ||
360 | uploadID, err := c.newUploadID(ctx, bucketName, objectName, opts) | ||
361 | if err != nil { | ||
362 | return UploadInfo{}, err | ||
363 | } | ||
364 | delete(opts.UserMetadata, "X-Amz-Checksum-Algorithm") | ||
365 | |||
366 | defer func() { | ||
367 | if err != nil { | ||
368 | c.abortMultipartUpload(ctx, bucketName, objectName, uploadID) | ||
369 | } | ||
370 | }() | ||
371 | |||
372 | // Part number always starts with '1'. | ||
373 | partNumber := 1 | ||
374 | |||
375 | // Initialize parts uploaded map. | ||
376 | partsInfo := make(map[int]ObjectPart) | ||
377 | |||
378 | // Create a buffer. | ||
379 | buf := make([]byte, partSize) | ||
380 | |||
381 | // Create checksums | ||
382 | // CRC32C is ~50% faster on AMD64 @ 30GB/s | ||
383 | var crcBytes []byte | ||
384 | customHeader := make(http.Header) | ||
385 | crc := crc32.New(crc32.MakeTable(crc32.Castagnoli)) | ||
386 | |||
387 | for partNumber <= totalPartsCount { | ||
388 | length, rerr := readFull(reader, buf) | ||
389 | if rerr == io.EOF && partNumber > 1 { | ||
390 | break | ||
391 | } | ||
392 | |||
393 | if rerr != nil && rerr != io.ErrUnexpectedEOF && rerr != io.EOF { | ||
394 | return UploadInfo{}, rerr | ||
395 | } | ||
396 | |||
397 | var md5Base64 string | ||
398 | if opts.SendContentMd5 { | ||
399 | // Calculate md5sum. | ||
400 | hash := c.md5Hasher() | ||
401 | hash.Write(buf[:length]) | ||
402 | md5Base64 = base64.StdEncoding.EncodeToString(hash.Sum(nil)) | ||
403 | hash.Close() | ||
404 | } else { | ||
405 | crc.Reset() | ||
406 | crc.Write(buf[:length]) | ||
407 | cSum := crc.Sum(nil) | ||
408 | customHeader.Set("x-amz-checksum-crc32c", base64.StdEncoding.EncodeToString(cSum)) | ||
409 | crcBytes = append(crcBytes, cSum...) | ||
410 | } | ||
411 | |||
412 | // Update progress reader appropriately to the latest offset | ||
413 | // as we read from the source. | ||
414 | rd := newHook(bytes.NewReader(buf[:length]), opts.Progress) | ||
415 | |||
416 | // Proceed to upload the part. | ||
417 | p := uploadPartParams{bucketName: bucketName, objectName: objectName, uploadID: uploadID, reader: rd, partNumber: partNumber, md5Base64: md5Base64, size: int64(length), sse: opts.ServerSideEncryption, streamSha256: !opts.DisableContentSha256, customHeader: customHeader} | ||
418 | objPart, uerr := c.uploadPart(ctx, p) | ||
419 | if uerr != nil { | ||
420 | return UploadInfo{}, uerr | ||
421 | } | ||
422 | |||
423 | // Save successfully uploaded part metadata. | ||
424 | partsInfo[partNumber] = objPart | ||
425 | |||
426 | // Save successfully uploaded size. | ||
427 | totalUploadedSize += int64(length) | ||
428 | |||
429 | // Increment part number. | ||
430 | partNumber++ | ||
431 | |||
432 | // For unknown size, Read EOF we break away. | ||
433 | // We do not have to upload till totalPartsCount. | ||
434 | if rerr == io.EOF { | ||
435 | break | ||
436 | } | ||
437 | } | ||
438 | |||
439 | // Loop over total uploaded parts to save them in | ||
440 | // Parts array before completing the multipart request. | ||
441 | for i := 1; i < partNumber; i++ { | ||
442 | part, ok := partsInfo[i] | ||
443 | if !ok { | ||
444 | return UploadInfo{}, errInvalidArgument(fmt.Sprintf("Missing part number %d", i)) | ||
445 | } | ||
446 | complMultipartUpload.Parts = append(complMultipartUpload.Parts, CompletePart{ | ||
447 | ETag: part.ETag, | ||
448 | PartNumber: part.PartNumber, | ||
449 | ChecksumCRC32: part.ChecksumCRC32, | ||
450 | ChecksumCRC32C: part.ChecksumCRC32C, | ||
451 | ChecksumSHA1: part.ChecksumSHA1, | ||
452 | ChecksumSHA256: part.ChecksumSHA256, | ||
453 | }) | ||
454 | } | ||
455 | |||
456 | // Sort all completed parts. | ||
457 | sort.Sort(completedParts(complMultipartUpload.Parts)) | ||
458 | |||
459 | opts = PutObjectOptions{} | ||
460 | if len(crcBytes) > 0 { | ||
461 | // Add hash of hashes. | ||
462 | crc.Reset() | ||
463 | crc.Write(crcBytes) | ||
464 | opts.UserMetadata = map[string]string{"X-Amz-Checksum-Crc32c": base64.StdEncoding.EncodeToString(crc.Sum(nil))} | ||
465 | } | ||
466 | uploadInfo, err := c.completeMultipartUpload(ctx, bucketName, objectName, uploadID, complMultipartUpload, opts) | ||
467 | if err != nil { | ||
468 | return UploadInfo{}, err | ||
469 | } | ||
470 | |||
471 | uploadInfo.Size = totalUploadedSize | ||
472 | return uploadInfo, nil | ||
473 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/api-putobject-snowball.go b/vendor/github.com/minio/minio-go/v7/api-putobject-snowball.go deleted file mode 100644 index eb4da41..0000000 --- a/vendor/github.com/minio/minio-go/v7/api-putobject-snowball.go +++ /dev/null | |||
@@ -1,246 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2021 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package minio | ||
19 | |||
20 | import ( | ||
21 | "archive/tar" | ||
22 | "bufio" | ||
23 | "bytes" | ||
24 | "context" | ||
25 | "fmt" | ||
26 | "io" | ||
27 | "net/http" | ||
28 | "os" | ||
29 | "strings" | ||
30 | "sync" | ||
31 | "time" | ||
32 | |||
33 | "github.com/klauspost/compress/s2" | ||
34 | ) | ||
35 | |||
36 | // SnowballOptions contains options for PutObjectsSnowball calls. | ||
37 | type SnowballOptions struct { | ||
38 | // Opts is options applied to all objects. | ||
39 | Opts PutObjectOptions | ||
40 | |||
41 | // Processing options: | ||
42 | |||
43 | // InMemory specifies that all objects should be collected in memory | ||
44 | // before they are uploaded. | ||
45 | // If false a temporary file will be created. | ||
46 | InMemory bool | ||
47 | |||
48 | // Compress enabled content compression before upload. | ||
49 | // Compression will typically reduce memory and network usage, | ||
50 | // Compression can safely be enabled with MinIO hosts. | ||
51 | Compress bool | ||
52 | |||
53 | // SkipErrs if enabled will skip any errors while reading the | ||
54 | // object content while creating the snowball archive | ||
55 | SkipErrs bool | ||
56 | } | ||
57 | |||
58 | // SnowballObject contains information about a single object to be added to the snowball. | ||
59 | type SnowballObject struct { | ||
60 | // Key is the destination key, including prefix. | ||
61 | Key string | ||
62 | |||
63 | // Size is the content size of this object. | ||
64 | Size int64 | ||
65 | |||
66 | // Modtime to apply to the object. | ||
67 | // If Modtime is the zero value current time will be used. | ||
68 | ModTime time.Time | ||
69 | |||
70 | // Content of the object. | ||
71 | // Exactly 'Size' number of bytes must be provided. | ||
72 | Content io.Reader | ||
73 | |||
74 | // VersionID of the object; if empty, a new versionID will be generated | ||
75 | VersionID string | ||
76 | |||
77 | // Headers contains more options for this object upload, the same as you | ||
78 | // would include in a regular PutObject operation, such as user metadata | ||
79 | // and content-disposition, expires, .. | ||
80 | Headers http.Header | ||
81 | |||
82 | // Close will be called when an object has finished processing. | ||
83 | // Note that if PutObjectsSnowball returns because of an error, | ||
84 | // objects not consumed from the input will NOT have been closed. | ||
85 | // Leave as nil for no callback. | ||
86 | Close func() | ||
87 | } | ||
88 | |||
89 | type nopReadSeekCloser struct { | ||
90 | io.ReadSeeker | ||
91 | } | ||
92 | |||
93 | func (n nopReadSeekCloser) Close() error { | ||
94 | return nil | ||
95 | } | ||
96 | |||
97 | // This is available as io.ReadSeekCloser from go1.16 | ||
98 | type readSeekCloser interface { | ||
99 | io.Reader | ||
100 | io.Closer | ||
101 | io.Seeker | ||
102 | } | ||
103 | |||
104 | // PutObjectsSnowball will put multiple objects with a single put call. | ||
105 | // A (compressed) TAR file will be created which will contain multiple objects. | ||
106 | // The key for each object will be used for the destination in the specified bucket. | ||
107 | // Total size should be < 5TB. | ||
108 | // This function blocks until 'objs' is closed and the content has been uploaded. | ||
109 | func (c Client) PutObjectsSnowball(ctx context.Context, bucketName string, opts SnowballOptions, objs <-chan SnowballObject) (err error) { | ||
110 | err = opts.Opts.validate() | ||
111 | if err != nil { | ||
112 | return err | ||
113 | } | ||
114 | var tmpWriter io.Writer | ||
115 | var getTmpReader func() (rc readSeekCloser, sz int64, err error) | ||
116 | if opts.InMemory { | ||
117 | b := bytes.NewBuffer(nil) | ||
118 | tmpWriter = b | ||
119 | getTmpReader = func() (readSeekCloser, int64, error) { | ||
120 | return nopReadSeekCloser{bytes.NewReader(b.Bytes())}, int64(b.Len()), nil | ||
121 | } | ||
122 | } else { | ||
123 | f, err := os.CreateTemp("", "s3-putsnowballobjects-*") | ||
124 | if err != nil { | ||
125 | return err | ||
126 | } | ||
127 | name := f.Name() | ||
128 | tmpWriter = f | ||
129 | var once sync.Once | ||
130 | defer once.Do(func() { | ||
131 | f.Close() | ||
132 | }) | ||
133 | defer os.Remove(name) | ||
134 | getTmpReader = func() (readSeekCloser, int64, error) { | ||
135 | once.Do(func() { | ||
136 | f.Close() | ||
137 | }) | ||
138 | f, err := os.Open(name) | ||
139 | if err != nil { | ||
140 | return nil, 0, err | ||
141 | } | ||
142 | st, err := f.Stat() | ||
143 | if err != nil { | ||
144 | return nil, 0, err | ||
145 | } | ||
146 | return f, st.Size(), nil | ||
147 | } | ||
148 | } | ||
149 | flush := func() error { return nil } | ||
150 | if !opts.Compress { | ||
151 | if !opts.InMemory { | ||
152 | // Insert buffer for writes. | ||
153 | buf := bufio.NewWriterSize(tmpWriter, 1<<20) | ||
154 | flush = buf.Flush | ||
155 | tmpWriter = buf | ||
156 | } | ||
157 | } else { | ||
158 | s2c := s2.NewWriter(tmpWriter, s2.WriterBetterCompression()) | ||
159 | flush = s2c.Close | ||
160 | defer s2c.Close() | ||
161 | tmpWriter = s2c | ||
162 | } | ||
163 | t := tar.NewWriter(tmpWriter) | ||
164 | |||
165 | objectLoop: | ||
166 | for { | ||
167 | select { | ||
168 | case <-ctx.Done(): | ||
169 | return ctx.Err() | ||
170 | case obj, ok := <-objs: | ||
171 | if !ok { | ||
172 | break objectLoop | ||
173 | } | ||
174 | |||
175 | closeObj := func() {} | ||
176 | if obj.Close != nil { | ||
177 | closeObj = obj.Close | ||
178 | } | ||
179 | |||
180 | // Trim accidental slash prefix. | ||
181 | obj.Key = strings.TrimPrefix(obj.Key, "/") | ||
182 | header := tar.Header{ | ||
183 | Typeflag: tar.TypeReg, | ||
184 | Name: obj.Key, | ||
185 | Size: obj.Size, | ||
186 | ModTime: obj.ModTime, | ||
187 | Format: tar.FormatPAX, | ||
188 | } | ||
189 | if header.ModTime.IsZero() { | ||
190 | header.ModTime = time.Now().UTC() | ||
191 | } | ||
192 | |||
193 | header.PAXRecords = make(map[string]string) | ||
194 | if obj.VersionID != "" { | ||
195 | header.PAXRecords["minio.versionId"] = obj.VersionID | ||
196 | } | ||
197 | for k, vals := range obj.Headers { | ||
198 | header.PAXRecords["minio.metadata."+k] = strings.Join(vals, ",") | ||
199 | } | ||
200 | |||
201 | if err := t.WriteHeader(&header); err != nil { | ||
202 | closeObj() | ||
203 | return err | ||
204 | } | ||
205 | n, err := io.Copy(t, obj.Content) | ||
206 | if err != nil { | ||
207 | closeObj() | ||
208 | if opts.SkipErrs { | ||
209 | continue | ||
210 | } | ||
211 | return err | ||
212 | } | ||
213 | if n != obj.Size { | ||
214 | closeObj() | ||
215 | if opts.SkipErrs { | ||
216 | continue | ||
217 | } | ||
218 | return io.ErrUnexpectedEOF | ||
219 | } | ||
220 | closeObj() | ||
221 | } | ||
222 | } | ||
223 | // Flush tar | ||
224 | err = t.Flush() | ||
225 | if err != nil { | ||
226 | return err | ||
227 | } | ||
228 | // Flush compression | ||
229 | err = flush() | ||
230 | if err != nil { | ||
231 | return err | ||
232 | } | ||
233 | if opts.Opts.UserMetadata == nil { | ||
234 | opts.Opts.UserMetadata = map[string]string{} | ||
235 | } | ||
236 | opts.Opts.UserMetadata["X-Amz-Meta-Snowball-Auto-Extract"] = "true" | ||
237 | opts.Opts.DisableMultipart = true | ||
238 | rc, sz, err := getTmpReader() | ||
239 | if err != nil { | ||
240 | return err | ||
241 | } | ||
242 | defer rc.Close() | ||
243 | rand := c.random.Uint64() | ||
244 | _, err = c.PutObject(ctx, bucketName, fmt.Sprintf("snowball-upload-%x.tar", rand), rc, sz, opts.Opts) | ||
245 | return err | ||
246 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/api-remove.go b/vendor/github.com/minio/minio-go/v7/api-remove.go deleted file mode 100644 index 9c0ac44..0000000 --- a/vendor/github.com/minio/minio-go/v7/api-remove.go +++ /dev/null | |||
@@ -1,548 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2015-2020 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package minio | ||
19 | |||
20 | import ( | ||
21 | "bytes" | ||
22 | "context" | ||
23 | "encoding/xml" | ||
24 | "io" | ||
25 | "net/http" | ||
26 | "net/url" | ||
27 | "time" | ||
28 | |||
29 | "github.com/minio/minio-go/v7/pkg/s3utils" | ||
30 | ) | ||
31 | |||
32 | //revive:disable | ||
33 | |||
34 | // Deprecated: BucketOptions will be renamed to RemoveBucketOptions in future versions. | ||
35 | type BucketOptions = RemoveBucketOptions | ||
36 | |||
37 | //revive:enable | ||
38 | |||
39 | // RemoveBucketOptions special headers to purge buckets, only | ||
40 | // useful when endpoint is MinIO | ||
41 | type RemoveBucketOptions struct { | ||
42 | ForceDelete bool | ||
43 | } | ||
44 | |||
45 | // RemoveBucketWithOptions deletes the bucket name. | ||
46 | // | ||
47 | // All objects (including all object versions and delete markers) | ||
48 | // in the bucket will be deleted forcibly if bucket options set | ||
49 | // ForceDelete to 'true'. | ||
50 | func (c *Client) RemoveBucketWithOptions(ctx context.Context, bucketName string, opts RemoveBucketOptions) error { | ||
51 | // Input validation. | ||
52 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
53 | return err | ||
54 | } | ||
55 | |||
56 | // Build headers. | ||
57 | headers := make(http.Header) | ||
58 | if opts.ForceDelete { | ||
59 | headers.Set(minIOForceDelete, "true") | ||
60 | } | ||
61 | |||
62 | // Execute DELETE on bucket. | ||
63 | resp, err := c.executeMethod(ctx, http.MethodDelete, requestMetadata{ | ||
64 | bucketName: bucketName, | ||
65 | contentSHA256Hex: emptySHA256Hex, | ||
66 | customHeader: headers, | ||
67 | }) | ||
68 | defer closeResponse(resp) | ||
69 | if err != nil { | ||
70 | return err | ||
71 | } | ||
72 | if resp != nil { | ||
73 | if resp.StatusCode != http.StatusNoContent { | ||
74 | return httpRespToErrorResponse(resp, bucketName, "") | ||
75 | } | ||
76 | } | ||
77 | |||
78 | // Remove the location from cache on a successful delete. | ||
79 | c.bucketLocCache.Delete(bucketName) | ||
80 | return nil | ||
81 | } | ||
82 | |||
83 | // RemoveBucket deletes the bucket name. | ||
84 | // | ||
85 | // All objects (including all object versions and delete markers). | ||
86 | // in the bucket must be deleted before successfully attempting this request. | ||
87 | func (c *Client) RemoveBucket(ctx context.Context, bucketName string) error { | ||
88 | // Input validation. | ||
89 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
90 | return err | ||
91 | } | ||
92 | // Execute DELETE on bucket. | ||
93 | resp, err := c.executeMethod(ctx, http.MethodDelete, requestMetadata{ | ||
94 | bucketName: bucketName, | ||
95 | contentSHA256Hex: emptySHA256Hex, | ||
96 | }) | ||
97 | defer closeResponse(resp) | ||
98 | if err != nil { | ||
99 | return err | ||
100 | } | ||
101 | if resp != nil { | ||
102 | if resp.StatusCode != http.StatusNoContent { | ||
103 | return httpRespToErrorResponse(resp, bucketName, "") | ||
104 | } | ||
105 | } | ||
106 | |||
107 | // Remove the location from cache on a successful delete. | ||
108 | c.bucketLocCache.Delete(bucketName) | ||
109 | |||
110 | return nil | ||
111 | } | ||
112 | |||
113 | // AdvancedRemoveOptions intended for internal use by replication | ||
114 | type AdvancedRemoveOptions struct { | ||
115 | ReplicationDeleteMarker bool | ||
116 | ReplicationStatus ReplicationStatus | ||
117 | ReplicationMTime time.Time | ||
118 | ReplicationRequest bool | ||
119 | ReplicationValidityCheck bool // check permissions | ||
120 | } | ||
121 | |||
122 | // RemoveObjectOptions represents options specified by user for RemoveObject call | ||
123 | type RemoveObjectOptions struct { | ||
124 | ForceDelete bool | ||
125 | GovernanceBypass bool | ||
126 | VersionID string | ||
127 | Internal AdvancedRemoveOptions | ||
128 | } | ||
129 | |||
130 | // RemoveObject removes an object from a bucket. | ||
131 | func (c *Client) RemoveObject(ctx context.Context, bucketName, objectName string, opts RemoveObjectOptions) error { | ||
132 | // Input validation. | ||
133 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
134 | return err | ||
135 | } | ||
136 | if err := s3utils.CheckValidObjectName(objectName); err != nil { | ||
137 | return err | ||
138 | } | ||
139 | |||
140 | res := c.removeObject(ctx, bucketName, objectName, opts) | ||
141 | return res.Err | ||
142 | } | ||
143 | |||
144 | func (c *Client) removeObject(ctx context.Context, bucketName, objectName string, opts RemoveObjectOptions) RemoveObjectResult { | ||
145 | // Get resources properly escaped and lined up before | ||
146 | // using them in http request. | ||
147 | urlValues := make(url.Values) | ||
148 | |||
149 | if opts.VersionID != "" { | ||
150 | urlValues.Set("versionId", opts.VersionID) | ||
151 | } | ||
152 | |||
153 | // Build headers. | ||
154 | headers := make(http.Header) | ||
155 | |||
156 | if opts.GovernanceBypass { | ||
157 | // Set the bypass goverenance retention header | ||
158 | headers.Set(amzBypassGovernance, "true") | ||
159 | } | ||
160 | if opts.Internal.ReplicationDeleteMarker { | ||
161 | headers.Set(minIOBucketReplicationDeleteMarker, "true") | ||
162 | } | ||
163 | if !opts.Internal.ReplicationMTime.IsZero() { | ||
164 | headers.Set(minIOBucketSourceMTime, opts.Internal.ReplicationMTime.Format(time.RFC3339Nano)) | ||
165 | } | ||
166 | if !opts.Internal.ReplicationStatus.Empty() { | ||
167 | headers.Set(amzBucketReplicationStatus, string(opts.Internal.ReplicationStatus)) | ||
168 | } | ||
169 | if opts.Internal.ReplicationRequest { | ||
170 | headers.Set(minIOBucketReplicationRequest, "true") | ||
171 | } | ||
172 | if opts.Internal.ReplicationValidityCheck { | ||
173 | headers.Set(minIOBucketReplicationCheck, "true") | ||
174 | } | ||
175 | if opts.ForceDelete { | ||
176 | headers.Set(minIOForceDelete, "true") | ||
177 | } | ||
178 | // Execute DELETE on objectName. | ||
179 | resp, err := c.executeMethod(ctx, http.MethodDelete, requestMetadata{ | ||
180 | bucketName: bucketName, | ||
181 | objectName: objectName, | ||
182 | contentSHA256Hex: emptySHA256Hex, | ||
183 | queryValues: urlValues, | ||
184 | customHeader: headers, | ||
185 | }) | ||
186 | defer closeResponse(resp) | ||
187 | if err != nil { | ||
188 | return RemoveObjectResult{Err: err} | ||
189 | } | ||
190 | if resp != nil { | ||
191 | // if some unexpected error happened and max retry is reached, we want to let client know | ||
192 | if resp.StatusCode != http.StatusNoContent { | ||
193 | err := httpRespToErrorResponse(resp, bucketName, objectName) | ||
194 | return RemoveObjectResult{Err: err} | ||
195 | } | ||
196 | } | ||
197 | |||
198 | // DeleteObject always responds with http '204' even for | ||
199 | // objects which do not exist. So no need to handle them | ||
200 | // specifically. | ||
201 | return RemoveObjectResult{ | ||
202 | ObjectName: objectName, | ||
203 | ObjectVersionID: opts.VersionID, | ||
204 | DeleteMarker: resp.Header.Get("x-amz-delete-marker") == "true", | ||
205 | DeleteMarkerVersionID: resp.Header.Get("x-amz-version-id"), | ||
206 | } | ||
207 | } | ||
208 | |||
209 | // RemoveObjectError - container of Multi Delete S3 API error | ||
210 | type RemoveObjectError struct { | ||
211 | ObjectName string | ||
212 | VersionID string | ||
213 | Err error | ||
214 | } | ||
215 | |||
216 | // RemoveObjectResult - container of Multi Delete S3 API result | ||
217 | type RemoveObjectResult struct { | ||
218 | ObjectName string | ||
219 | ObjectVersionID string | ||
220 | |||
221 | DeleteMarker bool | ||
222 | DeleteMarkerVersionID string | ||
223 | |||
224 | Err error | ||
225 | } | ||
226 | |||
227 | // generateRemoveMultiObjects - generate the XML request for remove multi objects request | ||
228 | func generateRemoveMultiObjectsRequest(objects []ObjectInfo) []byte { | ||
229 | delObjects := []deleteObject{} | ||
230 | for _, obj := range objects { | ||
231 | delObjects = append(delObjects, deleteObject{ | ||
232 | Key: obj.Key, | ||
233 | VersionID: obj.VersionID, | ||
234 | }) | ||
235 | } | ||
236 | xmlBytes, _ := xml.Marshal(deleteMultiObjects{Objects: delObjects, Quiet: false}) | ||
237 | return xmlBytes | ||
238 | } | ||
239 | |||
240 | // processRemoveMultiObjectsResponse - parse the remove multi objects web service | ||
241 | // and return the success/failure result status for each object | ||
242 | func processRemoveMultiObjectsResponse(body io.Reader, resultCh chan<- RemoveObjectResult) { | ||
243 | // Parse multi delete XML response | ||
244 | rmResult := &deleteMultiObjectsResult{} | ||
245 | err := xmlDecoder(body, rmResult) | ||
246 | if err != nil { | ||
247 | resultCh <- RemoveObjectResult{ObjectName: "", Err: err} | ||
248 | return | ||
249 | } | ||
250 | |||
251 | // Fill deletion that returned success | ||
252 | for _, obj := range rmResult.DeletedObjects { | ||
253 | resultCh <- RemoveObjectResult{ | ||
254 | ObjectName: obj.Key, | ||
255 | // Only filled with versioned buckets | ||
256 | ObjectVersionID: obj.VersionID, | ||
257 | DeleteMarker: obj.DeleteMarker, | ||
258 | DeleteMarkerVersionID: obj.DeleteMarkerVersionID, | ||
259 | } | ||
260 | } | ||
261 | |||
262 | // Fill deletion that returned an error. | ||
263 | for _, obj := range rmResult.UnDeletedObjects { | ||
264 | // Version does not exist is not an error ignore and continue. | ||
265 | switch obj.Code { | ||
266 | case "InvalidArgument", "NoSuchVersion": | ||
267 | continue | ||
268 | } | ||
269 | resultCh <- RemoveObjectResult{ | ||
270 | ObjectName: obj.Key, | ||
271 | ObjectVersionID: obj.VersionID, | ||
272 | Err: ErrorResponse{ | ||
273 | Code: obj.Code, | ||
274 | Message: obj.Message, | ||
275 | }, | ||
276 | } | ||
277 | } | ||
278 | } | ||
279 | |||
280 | // RemoveObjectsOptions represents options specified by user for RemoveObjects call | ||
281 | type RemoveObjectsOptions struct { | ||
282 | GovernanceBypass bool | ||
283 | } | ||
284 | |||
285 | // RemoveObjects removes multiple objects from a bucket while | ||
286 | // it is possible to specify objects versions which are received from | ||
287 | // objectsCh. Remove failures are sent back via error channel. | ||
288 | func (c *Client) RemoveObjects(ctx context.Context, bucketName string, objectsCh <-chan ObjectInfo, opts RemoveObjectsOptions) <-chan RemoveObjectError { | ||
289 | errorCh := make(chan RemoveObjectError, 1) | ||
290 | |||
291 | // Validate if bucket name is valid. | ||
292 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
293 | defer close(errorCh) | ||
294 | errorCh <- RemoveObjectError{ | ||
295 | Err: err, | ||
296 | } | ||
297 | return errorCh | ||
298 | } | ||
299 | // Validate objects channel to be properly allocated. | ||
300 | if objectsCh == nil { | ||
301 | defer close(errorCh) | ||
302 | errorCh <- RemoveObjectError{ | ||
303 | Err: errInvalidArgument("Objects channel cannot be nil"), | ||
304 | } | ||
305 | return errorCh | ||
306 | } | ||
307 | |||
308 | resultCh := make(chan RemoveObjectResult, 1) | ||
309 | go c.removeObjects(ctx, bucketName, objectsCh, resultCh, opts) | ||
310 | go func() { | ||
311 | defer close(errorCh) | ||
312 | for res := range resultCh { | ||
313 | // Send only errors to the error channel | ||
314 | if res.Err == nil { | ||
315 | continue | ||
316 | } | ||
317 | errorCh <- RemoveObjectError{ | ||
318 | ObjectName: res.ObjectName, | ||
319 | VersionID: res.ObjectVersionID, | ||
320 | Err: res.Err, | ||
321 | } | ||
322 | } | ||
323 | }() | ||
324 | |||
325 | return errorCh | ||
326 | } | ||
327 | |||
328 | // RemoveObjectsWithResult removes multiple objects from a bucket while | ||
329 | // it is possible to specify objects versions which are received from | ||
330 | // objectsCh. Remove results, successes and failures are sent back via | ||
331 | // RemoveObjectResult channel | ||
332 | func (c *Client) RemoveObjectsWithResult(ctx context.Context, bucketName string, objectsCh <-chan ObjectInfo, opts RemoveObjectsOptions) <-chan RemoveObjectResult { | ||
333 | resultCh := make(chan RemoveObjectResult, 1) | ||
334 | |||
335 | // Validate if bucket name is valid. | ||
336 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
337 | defer close(resultCh) | ||
338 | resultCh <- RemoveObjectResult{ | ||
339 | Err: err, | ||
340 | } | ||
341 | return resultCh | ||
342 | } | ||
343 | // Validate objects channel to be properly allocated. | ||
344 | if objectsCh == nil { | ||
345 | defer close(resultCh) | ||
346 | resultCh <- RemoveObjectResult{ | ||
347 | Err: errInvalidArgument("Objects channel cannot be nil"), | ||
348 | } | ||
349 | return resultCh | ||
350 | } | ||
351 | |||
352 | go c.removeObjects(ctx, bucketName, objectsCh, resultCh, opts) | ||
353 | return resultCh | ||
354 | } | ||
355 | |||
356 | // Return true if the character is within the allowed characters in an XML 1.0 document | ||
357 | // The list of allowed characters can be found here: https://www.w3.org/TR/xml/#charsets | ||
358 | func validXMLChar(r rune) (ok bool) { | ||
359 | return r == 0x09 || | ||
360 | r == 0x0A || | ||
361 | r == 0x0D || | ||
362 | r >= 0x20 && r <= 0xD7FF || | ||
363 | r >= 0xE000 && r <= 0xFFFD || | ||
364 | r >= 0x10000 && r <= 0x10FFFF | ||
365 | } | ||
366 | |||
367 | func hasInvalidXMLChar(str string) bool { | ||
368 | for _, s := range str { | ||
369 | if !validXMLChar(s) { | ||
370 | return true | ||
371 | } | ||
372 | } | ||
373 | return false | ||
374 | } | ||
375 | |||
376 | // Generate and call MultiDelete S3 requests based on entries received from objectsCh | ||
377 | func (c *Client) removeObjects(ctx context.Context, bucketName string, objectsCh <-chan ObjectInfo, resultCh chan<- RemoveObjectResult, opts RemoveObjectsOptions) { | ||
378 | maxEntries := 1000 | ||
379 | finish := false | ||
380 | urlValues := make(url.Values) | ||
381 | urlValues.Set("delete", "") | ||
382 | |||
383 | // Close result channel when Multi delete finishes. | ||
384 | defer close(resultCh) | ||
385 | |||
386 | // Loop over entries by 1000 and call MultiDelete requests | ||
387 | for { | ||
388 | if finish { | ||
389 | break | ||
390 | } | ||
391 | count := 0 | ||
392 | var batch []ObjectInfo | ||
393 | |||
394 | // Try to gather 1000 entries | ||
395 | for object := range objectsCh { | ||
396 | if hasInvalidXMLChar(object.Key) { | ||
397 | // Use single DELETE so the object name will be in the request URL instead of the multi-delete XML document. | ||
398 | removeResult := c.removeObject(ctx, bucketName, object.Key, RemoveObjectOptions{ | ||
399 | VersionID: object.VersionID, | ||
400 | GovernanceBypass: opts.GovernanceBypass, | ||
401 | }) | ||
402 | if err := removeResult.Err; err != nil { | ||
403 | // Version does not exist is not an error ignore and continue. | ||
404 | switch ToErrorResponse(err).Code { | ||
405 | case "InvalidArgument", "NoSuchVersion": | ||
406 | continue | ||
407 | } | ||
408 | resultCh <- removeResult | ||
409 | } | ||
410 | |||
411 | resultCh <- removeResult | ||
412 | continue | ||
413 | } | ||
414 | |||
415 | batch = append(batch, object) | ||
416 | if count++; count >= maxEntries { | ||
417 | break | ||
418 | } | ||
419 | } | ||
420 | if count == 0 { | ||
421 | // Multi Objects Delete API doesn't accept empty object list, quit immediately | ||
422 | break | ||
423 | } | ||
424 | if count < maxEntries { | ||
425 | // We didn't have 1000 entries, so this is the last batch | ||
426 | finish = true | ||
427 | } | ||
428 | |||
429 | // Build headers. | ||
430 | headers := make(http.Header) | ||
431 | if opts.GovernanceBypass { | ||
432 | // Set the bypass goverenance retention header | ||
433 | headers.Set(amzBypassGovernance, "true") | ||
434 | } | ||
435 | |||
436 | // Generate remove multi objects XML request | ||
437 | removeBytes := generateRemoveMultiObjectsRequest(batch) | ||
438 | // Execute GET on bucket to list objects. | ||
439 | resp, err := c.executeMethod(ctx, http.MethodPost, requestMetadata{ | ||
440 | bucketName: bucketName, | ||
441 | queryValues: urlValues, | ||
442 | contentBody: bytes.NewReader(removeBytes), | ||
443 | contentLength: int64(len(removeBytes)), | ||
444 | contentMD5Base64: sumMD5Base64(removeBytes), | ||
445 | contentSHA256Hex: sum256Hex(removeBytes), | ||
446 | customHeader: headers, | ||
447 | }) | ||
448 | if resp != nil { | ||
449 | if resp.StatusCode != http.StatusOK { | ||
450 | e := httpRespToErrorResponse(resp, bucketName, "") | ||
451 | resultCh <- RemoveObjectResult{ObjectName: "", Err: e} | ||
452 | } | ||
453 | } | ||
454 | if err != nil { | ||
455 | for _, b := range batch { | ||
456 | resultCh <- RemoveObjectResult{ | ||
457 | ObjectName: b.Key, | ||
458 | ObjectVersionID: b.VersionID, | ||
459 | Err: err, | ||
460 | } | ||
461 | } | ||
462 | continue | ||
463 | } | ||
464 | |||
465 | // Process multiobjects remove xml response | ||
466 | processRemoveMultiObjectsResponse(resp.Body, resultCh) | ||
467 | |||
468 | closeResponse(resp) | ||
469 | } | ||
470 | } | ||
471 | |||
472 | // RemoveIncompleteUpload aborts an partially uploaded object. | ||
473 | func (c *Client) RemoveIncompleteUpload(ctx context.Context, bucketName, objectName string) error { | ||
474 | // Input validation. | ||
475 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
476 | return err | ||
477 | } | ||
478 | if err := s3utils.CheckValidObjectName(objectName); err != nil { | ||
479 | return err | ||
480 | } | ||
481 | // Find multipart upload ids of the object to be aborted. | ||
482 | uploadIDs, err := c.findUploadIDs(ctx, bucketName, objectName) | ||
483 | if err != nil { | ||
484 | return err | ||
485 | } | ||
486 | |||
487 | for _, uploadID := range uploadIDs { | ||
488 | // abort incomplete multipart upload, based on the upload id passed. | ||
489 | err := c.abortMultipartUpload(ctx, bucketName, objectName, uploadID) | ||
490 | if err != nil { | ||
491 | return err | ||
492 | } | ||
493 | } | ||
494 | |||
495 | return nil | ||
496 | } | ||
497 | |||
498 | // abortMultipartUpload aborts a multipart upload for the given | ||
499 | // uploadID, all previously uploaded parts are deleted. | ||
500 | func (c *Client) abortMultipartUpload(ctx context.Context, bucketName, objectName, uploadID string) error { | ||
501 | // Input validation. | ||
502 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
503 | return err | ||
504 | } | ||
505 | if err := s3utils.CheckValidObjectName(objectName); err != nil { | ||
506 | return err | ||
507 | } | ||
508 | |||
509 | // Initialize url queries. | ||
510 | urlValues := make(url.Values) | ||
511 | urlValues.Set("uploadId", uploadID) | ||
512 | |||
513 | // Execute DELETE on multipart upload. | ||
514 | resp, err := c.executeMethod(ctx, http.MethodDelete, requestMetadata{ | ||
515 | bucketName: bucketName, | ||
516 | objectName: objectName, | ||
517 | queryValues: urlValues, | ||
518 | contentSHA256Hex: emptySHA256Hex, | ||
519 | }) | ||
520 | defer closeResponse(resp) | ||
521 | if err != nil { | ||
522 | return err | ||
523 | } | ||
524 | if resp != nil { | ||
525 | if resp.StatusCode != http.StatusNoContent { | ||
526 | // Abort has no response body, handle it for any errors. | ||
527 | var errorResponse ErrorResponse | ||
528 | switch resp.StatusCode { | ||
529 | case http.StatusNotFound: | ||
530 | // This is needed specifically for abort and it cannot | ||
531 | // be converged into default case. | ||
532 | errorResponse = ErrorResponse{ | ||
533 | Code: "NoSuchUpload", | ||
534 | Message: "The specified multipart upload does not exist.", | ||
535 | BucketName: bucketName, | ||
536 | Key: objectName, | ||
537 | RequestID: resp.Header.Get("x-amz-request-id"), | ||
538 | HostID: resp.Header.Get("x-amz-id-2"), | ||
539 | Region: resp.Header.Get("x-amz-bucket-region"), | ||
540 | } | ||
541 | default: | ||
542 | return httpRespToErrorResponse(resp, bucketName, objectName) | ||
543 | } | ||
544 | return errorResponse | ||
545 | } | ||
546 | } | ||
547 | return nil | ||
548 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/api-restore.go b/vendor/github.com/minio/minio-go/v7/api-restore.go deleted file mode 100644 index 9ec8f4f..0000000 --- a/vendor/github.com/minio/minio-go/v7/api-restore.go +++ /dev/null | |||
@@ -1,182 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * (C) 2018-2021 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package minio | ||
19 | |||
20 | import ( | ||
21 | "bytes" | ||
22 | "context" | ||
23 | "encoding/xml" | ||
24 | "net/http" | ||
25 | "net/url" | ||
26 | |||
27 | "github.com/minio/minio-go/v7/pkg/s3utils" | ||
28 | "github.com/minio/minio-go/v7/pkg/tags" | ||
29 | ) | ||
30 | |||
31 | // RestoreType represents the restore request type | ||
32 | type RestoreType string | ||
33 | |||
34 | const ( | ||
35 | // RestoreSelect represents the restore SELECT operation | ||
36 | RestoreSelect = RestoreType("SELECT") | ||
37 | ) | ||
38 | |||
39 | // TierType represents a retrieval tier | ||
40 | type TierType string | ||
41 | |||
42 | const ( | ||
43 | // TierStandard is the standard retrieval tier | ||
44 | TierStandard = TierType("Standard") | ||
45 | // TierBulk is the bulk retrieval tier | ||
46 | TierBulk = TierType("Bulk") | ||
47 | // TierExpedited is the expedited retrieval tier | ||
48 | TierExpedited = TierType("Expedited") | ||
49 | ) | ||
50 | |||
51 | // GlacierJobParameters represents the retrieval tier parameter | ||
52 | type GlacierJobParameters struct { | ||
53 | Tier TierType | ||
54 | } | ||
55 | |||
56 | // Encryption contains the type of server-side encryption used during object retrieval | ||
57 | type Encryption struct { | ||
58 | EncryptionType string | ||
59 | KMSContext string | ||
60 | KMSKeyID string `xml:"KMSKeyId"` | ||
61 | } | ||
62 | |||
63 | // MetadataEntry represents a metadata information of the restored object. | ||
64 | type MetadataEntry struct { | ||
65 | Name string | ||
66 | Value string | ||
67 | } | ||
68 | |||
69 | // S3 holds properties of the copy of the archived object | ||
70 | type S3 struct { | ||
71 | AccessControlList *AccessControlList `xml:"AccessControlList,omitempty"` | ||
72 | BucketName string | ||
73 | Prefix string | ||
74 | CannedACL *string `xml:"CannedACL,omitempty"` | ||
75 | Encryption *Encryption `xml:"Encryption,omitempty"` | ||
76 | StorageClass *string `xml:"StorageClass,omitempty"` | ||
77 | Tagging *tags.Tags `xml:"Tagging,omitempty"` | ||
78 | UserMetadata *MetadataEntry `xml:"UserMetadata,omitempty"` | ||
79 | } | ||
80 | |||
81 | // SelectParameters holds the select request parameters | ||
82 | type SelectParameters struct { | ||
83 | XMLName xml.Name `xml:"SelectParameters"` | ||
84 | ExpressionType QueryExpressionType | ||
85 | Expression string | ||
86 | InputSerialization SelectObjectInputSerialization | ||
87 | OutputSerialization SelectObjectOutputSerialization | ||
88 | } | ||
89 | |||
90 | // OutputLocation holds properties of the copy of the archived object | ||
91 | type OutputLocation struct { | ||
92 | XMLName xml.Name `xml:"OutputLocation"` | ||
93 | S3 S3 `xml:"S3"` | ||
94 | } | ||
95 | |||
96 | // RestoreRequest holds properties of the restore object request | ||
97 | type RestoreRequest struct { | ||
98 | XMLName xml.Name `xml:"http://s3.amazonaws.com/doc/2006-03-01/ RestoreRequest"` | ||
99 | Type *RestoreType `xml:"Type,omitempty"` | ||
100 | Tier *TierType `xml:"Tier,omitempty"` | ||
101 | Days *int `xml:"Days,omitempty"` | ||
102 | GlacierJobParameters *GlacierJobParameters `xml:"GlacierJobParameters,omitempty"` | ||
103 | Description *string `xml:"Description,omitempty"` | ||
104 | SelectParameters *SelectParameters `xml:"SelectParameters,omitempty"` | ||
105 | OutputLocation *OutputLocation `xml:"OutputLocation,omitempty"` | ||
106 | } | ||
107 | |||
108 | // SetDays sets the days parameter of the restore request | ||
109 | func (r *RestoreRequest) SetDays(v int) { | ||
110 | r.Days = &v | ||
111 | } | ||
112 | |||
113 | // SetGlacierJobParameters sets the GlacierJobParameters of the restore request | ||
114 | func (r *RestoreRequest) SetGlacierJobParameters(v GlacierJobParameters) { | ||
115 | r.GlacierJobParameters = &v | ||
116 | } | ||
117 | |||
118 | // SetType sets the type of the restore request | ||
119 | func (r *RestoreRequest) SetType(v RestoreType) { | ||
120 | r.Type = &v | ||
121 | } | ||
122 | |||
123 | // SetTier sets the retrieval tier of the restore request | ||
124 | func (r *RestoreRequest) SetTier(v TierType) { | ||
125 | r.Tier = &v | ||
126 | } | ||
127 | |||
128 | // SetDescription sets the description of the restore request | ||
129 | func (r *RestoreRequest) SetDescription(v string) { | ||
130 | r.Description = &v | ||
131 | } | ||
132 | |||
133 | // SetSelectParameters sets SelectParameters of the restore select request | ||
134 | func (r *RestoreRequest) SetSelectParameters(v SelectParameters) { | ||
135 | r.SelectParameters = &v | ||
136 | } | ||
137 | |||
138 | // SetOutputLocation sets the properties of the copy of the archived object | ||
139 | func (r *RestoreRequest) SetOutputLocation(v OutputLocation) { | ||
140 | r.OutputLocation = &v | ||
141 | } | ||
142 | |||
143 | // RestoreObject is a implementation of https://docs.aws.amazon.com/AmazonS3/latest/API/API_RestoreObject.html AWS S3 API | ||
144 | func (c *Client) RestoreObject(ctx context.Context, bucketName, objectName, versionID string, req RestoreRequest) error { | ||
145 | // Input validation. | ||
146 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
147 | return err | ||
148 | } | ||
149 | if err := s3utils.CheckValidObjectName(objectName); err != nil { | ||
150 | return err | ||
151 | } | ||
152 | |||
153 | restoreRequestBytes, err := xml.Marshal(req) | ||
154 | if err != nil { | ||
155 | return err | ||
156 | } | ||
157 | |||
158 | urlValues := make(url.Values) | ||
159 | urlValues.Set("restore", "") | ||
160 | if versionID != "" { | ||
161 | urlValues.Set("versionId", versionID) | ||
162 | } | ||
163 | |||
164 | // Execute POST on bucket/object. | ||
165 | resp, err := c.executeMethod(ctx, http.MethodPost, requestMetadata{ | ||
166 | bucketName: bucketName, | ||
167 | objectName: objectName, | ||
168 | queryValues: urlValues, | ||
169 | contentMD5Base64: sumMD5Base64(restoreRequestBytes), | ||
170 | contentSHA256Hex: sum256Hex(restoreRequestBytes), | ||
171 | contentBody: bytes.NewReader(restoreRequestBytes), | ||
172 | contentLength: int64(len(restoreRequestBytes)), | ||
173 | }) | ||
174 | defer closeResponse(resp) | ||
175 | if err != nil { | ||
176 | return err | ||
177 | } | ||
178 | if resp.StatusCode != http.StatusAccepted && resp.StatusCode != http.StatusOK { | ||
179 | return httpRespToErrorResponse(resp, bucketName, "") | ||
180 | } | ||
181 | return nil | ||
182 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/api-s3-datatypes.go b/vendor/github.com/minio/minio-go/v7/api-s3-datatypes.go deleted file mode 100644 index 1527b74..0000000 --- a/vendor/github.com/minio/minio-go/v7/api-s3-datatypes.go +++ /dev/null | |||
@@ -1,390 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2015-2020 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package minio | ||
19 | |||
20 | import ( | ||
21 | "encoding/xml" | ||
22 | "errors" | ||
23 | "io" | ||
24 | "reflect" | ||
25 | "time" | ||
26 | ) | ||
27 | |||
28 | // listAllMyBucketsResult container for listBuckets response. | ||
29 | type listAllMyBucketsResult struct { | ||
30 | // Container for one or more buckets. | ||
31 | Buckets struct { | ||
32 | Bucket []BucketInfo | ||
33 | } | ||
34 | Owner owner | ||
35 | } | ||
36 | |||
37 | // owner container for bucket owner information. | ||
38 | type owner struct { | ||
39 | DisplayName string | ||
40 | ID string | ||
41 | } | ||
42 | |||
43 | // CommonPrefix container for prefix response. | ||
44 | type CommonPrefix struct { | ||
45 | Prefix string | ||
46 | } | ||
47 | |||
48 | // ListBucketV2Result container for listObjects response version 2. | ||
49 | type ListBucketV2Result struct { | ||
50 | // A response can contain CommonPrefixes only if you have | ||
51 | // specified a delimiter. | ||
52 | CommonPrefixes []CommonPrefix | ||
53 | // Metadata about each object returned. | ||
54 | Contents []ObjectInfo | ||
55 | Delimiter string | ||
56 | |||
57 | // Encoding type used to encode object keys in the response. | ||
58 | EncodingType string | ||
59 | |||
60 | // A flag that indicates whether or not ListObjects returned all of the results | ||
61 | // that satisfied the search criteria. | ||
62 | IsTruncated bool | ||
63 | MaxKeys int64 | ||
64 | Name string | ||
65 | |||
66 | // Hold the token that will be sent in the next request to fetch the next group of keys | ||
67 | NextContinuationToken string | ||
68 | |||
69 | ContinuationToken string | ||
70 | Prefix string | ||
71 | |||
72 | // FetchOwner and StartAfter are currently not used | ||
73 | FetchOwner string | ||
74 | StartAfter string | ||
75 | } | ||
76 | |||
77 | // Version is an element in the list object versions response | ||
78 | type Version struct { | ||
79 | ETag string | ||
80 | IsLatest bool | ||
81 | Key string | ||
82 | LastModified time.Time | ||
83 | Owner Owner | ||
84 | Size int64 | ||
85 | StorageClass string | ||
86 | VersionID string `xml:"VersionId"` | ||
87 | |||
88 | // x-amz-meta-* headers stripped "x-amz-meta-" prefix containing the first value. | ||
89 | // Only returned by MinIO servers. | ||
90 | UserMetadata StringMap `json:"userMetadata,omitempty"` | ||
91 | |||
92 | // x-amz-tagging values in their k/v values. | ||
93 | // Only returned by MinIO servers. | ||
94 | UserTags URLMap `json:"userTags,omitempty" xml:"UserTags"` | ||
95 | |||
96 | Internal *struct { | ||
97 | K int // Data blocks | ||
98 | M int // Parity blocks | ||
99 | } `xml:"Internal"` | ||
100 | |||
101 | isDeleteMarker bool | ||
102 | } | ||
103 | |||
104 | // ListVersionsResult is an element in the list object versions response | ||
105 | // and has a special Unmarshaler because we need to preserver the order | ||
106 | // of <Version> and <DeleteMarker> in ListVersionsResult.Versions slice | ||
107 | type ListVersionsResult struct { | ||
108 | Versions []Version | ||
109 | |||
110 | CommonPrefixes []CommonPrefix | ||
111 | Name string | ||
112 | Prefix string | ||
113 | Delimiter string | ||
114 | MaxKeys int64 | ||
115 | EncodingType string | ||
116 | IsTruncated bool | ||
117 | KeyMarker string | ||
118 | VersionIDMarker string | ||
119 | NextKeyMarker string | ||
120 | NextVersionIDMarker string | ||
121 | } | ||
122 | |||
123 | // UnmarshalXML is a custom unmarshal code for the response of ListObjectVersions, the custom | ||
124 | // code will unmarshal <Version> and <DeleteMarker> tags and save them in Versions field to | ||
125 | // preserve the lexical order of the listing. | ||
126 | func (l *ListVersionsResult) UnmarshalXML(d *xml.Decoder, _ xml.StartElement) (err error) { | ||
127 | for { | ||
128 | // Read tokens from the XML document in a stream. | ||
129 | t, err := d.Token() | ||
130 | if err != nil { | ||
131 | if err == io.EOF { | ||
132 | break | ||
133 | } | ||
134 | return err | ||
135 | } | ||
136 | |||
137 | se, ok := t.(xml.StartElement) | ||
138 | if ok { | ||
139 | tagName := se.Name.Local | ||
140 | switch tagName { | ||
141 | case "Name", "Prefix", | ||
142 | "Delimiter", "EncodingType", | ||
143 | "KeyMarker", "NextKeyMarker": | ||
144 | var s string | ||
145 | if err = d.DecodeElement(&s, &se); err != nil { | ||
146 | return err | ||
147 | } | ||
148 | v := reflect.ValueOf(l).Elem().FieldByName(tagName) | ||
149 | if v.IsValid() { | ||
150 | v.SetString(s) | ||
151 | } | ||
152 | case "VersionIdMarker": | ||
153 | // VersionIdMarker is a special case because of 'Id' instead of 'ID' in field name | ||
154 | var s string | ||
155 | if err = d.DecodeElement(&s, &se); err != nil { | ||
156 | return err | ||
157 | } | ||
158 | l.VersionIDMarker = s | ||
159 | case "NextVersionIdMarker": | ||
160 | // NextVersionIdMarker is a special case because of 'Id' instead of 'ID' in field name | ||
161 | var s string | ||
162 | if err = d.DecodeElement(&s, &se); err != nil { | ||
163 | return err | ||
164 | } | ||
165 | l.NextVersionIDMarker = s | ||
166 | case "IsTruncated": // bool | ||
167 | var b bool | ||
168 | if err = d.DecodeElement(&b, &se); err != nil { | ||
169 | return err | ||
170 | } | ||
171 | l.IsTruncated = b | ||
172 | case "MaxKeys": // int64 | ||
173 | var i int64 | ||
174 | if err = d.DecodeElement(&i, &se); err != nil { | ||
175 | return err | ||
176 | } | ||
177 | l.MaxKeys = i | ||
178 | case "CommonPrefixes": | ||
179 | var cp CommonPrefix | ||
180 | if err = d.DecodeElement(&cp, &se); err != nil { | ||
181 | return err | ||
182 | } | ||
183 | l.CommonPrefixes = append(l.CommonPrefixes, cp) | ||
184 | case "DeleteMarker", "Version": | ||
185 | var v Version | ||
186 | if err = d.DecodeElement(&v, &se); err != nil { | ||
187 | return err | ||
188 | } | ||
189 | if tagName == "DeleteMarker" { | ||
190 | v.isDeleteMarker = true | ||
191 | } | ||
192 | l.Versions = append(l.Versions, v) | ||
193 | default: | ||
194 | return errors.New("unrecognized option:" + tagName) | ||
195 | } | ||
196 | |||
197 | } | ||
198 | } | ||
199 | return nil | ||
200 | } | ||
201 | |||
202 | // ListBucketResult container for listObjects response. | ||
203 | type ListBucketResult struct { | ||
204 | // A response can contain CommonPrefixes only if you have | ||
205 | // specified a delimiter. | ||
206 | CommonPrefixes []CommonPrefix | ||
207 | // Metadata about each object returned. | ||
208 | Contents []ObjectInfo | ||
209 | Delimiter string | ||
210 | |||
211 | // Encoding type used to encode object keys in the response. | ||
212 | EncodingType string | ||
213 | |||
214 | // A flag that indicates whether or not ListObjects returned all of the results | ||
215 | // that satisfied the search criteria. | ||
216 | IsTruncated bool | ||
217 | Marker string | ||
218 | MaxKeys int64 | ||
219 | Name string | ||
220 | |||
221 | // When response is truncated (the IsTruncated element value in | ||
222 | // the response is true), you can use the key name in this field | ||
223 | // as marker in the subsequent request to get next set of objects. | ||
224 | // Object storage lists objects in alphabetical order Note: This | ||
225 | // element is returned only if you have delimiter request | ||
226 | // parameter specified. If response does not include the NextMaker | ||
227 | // and it is truncated, you can use the value of the last Key in | ||
228 | // the response as the marker in the subsequent request to get the | ||
229 | // next set of object keys. | ||
230 | NextMarker string | ||
231 | Prefix string | ||
232 | } | ||
233 | |||
234 | // ListMultipartUploadsResult container for ListMultipartUploads response | ||
235 | type ListMultipartUploadsResult struct { | ||
236 | Bucket string | ||
237 | KeyMarker string | ||
238 | UploadIDMarker string `xml:"UploadIdMarker"` | ||
239 | NextKeyMarker string | ||
240 | NextUploadIDMarker string `xml:"NextUploadIdMarker"` | ||
241 | EncodingType string | ||
242 | MaxUploads int64 | ||
243 | IsTruncated bool | ||
244 | Uploads []ObjectMultipartInfo `xml:"Upload"` | ||
245 | Prefix string | ||
246 | Delimiter string | ||
247 | // A response can contain CommonPrefixes only if you specify a delimiter. | ||
248 | CommonPrefixes []CommonPrefix | ||
249 | } | ||
250 | |||
251 | // initiator container for who initiated multipart upload. | ||
252 | type initiator struct { | ||
253 | ID string | ||
254 | DisplayName string | ||
255 | } | ||
256 | |||
257 | // copyObjectResult container for copy object response. | ||
258 | type copyObjectResult struct { | ||
259 | ETag string | ||
260 | LastModified time.Time // time string format "2006-01-02T15:04:05.000Z" | ||
261 | } | ||
262 | |||
263 | // ObjectPart container for particular part of an object. | ||
264 | type ObjectPart struct { | ||
265 | // Part number identifies the part. | ||
266 | PartNumber int | ||
267 | |||
268 | // Date and time the part was uploaded. | ||
269 | LastModified time.Time | ||
270 | |||
271 | // Entity tag returned when the part was uploaded, usually md5sum | ||
272 | // of the part. | ||
273 | ETag string | ||
274 | |||
275 | // Size of the uploaded part data. | ||
276 | Size int64 | ||
277 | |||
278 | // Checksum values of each part. | ||
279 | ChecksumCRC32 string | ||
280 | ChecksumCRC32C string | ||
281 | ChecksumSHA1 string | ||
282 | ChecksumSHA256 string | ||
283 | } | ||
284 | |||
285 | // ListObjectPartsResult container for ListObjectParts response. | ||
286 | type ListObjectPartsResult struct { | ||
287 | Bucket string | ||
288 | Key string | ||
289 | UploadID string `xml:"UploadId"` | ||
290 | |||
291 | Initiator initiator | ||
292 | Owner owner | ||
293 | |||
294 | StorageClass string | ||
295 | PartNumberMarker int | ||
296 | NextPartNumberMarker int | ||
297 | MaxParts int | ||
298 | |||
299 | // Indicates whether the returned list of parts is truncated. | ||
300 | IsTruncated bool | ||
301 | ObjectParts []ObjectPart `xml:"Part"` | ||
302 | |||
303 | EncodingType string | ||
304 | } | ||
305 | |||
306 | // initiateMultipartUploadResult container for InitiateMultiPartUpload | ||
307 | // response. | ||
308 | type initiateMultipartUploadResult struct { | ||
309 | Bucket string | ||
310 | Key string | ||
311 | UploadID string `xml:"UploadId"` | ||
312 | } | ||
313 | |||
314 | // completeMultipartUploadResult container for completed multipart | ||
315 | // upload response. | ||
316 | type completeMultipartUploadResult struct { | ||
317 | Location string | ||
318 | Bucket string | ||
319 | Key string | ||
320 | ETag string | ||
321 | |||
322 | // Checksum values, hash of hashes of parts. | ||
323 | ChecksumCRC32 string | ||
324 | ChecksumCRC32C string | ||
325 | ChecksumSHA1 string | ||
326 | ChecksumSHA256 string | ||
327 | } | ||
328 | |||
329 | // CompletePart sub container lists individual part numbers and their | ||
330 | // md5sum, part of completeMultipartUpload. | ||
331 | type CompletePart struct { | ||
332 | // Part number identifies the part. | ||
333 | PartNumber int | ||
334 | ETag string | ||
335 | |||
336 | // Checksum values | ||
337 | ChecksumCRC32 string `xml:"ChecksumCRC32,omitempty"` | ||
338 | ChecksumCRC32C string `xml:"ChecksumCRC32C,omitempty"` | ||
339 | ChecksumSHA1 string `xml:"ChecksumSHA1,omitempty"` | ||
340 | ChecksumSHA256 string `xml:"ChecksumSHA256,omitempty"` | ||
341 | } | ||
342 | |||
343 | // completeMultipartUpload container for completing multipart upload. | ||
344 | type completeMultipartUpload struct { | ||
345 | XMLName xml.Name `xml:"http://s3.amazonaws.com/doc/2006-03-01/ CompleteMultipartUpload" json:"-"` | ||
346 | Parts []CompletePart `xml:"Part"` | ||
347 | } | ||
348 | |||
349 | // createBucketConfiguration container for bucket configuration. | ||
350 | type createBucketConfiguration struct { | ||
351 | XMLName xml.Name `xml:"http://s3.amazonaws.com/doc/2006-03-01/ CreateBucketConfiguration" json:"-"` | ||
352 | Location string `xml:"LocationConstraint"` | ||
353 | } | ||
354 | |||
355 | // deleteObject container for Delete element in MultiObjects Delete XML request | ||
356 | type deleteObject struct { | ||
357 | Key string | ||
358 | VersionID string `xml:"VersionId,omitempty"` | ||
359 | } | ||
360 | |||
361 | // deletedObject container for Deleted element in MultiObjects Delete XML response | ||
362 | type deletedObject struct { | ||
363 | Key string | ||
364 | VersionID string `xml:"VersionId,omitempty"` | ||
365 | // These fields are ignored. | ||
366 | DeleteMarker bool | ||
367 | DeleteMarkerVersionID string `xml:"DeleteMarkerVersionId,omitempty"` | ||
368 | } | ||
369 | |||
370 | // nonDeletedObject container for Error element (failed deletion) in MultiObjects Delete XML response | ||
371 | type nonDeletedObject struct { | ||
372 | Key string | ||
373 | Code string | ||
374 | Message string | ||
375 | VersionID string `xml:"VersionId"` | ||
376 | } | ||
377 | |||
378 | // deletedMultiObjects container for MultiObjects Delete XML request | ||
379 | type deleteMultiObjects struct { | ||
380 | XMLName xml.Name `xml:"Delete"` | ||
381 | Quiet bool | ||
382 | Objects []deleteObject `xml:"Object"` | ||
383 | } | ||
384 | |||
385 | // deletedMultiObjectsResult container for MultiObjects Delete XML response | ||
386 | type deleteMultiObjectsResult struct { | ||
387 | XMLName xml.Name `xml:"DeleteResult"` | ||
388 | DeletedObjects []deletedObject `xml:"Deleted"` | ||
389 | UnDeletedObjects []nonDeletedObject `xml:"Error"` | ||
390 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/api-select.go b/vendor/github.com/minio/minio-go/v7/api-select.go deleted file mode 100644 index 628d967..0000000 --- a/vendor/github.com/minio/minio-go/v7/api-select.go +++ /dev/null | |||
@@ -1,757 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * (C) 2018-2020 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package minio | ||
19 | |||
20 | import ( | ||
21 | "bytes" | ||
22 | "context" | ||
23 | "encoding/binary" | ||
24 | "encoding/xml" | ||
25 | "errors" | ||
26 | "fmt" | ||
27 | "hash" | ||
28 | "hash/crc32" | ||
29 | "io" | ||
30 | "net/http" | ||
31 | "net/url" | ||
32 | "strings" | ||
33 | |||
34 | "github.com/minio/minio-go/v7/pkg/encrypt" | ||
35 | "github.com/minio/minio-go/v7/pkg/s3utils" | ||
36 | ) | ||
37 | |||
38 | // CSVFileHeaderInfo - is the parameter for whether to utilize headers. | ||
39 | type CSVFileHeaderInfo string | ||
40 | |||
41 | // Constants for file header info. | ||
42 | const ( | ||
43 | CSVFileHeaderInfoNone CSVFileHeaderInfo = "NONE" | ||
44 | CSVFileHeaderInfoIgnore CSVFileHeaderInfo = "IGNORE" | ||
45 | CSVFileHeaderInfoUse CSVFileHeaderInfo = "USE" | ||
46 | ) | ||
47 | |||
48 | // SelectCompressionType - is the parameter for what type of compression is | ||
49 | // present | ||
50 | type SelectCompressionType string | ||
51 | |||
52 | // Constants for compression types under select API. | ||
53 | const ( | ||
54 | SelectCompressionNONE SelectCompressionType = "NONE" | ||
55 | SelectCompressionGZIP SelectCompressionType = "GZIP" | ||
56 | SelectCompressionBZIP SelectCompressionType = "BZIP2" | ||
57 | |||
58 | // Non-standard compression schemes, supported by MinIO hosts: | ||
59 | |||
60 | SelectCompressionZSTD SelectCompressionType = "ZSTD" // Zstandard compression. | ||
61 | SelectCompressionLZ4 SelectCompressionType = "LZ4" // LZ4 Stream | ||
62 | SelectCompressionS2 SelectCompressionType = "S2" // S2 Stream | ||
63 | SelectCompressionSNAPPY SelectCompressionType = "SNAPPY" // Snappy stream | ||
64 | ) | ||
65 | |||
66 | // CSVQuoteFields - is the parameter for how CSV fields are quoted. | ||
67 | type CSVQuoteFields string | ||
68 | |||
69 | // Constants for csv quote styles. | ||
70 | const ( | ||
71 | CSVQuoteFieldsAlways CSVQuoteFields = "Always" | ||
72 | CSVQuoteFieldsAsNeeded CSVQuoteFields = "AsNeeded" | ||
73 | ) | ||
74 | |||
75 | // QueryExpressionType - is of what syntax the expression is, this should only | ||
76 | // be SQL | ||
77 | type QueryExpressionType string | ||
78 | |||
79 | // Constants for expression type. | ||
80 | const ( | ||
81 | QueryExpressionTypeSQL QueryExpressionType = "SQL" | ||
82 | ) | ||
83 | |||
84 | // JSONType determines json input serialization type. | ||
85 | type JSONType string | ||
86 | |||
87 | // Constants for JSONTypes. | ||
88 | const ( | ||
89 | JSONDocumentType JSONType = "DOCUMENT" | ||
90 | JSONLinesType JSONType = "LINES" | ||
91 | ) | ||
92 | |||
93 | // ParquetInputOptions parquet input specific options | ||
94 | type ParquetInputOptions struct{} | ||
95 | |||
96 | // CSVInputOptions csv input specific options | ||
97 | type CSVInputOptions struct { | ||
98 | FileHeaderInfo CSVFileHeaderInfo | ||
99 | fileHeaderInfoSet bool | ||
100 | |||
101 | RecordDelimiter string | ||
102 | recordDelimiterSet bool | ||
103 | |||
104 | FieldDelimiter string | ||
105 | fieldDelimiterSet bool | ||
106 | |||
107 | QuoteCharacter string | ||
108 | quoteCharacterSet bool | ||
109 | |||
110 | QuoteEscapeCharacter string | ||
111 | quoteEscapeCharacterSet bool | ||
112 | |||
113 | Comments string | ||
114 | commentsSet bool | ||
115 | } | ||
116 | |||
117 | // SetFileHeaderInfo sets the file header info in the CSV input options | ||
118 | func (c *CSVInputOptions) SetFileHeaderInfo(val CSVFileHeaderInfo) { | ||
119 | c.FileHeaderInfo = val | ||
120 | c.fileHeaderInfoSet = true | ||
121 | } | ||
122 | |||
123 | // SetRecordDelimiter sets the record delimiter in the CSV input options | ||
124 | func (c *CSVInputOptions) SetRecordDelimiter(val string) { | ||
125 | c.RecordDelimiter = val | ||
126 | c.recordDelimiterSet = true | ||
127 | } | ||
128 | |||
129 | // SetFieldDelimiter sets the field delimiter in the CSV input options | ||
130 | func (c *CSVInputOptions) SetFieldDelimiter(val string) { | ||
131 | c.FieldDelimiter = val | ||
132 | c.fieldDelimiterSet = true | ||
133 | } | ||
134 | |||
135 | // SetQuoteCharacter sets the quote character in the CSV input options | ||
136 | func (c *CSVInputOptions) SetQuoteCharacter(val string) { | ||
137 | c.QuoteCharacter = val | ||
138 | c.quoteCharacterSet = true | ||
139 | } | ||
140 | |||
141 | // SetQuoteEscapeCharacter sets the quote escape character in the CSV input options | ||
142 | func (c *CSVInputOptions) SetQuoteEscapeCharacter(val string) { | ||
143 | c.QuoteEscapeCharacter = val | ||
144 | c.quoteEscapeCharacterSet = true | ||
145 | } | ||
146 | |||
147 | // SetComments sets the comments character in the CSV input options | ||
148 | func (c *CSVInputOptions) SetComments(val string) { | ||
149 | c.Comments = val | ||
150 | c.commentsSet = true | ||
151 | } | ||
152 | |||
153 | // MarshalXML - produces the xml representation of the CSV input options struct | ||
154 | func (c CSVInputOptions) MarshalXML(e *xml.Encoder, start xml.StartElement) error { | ||
155 | if err := e.EncodeToken(start); err != nil { | ||
156 | return err | ||
157 | } | ||
158 | if c.FileHeaderInfo != "" || c.fileHeaderInfoSet { | ||
159 | if err := e.EncodeElement(c.FileHeaderInfo, xml.StartElement{Name: xml.Name{Local: "FileHeaderInfo"}}); err != nil { | ||
160 | return err | ||
161 | } | ||
162 | } | ||
163 | |||
164 | if c.RecordDelimiter != "" || c.recordDelimiterSet { | ||
165 | if err := e.EncodeElement(c.RecordDelimiter, xml.StartElement{Name: xml.Name{Local: "RecordDelimiter"}}); err != nil { | ||
166 | return err | ||
167 | } | ||
168 | } | ||
169 | |||
170 | if c.FieldDelimiter != "" || c.fieldDelimiterSet { | ||
171 | if err := e.EncodeElement(c.FieldDelimiter, xml.StartElement{Name: xml.Name{Local: "FieldDelimiter"}}); err != nil { | ||
172 | return err | ||
173 | } | ||
174 | } | ||
175 | |||
176 | if c.QuoteCharacter != "" || c.quoteCharacterSet { | ||
177 | if err := e.EncodeElement(c.QuoteCharacter, xml.StartElement{Name: xml.Name{Local: "QuoteCharacter"}}); err != nil { | ||
178 | return err | ||
179 | } | ||
180 | } | ||
181 | |||
182 | if c.QuoteEscapeCharacter != "" || c.quoteEscapeCharacterSet { | ||
183 | if err := e.EncodeElement(c.QuoteEscapeCharacter, xml.StartElement{Name: xml.Name{Local: "QuoteEscapeCharacter"}}); err != nil { | ||
184 | return err | ||
185 | } | ||
186 | } | ||
187 | |||
188 | if c.Comments != "" || c.commentsSet { | ||
189 | if err := e.EncodeElement(c.Comments, xml.StartElement{Name: xml.Name{Local: "Comments"}}); err != nil { | ||
190 | return err | ||
191 | } | ||
192 | } | ||
193 | |||
194 | return e.EncodeToken(xml.EndElement{Name: start.Name}) | ||
195 | } | ||
196 | |||
197 | // CSVOutputOptions csv output specific options | ||
198 | type CSVOutputOptions struct { | ||
199 | QuoteFields CSVQuoteFields | ||
200 | quoteFieldsSet bool | ||
201 | |||
202 | RecordDelimiter string | ||
203 | recordDelimiterSet bool | ||
204 | |||
205 | FieldDelimiter string | ||
206 | fieldDelimiterSet bool | ||
207 | |||
208 | QuoteCharacter string | ||
209 | quoteCharacterSet bool | ||
210 | |||
211 | QuoteEscapeCharacter string | ||
212 | quoteEscapeCharacterSet bool | ||
213 | } | ||
214 | |||
215 | // SetQuoteFields sets the quote field parameter in the CSV output options | ||
216 | func (c *CSVOutputOptions) SetQuoteFields(val CSVQuoteFields) { | ||
217 | c.QuoteFields = val | ||
218 | c.quoteFieldsSet = true | ||
219 | } | ||
220 | |||
221 | // SetRecordDelimiter sets the record delimiter character in the CSV output options | ||
222 | func (c *CSVOutputOptions) SetRecordDelimiter(val string) { | ||
223 | c.RecordDelimiter = val | ||
224 | c.recordDelimiterSet = true | ||
225 | } | ||
226 | |||
227 | // SetFieldDelimiter sets the field delimiter character in the CSV output options | ||
228 | func (c *CSVOutputOptions) SetFieldDelimiter(val string) { | ||
229 | c.FieldDelimiter = val | ||
230 | c.fieldDelimiterSet = true | ||
231 | } | ||
232 | |||
233 | // SetQuoteCharacter sets the quote character in the CSV output options | ||
234 | func (c *CSVOutputOptions) SetQuoteCharacter(val string) { | ||
235 | c.QuoteCharacter = val | ||
236 | c.quoteCharacterSet = true | ||
237 | } | ||
238 | |||
239 | // SetQuoteEscapeCharacter sets the quote escape character in the CSV output options | ||
240 | func (c *CSVOutputOptions) SetQuoteEscapeCharacter(val string) { | ||
241 | c.QuoteEscapeCharacter = val | ||
242 | c.quoteEscapeCharacterSet = true | ||
243 | } | ||
244 | |||
245 | // MarshalXML - produces the xml representation of the CSVOutputOptions struct | ||
246 | func (c CSVOutputOptions) MarshalXML(e *xml.Encoder, start xml.StartElement) error { | ||
247 | if err := e.EncodeToken(start); err != nil { | ||
248 | return err | ||
249 | } | ||
250 | |||
251 | if c.QuoteFields != "" || c.quoteFieldsSet { | ||
252 | if err := e.EncodeElement(c.QuoteFields, xml.StartElement{Name: xml.Name{Local: "QuoteFields"}}); err != nil { | ||
253 | return err | ||
254 | } | ||
255 | } | ||
256 | |||
257 | if c.RecordDelimiter != "" || c.recordDelimiterSet { | ||
258 | if err := e.EncodeElement(c.RecordDelimiter, xml.StartElement{Name: xml.Name{Local: "RecordDelimiter"}}); err != nil { | ||
259 | return err | ||
260 | } | ||
261 | } | ||
262 | |||
263 | if c.FieldDelimiter != "" || c.fieldDelimiterSet { | ||
264 | if err := e.EncodeElement(c.FieldDelimiter, xml.StartElement{Name: xml.Name{Local: "FieldDelimiter"}}); err != nil { | ||
265 | return err | ||
266 | } | ||
267 | } | ||
268 | |||
269 | if c.QuoteCharacter != "" || c.quoteCharacterSet { | ||
270 | if err := e.EncodeElement(c.QuoteCharacter, xml.StartElement{Name: xml.Name{Local: "QuoteCharacter"}}); err != nil { | ||
271 | return err | ||
272 | } | ||
273 | } | ||
274 | |||
275 | if c.QuoteEscapeCharacter != "" || c.quoteEscapeCharacterSet { | ||
276 | if err := e.EncodeElement(c.QuoteEscapeCharacter, xml.StartElement{Name: xml.Name{Local: "QuoteEscapeCharacter"}}); err != nil { | ||
277 | return err | ||
278 | } | ||
279 | } | ||
280 | |||
281 | return e.EncodeToken(xml.EndElement{Name: start.Name}) | ||
282 | } | ||
283 | |||
284 | // JSONInputOptions json input specific options | ||
285 | type JSONInputOptions struct { | ||
286 | Type JSONType | ||
287 | typeSet bool | ||
288 | } | ||
289 | |||
290 | // SetType sets the JSON type in the JSON input options | ||
291 | func (j *JSONInputOptions) SetType(typ JSONType) { | ||
292 | j.Type = typ | ||
293 | j.typeSet = true | ||
294 | } | ||
295 | |||
296 | // MarshalXML - produces the xml representation of the JSONInputOptions struct | ||
297 | func (j JSONInputOptions) MarshalXML(e *xml.Encoder, start xml.StartElement) error { | ||
298 | if err := e.EncodeToken(start); err != nil { | ||
299 | return err | ||
300 | } | ||
301 | |||
302 | if j.Type != "" || j.typeSet { | ||
303 | if err := e.EncodeElement(j.Type, xml.StartElement{Name: xml.Name{Local: "Type"}}); err != nil { | ||
304 | return err | ||
305 | } | ||
306 | } | ||
307 | |||
308 | return e.EncodeToken(xml.EndElement{Name: start.Name}) | ||
309 | } | ||
310 | |||
311 | // JSONOutputOptions - json output specific options | ||
312 | type JSONOutputOptions struct { | ||
313 | RecordDelimiter string | ||
314 | recordDelimiterSet bool | ||
315 | } | ||
316 | |||
317 | // SetRecordDelimiter sets the record delimiter in the JSON output options | ||
318 | func (j *JSONOutputOptions) SetRecordDelimiter(val string) { | ||
319 | j.RecordDelimiter = val | ||
320 | j.recordDelimiterSet = true | ||
321 | } | ||
322 | |||
323 | // MarshalXML - produces the xml representation of the JSONOutputOptions struct | ||
324 | func (j JSONOutputOptions) MarshalXML(e *xml.Encoder, start xml.StartElement) error { | ||
325 | if err := e.EncodeToken(start); err != nil { | ||
326 | return err | ||
327 | } | ||
328 | |||
329 | if j.RecordDelimiter != "" || j.recordDelimiterSet { | ||
330 | if err := e.EncodeElement(j.RecordDelimiter, xml.StartElement{Name: xml.Name{Local: "RecordDelimiter"}}); err != nil { | ||
331 | return err | ||
332 | } | ||
333 | } | ||
334 | |||
335 | return e.EncodeToken(xml.EndElement{Name: start.Name}) | ||
336 | } | ||
337 | |||
338 | // SelectObjectInputSerialization - input serialization parameters | ||
339 | type SelectObjectInputSerialization struct { | ||
340 | CompressionType SelectCompressionType `xml:"CompressionType,omitempty"` | ||
341 | Parquet *ParquetInputOptions `xml:"Parquet,omitempty"` | ||
342 | CSV *CSVInputOptions `xml:"CSV,omitempty"` | ||
343 | JSON *JSONInputOptions `xml:"JSON,omitempty"` | ||
344 | } | ||
345 | |||
346 | // SelectObjectOutputSerialization - output serialization parameters. | ||
347 | type SelectObjectOutputSerialization struct { | ||
348 | CSV *CSVOutputOptions `xml:"CSV,omitempty"` | ||
349 | JSON *JSONOutputOptions `xml:"JSON,omitempty"` | ||
350 | } | ||
351 | |||
352 | // SelectObjectOptions - represents the input select body | ||
353 | type SelectObjectOptions struct { | ||
354 | XMLName xml.Name `xml:"SelectObjectContentRequest" json:"-"` | ||
355 | ServerSideEncryption encrypt.ServerSide `xml:"-"` | ||
356 | Expression string | ||
357 | ExpressionType QueryExpressionType | ||
358 | InputSerialization SelectObjectInputSerialization | ||
359 | OutputSerialization SelectObjectOutputSerialization | ||
360 | RequestProgress struct { | ||
361 | Enabled bool | ||
362 | } | ||
363 | } | ||
364 | |||
365 | // Header returns the http.Header representation of the SelectObject options. | ||
366 | func (o SelectObjectOptions) Header() http.Header { | ||
367 | headers := make(http.Header) | ||
368 | if o.ServerSideEncryption != nil && o.ServerSideEncryption.Type() == encrypt.SSEC { | ||
369 | o.ServerSideEncryption.Marshal(headers) | ||
370 | } | ||
371 | return headers | ||
372 | } | ||
373 | |||
374 | // SelectObjectType - is the parameter which defines what type of object the | ||
375 | // operation is being performed on. | ||
376 | type SelectObjectType string | ||
377 | |||
378 | // Constants for input data types. | ||
379 | const ( | ||
380 | SelectObjectTypeCSV SelectObjectType = "CSV" | ||
381 | SelectObjectTypeJSON SelectObjectType = "JSON" | ||
382 | SelectObjectTypeParquet SelectObjectType = "Parquet" | ||
383 | ) | ||
384 | |||
385 | // preludeInfo is used for keeping track of necessary information from the | ||
386 | // prelude. | ||
387 | type preludeInfo struct { | ||
388 | totalLen uint32 | ||
389 | headerLen uint32 | ||
390 | } | ||
391 | |||
392 | // SelectResults is used for the streaming responses from the server. | ||
393 | type SelectResults struct { | ||
394 | pipeReader *io.PipeReader | ||
395 | resp *http.Response | ||
396 | stats *StatsMessage | ||
397 | progress *ProgressMessage | ||
398 | } | ||
399 | |||
400 | // ProgressMessage is a struct for progress xml message. | ||
401 | type ProgressMessage struct { | ||
402 | XMLName xml.Name `xml:"Progress" json:"-"` | ||
403 | StatsMessage | ||
404 | } | ||
405 | |||
406 | // StatsMessage is a struct for stat xml message. | ||
407 | type StatsMessage struct { | ||
408 | XMLName xml.Name `xml:"Stats" json:"-"` | ||
409 | BytesScanned int64 | ||
410 | BytesProcessed int64 | ||
411 | BytesReturned int64 | ||
412 | } | ||
413 | |||
414 | // messageType represents the type of message. | ||
415 | type messageType string | ||
416 | |||
417 | const ( | ||
418 | errorMsg messageType = "error" | ||
419 | commonMsg messageType = "event" | ||
420 | ) | ||
421 | |||
422 | // eventType represents the type of event. | ||
423 | type eventType string | ||
424 | |||
425 | // list of event-types returned by Select API. | ||
426 | const ( | ||
427 | endEvent eventType = "End" | ||
428 | recordsEvent eventType = "Records" | ||
429 | progressEvent eventType = "Progress" | ||
430 | statsEvent eventType = "Stats" | ||
431 | ) | ||
432 | |||
433 | // contentType represents content type of event. | ||
434 | type contentType string | ||
435 | |||
436 | const ( | ||
437 | xmlContent contentType = "text/xml" | ||
438 | ) | ||
439 | |||
440 | // SelectObjectContent is a implementation of http://docs.aws.amazon.com/AmazonS3/latest/API/RESTObjectSELECTContent.html AWS S3 API. | ||
441 | func (c *Client) SelectObjectContent(ctx context.Context, bucketName, objectName string, opts SelectObjectOptions) (*SelectResults, error) { | ||
442 | // Input validation. | ||
443 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
444 | return nil, err | ||
445 | } | ||
446 | if err := s3utils.CheckValidObjectName(objectName); err != nil { | ||
447 | return nil, err | ||
448 | } | ||
449 | |||
450 | selectReqBytes, err := xml.Marshal(opts) | ||
451 | if err != nil { | ||
452 | return nil, err | ||
453 | } | ||
454 | |||
455 | urlValues := make(url.Values) | ||
456 | urlValues.Set("select", "") | ||
457 | urlValues.Set("select-type", "2") | ||
458 | |||
459 | // Execute POST on bucket/object. | ||
460 | resp, err := c.executeMethod(ctx, http.MethodPost, requestMetadata{ | ||
461 | bucketName: bucketName, | ||
462 | objectName: objectName, | ||
463 | queryValues: urlValues, | ||
464 | customHeader: opts.Header(), | ||
465 | contentMD5Base64: sumMD5Base64(selectReqBytes), | ||
466 | contentSHA256Hex: sum256Hex(selectReqBytes), | ||
467 | contentBody: bytes.NewReader(selectReqBytes), | ||
468 | contentLength: int64(len(selectReqBytes)), | ||
469 | }) | ||
470 | if err != nil { | ||
471 | return nil, err | ||
472 | } | ||
473 | |||
474 | return NewSelectResults(resp, bucketName) | ||
475 | } | ||
476 | |||
477 | // NewSelectResults creates a Select Result parser that parses the response | ||
478 | // and returns a Reader that will return parsed and assembled select output. | ||
479 | func NewSelectResults(resp *http.Response, bucketName string) (*SelectResults, error) { | ||
480 | if resp.StatusCode != http.StatusOK { | ||
481 | return nil, httpRespToErrorResponse(resp, bucketName, "") | ||
482 | } | ||
483 | |||
484 | pipeReader, pipeWriter := io.Pipe() | ||
485 | streamer := &SelectResults{ | ||
486 | resp: resp, | ||
487 | stats: &StatsMessage{}, | ||
488 | progress: &ProgressMessage{}, | ||
489 | pipeReader: pipeReader, | ||
490 | } | ||
491 | streamer.start(pipeWriter) | ||
492 | return streamer, nil | ||
493 | } | ||
494 | |||
495 | // Close - closes the underlying response body and the stream reader. | ||
496 | func (s *SelectResults) Close() error { | ||
497 | defer closeResponse(s.resp) | ||
498 | return s.pipeReader.Close() | ||
499 | } | ||
500 | |||
501 | // Read - is a reader compatible implementation for SelectObjectContent records. | ||
502 | func (s *SelectResults) Read(b []byte) (n int, err error) { | ||
503 | return s.pipeReader.Read(b) | ||
504 | } | ||
505 | |||
506 | // Stats - information about a request's stats when processing is complete. | ||
507 | func (s *SelectResults) Stats() *StatsMessage { | ||
508 | return s.stats | ||
509 | } | ||
510 | |||
511 | // Progress - information about the progress of a request. | ||
512 | func (s *SelectResults) Progress() *ProgressMessage { | ||
513 | return s.progress | ||
514 | } | ||
515 | |||
516 | // start is the main function that decodes the large byte array into | ||
517 | // several events that are sent through the eventstream. | ||
518 | func (s *SelectResults) start(pipeWriter *io.PipeWriter) { | ||
519 | go func() { | ||
520 | for { | ||
521 | var prelude preludeInfo | ||
522 | headers := make(http.Header) | ||
523 | var err error | ||
524 | |||
525 | // Create CRC code | ||
526 | crc := crc32.New(crc32.IEEETable) | ||
527 | crcReader := io.TeeReader(s.resp.Body, crc) | ||
528 | |||
529 | // Extract the prelude(12 bytes) into a struct to extract relevant information. | ||
530 | prelude, err = processPrelude(crcReader, crc) | ||
531 | if err != nil { | ||
532 | pipeWriter.CloseWithError(err) | ||
533 | closeResponse(s.resp) | ||
534 | return | ||
535 | } | ||
536 | |||
537 | // Extract the headers(variable bytes) into a struct to extract relevant information | ||
538 | if prelude.headerLen > 0 { | ||
539 | if err = extractHeader(io.LimitReader(crcReader, int64(prelude.headerLen)), headers); err != nil { | ||
540 | pipeWriter.CloseWithError(err) | ||
541 | closeResponse(s.resp) | ||
542 | return | ||
543 | } | ||
544 | } | ||
545 | |||
546 | // Get the actual payload length so that the appropriate amount of | ||
547 | // bytes can be read or parsed. | ||
548 | payloadLen := prelude.PayloadLen() | ||
549 | |||
550 | m := messageType(headers.Get("message-type")) | ||
551 | |||
552 | switch m { | ||
553 | case errorMsg: | ||
554 | pipeWriter.CloseWithError(errors.New(headers.Get("error-code") + ":\"" + headers.Get("error-message") + "\"")) | ||
555 | closeResponse(s.resp) | ||
556 | return | ||
557 | case commonMsg: | ||
558 | // Get content-type of the payload. | ||
559 | c := contentType(headers.Get("content-type")) | ||
560 | |||
561 | // Get event type of the payload. | ||
562 | e := eventType(headers.Get("event-type")) | ||
563 | |||
564 | // Handle all supported events. | ||
565 | switch e { | ||
566 | case endEvent: | ||
567 | pipeWriter.Close() | ||
568 | closeResponse(s.resp) | ||
569 | return | ||
570 | case recordsEvent: | ||
571 | if _, err = io.Copy(pipeWriter, io.LimitReader(crcReader, payloadLen)); err != nil { | ||
572 | pipeWriter.CloseWithError(err) | ||
573 | closeResponse(s.resp) | ||
574 | return | ||
575 | } | ||
576 | case progressEvent: | ||
577 | switch c { | ||
578 | case xmlContent: | ||
579 | if err = xmlDecoder(io.LimitReader(crcReader, payloadLen), s.progress); err != nil { | ||
580 | pipeWriter.CloseWithError(err) | ||
581 | closeResponse(s.resp) | ||
582 | return | ||
583 | } | ||
584 | default: | ||
585 | pipeWriter.CloseWithError(fmt.Errorf("Unexpected content-type %s sent for event-type %s", c, progressEvent)) | ||
586 | closeResponse(s.resp) | ||
587 | return | ||
588 | } | ||
589 | case statsEvent: | ||
590 | switch c { | ||
591 | case xmlContent: | ||
592 | if err = xmlDecoder(io.LimitReader(crcReader, payloadLen), s.stats); err != nil { | ||
593 | pipeWriter.CloseWithError(err) | ||
594 | closeResponse(s.resp) | ||
595 | return | ||
596 | } | ||
597 | default: | ||
598 | pipeWriter.CloseWithError(fmt.Errorf("Unexpected content-type %s sent for event-type %s", c, statsEvent)) | ||
599 | closeResponse(s.resp) | ||
600 | return | ||
601 | } | ||
602 | } | ||
603 | } | ||
604 | |||
605 | // Ensures that the full message's CRC is correct and | ||
606 | // that the message is not corrupted | ||
607 | if err := checkCRC(s.resp.Body, crc.Sum32()); err != nil { | ||
608 | pipeWriter.CloseWithError(err) | ||
609 | closeResponse(s.resp) | ||
610 | return | ||
611 | } | ||
612 | |||
613 | } | ||
614 | }() | ||
615 | } | ||
616 | |||
617 | // PayloadLen is a function that calculates the length of the payload. | ||
618 | func (p preludeInfo) PayloadLen() int64 { | ||
619 | return int64(p.totalLen - p.headerLen - 16) | ||
620 | } | ||
621 | |||
622 | // processPrelude is the function that reads the 12 bytes of the prelude and | ||
623 | // ensures the CRC is correct while also extracting relevant information into | ||
624 | // the struct, | ||
625 | func processPrelude(prelude io.Reader, crc hash.Hash32) (preludeInfo, error) { | ||
626 | var err error | ||
627 | pInfo := preludeInfo{} | ||
628 | |||
629 | // reads total length of the message (first 4 bytes) | ||
630 | pInfo.totalLen, err = extractUint32(prelude) | ||
631 | if err != nil { | ||
632 | return pInfo, err | ||
633 | } | ||
634 | |||
635 | // reads total header length of the message (2nd 4 bytes) | ||
636 | pInfo.headerLen, err = extractUint32(prelude) | ||
637 | if err != nil { | ||
638 | return pInfo, err | ||
639 | } | ||
640 | |||
641 | // checks that the CRC is correct (3rd 4 bytes) | ||
642 | preCRC := crc.Sum32() | ||
643 | if err := checkCRC(prelude, preCRC); err != nil { | ||
644 | return pInfo, err | ||
645 | } | ||
646 | |||
647 | return pInfo, nil | ||
648 | } | ||
649 | |||
650 | // extracts the relevant information from the Headers. | ||
651 | func extractHeader(body io.Reader, myHeaders http.Header) error { | ||
652 | for { | ||
653 | // extracts the first part of the header, | ||
654 | headerTypeName, err := extractHeaderType(body) | ||
655 | if err != nil { | ||
656 | // Since end of file, we have read all of our headers | ||
657 | if err == io.EOF { | ||
658 | break | ||
659 | } | ||
660 | return err | ||
661 | } | ||
662 | |||
663 | // reads the 7 present in the header and ignores it. | ||
664 | extractUint8(body) | ||
665 | |||
666 | headerValueName, err := extractHeaderValue(body) | ||
667 | if err != nil { | ||
668 | return err | ||
669 | } | ||
670 | |||
671 | myHeaders.Set(headerTypeName, headerValueName) | ||
672 | |||
673 | } | ||
674 | return nil | ||
675 | } | ||
676 | |||
677 | // extractHeaderType extracts the first half of the header message, the header type. | ||
678 | func extractHeaderType(body io.Reader) (string, error) { | ||
679 | // extracts 2 bit integer | ||
680 | headerNameLen, err := extractUint8(body) | ||
681 | if err != nil { | ||
682 | return "", err | ||
683 | } | ||
684 | // extracts the string with the appropriate number of bytes | ||
685 | headerName, err := extractString(body, int(headerNameLen)) | ||
686 | if err != nil { | ||
687 | return "", err | ||
688 | } | ||
689 | return strings.TrimPrefix(headerName, ":"), nil | ||
690 | } | ||
691 | |||
692 | // extractsHeaderValue extracts the second half of the header message, the | ||
693 | // header value | ||
694 | func extractHeaderValue(body io.Reader) (string, error) { | ||
695 | bodyLen, err := extractUint16(body) | ||
696 | if err != nil { | ||
697 | return "", err | ||
698 | } | ||
699 | bodyName, err := extractString(body, int(bodyLen)) | ||
700 | if err != nil { | ||
701 | return "", err | ||
702 | } | ||
703 | return bodyName, nil | ||
704 | } | ||
705 | |||
706 | // extracts a string from byte array of a particular number of bytes. | ||
707 | func extractString(source io.Reader, lenBytes int) (string, error) { | ||
708 | myVal := make([]byte, lenBytes) | ||
709 | _, err := source.Read(myVal) | ||
710 | if err != nil { | ||
711 | return "", err | ||
712 | } | ||
713 | return string(myVal), nil | ||
714 | } | ||
715 | |||
716 | // extractUint32 extracts a 4 byte integer from the byte array. | ||
717 | func extractUint32(r io.Reader) (uint32, error) { | ||
718 | buf := make([]byte, 4) | ||
719 | _, err := readFull(r, buf) | ||
720 | if err != nil { | ||
721 | return 0, err | ||
722 | } | ||
723 | return binary.BigEndian.Uint32(buf), nil | ||
724 | } | ||
725 | |||
726 | // extractUint16 extracts a 2 byte integer from the byte array. | ||
727 | func extractUint16(r io.Reader) (uint16, error) { | ||
728 | buf := make([]byte, 2) | ||
729 | _, err := readFull(r, buf) | ||
730 | if err != nil { | ||
731 | return 0, err | ||
732 | } | ||
733 | return binary.BigEndian.Uint16(buf), nil | ||
734 | } | ||
735 | |||
736 | // extractUint8 extracts a 1 byte integer from the byte array. | ||
737 | func extractUint8(r io.Reader) (uint8, error) { | ||
738 | buf := make([]byte, 1) | ||
739 | _, err := readFull(r, buf) | ||
740 | if err != nil { | ||
741 | return 0, err | ||
742 | } | ||
743 | return buf[0], nil | ||
744 | } | ||
745 | |||
746 | // checkCRC ensures that the CRC matches with the one from the reader. | ||
747 | func checkCRC(r io.Reader, expect uint32) error { | ||
748 | msgCRC, err := extractUint32(r) | ||
749 | if err != nil { | ||
750 | return err | ||
751 | } | ||
752 | |||
753 | if msgCRC != expect { | ||
754 | return fmt.Errorf("Checksum Mismatch, MessageCRC of 0x%X does not equal expected CRC of 0x%X", msgCRC, expect) | ||
755 | } | ||
756 | return nil | ||
757 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/api-stat.go b/vendor/github.com/minio/minio-go/v7/api-stat.go deleted file mode 100644 index b043dc4..0000000 --- a/vendor/github.com/minio/minio-go/v7/api-stat.go +++ /dev/null | |||
@@ -1,116 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2015-2020 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package minio | ||
19 | |||
20 | import ( | ||
21 | "context" | ||
22 | "net/http" | ||
23 | |||
24 | "github.com/minio/minio-go/v7/pkg/s3utils" | ||
25 | ) | ||
26 | |||
27 | // BucketExists verifies if bucket exists and you have permission to access it. Allows for a Context to | ||
28 | // control cancellations and timeouts. | ||
29 | func (c *Client) BucketExists(ctx context.Context, bucketName string) (bool, error) { | ||
30 | // Input validation. | ||
31 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
32 | return false, err | ||
33 | } | ||
34 | |||
35 | // Execute HEAD on bucketName. | ||
36 | resp, err := c.executeMethod(ctx, http.MethodHead, requestMetadata{ | ||
37 | bucketName: bucketName, | ||
38 | contentSHA256Hex: emptySHA256Hex, | ||
39 | }) | ||
40 | defer closeResponse(resp) | ||
41 | if err != nil { | ||
42 | if ToErrorResponse(err).Code == "NoSuchBucket" { | ||
43 | return false, nil | ||
44 | } | ||
45 | return false, err | ||
46 | } | ||
47 | if resp != nil { | ||
48 | resperr := httpRespToErrorResponse(resp, bucketName, "") | ||
49 | if ToErrorResponse(resperr).Code == "NoSuchBucket" { | ||
50 | return false, nil | ||
51 | } | ||
52 | if resp.StatusCode != http.StatusOK { | ||
53 | return false, httpRespToErrorResponse(resp, bucketName, "") | ||
54 | } | ||
55 | } | ||
56 | return true, nil | ||
57 | } | ||
58 | |||
59 | // StatObject verifies if object exists, you have permission to access it | ||
60 | // and returns information about the object. | ||
61 | func (c *Client) StatObject(ctx context.Context, bucketName, objectName string, opts StatObjectOptions) (ObjectInfo, error) { | ||
62 | // Input validation. | ||
63 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
64 | return ObjectInfo{}, err | ||
65 | } | ||
66 | if err := s3utils.CheckValidObjectName(objectName); err != nil { | ||
67 | return ObjectInfo{}, err | ||
68 | } | ||
69 | headers := opts.Header() | ||
70 | if opts.Internal.ReplicationDeleteMarker { | ||
71 | headers.Set(minIOBucketReplicationDeleteMarker, "true") | ||
72 | } | ||
73 | if opts.Internal.IsReplicationReadyForDeleteMarker { | ||
74 | headers.Set(isMinioTgtReplicationReady, "true") | ||
75 | } | ||
76 | |||
77 | // Execute HEAD on objectName. | ||
78 | resp, err := c.executeMethod(ctx, http.MethodHead, requestMetadata{ | ||
79 | bucketName: bucketName, | ||
80 | objectName: objectName, | ||
81 | queryValues: opts.toQueryValues(), | ||
82 | contentSHA256Hex: emptySHA256Hex, | ||
83 | customHeader: headers, | ||
84 | }) | ||
85 | defer closeResponse(resp) | ||
86 | if err != nil { | ||
87 | return ObjectInfo{}, err | ||
88 | } | ||
89 | |||
90 | if resp != nil { | ||
91 | deleteMarker := resp.Header.Get(amzDeleteMarker) == "true" | ||
92 | replicationReady := resp.Header.Get(minioTgtReplicationReady) == "true" | ||
93 | if resp.StatusCode != http.StatusOK && resp.StatusCode != http.StatusPartialContent { | ||
94 | if resp.StatusCode == http.StatusMethodNotAllowed && opts.VersionID != "" && deleteMarker { | ||
95 | errResp := ErrorResponse{ | ||
96 | StatusCode: resp.StatusCode, | ||
97 | Code: "MethodNotAllowed", | ||
98 | Message: "The specified method is not allowed against this resource.", | ||
99 | BucketName: bucketName, | ||
100 | Key: objectName, | ||
101 | } | ||
102 | return ObjectInfo{ | ||
103 | VersionID: resp.Header.Get(amzVersionID), | ||
104 | IsDeleteMarker: deleteMarker, | ||
105 | }, errResp | ||
106 | } | ||
107 | return ObjectInfo{ | ||
108 | VersionID: resp.Header.Get(amzVersionID), | ||
109 | IsDeleteMarker: deleteMarker, | ||
110 | ReplicationReady: replicationReady, // whether delete marker can be replicated | ||
111 | }, httpRespToErrorResponse(resp, bucketName, objectName) | ||
112 | } | ||
113 | } | ||
114 | |||
115 | return ToObjectInfo(bucketName, objectName, resp.Header) | ||
116 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/api.go b/vendor/github.com/minio/minio-go/v7/api.go deleted file mode 100644 index f8a9b34..0000000 --- a/vendor/github.com/minio/minio-go/v7/api.go +++ /dev/null | |||
@@ -1,995 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2015-2023 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package minio | ||
19 | |||
20 | import ( | ||
21 | "bytes" | ||
22 | "context" | ||
23 | "encoding/base64" | ||
24 | "errors" | ||
25 | "fmt" | ||
26 | "hash/crc32" | ||
27 | "io" | ||
28 | "math/rand" | ||
29 | "net" | ||
30 | "net/http" | ||
31 | "net/http/cookiejar" | ||
32 | "net/http/httptrace" | ||
33 | "net/http/httputil" | ||
34 | "net/url" | ||
35 | "os" | ||
36 | "runtime" | ||
37 | "strings" | ||
38 | "sync" | ||
39 | "sync/atomic" | ||
40 | "time" | ||
41 | |||
42 | md5simd "github.com/minio/md5-simd" | ||
43 | "github.com/minio/minio-go/v7/pkg/credentials" | ||
44 | "github.com/minio/minio-go/v7/pkg/s3utils" | ||
45 | "github.com/minio/minio-go/v7/pkg/signer" | ||
46 | "golang.org/x/net/publicsuffix" | ||
47 | ) | ||
48 | |||
49 | // Client implements Amazon S3 compatible methods. | ||
50 | type Client struct { | ||
51 | // Standard options. | ||
52 | |||
53 | // Parsed endpoint url provided by the user. | ||
54 | endpointURL *url.URL | ||
55 | |||
56 | // Holds various credential providers. | ||
57 | credsProvider *credentials.Credentials | ||
58 | |||
59 | // Custom signerType value overrides all credentials. | ||
60 | overrideSignerType credentials.SignatureType | ||
61 | |||
62 | // User supplied. | ||
63 | appInfo struct { | ||
64 | appName string | ||
65 | appVersion string | ||
66 | } | ||
67 | |||
68 | // Indicate whether we are using https or not | ||
69 | secure bool | ||
70 | |||
71 | // Needs allocation. | ||
72 | httpClient *http.Client | ||
73 | httpTrace *httptrace.ClientTrace | ||
74 | bucketLocCache *bucketLocationCache | ||
75 | |||
76 | // Advanced functionality. | ||
77 | isTraceEnabled bool | ||
78 | traceErrorsOnly bool | ||
79 | traceOutput io.Writer | ||
80 | |||
81 | // S3 specific accelerated endpoint. | ||
82 | s3AccelerateEndpoint string | ||
83 | |||
84 | // Region endpoint | ||
85 | region string | ||
86 | |||
87 | // Random seed. | ||
88 | random *rand.Rand | ||
89 | |||
90 | // lookup indicates type of url lookup supported by server. If not specified, | ||
91 | // default to Auto. | ||
92 | lookup BucketLookupType | ||
93 | |||
94 | // Factory for MD5 hash functions. | ||
95 | md5Hasher func() md5simd.Hasher | ||
96 | sha256Hasher func() md5simd.Hasher | ||
97 | |||
98 | healthStatus int32 | ||
99 | |||
100 | trailingHeaderSupport bool | ||
101 | } | ||
102 | |||
103 | // Options for New method | ||
104 | type Options struct { | ||
105 | Creds *credentials.Credentials | ||
106 | Secure bool | ||
107 | Transport http.RoundTripper | ||
108 | Trace *httptrace.ClientTrace | ||
109 | Region string | ||
110 | BucketLookup BucketLookupType | ||
111 | |||
112 | // Allows setting a custom region lookup based on URL pattern | ||
113 | // not all URL patterns are covered by this library so if you | ||
114 | // have a custom endpoints with many regions you can use this | ||
115 | // function to perform region lookups appropriately. | ||
116 | CustomRegionViaURL func(u url.URL) string | ||
117 | |||
118 | // TrailingHeaders indicates server support of trailing headers. | ||
119 | // Only supported for v4 signatures. | ||
120 | TrailingHeaders bool | ||
121 | |||
122 | // Custom hash routines. Leave nil to use standard. | ||
123 | CustomMD5 func() md5simd.Hasher | ||
124 | CustomSHA256 func() md5simd.Hasher | ||
125 | } | ||
126 | |||
127 | // Global constants. | ||
128 | const ( | ||
129 | libraryName = "minio-go" | ||
130 | libraryVersion = "v7.0.66" | ||
131 | ) | ||
132 | |||
133 | // User Agent should always following the below style. | ||
134 | // Please open an issue to discuss any new changes here. | ||
135 | // | ||
136 | // MinIO (OS; ARCH) LIB/VER APP/VER | ||
137 | const ( | ||
138 | libraryUserAgentPrefix = "MinIO (" + runtime.GOOS + "; " + runtime.GOARCH + ") " | ||
139 | libraryUserAgent = libraryUserAgentPrefix + libraryName + "/" + libraryVersion | ||
140 | ) | ||
141 | |||
142 | // BucketLookupType is type of url lookup supported by server. | ||
143 | type BucketLookupType int | ||
144 | |||
145 | // Different types of url lookup supported by the server.Initialized to BucketLookupAuto | ||
146 | const ( | ||
147 | BucketLookupAuto BucketLookupType = iota | ||
148 | BucketLookupDNS | ||
149 | BucketLookupPath | ||
150 | ) | ||
151 | |||
152 | // New - instantiate minio client with options | ||
153 | func New(endpoint string, opts *Options) (*Client, error) { | ||
154 | if opts == nil { | ||
155 | return nil, errors.New("no options provided") | ||
156 | } | ||
157 | clnt, err := privateNew(endpoint, opts) | ||
158 | if err != nil { | ||
159 | return nil, err | ||
160 | } | ||
161 | // If Amazon S3 set to signature v4. | ||
162 | if s3utils.IsAmazonEndpoint(*clnt.endpointURL) { | ||
163 | clnt.overrideSignerType = credentials.SignatureV4 | ||
164 | } | ||
165 | |||
166 | return clnt, nil | ||
167 | } | ||
168 | |||
169 | // EndpointURL returns the URL of the S3 endpoint. | ||
170 | func (c *Client) EndpointURL() *url.URL { | ||
171 | endpoint := *c.endpointURL // copy to prevent callers from modifying internal state | ||
172 | return &endpoint | ||
173 | } | ||
174 | |||
175 | // lockedRandSource provides protected rand source, implements rand.Source interface. | ||
176 | type lockedRandSource struct { | ||
177 | lk sync.Mutex | ||
178 | src rand.Source | ||
179 | } | ||
180 | |||
181 | // Int63 returns a non-negative pseudo-random 63-bit integer as an int64. | ||
182 | func (r *lockedRandSource) Int63() (n int64) { | ||
183 | r.lk.Lock() | ||
184 | n = r.src.Int63() | ||
185 | r.lk.Unlock() | ||
186 | return | ||
187 | } | ||
188 | |||
189 | // Seed uses the provided seed value to initialize the generator to a | ||
190 | // deterministic state. | ||
191 | func (r *lockedRandSource) Seed(seed int64) { | ||
192 | r.lk.Lock() | ||
193 | r.src.Seed(seed) | ||
194 | r.lk.Unlock() | ||
195 | } | ||
196 | |||
197 | func privateNew(endpoint string, opts *Options) (*Client, error) { | ||
198 | // construct endpoint. | ||
199 | endpointURL, err := getEndpointURL(endpoint, opts.Secure) | ||
200 | if err != nil { | ||
201 | return nil, err | ||
202 | } | ||
203 | |||
204 | // Initialize cookies to preserve server sent cookies if any and replay | ||
205 | // them upon each request. | ||
206 | jar, err := cookiejar.New(&cookiejar.Options{PublicSuffixList: publicsuffix.List}) | ||
207 | if err != nil { | ||
208 | return nil, err | ||
209 | } | ||
210 | |||
211 | // instantiate new Client. | ||
212 | clnt := new(Client) | ||
213 | |||
214 | // Save the credentials. | ||
215 | clnt.credsProvider = opts.Creds | ||
216 | |||
217 | // Remember whether we are using https or not | ||
218 | clnt.secure = opts.Secure | ||
219 | |||
220 | // Save endpoint URL, user agent for future uses. | ||
221 | clnt.endpointURL = endpointURL | ||
222 | |||
223 | transport := opts.Transport | ||
224 | if transport == nil { | ||
225 | transport, err = DefaultTransport(opts.Secure) | ||
226 | if err != nil { | ||
227 | return nil, err | ||
228 | } | ||
229 | } | ||
230 | |||
231 | clnt.httpTrace = opts.Trace | ||
232 | |||
233 | // Instantiate http client and bucket location cache. | ||
234 | clnt.httpClient = &http.Client{ | ||
235 | Jar: jar, | ||
236 | Transport: transport, | ||
237 | CheckRedirect: func(req *http.Request, via []*http.Request) error { | ||
238 | return http.ErrUseLastResponse | ||
239 | }, | ||
240 | } | ||
241 | |||
242 | // Sets custom region, if region is empty bucket location cache is used automatically. | ||
243 | if opts.Region == "" { | ||
244 | if opts.CustomRegionViaURL != nil { | ||
245 | opts.Region = opts.CustomRegionViaURL(*clnt.endpointURL) | ||
246 | } else { | ||
247 | opts.Region = s3utils.GetRegionFromURL(*clnt.endpointURL) | ||
248 | } | ||
249 | } | ||
250 | clnt.region = opts.Region | ||
251 | |||
252 | // Instantiate bucket location cache. | ||
253 | clnt.bucketLocCache = newBucketLocationCache() | ||
254 | |||
255 | // Introduce a new locked random seed. | ||
256 | clnt.random = rand.New(&lockedRandSource{src: rand.NewSource(time.Now().UTC().UnixNano())}) | ||
257 | |||
258 | // Add default md5 hasher. | ||
259 | clnt.md5Hasher = opts.CustomMD5 | ||
260 | clnt.sha256Hasher = opts.CustomSHA256 | ||
261 | if clnt.md5Hasher == nil { | ||
262 | clnt.md5Hasher = newMd5Hasher | ||
263 | } | ||
264 | if clnt.sha256Hasher == nil { | ||
265 | clnt.sha256Hasher = newSHA256Hasher | ||
266 | } | ||
267 | |||
268 | clnt.trailingHeaderSupport = opts.TrailingHeaders && clnt.overrideSignerType.IsV4() | ||
269 | |||
270 | // Sets bucket lookup style, whether server accepts DNS or Path lookup. Default is Auto - determined | ||
271 | // by the SDK. When Auto is specified, DNS lookup is used for Amazon/Google cloud endpoints and Path for all other endpoints. | ||
272 | clnt.lookup = opts.BucketLookup | ||
273 | |||
274 | // healthcheck is not initialized | ||
275 | clnt.healthStatus = unknown | ||
276 | |||
277 | // Return. | ||
278 | return clnt, nil | ||
279 | } | ||
280 | |||
281 | // SetAppInfo - add application details to user agent. | ||
282 | func (c *Client) SetAppInfo(appName, appVersion string) { | ||
283 | // if app name and version not set, we do not set a new user agent. | ||
284 | if appName != "" && appVersion != "" { | ||
285 | c.appInfo.appName = appName | ||
286 | c.appInfo.appVersion = appVersion | ||
287 | } | ||
288 | } | ||
289 | |||
290 | // TraceOn - enable HTTP tracing. | ||
291 | func (c *Client) TraceOn(outputStream io.Writer) { | ||
292 | // if outputStream is nil then default to os.Stdout. | ||
293 | if outputStream == nil { | ||
294 | outputStream = os.Stdout | ||
295 | } | ||
296 | // Sets a new output stream. | ||
297 | c.traceOutput = outputStream | ||
298 | |||
299 | // Enable tracing. | ||
300 | c.isTraceEnabled = true | ||
301 | } | ||
302 | |||
303 | // TraceErrorsOnlyOn - same as TraceOn, but only errors will be traced. | ||
304 | func (c *Client) TraceErrorsOnlyOn(outputStream io.Writer) { | ||
305 | c.TraceOn(outputStream) | ||
306 | c.traceErrorsOnly = true | ||
307 | } | ||
308 | |||
309 | // TraceErrorsOnlyOff - Turns off the errors only tracing and everything will be traced after this call. | ||
310 | // If all tracing needs to be turned off, call TraceOff(). | ||
311 | func (c *Client) TraceErrorsOnlyOff() { | ||
312 | c.traceErrorsOnly = false | ||
313 | } | ||
314 | |||
315 | // TraceOff - disable HTTP tracing. | ||
316 | func (c *Client) TraceOff() { | ||
317 | // Disable tracing. | ||
318 | c.isTraceEnabled = false | ||
319 | c.traceErrorsOnly = false | ||
320 | } | ||
321 | |||
322 | // SetS3TransferAccelerate - turns s3 accelerated endpoint on or off for all your | ||
323 | // requests. This feature is only specific to S3 for all other endpoints this | ||
324 | // function does nothing. To read further details on s3 transfer acceleration | ||
325 | // please vist - | ||
326 | // http://docs.aws.amazon.com/AmazonS3/latest/dev/transfer-acceleration.html | ||
327 | func (c *Client) SetS3TransferAccelerate(accelerateEndpoint string) { | ||
328 | if s3utils.IsAmazonEndpoint(*c.endpointURL) { | ||
329 | c.s3AccelerateEndpoint = accelerateEndpoint | ||
330 | } | ||
331 | } | ||
332 | |||
333 | // Hash materials provides relevant initialized hash algo writers | ||
334 | // based on the expected signature type. | ||
335 | // | ||
336 | // - For signature v4 request if the connection is insecure compute only sha256. | ||
337 | // - For signature v4 request if the connection is secure compute only md5. | ||
338 | // - For anonymous request compute md5. | ||
339 | func (c *Client) hashMaterials(isMd5Requested, isSha256Requested bool) (hashAlgos map[string]md5simd.Hasher, hashSums map[string][]byte) { | ||
340 | hashSums = make(map[string][]byte) | ||
341 | hashAlgos = make(map[string]md5simd.Hasher) | ||
342 | if c.overrideSignerType.IsV4() { | ||
343 | if c.secure { | ||
344 | hashAlgos["md5"] = c.md5Hasher() | ||
345 | } else { | ||
346 | if isSha256Requested { | ||
347 | hashAlgos["sha256"] = c.sha256Hasher() | ||
348 | } | ||
349 | } | ||
350 | } else { | ||
351 | if c.overrideSignerType.IsAnonymous() { | ||
352 | hashAlgos["md5"] = c.md5Hasher() | ||
353 | } | ||
354 | } | ||
355 | if isMd5Requested { | ||
356 | hashAlgos["md5"] = c.md5Hasher() | ||
357 | } | ||
358 | return hashAlgos, hashSums | ||
359 | } | ||
360 | |||
361 | const ( | ||
362 | unknown = -1 | ||
363 | offline = 0 | ||
364 | online = 1 | ||
365 | ) | ||
366 | |||
367 | // IsOnline returns true if healthcheck enabled and client is online. | ||
368 | // If HealthCheck function has not been called this will always return true. | ||
369 | func (c *Client) IsOnline() bool { | ||
370 | return !c.IsOffline() | ||
371 | } | ||
372 | |||
373 | // sets online healthStatus to offline | ||
374 | func (c *Client) markOffline() { | ||
375 | atomic.CompareAndSwapInt32(&c.healthStatus, online, offline) | ||
376 | } | ||
377 | |||
378 | // IsOffline returns true if healthcheck enabled and client is offline | ||
379 | // If HealthCheck function has not been called this will always return false. | ||
380 | func (c *Client) IsOffline() bool { | ||
381 | return atomic.LoadInt32(&c.healthStatus) == offline | ||
382 | } | ||
383 | |||
384 | // HealthCheck starts a healthcheck to see if endpoint is up. | ||
385 | // Returns a context cancellation function, to stop the health check, | ||
386 | // and an error if health check is already started. | ||
387 | func (c *Client) HealthCheck(hcDuration time.Duration) (context.CancelFunc, error) { | ||
388 | if atomic.LoadInt32(&c.healthStatus) != unknown { | ||
389 | return nil, fmt.Errorf("health check is running") | ||
390 | } | ||
391 | if hcDuration < 1*time.Second { | ||
392 | return nil, fmt.Errorf("health check duration should be at least 1 second") | ||
393 | } | ||
394 | probeBucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "probe-health-") | ||
395 | ctx, cancelFn := context.WithCancel(context.Background()) | ||
396 | atomic.StoreInt32(&c.healthStatus, offline) | ||
397 | { | ||
398 | // Change to online, if we can connect. | ||
399 | gctx, gcancel := context.WithTimeout(ctx, 3*time.Second) | ||
400 | _, err := c.getBucketLocation(gctx, probeBucketName) | ||
401 | gcancel() | ||
402 | if !IsNetworkOrHostDown(err, false) { | ||
403 | switch ToErrorResponse(err).Code { | ||
404 | case "NoSuchBucket", "AccessDenied", "": | ||
405 | atomic.CompareAndSwapInt32(&c.healthStatus, offline, online) | ||
406 | } | ||
407 | } | ||
408 | } | ||
409 | |||
410 | go func(duration time.Duration) { | ||
411 | timer := time.NewTimer(duration) | ||
412 | defer timer.Stop() | ||
413 | for { | ||
414 | select { | ||
415 | case <-ctx.Done(): | ||
416 | atomic.StoreInt32(&c.healthStatus, unknown) | ||
417 | return | ||
418 | case <-timer.C: | ||
419 | // Do health check the first time and ONLY if the connection is marked offline | ||
420 | if c.IsOffline() { | ||
421 | gctx, gcancel := context.WithTimeout(context.Background(), 3*time.Second) | ||
422 | _, err := c.getBucketLocation(gctx, probeBucketName) | ||
423 | gcancel() | ||
424 | if !IsNetworkOrHostDown(err, false) { | ||
425 | switch ToErrorResponse(err).Code { | ||
426 | case "NoSuchBucket", "AccessDenied", "": | ||
427 | atomic.CompareAndSwapInt32(&c.healthStatus, offline, online) | ||
428 | } | ||
429 | } | ||
430 | } | ||
431 | |||
432 | timer.Reset(duration) | ||
433 | } | ||
434 | } | ||
435 | }(hcDuration) | ||
436 | return cancelFn, nil | ||
437 | } | ||
438 | |||
439 | // requestMetadata - is container for all the values to make a request. | ||
440 | type requestMetadata struct { | ||
441 | // If set newRequest presigns the URL. | ||
442 | presignURL bool | ||
443 | |||
444 | // User supplied. | ||
445 | bucketName string | ||
446 | objectName string | ||
447 | queryValues url.Values | ||
448 | customHeader http.Header | ||
449 | extraPresignHeader http.Header | ||
450 | expires int64 | ||
451 | |||
452 | // Generated by our internal code. | ||
453 | bucketLocation string | ||
454 | contentBody io.Reader | ||
455 | contentLength int64 | ||
456 | contentMD5Base64 string // carries base64 encoded md5sum | ||
457 | contentSHA256Hex string // carries hex encoded sha256sum | ||
458 | streamSha256 bool | ||
459 | addCrc bool | ||
460 | trailer http.Header // (http.Request).Trailer. Requires v4 signature. | ||
461 | } | ||
462 | |||
463 | // dumpHTTP - dump HTTP request and response. | ||
464 | func (c *Client) dumpHTTP(req *http.Request, resp *http.Response) error { | ||
465 | // Starts http dump. | ||
466 | _, err := fmt.Fprintln(c.traceOutput, "---------START-HTTP---------") | ||
467 | if err != nil { | ||
468 | return err | ||
469 | } | ||
470 | |||
471 | // Filter out Signature field from Authorization header. | ||
472 | origAuth := req.Header.Get("Authorization") | ||
473 | if origAuth != "" { | ||
474 | req.Header.Set("Authorization", redactSignature(origAuth)) | ||
475 | } | ||
476 | |||
477 | // Only display request header. | ||
478 | reqTrace, err := httputil.DumpRequestOut(req, false) | ||
479 | if err != nil { | ||
480 | return err | ||
481 | } | ||
482 | |||
483 | // Write request to trace output. | ||
484 | _, err = fmt.Fprint(c.traceOutput, string(reqTrace)) | ||
485 | if err != nil { | ||
486 | return err | ||
487 | } | ||
488 | |||
489 | // Only display response header. | ||
490 | var respTrace []byte | ||
491 | |||
492 | // For errors we make sure to dump response body as well. | ||
493 | if resp.StatusCode != http.StatusOK && | ||
494 | resp.StatusCode != http.StatusPartialContent && | ||
495 | resp.StatusCode != http.StatusNoContent { | ||
496 | respTrace, err = httputil.DumpResponse(resp, true) | ||
497 | if err != nil { | ||
498 | return err | ||
499 | } | ||
500 | } else { | ||
501 | respTrace, err = httputil.DumpResponse(resp, false) | ||
502 | if err != nil { | ||
503 | return err | ||
504 | } | ||
505 | } | ||
506 | |||
507 | // Write response to trace output. | ||
508 | _, err = fmt.Fprint(c.traceOutput, strings.TrimSuffix(string(respTrace), "\r\n")) | ||
509 | if err != nil { | ||
510 | return err | ||
511 | } | ||
512 | |||
513 | // Ends the http dump. | ||
514 | _, err = fmt.Fprintln(c.traceOutput, "---------END-HTTP---------") | ||
515 | if err != nil { | ||
516 | return err | ||
517 | } | ||
518 | |||
519 | // Returns success. | ||
520 | return nil | ||
521 | } | ||
522 | |||
523 | // do - execute http request. | ||
524 | func (c *Client) do(req *http.Request) (resp *http.Response, err error) { | ||
525 | defer func() { | ||
526 | if IsNetworkOrHostDown(err, false) { | ||
527 | c.markOffline() | ||
528 | } | ||
529 | }() | ||
530 | |||
531 | resp, err = c.httpClient.Do(req) | ||
532 | if err != nil { | ||
533 | // Handle this specifically for now until future Golang versions fix this issue properly. | ||
534 | if urlErr, ok := err.(*url.Error); ok { | ||
535 | if strings.Contains(urlErr.Err.Error(), "EOF") { | ||
536 | return nil, &url.Error{ | ||
537 | Op: urlErr.Op, | ||
538 | URL: urlErr.URL, | ||
539 | Err: errors.New("Connection closed by foreign host " + urlErr.URL + ". Retry again."), | ||
540 | } | ||
541 | } | ||
542 | } | ||
543 | return nil, err | ||
544 | } | ||
545 | |||
546 | // Response cannot be non-nil, report error if thats the case. | ||
547 | if resp == nil { | ||
548 | msg := "Response is empty. " + reportIssue | ||
549 | return nil, errInvalidArgument(msg) | ||
550 | } | ||
551 | |||
552 | // If trace is enabled, dump http request and response, | ||
553 | // except when the traceErrorsOnly enabled and the response's status code is ok | ||
554 | if c.isTraceEnabled && !(c.traceErrorsOnly && resp.StatusCode == http.StatusOK) { | ||
555 | err = c.dumpHTTP(req, resp) | ||
556 | if err != nil { | ||
557 | return nil, err | ||
558 | } | ||
559 | } | ||
560 | |||
561 | return resp, nil | ||
562 | } | ||
563 | |||
564 | // List of success status. | ||
565 | var successStatus = []int{ | ||
566 | http.StatusOK, | ||
567 | http.StatusNoContent, | ||
568 | http.StatusPartialContent, | ||
569 | } | ||
570 | |||
571 | // executeMethod - instantiates a given method, and retries the | ||
572 | // request upon any error up to maxRetries attempts in a binomially | ||
573 | // delayed manner using a standard back off algorithm. | ||
574 | func (c *Client) executeMethod(ctx context.Context, method string, metadata requestMetadata) (res *http.Response, err error) { | ||
575 | if c.IsOffline() { | ||
576 | return nil, errors.New(c.endpointURL.String() + " is offline.") | ||
577 | } | ||
578 | |||
579 | var retryable bool // Indicates if request can be retried. | ||
580 | var bodySeeker io.Seeker // Extracted seeker from io.Reader. | ||
581 | reqRetry := MaxRetry // Indicates how many times we can retry the request | ||
582 | |||
583 | if metadata.contentBody != nil { | ||
584 | // Check if body is seekable then it is retryable. | ||
585 | bodySeeker, retryable = metadata.contentBody.(io.Seeker) | ||
586 | switch bodySeeker { | ||
587 | case os.Stdin, os.Stdout, os.Stderr: | ||
588 | retryable = false | ||
589 | } | ||
590 | // Retry only when reader is seekable | ||
591 | if !retryable { | ||
592 | reqRetry = 1 | ||
593 | } | ||
594 | |||
595 | // Figure out if the body can be closed - if yes | ||
596 | // we will definitely close it upon the function | ||
597 | // return. | ||
598 | bodyCloser, ok := metadata.contentBody.(io.Closer) | ||
599 | if ok { | ||
600 | defer bodyCloser.Close() | ||
601 | } | ||
602 | } | ||
603 | |||
604 | // Create cancel context to control 'newRetryTimer' go routine. | ||
605 | retryCtx, cancel := context.WithCancel(ctx) | ||
606 | |||
607 | // Indicate to our routine to exit cleanly upon return. | ||
608 | defer cancel() | ||
609 | |||
610 | for range c.newRetryTimer(retryCtx, reqRetry, DefaultRetryUnit, DefaultRetryCap, MaxJitter) { | ||
611 | // Retry executes the following function body if request has an | ||
612 | // error until maxRetries have been exhausted, retry attempts are | ||
613 | // performed after waiting for a given period of time in a | ||
614 | // binomial fashion. | ||
615 | if retryable { | ||
616 | // Seek back to beginning for each attempt. | ||
617 | if _, err = bodySeeker.Seek(0, 0); err != nil { | ||
618 | // If seek failed, no need to retry. | ||
619 | return nil, err | ||
620 | } | ||
621 | } | ||
622 | |||
623 | if metadata.addCrc { | ||
624 | if metadata.trailer == nil { | ||
625 | metadata.trailer = make(http.Header, 1) | ||
626 | } | ||
627 | crc := crc32.New(crc32.MakeTable(crc32.Castagnoli)) | ||
628 | metadata.contentBody = newHashReaderWrapper(metadata.contentBody, crc, func(hash []byte) { | ||
629 | // Update trailer when done. | ||
630 | metadata.trailer.Set("x-amz-checksum-crc32c", base64.StdEncoding.EncodeToString(hash)) | ||
631 | }) | ||
632 | metadata.trailer.Set("x-amz-checksum-crc32c", base64.StdEncoding.EncodeToString(crc.Sum(nil))) | ||
633 | } | ||
634 | // Instantiate a new request. | ||
635 | var req *http.Request | ||
636 | req, err = c.newRequest(ctx, method, metadata) | ||
637 | if err != nil { | ||
638 | errResponse := ToErrorResponse(err) | ||
639 | if isS3CodeRetryable(errResponse.Code) { | ||
640 | continue // Retry. | ||
641 | } | ||
642 | |||
643 | return nil, err | ||
644 | } | ||
645 | |||
646 | // Initiate the request. | ||
647 | res, err = c.do(req) | ||
648 | if err != nil { | ||
649 | if isRequestErrorRetryable(err) { | ||
650 | // Retry the request | ||
651 | continue | ||
652 | } | ||
653 | return nil, err | ||
654 | } | ||
655 | |||
656 | // For any known successful http status, return quickly. | ||
657 | for _, httpStatus := range successStatus { | ||
658 | if httpStatus == res.StatusCode { | ||
659 | return res, nil | ||
660 | } | ||
661 | } | ||
662 | |||
663 | // Read the body to be saved later. | ||
664 | errBodyBytes, err := io.ReadAll(res.Body) | ||
665 | // res.Body should be closed | ||
666 | closeResponse(res) | ||
667 | if err != nil { | ||
668 | return nil, err | ||
669 | } | ||
670 | |||
671 | // Save the body. | ||
672 | errBodySeeker := bytes.NewReader(errBodyBytes) | ||
673 | res.Body = io.NopCloser(errBodySeeker) | ||
674 | |||
675 | // For errors verify if its retryable otherwise fail quickly. | ||
676 | errResponse := ToErrorResponse(httpRespToErrorResponse(res, metadata.bucketName, metadata.objectName)) | ||
677 | |||
678 | // Save the body back again. | ||
679 | errBodySeeker.Seek(0, 0) // Seek back to starting point. | ||
680 | res.Body = io.NopCloser(errBodySeeker) | ||
681 | |||
682 | // Bucket region if set in error response and the error | ||
683 | // code dictates invalid region, we can retry the request | ||
684 | // with the new region. | ||
685 | // | ||
686 | // Additionally, we should only retry if bucketLocation and custom | ||
687 | // region is empty. | ||
688 | if c.region == "" { | ||
689 | switch errResponse.Code { | ||
690 | case "AuthorizationHeaderMalformed": | ||
691 | fallthrough | ||
692 | case "InvalidRegion": | ||
693 | fallthrough | ||
694 | case "AccessDenied": | ||
695 | if errResponse.Region == "" { | ||
696 | // Region is empty we simply return the error. | ||
697 | return res, err | ||
698 | } | ||
699 | // Region is not empty figure out a way to | ||
700 | // handle this appropriately. | ||
701 | if metadata.bucketName != "" { | ||
702 | // Gather Cached location only if bucketName is present. | ||
703 | if location, cachedOk := c.bucketLocCache.Get(metadata.bucketName); cachedOk && location != errResponse.Region { | ||
704 | c.bucketLocCache.Set(metadata.bucketName, errResponse.Region) | ||
705 | continue // Retry. | ||
706 | } | ||
707 | } else { | ||
708 | // This is for ListBuckets() fallback. | ||
709 | if errResponse.Region != metadata.bucketLocation { | ||
710 | // Retry if the error response has a different region | ||
711 | // than the request we just made. | ||
712 | metadata.bucketLocation = errResponse.Region | ||
713 | continue // Retry | ||
714 | } | ||
715 | } | ||
716 | } | ||
717 | } | ||
718 | |||
719 | // Verify if error response code is retryable. | ||
720 | if isS3CodeRetryable(errResponse.Code) { | ||
721 | continue // Retry. | ||
722 | } | ||
723 | |||
724 | // Verify if http status code is retryable. | ||
725 | if isHTTPStatusRetryable(res.StatusCode) { | ||
726 | continue // Retry. | ||
727 | } | ||
728 | |||
729 | // For all other cases break out of the retry loop. | ||
730 | break | ||
731 | } | ||
732 | |||
733 | // Return an error when retry is canceled or deadlined | ||
734 | if e := retryCtx.Err(); e != nil { | ||
735 | return nil, e | ||
736 | } | ||
737 | |||
738 | return res, err | ||
739 | } | ||
740 | |||
741 | // newRequest - instantiate a new HTTP request for a given method. | ||
742 | func (c *Client) newRequest(ctx context.Context, method string, metadata requestMetadata) (req *http.Request, err error) { | ||
743 | // If no method is supplied default to 'POST'. | ||
744 | if method == "" { | ||
745 | method = http.MethodPost | ||
746 | } | ||
747 | |||
748 | location := metadata.bucketLocation | ||
749 | if location == "" { | ||
750 | if metadata.bucketName != "" { | ||
751 | // Gather location only if bucketName is present. | ||
752 | location, err = c.getBucketLocation(ctx, metadata.bucketName) | ||
753 | if err != nil { | ||
754 | return nil, err | ||
755 | } | ||
756 | } | ||
757 | if location == "" { | ||
758 | location = getDefaultLocation(*c.endpointURL, c.region) | ||
759 | } | ||
760 | } | ||
761 | |||
762 | // Look if target url supports virtual host. | ||
763 | // We explicitly disallow MakeBucket calls to not use virtual DNS style, | ||
764 | // since the resolution may fail. | ||
765 | isMakeBucket := (metadata.objectName == "" && method == http.MethodPut && len(metadata.queryValues) == 0) | ||
766 | isVirtualHost := c.isVirtualHostStyleRequest(*c.endpointURL, metadata.bucketName) && !isMakeBucket | ||
767 | |||
768 | // Construct a new target URL. | ||
769 | targetURL, err := c.makeTargetURL(metadata.bucketName, metadata.objectName, location, | ||
770 | isVirtualHost, metadata.queryValues) | ||
771 | if err != nil { | ||
772 | return nil, err | ||
773 | } | ||
774 | |||
775 | if c.httpTrace != nil { | ||
776 | ctx = httptrace.WithClientTrace(ctx, c.httpTrace) | ||
777 | } | ||
778 | |||
779 | // Initialize a new HTTP request for the method. | ||
780 | req, err = http.NewRequestWithContext(ctx, method, targetURL.String(), nil) | ||
781 | if err != nil { | ||
782 | return nil, err | ||
783 | } | ||
784 | |||
785 | // Get credentials from the configured credentials provider. | ||
786 | value, err := c.credsProvider.Get() | ||
787 | if err != nil { | ||
788 | return nil, err | ||
789 | } | ||
790 | |||
791 | var ( | ||
792 | signerType = value.SignerType | ||
793 | accessKeyID = value.AccessKeyID | ||
794 | secretAccessKey = value.SecretAccessKey | ||
795 | sessionToken = value.SessionToken | ||
796 | ) | ||
797 | |||
798 | // Custom signer set then override the behavior. | ||
799 | if c.overrideSignerType != credentials.SignatureDefault { | ||
800 | signerType = c.overrideSignerType | ||
801 | } | ||
802 | |||
803 | // If signerType returned by credentials helper is anonymous, | ||
804 | // then do not sign regardless of signerType override. | ||
805 | if value.SignerType == credentials.SignatureAnonymous { | ||
806 | signerType = credentials.SignatureAnonymous | ||
807 | } | ||
808 | |||
809 | // Generate presign url if needed, return right here. | ||
810 | if metadata.expires != 0 && metadata.presignURL { | ||
811 | if signerType.IsAnonymous() { | ||
812 | return nil, errInvalidArgument("Presigned URLs cannot be generated with anonymous credentials.") | ||
813 | } | ||
814 | if metadata.extraPresignHeader != nil { | ||
815 | if signerType.IsV2() { | ||
816 | return nil, errInvalidArgument("Extra signed headers for Presign with Signature V2 is not supported.") | ||
817 | } | ||
818 | for k, v := range metadata.extraPresignHeader { | ||
819 | req.Header.Set(k, v[0]) | ||
820 | } | ||
821 | } | ||
822 | if signerType.IsV2() { | ||
823 | // Presign URL with signature v2. | ||
824 | req = signer.PreSignV2(*req, accessKeyID, secretAccessKey, metadata.expires, isVirtualHost) | ||
825 | } else if signerType.IsV4() { | ||
826 | // Presign URL with signature v4. | ||
827 | req = signer.PreSignV4(*req, accessKeyID, secretAccessKey, sessionToken, location, metadata.expires) | ||
828 | } | ||
829 | return req, nil | ||
830 | } | ||
831 | |||
832 | // Set 'User-Agent' header for the request. | ||
833 | c.setUserAgent(req) | ||
834 | |||
835 | // Set all headers. | ||
836 | for k, v := range metadata.customHeader { | ||
837 | req.Header.Set(k, v[0]) | ||
838 | } | ||
839 | |||
840 | // Go net/http notoriously closes the request body. | ||
841 | // - The request Body, if non-nil, will be closed by the underlying Transport, even on errors. | ||
842 | // This can cause underlying *os.File seekers to fail, avoid that | ||
843 | // by making sure to wrap the closer as a nop. | ||
844 | if metadata.contentLength == 0 { | ||
845 | req.Body = nil | ||
846 | } else { | ||
847 | req.Body = io.NopCloser(metadata.contentBody) | ||
848 | } | ||
849 | |||
850 | // Set incoming content-length. | ||
851 | req.ContentLength = metadata.contentLength | ||
852 | if req.ContentLength <= -1 { | ||
853 | // For unknown content length, we upload using transfer-encoding: chunked. | ||
854 | req.TransferEncoding = []string{"chunked"} | ||
855 | } | ||
856 | |||
857 | // set md5Sum for content protection. | ||
858 | if len(metadata.contentMD5Base64) > 0 { | ||
859 | req.Header.Set("Content-Md5", metadata.contentMD5Base64) | ||
860 | } | ||
861 | |||
862 | // For anonymous requests just return. | ||
863 | if signerType.IsAnonymous() { | ||
864 | return req, nil | ||
865 | } | ||
866 | |||
867 | switch { | ||
868 | case signerType.IsV2(): | ||
869 | // Add signature version '2' authorization header. | ||
870 | req = signer.SignV2(*req, accessKeyID, secretAccessKey, isVirtualHost) | ||
871 | case metadata.streamSha256 && !c.secure: | ||
872 | if len(metadata.trailer) > 0 { | ||
873 | req.Trailer = metadata.trailer | ||
874 | } | ||
875 | // Streaming signature is used by default for a PUT object request. | ||
876 | // Additionally, we also look if the initialized client is secure, | ||
877 | // if yes then we don't need to perform streaming signature. | ||
878 | req = signer.StreamingSignV4(req, accessKeyID, | ||
879 | secretAccessKey, sessionToken, location, metadata.contentLength, time.Now().UTC(), c.sha256Hasher()) | ||
880 | default: | ||
881 | // Set sha256 sum for signature calculation only with signature version '4'. | ||
882 | shaHeader := unsignedPayload | ||
883 | if metadata.contentSHA256Hex != "" { | ||
884 | shaHeader = metadata.contentSHA256Hex | ||
885 | if len(metadata.trailer) > 0 { | ||
886 | // Sanity check, we should not end up here if upstream is sane. | ||
887 | return nil, errors.New("internal error: contentSHA256Hex with trailer not supported") | ||
888 | } | ||
889 | } else if len(metadata.trailer) > 0 { | ||
890 | shaHeader = unsignedPayloadTrailer | ||
891 | } | ||
892 | req.Header.Set("X-Amz-Content-Sha256", shaHeader) | ||
893 | |||
894 | // Add signature version '4' authorization header. | ||
895 | req = signer.SignV4Trailer(*req, accessKeyID, secretAccessKey, sessionToken, location, metadata.trailer) | ||
896 | } | ||
897 | |||
898 | // Return request. | ||
899 | return req, nil | ||
900 | } | ||
901 | |||
902 | // set User agent. | ||
903 | func (c *Client) setUserAgent(req *http.Request) { | ||
904 | req.Header.Set("User-Agent", libraryUserAgent) | ||
905 | if c.appInfo.appName != "" && c.appInfo.appVersion != "" { | ||
906 | req.Header.Set("User-Agent", libraryUserAgent+" "+c.appInfo.appName+"/"+c.appInfo.appVersion) | ||
907 | } | ||
908 | } | ||
909 | |||
910 | // makeTargetURL make a new target url. | ||
911 | func (c *Client) makeTargetURL(bucketName, objectName, bucketLocation string, isVirtualHostStyle bool, queryValues url.Values) (*url.URL, error) { | ||
912 | host := c.endpointURL.Host | ||
913 | // For Amazon S3 endpoint, try to fetch location based endpoint. | ||
914 | if s3utils.IsAmazonEndpoint(*c.endpointURL) { | ||
915 | if c.s3AccelerateEndpoint != "" && bucketName != "" { | ||
916 | // http://docs.aws.amazon.com/AmazonS3/latest/dev/transfer-acceleration.html | ||
917 | // Disable transfer acceleration for non-compliant bucket names. | ||
918 | if strings.Contains(bucketName, ".") { | ||
919 | return nil, errTransferAccelerationBucket(bucketName) | ||
920 | } | ||
921 | // If transfer acceleration is requested set new host. | ||
922 | // For more details about enabling transfer acceleration read here. | ||
923 | // http://docs.aws.amazon.com/AmazonS3/latest/dev/transfer-acceleration.html | ||
924 | host = c.s3AccelerateEndpoint | ||
925 | } else { | ||
926 | // Do not change the host if the endpoint URL is a FIPS S3 endpoint or a S3 PrivateLink interface endpoint | ||
927 | if !s3utils.IsAmazonFIPSEndpoint(*c.endpointURL) && !s3utils.IsAmazonPrivateLinkEndpoint(*c.endpointURL) { | ||
928 | // Fetch new host based on the bucket location. | ||
929 | host = getS3Endpoint(bucketLocation) | ||
930 | } | ||
931 | } | ||
932 | } | ||
933 | |||
934 | // Save scheme. | ||
935 | scheme := c.endpointURL.Scheme | ||
936 | |||
937 | // Strip port 80 and 443 so we won't send these ports in Host header. | ||
938 | // The reason is that browsers and curl automatically remove :80 and :443 | ||
939 | // with the generated presigned urls, then a signature mismatch error. | ||
940 | if h, p, err := net.SplitHostPort(host); err == nil { | ||
941 | if scheme == "http" && p == "80" || scheme == "https" && p == "443" { | ||
942 | host = h | ||
943 | if ip := net.ParseIP(h); ip != nil && ip.To4() == nil { | ||
944 | host = "[" + h + "]" | ||
945 | } | ||
946 | } | ||
947 | } | ||
948 | |||
949 | urlStr := scheme + "://" + host + "/" | ||
950 | |||
951 | // Make URL only if bucketName is available, otherwise use the | ||
952 | // endpoint URL. | ||
953 | if bucketName != "" { | ||
954 | // If endpoint supports virtual host style use that always. | ||
955 | // Currently only S3 and Google Cloud Storage would support | ||
956 | // virtual host style. | ||
957 | if isVirtualHostStyle { | ||
958 | urlStr = scheme + "://" + bucketName + "." + host + "/" | ||
959 | if objectName != "" { | ||
960 | urlStr += s3utils.EncodePath(objectName) | ||
961 | } | ||
962 | } else { | ||
963 | // If not fall back to using path style. | ||
964 | urlStr = urlStr + bucketName + "/" | ||
965 | if objectName != "" { | ||
966 | urlStr += s3utils.EncodePath(objectName) | ||
967 | } | ||
968 | } | ||
969 | } | ||
970 | |||
971 | // If there are any query values, add them to the end. | ||
972 | if len(queryValues) > 0 { | ||
973 | urlStr = urlStr + "?" + s3utils.QueryEncode(queryValues) | ||
974 | } | ||
975 | |||
976 | return url.Parse(urlStr) | ||
977 | } | ||
978 | |||
979 | // returns true if virtual hosted style requests are to be used. | ||
980 | func (c *Client) isVirtualHostStyleRequest(url url.URL, bucketName string) bool { | ||
981 | if bucketName == "" { | ||
982 | return false | ||
983 | } | ||
984 | |||
985 | if c.lookup == BucketLookupDNS { | ||
986 | return true | ||
987 | } | ||
988 | if c.lookup == BucketLookupPath { | ||
989 | return false | ||
990 | } | ||
991 | |||
992 | // default to virtual only for Amazon/Google storage. In all other cases use | ||
993 | // path style requests | ||
994 | return s3utils.IsVirtualHostSupported(url, bucketName) | ||
995 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/bucket-cache.go b/vendor/github.com/minio/minio-go/v7/bucket-cache.go deleted file mode 100644 index b1d3b38..0000000 --- a/vendor/github.com/minio/minio-go/v7/bucket-cache.go +++ /dev/null | |||
@@ -1,256 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2015-2017 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package minio | ||
19 | |||
20 | import ( | ||
21 | "context" | ||
22 | "net" | ||
23 | "net/http" | ||
24 | "net/url" | ||
25 | "path" | ||
26 | "sync" | ||
27 | |||
28 | "github.com/minio/minio-go/v7/pkg/credentials" | ||
29 | "github.com/minio/minio-go/v7/pkg/s3utils" | ||
30 | "github.com/minio/minio-go/v7/pkg/signer" | ||
31 | ) | ||
32 | |||
33 | // bucketLocationCache - Provides simple mechanism to hold bucket | ||
34 | // locations in memory. | ||
35 | type bucketLocationCache struct { | ||
36 | // mutex is used for handling the concurrent | ||
37 | // read/write requests for cache. | ||
38 | sync.RWMutex | ||
39 | |||
40 | // items holds the cached bucket locations. | ||
41 | items map[string]string | ||
42 | } | ||
43 | |||
44 | // newBucketLocationCache - Provides a new bucket location cache to be | ||
45 | // used internally with the client object. | ||
46 | func newBucketLocationCache() *bucketLocationCache { | ||
47 | return &bucketLocationCache{ | ||
48 | items: make(map[string]string), | ||
49 | } | ||
50 | } | ||
51 | |||
52 | // Get - Returns a value of a given key if it exists. | ||
53 | func (r *bucketLocationCache) Get(bucketName string) (location string, ok bool) { | ||
54 | r.RLock() | ||
55 | defer r.RUnlock() | ||
56 | location, ok = r.items[bucketName] | ||
57 | return | ||
58 | } | ||
59 | |||
60 | // Set - Will persist a value into cache. | ||
61 | func (r *bucketLocationCache) Set(bucketName, location string) { | ||
62 | r.Lock() | ||
63 | defer r.Unlock() | ||
64 | r.items[bucketName] = location | ||
65 | } | ||
66 | |||
67 | // Delete - Deletes a bucket name from cache. | ||
68 | func (r *bucketLocationCache) Delete(bucketName string) { | ||
69 | r.Lock() | ||
70 | defer r.Unlock() | ||
71 | delete(r.items, bucketName) | ||
72 | } | ||
73 | |||
74 | // GetBucketLocation - get location for the bucket name from location cache, if not | ||
75 | // fetch freshly by making a new request. | ||
76 | func (c *Client) GetBucketLocation(ctx context.Context, bucketName string) (string, error) { | ||
77 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
78 | return "", err | ||
79 | } | ||
80 | return c.getBucketLocation(ctx, bucketName) | ||
81 | } | ||
82 | |||
83 | // getBucketLocation - Get location for the bucketName from location map cache, if not | ||
84 | // fetch freshly by making a new request. | ||
85 | func (c *Client) getBucketLocation(ctx context.Context, bucketName string) (string, error) { | ||
86 | if err := s3utils.CheckValidBucketName(bucketName); err != nil { | ||
87 | return "", err | ||
88 | } | ||
89 | |||
90 | // Region set then no need to fetch bucket location. | ||
91 | if c.region != "" { | ||
92 | return c.region, nil | ||
93 | } | ||
94 | |||
95 | if location, ok := c.bucketLocCache.Get(bucketName); ok { | ||
96 | return location, nil | ||
97 | } | ||
98 | |||
99 | // Initialize a new request. | ||
100 | req, err := c.getBucketLocationRequest(ctx, bucketName) | ||
101 | if err != nil { | ||
102 | return "", err | ||
103 | } | ||
104 | |||
105 | // Initiate the request. | ||
106 | resp, err := c.do(req) | ||
107 | defer closeResponse(resp) | ||
108 | if err != nil { | ||
109 | return "", err | ||
110 | } | ||
111 | location, err := processBucketLocationResponse(resp, bucketName) | ||
112 | if err != nil { | ||
113 | return "", err | ||
114 | } | ||
115 | c.bucketLocCache.Set(bucketName, location) | ||
116 | return location, nil | ||
117 | } | ||
118 | |||
119 | // processes the getBucketLocation http response from the server. | ||
120 | func processBucketLocationResponse(resp *http.Response, bucketName string) (bucketLocation string, err error) { | ||
121 | if resp != nil { | ||
122 | if resp.StatusCode != http.StatusOK { | ||
123 | err = httpRespToErrorResponse(resp, bucketName, "") | ||
124 | errResp := ToErrorResponse(err) | ||
125 | // For access denied error, it could be an anonymous | ||
126 | // request. Move forward and let the top level callers | ||
127 | // succeed if possible based on their policy. | ||
128 | switch errResp.Code { | ||
129 | case "NotImplemented": | ||
130 | switch errResp.Server { | ||
131 | case "AmazonSnowball": | ||
132 | return "snowball", nil | ||
133 | case "cloudflare": | ||
134 | return "us-east-1", nil | ||
135 | } | ||
136 | case "AuthorizationHeaderMalformed": | ||
137 | fallthrough | ||
138 | case "InvalidRegion": | ||
139 | fallthrough | ||
140 | case "AccessDenied": | ||
141 | if errResp.Region == "" { | ||
142 | return "us-east-1", nil | ||
143 | } | ||
144 | return errResp.Region, nil | ||
145 | } | ||
146 | return "", err | ||
147 | } | ||
148 | } | ||
149 | |||
150 | // Extract location. | ||
151 | var locationConstraint string | ||
152 | err = xmlDecoder(resp.Body, &locationConstraint) | ||
153 | if err != nil { | ||
154 | return "", err | ||
155 | } | ||
156 | |||
157 | location := locationConstraint | ||
158 | // Location is empty will be 'us-east-1'. | ||
159 | if location == "" { | ||
160 | location = "us-east-1" | ||
161 | } | ||
162 | |||
163 | // Location can be 'EU' convert it to meaningful 'eu-west-1'. | ||
164 | if location == "EU" { | ||
165 | location = "eu-west-1" | ||
166 | } | ||
167 | |||
168 | // Save the location into cache. | ||
169 | |||
170 | // Return. | ||
171 | return location, nil | ||
172 | } | ||
173 | |||
174 | // getBucketLocationRequest - Wrapper creates a new getBucketLocation request. | ||
175 | func (c *Client) getBucketLocationRequest(ctx context.Context, bucketName string) (*http.Request, error) { | ||
176 | // Set location query. | ||
177 | urlValues := make(url.Values) | ||
178 | urlValues.Set("location", "") | ||
179 | |||
180 | // Set get bucket location always as path style. | ||
181 | targetURL := *c.endpointURL | ||
182 | |||
183 | // as it works in makeTargetURL method from api.go file | ||
184 | if h, p, err := net.SplitHostPort(targetURL.Host); err == nil { | ||
185 | if targetURL.Scheme == "http" && p == "80" || targetURL.Scheme == "https" && p == "443" { | ||
186 | targetURL.Host = h | ||
187 | if ip := net.ParseIP(h); ip != nil && ip.To16() != nil { | ||
188 | targetURL.Host = "[" + h + "]" | ||
189 | } | ||
190 | } | ||
191 | } | ||
192 | |||
193 | isVirtualStyle := c.isVirtualHostStyleRequest(targetURL, bucketName) | ||
194 | |||
195 | var urlStr string | ||
196 | |||
197 | if isVirtualStyle { | ||
198 | urlStr = c.endpointURL.Scheme + "://" + bucketName + "." + targetURL.Host + "/?location" | ||
199 | } else { | ||
200 | targetURL.Path = path.Join(bucketName, "") + "/" | ||
201 | targetURL.RawQuery = urlValues.Encode() | ||
202 | urlStr = targetURL.String() | ||
203 | } | ||
204 | |||
205 | // Get a new HTTP request for the method. | ||
206 | req, err := http.NewRequestWithContext(ctx, http.MethodGet, urlStr, nil) | ||
207 | if err != nil { | ||
208 | return nil, err | ||
209 | } | ||
210 | |||
211 | // Set UserAgent for the request. | ||
212 | c.setUserAgent(req) | ||
213 | |||
214 | // Get credentials from the configured credentials provider. | ||
215 | value, err := c.credsProvider.Get() | ||
216 | if err != nil { | ||
217 | return nil, err | ||
218 | } | ||
219 | |||
220 | var ( | ||
221 | signerType = value.SignerType | ||
222 | accessKeyID = value.AccessKeyID | ||
223 | secretAccessKey = value.SecretAccessKey | ||
224 | sessionToken = value.SessionToken | ||
225 | ) | ||
226 | |||
227 | // Custom signer set then override the behavior. | ||
228 | if c.overrideSignerType != credentials.SignatureDefault { | ||
229 | signerType = c.overrideSignerType | ||
230 | } | ||
231 | |||
232 | // If signerType returned by credentials helper is anonymous, | ||
233 | // then do not sign regardless of signerType override. | ||
234 | if value.SignerType == credentials.SignatureAnonymous { | ||
235 | signerType = credentials.SignatureAnonymous | ||
236 | } | ||
237 | |||
238 | if signerType.IsAnonymous() { | ||
239 | return req, nil | ||
240 | } | ||
241 | |||
242 | if signerType.IsV2() { | ||
243 | req = signer.SignV2(*req, accessKeyID, secretAccessKey, isVirtualStyle) | ||
244 | return req, nil | ||
245 | } | ||
246 | |||
247 | // Set sha256 sum for signature calculation only with signature version '4'. | ||
248 | contentSha256 := emptySHA256Hex | ||
249 | if c.secure { | ||
250 | contentSha256 = unsignedPayload | ||
251 | } | ||
252 | |||
253 | req.Header.Set("X-Amz-Content-Sha256", contentSha256) | ||
254 | req = signer.SignV4(*req, accessKeyID, secretAccessKey, sessionToken, "us-east-1") | ||
255 | return req, nil | ||
256 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/checksum.go b/vendor/github.com/minio/minio-go/v7/checksum.go deleted file mode 100644 index a1f6f43..0000000 --- a/vendor/github.com/minio/minio-go/v7/checksum.go +++ /dev/null | |||
@@ -1,210 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2015-2023 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package minio | ||
19 | |||
20 | import ( | ||
21 | "crypto/sha1" | ||
22 | "crypto/sha256" | ||
23 | "encoding/base64" | ||
24 | "hash" | ||
25 | "hash/crc32" | ||
26 | "io" | ||
27 | "math/bits" | ||
28 | ) | ||
29 | |||
30 | // ChecksumType contains information about the checksum type. | ||
31 | type ChecksumType uint32 | ||
32 | |||
33 | const ( | ||
34 | |||
35 | // ChecksumSHA256 indicates a SHA256 checksum. | ||
36 | ChecksumSHA256 ChecksumType = 1 << iota | ||
37 | // ChecksumSHA1 indicates a SHA-1 checksum. | ||
38 | ChecksumSHA1 | ||
39 | // ChecksumCRC32 indicates a CRC32 checksum with IEEE table. | ||
40 | ChecksumCRC32 | ||
41 | // ChecksumCRC32C indicates a CRC32 checksum with Castagnoli table. | ||
42 | ChecksumCRC32C | ||
43 | |||
44 | // Keep after all valid checksums | ||
45 | checksumLast | ||
46 | |||
47 | // checksumMask is a mask for valid checksum types. | ||
48 | checksumMask = checksumLast - 1 | ||
49 | |||
50 | // ChecksumNone indicates no checksum. | ||
51 | ChecksumNone ChecksumType = 0 | ||
52 | |||
53 | amzChecksumAlgo = "x-amz-checksum-algorithm" | ||
54 | amzChecksumCRC32 = "x-amz-checksum-crc32" | ||
55 | amzChecksumCRC32C = "x-amz-checksum-crc32c" | ||
56 | amzChecksumSHA1 = "x-amz-checksum-sha1" | ||
57 | amzChecksumSHA256 = "x-amz-checksum-sha256" | ||
58 | ) | ||
59 | |||
60 | // Is returns if c is all of t. | ||
61 | func (c ChecksumType) Is(t ChecksumType) bool { | ||
62 | return c&t == t | ||
63 | } | ||
64 | |||
65 | // Key returns the header key. | ||
66 | // returns empty string if invalid or none. | ||
67 | func (c ChecksumType) Key() string { | ||
68 | switch c & checksumMask { | ||
69 | case ChecksumCRC32: | ||
70 | return amzChecksumCRC32 | ||
71 | case ChecksumCRC32C: | ||
72 | return amzChecksumCRC32C | ||
73 | case ChecksumSHA1: | ||
74 | return amzChecksumSHA1 | ||
75 | case ChecksumSHA256: | ||
76 | return amzChecksumSHA256 | ||
77 | } | ||
78 | return "" | ||
79 | } | ||
80 | |||
81 | // RawByteLen returns the size of the un-encoded checksum. | ||
82 | func (c ChecksumType) RawByteLen() int { | ||
83 | switch c & checksumMask { | ||
84 | case ChecksumCRC32, ChecksumCRC32C: | ||
85 | return 4 | ||
86 | case ChecksumSHA1: | ||
87 | return sha1.Size | ||
88 | case ChecksumSHA256: | ||
89 | return sha256.Size | ||
90 | } | ||
91 | return 0 | ||
92 | } | ||
93 | |||
94 | // Hasher returns a hasher corresponding to the checksum type. | ||
95 | // Returns nil if no checksum. | ||
96 | func (c ChecksumType) Hasher() hash.Hash { | ||
97 | switch c & checksumMask { | ||
98 | case ChecksumCRC32: | ||
99 | return crc32.NewIEEE() | ||
100 | case ChecksumCRC32C: | ||
101 | return crc32.New(crc32.MakeTable(crc32.Castagnoli)) | ||
102 | case ChecksumSHA1: | ||
103 | return sha1.New() | ||
104 | case ChecksumSHA256: | ||
105 | return sha256.New() | ||
106 | } | ||
107 | return nil | ||
108 | } | ||
109 | |||
110 | // IsSet returns whether the type is valid and known. | ||
111 | func (c ChecksumType) IsSet() bool { | ||
112 | return bits.OnesCount32(uint32(c)) == 1 | ||
113 | } | ||
114 | |||
115 | // String returns the type as a string. | ||
116 | // CRC32, CRC32C, SHA1, and SHA256 for valid values. | ||
117 | // Empty string for unset and "<invalid>" if not valid. | ||
118 | func (c ChecksumType) String() string { | ||
119 | switch c & checksumMask { | ||
120 | case ChecksumCRC32: | ||
121 | return "CRC32" | ||
122 | case ChecksumCRC32C: | ||
123 | return "CRC32C" | ||
124 | case ChecksumSHA1: | ||
125 | return "SHA1" | ||
126 | case ChecksumSHA256: | ||
127 | return "SHA256" | ||
128 | case ChecksumNone: | ||
129 | return "" | ||
130 | } | ||
131 | return "<invalid>" | ||
132 | } | ||
133 | |||
134 | // ChecksumReader reads all of r and returns a checksum of type c. | ||
135 | // Returns any error that may have occurred while reading. | ||
136 | func (c ChecksumType) ChecksumReader(r io.Reader) (Checksum, error) { | ||
137 | h := c.Hasher() | ||
138 | if h == nil { | ||
139 | return Checksum{}, nil | ||
140 | } | ||
141 | _, err := io.Copy(h, r) | ||
142 | if err != nil { | ||
143 | return Checksum{}, err | ||
144 | } | ||
145 | return NewChecksum(c, h.Sum(nil)), nil | ||
146 | } | ||
147 | |||
148 | // ChecksumBytes returns a checksum of the content b with type c. | ||
149 | func (c ChecksumType) ChecksumBytes(b []byte) Checksum { | ||
150 | h := c.Hasher() | ||
151 | if h == nil { | ||
152 | return Checksum{} | ||
153 | } | ||
154 | n, err := h.Write(b) | ||
155 | if err != nil || n != len(b) { | ||
156 | // Shouldn't happen with these checksummers. | ||
157 | return Checksum{} | ||
158 | } | ||
159 | return NewChecksum(c, h.Sum(nil)) | ||
160 | } | ||
161 | |||
162 | // Checksum is a type and encoded value. | ||
163 | type Checksum struct { | ||
164 | Type ChecksumType | ||
165 | r []byte | ||
166 | } | ||
167 | |||
168 | // NewChecksum sets the checksum to the value of b, | ||
169 | // which is the raw hash output. | ||
170 | // If the length of c does not match t.RawByteLen, | ||
171 | // a checksum with ChecksumNone is returned. | ||
172 | func NewChecksum(t ChecksumType, b []byte) Checksum { | ||
173 | if t.IsSet() && len(b) == t.RawByteLen() { | ||
174 | return Checksum{Type: t, r: b} | ||
175 | } | ||
176 | return Checksum{} | ||
177 | } | ||
178 | |||
179 | // NewChecksumString sets the checksum to the value of s, | ||
180 | // which is the base 64 encoded raw hash output. | ||
181 | // If the length of c does not match t.RawByteLen, it is not added. | ||
182 | func NewChecksumString(t ChecksumType, s string) Checksum { | ||
183 | b, _ := base64.StdEncoding.DecodeString(s) | ||
184 | if t.IsSet() && len(b) == t.RawByteLen() { | ||
185 | return Checksum{Type: t, r: b} | ||
186 | } | ||
187 | return Checksum{} | ||
188 | } | ||
189 | |||
190 | // IsSet returns whether the checksum is valid and known. | ||
191 | func (c Checksum) IsSet() bool { | ||
192 | return c.Type.IsSet() && len(c.r) == c.Type.RawByteLen() | ||
193 | } | ||
194 | |||
195 | // Encoded returns the encoded value. | ||
196 | // Returns the empty string if not set or valid. | ||
197 | func (c Checksum) Encoded() string { | ||
198 | if !c.IsSet() { | ||
199 | return "" | ||
200 | } | ||
201 | return base64.StdEncoding.EncodeToString(c.r) | ||
202 | } | ||
203 | |||
204 | // Raw returns the raw checksum value if set. | ||
205 | func (c Checksum) Raw() []byte { | ||
206 | if !c.IsSet() { | ||
207 | return nil | ||
208 | } | ||
209 | return c.r | ||
210 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/code_of_conduct.md b/vendor/github.com/minio/minio-go/v7/code_of_conduct.md deleted file mode 100644 index cb232c3..0000000 --- a/vendor/github.com/minio/minio-go/v7/code_of_conduct.md +++ /dev/null | |||
@@ -1,80 +0,0 @@ | |||
1 | # Contributor Covenant Code of Conduct | ||
2 | |||
3 | ## Our Pledge | ||
4 | |||
5 | In the interest of fostering an open and welcoming environment, we as | ||
6 | contributors and maintainers pledge to making participation in our project and | ||
7 | our community a harassment-free experience for everyone, regardless of age, body | ||
8 | size, disability, ethnicity, gender identity and expression, level of experience, | ||
9 | nationality, personal appearance, race, religion, or sexual identity and | ||
10 | orientation. | ||
11 | |||
12 | ## Our Standards | ||
13 | |||
14 | Examples of behavior that contributes to creating a positive environment | ||
15 | include: | ||
16 | |||
17 | * Using welcoming and inclusive language | ||
18 | * Being respectful of differing viewpoints and experiences | ||
19 | * Gracefully accepting constructive criticism | ||
20 | * Focusing on what is best for the community | ||
21 | * Showing empathy towards other community members | ||
22 | |||
23 | Examples of unacceptable behavior by participants include: | ||
24 | |||
25 | * The use of sexualized language or imagery and unwelcome sexual attention or | ||
26 | advances | ||
27 | * Trolling, insulting/derogatory comments, and personal or political attacks | ||
28 | * Public or private harassment | ||
29 | * Publishing others' private information, such as a physical or electronic | ||
30 | address, without explicit permission | ||
31 | * Other conduct which could reasonably be considered inappropriate in a | ||
32 | professional setting | ||
33 | |||
34 | ## Our Responsibilities | ||
35 | |||
36 | Project maintainers are responsible for clarifying the standards of acceptable | ||
37 | behavior and are expected to take appropriate and fair corrective action in | ||
38 | response to any instances of unacceptable behavior, in compliance with the | ||
39 | licensing terms applying to the Project developments. | ||
40 | |||
41 | Project maintainers have the right and responsibility to remove, edit, or | ||
42 | reject comments, commits, code, wiki edits, issues, and other contributions | ||
43 | that are not aligned to this Code of Conduct, or to ban temporarily or | ||
44 | permanently any contributor for other behaviors that they deem inappropriate, | ||
45 | threatening, offensive, or harmful. However, these actions shall respect the | ||
46 | licensing terms of the Project Developments that will always supersede such | ||
47 | Code of Conduct. | ||
48 | |||
49 | ## Scope | ||
50 | |||
51 | This Code of Conduct applies both within project spaces and in public spaces | ||
52 | when an individual is representing the project or its community. Examples of | ||
53 | representing a project or community include using an official project e-mail | ||
54 | address, posting via an official social media account, or acting as an appointed | ||
55 | representative at an online or offline event. Representation of a project may be | ||
56 | further defined and clarified by project maintainers. | ||
57 | |||
58 | ## Enforcement | ||
59 | |||
60 | Instances of abusive, harassing, or otherwise unacceptable behavior may be | ||
61 | reported by contacting the project team at [email protected]. The project team | ||
62 | will review and investigate all complaints, and will respond in a way that it deems | ||
63 | appropriate to the circumstances. The project team is obligated to maintain | ||
64 | confidentiality with regard to the reporter of an incident. | ||
65 | Further details of specific enforcement policies may be posted separately. | ||
66 | |||
67 | Project maintainers who do not follow or enforce the Code of Conduct in good | ||
68 | faith may face temporary or permanent repercussions as determined by other | ||
69 | members of the project's leadership. | ||
70 | |||
71 | ## Attribution | ||
72 | |||
73 | This Code of Conduct is adapted from the [Contributor Covenant][homepage], version 1.4, | ||
74 | available at [http://contributor-covenant.org/version/1/4][version] | ||
75 | |||
76 | This version includes a clarification to ensure that the code of conduct is in | ||
77 | compliance with the free software licensing terms of the project. | ||
78 | |||
79 | [homepage]: http://contributor-covenant.org | ||
80 | [version]: http://contributor-covenant.org/version/1/4/ | ||
diff --git a/vendor/github.com/minio/minio-go/v7/constants.go b/vendor/github.com/minio/minio-go/v7/constants.go deleted file mode 100644 index 401d2a7..0000000 --- a/vendor/github.com/minio/minio-go/v7/constants.go +++ /dev/null | |||
@@ -1,110 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2015-2017 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package minio | ||
19 | |||
20 | // Multipart upload defaults. | ||
21 | |||
22 | // absMinPartSize - absolute minimum part size (5 MiB) below which | ||
23 | // a part in a multipart upload may not be uploaded. | ||
24 | const absMinPartSize = 1024 * 1024 * 5 | ||
25 | |||
26 | // minPartSize - minimum part size 16MiB per object after which | ||
27 | // putObject behaves internally as multipart. | ||
28 | const minPartSize = 1024 * 1024 * 16 | ||
29 | |||
30 | // maxPartsCount - maximum number of parts for a single multipart session. | ||
31 | const maxPartsCount = 10000 | ||
32 | |||
33 | // maxPartSize - maximum part size 5GiB for a single multipart upload | ||
34 | // operation. | ||
35 | const maxPartSize = 1024 * 1024 * 1024 * 5 | ||
36 | |||
37 | // maxSinglePutObjectSize - maximum size 5GiB of object per PUT | ||
38 | // operation. | ||
39 | const maxSinglePutObjectSize = 1024 * 1024 * 1024 * 5 | ||
40 | |||
41 | // maxMultipartPutObjectSize - maximum size 5TiB of object for | ||
42 | // Multipart operation. | ||
43 | const maxMultipartPutObjectSize = 1024 * 1024 * 1024 * 1024 * 5 | ||
44 | |||
45 | // unsignedPayload - value to be set to X-Amz-Content-Sha256 header when | ||
46 | // we don't want to sign the request payload | ||
47 | const unsignedPayload = "UNSIGNED-PAYLOAD" | ||
48 | |||
49 | // unsignedPayloadTrailer value to be set to X-Amz-Content-Sha256 header when | ||
50 | // we don't want to sign the request payload, but have a trailer. | ||
51 | const unsignedPayloadTrailer = "STREAMING-UNSIGNED-PAYLOAD-TRAILER" | ||
52 | |||
53 | // Total number of parallel workers used for multipart operation. | ||
54 | const totalWorkers = 4 | ||
55 | |||
56 | // Signature related constants. | ||
57 | const ( | ||
58 | signV4Algorithm = "AWS4-HMAC-SHA256" | ||
59 | iso8601DateFormat = "20060102T150405Z" | ||
60 | ) | ||
61 | |||
62 | const ( | ||
63 | // Storage class header. | ||
64 | amzStorageClass = "X-Amz-Storage-Class" | ||
65 | |||
66 | // Website redirect location header | ||
67 | amzWebsiteRedirectLocation = "X-Amz-Website-Redirect-Location" | ||
68 | |||
69 | // Object Tagging headers | ||
70 | amzTaggingHeader = "X-Amz-Tagging" | ||
71 | amzTaggingHeaderDirective = "X-Amz-Tagging-Directive" | ||
72 | |||
73 | amzVersionID = "X-Amz-Version-Id" | ||
74 | amzTaggingCount = "X-Amz-Tagging-Count" | ||
75 | amzExpiration = "X-Amz-Expiration" | ||
76 | amzRestore = "X-Amz-Restore" | ||
77 | amzReplicationStatus = "X-Amz-Replication-Status" | ||
78 | amzDeleteMarker = "X-Amz-Delete-Marker" | ||
79 | |||
80 | // Object legal hold header | ||
81 | amzLegalHoldHeader = "X-Amz-Object-Lock-Legal-Hold" | ||
82 | |||
83 | // Object retention header | ||
84 | amzLockMode = "X-Amz-Object-Lock-Mode" | ||
85 | amzLockRetainUntil = "X-Amz-Object-Lock-Retain-Until-Date" | ||
86 | amzBypassGovernance = "X-Amz-Bypass-Governance-Retention" | ||
87 | |||
88 | // Replication status | ||
89 | amzBucketReplicationStatus = "X-Amz-Replication-Status" | ||
90 | // Minio specific Replication/lifecycle transition extension | ||
91 | minIOBucketSourceMTime = "X-Minio-Source-Mtime" | ||
92 | |||
93 | minIOBucketSourceETag = "X-Minio-Source-Etag" | ||
94 | minIOBucketReplicationDeleteMarker = "X-Minio-Source-DeleteMarker" | ||
95 | minIOBucketReplicationProxyRequest = "X-Minio-Source-Proxy-Request" | ||
96 | minIOBucketReplicationRequest = "X-Minio-Source-Replication-Request" | ||
97 | minIOBucketReplicationCheck = "X-Minio-Source-Replication-Check" | ||
98 | |||
99 | // Header indicates last tag update time on source | ||
100 | minIOBucketReplicationTaggingTimestamp = "X-Minio-Source-Replication-Tagging-Timestamp" | ||
101 | // Header indicates last retention update time on source | ||
102 | minIOBucketReplicationObjectRetentionTimestamp = "X-Minio-Source-Replication-Retention-Timestamp" | ||
103 | // Header indicates last legalhold update time on source | ||
104 | minIOBucketReplicationObjectLegalHoldTimestamp = "X-Minio-Source-Replication-LegalHold-Timestamp" | ||
105 | minIOForceDelete = "x-minio-force-delete" | ||
106 | // Header indicates delete marker replication request can be sent by source now. | ||
107 | minioTgtReplicationReady = "X-Minio-Replication-Ready" | ||
108 | // Header asks if delete marker replication request can be sent by source now. | ||
109 | isMinioTgtReplicationReady = "X-Minio-Check-Replication-Ready" | ||
110 | ) | ||
diff --git a/vendor/github.com/minio/minio-go/v7/core.go b/vendor/github.com/minio/minio-go/v7/core.go deleted file mode 100644 index 132ea70..0000000 --- a/vendor/github.com/minio/minio-go/v7/core.go +++ /dev/null | |||
@@ -1,150 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2015-2017 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package minio | ||
19 | |||
20 | import ( | ||
21 | "context" | ||
22 | "io" | ||
23 | "net/http" | ||
24 | |||
25 | "github.com/minio/minio-go/v7/pkg/encrypt" | ||
26 | ) | ||
27 | |||
28 | // Core - Inherits Client and adds new methods to expose the low level S3 APIs. | ||
29 | type Core struct { | ||
30 | *Client | ||
31 | } | ||
32 | |||
33 | // NewCore - Returns new initialized a Core client, this CoreClient should be | ||
34 | // only used under special conditions such as need to access lower primitives | ||
35 | // and being able to use them to write your own wrappers. | ||
36 | func NewCore(endpoint string, opts *Options) (*Core, error) { | ||
37 | var s3Client Core | ||
38 | client, err := New(endpoint, opts) | ||
39 | if err != nil { | ||
40 | return nil, err | ||
41 | } | ||
42 | s3Client.Client = client | ||
43 | return &s3Client, nil | ||
44 | } | ||
45 | |||
46 | // ListObjects - List all the objects at a prefix, optionally with marker and delimiter | ||
47 | // you can further filter the results. | ||
48 | func (c Core) ListObjects(bucket, prefix, marker, delimiter string, maxKeys int) (result ListBucketResult, err error) { | ||
49 | return c.listObjectsQuery(context.Background(), bucket, prefix, marker, delimiter, maxKeys, nil) | ||
50 | } | ||
51 | |||
52 | // ListObjectsV2 - Lists all the objects at a prefix, similar to ListObjects() but uses | ||
53 | // continuationToken instead of marker to support iteration over the results. | ||
54 | func (c Core) ListObjectsV2(bucketName, objectPrefix, startAfter, continuationToken, delimiter string, maxkeys int) (ListBucketV2Result, error) { | ||
55 | return c.listObjectsV2Query(context.Background(), bucketName, objectPrefix, continuationToken, true, false, delimiter, startAfter, maxkeys, nil) | ||
56 | } | ||
57 | |||
58 | // CopyObject - copies an object from source object to destination object on server side. | ||
59 | func (c Core) CopyObject(ctx context.Context, sourceBucket, sourceObject, destBucket, destObject string, metadata map[string]string, srcOpts CopySrcOptions, dstOpts PutObjectOptions) (ObjectInfo, error) { | ||
60 | return c.copyObjectDo(ctx, sourceBucket, sourceObject, destBucket, destObject, metadata, srcOpts, dstOpts) | ||
61 | } | ||
62 | |||
63 | // CopyObjectPart - creates a part in a multipart upload by copying (a | ||
64 | // part of) an existing object. | ||
65 | func (c Core) CopyObjectPart(ctx context.Context, srcBucket, srcObject, destBucket, destObject, uploadID string, | ||
66 | partID int, startOffset, length int64, metadata map[string]string, | ||
67 | ) (p CompletePart, err error) { | ||
68 | return c.copyObjectPartDo(ctx, srcBucket, srcObject, destBucket, destObject, uploadID, | ||
69 | partID, startOffset, length, metadata) | ||
70 | } | ||
71 | |||
72 | // PutObject - Upload object. Uploads using single PUT call. | ||
73 | func (c Core) PutObject(ctx context.Context, bucket, object string, data io.Reader, size int64, md5Base64, sha256Hex string, opts PutObjectOptions) (UploadInfo, error) { | ||
74 | hookReader := newHook(data, opts.Progress) | ||
75 | return c.putObjectDo(ctx, bucket, object, hookReader, md5Base64, sha256Hex, size, opts) | ||
76 | } | ||
77 | |||
78 | // NewMultipartUpload - Initiates new multipart upload and returns the new uploadID. | ||
79 | func (c Core) NewMultipartUpload(ctx context.Context, bucket, object string, opts PutObjectOptions) (uploadID string, err error) { | ||
80 | result, err := c.initiateMultipartUpload(ctx, bucket, object, opts) | ||
81 | return result.UploadID, err | ||
82 | } | ||
83 | |||
84 | // ListMultipartUploads - List incomplete uploads. | ||
85 | func (c Core) ListMultipartUploads(ctx context.Context, bucket, prefix, keyMarker, uploadIDMarker, delimiter string, maxUploads int) (result ListMultipartUploadsResult, err error) { | ||
86 | return c.listMultipartUploadsQuery(ctx, bucket, keyMarker, uploadIDMarker, prefix, delimiter, maxUploads) | ||
87 | } | ||
88 | |||
89 | // PutObjectPartOptions contains options for PutObjectPart API | ||
90 | type PutObjectPartOptions struct { | ||
91 | Md5Base64, Sha256Hex string | ||
92 | SSE encrypt.ServerSide | ||
93 | CustomHeader, Trailer http.Header | ||
94 | } | ||
95 | |||
96 | // PutObjectPart - Upload an object part. | ||
97 | func (c Core) PutObjectPart(ctx context.Context, bucket, object, uploadID string, partID int, | ||
98 | data io.Reader, size int64, opts PutObjectPartOptions, | ||
99 | ) (ObjectPart, error) { | ||
100 | p := uploadPartParams{ | ||
101 | bucketName: bucket, | ||
102 | objectName: object, | ||
103 | uploadID: uploadID, | ||
104 | reader: data, | ||
105 | partNumber: partID, | ||
106 | md5Base64: opts.Md5Base64, | ||
107 | sha256Hex: opts.Sha256Hex, | ||
108 | size: size, | ||
109 | sse: opts.SSE, | ||
110 | streamSha256: true, | ||
111 | customHeader: opts.CustomHeader, | ||
112 | trailer: opts.Trailer, | ||
113 | } | ||
114 | return c.uploadPart(ctx, p) | ||
115 | } | ||
116 | |||
117 | // ListObjectParts - List uploaded parts of an incomplete upload.x | ||
118 | func (c Core) ListObjectParts(ctx context.Context, bucket, object, uploadID string, partNumberMarker, maxParts int) (result ListObjectPartsResult, err error) { | ||
119 | return c.listObjectPartsQuery(ctx, bucket, object, uploadID, partNumberMarker, maxParts) | ||
120 | } | ||
121 | |||
122 | // CompleteMultipartUpload - Concatenate uploaded parts and commit to an object. | ||
123 | func (c Core) CompleteMultipartUpload(ctx context.Context, bucket, object, uploadID string, parts []CompletePart, opts PutObjectOptions) (UploadInfo, error) { | ||
124 | res, err := c.completeMultipartUpload(ctx, bucket, object, uploadID, completeMultipartUpload{ | ||
125 | Parts: parts, | ||
126 | }, opts) | ||
127 | return res, err | ||
128 | } | ||
129 | |||
130 | // AbortMultipartUpload - Abort an incomplete upload. | ||
131 | func (c Core) AbortMultipartUpload(ctx context.Context, bucket, object, uploadID string) error { | ||
132 | return c.abortMultipartUpload(ctx, bucket, object, uploadID) | ||
133 | } | ||
134 | |||
135 | // GetBucketPolicy - fetches bucket access policy for a given bucket. | ||
136 | func (c Core) GetBucketPolicy(ctx context.Context, bucket string) (string, error) { | ||
137 | return c.getBucketPolicy(ctx, bucket) | ||
138 | } | ||
139 | |||
140 | // PutBucketPolicy - applies a new bucket access policy for a given bucket. | ||
141 | func (c Core) PutBucketPolicy(ctx context.Context, bucket, bucketPolicy string) error { | ||
142 | return c.putBucketPolicy(ctx, bucket, bucketPolicy) | ||
143 | } | ||
144 | |||
145 | // GetObject is a lower level API implemented to support reading | ||
146 | // partial objects and also downloading objects with special conditions | ||
147 | // matching etag, modtime etc. | ||
148 | func (c Core) GetObject(ctx context.Context, bucketName, objectName string, opts GetObjectOptions) (io.ReadCloser, ObjectInfo, http.Header, error) { | ||
149 | return c.getObject(ctx, bucketName, objectName, opts) | ||
150 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/functional_tests.go b/vendor/github.com/minio/minio-go/v7/functional_tests.go deleted file mode 100644 index f951cd0..0000000 --- a/vendor/github.com/minio/minio-go/v7/functional_tests.go +++ /dev/null | |||
@@ -1,13004 +0,0 @@ | |||
1 | //go:build mint | ||
2 | // +build mint | ||
3 | |||
4 | /* | ||
5 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
6 | * Copyright 2015-2020 MinIO, Inc. | ||
7 | * | ||
8 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
9 | * you may not use this file except in compliance with the License. | ||
10 | * You may obtain a copy of the License at | ||
11 | * | ||
12 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
13 | * | ||
14 | * Unless required by applicable law or agreed to in writing, software | ||
15 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
16 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
17 | * See the License for the specific language governing permissions and | ||
18 | * limitations under the License. | ||
19 | */ | ||
20 | |||
21 | package main | ||
22 | |||
23 | import ( | ||
24 | "archive/zip" | ||
25 | "bytes" | ||
26 | "context" | ||
27 | "crypto/sha1" | ||
28 | "encoding/base64" | ||
29 | "errors" | ||
30 | "fmt" | ||
31 | "hash" | ||
32 | "hash/crc32" | ||
33 | "io" | ||
34 | "math/rand" | ||
35 | "mime/multipart" | ||
36 | "net/http" | ||
37 | "net/url" | ||
38 | "os" | ||
39 | "path" | ||
40 | "path/filepath" | ||
41 | "reflect" | ||
42 | "runtime" | ||
43 | "sort" | ||
44 | "strconv" | ||
45 | "strings" | ||
46 | "sync" | ||
47 | "time" | ||
48 | |||
49 | "github.com/dustin/go-humanize" | ||
50 | jsoniter "github.com/json-iterator/go" | ||
51 | "github.com/minio/sha256-simd" | ||
52 | log "github.com/sirupsen/logrus" | ||
53 | |||
54 | "github.com/minio/minio-go/v7" | ||
55 | "github.com/minio/minio-go/v7/pkg/credentials" | ||
56 | "github.com/minio/minio-go/v7/pkg/encrypt" | ||
57 | "github.com/minio/minio-go/v7/pkg/notification" | ||
58 | "github.com/minio/minio-go/v7/pkg/tags" | ||
59 | ) | ||
60 | |||
61 | const letterBytes = "abcdefghijklmnopqrstuvwxyz01234569" | ||
62 | const ( | ||
63 | letterIdxBits = 6 // 6 bits to represent a letter index | ||
64 | letterIdxMask = 1<<letterIdxBits - 1 // All 1-bits, as many as letterIdxBits | ||
65 | letterIdxMax = 63 / letterIdxBits // # of letter indices fitting in 63 bits | ||
66 | ) | ||
67 | |||
68 | const ( | ||
69 | serverEndpoint = "SERVER_ENDPOINT" | ||
70 | accessKey = "ACCESS_KEY" | ||
71 | secretKey = "SECRET_KEY" | ||
72 | enableHTTPS = "ENABLE_HTTPS" | ||
73 | enableKMS = "ENABLE_KMS" | ||
74 | ) | ||
75 | |||
76 | type mintJSONFormatter struct{} | ||
77 | |||
78 | func (f *mintJSONFormatter) Format(entry *log.Entry) ([]byte, error) { | ||
79 | data := make(log.Fields, len(entry.Data)) | ||
80 | for k, v := range entry.Data { | ||
81 | switch v := v.(type) { | ||
82 | case error: | ||
83 | // Otherwise errors are ignored by `encoding/json` | ||
84 | // https://github.com/sirupsen/logrus/issues/137 | ||
85 | data[k] = v.Error() | ||
86 | default: | ||
87 | data[k] = v | ||
88 | } | ||
89 | } | ||
90 | json := jsoniter.ConfigCompatibleWithStandardLibrary | ||
91 | serialized, err := json.Marshal(data) | ||
92 | if err != nil { | ||
93 | return nil, fmt.Errorf("Failed to marshal fields to JSON, %v", err) | ||
94 | } | ||
95 | return append(serialized, '\n'), nil | ||
96 | } | ||
97 | |||
98 | var readFull = func(r io.Reader, buf []byte) (n int, err error) { | ||
99 | // ReadFull reads exactly len(buf) bytes from r into buf. | ||
100 | // It returns the number of bytes copied and an error if | ||
101 | // fewer bytes were read. The error is EOF only if no bytes | ||
102 | // were read. If an EOF happens after reading some but not | ||
103 | // all the bytes, ReadFull returns ErrUnexpectedEOF. | ||
104 | // On return, n == len(buf) if and only if err == nil. | ||
105 | // If r returns an error having read at least len(buf) bytes, | ||
106 | // the error is dropped. | ||
107 | for n < len(buf) && err == nil { | ||
108 | var nn int | ||
109 | nn, err = r.Read(buf[n:]) | ||
110 | // Some spurious io.Reader's return | ||
111 | // io.ErrUnexpectedEOF when nn == 0 | ||
112 | // this behavior is undocumented | ||
113 | // so we are on purpose not using io.ReadFull | ||
114 | // implementation because this can lead | ||
115 | // to custom handling, to avoid that | ||
116 | // we simply modify the original io.ReadFull | ||
117 | // implementation to avoid this issue. | ||
118 | // io.ErrUnexpectedEOF with nn == 0 really | ||
119 | // means that io.EOF | ||
120 | if err == io.ErrUnexpectedEOF && nn == 0 { | ||
121 | err = io.EOF | ||
122 | } | ||
123 | n += nn | ||
124 | } | ||
125 | if n >= len(buf) { | ||
126 | err = nil | ||
127 | } else if n > 0 && err == io.EOF { | ||
128 | err = io.ErrUnexpectedEOF | ||
129 | } | ||
130 | return | ||
131 | } | ||
132 | |||
133 | func cleanEmptyEntries(fields log.Fields) log.Fields { | ||
134 | cleanFields := log.Fields{} | ||
135 | for k, v := range fields { | ||
136 | if v != "" { | ||
137 | cleanFields[k] = v | ||
138 | } | ||
139 | } | ||
140 | return cleanFields | ||
141 | } | ||
142 | |||
143 | // log successful test runs | ||
144 | func successLogger(testName, function string, args map[string]interface{}, startTime time.Time) *log.Entry { | ||
145 | // calculate the test case duration | ||
146 | duration := time.Since(startTime) | ||
147 | // log with the fields as per mint | ||
148 | fields := log.Fields{"name": "minio-go: " + testName, "function": function, "args": args, "duration": duration.Nanoseconds() / 1000000, "status": "PASS"} | ||
149 | return log.WithFields(cleanEmptyEntries(fields)) | ||
150 | } | ||
151 | |||
152 | // As few of the features are not available in Gateway(s) currently, Check if err value is NotImplemented, | ||
153 | // and log as NA in that case and continue execution. Otherwise log as failure and return | ||
154 | func logError(testName, function string, args map[string]interface{}, startTime time.Time, alert, message string, err error) { | ||
155 | // If server returns NotImplemented we assume it is gateway mode and hence log it as info and move on to next tests | ||
156 | // Special case for ComposeObject API as it is implemented on client side and adds specific error details like `Error in upload-part-copy` in | ||
157 | // addition to NotImplemented error returned from server | ||
158 | if isErrNotImplemented(err) { | ||
159 | ignoredLog(testName, function, args, startTime, message).Info() | ||
160 | } else if isRunOnFail() { | ||
161 | failureLog(testName, function, args, startTime, alert, message, err).Error() | ||
162 | } else { | ||
163 | failureLog(testName, function, args, startTime, alert, message, err).Fatal() | ||
164 | } | ||
165 | } | ||
166 | |||
167 | // log failed test runs | ||
168 | func failureLog(testName, function string, args map[string]interface{}, startTime time.Time, alert, message string, err error) *log.Entry { | ||
169 | // calculate the test case duration | ||
170 | duration := time.Since(startTime) | ||
171 | var fields log.Fields | ||
172 | // log with the fields as per mint | ||
173 | if err != nil { | ||
174 | fields = log.Fields{ | ||
175 | "name": "minio-go: " + testName, "function": function, "args": args, | ||
176 | "duration": duration.Nanoseconds() / 1000000, "status": "FAIL", "alert": alert, "message": message, "error": err, | ||
177 | } | ||
178 | } else { | ||
179 | fields = log.Fields{ | ||
180 | "name": "minio-go: " + testName, "function": function, "args": args, | ||
181 | "duration": duration.Nanoseconds() / 1000000, "status": "FAIL", "alert": alert, "message": message, | ||
182 | } | ||
183 | } | ||
184 | return log.WithFields(cleanEmptyEntries(fields)) | ||
185 | } | ||
186 | |||
187 | // log not applicable test runs | ||
188 | func ignoredLog(testName, function string, args map[string]interface{}, startTime time.Time, alert string) *log.Entry { | ||
189 | // calculate the test case duration | ||
190 | duration := time.Since(startTime) | ||
191 | // log with the fields as per mint | ||
192 | fields := log.Fields{ | ||
193 | "name": "minio-go: " + testName, "function": function, "args": args, | ||
194 | "duration": duration.Nanoseconds() / 1000000, "status": "NA", "alert": strings.Split(alert, " ")[0] + " is NotImplemented", | ||
195 | } | ||
196 | return log.WithFields(cleanEmptyEntries(fields)) | ||
197 | } | ||
198 | |||
199 | // Delete objects in given bucket, recursively | ||
200 | func cleanupBucket(bucketName string, c *minio.Client) error { | ||
201 | // Create a done channel to control 'ListObjectsV2' go routine. | ||
202 | doneCh := make(chan struct{}) | ||
203 | // Exit cleanly upon return. | ||
204 | defer close(doneCh) | ||
205 | // Iterate over all objects in the bucket via listObjectsV2 and delete | ||
206 | for objCh := range c.ListObjects(context.Background(), bucketName, minio.ListObjectsOptions{Recursive: true}) { | ||
207 | if objCh.Err != nil { | ||
208 | return objCh.Err | ||
209 | } | ||
210 | if objCh.Key != "" { | ||
211 | err := c.RemoveObject(context.Background(), bucketName, objCh.Key, minio.RemoveObjectOptions{}) | ||
212 | if err != nil { | ||
213 | return err | ||
214 | } | ||
215 | } | ||
216 | } | ||
217 | for objPartInfo := range c.ListIncompleteUploads(context.Background(), bucketName, "", true) { | ||
218 | if objPartInfo.Err != nil { | ||
219 | return objPartInfo.Err | ||
220 | } | ||
221 | if objPartInfo.Key != "" { | ||
222 | err := c.RemoveIncompleteUpload(context.Background(), bucketName, objPartInfo.Key) | ||
223 | if err != nil { | ||
224 | return err | ||
225 | } | ||
226 | } | ||
227 | } | ||
228 | // objects are already deleted, clear the buckets now | ||
229 | err := c.RemoveBucket(context.Background(), bucketName) | ||
230 | if err != nil { | ||
231 | return err | ||
232 | } | ||
233 | return err | ||
234 | } | ||
235 | |||
236 | func cleanupVersionedBucket(bucketName string, c *minio.Client) error { | ||
237 | doneCh := make(chan struct{}) | ||
238 | defer close(doneCh) | ||
239 | for obj := range c.ListObjects(context.Background(), bucketName, minio.ListObjectsOptions{WithVersions: true, Recursive: true}) { | ||
240 | if obj.Err != nil { | ||
241 | return obj.Err | ||
242 | } | ||
243 | if obj.Key != "" { | ||
244 | err := c.RemoveObject(context.Background(), bucketName, obj.Key, | ||
245 | minio.RemoveObjectOptions{VersionID: obj.VersionID, GovernanceBypass: true}) | ||
246 | if err != nil { | ||
247 | return err | ||
248 | } | ||
249 | } | ||
250 | } | ||
251 | for objPartInfo := range c.ListIncompleteUploads(context.Background(), bucketName, "", true) { | ||
252 | if objPartInfo.Err != nil { | ||
253 | return objPartInfo.Err | ||
254 | } | ||
255 | if objPartInfo.Key != "" { | ||
256 | err := c.RemoveIncompleteUpload(context.Background(), bucketName, objPartInfo.Key) | ||
257 | if err != nil { | ||
258 | return err | ||
259 | } | ||
260 | } | ||
261 | } | ||
262 | // objects are already deleted, clear the buckets now | ||
263 | err := c.RemoveBucket(context.Background(), bucketName) | ||
264 | if err != nil { | ||
265 | for obj := range c.ListObjects(context.Background(), bucketName, minio.ListObjectsOptions{WithVersions: true, Recursive: true}) { | ||
266 | log.Println("found", obj.Key, obj.VersionID) | ||
267 | } | ||
268 | return err | ||
269 | } | ||
270 | return err | ||
271 | } | ||
272 | |||
273 | func isErrNotImplemented(err error) bool { | ||
274 | return minio.ToErrorResponse(err).Code == "NotImplemented" | ||
275 | } | ||
276 | |||
277 | func isRunOnFail() bool { | ||
278 | return os.Getenv("RUN_ON_FAIL") == "1" | ||
279 | } | ||
280 | |||
281 | func init() { | ||
282 | // If server endpoint is not set, all tests default to | ||
283 | // using https://play.min.io | ||
284 | if os.Getenv(serverEndpoint) == "" { | ||
285 | os.Setenv(serverEndpoint, "play.min.io") | ||
286 | os.Setenv(accessKey, "Q3AM3UQ867SPQQA43P2F") | ||
287 | os.Setenv(secretKey, "zuf+tfteSlswRu7BJ86wekitnifILbZam1KYY3TG") | ||
288 | os.Setenv(enableHTTPS, "1") | ||
289 | } | ||
290 | } | ||
291 | |||
292 | var mintDataDir = os.Getenv("MINT_DATA_DIR") | ||
293 | |||
294 | func getMintDataDirFilePath(filename string) (fp string) { | ||
295 | if mintDataDir == "" { | ||
296 | return | ||
297 | } | ||
298 | return filepath.Join(mintDataDir, filename) | ||
299 | } | ||
300 | |||
301 | func newRandomReader(seed, size int64) io.Reader { | ||
302 | return io.LimitReader(rand.New(rand.NewSource(seed)), size) | ||
303 | } | ||
304 | |||
305 | func mustCrcReader(r io.Reader) uint32 { | ||
306 | crc := crc32.NewIEEE() | ||
307 | _, err := io.Copy(crc, r) | ||
308 | if err != nil { | ||
309 | panic(err) | ||
310 | } | ||
311 | return crc.Sum32() | ||
312 | } | ||
313 | |||
314 | func crcMatches(r io.Reader, want uint32) error { | ||
315 | crc := crc32.NewIEEE() | ||
316 | _, err := io.Copy(crc, r) | ||
317 | if err != nil { | ||
318 | panic(err) | ||
319 | } | ||
320 | got := crc.Sum32() | ||
321 | if got != want { | ||
322 | return fmt.Errorf("crc mismatch, want %x, got %x", want, got) | ||
323 | } | ||
324 | return nil | ||
325 | } | ||
326 | |||
327 | func crcMatchesName(r io.Reader, name string) error { | ||
328 | want := dataFileCRC32[name] | ||
329 | crc := crc32.NewIEEE() | ||
330 | _, err := io.Copy(crc, r) | ||
331 | if err != nil { | ||
332 | panic(err) | ||
333 | } | ||
334 | got := crc.Sum32() | ||
335 | if got != want { | ||
336 | return fmt.Errorf("crc mismatch, want %x, got %x", want, got) | ||
337 | } | ||
338 | return nil | ||
339 | } | ||
340 | |||
341 | // read data from file if it exists or optionally create a buffer of particular size | ||
342 | func getDataReader(fileName string) io.ReadCloser { | ||
343 | if mintDataDir == "" { | ||
344 | size := int64(dataFileMap[fileName]) | ||
345 | if _, ok := dataFileCRC32[fileName]; !ok { | ||
346 | dataFileCRC32[fileName] = mustCrcReader(newRandomReader(size, size)) | ||
347 | } | ||
348 | return io.NopCloser(newRandomReader(size, size)) | ||
349 | } | ||
350 | reader, _ := os.Open(getMintDataDirFilePath(fileName)) | ||
351 | if _, ok := dataFileCRC32[fileName]; !ok { | ||
352 | dataFileCRC32[fileName] = mustCrcReader(reader) | ||
353 | reader.Close() | ||
354 | reader, _ = os.Open(getMintDataDirFilePath(fileName)) | ||
355 | } | ||
356 | return reader | ||
357 | } | ||
358 | |||
359 | // randString generates random names and prepends them with a known prefix. | ||
360 | func randString(n int, src rand.Source, prefix string) string { | ||
361 | b := make([]byte, n) | ||
362 | // A rand.Int63() generates 63 random bits, enough for letterIdxMax letters! | ||
363 | for i, cache, remain := n-1, src.Int63(), letterIdxMax; i >= 0; { | ||
364 | if remain == 0 { | ||
365 | cache, remain = src.Int63(), letterIdxMax | ||
366 | } | ||
367 | if idx := int(cache & letterIdxMask); idx < len(letterBytes) { | ||
368 | b[i] = letterBytes[idx] | ||
369 | i-- | ||
370 | } | ||
371 | cache >>= letterIdxBits | ||
372 | remain-- | ||
373 | } | ||
374 | return prefix + string(b[0:30-len(prefix)]) | ||
375 | } | ||
376 | |||
377 | var dataFileMap = map[string]int{ | ||
378 | "datafile-0-b": 0, | ||
379 | "datafile-1-b": 1, | ||
380 | "datafile-1-kB": 1 * humanize.KiByte, | ||
381 | "datafile-10-kB": 10 * humanize.KiByte, | ||
382 | "datafile-33-kB": 33 * humanize.KiByte, | ||
383 | "datafile-100-kB": 100 * humanize.KiByte, | ||
384 | "datafile-1.03-MB": 1056 * humanize.KiByte, | ||
385 | "datafile-1-MB": 1 * humanize.MiByte, | ||
386 | "datafile-5-MB": 5 * humanize.MiByte, | ||
387 | "datafile-6-MB": 6 * humanize.MiByte, | ||
388 | "datafile-11-MB": 11 * humanize.MiByte, | ||
389 | "datafile-65-MB": 65 * humanize.MiByte, | ||
390 | "datafile-129-MB": 129 * humanize.MiByte, | ||
391 | } | ||
392 | |||
393 | var dataFileCRC32 = map[string]uint32{} | ||
394 | |||
395 | func isFullMode() bool { | ||
396 | return os.Getenv("MINT_MODE") == "full" | ||
397 | } | ||
398 | |||
399 | func getFuncName() string { | ||
400 | return getFuncNameLoc(2) | ||
401 | } | ||
402 | |||
403 | func getFuncNameLoc(caller int) string { | ||
404 | pc, _, _, _ := runtime.Caller(caller) | ||
405 | return strings.TrimPrefix(runtime.FuncForPC(pc).Name(), "main.") | ||
406 | } | ||
407 | |||
408 | // Tests bucket re-create errors. | ||
409 | func testMakeBucketError() { | ||
410 | region := "eu-central-1" | ||
411 | |||
412 | // initialize logging params | ||
413 | startTime := time.Now() | ||
414 | testName := getFuncName() | ||
415 | function := "MakeBucket(bucketName, region)" | ||
416 | // initialize logging params | ||
417 | args := map[string]interface{}{ | ||
418 | "bucketName": "", | ||
419 | "region": region, | ||
420 | } | ||
421 | |||
422 | // Seed random based on current time. | ||
423 | rand.Seed(time.Now().Unix()) | ||
424 | |||
425 | // Instantiate new minio client object. | ||
426 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
427 | &minio.Options{ | ||
428 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
429 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
430 | }) | ||
431 | if err != nil { | ||
432 | logError(testName, function, args, startTime, "", "MinIO client creation failed", err) | ||
433 | return | ||
434 | } | ||
435 | |||
436 | // Enable tracing, write to stderr. | ||
437 | // c.TraceOn(os.Stderr) | ||
438 | |||
439 | // Set user agent. | ||
440 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
441 | |||
442 | // Generate a new random bucket name. | ||
443 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
444 | args["bucketName"] = bucketName | ||
445 | |||
446 | // Make a new bucket in 'eu-central-1'. | ||
447 | if err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: region}); err != nil { | ||
448 | logError(testName, function, args, startTime, "", "MakeBucket Failed", err) | ||
449 | return | ||
450 | } | ||
451 | defer cleanupBucket(bucketName, c) | ||
452 | |||
453 | if err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: region}); err == nil { | ||
454 | logError(testName, function, args, startTime, "", "Bucket already exists", err) | ||
455 | return | ||
456 | } | ||
457 | // Verify valid error response from server. | ||
458 | if minio.ToErrorResponse(err).Code != "BucketAlreadyExists" && | ||
459 | minio.ToErrorResponse(err).Code != "BucketAlreadyOwnedByYou" { | ||
460 | logError(testName, function, args, startTime, "", "Invalid error returned by server", err) | ||
461 | return | ||
462 | } | ||
463 | |||
464 | successLogger(testName, function, args, startTime).Info() | ||
465 | } | ||
466 | |||
467 | func testMetadataSizeLimit() { | ||
468 | startTime := time.Now() | ||
469 | testName := getFuncName() | ||
470 | function := "PutObject(bucketName, objectName, reader, objectSize, opts)" | ||
471 | args := map[string]interface{}{ | ||
472 | "bucketName": "", | ||
473 | "objectName": "", | ||
474 | "opts.UserMetadata": "", | ||
475 | } | ||
476 | rand.Seed(startTime.Unix()) | ||
477 | |||
478 | // Instantiate new minio client object. | ||
479 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
480 | &minio.Options{ | ||
481 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
482 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
483 | }) | ||
484 | if err != nil { | ||
485 | logError(testName, function, args, startTime, "", "MinIO client creation failed", err) | ||
486 | return | ||
487 | } | ||
488 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
489 | |||
490 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
491 | args["bucketName"] = bucketName | ||
492 | |||
493 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
494 | args["objectName"] = objectName | ||
495 | |||
496 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
497 | if err != nil { | ||
498 | logError(testName, function, args, startTime, "", "Make bucket failed", err) | ||
499 | return | ||
500 | } | ||
501 | |||
502 | defer cleanupBucket(bucketName, c) | ||
503 | |||
504 | const HeaderSizeLimit = 8 * 1024 | ||
505 | const UserMetadataLimit = 2 * 1024 | ||
506 | |||
507 | // Meta-data greater than the 2 KB limit of AWS - PUT calls with this meta-data should fail | ||
508 | metadata := make(map[string]string) | ||
509 | metadata["X-Amz-Meta-Mint-Test"] = string(bytes.Repeat([]byte("m"), 1+UserMetadataLimit-len("X-Amz-Meta-Mint-Test"))) | ||
510 | args["metadata"] = fmt.Sprint(metadata) | ||
511 | |||
512 | _, err = c.PutObject(context.Background(), bucketName, objectName, bytes.NewReader(nil), 0, minio.PutObjectOptions{UserMetadata: metadata}) | ||
513 | if err == nil { | ||
514 | logError(testName, function, args, startTime, "", "Created object with user-defined metadata exceeding metadata size limits", nil) | ||
515 | return | ||
516 | } | ||
517 | |||
518 | // Meta-data (headers) greater than the 8 KB limit of AWS - PUT calls with this meta-data should fail | ||
519 | metadata = make(map[string]string) | ||
520 | metadata["X-Amz-Mint-Test"] = string(bytes.Repeat([]byte("m"), 1+HeaderSizeLimit-len("X-Amz-Mint-Test"))) | ||
521 | args["metadata"] = fmt.Sprint(metadata) | ||
522 | _, err = c.PutObject(context.Background(), bucketName, objectName, bytes.NewReader(nil), 0, minio.PutObjectOptions{UserMetadata: metadata}) | ||
523 | if err == nil { | ||
524 | logError(testName, function, args, startTime, "", "Created object with headers exceeding header size limits", nil) | ||
525 | return | ||
526 | } | ||
527 | |||
528 | successLogger(testName, function, args, startTime).Info() | ||
529 | } | ||
530 | |||
531 | // Tests various bucket supported formats. | ||
532 | func testMakeBucketRegions() { | ||
533 | region := "eu-central-1" | ||
534 | // initialize logging params | ||
535 | startTime := time.Now() | ||
536 | testName := getFuncName() | ||
537 | function := "MakeBucket(bucketName, region)" | ||
538 | // initialize logging params | ||
539 | args := map[string]interface{}{ | ||
540 | "bucketName": "", | ||
541 | "region": region, | ||
542 | } | ||
543 | |||
544 | // Seed random based on current time. | ||
545 | rand.Seed(time.Now().Unix()) | ||
546 | |||
547 | // Instantiate new minio client object. | ||
548 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
549 | &minio.Options{ | ||
550 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
551 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
552 | }) | ||
553 | if err != nil { | ||
554 | logError(testName, function, args, startTime, "", "MinIO client creation failed", err) | ||
555 | return | ||
556 | } | ||
557 | |||
558 | // Enable tracing, write to stderr. | ||
559 | // c.TraceOn(os.Stderr) | ||
560 | |||
561 | // Set user agent. | ||
562 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
563 | |||
564 | // Generate a new random bucket name. | ||
565 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
566 | args["bucketName"] = bucketName | ||
567 | |||
568 | // Make a new bucket in 'eu-central-1'. | ||
569 | if err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: region}); err != nil { | ||
570 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
571 | return | ||
572 | } | ||
573 | |||
574 | // Delete all objects and buckets | ||
575 | if err = cleanupBucket(bucketName, c); err != nil { | ||
576 | logError(testName, function, args, startTime, "", "CleanupBucket failed", err) | ||
577 | return | ||
578 | } | ||
579 | |||
580 | // Make a new bucket with '.' in its name, in 'us-west-2'. This | ||
581 | // request is internally staged into a path style instead of | ||
582 | // virtual host style. | ||
583 | region = "us-west-2" | ||
584 | args["region"] = region | ||
585 | if err = c.MakeBucket(context.Background(), bucketName+".withperiod", minio.MakeBucketOptions{Region: region}); err != nil { | ||
586 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
587 | return | ||
588 | } | ||
589 | |||
590 | // Delete all objects and buckets | ||
591 | if err = cleanupBucket(bucketName+".withperiod", c); err != nil { | ||
592 | logError(testName, function, args, startTime, "", "CleanupBucket failed", err) | ||
593 | return | ||
594 | } | ||
595 | successLogger(testName, function, args, startTime).Info() | ||
596 | } | ||
597 | |||
598 | // Test PutObject using a large data to trigger multipart readat | ||
599 | func testPutObjectReadAt() { | ||
600 | // initialize logging params | ||
601 | startTime := time.Now() | ||
602 | testName := getFuncName() | ||
603 | function := "PutObject(bucketName, objectName, reader, opts)" | ||
604 | args := map[string]interface{}{ | ||
605 | "bucketName": "", | ||
606 | "objectName": "", | ||
607 | "opts": "objectContentType", | ||
608 | } | ||
609 | |||
610 | // Seed random based on current time. | ||
611 | rand.Seed(time.Now().Unix()) | ||
612 | |||
613 | // Instantiate new minio client object. | ||
614 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
615 | &minio.Options{ | ||
616 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
617 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
618 | }) | ||
619 | if err != nil { | ||
620 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
621 | return | ||
622 | } | ||
623 | |||
624 | // Enable tracing, write to stderr. | ||
625 | // c.TraceOn(os.Stderr) | ||
626 | |||
627 | // Set user agent. | ||
628 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
629 | |||
630 | // Generate a new random bucket name. | ||
631 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
632 | args["bucketName"] = bucketName | ||
633 | |||
634 | // Make a new bucket. | ||
635 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
636 | if err != nil { | ||
637 | logError(testName, function, args, startTime, "", "Make bucket failed", err) | ||
638 | return | ||
639 | } | ||
640 | |||
641 | defer cleanupBucket(bucketName, c) | ||
642 | |||
643 | bufSize := dataFileMap["datafile-129-MB"] | ||
644 | reader := getDataReader("datafile-129-MB") | ||
645 | defer reader.Close() | ||
646 | |||
647 | // Save the data | ||
648 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
649 | args["objectName"] = objectName | ||
650 | |||
651 | // Object content type | ||
652 | objectContentType := "binary/octet-stream" | ||
653 | args["objectContentType"] = objectContentType | ||
654 | |||
655 | _, err = c.PutObject(context.Background(), bucketName, objectName, reader, int64(bufSize), minio.PutObjectOptions{ContentType: objectContentType}) | ||
656 | if err != nil { | ||
657 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
658 | return | ||
659 | } | ||
660 | |||
661 | // Read the data back | ||
662 | r, err := c.GetObject(context.Background(), bucketName, objectName, minio.GetObjectOptions{}) | ||
663 | if err != nil { | ||
664 | logError(testName, function, args, startTime, "", "Get Object failed", err) | ||
665 | return | ||
666 | } | ||
667 | |||
668 | st, err := r.Stat() | ||
669 | if err != nil { | ||
670 | logError(testName, function, args, startTime, "", "Stat Object failed", err) | ||
671 | return | ||
672 | } | ||
673 | if st.Size != int64(bufSize) { | ||
674 | logError(testName, function, args, startTime, "", fmt.Sprintf("Number of bytes in stat does not match, expected %d got %d", bufSize, st.Size), err) | ||
675 | return | ||
676 | } | ||
677 | if st.ContentType != objectContentType && st.ContentType != "application/octet-stream" { | ||
678 | logError(testName, function, args, startTime, "", "Content types don't match", err) | ||
679 | return | ||
680 | } | ||
681 | if err := crcMatchesName(r, "datafile-129-MB"); err != nil { | ||
682 | logError(testName, function, args, startTime, "", "data CRC check failed", err) | ||
683 | return | ||
684 | } | ||
685 | if err := r.Close(); err != nil { | ||
686 | logError(testName, function, args, startTime, "", "Object Close failed", err) | ||
687 | return | ||
688 | } | ||
689 | if err := r.Close(); err == nil { | ||
690 | logError(testName, function, args, startTime, "", "Object is already closed, didn't return error on Close", err) | ||
691 | return | ||
692 | } | ||
693 | |||
694 | successLogger(testName, function, args, startTime).Info() | ||
695 | } | ||
696 | |||
697 | func testListObjectVersions() { | ||
698 | // initialize logging params | ||
699 | startTime := time.Now() | ||
700 | testName := getFuncName() | ||
701 | function := "ListObjectVersions(bucketName, prefix, recursive)" | ||
702 | args := map[string]interface{}{ | ||
703 | "bucketName": "", | ||
704 | "prefix": "", | ||
705 | "recursive": "", | ||
706 | } | ||
707 | |||
708 | // Seed random based on current time. | ||
709 | rand.Seed(time.Now().Unix()) | ||
710 | |||
711 | // Instantiate new minio client object. | ||
712 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
713 | &minio.Options{ | ||
714 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
715 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
716 | }) | ||
717 | if err != nil { | ||
718 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
719 | return | ||
720 | } | ||
721 | |||
722 | // Enable tracing, write to stderr. | ||
723 | // c.TraceOn(os.Stderr) | ||
724 | |||
725 | // Set user agent. | ||
726 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
727 | |||
728 | // Generate a new random bucket name. | ||
729 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
730 | args["bucketName"] = bucketName | ||
731 | |||
732 | // Make a new bucket. | ||
733 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1", ObjectLocking: true}) | ||
734 | if err != nil { | ||
735 | logError(testName, function, args, startTime, "", "Make bucket failed", err) | ||
736 | return | ||
737 | } | ||
738 | |||
739 | err = c.EnableVersioning(context.Background(), bucketName) | ||
740 | if err != nil { | ||
741 | logError(testName, function, args, startTime, "", "Enable versioning failed", err) | ||
742 | return | ||
743 | } | ||
744 | |||
745 | // Save the data | ||
746 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
747 | args["objectName"] = objectName | ||
748 | |||
749 | bufSize := dataFileMap["datafile-10-kB"] | ||
750 | reader := getDataReader("datafile-10-kB") | ||
751 | |||
752 | _, err = c.PutObject(context.Background(), bucketName, objectName, reader, int64(bufSize), minio.PutObjectOptions{}) | ||
753 | if err != nil { | ||
754 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
755 | return | ||
756 | } | ||
757 | reader.Close() | ||
758 | |||
759 | bufSize = dataFileMap["datafile-1-b"] | ||
760 | reader = getDataReader("datafile-1-b") | ||
761 | _, err = c.PutObject(context.Background(), bucketName, objectName, reader, int64(bufSize), minio.PutObjectOptions{}) | ||
762 | if err != nil { | ||
763 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
764 | return | ||
765 | } | ||
766 | reader.Close() | ||
767 | |||
768 | err = c.RemoveObject(context.Background(), bucketName, objectName, minio.RemoveObjectOptions{}) | ||
769 | if err != nil { | ||
770 | logError(testName, function, args, startTime, "", "Unexpected object deletion", err) | ||
771 | return | ||
772 | } | ||
773 | |||
774 | var deleteMarkers, versions int | ||
775 | |||
776 | objectsInfo := c.ListObjects(context.Background(), bucketName, minio.ListObjectsOptions{WithVersions: true, Recursive: true}) | ||
777 | for info := range objectsInfo { | ||
778 | if info.Err != nil { | ||
779 | logError(testName, function, args, startTime, "", "Unexpected error during listing objects", err) | ||
780 | return | ||
781 | } | ||
782 | if info.Key != objectName { | ||
783 | logError(testName, function, args, startTime, "", "Unexpected object name in listing objects", nil) | ||
784 | return | ||
785 | } | ||
786 | if info.VersionID == "" { | ||
787 | logError(testName, function, args, startTime, "", "Unexpected version id in listing objects", nil) | ||
788 | return | ||
789 | } | ||
790 | if info.IsDeleteMarker { | ||
791 | deleteMarkers++ | ||
792 | if !info.IsLatest { | ||
793 | logError(testName, function, args, startTime, "", "Unexpected IsLatest field in listing objects", nil) | ||
794 | return | ||
795 | } | ||
796 | } else { | ||
797 | versions++ | ||
798 | } | ||
799 | } | ||
800 | |||
801 | if deleteMarkers != 1 { | ||
802 | logError(testName, function, args, startTime, "", "Unexpected number of DeleteMarker elements in listing objects", nil) | ||
803 | return | ||
804 | } | ||
805 | |||
806 | if versions != 2 { | ||
807 | logError(testName, function, args, startTime, "", "Unexpected number of Version elements in listing objects", nil) | ||
808 | return | ||
809 | } | ||
810 | |||
811 | // Delete all objects and their versions as long as the bucket itself | ||
812 | if err = cleanupVersionedBucket(bucketName, c); err != nil { | ||
813 | logError(testName, function, args, startTime, "", "CleanupBucket failed", err) | ||
814 | return | ||
815 | } | ||
816 | |||
817 | successLogger(testName, function, args, startTime).Info() | ||
818 | } | ||
819 | |||
820 | func testStatObjectWithVersioning() { | ||
821 | // initialize logging params | ||
822 | startTime := time.Now() | ||
823 | testName := getFuncName() | ||
824 | function := "StatObject" | ||
825 | args := map[string]interface{}{} | ||
826 | |||
827 | // Seed random based on current time. | ||
828 | rand.Seed(time.Now().Unix()) | ||
829 | |||
830 | // Instantiate new minio client object. | ||
831 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
832 | &minio.Options{ | ||
833 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
834 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
835 | }) | ||
836 | if err != nil { | ||
837 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
838 | return | ||
839 | } | ||
840 | |||
841 | // Enable tracing, write to stderr. | ||
842 | // c.TraceOn(os.Stderr) | ||
843 | |||
844 | // Set user agent. | ||
845 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
846 | |||
847 | // Generate a new random bucket name. | ||
848 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
849 | args["bucketName"] = bucketName | ||
850 | |||
851 | // Make a new bucket. | ||
852 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1", ObjectLocking: true}) | ||
853 | if err != nil { | ||
854 | logError(testName, function, args, startTime, "", "Make bucket failed", err) | ||
855 | return | ||
856 | } | ||
857 | |||
858 | err = c.EnableVersioning(context.Background(), bucketName) | ||
859 | if err != nil { | ||
860 | logError(testName, function, args, startTime, "", "Enable versioning failed", err) | ||
861 | return | ||
862 | } | ||
863 | |||
864 | // Save the data | ||
865 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
866 | args["objectName"] = objectName | ||
867 | |||
868 | bufSize := dataFileMap["datafile-10-kB"] | ||
869 | reader := getDataReader("datafile-10-kB") | ||
870 | |||
871 | _, err = c.PutObject(context.Background(), bucketName, objectName, reader, int64(bufSize), minio.PutObjectOptions{}) | ||
872 | if err != nil { | ||
873 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
874 | return | ||
875 | } | ||
876 | reader.Close() | ||
877 | |||
878 | bufSize = dataFileMap["datafile-1-b"] | ||
879 | reader = getDataReader("datafile-1-b") | ||
880 | _, err = c.PutObject(context.Background(), bucketName, objectName, reader, int64(bufSize), minio.PutObjectOptions{}) | ||
881 | if err != nil { | ||
882 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
883 | return | ||
884 | } | ||
885 | reader.Close() | ||
886 | |||
887 | objectsInfo := c.ListObjects(context.Background(), bucketName, minio.ListObjectsOptions{WithVersions: true, Recursive: true}) | ||
888 | |||
889 | var results []minio.ObjectInfo | ||
890 | for info := range objectsInfo { | ||
891 | if info.Err != nil { | ||
892 | logError(testName, function, args, startTime, "", "Unexpected error during listing objects", err) | ||
893 | return | ||
894 | } | ||
895 | results = append(results, info) | ||
896 | } | ||
897 | |||
898 | if len(results) != 2 { | ||
899 | logError(testName, function, args, startTime, "", "Unexpected number of Version elements in listing objects", nil) | ||
900 | return | ||
901 | } | ||
902 | |||
903 | for i := 0; i < len(results); i++ { | ||
904 | opts := minio.StatObjectOptions{VersionID: results[i].VersionID} | ||
905 | statInfo, err := c.StatObject(context.Background(), bucketName, objectName, opts) | ||
906 | if err != nil { | ||
907 | logError(testName, function, args, startTime, "", "error during HEAD object", err) | ||
908 | return | ||
909 | } | ||
910 | if statInfo.VersionID == "" || statInfo.VersionID != results[i].VersionID { | ||
911 | logError(testName, function, args, startTime, "", "error during HEAD object, unexpected version id", err) | ||
912 | return | ||
913 | } | ||
914 | if statInfo.ETag != results[i].ETag { | ||
915 | logError(testName, function, args, startTime, "", "error during HEAD object, unexpected ETag", err) | ||
916 | return | ||
917 | } | ||
918 | if statInfo.LastModified.Unix() != results[i].LastModified.Unix() { | ||
919 | logError(testName, function, args, startTime, "", "error during HEAD object, unexpected Last-Modified", err) | ||
920 | return | ||
921 | } | ||
922 | if statInfo.Size != results[i].Size { | ||
923 | logError(testName, function, args, startTime, "", "error during HEAD object, unexpected Content-Length", err) | ||
924 | return | ||
925 | } | ||
926 | } | ||
927 | |||
928 | // Delete all objects and their versions as long as the bucket itself | ||
929 | if err = cleanupVersionedBucket(bucketName, c); err != nil { | ||
930 | logError(testName, function, args, startTime, "", "CleanupBucket failed", err) | ||
931 | return | ||
932 | } | ||
933 | |||
934 | successLogger(testName, function, args, startTime).Info() | ||
935 | } | ||
936 | |||
937 | func testGetObjectWithVersioning() { | ||
938 | // initialize logging params | ||
939 | startTime := time.Now() | ||
940 | testName := getFuncName() | ||
941 | function := "GetObject()" | ||
942 | args := map[string]interface{}{} | ||
943 | |||
944 | // Seed random based on current time. | ||
945 | rand.Seed(time.Now().Unix()) | ||
946 | |||
947 | // Instantiate new minio client object. | ||
948 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
949 | &minio.Options{ | ||
950 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
951 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
952 | }) | ||
953 | if err != nil { | ||
954 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
955 | return | ||
956 | } | ||
957 | |||
958 | // Enable tracing, write to stderr. | ||
959 | // c.TraceOn(os.Stderr) | ||
960 | |||
961 | // Set user agent. | ||
962 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
963 | |||
964 | // Generate a new random bucket name. | ||
965 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
966 | args["bucketName"] = bucketName | ||
967 | |||
968 | // Make a new bucket. | ||
969 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1", ObjectLocking: true}) | ||
970 | if err != nil { | ||
971 | logError(testName, function, args, startTime, "", "Make bucket failed", err) | ||
972 | return | ||
973 | } | ||
974 | |||
975 | err = c.EnableVersioning(context.Background(), bucketName) | ||
976 | if err != nil { | ||
977 | logError(testName, function, args, startTime, "", "Enable versioning failed", err) | ||
978 | return | ||
979 | } | ||
980 | |||
981 | // Save the data | ||
982 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
983 | args["objectName"] = objectName | ||
984 | |||
985 | // Save the contents of datafiles to check with GetObject() reader output later | ||
986 | var buffers [][]byte | ||
987 | testFiles := []string{"datafile-1-b", "datafile-10-kB"} | ||
988 | |||
989 | for _, testFile := range testFiles { | ||
990 | r := getDataReader(testFile) | ||
991 | buf, err := io.ReadAll(r) | ||
992 | if err != nil { | ||
993 | logError(testName, function, args, startTime, "", "unexpected failure", err) | ||
994 | return | ||
995 | } | ||
996 | r.Close() | ||
997 | _, err = c.PutObject(context.Background(), bucketName, objectName, bytes.NewReader(buf), int64(len(buf)), minio.PutObjectOptions{}) | ||
998 | if err != nil { | ||
999 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
1000 | return | ||
1001 | } | ||
1002 | buffers = append(buffers, buf) | ||
1003 | } | ||
1004 | |||
1005 | objectsInfo := c.ListObjects(context.Background(), bucketName, minio.ListObjectsOptions{WithVersions: true, Recursive: true}) | ||
1006 | |||
1007 | var results []minio.ObjectInfo | ||
1008 | for info := range objectsInfo { | ||
1009 | if info.Err != nil { | ||
1010 | logError(testName, function, args, startTime, "", "Unexpected error during listing objects", err) | ||
1011 | return | ||
1012 | } | ||
1013 | results = append(results, info) | ||
1014 | } | ||
1015 | |||
1016 | if len(results) != 2 { | ||
1017 | logError(testName, function, args, startTime, "", "Unexpected number of Version elements in listing objects", nil) | ||
1018 | return | ||
1019 | } | ||
1020 | |||
1021 | sort.SliceStable(results, func(i, j int) bool { | ||
1022 | return results[i].Size < results[j].Size | ||
1023 | }) | ||
1024 | |||
1025 | sort.SliceStable(buffers, func(i, j int) bool { | ||
1026 | return len(buffers[i]) < len(buffers[j]) | ||
1027 | }) | ||
1028 | |||
1029 | for i := 0; i < len(results); i++ { | ||
1030 | opts := minio.GetObjectOptions{VersionID: results[i].VersionID} | ||
1031 | reader, err := c.GetObject(context.Background(), bucketName, objectName, opts) | ||
1032 | if err != nil { | ||
1033 | logError(testName, function, args, startTime, "", "error during GET object", err) | ||
1034 | return | ||
1035 | } | ||
1036 | statInfo, err := reader.Stat() | ||
1037 | if err != nil { | ||
1038 | logError(testName, function, args, startTime, "", "error during calling reader.Stat()", err) | ||
1039 | return | ||
1040 | } | ||
1041 | if statInfo.ETag != results[i].ETag { | ||
1042 | logError(testName, function, args, startTime, "", "error during HEAD object, unexpected ETag", err) | ||
1043 | return | ||
1044 | } | ||
1045 | if statInfo.LastModified.Unix() != results[i].LastModified.Unix() { | ||
1046 | logError(testName, function, args, startTime, "", "error during HEAD object, unexpected Last-Modified", err) | ||
1047 | return | ||
1048 | } | ||
1049 | if statInfo.Size != results[i].Size { | ||
1050 | logError(testName, function, args, startTime, "", "error during HEAD object, unexpected Content-Length", err) | ||
1051 | return | ||
1052 | } | ||
1053 | |||
1054 | tmpBuffer := bytes.NewBuffer([]byte{}) | ||
1055 | _, err = io.Copy(tmpBuffer, reader) | ||
1056 | if err != nil { | ||
1057 | logError(testName, function, args, startTime, "", "unexpected io.Copy()", err) | ||
1058 | return | ||
1059 | } | ||
1060 | |||
1061 | if !bytes.Equal(tmpBuffer.Bytes(), buffers[i]) { | ||
1062 | logError(testName, function, args, startTime, "", "unexpected content of GetObject()", err) | ||
1063 | return | ||
1064 | } | ||
1065 | } | ||
1066 | |||
1067 | // Delete all objects and their versions as long as the bucket itself | ||
1068 | if err = cleanupVersionedBucket(bucketName, c); err != nil { | ||
1069 | logError(testName, function, args, startTime, "", "CleanupBucket failed", err) | ||
1070 | return | ||
1071 | } | ||
1072 | |||
1073 | successLogger(testName, function, args, startTime).Info() | ||
1074 | } | ||
1075 | |||
1076 | func testPutObjectWithVersioning() { | ||
1077 | // initialize logging params | ||
1078 | startTime := time.Now() | ||
1079 | testName := getFuncName() | ||
1080 | function := "GetObject()" | ||
1081 | args := map[string]interface{}{} | ||
1082 | |||
1083 | // Seed random based on current time. | ||
1084 | rand.Seed(time.Now().Unix()) | ||
1085 | |||
1086 | // Instantiate new minio client object. | ||
1087 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
1088 | &minio.Options{ | ||
1089 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
1090 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
1091 | }) | ||
1092 | if err != nil { | ||
1093 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
1094 | return | ||
1095 | } | ||
1096 | |||
1097 | // Enable tracing, write to stderr. | ||
1098 | // c.TraceOn(os.Stderr) | ||
1099 | |||
1100 | // Set user agent. | ||
1101 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
1102 | |||
1103 | // Generate a new random bucket name. | ||
1104 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
1105 | args["bucketName"] = bucketName | ||
1106 | |||
1107 | // Make a new bucket. | ||
1108 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1", ObjectLocking: true}) | ||
1109 | if err != nil { | ||
1110 | logError(testName, function, args, startTime, "", "Make bucket failed", err) | ||
1111 | return | ||
1112 | } | ||
1113 | |||
1114 | err = c.EnableVersioning(context.Background(), bucketName) | ||
1115 | if err != nil { | ||
1116 | logError(testName, function, args, startTime, "", "Enable versioning failed", err) | ||
1117 | return | ||
1118 | } | ||
1119 | |||
1120 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
1121 | args["objectName"] = objectName | ||
1122 | |||
1123 | const n = 10 | ||
1124 | // Read input... | ||
1125 | |||
1126 | // Save the data concurrently. | ||
1127 | var wg sync.WaitGroup | ||
1128 | wg.Add(n) | ||
1129 | buffers := make([][]byte, n) | ||
1130 | var errs [n]error | ||
1131 | for i := 0; i < n; i++ { | ||
1132 | r := newRandomReader(int64((1<<20)*i+i), int64(i)) | ||
1133 | buf, err := io.ReadAll(r) | ||
1134 | if err != nil { | ||
1135 | logError(testName, function, args, startTime, "", "unexpected failure", err) | ||
1136 | return | ||
1137 | } | ||
1138 | buffers[i] = buf | ||
1139 | |||
1140 | go func(i int) { | ||
1141 | defer wg.Done() | ||
1142 | _, errs[i] = c.PutObject(context.Background(), bucketName, objectName, bytes.NewReader(buf), int64(len(buf)), minio.PutObjectOptions{PartSize: 5 << 20}) | ||
1143 | }(i) | ||
1144 | } | ||
1145 | wg.Wait() | ||
1146 | for _, err := range errs { | ||
1147 | if err != nil { | ||
1148 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
1149 | return | ||
1150 | } | ||
1151 | } | ||
1152 | |||
1153 | objectsInfo := c.ListObjects(context.Background(), bucketName, minio.ListObjectsOptions{WithVersions: true, Recursive: true}) | ||
1154 | var results []minio.ObjectInfo | ||
1155 | for info := range objectsInfo { | ||
1156 | if info.Err != nil { | ||
1157 | logError(testName, function, args, startTime, "", "Unexpected error during listing objects", err) | ||
1158 | return | ||
1159 | } | ||
1160 | results = append(results, info) | ||
1161 | } | ||
1162 | |||
1163 | if len(results) != n { | ||
1164 | logError(testName, function, args, startTime, "", "Unexpected number of Version elements in listing objects", nil) | ||
1165 | return | ||
1166 | } | ||
1167 | |||
1168 | sort.Slice(results, func(i, j int) bool { | ||
1169 | return results[i].Size < results[j].Size | ||
1170 | }) | ||
1171 | |||
1172 | sort.Slice(buffers, func(i, j int) bool { | ||
1173 | return len(buffers[i]) < len(buffers[j]) | ||
1174 | }) | ||
1175 | |||
1176 | for i := 0; i < len(results); i++ { | ||
1177 | opts := minio.GetObjectOptions{VersionID: results[i].VersionID} | ||
1178 | reader, err := c.GetObject(context.Background(), bucketName, objectName, opts) | ||
1179 | if err != nil { | ||
1180 | logError(testName, function, args, startTime, "", "error during GET object", err) | ||
1181 | return | ||
1182 | } | ||
1183 | statInfo, err := reader.Stat() | ||
1184 | if err != nil { | ||
1185 | logError(testName, function, args, startTime, "", "error during calling reader.Stat()", err) | ||
1186 | return | ||
1187 | } | ||
1188 | if statInfo.ETag != results[i].ETag { | ||
1189 | logError(testName, function, args, startTime, "", "error during HEAD object, unexpected ETag", err) | ||
1190 | return | ||
1191 | } | ||
1192 | if statInfo.LastModified.Unix() != results[i].LastModified.Unix() { | ||
1193 | logError(testName, function, args, startTime, "", "error during HEAD object, unexpected Last-Modified", err) | ||
1194 | return | ||
1195 | } | ||
1196 | if statInfo.Size != results[i].Size { | ||
1197 | logError(testName, function, args, startTime, "", "error during HEAD object, unexpected Content-Length", err) | ||
1198 | return | ||
1199 | } | ||
1200 | |||
1201 | tmpBuffer := bytes.NewBuffer([]byte{}) | ||
1202 | _, err = io.Copy(tmpBuffer, reader) | ||
1203 | if err != nil { | ||
1204 | logError(testName, function, args, startTime, "", "unexpected io.Copy()", err) | ||
1205 | return | ||
1206 | } | ||
1207 | |||
1208 | if !bytes.Equal(tmpBuffer.Bytes(), buffers[i]) { | ||
1209 | logError(testName, function, args, startTime, "", "unexpected content of GetObject()", err) | ||
1210 | return | ||
1211 | } | ||
1212 | } | ||
1213 | |||
1214 | // Delete all objects and their versions as long as the bucket itself | ||
1215 | if err = cleanupVersionedBucket(bucketName, c); err != nil { | ||
1216 | logError(testName, function, args, startTime, "", "CleanupBucket failed", err) | ||
1217 | return | ||
1218 | } | ||
1219 | |||
1220 | successLogger(testName, function, args, startTime).Info() | ||
1221 | } | ||
1222 | |||
1223 | func testCopyObjectWithVersioning() { | ||
1224 | // initialize logging params | ||
1225 | startTime := time.Now() | ||
1226 | testName := getFuncName() | ||
1227 | function := "CopyObject()" | ||
1228 | args := map[string]interface{}{} | ||
1229 | |||
1230 | // Seed random based on current time. | ||
1231 | rand.Seed(time.Now().Unix()) | ||
1232 | |||
1233 | // Instantiate new minio client object. | ||
1234 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
1235 | &minio.Options{ | ||
1236 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
1237 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
1238 | }) | ||
1239 | if err != nil { | ||
1240 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
1241 | return | ||
1242 | } | ||
1243 | |||
1244 | // Enable tracing, write to stderr. | ||
1245 | // c.TraceOn(os.Stderr) | ||
1246 | |||
1247 | // Set user agent. | ||
1248 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
1249 | |||
1250 | // Generate a new random bucket name. | ||
1251 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
1252 | args["bucketName"] = bucketName | ||
1253 | |||
1254 | // Make a new bucket. | ||
1255 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1", ObjectLocking: true}) | ||
1256 | if err != nil { | ||
1257 | logError(testName, function, args, startTime, "", "Make bucket failed", err) | ||
1258 | return | ||
1259 | } | ||
1260 | |||
1261 | err = c.EnableVersioning(context.Background(), bucketName) | ||
1262 | if err != nil { | ||
1263 | logError(testName, function, args, startTime, "", "Enable versioning failed", err) | ||
1264 | return | ||
1265 | } | ||
1266 | |||
1267 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
1268 | args["objectName"] = objectName | ||
1269 | |||
1270 | testFiles := []string{"datafile-1-b", "datafile-10-kB"} | ||
1271 | for _, testFile := range testFiles { | ||
1272 | r := getDataReader(testFile) | ||
1273 | buf, err := io.ReadAll(r) | ||
1274 | if err != nil { | ||
1275 | logError(testName, function, args, startTime, "", "unexpected failure", err) | ||
1276 | return | ||
1277 | } | ||
1278 | r.Close() | ||
1279 | _, err = c.PutObject(context.Background(), bucketName, objectName, bytes.NewReader(buf), int64(len(buf)), minio.PutObjectOptions{}) | ||
1280 | if err != nil { | ||
1281 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
1282 | return | ||
1283 | } | ||
1284 | } | ||
1285 | |||
1286 | objectsInfo := c.ListObjects(context.Background(), bucketName, minio.ListObjectsOptions{WithVersions: true, Recursive: true}) | ||
1287 | var infos []minio.ObjectInfo | ||
1288 | for info := range objectsInfo { | ||
1289 | if info.Err != nil { | ||
1290 | logError(testName, function, args, startTime, "", "Unexpected error during listing objects", err) | ||
1291 | return | ||
1292 | } | ||
1293 | infos = append(infos, info) | ||
1294 | } | ||
1295 | |||
1296 | sort.Slice(infos, func(i, j int) bool { | ||
1297 | return infos[i].Size < infos[j].Size | ||
1298 | }) | ||
1299 | |||
1300 | reader, err := c.GetObject(context.Background(), bucketName, objectName, minio.GetObjectOptions{VersionID: infos[0].VersionID}) | ||
1301 | if err != nil { | ||
1302 | logError(testName, function, args, startTime, "", "GetObject of the oldest version content failed", err) | ||
1303 | return | ||
1304 | } | ||
1305 | |||
1306 | oldestContent, err := io.ReadAll(reader) | ||
1307 | if err != nil { | ||
1308 | logError(testName, function, args, startTime, "", "Reading the oldest object version failed", err) | ||
1309 | return | ||
1310 | } | ||
1311 | |||
1312 | // Copy Source | ||
1313 | srcOpts := minio.CopySrcOptions{ | ||
1314 | Bucket: bucketName, | ||
1315 | Object: objectName, | ||
1316 | VersionID: infos[0].VersionID, | ||
1317 | } | ||
1318 | args["src"] = srcOpts | ||
1319 | |||
1320 | dstOpts := minio.CopyDestOptions{ | ||
1321 | Bucket: bucketName, | ||
1322 | Object: objectName + "-copy", | ||
1323 | } | ||
1324 | args["dst"] = dstOpts | ||
1325 | |||
1326 | // Perform the Copy | ||
1327 | if _, err = c.CopyObject(context.Background(), dstOpts, srcOpts); err != nil { | ||
1328 | logError(testName, function, args, startTime, "", "CopyObject failed", err) | ||
1329 | return | ||
1330 | } | ||
1331 | |||
1332 | // Destination object | ||
1333 | readerCopy, err := c.GetObject(context.Background(), bucketName, objectName+"-copy", minio.GetObjectOptions{}) | ||
1334 | if err != nil { | ||
1335 | logError(testName, function, args, startTime, "", "GetObject failed", err) | ||
1336 | return | ||
1337 | } | ||
1338 | defer readerCopy.Close() | ||
1339 | |||
1340 | newestContent, err := io.ReadAll(readerCopy) | ||
1341 | if err != nil { | ||
1342 | logError(testName, function, args, startTime, "", "Reading from GetObject reader failed", err) | ||
1343 | return | ||
1344 | } | ||
1345 | |||
1346 | if len(newestContent) == 0 || !bytes.Equal(oldestContent, newestContent) { | ||
1347 | logError(testName, function, args, startTime, "", "Unexpected destination object content", err) | ||
1348 | return | ||
1349 | } | ||
1350 | |||
1351 | // Delete all objects and their versions as long as the bucket itself | ||
1352 | if err = cleanupVersionedBucket(bucketName, c); err != nil { | ||
1353 | logError(testName, function, args, startTime, "", "CleanupBucket failed", err) | ||
1354 | return | ||
1355 | } | ||
1356 | |||
1357 | successLogger(testName, function, args, startTime).Info() | ||
1358 | } | ||
1359 | |||
1360 | func testConcurrentCopyObjectWithVersioning() { | ||
1361 | // initialize logging params | ||
1362 | startTime := time.Now() | ||
1363 | testName := getFuncName() | ||
1364 | function := "CopyObject()" | ||
1365 | args := map[string]interface{}{} | ||
1366 | |||
1367 | // Seed random based on current time. | ||
1368 | rand.Seed(time.Now().Unix()) | ||
1369 | |||
1370 | // Instantiate new minio client object. | ||
1371 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
1372 | &minio.Options{ | ||
1373 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
1374 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
1375 | }) | ||
1376 | if err != nil { | ||
1377 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
1378 | return | ||
1379 | } | ||
1380 | |||
1381 | // Enable tracing, write to stderr. | ||
1382 | // c.TraceOn(os.Stderr) | ||
1383 | |||
1384 | // Set user agent. | ||
1385 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
1386 | |||
1387 | // Generate a new random bucket name. | ||
1388 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
1389 | args["bucketName"] = bucketName | ||
1390 | |||
1391 | // Make a new bucket. | ||
1392 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1", ObjectLocking: true}) | ||
1393 | if err != nil { | ||
1394 | logError(testName, function, args, startTime, "", "Make bucket failed", err) | ||
1395 | return | ||
1396 | } | ||
1397 | |||
1398 | err = c.EnableVersioning(context.Background(), bucketName) | ||
1399 | if err != nil { | ||
1400 | logError(testName, function, args, startTime, "", "Enable versioning failed", err) | ||
1401 | return | ||
1402 | } | ||
1403 | |||
1404 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
1405 | args["objectName"] = objectName | ||
1406 | |||
1407 | testFiles := []string{"datafile-10-kB"} | ||
1408 | for _, testFile := range testFiles { | ||
1409 | r := getDataReader(testFile) | ||
1410 | buf, err := io.ReadAll(r) | ||
1411 | if err != nil { | ||
1412 | logError(testName, function, args, startTime, "", "unexpected failure", err) | ||
1413 | return | ||
1414 | } | ||
1415 | r.Close() | ||
1416 | _, err = c.PutObject(context.Background(), bucketName, objectName, bytes.NewReader(buf), int64(len(buf)), minio.PutObjectOptions{}) | ||
1417 | if err != nil { | ||
1418 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
1419 | return | ||
1420 | } | ||
1421 | } | ||
1422 | |||
1423 | objectsInfo := c.ListObjects(context.Background(), bucketName, minio.ListObjectsOptions{WithVersions: true, Recursive: true}) | ||
1424 | var infos []minio.ObjectInfo | ||
1425 | for info := range objectsInfo { | ||
1426 | if info.Err != nil { | ||
1427 | logError(testName, function, args, startTime, "", "Unexpected error during listing objects", err) | ||
1428 | return | ||
1429 | } | ||
1430 | infos = append(infos, info) | ||
1431 | } | ||
1432 | |||
1433 | sort.Slice(infos, func(i, j int) bool { | ||
1434 | return infos[i].Size < infos[j].Size | ||
1435 | }) | ||
1436 | |||
1437 | reader, err := c.GetObject(context.Background(), bucketName, objectName, minio.GetObjectOptions{VersionID: infos[0].VersionID}) | ||
1438 | if err != nil { | ||
1439 | logError(testName, function, args, startTime, "", "GetObject of the oldest version content failed", err) | ||
1440 | return | ||
1441 | } | ||
1442 | |||
1443 | oldestContent, err := io.ReadAll(reader) | ||
1444 | if err != nil { | ||
1445 | logError(testName, function, args, startTime, "", "Reading the oldest object version failed", err) | ||
1446 | return | ||
1447 | } | ||
1448 | |||
1449 | // Copy Source | ||
1450 | srcOpts := minio.CopySrcOptions{ | ||
1451 | Bucket: bucketName, | ||
1452 | Object: objectName, | ||
1453 | VersionID: infos[0].VersionID, | ||
1454 | } | ||
1455 | args["src"] = srcOpts | ||
1456 | |||
1457 | dstOpts := minio.CopyDestOptions{ | ||
1458 | Bucket: bucketName, | ||
1459 | Object: objectName + "-copy", | ||
1460 | } | ||
1461 | args["dst"] = dstOpts | ||
1462 | |||
1463 | // Perform the Copy concurrently | ||
1464 | const n = 10 | ||
1465 | var wg sync.WaitGroup | ||
1466 | wg.Add(n) | ||
1467 | var errs [n]error | ||
1468 | for i := 0; i < n; i++ { | ||
1469 | go func(i int) { | ||
1470 | defer wg.Done() | ||
1471 | _, errs[i] = c.CopyObject(context.Background(), dstOpts, srcOpts) | ||
1472 | }(i) | ||
1473 | } | ||
1474 | wg.Wait() | ||
1475 | for _, err := range errs { | ||
1476 | if err != nil { | ||
1477 | logError(testName, function, args, startTime, "", "CopyObject failed", err) | ||
1478 | return | ||
1479 | } | ||
1480 | } | ||
1481 | |||
1482 | objectsInfo = c.ListObjects(context.Background(), bucketName, minio.ListObjectsOptions{WithVersions: true, Recursive: false, Prefix: dstOpts.Object}) | ||
1483 | infos = []minio.ObjectInfo{} | ||
1484 | for info := range objectsInfo { | ||
1485 | // Destination object | ||
1486 | readerCopy, err := c.GetObject(context.Background(), bucketName, objectName+"-copy", minio.GetObjectOptions{VersionID: info.VersionID}) | ||
1487 | if err != nil { | ||
1488 | logError(testName, function, args, startTime, "", "GetObject failed", err) | ||
1489 | return | ||
1490 | } | ||
1491 | defer readerCopy.Close() | ||
1492 | |||
1493 | newestContent, err := io.ReadAll(readerCopy) | ||
1494 | if err != nil { | ||
1495 | logError(testName, function, args, startTime, "", "Reading from GetObject reader failed", err) | ||
1496 | return | ||
1497 | } | ||
1498 | |||
1499 | if len(newestContent) == 0 || !bytes.Equal(oldestContent, newestContent) { | ||
1500 | logError(testName, function, args, startTime, "", "Unexpected destination object content", err) | ||
1501 | return | ||
1502 | } | ||
1503 | infos = append(infos, info) | ||
1504 | } | ||
1505 | |||
1506 | if len(infos) != n { | ||
1507 | logError(testName, function, args, startTime, "", "Unexpected number of Version elements in listing objects", nil) | ||
1508 | return | ||
1509 | } | ||
1510 | |||
1511 | // Delete all objects and their versions as long as the bucket itself | ||
1512 | if err = cleanupVersionedBucket(bucketName, c); err != nil { | ||
1513 | logError(testName, function, args, startTime, "", "CleanupBucket failed", err) | ||
1514 | return | ||
1515 | } | ||
1516 | |||
1517 | successLogger(testName, function, args, startTime).Info() | ||
1518 | } | ||
1519 | |||
1520 | func testComposeObjectWithVersioning() { | ||
1521 | // initialize logging params | ||
1522 | startTime := time.Now() | ||
1523 | testName := getFuncName() | ||
1524 | function := "ComposeObject()" | ||
1525 | args := map[string]interface{}{} | ||
1526 | |||
1527 | // Seed random based on current time. | ||
1528 | rand.Seed(time.Now().Unix()) | ||
1529 | |||
1530 | // Instantiate new minio client object. | ||
1531 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
1532 | &minio.Options{ | ||
1533 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
1534 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
1535 | }) | ||
1536 | if err != nil { | ||
1537 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
1538 | return | ||
1539 | } | ||
1540 | |||
1541 | // Enable tracing, write to stderr. | ||
1542 | // c.TraceOn(os.Stderr) | ||
1543 | |||
1544 | // Set user agent. | ||
1545 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
1546 | |||
1547 | // Generate a new random bucket name. | ||
1548 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
1549 | args["bucketName"] = bucketName | ||
1550 | |||
1551 | // Make a new bucket. | ||
1552 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1", ObjectLocking: true}) | ||
1553 | if err != nil { | ||
1554 | logError(testName, function, args, startTime, "", "Make bucket failed", err) | ||
1555 | return | ||
1556 | } | ||
1557 | |||
1558 | err = c.EnableVersioning(context.Background(), bucketName) | ||
1559 | if err != nil { | ||
1560 | logError(testName, function, args, startTime, "", "Enable versioning failed", err) | ||
1561 | return | ||
1562 | } | ||
1563 | |||
1564 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
1565 | args["objectName"] = objectName | ||
1566 | |||
1567 | // var testFiles = []string{"datafile-5-MB", "datafile-10-kB"} | ||
1568 | testFiles := []string{"datafile-5-MB", "datafile-10-kB"} | ||
1569 | var testFilesBytes [][]byte | ||
1570 | |||
1571 | for _, testFile := range testFiles { | ||
1572 | r := getDataReader(testFile) | ||
1573 | buf, err := io.ReadAll(r) | ||
1574 | if err != nil { | ||
1575 | logError(testName, function, args, startTime, "", "unexpected failure", err) | ||
1576 | return | ||
1577 | } | ||
1578 | r.Close() | ||
1579 | _, err = c.PutObject(context.Background(), bucketName, objectName, bytes.NewReader(buf), int64(len(buf)), minio.PutObjectOptions{}) | ||
1580 | if err != nil { | ||
1581 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
1582 | return | ||
1583 | } | ||
1584 | testFilesBytes = append(testFilesBytes, buf) | ||
1585 | } | ||
1586 | |||
1587 | objectsInfo := c.ListObjects(context.Background(), bucketName, minio.ListObjectsOptions{WithVersions: true, Recursive: true}) | ||
1588 | |||
1589 | var results []minio.ObjectInfo | ||
1590 | for info := range objectsInfo { | ||
1591 | if info.Err != nil { | ||
1592 | logError(testName, function, args, startTime, "", "Unexpected error during listing objects", err) | ||
1593 | return | ||
1594 | } | ||
1595 | results = append(results, info) | ||
1596 | } | ||
1597 | |||
1598 | sort.SliceStable(results, func(i, j int) bool { | ||
1599 | return results[i].Size > results[j].Size | ||
1600 | }) | ||
1601 | |||
1602 | // Source objects to concatenate. We also specify decryption | ||
1603 | // key for each | ||
1604 | src1 := minio.CopySrcOptions{ | ||
1605 | Bucket: bucketName, | ||
1606 | Object: objectName, | ||
1607 | VersionID: results[0].VersionID, | ||
1608 | } | ||
1609 | |||
1610 | src2 := minio.CopySrcOptions{ | ||
1611 | Bucket: bucketName, | ||
1612 | Object: objectName, | ||
1613 | VersionID: results[1].VersionID, | ||
1614 | } | ||
1615 | |||
1616 | dst := minio.CopyDestOptions{ | ||
1617 | Bucket: bucketName, | ||
1618 | Object: objectName + "-copy", | ||
1619 | } | ||
1620 | |||
1621 | _, err = c.ComposeObject(context.Background(), dst, src1, src2) | ||
1622 | if err != nil { | ||
1623 | logError(testName, function, args, startTime, "", "ComposeObject failed", err) | ||
1624 | return | ||
1625 | } | ||
1626 | |||
1627 | // Destination object | ||
1628 | readerCopy, err := c.GetObject(context.Background(), bucketName, objectName+"-copy", minio.GetObjectOptions{}) | ||
1629 | if err != nil { | ||
1630 | logError(testName, function, args, startTime, "", "GetObject of the copy object failed", err) | ||
1631 | return | ||
1632 | } | ||
1633 | defer readerCopy.Close() | ||
1634 | |||
1635 | copyContentBytes, err := io.ReadAll(readerCopy) | ||
1636 | if err != nil { | ||
1637 | logError(testName, function, args, startTime, "", "Reading from the copy object reader failed", err) | ||
1638 | return | ||
1639 | } | ||
1640 | |||
1641 | var expectedContent []byte | ||
1642 | for _, fileBytes := range testFilesBytes { | ||
1643 | expectedContent = append(expectedContent, fileBytes...) | ||
1644 | } | ||
1645 | |||
1646 | if len(copyContentBytes) == 0 || !bytes.Equal(copyContentBytes, expectedContent) { | ||
1647 | logError(testName, function, args, startTime, "", "Unexpected destination object content", err) | ||
1648 | return | ||
1649 | } | ||
1650 | |||
1651 | // Delete all objects and their versions as long as the bucket itself | ||
1652 | if err = cleanupVersionedBucket(bucketName, c); err != nil { | ||
1653 | logError(testName, function, args, startTime, "", "CleanupBucket failed", err) | ||
1654 | return | ||
1655 | } | ||
1656 | |||
1657 | successLogger(testName, function, args, startTime).Info() | ||
1658 | } | ||
1659 | |||
1660 | func testRemoveObjectWithVersioning() { | ||
1661 | // initialize logging params | ||
1662 | startTime := time.Now() | ||
1663 | testName := getFuncName() | ||
1664 | function := "DeleteObject()" | ||
1665 | args := map[string]interface{}{} | ||
1666 | |||
1667 | // Seed random based on current time. | ||
1668 | rand.Seed(time.Now().Unix()) | ||
1669 | |||
1670 | // Instantiate new minio client object. | ||
1671 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
1672 | &minio.Options{ | ||
1673 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
1674 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
1675 | }) | ||
1676 | if err != nil { | ||
1677 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
1678 | return | ||
1679 | } | ||
1680 | |||
1681 | // Enable tracing, write to stderr. | ||
1682 | // c.TraceOn(os.Stderr) | ||
1683 | |||
1684 | // Set user agent. | ||
1685 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
1686 | |||
1687 | // Generate a new random bucket name. | ||
1688 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
1689 | args["bucketName"] = bucketName | ||
1690 | |||
1691 | // Make a new bucket. | ||
1692 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1", ObjectLocking: true}) | ||
1693 | if err != nil { | ||
1694 | logError(testName, function, args, startTime, "", "Make bucket failed", err) | ||
1695 | return | ||
1696 | } | ||
1697 | |||
1698 | err = c.EnableVersioning(context.Background(), bucketName) | ||
1699 | if err != nil { | ||
1700 | logError(testName, function, args, startTime, "", "Enable versioning failed", err) | ||
1701 | return | ||
1702 | } | ||
1703 | |||
1704 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
1705 | args["objectName"] = objectName | ||
1706 | |||
1707 | _, err = c.PutObject(context.Background(), bucketName, objectName, getDataReader("datafile-10-kB"), int64(dataFileMap["datafile-10-kB"]), minio.PutObjectOptions{}) | ||
1708 | if err != nil { | ||
1709 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
1710 | return | ||
1711 | } | ||
1712 | |||
1713 | objectsInfo := c.ListObjects(context.Background(), bucketName, minio.ListObjectsOptions{WithVersions: true, Recursive: true}) | ||
1714 | var version minio.ObjectInfo | ||
1715 | for info := range objectsInfo { | ||
1716 | if info.Err != nil { | ||
1717 | logError(testName, function, args, startTime, "", "Unexpected error during listing objects", err) | ||
1718 | return | ||
1719 | } | ||
1720 | version = info | ||
1721 | break | ||
1722 | } | ||
1723 | |||
1724 | err = c.RemoveObject(context.Background(), bucketName, objectName, minio.RemoveObjectOptions{VersionID: version.VersionID}) | ||
1725 | if err != nil { | ||
1726 | logError(testName, function, args, startTime, "", "DeleteObject failed", err) | ||
1727 | return | ||
1728 | } | ||
1729 | |||
1730 | objectsInfo = c.ListObjects(context.Background(), bucketName, minio.ListObjectsOptions{WithVersions: true, Recursive: true}) | ||
1731 | for range objectsInfo { | ||
1732 | logError(testName, function, args, startTime, "", "Unexpected versioning info, should not have any one ", err) | ||
1733 | return | ||
1734 | } | ||
1735 | // test delete marker version id is non-null | ||
1736 | _, err = c.PutObject(context.Background(), bucketName, objectName, getDataReader("datafile-10-kB"), int64(dataFileMap["datafile-10-kB"]), minio.PutObjectOptions{}) | ||
1737 | if err != nil { | ||
1738 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
1739 | return | ||
1740 | } | ||
1741 | // create delete marker | ||
1742 | err = c.RemoveObject(context.Background(), bucketName, objectName, minio.RemoveObjectOptions{}) | ||
1743 | if err != nil { | ||
1744 | logError(testName, function, args, startTime, "", "DeleteObject failed", err) | ||
1745 | return | ||
1746 | } | ||
1747 | objectsInfo = c.ListObjects(context.Background(), bucketName, minio.ListObjectsOptions{WithVersions: true, Recursive: true}) | ||
1748 | idx := 0 | ||
1749 | for info := range objectsInfo { | ||
1750 | if info.Err != nil { | ||
1751 | logError(testName, function, args, startTime, "", "Unexpected error during listing objects", err) | ||
1752 | return | ||
1753 | } | ||
1754 | if idx == 0 { | ||
1755 | if !info.IsDeleteMarker { | ||
1756 | logError(testName, function, args, startTime, "", "Unexpected error - expected delete marker to have been created", err) | ||
1757 | return | ||
1758 | } | ||
1759 | if info.VersionID == "" { | ||
1760 | logError(testName, function, args, startTime, "", "Unexpected error - expected delete marker to be versioned", err) | ||
1761 | return | ||
1762 | } | ||
1763 | } | ||
1764 | idx++ | ||
1765 | } | ||
1766 | |||
1767 | defer cleanupBucket(bucketName, c) | ||
1768 | |||
1769 | successLogger(testName, function, args, startTime).Info() | ||
1770 | } | ||
1771 | |||
1772 | func testRemoveObjectsWithVersioning() { | ||
1773 | // initialize logging params | ||
1774 | startTime := time.Now() | ||
1775 | testName := getFuncName() | ||
1776 | function := "DeleteObjects()" | ||
1777 | args := map[string]interface{}{} | ||
1778 | |||
1779 | // Seed random based on current time. | ||
1780 | rand.Seed(time.Now().Unix()) | ||
1781 | |||
1782 | // Instantiate new minio client object. | ||
1783 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
1784 | &minio.Options{ | ||
1785 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
1786 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
1787 | }) | ||
1788 | if err != nil { | ||
1789 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
1790 | return | ||
1791 | } | ||
1792 | |||
1793 | // Enable tracing, write to stderr. | ||
1794 | // c.TraceOn(os.Stderr) | ||
1795 | |||
1796 | // Set user agent. | ||
1797 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
1798 | |||
1799 | // Generate a new random bucket name. | ||
1800 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
1801 | args["bucketName"] = bucketName | ||
1802 | |||
1803 | // Make a new bucket. | ||
1804 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1", ObjectLocking: true}) | ||
1805 | if err != nil { | ||
1806 | logError(testName, function, args, startTime, "", "Make bucket failed", err) | ||
1807 | return | ||
1808 | } | ||
1809 | |||
1810 | err = c.EnableVersioning(context.Background(), bucketName) | ||
1811 | if err != nil { | ||
1812 | logError(testName, function, args, startTime, "", "Enable versioning failed", err) | ||
1813 | return | ||
1814 | } | ||
1815 | |||
1816 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
1817 | args["objectName"] = objectName | ||
1818 | |||
1819 | _, err = c.PutObject(context.Background(), bucketName, objectName, getDataReader("datafile-10-kB"), int64(dataFileMap["datafile-10-kB"]), minio.PutObjectOptions{}) | ||
1820 | if err != nil { | ||
1821 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
1822 | return | ||
1823 | } | ||
1824 | |||
1825 | objectsVersions := make(chan minio.ObjectInfo) | ||
1826 | go func() { | ||
1827 | objectsVersionsInfo := c.ListObjects(context.Background(), bucketName, | ||
1828 | minio.ListObjectsOptions{WithVersions: true, Recursive: true}) | ||
1829 | for info := range objectsVersionsInfo { | ||
1830 | if info.Err != nil { | ||
1831 | logError(testName, function, args, startTime, "", "Unexpected error during listing objects", err) | ||
1832 | return | ||
1833 | } | ||
1834 | objectsVersions <- info | ||
1835 | } | ||
1836 | close(objectsVersions) | ||
1837 | }() | ||
1838 | |||
1839 | removeErrors := c.RemoveObjects(context.Background(), bucketName, objectsVersions, minio.RemoveObjectsOptions{}) | ||
1840 | if err != nil { | ||
1841 | logError(testName, function, args, startTime, "", "DeleteObjects call failed", err) | ||
1842 | return | ||
1843 | } | ||
1844 | |||
1845 | for e := range removeErrors { | ||
1846 | if e.Err != nil { | ||
1847 | logError(testName, function, args, startTime, "", "Single delete operation failed", err) | ||
1848 | return | ||
1849 | } | ||
1850 | } | ||
1851 | |||
1852 | objectsVersionsInfo := c.ListObjects(context.Background(), bucketName, minio.ListObjectsOptions{WithVersions: true, Recursive: true}) | ||
1853 | for range objectsVersionsInfo { | ||
1854 | logError(testName, function, args, startTime, "", "Unexpected versioning info, should not have any one ", err) | ||
1855 | return | ||
1856 | } | ||
1857 | |||
1858 | err = c.RemoveBucket(context.Background(), bucketName) | ||
1859 | if err != nil { | ||
1860 | logError(testName, function, args, startTime, "", "CleanupBucket failed", err) | ||
1861 | return | ||
1862 | } | ||
1863 | |||
1864 | successLogger(testName, function, args, startTime).Info() | ||
1865 | } | ||
1866 | |||
1867 | func testObjectTaggingWithVersioning() { | ||
1868 | // initialize logging params | ||
1869 | startTime := time.Now() | ||
1870 | testName := getFuncName() | ||
1871 | function := "{Get,Set,Remove}ObjectTagging()" | ||
1872 | args := map[string]interface{}{} | ||
1873 | |||
1874 | // Seed random based on current time. | ||
1875 | rand.Seed(time.Now().Unix()) | ||
1876 | |||
1877 | // Instantiate new minio client object. | ||
1878 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
1879 | &minio.Options{ | ||
1880 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
1881 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
1882 | }) | ||
1883 | if err != nil { | ||
1884 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
1885 | return | ||
1886 | } | ||
1887 | |||
1888 | // Enable tracing, write to stderr. | ||
1889 | // c.TraceOn(os.Stderr) | ||
1890 | |||
1891 | // Set user agent. | ||
1892 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
1893 | |||
1894 | // Generate a new random bucket name. | ||
1895 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
1896 | args["bucketName"] = bucketName | ||
1897 | |||
1898 | // Make a new bucket. | ||
1899 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1", ObjectLocking: true}) | ||
1900 | if err != nil { | ||
1901 | logError(testName, function, args, startTime, "", "Make bucket failed", err) | ||
1902 | return | ||
1903 | } | ||
1904 | |||
1905 | err = c.EnableVersioning(context.Background(), bucketName) | ||
1906 | if err != nil { | ||
1907 | logError(testName, function, args, startTime, "", "Enable versioning failed", err) | ||
1908 | return | ||
1909 | } | ||
1910 | |||
1911 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
1912 | args["objectName"] = objectName | ||
1913 | |||
1914 | for _, file := range []string{"datafile-1-b", "datafile-10-kB"} { | ||
1915 | _, err = c.PutObject(context.Background(), bucketName, objectName, getDataReader(file), int64(dataFileMap[file]), minio.PutObjectOptions{}) | ||
1916 | if err != nil { | ||
1917 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
1918 | return | ||
1919 | } | ||
1920 | } | ||
1921 | |||
1922 | versionsInfo := c.ListObjects(context.Background(), bucketName, minio.ListObjectsOptions{WithVersions: true, Recursive: true}) | ||
1923 | |||
1924 | var versions []minio.ObjectInfo | ||
1925 | for info := range versionsInfo { | ||
1926 | if info.Err != nil { | ||
1927 | logError(testName, function, args, startTime, "", "Unexpected error during listing objects", err) | ||
1928 | return | ||
1929 | } | ||
1930 | versions = append(versions, info) | ||
1931 | } | ||
1932 | |||
1933 | sort.SliceStable(versions, func(i, j int) bool { | ||
1934 | return versions[i].Size < versions[j].Size | ||
1935 | }) | ||
1936 | |||
1937 | tagsV1 := map[string]string{"key1": "val1"} | ||
1938 | t1, err := tags.MapToObjectTags(tagsV1) | ||
1939 | if err != nil { | ||
1940 | logError(testName, function, args, startTime, "", "PutObjectTagging (1) failed", err) | ||
1941 | return | ||
1942 | } | ||
1943 | |||
1944 | err = c.PutObjectTagging(context.Background(), bucketName, objectName, t1, minio.PutObjectTaggingOptions{VersionID: versions[0].VersionID}) | ||
1945 | if err != nil { | ||
1946 | logError(testName, function, args, startTime, "", "PutObjectTagging (1) failed", err) | ||
1947 | return | ||
1948 | } | ||
1949 | |||
1950 | tagsV2 := map[string]string{"key2": "val2"} | ||
1951 | t2, err := tags.MapToObjectTags(tagsV2) | ||
1952 | if err != nil { | ||
1953 | logError(testName, function, args, startTime, "", "PutObjectTagging (1) failed", err) | ||
1954 | return | ||
1955 | } | ||
1956 | |||
1957 | err = c.PutObjectTagging(context.Background(), bucketName, objectName, t2, minio.PutObjectTaggingOptions{VersionID: versions[1].VersionID}) | ||
1958 | if err != nil { | ||
1959 | logError(testName, function, args, startTime, "", "PutObjectTagging (2) failed", err) | ||
1960 | return | ||
1961 | } | ||
1962 | |||
1963 | tagsEqual := func(tags1, tags2 map[string]string) bool { | ||
1964 | for k1, v1 := range tags1 { | ||
1965 | v2, found := tags2[k1] | ||
1966 | if found { | ||
1967 | if v1 != v2 { | ||
1968 | return false | ||
1969 | } | ||
1970 | } | ||
1971 | } | ||
1972 | return true | ||
1973 | } | ||
1974 | |||
1975 | gotTagsV1, err := c.GetObjectTagging(context.Background(), bucketName, objectName, minio.GetObjectTaggingOptions{VersionID: versions[0].VersionID}) | ||
1976 | if err != nil { | ||
1977 | logError(testName, function, args, startTime, "", "GetObjectTagging failed", err) | ||
1978 | return | ||
1979 | } | ||
1980 | |||
1981 | if !tagsEqual(t1.ToMap(), gotTagsV1.ToMap()) { | ||
1982 | logError(testName, function, args, startTime, "", "Unexpected tags content (1)", err) | ||
1983 | return | ||
1984 | } | ||
1985 | |||
1986 | gotTagsV2, err := c.GetObjectTagging(context.Background(), bucketName, objectName, minio.GetObjectTaggingOptions{}) | ||
1987 | if err != nil { | ||
1988 | logError(testName, function, args, startTime, "", "GetObjectTaggingContext failed", err) | ||
1989 | return | ||
1990 | } | ||
1991 | |||
1992 | if !tagsEqual(t2.ToMap(), gotTagsV2.ToMap()) { | ||
1993 | logError(testName, function, args, startTime, "", "Unexpected tags content (2)", err) | ||
1994 | return | ||
1995 | } | ||
1996 | |||
1997 | err = c.RemoveObjectTagging(context.Background(), bucketName, objectName, minio.RemoveObjectTaggingOptions{VersionID: versions[0].VersionID}) | ||
1998 | if err != nil { | ||
1999 | logError(testName, function, args, startTime, "", "PutObjectTagging (2) failed", err) | ||
2000 | return | ||
2001 | } | ||
2002 | |||
2003 | emptyTags, err := c.GetObjectTagging(context.Background(), bucketName, objectName, | ||
2004 | minio.GetObjectTaggingOptions{VersionID: versions[0].VersionID}) | ||
2005 | if err != nil { | ||
2006 | logError(testName, function, args, startTime, "", "GetObjectTagging failed", err) | ||
2007 | return | ||
2008 | } | ||
2009 | |||
2010 | if len(emptyTags.ToMap()) != 0 { | ||
2011 | logError(testName, function, args, startTime, "", "Unexpected tags content (2)", err) | ||
2012 | return | ||
2013 | } | ||
2014 | |||
2015 | // Delete all objects and their versions as long as the bucket itself | ||
2016 | if err = cleanupVersionedBucket(bucketName, c); err != nil { | ||
2017 | logError(testName, function, args, startTime, "", "CleanupBucket failed", err) | ||
2018 | return | ||
2019 | } | ||
2020 | |||
2021 | successLogger(testName, function, args, startTime).Info() | ||
2022 | } | ||
2023 | |||
2024 | // Test PutObject with custom checksums. | ||
2025 | func testPutObjectWithChecksums() { | ||
2026 | // initialize logging params | ||
2027 | startTime := time.Now() | ||
2028 | testName := getFuncName() | ||
2029 | function := "PutObject(bucketName, objectName, reader,size, opts)" | ||
2030 | args := map[string]interface{}{ | ||
2031 | "bucketName": "", | ||
2032 | "objectName": "", | ||
2033 | "opts": "minio.PutObjectOptions{UserMetadata: metadata, Progress: progress}", | ||
2034 | } | ||
2035 | |||
2036 | if !isFullMode() { | ||
2037 | ignoredLog(testName, function, args, startTime, "Skipping functional tests for short/quick runs").Info() | ||
2038 | return | ||
2039 | } | ||
2040 | |||
2041 | // Seed random based on current time. | ||
2042 | rand.Seed(time.Now().Unix()) | ||
2043 | |||
2044 | // Instantiate new minio client object. | ||
2045 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
2046 | &minio.Options{ | ||
2047 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
2048 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
2049 | }) | ||
2050 | if err != nil { | ||
2051 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
2052 | return | ||
2053 | } | ||
2054 | |||
2055 | // Enable tracing, write to stderr. | ||
2056 | // c.TraceOn(os.Stderr) | ||
2057 | |||
2058 | // Set user agent. | ||
2059 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
2060 | |||
2061 | // Generate a new random bucket name. | ||
2062 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
2063 | args["bucketName"] = bucketName | ||
2064 | |||
2065 | // Make a new bucket. | ||
2066 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
2067 | if err != nil { | ||
2068 | logError(testName, function, args, startTime, "", "Make bucket failed", err) | ||
2069 | return | ||
2070 | } | ||
2071 | |||
2072 | defer cleanupBucket(bucketName, c) | ||
2073 | tests := []struct { | ||
2074 | header string | ||
2075 | hasher hash.Hash | ||
2076 | |||
2077 | // Checksum values | ||
2078 | ChecksumCRC32 string | ||
2079 | ChecksumCRC32C string | ||
2080 | ChecksumSHA1 string | ||
2081 | ChecksumSHA256 string | ||
2082 | }{ | ||
2083 | {header: "x-amz-checksum-crc32", hasher: crc32.NewIEEE()}, | ||
2084 | {header: "x-amz-checksum-crc32c", hasher: crc32.New(crc32.MakeTable(crc32.Castagnoli))}, | ||
2085 | {header: "x-amz-checksum-sha1", hasher: sha1.New()}, | ||
2086 | {header: "x-amz-checksum-sha256", hasher: sha256.New()}, | ||
2087 | } | ||
2088 | |||
2089 | for i, test := range tests { | ||
2090 | bufSize := dataFileMap["datafile-10-kB"] | ||
2091 | |||
2092 | // Save the data | ||
2093 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
2094 | args["objectName"] = objectName | ||
2095 | |||
2096 | cmpChecksum := func(got, want string) { | ||
2097 | if want != got { | ||
2098 | logError(testName, function, args, startTime, "", "checksum mismatch", fmt.Errorf("want %s, got %s", want, got)) | ||
2099 | return | ||
2100 | } | ||
2101 | } | ||
2102 | |||
2103 | meta := map[string]string{} | ||
2104 | reader := getDataReader("datafile-10-kB") | ||
2105 | b, err := io.ReadAll(reader) | ||
2106 | if err != nil { | ||
2107 | logError(testName, function, args, startTime, "", "Read failed", err) | ||
2108 | return | ||
2109 | } | ||
2110 | h := test.hasher | ||
2111 | h.Reset() | ||
2112 | // Wrong CRC. | ||
2113 | meta[test.header] = base64.StdEncoding.EncodeToString(h.Sum(nil)) | ||
2114 | args["metadata"] = meta | ||
2115 | args["range"] = "false" | ||
2116 | |||
2117 | resp, err := c.PutObject(context.Background(), bucketName, objectName, bytes.NewReader(b), int64(bufSize), minio.PutObjectOptions{ | ||
2118 | DisableMultipart: true, | ||
2119 | UserMetadata: meta, | ||
2120 | }) | ||
2121 | if err == nil { | ||
2122 | if i == 0 && resp.ChecksumCRC32 == "" { | ||
2123 | ignoredLog(testName, function, args, startTime, "Checksums does not appear to be supported by backend").Info() | ||
2124 | return | ||
2125 | } | ||
2126 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
2127 | return | ||
2128 | } | ||
2129 | |||
2130 | // Set correct CRC. | ||
2131 | h.Write(b) | ||
2132 | meta[test.header] = base64.StdEncoding.EncodeToString(h.Sum(nil)) | ||
2133 | reader.Close() | ||
2134 | |||
2135 | resp, err = c.PutObject(context.Background(), bucketName, objectName, bytes.NewReader(b), int64(bufSize), minio.PutObjectOptions{ | ||
2136 | DisableMultipart: true, | ||
2137 | DisableContentSha256: true, | ||
2138 | UserMetadata: meta, | ||
2139 | }) | ||
2140 | if err != nil { | ||
2141 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
2142 | return | ||
2143 | } | ||
2144 | cmpChecksum(resp.ChecksumSHA256, meta["x-amz-checksum-sha256"]) | ||
2145 | cmpChecksum(resp.ChecksumSHA1, meta["x-amz-checksum-sha1"]) | ||
2146 | cmpChecksum(resp.ChecksumCRC32, meta["x-amz-checksum-crc32"]) | ||
2147 | cmpChecksum(resp.ChecksumCRC32C, meta["x-amz-checksum-crc32c"]) | ||
2148 | |||
2149 | // Read the data back | ||
2150 | gopts := minio.GetObjectOptions{Checksum: true} | ||
2151 | |||
2152 | r, err := c.GetObject(context.Background(), bucketName, objectName, gopts) | ||
2153 | if err != nil { | ||
2154 | logError(testName, function, args, startTime, "", "GetObject failed", err) | ||
2155 | return | ||
2156 | } | ||
2157 | |||
2158 | st, err := r.Stat() | ||
2159 | if err != nil { | ||
2160 | logError(testName, function, args, startTime, "", "Stat failed", err) | ||
2161 | return | ||
2162 | } | ||
2163 | cmpChecksum(st.ChecksumSHA256, meta["x-amz-checksum-sha256"]) | ||
2164 | cmpChecksum(st.ChecksumSHA1, meta["x-amz-checksum-sha1"]) | ||
2165 | cmpChecksum(st.ChecksumCRC32, meta["x-amz-checksum-crc32"]) | ||
2166 | cmpChecksum(st.ChecksumCRC32C, meta["x-amz-checksum-crc32c"]) | ||
2167 | |||
2168 | if st.Size != int64(bufSize) { | ||
2169 | logError(testName, function, args, startTime, "", "Number of bytes returned by PutObject does not match GetObject, expected "+string(bufSize)+" got "+string(st.Size), err) | ||
2170 | return | ||
2171 | } | ||
2172 | |||
2173 | if err := r.Close(); err != nil { | ||
2174 | logError(testName, function, args, startTime, "", "Object Close failed", err) | ||
2175 | return | ||
2176 | } | ||
2177 | if err := r.Close(); err == nil { | ||
2178 | logError(testName, function, args, startTime, "", "Object already closed, should respond with error", err) | ||
2179 | return | ||
2180 | } | ||
2181 | |||
2182 | args["range"] = "true" | ||
2183 | err = gopts.SetRange(100, 1000) | ||
2184 | if err != nil { | ||
2185 | logError(testName, function, args, startTime, "", "SetRange failed", err) | ||
2186 | return | ||
2187 | } | ||
2188 | r, err = c.GetObject(context.Background(), bucketName, objectName, gopts) | ||
2189 | if err != nil { | ||
2190 | logError(testName, function, args, startTime, "", "GetObject failed", err) | ||
2191 | return | ||
2192 | } | ||
2193 | |||
2194 | b, err = io.ReadAll(r) | ||
2195 | if err != nil { | ||
2196 | logError(testName, function, args, startTime, "", "Read failed", err) | ||
2197 | return | ||
2198 | } | ||
2199 | st, err = r.Stat() | ||
2200 | if err != nil { | ||
2201 | logError(testName, function, args, startTime, "", "Stat failed", err) | ||
2202 | return | ||
2203 | } | ||
2204 | |||
2205 | // Range requests should return empty checksums... | ||
2206 | cmpChecksum(st.ChecksumSHA256, "") | ||
2207 | cmpChecksum(st.ChecksumSHA1, "") | ||
2208 | cmpChecksum(st.ChecksumCRC32, "") | ||
2209 | cmpChecksum(st.ChecksumCRC32C, "") | ||
2210 | |||
2211 | delete(args, "range") | ||
2212 | delete(args, "metadata") | ||
2213 | } | ||
2214 | |||
2215 | successLogger(testName, function, args, startTime).Info() | ||
2216 | } | ||
2217 | |||
2218 | // Test PutObject with custom checksums. | ||
2219 | func testPutMultipartObjectWithChecksums() { | ||
2220 | // initialize logging params | ||
2221 | startTime := time.Now() | ||
2222 | testName := getFuncName() | ||
2223 | function := "PutObject(bucketName, objectName, reader,size, opts)" | ||
2224 | args := map[string]interface{}{ | ||
2225 | "bucketName": "", | ||
2226 | "objectName": "", | ||
2227 | "opts": "minio.PutObjectOptions{UserMetadata: metadata, Progress: progress}", | ||
2228 | } | ||
2229 | |||
2230 | if !isFullMode() { | ||
2231 | ignoredLog(testName, function, args, startTime, "Skipping functional tests for short/quick runs").Info() | ||
2232 | return | ||
2233 | } | ||
2234 | |||
2235 | // Seed random based on current time. | ||
2236 | rand.Seed(time.Now().Unix()) | ||
2237 | |||
2238 | // Instantiate new minio client object. | ||
2239 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
2240 | &minio.Options{ | ||
2241 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
2242 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
2243 | }) | ||
2244 | if err != nil { | ||
2245 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
2246 | return | ||
2247 | } | ||
2248 | |||
2249 | // Enable tracing, write to stderr. | ||
2250 | // c.TraceOn(os.Stderr) | ||
2251 | |||
2252 | // Set user agent. | ||
2253 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
2254 | |||
2255 | // Generate a new random bucket name. | ||
2256 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
2257 | args["bucketName"] = bucketName | ||
2258 | |||
2259 | // Make a new bucket. | ||
2260 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
2261 | if err != nil { | ||
2262 | logError(testName, function, args, startTime, "", "Make bucket failed", err) | ||
2263 | return | ||
2264 | } | ||
2265 | |||
2266 | hashMultiPart := func(b []byte, partSize int, hasher hash.Hash) string { | ||
2267 | r := bytes.NewReader(b) | ||
2268 | tmp := make([]byte, partSize) | ||
2269 | parts := 0 | ||
2270 | var all []byte | ||
2271 | for { | ||
2272 | n, err := io.ReadFull(r, tmp) | ||
2273 | if err != nil && err != io.ErrUnexpectedEOF { | ||
2274 | logError(testName, function, args, startTime, "", "Calc crc failed", err) | ||
2275 | } | ||
2276 | if n == 0 { | ||
2277 | break | ||
2278 | } | ||
2279 | parts++ | ||
2280 | hasher.Reset() | ||
2281 | hasher.Write(tmp[:n]) | ||
2282 | all = append(all, hasher.Sum(nil)...) | ||
2283 | if err != nil { | ||
2284 | break | ||
2285 | } | ||
2286 | } | ||
2287 | hasher.Reset() | ||
2288 | hasher.Write(all) | ||
2289 | return fmt.Sprintf("%s-%d", base64.StdEncoding.EncodeToString(hasher.Sum(nil)), parts) | ||
2290 | } | ||
2291 | defer cleanupBucket(bucketName, c) | ||
2292 | tests := []struct { | ||
2293 | header string | ||
2294 | hasher hash.Hash | ||
2295 | |||
2296 | // Checksum values | ||
2297 | ChecksumCRC32 string | ||
2298 | ChecksumCRC32C string | ||
2299 | ChecksumSHA1 string | ||
2300 | ChecksumSHA256 string | ||
2301 | }{ | ||
2302 | // Currently there is no way to override the checksum type. | ||
2303 | {header: "x-amz-checksum-crc32c", hasher: crc32.New(crc32.MakeTable(crc32.Castagnoli)), ChecksumCRC32C: "OpEx0Q==-13"}, | ||
2304 | } | ||
2305 | |||
2306 | for _, test := range tests { | ||
2307 | bufSize := dataFileMap["datafile-129-MB"] | ||
2308 | |||
2309 | // Save the data | ||
2310 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
2311 | args["objectName"] = objectName | ||
2312 | |||
2313 | cmpChecksum := func(got, want string) { | ||
2314 | if want != got { | ||
2315 | // logError(testName, function, args, startTime, "", "checksum mismatch", fmt.Errorf("want %s, got %s", want, got)) | ||
2316 | fmt.Printf("want %s, got %s\n", want, got) | ||
2317 | return | ||
2318 | } | ||
2319 | } | ||
2320 | |||
2321 | const partSize = 10 << 20 | ||
2322 | reader := getDataReader("datafile-129-MB") | ||
2323 | b, err := io.ReadAll(reader) | ||
2324 | if err != nil { | ||
2325 | logError(testName, function, args, startTime, "", "Read failed", err) | ||
2326 | return | ||
2327 | } | ||
2328 | reader.Close() | ||
2329 | h := test.hasher | ||
2330 | h.Reset() | ||
2331 | test.ChecksumCRC32C = hashMultiPart(b, partSize, test.hasher) | ||
2332 | |||
2333 | // Set correct CRC. | ||
2334 | |||
2335 | resp, err := c.PutObject(context.Background(), bucketName, objectName, io.NopCloser(bytes.NewReader(b)), int64(bufSize), minio.PutObjectOptions{ | ||
2336 | DisableContentSha256: true, | ||
2337 | DisableMultipart: false, | ||
2338 | UserMetadata: nil, | ||
2339 | PartSize: partSize, | ||
2340 | }) | ||
2341 | if err != nil { | ||
2342 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
2343 | return | ||
2344 | } | ||
2345 | cmpChecksum(resp.ChecksumSHA256, test.ChecksumSHA256) | ||
2346 | cmpChecksum(resp.ChecksumSHA1, test.ChecksumSHA1) | ||
2347 | cmpChecksum(resp.ChecksumCRC32, test.ChecksumCRC32) | ||
2348 | cmpChecksum(resp.ChecksumCRC32C, test.ChecksumCRC32C) | ||
2349 | |||
2350 | // Read the data back | ||
2351 | gopts := minio.GetObjectOptions{Checksum: true} | ||
2352 | gopts.PartNumber = 2 | ||
2353 | |||
2354 | // We cannot use StatObject, since it ignores partnumber. | ||
2355 | r, err := c.GetObject(context.Background(), bucketName, objectName, gopts) | ||
2356 | if err != nil { | ||
2357 | logError(testName, function, args, startTime, "", "GetObject failed", err) | ||
2358 | return | ||
2359 | } | ||
2360 | io.Copy(io.Discard, r) | ||
2361 | st, err := r.Stat() | ||
2362 | if err != nil { | ||
2363 | logError(testName, function, args, startTime, "", "Stat failed", err) | ||
2364 | return | ||
2365 | } | ||
2366 | |||
2367 | // Test part 2 checksum... | ||
2368 | h.Reset() | ||
2369 | h.Write(b[partSize : 2*partSize]) | ||
2370 | got := base64.StdEncoding.EncodeToString(h.Sum(nil)) | ||
2371 | if test.ChecksumSHA256 != "" { | ||
2372 | cmpChecksum(st.ChecksumSHA256, got) | ||
2373 | } | ||
2374 | if test.ChecksumSHA1 != "" { | ||
2375 | cmpChecksum(st.ChecksumSHA1, got) | ||
2376 | } | ||
2377 | if test.ChecksumCRC32 != "" { | ||
2378 | cmpChecksum(st.ChecksumCRC32, got) | ||
2379 | } | ||
2380 | if test.ChecksumCRC32C != "" { | ||
2381 | cmpChecksum(st.ChecksumCRC32C, got) | ||
2382 | } | ||
2383 | |||
2384 | delete(args, "metadata") | ||
2385 | } | ||
2386 | |||
2387 | successLogger(testName, function, args, startTime).Info() | ||
2388 | } | ||
2389 | |||
2390 | // Test PutObject with trailing checksums. | ||
2391 | func testTrailingChecksums() { | ||
2392 | // initialize logging params | ||
2393 | startTime := time.Now() | ||
2394 | testName := getFuncName() | ||
2395 | function := "PutObject(bucketName, objectName, reader,size, opts)" | ||
2396 | args := map[string]interface{}{ | ||
2397 | "bucketName": "", | ||
2398 | "objectName": "", | ||
2399 | "opts": "minio.PutObjectOptions{UserMetadata: metadata, Progress: progress}", | ||
2400 | } | ||
2401 | |||
2402 | if !isFullMode() { | ||
2403 | ignoredLog(testName, function, args, startTime, "Skipping functional tests for short/quick runs").Info() | ||
2404 | return | ||
2405 | } | ||
2406 | |||
2407 | // Instantiate new minio client object. | ||
2408 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
2409 | &minio.Options{ | ||
2410 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
2411 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
2412 | TrailingHeaders: true, | ||
2413 | }) | ||
2414 | if err != nil { | ||
2415 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
2416 | return | ||
2417 | } | ||
2418 | |||
2419 | // Enable tracing, write to stderr. | ||
2420 | // c.TraceOn(os.Stderr) | ||
2421 | |||
2422 | // Set user agent. | ||
2423 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
2424 | |||
2425 | // Generate a new random bucket name. | ||
2426 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
2427 | args["bucketName"] = bucketName | ||
2428 | |||
2429 | // Make a new bucket. | ||
2430 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
2431 | if err != nil { | ||
2432 | logError(testName, function, args, startTime, "", "Make bucket failed", err) | ||
2433 | return | ||
2434 | } | ||
2435 | |||
2436 | hashMultiPart := func(b []byte, partSize int, hasher hash.Hash) string { | ||
2437 | r := bytes.NewReader(b) | ||
2438 | tmp := make([]byte, partSize) | ||
2439 | parts := 0 | ||
2440 | var all []byte | ||
2441 | for { | ||
2442 | n, err := io.ReadFull(r, tmp) | ||
2443 | if err != nil && err != io.ErrUnexpectedEOF { | ||
2444 | logError(testName, function, args, startTime, "", "Calc crc failed", err) | ||
2445 | } | ||
2446 | if n == 0 { | ||
2447 | break | ||
2448 | } | ||
2449 | parts++ | ||
2450 | hasher.Reset() | ||
2451 | hasher.Write(tmp[:n]) | ||
2452 | all = append(all, hasher.Sum(nil)...) | ||
2453 | if err != nil { | ||
2454 | break | ||
2455 | } | ||
2456 | } | ||
2457 | hasher.Reset() | ||
2458 | hasher.Write(all) | ||
2459 | return fmt.Sprintf("%s-%d", base64.StdEncoding.EncodeToString(hasher.Sum(nil)), parts) | ||
2460 | } | ||
2461 | defer cleanupBucket(bucketName, c) | ||
2462 | tests := []struct { | ||
2463 | header string | ||
2464 | hasher hash.Hash | ||
2465 | |||
2466 | // Checksum values | ||
2467 | ChecksumCRC32 string | ||
2468 | ChecksumCRC32C string | ||
2469 | ChecksumSHA1 string | ||
2470 | ChecksumSHA256 string | ||
2471 | PO minio.PutObjectOptions | ||
2472 | }{ | ||
2473 | // Currently there is no way to override the checksum type. | ||
2474 | { | ||
2475 | header: "x-amz-checksum-crc32c", | ||
2476 | hasher: crc32.New(crc32.MakeTable(crc32.Castagnoli)), | ||
2477 | ChecksumCRC32C: "set", | ||
2478 | PO: minio.PutObjectOptions{ | ||
2479 | DisableContentSha256: true, | ||
2480 | DisableMultipart: false, | ||
2481 | UserMetadata: nil, | ||
2482 | PartSize: 5 << 20, | ||
2483 | }, | ||
2484 | }, | ||
2485 | { | ||
2486 | header: "x-amz-checksum-crc32c", | ||
2487 | hasher: crc32.New(crc32.MakeTable(crc32.Castagnoli)), | ||
2488 | ChecksumCRC32C: "set", | ||
2489 | PO: minio.PutObjectOptions{ | ||
2490 | DisableContentSha256: true, | ||
2491 | DisableMultipart: false, | ||
2492 | UserMetadata: nil, | ||
2493 | PartSize: 6_645_654, // Rather arbitrary size | ||
2494 | }, | ||
2495 | }, | ||
2496 | { | ||
2497 | header: "x-amz-checksum-crc32c", | ||
2498 | hasher: crc32.New(crc32.MakeTable(crc32.Castagnoli)), | ||
2499 | ChecksumCRC32C: "set", | ||
2500 | PO: minio.PutObjectOptions{ | ||
2501 | DisableContentSha256: false, | ||
2502 | DisableMultipart: false, | ||
2503 | UserMetadata: nil, | ||
2504 | PartSize: 5 << 20, | ||
2505 | }, | ||
2506 | }, | ||
2507 | { | ||
2508 | header: "x-amz-checksum-crc32c", | ||
2509 | hasher: crc32.New(crc32.MakeTable(crc32.Castagnoli)), | ||
2510 | ChecksumCRC32C: "set", | ||
2511 | PO: minio.PutObjectOptions{ | ||
2512 | DisableContentSha256: false, | ||
2513 | DisableMultipart: false, | ||
2514 | UserMetadata: nil, | ||
2515 | PartSize: 6_645_654, // Rather arbitrary size | ||
2516 | }, | ||
2517 | }, | ||
2518 | } | ||
2519 | |||
2520 | for _, test := range tests { | ||
2521 | bufSize := dataFileMap["datafile-11-MB"] | ||
2522 | |||
2523 | // Save the data | ||
2524 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
2525 | args["objectName"] = objectName | ||
2526 | |||
2527 | cmpChecksum := func(got, want string) { | ||
2528 | if want != got { | ||
2529 | logError(testName, function, args, startTime, "", "checksum mismatch", fmt.Errorf("want %q, got %q", want, got)) | ||
2530 | return | ||
2531 | } | ||
2532 | } | ||
2533 | |||
2534 | reader := getDataReader("datafile-11-MB") | ||
2535 | b, err := io.ReadAll(reader) | ||
2536 | if err != nil { | ||
2537 | logError(testName, function, args, startTime, "", "Read failed", err) | ||
2538 | return | ||
2539 | } | ||
2540 | reader.Close() | ||
2541 | h := test.hasher | ||
2542 | h.Reset() | ||
2543 | test.ChecksumCRC32C = hashMultiPart(b, int(test.PO.PartSize), test.hasher) | ||
2544 | |||
2545 | // Set correct CRC. | ||
2546 | // c.TraceOn(os.Stderr) | ||
2547 | resp, err := c.PutObject(context.Background(), bucketName, objectName, bytes.NewReader(b), int64(bufSize), test.PO) | ||
2548 | if err != nil { | ||
2549 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
2550 | return | ||
2551 | } | ||
2552 | // c.TraceOff() | ||
2553 | cmpChecksum(resp.ChecksumSHA256, test.ChecksumSHA256) | ||
2554 | cmpChecksum(resp.ChecksumSHA1, test.ChecksumSHA1) | ||
2555 | cmpChecksum(resp.ChecksumCRC32, test.ChecksumCRC32) | ||
2556 | cmpChecksum(resp.ChecksumCRC32C, test.ChecksumCRC32C) | ||
2557 | |||
2558 | // Read the data back | ||
2559 | gopts := minio.GetObjectOptions{Checksum: true} | ||
2560 | gopts.PartNumber = 2 | ||
2561 | |||
2562 | // We cannot use StatObject, since it ignores partnumber. | ||
2563 | r, err := c.GetObject(context.Background(), bucketName, objectName, gopts) | ||
2564 | if err != nil { | ||
2565 | logError(testName, function, args, startTime, "", "GetObject failed", err) | ||
2566 | return | ||
2567 | } | ||
2568 | io.Copy(io.Discard, r) | ||
2569 | st, err := r.Stat() | ||
2570 | if err != nil { | ||
2571 | logError(testName, function, args, startTime, "", "Stat failed", err) | ||
2572 | return | ||
2573 | } | ||
2574 | |||
2575 | // Test part 2 checksum... | ||
2576 | h.Reset() | ||
2577 | p2 := b[test.PO.PartSize:] | ||
2578 | if len(p2) > int(test.PO.PartSize) { | ||
2579 | p2 = p2[:test.PO.PartSize] | ||
2580 | } | ||
2581 | h.Write(p2) | ||
2582 | got := base64.StdEncoding.EncodeToString(h.Sum(nil)) | ||
2583 | if test.ChecksumSHA256 != "" { | ||
2584 | cmpChecksum(st.ChecksumSHA256, got) | ||
2585 | } | ||
2586 | if test.ChecksumSHA1 != "" { | ||
2587 | cmpChecksum(st.ChecksumSHA1, got) | ||
2588 | } | ||
2589 | if test.ChecksumCRC32 != "" { | ||
2590 | cmpChecksum(st.ChecksumCRC32, got) | ||
2591 | } | ||
2592 | if test.ChecksumCRC32C != "" { | ||
2593 | cmpChecksum(st.ChecksumCRC32C, got) | ||
2594 | } | ||
2595 | |||
2596 | delete(args, "metadata") | ||
2597 | } | ||
2598 | } | ||
2599 | |||
2600 | // Test PutObject with custom checksums. | ||
2601 | func testPutObjectWithAutomaticChecksums() { | ||
2602 | // initialize logging params | ||
2603 | startTime := time.Now() | ||
2604 | testName := getFuncName() | ||
2605 | function := "PutObject(bucketName, objectName, reader,size, opts)" | ||
2606 | args := map[string]interface{}{ | ||
2607 | "bucketName": "", | ||
2608 | "objectName": "", | ||
2609 | "opts": "minio.PutObjectOptions{UserMetadata: metadata, Progress: progress}", | ||
2610 | } | ||
2611 | |||
2612 | if !isFullMode() { | ||
2613 | ignoredLog(testName, function, args, startTime, "Skipping functional tests for short/quick runs").Info() | ||
2614 | return | ||
2615 | } | ||
2616 | |||
2617 | // Seed random based on current time. | ||
2618 | rand.Seed(time.Now().Unix()) | ||
2619 | |||
2620 | // Instantiate new minio client object. | ||
2621 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
2622 | &minio.Options{ | ||
2623 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
2624 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
2625 | TrailingHeaders: true, | ||
2626 | }) | ||
2627 | if err != nil { | ||
2628 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
2629 | return | ||
2630 | } | ||
2631 | |||
2632 | // Set user agent. | ||
2633 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
2634 | |||
2635 | // Generate a new random bucket name. | ||
2636 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
2637 | args["bucketName"] = bucketName | ||
2638 | |||
2639 | // Make a new bucket. | ||
2640 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
2641 | if err != nil { | ||
2642 | logError(testName, function, args, startTime, "", "Make bucket failed", err) | ||
2643 | return | ||
2644 | } | ||
2645 | |||
2646 | defer cleanupBucket(bucketName, c) | ||
2647 | tests := []struct { | ||
2648 | header string | ||
2649 | hasher hash.Hash | ||
2650 | |||
2651 | // Checksum values | ||
2652 | ChecksumCRC32 string | ||
2653 | ChecksumCRC32C string | ||
2654 | ChecksumSHA1 string | ||
2655 | ChecksumSHA256 string | ||
2656 | }{ | ||
2657 | // Built-in will only add crc32c, when no MD5 nor SHA256. | ||
2658 | {header: "x-amz-checksum-crc32c", hasher: crc32.New(crc32.MakeTable(crc32.Castagnoli))}, | ||
2659 | } | ||
2660 | |||
2661 | // Enable tracing, write to stderr. | ||
2662 | // c.TraceOn(os.Stderr) | ||
2663 | // defer c.TraceOff() | ||
2664 | |||
2665 | for i, test := range tests { | ||
2666 | bufSize := dataFileMap["datafile-10-kB"] | ||
2667 | |||
2668 | // Save the data | ||
2669 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
2670 | args["objectName"] = objectName | ||
2671 | |||
2672 | cmpChecksum := func(got, want string) { | ||
2673 | if want != got { | ||
2674 | logError(testName, function, args, startTime, "", "checksum mismatch", fmt.Errorf("want %s, got %s", want, got)) | ||
2675 | return | ||
2676 | } | ||
2677 | } | ||
2678 | |||
2679 | meta := map[string]string{} | ||
2680 | reader := getDataReader("datafile-10-kB") | ||
2681 | b, err := io.ReadAll(reader) | ||
2682 | if err != nil { | ||
2683 | logError(testName, function, args, startTime, "", "Read failed", err) | ||
2684 | return | ||
2685 | } | ||
2686 | |||
2687 | h := test.hasher | ||
2688 | h.Reset() | ||
2689 | h.Write(b) | ||
2690 | meta[test.header] = base64.StdEncoding.EncodeToString(h.Sum(nil)) | ||
2691 | args["metadata"] = meta | ||
2692 | |||
2693 | resp, err := c.PutObject(context.Background(), bucketName, objectName, bytes.NewReader(b), int64(bufSize), minio.PutObjectOptions{ | ||
2694 | DisableMultipart: true, | ||
2695 | UserMetadata: nil, | ||
2696 | DisableContentSha256: true, | ||
2697 | SendContentMd5: false, | ||
2698 | }) | ||
2699 | if err == nil { | ||
2700 | if i == 0 && resp.ChecksumCRC32C == "" { | ||
2701 | ignoredLog(testName, function, args, startTime, "Checksums does not appear to be supported by backend").Info() | ||
2702 | return | ||
2703 | } | ||
2704 | } else { | ||
2705 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
2706 | return | ||
2707 | } | ||
2708 | cmpChecksum(resp.ChecksumSHA256, meta["x-amz-checksum-sha256"]) | ||
2709 | cmpChecksum(resp.ChecksumSHA1, meta["x-amz-checksum-sha1"]) | ||
2710 | cmpChecksum(resp.ChecksumCRC32, meta["x-amz-checksum-crc32"]) | ||
2711 | cmpChecksum(resp.ChecksumCRC32C, meta["x-amz-checksum-crc32c"]) | ||
2712 | |||
2713 | // Usually this will be the same as above, since we skip automatic checksum when SHA256 content is sent. | ||
2714 | // When/if we add a checksum control to PutObjectOptions this will make more sense. | ||
2715 | resp, err = c.PutObject(context.Background(), bucketName, objectName, bytes.NewReader(b), int64(bufSize), minio.PutObjectOptions{ | ||
2716 | DisableMultipart: true, | ||
2717 | UserMetadata: nil, | ||
2718 | DisableContentSha256: false, | ||
2719 | SendContentMd5: false, | ||
2720 | }) | ||
2721 | if err != nil { | ||
2722 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
2723 | return | ||
2724 | } | ||
2725 | // The checksum will not be enabled on HTTP, since it uses SHA256 blocks. | ||
2726 | if mustParseBool(os.Getenv(enableHTTPS)) { | ||
2727 | cmpChecksum(resp.ChecksumSHA256, meta["x-amz-checksum-sha256"]) | ||
2728 | cmpChecksum(resp.ChecksumSHA1, meta["x-amz-checksum-sha1"]) | ||
2729 | cmpChecksum(resp.ChecksumCRC32, meta["x-amz-checksum-crc32"]) | ||
2730 | cmpChecksum(resp.ChecksumCRC32C, meta["x-amz-checksum-crc32c"]) | ||
2731 | } | ||
2732 | |||
2733 | // Set SHA256 header manually | ||
2734 | sh256 := sha256.Sum256(b) | ||
2735 | meta = map[string]string{"x-amz-checksum-sha256": base64.StdEncoding.EncodeToString(sh256[:])} | ||
2736 | args["metadata"] = meta | ||
2737 | resp, err = c.PutObject(context.Background(), bucketName, objectName, bytes.NewReader(b), int64(bufSize), minio.PutObjectOptions{ | ||
2738 | DisableMultipart: true, | ||
2739 | UserMetadata: meta, | ||
2740 | DisableContentSha256: true, | ||
2741 | SendContentMd5: false, | ||
2742 | }) | ||
2743 | if err != nil { | ||
2744 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
2745 | return | ||
2746 | } | ||
2747 | cmpChecksum(resp.ChecksumSHA256, meta["x-amz-checksum-sha256"]) | ||
2748 | cmpChecksum(resp.ChecksumSHA1, meta["x-amz-checksum-sha1"]) | ||
2749 | cmpChecksum(resp.ChecksumCRC32, meta["x-amz-checksum-crc32"]) | ||
2750 | cmpChecksum(resp.ChecksumCRC32C, meta["x-amz-checksum-crc32c"]) | ||
2751 | delete(args, "metadata") | ||
2752 | } | ||
2753 | |||
2754 | successLogger(testName, function, args, startTime).Info() | ||
2755 | } | ||
2756 | |||
2757 | // Test PutObject using a large data to trigger multipart readat | ||
2758 | func testPutObjectWithMetadata() { | ||
2759 | // initialize logging params | ||
2760 | startTime := time.Now() | ||
2761 | testName := getFuncName() | ||
2762 | function := "PutObject(bucketName, objectName, reader,size, opts)" | ||
2763 | args := map[string]interface{}{ | ||
2764 | "bucketName": "", | ||
2765 | "objectName": "", | ||
2766 | "opts": "minio.PutObjectOptions{UserMetadata: metadata, Progress: progress}", | ||
2767 | } | ||
2768 | |||
2769 | if !isFullMode() { | ||
2770 | ignoredLog(testName, function, args, startTime, "Skipping functional tests for short/quick runs").Info() | ||
2771 | return | ||
2772 | } | ||
2773 | |||
2774 | // Seed random based on current time. | ||
2775 | rand.Seed(time.Now().Unix()) | ||
2776 | |||
2777 | // Instantiate new minio client object. | ||
2778 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
2779 | &minio.Options{ | ||
2780 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
2781 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
2782 | }) | ||
2783 | if err != nil { | ||
2784 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
2785 | return | ||
2786 | } | ||
2787 | |||
2788 | // Enable tracing, write to stderr. | ||
2789 | // c.TraceOn(os.Stderr) | ||
2790 | |||
2791 | // Set user agent. | ||
2792 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
2793 | |||
2794 | // Generate a new random bucket name. | ||
2795 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
2796 | args["bucketName"] = bucketName | ||
2797 | |||
2798 | // Make a new bucket. | ||
2799 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
2800 | if err != nil { | ||
2801 | logError(testName, function, args, startTime, "", "Make bucket failed", err) | ||
2802 | return | ||
2803 | } | ||
2804 | |||
2805 | defer cleanupBucket(bucketName, c) | ||
2806 | |||
2807 | bufSize := dataFileMap["datafile-129-MB"] | ||
2808 | reader := getDataReader("datafile-129-MB") | ||
2809 | defer reader.Close() | ||
2810 | |||
2811 | // Save the data | ||
2812 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
2813 | args["objectName"] = objectName | ||
2814 | |||
2815 | // Object custom metadata | ||
2816 | customContentType := "custom/contenttype" | ||
2817 | |||
2818 | args["metadata"] = map[string][]string{ | ||
2819 | "Content-Type": {customContentType}, | ||
2820 | "X-Amz-Meta-CustomKey": {"extra spaces in value"}, | ||
2821 | } | ||
2822 | |||
2823 | _, err = c.PutObject(context.Background(), bucketName, objectName, reader, int64(bufSize), minio.PutObjectOptions{ | ||
2824 | ContentType: customContentType, | ||
2825 | }) | ||
2826 | if err != nil { | ||
2827 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
2828 | return | ||
2829 | } | ||
2830 | |||
2831 | // Read the data back | ||
2832 | r, err := c.GetObject(context.Background(), bucketName, objectName, minio.GetObjectOptions{}) | ||
2833 | if err != nil { | ||
2834 | logError(testName, function, args, startTime, "", "GetObject failed", err) | ||
2835 | return | ||
2836 | } | ||
2837 | |||
2838 | st, err := r.Stat() | ||
2839 | if err != nil { | ||
2840 | logError(testName, function, args, startTime, "", "Stat failed", err) | ||
2841 | return | ||
2842 | } | ||
2843 | if st.Size != int64(bufSize) { | ||
2844 | logError(testName, function, args, startTime, "", "Number of bytes returned by PutObject does not match GetObject, expected "+string(bufSize)+" got "+string(st.Size), err) | ||
2845 | return | ||
2846 | } | ||
2847 | if st.ContentType != customContentType && st.ContentType != "application/octet-stream" { | ||
2848 | logError(testName, function, args, startTime, "", "ContentType does not match, expected "+customContentType+" got "+st.ContentType, err) | ||
2849 | return | ||
2850 | } | ||
2851 | if err := crcMatchesName(r, "datafile-129-MB"); err != nil { | ||
2852 | logError(testName, function, args, startTime, "", "data CRC check failed", err) | ||
2853 | return | ||
2854 | } | ||
2855 | if err := r.Close(); err != nil { | ||
2856 | logError(testName, function, args, startTime, "", "Object Close failed", err) | ||
2857 | return | ||
2858 | } | ||
2859 | if err := r.Close(); err == nil { | ||
2860 | logError(testName, function, args, startTime, "", "Object already closed, should respond with error", err) | ||
2861 | return | ||
2862 | } | ||
2863 | |||
2864 | successLogger(testName, function, args, startTime).Info() | ||
2865 | } | ||
2866 | |||
2867 | func testPutObjectWithContentLanguage() { | ||
2868 | // initialize logging params | ||
2869 | objectName := "test-object" | ||
2870 | startTime := time.Now() | ||
2871 | testName := getFuncName() | ||
2872 | function := "PutObject(bucketName, objectName, reader, size, opts)" | ||
2873 | args := map[string]interface{}{ | ||
2874 | "bucketName": "", | ||
2875 | "objectName": objectName, | ||
2876 | "size": -1, | ||
2877 | "opts": "", | ||
2878 | } | ||
2879 | |||
2880 | // Seed random based on current time. | ||
2881 | rand.Seed(time.Now().Unix()) | ||
2882 | |||
2883 | // Instantiate new minio client object. | ||
2884 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
2885 | &minio.Options{ | ||
2886 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
2887 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
2888 | }) | ||
2889 | if err != nil { | ||
2890 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
2891 | return | ||
2892 | } | ||
2893 | |||
2894 | // Enable tracing, write to stderr. | ||
2895 | // c.TraceOn(os.Stderr) | ||
2896 | |||
2897 | // Set user agent. | ||
2898 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
2899 | |||
2900 | // Generate a new random bucket name. | ||
2901 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
2902 | args["bucketName"] = bucketName | ||
2903 | // Make a new bucket. | ||
2904 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
2905 | if err != nil { | ||
2906 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
2907 | return | ||
2908 | } | ||
2909 | |||
2910 | defer cleanupBucket(bucketName, c) | ||
2911 | |||
2912 | data := []byte{} | ||
2913 | _, err = c.PutObject(context.Background(), bucketName, objectName, bytes.NewReader(data), int64(0), minio.PutObjectOptions{ | ||
2914 | ContentLanguage: "en", | ||
2915 | }) | ||
2916 | if err != nil { | ||
2917 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
2918 | return | ||
2919 | } | ||
2920 | |||
2921 | objInfo, err := c.StatObject(context.Background(), bucketName, objectName, minio.StatObjectOptions{}) | ||
2922 | if err != nil { | ||
2923 | logError(testName, function, args, startTime, "", "StatObject failed", err) | ||
2924 | return | ||
2925 | } | ||
2926 | |||
2927 | if objInfo.Metadata.Get("Content-Language") != "en" { | ||
2928 | logError(testName, function, args, startTime, "", "Expected content-language 'en' doesn't match with StatObject return value", err) | ||
2929 | return | ||
2930 | } | ||
2931 | |||
2932 | successLogger(testName, function, args, startTime).Info() | ||
2933 | } | ||
2934 | |||
2935 | // Test put object with streaming signature. | ||
2936 | func testPutObjectStreaming() { | ||
2937 | // initialize logging params | ||
2938 | objectName := "test-object" | ||
2939 | startTime := time.Now() | ||
2940 | testName := getFuncName() | ||
2941 | function := "PutObject(bucketName, objectName, reader,size,opts)" | ||
2942 | args := map[string]interface{}{ | ||
2943 | "bucketName": "", | ||
2944 | "objectName": objectName, | ||
2945 | "size": -1, | ||
2946 | "opts": "", | ||
2947 | } | ||
2948 | |||
2949 | // Seed random based on current time. | ||
2950 | rand.Seed(time.Now().Unix()) | ||
2951 | |||
2952 | // Instantiate new minio client object. | ||
2953 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
2954 | &minio.Options{ | ||
2955 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
2956 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
2957 | }) | ||
2958 | if err != nil { | ||
2959 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
2960 | return | ||
2961 | } | ||
2962 | |||
2963 | // Enable tracing, write to stderr. | ||
2964 | // c.TraceOn(os.Stderr) | ||
2965 | |||
2966 | // Set user agent. | ||
2967 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
2968 | |||
2969 | // Generate a new random bucket name. | ||
2970 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
2971 | args["bucketName"] = bucketName | ||
2972 | // Make a new bucket. | ||
2973 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
2974 | if err != nil { | ||
2975 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
2976 | return | ||
2977 | } | ||
2978 | |||
2979 | defer cleanupBucket(bucketName, c) | ||
2980 | |||
2981 | // Upload an object. | ||
2982 | sizes := []int64{0, 64*1024 - 1, 64 * 1024} | ||
2983 | |||
2984 | for _, size := range sizes { | ||
2985 | data := newRandomReader(size, size) | ||
2986 | ui, err := c.PutObject(context.Background(), bucketName, objectName, data, int64(size), minio.PutObjectOptions{}) | ||
2987 | if err != nil { | ||
2988 | logError(testName, function, args, startTime, "", "PutObjectStreaming failed", err) | ||
2989 | return | ||
2990 | } | ||
2991 | |||
2992 | if ui.Size != size { | ||
2993 | logError(testName, function, args, startTime, "", "PutObjectStreaming result has unexpected size", nil) | ||
2994 | return | ||
2995 | } | ||
2996 | |||
2997 | objInfo, err := c.StatObject(context.Background(), bucketName, objectName, minio.StatObjectOptions{}) | ||
2998 | if err != nil { | ||
2999 | logError(testName, function, args, startTime, "", "StatObject failed", err) | ||
3000 | return | ||
3001 | } | ||
3002 | if objInfo.Size != size { | ||
3003 | logError(testName, function, args, startTime, "", "Unexpected size", err) | ||
3004 | return | ||
3005 | } | ||
3006 | |||
3007 | } | ||
3008 | |||
3009 | successLogger(testName, function, args, startTime).Info() | ||
3010 | } | ||
3011 | |||
3012 | // Test get object seeker from the end, using whence set to '2'. | ||
3013 | func testGetObjectSeekEnd() { | ||
3014 | // initialize logging params | ||
3015 | startTime := time.Now() | ||
3016 | testName := getFuncName() | ||
3017 | function := "GetObject(bucketName, objectName)" | ||
3018 | args := map[string]interface{}{} | ||
3019 | |||
3020 | // Seed random based on current time. | ||
3021 | rand.Seed(time.Now().Unix()) | ||
3022 | |||
3023 | // Instantiate new minio client object. | ||
3024 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
3025 | &minio.Options{ | ||
3026 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
3027 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
3028 | }) | ||
3029 | if err != nil { | ||
3030 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
3031 | return | ||
3032 | } | ||
3033 | |||
3034 | // Enable tracing, write to stderr. | ||
3035 | // c.TraceOn(os.Stderr) | ||
3036 | |||
3037 | // Set user agent. | ||
3038 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
3039 | |||
3040 | // Generate a new random bucket name. | ||
3041 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
3042 | args["bucketName"] = bucketName | ||
3043 | |||
3044 | // Make a new bucket. | ||
3045 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
3046 | if err != nil { | ||
3047 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
3048 | return | ||
3049 | } | ||
3050 | |||
3051 | defer cleanupBucket(bucketName, c) | ||
3052 | |||
3053 | // Generate 33K of data. | ||
3054 | bufSize := dataFileMap["datafile-33-kB"] | ||
3055 | reader := getDataReader("datafile-33-kB") | ||
3056 | defer reader.Close() | ||
3057 | |||
3058 | // Save the data | ||
3059 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
3060 | args["objectName"] = objectName | ||
3061 | |||
3062 | buf, err := io.ReadAll(reader) | ||
3063 | if err != nil { | ||
3064 | logError(testName, function, args, startTime, "", "ReadAll failed", err) | ||
3065 | return | ||
3066 | } | ||
3067 | |||
3068 | _, err = c.PutObject(context.Background(), bucketName, objectName, bytes.NewReader(buf), int64(len(buf)), minio.PutObjectOptions{ContentType: "binary/octet-stream"}) | ||
3069 | if err != nil { | ||
3070 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
3071 | return | ||
3072 | } | ||
3073 | |||
3074 | // Read the data back | ||
3075 | r, err := c.GetObject(context.Background(), bucketName, objectName, minio.GetObjectOptions{}) | ||
3076 | if err != nil { | ||
3077 | logError(testName, function, args, startTime, "", "GetObject failed", err) | ||
3078 | return | ||
3079 | } | ||
3080 | |||
3081 | st, err := r.Stat() | ||
3082 | if err != nil { | ||
3083 | logError(testName, function, args, startTime, "", "Stat failed", err) | ||
3084 | return | ||
3085 | } | ||
3086 | |||
3087 | if st.Size != int64(bufSize) { | ||
3088 | logError(testName, function, args, startTime, "", "Number of bytes read does not match, expected "+string(int64(bufSize))+" got "+string(st.Size), err) | ||
3089 | return | ||
3090 | } | ||
3091 | |||
3092 | pos, err := r.Seek(-100, 2) | ||
3093 | if err != nil { | ||
3094 | logError(testName, function, args, startTime, "", "Object Seek failed", err) | ||
3095 | return | ||
3096 | } | ||
3097 | if pos != st.Size-100 { | ||
3098 | logError(testName, function, args, startTime, "", "Incorrect position", err) | ||
3099 | return | ||
3100 | } | ||
3101 | buf2 := make([]byte, 100) | ||
3102 | m, err := readFull(r, buf2) | ||
3103 | if err != nil { | ||
3104 | logError(testName, function, args, startTime, "", "Error reading through readFull", err) | ||
3105 | return | ||
3106 | } | ||
3107 | if m != len(buf2) { | ||
3108 | logError(testName, function, args, startTime, "", "Number of bytes dont match, expected "+string(len(buf2))+" got "+string(m), err) | ||
3109 | return | ||
3110 | } | ||
3111 | hexBuf1 := fmt.Sprintf("%02x", buf[len(buf)-100:]) | ||
3112 | hexBuf2 := fmt.Sprintf("%02x", buf2[:m]) | ||
3113 | if hexBuf1 != hexBuf2 { | ||
3114 | logError(testName, function, args, startTime, "", "Values at same index dont match", err) | ||
3115 | return | ||
3116 | } | ||
3117 | pos, err = r.Seek(-100, 2) | ||
3118 | if err != nil { | ||
3119 | logError(testName, function, args, startTime, "", "Object Seek failed", err) | ||
3120 | return | ||
3121 | } | ||
3122 | if pos != st.Size-100 { | ||
3123 | logError(testName, function, args, startTime, "", "Incorrect position", err) | ||
3124 | return | ||
3125 | } | ||
3126 | if err = r.Close(); err != nil { | ||
3127 | logError(testName, function, args, startTime, "", "ObjectClose failed", err) | ||
3128 | return | ||
3129 | } | ||
3130 | |||
3131 | successLogger(testName, function, args, startTime).Info() | ||
3132 | } | ||
3133 | |||
3134 | // Test get object reader to not throw error on being closed twice. | ||
3135 | func testGetObjectClosedTwice() { | ||
3136 | // initialize logging params | ||
3137 | startTime := time.Now() | ||
3138 | testName := getFuncName() | ||
3139 | function := "GetObject(bucketName, objectName)" | ||
3140 | args := map[string]interface{}{} | ||
3141 | |||
3142 | // Seed random based on current time. | ||
3143 | rand.Seed(time.Now().Unix()) | ||
3144 | |||
3145 | // Instantiate new minio client object. | ||
3146 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
3147 | &minio.Options{ | ||
3148 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
3149 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
3150 | }) | ||
3151 | if err != nil { | ||
3152 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
3153 | return | ||
3154 | } | ||
3155 | |||
3156 | // Enable tracing, write to stderr. | ||
3157 | // c.TraceOn(os.Stderr) | ||
3158 | |||
3159 | // Set user agent. | ||
3160 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
3161 | |||
3162 | // Generate a new random bucket name. | ||
3163 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
3164 | args["bucketName"] = bucketName | ||
3165 | |||
3166 | // Make a new bucket. | ||
3167 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
3168 | if err != nil { | ||
3169 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
3170 | return | ||
3171 | } | ||
3172 | |||
3173 | defer cleanupBucket(bucketName, c) | ||
3174 | |||
3175 | // Generate 33K of data. | ||
3176 | bufSize := dataFileMap["datafile-33-kB"] | ||
3177 | reader := getDataReader("datafile-33-kB") | ||
3178 | defer reader.Close() | ||
3179 | |||
3180 | // Save the data | ||
3181 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
3182 | args["objectName"] = objectName | ||
3183 | |||
3184 | _, err = c.PutObject(context.Background(), bucketName, objectName, reader, int64(bufSize), minio.PutObjectOptions{ContentType: "binary/octet-stream"}) | ||
3185 | if err != nil { | ||
3186 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
3187 | return | ||
3188 | } | ||
3189 | |||
3190 | // Read the data back | ||
3191 | r, err := c.GetObject(context.Background(), bucketName, objectName, minio.GetObjectOptions{}) | ||
3192 | if err != nil { | ||
3193 | logError(testName, function, args, startTime, "", "GetObject failed", err) | ||
3194 | return | ||
3195 | } | ||
3196 | |||
3197 | st, err := r.Stat() | ||
3198 | if err != nil { | ||
3199 | logError(testName, function, args, startTime, "", "Stat failed", err) | ||
3200 | return | ||
3201 | } | ||
3202 | if st.Size != int64(bufSize) { | ||
3203 | logError(testName, function, args, startTime, "", "Number of bytes in stat does not match, expected "+string(int64(bufSize))+" got "+string(st.Size), err) | ||
3204 | return | ||
3205 | } | ||
3206 | if err := crcMatchesName(r, "datafile-33-kB"); err != nil { | ||
3207 | logError(testName, function, args, startTime, "", "data CRC check failed", err) | ||
3208 | return | ||
3209 | } | ||
3210 | if err := r.Close(); err != nil { | ||
3211 | logError(testName, function, args, startTime, "", "Object Close failed", err) | ||
3212 | return | ||
3213 | } | ||
3214 | if err := r.Close(); err == nil { | ||
3215 | logError(testName, function, args, startTime, "", "Already closed object. No error returned", err) | ||
3216 | return | ||
3217 | } | ||
3218 | |||
3219 | successLogger(testName, function, args, startTime).Info() | ||
3220 | } | ||
3221 | |||
3222 | // Test RemoveObjects request where context cancels after timeout | ||
3223 | func testRemoveObjectsContext() { | ||
3224 | // Initialize logging params. | ||
3225 | startTime := time.Now() | ||
3226 | testName := getFuncName() | ||
3227 | function := "RemoveObjects(ctx, bucketName, objectsCh)" | ||
3228 | args := map[string]interface{}{ | ||
3229 | "bucketName": "", | ||
3230 | } | ||
3231 | |||
3232 | // Seed random based on current tie. | ||
3233 | rand.Seed(time.Now().Unix()) | ||
3234 | |||
3235 | // Instantiate new minio client. | ||
3236 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
3237 | &minio.Options{ | ||
3238 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
3239 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
3240 | }) | ||
3241 | if err != nil { | ||
3242 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
3243 | return | ||
3244 | } | ||
3245 | |||
3246 | // Set user agent. | ||
3247 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
3248 | // Enable tracing, write to stdout. | ||
3249 | // c.TraceOn(os.Stderr) | ||
3250 | |||
3251 | // Generate a new random bucket name. | ||
3252 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
3253 | args["bucketName"] = bucketName | ||
3254 | |||
3255 | // Make a new bucket. | ||
3256 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
3257 | if err != nil { | ||
3258 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
3259 | return | ||
3260 | } | ||
3261 | |||
3262 | defer cleanupBucket(bucketName, c) | ||
3263 | |||
3264 | // Generate put data. | ||
3265 | r := bytes.NewReader(bytes.Repeat([]byte("a"), 8)) | ||
3266 | |||
3267 | // Multi remove of 20 objects. | ||
3268 | nrObjects := 20 | ||
3269 | objectsCh := make(chan minio.ObjectInfo) | ||
3270 | go func() { | ||
3271 | defer close(objectsCh) | ||
3272 | for i := 0; i < nrObjects; i++ { | ||
3273 | objectName := "sample" + strconv.Itoa(i) + ".txt" | ||
3274 | info, err := c.PutObject(context.Background(), bucketName, objectName, r, 8, | ||
3275 | minio.PutObjectOptions{ContentType: "application/octet-stream"}) | ||
3276 | if err != nil { | ||
3277 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
3278 | continue | ||
3279 | } | ||
3280 | objectsCh <- minio.ObjectInfo{ | ||
3281 | Key: info.Key, | ||
3282 | VersionID: info.VersionID, | ||
3283 | } | ||
3284 | } | ||
3285 | }() | ||
3286 | // Set context to cancel in 1 nanosecond. | ||
3287 | ctx, cancel := context.WithTimeout(context.Background(), 1*time.Nanosecond) | ||
3288 | args["ctx"] = ctx | ||
3289 | defer cancel() | ||
3290 | |||
3291 | // Call RemoveObjects API with short timeout. | ||
3292 | errorCh := c.RemoveObjects(ctx, bucketName, objectsCh, minio.RemoveObjectsOptions{}) | ||
3293 | // Check for error. | ||
3294 | select { | ||
3295 | case r := <-errorCh: | ||
3296 | if r.Err == nil { | ||
3297 | logError(testName, function, args, startTime, "", "RemoveObjects should fail on short timeout", err) | ||
3298 | return | ||
3299 | } | ||
3300 | } | ||
3301 | // Set context with longer timeout. | ||
3302 | ctx, cancel = context.WithTimeout(context.Background(), 1*time.Hour) | ||
3303 | args["ctx"] = ctx | ||
3304 | defer cancel() | ||
3305 | // Perform RemoveObjects with the longer timeout. Expect the removals to succeed. | ||
3306 | errorCh = c.RemoveObjects(ctx, bucketName, objectsCh, minio.RemoveObjectsOptions{}) | ||
3307 | select { | ||
3308 | case r, more := <-errorCh: | ||
3309 | if more || r.Err != nil { | ||
3310 | logError(testName, function, args, startTime, "", "Unexpected error", r.Err) | ||
3311 | return | ||
3312 | } | ||
3313 | } | ||
3314 | |||
3315 | successLogger(testName, function, args, startTime).Info() | ||
3316 | } | ||
3317 | |||
3318 | // Test removing multiple objects with Remove API | ||
3319 | func testRemoveMultipleObjects() { | ||
3320 | // initialize logging params | ||
3321 | startTime := time.Now() | ||
3322 | testName := getFuncName() | ||
3323 | function := "RemoveObjects(bucketName, objectsCh)" | ||
3324 | args := map[string]interface{}{ | ||
3325 | "bucketName": "", | ||
3326 | } | ||
3327 | |||
3328 | // Seed random based on current time. | ||
3329 | rand.Seed(time.Now().Unix()) | ||
3330 | |||
3331 | // Instantiate new minio client object. | ||
3332 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
3333 | &minio.Options{ | ||
3334 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
3335 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
3336 | }) | ||
3337 | if err != nil { | ||
3338 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
3339 | return | ||
3340 | } | ||
3341 | |||
3342 | // Set user agent. | ||
3343 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
3344 | |||
3345 | // Enable tracing, write to stdout. | ||
3346 | // c.TraceOn(os.Stderr) | ||
3347 | |||
3348 | // Generate a new random bucket name. | ||
3349 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
3350 | args["bucketName"] = bucketName | ||
3351 | |||
3352 | // Make a new bucket. | ||
3353 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
3354 | if err != nil { | ||
3355 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
3356 | return | ||
3357 | } | ||
3358 | |||
3359 | defer cleanupBucket(bucketName, c) | ||
3360 | |||
3361 | r := bytes.NewReader(bytes.Repeat([]byte("a"), 8)) | ||
3362 | |||
3363 | // Multi remove of 1100 objects | ||
3364 | nrObjects := 200 | ||
3365 | |||
3366 | objectsCh := make(chan minio.ObjectInfo) | ||
3367 | |||
3368 | go func() { | ||
3369 | defer close(objectsCh) | ||
3370 | // Upload objects and send them to objectsCh | ||
3371 | for i := 0; i < nrObjects; i++ { | ||
3372 | objectName := "sample" + strconv.Itoa(i) + ".txt" | ||
3373 | info, err := c.PutObject(context.Background(), bucketName, objectName, r, 8, | ||
3374 | minio.PutObjectOptions{ContentType: "application/octet-stream"}) | ||
3375 | if err != nil { | ||
3376 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
3377 | continue | ||
3378 | } | ||
3379 | objectsCh <- minio.ObjectInfo{ | ||
3380 | Key: info.Key, | ||
3381 | VersionID: info.VersionID, | ||
3382 | } | ||
3383 | } | ||
3384 | }() | ||
3385 | |||
3386 | // Call RemoveObjects API | ||
3387 | errorCh := c.RemoveObjects(context.Background(), bucketName, objectsCh, minio.RemoveObjectsOptions{}) | ||
3388 | |||
3389 | // Check if errorCh doesn't receive any error | ||
3390 | select { | ||
3391 | case r, more := <-errorCh: | ||
3392 | if more { | ||
3393 | logError(testName, function, args, startTime, "", "Unexpected error", r.Err) | ||
3394 | return | ||
3395 | } | ||
3396 | } | ||
3397 | |||
3398 | successLogger(testName, function, args, startTime).Info() | ||
3399 | } | ||
3400 | |||
3401 | // Test removing multiple objects and check for results | ||
3402 | func testRemoveMultipleObjectsWithResult() { | ||
3403 | // initialize logging params | ||
3404 | startTime := time.Now() | ||
3405 | testName := getFuncName() | ||
3406 | function := "RemoveObjects(bucketName, objectsCh)" | ||
3407 | args := map[string]interface{}{ | ||
3408 | "bucketName": "", | ||
3409 | } | ||
3410 | |||
3411 | // Seed random based on current time. | ||
3412 | rand.Seed(time.Now().Unix()) | ||
3413 | |||
3414 | // Instantiate new minio client object. | ||
3415 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
3416 | &minio.Options{ | ||
3417 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
3418 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
3419 | }) | ||
3420 | if err != nil { | ||
3421 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
3422 | return | ||
3423 | } | ||
3424 | |||
3425 | // Set user agent. | ||
3426 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
3427 | |||
3428 | // Enable tracing, write to stdout. | ||
3429 | // c.TraceOn(os.Stderr) | ||
3430 | |||
3431 | // Generate a new random bucket name. | ||
3432 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
3433 | args["bucketName"] = bucketName | ||
3434 | |||
3435 | // Make a new bucket. | ||
3436 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1", ObjectLocking: true}) | ||
3437 | if err != nil { | ||
3438 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
3439 | return | ||
3440 | } | ||
3441 | |||
3442 | defer cleanupVersionedBucket(bucketName, c) | ||
3443 | |||
3444 | r := bytes.NewReader(bytes.Repeat([]byte("a"), 8)) | ||
3445 | |||
3446 | nrObjects := 10 | ||
3447 | nrLockedObjects := 5 | ||
3448 | |||
3449 | objectsCh := make(chan minio.ObjectInfo) | ||
3450 | |||
3451 | go func() { | ||
3452 | defer close(objectsCh) | ||
3453 | // Upload objects and send them to objectsCh | ||
3454 | for i := 0; i < nrObjects; i++ { | ||
3455 | objectName := "sample" + strconv.Itoa(i) + ".txt" | ||
3456 | info, err := c.PutObject(context.Background(), bucketName, objectName, r, 8, | ||
3457 | minio.PutObjectOptions{ContentType: "application/octet-stream"}) | ||
3458 | if err != nil { | ||
3459 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
3460 | return | ||
3461 | } | ||
3462 | if i < nrLockedObjects { | ||
3463 | // t := time.Date(2130, time.April, 25, 14, 0, 0, 0, time.UTC) | ||
3464 | t := time.Now().Add(5 * time.Minute) | ||
3465 | m := minio.RetentionMode(minio.Governance) | ||
3466 | opts := minio.PutObjectRetentionOptions{ | ||
3467 | GovernanceBypass: false, | ||
3468 | RetainUntilDate: &t, | ||
3469 | Mode: &m, | ||
3470 | VersionID: info.VersionID, | ||
3471 | } | ||
3472 | err = c.PutObjectRetention(context.Background(), bucketName, objectName, opts) | ||
3473 | if err != nil { | ||
3474 | logError(testName, function, args, startTime, "", "Error setting retention", err) | ||
3475 | return | ||
3476 | } | ||
3477 | } | ||
3478 | |||
3479 | objectsCh <- minio.ObjectInfo{ | ||
3480 | Key: info.Key, | ||
3481 | VersionID: info.VersionID, | ||
3482 | } | ||
3483 | } | ||
3484 | }() | ||
3485 | |||
3486 | // Call RemoveObjects API | ||
3487 | resultCh := c.RemoveObjectsWithResult(context.Background(), bucketName, objectsCh, minio.RemoveObjectsOptions{}) | ||
3488 | |||
3489 | var foundNil, foundErr int | ||
3490 | |||
3491 | for { | ||
3492 | // Check if errorCh doesn't receive any error | ||
3493 | select { | ||
3494 | case deleteRes, ok := <-resultCh: | ||
3495 | if !ok { | ||
3496 | goto out | ||
3497 | } | ||
3498 | if deleteRes.ObjectName == "" { | ||
3499 | logError(testName, function, args, startTime, "", "Unexpected object name", nil) | ||
3500 | return | ||
3501 | } | ||
3502 | if deleteRes.ObjectVersionID == "" { | ||
3503 | logError(testName, function, args, startTime, "", "Unexpected object version ID", nil) | ||
3504 | return | ||
3505 | } | ||
3506 | |||
3507 | if deleteRes.Err == nil { | ||
3508 | foundNil++ | ||
3509 | } else { | ||
3510 | foundErr++ | ||
3511 | } | ||
3512 | } | ||
3513 | } | ||
3514 | out: | ||
3515 | if foundNil+foundErr != nrObjects { | ||
3516 | logError(testName, function, args, startTime, "", "Unexpected number of results", nil) | ||
3517 | return | ||
3518 | } | ||
3519 | |||
3520 | if foundNil != nrObjects-nrLockedObjects { | ||
3521 | logError(testName, function, args, startTime, "", "Unexpected number of nil errors", nil) | ||
3522 | return | ||
3523 | } | ||
3524 | |||
3525 | if foundErr != nrLockedObjects { | ||
3526 | logError(testName, function, args, startTime, "", "Unexpected number of errors", nil) | ||
3527 | return | ||
3528 | } | ||
3529 | |||
3530 | successLogger(testName, function, args, startTime).Info() | ||
3531 | } | ||
3532 | |||
3533 | // Tests FPutObject of a big file to trigger multipart | ||
3534 | func testFPutObjectMultipart() { | ||
3535 | // initialize logging params | ||
3536 | startTime := time.Now() | ||
3537 | testName := getFuncName() | ||
3538 | function := "FPutObject(bucketName, objectName, fileName, opts)" | ||
3539 | args := map[string]interface{}{ | ||
3540 | "bucketName": "", | ||
3541 | "objectName": "", | ||
3542 | "fileName": "", | ||
3543 | "opts": "", | ||
3544 | } | ||
3545 | |||
3546 | // Seed random based on current time. | ||
3547 | rand.Seed(time.Now().Unix()) | ||
3548 | |||
3549 | // Instantiate new minio client object. | ||
3550 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
3551 | &minio.Options{ | ||
3552 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
3553 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
3554 | }) | ||
3555 | if err != nil { | ||
3556 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
3557 | return | ||
3558 | } | ||
3559 | |||
3560 | // Enable tracing, write to stderr. | ||
3561 | // c.TraceOn(os.Stderr) | ||
3562 | |||
3563 | // Set user agent. | ||
3564 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
3565 | |||
3566 | // Generate a new random bucket name. | ||
3567 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
3568 | args["bucketName"] = bucketName | ||
3569 | |||
3570 | // Make a new bucket. | ||
3571 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
3572 | if err != nil { | ||
3573 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
3574 | return | ||
3575 | } | ||
3576 | |||
3577 | defer cleanupBucket(bucketName, c) | ||
3578 | |||
3579 | // Upload 4 parts to utilize all 3 'workers' in multipart and still have a part to upload. | ||
3580 | fileName := getMintDataDirFilePath("datafile-129-MB") | ||
3581 | if fileName == "" { | ||
3582 | // Make a temp file with minPartSize bytes of data. | ||
3583 | file, err := os.CreateTemp(os.TempDir(), "FPutObjectTest") | ||
3584 | if err != nil { | ||
3585 | logError(testName, function, args, startTime, "", "TempFile creation failed", err) | ||
3586 | return | ||
3587 | } | ||
3588 | // Upload 2 parts to utilize all 3 'workers' in multipart and still have a part to upload. | ||
3589 | if _, err = io.Copy(file, getDataReader("datafile-129-MB")); err != nil { | ||
3590 | logError(testName, function, args, startTime, "", "Copy failed", err) | ||
3591 | return | ||
3592 | } | ||
3593 | if err = file.Close(); err != nil { | ||
3594 | logError(testName, function, args, startTime, "", "File Close failed", err) | ||
3595 | return | ||
3596 | } | ||
3597 | fileName = file.Name() | ||
3598 | args["fileName"] = fileName | ||
3599 | } | ||
3600 | totalSize := dataFileMap["datafile-129-MB"] | ||
3601 | // Set base object name | ||
3602 | objectName := bucketName + "FPutObject" + "-standard" | ||
3603 | args["objectName"] = objectName | ||
3604 | |||
3605 | objectContentType := "testapplication/octet-stream" | ||
3606 | args["objectContentType"] = objectContentType | ||
3607 | |||
3608 | // Perform standard FPutObject with contentType provided (Expecting application/octet-stream) | ||
3609 | _, err = c.FPutObject(context.Background(), bucketName, objectName, fileName, minio.PutObjectOptions{ContentType: objectContentType}) | ||
3610 | if err != nil { | ||
3611 | logError(testName, function, args, startTime, "", "FPutObject failed", err) | ||
3612 | return | ||
3613 | } | ||
3614 | |||
3615 | r, err := c.GetObject(context.Background(), bucketName, objectName, minio.GetObjectOptions{}) | ||
3616 | if err != nil { | ||
3617 | logError(testName, function, args, startTime, "", "GetObject failed", err) | ||
3618 | return | ||
3619 | } | ||
3620 | objInfo, err := r.Stat() | ||
3621 | if err != nil { | ||
3622 | logError(testName, function, args, startTime, "", "Unexpected error", err) | ||
3623 | return | ||
3624 | } | ||
3625 | if objInfo.Size != int64(totalSize) { | ||
3626 | logError(testName, function, args, startTime, "", "Number of bytes does not match, expected "+string(int64(totalSize))+" got "+string(objInfo.Size), err) | ||
3627 | return | ||
3628 | } | ||
3629 | if objInfo.ContentType != objectContentType && objInfo.ContentType != "application/octet-stream" { | ||
3630 | logError(testName, function, args, startTime, "", "ContentType doesn't match", err) | ||
3631 | return | ||
3632 | } | ||
3633 | |||
3634 | successLogger(testName, function, args, startTime).Info() | ||
3635 | } | ||
3636 | |||
3637 | // Tests FPutObject with null contentType (default = application/octet-stream) | ||
3638 | func testFPutObject() { | ||
3639 | // initialize logging params | ||
3640 | startTime := time.Now() | ||
3641 | testName := getFuncName() | ||
3642 | function := "FPutObject(bucketName, objectName, fileName, opts)" | ||
3643 | |||
3644 | args := map[string]interface{}{ | ||
3645 | "bucketName": "", | ||
3646 | "objectName": "", | ||
3647 | "fileName": "", | ||
3648 | "opts": "", | ||
3649 | } | ||
3650 | |||
3651 | // Seed random based on current time. | ||
3652 | rand.Seed(time.Now().Unix()) | ||
3653 | |||
3654 | // Instantiate new minio client object. | ||
3655 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
3656 | &minio.Options{ | ||
3657 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
3658 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
3659 | }) | ||
3660 | if err != nil { | ||
3661 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
3662 | return | ||
3663 | } | ||
3664 | |||
3665 | // Enable tracing, write to stderr. | ||
3666 | // c.TraceOn(os.Stderr) | ||
3667 | |||
3668 | // Set user agent. | ||
3669 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
3670 | |||
3671 | // Generate a new random bucket name. | ||
3672 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
3673 | location := "us-east-1" | ||
3674 | |||
3675 | // Make a new bucket. | ||
3676 | args["bucketName"] = bucketName | ||
3677 | args["location"] = location | ||
3678 | function = "MakeBucket(bucketName, location)" | ||
3679 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: location}) | ||
3680 | if err != nil { | ||
3681 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
3682 | return | ||
3683 | } | ||
3684 | |||
3685 | defer cleanupBucket(bucketName, c) | ||
3686 | |||
3687 | // Upload 3 parts worth of data to use all 3 of multiparts 'workers' and have an extra part. | ||
3688 | // Use different data in part for multipart tests to check parts are uploaded in correct order. | ||
3689 | fName := getMintDataDirFilePath("datafile-129-MB") | ||
3690 | if fName == "" { | ||
3691 | // Make a temp file with minPartSize bytes of data. | ||
3692 | file, err := os.CreateTemp(os.TempDir(), "FPutObjectTest") | ||
3693 | if err != nil { | ||
3694 | logError(testName, function, args, startTime, "", "TempFile creation failed", err) | ||
3695 | return | ||
3696 | } | ||
3697 | |||
3698 | // Upload 3 parts to utilize all 3 'workers' in multipart and still have a part to upload. | ||
3699 | if _, err = io.Copy(file, getDataReader("datafile-129-MB")); err != nil { | ||
3700 | logError(testName, function, args, startTime, "", "File copy failed", err) | ||
3701 | return | ||
3702 | } | ||
3703 | // Close the file pro-actively for windows. | ||
3704 | if err = file.Close(); err != nil { | ||
3705 | logError(testName, function, args, startTime, "", "File close failed", err) | ||
3706 | return | ||
3707 | } | ||
3708 | defer os.Remove(file.Name()) | ||
3709 | fName = file.Name() | ||
3710 | } | ||
3711 | |||
3712 | // Set base object name | ||
3713 | function = "FPutObject(bucketName, objectName, fileName, opts)" | ||
3714 | objectName := bucketName + "FPutObject" | ||
3715 | args["objectName"] = objectName + "-standard" | ||
3716 | args["fileName"] = fName | ||
3717 | args["opts"] = minio.PutObjectOptions{ContentType: "application/octet-stream"} | ||
3718 | |||
3719 | // Perform standard FPutObject with contentType provided (Expecting application/octet-stream) | ||
3720 | ui, err := c.FPutObject(context.Background(), bucketName, objectName+"-standard", fName, minio.PutObjectOptions{ContentType: "application/octet-stream"}) | ||
3721 | if err != nil { | ||
3722 | logError(testName, function, args, startTime, "", "FPutObject failed", err) | ||
3723 | return | ||
3724 | } | ||
3725 | |||
3726 | if ui.Size != int64(dataFileMap["datafile-129-MB"]) { | ||
3727 | logError(testName, function, args, startTime, "", "FPutObject returned an unexpected upload size", err) | ||
3728 | return | ||
3729 | } | ||
3730 | |||
3731 | // Perform FPutObject with no contentType provided (Expecting application/octet-stream) | ||
3732 | args["objectName"] = objectName + "-Octet" | ||
3733 | _, err = c.FPutObject(context.Background(), bucketName, objectName+"-Octet", fName, minio.PutObjectOptions{}) | ||
3734 | if err != nil { | ||
3735 | logError(testName, function, args, startTime, "", "File close failed", err) | ||
3736 | return | ||
3737 | } | ||
3738 | |||
3739 | srcFile, err := os.Open(fName) | ||
3740 | if err != nil { | ||
3741 | logError(testName, function, args, startTime, "", "File open failed", err) | ||
3742 | return | ||
3743 | } | ||
3744 | defer srcFile.Close() | ||
3745 | // Add extension to temp file name | ||
3746 | tmpFile, err := os.Create(fName + ".gtar") | ||
3747 | if err != nil { | ||
3748 | logError(testName, function, args, startTime, "", "File create failed", err) | ||
3749 | return | ||
3750 | } | ||
3751 | _, err = io.Copy(tmpFile, srcFile) | ||
3752 | if err != nil { | ||
3753 | logError(testName, function, args, startTime, "", "File copy failed", err) | ||
3754 | return | ||
3755 | } | ||
3756 | tmpFile.Close() | ||
3757 | |||
3758 | // Perform FPutObject with no contentType provided (Expecting application/x-gtar) | ||
3759 | args["objectName"] = objectName + "-GTar" | ||
3760 | args["opts"] = minio.PutObjectOptions{} | ||
3761 | _, err = c.FPutObject(context.Background(), bucketName, objectName+"-GTar", fName+".gtar", minio.PutObjectOptions{}) | ||
3762 | if err != nil { | ||
3763 | logError(testName, function, args, startTime, "", "FPutObject failed", err) | ||
3764 | return | ||
3765 | } | ||
3766 | |||
3767 | // Check headers | ||
3768 | function = "StatObject(bucketName, objectName, opts)" | ||
3769 | args["objectName"] = objectName + "-standard" | ||
3770 | rStandard, err := c.StatObject(context.Background(), bucketName, objectName+"-standard", minio.StatObjectOptions{}) | ||
3771 | if err != nil { | ||
3772 | logError(testName, function, args, startTime, "", "StatObject failed", err) | ||
3773 | return | ||
3774 | } | ||
3775 | if rStandard.ContentType != "application/octet-stream" { | ||
3776 | logError(testName, function, args, startTime, "", "ContentType does not match, expected application/octet-stream, got "+rStandard.ContentType, err) | ||
3777 | return | ||
3778 | } | ||
3779 | |||
3780 | function = "StatObject(bucketName, objectName, opts)" | ||
3781 | args["objectName"] = objectName + "-Octet" | ||
3782 | rOctet, err := c.StatObject(context.Background(), bucketName, objectName+"-Octet", minio.StatObjectOptions{}) | ||
3783 | if err != nil { | ||
3784 | logError(testName, function, args, startTime, "", "StatObject failed", err) | ||
3785 | return | ||
3786 | } | ||
3787 | if rOctet.ContentType != "application/octet-stream" { | ||
3788 | logError(testName, function, args, startTime, "", "ContentType does not match, expected application/octet-stream, got "+rOctet.ContentType, err) | ||
3789 | return | ||
3790 | } | ||
3791 | |||
3792 | function = "StatObject(bucketName, objectName, opts)" | ||
3793 | args["objectName"] = objectName + "-GTar" | ||
3794 | rGTar, err := c.StatObject(context.Background(), bucketName, objectName+"-GTar", minio.StatObjectOptions{}) | ||
3795 | if err != nil { | ||
3796 | logError(testName, function, args, startTime, "", "StatObject failed", err) | ||
3797 | return | ||
3798 | } | ||
3799 | if rGTar.ContentType != "application/x-gtar" && rGTar.ContentType != "application/octet-stream" && rGTar.ContentType != "application/x-tar" { | ||
3800 | logError(testName, function, args, startTime, "", "ContentType does not match, expected application/x-tar or application/octet-stream, got "+rGTar.ContentType, err) | ||
3801 | return | ||
3802 | } | ||
3803 | |||
3804 | os.Remove(fName + ".gtar") | ||
3805 | successLogger(testName, function, args, startTime).Info() | ||
3806 | } | ||
3807 | |||
3808 | // Tests FPutObject request when context cancels after timeout | ||
3809 | func testFPutObjectContext() { | ||
3810 | // initialize logging params | ||
3811 | startTime := time.Now() | ||
3812 | testName := getFuncName() | ||
3813 | function := "FPutObject(bucketName, objectName, fileName, opts)" | ||
3814 | args := map[string]interface{}{ | ||
3815 | "bucketName": "", | ||
3816 | "objectName": "", | ||
3817 | "fileName": "", | ||
3818 | "opts": "", | ||
3819 | } | ||
3820 | // Seed random based on current time. | ||
3821 | rand.Seed(time.Now().Unix()) | ||
3822 | |||
3823 | // Instantiate new minio client object. | ||
3824 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
3825 | &minio.Options{ | ||
3826 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
3827 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
3828 | }) | ||
3829 | if err != nil { | ||
3830 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
3831 | return | ||
3832 | } | ||
3833 | |||
3834 | // Enable tracing, write to stderr. | ||
3835 | // c.TraceOn(os.Stderr) | ||
3836 | |||
3837 | // Set user agent. | ||
3838 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
3839 | |||
3840 | // Generate a new random bucket name. | ||
3841 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
3842 | args["bucketName"] = bucketName | ||
3843 | |||
3844 | // Make a new bucket. | ||
3845 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
3846 | if err != nil { | ||
3847 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
3848 | return | ||
3849 | } | ||
3850 | |||
3851 | defer cleanupBucket(bucketName, c) | ||
3852 | |||
3853 | // Upload 1 parts worth of data to use multipart upload. | ||
3854 | // Use different data in part for multipart tests to check parts are uploaded in correct order. | ||
3855 | fName := getMintDataDirFilePath("datafile-1-MB") | ||
3856 | if fName == "" { | ||
3857 | // Make a temp file with 1 MiB bytes of data. | ||
3858 | file, err := os.CreateTemp(os.TempDir(), "FPutObjectContextTest") | ||
3859 | if err != nil { | ||
3860 | logError(testName, function, args, startTime, "", "TempFile creation failed", err) | ||
3861 | return | ||
3862 | } | ||
3863 | |||
3864 | // Upload 1 parts to trigger multipart upload | ||
3865 | if _, err = io.Copy(file, getDataReader("datafile-1-MB")); err != nil { | ||
3866 | logError(testName, function, args, startTime, "", "File copy failed", err) | ||
3867 | return | ||
3868 | } | ||
3869 | // Close the file pro-actively for windows. | ||
3870 | if err = file.Close(); err != nil { | ||
3871 | logError(testName, function, args, startTime, "", "File close failed", err) | ||
3872 | return | ||
3873 | } | ||
3874 | defer os.Remove(file.Name()) | ||
3875 | fName = file.Name() | ||
3876 | } | ||
3877 | |||
3878 | // Set base object name | ||
3879 | objectName := bucketName + "FPutObjectContext" | ||
3880 | args["objectName"] = objectName | ||
3881 | ctx, cancel := context.WithTimeout(context.Background(), 1*time.Nanosecond) | ||
3882 | args["ctx"] = ctx | ||
3883 | defer cancel() | ||
3884 | |||
3885 | // Perform FPutObject with contentType provided (Expecting application/octet-stream) | ||
3886 | _, err = c.FPutObject(ctx, bucketName, objectName+"-Shorttimeout", fName, minio.PutObjectOptions{ContentType: "application/octet-stream"}) | ||
3887 | if err == nil { | ||
3888 | logError(testName, function, args, startTime, "", "FPutObject should fail on short timeout", err) | ||
3889 | return | ||
3890 | } | ||
3891 | ctx, cancel = context.WithTimeout(context.Background(), 1*time.Hour) | ||
3892 | defer cancel() | ||
3893 | // Perform FPutObject with a long timeout. Expect the put object to succeed | ||
3894 | _, err = c.FPutObject(ctx, bucketName, objectName+"-Longtimeout", fName, minio.PutObjectOptions{}) | ||
3895 | if err != nil { | ||
3896 | logError(testName, function, args, startTime, "", "FPutObject shouldn't fail on long timeout", err) | ||
3897 | return | ||
3898 | } | ||
3899 | |||
3900 | _, err = c.StatObject(context.Background(), bucketName, objectName+"-Longtimeout", minio.StatObjectOptions{}) | ||
3901 | if err != nil { | ||
3902 | logError(testName, function, args, startTime, "", "StatObject failed", err) | ||
3903 | return | ||
3904 | } | ||
3905 | |||
3906 | successLogger(testName, function, args, startTime).Info() | ||
3907 | } | ||
3908 | |||
3909 | // Tests FPutObject request when context cancels after timeout | ||
3910 | func testFPutObjectContextV2() { | ||
3911 | // initialize logging params | ||
3912 | startTime := time.Now() | ||
3913 | testName := getFuncName() | ||
3914 | function := "FPutObjectContext(ctx, bucketName, objectName, fileName, opts)" | ||
3915 | args := map[string]interface{}{ | ||
3916 | "bucketName": "", | ||
3917 | "objectName": "", | ||
3918 | "opts": "minio.PutObjectOptions{ContentType:objectContentType}", | ||
3919 | } | ||
3920 | // Seed random based on current time. | ||
3921 | rand.Seed(time.Now().Unix()) | ||
3922 | |||
3923 | // Instantiate new minio client object. | ||
3924 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
3925 | &minio.Options{ | ||
3926 | Creds: credentials.NewStaticV2(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
3927 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
3928 | }) | ||
3929 | if err != nil { | ||
3930 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
3931 | return | ||
3932 | } | ||
3933 | |||
3934 | // Enable tracing, write to stderr. | ||
3935 | // c.TraceOn(os.Stderr) | ||
3936 | |||
3937 | // Set user agent. | ||
3938 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
3939 | |||
3940 | // Generate a new random bucket name. | ||
3941 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
3942 | args["bucketName"] = bucketName | ||
3943 | |||
3944 | // Make a new bucket. | ||
3945 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
3946 | if err != nil { | ||
3947 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
3948 | return | ||
3949 | } | ||
3950 | |||
3951 | defer cleanupBucket(bucketName, c) | ||
3952 | |||
3953 | // Upload 1 parts worth of data to use multipart upload. | ||
3954 | // Use different data in part for multipart tests to check parts are uploaded in correct order. | ||
3955 | fName := getMintDataDirFilePath("datafile-1-MB") | ||
3956 | if fName == "" { | ||
3957 | // Make a temp file with 1 MiB bytes of data. | ||
3958 | file, err := os.CreateTemp(os.TempDir(), "FPutObjectContextTest") | ||
3959 | if err != nil { | ||
3960 | logError(testName, function, args, startTime, "", "Temp file creation failed", err) | ||
3961 | return | ||
3962 | } | ||
3963 | |||
3964 | // Upload 1 parts to trigger multipart upload | ||
3965 | if _, err = io.Copy(file, getDataReader("datafile-1-MB")); err != nil { | ||
3966 | logError(testName, function, args, startTime, "", "File copy failed", err) | ||
3967 | return | ||
3968 | } | ||
3969 | |||
3970 | // Close the file pro-actively for windows. | ||
3971 | if err = file.Close(); err != nil { | ||
3972 | logError(testName, function, args, startTime, "", "File close failed", err) | ||
3973 | return | ||
3974 | } | ||
3975 | defer os.Remove(file.Name()) | ||
3976 | fName = file.Name() | ||
3977 | } | ||
3978 | |||
3979 | // Set base object name | ||
3980 | objectName := bucketName + "FPutObjectContext" | ||
3981 | args["objectName"] = objectName | ||
3982 | |||
3983 | ctx, cancel := context.WithTimeout(context.Background(), 1*time.Nanosecond) | ||
3984 | args["ctx"] = ctx | ||
3985 | defer cancel() | ||
3986 | |||
3987 | // Perform FPutObject with contentType provided (Expecting application/octet-stream) | ||
3988 | _, err = c.FPutObject(ctx, bucketName, objectName+"-Shorttimeout", fName, minio.PutObjectOptions{ContentType: "application/octet-stream"}) | ||
3989 | if err == nil { | ||
3990 | logError(testName, function, args, startTime, "", "FPutObject should fail on short timeout", err) | ||
3991 | return | ||
3992 | } | ||
3993 | ctx, cancel = context.WithTimeout(context.Background(), 1*time.Hour) | ||
3994 | defer cancel() | ||
3995 | // Perform FPutObject with a long timeout. Expect the put object to succeed | ||
3996 | _, err = c.FPutObject(ctx, bucketName, objectName+"-Longtimeout", fName, minio.PutObjectOptions{}) | ||
3997 | if err != nil { | ||
3998 | logError(testName, function, args, startTime, "", "FPutObject shouldn't fail on longer timeout", err) | ||
3999 | return | ||
4000 | } | ||
4001 | |||
4002 | _, err = c.StatObject(context.Background(), bucketName, objectName+"-Longtimeout", minio.StatObjectOptions{}) | ||
4003 | if err != nil { | ||
4004 | logError(testName, function, args, startTime, "", "StatObject failed", err) | ||
4005 | return | ||
4006 | } | ||
4007 | |||
4008 | successLogger(testName, function, args, startTime).Info() | ||
4009 | } | ||
4010 | |||
4011 | // Test validates putObject with context to see if request cancellation is honored. | ||
4012 | func testPutObjectContext() { | ||
4013 | // initialize logging params | ||
4014 | startTime := time.Now() | ||
4015 | testName := getFuncName() | ||
4016 | function := "PutObject(ctx, bucketName, objectName, fileName, opts)" | ||
4017 | args := map[string]interface{}{ | ||
4018 | "ctx": "", | ||
4019 | "bucketName": "", | ||
4020 | "objectName": "", | ||
4021 | "opts": "", | ||
4022 | } | ||
4023 | |||
4024 | // Instantiate new minio client object. | ||
4025 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
4026 | &minio.Options{ | ||
4027 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
4028 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
4029 | }) | ||
4030 | if err != nil { | ||
4031 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
4032 | return | ||
4033 | } | ||
4034 | |||
4035 | // Enable tracing, write to stderr. | ||
4036 | // c.TraceOn(os.Stderr) | ||
4037 | |||
4038 | // Set user agent. | ||
4039 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
4040 | |||
4041 | // Make a new bucket. | ||
4042 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
4043 | args["bucketName"] = bucketName | ||
4044 | |||
4045 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
4046 | if err != nil { | ||
4047 | logError(testName, function, args, startTime, "", "MakeBucket call failed", err) | ||
4048 | return | ||
4049 | } | ||
4050 | |||
4051 | defer cleanupBucket(bucketName, c) | ||
4052 | |||
4053 | bufSize := dataFileMap["datafile-33-kB"] | ||
4054 | reader := getDataReader("datafile-33-kB") | ||
4055 | defer reader.Close() | ||
4056 | objectName := fmt.Sprintf("test-file-%v", rand.Uint32()) | ||
4057 | args["objectName"] = objectName | ||
4058 | |||
4059 | ctx, cancel := context.WithTimeout(context.Background(), 1*time.Nanosecond) | ||
4060 | cancel() | ||
4061 | args["ctx"] = ctx | ||
4062 | args["opts"] = minio.PutObjectOptions{ContentType: "binary/octet-stream"} | ||
4063 | |||
4064 | _, err = c.PutObject(ctx, bucketName, objectName, reader, int64(bufSize), minio.PutObjectOptions{ContentType: "binary/octet-stream"}) | ||
4065 | if err == nil { | ||
4066 | logError(testName, function, args, startTime, "", "PutObject should fail on short timeout", err) | ||
4067 | return | ||
4068 | } | ||
4069 | |||
4070 | ctx, cancel = context.WithTimeout(context.Background(), 1*time.Hour) | ||
4071 | args["ctx"] = ctx | ||
4072 | |||
4073 | defer cancel() | ||
4074 | reader = getDataReader("datafile-33-kB") | ||
4075 | defer reader.Close() | ||
4076 | _, err = c.PutObject(ctx, bucketName, objectName, reader, int64(bufSize), minio.PutObjectOptions{ContentType: "binary/octet-stream"}) | ||
4077 | if err != nil { | ||
4078 | logError(testName, function, args, startTime, "", "PutObject with long timeout failed", err) | ||
4079 | return | ||
4080 | } | ||
4081 | |||
4082 | successLogger(testName, function, args, startTime).Info() | ||
4083 | } | ||
4084 | |||
4085 | // Tests get object with s3zip extensions. | ||
4086 | func testGetObjectS3Zip() { | ||
4087 | // initialize logging params | ||
4088 | startTime := time.Now() | ||
4089 | testName := getFuncName() | ||
4090 | function := "GetObject(bucketName, objectName)" | ||
4091 | args := map[string]interface{}{"x-minio-extract": true} | ||
4092 | |||
4093 | // Seed random based on current time. | ||
4094 | rand.Seed(time.Now().Unix()) | ||
4095 | |||
4096 | // Instantiate new minio client object. | ||
4097 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
4098 | &minio.Options{ | ||
4099 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
4100 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
4101 | }) | ||
4102 | if err != nil { | ||
4103 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
4104 | return | ||
4105 | } | ||
4106 | |||
4107 | // Enable tracing, write to stderr. | ||
4108 | // c.TraceOn(os.Stderr) | ||
4109 | |||
4110 | // Set user agent. | ||
4111 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
4112 | |||
4113 | // Generate a new random bucket name. | ||
4114 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
4115 | args["bucketName"] = bucketName | ||
4116 | |||
4117 | // Make a new bucket. | ||
4118 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
4119 | if err != nil { | ||
4120 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
4121 | return | ||
4122 | } | ||
4123 | |||
4124 | defer func() { | ||
4125 | // Delete all objects and buckets | ||
4126 | if err = cleanupBucket(bucketName, c); err != nil { | ||
4127 | logError(testName, function, args, startTime, "", "CleanupBucket failed", err) | ||
4128 | return | ||
4129 | } | ||
4130 | }() | ||
4131 | |||
4132 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") + ".zip" | ||
4133 | args["objectName"] = objectName | ||
4134 | |||
4135 | var zipFile bytes.Buffer | ||
4136 | zw := zip.NewWriter(&zipFile) | ||
4137 | rng := rand.New(rand.NewSource(0xc0cac01a)) | ||
4138 | const nFiles = 500 | ||
4139 | for i := 0; i <= nFiles; i++ { | ||
4140 | if i == nFiles { | ||
4141 | // Make one large, compressible file. | ||
4142 | i = 1000000 | ||
4143 | } | ||
4144 | b := make([]byte, i) | ||
4145 | if i < nFiles { | ||
4146 | rng.Read(b) | ||
4147 | } | ||
4148 | wc, err := zw.Create(fmt.Sprintf("test/small/file-%d.bin", i)) | ||
4149 | if err != nil { | ||
4150 | logError(testName, function, args, startTime, "", "zw.Create failed", err) | ||
4151 | return | ||
4152 | } | ||
4153 | wc.Write(b) | ||
4154 | } | ||
4155 | err = zw.Close() | ||
4156 | if err != nil { | ||
4157 | logError(testName, function, args, startTime, "", "zw.Close failed", err) | ||
4158 | return | ||
4159 | } | ||
4160 | buf := zipFile.Bytes() | ||
4161 | |||
4162 | // Save the data | ||
4163 | _, err = c.PutObject(context.Background(), bucketName, objectName, bytes.NewReader(buf), int64(len(buf)), minio.PutObjectOptions{ContentType: "binary/octet-stream"}) | ||
4164 | if err != nil { | ||
4165 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
4166 | return | ||
4167 | } | ||
4168 | |||
4169 | // Read the data back | ||
4170 | r, err := c.GetObject(context.Background(), bucketName, objectName, minio.GetObjectOptions{}) | ||
4171 | if err != nil { | ||
4172 | logError(testName, function, args, startTime, "", "GetObject failed", err) | ||
4173 | return | ||
4174 | } | ||
4175 | |||
4176 | st, err := r.Stat() | ||
4177 | if err != nil { | ||
4178 | logError(testName, function, args, startTime, "", "Stat object failed", err) | ||
4179 | return | ||
4180 | } | ||
4181 | |||
4182 | if st.Size != int64(len(buf)) { | ||
4183 | logError(testName, function, args, startTime, "", "Number of bytes does not match, expected "+string(len(buf))+", got "+string(st.Size), err) | ||
4184 | return | ||
4185 | } | ||
4186 | r.Close() | ||
4187 | |||
4188 | zr, err := zip.NewReader(bytes.NewReader(buf), int64(len(buf))) | ||
4189 | if err != nil { | ||
4190 | logError(testName, function, args, startTime, "", "zip.NewReader failed", err) | ||
4191 | return | ||
4192 | } | ||
4193 | lOpts := minio.ListObjectsOptions{} | ||
4194 | lOpts.Set("x-minio-extract", "true") | ||
4195 | lOpts.Prefix = objectName + "/" | ||
4196 | lOpts.Recursive = true | ||
4197 | list := c.ListObjects(context.Background(), bucketName, lOpts) | ||
4198 | listed := map[string]minio.ObjectInfo{} | ||
4199 | for item := range list { | ||
4200 | if item.Err != nil { | ||
4201 | break | ||
4202 | } | ||
4203 | listed[item.Key] = item | ||
4204 | } | ||
4205 | if len(listed) == 0 { | ||
4206 | // Assume we are running against non-minio. | ||
4207 | args["SKIPPED"] = true | ||
4208 | ignoredLog(testName, function, args, startTime, "s3zip does not appear to be present").Info() | ||
4209 | return | ||
4210 | } | ||
4211 | |||
4212 | for _, file := range zr.File { | ||
4213 | if file.FileInfo().IsDir() { | ||
4214 | continue | ||
4215 | } | ||
4216 | args["zipfile"] = file.Name | ||
4217 | zfr, err := file.Open() | ||
4218 | if err != nil { | ||
4219 | logError(testName, function, args, startTime, "", "file.Open failed", err) | ||
4220 | return | ||
4221 | } | ||
4222 | want, err := io.ReadAll(zfr) | ||
4223 | if err != nil { | ||
4224 | logError(testName, function, args, startTime, "", "fzip file read failed", err) | ||
4225 | return | ||
4226 | } | ||
4227 | |||
4228 | opts := minio.GetObjectOptions{} | ||
4229 | opts.Set("x-minio-extract", "true") | ||
4230 | key := path.Join(objectName, file.Name) | ||
4231 | r, err = c.GetObject(context.Background(), bucketName, key, opts) | ||
4232 | if err != nil { | ||
4233 | terr := minio.ToErrorResponse(err) | ||
4234 | if terr.StatusCode != http.StatusNotFound { | ||
4235 | logError(testName, function, args, startTime, "", "GetObject failed", err) | ||
4236 | } | ||
4237 | return | ||
4238 | } | ||
4239 | got, err := io.ReadAll(r) | ||
4240 | if err != nil { | ||
4241 | logError(testName, function, args, startTime, "", "ReadAll failed", err) | ||
4242 | return | ||
4243 | } | ||
4244 | r.Close() | ||
4245 | if !bytes.Equal(want, got) { | ||
4246 | logError(testName, function, args, startTime, "", "Content mismatch", err) | ||
4247 | return | ||
4248 | } | ||
4249 | oi, ok := listed[key] | ||
4250 | if !ok { | ||
4251 | logError(testName, function, args, startTime, "", "Object Missing", fmt.Errorf("%s not present in listing", key)) | ||
4252 | return | ||
4253 | } | ||
4254 | if int(oi.Size) != len(got) { | ||
4255 | logError(testName, function, args, startTime, "", "Object Size Incorrect", fmt.Errorf("listing %d, read %d", oi.Size, len(got))) | ||
4256 | return | ||
4257 | } | ||
4258 | delete(listed, key) | ||
4259 | } | ||
4260 | delete(args, "zipfile") | ||
4261 | if len(listed) > 0 { | ||
4262 | logError(testName, function, args, startTime, "", "Extra listed objects", fmt.Errorf("left over: %v", listed)) | ||
4263 | return | ||
4264 | } | ||
4265 | successLogger(testName, function, args, startTime).Info() | ||
4266 | } | ||
4267 | |||
4268 | // Tests get object ReaderSeeker interface methods. | ||
4269 | func testGetObjectReadSeekFunctional() { | ||
4270 | // initialize logging params | ||
4271 | startTime := time.Now() | ||
4272 | testName := getFuncName() | ||
4273 | function := "GetObject(bucketName, objectName)" | ||
4274 | args := map[string]interface{}{} | ||
4275 | |||
4276 | // Seed random based on current time. | ||
4277 | rand.Seed(time.Now().Unix()) | ||
4278 | |||
4279 | // Instantiate new minio client object. | ||
4280 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
4281 | &minio.Options{ | ||
4282 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
4283 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
4284 | }) | ||
4285 | if err != nil { | ||
4286 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
4287 | return | ||
4288 | } | ||
4289 | |||
4290 | // Enable tracing, write to stderr. | ||
4291 | // c.TraceOn(os.Stderr) | ||
4292 | |||
4293 | // Set user agent. | ||
4294 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
4295 | |||
4296 | // Generate a new random bucket name. | ||
4297 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
4298 | args["bucketName"] = bucketName | ||
4299 | |||
4300 | // Make a new bucket. | ||
4301 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
4302 | if err != nil { | ||
4303 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
4304 | return | ||
4305 | } | ||
4306 | |||
4307 | defer func() { | ||
4308 | // Delete all objects and buckets | ||
4309 | if err = cleanupBucket(bucketName, c); err != nil { | ||
4310 | logError(testName, function, args, startTime, "", "CleanupBucket failed", err) | ||
4311 | return | ||
4312 | } | ||
4313 | }() | ||
4314 | |||
4315 | // Generate 33K of data. | ||
4316 | bufSize := dataFileMap["datafile-33-kB"] | ||
4317 | reader := getDataReader("datafile-33-kB") | ||
4318 | defer reader.Close() | ||
4319 | |||
4320 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
4321 | args["objectName"] = objectName | ||
4322 | |||
4323 | buf, err := io.ReadAll(reader) | ||
4324 | if err != nil { | ||
4325 | logError(testName, function, args, startTime, "", "ReadAll failed", err) | ||
4326 | return | ||
4327 | } | ||
4328 | |||
4329 | // Save the data | ||
4330 | _, err = c.PutObject(context.Background(), bucketName, objectName, bytes.NewReader(buf), int64(len(buf)), minio.PutObjectOptions{ContentType: "binary/octet-stream"}) | ||
4331 | if err != nil { | ||
4332 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
4333 | return | ||
4334 | } | ||
4335 | |||
4336 | // Read the data back | ||
4337 | r, err := c.GetObject(context.Background(), bucketName, objectName, minio.GetObjectOptions{}) | ||
4338 | if err != nil { | ||
4339 | logError(testName, function, args, startTime, "", "GetObject failed", err) | ||
4340 | return | ||
4341 | } | ||
4342 | |||
4343 | st, err := r.Stat() | ||
4344 | if err != nil { | ||
4345 | logError(testName, function, args, startTime, "", "Stat object failed", err) | ||
4346 | return | ||
4347 | } | ||
4348 | |||
4349 | if st.Size != int64(bufSize) { | ||
4350 | logError(testName, function, args, startTime, "", "Number of bytes does not match, expected "+string(int64(bufSize))+", got "+string(st.Size), err) | ||
4351 | return | ||
4352 | } | ||
4353 | |||
4354 | // This following function helps us to compare data from the reader after seek | ||
4355 | // with the data from the original buffer | ||
4356 | cmpData := func(r io.Reader, start, end int) { | ||
4357 | if end-start == 0 { | ||
4358 | return | ||
4359 | } | ||
4360 | buffer := bytes.NewBuffer([]byte{}) | ||
4361 | if _, err := io.CopyN(buffer, r, int64(bufSize)); err != nil { | ||
4362 | if err != io.EOF { | ||
4363 | logError(testName, function, args, startTime, "", "CopyN failed", err) | ||
4364 | return | ||
4365 | } | ||
4366 | } | ||
4367 | if !bytes.Equal(buf[start:end], buffer.Bytes()) { | ||
4368 | logError(testName, function, args, startTime, "", "Incorrect read bytes v/s original buffer", err) | ||
4369 | return | ||
4370 | } | ||
4371 | } | ||
4372 | |||
4373 | // Generic seek error for errors other than io.EOF | ||
4374 | seekErr := errors.New("seek error") | ||
4375 | |||
4376 | testCases := []struct { | ||
4377 | offset int64 | ||
4378 | whence int | ||
4379 | pos int64 | ||
4380 | err error | ||
4381 | shouldCmp bool | ||
4382 | start int | ||
4383 | end int | ||
4384 | }{ | ||
4385 | // Start from offset 0, fetch data and compare | ||
4386 | {0, 0, 0, nil, true, 0, 0}, | ||
4387 | // Start from offset 2048, fetch data and compare | ||
4388 | {2048, 0, 2048, nil, true, 2048, bufSize}, | ||
4389 | // Start from offset larger than possible | ||
4390 | {int64(bufSize) + 1024, 0, 0, seekErr, false, 0, 0}, | ||
4391 | // Move to offset 0 without comparing | ||
4392 | {0, 0, 0, nil, false, 0, 0}, | ||
4393 | // Move one step forward and compare | ||
4394 | {1, 1, 1, nil, true, 1, bufSize}, | ||
4395 | // Move larger than possible | ||
4396 | {int64(bufSize), 1, 0, seekErr, false, 0, 0}, | ||
4397 | // Provide negative offset with CUR_SEEK | ||
4398 | {int64(-1), 1, 0, seekErr, false, 0, 0}, | ||
4399 | // Test with whence SEEK_END and with positive offset | ||
4400 | {1024, 2, int64(bufSize) - 1024, io.EOF, true, 0, 0}, | ||
4401 | // Test with whence SEEK_END and with negative offset | ||
4402 | {-1024, 2, int64(bufSize) - 1024, nil, true, bufSize - 1024, bufSize}, | ||
4403 | // Test with whence SEEK_END and with large negative offset | ||
4404 | {-int64(bufSize) * 2, 2, 0, seekErr, true, 0, 0}, | ||
4405 | } | ||
4406 | |||
4407 | for i, testCase := range testCases { | ||
4408 | // Perform seek operation | ||
4409 | n, err := r.Seek(testCase.offset, testCase.whence) | ||
4410 | // We expect an error | ||
4411 | if testCase.err == seekErr && err == nil { | ||
4412 | logError(testName, function, args, startTime, "", "Test "+string(i+1)+", unexpected err value: expected: "+testCase.err.Error()+", found: "+err.Error(), err) | ||
4413 | return | ||
4414 | } | ||
4415 | // We expect a specific error | ||
4416 | if testCase.err != seekErr && testCase.err != err { | ||
4417 | logError(testName, function, args, startTime, "", "Test "+string(i+1)+", unexpected err value: expected: "+testCase.err.Error()+", found: "+err.Error(), err) | ||
4418 | return | ||
4419 | } | ||
4420 | // If we expect an error go to the next loop | ||
4421 | if testCase.err != nil { | ||
4422 | continue | ||
4423 | } | ||
4424 | // Check the returned seek pos | ||
4425 | if n != testCase.pos { | ||
4426 | logError(testName, function, args, startTime, "", "Test "+string(i+1)+", number of bytes seeked does not match, expected "+string(testCase.pos)+", got "+string(n), err) | ||
4427 | return | ||
4428 | } | ||
4429 | // Compare only if shouldCmp is activated | ||
4430 | if testCase.shouldCmp { | ||
4431 | cmpData(r, testCase.start, testCase.end) | ||
4432 | } | ||
4433 | } | ||
4434 | successLogger(testName, function, args, startTime).Info() | ||
4435 | } | ||
4436 | |||
4437 | // Tests get object ReaderAt interface methods. | ||
4438 | func testGetObjectReadAtFunctional() { | ||
4439 | // initialize logging params | ||
4440 | startTime := time.Now() | ||
4441 | testName := getFuncName() | ||
4442 | function := "GetObject(bucketName, objectName)" | ||
4443 | args := map[string]interface{}{} | ||
4444 | |||
4445 | // Seed random based on current time. | ||
4446 | rand.Seed(time.Now().Unix()) | ||
4447 | |||
4448 | // Instantiate new minio client object. | ||
4449 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
4450 | &minio.Options{ | ||
4451 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
4452 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
4453 | }) | ||
4454 | if err != nil { | ||
4455 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
4456 | return | ||
4457 | } | ||
4458 | |||
4459 | // Enable tracing, write to stderr. | ||
4460 | // c.TraceOn(os.Stderr) | ||
4461 | |||
4462 | // Set user agent. | ||
4463 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
4464 | |||
4465 | // Generate a new random bucket name. | ||
4466 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
4467 | args["bucketName"] = bucketName | ||
4468 | |||
4469 | // Make a new bucket. | ||
4470 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
4471 | if err != nil { | ||
4472 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
4473 | return | ||
4474 | } | ||
4475 | |||
4476 | defer cleanupBucket(bucketName, c) | ||
4477 | |||
4478 | // Generate 33K of data. | ||
4479 | bufSize := dataFileMap["datafile-33-kB"] | ||
4480 | reader := getDataReader("datafile-33-kB") | ||
4481 | defer reader.Close() | ||
4482 | |||
4483 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
4484 | args["objectName"] = objectName | ||
4485 | |||
4486 | buf, err := io.ReadAll(reader) | ||
4487 | if err != nil { | ||
4488 | logError(testName, function, args, startTime, "", "ReadAll failed", err) | ||
4489 | return | ||
4490 | } | ||
4491 | |||
4492 | // Save the data | ||
4493 | _, err = c.PutObject(context.Background(), bucketName, objectName, bytes.NewReader(buf), int64(len(buf)), minio.PutObjectOptions{ContentType: "binary/octet-stream"}) | ||
4494 | if err != nil { | ||
4495 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
4496 | return | ||
4497 | } | ||
4498 | |||
4499 | // read the data back | ||
4500 | r, err := c.GetObject(context.Background(), bucketName, objectName, minio.GetObjectOptions{}) | ||
4501 | if err != nil { | ||
4502 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
4503 | return | ||
4504 | } | ||
4505 | offset := int64(2048) | ||
4506 | |||
4507 | // read directly | ||
4508 | buf1 := make([]byte, 512) | ||
4509 | buf2 := make([]byte, 512) | ||
4510 | buf3 := make([]byte, 512) | ||
4511 | buf4 := make([]byte, 512) | ||
4512 | |||
4513 | // Test readAt before stat is called such that objectInfo doesn't change. | ||
4514 | m, err := r.ReadAt(buf1, offset) | ||
4515 | if err != nil { | ||
4516 | logError(testName, function, args, startTime, "", "ReadAt failed", err) | ||
4517 | return | ||
4518 | } | ||
4519 | if m != len(buf1) { | ||
4520 | logError(testName, function, args, startTime, "", "ReadAt read shorter bytes before reaching EOF, expected "+string(len(buf1))+", got "+string(m), err) | ||
4521 | return | ||
4522 | } | ||
4523 | if !bytes.Equal(buf1, buf[offset:offset+512]) { | ||
4524 | logError(testName, function, args, startTime, "", "Incorrect read between two ReadAt from same offset", err) | ||
4525 | return | ||
4526 | } | ||
4527 | offset += 512 | ||
4528 | |||
4529 | st, err := r.Stat() | ||
4530 | if err != nil { | ||
4531 | logError(testName, function, args, startTime, "", "Stat failed", err) | ||
4532 | return | ||
4533 | } | ||
4534 | |||
4535 | if st.Size != int64(bufSize) { | ||
4536 | logError(testName, function, args, startTime, "", "Number of bytes in stat does not match, expected "+string(int64(bufSize))+", got "+string(st.Size), err) | ||
4537 | return | ||
4538 | } | ||
4539 | |||
4540 | m, err = r.ReadAt(buf2, offset) | ||
4541 | if err != nil { | ||
4542 | logError(testName, function, args, startTime, "", "ReadAt failed", err) | ||
4543 | return | ||
4544 | } | ||
4545 | if m != len(buf2) { | ||
4546 | logError(testName, function, args, startTime, "", "ReadAt read shorter bytes before reaching EOF, expected "+string(len(buf2))+", got "+string(m), err) | ||
4547 | return | ||
4548 | } | ||
4549 | if !bytes.Equal(buf2, buf[offset:offset+512]) { | ||
4550 | logError(testName, function, args, startTime, "", "Incorrect read between two ReadAt from same offset", err) | ||
4551 | return | ||
4552 | } | ||
4553 | |||
4554 | offset += 512 | ||
4555 | m, err = r.ReadAt(buf3, offset) | ||
4556 | if err != nil { | ||
4557 | logError(testName, function, args, startTime, "", "ReadAt failed", err) | ||
4558 | return | ||
4559 | } | ||
4560 | if m != len(buf3) { | ||
4561 | logError(testName, function, args, startTime, "", "ReadAt read shorter bytes before reaching EOF, expected "+string(len(buf3))+", got "+string(m), err) | ||
4562 | return | ||
4563 | } | ||
4564 | if !bytes.Equal(buf3, buf[offset:offset+512]) { | ||
4565 | logError(testName, function, args, startTime, "", "Incorrect read between two ReadAt from same offset", err) | ||
4566 | return | ||
4567 | } | ||
4568 | offset += 512 | ||
4569 | m, err = r.ReadAt(buf4, offset) | ||
4570 | if err != nil { | ||
4571 | logError(testName, function, args, startTime, "", "ReadAt failed", err) | ||
4572 | return | ||
4573 | } | ||
4574 | if m != len(buf4) { | ||
4575 | logError(testName, function, args, startTime, "", "ReadAt read shorter bytes before reaching EOF, expected "+string(len(buf4))+", got "+string(m), err) | ||
4576 | return | ||
4577 | } | ||
4578 | if !bytes.Equal(buf4, buf[offset:offset+512]) { | ||
4579 | logError(testName, function, args, startTime, "", "Incorrect read between two ReadAt from same offset", err) | ||
4580 | return | ||
4581 | } | ||
4582 | |||
4583 | buf5 := make([]byte, len(buf)) | ||
4584 | // Read the whole object. | ||
4585 | m, err = r.ReadAt(buf5, 0) | ||
4586 | if err != nil { | ||
4587 | if err != io.EOF { | ||
4588 | logError(testName, function, args, startTime, "", "ReadAt failed", err) | ||
4589 | return | ||
4590 | } | ||
4591 | } | ||
4592 | if m != len(buf5) { | ||
4593 | logError(testName, function, args, startTime, "", "ReadAt read shorter bytes before reaching EOF, expected "+string(len(buf5))+", got "+string(m), err) | ||
4594 | return | ||
4595 | } | ||
4596 | if !bytes.Equal(buf, buf5) { | ||
4597 | logError(testName, function, args, startTime, "", "Incorrect data read in GetObject, than what was previously uploaded", err) | ||
4598 | return | ||
4599 | } | ||
4600 | |||
4601 | buf6 := make([]byte, len(buf)+1) | ||
4602 | // Read the whole object and beyond. | ||
4603 | _, err = r.ReadAt(buf6, 0) | ||
4604 | if err != nil { | ||
4605 | if err != io.EOF { | ||
4606 | logError(testName, function, args, startTime, "", "ReadAt failed", err) | ||
4607 | return | ||
4608 | } | ||
4609 | } | ||
4610 | |||
4611 | successLogger(testName, function, args, startTime).Info() | ||
4612 | } | ||
4613 | |||
4614 | // Reproduces issue https://github.com/minio/minio-go/issues/1137 | ||
4615 | func testGetObjectReadAtWhenEOFWasReached() { | ||
4616 | // initialize logging params | ||
4617 | startTime := time.Now() | ||
4618 | testName := getFuncName() | ||
4619 | function := "GetObject(bucketName, objectName)" | ||
4620 | args := map[string]interface{}{} | ||
4621 | |||
4622 | // Seed random based on current time. | ||
4623 | rand.Seed(time.Now().Unix()) | ||
4624 | |||
4625 | // Instantiate new minio client object. | ||
4626 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
4627 | &minio.Options{ | ||
4628 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
4629 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
4630 | }) | ||
4631 | if err != nil { | ||
4632 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
4633 | return | ||
4634 | } | ||
4635 | |||
4636 | // Enable tracing, write to stderr. | ||
4637 | // c.TraceOn(os.Stderr) | ||
4638 | |||
4639 | // Set user agent. | ||
4640 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
4641 | |||
4642 | // Generate a new random bucket name. | ||
4643 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
4644 | args["bucketName"] = bucketName | ||
4645 | |||
4646 | // Make a new bucket. | ||
4647 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
4648 | if err != nil { | ||
4649 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
4650 | return | ||
4651 | } | ||
4652 | |||
4653 | defer cleanupBucket(bucketName, c) | ||
4654 | |||
4655 | // Generate 33K of data. | ||
4656 | bufSize := dataFileMap["datafile-33-kB"] | ||
4657 | reader := getDataReader("datafile-33-kB") | ||
4658 | defer reader.Close() | ||
4659 | |||
4660 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
4661 | args["objectName"] = objectName | ||
4662 | |||
4663 | buf, err := io.ReadAll(reader) | ||
4664 | if err != nil { | ||
4665 | logError(testName, function, args, startTime, "", "ReadAll failed", err) | ||
4666 | return | ||
4667 | } | ||
4668 | |||
4669 | // Save the data | ||
4670 | _, err = c.PutObject(context.Background(), bucketName, objectName, bytes.NewReader(buf), int64(len(buf)), minio.PutObjectOptions{ContentType: "binary/octet-stream"}) | ||
4671 | if err != nil { | ||
4672 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
4673 | return | ||
4674 | } | ||
4675 | |||
4676 | // read the data back | ||
4677 | r, err := c.GetObject(context.Background(), bucketName, objectName, minio.GetObjectOptions{}) | ||
4678 | if err != nil { | ||
4679 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
4680 | return | ||
4681 | } | ||
4682 | |||
4683 | // read directly | ||
4684 | buf1 := make([]byte, len(buf)) | ||
4685 | buf2 := make([]byte, 512) | ||
4686 | |||
4687 | m, err := r.Read(buf1) | ||
4688 | if err != nil { | ||
4689 | if err != io.EOF { | ||
4690 | logError(testName, function, args, startTime, "", "Read failed", err) | ||
4691 | return | ||
4692 | } | ||
4693 | } | ||
4694 | if m != len(buf1) { | ||
4695 | logError(testName, function, args, startTime, "", "Read read shorter bytes before reaching EOF, expected "+string(len(buf1))+", got "+string(m), err) | ||
4696 | return | ||
4697 | } | ||
4698 | if !bytes.Equal(buf1, buf) { | ||
4699 | logError(testName, function, args, startTime, "", "Incorrect count of Read data", err) | ||
4700 | return | ||
4701 | } | ||
4702 | |||
4703 | st, err := r.Stat() | ||
4704 | if err != nil { | ||
4705 | logError(testName, function, args, startTime, "", "Stat failed", err) | ||
4706 | return | ||
4707 | } | ||
4708 | |||
4709 | if st.Size != int64(bufSize) { | ||
4710 | logError(testName, function, args, startTime, "", "Number of bytes in stat does not match, expected "+string(int64(bufSize))+", got "+string(st.Size), err) | ||
4711 | return | ||
4712 | } | ||
4713 | |||
4714 | m, err = r.ReadAt(buf2, 512) | ||
4715 | if err != nil { | ||
4716 | logError(testName, function, args, startTime, "", "ReadAt failed", err) | ||
4717 | return | ||
4718 | } | ||
4719 | if m != len(buf2) { | ||
4720 | logError(testName, function, args, startTime, "", "ReadAt read shorter bytes before reaching EOF, expected "+string(len(buf2))+", got "+string(m), err) | ||
4721 | return | ||
4722 | } | ||
4723 | if !bytes.Equal(buf2, buf[512:1024]) { | ||
4724 | logError(testName, function, args, startTime, "", "Incorrect count of ReadAt data", err) | ||
4725 | return | ||
4726 | } | ||
4727 | |||
4728 | successLogger(testName, function, args, startTime).Info() | ||
4729 | } | ||
4730 | |||
4731 | // Test Presigned Post Policy | ||
4732 | func testPresignedPostPolicy() { | ||
4733 | // initialize logging params | ||
4734 | startTime := time.Now() | ||
4735 | testName := getFuncName() | ||
4736 | function := "PresignedPostPolicy(policy)" | ||
4737 | args := map[string]interface{}{ | ||
4738 | "policy": "", | ||
4739 | } | ||
4740 | |||
4741 | // Seed random based on current time. | ||
4742 | rand.Seed(time.Now().Unix()) | ||
4743 | |||
4744 | // Instantiate new minio client object | ||
4745 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
4746 | &minio.Options{ | ||
4747 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
4748 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
4749 | }) | ||
4750 | if err != nil { | ||
4751 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
4752 | return | ||
4753 | } | ||
4754 | |||
4755 | // Enable tracing, write to stderr. | ||
4756 | // c.TraceOn(os.Stderr) | ||
4757 | |||
4758 | // Set user agent. | ||
4759 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
4760 | |||
4761 | // Generate a new random bucket name. | ||
4762 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
4763 | |||
4764 | // Make a new bucket in 'us-east-1' (source bucket). | ||
4765 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
4766 | if err != nil { | ||
4767 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
4768 | return | ||
4769 | } | ||
4770 | |||
4771 | defer cleanupBucket(bucketName, c) | ||
4772 | |||
4773 | // Generate 33K of data. | ||
4774 | reader := getDataReader("datafile-33-kB") | ||
4775 | defer reader.Close() | ||
4776 | |||
4777 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
4778 | // Azure requires the key to not start with a number | ||
4779 | metadataKey := randString(60, rand.NewSource(time.Now().UnixNano()), "user") | ||
4780 | metadataValue := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
4781 | |||
4782 | buf, err := io.ReadAll(reader) | ||
4783 | if err != nil { | ||
4784 | logError(testName, function, args, startTime, "", "ReadAll failed", err) | ||
4785 | return | ||
4786 | } | ||
4787 | |||
4788 | // Save the data | ||
4789 | _, err = c.PutObject(context.Background(), bucketName, objectName, bytes.NewReader(buf), int64(len(buf)), minio.PutObjectOptions{ContentType: "binary/octet-stream"}) | ||
4790 | if err != nil { | ||
4791 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
4792 | return | ||
4793 | } | ||
4794 | |||
4795 | policy := minio.NewPostPolicy() | ||
4796 | |||
4797 | if err := policy.SetBucket(""); err == nil { | ||
4798 | logError(testName, function, args, startTime, "", "SetBucket did not fail for invalid conditions", err) | ||
4799 | return | ||
4800 | } | ||
4801 | if err := policy.SetKey(""); err == nil { | ||
4802 | logError(testName, function, args, startTime, "", "SetKey did not fail for invalid conditions", err) | ||
4803 | return | ||
4804 | } | ||
4805 | if err := policy.SetExpires(time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC)); err == nil { | ||
4806 | logError(testName, function, args, startTime, "", "SetExpires did not fail for invalid conditions", err) | ||
4807 | return | ||
4808 | } | ||
4809 | if err := policy.SetContentType(""); err == nil { | ||
4810 | logError(testName, function, args, startTime, "", "SetContentType did not fail for invalid conditions", err) | ||
4811 | return | ||
4812 | } | ||
4813 | if err := policy.SetContentLengthRange(1024*1024, 1024); err == nil { | ||
4814 | logError(testName, function, args, startTime, "", "SetContentLengthRange did not fail for invalid conditions", err) | ||
4815 | return | ||
4816 | } | ||
4817 | if err := policy.SetUserMetadata("", ""); err == nil { | ||
4818 | logError(testName, function, args, startTime, "", "SetUserMetadata did not fail for invalid conditions", err) | ||
4819 | return | ||
4820 | } | ||
4821 | |||
4822 | policy.SetBucket(bucketName) | ||
4823 | policy.SetKey(objectName) | ||
4824 | policy.SetExpires(time.Now().UTC().AddDate(0, 0, 10)) // expires in 10 days | ||
4825 | policy.SetContentType("binary/octet-stream") | ||
4826 | policy.SetContentLengthRange(10, 1024*1024) | ||
4827 | policy.SetUserMetadata(metadataKey, metadataValue) | ||
4828 | |||
4829 | // Add CRC32C | ||
4830 | checksum := minio.ChecksumCRC32C.ChecksumBytes(buf) | ||
4831 | policy.SetChecksum(checksum) | ||
4832 | |||
4833 | args["policy"] = policy.String() | ||
4834 | |||
4835 | presignedPostPolicyURL, formData, err := c.PresignedPostPolicy(context.Background(), policy) | ||
4836 | if err != nil { | ||
4837 | logError(testName, function, args, startTime, "", "PresignedPostPolicy failed", err) | ||
4838 | return | ||
4839 | } | ||
4840 | |||
4841 | var formBuf bytes.Buffer | ||
4842 | writer := multipart.NewWriter(&formBuf) | ||
4843 | for k, v := range formData { | ||
4844 | writer.WriteField(k, v) | ||
4845 | } | ||
4846 | |||
4847 | // Get a 33KB file to upload and test if set post policy works | ||
4848 | filePath := getMintDataDirFilePath("datafile-33-kB") | ||
4849 | if filePath == "" { | ||
4850 | // Make a temp file with 33 KB data. | ||
4851 | file, err := os.CreateTemp(os.TempDir(), "PresignedPostPolicyTest") | ||
4852 | if err != nil { | ||
4853 | logError(testName, function, args, startTime, "", "TempFile creation failed", err) | ||
4854 | return | ||
4855 | } | ||
4856 | if _, err = io.Copy(file, getDataReader("datafile-33-kB")); err != nil { | ||
4857 | logError(testName, function, args, startTime, "", "Copy failed", err) | ||
4858 | return | ||
4859 | } | ||
4860 | if err = file.Close(); err != nil { | ||
4861 | logError(testName, function, args, startTime, "", "File Close failed", err) | ||
4862 | return | ||
4863 | } | ||
4864 | filePath = file.Name() | ||
4865 | } | ||
4866 | |||
4867 | // add file to post request | ||
4868 | f, err := os.Open(filePath) | ||
4869 | defer f.Close() | ||
4870 | if err != nil { | ||
4871 | logError(testName, function, args, startTime, "", "File open failed", err) | ||
4872 | return | ||
4873 | } | ||
4874 | w, err := writer.CreateFormFile("file", filePath) | ||
4875 | if err != nil { | ||
4876 | logError(testName, function, args, startTime, "", "CreateFormFile failed", err) | ||
4877 | return | ||
4878 | } | ||
4879 | |||
4880 | _, err = io.Copy(w, f) | ||
4881 | if err != nil { | ||
4882 | logError(testName, function, args, startTime, "", "Copy failed", err) | ||
4883 | return | ||
4884 | } | ||
4885 | writer.Close() | ||
4886 | |||
4887 | transport, err := minio.DefaultTransport(mustParseBool(os.Getenv(enableHTTPS))) | ||
4888 | if err != nil { | ||
4889 | logError(testName, function, args, startTime, "", "DefaultTransport failed", err) | ||
4890 | return | ||
4891 | } | ||
4892 | |||
4893 | httpClient := &http.Client{ | ||
4894 | // Setting a sensible time out of 30secs to wait for response | ||
4895 | // headers. Request is pro-actively canceled after 30secs | ||
4896 | // with no response. | ||
4897 | Timeout: 30 * time.Second, | ||
4898 | Transport: transport, | ||
4899 | } | ||
4900 | args["url"] = presignedPostPolicyURL.String() | ||
4901 | |||
4902 | req, err := http.NewRequest(http.MethodPost, presignedPostPolicyURL.String(), bytes.NewReader(formBuf.Bytes())) | ||
4903 | if err != nil { | ||
4904 | logError(testName, function, args, startTime, "", "Http request failed", err) | ||
4905 | return | ||
4906 | } | ||
4907 | |||
4908 | req.Header.Set("Content-Type", writer.FormDataContentType()) | ||
4909 | |||
4910 | // make post request with correct form data | ||
4911 | res, err := httpClient.Do(req) | ||
4912 | if err != nil { | ||
4913 | logError(testName, function, args, startTime, "", "Http request failed", err) | ||
4914 | return | ||
4915 | } | ||
4916 | defer res.Body.Close() | ||
4917 | if res.StatusCode != http.StatusNoContent { | ||
4918 | logError(testName, function, args, startTime, "", "Http request failed", errors.New(res.Status)) | ||
4919 | return | ||
4920 | } | ||
4921 | |||
4922 | // expected path should be absolute path of the object | ||
4923 | var scheme string | ||
4924 | if mustParseBool(os.Getenv(enableHTTPS)) { | ||
4925 | scheme = "https://" | ||
4926 | } else { | ||
4927 | scheme = "http://" | ||
4928 | } | ||
4929 | |||
4930 | expectedLocation := scheme + os.Getenv(serverEndpoint) + "/" + bucketName + "/" + objectName | ||
4931 | expectedLocationBucketDNS := scheme + bucketName + "." + os.Getenv(serverEndpoint) + "/" + objectName | ||
4932 | |||
4933 | if !strings.Contains(expectedLocation, "s3.amazonaws.com/") { | ||
4934 | // Test when not against AWS S3. | ||
4935 | if val, ok := res.Header["Location"]; ok { | ||
4936 | if val[0] != expectedLocation && val[0] != expectedLocationBucketDNS { | ||
4937 | logError(testName, function, args, startTime, "", fmt.Sprintf("Location in header response is incorrect. Want %q or %q, got %q", expectedLocation, expectedLocationBucketDNS, val[0]), err) | ||
4938 | return | ||
4939 | } | ||
4940 | } else { | ||
4941 | logError(testName, function, args, startTime, "", "Location not found in header response", err) | ||
4942 | return | ||
4943 | } | ||
4944 | } | ||
4945 | want := checksum.Encoded() | ||
4946 | if got := res.Header.Get("X-Amz-Checksum-Crc32c"); got != want { | ||
4947 | logError(testName, function, args, startTime, "", fmt.Sprintf("Want checksum %q, got %q", want, got), nil) | ||
4948 | return | ||
4949 | } | ||
4950 | |||
4951 | successLogger(testName, function, args, startTime).Info() | ||
4952 | } | ||
4953 | |||
4954 | // Tests copy object | ||
4955 | func testCopyObject() { | ||
4956 | // initialize logging params | ||
4957 | startTime := time.Now() | ||
4958 | testName := getFuncName() | ||
4959 | function := "CopyObject(dst, src)" | ||
4960 | args := map[string]interface{}{} | ||
4961 | |||
4962 | // Seed random based on current time. | ||
4963 | rand.Seed(time.Now().Unix()) | ||
4964 | |||
4965 | // Instantiate new minio client object | ||
4966 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
4967 | &minio.Options{ | ||
4968 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
4969 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
4970 | }) | ||
4971 | if err != nil { | ||
4972 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
4973 | return | ||
4974 | } | ||
4975 | |||
4976 | // Enable tracing, write to stderr. | ||
4977 | // c.TraceOn(os.Stderr) | ||
4978 | |||
4979 | // Set user agent. | ||
4980 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
4981 | |||
4982 | // Generate a new random bucket name. | ||
4983 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
4984 | |||
4985 | // Make a new bucket in 'us-east-1' (source bucket). | ||
4986 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
4987 | if err != nil { | ||
4988 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
4989 | return | ||
4990 | } | ||
4991 | |||
4992 | defer cleanupBucket(bucketName, c) | ||
4993 | |||
4994 | // Make a new bucket in 'us-east-1' (destination bucket). | ||
4995 | err = c.MakeBucket(context.Background(), bucketName+"-copy", minio.MakeBucketOptions{Region: "us-east-1"}) | ||
4996 | if err != nil { | ||
4997 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
4998 | return | ||
4999 | } | ||
5000 | defer cleanupBucket(bucketName+"-copy", c) | ||
5001 | |||
5002 | // Generate 33K of data. | ||
5003 | bufSize := dataFileMap["datafile-33-kB"] | ||
5004 | reader := getDataReader("datafile-33-kB") | ||
5005 | |||
5006 | // Save the data | ||
5007 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
5008 | _, err = c.PutObject(context.Background(), bucketName, objectName, reader, int64(bufSize), minio.PutObjectOptions{ContentType: "binary/octet-stream"}) | ||
5009 | if err != nil { | ||
5010 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
5011 | return | ||
5012 | } | ||
5013 | |||
5014 | r, err := c.GetObject(context.Background(), bucketName, objectName, minio.GetObjectOptions{}) | ||
5015 | if err != nil { | ||
5016 | logError(testName, function, args, startTime, "", "GetObject failed", err) | ||
5017 | return | ||
5018 | } | ||
5019 | // Check the various fields of source object against destination object. | ||
5020 | objInfo, err := r.Stat() | ||
5021 | if err != nil { | ||
5022 | logError(testName, function, args, startTime, "", "Stat failed", err) | ||
5023 | return | ||
5024 | } | ||
5025 | |||
5026 | // Copy Source | ||
5027 | src := minio.CopySrcOptions{ | ||
5028 | Bucket: bucketName, | ||
5029 | Object: objectName, | ||
5030 | // Set copy conditions. | ||
5031 | MatchETag: objInfo.ETag, | ||
5032 | MatchModifiedSince: time.Date(2014, time.April, 0, 0, 0, 0, 0, time.UTC), | ||
5033 | } | ||
5034 | args["src"] = src | ||
5035 | |||
5036 | dst := minio.CopyDestOptions{ | ||
5037 | Bucket: bucketName + "-copy", | ||
5038 | Object: objectName + "-copy", | ||
5039 | } | ||
5040 | |||
5041 | // Perform the Copy | ||
5042 | if _, err = c.CopyObject(context.Background(), dst, src); err != nil { | ||
5043 | logError(testName, function, args, startTime, "", "CopyObject failed", err) | ||
5044 | return | ||
5045 | } | ||
5046 | |||
5047 | // Source object | ||
5048 | r, err = c.GetObject(context.Background(), bucketName, objectName, minio.GetObjectOptions{}) | ||
5049 | if err != nil { | ||
5050 | logError(testName, function, args, startTime, "", "GetObject failed", err) | ||
5051 | return | ||
5052 | } | ||
5053 | |||
5054 | // Destination object | ||
5055 | readerCopy, err := c.GetObject(context.Background(), bucketName+"-copy", objectName+"-copy", minio.GetObjectOptions{}) | ||
5056 | if err != nil { | ||
5057 | logError(testName, function, args, startTime, "", "GetObject failed", err) | ||
5058 | return | ||
5059 | } | ||
5060 | |||
5061 | // Check the various fields of source object against destination object. | ||
5062 | objInfo, err = r.Stat() | ||
5063 | if err != nil { | ||
5064 | logError(testName, function, args, startTime, "", "Stat failed", err) | ||
5065 | return | ||
5066 | } | ||
5067 | objInfoCopy, err := readerCopy.Stat() | ||
5068 | if err != nil { | ||
5069 | logError(testName, function, args, startTime, "", "Stat failed", err) | ||
5070 | return | ||
5071 | } | ||
5072 | if objInfo.Size != objInfoCopy.Size { | ||
5073 | logError(testName, function, args, startTime, "", "Number of bytes does not match, expected "+string(objInfoCopy.Size)+", got "+string(objInfo.Size), err) | ||
5074 | return | ||
5075 | } | ||
5076 | |||
5077 | if err := crcMatchesName(r, "datafile-33-kB"); err != nil { | ||
5078 | logError(testName, function, args, startTime, "", "data CRC check failed", err) | ||
5079 | return | ||
5080 | } | ||
5081 | if err := crcMatchesName(readerCopy, "datafile-33-kB"); err != nil { | ||
5082 | logError(testName, function, args, startTime, "", "copy data CRC check failed", err) | ||
5083 | return | ||
5084 | } | ||
5085 | // Close all the get readers before proceeding with CopyObject operations. | ||
5086 | r.Close() | ||
5087 | readerCopy.Close() | ||
5088 | |||
5089 | // CopyObject again but with wrong conditions | ||
5090 | src = minio.CopySrcOptions{ | ||
5091 | Bucket: bucketName, | ||
5092 | Object: objectName, | ||
5093 | MatchUnmodifiedSince: time.Date(2014, time.April, 0, 0, 0, 0, 0, time.UTC), | ||
5094 | NoMatchETag: objInfo.ETag, | ||
5095 | } | ||
5096 | |||
5097 | // Perform the Copy which should fail | ||
5098 | _, err = c.CopyObject(context.Background(), dst, src) | ||
5099 | if err == nil { | ||
5100 | logError(testName, function, args, startTime, "", "CopyObject did not fail for invalid conditions", err) | ||
5101 | return | ||
5102 | } | ||
5103 | |||
5104 | src = minio.CopySrcOptions{ | ||
5105 | Bucket: bucketName, | ||
5106 | Object: objectName, | ||
5107 | } | ||
5108 | |||
5109 | dst = minio.CopyDestOptions{ | ||
5110 | Bucket: bucketName, | ||
5111 | Object: objectName, | ||
5112 | ReplaceMetadata: true, | ||
5113 | UserMetadata: map[string]string{ | ||
5114 | "Copy": "should be same", | ||
5115 | }, | ||
5116 | } | ||
5117 | args["dst"] = dst | ||
5118 | args["src"] = src | ||
5119 | |||
5120 | _, err = c.CopyObject(context.Background(), dst, src) | ||
5121 | if err != nil { | ||
5122 | logError(testName, function, args, startTime, "", "CopyObject shouldn't fail", err) | ||
5123 | return | ||
5124 | } | ||
5125 | |||
5126 | oi, err := c.StatObject(context.Background(), bucketName, objectName, minio.StatObjectOptions{}) | ||
5127 | if err != nil { | ||
5128 | logError(testName, function, args, startTime, "", "StatObject failed", err) | ||
5129 | return | ||
5130 | } | ||
5131 | |||
5132 | stOpts := minio.StatObjectOptions{} | ||
5133 | stOpts.SetMatchETag(oi.ETag) | ||
5134 | objInfo, err = c.StatObject(context.Background(), bucketName, objectName, stOpts) | ||
5135 | if err != nil { | ||
5136 | logError(testName, function, args, startTime, "", "CopyObject ETag should match and not fail", err) | ||
5137 | return | ||
5138 | } | ||
5139 | |||
5140 | if objInfo.Metadata.Get("x-amz-meta-copy") != "should be same" { | ||
5141 | logError(testName, function, args, startTime, "", "CopyObject modified metadata should match", err) | ||
5142 | return | ||
5143 | } | ||
5144 | |||
5145 | successLogger(testName, function, args, startTime).Info() | ||
5146 | } | ||
5147 | |||
5148 | // Tests SSE-C get object ReaderSeeker interface methods. | ||
5149 | func testSSECEncryptedGetObjectReadSeekFunctional() { | ||
5150 | // initialize logging params | ||
5151 | startTime := time.Now() | ||
5152 | testName := getFuncName() | ||
5153 | function := "GetObject(bucketName, objectName)" | ||
5154 | args := map[string]interface{}{} | ||
5155 | |||
5156 | // Seed random based on current time. | ||
5157 | rand.Seed(time.Now().Unix()) | ||
5158 | |||
5159 | // Instantiate new minio client object. | ||
5160 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
5161 | &minio.Options{ | ||
5162 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
5163 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
5164 | }) | ||
5165 | if err != nil { | ||
5166 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
5167 | return | ||
5168 | } | ||
5169 | |||
5170 | // Enable tracing, write to stderr. | ||
5171 | // c.TraceOn(os.Stderr) | ||
5172 | |||
5173 | // Set user agent. | ||
5174 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
5175 | |||
5176 | // Generate a new random bucket name. | ||
5177 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
5178 | args["bucketName"] = bucketName | ||
5179 | |||
5180 | // Make a new bucket. | ||
5181 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
5182 | if err != nil { | ||
5183 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
5184 | return | ||
5185 | } | ||
5186 | |||
5187 | defer func() { | ||
5188 | // Delete all objects and buckets | ||
5189 | if err = cleanupBucket(bucketName, c); err != nil { | ||
5190 | logError(testName, function, args, startTime, "", "CleanupBucket failed", err) | ||
5191 | return | ||
5192 | } | ||
5193 | }() | ||
5194 | |||
5195 | // Generate 129MiB of data. | ||
5196 | bufSize := dataFileMap["datafile-129-MB"] | ||
5197 | reader := getDataReader("datafile-129-MB") | ||
5198 | defer reader.Close() | ||
5199 | |||
5200 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
5201 | args["objectName"] = objectName | ||
5202 | |||
5203 | buf, err := io.ReadAll(reader) | ||
5204 | if err != nil { | ||
5205 | logError(testName, function, args, startTime, "", "ReadAll failed", err) | ||
5206 | return | ||
5207 | } | ||
5208 | |||
5209 | // Save the data | ||
5210 | _, err = c.PutObject(context.Background(), bucketName, objectName, bytes.NewReader(buf), int64(len(buf)), minio.PutObjectOptions{ | ||
5211 | ContentType: "binary/octet-stream", | ||
5212 | ServerSideEncryption: encrypt.DefaultPBKDF([]byte("correct horse battery staple"), []byte(bucketName+objectName)), | ||
5213 | }) | ||
5214 | if err != nil { | ||
5215 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
5216 | return | ||
5217 | } | ||
5218 | |||
5219 | // Read the data back | ||
5220 | r, err := c.GetObject(context.Background(), bucketName, objectName, minio.GetObjectOptions{ | ||
5221 | ServerSideEncryption: encrypt.DefaultPBKDF([]byte("correct horse battery staple"), []byte(bucketName+objectName)), | ||
5222 | }) | ||
5223 | if err != nil { | ||
5224 | logError(testName, function, args, startTime, "", "GetObject failed", err) | ||
5225 | return | ||
5226 | } | ||
5227 | defer r.Close() | ||
5228 | |||
5229 | st, err := r.Stat() | ||
5230 | if err != nil { | ||
5231 | logError(testName, function, args, startTime, "", "Stat object failed", err) | ||
5232 | return | ||
5233 | } | ||
5234 | |||
5235 | if st.Size != int64(bufSize) { | ||
5236 | logError(testName, function, args, startTime, "", "Number of bytes does not match, expected "+string(int64(bufSize))+", got "+string(st.Size), err) | ||
5237 | return | ||
5238 | } | ||
5239 | |||
5240 | // This following function helps us to compare data from the reader after seek | ||
5241 | // with the data from the original buffer | ||
5242 | cmpData := func(r io.Reader, start, end int) { | ||
5243 | if end-start == 0 { | ||
5244 | return | ||
5245 | } | ||
5246 | buffer := bytes.NewBuffer([]byte{}) | ||
5247 | if _, err := io.CopyN(buffer, r, int64(bufSize)); err != nil { | ||
5248 | if err != io.EOF { | ||
5249 | logError(testName, function, args, startTime, "", "CopyN failed", err) | ||
5250 | return | ||
5251 | } | ||
5252 | } | ||
5253 | if !bytes.Equal(buf[start:end], buffer.Bytes()) { | ||
5254 | logError(testName, function, args, startTime, "", "Incorrect read bytes v/s original buffer", err) | ||
5255 | return | ||
5256 | } | ||
5257 | } | ||
5258 | |||
5259 | testCases := []struct { | ||
5260 | offset int64 | ||
5261 | whence int | ||
5262 | pos int64 | ||
5263 | err error | ||
5264 | shouldCmp bool | ||
5265 | start int | ||
5266 | end int | ||
5267 | }{ | ||
5268 | // Start from offset 0, fetch data and compare | ||
5269 | {0, 0, 0, nil, true, 0, 0}, | ||
5270 | // Start from offset 2048, fetch data and compare | ||
5271 | {2048, 0, 2048, nil, true, 2048, bufSize}, | ||
5272 | // Start from offset larger than possible | ||
5273 | {int64(bufSize) + 1024, 0, 0, io.EOF, false, 0, 0}, | ||
5274 | // Move to offset 0 without comparing | ||
5275 | {0, 0, 0, nil, false, 0, 0}, | ||
5276 | // Move one step forward and compare | ||
5277 | {1, 1, 1, nil, true, 1, bufSize}, | ||
5278 | // Move larger than possible | ||
5279 | {int64(bufSize), 1, 0, io.EOF, false, 0, 0}, | ||
5280 | // Provide negative offset with CUR_SEEK | ||
5281 | {int64(-1), 1, 0, fmt.Errorf("Negative position not allowed for 1"), false, 0, 0}, | ||
5282 | // Test with whence SEEK_END and with positive offset | ||
5283 | {1024, 2, 0, io.EOF, false, 0, 0}, | ||
5284 | // Test with whence SEEK_END and with negative offset | ||
5285 | {-1024, 2, int64(bufSize) - 1024, nil, true, bufSize - 1024, bufSize}, | ||
5286 | // Test with whence SEEK_END and with large negative offset | ||
5287 | {-int64(bufSize) * 2, 2, 0, fmt.Errorf("Seeking at negative offset not allowed for 2"), false, 0, 0}, | ||
5288 | // Test with invalid whence | ||
5289 | {0, 3, 0, fmt.Errorf("Invalid whence 3"), false, 0, 0}, | ||
5290 | } | ||
5291 | |||
5292 | for i, testCase := range testCases { | ||
5293 | // Perform seek operation | ||
5294 | n, err := r.Seek(testCase.offset, testCase.whence) | ||
5295 | if err != nil && testCase.err == nil { | ||
5296 | // We expected success. | ||
5297 | logError(testName, function, args, startTime, "", | ||
5298 | fmt.Sprintf("Test %d, unexpected err value: expected: %s, found: %s", i+1, testCase.err, err), err) | ||
5299 | return | ||
5300 | } | ||
5301 | if err == nil && testCase.err != nil { | ||
5302 | // We expected failure, but got success. | ||
5303 | logError(testName, function, args, startTime, "", | ||
5304 | fmt.Sprintf("Test %d, unexpected err value: expected: %s, found: %s", i+1, testCase.err, err), err) | ||
5305 | return | ||
5306 | } | ||
5307 | if err != nil && testCase.err != nil { | ||
5308 | if err.Error() != testCase.err.Error() { | ||
5309 | // We expect a specific error | ||
5310 | logError(testName, function, args, startTime, "", | ||
5311 | fmt.Sprintf("Test %d, unexpected err value: expected: %s, found: %s", i+1, testCase.err, err), err) | ||
5312 | return | ||
5313 | } | ||
5314 | } | ||
5315 | // Check the returned seek pos | ||
5316 | if n != testCase.pos { | ||
5317 | logError(testName, function, args, startTime, "", | ||
5318 | fmt.Sprintf("Test %d, number of bytes seeked does not match, expected %d, got %d", i+1, testCase.pos, n), err) | ||
5319 | return | ||
5320 | } | ||
5321 | // Compare only if shouldCmp is activated | ||
5322 | if testCase.shouldCmp { | ||
5323 | cmpData(r, testCase.start, testCase.end) | ||
5324 | } | ||
5325 | } | ||
5326 | |||
5327 | successLogger(testName, function, args, startTime).Info() | ||
5328 | } | ||
5329 | |||
5330 | // Tests SSE-S3 get object ReaderSeeker interface methods. | ||
5331 | func testSSES3EncryptedGetObjectReadSeekFunctional() { | ||
5332 | // initialize logging params | ||
5333 | startTime := time.Now() | ||
5334 | testName := getFuncName() | ||
5335 | function := "GetObject(bucketName, objectName)" | ||
5336 | args := map[string]interface{}{} | ||
5337 | |||
5338 | // Seed random based on current time. | ||
5339 | rand.Seed(time.Now().Unix()) | ||
5340 | |||
5341 | // Instantiate new minio client object. | ||
5342 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
5343 | &minio.Options{ | ||
5344 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
5345 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
5346 | }) | ||
5347 | if err != nil { | ||
5348 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
5349 | return | ||
5350 | } | ||
5351 | |||
5352 | // Enable tracing, write to stderr. | ||
5353 | // c.TraceOn(os.Stderr) | ||
5354 | |||
5355 | // Set user agent. | ||
5356 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
5357 | |||
5358 | // Generate a new random bucket name. | ||
5359 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
5360 | args["bucketName"] = bucketName | ||
5361 | |||
5362 | // Make a new bucket. | ||
5363 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
5364 | if err != nil { | ||
5365 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
5366 | return | ||
5367 | } | ||
5368 | |||
5369 | defer func() { | ||
5370 | // Delete all objects and buckets | ||
5371 | if err = cleanupBucket(bucketName, c); err != nil { | ||
5372 | logError(testName, function, args, startTime, "", "CleanupBucket failed", err) | ||
5373 | return | ||
5374 | } | ||
5375 | }() | ||
5376 | |||
5377 | // Generate 129MiB of data. | ||
5378 | bufSize := dataFileMap["datafile-129-MB"] | ||
5379 | reader := getDataReader("datafile-129-MB") | ||
5380 | defer reader.Close() | ||
5381 | |||
5382 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
5383 | args["objectName"] = objectName | ||
5384 | |||
5385 | buf, err := io.ReadAll(reader) | ||
5386 | if err != nil { | ||
5387 | logError(testName, function, args, startTime, "", "ReadAll failed", err) | ||
5388 | return | ||
5389 | } | ||
5390 | |||
5391 | // Save the data | ||
5392 | _, err = c.PutObject(context.Background(), bucketName, objectName, bytes.NewReader(buf), int64(len(buf)), minio.PutObjectOptions{ | ||
5393 | ContentType: "binary/octet-stream", | ||
5394 | ServerSideEncryption: encrypt.NewSSE(), | ||
5395 | }) | ||
5396 | if err != nil { | ||
5397 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
5398 | return | ||
5399 | } | ||
5400 | |||
5401 | // Read the data back | ||
5402 | r, err := c.GetObject(context.Background(), bucketName, objectName, minio.GetObjectOptions{}) | ||
5403 | if err != nil { | ||
5404 | logError(testName, function, args, startTime, "", "GetObject failed", err) | ||
5405 | return | ||
5406 | } | ||
5407 | defer r.Close() | ||
5408 | |||
5409 | st, err := r.Stat() | ||
5410 | if err != nil { | ||
5411 | logError(testName, function, args, startTime, "", "Stat object failed", err) | ||
5412 | return | ||
5413 | } | ||
5414 | |||
5415 | if st.Size != int64(bufSize) { | ||
5416 | logError(testName, function, args, startTime, "", "Number of bytes does not match, expected "+string(int64(bufSize))+", got "+string(st.Size), err) | ||
5417 | return | ||
5418 | } | ||
5419 | |||
5420 | // This following function helps us to compare data from the reader after seek | ||
5421 | // with the data from the original buffer | ||
5422 | cmpData := func(r io.Reader, start, end int) { | ||
5423 | if end-start == 0 { | ||
5424 | return | ||
5425 | } | ||
5426 | buffer := bytes.NewBuffer([]byte{}) | ||
5427 | if _, err := io.CopyN(buffer, r, int64(bufSize)); err != nil { | ||
5428 | if err != io.EOF { | ||
5429 | logError(testName, function, args, startTime, "", "CopyN failed", err) | ||
5430 | return | ||
5431 | } | ||
5432 | } | ||
5433 | if !bytes.Equal(buf[start:end], buffer.Bytes()) { | ||
5434 | logError(testName, function, args, startTime, "", "Incorrect read bytes v/s original buffer", err) | ||
5435 | return | ||
5436 | } | ||
5437 | } | ||
5438 | |||
5439 | testCases := []struct { | ||
5440 | offset int64 | ||
5441 | whence int | ||
5442 | pos int64 | ||
5443 | err error | ||
5444 | shouldCmp bool | ||
5445 | start int | ||
5446 | end int | ||
5447 | }{ | ||
5448 | // Start from offset 0, fetch data and compare | ||
5449 | {0, 0, 0, nil, true, 0, 0}, | ||
5450 | // Start from offset 2048, fetch data and compare | ||
5451 | {2048, 0, 2048, nil, true, 2048, bufSize}, | ||
5452 | // Start from offset larger than possible | ||
5453 | {int64(bufSize) + 1024, 0, 0, io.EOF, false, 0, 0}, | ||
5454 | // Move to offset 0 without comparing | ||
5455 | {0, 0, 0, nil, false, 0, 0}, | ||
5456 | // Move one step forward and compare | ||
5457 | {1, 1, 1, nil, true, 1, bufSize}, | ||
5458 | // Move larger than possible | ||
5459 | {int64(bufSize), 1, 0, io.EOF, false, 0, 0}, | ||
5460 | // Provide negative offset with CUR_SEEK | ||
5461 | {int64(-1), 1, 0, fmt.Errorf("Negative position not allowed for 1"), false, 0, 0}, | ||
5462 | // Test with whence SEEK_END and with positive offset | ||
5463 | {1024, 2, 0, io.EOF, false, 0, 0}, | ||
5464 | // Test with whence SEEK_END and with negative offset | ||
5465 | {-1024, 2, int64(bufSize) - 1024, nil, true, bufSize - 1024, bufSize}, | ||
5466 | // Test with whence SEEK_END and with large negative offset | ||
5467 | {-int64(bufSize) * 2, 2, 0, fmt.Errorf("Seeking at negative offset not allowed for 2"), false, 0, 0}, | ||
5468 | // Test with invalid whence | ||
5469 | {0, 3, 0, fmt.Errorf("Invalid whence 3"), false, 0, 0}, | ||
5470 | } | ||
5471 | |||
5472 | for i, testCase := range testCases { | ||
5473 | // Perform seek operation | ||
5474 | n, err := r.Seek(testCase.offset, testCase.whence) | ||
5475 | if err != nil && testCase.err == nil { | ||
5476 | // We expected success. | ||
5477 | logError(testName, function, args, startTime, "", | ||
5478 | fmt.Sprintf("Test %d, unexpected err value: expected: %s, found: %s", i+1, testCase.err, err), err) | ||
5479 | return | ||
5480 | } | ||
5481 | if err == nil && testCase.err != nil { | ||
5482 | // We expected failure, but got success. | ||
5483 | logError(testName, function, args, startTime, "", | ||
5484 | fmt.Sprintf("Test %d, unexpected err value: expected: %s, found: %s", i+1, testCase.err, err), err) | ||
5485 | return | ||
5486 | } | ||
5487 | if err != nil && testCase.err != nil { | ||
5488 | if err.Error() != testCase.err.Error() { | ||
5489 | // We expect a specific error | ||
5490 | logError(testName, function, args, startTime, "", | ||
5491 | fmt.Sprintf("Test %d, unexpected err value: expected: %s, found: %s", i+1, testCase.err, err), err) | ||
5492 | return | ||
5493 | } | ||
5494 | } | ||
5495 | // Check the returned seek pos | ||
5496 | if n != testCase.pos { | ||
5497 | logError(testName, function, args, startTime, "", | ||
5498 | fmt.Sprintf("Test %d, number of bytes seeked does not match, expected %d, got %d", i+1, testCase.pos, n), err) | ||
5499 | return | ||
5500 | } | ||
5501 | // Compare only if shouldCmp is activated | ||
5502 | if testCase.shouldCmp { | ||
5503 | cmpData(r, testCase.start, testCase.end) | ||
5504 | } | ||
5505 | } | ||
5506 | |||
5507 | successLogger(testName, function, args, startTime).Info() | ||
5508 | } | ||
5509 | |||
5510 | // Tests SSE-C get object ReaderAt interface methods. | ||
5511 | func testSSECEncryptedGetObjectReadAtFunctional() { | ||
5512 | // initialize logging params | ||
5513 | startTime := time.Now() | ||
5514 | testName := getFuncName() | ||
5515 | function := "GetObject(bucketName, objectName)" | ||
5516 | args := map[string]interface{}{} | ||
5517 | |||
5518 | // Seed random based on current time. | ||
5519 | rand.Seed(time.Now().Unix()) | ||
5520 | |||
5521 | // Instantiate new minio client object. | ||
5522 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
5523 | &minio.Options{ | ||
5524 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
5525 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
5526 | }) | ||
5527 | if err != nil { | ||
5528 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
5529 | return | ||
5530 | } | ||
5531 | |||
5532 | // Enable tracing, write to stderr. | ||
5533 | // c.TraceOn(os.Stderr) | ||
5534 | |||
5535 | // Set user agent. | ||
5536 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
5537 | |||
5538 | // Generate a new random bucket name. | ||
5539 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
5540 | args["bucketName"] = bucketName | ||
5541 | |||
5542 | // Make a new bucket. | ||
5543 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
5544 | if err != nil { | ||
5545 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
5546 | return | ||
5547 | } | ||
5548 | |||
5549 | defer cleanupBucket(bucketName, c) | ||
5550 | |||
5551 | // Generate 129MiB of data. | ||
5552 | bufSize := dataFileMap["datafile-129-MB"] | ||
5553 | reader := getDataReader("datafile-129-MB") | ||
5554 | defer reader.Close() | ||
5555 | |||
5556 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
5557 | args["objectName"] = objectName | ||
5558 | |||
5559 | buf, err := io.ReadAll(reader) | ||
5560 | if err != nil { | ||
5561 | logError(testName, function, args, startTime, "", "ReadAll failed", err) | ||
5562 | return | ||
5563 | } | ||
5564 | |||
5565 | // Save the data | ||
5566 | _, err = c.PutObject(context.Background(), bucketName, objectName, bytes.NewReader(buf), int64(len(buf)), minio.PutObjectOptions{ | ||
5567 | ContentType: "binary/octet-stream", | ||
5568 | ServerSideEncryption: encrypt.DefaultPBKDF([]byte("correct horse battery staple"), []byte(bucketName+objectName)), | ||
5569 | }) | ||
5570 | if err != nil { | ||
5571 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
5572 | return | ||
5573 | } | ||
5574 | |||
5575 | // read the data back | ||
5576 | r, err := c.GetObject(context.Background(), bucketName, objectName, minio.GetObjectOptions{ | ||
5577 | ServerSideEncryption: encrypt.DefaultPBKDF([]byte("correct horse battery staple"), []byte(bucketName+objectName)), | ||
5578 | }) | ||
5579 | if err != nil { | ||
5580 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
5581 | return | ||
5582 | } | ||
5583 | defer r.Close() | ||
5584 | |||
5585 | offset := int64(2048) | ||
5586 | |||
5587 | // read directly | ||
5588 | buf1 := make([]byte, 512) | ||
5589 | buf2 := make([]byte, 512) | ||
5590 | buf3 := make([]byte, 512) | ||
5591 | buf4 := make([]byte, 512) | ||
5592 | |||
5593 | // Test readAt before stat is called such that objectInfo doesn't change. | ||
5594 | m, err := r.ReadAt(buf1, offset) | ||
5595 | if err != nil { | ||
5596 | logError(testName, function, args, startTime, "", "ReadAt failed", err) | ||
5597 | return | ||
5598 | } | ||
5599 | if m != len(buf1) { | ||
5600 | logError(testName, function, args, startTime, "", "ReadAt read shorter bytes before reaching EOF, expected "+string(len(buf1))+", got "+string(m), err) | ||
5601 | return | ||
5602 | } | ||
5603 | if !bytes.Equal(buf1, buf[offset:offset+512]) { | ||
5604 | logError(testName, function, args, startTime, "", "Incorrect read between two ReadAt from same offset", err) | ||
5605 | return | ||
5606 | } | ||
5607 | offset += 512 | ||
5608 | |||
5609 | st, err := r.Stat() | ||
5610 | if err != nil { | ||
5611 | logError(testName, function, args, startTime, "", "Stat failed", err) | ||
5612 | return | ||
5613 | } | ||
5614 | |||
5615 | if st.Size != int64(bufSize) { | ||
5616 | logError(testName, function, args, startTime, "", "Number of bytes in stat does not match, expected "+string(int64(bufSize))+", got "+string(st.Size), err) | ||
5617 | return | ||
5618 | } | ||
5619 | |||
5620 | m, err = r.ReadAt(buf2, offset) | ||
5621 | if err != nil { | ||
5622 | logError(testName, function, args, startTime, "", "ReadAt failed", err) | ||
5623 | return | ||
5624 | } | ||
5625 | if m != len(buf2) { | ||
5626 | logError(testName, function, args, startTime, "", "ReadAt read shorter bytes before reaching EOF, expected "+string(len(buf2))+", got "+string(m), err) | ||
5627 | return | ||
5628 | } | ||
5629 | if !bytes.Equal(buf2, buf[offset:offset+512]) { | ||
5630 | logError(testName, function, args, startTime, "", "Incorrect read between two ReadAt from same offset", err) | ||
5631 | return | ||
5632 | } | ||
5633 | offset += 512 | ||
5634 | m, err = r.ReadAt(buf3, offset) | ||
5635 | if err != nil { | ||
5636 | logError(testName, function, args, startTime, "", "ReadAt failed", err) | ||
5637 | return | ||
5638 | } | ||
5639 | if m != len(buf3) { | ||
5640 | logError(testName, function, args, startTime, "", "ReadAt read shorter bytes before reaching EOF, expected "+string(len(buf3))+", got "+string(m), err) | ||
5641 | return | ||
5642 | } | ||
5643 | if !bytes.Equal(buf3, buf[offset:offset+512]) { | ||
5644 | logError(testName, function, args, startTime, "", "Incorrect read between two ReadAt from same offset", err) | ||
5645 | return | ||
5646 | } | ||
5647 | offset += 512 | ||
5648 | m, err = r.ReadAt(buf4, offset) | ||
5649 | if err != nil { | ||
5650 | logError(testName, function, args, startTime, "", "ReadAt failed", err) | ||
5651 | return | ||
5652 | } | ||
5653 | if m != len(buf4) { | ||
5654 | logError(testName, function, args, startTime, "", "ReadAt read shorter bytes before reaching EOF, expected "+string(len(buf4))+", got "+string(m), err) | ||
5655 | return | ||
5656 | } | ||
5657 | if !bytes.Equal(buf4, buf[offset:offset+512]) { | ||
5658 | logError(testName, function, args, startTime, "", "Incorrect read between two ReadAt from same offset", err) | ||
5659 | return | ||
5660 | } | ||
5661 | |||
5662 | buf5 := make([]byte, len(buf)) | ||
5663 | // Read the whole object. | ||
5664 | m, err = r.ReadAt(buf5, 0) | ||
5665 | if err != nil { | ||
5666 | if err != io.EOF { | ||
5667 | logError(testName, function, args, startTime, "", "ReadAt failed", err) | ||
5668 | return | ||
5669 | } | ||
5670 | } | ||
5671 | if m != len(buf5) { | ||
5672 | logError(testName, function, args, startTime, "", "ReadAt read shorter bytes before reaching EOF, expected "+string(len(buf5))+", got "+string(m), err) | ||
5673 | return | ||
5674 | } | ||
5675 | if !bytes.Equal(buf, buf5) { | ||
5676 | logError(testName, function, args, startTime, "", "Incorrect data read in GetObject, than what was previously uploaded", err) | ||
5677 | return | ||
5678 | } | ||
5679 | |||
5680 | buf6 := make([]byte, len(buf)+1) | ||
5681 | // Read the whole object and beyond. | ||
5682 | _, err = r.ReadAt(buf6, 0) | ||
5683 | if err != nil { | ||
5684 | if err != io.EOF { | ||
5685 | logError(testName, function, args, startTime, "", "ReadAt failed", err) | ||
5686 | return | ||
5687 | } | ||
5688 | } | ||
5689 | |||
5690 | successLogger(testName, function, args, startTime).Info() | ||
5691 | } | ||
5692 | |||
5693 | // Tests SSE-S3 get object ReaderAt interface methods. | ||
5694 | func testSSES3EncryptedGetObjectReadAtFunctional() { | ||
5695 | // initialize logging params | ||
5696 | startTime := time.Now() | ||
5697 | testName := getFuncName() | ||
5698 | function := "GetObject(bucketName, objectName)" | ||
5699 | args := map[string]interface{}{} | ||
5700 | |||
5701 | // Seed random based on current time. | ||
5702 | rand.Seed(time.Now().Unix()) | ||
5703 | |||
5704 | // Instantiate new minio client object. | ||
5705 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
5706 | &minio.Options{ | ||
5707 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
5708 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
5709 | }) | ||
5710 | if err != nil { | ||
5711 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
5712 | return | ||
5713 | } | ||
5714 | |||
5715 | // Enable tracing, write to stderr. | ||
5716 | // c.TraceOn(os.Stderr) | ||
5717 | |||
5718 | // Set user agent. | ||
5719 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
5720 | |||
5721 | // Generate a new random bucket name. | ||
5722 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
5723 | args["bucketName"] = bucketName | ||
5724 | |||
5725 | // Make a new bucket. | ||
5726 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
5727 | if err != nil { | ||
5728 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
5729 | return | ||
5730 | } | ||
5731 | |||
5732 | defer cleanupBucket(bucketName, c) | ||
5733 | |||
5734 | // Generate 129MiB of data. | ||
5735 | bufSize := dataFileMap["datafile-129-MB"] | ||
5736 | reader := getDataReader("datafile-129-MB") | ||
5737 | defer reader.Close() | ||
5738 | |||
5739 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
5740 | args["objectName"] = objectName | ||
5741 | |||
5742 | buf, err := io.ReadAll(reader) | ||
5743 | if err != nil { | ||
5744 | logError(testName, function, args, startTime, "", "ReadAll failed", err) | ||
5745 | return | ||
5746 | } | ||
5747 | |||
5748 | // Save the data | ||
5749 | _, err = c.PutObject(context.Background(), bucketName, objectName, bytes.NewReader(buf), int64(len(buf)), minio.PutObjectOptions{ | ||
5750 | ContentType: "binary/octet-stream", | ||
5751 | ServerSideEncryption: encrypt.NewSSE(), | ||
5752 | }) | ||
5753 | if err != nil { | ||
5754 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
5755 | return | ||
5756 | } | ||
5757 | |||
5758 | // read the data back | ||
5759 | r, err := c.GetObject(context.Background(), bucketName, objectName, minio.GetObjectOptions{}) | ||
5760 | if err != nil { | ||
5761 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
5762 | return | ||
5763 | } | ||
5764 | defer r.Close() | ||
5765 | |||
5766 | offset := int64(2048) | ||
5767 | |||
5768 | // read directly | ||
5769 | buf1 := make([]byte, 512) | ||
5770 | buf2 := make([]byte, 512) | ||
5771 | buf3 := make([]byte, 512) | ||
5772 | buf4 := make([]byte, 512) | ||
5773 | |||
5774 | // Test readAt before stat is called such that objectInfo doesn't change. | ||
5775 | m, err := r.ReadAt(buf1, offset) | ||
5776 | if err != nil { | ||
5777 | logError(testName, function, args, startTime, "", "ReadAt failed", err) | ||
5778 | return | ||
5779 | } | ||
5780 | if m != len(buf1) { | ||
5781 | logError(testName, function, args, startTime, "", "ReadAt read shorter bytes before reaching EOF, expected "+string(len(buf1))+", got "+string(m), err) | ||
5782 | return | ||
5783 | } | ||
5784 | if !bytes.Equal(buf1, buf[offset:offset+512]) { | ||
5785 | logError(testName, function, args, startTime, "", "Incorrect read between two ReadAt from same offset", err) | ||
5786 | return | ||
5787 | } | ||
5788 | offset += 512 | ||
5789 | |||
5790 | st, err := r.Stat() | ||
5791 | if err != nil { | ||
5792 | logError(testName, function, args, startTime, "", "Stat failed", err) | ||
5793 | return | ||
5794 | } | ||
5795 | |||
5796 | if st.Size != int64(bufSize) { | ||
5797 | logError(testName, function, args, startTime, "", "Number of bytes in stat does not match, expected "+string(int64(bufSize))+", got "+string(st.Size), err) | ||
5798 | return | ||
5799 | } | ||
5800 | |||
5801 | m, err = r.ReadAt(buf2, offset) | ||
5802 | if err != nil { | ||
5803 | logError(testName, function, args, startTime, "", "ReadAt failed", err) | ||
5804 | return | ||
5805 | } | ||
5806 | if m != len(buf2) { | ||
5807 | logError(testName, function, args, startTime, "", "ReadAt read shorter bytes before reaching EOF, expected "+string(len(buf2))+", got "+string(m), err) | ||
5808 | return | ||
5809 | } | ||
5810 | if !bytes.Equal(buf2, buf[offset:offset+512]) { | ||
5811 | logError(testName, function, args, startTime, "", "Incorrect read between two ReadAt from same offset", err) | ||
5812 | return | ||
5813 | } | ||
5814 | offset += 512 | ||
5815 | m, err = r.ReadAt(buf3, offset) | ||
5816 | if err != nil { | ||
5817 | logError(testName, function, args, startTime, "", "ReadAt failed", err) | ||
5818 | return | ||
5819 | } | ||
5820 | if m != len(buf3) { | ||
5821 | logError(testName, function, args, startTime, "", "ReadAt read shorter bytes before reaching EOF, expected "+string(len(buf3))+", got "+string(m), err) | ||
5822 | return | ||
5823 | } | ||
5824 | if !bytes.Equal(buf3, buf[offset:offset+512]) { | ||
5825 | logError(testName, function, args, startTime, "", "Incorrect read between two ReadAt from same offset", err) | ||
5826 | return | ||
5827 | } | ||
5828 | offset += 512 | ||
5829 | m, err = r.ReadAt(buf4, offset) | ||
5830 | if err != nil { | ||
5831 | logError(testName, function, args, startTime, "", "ReadAt failed", err) | ||
5832 | return | ||
5833 | } | ||
5834 | if m != len(buf4) { | ||
5835 | logError(testName, function, args, startTime, "", "ReadAt read shorter bytes before reaching EOF, expected "+string(len(buf4))+", got "+string(m), err) | ||
5836 | return | ||
5837 | } | ||
5838 | if !bytes.Equal(buf4, buf[offset:offset+512]) { | ||
5839 | logError(testName, function, args, startTime, "", "Incorrect read between two ReadAt from same offset", err) | ||
5840 | return | ||
5841 | } | ||
5842 | |||
5843 | buf5 := make([]byte, len(buf)) | ||
5844 | // Read the whole object. | ||
5845 | m, err = r.ReadAt(buf5, 0) | ||
5846 | if err != nil { | ||
5847 | if err != io.EOF { | ||
5848 | logError(testName, function, args, startTime, "", "ReadAt failed", err) | ||
5849 | return | ||
5850 | } | ||
5851 | } | ||
5852 | if m != len(buf5) { | ||
5853 | logError(testName, function, args, startTime, "", "ReadAt read shorter bytes before reaching EOF, expected "+string(len(buf5))+", got "+string(m), err) | ||
5854 | return | ||
5855 | } | ||
5856 | if !bytes.Equal(buf, buf5) { | ||
5857 | logError(testName, function, args, startTime, "", "Incorrect data read in GetObject, than what was previously uploaded", err) | ||
5858 | return | ||
5859 | } | ||
5860 | |||
5861 | buf6 := make([]byte, len(buf)+1) | ||
5862 | // Read the whole object and beyond. | ||
5863 | _, err = r.ReadAt(buf6, 0) | ||
5864 | if err != nil { | ||
5865 | if err != io.EOF { | ||
5866 | logError(testName, function, args, startTime, "", "ReadAt failed", err) | ||
5867 | return | ||
5868 | } | ||
5869 | } | ||
5870 | |||
5871 | successLogger(testName, function, args, startTime).Info() | ||
5872 | } | ||
5873 | |||
5874 | // testSSECEncryptionPutGet tests encryption with customer provided encryption keys | ||
5875 | func testSSECEncryptionPutGet() { | ||
5876 | // initialize logging params | ||
5877 | startTime := time.Now() | ||
5878 | testName := getFuncName() | ||
5879 | function := "PutEncryptedObject(bucketName, objectName, reader, sse)" | ||
5880 | args := map[string]interface{}{ | ||
5881 | "bucketName": "", | ||
5882 | "objectName": "", | ||
5883 | "sse": "", | ||
5884 | } | ||
5885 | // Seed random based on current time. | ||
5886 | rand.Seed(time.Now().Unix()) | ||
5887 | |||
5888 | // Instantiate new minio client object | ||
5889 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
5890 | &minio.Options{ | ||
5891 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
5892 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
5893 | }) | ||
5894 | if err != nil { | ||
5895 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
5896 | return | ||
5897 | } | ||
5898 | |||
5899 | // Enable tracing, write to stderr. | ||
5900 | // c.TraceOn(os.Stderr) | ||
5901 | |||
5902 | // Set user agent. | ||
5903 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
5904 | |||
5905 | // Generate a new random bucket name. | ||
5906 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
5907 | args["bucketName"] = bucketName | ||
5908 | |||
5909 | // Make a new bucket. | ||
5910 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
5911 | if err != nil { | ||
5912 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
5913 | return | ||
5914 | } | ||
5915 | |||
5916 | defer cleanupBucket(bucketName, c) | ||
5917 | |||
5918 | testCases := []struct { | ||
5919 | buf []byte | ||
5920 | }{ | ||
5921 | {buf: bytes.Repeat([]byte("F"), 1)}, | ||
5922 | {buf: bytes.Repeat([]byte("F"), 15)}, | ||
5923 | {buf: bytes.Repeat([]byte("F"), 16)}, | ||
5924 | {buf: bytes.Repeat([]byte("F"), 17)}, | ||
5925 | {buf: bytes.Repeat([]byte("F"), 31)}, | ||
5926 | {buf: bytes.Repeat([]byte("F"), 32)}, | ||
5927 | {buf: bytes.Repeat([]byte("F"), 33)}, | ||
5928 | {buf: bytes.Repeat([]byte("F"), 1024)}, | ||
5929 | {buf: bytes.Repeat([]byte("F"), 1024*2)}, | ||
5930 | {buf: bytes.Repeat([]byte("F"), 1024*1024)}, | ||
5931 | } | ||
5932 | |||
5933 | const password = "correct horse battery staple" // https://xkcd.com/936/ | ||
5934 | |||
5935 | for i, testCase := range testCases { | ||
5936 | // Generate a random object name | ||
5937 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
5938 | args["objectName"] = objectName | ||
5939 | |||
5940 | // Secured object | ||
5941 | sse := encrypt.DefaultPBKDF([]byte(password), []byte(bucketName+objectName)) | ||
5942 | args["sse"] = sse | ||
5943 | |||
5944 | // Put encrypted data | ||
5945 | _, err = c.PutObject(context.Background(), bucketName, objectName, bytes.NewReader(testCase.buf), int64(len(testCase.buf)), minio.PutObjectOptions{ServerSideEncryption: sse}) | ||
5946 | if err != nil { | ||
5947 | logError(testName, function, args, startTime, "", "PutEncryptedObject failed", err) | ||
5948 | return | ||
5949 | } | ||
5950 | |||
5951 | // Read the data back | ||
5952 | r, err := c.GetObject(context.Background(), bucketName, objectName, minio.GetObjectOptions{ServerSideEncryption: sse}) | ||
5953 | if err != nil { | ||
5954 | logError(testName, function, args, startTime, "", "GetEncryptedObject failed", err) | ||
5955 | return | ||
5956 | } | ||
5957 | defer r.Close() | ||
5958 | |||
5959 | // Compare the sent object with the received one | ||
5960 | recvBuffer := bytes.NewBuffer([]byte{}) | ||
5961 | if _, err = io.Copy(recvBuffer, r); err != nil { | ||
5962 | logError(testName, function, args, startTime, "", "Test "+string(i+1)+", error: "+err.Error(), err) | ||
5963 | return | ||
5964 | } | ||
5965 | if recvBuffer.Len() != len(testCase.buf) { | ||
5966 | logError(testName, function, args, startTime, "", "Test "+string(i+1)+", Number of bytes of received object does not match, expected "+string(len(testCase.buf))+", got "+string(recvBuffer.Len()), err) | ||
5967 | return | ||
5968 | } | ||
5969 | if !bytes.Equal(testCase.buf, recvBuffer.Bytes()) { | ||
5970 | logError(testName, function, args, startTime, "", "Test "+string(i+1)+", Encrypted sent is not equal to decrypted, expected "+string(testCase.buf)+", got "+string(recvBuffer.Bytes()), err) | ||
5971 | return | ||
5972 | } | ||
5973 | |||
5974 | successLogger(testName, function, args, startTime).Info() | ||
5975 | |||
5976 | } | ||
5977 | |||
5978 | successLogger(testName, function, args, startTime).Info() | ||
5979 | } | ||
5980 | |||
5981 | // TestEncryptionFPut tests encryption with customer specified encryption keys | ||
5982 | func testSSECEncryptionFPut() { | ||
5983 | // initialize logging params | ||
5984 | startTime := time.Now() | ||
5985 | testName := getFuncName() | ||
5986 | function := "FPutEncryptedObject(bucketName, objectName, filePath, contentType, sse)" | ||
5987 | args := map[string]interface{}{ | ||
5988 | "bucketName": "", | ||
5989 | "objectName": "", | ||
5990 | "filePath": "", | ||
5991 | "contentType": "", | ||
5992 | "sse": "", | ||
5993 | } | ||
5994 | // Seed random based on current time. | ||
5995 | rand.Seed(time.Now().Unix()) | ||
5996 | |||
5997 | // Instantiate new minio client object | ||
5998 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
5999 | &minio.Options{ | ||
6000 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
6001 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
6002 | }) | ||
6003 | if err != nil { | ||
6004 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
6005 | return | ||
6006 | } | ||
6007 | |||
6008 | // Enable tracing, write to stderr. | ||
6009 | // c.TraceOn(os.Stderr) | ||
6010 | |||
6011 | // Set user agent. | ||
6012 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
6013 | |||
6014 | // Generate a new random bucket name. | ||
6015 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
6016 | args["bucketName"] = bucketName | ||
6017 | |||
6018 | // Make a new bucket. | ||
6019 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
6020 | if err != nil { | ||
6021 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
6022 | return | ||
6023 | } | ||
6024 | |||
6025 | defer cleanupBucket(bucketName, c) | ||
6026 | |||
6027 | // Object custom metadata | ||
6028 | customContentType := "custom/contenttype" | ||
6029 | args["metadata"] = customContentType | ||
6030 | |||
6031 | testCases := []struct { | ||
6032 | buf []byte | ||
6033 | }{ | ||
6034 | {buf: bytes.Repeat([]byte("F"), 0)}, | ||
6035 | {buf: bytes.Repeat([]byte("F"), 1)}, | ||
6036 | {buf: bytes.Repeat([]byte("F"), 15)}, | ||
6037 | {buf: bytes.Repeat([]byte("F"), 16)}, | ||
6038 | {buf: bytes.Repeat([]byte("F"), 17)}, | ||
6039 | {buf: bytes.Repeat([]byte("F"), 31)}, | ||
6040 | {buf: bytes.Repeat([]byte("F"), 32)}, | ||
6041 | {buf: bytes.Repeat([]byte("F"), 33)}, | ||
6042 | {buf: bytes.Repeat([]byte("F"), 1024)}, | ||
6043 | {buf: bytes.Repeat([]byte("F"), 1024*2)}, | ||
6044 | {buf: bytes.Repeat([]byte("F"), 1024*1024)}, | ||
6045 | } | ||
6046 | |||
6047 | const password = "correct horse battery staple" // https://xkcd.com/936/ | ||
6048 | for i, testCase := range testCases { | ||
6049 | // Generate a random object name | ||
6050 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
6051 | args["objectName"] = objectName | ||
6052 | |||
6053 | // Secured object | ||
6054 | sse := encrypt.DefaultPBKDF([]byte(password), []byte(bucketName+objectName)) | ||
6055 | args["sse"] = sse | ||
6056 | |||
6057 | // Generate a random file name. | ||
6058 | fileName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
6059 | file, err := os.Create(fileName) | ||
6060 | if err != nil { | ||
6061 | logError(testName, function, args, startTime, "", "file create failed", err) | ||
6062 | return | ||
6063 | } | ||
6064 | _, err = file.Write(testCase.buf) | ||
6065 | if err != nil { | ||
6066 | logError(testName, function, args, startTime, "", "file write failed", err) | ||
6067 | return | ||
6068 | } | ||
6069 | file.Close() | ||
6070 | // Put encrypted data | ||
6071 | if _, err = c.FPutObject(context.Background(), bucketName, objectName, fileName, minio.PutObjectOptions{ServerSideEncryption: sse}); err != nil { | ||
6072 | logError(testName, function, args, startTime, "", "FPutEncryptedObject failed", err) | ||
6073 | return | ||
6074 | } | ||
6075 | |||
6076 | // Read the data back | ||
6077 | r, err := c.GetObject(context.Background(), bucketName, objectName, minio.GetObjectOptions{ServerSideEncryption: sse}) | ||
6078 | if err != nil { | ||
6079 | logError(testName, function, args, startTime, "", "GetEncryptedObject failed", err) | ||
6080 | return | ||
6081 | } | ||
6082 | defer r.Close() | ||
6083 | |||
6084 | // Compare the sent object with the received one | ||
6085 | recvBuffer := bytes.NewBuffer([]byte{}) | ||
6086 | if _, err = io.Copy(recvBuffer, r); err != nil { | ||
6087 | logError(testName, function, args, startTime, "", "Test "+string(i+1)+", error: "+err.Error(), err) | ||
6088 | return | ||
6089 | } | ||
6090 | if recvBuffer.Len() != len(testCase.buf) { | ||
6091 | logError(testName, function, args, startTime, "", "Test "+string(i+1)+", Number of bytes of received object does not match, expected "+string(len(testCase.buf))+", got "+string(recvBuffer.Len()), err) | ||
6092 | return | ||
6093 | } | ||
6094 | if !bytes.Equal(testCase.buf, recvBuffer.Bytes()) { | ||
6095 | logError(testName, function, args, startTime, "", "Test "+string(i+1)+", Encrypted sent is not equal to decrypted, expected "+string(testCase.buf)+", got "+string(recvBuffer.Bytes()), err) | ||
6096 | return | ||
6097 | } | ||
6098 | |||
6099 | os.Remove(fileName) | ||
6100 | } | ||
6101 | |||
6102 | successLogger(testName, function, args, startTime).Info() | ||
6103 | } | ||
6104 | |||
6105 | // testSSES3EncryptionPutGet tests SSE-S3 encryption | ||
6106 | func testSSES3EncryptionPutGet() { | ||
6107 | // initialize logging params | ||
6108 | startTime := time.Now() | ||
6109 | testName := getFuncName() | ||
6110 | function := "PutEncryptedObject(bucketName, objectName, reader, sse)" | ||
6111 | args := map[string]interface{}{ | ||
6112 | "bucketName": "", | ||
6113 | "objectName": "", | ||
6114 | "sse": "", | ||
6115 | } | ||
6116 | // Seed random based on current time. | ||
6117 | rand.Seed(time.Now().Unix()) | ||
6118 | |||
6119 | // Instantiate new minio client object | ||
6120 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
6121 | &minio.Options{ | ||
6122 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
6123 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
6124 | }) | ||
6125 | if err != nil { | ||
6126 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
6127 | return | ||
6128 | } | ||
6129 | |||
6130 | // Enable tracing, write to stderr. | ||
6131 | // c.TraceOn(os.Stderr) | ||
6132 | |||
6133 | // Set user agent. | ||
6134 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
6135 | |||
6136 | // Generate a new random bucket name. | ||
6137 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
6138 | args["bucketName"] = bucketName | ||
6139 | |||
6140 | // Make a new bucket. | ||
6141 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
6142 | if err != nil { | ||
6143 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
6144 | return | ||
6145 | } | ||
6146 | |||
6147 | defer cleanupBucket(bucketName, c) | ||
6148 | |||
6149 | testCases := []struct { | ||
6150 | buf []byte | ||
6151 | }{ | ||
6152 | {buf: bytes.Repeat([]byte("F"), 1)}, | ||
6153 | {buf: bytes.Repeat([]byte("F"), 15)}, | ||
6154 | {buf: bytes.Repeat([]byte("F"), 16)}, | ||
6155 | {buf: bytes.Repeat([]byte("F"), 17)}, | ||
6156 | {buf: bytes.Repeat([]byte("F"), 31)}, | ||
6157 | {buf: bytes.Repeat([]byte("F"), 32)}, | ||
6158 | {buf: bytes.Repeat([]byte("F"), 33)}, | ||
6159 | {buf: bytes.Repeat([]byte("F"), 1024)}, | ||
6160 | {buf: bytes.Repeat([]byte("F"), 1024*2)}, | ||
6161 | {buf: bytes.Repeat([]byte("F"), 1024*1024)}, | ||
6162 | } | ||
6163 | |||
6164 | for i, testCase := range testCases { | ||
6165 | // Generate a random object name | ||
6166 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
6167 | args["objectName"] = objectName | ||
6168 | |||
6169 | // Secured object | ||
6170 | sse := encrypt.NewSSE() | ||
6171 | args["sse"] = sse | ||
6172 | |||
6173 | // Put encrypted data | ||
6174 | _, err = c.PutObject(context.Background(), bucketName, objectName, bytes.NewReader(testCase.buf), int64(len(testCase.buf)), minio.PutObjectOptions{ServerSideEncryption: sse}) | ||
6175 | if err != nil { | ||
6176 | logError(testName, function, args, startTime, "", "PutEncryptedObject failed", err) | ||
6177 | return | ||
6178 | } | ||
6179 | |||
6180 | // Read the data back without any encryption headers | ||
6181 | r, err := c.GetObject(context.Background(), bucketName, objectName, minio.GetObjectOptions{}) | ||
6182 | if err != nil { | ||
6183 | logError(testName, function, args, startTime, "", "GetEncryptedObject failed", err) | ||
6184 | return | ||
6185 | } | ||
6186 | defer r.Close() | ||
6187 | |||
6188 | // Compare the sent object with the received one | ||
6189 | recvBuffer := bytes.NewBuffer([]byte{}) | ||
6190 | if _, err = io.Copy(recvBuffer, r); err != nil { | ||
6191 | logError(testName, function, args, startTime, "", "Test "+string(i+1)+", error: "+err.Error(), err) | ||
6192 | return | ||
6193 | } | ||
6194 | if recvBuffer.Len() != len(testCase.buf) { | ||
6195 | logError(testName, function, args, startTime, "", "Test "+string(i+1)+", Number of bytes of received object does not match, expected "+string(len(testCase.buf))+", got "+string(recvBuffer.Len()), err) | ||
6196 | return | ||
6197 | } | ||
6198 | if !bytes.Equal(testCase.buf, recvBuffer.Bytes()) { | ||
6199 | logError(testName, function, args, startTime, "", "Test "+string(i+1)+", Encrypted sent is not equal to decrypted, expected "+string(testCase.buf)+", got "+string(recvBuffer.Bytes()), err) | ||
6200 | return | ||
6201 | } | ||
6202 | |||
6203 | successLogger(testName, function, args, startTime).Info() | ||
6204 | |||
6205 | } | ||
6206 | |||
6207 | successLogger(testName, function, args, startTime).Info() | ||
6208 | } | ||
6209 | |||
6210 | // TestSSES3EncryptionFPut tests server side encryption | ||
6211 | func testSSES3EncryptionFPut() { | ||
6212 | // initialize logging params | ||
6213 | startTime := time.Now() | ||
6214 | testName := getFuncName() | ||
6215 | function := "FPutEncryptedObject(bucketName, objectName, filePath, contentType, sse)" | ||
6216 | args := map[string]interface{}{ | ||
6217 | "bucketName": "", | ||
6218 | "objectName": "", | ||
6219 | "filePath": "", | ||
6220 | "contentType": "", | ||
6221 | "sse": "", | ||
6222 | } | ||
6223 | // Seed random based on current time. | ||
6224 | rand.Seed(time.Now().Unix()) | ||
6225 | |||
6226 | // Instantiate new minio client object | ||
6227 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
6228 | &minio.Options{ | ||
6229 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
6230 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
6231 | }) | ||
6232 | if err != nil { | ||
6233 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
6234 | return | ||
6235 | } | ||
6236 | |||
6237 | // Enable tracing, write to stderr. | ||
6238 | // c.TraceOn(os.Stderr) | ||
6239 | |||
6240 | // Set user agent. | ||
6241 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
6242 | |||
6243 | // Generate a new random bucket name. | ||
6244 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
6245 | args["bucketName"] = bucketName | ||
6246 | |||
6247 | // Make a new bucket. | ||
6248 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
6249 | if err != nil { | ||
6250 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
6251 | return | ||
6252 | } | ||
6253 | |||
6254 | defer cleanupBucket(bucketName, c) | ||
6255 | |||
6256 | // Object custom metadata | ||
6257 | customContentType := "custom/contenttype" | ||
6258 | args["metadata"] = customContentType | ||
6259 | |||
6260 | testCases := []struct { | ||
6261 | buf []byte | ||
6262 | }{ | ||
6263 | {buf: bytes.Repeat([]byte("F"), 0)}, | ||
6264 | {buf: bytes.Repeat([]byte("F"), 1)}, | ||
6265 | {buf: bytes.Repeat([]byte("F"), 15)}, | ||
6266 | {buf: bytes.Repeat([]byte("F"), 16)}, | ||
6267 | {buf: bytes.Repeat([]byte("F"), 17)}, | ||
6268 | {buf: bytes.Repeat([]byte("F"), 31)}, | ||
6269 | {buf: bytes.Repeat([]byte("F"), 32)}, | ||
6270 | {buf: bytes.Repeat([]byte("F"), 33)}, | ||
6271 | {buf: bytes.Repeat([]byte("F"), 1024)}, | ||
6272 | {buf: bytes.Repeat([]byte("F"), 1024*2)}, | ||
6273 | {buf: bytes.Repeat([]byte("F"), 1024*1024)}, | ||
6274 | } | ||
6275 | |||
6276 | for i, testCase := range testCases { | ||
6277 | // Generate a random object name | ||
6278 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
6279 | args["objectName"] = objectName | ||
6280 | |||
6281 | // Secured object | ||
6282 | sse := encrypt.NewSSE() | ||
6283 | args["sse"] = sse | ||
6284 | |||
6285 | // Generate a random file name. | ||
6286 | fileName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
6287 | file, err := os.Create(fileName) | ||
6288 | if err != nil { | ||
6289 | logError(testName, function, args, startTime, "", "file create failed", err) | ||
6290 | return | ||
6291 | } | ||
6292 | _, err = file.Write(testCase.buf) | ||
6293 | if err != nil { | ||
6294 | logError(testName, function, args, startTime, "", "file write failed", err) | ||
6295 | return | ||
6296 | } | ||
6297 | file.Close() | ||
6298 | // Put encrypted data | ||
6299 | if _, err = c.FPutObject(context.Background(), bucketName, objectName, fileName, minio.PutObjectOptions{ServerSideEncryption: sse}); err != nil { | ||
6300 | logError(testName, function, args, startTime, "", "FPutEncryptedObject failed", err) | ||
6301 | return | ||
6302 | } | ||
6303 | |||
6304 | // Read the data back | ||
6305 | r, err := c.GetObject(context.Background(), bucketName, objectName, minio.GetObjectOptions{}) | ||
6306 | if err != nil { | ||
6307 | logError(testName, function, args, startTime, "", "GetEncryptedObject failed", err) | ||
6308 | return | ||
6309 | } | ||
6310 | defer r.Close() | ||
6311 | |||
6312 | // Compare the sent object with the received one | ||
6313 | recvBuffer := bytes.NewBuffer([]byte{}) | ||
6314 | if _, err = io.Copy(recvBuffer, r); err != nil { | ||
6315 | logError(testName, function, args, startTime, "", "Test "+string(i+1)+", error: "+err.Error(), err) | ||
6316 | return | ||
6317 | } | ||
6318 | if recvBuffer.Len() != len(testCase.buf) { | ||
6319 | logError(testName, function, args, startTime, "", "Test "+string(i+1)+", Number of bytes of received object does not match, expected "+string(len(testCase.buf))+", got "+string(recvBuffer.Len()), err) | ||
6320 | return | ||
6321 | } | ||
6322 | if !bytes.Equal(testCase.buf, recvBuffer.Bytes()) { | ||
6323 | logError(testName, function, args, startTime, "", "Test "+string(i+1)+", Encrypted sent is not equal to decrypted, expected "+string(testCase.buf)+", got "+string(recvBuffer.Bytes()), err) | ||
6324 | return | ||
6325 | } | ||
6326 | |||
6327 | os.Remove(fileName) | ||
6328 | } | ||
6329 | |||
6330 | successLogger(testName, function, args, startTime).Info() | ||
6331 | } | ||
6332 | |||
6333 | func testBucketNotification() { | ||
6334 | // initialize logging params | ||
6335 | startTime := time.Now() | ||
6336 | testName := getFuncName() | ||
6337 | function := "SetBucketNotification(bucketName)" | ||
6338 | args := map[string]interface{}{ | ||
6339 | "bucketName": "", | ||
6340 | } | ||
6341 | |||
6342 | if os.Getenv("NOTIFY_BUCKET") == "" || | ||
6343 | os.Getenv("NOTIFY_SERVICE") == "" || | ||
6344 | os.Getenv("NOTIFY_REGION") == "" || | ||
6345 | os.Getenv("NOTIFY_ACCOUNTID") == "" || | ||
6346 | os.Getenv("NOTIFY_RESOURCE") == "" { | ||
6347 | ignoredLog(testName, function, args, startTime, "Skipped notification test as it is not configured").Info() | ||
6348 | return | ||
6349 | } | ||
6350 | |||
6351 | // Seed random based on current time. | ||
6352 | rand.Seed(time.Now().Unix()) | ||
6353 | |||
6354 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
6355 | &minio.Options{ | ||
6356 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
6357 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
6358 | }) | ||
6359 | if err != nil { | ||
6360 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
6361 | return | ||
6362 | } | ||
6363 | |||
6364 | // Enable to debug | ||
6365 | // c.TraceOn(os.Stderr) | ||
6366 | |||
6367 | // Set user agent. | ||
6368 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
6369 | |||
6370 | bucketName := os.Getenv("NOTIFY_BUCKET") | ||
6371 | args["bucketName"] = bucketName | ||
6372 | |||
6373 | topicArn := notification.NewArn("aws", os.Getenv("NOTIFY_SERVICE"), os.Getenv("NOTIFY_REGION"), os.Getenv("NOTIFY_ACCOUNTID"), os.Getenv("NOTIFY_RESOURCE")) | ||
6374 | queueArn := notification.NewArn("aws", "dummy-service", "dummy-region", "dummy-accountid", "dummy-resource") | ||
6375 | |||
6376 | topicConfig := notification.NewConfig(topicArn) | ||
6377 | topicConfig.AddEvents(notification.ObjectCreatedAll, notification.ObjectRemovedAll) | ||
6378 | topicConfig.AddFilterSuffix("jpg") | ||
6379 | |||
6380 | queueConfig := notification.NewConfig(queueArn) | ||
6381 | queueConfig.AddEvents(notification.ObjectCreatedAll) | ||
6382 | queueConfig.AddFilterPrefix("photos/") | ||
6383 | |||
6384 | config := notification.Configuration{} | ||
6385 | config.AddTopic(topicConfig) | ||
6386 | |||
6387 | // Add the same topicConfig again, should have no effect | ||
6388 | // because it is duplicated | ||
6389 | config.AddTopic(topicConfig) | ||
6390 | if len(config.TopicConfigs) != 1 { | ||
6391 | logError(testName, function, args, startTime, "", "Duplicate entry added", err) | ||
6392 | return | ||
6393 | } | ||
6394 | |||
6395 | // Add and remove a queue config | ||
6396 | config.AddQueue(queueConfig) | ||
6397 | config.RemoveQueueByArn(queueArn) | ||
6398 | |||
6399 | err = c.SetBucketNotification(context.Background(), bucketName, config) | ||
6400 | if err != nil { | ||
6401 | logError(testName, function, args, startTime, "", "SetBucketNotification failed", err) | ||
6402 | return | ||
6403 | } | ||
6404 | |||
6405 | config, err = c.GetBucketNotification(context.Background(), bucketName) | ||
6406 | if err != nil { | ||
6407 | logError(testName, function, args, startTime, "", "GetBucketNotification failed", err) | ||
6408 | return | ||
6409 | } | ||
6410 | |||
6411 | if len(config.TopicConfigs) != 1 { | ||
6412 | logError(testName, function, args, startTime, "", "Topic config is empty", err) | ||
6413 | return | ||
6414 | } | ||
6415 | |||
6416 | if config.TopicConfigs[0].Filter.S3Key.FilterRules[0].Value != "jpg" { | ||
6417 | logError(testName, function, args, startTime, "", "Couldn't get the suffix", err) | ||
6418 | return | ||
6419 | } | ||
6420 | |||
6421 | err = c.RemoveAllBucketNotification(context.Background(), bucketName) | ||
6422 | if err != nil { | ||
6423 | logError(testName, function, args, startTime, "", "RemoveAllBucketNotification failed", err) | ||
6424 | return | ||
6425 | } | ||
6426 | |||
6427 | // Delete all objects and buckets | ||
6428 | if err = cleanupBucket(bucketName, c); err != nil { | ||
6429 | logError(testName, function, args, startTime, "", "CleanupBucket failed", err) | ||
6430 | return | ||
6431 | } | ||
6432 | |||
6433 | successLogger(testName, function, args, startTime).Info() | ||
6434 | } | ||
6435 | |||
6436 | // Tests comprehensive list of all methods. | ||
6437 | func testFunctional() { | ||
6438 | // initialize logging params | ||
6439 | startTime := time.Now() | ||
6440 | testName := getFuncName() | ||
6441 | function := "testFunctional()" | ||
6442 | functionAll := "" | ||
6443 | args := map[string]interface{}{} | ||
6444 | |||
6445 | // Seed random based on current time. | ||
6446 | rand.Seed(time.Now().Unix()) | ||
6447 | |||
6448 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
6449 | &minio.Options{ | ||
6450 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
6451 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
6452 | }) | ||
6453 | if err != nil { | ||
6454 | logError(testName, function, nil, startTime, "", "MinIO client object creation failed", err) | ||
6455 | return | ||
6456 | } | ||
6457 | |||
6458 | // Enable to debug | ||
6459 | // c.TraceOn(os.Stderr) | ||
6460 | |||
6461 | // Set user agent. | ||
6462 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
6463 | |||
6464 | // Generate a new random bucket name. | ||
6465 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
6466 | |||
6467 | // Make a new bucket. | ||
6468 | function = "MakeBucket(bucketName, region)" | ||
6469 | functionAll = "MakeBucket(bucketName, region)" | ||
6470 | args["bucketName"] = bucketName | ||
6471 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
6472 | |||
6473 | defer cleanupBucket(bucketName, c) | ||
6474 | if err != nil { | ||
6475 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
6476 | return | ||
6477 | } | ||
6478 | |||
6479 | // Generate a random file name. | ||
6480 | fileName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
6481 | file, err := os.Create(fileName) | ||
6482 | if err != nil { | ||
6483 | logError(testName, function, args, startTime, "", "File creation failed", err) | ||
6484 | return | ||
6485 | } | ||
6486 | for i := 0; i < 3; i++ { | ||
6487 | buf := make([]byte, rand.Intn(1<<19)) | ||
6488 | _, err = file.Write(buf) | ||
6489 | if err != nil { | ||
6490 | logError(testName, function, args, startTime, "", "File write failed", err) | ||
6491 | return | ||
6492 | } | ||
6493 | } | ||
6494 | file.Close() | ||
6495 | |||
6496 | // Verify if bucket exits and you have access. | ||
6497 | var exists bool | ||
6498 | function = "BucketExists(bucketName)" | ||
6499 | functionAll += ", " + function | ||
6500 | args = map[string]interface{}{ | ||
6501 | "bucketName": bucketName, | ||
6502 | } | ||
6503 | exists, err = c.BucketExists(context.Background(), bucketName) | ||
6504 | |||
6505 | if err != nil { | ||
6506 | logError(testName, function, args, startTime, "", "BucketExists failed", err) | ||
6507 | return | ||
6508 | } | ||
6509 | if !exists { | ||
6510 | logError(testName, function, args, startTime, "", "Could not find the bucket", err) | ||
6511 | return | ||
6512 | } | ||
6513 | |||
6514 | // Asserting the default bucket policy. | ||
6515 | function = "GetBucketPolicy(ctx, bucketName)" | ||
6516 | functionAll += ", " + function | ||
6517 | args = map[string]interface{}{ | ||
6518 | "bucketName": bucketName, | ||
6519 | } | ||
6520 | nilPolicy, err := c.GetBucketPolicy(context.Background(), bucketName) | ||
6521 | if err != nil { | ||
6522 | logError(testName, function, args, startTime, "", "GetBucketPolicy failed", err) | ||
6523 | return | ||
6524 | } | ||
6525 | if nilPolicy != "" { | ||
6526 | logError(testName, function, args, startTime, "", "policy should be set to nil", err) | ||
6527 | return | ||
6528 | } | ||
6529 | |||
6530 | // Set the bucket policy to 'public readonly'. | ||
6531 | function = "SetBucketPolicy(bucketName, readOnlyPolicy)" | ||
6532 | functionAll += ", " + function | ||
6533 | |||
6534 | readOnlyPolicy := `{"Version":"2012-10-17","Statement":[{"Effect":"Allow","Principal":{"AWS":["*"]},"Action":["s3:ListBucket"],"Resource":["arn:aws:s3:::` + bucketName + `"]}]}` | ||
6535 | args = map[string]interface{}{ | ||
6536 | "bucketName": bucketName, | ||
6537 | "bucketPolicy": readOnlyPolicy, | ||
6538 | } | ||
6539 | |||
6540 | err = c.SetBucketPolicy(context.Background(), bucketName, readOnlyPolicy) | ||
6541 | if err != nil { | ||
6542 | logError(testName, function, args, startTime, "", "SetBucketPolicy failed", err) | ||
6543 | return | ||
6544 | } | ||
6545 | // should return policy `readonly`. | ||
6546 | function = "GetBucketPolicy(ctx, bucketName)" | ||
6547 | functionAll += ", " + function | ||
6548 | args = map[string]interface{}{ | ||
6549 | "bucketName": bucketName, | ||
6550 | } | ||
6551 | _, err = c.GetBucketPolicy(context.Background(), bucketName) | ||
6552 | if err != nil { | ||
6553 | logError(testName, function, args, startTime, "", "GetBucketPolicy failed", err) | ||
6554 | return | ||
6555 | } | ||
6556 | |||
6557 | // Make the bucket 'public writeonly'. | ||
6558 | function = "SetBucketPolicy(bucketName, writeOnlyPolicy)" | ||
6559 | functionAll += ", " + function | ||
6560 | |||
6561 | writeOnlyPolicy := `{"Version":"2012-10-17","Statement":[{"Effect":"Allow","Principal":{"AWS":["*"]},"Action":["s3:ListBucketMultipartUploads"],"Resource":["arn:aws:s3:::` + bucketName + `"]}]}` | ||
6562 | args = map[string]interface{}{ | ||
6563 | "bucketName": bucketName, | ||
6564 | "bucketPolicy": writeOnlyPolicy, | ||
6565 | } | ||
6566 | err = c.SetBucketPolicy(context.Background(), bucketName, writeOnlyPolicy) | ||
6567 | |||
6568 | if err != nil { | ||
6569 | logError(testName, function, args, startTime, "", "SetBucketPolicy failed", err) | ||
6570 | return | ||
6571 | } | ||
6572 | // should return policy `writeonly`. | ||
6573 | function = "GetBucketPolicy(ctx, bucketName)" | ||
6574 | functionAll += ", " + function | ||
6575 | args = map[string]interface{}{ | ||
6576 | "bucketName": bucketName, | ||
6577 | } | ||
6578 | |||
6579 | _, err = c.GetBucketPolicy(context.Background(), bucketName) | ||
6580 | if err != nil { | ||
6581 | logError(testName, function, args, startTime, "", "GetBucketPolicy failed", err) | ||
6582 | return | ||
6583 | } | ||
6584 | |||
6585 | // Make the bucket 'public read/write'. | ||
6586 | function = "SetBucketPolicy(bucketName, readWritePolicy)" | ||
6587 | functionAll += ", " + function | ||
6588 | |||
6589 | readWritePolicy := `{"Version":"2012-10-17","Statement":[{"Effect":"Allow","Principal":{"AWS":["*"]},"Action":["s3:ListBucket","s3:ListBucketMultipartUploads"],"Resource":["arn:aws:s3:::` + bucketName + `"]}]}` | ||
6590 | |||
6591 | args = map[string]interface{}{ | ||
6592 | "bucketName": bucketName, | ||
6593 | "bucketPolicy": readWritePolicy, | ||
6594 | } | ||
6595 | err = c.SetBucketPolicy(context.Background(), bucketName, readWritePolicy) | ||
6596 | |||
6597 | if err != nil { | ||
6598 | logError(testName, function, args, startTime, "", "SetBucketPolicy failed", err) | ||
6599 | return | ||
6600 | } | ||
6601 | // should return policy `readwrite`. | ||
6602 | function = "GetBucketPolicy(bucketName)" | ||
6603 | functionAll += ", " + function | ||
6604 | args = map[string]interface{}{ | ||
6605 | "bucketName": bucketName, | ||
6606 | } | ||
6607 | _, err = c.GetBucketPolicy(context.Background(), bucketName) | ||
6608 | if err != nil { | ||
6609 | logError(testName, function, args, startTime, "", "GetBucketPolicy failed", err) | ||
6610 | return | ||
6611 | } | ||
6612 | |||
6613 | // List all buckets. | ||
6614 | function = "ListBuckets()" | ||
6615 | functionAll += ", " + function | ||
6616 | args = nil | ||
6617 | buckets, err := c.ListBuckets(context.Background()) | ||
6618 | |||
6619 | if len(buckets) == 0 { | ||
6620 | logError(testName, function, args, startTime, "", "Found bucket list to be empty", err) | ||
6621 | return | ||
6622 | } | ||
6623 | if err != nil { | ||
6624 | logError(testName, function, args, startTime, "", "ListBuckets failed", err) | ||
6625 | return | ||
6626 | } | ||
6627 | |||
6628 | // Verify if previously created bucket is listed in list buckets. | ||
6629 | bucketFound := false | ||
6630 | for _, bucket := range buckets { | ||
6631 | if bucket.Name == bucketName { | ||
6632 | bucketFound = true | ||
6633 | } | ||
6634 | } | ||
6635 | |||
6636 | // If bucket not found error out. | ||
6637 | if !bucketFound { | ||
6638 | logError(testName, function, args, startTime, "", "Bucket: "+bucketName+" not found", err) | ||
6639 | return | ||
6640 | } | ||
6641 | |||
6642 | objectName := bucketName + "unique" | ||
6643 | |||
6644 | // Generate data | ||
6645 | buf := bytes.Repeat([]byte("f"), 1<<19) | ||
6646 | |||
6647 | function = "PutObject(bucketName, objectName, reader, contentType)" | ||
6648 | functionAll += ", " + function | ||
6649 | args = map[string]interface{}{ | ||
6650 | "bucketName": bucketName, | ||
6651 | "objectName": objectName, | ||
6652 | "contentType": "", | ||
6653 | } | ||
6654 | |||
6655 | _, err = c.PutObject(context.Background(), bucketName, objectName, bytes.NewReader(buf), int64(len(buf)), minio.PutObjectOptions{}) | ||
6656 | if err != nil { | ||
6657 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
6658 | return | ||
6659 | } | ||
6660 | |||
6661 | args = map[string]interface{}{ | ||
6662 | "bucketName": bucketName, | ||
6663 | "objectName": objectName + "-nolength", | ||
6664 | "contentType": "binary/octet-stream", | ||
6665 | } | ||
6666 | |||
6667 | _, err = c.PutObject(context.Background(), bucketName, objectName+"-nolength", bytes.NewReader(buf), int64(len(buf)), minio.PutObjectOptions{ContentType: "binary/octet-stream"}) | ||
6668 | if err != nil { | ||
6669 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
6670 | return | ||
6671 | } | ||
6672 | |||
6673 | // Instantiate a done channel to close all listing. | ||
6674 | doneCh := make(chan struct{}) | ||
6675 | defer close(doneCh) | ||
6676 | |||
6677 | objFound := false | ||
6678 | isRecursive := true // Recursive is true. | ||
6679 | |||
6680 | function = "ListObjects(bucketName, objectName, isRecursive, doneCh)" | ||
6681 | functionAll += ", " + function | ||
6682 | args = map[string]interface{}{ | ||
6683 | "bucketName": bucketName, | ||
6684 | "objectName": objectName, | ||
6685 | "isRecursive": isRecursive, | ||
6686 | } | ||
6687 | |||
6688 | for obj := range c.ListObjects(context.Background(), bucketName, minio.ListObjectsOptions{UseV1: true, Prefix: objectName, Recursive: true}) { | ||
6689 | if obj.Key == objectName { | ||
6690 | objFound = true | ||
6691 | break | ||
6692 | } | ||
6693 | } | ||
6694 | if !objFound { | ||
6695 | logError(testName, function, args, startTime, "", "Object "+objectName+" not found", err) | ||
6696 | return | ||
6697 | } | ||
6698 | |||
6699 | objFound = false | ||
6700 | isRecursive = true // Recursive is true. | ||
6701 | function = "ListObjects()" | ||
6702 | functionAll += ", " + function | ||
6703 | args = map[string]interface{}{ | ||
6704 | "bucketName": bucketName, | ||
6705 | "objectName": objectName, | ||
6706 | "isRecursive": isRecursive, | ||
6707 | } | ||
6708 | |||
6709 | for obj := range c.ListObjects(context.Background(), bucketName, minio.ListObjectsOptions{Prefix: objectName, Recursive: isRecursive}) { | ||
6710 | if obj.Key == objectName { | ||
6711 | objFound = true | ||
6712 | break | ||
6713 | } | ||
6714 | } | ||
6715 | if !objFound { | ||
6716 | logError(testName, function, args, startTime, "", "Object "+objectName+" not found", err) | ||
6717 | return | ||
6718 | } | ||
6719 | |||
6720 | incompObjNotFound := true | ||
6721 | |||
6722 | function = "ListIncompleteUploads(bucketName, objectName, isRecursive, doneCh)" | ||
6723 | functionAll += ", " + function | ||
6724 | args = map[string]interface{}{ | ||
6725 | "bucketName": bucketName, | ||
6726 | "objectName": objectName, | ||
6727 | "isRecursive": isRecursive, | ||
6728 | } | ||
6729 | |||
6730 | for objIncompl := range c.ListIncompleteUploads(context.Background(), bucketName, objectName, isRecursive) { | ||
6731 | if objIncompl.Key != "" { | ||
6732 | incompObjNotFound = false | ||
6733 | break | ||
6734 | } | ||
6735 | } | ||
6736 | if !incompObjNotFound { | ||
6737 | logError(testName, function, args, startTime, "", "Unexpected dangling incomplete upload found", err) | ||
6738 | return | ||
6739 | } | ||
6740 | |||
6741 | function = "GetObject(bucketName, objectName)" | ||
6742 | functionAll += ", " + function | ||
6743 | args = map[string]interface{}{ | ||
6744 | "bucketName": bucketName, | ||
6745 | "objectName": objectName, | ||
6746 | } | ||
6747 | newReader, err := c.GetObject(context.Background(), bucketName, objectName, minio.GetObjectOptions{}) | ||
6748 | if err != nil { | ||
6749 | logError(testName, function, args, startTime, "", "GetObject failed", err) | ||
6750 | return | ||
6751 | } | ||
6752 | |||
6753 | newReadBytes, err := io.ReadAll(newReader) | ||
6754 | if err != nil { | ||
6755 | logError(testName, function, args, startTime, "", "ReadAll failed", err) | ||
6756 | return | ||
6757 | } | ||
6758 | |||
6759 | if !bytes.Equal(newReadBytes, buf) { | ||
6760 | logError(testName, function, args, startTime, "", "GetObject bytes mismatch", err) | ||
6761 | return | ||
6762 | } | ||
6763 | newReader.Close() | ||
6764 | |||
6765 | function = "FGetObject(bucketName, objectName, fileName)" | ||
6766 | functionAll += ", " + function | ||
6767 | args = map[string]interface{}{ | ||
6768 | "bucketName": bucketName, | ||
6769 | "objectName": objectName, | ||
6770 | "fileName": fileName + "-f", | ||
6771 | } | ||
6772 | err = c.FGetObject(context.Background(), bucketName, objectName, fileName+"-f", minio.GetObjectOptions{}) | ||
6773 | |||
6774 | if err != nil { | ||
6775 | logError(testName, function, args, startTime, "", "FGetObject failed", err) | ||
6776 | return | ||
6777 | } | ||
6778 | |||
6779 | function = "PresignedHeadObject(bucketName, objectName, expires, reqParams)" | ||
6780 | functionAll += ", " + function | ||
6781 | args = map[string]interface{}{ | ||
6782 | "bucketName": bucketName, | ||
6783 | "objectName": "", | ||
6784 | "expires": 3600 * time.Second, | ||
6785 | } | ||
6786 | if _, err = c.PresignedHeadObject(context.Background(), bucketName, "", 3600*time.Second, nil); err == nil { | ||
6787 | logError(testName, function, args, startTime, "", "PresignedHeadObject success", err) | ||
6788 | return | ||
6789 | } | ||
6790 | |||
6791 | // Generate presigned HEAD object url. | ||
6792 | function = "PresignedHeadObject(bucketName, objectName, expires, reqParams)" | ||
6793 | functionAll += ", " + function | ||
6794 | args = map[string]interface{}{ | ||
6795 | "bucketName": bucketName, | ||
6796 | "objectName": objectName, | ||
6797 | "expires": 3600 * time.Second, | ||
6798 | } | ||
6799 | presignedHeadURL, err := c.PresignedHeadObject(context.Background(), bucketName, objectName, 3600*time.Second, nil) | ||
6800 | if err != nil { | ||
6801 | logError(testName, function, args, startTime, "", "PresignedHeadObject failed", err) | ||
6802 | return | ||
6803 | } | ||
6804 | |||
6805 | transport, err := minio.DefaultTransport(mustParseBool(os.Getenv(enableHTTPS))) | ||
6806 | if err != nil { | ||
6807 | logError(testName, function, args, startTime, "", "DefaultTransport failed", err) | ||
6808 | return | ||
6809 | } | ||
6810 | |||
6811 | httpClient := &http.Client{ | ||
6812 | // Setting a sensible time out of 30secs to wait for response | ||
6813 | // headers. Request is pro-actively canceled after 30secs | ||
6814 | // with no response. | ||
6815 | Timeout: 30 * time.Second, | ||
6816 | Transport: transport, | ||
6817 | } | ||
6818 | |||
6819 | req, err := http.NewRequest(http.MethodHead, presignedHeadURL.String(), nil) | ||
6820 | if err != nil { | ||
6821 | logError(testName, function, args, startTime, "", "PresignedHeadObject request was incorrect", err) | ||
6822 | return | ||
6823 | } | ||
6824 | |||
6825 | // Verify if presigned url works. | ||
6826 | resp, err := httpClient.Do(req) | ||
6827 | if err != nil { | ||
6828 | logError(testName, function, args, startTime, "", "PresignedHeadObject response incorrect", err) | ||
6829 | return | ||
6830 | } | ||
6831 | if resp.StatusCode != http.StatusOK { | ||
6832 | logError(testName, function, args, startTime, "", "PresignedHeadObject response incorrect, status "+string(resp.StatusCode), err) | ||
6833 | return | ||
6834 | } | ||
6835 | if resp.Header.Get("ETag") == "" { | ||
6836 | logError(testName, function, args, startTime, "", "PresignedHeadObject response incorrect", err) | ||
6837 | return | ||
6838 | } | ||
6839 | resp.Body.Close() | ||
6840 | |||
6841 | function = "PresignedGetObject(bucketName, objectName, expires, reqParams)" | ||
6842 | functionAll += ", " + function | ||
6843 | args = map[string]interface{}{ | ||
6844 | "bucketName": bucketName, | ||
6845 | "objectName": "", | ||
6846 | "expires": 3600 * time.Second, | ||
6847 | } | ||
6848 | _, err = c.PresignedGetObject(context.Background(), bucketName, "", 3600*time.Second, nil) | ||
6849 | if err == nil { | ||
6850 | logError(testName, function, args, startTime, "", "PresignedGetObject success", err) | ||
6851 | return | ||
6852 | } | ||
6853 | |||
6854 | // Generate presigned GET object url. | ||
6855 | function = "PresignedGetObject(bucketName, objectName, expires, reqParams)" | ||
6856 | functionAll += ", " + function | ||
6857 | args = map[string]interface{}{ | ||
6858 | "bucketName": bucketName, | ||
6859 | "objectName": objectName, | ||
6860 | "expires": 3600 * time.Second, | ||
6861 | } | ||
6862 | presignedGetURL, err := c.PresignedGetObject(context.Background(), bucketName, objectName, 3600*time.Second, nil) | ||
6863 | if err != nil { | ||
6864 | logError(testName, function, args, startTime, "", "PresignedGetObject failed", err) | ||
6865 | return | ||
6866 | } | ||
6867 | |||
6868 | // Verify if presigned url works. | ||
6869 | req, err = http.NewRequest(http.MethodGet, presignedGetURL.String(), nil) | ||
6870 | if err != nil { | ||
6871 | logError(testName, function, args, startTime, "", "PresignedGetObject request incorrect", err) | ||
6872 | return | ||
6873 | } | ||
6874 | |||
6875 | resp, err = httpClient.Do(req) | ||
6876 | if err != nil { | ||
6877 | logError(testName, function, args, startTime, "", "PresignedGetObject response incorrect", err) | ||
6878 | return | ||
6879 | } | ||
6880 | if resp.StatusCode != http.StatusOK { | ||
6881 | logError(testName, function, args, startTime, "", "PresignedGetObject response incorrect, status "+string(resp.StatusCode), err) | ||
6882 | return | ||
6883 | } | ||
6884 | newPresignedBytes, err := io.ReadAll(resp.Body) | ||
6885 | if err != nil { | ||
6886 | logError(testName, function, args, startTime, "", "PresignedGetObject response incorrect", err) | ||
6887 | return | ||
6888 | } | ||
6889 | resp.Body.Close() | ||
6890 | if !bytes.Equal(newPresignedBytes, buf) { | ||
6891 | logError(testName, function, args, startTime, "", "PresignedGetObject response incorrect", err) | ||
6892 | return | ||
6893 | } | ||
6894 | |||
6895 | // Set request parameters. | ||
6896 | reqParams := make(url.Values) | ||
6897 | reqParams.Set("response-content-disposition", "attachment; filename=\"test.txt\"") | ||
6898 | args = map[string]interface{}{ | ||
6899 | "bucketName": bucketName, | ||
6900 | "objectName": objectName, | ||
6901 | "expires": 3600 * time.Second, | ||
6902 | "reqParams": reqParams, | ||
6903 | } | ||
6904 | presignedGetURL, err = c.PresignedGetObject(context.Background(), bucketName, objectName, 3600*time.Second, reqParams) | ||
6905 | |||
6906 | if err != nil { | ||
6907 | logError(testName, function, args, startTime, "", "PresignedGetObject failed", err) | ||
6908 | return | ||
6909 | } | ||
6910 | |||
6911 | // Verify if presigned url works. | ||
6912 | req, err = http.NewRequest(http.MethodGet, presignedGetURL.String(), nil) | ||
6913 | if err != nil { | ||
6914 | logError(testName, function, args, startTime, "", "PresignedGetObject request incorrect", err) | ||
6915 | return | ||
6916 | } | ||
6917 | |||
6918 | resp, err = httpClient.Do(req) | ||
6919 | if err != nil { | ||
6920 | logError(testName, function, args, startTime, "", "PresignedGetObject response incorrect", err) | ||
6921 | return | ||
6922 | } | ||
6923 | if resp.StatusCode != http.StatusOK { | ||
6924 | logError(testName, function, args, startTime, "", "PresignedGetObject response incorrect, status "+string(resp.StatusCode), err) | ||
6925 | return | ||
6926 | } | ||
6927 | newPresignedBytes, err = io.ReadAll(resp.Body) | ||
6928 | if err != nil { | ||
6929 | logError(testName, function, args, startTime, "", "PresignedGetObject response incorrect", err) | ||
6930 | return | ||
6931 | } | ||
6932 | if !bytes.Equal(newPresignedBytes, buf) { | ||
6933 | logError(testName, function, args, startTime, "", "Bytes mismatch for presigned GET URL", err) | ||
6934 | return | ||
6935 | } | ||
6936 | if resp.Header.Get("Content-Disposition") != "attachment; filename=\"test.txt\"" { | ||
6937 | logError(testName, function, args, startTime, "", "wrong Content-Disposition received "+string(resp.Header.Get("Content-Disposition")), err) | ||
6938 | return | ||
6939 | } | ||
6940 | |||
6941 | function = "PresignedPutObject(bucketName, objectName, expires)" | ||
6942 | functionAll += ", " + function | ||
6943 | args = map[string]interface{}{ | ||
6944 | "bucketName": bucketName, | ||
6945 | "objectName": "", | ||
6946 | "expires": 3600 * time.Second, | ||
6947 | } | ||
6948 | _, err = c.PresignedPutObject(context.Background(), bucketName, "", 3600*time.Second) | ||
6949 | if err == nil { | ||
6950 | logError(testName, function, args, startTime, "", "PresignedPutObject success", err) | ||
6951 | return | ||
6952 | } | ||
6953 | |||
6954 | function = "PresignedPutObject(bucketName, objectName, expires)" | ||
6955 | functionAll += ", " + function | ||
6956 | args = map[string]interface{}{ | ||
6957 | "bucketName": bucketName, | ||
6958 | "objectName": objectName + "-presigned", | ||
6959 | "expires": 3600 * time.Second, | ||
6960 | } | ||
6961 | presignedPutURL, err := c.PresignedPutObject(context.Background(), bucketName, objectName+"-presigned", 3600*time.Second) | ||
6962 | if err != nil { | ||
6963 | logError(testName, function, args, startTime, "", "PresignedPutObject failed", err) | ||
6964 | return | ||
6965 | } | ||
6966 | |||
6967 | buf = bytes.Repeat([]byte("g"), 1<<19) | ||
6968 | |||
6969 | req, err = http.NewRequest(http.MethodPut, presignedPutURL.String(), bytes.NewReader(buf)) | ||
6970 | if err != nil { | ||
6971 | logError(testName, function, args, startTime, "", "Couldn't make HTTP request with PresignedPutObject URL", err) | ||
6972 | return | ||
6973 | } | ||
6974 | |||
6975 | resp, err = httpClient.Do(req) | ||
6976 | if err != nil { | ||
6977 | logError(testName, function, args, startTime, "", "PresignedPutObject failed", err) | ||
6978 | return | ||
6979 | } | ||
6980 | |||
6981 | newReader, err = c.GetObject(context.Background(), bucketName, objectName+"-presigned", minio.GetObjectOptions{}) | ||
6982 | if err != nil { | ||
6983 | logError(testName, function, args, startTime, "", "GetObject after PresignedPutObject failed", err) | ||
6984 | return | ||
6985 | } | ||
6986 | |||
6987 | newReadBytes, err = io.ReadAll(newReader) | ||
6988 | if err != nil { | ||
6989 | logError(testName, function, args, startTime, "", "ReadAll after GetObject failed", err) | ||
6990 | return | ||
6991 | } | ||
6992 | |||
6993 | if !bytes.Equal(newReadBytes, buf) { | ||
6994 | logError(testName, function, args, startTime, "", "Bytes mismatch", err) | ||
6995 | return | ||
6996 | } | ||
6997 | |||
6998 | function = "PresignHeader(method, bucketName, objectName, expires, reqParams, extraHeaders)" | ||
6999 | functionAll += ", " + function | ||
7000 | presignExtraHeaders := map[string][]string{ | ||
7001 | "mysecret": {"abcxxx"}, | ||
7002 | } | ||
7003 | args = map[string]interface{}{ | ||
7004 | "method": "PUT", | ||
7005 | "bucketName": bucketName, | ||
7006 | "objectName": objectName + "-presign-custom", | ||
7007 | "expires": 3600 * time.Second, | ||
7008 | "extraHeaders": presignExtraHeaders, | ||
7009 | } | ||
7010 | presignedURL, err := c.PresignHeader(context.Background(), "PUT", bucketName, objectName+"-presign-custom", 3600*time.Second, nil, presignExtraHeaders) | ||
7011 | if err != nil { | ||
7012 | logError(testName, function, args, startTime, "", "Presigned failed", err) | ||
7013 | return | ||
7014 | } | ||
7015 | |||
7016 | // Generate data more than 32K | ||
7017 | buf = bytes.Repeat([]byte("1"), rand.Intn(1<<10)+32*1024) | ||
7018 | |||
7019 | req, err = http.NewRequest(http.MethodPut, presignedURL.String(), bytes.NewReader(buf)) | ||
7020 | if err != nil { | ||
7021 | logError(testName, function, args, startTime, "", "HTTP request to Presigned URL failed", err) | ||
7022 | return | ||
7023 | } | ||
7024 | |||
7025 | req.Header.Add("mysecret", "abcxxx") | ||
7026 | resp, err = httpClient.Do(req) | ||
7027 | if err != nil { | ||
7028 | logError(testName, function, args, startTime, "", "HTTP request to Presigned URL failed", err) | ||
7029 | return | ||
7030 | } | ||
7031 | |||
7032 | // Download the uploaded object to verify | ||
7033 | args = map[string]interface{}{ | ||
7034 | "bucketName": bucketName, | ||
7035 | "objectName": objectName + "-presign-custom", | ||
7036 | } | ||
7037 | newReader, err = c.GetObject(context.Background(), bucketName, objectName+"-presign-custom", minio.GetObjectOptions{}) | ||
7038 | if err != nil { | ||
7039 | logError(testName, function, args, startTime, "", "GetObject of uploaded custom-presigned object failed", err) | ||
7040 | return | ||
7041 | } | ||
7042 | |||
7043 | newReadBytes, err = io.ReadAll(newReader) | ||
7044 | if err != nil { | ||
7045 | logError(testName, function, args, startTime, "", "ReadAll failed during get on custom-presigned put object", err) | ||
7046 | return | ||
7047 | } | ||
7048 | newReader.Close() | ||
7049 | |||
7050 | if !bytes.Equal(newReadBytes, buf) { | ||
7051 | logError(testName, function, args, startTime, "", "Bytes mismatch on custom-presigned object upload verification", err) | ||
7052 | return | ||
7053 | } | ||
7054 | |||
7055 | function = "RemoveObject(bucketName, objectName)" | ||
7056 | functionAll += ", " + function | ||
7057 | args = map[string]interface{}{ | ||
7058 | "bucketName": bucketName, | ||
7059 | "objectName": objectName, | ||
7060 | } | ||
7061 | err = c.RemoveObject(context.Background(), bucketName, objectName, minio.RemoveObjectOptions{}) | ||
7062 | |||
7063 | if err != nil { | ||
7064 | logError(testName, function, args, startTime, "", "RemoveObject failed", err) | ||
7065 | return | ||
7066 | } | ||
7067 | args["objectName"] = objectName + "-f" | ||
7068 | err = c.RemoveObject(context.Background(), bucketName, objectName+"-f", minio.RemoveObjectOptions{}) | ||
7069 | |||
7070 | if err != nil { | ||
7071 | logError(testName, function, args, startTime, "", "RemoveObject failed", err) | ||
7072 | return | ||
7073 | } | ||
7074 | |||
7075 | args["objectName"] = objectName + "-nolength" | ||
7076 | err = c.RemoveObject(context.Background(), bucketName, objectName+"-nolength", minio.RemoveObjectOptions{}) | ||
7077 | |||
7078 | if err != nil { | ||
7079 | logError(testName, function, args, startTime, "", "RemoveObject failed", err) | ||
7080 | return | ||
7081 | } | ||
7082 | |||
7083 | args["objectName"] = objectName + "-presigned" | ||
7084 | err = c.RemoveObject(context.Background(), bucketName, objectName+"-presigned", minio.RemoveObjectOptions{}) | ||
7085 | |||
7086 | if err != nil { | ||
7087 | logError(testName, function, args, startTime, "", "RemoveObject failed", err) | ||
7088 | return | ||
7089 | } | ||
7090 | |||
7091 | args["objectName"] = objectName + "-presign-custom" | ||
7092 | err = c.RemoveObject(context.Background(), bucketName, objectName+"-presign-custom", minio.RemoveObjectOptions{}) | ||
7093 | |||
7094 | if err != nil { | ||
7095 | logError(testName, function, args, startTime, "", "RemoveObject failed", err) | ||
7096 | return | ||
7097 | } | ||
7098 | |||
7099 | function = "RemoveBucket(bucketName)" | ||
7100 | functionAll += ", " + function | ||
7101 | args = map[string]interface{}{ | ||
7102 | "bucketName": bucketName, | ||
7103 | } | ||
7104 | err = c.RemoveBucket(context.Background(), bucketName) | ||
7105 | |||
7106 | if err != nil { | ||
7107 | logError(testName, function, args, startTime, "", "RemoveBucket failed", err) | ||
7108 | return | ||
7109 | } | ||
7110 | err = c.RemoveBucket(context.Background(), bucketName) | ||
7111 | if err == nil { | ||
7112 | logError(testName, function, args, startTime, "", "RemoveBucket did not fail for invalid bucket name", err) | ||
7113 | return | ||
7114 | } | ||
7115 | if err.Error() != "The specified bucket does not exist" { | ||
7116 | logError(testName, function, args, startTime, "", "RemoveBucket failed", err) | ||
7117 | return | ||
7118 | } | ||
7119 | |||
7120 | os.Remove(fileName) | ||
7121 | os.Remove(fileName + "-f") | ||
7122 | successLogger(testName, functionAll, args, startTime).Info() | ||
7123 | } | ||
7124 | |||
7125 | // Test for validating GetObject Reader* methods functioning when the | ||
7126 | // object is modified in the object store. | ||
7127 | func testGetObjectModified() { | ||
7128 | // initialize logging params | ||
7129 | startTime := time.Now() | ||
7130 | testName := getFuncName() | ||
7131 | function := "GetObject(bucketName, objectName)" | ||
7132 | args := map[string]interface{}{} | ||
7133 | |||
7134 | // Instantiate new minio client object. | ||
7135 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
7136 | &minio.Options{ | ||
7137 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
7138 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
7139 | }) | ||
7140 | if err != nil { | ||
7141 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
7142 | return | ||
7143 | } | ||
7144 | |||
7145 | // Enable tracing, write to stderr. | ||
7146 | // c.TraceOn(os.Stderr) | ||
7147 | |||
7148 | // Set user agent. | ||
7149 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
7150 | |||
7151 | // Make a new bucket. | ||
7152 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
7153 | args["bucketName"] = bucketName | ||
7154 | |||
7155 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
7156 | if err != nil { | ||
7157 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
7158 | return | ||
7159 | } | ||
7160 | |||
7161 | defer cleanupBucket(bucketName, c) | ||
7162 | |||
7163 | // Upload an object. | ||
7164 | objectName := "myobject" | ||
7165 | args["objectName"] = objectName | ||
7166 | content := "helloworld" | ||
7167 | _, err = c.PutObject(context.Background(), bucketName, objectName, strings.NewReader(content), int64(len(content)), minio.PutObjectOptions{ContentType: "application/text"}) | ||
7168 | if err != nil { | ||
7169 | logError(testName, function, args, startTime, "", "Failed to upload "+objectName+", to bucket "+bucketName, err) | ||
7170 | return | ||
7171 | } | ||
7172 | |||
7173 | defer c.RemoveObject(context.Background(), bucketName, objectName, minio.RemoveObjectOptions{}) | ||
7174 | |||
7175 | reader, err := c.GetObject(context.Background(), bucketName, objectName, minio.GetObjectOptions{}) | ||
7176 | if err != nil { | ||
7177 | logError(testName, function, args, startTime, "", "Failed to GetObject "+objectName+", from bucket "+bucketName, err) | ||
7178 | return | ||
7179 | } | ||
7180 | defer reader.Close() | ||
7181 | |||
7182 | // Read a few bytes of the object. | ||
7183 | b := make([]byte, 5) | ||
7184 | n, err := reader.ReadAt(b, 0) | ||
7185 | if err != nil { | ||
7186 | logError(testName, function, args, startTime, "", "Failed to read object "+objectName+", from bucket "+bucketName+" at an offset", err) | ||
7187 | return | ||
7188 | } | ||
7189 | |||
7190 | // Upload different contents to the same object while object is being read. | ||
7191 | newContent := "goodbyeworld" | ||
7192 | _, err = c.PutObject(context.Background(), bucketName, objectName, strings.NewReader(newContent), int64(len(newContent)), minio.PutObjectOptions{ContentType: "application/text"}) | ||
7193 | if err != nil { | ||
7194 | logError(testName, function, args, startTime, "", "Failed to upload "+objectName+", to bucket "+bucketName, err) | ||
7195 | return | ||
7196 | } | ||
7197 | |||
7198 | // Confirm that a Stat() call in between doesn't change the Object's cached etag. | ||
7199 | _, err = reader.Stat() | ||
7200 | expectedError := "At least one of the pre-conditions you specified did not hold" | ||
7201 | if err.Error() != expectedError { | ||
7202 | logError(testName, function, args, startTime, "", "Expected Stat to fail with error "+expectedError+", but received "+err.Error(), err) | ||
7203 | return | ||
7204 | } | ||
7205 | |||
7206 | // Read again only to find object contents have been modified since last read. | ||
7207 | _, err = reader.ReadAt(b, int64(n)) | ||
7208 | if err.Error() != expectedError { | ||
7209 | logError(testName, function, args, startTime, "", "Expected ReadAt to fail with error "+expectedError+", but received "+err.Error(), err) | ||
7210 | return | ||
7211 | } | ||
7212 | |||
7213 | successLogger(testName, function, args, startTime).Info() | ||
7214 | } | ||
7215 | |||
7216 | // Test validates putObject to upload a file seeked at a given offset. | ||
7217 | func testPutObjectUploadSeekedObject() { | ||
7218 | // initialize logging params | ||
7219 | startTime := time.Now() | ||
7220 | testName := getFuncName() | ||
7221 | function := "PutObject(bucketName, objectName, fileToUpload, contentType)" | ||
7222 | args := map[string]interface{}{ | ||
7223 | "bucketName": "", | ||
7224 | "objectName": "", | ||
7225 | "fileToUpload": "", | ||
7226 | "contentType": "binary/octet-stream", | ||
7227 | } | ||
7228 | |||
7229 | // Instantiate new minio client object. | ||
7230 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
7231 | &minio.Options{ | ||
7232 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
7233 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
7234 | }) | ||
7235 | if err != nil { | ||
7236 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
7237 | return | ||
7238 | } | ||
7239 | |||
7240 | // Enable tracing, write to stderr. | ||
7241 | // c.TraceOn(os.Stderr) | ||
7242 | |||
7243 | // Set user agent. | ||
7244 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
7245 | |||
7246 | // Make a new bucket. | ||
7247 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
7248 | args["bucketName"] = bucketName | ||
7249 | |||
7250 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
7251 | if err != nil { | ||
7252 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
7253 | return | ||
7254 | } | ||
7255 | defer cleanupBucket(bucketName, c) | ||
7256 | |||
7257 | var tempfile *os.File | ||
7258 | |||
7259 | if fileName := getMintDataDirFilePath("datafile-100-kB"); fileName != "" { | ||
7260 | tempfile, err = os.Open(fileName) | ||
7261 | if err != nil { | ||
7262 | logError(testName, function, args, startTime, "", "File open failed", err) | ||
7263 | return | ||
7264 | } | ||
7265 | args["fileToUpload"] = fileName | ||
7266 | } else { | ||
7267 | tempfile, err = os.CreateTemp("", "minio-go-upload-test-") | ||
7268 | if err != nil { | ||
7269 | logError(testName, function, args, startTime, "", "TempFile create failed", err) | ||
7270 | return | ||
7271 | } | ||
7272 | args["fileToUpload"] = tempfile.Name() | ||
7273 | |||
7274 | // Generate 100kB data | ||
7275 | if _, err = io.Copy(tempfile, getDataReader("datafile-100-kB")); err != nil { | ||
7276 | logError(testName, function, args, startTime, "", "File copy failed", err) | ||
7277 | return | ||
7278 | } | ||
7279 | |||
7280 | defer os.Remove(tempfile.Name()) | ||
7281 | |||
7282 | // Seek back to the beginning of the file. | ||
7283 | tempfile.Seek(0, 0) | ||
7284 | } | ||
7285 | length := 100 * humanize.KiByte | ||
7286 | objectName := fmt.Sprintf("test-file-%v", rand.Uint32()) | ||
7287 | args["objectName"] = objectName | ||
7288 | |||
7289 | offset := length / 2 | ||
7290 | if _, err = tempfile.Seek(int64(offset), 0); err != nil { | ||
7291 | logError(testName, function, args, startTime, "", "TempFile seek failed", err) | ||
7292 | return | ||
7293 | } | ||
7294 | |||
7295 | _, err = c.PutObject(context.Background(), bucketName, objectName, tempfile, int64(length-offset), minio.PutObjectOptions{ContentType: "binary/octet-stream"}) | ||
7296 | if err != nil { | ||
7297 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
7298 | return | ||
7299 | } | ||
7300 | tempfile.Close() | ||
7301 | |||
7302 | obj, err := c.GetObject(context.Background(), bucketName, objectName, minio.GetObjectOptions{}) | ||
7303 | if err != nil { | ||
7304 | logError(testName, function, args, startTime, "", "GetObject failed", err) | ||
7305 | return | ||
7306 | } | ||
7307 | defer obj.Close() | ||
7308 | |||
7309 | n, err := obj.Seek(int64(offset), 0) | ||
7310 | if err != nil { | ||
7311 | logError(testName, function, args, startTime, "", "Seek failed", err) | ||
7312 | return | ||
7313 | } | ||
7314 | if n != int64(offset) { | ||
7315 | logError(testName, function, args, startTime, "", fmt.Sprintf("Invalid offset returned, expected %d got %d", int64(offset), n), err) | ||
7316 | return | ||
7317 | } | ||
7318 | |||
7319 | _, err = c.PutObject(context.Background(), bucketName, objectName+"getobject", obj, int64(length-offset), minio.PutObjectOptions{ContentType: "binary/octet-stream"}) | ||
7320 | if err != nil { | ||
7321 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
7322 | return | ||
7323 | } | ||
7324 | st, err := c.StatObject(context.Background(), bucketName, objectName+"getobject", minio.StatObjectOptions{}) | ||
7325 | if err != nil { | ||
7326 | logError(testName, function, args, startTime, "", "StatObject failed", err) | ||
7327 | return | ||
7328 | } | ||
7329 | if st.Size != int64(length-offset) { | ||
7330 | logError(testName, function, args, startTime, "", fmt.Sprintf("Invalid offset returned, expected %d got %d", int64(length-offset), n), err) | ||
7331 | return | ||
7332 | } | ||
7333 | |||
7334 | successLogger(testName, function, args, startTime).Info() | ||
7335 | } | ||
7336 | |||
7337 | // Tests bucket re-create errors. | ||
7338 | func testMakeBucketErrorV2() { | ||
7339 | // initialize logging params | ||
7340 | startTime := time.Now() | ||
7341 | testName := getFuncName() | ||
7342 | function := "MakeBucket(bucketName, region)" | ||
7343 | args := map[string]interface{}{ | ||
7344 | "bucketName": "", | ||
7345 | "region": "eu-west-1", | ||
7346 | } | ||
7347 | |||
7348 | // Seed random based on current time. | ||
7349 | rand.Seed(time.Now().Unix()) | ||
7350 | |||
7351 | // Instantiate new minio client object. | ||
7352 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
7353 | &minio.Options{ | ||
7354 | Creds: credentials.NewStaticV2(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
7355 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
7356 | }) | ||
7357 | if err != nil { | ||
7358 | logError(testName, function, args, startTime, "", "MinIO v2 client object creation failed", err) | ||
7359 | return | ||
7360 | } | ||
7361 | |||
7362 | // Enable tracing, write to stderr. | ||
7363 | // c.TraceOn(os.Stderr) | ||
7364 | |||
7365 | // Set user agent. | ||
7366 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
7367 | |||
7368 | // Generate a new random bucket name. | ||
7369 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
7370 | region := "eu-west-1" | ||
7371 | args["bucketName"] = bucketName | ||
7372 | args["region"] = region | ||
7373 | |||
7374 | // Make a new bucket in 'eu-west-1'. | ||
7375 | if err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: region}); err != nil { | ||
7376 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
7377 | return | ||
7378 | } | ||
7379 | |||
7380 | defer cleanupBucket(bucketName, c) | ||
7381 | |||
7382 | if err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: region}); err == nil { | ||
7383 | logError(testName, function, args, startTime, "", "MakeBucket did not fail for existing bucket name", err) | ||
7384 | return | ||
7385 | } | ||
7386 | // Verify valid error response from server. | ||
7387 | if minio.ToErrorResponse(err).Code != "BucketAlreadyExists" && | ||
7388 | minio.ToErrorResponse(err).Code != "BucketAlreadyOwnedByYou" { | ||
7389 | logError(testName, function, args, startTime, "", "Invalid error returned by server", err) | ||
7390 | return | ||
7391 | } | ||
7392 | |||
7393 | successLogger(testName, function, args, startTime).Info() | ||
7394 | } | ||
7395 | |||
7396 | // Test get object reader to not throw error on being closed twice. | ||
7397 | func testGetObjectClosedTwiceV2() { | ||
7398 | // initialize logging params | ||
7399 | startTime := time.Now() | ||
7400 | testName := getFuncName() | ||
7401 | function := "MakeBucket(bucketName, region)" | ||
7402 | args := map[string]interface{}{ | ||
7403 | "bucketName": "", | ||
7404 | "region": "eu-west-1", | ||
7405 | } | ||
7406 | |||
7407 | // Seed random based on current time. | ||
7408 | rand.Seed(time.Now().Unix()) | ||
7409 | |||
7410 | // Instantiate new minio client object. | ||
7411 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
7412 | &minio.Options{ | ||
7413 | Creds: credentials.NewStaticV2(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
7414 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
7415 | }) | ||
7416 | if err != nil { | ||
7417 | logError(testName, function, args, startTime, "", "MinIO v2 client object creation failed", err) | ||
7418 | return | ||
7419 | } | ||
7420 | |||
7421 | // Enable tracing, write to stderr. | ||
7422 | // c.TraceOn(os.Stderr) | ||
7423 | |||
7424 | // Set user agent. | ||
7425 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
7426 | |||
7427 | // Generate a new random bucket name. | ||
7428 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
7429 | args["bucketName"] = bucketName | ||
7430 | |||
7431 | // Make a new bucket. | ||
7432 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
7433 | if err != nil { | ||
7434 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
7435 | return | ||
7436 | } | ||
7437 | |||
7438 | defer cleanupBucket(bucketName, c) | ||
7439 | |||
7440 | // Generate 33K of data. | ||
7441 | bufSize := dataFileMap["datafile-33-kB"] | ||
7442 | reader := getDataReader("datafile-33-kB") | ||
7443 | defer reader.Close() | ||
7444 | |||
7445 | // Save the data | ||
7446 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
7447 | args["objectName"] = objectName | ||
7448 | |||
7449 | _, err = c.PutObject(context.Background(), bucketName, objectName, reader, int64(bufSize), minio.PutObjectOptions{ContentType: "binary/octet-stream"}) | ||
7450 | if err != nil { | ||
7451 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
7452 | return | ||
7453 | } | ||
7454 | |||
7455 | // Read the data back | ||
7456 | r, err := c.GetObject(context.Background(), bucketName, objectName, minio.GetObjectOptions{}) | ||
7457 | if err != nil { | ||
7458 | logError(testName, function, args, startTime, "", "GetObject failed", err) | ||
7459 | return | ||
7460 | } | ||
7461 | |||
7462 | st, err := r.Stat() | ||
7463 | if err != nil { | ||
7464 | logError(testName, function, args, startTime, "", "Stat failed", err) | ||
7465 | return | ||
7466 | } | ||
7467 | |||
7468 | if st.Size != int64(bufSize) { | ||
7469 | logError(testName, function, args, startTime, "", "Number of bytes does not match, expected "+string(bufSize)+" got "+string(st.Size), err) | ||
7470 | return | ||
7471 | } | ||
7472 | if err := r.Close(); err != nil { | ||
7473 | logError(testName, function, args, startTime, "", "Stat failed", err) | ||
7474 | return | ||
7475 | } | ||
7476 | if err := r.Close(); err == nil { | ||
7477 | logError(testName, function, args, startTime, "", "Object is already closed, should return error", err) | ||
7478 | return | ||
7479 | } | ||
7480 | |||
7481 | successLogger(testName, function, args, startTime).Info() | ||
7482 | } | ||
7483 | |||
7484 | // Tests FPutObject hidden contentType setting | ||
7485 | func testFPutObjectV2() { | ||
7486 | // initialize logging params | ||
7487 | startTime := time.Now() | ||
7488 | testName := getFuncName() | ||
7489 | function := "FPutObject(bucketName, objectName, fileName, opts)" | ||
7490 | args := map[string]interface{}{ | ||
7491 | "bucketName": "", | ||
7492 | "objectName": "", | ||
7493 | "fileName": "", | ||
7494 | "opts": "", | ||
7495 | } | ||
7496 | |||
7497 | // Seed random based on current time. | ||
7498 | rand.Seed(time.Now().Unix()) | ||
7499 | |||
7500 | // Instantiate new minio client object. | ||
7501 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
7502 | &minio.Options{ | ||
7503 | Creds: credentials.NewStaticV2(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
7504 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
7505 | }) | ||
7506 | if err != nil { | ||
7507 | logError(testName, function, args, startTime, "", "MinIO v2 client object creation failed", err) | ||
7508 | return | ||
7509 | } | ||
7510 | |||
7511 | // Enable tracing, write to stderr. | ||
7512 | // c.TraceOn(os.Stderr) | ||
7513 | |||
7514 | // Set user agent. | ||
7515 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
7516 | |||
7517 | // Generate a new random bucket name. | ||
7518 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
7519 | args["bucketName"] = bucketName | ||
7520 | |||
7521 | // Make a new bucket. | ||
7522 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
7523 | if err != nil { | ||
7524 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
7525 | return | ||
7526 | } | ||
7527 | |||
7528 | defer cleanupBucket(bucketName, c) | ||
7529 | |||
7530 | // Make a temp file with 11*1024*1024 bytes of data. | ||
7531 | file, err := os.CreateTemp(os.TempDir(), "FPutObjectTest") | ||
7532 | if err != nil { | ||
7533 | logError(testName, function, args, startTime, "", "TempFile creation failed", err) | ||
7534 | return | ||
7535 | } | ||
7536 | |||
7537 | r := bytes.NewReader(bytes.Repeat([]byte("b"), 11*1024*1024)) | ||
7538 | n, err := io.CopyN(file, r, 11*1024*1024) | ||
7539 | if err != nil { | ||
7540 | logError(testName, function, args, startTime, "", "Copy failed", err) | ||
7541 | return | ||
7542 | } | ||
7543 | if n != int64(11*1024*1024) { | ||
7544 | logError(testName, function, args, startTime, "", "Number of bytes does not match, expected "+string(int64(11*1024*1024))+" got "+string(n), err) | ||
7545 | return | ||
7546 | } | ||
7547 | |||
7548 | // Close the file pro-actively for windows. | ||
7549 | err = file.Close() | ||
7550 | if err != nil { | ||
7551 | logError(testName, function, args, startTime, "", "File close failed", err) | ||
7552 | return | ||
7553 | } | ||
7554 | |||
7555 | // Set base object name | ||
7556 | objectName := bucketName + "FPutObject" | ||
7557 | args["objectName"] = objectName | ||
7558 | args["fileName"] = file.Name() | ||
7559 | |||
7560 | // Perform standard FPutObject with contentType provided (Expecting application/octet-stream) | ||
7561 | _, err = c.FPutObject(context.Background(), bucketName, objectName+"-standard", file.Name(), minio.PutObjectOptions{ContentType: "application/octet-stream"}) | ||
7562 | if err != nil { | ||
7563 | logError(testName, function, args, startTime, "", "FPutObject failed", err) | ||
7564 | return | ||
7565 | } | ||
7566 | |||
7567 | // Perform FPutObject with no contentType provided (Expecting application/octet-stream) | ||
7568 | args["objectName"] = objectName + "-Octet" | ||
7569 | args["contentType"] = "" | ||
7570 | |||
7571 | _, err = c.FPutObject(context.Background(), bucketName, objectName+"-Octet", file.Name(), minio.PutObjectOptions{}) | ||
7572 | if err != nil { | ||
7573 | logError(testName, function, args, startTime, "", "FPutObject failed", err) | ||
7574 | return | ||
7575 | } | ||
7576 | |||
7577 | // Add extension to temp file name | ||
7578 | fileName := file.Name() | ||
7579 | err = os.Rename(fileName, fileName+".gtar") | ||
7580 | if err != nil { | ||
7581 | logError(testName, function, args, startTime, "", "Rename failed", err) | ||
7582 | return | ||
7583 | } | ||
7584 | |||
7585 | // Perform FPutObject with no contentType provided (Expecting application/x-gtar) | ||
7586 | args["objectName"] = objectName + "-Octet" | ||
7587 | args["contentType"] = "" | ||
7588 | args["fileName"] = fileName + ".gtar" | ||
7589 | |||
7590 | _, err = c.FPutObject(context.Background(), bucketName, objectName+"-GTar", fileName+".gtar", minio.PutObjectOptions{}) | ||
7591 | if err != nil { | ||
7592 | logError(testName, function, args, startTime, "", "FPutObject failed", err) | ||
7593 | return | ||
7594 | } | ||
7595 | |||
7596 | // Check headers and sizes | ||
7597 | rStandard, err := c.StatObject(context.Background(), bucketName, objectName+"-standard", minio.StatObjectOptions{}) | ||
7598 | if err != nil { | ||
7599 | logError(testName, function, args, startTime, "", "StatObject failed", err) | ||
7600 | return | ||
7601 | } | ||
7602 | |||
7603 | if rStandard.Size != 11*1024*1024 { | ||
7604 | logError(testName, function, args, startTime, "", "Unexpected size", nil) | ||
7605 | return | ||
7606 | } | ||
7607 | |||
7608 | if rStandard.ContentType != "application/octet-stream" { | ||
7609 | logError(testName, function, args, startTime, "", "Content-Type headers mismatched, expected: application/octet-stream , got "+rStandard.ContentType, err) | ||
7610 | return | ||
7611 | } | ||
7612 | |||
7613 | rOctet, err := c.StatObject(context.Background(), bucketName, objectName+"-Octet", minio.StatObjectOptions{}) | ||
7614 | if err != nil { | ||
7615 | logError(testName, function, args, startTime, "", "StatObject failed", err) | ||
7616 | return | ||
7617 | } | ||
7618 | if rOctet.ContentType != "application/octet-stream" { | ||
7619 | logError(testName, function, args, startTime, "", "Content-Type headers mismatched, expected: application/octet-stream , got "+rOctet.ContentType, err) | ||
7620 | return | ||
7621 | } | ||
7622 | |||
7623 | if rOctet.Size != 11*1024*1024 { | ||
7624 | logError(testName, function, args, startTime, "", "Unexpected size", nil) | ||
7625 | return | ||
7626 | } | ||
7627 | |||
7628 | rGTar, err := c.StatObject(context.Background(), bucketName, objectName+"-GTar", minio.StatObjectOptions{}) | ||
7629 | if err != nil { | ||
7630 | logError(testName, function, args, startTime, "", "StatObject failed", err) | ||
7631 | return | ||
7632 | } | ||
7633 | if rGTar.Size != 11*1024*1024 { | ||
7634 | logError(testName, function, args, startTime, "", "Unexpected size", nil) | ||
7635 | return | ||
7636 | } | ||
7637 | if rGTar.ContentType != "application/x-gtar" && rGTar.ContentType != "application/octet-stream" && rGTar.ContentType != "application/x-tar" { | ||
7638 | logError(testName, function, args, startTime, "", "Content-Type headers mismatched, expected: application/x-tar , got "+rGTar.ContentType, err) | ||
7639 | return | ||
7640 | } | ||
7641 | |||
7642 | os.Remove(fileName + ".gtar") | ||
7643 | successLogger(testName, function, args, startTime).Info() | ||
7644 | } | ||
7645 | |||
7646 | // Tests various bucket supported formats. | ||
7647 | func testMakeBucketRegionsV2() { | ||
7648 | // initialize logging params | ||
7649 | startTime := time.Now() | ||
7650 | testName := getFuncName() | ||
7651 | function := "MakeBucket(bucketName, region)" | ||
7652 | args := map[string]interface{}{ | ||
7653 | "bucketName": "", | ||
7654 | "region": "eu-west-1", | ||
7655 | } | ||
7656 | |||
7657 | // Seed random based on current time. | ||
7658 | rand.Seed(time.Now().Unix()) | ||
7659 | |||
7660 | // Instantiate new minio client object. | ||
7661 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
7662 | &minio.Options{ | ||
7663 | Creds: credentials.NewStaticV2(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
7664 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
7665 | }) | ||
7666 | if err != nil { | ||
7667 | logError(testName, function, args, startTime, "", "MinIO v2 client object creation failed", err) | ||
7668 | return | ||
7669 | } | ||
7670 | |||
7671 | // Enable tracing, write to stderr. | ||
7672 | // c.TraceOn(os.Stderr) | ||
7673 | |||
7674 | // Set user agent. | ||
7675 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
7676 | |||
7677 | // Generate a new random bucket name. | ||
7678 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
7679 | args["bucketName"] = bucketName | ||
7680 | |||
7681 | // Make a new bucket in 'eu-central-1'. | ||
7682 | if err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "eu-west-1"}); err != nil { | ||
7683 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
7684 | return | ||
7685 | } | ||
7686 | |||
7687 | if err = cleanupBucket(bucketName, c); err != nil { | ||
7688 | logError(testName, function, args, startTime, "", "CleanupBucket failed while removing bucket recursively", err) | ||
7689 | return | ||
7690 | } | ||
7691 | |||
7692 | // Make a new bucket with '.' in its name, in 'us-west-2'. This | ||
7693 | // request is internally staged into a path style instead of | ||
7694 | // virtual host style. | ||
7695 | if err = c.MakeBucket(context.Background(), bucketName+".withperiod", minio.MakeBucketOptions{Region: "us-west-2"}); err != nil { | ||
7696 | args["bucketName"] = bucketName + ".withperiod" | ||
7697 | args["region"] = "us-west-2" | ||
7698 | logError(testName, function, args, startTime, "", "MakeBucket test with a bucket name with period, '.', failed", err) | ||
7699 | return | ||
7700 | } | ||
7701 | |||
7702 | // Delete all objects and buckets | ||
7703 | if err = cleanupBucket(bucketName+".withperiod", c); err != nil { | ||
7704 | logError(testName, function, args, startTime, "", "CleanupBucket failed while removing bucket recursively", err) | ||
7705 | return | ||
7706 | } | ||
7707 | |||
7708 | successLogger(testName, function, args, startTime).Info() | ||
7709 | } | ||
7710 | |||
7711 | // Tests get object ReaderSeeker interface methods. | ||
7712 | func testGetObjectReadSeekFunctionalV2() { | ||
7713 | // initialize logging params | ||
7714 | startTime := time.Now() | ||
7715 | testName := getFuncName() | ||
7716 | function := "GetObject(bucketName, objectName)" | ||
7717 | args := map[string]interface{}{} | ||
7718 | |||
7719 | // Seed random based on current time. | ||
7720 | rand.Seed(time.Now().Unix()) | ||
7721 | |||
7722 | // Instantiate new minio client object. | ||
7723 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
7724 | &minio.Options{ | ||
7725 | Creds: credentials.NewStaticV2(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
7726 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
7727 | }) | ||
7728 | if err != nil { | ||
7729 | logError(testName, function, args, startTime, "", "MinIO v2 client object creation failed", err) | ||
7730 | return | ||
7731 | } | ||
7732 | |||
7733 | // Enable tracing, write to stderr. | ||
7734 | // c.TraceOn(os.Stderr) | ||
7735 | |||
7736 | // Set user agent. | ||
7737 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
7738 | |||
7739 | // Generate a new random bucket name. | ||
7740 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
7741 | args["bucketName"] = bucketName | ||
7742 | |||
7743 | // Make a new bucket. | ||
7744 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
7745 | if err != nil { | ||
7746 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
7747 | return | ||
7748 | } | ||
7749 | |||
7750 | defer cleanupBucket(bucketName, c) | ||
7751 | |||
7752 | // Generate 33K of data. | ||
7753 | bufSize := dataFileMap["datafile-33-kB"] | ||
7754 | reader := getDataReader("datafile-33-kB") | ||
7755 | defer reader.Close() | ||
7756 | |||
7757 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
7758 | args["objectName"] = objectName | ||
7759 | |||
7760 | buf, err := io.ReadAll(reader) | ||
7761 | if err != nil { | ||
7762 | logError(testName, function, args, startTime, "", "ReadAll failed", err) | ||
7763 | return | ||
7764 | } | ||
7765 | |||
7766 | // Save the data. | ||
7767 | _, err = c.PutObject(context.Background(), bucketName, objectName, bytes.NewReader(buf), int64(bufSize), minio.PutObjectOptions{ContentType: "binary/octet-stream"}) | ||
7768 | if err != nil { | ||
7769 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
7770 | return | ||
7771 | } | ||
7772 | |||
7773 | // Read the data back | ||
7774 | r, err := c.GetObject(context.Background(), bucketName, objectName, minio.GetObjectOptions{}) | ||
7775 | if err != nil { | ||
7776 | logError(testName, function, args, startTime, "", "GetObject failed", err) | ||
7777 | return | ||
7778 | } | ||
7779 | defer r.Close() | ||
7780 | |||
7781 | st, err := r.Stat() | ||
7782 | if err != nil { | ||
7783 | logError(testName, function, args, startTime, "", "Stat failed", err) | ||
7784 | return | ||
7785 | } | ||
7786 | |||
7787 | if st.Size != int64(bufSize) { | ||
7788 | logError(testName, function, args, startTime, "", "Number of bytes in stat does not match, expected "+string(int64(bufSize))+" got "+string(st.Size), err) | ||
7789 | return | ||
7790 | } | ||
7791 | |||
7792 | offset := int64(2048) | ||
7793 | n, err := r.Seek(offset, 0) | ||
7794 | if err != nil { | ||
7795 | logError(testName, function, args, startTime, "", "Seek failed", err) | ||
7796 | return | ||
7797 | } | ||
7798 | if n != offset { | ||
7799 | logError(testName, function, args, startTime, "", "Number of seeked bytes does not match, expected "+string(offset)+" got "+string(n), err) | ||
7800 | return | ||
7801 | } | ||
7802 | n, err = r.Seek(0, 1) | ||
7803 | if err != nil { | ||
7804 | logError(testName, function, args, startTime, "", "Seek failed", err) | ||
7805 | return | ||
7806 | } | ||
7807 | if n != offset { | ||
7808 | logError(testName, function, args, startTime, "", "Number of seeked bytes does not match, expected "+string(offset)+" got "+string(n), err) | ||
7809 | return | ||
7810 | } | ||
7811 | _, err = r.Seek(offset, 2) | ||
7812 | if err == nil { | ||
7813 | logError(testName, function, args, startTime, "", "Seek on positive offset for whence '2' should error out", err) | ||
7814 | return | ||
7815 | } | ||
7816 | n, err = r.Seek(-offset, 2) | ||
7817 | if err != nil { | ||
7818 | logError(testName, function, args, startTime, "", "Seek failed", err) | ||
7819 | return | ||
7820 | } | ||
7821 | if n != st.Size-offset { | ||
7822 | logError(testName, function, args, startTime, "", "Number of seeked bytes does not match, expected "+string(st.Size-offset)+" got "+string(n), err) | ||
7823 | return | ||
7824 | } | ||
7825 | |||
7826 | var buffer1 bytes.Buffer | ||
7827 | if _, err = io.CopyN(&buffer1, r, st.Size); err != nil { | ||
7828 | if err != io.EOF { | ||
7829 | logError(testName, function, args, startTime, "", "Copy failed", err) | ||
7830 | return | ||
7831 | } | ||
7832 | } | ||
7833 | if !bytes.Equal(buf[len(buf)-int(offset):], buffer1.Bytes()) { | ||
7834 | logError(testName, function, args, startTime, "", "Incorrect read bytes v/s original buffer", err) | ||
7835 | return | ||
7836 | } | ||
7837 | |||
7838 | // Seek again and read again. | ||
7839 | n, err = r.Seek(offset-1, 0) | ||
7840 | if err != nil { | ||
7841 | logError(testName, function, args, startTime, "", "Seek failed", err) | ||
7842 | return | ||
7843 | } | ||
7844 | if n != (offset - 1) { | ||
7845 | logError(testName, function, args, startTime, "", "Number of seeked bytes does not match, expected "+string(offset-1)+" got "+string(n), err) | ||
7846 | return | ||
7847 | } | ||
7848 | |||
7849 | var buffer2 bytes.Buffer | ||
7850 | if _, err = io.CopyN(&buffer2, r, st.Size); err != nil { | ||
7851 | if err != io.EOF { | ||
7852 | logError(testName, function, args, startTime, "", "Copy failed", err) | ||
7853 | return | ||
7854 | } | ||
7855 | } | ||
7856 | // Verify now lesser bytes. | ||
7857 | if !bytes.Equal(buf[2047:], buffer2.Bytes()) { | ||
7858 | logError(testName, function, args, startTime, "", "Incorrect read bytes v/s original buffer", err) | ||
7859 | return | ||
7860 | } | ||
7861 | |||
7862 | successLogger(testName, function, args, startTime).Info() | ||
7863 | } | ||
7864 | |||
7865 | // Tests get object ReaderAt interface methods. | ||
7866 | func testGetObjectReadAtFunctionalV2() { | ||
7867 | // initialize logging params | ||
7868 | startTime := time.Now() | ||
7869 | testName := getFuncName() | ||
7870 | function := "GetObject(bucketName, objectName)" | ||
7871 | args := map[string]interface{}{} | ||
7872 | |||
7873 | // Seed random based on current time. | ||
7874 | rand.Seed(time.Now().Unix()) | ||
7875 | |||
7876 | // Instantiate new minio client object. | ||
7877 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
7878 | &minio.Options{ | ||
7879 | Creds: credentials.NewStaticV2(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
7880 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
7881 | }) | ||
7882 | if err != nil { | ||
7883 | logError(testName, function, args, startTime, "", "MinIO v2 client object creation failed", err) | ||
7884 | return | ||
7885 | } | ||
7886 | |||
7887 | // Enable tracing, write to stderr. | ||
7888 | // c.TraceOn(os.Stderr) | ||
7889 | |||
7890 | // Set user agent. | ||
7891 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
7892 | |||
7893 | // Generate a new random bucket name. | ||
7894 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
7895 | args["bucketName"] = bucketName | ||
7896 | |||
7897 | // Make a new bucket. | ||
7898 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
7899 | if err != nil { | ||
7900 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
7901 | return | ||
7902 | } | ||
7903 | |||
7904 | defer cleanupBucket(bucketName, c) | ||
7905 | |||
7906 | // Generate 33K of data. | ||
7907 | bufSize := dataFileMap["datafile-33-kB"] | ||
7908 | reader := getDataReader("datafile-33-kB") | ||
7909 | defer reader.Close() | ||
7910 | |||
7911 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
7912 | args["objectName"] = objectName | ||
7913 | |||
7914 | buf, err := io.ReadAll(reader) | ||
7915 | if err != nil { | ||
7916 | logError(testName, function, args, startTime, "", "ReadAll failed", err) | ||
7917 | return | ||
7918 | } | ||
7919 | |||
7920 | // Save the data | ||
7921 | _, err = c.PutObject(context.Background(), bucketName, objectName, bytes.NewReader(buf), int64(bufSize), minio.PutObjectOptions{ContentType: "binary/octet-stream"}) | ||
7922 | if err != nil { | ||
7923 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
7924 | return | ||
7925 | } | ||
7926 | |||
7927 | // Read the data back | ||
7928 | r, err := c.GetObject(context.Background(), bucketName, objectName, minio.GetObjectOptions{}) | ||
7929 | if err != nil { | ||
7930 | logError(testName, function, args, startTime, "", "GetObject failed", err) | ||
7931 | return | ||
7932 | } | ||
7933 | defer r.Close() | ||
7934 | |||
7935 | st, err := r.Stat() | ||
7936 | if err != nil { | ||
7937 | logError(testName, function, args, startTime, "", "Stat failed", err) | ||
7938 | return | ||
7939 | } | ||
7940 | |||
7941 | if st.Size != int64(bufSize) { | ||
7942 | logError(testName, function, args, startTime, "", "Number of bytes does not match, expected "+string(bufSize)+" got "+string(st.Size), err) | ||
7943 | return | ||
7944 | } | ||
7945 | |||
7946 | offset := int64(2048) | ||
7947 | |||
7948 | // Read directly | ||
7949 | buf2 := make([]byte, 512) | ||
7950 | buf3 := make([]byte, 512) | ||
7951 | buf4 := make([]byte, 512) | ||
7952 | |||
7953 | m, err := r.ReadAt(buf2, offset) | ||
7954 | if err != nil { | ||
7955 | logError(testName, function, args, startTime, "", "ReadAt failed", err) | ||
7956 | return | ||
7957 | } | ||
7958 | if m != len(buf2) { | ||
7959 | logError(testName, function, args, startTime, "", "ReadAt read shorter bytes before reaching EOF, expected "+string(len(buf2))+" got "+string(m), err) | ||
7960 | return | ||
7961 | } | ||
7962 | if !bytes.Equal(buf2, buf[offset:offset+512]) { | ||
7963 | logError(testName, function, args, startTime, "", "Incorrect read between two ReadAt from same offset", err) | ||
7964 | return | ||
7965 | } | ||
7966 | offset += 512 | ||
7967 | m, err = r.ReadAt(buf3, offset) | ||
7968 | if err != nil { | ||
7969 | logError(testName, function, args, startTime, "", "ReadAt failed", err) | ||
7970 | return | ||
7971 | } | ||
7972 | if m != len(buf3) { | ||
7973 | logError(testName, function, args, startTime, "", "ReadAt read shorter bytes before reaching EOF, expected "+string(len(buf3))+" got "+string(m), err) | ||
7974 | return | ||
7975 | } | ||
7976 | if !bytes.Equal(buf3, buf[offset:offset+512]) { | ||
7977 | logError(testName, function, args, startTime, "", "Incorrect read between two ReadAt from same offset", err) | ||
7978 | return | ||
7979 | } | ||
7980 | offset += 512 | ||
7981 | m, err = r.ReadAt(buf4, offset) | ||
7982 | if err != nil { | ||
7983 | logError(testName, function, args, startTime, "", "ReadAt failed", err) | ||
7984 | return | ||
7985 | } | ||
7986 | if m != len(buf4) { | ||
7987 | logError(testName, function, args, startTime, "", "ReadAt read shorter bytes before reaching EOF, expected "+string(len(buf4))+" got "+string(m), err) | ||
7988 | return | ||
7989 | } | ||
7990 | if !bytes.Equal(buf4, buf[offset:offset+512]) { | ||
7991 | logError(testName, function, args, startTime, "", "Incorrect read between two ReadAt from same offset", err) | ||
7992 | return | ||
7993 | } | ||
7994 | |||
7995 | buf5 := make([]byte, bufSize) | ||
7996 | // Read the whole object. | ||
7997 | m, err = r.ReadAt(buf5, 0) | ||
7998 | if err != nil { | ||
7999 | if err != io.EOF { | ||
8000 | logError(testName, function, args, startTime, "", "ReadAt failed", err) | ||
8001 | return | ||
8002 | } | ||
8003 | } | ||
8004 | if m != len(buf5) { | ||
8005 | logError(testName, function, args, startTime, "", "ReadAt read shorter bytes before reaching EOF, expected "+string(len(buf5))+" got "+string(m), err) | ||
8006 | return | ||
8007 | } | ||
8008 | if !bytes.Equal(buf, buf5) { | ||
8009 | logError(testName, function, args, startTime, "", "Incorrect data read in GetObject, than what was previously uploaded", err) | ||
8010 | return | ||
8011 | } | ||
8012 | |||
8013 | buf6 := make([]byte, bufSize+1) | ||
8014 | // Read the whole object and beyond. | ||
8015 | _, err = r.ReadAt(buf6, 0) | ||
8016 | if err != nil { | ||
8017 | if err != io.EOF { | ||
8018 | logError(testName, function, args, startTime, "", "ReadAt failed", err) | ||
8019 | return | ||
8020 | } | ||
8021 | } | ||
8022 | |||
8023 | successLogger(testName, function, args, startTime).Info() | ||
8024 | } | ||
8025 | |||
8026 | // Tests copy object | ||
8027 | func testCopyObjectV2() { | ||
8028 | // initialize logging params | ||
8029 | startTime := time.Now() | ||
8030 | testName := getFuncName() | ||
8031 | function := "CopyObject(destination, source)" | ||
8032 | args := map[string]interface{}{} | ||
8033 | |||
8034 | // Seed random based on current time. | ||
8035 | rand.Seed(time.Now().Unix()) | ||
8036 | |||
8037 | // Instantiate new minio client object | ||
8038 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
8039 | &minio.Options{ | ||
8040 | Creds: credentials.NewStaticV2(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
8041 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
8042 | }) | ||
8043 | if err != nil { | ||
8044 | logError(testName, function, args, startTime, "", "MinIO v2 client object creation failed", err) | ||
8045 | return | ||
8046 | } | ||
8047 | |||
8048 | // Enable tracing, write to stderr. | ||
8049 | // c.TraceOn(os.Stderr) | ||
8050 | |||
8051 | // Set user agent. | ||
8052 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
8053 | |||
8054 | // Generate a new random bucket name. | ||
8055 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
8056 | |||
8057 | // Make a new bucket in 'us-east-1' (source bucket). | ||
8058 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
8059 | if err != nil { | ||
8060 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
8061 | return | ||
8062 | } | ||
8063 | defer cleanupBucket(bucketName, c) | ||
8064 | |||
8065 | // Make a new bucket in 'us-east-1' (destination bucket). | ||
8066 | err = c.MakeBucket(context.Background(), bucketName+"-copy", minio.MakeBucketOptions{Region: "us-east-1"}) | ||
8067 | if err != nil { | ||
8068 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
8069 | return | ||
8070 | } | ||
8071 | defer cleanupBucket(bucketName+"-copy", c) | ||
8072 | |||
8073 | // Generate 33K of data. | ||
8074 | bufSize := dataFileMap["datafile-33-kB"] | ||
8075 | reader := getDataReader("datafile-33-kB") | ||
8076 | defer reader.Close() | ||
8077 | |||
8078 | // Save the data | ||
8079 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
8080 | _, err = c.PutObject(context.Background(), bucketName, objectName, reader, int64(bufSize), minio.PutObjectOptions{ContentType: "binary/octet-stream"}) | ||
8081 | if err != nil { | ||
8082 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
8083 | return | ||
8084 | } | ||
8085 | |||
8086 | r, err := c.GetObject(context.Background(), bucketName, objectName, minio.GetObjectOptions{}) | ||
8087 | if err != nil { | ||
8088 | logError(testName, function, args, startTime, "", "GetObject failed", err) | ||
8089 | return | ||
8090 | } | ||
8091 | // Check the various fields of source object against destination object. | ||
8092 | objInfo, err := r.Stat() | ||
8093 | if err != nil { | ||
8094 | logError(testName, function, args, startTime, "", "Stat failed", err) | ||
8095 | return | ||
8096 | } | ||
8097 | r.Close() | ||
8098 | |||
8099 | // Copy Source | ||
8100 | src := minio.CopySrcOptions{ | ||
8101 | Bucket: bucketName, | ||
8102 | Object: objectName, | ||
8103 | MatchModifiedSince: time.Date(2014, time.April, 0, 0, 0, 0, 0, time.UTC), | ||
8104 | MatchETag: objInfo.ETag, | ||
8105 | } | ||
8106 | args["source"] = src | ||
8107 | |||
8108 | // Set copy conditions. | ||
8109 | dst := minio.CopyDestOptions{ | ||
8110 | Bucket: bucketName + "-copy", | ||
8111 | Object: objectName + "-copy", | ||
8112 | } | ||
8113 | args["destination"] = dst | ||
8114 | |||
8115 | // Perform the Copy | ||
8116 | _, err = c.CopyObject(context.Background(), dst, src) | ||
8117 | if err != nil { | ||
8118 | logError(testName, function, args, startTime, "", "CopyObject failed", err) | ||
8119 | return | ||
8120 | } | ||
8121 | |||
8122 | // Source object | ||
8123 | r, err = c.GetObject(context.Background(), bucketName, objectName, minio.GetObjectOptions{}) | ||
8124 | if err != nil { | ||
8125 | logError(testName, function, args, startTime, "", "GetObject failed", err) | ||
8126 | return | ||
8127 | } | ||
8128 | // Destination object | ||
8129 | readerCopy, err := c.GetObject(context.Background(), bucketName+"-copy", objectName+"-copy", minio.GetObjectOptions{}) | ||
8130 | if err != nil { | ||
8131 | logError(testName, function, args, startTime, "", "GetObject failed", err) | ||
8132 | return | ||
8133 | } | ||
8134 | // Check the various fields of source object against destination object. | ||
8135 | objInfo, err = r.Stat() | ||
8136 | if err != nil { | ||
8137 | logError(testName, function, args, startTime, "", "Stat failed", err) | ||
8138 | return | ||
8139 | } | ||
8140 | objInfoCopy, err := readerCopy.Stat() | ||
8141 | if err != nil { | ||
8142 | logError(testName, function, args, startTime, "", "Stat failed", err) | ||
8143 | return | ||
8144 | } | ||
8145 | if objInfo.Size != objInfoCopy.Size { | ||
8146 | logError(testName, function, args, startTime, "", "Number of bytes does not match, expected "+string(objInfoCopy.Size)+" got "+string(objInfo.Size), err) | ||
8147 | return | ||
8148 | } | ||
8149 | |||
8150 | // Close all the readers. | ||
8151 | r.Close() | ||
8152 | readerCopy.Close() | ||
8153 | |||
8154 | // CopyObject again but with wrong conditions | ||
8155 | src = minio.CopySrcOptions{ | ||
8156 | Bucket: bucketName, | ||
8157 | Object: objectName, | ||
8158 | MatchUnmodifiedSince: time.Date(2014, time.April, 0, 0, 0, 0, 0, time.UTC), | ||
8159 | NoMatchETag: objInfo.ETag, | ||
8160 | } | ||
8161 | |||
8162 | // Perform the Copy which should fail | ||
8163 | _, err = c.CopyObject(context.Background(), dst, src) | ||
8164 | if err == nil { | ||
8165 | logError(testName, function, args, startTime, "", "CopyObject did not fail for invalid conditions", err) | ||
8166 | return | ||
8167 | } | ||
8168 | |||
8169 | successLogger(testName, function, args, startTime).Info() | ||
8170 | } | ||
8171 | |||
8172 | func testComposeObjectErrorCasesWrapper(c *minio.Client) { | ||
8173 | // initialize logging params | ||
8174 | startTime := time.Now() | ||
8175 | testName := getFuncName() | ||
8176 | function := "ComposeObject(destination, sourceList)" | ||
8177 | args := map[string]interface{}{} | ||
8178 | |||
8179 | // Generate a new random bucket name. | ||
8180 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
8181 | |||
8182 | // Make a new bucket in 'us-east-1' (source bucket). | ||
8183 | err := c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
8184 | if err != nil { | ||
8185 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
8186 | return | ||
8187 | } | ||
8188 | |||
8189 | defer cleanupBucket(bucketName, c) | ||
8190 | |||
8191 | // Test that more than 10K source objects cannot be | ||
8192 | // concatenated. | ||
8193 | srcArr := [10001]minio.CopySrcOptions{} | ||
8194 | srcSlice := srcArr[:] | ||
8195 | dst := minio.CopyDestOptions{ | ||
8196 | Bucket: bucketName, | ||
8197 | Object: "object", | ||
8198 | } | ||
8199 | |||
8200 | args["destination"] = dst | ||
8201 | // Just explain about srcArr in args["sourceList"] | ||
8202 | // to stop having 10,001 null headers logged | ||
8203 | args["sourceList"] = "source array of 10,001 elements" | ||
8204 | if _, err := c.ComposeObject(context.Background(), dst, srcSlice...); err == nil { | ||
8205 | logError(testName, function, args, startTime, "", "Expected error in ComposeObject", err) | ||
8206 | return | ||
8207 | } else if err.Error() != "There must be as least one and up to 10000 source objects." { | ||
8208 | logError(testName, function, args, startTime, "", "Got unexpected error", err) | ||
8209 | return | ||
8210 | } | ||
8211 | |||
8212 | // Create a source with invalid offset spec and check that | ||
8213 | // error is returned: | ||
8214 | // 1. Create the source object. | ||
8215 | const badSrcSize = 5 * 1024 * 1024 | ||
8216 | buf := bytes.Repeat([]byte("1"), badSrcSize) | ||
8217 | _, err = c.PutObject(context.Background(), bucketName, "badObject", bytes.NewReader(buf), int64(len(buf)), minio.PutObjectOptions{}) | ||
8218 | if err != nil { | ||
8219 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
8220 | return | ||
8221 | } | ||
8222 | // 2. Set invalid range spec on the object (going beyond | ||
8223 | // object size) | ||
8224 | badSrc := minio.CopySrcOptions{ | ||
8225 | Bucket: bucketName, | ||
8226 | Object: "badObject", | ||
8227 | MatchRange: true, | ||
8228 | Start: 1, | ||
8229 | End: badSrcSize, | ||
8230 | } | ||
8231 | |||
8232 | // 3. ComposeObject call should fail. | ||
8233 | if _, err := c.ComposeObject(context.Background(), dst, badSrc); err == nil { | ||
8234 | logError(testName, function, args, startTime, "", "ComposeObject expected to fail", err) | ||
8235 | return | ||
8236 | } else if !strings.Contains(err.Error(), "has invalid segment-to-copy") { | ||
8237 | logError(testName, function, args, startTime, "", "Got invalid error", err) | ||
8238 | return | ||
8239 | } | ||
8240 | |||
8241 | successLogger(testName, function, args, startTime).Info() | ||
8242 | } | ||
8243 | |||
8244 | // Test expected error cases | ||
8245 | func testComposeObjectErrorCasesV2() { | ||
8246 | // initialize logging params | ||
8247 | startTime := time.Now() | ||
8248 | testName := getFuncName() | ||
8249 | function := "ComposeObject(destination, sourceList)" | ||
8250 | args := map[string]interface{}{} | ||
8251 | |||
8252 | // Instantiate new minio client object | ||
8253 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
8254 | &minio.Options{ | ||
8255 | Creds: credentials.NewStaticV2(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
8256 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
8257 | }) | ||
8258 | if err != nil { | ||
8259 | logError(testName, function, args, startTime, "", "MinIO v2 client object creation failed", err) | ||
8260 | return | ||
8261 | } | ||
8262 | |||
8263 | testComposeObjectErrorCasesWrapper(c) | ||
8264 | } | ||
8265 | |||
8266 | func testComposeMultipleSources(c *minio.Client) { | ||
8267 | // initialize logging params | ||
8268 | startTime := time.Now() | ||
8269 | testName := getFuncName() | ||
8270 | function := "ComposeObject(destination, sourceList)" | ||
8271 | args := map[string]interface{}{ | ||
8272 | "destination": "", | ||
8273 | "sourceList": "", | ||
8274 | } | ||
8275 | |||
8276 | // Generate a new random bucket name. | ||
8277 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
8278 | // Make a new bucket in 'us-east-1' (source bucket). | ||
8279 | err := c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
8280 | if err != nil { | ||
8281 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
8282 | return | ||
8283 | } | ||
8284 | |||
8285 | defer cleanupBucket(bucketName, c) | ||
8286 | |||
8287 | // Upload a small source object | ||
8288 | const srcSize = 1024 * 1024 * 5 | ||
8289 | buf := bytes.Repeat([]byte("1"), srcSize) | ||
8290 | _, err = c.PutObject(context.Background(), bucketName, "srcObject", bytes.NewReader(buf), int64(srcSize), minio.PutObjectOptions{ContentType: "binary/octet-stream"}) | ||
8291 | if err != nil { | ||
8292 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
8293 | return | ||
8294 | } | ||
8295 | |||
8296 | // We will append 10 copies of the object. | ||
8297 | srcs := []minio.CopySrcOptions{} | ||
8298 | for i := 0; i < 10; i++ { | ||
8299 | srcs = append(srcs, minio.CopySrcOptions{ | ||
8300 | Bucket: bucketName, | ||
8301 | Object: "srcObject", | ||
8302 | }) | ||
8303 | } | ||
8304 | |||
8305 | // make the last part very small | ||
8306 | srcs[9].MatchRange = true | ||
8307 | |||
8308 | args["sourceList"] = srcs | ||
8309 | |||
8310 | dst := minio.CopyDestOptions{ | ||
8311 | Bucket: bucketName, | ||
8312 | Object: "dstObject", | ||
8313 | } | ||
8314 | args["destination"] = dst | ||
8315 | |||
8316 | ui, err := c.ComposeObject(context.Background(), dst, srcs...) | ||
8317 | if err != nil { | ||
8318 | logError(testName, function, args, startTime, "", "ComposeObject failed", err) | ||
8319 | return | ||
8320 | } | ||
8321 | |||
8322 | if ui.Size != 9*srcSize+1 { | ||
8323 | logError(testName, function, args, startTime, "", "ComposeObject returned unexpected size", err) | ||
8324 | return | ||
8325 | } | ||
8326 | |||
8327 | objProps, err := c.StatObject(context.Background(), bucketName, "dstObject", minio.StatObjectOptions{}) | ||
8328 | if err != nil { | ||
8329 | logError(testName, function, args, startTime, "", "StatObject failed", err) | ||
8330 | return | ||
8331 | } | ||
8332 | |||
8333 | if objProps.Size != 9*srcSize+1 { | ||
8334 | logError(testName, function, args, startTime, "", "Size mismatched! Expected "+string(10000*srcSize)+" got "+string(objProps.Size), err) | ||
8335 | return | ||
8336 | } | ||
8337 | |||
8338 | successLogger(testName, function, args, startTime).Info() | ||
8339 | } | ||
8340 | |||
8341 | // Test concatenating multiple 10K objects V2 | ||
8342 | func testCompose10KSourcesV2() { | ||
8343 | // initialize logging params | ||
8344 | startTime := time.Now() | ||
8345 | testName := getFuncName() | ||
8346 | function := "ComposeObject(destination, sourceList)" | ||
8347 | args := map[string]interface{}{} | ||
8348 | |||
8349 | // Instantiate new minio client object | ||
8350 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
8351 | &minio.Options{ | ||
8352 | Creds: credentials.NewStaticV2(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
8353 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
8354 | }) | ||
8355 | if err != nil { | ||
8356 | logError(testName, function, args, startTime, "", "MinIO v2 client object creation failed", err) | ||
8357 | return | ||
8358 | } | ||
8359 | |||
8360 | testComposeMultipleSources(c) | ||
8361 | } | ||
8362 | |||
8363 | func testEncryptedEmptyObject() { | ||
8364 | // initialize logging params | ||
8365 | startTime := time.Now() | ||
8366 | testName := getFuncName() | ||
8367 | function := "PutObject(bucketName, objectName, reader, objectSize, opts)" | ||
8368 | args := map[string]interface{}{} | ||
8369 | |||
8370 | // Instantiate new minio client object | ||
8371 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
8372 | &minio.Options{ | ||
8373 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
8374 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
8375 | }) | ||
8376 | if err != nil { | ||
8377 | logError(testName, function, args, startTime, "", "MinIO v4 client object creation failed", err) | ||
8378 | return | ||
8379 | } | ||
8380 | |||
8381 | // Generate a new random bucket name. | ||
8382 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
8383 | args["bucketName"] = bucketName | ||
8384 | // Make a new bucket in 'us-east-1' (source bucket). | ||
8385 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
8386 | if err != nil { | ||
8387 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
8388 | return | ||
8389 | } | ||
8390 | |||
8391 | defer cleanupBucket(bucketName, c) | ||
8392 | |||
8393 | sse := encrypt.DefaultPBKDF([]byte("correct horse battery staple"), []byte(bucketName+"object")) | ||
8394 | |||
8395 | // 1. create an sse-c encrypted object to copy by uploading | ||
8396 | const srcSize = 0 | ||
8397 | var buf []byte // Empty buffer | ||
8398 | args["objectName"] = "object" | ||
8399 | _, err = c.PutObject(context.Background(), bucketName, "object", bytes.NewReader(buf), int64(len(buf)), minio.PutObjectOptions{ServerSideEncryption: sse}) | ||
8400 | if err != nil { | ||
8401 | logError(testName, function, args, startTime, "", "PutObject call failed", err) | ||
8402 | return | ||
8403 | } | ||
8404 | |||
8405 | // 2. Test CopyObject for an empty object | ||
8406 | src := minio.CopySrcOptions{ | ||
8407 | Bucket: bucketName, | ||
8408 | Object: "object", | ||
8409 | Encryption: sse, | ||
8410 | } | ||
8411 | |||
8412 | dst := minio.CopyDestOptions{ | ||
8413 | Bucket: bucketName, | ||
8414 | Object: "new-object", | ||
8415 | Encryption: sse, | ||
8416 | } | ||
8417 | |||
8418 | if _, err = c.CopyObject(context.Background(), dst, src); err != nil { | ||
8419 | function = "CopyObject(dst, src)" | ||
8420 | logError(testName, function, map[string]interface{}{}, startTime, "", "CopyObject failed", err) | ||
8421 | return | ||
8422 | } | ||
8423 | |||
8424 | // 3. Test Key rotation | ||
8425 | newSSE := encrypt.DefaultPBKDF([]byte("Don't Panic"), []byte(bucketName+"new-object")) | ||
8426 | src = minio.CopySrcOptions{ | ||
8427 | Bucket: bucketName, | ||
8428 | Object: "new-object", | ||
8429 | Encryption: sse, | ||
8430 | } | ||
8431 | |||
8432 | dst = minio.CopyDestOptions{ | ||
8433 | Bucket: bucketName, | ||
8434 | Object: "new-object", | ||
8435 | Encryption: newSSE, | ||
8436 | } | ||
8437 | |||
8438 | if _, err = c.CopyObject(context.Background(), dst, src); err != nil { | ||
8439 | function = "CopyObject(dst, src)" | ||
8440 | logError(testName, function, map[string]interface{}{}, startTime, "", "CopyObject with key rotation failed", err) | ||
8441 | return | ||
8442 | } | ||
8443 | |||
8444 | // 4. Download the object. | ||
8445 | reader, err := c.GetObject(context.Background(), bucketName, "new-object", minio.GetObjectOptions{ServerSideEncryption: newSSE}) | ||
8446 | if err != nil { | ||
8447 | logError(testName, function, args, startTime, "", "GetObject failed", err) | ||
8448 | return | ||
8449 | } | ||
8450 | defer reader.Close() | ||
8451 | |||
8452 | decBytes, err := io.ReadAll(reader) | ||
8453 | if err != nil { | ||
8454 | logError(testName, function, map[string]interface{}{}, startTime, "", "ReadAll failed", err) | ||
8455 | return | ||
8456 | } | ||
8457 | if !bytes.Equal(decBytes, buf) { | ||
8458 | logError(testName, function, map[string]interface{}{}, startTime, "", "Downloaded object doesn't match the empty encrypted object", err) | ||
8459 | return | ||
8460 | } | ||
8461 | |||
8462 | delete(args, "objectName") | ||
8463 | successLogger(testName, function, args, startTime).Info() | ||
8464 | } | ||
8465 | |||
8466 | func testEncryptedCopyObjectWrapper(c *minio.Client, bucketName string, sseSrc, sseDst encrypt.ServerSide) { | ||
8467 | // initialize logging params | ||
8468 | startTime := time.Now() | ||
8469 | testName := getFuncNameLoc(2) | ||
8470 | function := "CopyObject(destination, source)" | ||
8471 | args := map[string]interface{}{} | ||
8472 | var srcEncryption, dstEncryption encrypt.ServerSide | ||
8473 | |||
8474 | // Make a new bucket in 'us-east-1' (source bucket). | ||
8475 | err := c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
8476 | if err != nil { | ||
8477 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
8478 | return | ||
8479 | } | ||
8480 | |||
8481 | defer cleanupBucket(bucketName, c) | ||
8482 | |||
8483 | // 1. create an sse-c encrypted object to copy by uploading | ||
8484 | const srcSize = 1024 * 1024 | ||
8485 | buf := bytes.Repeat([]byte("abcde"), srcSize) // gives a buffer of 5MiB | ||
8486 | _, err = c.PutObject(context.Background(), bucketName, "srcObject", bytes.NewReader(buf), int64(len(buf)), minio.PutObjectOptions{ | ||
8487 | ServerSideEncryption: sseSrc, | ||
8488 | }) | ||
8489 | if err != nil { | ||
8490 | logError(testName, function, args, startTime, "", "PutObject call failed", err) | ||
8491 | return | ||
8492 | } | ||
8493 | |||
8494 | if sseSrc != nil && sseSrc.Type() != encrypt.S3 { | ||
8495 | srcEncryption = sseSrc | ||
8496 | } | ||
8497 | |||
8498 | // 2. copy object and change encryption key | ||
8499 | src := minio.CopySrcOptions{ | ||
8500 | Bucket: bucketName, | ||
8501 | Object: "srcObject", | ||
8502 | Encryption: srcEncryption, | ||
8503 | } | ||
8504 | args["source"] = src | ||
8505 | |||
8506 | dst := minio.CopyDestOptions{ | ||
8507 | Bucket: bucketName, | ||
8508 | Object: "dstObject", | ||
8509 | Encryption: sseDst, | ||
8510 | } | ||
8511 | args["destination"] = dst | ||
8512 | |||
8513 | _, err = c.CopyObject(context.Background(), dst, src) | ||
8514 | if err != nil { | ||
8515 | logError(testName, function, args, startTime, "", "CopyObject failed", err) | ||
8516 | return | ||
8517 | } | ||
8518 | |||
8519 | if sseDst != nil && sseDst.Type() != encrypt.S3 { | ||
8520 | dstEncryption = sseDst | ||
8521 | } | ||
8522 | // 3. get copied object and check if content is equal | ||
8523 | coreClient := minio.Core{c} | ||
8524 | reader, _, _, err := coreClient.GetObject(context.Background(), bucketName, "dstObject", minio.GetObjectOptions{ServerSideEncryption: dstEncryption}) | ||
8525 | if err != nil { | ||
8526 | logError(testName, function, args, startTime, "", "GetObject failed", err) | ||
8527 | return | ||
8528 | } | ||
8529 | |||
8530 | decBytes, err := io.ReadAll(reader) | ||
8531 | if err != nil { | ||
8532 | logError(testName, function, args, startTime, "", "ReadAll failed", err) | ||
8533 | return | ||
8534 | } | ||
8535 | if !bytes.Equal(decBytes, buf) { | ||
8536 | logError(testName, function, args, startTime, "", "Downloaded object mismatched for encrypted object", err) | ||
8537 | return | ||
8538 | } | ||
8539 | reader.Close() | ||
8540 | |||
8541 | // Test key rotation for source object in-place. | ||
8542 | var newSSE encrypt.ServerSide | ||
8543 | if sseSrc != nil && sseSrc.Type() == encrypt.SSEC { | ||
8544 | newSSE = encrypt.DefaultPBKDF([]byte("Don't Panic"), []byte(bucketName+"srcObject")) // replace key | ||
8545 | } | ||
8546 | if sseSrc != nil && sseSrc.Type() == encrypt.S3 { | ||
8547 | newSSE = encrypt.NewSSE() | ||
8548 | } | ||
8549 | if newSSE != nil { | ||
8550 | dst = minio.CopyDestOptions{ | ||
8551 | Bucket: bucketName, | ||
8552 | Object: "srcObject", | ||
8553 | Encryption: newSSE, | ||
8554 | } | ||
8555 | args["destination"] = dst | ||
8556 | |||
8557 | _, err = c.CopyObject(context.Background(), dst, src) | ||
8558 | if err != nil { | ||
8559 | logError(testName, function, args, startTime, "", "CopyObject failed", err) | ||
8560 | return | ||
8561 | } | ||
8562 | |||
8563 | // Get copied object and check if content is equal | ||
8564 | reader, _, _, err = coreClient.GetObject(context.Background(), bucketName, "srcObject", minio.GetObjectOptions{ServerSideEncryption: newSSE}) | ||
8565 | if err != nil { | ||
8566 | logError(testName, function, args, startTime, "", "GetObject failed", err) | ||
8567 | return | ||
8568 | } | ||
8569 | |||
8570 | decBytes, err = io.ReadAll(reader) | ||
8571 | if err != nil { | ||
8572 | logError(testName, function, args, startTime, "", "ReadAll failed", err) | ||
8573 | return | ||
8574 | } | ||
8575 | if !bytes.Equal(decBytes, buf) { | ||
8576 | logError(testName, function, args, startTime, "", "Downloaded object mismatched for encrypted object", err) | ||
8577 | return | ||
8578 | } | ||
8579 | reader.Close() | ||
8580 | |||
8581 | // Test in-place decryption. | ||
8582 | dst = minio.CopyDestOptions{ | ||
8583 | Bucket: bucketName, | ||
8584 | Object: "srcObject", | ||
8585 | } | ||
8586 | args["destination"] = dst | ||
8587 | |||
8588 | src = minio.CopySrcOptions{ | ||
8589 | Bucket: bucketName, | ||
8590 | Object: "srcObject", | ||
8591 | Encryption: newSSE, | ||
8592 | } | ||
8593 | args["source"] = src | ||
8594 | _, err = c.CopyObject(context.Background(), dst, src) | ||
8595 | if err != nil { | ||
8596 | logError(testName, function, args, startTime, "", "CopyObject Key rotation failed", err) | ||
8597 | return | ||
8598 | } | ||
8599 | } | ||
8600 | |||
8601 | // Get copied decrypted object and check if content is equal | ||
8602 | reader, _, _, err = coreClient.GetObject(context.Background(), bucketName, "srcObject", minio.GetObjectOptions{}) | ||
8603 | if err != nil { | ||
8604 | logError(testName, function, args, startTime, "", "GetObject failed", err) | ||
8605 | return | ||
8606 | } | ||
8607 | defer reader.Close() | ||
8608 | |||
8609 | decBytes, err = io.ReadAll(reader) | ||
8610 | if err != nil { | ||
8611 | logError(testName, function, args, startTime, "", "ReadAll failed", err) | ||
8612 | return | ||
8613 | } | ||
8614 | if !bytes.Equal(decBytes, buf) { | ||
8615 | logError(testName, function, args, startTime, "", "Downloaded object mismatched for encrypted object", err) | ||
8616 | return | ||
8617 | } | ||
8618 | |||
8619 | successLogger(testName, function, args, startTime).Info() | ||
8620 | } | ||
8621 | |||
8622 | // Test encrypted copy object | ||
8623 | func testUnencryptedToSSECCopyObject() { | ||
8624 | // initialize logging params | ||
8625 | startTime := time.Now() | ||
8626 | testName := getFuncName() | ||
8627 | function := "CopyObject(destination, source)" | ||
8628 | args := map[string]interface{}{} | ||
8629 | |||
8630 | // Instantiate new minio client object | ||
8631 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
8632 | &minio.Options{ | ||
8633 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
8634 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
8635 | }) | ||
8636 | if err != nil { | ||
8637 | logError(testName, function, args, startTime, "", "MinIO v2 client object creation failed", err) | ||
8638 | return | ||
8639 | } | ||
8640 | // Generate a new random bucket name. | ||
8641 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
8642 | |||
8643 | sseDst := encrypt.DefaultPBKDF([]byte("correct horse battery staple"), []byte(bucketName+"dstObject")) | ||
8644 | // c.TraceOn(os.Stderr) | ||
8645 | testEncryptedCopyObjectWrapper(c, bucketName, nil, sseDst) | ||
8646 | } | ||
8647 | |||
8648 | // Test encrypted copy object | ||
8649 | func testUnencryptedToSSES3CopyObject() { | ||
8650 | // initialize logging params | ||
8651 | startTime := time.Now() | ||
8652 | testName := getFuncName() | ||
8653 | function := "CopyObject(destination, source)" | ||
8654 | args := map[string]interface{}{} | ||
8655 | |||
8656 | // Instantiate new minio client object | ||
8657 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
8658 | &minio.Options{ | ||
8659 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
8660 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
8661 | }) | ||
8662 | if err != nil { | ||
8663 | logError(testName, function, args, startTime, "", "MinIO v2 client object creation failed", err) | ||
8664 | return | ||
8665 | } | ||
8666 | // Generate a new random bucket name. | ||
8667 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
8668 | |||
8669 | var sseSrc encrypt.ServerSide | ||
8670 | sseDst := encrypt.NewSSE() | ||
8671 | // c.TraceOn(os.Stderr) | ||
8672 | testEncryptedCopyObjectWrapper(c, bucketName, sseSrc, sseDst) | ||
8673 | } | ||
8674 | |||
8675 | // Test encrypted copy object | ||
8676 | func testUnencryptedToUnencryptedCopyObject() { | ||
8677 | // initialize logging params | ||
8678 | startTime := time.Now() | ||
8679 | testName := getFuncName() | ||
8680 | function := "CopyObject(destination, source)" | ||
8681 | args := map[string]interface{}{} | ||
8682 | |||
8683 | // Instantiate new minio client object | ||
8684 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
8685 | &minio.Options{ | ||
8686 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
8687 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
8688 | }) | ||
8689 | if err != nil { | ||
8690 | logError(testName, function, args, startTime, "", "MinIO v2 client object creation failed", err) | ||
8691 | return | ||
8692 | } | ||
8693 | // Generate a new random bucket name. | ||
8694 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
8695 | |||
8696 | var sseSrc, sseDst encrypt.ServerSide | ||
8697 | // c.TraceOn(os.Stderr) | ||
8698 | testEncryptedCopyObjectWrapper(c, bucketName, sseSrc, sseDst) | ||
8699 | } | ||
8700 | |||
8701 | // Test encrypted copy object | ||
8702 | func testEncryptedSSECToSSECCopyObject() { | ||
8703 | // initialize logging params | ||
8704 | startTime := time.Now() | ||
8705 | testName := getFuncName() | ||
8706 | function := "CopyObject(destination, source)" | ||
8707 | args := map[string]interface{}{} | ||
8708 | |||
8709 | // Instantiate new minio client object | ||
8710 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
8711 | &minio.Options{ | ||
8712 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
8713 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
8714 | }) | ||
8715 | if err != nil { | ||
8716 | logError(testName, function, args, startTime, "", "MinIO v2 client object creation failed", err) | ||
8717 | return | ||
8718 | } | ||
8719 | // Generate a new random bucket name. | ||
8720 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
8721 | |||
8722 | sseSrc := encrypt.DefaultPBKDF([]byte("correct horse battery staple"), []byte(bucketName+"srcObject")) | ||
8723 | sseDst := encrypt.DefaultPBKDF([]byte("correct horse battery staple"), []byte(bucketName+"dstObject")) | ||
8724 | // c.TraceOn(os.Stderr) | ||
8725 | testEncryptedCopyObjectWrapper(c, bucketName, sseSrc, sseDst) | ||
8726 | } | ||
8727 | |||
8728 | // Test encrypted copy object | ||
8729 | func testEncryptedSSECToSSES3CopyObject() { | ||
8730 | // initialize logging params | ||
8731 | startTime := time.Now() | ||
8732 | testName := getFuncName() | ||
8733 | function := "CopyObject(destination, source)" | ||
8734 | args := map[string]interface{}{} | ||
8735 | |||
8736 | // Instantiate new minio client object | ||
8737 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
8738 | &minio.Options{ | ||
8739 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
8740 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
8741 | }) | ||
8742 | if err != nil { | ||
8743 | logError(testName, function, args, startTime, "", "MinIO v2 client object creation failed", err) | ||
8744 | return | ||
8745 | } | ||
8746 | // Generate a new random bucket name. | ||
8747 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
8748 | |||
8749 | sseSrc := encrypt.DefaultPBKDF([]byte("correct horse battery staple"), []byte(bucketName+"srcObject")) | ||
8750 | sseDst := encrypt.NewSSE() | ||
8751 | // c.TraceOn(os.Stderr) | ||
8752 | testEncryptedCopyObjectWrapper(c, bucketName, sseSrc, sseDst) | ||
8753 | } | ||
8754 | |||
8755 | // Test encrypted copy object | ||
8756 | func testEncryptedSSECToUnencryptedCopyObject() { | ||
8757 | // initialize logging params | ||
8758 | startTime := time.Now() | ||
8759 | testName := getFuncName() | ||
8760 | function := "CopyObject(destination, source)" | ||
8761 | args := map[string]interface{}{} | ||
8762 | |||
8763 | // Instantiate new minio client object | ||
8764 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
8765 | &minio.Options{ | ||
8766 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
8767 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
8768 | }) | ||
8769 | if err != nil { | ||
8770 | logError(testName, function, args, startTime, "", "MinIO v2 client object creation failed", err) | ||
8771 | return | ||
8772 | } | ||
8773 | // Generate a new random bucket name. | ||
8774 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
8775 | |||
8776 | sseSrc := encrypt.DefaultPBKDF([]byte("correct horse battery staple"), []byte(bucketName+"srcObject")) | ||
8777 | var sseDst encrypt.ServerSide | ||
8778 | // c.TraceOn(os.Stderr) | ||
8779 | testEncryptedCopyObjectWrapper(c, bucketName, sseSrc, sseDst) | ||
8780 | } | ||
8781 | |||
8782 | // Test encrypted copy object | ||
8783 | func testEncryptedSSES3ToSSECCopyObject() { | ||
8784 | // initialize logging params | ||
8785 | startTime := time.Now() | ||
8786 | testName := getFuncName() | ||
8787 | function := "CopyObject(destination, source)" | ||
8788 | args := map[string]interface{}{} | ||
8789 | |||
8790 | // Instantiate new minio client object | ||
8791 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
8792 | &minio.Options{ | ||
8793 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
8794 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
8795 | }) | ||
8796 | if err != nil { | ||
8797 | logError(testName, function, args, startTime, "", "MinIO v2 client object creation failed", err) | ||
8798 | return | ||
8799 | } | ||
8800 | // Generate a new random bucket name. | ||
8801 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
8802 | |||
8803 | sseSrc := encrypt.NewSSE() | ||
8804 | sseDst := encrypt.DefaultPBKDF([]byte("correct horse battery staple"), []byte(bucketName+"dstObject")) | ||
8805 | // c.TraceOn(os.Stderr) | ||
8806 | testEncryptedCopyObjectWrapper(c, bucketName, sseSrc, sseDst) | ||
8807 | } | ||
8808 | |||
8809 | // Test encrypted copy object | ||
8810 | func testEncryptedSSES3ToSSES3CopyObject() { | ||
8811 | // initialize logging params | ||
8812 | startTime := time.Now() | ||
8813 | testName := getFuncName() | ||
8814 | function := "CopyObject(destination, source)" | ||
8815 | args := map[string]interface{}{} | ||
8816 | |||
8817 | // Instantiate new minio client object | ||
8818 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
8819 | &minio.Options{ | ||
8820 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
8821 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
8822 | }) | ||
8823 | if err != nil { | ||
8824 | logError(testName, function, args, startTime, "", "MinIO v2 client object creation failed", err) | ||
8825 | return | ||
8826 | } | ||
8827 | // Generate a new random bucket name. | ||
8828 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
8829 | |||
8830 | sseSrc := encrypt.NewSSE() | ||
8831 | sseDst := encrypt.NewSSE() | ||
8832 | // c.TraceOn(os.Stderr) | ||
8833 | testEncryptedCopyObjectWrapper(c, bucketName, sseSrc, sseDst) | ||
8834 | } | ||
8835 | |||
8836 | // Test encrypted copy object | ||
8837 | func testEncryptedSSES3ToUnencryptedCopyObject() { | ||
8838 | // initialize logging params | ||
8839 | startTime := time.Now() | ||
8840 | testName := getFuncName() | ||
8841 | function := "CopyObject(destination, source)" | ||
8842 | args := map[string]interface{}{} | ||
8843 | |||
8844 | // Instantiate new minio client object | ||
8845 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
8846 | &minio.Options{ | ||
8847 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
8848 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
8849 | }) | ||
8850 | if err != nil { | ||
8851 | logError(testName, function, args, startTime, "", "MinIO v2 client object creation failed", err) | ||
8852 | return | ||
8853 | } | ||
8854 | // Generate a new random bucket name. | ||
8855 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
8856 | |||
8857 | sseSrc := encrypt.NewSSE() | ||
8858 | var sseDst encrypt.ServerSide | ||
8859 | // c.TraceOn(os.Stderr) | ||
8860 | testEncryptedCopyObjectWrapper(c, bucketName, sseSrc, sseDst) | ||
8861 | } | ||
8862 | |||
8863 | // Test encrypted copy object | ||
8864 | func testEncryptedCopyObjectV2() { | ||
8865 | // initialize logging params | ||
8866 | startTime := time.Now() | ||
8867 | testName := getFuncName() | ||
8868 | function := "CopyObject(destination, source)" | ||
8869 | args := map[string]interface{}{} | ||
8870 | |||
8871 | // Instantiate new minio client object | ||
8872 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
8873 | &minio.Options{ | ||
8874 | Creds: credentials.NewStaticV2(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
8875 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
8876 | }) | ||
8877 | if err != nil { | ||
8878 | logError(testName, function, args, startTime, "", "MinIO v2 client object creation failed", err) | ||
8879 | return | ||
8880 | } | ||
8881 | // Generate a new random bucket name. | ||
8882 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
8883 | |||
8884 | sseSrc := encrypt.DefaultPBKDF([]byte("correct horse battery staple"), []byte(bucketName+"srcObject")) | ||
8885 | sseDst := encrypt.DefaultPBKDF([]byte("correct horse battery staple"), []byte(bucketName+"dstObject")) | ||
8886 | // c.TraceOn(os.Stderr) | ||
8887 | testEncryptedCopyObjectWrapper(c, bucketName, sseSrc, sseDst) | ||
8888 | } | ||
8889 | |||
8890 | func testDecryptedCopyObject() { | ||
8891 | // initialize logging params | ||
8892 | startTime := time.Now() | ||
8893 | testName := getFuncName() | ||
8894 | function := "CopyObject(destination, source)" | ||
8895 | args := map[string]interface{}{} | ||
8896 | |||
8897 | // Instantiate new minio client object | ||
8898 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
8899 | &minio.Options{ | ||
8900 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
8901 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
8902 | }) | ||
8903 | if err != nil { | ||
8904 | logError(testName, function, args, startTime, "", "MinIO v2 client object creation failed", err) | ||
8905 | return | ||
8906 | } | ||
8907 | |||
8908 | bucketName, objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-"), "object" | ||
8909 | if err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}); err != nil { | ||
8910 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
8911 | return | ||
8912 | } | ||
8913 | |||
8914 | defer cleanupBucket(bucketName, c) | ||
8915 | |||
8916 | encryption := encrypt.DefaultPBKDF([]byte("correct horse battery staple"), []byte(bucketName+objectName)) | ||
8917 | _, err = c.PutObject(context.Background(), bucketName, objectName, bytes.NewReader(bytes.Repeat([]byte("a"), 1024*1024)), 1024*1024, minio.PutObjectOptions{ | ||
8918 | ServerSideEncryption: encryption, | ||
8919 | }) | ||
8920 | if err != nil { | ||
8921 | logError(testName, function, args, startTime, "", "PutObject call failed", err) | ||
8922 | return | ||
8923 | } | ||
8924 | |||
8925 | src := minio.CopySrcOptions{ | ||
8926 | Bucket: bucketName, | ||
8927 | Object: objectName, | ||
8928 | Encryption: encrypt.SSECopy(encryption), | ||
8929 | } | ||
8930 | args["source"] = src | ||
8931 | |||
8932 | dst := minio.CopyDestOptions{ | ||
8933 | Bucket: bucketName, | ||
8934 | Object: "decrypted-" + objectName, | ||
8935 | } | ||
8936 | args["destination"] = dst | ||
8937 | |||
8938 | if _, err = c.CopyObject(context.Background(), dst, src); err != nil { | ||
8939 | logError(testName, function, args, startTime, "", "CopyObject failed", err) | ||
8940 | return | ||
8941 | } | ||
8942 | if _, err = c.GetObject(context.Background(), bucketName, "decrypted-"+objectName, minio.GetObjectOptions{}); err != nil { | ||
8943 | logError(testName, function, args, startTime, "", "GetObject failed", err) | ||
8944 | return | ||
8945 | } | ||
8946 | successLogger(testName, function, args, startTime).Info() | ||
8947 | } | ||
8948 | |||
8949 | func testSSECMultipartEncryptedToSSECCopyObjectPart() { | ||
8950 | // initialize logging params | ||
8951 | startTime := time.Now() | ||
8952 | testName := getFuncName() | ||
8953 | function := "CopyObjectPart(destination, source)" | ||
8954 | args := map[string]interface{}{} | ||
8955 | |||
8956 | // Instantiate new minio client object | ||
8957 | client, err := minio.New(os.Getenv(serverEndpoint), | ||
8958 | &minio.Options{ | ||
8959 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
8960 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
8961 | }) | ||
8962 | if err != nil { | ||
8963 | logError(testName, function, args, startTime, "", "MinIO v4 client object creation failed", err) | ||
8964 | return | ||
8965 | } | ||
8966 | |||
8967 | // Instantiate new core client object. | ||
8968 | c := minio.Core{client} | ||
8969 | |||
8970 | // Enable tracing, write to stderr. | ||
8971 | // c.TraceOn(os.Stderr) | ||
8972 | |||
8973 | // Set user agent. | ||
8974 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
8975 | |||
8976 | // Generate a new random bucket name. | ||
8977 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test") | ||
8978 | |||
8979 | // Make a new bucket. | ||
8980 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
8981 | if err != nil { | ||
8982 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
8983 | return | ||
8984 | } | ||
8985 | defer cleanupBucket(bucketName, client) | ||
8986 | // Make a buffer with 6MB of data | ||
8987 | buf := bytes.Repeat([]byte("abcdef"), 1024*1024) | ||
8988 | |||
8989 | // Save the data | ||
8990 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
8991 | password := "correct horse battery staple" | ||
8992 | srcencryption := encrypt.DefaultPBKDF([]byte(password), []byte(bucketName+objectName)) | ||
8993 | |||
8994 | // Upload a 6MB object using multipart mechanism | ||
8995 | uploadID, err := c.NewMultipartUpload(context.Background(), bucketName, objectName, minio.PutObjectOptions{ServerSideEncryption: srcencryption}) | ||
8996 | if err != nil { | ||
8997 | logError(testName, function, args, startTime, "", "NewMultipartUpload call failed", err) | ||
8998 | return | ||
8999 | } | ||
9000 | |||
9001 | var completeParts []minio.CompletePart | ||
9002 | |||
9003 | part, err := c.PutObjectPart(context.Background(), bucketName, objectName, uploadID, 1, | ||
9004 | bytes.NewReader(buf[:5*1024*1024]), 5*1024*1024, | ||
9005 | minio.PutObjectPartOptions{SSE: srcencryption}, | ||
9006 | ) | ||
9007 | if err != nil { | ||
9008 | logError(testName, function, args, startTime, "", "PutObjectPart call failed", err) | ||
9009 | return | ||
9010 | } | ||
9011 | completeParts = append(completeParts, minio.CompletePart{PartNumber: part.PartNumber, ETag: part.ETag}) | ||
9012 | |||
9013 | part, err = c.PutObjectPart(context.Background(), bucketName, objectName, uploadID, 2, | ||
9014 | bytes.NewReader(buf[5*1024*1024:]), 1024*1024, | ||
9015 | minio.PutObjectPartOptions{SSE: srcencryption}, | ||
9016 | ) | ||
9017 | if err != nil { | ||
9018 | logError(testName, function, args, startTime, "", "PutObjectPart call failed", err) | ||
9019 | return | ||
9020 | } | ||
9021 | completeParts = append(completeParts, minio.CompletePart{PartNumber: part.PartNumber, ETag: part.ETag}) | ||
9022 | |||
9023 | // Complete the multipart upload | ||
9024 | _, err = c.CompleteMultipartUpload(context.Background(), bucketName, objectName, uploadID, completeParts, minio.PutObjectOptions{}) | ||
9025 | if err != nil { | ||
9026 | logError(testName, function, args, startTime, "", "CompleteMultipartUpload call failed", err) | ||
9027 | return | ||
9028 | } | ||
9029 | |||
9030 | // Stat the object and check its length matches | ||
9031 | objInfo, err := c.StatObject(context.Background(), bucketName, objectName, minio.StatObjectOptions{ServerSideEncryption: srcencryption}) | ||
9032 | if err != nil { | ||
9033 | logError(testName, function, args, startTime, "", "StatObject call failed", err) | ||
9034 | return | ||
9035 | } | ||
9036 | |||
9037 | destBucketName := bucketName | ||
9038 | destObjectName := objectName + "-dest" | ||
9039 | dstencryption := encrypt.DefaultPBKDF([]byte(password), []byte(destBucketName+destObjectName)) | ||
9040 | |||
9041 | uploadID, err = c.NewMultipartUpload(context.Background(), destBucketName, destObjectName, minio.PutObjectOptions{ServerSideEncryption: dstencryption}) | ||
9042 | if err != nil { | ||
9043 | logError(testName, function, args, startTime, "", "NewMultipartUpload call failed", err) | ||
9044 | return | ||
9045 | } | ||
9046 | |||
9047 | // Content of the destination object will be two copies of | ||
9048 | // `objectName` concatenated, followed by first byte of | ||
9049 | // `objectName`. | ||
9050 | metadata := make(map[string]string) | ||
9051 | header := make(http.Header) | ||
9052 | encrypt.SSECopy(srcencryption).Marshal(header) | ||
9053 | dstencryption.Marshal(header) | ||
9054 | for k, v := range header { | ||
9055 | metadata[k] = v[0] | ||
9056 | } | ||
9057 | |||
9058 | metadata["x-amz-copy-source-if-match"] = objInfo.ETag | ||
9059 | |||
9060 | // First of three parts | ||
9061 | fstPart, err := c.CopyObjectPart(context.Background(), bucketName, objectName, destBucketName, destObjectName, uploadID, 1, 0, -1, metadata) | ||
9062 | if err != nil { | ||
9063 | logError(testName, function, args, startTime, "", "CopyObjectPart call failed", err) | ||
9064 | return | ||
9065 | } | ||
9066 | |||
9067 | // Second of three parts | ||
9068 | sndPart, err := c.CopyObjectPart(context.Background(), bucketName, objectName, destBucketName, destObjectName, uploadID, 2, 0, -1, metadata) | ||
9069 | if err != nil { | ||
9070 | logError(testName, function, args, startTime, "", "CopyObjectPart call failed", err) | ||
9071 | return | ||
9072 | } | ||
9073 | |||
9074 | // Last of three parts | ||
9075 | lstPart, err := c.CopyObjectPart(context.Background(), bucketName, objectName, destBucketName, destObjectName, uploadID, 3, 0, 1, metadata) | ||
9076 | if err != nil { | ||
9077 | logError(testName, function, args, startTime, "", "CopyObjectPart call failed", err) | ||
9078 | return | ||
9079 | } | ||
9080 | |||
9081 | // Complete the multipart upload | ||
9082 | _, err = c.CompleteMultipartUpload(context.Background(), destBucketName, destObjectName, uploadID, []minio.CompletePart{fstPart, sndPart, lstPart}, minio.PutObjectOptions{}) | ||
9083 | if err != nil { | ||
9084 | logError(testName, function, args, startTime, "", "CompleteMultipartUpload call failed", err) | ||
9085 | return | ||
9086 | } | ||
9087 | |||
9088 | // Stat the object and check its length matches | ||
9089 | objInfo, err = c.StatObject(context.Background(), destBucketName, destObjectName, minio.StatObjectOptions{ServerSideEncryption: dstencryption}) | ||
9090 | if err != nil { | ||
9091 | logError(testName, function, args, startTime, "", "StatObject call failed", err) | ||
9092 | return | ||
9093 | } | ||
9094 | |||
9095 | if objInfo.Size != (6*1024*1024)*2+1 { | ||
9096 | logError(testName, function, args, startTime, "", "Destination object has incorrect size!", err) | ||
9097 | return | ||
9098 | } | ||
9099 | |||
9100 | // Now we read the data back | ||
9101 | getOpts := minio.GetObjectOptions{ServerSideEncryption: dstencryption} | ||
9102 | getOpts.SetRange(0, 6*1024*1024-1) | ||
9103 | r, _, _, err := c.GetObject(context.Background(), destBucketName, destObjectName, getOpts) | ||
9104 | if err != nil { | ||
9105 | logError(testName, function, args, startTime, "", "GetObject call failed", err) | ||
9106 | return | ||
9107 | } | ||
9108 | getBuf := make([]byte, 6*1024*1024) | ||
9109 | _, err = readFull(r, getBuf) | ||
9110 | if err != nil { | ||
9111 | logError(testName, function, args, startTime, "", "Read buffer failed", err) | ||
9112 | return | ||
9113 | } | ||
9114 | if !bytes.Equal(getBuf, buf) { | ||
9115 | logError(testName, function, args, startTime, "", "Got unexpected data in first 6MB", err) | ||
9116 | return | ||
9117 | } | ||
9118 | |||
9119 | getOpts.SetRange(6*1024*1024, 0) | ||
9120 | r, _, _, err = c.GetObject(context.Background(), destBucketName, destObjectName, getOpts) | ||
9121 | if err != nil { | ||
9122 | logError(testName, function, args, startTime, "", "GetObject call failed", err) | ||
9123 | return | ||
9124 | } | ||
9125 | getBuf = make([]byte, 6*1024*1024+1) | ||
9126 | _, err = readFull(r, getBuf) | ||
9127 | if err != nil { | ||
9128 | logError(testName, function, args, startTime, "", "Read buffer failed", err) | ||
9129 | return | ||
9130 | } | ||
9131 | if !bytes.Equal(getBuf[:6*1024*1024], buf) { | ||
9132 | logError(testName, function, args, startTime, "", "Got unexpected data in second 6MB", err) | ||
9133 | return | ||
9134 | } | ||
9135 | if getBuf[6*1024*1024] != buf[0] { | ||
9136 | logError(testName, function, args, startTime, "", "Got unexpected data in last byte of copied object!", err) | ||
9137 | return | ||
9138 | } | ||
9139 | |||
9140 | successLogger(testName, function, args, startTime).Info() | ||
9141 | |||
9142 | // Do not need to remove destBucketName its same as bucketName. | ||
9143 | } | ||
9144 | |||
9145 | // Test Core CopyObjectPart implementation | ||
9146 | func testSSECEncryptedToSSECCopyObjectPart() { | ||
9147 | // initialize logging params | ||
9148 | startTime := time.Now() | ||
9149 | testName := getFuncName() | ||
9150 | function := "CopyObjectPart(destination, source)" | ||
9151 | args := map[string]interface{}{} | ||
9152 | |||
9153 | // Instantiate new minio client object | ||
9154 | client, err := minio.New(os.Getenv(serverEndpoint), | ||
9155 | &minio.Options{ | ||
9156 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
9157 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
9158 | }) | ||
9159 | if err != nil { | ||
9160 | logError(testName, function, args, startTime, "", "MinIO v4 client object creation failed", err) | ||
9161 | return | ||
9162 | } | ||
9163 | |||
9164 | // Instantiate new core client object. | ||
9165 | c := minio.Core{client} | ||
9166 | |||
9167 | // Enable tracing, write to stderr. | ||
9168 | // c.TraceOn(os.Stderr) | ||
9169 | |||
9170 | // Set user agent. | ||
9171 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
9172 | |||
9173 | // Generate a new random bucket name. | ||
9174 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test") | ||
9175 | |||
9176 | // Make a new bucket. | ||
9177 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
9178 | if err != nil { | ||
9179 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
9180 | return | ||
9181 | } | ||
9182 | defer cleanupBucket(bucketName, client) | ||
9183 | // Make a buffer with 5MB of data | ||
9184 | buf := bytes.Repeat([]byte("abcde"), 1024*1024) | ||
9185 | |||
9186 | // Save the data | ||
9187 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
9188 | password := "correct horse battery staple" | ||
9189 | srcencryption := encrypt.DefaultPBKDF([]byte(password), []byte(bucketName+objectName)) | ||
9190 | putmetadata := map[string]string{ | ||
9191 | "Content-Type": "binary/octet-stream", | ||
9192 | } | ||
9193 | opts := minio.PutObjectOptions{ | ||
9194 | UserMetadata: putmetadata, | ||
9195 | ServerSideEncryption: srcencryption, | ||
9196 | } | ||
9197 | uploadInfo, err := c.PutObject(context.Background(), bucketName, objectName, bytes.NewReader(buf), int64(len(buf)), "", "", opts) | ||
9198 | if err != nil { | ||
9199 | logError(testName, function, args, startTime, "", "PutObject call failed", err) | ||
9200 | return | ||
9201 | } | ||
9202 | |||
9203 | st, err := c.StatObject(context.Background(), bucketName, objectName, minio.StatObjectOptions{ServerSideEncryption: srcencryption}) | ||
9204 | if err != nil { | ||
9205 | logError(testName, function, args, startTime, "", "StatObject call failed", err) | ||
9206 | return | ||
9207 | } | ||
9208 | |||
9209 | if st.Size != int64(len(buf)) { | ||
9210 | logError(testName, function, args, startTime, "", fmt.Sprintf("Error: number of bytes does not match, want %v, got %v\n", len(buf), st.Size), err) | ||
9211 | return | ||
9212 | } | ||
9213 | |||
9214 | destBucketName := bucketName | ||
9215 | destObjectName := objectName + "-dest" | ||
9216 | dstencryption := encrypt.DefaultPBKDF([]byte(password), []byte(destBucketName+destObjectName)) | ||
9217 | |||
9218 | uploadID, err := c.NewMultipartUpload(context.Background(), destBucketName, destObjectName, minio.PutObjectOptions{ServerSideEncryption: dstencryption}) | ||
9219 | if err != nil { | ||
9220 | logError(testName, function, args, startTime, "", "NewMultipartUpload call failed", err) | ||
9221 | return | ||
9222 | } | ||
9223 | |||
9224 | // Content of the destination object will be two copies of | ||
9225 | // `objectName` concatenated, followed by first byte of | ||
9226 | // `objectName`. | ||
9227 | metadata := make(map[string]string) | ||
9228 | header := make(http.Header) | ||
9229 | encrypt.SSECopy(srcencryption).Marshal(header) | ||
9230 | dstencryption.Marshal(header) | ||
9231 | for k, v := range header { | ||
9232 | metadata[k] = v[0] | ||
9233 | } | ||
9234 | |||
9235 | metadata["x-amz-copy-source-if-match"] = uploadInfo.ETag | ||
9236 | |||
9237 | // First of three parts | ||
9238 | fstPart, err := c.CopyObjectPart(context.Background(), bucketName, objectName, destBucketName, destObjectName, uploadID, 1, 0, -1, metadata) | ||
9239 | if err != nil { | ||
9240 | logError(testName, function, args, startTime, "", "CopyObjectPart call failed", err) | ||
9241 | return | ||
9242 | } | ||
9243 | |||
9244 | // Second of three parts | ||
9245 | sndPart, err := c.CopyObjectPart(context.Background(), bucketName, objectName, destBucketName, destObjectName, uploadID, 2, 0, -1, metadata) | ||
9246 | if err != nil { | ||
9247 | logError(testName, function, args, startTime, "", "CopyObjectPart call failed", err) | ||
9248 | return | ||
9249 | } | ||
9250 | |||
9251 | // Last of three parts | ||
9252 | lstPart, err := c.CopyObjectPart(context.Background(), bucketName, objectName, destBucketName, destObjectName, uploadID, 3, 0, 1, metadata) | ||
9253 | if err != nil { | ||
9254 | logError(testName, function, args, startTime, "", "CopyObjectPart call failed", err) | ||
9255 | return | ||
9256 | } | ||
9257 | |||
9258 | // Complete the multipart upload | ||
9259 | _, err = c.CompleteMultipartUpload(context.Background(), destBucketName, destObjectName, uploadID, []minio.CompletePart{fstPart, sndPart, lstPart}, minio.PutObjectOptions{}) | ||
9260 | if err != nil { | ||
9261 | logError(testName, function, args, startTime, "", "CompleteMultipartUpload call failed", err) | ||
9262 | return | ||
9263 | } | ||
9264 | |||
9265 | // Stat the object and check its length matches | ||
9266 | objInfo, err := c.StatObject(context.Background(), destBucketName, destObjectName, minio.StatObjectOptions{ServerSideEncryption: dstencryption}) | ||
9267 | if err != nil { | ||
9268 | logError(testName, function, args, startTime, "", "StatObject call failed", err) | ||
9269 | return | ||
9270 | } | ||
9271 | |||
9272 | if objInfo.Size != (5*1024*1024)*2+1 { | ||
9273 | logError(testName, function, args, startTime, "", "Destination object has incorrect size!", err) | ||
9274 | return | ||
9275 | } | ||
9276 | |||
9277 | // Now we read the data back | ||
9278 | getOpts := minio.GetObjectOptions{ServerSideEncryption: dstencryption} | ||
9279 | getOpts.SetRange(0, 5*1024*1024-1) | ||
9280 | r, _, _, err := c.GetObject(context.Background(), destBucketName, destObjectName, getOpts) | ||
9281 | if err != nil { | ||
9282 | logError(testName, function, args, startTime, "", "GetObject call failed", err) | ||
9283 | return | ||
9284 | } | ||
9285 | getBuf := make([]byte, 5*1024*1024) | ||
9286 | _, err = readFull(r, getBuf) | ||
9287 | if err != nil { | ||
9288 | logError(testName, function, args, startTime, "", "Read buffer failed", err) | ||
9289 | return | ||
9290 | } | ||
9291 | if !bytes.Equal(getBuf, buf) { | ||
9292 | logError(testName, function, args, startTime, "", "Got unexpected data in first 5MB", err) | ||
9293 | return | ||
9294 | } | ||
9295 | |||
9296 | getOpts.SetRange(5*1024*1024, 0) | ||
9297 | r, _, _, err = c.GetObject(context.Background(), destBucketName, destObjectName, getOpts) | ||
9298 | if err != nil { | ||
9299 | logError(testName, function, args, startTime, "", "GetObject call failed", err) | ||
9300 | return | ||
9301 | } | ||
9302 | getBuf = make([]byte, 5*1024*1024+1) | ||
9303 | _, err = readFull(r, getBuf) | ||
9304 | if err != nil { | ||
9305 | logError(testName, function, args, startTime, "", "Read buffer failed", err) | ||
9306 | return | ||
9307 | } | ||
9308 | if !bytes.Equal(getBuf[:5*1024*1024], buf) { | ||
9309 | logError(testName, function, args, startTime, "", "Got unexpected data in second 5MB", err) | ||
9310 | return | ||
9311 | } | ||
9312 | if getBuf[5*1024*1024] != buf[0] { | ||
9313 | logError(testName, function, args, startTime, "", "Got unexpected data in last byte of copied object!", err) | ||
9314 | return | ||
9315 | } | ||
9316 | |||
9317 | successLogger(testName, function, args, startTime).Info() | ||
9318 | |||
9319 | // Do not need to remove destBucketName its same as bucketName. | ||
9320 | } | ||
9321 | |||
9322 | // Test Core CopyObjectPart implementation for SSEC encrypted to unencrypted copy | ||
9323 | func testSSECEncryptedToUnencryptedCopyPart() { | ||
9324 | // initialize logging params | ||
9325 | startTime := time.Now() | ||
9326 | testName := getFuncName() | ||
9327 | function := "CopyObjectPart(destination, source)" | ||
9328 | args := map[string]interface{}{} | ||
9329 | |||
9330 | // Instantiate new minio client object | ||
9331 | client, err := minio.New(os.Getenv(serverEndpoint), | ||
9332 | &minio.Options{ | ||
9333 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
9334 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
9335 | }) | ||
9336 | if err != nil { | ||
9337 | logError(testName, function, args, startTime, "", "MinIO v4 client object creation failed", err) | ||
9338 | return | ||
9339 | } | ||
9340 | |||
9341 | // Instantiate new core client object. | ||
9342 | c := minio.Core{client} | ||
9343 | |||
9344 | // Enable tracing, write to stderr. | ||
9345 | // c.TraceOn(os.Stderr) | ||
9346 | |||
9347 | // Set user agent. | ||
9348 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
9349 | |||
9350 | // Generate a new random bucket name. | ||
9351 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test") | ||
9352 | |||
9353 | // Make a new bucket. | ||
9354 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
9355 | if err != nil { | ||
9356 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
9357 | return | ||
9358 | } | ||
9359 | defer cleanupBucket(bucketName, client) | ||
9360 | // Make a buffer with 5MB of data | ||
9361 | buf := bytes.Repeat([]byte("abcde"), 1024*1024) | ||
9362 | |||
9363 | // Save the data | ||
9364 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
9365 | password := "correct horse battery staple" | ||
9366 | srcencryption := encrypt.DefaultPBKDF([]byte(password), []byte(bucketName+objectName)) | ||
9367 | |||
9368 | opts := minio.PutObjectOptions{ | ||
9369 | UserMetadata: map[string]string{ | ||
9370 | "Content-Type": "binary/octet-stream", | ||
9371 | }, | ||
9372 | ServerSideEncryption: srcencryption, | ||
9373 | } | ||
9374 | uploadInfo, err := c.PutObject(context.Background(), bucketName, objectName, bytes.NewReader(buf), int64(len(buf)), "", "", opts) | ||
9375 | if err != nil { | ||
9376 | logError(testName, function, args, startTime, "", "PutObject call failed", err) | ||
9377 | return | ||
9378 | } | ||
9379 | |||
9380 | st, err := c.StatObject(context.Background(), bucketName, objectName, minio.StatObjectOptions{ServerSideEncryption: srcencryption}) | ||
9381 | if err != nil { | ||
9382 | logError(testName, function, args, startTime, "", "StatObject call failed", err) | ||
9383 | return | ||
9384 | } | ||
9385 | |||
9386 | if st.Size != int64(len(buf)) { | ||
9387 | logError(testName, function, args, startTime, "", fmt.Sprintf("Error: number of bytes does not match, want %v, got %v\n", len(buf), st.Size), err) | ||
9388 | return | ||
9389 | } | ||
9390 | |||
9391 | destBucketName := bucketName | ||
9392 | destObjectName := objectName + "-dest" | ||
9393 | var dstencryption encrypt.ServerSide | ||
9394 | |||
9395 | uploadID, err := c.NewMultipartUpload(context.Background(), destBucketName, destObjectName, minio.PutObjectOptions{ServerSideEncryption: dstencryption}) | ||
9396 | if err != nil { | ||
9397 | logError(testName, function, args, startTime, "", "NewMultipartUpload call failed", err) | ||
9398 | return | ||
9399 | } | ||
9400 | |||
9401 | // Content of the destination object will be two copies of | ||
9402 | // `objectName` concatenated, followed by first byte of | ||
9403 | // `objectName`. | ||
9404 | metadata := make(map[string]string) | ||
9405 | header := make(http.Header) | ||
9406 | encrypt.SSECopy(srcencryption).Marshal(header) | ||
9407 | for k, v := range header { | ||
9408 | metadata[k] = v[0] | ||
9409 | } | ||
9410 | |||
9411 | metadata["x-amz-copy-source-if-match"] = uploadInfo.ETag | ||
9412 | |||
9413 | // First of three parts | ||
9414 | fstPart, err := c.CopyObjectPart(context.Background(), bucketName, objectName, destBucketName, destObjectName, uploadID, 1, 0, -1, metadata) | ||
9415 | if err != nil { | ||
9416 | logError(testName, function, args, startTime, "", "CopyObjectPart call failed", err) | ||
9417 | return | ||
9418 | } | ||
9419 | |||
9420 | // Second of three parts | ||
9421 | sndPart, err := c.CopyObjectPart(context.Background(), bucketName, objectName, destBucketName, destObjectName, uploadID, 2, 0, -1, metadata) | ||
9422 | if err != nil { | ||
9423 | logError(testName, function, args, startTime, "", "CopyObjectPart call failed", err) | ||
9424 | return | ||
9425 | } | ||
9426 | |||
9427 | // Last of three parts | ||
9428 | lstPart, err := c.CopyObjectPart(context.Background(), bucketName, objectName, destBucketName, destObjectName, uploadID, 3, 0, 1, metadata) | ||
9429 | if err != nil { | ||
9430 | logError(testName, function, args, startTime, "", "CopyObjectPart call failed", err) | ||
9431 | return | ||
9432 | } | ||
9433 | |||
9434 | // Complete the multipart upload | ||
9435 | _, err = c.CompleteMultipartUpload(context.Background(), destBucketName, destObjectName, uploadID, []minio.CompletePart{fstPart, sndPart, lstPart}, minio.PutObjectOptions{}) | ||
9436 | if err != nil { | ||
9437 | logError(testName, function, args, startTime, "", "CompleteMultipartUpload call failed", err) | ||
9438 | return | ||
9439 | } | ||
9440 | |||
9441 | // Stat the object and check its length matches | ||
9442 | objInfo, err := c.StatObject(context.Background(), destBucketName, destObjectName, minio.StatObjectOptions{}) | ||
9443 | if err != nil { | ||
9444 | logError(testName, function, args, startTime, "", "StatObject call failed", err) | ||
9445 | return | ||
9446 | } | ||
9447 | |||
9448 | if objInfo.Size != (5*1024*1024)*2+1 { | ||
9449 | logError(testName, function, args, startTime, "", "Destination object has incorrect size!", err) | ||
9450 | return | ||
9451 | } | ||
9452 | |||
9453 | // Now we read the data back | ||
9454 | getOpts := minio.GetObjectOptions{} | ||
9455 | getOpts.SetRange(0, 5*1024*1024-1) | ||
9456 | r, _, _, err := c.GetObject(context.Background(), destBucketName, destObjectName, getOpts) | ||
9457 | if err != nil { | ||
9458 | logError(testName, function, args, startTime, "", "GetObject call failed", err) | ||
9459 | return | ||
9460 | } | ||
9461 | getBuf := make([]byte, 5*1024*1024) | ||
9462 | _, err = readFull(r, getBuf) | ||
9463 | if err != nil { | ||
9464 | logError(testName, function, args, startTime, "", "Read buffer failed", err) | ||
9465 | return | ||
9466 | } | ||
9467 | if !bytes.Equal(getBuf, buf) { | ||
9468 | logError(testName, function, args, startTime, "", "Got unexpected data in first 5MB", err) | ||
9469 | return | ||
9470 | } | ||
9471 | |||
9472 | getOpts.SetRange(5*1024*1024, 0) | ||
9473 | r, _, _, err = c.GetObject(context.Background(), destBucketName, destObjectName, getOpts) | ||
9474 | if err != nil { | ||
9475 | logError(testName, function, args, startTime, "", "GetObject call failed", err) | ||
9476 | return | ||
9477 | } | ||
9478 | getBuf = make([]byte, 5*1024*1024+1) | ||
9479 | _, err = readFull(r, getBuf) | ||
9480 | if err != nil { | ||
9481 | logError(testName, function, args, startTime, "", "Read buffer failed", err) | ||
9482 | return | ||
9483 | } | ||
9484 | if !bytes.Equal(getBuf[:5*1024*1024], buf) { | ||
9485 | logError(testName, function, args, startTime, "", "Got unexpected data in second 5MB", err) | ||
9486 | return | ||
9487 | } | ||
9488 | if getBuf[5*1024*1024] != buf[0] { | ||
9489 | logError(testName, function, args, startTime, "", "Got unexpected data in last byte of copied object!", err) | ||
9490 | return | ||
9491 | } | ||
9492 | |||
9493 | successLogger(testName, function, args, startTime).Info() | ||
9494 | |||
9495 | // Do not need to remove destBucketName its same as bucketName. | ||
9496 | } | ||
9497 | |||
9498 | // Test Core CopyObjectPart implementation for SSEC encrypted to SSE-S3 encrypted copy | ||
9499 | func testSSECEncryptedToSSES3CopyObjectPart() { | ||
9500 | // initialize logging params | ||
9501 | startTime := time.Now() | ||
9502 | testName := getFuncName() | ||
9503 | function := "CopyObjectPart(destination, source)" | ||
9504 | args := map[string]interface{}{} | ||
9505 | |||
9506 | // Instantiate new minio client object | ||
9507 | client, err := minio.New(os.Getenv(serverEndpoint), | ||
9508 | &minio.Options{ | ||
9509 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
9510 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
9511 | }) | ||
9512 | if err != nil { | ||
9513 | logError(testName, function, args, startTime, "", "MinIO v4 client object creation failed", err) | ||
9514 | return | ||
9515 | } | ||
9516 | |||
9517 | // Instantiate new core client object. | ||
9518 | c := minio.Core{client} | ||
9519 | |||
9520 | // Enable tracing, write to stderr. | ||
9521 | // c.TraceOn(os.Stderr) | ||
9522 | |||
9523 | // Set user agent. | ||
9524 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
9525 | |||
9526 | // Generate a new random bucket name. | ||
9527 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test") | ||
9528 | |||
9529 | // Make a new bucket. | ||
9530 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
9531 | if err != nil { | ||
9532 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
9533 | return | ||
9534 | } | ||
9535 | defer cleanupBucket(bucketName, client) | ||
9536 | // Make a buffer with 5MB of data | ||
9537 | buf := bytes.Repeat([]byte("abcde"), 1024*1024) | ||
9538 | |||
9539 | // Save the data | ||
9540 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
9541 | password := "correct horse battery staple" | ||
9542 | srcencryption := encrypt.DefaultPBKDF([]byte(password), []byte(bucketName+objectName)) | ||
9543 | putmetadata := map[string]string{ | ||
9544 | "Content-Type": "binary/octet-stream", | ||
9545 | } | ||
9546 | opts := minio.PutObjectOptions{ | ||
9547 | UserMetadata: putmetadata, | ||
9548 | ServerSideEncryption: srcencryption, | ||
9549 | } | ||
9550 | |||
9551 | uploadInfo, err := c.PutObject(context.Background(), bucketName, objectName, bytes.NewReader(buf), int64(len(buf)), "", "", opts) | ||
9552 | if err != nil { | ||
9553 | logError(testName, function, args, startTime, "", "PutObject call failed", err) | ||
9554 | return | ||
9555 | } | ||
9556 | |||
9557 | st, err := c.StatObject(context.Background(), bucketName, objectName, minio.StatObjectOptions{ServerSideEncryption: srcencryption}) | ||
9558 | if err != nil { | ||
9559 | logError(testName, function, args, startTime, "", "StatObject call failed", err) | ||
9560 | return | ||
9561 | } | ||
9562 | |||
9563 | if st.Size != int64(len(buf)) { | ||
9564 | logError(testName, function, args, startTime, "", fmt.Sprintf("Error: number of bytes does not match, want %v, got %v\n", len(buf), st.Size), err) | ||
9565 | return | ||
9566 | } | ||
9567 | |||
9568 | destBucketName := bucketName | ||
9569 | destObjectName := objectName + "-dest" | ||
9570 | dstencryption := encrypt.NewSSE() | ||
9571 | |||
9572 | uploadID, err := c.NewMultipartUpload(context.Background(), destBucketName, destObjectName, minio.PutObjectOptions{ServerSideEncryption: dstencryption}) | ||
9573 | if err != nil { | ||
9574 | logError(testName, function, args, startTime, "", "NewMultipartUpload call failed", err) | ||
9575 | return | ||
9576 | } | ||
9577 | |||
9578 | // Content of the destination object will be two copies of | ||
9579 | // `objectName` concatenated, followed by first byte of | ||
9580 | // `objectName`. | ||
9581 | metadata := make(map[string]string) | ||
9582 | header := make(http.Header) | ||
9583 | encrypt.SSECopy(srcencryption).Marshal(header) | ||
9584 | dstencryption.Marshal(header) | ||
9585 | |||
9586 | for k, v := range header { | ||
9587 | metadata[k] = v[0] | ||
9588 | } | ||
9589 | |||
9590 | metadata["x-amz-copy-source-if-match"] = uploadInfo.ETag | ||
9591 | |||
9592 | // First of three parts | ||
9593 | fstPart, err := c.CopyObjectPart(context.Background(), bucketName, objectName, destBucketName, destObjectName, uploadID, 1, 0, -1, metadata) | ||
9594 | if err != nil { | ||
9595 | logError(testName, function, args, startTime, "", "CopyObjectPart call failed", err) | ||
9596 | return | ||
9597 | } | ||
9598 | |||
9599 | // Second of three parts | ||
9600 | sndPart, err := c.CopyObjectPart(context.Background(), bucketName, objectName, destBucketName, destObjectName, uploadID, 2, 0, -1, metadata) | ||
9601 | if err != nil { | ||
9602 | logError(testName, function, args, startTime, "", "CopyObjectPart call failed", err) | ||
9603 | return | ||
9604 | } | ||
9605 | |||
9606 | // Last of three parts | ||
9607 | lstPart, err := c.CopyObjectPart(context.Background(), bucketName, objectName, destBucketName, destObjectName, uploadID, 3, 0, 1, metadata) | ||
9608 | if err != nil { | ||
9609 | logError(testName, function, args, startTime, "", "CopyObjectPart call failed", err) | ||
9610 | return | ||
9611 | } | ||
9612 | |||
9613 | // Complete the multipart upload | ||
9614 | _, err = c.CompleteMultipartUpload(context.Background(), destBucketName, destObjectName, uploadID, []minio.CompletePart{fstPart, sndPart, lstPart}, minio.PutObjectOptions{}) | ||
9615 | if err != nil { | ||
9616 | logError(testName, function, args, startTime, "", "CompleteMultipartUpload call failed", err) | ||
9617 | return | ||
9618 | } | ||
9619 | |||
9620 | // Stat the object and check its length matches | ||
9621 | objInfo, err := c.StatObject(context.Background(), destBucketName, destObjectName, minio.StatObjectOptions{}) | ||
9622 | if err != nil { | ||
9623 | logError(testName, function, args, startTime, "", "StatObject call failed", err) | ||
9624 | return | ||
9625 | } | ||
9626 | |||
9627 | if objInfo.Size != (5*1024*1024)*2+1 { | ||
9628 | logError(testName, function, args, startTime, "", "Destination object has incorrect size!", err) | ||
9629 | return | ||
9630 | } | ||
9631 | |||
9632 | // Now we read the data back | ||
9633 | getOpts := minio.GetObjectOptions{} | ||
9634 | getOpts.SetRange(0, 5*1024*1024-1) | ||
9635 | r, _, _, err := c.GetObject(context.Background(), destBucketName, destObjectName, getOpts) | ||
9636 | if err != nil { | ||
9637 | logError(testName, function, args, startTime, "", "GetObject call failed", err) | ||
9638 | return | ||
9639 | } | ||
9640 | getBuf := make([]byte, 5*1024*1024) | ||
9641 | _, err = readFull(r, getBuf) | ||
9642 | if err != nil { | ||
9643 | logError(testName, function, args, startTime, "", "Read buffer failed", err) | ||
9644 | return | ||
9645 | } | ||
9646 | if !bytes.Equal(getBuf, buf) { | ||
9647 | logError(testName, function, args, startTime, "", "Got unexpected data in first 5MB", err) | ||
9648 | return | ||
9649 | } | ||
9650 | |||
9651 | getOpts.SetRange(5*1024*1024, 0) | ||
9652 | r, _, _, err = c.GetObject(context.Background(), destBucketName, destObjectName, getOpts) | ||
9653 | if err != nil { | ||
9654 | logError(testName, function, args, startTime, "", "GetObject call failed", err) | ||
9655 | return | ||
9656 | } | ||
9657 | getBuf = make([]byte, 5*1024*1024+1) | ||
9658 | _, err = readFull(r, getBuf) | ||
9659 | if err != nil { | ||
9660 | logError(testName, function, args, startTime, "", "Read buffer failed", err) | ||
9661 | return | ||
9662 | } | ||
9663 | if !bytes.Equal(getBuf[:5*1024*1024], buf) { | ||
9664 | logError(testName, function, args, startTime, "", "Got unexpected data in second 5MB", err) | ||
9665 | return | ||
9666 | } | ||
9667 | if getBuf[5*1024*1024] != buf[0] { | ||
9668 | logError(testName, function, args, startTime, "", "Got unexpected data in last byte of copied object!", err) | ||
9669 | return | ||
9670 | } | ||
9671 | |||
9672 | successLogger(testName, function, args, startTime).Info() | ||
9673 | |||
9674 | // Do not need to remove destBucketName its same as bucketName. | ||
9675 | } | ||
9676 | |||
9677 | // Test Core CopyObjectPart implementation for unencrypted to SSEC encryption copy part | ||
9678 | func testUnencryptedToSSECCopyObjectPart() { | ||
9679 | // initialize logging params | ||
9680 | startTime := time.Now() | ||
9681 | testName := getFuncName() | ||
9682 | function := "CopyObjectPart(destination, source)" | ||
9683 | args := map[string]interface{}{} | ||
9684 | |||
9685 | // Instantiate new minio client object | ||
9686 | client, err := minio.New(os.Getenv(serverEndpoint), | ||
9687 | &minio.Options{ | ||
9688 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
9689 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
9690 | }) | ||
9691 | if err != nil { | ||
9692 | logError(testName, function, args, startTime, "", "MinIO v4 client object creation failed", err) | ||
9693 | return | ||
9694 | } | ||
9695 | |||
9696 | // Instantiate new core client object. | ||
9697 | c := minio.Core{client} | ||
9698 | |||
9699 | // Enable tracing, write to stderr. | ||
9700 | // c.TraceOn(os.Stderr) | ||
9701 | |||
9702 | // Set user agent. | ||
9703 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
9704 | |||
9705 | // Generate a new random bucket name. | ||
9706 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test") | ||
9707 | |||
9708 | // Make a new bucket. | ||
9709 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
9710 | if err != nil { | ||
9711 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
9712 | return | ||
9713 | } | ||
9714 | defer cleanupBucket(bucketName, client) | ||
9715 | // Make a buffer with 5MB of data | ||
9716 | buf := bytes.Repeat([]byte("abcde"), 1024*1024) | ||
9717 | |||
9718 | // Save the data | ||
9719 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
9720 | password := "correct horse battery staple" | ||
9721 | putmetadata := map[string]string{ | ||
9722 | "Content-Type": "binary/octet-stream", | ||
9723 | } | ||
9724 | opts := minio.PutObjectOptions{ | ||
9725 | UserMetadata: putmetadata, | ||
9726 | } | ||
9727 | uploadInfo, err := c.PutObject(context.Background(), bucketName, objectName, bytes.NewReader(buf), int64(len(buf)), "", "", opts) | ||
9728 | if err != nil { | ||
9729 | logError(testName, function, args, startTime, "", "PutObject call failed", err) | ||
9730 | return | ||
9731 | } | ||
9732 | |||
9733 | st, err := c.StatObject(context.Background(), bucketName, objectName, minio.StatObjectOptions{}) | ||
9734 | if err != nil { | ||
9735 | logError(testName, function, args, startTime, "", "StatObject call failed", err) | ||
9736 | return | ||
9737 | } | ||
9738 | |||
9739 | if st.Size != int64(len(buf)) { | ||
9740 | logError(testName, function, args, startTime, "", fmt.Sprintf("Error: number of bytes does not match, want %v, got %v\n", len(buf), st.Size), err) | ||
9741 | return | ||
9742 | } | ||
9743 | |||
9744 | destBucketName := bucketName | ||
9745 | destObjectName := objectName + "-dest" | ||
9746 | dstencryption := encrypt.DefaultPBKDF([]byte(password), []byte(destBucketName+destObjectName)) | ||
9747 | |||
9748 | uploadID, err := c.NewMultipartUpload(context.Background(), destBucketName, destObjectName, minio.PutObjectOptions{ServerSideEncryption: dstencryption}) | ||
9749 | if err != nil { | ||
9750 | logError(testName, function, args, startTime, "", "NewMultipartUpload call failed", err) | ||
9751 | return | ||
9752 | } | ||
9753 | |||
9754 | // Content of the destination object will be two copies of | ||
9755 | // `objectName` concatenated, followed by first byte of | ||
9756 | // `objectName`. | ||
9757 | metadata := make(map[string]string) | ||
9758 | header := make(http.Header) | ||
9759 | dstencryption.Marshal(header) | ||
9760 | for k, v := range header { | ||
9761 | metadata[k] = v[0] | ||
9762 | } | ||
9763 | |||
9764 | metadata["x-amz-copy-source-if-match"] = uploadInfo.ETag | ||
9765 | |||
9766 | // First of three parts | ||
9767 | fstPart, err := c.CopyObjectPart(context.Background(), bucketName, objectName, destBucketName, destObjectName, uploadID, 1, 0, -1, metadata) | ||
9768 | if err != nil { | ||
9769 | logError(testName, function, args, startTime, "", "CopyObjectPart call failed", err) | ||
9770 | return | ||
9771 | } | ||
9772 | |||
9773 | // Second of three parts | ||
9774 | sndPart, err := c.CopyObjectPart(context.Background(), bucketName, objectName, destBucketName, destObjectName, uploadID, 2, 0, -1, metadata) | ||
9775 | if err != nil { | ||
9776 | logError(testName, function, args, startTime, "", "CopyObjectPart call failed", err) | ||
9777 | return | ||
9778 | } | ||
9779 | |||
9780 | // Last of three parts | ||
9781 | lstPart, err := c.CopyObjectPart(context.Background(), bucketName, objectName, destBucketName, destObjectName, uploadID, 3, 0, 1, metadata) | ||
9782 | if err != nil { | ||
9783 | logError(testName, function, args, startTime, "", "CopyObjectPart call failed", err) | ||
9784 | return | ||
9785 | } | ||
9786 | |||
9787 | // Complete the multipart upload | ||
9788 | _, err = c.CompleteMultipartUpload(context.Background(), destBucketName, destObjectName, uploadID, []minio.CompletePart{fstPart, sndPart, lstPart}, minio.PutObjectOptions{}) | ||
9789 | if err != nil { | ||
9790 | logError(testName, function, args, startTime, "", "CompleteMultipartUpload call failed", err) | ||
9791 | return | ||
9792 | } | ||
9793 | |||
9794 | // Stat the object and check its length matches | ||
9795 | objInfo, err := c.StatObject(context.Background(), destBucketName, destObjectName, minio.StatObjectOptions{ServerSideEncryption: dstencryption}) | ||
9796 | if err != nil { | ||
9797 | logError(testName, function, args, startTime, "", "StatObject call failed", err) | ||
9798 | return | ||
9799 | } | ||
9800 | |||
9801 | if objInfo.Size != (5*1024*1024)*2+1 { | ||
9802 | logError(testName, function, args, startTime, "", "Destination object has incorrect size!", err) | ||
9803 | return | ||
9804 | } | ||
9805 | |||
9806 | // Now we read the data back | ||
9807 | getOpts := minio.GetObjectOptions{ServerSideEncryption: dstencryption} | ||
9808 | getOpts.SetRange(0, 5*1024*1024-1) | ||
9809 | r, _, _, err := c.GetObject(context.Background(), destBucketName, destObjectName, getOpts) | ||
9810 | if err != nil { | ||
9811 | logError(testName, function, args, startTime, "", "GetObject call failed", err) | ||
9812 | return | ||
9813 | } | ||
9814 | getBuf := make([]byte, 5*1024*1024) | ||
9815 | _, err = readFull(r, getBuf) | ||
9816 | if err != nil { | ||
9817 | logError(testName, function, args, startTime, "", "Read buffer failed", err) | ||
9818 | return | ||
9819 | } | ||
9820 | if !bytes.Equal(getBuf, buf) { | ||
9821 | logError(testName, function, args, startTime, "", "Got unexpected data in first 5MB", err) | ||
9822 | return | ||
9823 | } | ||
9824 | |||
9825 | getOpts.SetRange(5*1024*1024, 0) | ||
9826 | r, _, _, err = c.GetObject(context.Background(), destBucketName, destObjectName, getOpts) | ||
9827 | if err != nil { | ||
9828 | logError(testName, function, args, startTime, "", "GetObject call failed", err) | ||
9829 | return | ||
9830 | } | ||
9831 | getBuf = make([]byte, 5*1024*1024+1) | ||
9832 | _, err = readFull(r, getBuf) | ||
9833 | if err != nil { | ||
9834 | logError(testName, function, args, startTime, "", "Read buffer failed", err) | ||
9835 | return | ||
9836 | } | ||
9837 | if !bytes.Equal(getBuf[:5*1024*1024], buf) { | ||
9838 | logError(testName, function, args, startTime, "", "Got unexpected data in second 5MB", err) | ||
9839 | return | ||
9840 | } | ||
9841 | if getBuf[5*1024*1024] != buf[0] { | ||
9842 | logError(testName, function, args, startTime, "", "Got unexpected data in last byte of copied object!", err) | ||
9843 | return | ||
9844 | } | ||
9845 | |||
9846 | successLogger(testName, function, args, startTime).Info() | ||
9847 | |||
9848 | // Do not need to remove destBucketName its same as bucketName. | ||
9849 | } | ||
9850 | |||
9851 | // Test Core CopyObjectPart implementation for unencrypted to unencrypted copy | ||
9852 | func testUnencryptedToUnencryptedCopyPart() { | ||
9853 | // initialize logging params | ||
9854 | startTime := time.Now() | ||
9855 | testName := getFuncName() | ||
9856 | function := "CopyObjectPart(destination, source)" | ||
9857 | args := map[string]interface{}{} | ||
9858 | |||
9859 | // Instantiate new minio client object | ||
9860 | client, err := minio.New(os.Getenv(serverEndpoint), | ||
9861 | &minio.Options{ | ||
9862 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
9863 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
9864 | }) | ||
9865 | if err != nil { | ||
9866 | logError(testName, function, args, startTime, "", "MinIO v4 client object creation failed", err) | ||
9867 | return | ||
9868 | } | ||
9869 | |||
9870 | // Instantiate new core client object. | ||
9871 | c := minio.Core{client} | ||
9872 | |||
9873 | // Enable tracing, write to stderr. | ||
9874 | // c.TraceOn(os.Stderr) | ||
9875 | |||
9876 | // Set user agent. | ||
9877 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
9878 | |||
9879 | // Generate a new random bucket name. | ||
9880 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test") | ||
9881 | |||
9882 | // Make a new bucket. | ||
9883 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
9884 | if err != nil { | ||
9885 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
9886 | return | ||
9887 | } | ||
9888 | defer cleanupBucket(bucketName, client) | ||
9889 | // Make a buffer with 5MB of data | ||
9890 | buf := bytes.Repeat([]byte("abcde"), 1024*1024) | ||
9891 | |||
9892 | // Save the data | ||
9893 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
9894 | putmetadata := map[string]string{ | ||
9895 | "Content-Type": "binary/octet-stream", | ||
9896 | } | ||
9897 | opts := minio.PutObjectOptions{ | ||
9898 | UserMetadata: putmetadata, | ||
9899 | } | ||
9900 | uploadInfo, err := c.PutObject(context.Background(), bucketName, objectName, bytes.NewReader(buf), int64(len(buf)), "", "", opts) | ||
9901 | if err != nil { | ||
9902 | logError(testName, function, args, startTime, "", "PutObject call failed", err) | ||
9903 | return | ||
9904 | } | ||
9905 | st, err := c.StatObject(context.Background(), bucketName, objectName, minio.StatObjectOptions{}) | ||
9906 | if err != nil { | ||
9907 | logError(testName, function, args, startTime, "", "StatObject call failed", err) | ||
9908 | return | ||
9909 | } | ||
9910 | |||
9911 | if st.Size != int64(len(buf)) { | ||
9912 | logError(testName, function, args, startTime, "", fmt.Sprintf("Error: number of bytes does not match, want %v, got %v\n", len(buf), st.Size), err) | ||
9913 | return | ||
9914 | } | ||
9915 | |||
9916 | destBucketName := bucketName | ||
9917 | destObjectName := objectName + "-dest" | ||
9918 | |||
9919 | uploadID, err := c.NewMultipartUpload(context.Background(), destBucketName, destObjectName, minio.PutObjectOptions{}) | ||
9920 | if err != nil { | ||
9921 | logError(testName, function, args, startTime, "", "NewMultipartUpload call failed", err) | ||
9922 | return | ||
9923 | } | ||
9924 | |||
9925 | // Content of the destination object will be two copies of | ||
9926 | // `objectName` concatenated, followed by first byte of | ||
9927 | // `objectName`. | ||
9928 | metadata := make(map[string]string) | ||
9929 | header := make(http.Header) | ||
9930 | for k, v := range header { | ||
9931 | metadata[k] = v[0] | ||
9932 | } | ||
9933 | |||
9934 | metadata["x-amz-copy-source-if-match"] = uploadInfo.ETag | ||
9935 | |||
9936 | // First of three parts | ||
9937 | fstPart, err := c.CopyObjectPart(context.Background(), bucketName, objectName, destBucketName, destObjectName, uploadID, 1, 0, -1, metadata) | ||
9938 | if err != nil { | ||
9939 | logError(testName, function, args, startTime, "", "CopyObjectPart call failed", err) | ||
9940 | return | ||
9941 | } | ||
9942 | |||
9943 | // Second of three parts | ||
9944 | sndPart, err := c.CopyObjectPart(context.Background(), bucketName, objectName, destBucketName, destObjectName, uploadID, 2, 0, -1, metadata) | ||
9945 | if err != nil { | ||
9946 | logError(testName, function, args, startTime, "", "CopyObjectPart call failed", err) | ||
9947 | return | ||
9948 | } | ||
9949 | |||
9950 | // Last of three parts | ||
9951 | lstPart, err := c.CopyObjectPart(context.Background(), bucketName, objectName, destBucketName, destObjectName, uploadID, 3, 0, 1, metadata) | ||
9952 | if err != nil { | ||
9953 | logError(testName, function, args, startTime, "", "CopyObjectPart call failed", err) | ||
9954 | return | ||
9955 | } | ||
9956 | |||
9957 | // Complete the multipart upload | ||
9958 | _, err = c.CompleteMultipartUpload(context.Background(), destBucketName, destObjectName, uploadID, []minio.CompletePart{fstPart, sndPart, lstPart}, minio.PutObjectOptions{}) | ||
9959 | if err != nil { | ||
9960 | logError(testName, function, args, startTime, "", "CompleteMultipartUpload call failed", err) | ||
9961 | return | ||
9962 | } | ||
9963 | |||
9964 | // Stat the object and check its length matches | ||
9965 | objInfo, err := c.StatObject(context.Background(), destBucketName, destObjectName, minio.StatObjectOptions{}) | ||
9966 | if err != nil { | ||
9967 | logError(testName, function, args, startTime, "", "StatObject call failed", err) | ||
9968 | return | ||
9969 | } | ||
9970 | |||
9971 | if objInfo.Size != (5*1024*1024)*2+1 { | ||
9972 | logError(testName, function, args, startTime, "", "Destination object has incorrect size!", err) | ||
9973 | return | ||
9974 | } | ||
9975 | |||
9976 | // Now we read the data back | ||
9977 | getOpts := minio.GetObjectOptions{} | ||
9978 | getOpts.SetRange(0, 5*1024*1024-1) | ||
9979 | r, _, _, err := c.GetObject(context.Background(), destBucketName, destObjectName, getOpts) | ||
9980 | if err != nil { | ||
9981 | logError(testName, function, args, startTime, "", "GetObject call failed", err) | ||
9982 | return | ||
9983 | } | ||
9984 | getBuf := make([]byte, 5*1024*1024) | ||
9985 | _, err = readFull(r, getBuf) | ||
9986 | if err != nil { | ||
9987 | logError(testName, function, args, startTime, "", "Read buffer failed", err) | ||
9988 | return | ||
9989 | } | ||
9990 | if !bytes.Equal(getBuf, buf) { | ||
9991 | logError(testName, function, args, startTime, "", "Got unexpected data in first 5MB", err) | ||
9992 | return | ||
9993 | } | ||
9994 | |||
9995 | getOpts.SetRange(5*1024*1024, 0) | ||
9996 | r, _, _, err = c.GetObject(context.Background(), destBucketName, destObjectName, getOpts) | ||
9997 | if err != nil { | ||
9998 | logError(testName, function, args, startTime, "", "GetObject call failed", err) | ||
9999 | return | ||
10000 | } | ||
10001 | getBuf = make([]byte, 5*1024*1024+1) | ||
10002 | _, err = readFull(r, getBuf) | ||
10003 | if err != nil { | ||
10004 | logError(testName, function, args, startTime, "", "Read buffer failed", err) | ||
10005 | return | ||
10006 | } | ||
10007 | if !bytes.Equal(getBuf[:5*1024*1024], buf) { | ||
10008 | logError(testName, function, args, startTime, "", "Got unexpected data in second 5MB", err) | ||
10009 | return | ||
10010 | } | ||
10011 | if getBuf[5*1024*1024] != buf[0] { | ||
10012 | logError(testName, function, args, startTime, "", "Got unexpected data in last byte of copied object!", err) | ||
10013 | return | ||
10014 | } | ||
10015 | |||
10016 | successLogger(testName, function, args, startTime).Info() | ||
10017 | |||
10018 | // Do not need to remove destBucketName its same as bucketName. | ||
10019 | } | ||
10020 | |||
10021 | // Test Core CopyObjectPart implementation for unencrypted to SSE-S3 encrypted copy | ||
10022 | func testUnencryptedToSSES3CopyObjectPart() { | ||
10023 | // initialize logging params | ||
10024 | startTime := time.Now() | ||
10025 | testName := getFuncName() | ||
10026 | function := "CopyObjectPart(destination, source)" | ||
10027 | args := map[string]interface{}{} | ||
10028 | |||
10029 | // Instantiate new minio client object | ||
10030 | client, err := minio.New(os.Getenv(serverEndpoint), | ||
10031 | &minio.Options{ | ||
10032 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
10033 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
10034 | }) | ||
10035 | if err != nil { | ||
10036 | logError(testName, function, args, startTime, "", "MinIO v4 client object creation failed", err) | ||
10037 | return | ||
10038 | } | ||
10039 | |||
10040 | // Instantiate new core client object. | ||
10041 | c := minio.Core{client} | ||
10042 | |||
10043 | // Enable tracing, write to stderr. | ||
10044 | // c.TraceOn(os.Stderr) | ||
10045 | |||
10046 | // Set user agent. | ||
10047 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
10048 | |||
10049 | // Generate a new random bucket name. | ||
10050 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test") | ||
10051 | |||
10052 | // Make a new bucket. | ||
10053 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
10054 | if err != nil { | ||
10055 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
10056 | return | ||
10057 | } | ||
10058 | defer cleanupBucket(bucketName, client) | ||
10059 | // Make a buffer with 5MB of data | ||
10060 | buf := bytes.Repeat([]byte("abcde"), 1024*1024) | ||
10061 | |||
10062 | // Save the data | ||
10063 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
10064 | opts := minio.PutObjectOptions{ | ||
10065 | UserMetadata: map[string]string{ | ||
10066 | "Content-Type": "binary/octet-stream", | ||
10067 | }, | ||
10068 | } | ||
10069 | uploadInfo, err := c.PutObject(context.Background(), bucketName, objectName, bytes.NewReader(buf), int64(len(buf)), "", "", opts) | ||
10070 | if err != nil { | ||
10071 | logError(testName, function, args, startTime, "", "PutObject call failed", err) | ||
10072 | return | ||
10073 | } | ||
10074 | st, err := c.StatObject(context.Background(), bucketName, objectName, minio.StatObjectOptions{}) | ||
10075 | if err != nil { | ||
10076 | logError(testName, function, args, startTime, "", "StatObject call failed", err) | ||
10077 | return | ||
10078 | } | ||
10079 | |||
10080 | if st.Size != int64(len(buf)) { | ||
10081 | logError(testName, function, args, startTime, "", fmt.Sprintf("Error: number of bytes does not match, want %v, got %v\n", len(buf), st.Size), err) | ||
10082 | return | ||
10083 | } | ||
10084 | |||
10085 | destBucketName := bucketName | ||
10086 | destObjectName := objectName + "-dest" | ||
10087 | dstencryption := encrypt.NewSSE() | ||
10088 | |||
10089 | uploadID, err := c.NewMultipartUpload(context.Background(), destBucketName, destObjectName, minio.PutObjectOptions{ServerSideEncryption: dstencryption}) | ||
10090 | if err != nil { | ||
10091 | logError(testName, function, args, startTime, "", "NewMultipartUpload call failed", err) | ||
10092 | return | ||
10093 | } | ||
10094 | |||
10095 | // Content of the destination object will be two copies of | ||
10096 | // `objectName` concatenated, followed by first byte of | ||
10097 | // `objectName`. | ||
10098 | metadata := make(map[string]string) | ||
10099 | header := make(http.Header) | ||
10100 | dstencryption.Marshal(header) | ||
10101 | |||
10102 | for k, v := range header { | ||
10103 | metadata[k] = v[0] | ||
10104 | } | ||
10105 | |||
10106 | metadata["x-amz-copy-source-if-match"] = uploadInfo.ETag | ||
10107 | |||
10108 | // First of three parts | ||
10109 | fstPart, err := c.CopyObjectPart(context.Background(), bucketName, objectName, destBucketName, destObjectName, uploadID, 1, 0, -1, metadata) | ||
10110 | if err != nil { | ||
10111 | logError(testName, function, args, startTime, "", "CopyObjectPart call failed", err) | ||
10112 | return | ||
10113 | } | ||
10114 | |||
10115 | // Second of three parts | ||
10116 | sndPart, err := c.CopyObjectPart(context.Background(), bucketName, objectName, destBucketName, destObjectName, uploadID, 2, 0, -1, metadata) | ||
10117 | if err != nil { | ||
10118 | logError(testName, function, args, startTime, "", "CopyObjectPart call failed", err) | ||
10119 | return | ||
10120 | } | ||
10121 | |||
10122 | // Last of three parts | ||
10123 | lstPart, err := c.CopyObjectPart(context.Background(), bucketName, objectName, destBucketName, destObjectName, uploadID, 3, 0, 1, metadata) | ||
10124 | if err != nil { | ||
10125 | logError(testName, function, args, startTime, "", "CopyObjectPart call failed", err) | ||
10126 | return | ||
10127 | } | ||
10128 | |||
10129 | // Complete the multipart upload | ||
10130 | _, err = c.CompleteMultipartUpload(context.Background(), destBucketName, destObjectName, uploadID, []minio.CompletePart{fstPart, sndPart, lstPart}, minio.PutObjectOptions{}) | ||
10131 | if err != nil { | ||
10132 | logError(testName, function, args, startTime, "", "CompleteMultipartUpload call failed", err) | ||
10133 | return | ||
10134 | } | ||
10135 | |||
10136 | // Stat the object and check its length matches | ||
10137 | objInfo, err := c.StatObject(context.Background(), destBucketName, destObjectName, minio.StatObjectOptions{}) | ||
10138 | if err != nil { | ||
10139 | logError(testName, function, args, startTime, "", "StatObject call failed", err) | ||
10140 | return | ||
10141 | } | ||
10142 | |||
10143 | if objInfo.Size != (5*1024*1024)*2+1 { | ||
10144 | logError(testName, function, args, startTime, "", "Destination object has incorrect size!", err) | ||
10145 | return | ||
10146 | } | ||
10147 | |||
10148 | // Now we read the data back | ||
10149 | getOpts := minio.GetObjectOptions{} | ||
10150 | getOpts.SetRange(0, 5*1024*1024-1) | ||
10151 | r, _, _, err := c.GetObject(context.Background(), destBucketName, destObjectName, getOpts) | ||
10152 | if err != nil { | ||
10153 | logError(testName, function, args, startTime, "", "GetObject call failed", err) | ||
10154 | return | ||
10155 | } | ||
10156 | getBuf := make([]byte, 5*1024*1024) | ||
10157 | _, err = readFull(r, getBuf) | ||
10158 | if err != nil { | ||
10159 | logError(testName, function, args, startTime, "", "Read buffer failed", err) | ||
10160 | return | ||
10161 | } | ||
10162 | if !bytes.Equal(getBuf, buf) { | ||
10163 | logError(testName, function, args, startTime, "", "Got unexpected data in first 5MB", err) | ||
10164 | return | ||
10165 | } | ||
10166 | |||
10167 | getOpts.SetRange(5*1024*1024, 0) | ||
10168 | r, _, _, err = c.GetObject(context.Background(), destBucketName, destObjectName, getOpts) | ||
10169 | if err != nil { | ||
10170 | logError(testName, function, args, startTime, "", "GetObject call failed", err) | ||
10171 | return | ||
10172 | } | ||
10173 | getBuf = make([]byte, 5*1024*1024+1) | ||
10174 | _, err = readFull(r, getBuf) | ||
10175 | if err != nil { | ||
10176 | logError(testName, function, args, startTime, "", "Read buffer failed", err) | ||
10177 | return | ||
10178 | } | ||
10179 | if !bytes.Equal(getBuf[:5*1024*1024], buf) { | ||
10180 | logError(testName, function, args, startTime, "", "Got unexpected data in second 5MB", err) | ||
10181 | return | ||
10182 | } | ||
10183 | if getBuf[5*1024*1024] != buf[0] { | ||
10184 | logError(testName, function, args, startTime, "", "Got unexpected data in last byte of copied object!", err) | ||
10185 | return | ||
10186 | } | ||
10187 | |||
10188 | successLogger(testName, function, args, startTime).Info() | ||
10189 | |||
10190 | // Do not need to remove destBucketName its same as bucketName. | ||
10191 | } | ||
10192 | |||
10193 | // Test Core CopyObjectPart implementation for SSE-S3 to SSEC encryption copy part | ||
10194 | func testSSES3EncryptedToSSECCopyObjectPart() { | ||
10195 | // initialize logging params | ||
10196 | startTime := time.Now() | ||
10197 | testName := getFuncName() | ||
10198 | function := "CopyObjectPart(destination, source)" | ||
10199 | args := map[string]interface{}{} | ||
10200 | |||
10201 | // Instantiate new minio client object | ||
10202 | client, err := minio.New(os.Getenv(serverEndpoint), | ||
10203 | &minio.Options{ | ||
10204 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
10205 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
10206 | }) | ||
10207 | if err != nil { | ||
10208 | logError(testName, function, args, startTime, "", "MinIO v4 client object creation failed", err) | ||
10209 | return | ||
10210 | } | ||
10211 | |||
10212 | // Instantiate new core client object. | ||
10213 | c := minio.Core{client} | ||
10214 | |||
10215 | // Enable tracing, write to stderr. | ||
10216 | // c.TraceOn(os.Stderr) | ||
10217 | |||
10218 | // Set user agent. | ||
10219 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
10220 | |||
10221 | // Generate a new random bucket name. | ||
10222 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test") | ||
10223 | |||
10224 | // Make a new bucket. | ||
10225 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
10226 | if err != nil { | ||
10227 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
10228 | return | ||
10229 | } | ||
10230 | defer cleanupBucket(bucketName, client) | ||
10231 | // Make a buffer with 5MB of data | ||
10232 | buf := bytes.Repeat([]byte("abcde"), 1024*1024) | ||
10233 | |||
10234 | // Save the data | ||
10235 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
10236 | password := "correct horse battery staple" | ||
10237 | srcEncryption := encrypt.NewSSE() | ||
10238 | opts := minio.PutObjectOptions{ | ||
10239 | UserMetadata: map[string]string{ | ||
10240 | "Content-Type": "binary/octet-stream", | ||
10241 | }, | ||
10242 | ServerSideEncryption: srcEncryption, | ||
10243 | } | ||
10244 | uploadInfo, err := c.PutObject(context.Background(), bucketName, objectName, bytes.NewReader(buf), int64(len(buf)), "", "", opts) | ||
10245 | if err != nil { | ||
10246 | logError(testName, function, args, startTime, "", "PutObject call failed", err) | ||
10247 | return | ||
10248 | } | ||
10249 | |||
10250 | st, err := c.StatObject(context.Background(), bucketName, objectName, minio.StatObjectOptions{ServerSideEncryption: srcEncryption}) | ||
10251 | if err != nil { | ||
10252 | logError(testName, function, args, startTime, "", "StatObject call failed", err) | ||
10253 | return | ||
10254 | } | ||
10255 | |||
10256 | if st.Size != int64(len(buf)) { | ||
10257 | logError(testName, function, args, startTime, "", fmt.Sprintf("Error: number of bytes does not match, want %v, got %v\n", len(buf), st.Size), err) | ||
10258 | return | ||
10259 | } | ||
10260 | |||
10261 | destBucketName := bucketName | ||
10262 | destObjectName := objectName + "-dest" | ||
10263 | dstencryption := encrypt.DefaultPBKDF([]byte(password), []byte(destBucketName+destObjectName)) | ||
10264 | |||
10265 | uploadID, err := c.NewMultipartUpload(context.Background(), destBucketName, destObjectName, minio.PutObjectOptions{ServerSideEncryption: dstencryption}) | ||
10266 | if err != nil { | ||
10267 | logError(testName, function, args, startTime, "", "NewMultipartUpload call failed", err) | ||
10268 | return | ||
10269 | } | ||
10270 | |||
10271 | // Content of the destination object will be two copies of | ||
10272 | // `objectName` concatenated, followed by first byte of | ||
10273 | // `objectName`. | ||
10274 | metadata := make(map[string]string) | ||
10275 | header := make(http.Header) | ||
10276 | dstencryption.Marshal(header) | ||
10277 | for k, v := range header { | ||
10278 | metadata[k] = v[0] | ||
10279 | } | ||
10280 | |||
10281 | metadata["x-amz-copy-source-if-match"] = uploadInfo.ETag | ||
10282 | |||
10283 | // First of three parts | ||
10284 | fstPart, err := c.CopyObjectPart(context.Background(), bucketName, objectName, destBucketName, destObjectName, uploadID, 1, 0, -1, metadata) | ||
10285 | if err != nil { | ||
10286 | logError(testName, function, args, startTime, "", "CopyObjectPart call failed", err) | ||
10287 | return | ||
10288 | } | ||
10289 | |||
10290 | // Second of three parts | ||
10291 | sndPart, err := c.CopyObjectPart(context.Background(), bucketName, objectName, destBucketName, destObjectName, uploadID, 2, 0, -1, metadata) | ||
10292 | if err != nil { | ||
10293 | logError(testName, function, args, startTime, "", "CopyObjectPart call failed", err) | ||
10294 | return | ||
10295 | } | ||
10296 | |||
10297 | // Last of three parts | ||
10298 | lstPart, err := c.CopyObjectPart(context.Background(), bucketName, objectName, destBucketName, destObjectName, uploadID, 3, 0, 1, metadata) | ||
10299 | if err != nil { | ||
10300 | logError(testName, function, args, startTime, "", "CopyObjectPart call failed", err) | ||
10301 | return | ||
10302 | } | ||
10303 | |||
10304 | // Complete the multipart upload | ||
10305 | _, err = c.CompleteMultipartUpload(context.Background(), destBucketName, destObjectName, uploadID, []minio.CompletePart{fstPart, sndPart, lstPart}, minio.PutObjectOptions{}) | ||
10306 | if err != nil { | ||
10307 | logError(testName, function, args, startTime, "", "CompleteMultipartUpload call failed", err) | ||
10308 | return | ||
10309 | } | ||
10310 | |||
10311 | // Stat the object and check its length matches | ||
10312 | objInfo, err := c.StatObject(context.Background(), destBucketName, destObjectName, minio.StatObjectOptions{ServerSideEncryption: dstencryption}) | ||
10313 | if err != nil { | ||
10314 | logError(testName, function, args, startTime, "", "StatObject call failed", err) | ||
10315 | return | ||
10316 | } | ||
10317 | |||
10318 | if objInfo.Size != (5*1024*1024)*2+1 { | ||
10319 | logError(testName, function, args, startTime, "", "Destination object has incorrect size!", err) | ||
10320 | return | ||
10321 | } | ||
10322 | |||
10323 | // Now we read the data back | ||
10324 | getOpts := minio.GetObjectOptions{ServerSideEncryption: dstencryption} | ||
10325 | getOpts.SetRange(0, 5*1024*1024-1) | ||
10326 | r, _, _, err := c.GetObject(context.Background(), destBucketName, destObjectName, getOpts) | ||
10327 | if err != nil { | ||
10328 | logError(testName, function, args, startTime, "", "GetObject call failed", err) | ||
10329 | return | ||
10330 | } | ||
10331 | getBuf := make([]byte, 5*1024*1024) | ||
10332 | _, err = readFull(r, getBuf) | ||
10333 | if err != nil { | ||
10334 | logError(testName, function, args, startTime, "", "Read buffer failed", err) | ||
10335 | return | ||
10336 | } | ||
10337 | if !bytes.Equal(getBuf, buf) { | ||
10338 | logError(testName, function, args, startTime, "", "Got unexpected data in first 5MB", err) | ||
10339 | return | ||
10340 | } | ||
10341 | |||
10342 | getOpts.SetRange(5*1024*1024, 0) | ||
10343 | r, _, _, err = c.GetObject(context.Background(), destBucketName, destObjectName, getOpts) | ||
10344 | if err != nil { | ||
10345 | logError(testName, function, args, startTime, "", "GetObject call failed", err) | ||
10346 | return | ||
10347 | } | ||
10348 | getBuf = make([]byte, 5*1024*1024+1) | ||
10349 | _, err = readFull(r, getBuf) | ||
10350 | if err != nil { | ||
10351 | logError(testName, function, args, startTime, "", "Read buffer failed", err) | ||
10352 | return | ||
10353 | } | ||
10354 | if !bytes.Equal(getBuf[:5*1024*1024], buf) { | ||
10355 | logError(testName, function, args, startTime, "", "Got unexpected data in second 5MB", err) | ||
10356 | return | ||
10357 | } | ||
10358 | if getBuf[5*1024*1024] != buf[0] { | ||
10359 | logError(testName, function, args, startTime, "", "Got unexpected data in last byte of copied object!", err) | ||
10360 | return | ||
10361 | } | ||
10362 | |||
10363 | successLogger(testName, function, args, startTime).Info() | ||
10364 | |||
10365 | // Do not need to remove destBucketName its same as bucketName. | ||
10366 | } | ||
10367 | |||
10368 | // Test Core CopyObjectPart implementation for unencrypted to unencrypted copy | ||
10369 | func testSSES3EncryptedToUnencryptedCopyPart() { | ||
10370 | // initialize logging params | ||
10371 | startTime := time.Now() | ||
10372 | testName := getFuncName() | ||
10373 | function := "CopyObjectPart(destination, source)" | ||
10374 | args := map[string]interface{}{} | ||
10375 | |||
10376 | // Instantiate new minio client object | ||
10377 | client, err := minio.New(os.Getenv(serverEndpoint), | ||
10378 | &minio.Options{ | ||
10379 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
10380 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
10381 | }) | ||
10382 | if err != nil { | ||
10383 | logError(testName, function, args, startTime, "", "MinIO v4 client object creation failed", err) | ||
10384 | return | ||
10385 | } | ||
10386 | |||
10387 | // Instantiate new core client object. | ||
10388 | c := minio.Core{client} | ||
10389 | |||
10390 | // Enable tracing, write to stderr. | ||
10391 | // c.TraceOn(os.Stderr) | ||
10392 | |||
10393 | // Set user agent. | ||
10394 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
10395 | |||
10396 | // Generate a new random bucket name. | ||
10397 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test") | ||
10398 | |||
10399 | // Make a new bucket. | ||
10400 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
10401 | if err != nil { | ||
10402 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
10403 | return | ||
10404 | } | ||
10405 | defer cleanupBucket(bucketName, client) | ||
10406 | // Make a buffer with 5MB of data | ||
10407 | buf := bytes.Repeat([]byte("abcde"), 1024*1024) | ||
10408 | |||
10409 | // Save the data | ||
10410 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
10411 | srcEncryption := encrypt.NewSSE() | ||
10412 | opts := minio.PutObjectOptions{ | ||
10413 | UserMetadata: map[string]string{ | ||
10414 | "Content-Type": "binary/octet-stream", | ||
10415 | }, | ||
10416 | ServerSideEncryption: srcEncryption, | ||
10417 | } | ||
10418 | uploadInfo, err := c.PutObject(context.Background(), bucketName, objectName, bytes.NewReader(buf), int64(len(buf)), "", "", opts) | ||
10419 | if err != nil { | ||
10420 | logError(testName, function, args, startTime, "", "PutObject call failed", err) | ||
10421 | return | ||
10422 | } | ||
10423 | st, err := c.StatObject(context.Background(), bucketName, objectName, minio.StatObjectOptions{ServerSideEncryption: srcEncryption}) | ||
10424 | if err != nil { | ||
10425 | logError(testName, function, args, startTime, "", "StatObject call failed", err) | ||
10426 | return | ||
10427 | } | ||
10428 | |||
10429 | if st.Size != int64(len(buf)) { | ||
10430 | logError(testName, function, args, startTime, "", fmt.Sprintf("Error: number of bytes does not match, want %v, got %v\n", len(buf), st.Size), err) | ||
10431 | return | ||
10432 | } | ||
10433 | |||
10434 | destBucketName := bucketName | ||
10435 | destObjectName := objectName + "-dest" | ||
10436 | |||
10437 | uploadID, err := c.NewMultipartUpload(context.Background(), destBucketName, destObjectName, minio.PutObjectOptions{}) | ||
10438 | if err != nil { | ||
10439 | logError(testName, function, args, startTime, "", "NewMultipartUpload call failed", err) | ||
10440 | return | ||
10441 | } | ||
10442 | |||
10443 | // Content of the destination object will be two copies of | ||
10444 | // `objectName` concatenated, followed by first byte of | ||
10445 | // `objectName`. | ||
10446 | metadata := make(map[string]string) | ||
10447 | header := make(http.Header) | ||
10448 | for k, v := range header { | ||
10449 | metadata[k] = v[0] | ||
10450 | } | ||
10451 | |||
10452 | metadata["x-amz-copy-source-if-match"] = uploadInfo.ETag | ||
10453 | |||
10454 | // First of three parts | ||
10455 | fstPart, err := c.CopyObjectPart(context.Background(), bucketName, objectName, destBucketName, destObjectName, uploadID, 1, 0, -1, metadata) | ||
10456 | if err != nil { | ||
10457 | logError(testName, function, args, startTime, "", "CopyObjectPart call failed", err) | ||
10458 | return | ||
10459 | } | ||
10460 | |||
10461 | // Second of three parts | ||
10462 | sndPart, err := c.CopyObjectPart(context.Background(), bucketName, objectName, destBucketName, destObjectName, uploadID, 2, 0, -1, metadata) | ||
10463 | if err != nil { | ||
10464 | logError(testName, function, args, startTime, "", "CopyObjectPart call failed", err) | ||
10465 | return | ||
10466 | } | ||
10467 | |||
10468 | // Last of three parts | ||
10469 | lstPart, err := c.CopyObjectPart(context.Background(), bucketName, objectName, destBucketName, destObjectName, uploadID, 3, 0, 1, metadata) | ||
10470 | if err != nil { | ||
10471 | logError(testName, function, args, startTime, "", "CopyObjectPart call failed", err) | ||
10472 | return | ||
10473 | } | ||
10474 | |||
10475 | // Complete the multipart upload | ||
10476 | _, err = c.CompleteMultipartUpload(context.Background(), destBucketName, destObjectName, uploadID, []minio.CompletePart{fstPart, sndPart, lstPart}, minio.PutObjectOptions{}) | ||
10477 | if err != nil { | ||
10478 | logError(testName, function, args, startTime, "", "CompleteMultipartUpload call failed", err) | ||
10479 | return | ||
10480 | } | ||
10481 | |||
10482 | // Stat the object and check its length matches | ||
10483 | objInfo, err := c.StatObject(context.Background(), destBucketName, destObjectName, minio.StatObjectOptions{}) | ||
10484 | if err != nil { | ||
10485 | logError(testName, function, args, startTime, "", "StatObject call failed", err) | ||
10486 | return | ||
10487 | } | ||
10488 | |||
10489 | if objInfo.Size != (5*1024*1024)*2+1 { | ||
10490 | logError(testName, function, args, startTime, "", "Destination object has incorrect size!", err) | ||
10491 | return | ||
10492 | } | ||
10493 | |||
10494 | // Now we read the data back | ||
10495 | getOpts := minio.GetObjectOptions{} | ||
10496 | getOpts.SetRange(0, 5*1024*1024-1) | ||
10497 | r, _, _, err := c.GetObject(context.Background(), destBucketName, destObjectName, getOpts) | ||
10498 | if err != nil { | ||
10499 | logError(testName, function, args, startTime, "", "GetObject call failed", err) | ||
10500 | return | ||
10501 | } | ||
10502 | getBuf := make([]byte, 5*1024*1024) | ||
10503 | _, err = readFull(r, getBuf) | ||
10504 | if err != nil { | ||
10505 | logError(testName, function, args, startTime, "", "Read buffer failed", err) | ||
10506 | return | ||
10507 | } | ||
10508 | if !bytes.Equal(getBuf, buf) { | ||
10509 | logError(testName, function, args, startTime, "", "Got unexpected data in first 5MB", err) | ||
10510 | return | ||
10511 | } | ||
10512 | |||
10513 | getOpts.SetRange(5*1024*1024, 0) | ||
10514 | r, _, _, err = c.GetObject(context.Background(), destBucketName, destObjectName, getOpts) | ||
10515 | if err != nil { | ||
10516 | logError(testName, function, args, startTime, "", "GetObject call failed", err) | ||
10517 | return | ||
10518 | } | ||
10519 | getBuf = make([]byte, 5*1024*1024+1) | ||
10520 | _, err = readFull(r, getBuf) | ||
10521 | if err != nil { | ||
10522 | logError(testName, function, args, startTime, "", "Read buffer failed", err) | ||
10523 | return | ||
10524 | } | ||
10525 | if !bytes.Equal(getBuf[:5*1024*1024], buf) { | ||
10526 | logError(testName, function, args, startTime, "", "Got unexpected data in second 5MB", err) | ||
10527 | return | ||
10528 | } | ||
10529 | if getBuf[5*1024*1024] != buf[0] { | ||
10530 | logError(testName, function, args, startTime, "", "Got unexpected data in last byte of copied object!", err) | ||
10531 | return | ||
10532 | } | ||
10533 | |||
10534 | successLogger(testName, function, args, startTime).Info() | ||
10535 | |||
10536 | // Do not need to remove destBucketName its same as bucketName. | ||
10537 | } | ||
10538 | |||
10539 | // Test Core CopyObjectPart implementation for unencrypted to SSE-S3 encrypted copy | ||
10540 | func testSSES3EncryptedToSSES3CopyObjectPart() { | ||
10541 | // initialize logging params | ||
10542 | startTime := time.Now() | ||
10543 | testName := getFuncName() | ||
10544 | function := "CopyObjectPart(destination, source)" | ||
10545 | args := map[string]interface{}{} | ||
10546 | |||
10547 | // Instantiate new minio client object | ||
10548 | client, err := minio.New(os.Getenv(serverEndpoint), | ||
10549 | &minio.Options{ | ||
10550 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
10551 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
10552 | }) | ||
10553 | if err != nil { | ||
10554 | logError(testName, function, args, startTime, "", "MinIO v4 client object creation failed", err) | ||
10555 | return | ||
10556 | } | ||
10557 | |||
10558 | // Instantiate new core client object. | ||
10559 | c := minio.Core{client} | ||
10560 | |||
10561 | // Enable tracing, write to stderr. | ||
10562 | // c.TraceOn(os.Stderr) | ||
10563 | |||
10564 | // Set user agent. | ||
10565 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
10566 | |||
10567 | // Generate a new random bucket name. | ||
10568 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test") | ||
10569 | |||
10570 | // Make a new bucket. | ||
10571 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
10572 | if err != nil { | ||
10573 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
10574 | return | ||
10575 | } | ||
10576 | defer cleanupBucket(bucketName, client) | ||
10577 | // Make a buffer with 5MB of data | ||
10578 | buf := bytes.Repeat([]byte("abcde"), 1024*1024) | ||
10579 | |||
10580 | // Save the data | ||
10581 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
10582 | srcEncryption := encrypt.NewSSE() | ||
10583 | opts := minio.PutObjectOptions{ | ||
10584 | UserMetadata: map[string]string{ | ||
10585 | "Content-Type": "binary/octet-stream", | ||
10586 | }, | ||
10587 | ServerSideEncryption: srcEncryption, | ||
10588 | } | ||
10589 | |||
10590 | uploadInfo, err := c.PutObject(context.Background(), bucketName, objectName, bytes.NewReader(buf), int64(len(buf)), "", "", opts) | ||
10591 | if err != nil { | ||
10592 | logError(testName, function, args, startTime, "", "PutObject call failed", err) | ||
10593 | return | ||
10594 | } | ||
10595 | st, err := c.StatObject(context.Background(), bucketName, objectName, minio.StatObjectOptions{ServerSideEncryption: srcEncryption}) | ||
10596 | if err != nil { | ||
10597 | logError(testName, function, args, startTime, "", "StatObject call failed", err) | ||
10598 | return | ||
10599 | } | ||
10600 | if st.Size != int64(len(buf)) { | ||
10601 | logError(testName, function, args, startTime, "", fmt.Sprintf("Error: number of bytes does not match, want %v, got %v\n", len(buf), st.Size), err) | ||
10602 | return | ||
10603 | } | ||
10604 | |||
10605 | destBucketName := bucketName | ||
10606 | destObjectName := objectName + "-dest" | ||
10607 | dstencryption := encrypt.NewSSE() | ||
10608 | |||
10609 | uploadID, err := c.NewMultipartUpload(context.Background(), destBucketName, destObjectName, minio.PutObjectOptions{ServerSideEncryption: dstencryption}) | ||
10610 | if err != nil { | ||
10611 | logError(testName, function, args, startTime, "", "NewMultipartUpload call failed", err) | ||
10612 | return | ||
10613 | } | ||
10614 | |||
10615 | // Content of the destination object will be two copies of | ||
10616 | // `objectName` concatenated, followed by first byte of | ||
10617 | // `objectName`. | ||
10618 | metadata := make(map[string]string) | ||
10619 | header := make(http.Header) | ||
10620 | dstencryption.Marshal(header) | ||
10621 | |||
10622 | for k, v := range header { | ||
10623 | metadata[k] = v[0] | ||
10624 | } | ||
10625 | |||
10626 | metadata["x-amz-copy-source-if-match"] = uploadInfo.ETag | ||
10627 | |||
10628 | // First of three parts | ||
10629 | fstPart, err := c.CopyObjectPart(context.Background(), bucketName, objectName, destBucketName, destObjectName, uploadID, 1, 0, -1, metadata) | ||
10630 | if err != nil { | ||
10631 | logError(testName, function, args, startTime, "", "CopyObjectPart call failed", err) | ||
10632 | return | ||
10633 | } | ||
10634 | |||
10635 | // Second of three parts | ||
10636 | sndPart, err := c.CopyObjectPart(context.Background(), bucketName, objectName, destBucketName, destObjectName, uploadID, 2, 0, -1, metadata) | ||
10637 | if err != nil { | ||
10638 | logError(testName, function, args, startTime, "", "CopyObjectPart call failed", err) | ||
10639 | return | ||
10640 | } | ||
10641 | |||
10642 | // Last of three parts | ||
10643 | lstPart, err := c.CopyObjectPart(context.Background(), bucketName, objectName, destBucketName, destObjectName, uploadID, 3, 0, 1, metadata) | ||
10644 | if err != nil { | ||
10645 | logError(testName, function, args, startTime, "", "CopyObjectPart call failed", err) | ||
10646 | return | ||
10647 | } | ||
10648 | |||
10649 | // Complete the multipart upload | ||
10650 | _, err = c.CompleteMultipartUpload(context.Background(), destBucketName, destObjectName, uploadID, []minio.CompletePart{fstPart, sndPart, lstPart}, minio.PutObjectOptions{}) | ||
10651 | if err != nil { | ||
10652 | logError(testName, function, args, startTime, "", "CompleteMultipartUpload call failed", err) | ||
10653 | return | ||
10654 | } | ||
10655 | |||
10656 | // Stat the object and check its length matches | ||
10657 | objInfo, err := c.StatObject(context.Background(), destBucketName, destObjectName, minio.StatObjectOptions{}) | ||
10658 | if err != nil { | ||
10659 | logError(testName, function, args, startTime, "", "StatObject call failed", err) | ||
10660 | return | ||
10661 | } | ||
10662 | |||
10663 | if objInfo.Size != (5*1024*1024)*2+1 { | ||
10664 | logError(testName, function, args, startTime, "", "Destination object has incorrect size!", err) | ||
10665 | return | ||
10666 | } | ||
10667 | |||
10668 | // Now we read the data back | ||
10669 | getOpts := minio.GetObjectOptions{} | ||
10670 | getOpts.SetRange(0, 5*1024*1024-1) | ||
10671 | r, _, _, err := c.GetObject(context.Background(), destBucketName, destObjectName, getOpts) | ||
10672 | if err != nil { | ||
10673 | logError(testName, function, args, startTime, "", "GetObject call failed", err) | ||
10674 | return | ||
10675 | } | ||
10676 | getBuf := make([]byte, 5*1024*1024) | ||
10677 | _, err = readFull(r, getBuf) | ||
10678 | if err != nil { | ||
10679 | logError(testName, function, args, startTime, "", "Read buffer failed", err) | ||
10680 | return | ||
10681 | } | ||
10682 | if !bytes.Equal(getBuf, buf) { | ||
10683 | logError(testName, function, args, startTime, "", "Got unexpected data in first 5MB", err) | ||
10684 | return | ||
10685 | } | ||
10686 | |||
10687 | getOpts.SetRange(5*1024*1024, 0) | ||
10688 | r, _, _, err = c.GetObject(context.Background(), destBucketName, destObjectName, getOpts) | ||
10689 | if err != nil { | ||
10690 | logError(testName, function, args, startTime, "", "GetObject call failed", err) | ||
10691 | return | ||
10692 | } | ||
10693 | getBuf = make([]byte, 5*1024*1024+1) | ||
10694 | _, err = readFull(r, getBuf) | ||
10695 | if err != nil { | ||
10696 | logError(testName, function, args, startTime, "", "Read buffer failed", err) | ||
10697 | return | ||
10698 | } | ||
10699 | if !bytes.Equal(getBuf[:5*1024*1024], buf) { | ||
10700 | logError(testName, function, args, startTime, "", "Got unexpected data in second 5MB", err) | ||
10701 | return | ||
10702 | } | ||
10703 | if getBuf[5*1024*1024] != buf[0] { | ||
10704 | logError(testName, function, args, startTime, "", "Got unexpected data in last byte of copied object!", err) | ||
10705 | return | ||
10706 | } | ||
10707 | |||
10708 | successLogger(testName, function, args, startTime).Info() | ||
10709 | |||
10710 | // Do not need to remove destBucketName its same as bucketName. | ||
10711 | } | ||
10712 | |||
10713 | func testUserMetadataCopying() { | ||
10714 | // initialize logging params | ||
10715 | startTime := time.Now() | ||
10716 | testName := getFuncName() | ||
10717 | function := "CopyObject(destination, source)" | ||
10718 | args := map[string]interface{}{} | ||
10719 | |||
10720 | // Instantiate new minio client object | ||
10721 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
10722 | &minio.Options{ | ||
10723 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
10724 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
10725 | }) | ||
10726 | if err != nil { | ||
10727 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
10728 | return | ||
10729 | } | ||
10730 | |||
10731 | // c.TraceOn(os.Stderr) | ||
10732 | testUserMetadataCopyingWrapper(c) | ||
10733 | } | ||
10734 | |||
10735 | func testUserMetadataCopyingWrapper(c *minio.Client) { | ||
10736 | // initialize logging params | ||
10737 | startTime := time.Now() | ||
10738 | testName := getFuncName() | ||
10739 | function := "CopyObject(destination, source)" | ||
10740 | args := map[string]interface{}{} | ||
10741 | |||
10742 | // Generate a new random bucket name. | ||
10743 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
10744 | // Make a new bucket in 'us-east-1' (source bucket). | ||
10745 | err := c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
10746 | if err != nil { | ||
10747 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
10748 | return | ||
10749 | } | ||
10750 | |||
10751 | defer cleanupBucket(bucketName, c) | ||
10752 | |||
10753 | fetchMeta := func(object string) (h http.Header) { | ||
10754 | objInfo, err := c.StatObject(context.Background(), bucketName, object, minio.StatObjectOptions{}) | ||
10755 | if err != nil { | ||
10756 | logError(testName, function, args, startTime, "", "Stat failed", err) | ||
10757 | return | ||
10758 | } | ||
10759 | h = make(http.Header) | ||
10760 | for k, vs := range objInfo.Metadata { | ||
10761 | if strings.HasPrefix(strings.ToLower(k), "x-amz-meta-") { | ||
10762 | h.Add(k, vs[0]) | ||
10763 | } | ||
10764 | } | ||
10765 | return h | ||
10766 | } | ||
10767 | |||
10768 | // 1. create a client encrypted object to copy by uploading | ||
10769 | const srcSize = 1024 * 1024 | ||
10770 | buf := bytes.Repeat([]byte("abcde"), srcSize) // gives a buffer of 5MiB | ||
10771 | metadata := make(http.Header) | ||
10772 | metadata.Set("x-amz-meta-myheader", "myvalue") | ||
10773 | m := make(map[string]string) | ||
10774 | m["x-amz-meta-myheader"] = "myvalue" | ||
10775 | _, err = c.PutObject(context.Background(), bucketName, "srcObject", | ||
10776 | bytes.NewReader(buf), int64(len(buf)), minio.PutObjectOptions{UserMetadata: m}) | ||
10777 | if err != nil { | ||
10778 | logError(testName, function, args, startTime, "", "PutObjectWithMetadata failed", err) | ||
10779 | return | ||
10780 | } | ||
10781 | if !reflect.DeepEqual(metadata, fetchMeta("srcObject")) { | ||
10782 | logError(testName, function, args, startTime, "", "Metadata match failed", err) | ||
10783 | return | ||
10784 | } | ||
10785 | |||
10786 | // 2. create source | ||
10787 | src := minio.CopySrcOptions{ | ||
10788 | Bucket: bucketName, | ||
10789 | Object: "srcObject", | ||
10790 | } | ||
10791 | |||
10792 | // 2.1 create destination with metadata set | ||
10793 | dst1 := minio.CopyDestOptions{ | ||
10794 | Bucket: bucketName, | ||
10795 | Object: "dstObject-1", | ||
10796 | UserMetadata: map[string]string{"notmyheader": "notmyvalue"}, | ||
10797 | ReplaceMetadata: true, | ||
10798 | } | ||
10799 | |||
10800 | // 3. Check that copying to an object with metadata set resets | ||
10801 | // the headers on the copy. | ||
10802 | args["source"] = src | ||
10803 | args["destination"] = dst1 | ||
10804 | _, err = c.CopyObject(context.Background(), dst1, src) | ||
10805 | if err != nil { | ||
10806 | logError(testName, function, args, startTime, "", "CopyObject failed", err) | ||
10807 | return | ||
10808 | } | ||
10809 | |||
10810 | expectedHeaders := make(http.Header) | ||
10811 | expectedHeaders.Set("x-amz-meta-notmyheader", "notmyvalue") | ||
10812 | if !reflect.DeepEqual(expectedHeaders, fetchMeta("dstObject-1")) { | ||
10813 | logError(testName, function, args, startTime, "", "Metadata match failed", err) | ||
10814 | return | ||
10815 | } | ||
10816 | |||
10817 | // 4. create destination with no metadata set and same source | ||
10818 | dst2 := minio.CopyDestOptions{ | ||
10819 | Bucket: bucketName, | ||
10820 | Object: "dstObject-2", | ||
10821 | } | ||
10822 | |||
10823 | // 5. Check that copying to an object with no metadata set, | ||
10824 | // copies metadata. | ||
10825 | args["source"] = src | ||
10826 | args["destination"] = dst2 | ||
10827 | _, err = c.CopyObject(context.Background(), dst2, src) | ||
10828 | if err != nil { | ||
10829 | logError(testName, function, args, startTime, "", "CopyObject failed", err) | ||
10830 | return | ||
10831 | } | ||
10832 | |||
10833 | expectedHeaders = metadata | ||
10834 | if !reflect.DeepEqual(expectedHeaders, fetchMeta("dstObject-2")) { | ||
10835 | logError(testName, function, args, startTime, "", "Metadata match failed", err) | ||
10836 | return | ||
10837 | } | ||
10838 | |||
10839 | // 6. Compose a pair of sources. | ||
10840 | dst3 := minio.CopyDestOptions{ | ||
10841 | Bucket: bucketName, | ||
10842 | Object: "dstObject-3", | ||
10843 | ReplaceMetadata: true, | ||
10844 | } | ||
10845 | |||
10846 | function = "ComposeObject(destination, sources)" | ||
10847 | args["source"] = []minio.CopySrcOptions{src, src} | ||
10848 | args["destination"] = dst3 | ||
10849 | _, err = c.ComposeObject(context.Background(), dst3, src, src) | ||
10850 | if err != nil { | ||
10851 | logError(testName, function, args, startTime, "", "ComposeObject failed", err) | ||
10852 | return | ||
10853 | } | ||
10854 | |||
10855 | // Check that no headers are copied in this case | ||
10856 | if !reflect.DeepEqual(make(http.Header), fetchMeta("dstObject-3")) { | ||
10857 | logError(testName, function, args, startTime, "", "Metadata match failed", err) | ||
10858 | return | ||
10859 | } | ||
10860 | |||
10861 | // 7. Compose a pair of sources with dest user metadata set. | ||
10862 | dst4 := minio.CopyDestOptions{ | ||
10863 | Bucket: bucketName, | ||
10864 | Object: "dstObject-4", | ||
10865 | UserMetadata: map[string]string{"notmyheader": "notmyvalue"}, | ||
10866 | ReplaceMetadata: true, | ||
10867 | } | ||
10868 | |||
10869 | function = "ComposeObject(destination, sources)" | ||
10870 | args["source"] = []minio.CopySrcOptions{src, src} | ||
10871 | args["destination"] = dst4 | ||
10872 | _, err = c.ComposeObject(context.Background(), dst4, src, src) | ||
10873 | if err != nil { | ||
10874 | logError(testName, function, args, startTime, "", "ComposeObject failed", err) | ||
10875 | return | ||
10876 | } | ||
10877 | |||
10878 | // Check that no headers are copied in this case | ||
10879 | expectedHeaders = make(http.Header) | ||
10880 | expectedHeaders.Set("x-amz-meta-notmyheader", "notmyvalue") | ||
10881 | if !reflect.DeepEqual(expectedHeaders, fetchMeta("dstObject-4")) { | ||
10882 | logError(testName, function, args, startTime, "", "Metadata match failed", err) | ||
10883 | return | ||
10884 | } | ||
10885 | |||
10886 | successLogger(testName, function, args, startTime).Info() | ||
10887 | } | ||
10888 | |||
10889 | func testUserMetadataCopyingV2() { | ||
10890 | // initialize logging params | ||
10891 | startTime := time.Now() | ||
10892 | testName := getFuncName() | ||
10893 | function := "CopyObject(destination, source)" | ||
10894 | args := map[string]interface{}{} | ||
10895 | |||
10896 | // Instantiate new minio client object | ||
10897 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
10898 | &minio.Options{ | ||
10899 | Creds: credentials.NewStaticV2(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
10900 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
10901 | }) | ||
10902 | if err != nil { | ||
10903 | logError(testName, function, args, startTime, "", "MinIO client v2 object creation failed", err) | ||
10904 | return | ||
10905 | } | ||
10906 | |||
10907 | // c.TraceOn(os.Stderr) | ||
10908 | testUserMetadataCopyingWrapper(c) | ||
10909 | } | ||
10910 | |||
10911 | func testStorageClassMetadataPutObject() { | ||
10912 | // initialize logging params | ||
10913 | startTime := time.Now() | ||
10914 | function := "testStorageClassMetadataPutObject()" | ||
10915 | args := map[string]interface{}{} | ||
10916 | testName := getFuncName() | ||
10917 | |||
10918 | // Instantiate new minio client object | ||
10919 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
10920 | &minio.Options{ | ||
10921 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
10922 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
10923 | }) | ||
10924 | if err != nil { | ||
10925 | logError(testName, function, args, startTime, "", "MinIO v4 client object creation failed", err) | ||
10926 | return | ||
10927 | } | ||
10928 | |||
10929 | // Generate a new random bucket name. | ||
10930 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test") | ||
10931 | // Make a new bucket in 'us-east-1' (source bucket). | ||
10932 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
10933 | if err != nil { | ||
10934 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
10935 | return | ||
10936 | } | ||
10937 | |||
10938 | defer cleanupBucket(bucketName, c) | ||
10939 | |||
10940 | fetchMeta := func(object string) (h http.Header) { | ||
10941 | objInfo, err := c.StatObject(context.Background(), bucketName, object, minio.StatObjectOptions{}) | ||
10942 | if err != nil { | ||
10943 | logError(testName, function, args, startTime, "", "Stat failed", err) | ||
10944 | return | ||
10945 | } | ||
10946 | h = make(http.Header) | ||
10947 | for k, vs := range objInfo.Metadata { | ||
10948 | if strings.HasPrefix(strings.ToLower(k), "x-amz-storage-class") { | ||
10949 | for _, v := range vs { | ||
10950 | h.Add(k, v) | ||
10951 | } | ||
10952 | } | ||
10953 | } | ||
10954 | return h | ||
10955 | } | ||
10956 | |||
10957 | metadata := make(http.Header) | ||
10958 | metadata.Set("x-amz-storage-class", "REDUCED_REDUNDANCY") | ||
10959 | |||
10960 | emptyMetadata := make(http.Header) | ||
10961 | |||
10962 | const srcSize = 1024 * 1024 | ||
10963 | buf := bytes.Repeat([]byte("abcde"), srcSize) // gives a buffer of 1MiB | ||
10964 | |||
10965 | _, err = c.PutObject(context.Background(), bucketName, "srcObjectRRSClass", | ||
10966 | bytes.NewReader(buf), int64(len(buf)), minio.PutObjectOptions{StorageClass: "REDUCED_REDUNDANCY"}) | ||
10967 | if err != nil { | ||
10968 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
10969 | return | ||
10970 | } | ||
10971 | |||
10972 | // Get the returned metadata | ||
10973 | returnedMeta := fetchMeta("srcObjectRRSClass") | ||
10974 | |||
10975 | // The response metada should either be equal to metadata (with REDUCED_REDUNDANCY) or emptyMetadata (in case of gateways) | ||
10976 | if !reflect.DeepEqual(metadata, returnedMeta) && !reflect.DeepEqual(emptyMetadata, returnedMeta) { | ||
10977 | logError(testName, function, args, startTime, "", "Metadata match failed", err) | ||
10978 | return | ||
10979 | } | ||
10980 | |||
10981 | metadata = make(http.Header) | ||
10982 | metadata.Set("x-amz-storage-class", "STANDARD") | ||
10983 | |||
10984 | _, err = c.PutObject(context.Background(), bucketName, "srcObjectSSClass", | ||
10985 | bytes.NewReader(buf), int64(len(buf)), minio.PutObjectOptions{StorageClass: "STANDARD"}) | ||
10986 | if err != nil { | ||
10987 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
10988 | return | ||
10989 | } | ||
10990 | if reflect.DeepEqual(metadata, fetchMeta("srcObjectSSClass")) { | ||
10991 | logError(testName, function, args, startTime, "", "Metadata verification failed, STANDARD storage class should not be a part of response metadata", err) | ||
10992 | return | ||
10993 | } | ||
10994 | |||
10995 | successLogger(testName, function, args, startTime).Info() | ||
10996 | } | ||
10997 | |||
10998 | func testStorageClassInvalidMetadataPutObject() { | ||
10999 | // initialize logging params | ||
11000 | startTime := time.Now() | ||
11001 | function := "testStorageClassInvalidMetadataPutObject()" | ||
11002 | args := map[string]interface{}{} | ||
11003 | testName := getFuncName() | ||
11004 | |||
11005 | // Instantiate new minio client object | ||
11006 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
11007 | &minio.Options{ | ||
11008 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
11009 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
11010 | }) | ||
11011 | if err != nil { | ||
11012 | logError(testName, function, args, startTime, "", "MinIO v4 client object creation failed", err) | ||
11013 | return | ||
11014 | } | ||
11015 | |||
11016 | // Generate a new random bucket name. | ||
11017 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test") | ||
11018 | // Make a new bucket in 'us-east-1' (source bucket). | ||
11019 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
11020 | if err != nil { | ||
11021 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
11022 | return | ||
11023 | } | ||
11024 | |||
11025 | defer cleanupBucket(bucketName, c) | ||
11026 | |||
11027 | const srcSize = 1024 * 1024 | ||
11028 | buf := bytes.Repeat([]byte("abcde"), srcSize) // gives a buffer of 1MiB | ||
11029 | |||
11030 | _, err = c.PutObject(context.Background(), bucketName, "srcObjectRRSClass", | ||
11031 | bytes.NewReader(buf), int64(len(buf)), minio.PutObjectOptions{StorageClass: "INVALID_STORAGE_CLASS"}) | ||
11032 | if err == nil { | ||
11033 | logError(testName, function, args, startTime, "", "PutObject with invalid storage class passed, was expected to fail", err) | ||
11034 | return | ||
11035 | } | ||
11036 | |||
11037 | successLogger(testName, function, args, startTime).Info() | ||
11038 | } | ||
11039 | |||
11040 | func testStorageClassMetadataCopyObject() { | ||
11041 | // initialize logging params | ||
11042 | startTime := time.Now() | ||
11043 | function := "testStorageClassMetadataCopyObject()" | ||
11044 | args := map[string]interface{}{} | ||
11045 | testName := getFuncName() | ||
11046 | |||
11047 | // Instantiate new minio client object | ||
11048 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
11049 | &minio.Options{ | ||
11050 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
11051 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
11052 | }) | ||
11053 | if err != nil { | ||
11054 | logError(testName, function, args, startTime, "", "MinIO v4 client object creation failed", err) | ||
11055 | return | ||
11056 | } | ||
11057 | |||
11058 | // Generate a new random bucket name. | ||
11059 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test") | ||
11060 | // Make a new bucket in 'us-east-1' (source bucket). | ||
11061 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
11062 | if err != nil { | ||
11063 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
11064 | return | ||
11065 | } | ||
11066 | |||
11067 | defer cleanupBucket(bucketName, c) | ||
11068 | |||
11069 | fetchMeta := func(object string) (h http.Header) { | ||
11070 | objInfo, err := c.StatObject(context.Background(), bucketName, object, minio.StatObjectOptions{}) | ||
11071 | args["bucket"] = bucketName | ||
11072 | args["object"] = object | ||
11073 | if err != nil { | ||
11074 | logError(testName, function, args, startTime, "", "Stat failed", err) | ||
11075 | return | ||
11076 | } | ||
11077 | h = make(http.Header) | ||
11078 | for k, vs := range objInfo.Metadata { | ||
11079 | if strings.HasPrefix(strings.ToLower(k), "x-amz-storage-class") { | ||
11080 | for _, v := range vs { | ||
11081 | h.Add(k, v) | ||
11082 | } | ||
11083 | } | ||
11084 | } | ||
11085 | return h | ||
11086 | } | ||
11087 | |||
11088 | metadata := make(http.Header) | ||
11089 | metadata.Set("x-amz-storage-class", "REDUCED_REDUNDANCY") | ||
11090 | |||
11091 | emptyMetadata := make(http.Header) | ||
11092 | |||
11093 | const srcSize = 1024 * 1024 | ||
11094 | buf := bytes.Repeat([]byte("abcde"), srcSize) | ||
11095 | |||
11096 | // Put an object with RRS Storage class | ||
11097 | _, err = c.PutObject(context.Background(), bucketName, "srcObjectRRSClass", | ||
11098 | bytes.NewReader(buf), int64(len(buf)), minio.PutObjectOptions{StorageClass: "REDUCED_REDUNDANCY"}) | ||
11099 | if err != nil { | ||
11100 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
11101 | return | ||
11102 | } | ||
11103 | |||
11104 | // Make server side copy of object uploaded in previous step | ||
11105 | src := minio.CopySrcOptions{ | ||
11106 | Bucket: bucketName, | ||
11107 | Object: "srcObjectRRSClass", | ||
11108 | } | ||
11109 | dst := minio.CopyDestOptions{ | ||
11110 | Bucket: bucketName, | ||
11111 | Object: "srcObjectRRSClassCopy", | ||
11112 | } | ||
11113 | if _, err = c.CopyObject(context.Background(), dst, src); err != nil { | ||
11114 | logError(testName, function, args, startTime, "", "CopyObject failed on RRS", err) | ||
11115 | return | ||
11116 | } | ||
11117 | |||
11118 | // Get the returned metadata | ||
11119 | returnedMeta := fetchMeta("srcObjectRRSClassCopy") | ||
11120 | |||
11121 | // The response metada should either be equal to metadata (with REDUCED_REDUNDANCY) or emptyMetadata (in case of gateways) | ||
11122 | if !reflect.DeepEqual(metadata, returnedMeta) && !reflect.DeepEqual(emptyMetadata, returnedMeta) { | ||
11123 | logError(testName, function, args, startTime, "", "Metadata match failed", err) | ||
11124 | return | ||
11125 | } | ||
11126 | |||
11127 | metadata = make(http.Header) | ||
11128 | metadata.Set("x-amz-storage-class", "STANDARD") | ||
11129 | |||
11130 | // Put an object with Standard Storage class | ||
11131 | _, err = c.PutObject(context.Background(), bucketName, "srcObjectSSClass", | ||
11132 | bytes.NewReader(buf), int64(len(buf)), minio.PutObjectOptions{StorageClass: "STANDARD"}) | ||
11133 | if err != nil { | ||
11134 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
11135 | return | ||
11136 | } | ||
11137 | |||
11138 | // Make server side copy of object uploaded in previous step | ||
11139 | src = minio.CopySrcOptions{ | ||
11140 | Bucket: bucketName, | ||
11141 | Object: "srcObjectSSClass", | ||
11142 | } | ||
11143 | dst = minio.CopyDestOptions{ | ||
11144 | Bucket: bucketName, | ||
11145 | Object: "srcObjectSSClassCopy", | ||
11146 | } | ||
11147 | if _, err = c.CopyObject(context.Background(), dst, src); err != nil { | ||
11148 | logError(testName, function, args, startTime, "", "CopyObject failed on SS", err) | ||
11149 | return | ||
11150 | } | ||
11151 | // Fetch the meta data of copied object | ||
11152 | if reflect.DeepEqual(metadata, fetchMeta("srcObjectSSClassCopy")) { | ||
11153 | logError(testName, function, args, startTime, "", "Metadata verification failed, STANDARD storage class should not be a part of response metadata", err) | ||
11154 | return | ||
11155 | } | ||
11156 | |||
11157 | successLogger(testName, function, args, startTime).Info() | ||
11158 | } | ||
11159 | |||
11160 | // Test put object with size -1 byte object. | ||
11161 | func testPutObjectNoLengthV2() { | ||
11162 | // initialize logging params | ||
11163 | startTime := time.Now() | ||
11164 | testName := getFuncName() | ||
11165 | function := "PutObject(bucketName, objectName, reader, size, opts)" | ||
11166 | args := map[string]interface{}{ | ||
11167 | "bucketName": "", | ||
11168 | "objectName": "", | ||
11169 | "size": -1, | ||
11170 | "opts": "", | ||
11171 | } | ||
11172 | |||
11173 | // Seed random based on current time. | ||
11174 | rand.Seed(time.Now().Unix()) | ||
11175 | |||
11176 | // Instantiate new minio client object. | ||
11177 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
11178 | &minio.Options{ | ||
11179 | Creds: credentials.NewStaticV2(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
11180 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
11181 | }) | ||
11182 | if err != nil { | ||
11183 | logError(testName, function, args, startTime, "", "MinIO client v2 object creation failed", err) | ||
11184 | return | ||
11185 | } | ||
11186 | |||
11187 | // Enable tracing, write to stderr. | ||
11188 | // c.TraceOn(os.Stderr) | ||
11189 | |||
11190 | // Set user agent. | ||
11191 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
11192 | |||
11193 | // Generate a new random bucket name. | ||
11194 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
11195 | args["bucketName"] = bucketName | ||
11196 | |||
11197 | // Make a new bucket. | ||
11198 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
11199 | if err != nil { | ||
11200 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
11201 | return | ||
11202 | } | ||
11203 | |||
11204 | defer cleanupBucket(bucketName, c) | ||
11205 | |||
11206 | objectName := bucketName + "unique" | ||
11207 | args["objectName"] = objectName | ||
11208 | |||
11209 | bufSize := dataFileMap["datafile-129-MB"] | ||
11210 | reader := getDataReader("datafile-129-MB") | ||
11211 | defer reader.Close() | ||
11212 | args["size"] = bufSize | ||
11213 | |||
11214 | // Upload an object. | ||
11215 | _, err = c.PutObject(context.Background(), bucketName, objectName, reader, -1, minio.PutObjectOptions{}) | ||
11216 | if err != nil { | ||
11217 | logError(testName, function, args, startTime, "", "PutObjectWithSize failed", err) | ||
11218 | return | ||
11219 | } | ||
11220 | |||
11221 | st, err := c.StatObject(context.Background(), bucketName, objectName, minio.StatObjectOptions{}) | ||
11222 | if err != nil { | ||
11223 | logError(testName, function, args, startTime, "", "StatObject failed", err) | ||
11224 | return | ||
11225 | } | ||
11226 | |||
11227 | if st.Size != int64(bufSize) { | ||
11228 | logError(testName, function, args, startTime, "", "Expected upload object size "+string(bufSize)+" got "+string(st.Size), err) | ||
11229 | return | ||
11230 | } | ||
11231 | |||
11232 | successLogger(testName, function, args, startTime).Info() | ||
11233 | } | ||
11234 | |||
11235 | // Test put objects of unknown size. | ||
11236 | func testPutObjectsUnknownV2() { | ||
11237 | // initialize logging params | ||
11238 | startTime := time.Now() | ||
11239 | testName := getFuncName() | ||
11240 | function := "PutObject(bucketName, objectName, reader,size,opts)" | ||
11241 | args := map[string]interface{}{ | ||
11242 | "bucketName": "", | ||
11243 | "objectName": "", | ||
11244 | "size": "", | ||
11245 | "opts": "", | ||
11246 | } | ||
11247 | |||
11248 | // Seed random based on current time. | ||
11249 | rand.Seed(time.Now().Unix()) | ||
11250 | |||
11251 | // Instantiate new minio client object. | ||
11252 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
11253 | &minio.Options{ | ||
11254 | Creds: credentials.NewStaticV2(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
11255 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
11256 | }) | ||
11257 | if err != nil { | ||
11258 | logError(testName, function, args, startTime, "", "MinIO client v2 object creation failed", err) | ||
11259 | return | ||
11260 | } | ||
11261 | |||
11262 | // Enable tracing, write to stderr. | ||
11263 | // c.TraceOn(os.Stderr) | ||
11264 | |||
11265 | // Set user agent. | ||
11266 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
11267 | |||
11268 | // Generate a new random bucket name. | ||
11269 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
11270 | args["bucketName"] = bucketName | ||
11271 | |||
11272 | // Make a new bucket. | ||
11273 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
11274 | if err != nil { | ||
11275 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
11276 | return | ||
11277 | } | ||
11278 | |||
11279 | defer cleanupBucket(bucketName, c) | ||
11280 | |||
11281 | // Issues are revealed by trying to upload multiple files of unknown size | ||
11282 | // sequentially (on 4GB machines) | ||
11283 | for i := 1; i <= 4; i++ { | ||
11284 | // Simulate that we could be receiving byte slices of data that we want | ||
11285 | // to upload as a file | ||
11286 | rpipe, wpipe := io.Pipe() | ||
11287 | defer rpipe.Close() | ||
11288 | go func() { | ||
11289 | b := []byte("test") | ||
11290 | wpipe.Write(b) | ||
11291 | wpipe.Close() | ||
11292 | }() | ||
11293 | |||
11294 | // Upload the object. | ||
11295 | objectName := fmt.Sprintf("%sunique%d", bucketName, i) | ||
11296 | args["objectName"] = objectName | ||
11297 | |||
11298 | ui, err := c.PutObject(context.Background(), bucketName, objectName, rpipe, -1, minio.PutObjectOptions{}) | ||
11299 | if err != nil { | ||
11300 | logError(testName, function, args, startTime, "", "PutObjectStreaming failed", err) | ||
11301 | return | ||
11302 | } | ||
11303 | |||
11304 | if ui.Size != 4 { | ||
11305 | logError(testName, function, args, startTime, "", "Expected upload object size "+string(4)+" got "+string(ui.Size), nil) | ||
11306 | return | ||
11307 | } | ||
11308 | |||
11309 | st, err := c.StatObject(context.Background(), bucketName, objectName, minio.StatObjectOptions{}) | ||
11310 | if err != nil { | ||
11311 | logError(testName, function, args, startTime, "", "StatObjectStreaming failed", err) | ||
11312 | return | ||
11313 | } | ||
11314 | |||
11315 | if st.Size != int64(4) { | ||
11316 | logError(testName, function, args, startTime, "", "Expected upload object size "+string(4)+" got "+string(st.Size), err) | ||
11317 | return | ||
11318 | } | ||
11319 | |||
11320 | } | ||
11321 | |||
11322 | successLogger(testName, function, args, startTime).Info() | ||
11323 | } | ||
11324 | |||
11325 | // Test put object with 0 byte object. | ||
11326 | func testPutObject0ByteV2() { | ||
11327 | // initialize logging params | ||
11328 | startTime := time.Now() | ||
11329 | testName := getFuncName() | ||
11330 | function := "PutObject(bucketName, objectName, reader, size, opts)" | ||
11331 | args := map[string]interface{}{ | ||
11332 | "bucketName": "", | ||
11333 | "objectName": "", | ||
11334 | "size": 0, | ||
11335 | "opts": "", | ||
11336 | } | ||
11337 | |||
11338 | // Seed random based on current time. | ||
11339 | rand.Seed(time.Now().Unix()) | ||
11340 | |||
11341 | // Instantiate new minio client object. | ||
11342 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
11343 | &minio.Options{ | ||
11344 | Creds: credentials.NewStaticV2(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
11345 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
11346 | }) | ||
11347 | if err != nil { | ||
11348 | logError(testName, function, args, startTime, "", "MinIO client v2 object creation failed", err) | ||
11349 | return | ||
11350 | } | ||
11351 | |||
11352 | // Enable tracing, write to stderr. | ||
11353 | // c.TraceOn(os.Stderr) | ||
11354 | |||
11355 | // Set user agent. | ||
11356 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
11357 | |||
11358 | // Generate a new random bucket name. | ||
11359 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
11360 | args["bucketName"] = bucketName | ||
11361 | |||
11362 | // Make a new bucket. | ||
11363 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
11364 | if err != nil { | ||
11365 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
11366 | return | ||
11367 | } | ||
11368 | |||
11369 | defer cleanupBucket(bucketName, c) | ||
11370 | |||
11371 | objectName := bucketName + "unique" | ||
11372 | args["objectName"] = objectName | ||
11373 | args["opts"] = minio.PutObjectOptions{} | ||
11374 | |||
11375 | // Upload an object. | ||
11376 | _, err = c.PutObject(context.Background(), bucketName, objectName, bytes.NewReader([]byte("")), 0, minio.PutObjectOptions{}) | ||
11377 | if err != nil { | ||
11378 | logError(testName, function, args, startTime, "", "PutObjectWithSize failed", err) | ||
11379 | return | ||
11380 | } | ||
11381 | st, err := c.StatObject(context.Background(), bucketName, objectName, minio.StatObjectOptions{}) | ||
11382 | if err != nil { | ||
11383 | logError(testName, function, args, startTime, "", "StatObjectWithSize failed", err) | ||
11384 | return | ||
11385 | } | ||
11386 | if st.Size != 0 { | ||
11387 | logError(testName, function, args, startTime, "", "Expected upload object size 0 but got "+string(st.Size), err) | ||
11388 | return | ||
11389 | } | ||
11390 | |||
11391 | successLogger(testName, function, args, startTime).Info() | ||
11392 | } | ||
11393 | |||
11394 | // Test expected error cases | ||
11395 | func testComposeObjectErrorCases() { | ||
11396 | // initialize logging params | ||
11397 | startTime := time.Now() | ||
11398 | testName := getFuncName() | ||
11399 | function := "ComposeObject(destination, sourceList)" | ||
11400 | args := map[string]interface{}{} | ||
11401 | |||
11402 | // Instantiate new minio client object | ||
11403 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
11404 | &minio.Options{ | ||
11405 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
11406 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
11407 | }) | ||
11408 | if err != nil { | ||
11409 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
11410 | return | ||
11411 | } | ||
11412 | |||
11413 | testComposeObjectErrorCasesWrapper(c) | ||
11414 | } | ||
11415 | |||
11416 | // Test concatenating multiple 10K objects V4 | ||
11417 | func testCompose10KSources() { | ||
11418 | // initialize logging params | ||
11419 | startTime := time.Now() | ||
11420 | testName := getFuncName() | ||
11421 | function := "ComposeObject(destination, sourceList)" | ||
11422 | args := map[string]interface{}{} | ||
11423 | |||
11424 | // Instantiate new minio client object | ||
11425 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
11426 | &minio.Options{ | ||
11427 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
11428 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
11429 | }) | ||
11430 | if err != nil { | ||
11431 | logError(testName, function, args, startTime, "", "MinIO client object creation failed", err) | ||
11432 | return | ||
11433 | } | ||
11434 | |||
11435 | testComposeMultipleSources(c) | ||
11436 | } | ||
11437 | |||
11438 | // Tests comprehensive list of all methods. | ||
11439 | func testFunctionalV2() { | ||
11440 | // initialize logging params | ||
11441 | startTime := time.Now() | ||
11442 | testName := getFuncName() | ||
11443 | function := "testFunctionalV2()" | ||
11444 | functionAll := "" | ||
11445 | args := map[string]interface{}{} | ||
11446 | |||
11447 | // Seed random based on current time. | ||
11448 | rand.Seed(time.Now().Unix()) | ||
11449 | |||
11450 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
11451 | &minio.Options{ | ||
11452 | Creds: credentials.NewStaticV2(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
11453 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
11454 | }) | ||
11455 | if err != nil { | ||
11456 | logError(testName, function, args, startTime, "", "MinIO client v2 object creation failed", err) | ||
11457 | return | ||
11458 | } | ||
11459 | |||
11460 | // Enable to debug | ||
11461 | // c.TraceOn(os.Stderr) | ||
11462 | |||
11463 | // Set user agent. | ||
11464 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
11465 | |||
11466 | // Generate a new random bucket name. | ||
11467 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
11468 | location := "us-east-1" | ||
11469 | // Make a new bucket. | ||
11470 | function = "MakeBucket(bucketName, location)" | ||
11471 | functionAll = "MakeBucket(bucketName, location)" | ||
11472 | args = map[string]interface{}{ | ||
11473 | "bucketName": bucketName, | ||
11474 | "location": location, | ||
11475 | } | ||
11476 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: location}) | ||
11477 | if err != nil { | ||
11478 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
11479 | return | ||
11480 | } | ||
11481 | |||
11482 | defer cleanupBucket(bucketName, c) | ||
11483 | |||
11484 | // Generate a random file name. | ||
11485 | fileName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
11486 | file, err := os.Create(fileName) | ||
11487 | if err != nil { | ||
11488 | logError(testName, function, args, startTime, "", "file create failed", err) | ||
11489 | return | ||
11490 | } | ||
11491 | for i := 0; i < 3; i++ { | ||
11492 | buf := make([]byte, rand.Intn(1<<19)) | ||
11493 | _, err = file.Write(buf) | ||
11494 | if err != nil { | ||
11495 | logError(testName, function, args, startTime, "", "file write failed", err) | ||
11496 | return | ||
11497 | } | ||
11498 | } | ||
11499 | file.Close() | ||
11500 | |||
11501 | // Verify if bucket exits and you have access. | ||
11502 | var exists bool | ||
11503 | function = "BucketExists(bucketName)" | ||
11504 | functionAll += ", " + function | ||
11505 | args = map[string]interface{}{ | ||
11506 | "bucketName": bucketName, | ||
11507 | } | ||
11508 | exists, err = c.BucketExists(context.Background(), bucketName) | ||
11509 | if err != nil { | ||
11510 | logError(testName, function, args, startTime, "", "BucketExists failed", err) | ||
11511 | return | ||
11512 | } | ||
11513 | if !exists { | ||
11514 | logError(testName, function, args, startTime, "", "Could not find existing bucket "+bucketName, err) | ||
11515 | return | ||
11516 | } | ||
11517 | |||
11518 | // Make the bucket 'public read/write'. | ||
11519 | function = "SetBucketPolicy(bucketName, bucketPolicy)" | ||
11520 | functionAll += ", " + function | ||
11521 | |||
11522 | readWritePolicy := `{"Version": "2012-10-17","Statement": [{"Action": ["s3:ListBucketMultipartUploads", "s3:ListBucket"],"Effect": "Allow","Principal": {"AWS": ["*"]},"Resource": ["arn:aws:s3:::` + bucketName + `"],"Sid": ""}]}` | ||
11523 | |||
11524 | args = map[string]interface{}{ | ||
11525 | "bucketName": bucketName, | ||
11526 | "bucketPolicy": readWritePolicy, | ||
11527 | } | ||
11528 | err = c.SetBucketPolicy(context.Background(), bucketName, readWritePolicy) | ||
11529 | |||
11530 | if err != nil { | ||
11531 | logError(testName, function, args, startTime, "", "SetBucketPolicy failed", err) | ||
11532 | return | ||
11533 | } | ||
11534 | |||
11535 | // List all buckets. | ||
11536 | function = "ListBuckets()" | ||
11537 | functionAll += ", " + function | ||
11538 | args = nil | ||
11539 | buckets, err := c.ListBuckets(context.Background()) | ||
11540 | if len(buckets) == 0 { | ||
11541 | logError(testName, function, args, startTime, "", "List buckets cannot be empty", err) | ||
11542 | return | ||
11543 | } | ||
11544 | if err != nil { | ||
11545 | logError(testName, function, args, startTime, "", "ListBuckets failed", err) | ||
11546 | return | ||
11547 | } | ||
11548 | |||
11549 | // Verify if previously created bucket is listed in list buckets. | ||
11550 | bucketFound := false | ||
11551 | for _, bucket := range buckets { | ||
11552 | if bucket.Name == bucketName { | ||
11553 | bucketFound = true | ||
11554 | } | ||
11555 | } | ||
11556 | |||
11557 | // If bucket not found error out. | ||
11558 | if !bucketFound { | ||
11559 | logError(testName, function, args, startTime, "", "Bucket "+bucketName+"not found", err) | ||
11560 | return | ||
11561 | } | ||
11562 | |||
11563 | objectName := bucketName + "unique" | ||
11564 | |||
11565 | // Generate data | ||
11566 | buf := bytes.Repeat([]byte("n"), rand.Intn(1<<19)) | ||
11567 | |||
11568 | args = map[string]interface{}{ | ||
11569 | "bucketName": bucketName, | ||
11570 | "objectName": objectName, | ||
11571 | "contentType": "", | ||
11572 | } | ||
11573 | _, err = c.PutObject(context.Background(), bucketName, objectName, bytes.NewReader(buf), int64(len(buf)), minio.PutObjectOptions{}) | ||
11574 | if err != nil { | ||
11575 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
11576 | return | ||
11577 | } | ||
11578 | |||
11579 | st, err := c.StatObject(context.Background(), bucketName, objectName, minio.StatObjectOptions{}) | ||
11580 | if err != nil { | ||
11581 | logError(testName, function, args, startTime, "", "StatObject failed", err) | ||
11582 | return | ||
11583 | } | ||
11584 | if st.Size != int64(len(buf)) { | ||
11585 | logError(testName, function, args, startTime, "", "Expected uploaded object length "+string(len(buf))+" got "+string(st.Size), err) | ||
11586 | return | ||
11587 | } | ||
11588 | |||
11589 | objectNameNoLength := objectName + "-nolength" | ||
11590 | args["objectName"] = objectNameNoLength | ||
11591 | _, err = c.PutObject(context.Background(), bucketName, objectNameNoLength, bytes.NewReader(buf), int64(len(buf)), minio.PutObjectOptions{ContentType: "binary/octet-stream"}) | ||
11592 | if err != nil { | ||
11593 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
11594 | return | ||
11595 | } | ||
11596 | st, err = c.StatObject(context.Background(), bucketName, objectNameNoLength, minio.StatObjectOptions{}) | ||
11597 | if err != nil { | ||
11598 | logError(testName, function, args, startTime, "", "StatObject failed", err) | ||
11599 | return | ||
11600 | } | ||
11601 | if st.Size != int64(len(buf)) { | ||
11602 | logError(testName, function, args, startTime, "", "Expected uploaded object length "+string(len(buf))+" got "+string(st.Size), err) | ||
11603 | return | ||
11604 | } | ||
11605 | |||
11606 | // Instantiate a done channel to close all listing. | ||
11607 | doneCh := make(chan struct{}) | ||
11608 | defer close(doneCh) | ||
11609 | |||
11610 | objFound := false | ||
11611 | isRecursive := true // Recursive is true. | ||
11612 | function = "ListObjects(bucketName, objectName, isRecursive, doneCh)" | ||
11613 | functionAll += ", " + function | ||
11614 | args = map[string]interface{}{ | ||
11615 | "bucketName": bucketName, | ||
11616 | "objectName": objectName, | ||
11617 | "isRecursive": isRecursive, | ||
11618 | } | ||
11619 | for obj := range c.ListObjects(context.Background(), bucketName, minio.ListObjectsOptions{UseV1: true, Prefix: objectName, Recursive: isRecursive}) { | ||
11620 | if obj.Key == objectName { | ||
11621 | objFound = true | ||
11622 | break | ||
11623 | } | ||
11624 | } | ||
11625 | if !objFound { | ||
11626 | logError(testName, function, args, startTime, "", "Could not find existing object "+objectName, err) | ||
11627 | return | ||
11628 | } | ||
11629 | |||
11630 | incompObjNotFound := true | ||
11631 | function = "ListIncompleteUploads(bucketName, objectName, isRecursive, doneCh)" | ||
11632 | functionAll += ", " + function | ||
11633 | args = map[string]interface{}{ | ||
11634 | "bucketName": bucketName, | ||
11635 | "objectName": objectName, | ||
11636 | "isRecursive": isRecursive, | ||
11637 | } | ||
11638 | for objIncompl := range c.ListIncompleteUploads(context.Background(), bucketName, objectName, isRecursive) { | ||
11639 | if objIncompl.Key != "" { | ||
11640 | incompObjNotFound = false | ||
11641 | break | ||
11642 | } | ||
11643 | } | ||
11644 | if !incompObjNotFound { | ||
11645 | logError(testName, function, args, startTime, "", "Unexpected dangling incomplete upload found", err) | ||
11646 | return | ||
11647 | } | ||
11648 | |||
11649 | function = "GetObject(bucketName, objectName)" | ||
11650 | functionAll += ", " + function | ||
11651 | args = map[string]interface{}{ | ||
11652 | "bucketName": bucketName, | ||
11653 | "objectName": objectName, | ||
11654 | } | ||
11655 | newReader, err := c.GetObject(context.Background(), bucketName, objectName, minio.GetObjectOptions{}) | ||
11656 | if err != nil { | ||
11657 | logError(testName, function, args, startTime, "", "GetObject failed", err) | ||
11658 | return | ||
11659 | } | ||
11660 | |||
11661 | newReadBytes, err := io.ReadAll(newReader) | ||
11662 | if err != nil { | ||
11663 | logError(testName, function, args, startTime, "", "ReadAll failed", err) | ||
11664 | return | ||
11665 | } | ||
11666 | newReader.Close() | ||
11667 | |||
11668 | if !bytes.Equal(newReadBytes, buf) { | ||
11669 | logError(testName, function, args, startTime, "", "Bytes mismatch", err) | ||
11670 | return | ||
11671 | } | ||
11672 | |||
11673 | function = "FGetObject(bucketName, objectName, fileName)" | ||
11674 | functionAll += ", " + function | ||
11675 | args = map[string]interface{}{ | ||
11676 | "bucketName": bucketName, | ||
11677 | "objectName": objectName, | ||
11678 | "fileName": fileName + "-f", | ||
11679 | } | ||
11680 | err = c.FGetObject(context.Background(), bucketName, objectName, fileName+"-f", minio.GetObjectOptions{}) | ||
11681 | if err != nil { | ||
11682 | logError(testName, function, args, startTime, "", "FgetObject failed", err) | ||
11683 | return | ||
11684 | } | ||
11685 | |||
11686 | // Generate presigned HEAD object url. | ||
11687 | function = "PresignedHeadObject(bucketName, objectName, expires, reqParams)" | ||
11688 | functionAll += ", " + function | ||
11689 | args = map[string]interface{}{ | ||
11690 | "bucketName": bucketName, | ||
11691 | "objectName": objectName, | ||
11692 | "expires": 3600 * time.Second, | ||
11693 | } | ||
11694 | presignedHeadURL, err := c.PresignedHeadObject(context.Background(), bucketName, objectName, 3600*time.Second, nil) | ||
11695 | if err != nil { | ||
11696 | logError(testName, function, args, startTime, "", "PresignedHeadObject failed", err) | ||
11697 | return | ||
11698 | } | ||
11699 | |||
11700 | transport, err := minio.DefaultTransport(mustParseBool(os.Getenv(enableHTTPS))) | ||
11701 | if err != nil { | ||
11702 | logError(testName, function, args, startTime, "", "DefaultTransport failed", err) | ||
11703 | return | ||
11704 | } | ||
11705 | |||
11706 | httpClient := &http.Client{ | ||
11707 | // Setting a sensible time out of 30secs to wait for response | ||
11708 | // headers. Request is pro-actively canceled after 30secs | ||
11709 | // with no response. | ||
11710 | Timeout: 30 * time.Second, | ||
11711 | Transport: transport, | ||
11712 | } | ||
11713 | |||
11714 | req, err := http.NewRequest(http.MethodHead, presignedHeadURL.String(), nil) | ||
11715 | if err != nil { | ||
11716 | logError(testName, function, args, startTime, "", "PresignedHeadObject URL head request failed", err) | ||
11717 | return | ||
11718 | } | ||
11719 | |||
11720 | // Verify if presigned url works. | ||
11721 | resp, err := httpClient.Do(req) | ||
11722 | if err != nil { | ||
11723 | logError(testName, function, args, startTime, "", "PresignedHeadObject URL head request failed", err) | ||
11724 | return | ||
11725 | } | ||
11726 | if resp.StatusCode != http.StatusOK { | ||
11727 | logError(testName, function, args, startTime, "", "PresignedHeadObject URL returns status "+string(resp.StatusCode), err) | ||
11728 | return | ||
11729 | } | ||
11730 | if resp.Header.Get("ETag") == "" { | ||
11731 | logError(testName, function, args, startTime, "", "Got empty ETag", err) | ||
11732 | return | ||
11733 | } | ||
11734 | resp.Body.Close() | ||
11735 | |||
11736 | // Generate presigned GET object url. | ||
11737 | function = "PresignedGetObject(bucketName, objectName, expires, reqParams)" | ||
11738 | functionAll += ", " + function | ||
11739 | args = map[string]interface{}{ | ||
11740 | "bucketName": bucketName, | ||
11741 | "objectName": objectName, | ||
11742 | "expires": 3600 * time.Second, | ||
11743 | } | ||
11744 | presignedGetURL, err := c.PresignedGetObject(context.Background(), bucketName, objectName, 3600*time.Second, nil) | ||
11745 | if err != nil { | ||
11746 | logError(testName, function, args, startTime, "", "PresignedGetObject failed", err) | ||
11747 | return | ||
11748 | } | ||
11749 | |||
11750 | // Verify if presigned url works. | ||
11751 | req, err = http.NewRequest(http.MethodGet, presignedGetURL.String(), nil) | ||
11752 | if err != nil { | ||
11753 | logError(testName, function, args, startTime, "", "PresignedGetObject request incorrect", err) | ||
11754 | return | ||
11755 | } | ||
11756 | |||
11757 | resp, err = httpClient.Do(req) | ||
11758 | if err != nil { | ||
11759 | logError(testName, function, args, startTime, "", "PresignedGetObject response incorrect", err) | ||
11760 | return | ||
11761 | } | ||
11762 | |||
11763 | if resp.StatusCode != http.StatusOK { | ||
11764 | logError(testName, function, args, startTime, "", "PresignedGetObject URL returns status "+string(resp.StatusCode), err) | ||
11765 | return | ||
11766 | } | ||
11767 | newPresignedBytes, err := io.ReadAll(resp.Body) | ||
11768 | if err != nil { | ||
11769 | logError(testName, function, args, startTime, "", "ReadAll failed", err) | ||
11770 | return | ||
11771 | } | ||
11772 | resp.Body.Close() | ||
11773 | if !bytes.Equal(newPresignedBytes, buf) { | ||
11774 | logError(testName, function, args, startTime, "", "Bytes mismatch", err) | ||
11775 | return | ||
11776 | } | ||
11777 | |||
11778 | // Set request parameters. | ||
11779 | reqParams := make(url.Values) | ||
11780 | reqParams.Set("response-content-disposition", "attachment; filename=\"test.txt\"") | ||
11781 | // Generate presigned GET object url. | ||
11782 | args["reqParams"] = reqParams | ||
11783 | presignedGetURL, err = c.PresignedGetObject(context.Background(), bucketName, objectName, 3600*time.Second, reqParams) | ||
11784 | if err != nil { | ||
11785 | logError(testName, function, args, startTime, "", "PresignedGetObject failed", err) | ||
11786 | return | ||
11787 | } | ||
11788 | |||
11789 | // Verify if presigned url works. | ||
11790 | req, err = http.NewRequest(http.MethodGet, presignedGetURL.String(), nil) | ||
11791 | if err != nil { | ||
11792 | logError(testName, function, args, startTime, "", "PresignedGetObject request incorrect", err) | ||
11793 | return | ||
11794 | } | ||
11795 | |||
11796 | resp, err = httpClient.Do(req) | ||
11797 | if err != nil { | ||
11798 | logError(testName, function, args, startTime, "", "PresignedGetObject response incorrect", err) | ||
11799 | return | ||
11800 | } | ||
11801 | |||
11802 | if resp.StatusCode != http.StatusOK { | ||
11803 | logError(testName, function, args, startTime, "", "PresignedGetObject URL returns status "+string(resp.StatusCode), err) | ||
11804 | return | ||
11805 | } | ||
11806 | newPresignedBytes, err = io.ReadAll(resp.Body) | ||
11807 | if err != nil { | ||
11808 | logError(testName, function, args, startTime, "", "ReadAll failed", err) | ||
11809 | return | ||
11810 | } | ||
11811 | if !bytes.Equal(newPresignedBytes, buf) { | ||
11812 | logError(testName, function, args, startTime, "", "Bytes mismatch", err) | ||
11813 | return | ||
11814 | } | ||
11815 | // Verify content disposition. | ||
11816 | if resp.Header.Get("Content-Disposition") != "attachment; filename=\"test.txt\"" { | ||
11817 | logError(testName, function, args, startTime, "", "wrong Content-Disposition received ", err) | ||
11818 | return | ||
11819 | } | ||
11820 | |||
11821 | function = "PresignedPutObject(bucketName, objectName, expires)" | ||
11822 | functionAll += ", " + function | ||
11823 | args = map[string]interface{}{ | ||
11824 | "bucketName": bucketName, | ||
11825 | "objectName": objectName + "-presigned", | ||
11826 | "expires": 3600 * time.Second, | ||
11827 | } | ||
11828 | presignedPutURL, err := c.PresignedPutObject(context.Background(), bucketName, objectName+"-presigned", 3600*time.Second) | ||
11829 | if err != nil { | ||
11830 | logError(testName, function, args, startTime, "", "PresignedPutObject failed", err) | ||
11831 | return | ||
11832 | } | ||
11833 | |||
11834 | // Generate data more than 32K | ||
11835 | buf = bytes.Repeat([]byte("1"), rand.Intn(1<<10)+32*1024) | ||
11836 | |||
11837 | req, err = http.NewRequest(http.MethodPut, presignedPutURL.String(), bytes.NewReader(buf)) | ||
11838 | if err != nil { | ||
11839 | logError(testName, function, args, startTime, "", "HTTP request to PresignedPutObject URL failed", err) | ||
11840 | return | ||
11841 | } | ||
11842 | |||
11843 | resp, err = httpClient.Do(req) | ||
11844 | if err != nil { | ||
11845 | logError(testName, function, args, startTime, "", "HTTP request to PresignedPutObject URL failed", err) | ||
11846 | return | ||
11847 | } | ||
11848 | |||
11849 | // Download the uploaded object to verify | ||
11850 | args = map[string]interface{}{ | ||
11851 | "bucketName": bucketName, | ||
11852 | "objectName": objectName + "-presigned", | ||
11853 | } | ||
11854 | newReader, err = c.GetObject(context.Background(), bucketName, objectName+"-presigned", minio.GetObjectOptions{}) | ||
11855 | if err != nil { | ||
11856 | logError(testName, function, args, startTime, "", "GetObject of uploaded presigned object failed", err) | ||
11857 | return | ||
11858 | } | ||
11859 | |||
11860 | newReadBytes, err = io.ReadAll(newReader) | ||
11861 | if err != nil { | ||
11862 | logError(testName, function, args, startTime, "", "ReadAll failed during get on presigned put object", err) | ||
11863 | return | ||
11864 | } | ||
11865 | newReader.Close() | ||
11866 | |||
11867 | if !bytes.Equal(newReadBytes, buf) { | ||
11868 | logError(testName, function, args, startTime, "", "Bytes mismatch on presigned object upload verification", err) | ||
11869 | return | ||
11870 | } | ||
11871 | |||
11872 | function = "PresignHeader(method, bucketName, objectName, expires, reqParams, extraHeaders)" | ||
11873 | functionAll += ", " + function | ||
11874 | presignExtraHeaders := map[string][]string{ | ||
11875 | "mysecret": {"abcxxx"}, | ||
11876 | } | ||
11877 | args = map[string]interface{}{ | ||
11878 | "method": "PUT", | ||
11879 | "bucketName": bucketName, | ||
11880 | "objectName": objectName + "-presign-custom", | ||
11881 | "expires": 3600 * time.Second, | ||
11882 | "extraHeaders": presignExtraHeaders, | ||
11883 | } | ||
11884 | _, err = c.PresignHeader(context.Background(), "PUT", bucketName, objectName+"-presign-custom", 3600*time.Second, nil, presignExtraHeaders) | ||
11885 | if err == nil { | ||
11886 | logError(testName, function, args, startTime, "", "Presigned with extra headers succeeded", err) | ||
11887 | return | ||
11888 | } | ||
11889 | |||
11890 | os.Remove(fileName) | ||
11891 | os.Remove(fileName + "-f") | ||
11892 | successLogger(testName, functionAll, args, startTime).Info() | ||
11893 | } | ||
11894 | |||
11895 | // Test get object with GetObject with context | ||
11896 | func testGetObjectContext() { | ||
11897 | // initialize logging params | ||
11898 | startTime := time.Now() | ||
11899 | testName := getFuncName() | ||
11900 | function := "GetObject(ctx, bucketName, objectName)" | ||
11901 | args := map[string]interface{}{ | ||
11902 | "ctx": "", | ||
11903 | "bucketName": "", | ||
11904 | "objectName": "", | ||
11905 | } | ||
11906 | // Seed random based on current time. | ||
11907 | rand.Seed(time.Now().Unix()) | ||
11908 | |||
11909 | // Instantiate new minio client object. | ||
11910 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
11911 | &minio.Options{ | ||
11912 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
11913 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
11914 | }) | ||
11915 | if err != nil { | ||
11916 | logError(testName, function, args, startTime, "", "MinIO client v4 object creation failed", err) | ||
11917 | return | ||
11918 | } | ||
11919 | |||
11920 | // Enable tracing, write to stderr. | ||
11921 | // c.TraceOn(os.Stderr) | ||
11922 | |||
11923 | // Set user agent. | ||
11924 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
11925 | |||
11926 | // Generate a new random bucket name. | ||
11927 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
11928 | args["bucketName"] = bucketName | ||
11929 | |||
11930 | // Make a new bucket. | ||
11931 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
11932 | if err != nil { | ||
11933 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
11934 | return | ||
11935 | } | ||
11936 | |||
11937 | defer cleanupBucket(bucketName, c) | ||
11938 | |||
11939 | bufSize := dataFileMap["datafile-33-kB"] | ||
11940 | reader := getDataReader("datafile-33-kB") | ||
11941 | defer reader.Close() | ||
11942 | // Save the data | ||
11943 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
11944 | args["objectName"] = objectName | ||
11945 | |||
11946 | _, err = c.PutObject(context.Background(), bucketName, objectName, reader, int64(bufSize), minio.PutObjectOptions{ContentType: "binary/octet-stream"}) | ||
11947 | if err != nil { | ||
11948 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
11949 | return | ||
11950 | } | ||
11951 | |||
11952 | ctx, cancel := context.WithTimeout(context.Background(), 1*time.Nanosecond) | ||
11953 | args["ctx"] = ctx | ||
11954 | cancel() | ||
11955 | |||
11956 | r, err := c.GetObject(ctx, bucketName, objectName, minio.GetObjectOptions{}) | ||
11957 | if err != nil { | ||
11958 | logError(testName, function, args, startTime, "", "GetObject failed unexpectedly", err) | ||
11959 | return | ||
11960 | } | ||
11961 | |||
11962 | if _, err = r.Stat(); err == nil { | ||
11963 | logError(testName, function, args, startTime, "", "GetObject should fail on short timeout", err) | ||
11964 | return | ||
11965 | } | ||
11966 | r.Close() | ||
11967 | |||
11968 | ctx, cancel = context.WithTimeout(context.Background(), 1*time.Hour) | ||
11969 | args["ctx"] = ctx | ||
11970 | defer cancel() | ||
11971 | |||
11972 | // Read the data back | ||
11973 | r, err = c.GetObject(ctx, bucketName, objectName, minio.GetObjectOptions{}) | ||
11974 | if err != nil { | ||
11975 | logError(testName, function, args, startTime, "", "GetObject failed", err) | ||
11976 | return | ||
11977 | } | ||
11978 | |||
11979 | st, err := r.Stat() | ||
11980 | if err != nil { | ||
11981 | logError(testName, function, args, startTime, "", "object Stat call failed", err) | ||
11982 | return | ||
11983 | } | ||
11984 | if st.Size != int64(bufSize) { | ||
11985 | logError(testName, function, args, startTime, "", "Number of bytes in stat does not match: want "+string(bufSize)+", got"+string(st.Size), err) | ||
11986 | return | ||
11987 | } | ||
11988 | if err := r.Close(); err != nil { | ||
11989 | logError(testName, function, args, startTime, "", "object Close() call failed", err) | ||
11990 | return | ||
11991 | } | ||
11992 | |||
11993 | successLogger(testName, function, args, startTime).Info() | ||
11994 | } | ||
11995 | |||
11996 | // Test get object with FGetObject with a user provided context | ||
11997 | func testFGetObjectContext() { | ||
11998 | // initialize logging params | ||
11999 | startTime := time.Now() | ||
12000 | testName := getFuncName() | ||
12001 | function := "FGetObject(ctx, bucketName, objectName, fileName)" | ||
12002 | args := map[string]interface{}{ | ||
12003 | "ctx": "", | ||
12004 | "bucketName": "", | ||
12005 | "objectName": "", | ||
12006 | "fileName": "", | ||
12007 | } | ||
12008 | // Seed random based on current time. | ||
12009 | rand.Seed(time.Now().Unix()) | ||
12010 | |||
12011 | // Instantiate new minio client object. | ||
12012 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
12013 | &minio.Options{ | ||
12014 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
12015 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
12016 | }) | ||
12017 | if err != nil { | ||
12018 | logError(testName, function, args, startTime, "", "MinIO client v4 object creation failed", err) | ||
12019 | return | ||
12020 | } | ||
12021 | |||
12022 | // Enable tracing, write to stderr. | ||
12023 | // c.TraceOn(os.Stderr) | ||
12024 | |||
12025 | // Set user agent. | ||
12026 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
12027 | |||
12028 | // Generate a new random bucket name. | ||
12029 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
12030 | args["bucketName"] = bucketName | ||
12031 | |||
12032 | // Make a new bucket. | ||
12033 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
12034 | if err != nil { | ||
12035 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
12036 | return | ||
12037 | } | ||
12038 | |||
12039 | defer cleanupBucket(bucketName, c) | ||
12040 | |||
12041 | bufSize := dataFileMap["datafile-1-MB"] | ||
12042 | reader := getDataReader("datafile-1-MB") | ||
12043 | defer reader.Close() | ||
12044 | // Save the data | ||
12045 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
12046 | args["objectName"] = objectName | ||
12047 | |||
12048 | _, err = c.PutObject(context.Background(), bucketName, objectName, reader, int64(bufSize), minio.PutObjectOptions{ContentType: "binary/octet-stream"}) | ||
12049 | if err != nil { | ||
12050 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
12051 | return | ||
12052 | } | ||
12053 | |||
12054 | ctx, cancel := context.WithTimeout(context.Background(), 1*time.Nanosecond) | ||
12055 | args["ctx"] = ctx | ||
12056 | defer cancel() | ||
12057 | |||
12058 | fileName := "tempfile-context" | ||
12059 | args["fileName"] = fileName | ||
12060 | // Read the data back | ||
12061 | err = c.FGetObject(ctx, bucketName, objectName, fileName+"-f", minio.GetObjectOptions{}) | ||
12062 | if err == nil { | ||
12063 | logError(testName, function, args, startTime, "", "FGetObject should fail on short timeout", err) | ||
12064 | return | ||
12065 | } | ||
12066 | ctx, cancel = context.WithTimeout(context.Background(), 1*time.Hour) | ||
12067 | defer cancel() | ||
12068 | |||
12069 | // Read the data back | ||
12070 | err = c.FGetObject(ctx, bucketName, objectName, fileName+"-fcontext", minio.GetObjectOptions{}) | ||
12071 | if err != nil { | ||
12072 | logError(testName, function, args, startTime, "", "FGetObject with long timeout failed", err) | ||
12073 | return | ||
12074 | } | ||
12075 | if err = os.Remove(fileName + "-fcontext"); err != nil { | ||
12076 | logError(testName, function, args, startTime, "", "Remove file failed", err) | ||
12077 | return | ||
12078 | } | ||
12079 | |||
12080 | successLogger(testName, function, args, startTime).Info() | ||
12081 | } | ||
12082 | |||
12083 | // Test get object with GetObject with a user provided context | ||
12084 | func testGetObjectRanges() { | ||
12085 | // initialize logging params | ||
12086 | startTime := time.Now() | ||
12087 | testName := getFuncName() | ||
12088 | function := "GetObject(ctx, bucketName, objectName, fileName)" | ||
12089 | args := map[string]interface{}{ | ||
12090 | "ctx": "", | ||
12091 | "bucketName": "", | ||
12092 | "objectName": "", | ||
12093 | "fileName": "", | ||
12094 | } | ||
12095 | ctx, cancel := context.WithTimeout(context.Background(), 5*time.Minute) | ||
12096 | defer cancel() | ||
12097 | |||
12098 | rng := rand.NewSource(time.Now().UnixNano()) | ||
12099 | // Instantiate new minio client object. | ||
12100 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
12101 | &minio.Options{ | ||
12102 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
12103 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
12104 | }) | ||
12105 | if err != nil { | ||
12106 | logError(testName, function, args, startTime, "", "MinIO client v4 object creation failed", err) | ||
12107 | return | ||
12108 | } | ||
12109 | |||
12110 | // Enable tracing, write to stderr. | ||
12111 | // c.TraceOn(os.Stderr) | ||
12112 | |||
12113 | // Set user agent. | ||
12114 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
12115 | |||
12116 | // Generate a new random bucket name. | ||
12117 | bucketName := randString(60, rng, "minio-go-test-") | ||
12118 | args["bucketName"] = bucketName | ||
12119 | |||
12120 | // Make a new bucket. | ||
12121 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
12122 | if err != nil { | ||
12123 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
12124 | return | ||
12125 | } | ||
12126 | |||
12127 | defer cleanupBucket(bucketName, c) | ||
12128 | |||
12129 | bufSize := dataFileMap["datafile-129-MB"] | ||
12130 | reader := getDataReader("datafile-129-MB") | ||
12131 | defer reader.Close() | ||
12132 | // Save the data | ||
12133 | objectName := randString(60, rng, "") | ||
12134 | args["objectName"] = objectName | ||
12135 | |||
12136 | _, err = c.PutObject(context.Background(), bucketName, objectName, reader, int64(bufSize), minio.PutObjectOptions{ContentType: "binary/octet-stream"}) | ||
12137 | if err != nil { | ||
12138 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
12139 | return | ||
12140 | } | ||
12141 | |||
12142 | // Read the data back | ||
12143 | tests := []struct { | ||
12144 | start int64 | ||
12145 | end int64 | ||
12146 | }{ | ||
12147 | { | ||
12148 | start: 1024, | ||
12149 | end: 1024 + 1<<20, | ||
12150 | }, | ||
12151 | { | ||
12152 | start: 20e6, | ||
12153 | end: 20e6 + 10000, | ||
12154 | }, | ||
12155 | { | ||
12156 | start: 40e6, | ||
12157 | end: 40e6 + 10000, | ||
12158 | }, | ||
12159 | { | ||
12160 | start: 60e6, | ||
12161 | end: 60e6 + 10000, | ||
12162 | }, | ||
12163 | { | ||
12164 | start: 80e6, | ||
12165 | end: 80e6 + 10000, | ||
12166 | }, | ||
12167 | { | ||
12168 | start: 120e6, | ||
12169 | end: int64(bufSize), | ||
12170 | }, | ||
12171 | } | ||
12172 | for _, test := range tests { | ||
12173 | wantRC := getDataReader("datafile-129-MB") | ||
12174 | io.CopyN(io.Discard, wantRC, test.start) | ||
12175 | want := mustCrcReader(io.LimitReader(wantRC, test.end-test.start+1)) | ||
12176 | opts := minio.GetObjectOptions{} | ||
12177 | opts.SetRange(test.start, test.end) | ||
12178 | args["opts"] = fmt.Sprintf("%+v", test) | ||
12179 | obj, err := c.GetObject(ctx, bucketName, objectName, opts) | ||
12180 | if err != nil { | ||
12181 | logError(testName, function, args, startTime, "", "FGetObject with long timeout failed", err) | ||
12182 | return | ||
12183 | } | ||
12184 | err = crcMatches(obj, want) | ||
12185 | if err != nil { | ||
12186 | logError(testName, function, args, startTime, "", fmt.Sprintf("GetObject offset %d -> %d", test.start, test.end), err) | ||
12187 | return | ||
12188 | } | ||
12189 | } | ||
12190 | |||
12191 | successLogger(testName, function, args, startTime).Info() | ||
12192 | } | ||
12193 | |||
12194 | // Test get object ACLs with GetObjectACL with custom provided context | ||
12195 | func testGetObjectACLContext() { | ||
12196 | // initialize logging params | ||
12197 | startTime := time.Now() | ||
12198 | testName := getFuncName() | ||
12199 | function := "GetObjectACL(ctx, bucketName, objectName)" | ||
12200 | args := map[string]interface{}{ | ||
12201 | "ctx": "", | ||
12202 | "bucketName": "", | ||
12203 | "objectName": "", | ||
12204 | } | ||
12205 | // Seed random based on current time. | ||
12206 | rand.Seed(time.Now().Unix()) | ||
12207 | |||
12208 | // Instantiate new minio client object. | ||
12209 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
12210 | &minio.Options{ | ||
12211 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
12212 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
12213 | }) | ||
12214 | if err != nil { | ||
12215 | logError(testName, function, args, startTime, "", "MinIO client v4 object creation failed", err) | ||
12216 | return | ||
12217 | } | ||
12218 | |||
12219 | // Enable tracing, write to stderr. | ||
12220 | // c.TraceOn(os.Stderr) | ||
12221 | |||
12222 | // Set user agent. | ||
12223 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
12224 | |||
12225 | // Generate a new random bucket name. | ||
12226 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
12227 | args["bucketName"] = bucketName | ||
12228 | |||
12229 | // Make a new bucket. | ||
12230 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
12231 | if err != nil { | ||
12232 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
12233 | return | ||
12234 | } | ||
12235 | |||
12236 | defer cleanupBucket(bucketName, c) | ||
12237 | |||
12238 | bufSize := dataFileMap["datafile-1-MB"] | ||
12239 | reader := getDataReader("datafile-1-MB") | ||
12240 | defer reader.Close() | ||
12241 | // Save the data | ||
12242 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
12243 | args["objectName"] = objectName | ||
12244 | |||
12245 | // Add meta data to add a canned acl | ||
12246 | metaData := map[string]string{ | ||
12247 | "X-Amz-Acl": "public-read-write", | ||
12248 | } | ||
12249 | |||
12250 | _, err = c.PutObject(context.Background(), bucketName, | ||
12251 | objectName, reader, int64(bufSize), | ||
12252 | minio.PutObjectOptions{ | ||
12253 | ContentType: "binary/octet-stream", | ||
12254 | UserMetadata: metaData, | ||
12255 | }) | ||
12256 | |||
12257 | if err != nil { | ||
12258 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
12259 | return | ||
12260 | } | ||
12261 | |||
12262 | ctx, cancel := context.WithTimeout(context.Background(), 10*time.Second) | ||
12263 | args["ctx"] = ctx | ||
12264 | defer cancel() | ||
12265 | |||
12266 | // Read the data back | ||
12267 | objectInfo, getObjectACLErr := c.GetObjectACL(ctx, bucketName, objectName) | ||
12268 | if getObjectACLErr != nil { | ||
12269 | logError(testName, function, args, startTime, "", "GetObjectACL failed. ", getObjectACLErr) | ||
12270 | return | ||
12271 | } | ||
12272 | |||
12273 | s, ok := objectInfo.Metadata["X-Amz-Acl"] | ||
12274 | if !ok { | ||
12275 | logError(testName, function, args, startTime, "", "GetObjectACL fail unable to find \"X-Amz-Acl\"", nil) | ||
12276 | return | ||
12277 | } | ||
12278 | |||
12279 | if len(s) != 1 { | ||
12280 | logError(testName, function, args, startTime, "", "GetObjectACL fail \"X-Amz-Acl\" canned acl expected \"1\" got "+fmt.Sprintf(`"%d"`, len(s)), nil) | ||
12281 | return | ||
12282 | } | ||
12283 | |||
12284 | // Do a very limited testing if this is not AWS S3 | ||
12285 | if os.Getenv(serverEndpoint) != "s3.amazonaws.com" { | ||
12286 | if s[0] != "private" { | ||
12287 | logError(testName, function, args, startTime, "", "GetObjectACL fail \"X-Amz-Acl\" expected \"private\" but got"+fmt.Sprintf("%q", s[0]), nil) | ||
12288 | return | ||
12289 | } | ||
12290 | |||
12291 | successLogger(testName, function, args, startTime).Info() | ||
12292 | return | ||
12293 | } | ||
12294 | |||
12295 | if s[0] != "public-read-write" { | ||
12296 | logError(testName, function, args, startTime, "", "GetObjectACL fail \"X-Amz-Acl\" expected \"public-read-write\" but got"+fmt.Sprintf("%q", s[0]), nil) | ||
12297 | return | ||
12298 | } | ||
12299 | |||
12300 | bufSize = dataFileMap["datafile-1-MB"] | ||
12301 | reader2 := getDataReader("datafile-1-MB") | ||
12302 | defer reader2.Close() | ||
12303 | // Save the data | ||
12304 | objectName = randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
12305 | args["objectName"] = objectName | ||
12306 | |||
12307 | // Add meta data to add a canned acl | ||
12308 | metaData = map[string]string{ | ||
12309 | "X-Amz-Grant-Read": "[email protected]", | ||
12310 | "X-Amz-Grant-Write": "[email protected]", | ||
12311 | } | ||
12312 | |||
12313 | _, err = c.PutObject(context.Background(), bucketName, objectName, reader2, int64(bufSize), minio.PutObjectOptions{ContentType: "binary/octet-stream", UserMetadata: metaData}) | ||
12314 | if err != nil { | ||
12315 | logError(testName, function, args, startTime, "", "PutObject failed", err) | ||
12316 | return | ||
12317 | } | ||
12318 | |||
12319 | ctx, cancel = context.WithTimeout(context.Background(), 10*time.Second) | ||
12320 | args["ctx"] = ctx | ||
12321 | defer cancel() | ||
12322 | |||
12323 | // Read the data back | ||
12324 | objectInfo, getObjectACLErr = c.GetObjectACL(ctx, bucketName, objectName) | ||
12325 | if getObjectACLErr == nil { | ||
12326 | logError(testName, function, args, startTime, "", "GetObjectACL fail", getObjectACLErr) | ||
12327 | return | ||
12328 | } | ||
12329 | |||
12330 | if len(objectInfo.Metadata) != 3 { | ||
12331 | logError(testName, function, args, startTime, "", "GetObjectACL fail expected \"3\" ACLs but got "+fmt.Sprintf(`"%d"`, len(objectInfo.Metadata)), nil) | ||
12332 | return | ||
12333 | } | ||
12334 | |||
12335 | s, ok = objectInfo.Metadata["X-Amz-Grant-Read"] | ||
12336 | if !ok { | ||
12337 | logError(testName, function, args, startTime, "", "GetObjectACL fail unable to find \"X-Amz-Grant-Read\"", nil) | ||
12338 | return | ||
12339 | } | ||
12340 | |||
12341 | if len(s) != 1 { | ||
12342 | logError(testName, function, args, startTime, "", "GetObjectACL fail \"X-Amz-Grant-Read\" acl expected \"1\" got "+fmt.Sprintf(`"%d"`, len(s)), nil) | ||
12343 | return | ||
12344 | } | ||
12345 | |||
12346 | if s[0] != "[email protected]" { | ||
12347 | logError(testName, function, args, startTime, "", "GetObjectACL fail \"X-Amz-Grant-Read\" acl expected \"[email protected]\" got "+fmt.Sprintf("%q", s), nil) | ||
12348 | return | ||
12349 | } | ||
12350 | |||
12351 | s, ok = objectInfo.Metadata["X-Amz-Grant-Write"] | ||
12352 | if !ok { | ||
12353 | logError(testName, function, args, startTime, "", "GetObjectACL fail unable to find \"X-Amz-Grant-Write\"", nil) | ||
12354 | return | ||
12355 | } | ||
12356 | |||
12357 | if len(s) != 1 { | ||
12358 | logError(testName, function, args, startTime, "", "GetObjectACL fail \"X-Amz-Grant-Write\" acl expected \"1\" got "+fmt.Sprintf(`"%d"`, len(s)), nil) | ||
12359 | return | ||
12360 | } | ||
12361 | |||
12362 | if s[0] != "[email protected]" { | ||
12363 | logError(testName, function, args, startTime, "", "GetObjectACL fail \"X-Amz-Grant-Write\" acl expected \"[email protected]\" got "+fmt.Sprintf("%q", s), nil) | ||
12364 | return | ||
12365 | } | ||
12366 | |||
12367 | successLogger(testName, function, args, startTime).Info() | ||
12368 | } | ||
12369 | |||
12370 | // Test validates putObject with context to see if request cancellation is honored for V2. | ||
12371 | func testPutObjectContextV2() { | ||
12372 | // initialize logging params | ||
12373 | startTime := time.Now() | ||
12374 | testName := getFuncName() | ||
12375 | function := "PutObject(ctx, bucketName, objectName, reader, size, opts)" | ||
12376 | args := map[string]interface{}{ | ||
12377 | "ctx": "", | ||
12378 | "bucketName": "", | ||
12379 | "objectName": "", | ||
12380 | "size": "", | ||
12381 | "opts": "", | ||
12382 | } | ||
12383 | // Instantiate new minio client object. | ||
12384 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
12385 | &minio.Options{ | ||
12386 | Creds: credentials.NewStaticV2(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
12387 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
12388 | }) | ||
12389 | if err != nil { | ||
12390 | logError(testName, function, args, startTime, "", "MinIO client v2 object creation failed", err) | ||
12391 | return | ||
12392 | } | ||
12393 | |||
12394 | // Enable tracing, write to stderr. | ||
12395 | // c.TraceOn(os.Stderr) | ||
12396 | |||
12397 | // Set user agent. | ||
12398 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
12399 | |||
12400 | // Make a new bucket. | ||
12401 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
12402 | args["bucketName"] = bucketName | ||
12403 | |||
12404 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
12405 | if err != nil { | ||
12406 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
12407 | return | ||
12408 | } | ||
12409 | defer cleanupBucket(bucketName, c) | ||
12410 | bufSize := dataFileMap["datatfile-33-kB"] | ||
12411 | reader := getDataReader("datafile-33-kB") | ||
12412 | defer reader.Close() | ||
12413 | |||
12414 | objectName := fmt.Sprintf("test-file-%v", rand.Uint32()) | ||
12415 | args["objectName"] = objectName | ||
12416 | |||
12417 | ctx, cancel := context.WithTimeout(context.Background(), 30*time.Second) | ||
12418 | args["ctx"] = ctx | ||
12419 | args["size"] = bufSize | ||
12420 | defer cancel() | ||
12421 | |||
12422 | _, err = c.PutObject(ctx, bucketName, objectName, reader, int64(bufSize), minio.PutObjectOptions{ContentType: "binary/octet-stream"}) | ||
12423 | if err != nil { | ||
12424 | logError(testName, function, args, startTime, "", "PutObject with short timeout failed", err) | ||
12425 | return | ||
12426 | } | ||
12427 | |||
12428 | ctx, cancel = context.WithTimeout(context.Background(), 1*time.Hour) | ||
12429 | args["ctx"] = ctx | ||
12430 | |||
12431 | defer cancel() | ||
12432 | reader = getDataReader("datafile-33-kB") | ||
12433 | defer reader.Close() | ||
12434 | _, err = c.PutObject(ctx, bucketName, objectName, reader, int64(bufSize), minio.PutObjectOptions{ContentType: "binary/octet-stream"}) | ||
12435 | if err != nil { | ||
12436 | logError(testName, function, args, startTime, "", "PutObject with long timeout failed", err) | ||
12437 | return | ||
12438 | } | ||
12439 | |||
12440 | successLogger(testName, function, args, startTime).Info() | ||
12441 | } | ||
12442 | |||
12443 | // Test get object with GetObject with custom context | ||
12444 | func testGetObjectContextV2() { | ||
12445 | // initialize logging params | ||
12446 | startTime := time.Now() | ||
12447 | testName := getFuncName() | ||
12448 | function := "GetObject(ctx, bucketName, objectName)" | ||
12449 | args := map[string]interface{}{ | ||
12450 | "ctx": "", | ||
12451 | "bucketName": "", | ||
12452 | "objectName": "", | ||
12453 | } | ||
12454 | // Seed random based on current time. | ||
12455 | rand.Seed(time.Now().Unix()) | ||
12456 | |||
12457 | // Instantiate new minio client object. | ||
12458 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
12459 | &minio.Options{ | ||
12460 | Creds: credentials.NewStaticV2(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
12461 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
12462 | }) | ||
12463 | if err != nil { | ||
12464 | logError(testName, function, args, startTime, "", "MinIO client v2 object creation failed", err) | ||
12465 | return | ||
12466 | } | ||
12467 | |||
12468 | // Enable tracing, write to stderr. | ||
12469 | // c.TraceOn(os.Stderr) | ||
12470 | |||
12471 | // Set user agent. | ||
12472 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
12473 | |||
12474 | // Generate a new random bucket name. | ||
12475 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
12476 | args["bucketName"] = bucketName | ||
12477 | |||
12478 | // Make a new bucket. | ||
12479 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
12480 | if err != nil { | ||
12481 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
12482 | return | ||
12483 | } | ||
12484 | |||
12485 | defer cleanupBucket(bucketName, c) | ||
12486 | |||
12487 | bufSize := dataFileMap["datafile-33-kB"] | ||
12488 | reader := getDataReader("datafile-33-kB") | ||
12489 | defer reader.Close() | ||
12490 | // Save the data | ||
12491 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
12492 | args["objectName"] = objectName | ||
12493 | |||
12494 | _, err = c.PutObject(context.Background(), bucketName, objectName, reader, int64(bufSize), minio.PutObjectOptions{ContentType: "binary/octet-stream"}) | ||
12495 | if err != nil { | ||
12496 | logError(testName, function, args, startTime, "", "PutObject call failed", err) | ||
12497 | return | ||
12498 | } | ||
12499 | |||
12500 | ctx, cancel := context.WithTimeout(context.Background(), 1*time.Nanosecond) | ||
12501 | args["ctx"] = ctx | ||
12502 | cancel() | ||
12503 | |||
12504 | r, err := c.GetObject(ctx, bucketName, objectName, minio.GetObjectOptions{}) | ||
12505 | if err != nil { | ||
12506 | logError(testName, function, args, startTime, "", "GetObject failed unexpectedly", err) | ||
12507 | return | ||
12508 | } | ||
12509 | if _, err = r.Stat(); err == nil { | ||
12510 | logError(testName, function, args, startTime, "", "GetObject should fail on short timeout", err) | ||
12511 | return | ||
12512 | } | ||
12513 | r.Close() | ||
12514 | |||
12515 | ctx, cancel = context.WithTimeout(context.Background(), 1*time.Hour) | ||
12516 | defer cancel() | ||
12517 | |||
12518 | // Read the data back | ||
12519 | r, err = c.GetObject(ctx, bucketName, objectName, minio.GetObjectOptions{}) | ||
12520 | if err != nil { | ||
12521 | logError(testName, function, args, startTime, "", "GetObject shouldn't fail on longer timeout", err) | ||
12522 | return | ||
12523 | } | ||
12524 | |||
12525 | st, err := r.Stat() | ||
12526 | if err != nil { | ||
12527 | logError(testName, function, args, startTime, "", "object Stat call failed", err) | ||
12528 | return | ||
12529 | } | ||
12530 | if st.Size != int64(bufSize) { | ||
12531 | logError(testName, function, args, startTime, "", "Number of bytes in stat does not match, expected "+string(bufSize)+" got "+string(st.Size), err) | ||
12532 | return | ||
12533 | } | ||
12534 | if err := r.Close(); err != nil { | ||
12535 | logError(testName, function, args, startTime, "", " object Close() call failed", err) | ||
12536 | return | ||
12537 | } | ||
12538 | |||
12539 | successLogger(testName, function, args, startTime).Info() | ||
12540 | } | ||
12541 | |||
12542 | // Test get object with FGetObject with custom context | ||
12543 | func testFGetObjectContextV2() { | ||
12544 | // initialize logging params | ||
12545 | startTime := time.Now() | ||
12546 | testName := getFuncName() | ||
12547 | function := "FGetObject(ctx, bucketName, objectName,fileName)" | ||
12548 | args := map[string]interface{}{ | ||
12549 | "ctx": "", | ||
12550 | "bucketName": "", | ||
12551 | "objectName": "", | ||
12552 | "fileName": "", | ||
12553 | } | ||
12554 | // Seed random based on current time. | ||
12555 | rand.Seed(time.Now().Unix()) | ||
12556 | |||
12557 | // Instantiate new minio client object. | ||
12558 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
12559 | &minio.Options{ | ||
12560 | Creds: credentials.NewStaticV2(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
12561 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
12562 | }) | ||
12563 | if err != nil { | ||
12564 | logError(testName, function, args, startTime, "", "MinIO client v2 object creation failed", err) | ||
12565 | return | ||
12566 | } | ||
12567 | |||
12568 | // Enable tracing, write to stderr. | ||
12569 | // c.TraceOn(os.Stderr) | ||
12570 | |||
12571 | // Set user agent. | ||
12572 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
12573 | |||
12574 | // Generate a new random bucket name. | ||
12575 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
12576 | args["bucketName"] = bucketName | ||
12577 | |||
12578 | // Make a new bucket. | ||
12579 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
12580 | if err != nil { | ||
12581 | logError(testName, function, args, startTime, "", "MakeBucket call failed", err) | ||
12582 | return | ||
12583 | } | ||
12584 | |||
12585 | defer cleanupBucket(bucketName, c) | ||
12586 | |||
12587 | bufSize := dataFileMap["datatfile-1-MB"] | ||
12588 | reader := getDataReader("datafile-1-MB") | ||
12589 | defer reader.Close() | ||
12590 | // Save the data | ||
12591 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
12592 | args["objectName"] = objectName | ||
12593 | |||
12594 | _, err = c.PutObject(context.Background(), bucketName, objectName, reader, int64(bufSize), minio.PutObjectOptions{ContentType: "binary/octet-stream"}) | ||
12595 | if err != nil { | ||
12596 | logError(testName, function, args, startTime, "", "PutObject call failed", err) | ||
12597 | return | ||
12598 | } | ||
12599 | |||
12600 | ctx, cancel := context.WithTimeout(context.Background(), 1*time.Nanosecond) | ||
12601 | args["ctx"] = ctx | ||
12602 | defer cancel() | ||
12603 | |||
12604 | fileName := "tempfile-context" | ||
12605 | args["fileName"] = fileName | ||
12606 | |||
12607 | // Read the data back | ||
12608 | err = c.FGetObject(ctx, bucketName, objectName, fileName+"-f", minio.GetObjectOptions{}) | ||
12609 | if err == nil { | ||
12610 | logError(testName, function, args, startTime, "", "FGetObject should fail on short timeout", err) | ||
12611 | return | ||
12612 | } | ||
12613 | ctx, cancel = context.WithTimeout(context.Background(), 1*time.Hour) | ||
12614 | defer cancel() | ||
12615 | |||
12616 | // Read the data back | ||
12617 | err = c.FGetObject(ctx, bucketName, objectName, fileName+"-fcontext", minio.GetObjectOptions{}) | ||
12618 | if err != nil { | ||
12619 | logError(testName, function, args, startTime, "", "FGetObject call shouldn't fail on long timeout", err) | ||
12620 | return | ||
12621 | } | ||
12622 | |||
12623 | if err = os.Remove(fileName + "-fcontext"); err != nil { | ||
12624 | logError(testName, function, args, startTime, "", "Remove file failed", err) | ||
12625 | return | ||
12626 | } | ||
12627 | |||
12628 | successLogger(testName, function, args, startTime).Info() | ||
12629 | } | ||
12630 | |||
12631 | // Test list object v1 and V2 | ||
12632 | func testListObjects() { | ||
12633 | // initialize logging params | ||
12634 | startTime := time.Now() | ||
12635 | testName := getFuncName() | ||
12636 | function := "ListObjects(bucketName, objectPrefix, recursive, doneCh)" | ||
12637 | args := map[string]interface{}{ | ||
12638 | "bucketName": "", | ||
12639 | "objectPrefix": "", | ||
12640 | "recursive": "true", | ||
12641 | } | ||
12642 | // Seed random based on current time. | ||
12643 | rand.Seed(time.Now().Unix()) | ||
12644 | |||
12645 | // Instantiate new minio client object. | ||
12646 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
12647 | &minio.Options{ | ||
12648 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
12649 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
12650 | }) | ||
12651 | if err != nil { | ||
12652 | logError(testName, function, args, startTime, "", "MinIO client v4 object creation failed", err) | ||
12653 | return | ||
12654 | } | ||
12655 | |||
12656 | // Enable tracing, write to stderr. | ||
12657 | // c.TraceOn(os.Stderr) | ||
12658 | |||
12659 | // Set user agent. | ||
12660 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
12661 | |||
12662 | // Generate a new random bucket name. | ||
12663 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
12664 | args["bucketName"] = bucketName | ||
12665 | |||
12666 | // Make a new bucket. | ||
12667 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1"}) | ||
12668 | if err != nil { | ||
12669 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
12670 | return | ||
12671 | } | ||
12672 | |||
12673 | defer cleanupBucket(bucketName, c) | ||
12674 | |||
12675 | testObjects := []struct { | ||
12676 | name string | ||
12677 | storageClass string | ||
12678 | }{ | ||
12679 | // Special characters | ||
12680 | {"foo bar", "STANDARD"}, | ||
12681 | {"foo-%", "STANDARD"}, | ||
12682 | {"random-object-1", "STANDARD"}, | ||
12683 | {"random-object-2", "REDUCED_REDUNDANCY"}, | ||
12684 | } | ||
12685 | |||
12686 | for i, object := range testObjects { | ||
12687 | bufSize := dataFileMap["datafile-33-kB"] | ||
12688 | reader := getDataReader("datafile-33-kB") | ||
12689 | defer reader.Close() | ||
12690 | _, err = c.PutObject(context.Background(), bucketName, object.name, reader, int64(bufSize), | ||
12691 | minio.PutObjectOptions{ContentType: "binary/octet-stream", StorageClass: object.storageClass}) | ||
12692 | if err != nil { | ||
12693 | logError(testName, function, args, startTime, "", fmt.Sprintf("PutObject %d call failed", i+1), err) | ||
12694 | return | ||
12695 | } | ||
12696 | } | ||
12697 | |||
12698 | testList := func(listFn func(context.Context, string, minio.ListObjectsOptions) <-chan minio.ObjectInfo, bucket string, opts minio.ListObjectsOptions) { | ||
12699 | var objCursor int | ||
12700 | |||
12701 | // check for object name and storage-class from listing object result | ||
12702 | for objInfo := range listFn(context.Background(), bucket, opts) { | ||
12703 | if objInfo.Err != nil { | ||
12704 | logError(testName, function, args, startTime, "", "ListObjects failed unexpectedly", err) | ||
12705 | return | ||
12706 | } | ||
12707 | if objInfo.Key != testObjects[objCursor].name { | ||
12708 | logError(testName, function, args, startTime, "", "ListObjects does not return expected object name", err) | ||
12709 | return | ||
12710 | } | ||
12711 | if objInfo.StorageClass != testObjects[objCursor].storageClass { | ||
12712 | // Ignored as Gateways (Azure/GCS etc) wont return storage class | ||
12713 | ignoredLog(testName, function, args, startTime, "ListObjects doesn't return expected storage class").Info() | ||
12714 | } | ||
12715 | objCursor++ | ||
12716 | } | ||
12717 | |||
12718 | if objCursor != len(testObjects) { | ||
12719 | logError(testName, function, args, startTime, "", "ListObjects returned unexpected number of items", errors.New("")) | ||
12720 | return | ||
12721 | } | ||
12722 | } | ||
12723 | |||
12724 | testList(c.ListObjects, bucketName, minio.ListObjectsOptions{Recursive: true, UseV1: true}) | ||
12725 | testList(c.ListObjects, bucketName, minio.ListObjectsOptions{Recursive: true}) | ||
12726 | testList(c.ListObjects, bucketName, minio.ListObjectsOptions{Recursive: true, WithMetadata: true}) | ||
12727 | |||
12728 | successLogger(testName, function, args, startTime).Info() | ||
12729 | } | ||
12730 | |||
12731 | // Test deleting multiple objects with object retention set in Governance mode | ||
12732 | func testRemoveObjects() { | ||
12733 | // initialize logging params | ||
12734 | startTime := time.Now() | ||
12735 | testName := getFuncName() | ||
12736 | function := "RemoveObjects(bucketName, objectsCh, opts)" | ||
12737 | args := map[string]interface{}{ | ||
12738 | "bucketName": "", | ||
12739 | "objectPrefix": "", | ||
12740 | "recursive": "true", | ||
12741 | } | ||
12742 | // Seed random based on current time. | ||
12743 | rand.Seed(time.Now().Unix()) | ||
12744 | |||
12745 | // Instantiate new minio client object. | ||
12746 | c, err := minio.New(os.Getenv(serverEndpoint), | ||
12747 | &minio.Options{ | ||
12748 | Creds: credentials.NewStaticV4(os.Getenv(accessKey), os.Getenv(secretKey), ""), | ||
12749 | Secure: mustParseBool(os.Getenv(enableHTTPS)), | ||
12750 | }) | ||
12751 | if err != nil { | ||
12752 | logError(testName, function, args, startTime, "", "MinIO client v4 object creation failed", err) | ||
12753 | return | ||
12754 | } | ||
12755 | |||
12756 | // Enable tracing, write to stderr. | ||
12757 | // c.TraceOn(os.Stderr) | ||
12758 | |||
12759 | // Set user agent. | ||
12760 | c.SetAppInfo("MinIO-go-FunctionalTest", "0.1.0") | ||
12761 | |||
12762 | // Generate a new random bucket name. | ||
12763 | bucketName := randString(60, rand.NewSource(time.Now().UnixNano()), "minio-go-test-") | ||
12764 | args["bucketName"] = bucketName | ||
12765 | objectName := randString(60, rand.NewSource(time.Now().UnixNano()), "") | ||
12766 | args["objectName"] = objectName | ||
12767 | |||
12768 | // Make a new bucket. | ||
12769 | err = c.MakeBucket(context.Background(), bucketName, minio.MakeBucketOptions{Region: "us-east-1", ObjectLocking: true}) | ||
12770 | if err != nil { | ||
12771 | logError(testName, function, args, startTime, "", "MakeBucket failed", err) | ||
12772 | return | ||
12773 | } | ||
12774 | |||
12775 | bufSize := dataFileMap["datafile-129-MB"] | ||
12776 | reader := getDataReader("datafile-129-MB") | ||
12777 | defer reader.Close() | ||
12778 | |||
12779 | _, err = c.PutObject(context.Background(), bucketName, objectName, reader, int64(bufSize), minio.PutObjectOptions{}) | ||
12780 | if err != nil { | ||
12781 | logError(testName, function, args, startTime, "", "Error uploading object", err) | ||
12782 | return | ||
12783 | } | ||
12784 | |||
12785 | // Replace with smaller... | ||
12786 | bufSize = dataFileMap["datafile-10-kB"] | ||
12787 | reader = getDataReader("datafile-10-kB") | ||
12788 | defer reader.Close() | ||
12789 | |||
12790 | _, err = c.PutObject(context.Background(), bucketName, objectName, reader, int64(bufSize), minio.PutObjectOptions{}) | ||
12791 | if err != nil { | ||
12792 | logError(testName, function, args, startTime, "", "Error uploading object", err) | ||
12793 | } | ||
12794 | |||
12795 | t := time.Date(2030, time.April, 25, 14, 0, 0, 0, time.UTC) | ||
12796 | m := minio.RetentionMode(minio.Governance) | ||
12797 | opts := minio.PutObjectRetentionOptions{ | ||
12798 | GovernanceBypass: false, | ||
12799 | RetainUntilDate: &t, | ||
12800 | Mode: &m, | ||
12801 | } | ||
12802 | err = c.PutObjectRetention(context.Background(), bucketName, objectName, opts) | ||
12803 | if err != nil { | ||
12804 | logError(testName, function, args, startTime, "", "Error setting retention", err) | ||
12805 | return | ||
12806 | } | ||
12807 | |||
12808 | objectsCh := make(chan minio.ObjectInfo) | ||
12809 | // Send object names that are needed to be removed to objectsCh | ||
12810 | go func() { | ||
12811 | defer close(objectsCh) | ||
12812 | // List all objects from a bucket-name with a matching prefix. | ||
12813 | for object := range c.ListObjects(context.Background(), bucketName, minio.ListObjectsOptions{UseV1: true, Recursive: true}) { | ||
12814 | if object.Err != nil { | ||
12815 | logError(testName, function, args, startTime, "", "Error listing objects", object.Err) | ||
12816 | return | ||
12817 | } | ||
12818 | objectsCh <- object | ||
12819 | } | ||
12820 | }() | ||
12821 | |||
12822 | for rErr := range c.RemoveObjects(context.Background(), bucketName, objectsCh, minio.RemoveObjectsOptions{}) { | ||
12823 | // Error is expected here because Retention is set on the object | ||
12824 | // and RemoveObjects is called without Bypass Governance | ||
12825 | if rErr.Err == nil { | ||
12826 | logError(testName, function, args, startTime, "", "Expected error during deletion", nil) | ||
12827 | return | ||
12828 | } | ||
12829 | } | ||
12830 | |||
12831 | objectsCh1 := make(chan minio.ObjectInfo) | ||
12832 | |||
12833 | // Send object names that are needed to be removed to objectsCh | ||
12834 | go func() { | ||
12835 | defer close(objectsCh1) | ||
12836 | // List all objects from a bucket-name with a matching prefix. | ||
12837 | for object := range c.ListObjects(context.Background(), bucketName, minio.ListObjectsOptions{UseV1: true, Recursive: true}) { | ||
12838 | if object.Err != nil { | ||
12839 | logError(testName, function, args, startTime, "", "Error listing objects", object.Err) | ||
12840 | return | ||
12841 | } | ||
12842 | objectsCh1 <- object | ||
12843 | } | ||
12844 | }() | ||
12845 | |||
12846 | opts1 := minio.RemoveObjectsOptions{ | ||
12847 | GovernanceBypass: true, | ||
12848 | } | ||
12849 | |||
12850 | for rErr := range c.RemoveObjects(context.Background(), bucketName, objectsCh1, opts1) { | ||
12851 | // Error is not expected here because Retention is set on the object | ||
12852 | // and RemoveObjects is called with Bypass Governance | ||
12853 | logError(testName, function, args, startTime, "", "Error detected during deletion", rErr.Err) | ||
12854 | return | ||
12855 | } | ||
12856 | |||
12857 | // Delete all objects and buckets | ||
12858 | if err = cleanupVersionedBucket(bucketName, c); err != nil { | ||
12859 | logError(testName, function, args, startTime, "", "CleanupBucket failed", err) | ||
12860 | return | ||
12861 | } | ||
12862 | |||
12863 | successLogger(testName, function, args, startTime).Info() | ||
12864 | } | ||
12865 | |||
12866 | // Convert string to bool and always return false if any error | ||
12867 | func mustParseBool(str string) bool { | ||
12868 | b, err := strconv.ParseBool(str) | ||
12869 | if err != nil { | ||
12870 | return false | ||
12871 | } | ||
12872 | return b | ||
12873 | } | ||
12874 | |||
12875 | func main() { | ||
12876 | // Output to stdout instead of the default stderr | ||
12877 | log.SetOutput(os.Stdout) | ||
12878 | // create custom formatter | ||
12879 | mintFormatter := mintJSONFormatter{} | ||
12880 | // set custom formatter | ||
12881 | log.SetFormatter(&mintFormatter) | ||
12882 | // log Info or above -- success cases are Info level, failures are Fatal level | ||
12883 | log.SetLevel(log.InfoLevel) | ||
12884 | |||
12885 | tls := mustParseBool(os.Getenv(enableHTTPS)) | ||
12886 | kms := mustParseBool(os.Getenv(enableKMS)) | ||
12887 | if os.Getenv(enableKMS) == "" { | ||
12888 | // Default to KMS tests. | ||
12889 | kms = true | ||
12890 | } | ||
12891 | |||
12892 | // execute tests | ||
12893 | if isFullMode() { | ||
12894 | testMakeBucketErrorV2() | ||
12895 | testGetObjectClosedTwiceV2() | ||
12896 | testFPutObjectV2() | ||
12897 | testMakeBucketRegionsV2() | ||
12898 | testGetObjectReadSeekFunctionalV2() | ||
12899 | testGetObjectReadAtFunctionalV2() | ||
12900 | testGetObjectRanges() | ||
12901 | testCopyObjectV2() | ||
12902 | testFunctionalV2() | ||
12903 | testComposeObjectErrorCasesV2() | ||
12904 | testCompose10KSourcesV2() | ||
12905 | testUserMetadataCopyingV2() | ||
12906 | testPutObjectWithChecksums() | ||
12907 | testPutMultipartObjectWithChecksums() | ||
12908 | testPutObject0ByteV2() | ||
12909 | testPutObjectNoLengthV2() | ||
12910 | testPutObjectsUnknownV2() | ||
12911 | testGetObjectContextV2() | ||
12912 | testFPutObjectContextV2() | ||
12913 | testFGetObjectContextV2() | ||
12914 | testPutObjectContextV2() | ||
12915 | testPutObjectWithVersioning() | ||
12916 | testMakeBucketError() | ||
12917 | testMakeBucketRegions() | ||
12918 | testPutObjectWithMetadata() | ||
12919 | testPutObjectReadAt() | ||
12920 | testPutObjectStreaming() | ||
12921 | testGetObjectSeekEnd() | ||
12922 | testGetObjectClosedTwice() | ||
12923 | testGetObjectS3Zip() | ||
12924 | testRemoveMultipleObjects() | ||
12925 | testRemoveMultipleObjectsWithResult() | ||
12926 | testFPutObjectMultipart() | ||
12927 | testFPutObject() | ||
12928 | testGetObjectReadSeekFunctional() | ||
12929 | testGetObjectReadAtFunctional() | ||
12930 | testGetObjectReadAtWhenEOFWasReached() | ||
12931 | testPresignedPostPolicy() | ||
12932 | testCopyObject() | ||
12933 | testComposeObjectErrorCases() | ||
12934 | testCompose10KSources() | ||
12935 | testUserMetadataCopying() | ||
12936 | testBucketNotification() | ||
12937 | testFunctional() | ||
12938 | testGetObjectModified() | ||
12939 | testPutObjectUploadSeekedObject() | ||
12940 | testGetObjectContext() | ||
12941 | testFPutObjectContext() | ||
12942 | testFGetObjectContext() | ||
12943 | testGetObjectACLContext() | ||
12944 | testPutObjectContext() | ||
12945 | testStorageClassMetadataPutObject() | ||
12946 | testStorageClassInvalidMetadataPutObject() | ||
12947 | testStorageClassMetadataCopyObject() | ||
12948 | testPutObjectWithContentLanguage() | ||
12949 | testListObjects() | ||
12950 | testRemoveObjects() | ||
12951 | testListObjectVersions() | ||
12952 | testStatObjectWithVersioning() | ||
12953 | testGetObjectWithVersioning() | ||
12954 | testCopyObjectWithVersioning() | ||
12955 | testConcurrentCopyObjectWithVersioning() | ||
12956 | testComposeObjectWithVersioning() | ||
12957 | testRemoveObjectWithVersioning() | ||
12958 | testRemoveObjectsWithVersioning() | ||
12959 | testObjectTaggingWithVersioning() | ||
12960 | testTrailingChecksums() | ||
12961 | testPutObjectWithAutomaticChecksums() | ||
12962 | |||
12963 | // SSE-C tests will only work over TLS connection. | ||
12964 | if tls { | ||
12965 | testSSECEncryptionPutGet() | ||
12966 | testSSECEncryptionFPut() | ||
12967 | testSSECEncryptedGetObjectReadAtFunctional() | ||
12968 | testSSECEncryptedGetObjectReadSeekFunctional() | ||
12969 | testEncryptedCopyObjectV2() | ||
12970 | testEncryptedSSECToSSECCopyObject() | ||
12971 | testEncryptedSSECToUnencryptedCopyObject() | ||
12972 | testUnencryptedToSSECCopyObject() | ||
12973 | testUnencryptedToUnencryptedCopyObject() | ||
12974 | testEncryptedEmptyObject() | ||
12975 | testDecryptedCopyObject() | ||
12976 | testSSECEncryptedToSSECCopyObjectPart() | ||
12977 | testSSECMultipartEncryptedToSSECCopyObjectPart() | ||
12978 | testSSECEncryptedToUnencryptedCopyPart() | ||
12979 | testUnencryptedToSSECCopyObjectPart() | ||
12980 | testUnencryptedToUnencryptedCopyPart() | ||
12981 | testEncryptedSSECToSSES3CopyObject() | ||
12982 | testEncryptedSSES3ToSSECCopyObject() | ||
12983 | testSSECEncryptedToSSES3CopyObjectPart() | ||
12984 | testSSES3EncryptedToSSECCopyObjectPart() | ||
12985 | } | ||
12986 | |||
12987 | // KMS tests | ||
12988 | if kms { | ||
12989 | testSSES3EncryptionPutGet() | ||
12990 | testSSES3EncryptionFPut() | ||
12991 | testSSES3EncryptedGetObjectReadAtFunctional() | ||
12992 | testSSES3EncryptedGetObjectReadSeekFunctional() | ||
12993 | testEncryptedSSES3ToSSES3CopyObject() | ||
12994 | testEncryptedSSES3ToUnencryptedCopyObject() | ||
12995 | testUnencryptedToSSES3CopyObject() | ||
12996 | testUnencryptedToSSES3CopyObjectPart() | ||
12997 | testSSES3EncryptedToUnencryptedCopyPart() | ||
12998 | testSSES3EncryptedToSSES3CopyObjectPart() | ||
12999 | } | ||
13000 | } else { | ||
13001 | testFunctional() | ||
13002 | testFunctionalV2() | ||
13003 | } | ||
13004 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/hook-reader.go b/vendor/github.com/minio/minio-go/v7/hook-reader.go deleted file mode 100644 index 07bc7db..0000000 --- a/vendor/github.com/minio/minio-go/v7/hook-reader.go +++ /dev/null | |||
@@ -1,101 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2015-2017 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package minio | ||
19 | |||
20 | import ( | ||
21 | "fmt" | ||
22 | "io" | ||
23 | "sync" | ||
24 | ) | ||
25 | |||
26 | // hookReader hooks additional reader in the source stream. It is | ||
27 | // useful for making progress bars. Second reader is appropriately | ||
28 | // notified about the exact number of bytes read from the primary | ||
29 | // source on each Read operation. | ||
30 | type hookReader struct { | ||
31 | mu sync.RWMutex | ||
32 | source io.Reader | ||
33 | hook io.Reader | ||
34 | } | ||
35 | |||
36 | // Seek implements io.Seeker. Seeks source first, and if necessary | ||
37 | // seeks hook if Seek method is appropriately found. | ||
38 | func (hr *hookReader) Seek(offset int64, whence int) (n int64, err error) { | ||
39 | hr.mu.Lock() | ||
40 | defer hr.mu.Unlock() | ||
41 | |||
42 | // Verify for source has embedded Seeker, use it. | ||
43 | sourceSeeker, ok := hr.source.(io.Seeker) | ||
44 | if ok { | ||
45 | n, err = sourceSeeker.Seek(offset, whence) | ||
46 | if err != nil { | ||
47 | return 0, err | ||
48 | } | ||
49 | } | ||
50 | |||
51 | if hr.hook != nil { | ||
52 | // Verify if hook has embedded Seeker, use it. | ||
53 | hookSeeker, ok := hr.hook.(io.Seeker) | ||
54 | if ok { | ||
55 | var m int64 | ||
56 | m, err = hookSeeker.Seek(offset, whence) | ||
57 | if err != nil { | ||
58 | return 0, err | ||
59 | } | ||
60 | if n != m { | ||
61 | return 0, fmt.Errorf("hook seeker seeked %d bytes, expected source %d bytes", m, n) | ||
62 | } | ||
63 | } | ||
64 | } | ||
65 | |||
66 | return n, nil | ||
67 | } | ||
68 | |||
69 | // Read implements io.Reader. Always reads from the source, the return | ||
70 | // value 'n' number of bytes are reported through the hook. Returns | ||
71 | // error for all non io.EOF conditions. | ||
72 | func (hr *hookReader) Read(b []byte) (n int, err error) { | ||
73 | hr.mu.RLock() | ||
74 | defer hr.mu.RUnlock() | ||
75 | |||
76 | n, err = hr.source.Read(b) | ||
77 | if err != nil && err != io.EOF { | ||
78 | return n, err | ||
79 | } | ||
80 | if hr.hook != nil { | ||
81 | // Progress the hook with the total read bytes from the source. | ||
82 | if _, herr := hr.hook.Read(b[:n]); herr != nil { | ||
83 | if herr != io.EOF { | ||
84 | return n, herr | ||
85 | } | ||
86 | } | ||
87 | } | ||
88 | return n, err | ||
89 | } | ||
90 | |||
91 | // newHook returns a io.ReadSeeker which implements hookReader that | ||
92 | // reports the data read from the source to the hook. | ||
93 | func newHook(source, hook io.Reader) io.Reader { | ||
94 | if hook == nil { | ||
95 | return &hookReader{source: source} | ||
96 | } | ||
97 | return &hookReader{ | ||
98 | source: source, | ||
99 | hook: hook, | ||
100 | } | ||
101 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/pkg/credentials/assume_role.go b/vendor/github.com/minio/minio-go/v7/pkg/credentials/assume_role.go deleted file mode 100644 index 800c4a2..0000000 --- a/vendor/github.com/minio/minio-go/v7/pkg/credentials/assume_role.go +++ /dev/null | |||
@@ -1,242 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2020 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package credentials | ||
19 | |||
20 | import ( | ||
21 | "bytes" | ||
22 | "crypto/sha256" | ||
23 | "encoding/hex" | ||
24 | "encoding/xml" | ||
25 | "errors" | ||
26 | "io" | ||
27 | "net/http" | ||
28 | "net/url" | ||
29 | "strconv" | ||
30 | "strings" | ||
31 | "time" | ||
32 | |||
33 | "github.com/minio/minio-go/v7/pkg/signer" | ||
34 | ) | ||
35 | |||
36 | // AssumeRoleResponse contains the result of successful AssumeRole request. | ||
37 | type AssumeRoleResponse struct { | ||
38 | XMLName xml.Name `xml:"https://sts.amazonaws.com/doc/2011-06-15/ AssumeRoleResponse" json:"-"` | ||
39 | |||
40 | Result AssumeRoleResult `xml:"AssumeRoleResult"` | ||
41 | ResponseMetadata struct { | ||
42 | RequestID string `xml:"RequestId,omitempty"` | ||
43 | } `xml:"ResponseMetadata,omitempty"` | ||
44 | } | ||
45 | |||
46 | // AssumeRoleResult - Contains the response to a successful AssumeRole | ||
47 | // request, including temporary credentials that can be used to make | ||
48 | // MinIO API requests. | ||
49 | type AssumeRoleResult struct { | ||
50 | // The identifiers for the temporary security credentials that the operation | ||
51 | // returns. | ||
52 | AssumedRoleUser AssumedRoleUser `xml:",omitempty"` | ||
53 | |||
54 | // The temporary security credentials, which include an access key ID, a secret | ||
55 | // access key, and a security (or session) token. | ||
56 | // | ||
57 | // Note: The size of the security token that STS APIs return is not fixed. We | ||
58 | // strongly recommend that you make no assumptions about the maximum size. As | ||
59 | // of this writing, the typical size is less than 4096 bytes, but that can vary. | ||
60 | // Also, future updates to AWS might require larger sizes. | ||
61 | Credentials struct { | ||
62 | AccessKey string `xml:"AccessKeyId" json:"accessKey,omitempty"` | ||
63 | SecretKey string `xml:"SecretAccessKey" json:"secretKey,omitempty"` | ||
64 | Expiration time.Time `xml:"Expiration" json:"expiration,omitempty"` | ||
65 | SessionToken string `xml:"SessionToken" json:"sessionToken,omitempty"` | ||
66 | } `xml:",omitempty"` | ||
67 | |||
68 | // A percentage value that indicates the size of the policy in packed form. | ||
69 | // The service rejects any policy with a packed size greater than 100 percent, | ||
70 | // which means the policy exceeded the allowed space. | ||
71 | PackedPolicySize int `xml:",omitempty"` | ||
72 | } | ||
73 | |||
74 | // A STSAssumeRole retrieves credentials from MinIO service, and keeps track if | ||
75 | // those credentials are expired. | ||
76 | type STSAssumeRole struct { | ||
77 | Expiry | ||
78 | |||
79 | // Required http Client to use when connecting to MinIO STS service. | ||
80 | Client *http.Client | ||
81 | |||
82 | // STS endpoint to fetch STS credentials. | ||
83 | STSEndpoint string | ||
84 | |||
85 | // various options for this request. | ||
86 | Options STSAssumeRoleOptions | ||
87 | } | ||
88 | |||
89 | // STSAssumeRoleOptions collection of various input options | ||
90 | // to obtain AssumeRole credentials. | ||
91 | type STSAssumeRoleOptions struct { | ||
92 | // Mandatory inputs. | ||
93 | AccessKey string | ||
94 | SecretKey string | ||
95 | |||
96 | SessionToken string // Optional if the first request is made with temporary credentials. | ||
97 | Policy string // Optional to assign a policy to the assumed role | ||
98 | |||
99 | Location string // Optional commonly needed with AWS STS. | ||
100 | DurationSeconds int // Optional defaults to 1 hour. | ||
101 | |||
102 | // Optional only valid if using with AWS STS | ||
103 | RoleARN string | ||
104 | RoleSessionName string | ||
105 | ExternalID string | ||
106 | } | ||
107 | |||
108 | // NewSTSAssumeRole returns a pointer to a new | ||
109 | // Credentials object wrapping the STSAssumeRole. | ||
110 | func NewSTSAssumeRole(stsEndpoint string, opts STSAssumeRoleOptions) (*Credentials, error) { | ||
111 | if stsEndpoint == "" { | ||
112 | return nil, errors.New("STS endpoint cannot be empty") | ||
113 | } | ||
114 | if opts.AccessKey == "" || opts.SecretKey == "" { | ||
115 | return nil, errors.New("AssumeRole credentials access/secretkey is mandatory") | ||
116 | } | ||
117 | return New(&STSAssumeRole{ | ||
118 | Client: &http.Client{ | ||
119 | Transport: http.DefaultTransport, | ||
120 | }, | ||
121 | STSEndpoint: stsEndpoint, | ||
122 | Options: opts, | ||
123 | }), nil | ||
124 | } | ||
125 | |||
126 | const defaultDurationSeconds = 3600 | ||
127 | |||
128 | // closeResponse close non nil response with any response Body. | ||
129 | // convenient wrapper to drain any remaining data on response body. | ||
130 | // | ||
131 | // Subsequently this allows golang http RoundTripper | ||
132 | // to re-use the same connection for future requests. | ||
133 | func closeResponse(resp *http.Response) { | ||
134 | // Callers should close resp.Body when done reading from it. | ||
135 | // If resp.Body is not closed, the Client's underlying RoundTripper | ||
136 | // (typically Transport) may not be able to re-use a persistent TCP | ||
137 | // connection to the server for a subsequent "keep-alive" request. | ||
138 | if resp != nil && resp.Body != nil { | ||
139 | // Drain any remaining Body and then close the connection. | ||
140 | // Without this closing connection would disallow re-using | ||
141 | // the same connection for future uses. | ||
142 | // - http://stackoverflow.com/a/17961593/4465767 | ||
143 | io.Copy(io.Discard, resp.Body) | ||
144 | resp.Body.Close() | ||
145 | } | ||
146 | } | ||
147 | |||
148 | func getAssumeRoleCredentials(clnt *http.Client, endpoint string, opts STSAssumeRoleOptions) (AssumeRoleResponse, error) { | ||
149 | v := url.Values{} | ||
150 | v.Set("Action", "AssumeRole") | ||
151 | v.Set("Version", STSVersion) | ||
152 | if opts.RoleARN != "" { | ||
153 | v.Set("RoleArn", opts.RoleARN) | ||
154 | } | ||
155 | if opts.RoleSessionName != "" { | ||
156 | v.Set("RoleSessionName", opts.RoleSessionName) | ||
157 | } | ||
158 | if opts.DurationSeconds > defaultDurationSeconds { | ||
159 | v.Set("DurationSeconds", strconv.Itoa(opts.DurationSeconds)) | ||
160 | } else { | ||
161 | v.Set("DurationSeconds", strconv.Itoa(defaultDurationSeconds)) | ||
162 | } | ||
163 | if opts.Policy != "" { | ||
164 | v.Set("Policy", opts.Policy) | ||
165 | } | ||
166 | if opts.ExternalID != "" { | ||
167 | v.Set("ExternalId", opts.ExternalID) | ||
168 | } | ||
169 | |||
170 | u, err := url.Parse(endpoint) | ||
171 | if err != nil { | ||
172 | return AssumeRoleResponse{}, err | ||
173 | } | ||
174 | u.Path = "/" | ||
175 | |||
176 | postBody := strings.NewReader(v.Encode()) | ||
177 | hash := sha256.New() | ||
178 | if _, err = io.Copy(hash, postBody); err != nil { | ||
179 | return AssumeRoleResponse{}, err | ||
180 | } | ||
181 | postBody.Seek(0, 0) | ||
182 | |||
183 | req, err := http.NewRequest(http.MethodPost, u.String(), postBody) | ||
184 | if err != nil { | ||
185 | return AssumeRoleResponse{}, err | ||
186 | } | ||
187 | req.Header.Set("Content-Type", "application/x-www-form-urlencoded") | ||
188 | req.Header.Set("X-Amz-Content-Sha256", hex.EncodeToString(hash.Sum(nil))) | ||
189 | if opts.SessionToken != "" { | ||
190 | req.Header.Set("X-Amz-Security-Token", opts.SessionToken) | ||
191 | } | ||
192 | req = signer.SignV4STS(*req, opts.AccessKey, opts.SecretKey, opts.Location) | ||
193 | |||
194 | resp, err := clnt.Do(req) | ||
195 | if err != nil { | ||
196 | return AssumeRoleResponse{}, err | ||
197 | } | ||
198 | defer closeResponse(resp) | ||
199 | if resp.StatusCode != http.StatusOK { | ||
200 | var errResp ErrorResponse | ||
201 | buf, err := io.ReadAll(resp.Body) | ||
202 | if err != nil { | ||
203 | return AssumeRoleResponse{}, err | ||
204 | } | ||
205 | _, err = xmlDecodeAndBody(bytes.NewReader(buf), &errResp) | ||
206 | if err != nil { | ||
207 | var s3Err Error | ||
208 | if _, err = xmlDecodeAndBody(bytes.NewReader(buf), &s3Err); err != nil { | ||
209 | return AssumeRoleResponse{}, err | ||
210 | } | ||
211 | errResp.RequestID = s3Err.RequestID | ||
212 | errResp.STSError.Code = s3Err.Code | ||
213 | errResp.STSError.Message = s3Err.Message | ||
214 | } | ||
215 | return AssumeRoleResponse{}, errResp | ||
216 | } | ||
217 | |||
218 | a := AssumeRoleResponse{} | ||
219 | if _, err = xmlDecodeAndBody(resp.Body, &a); err != nil { | ||
220 | return AssumeRoleResponse{}, err | ||
221 | } | ||
222 | return a, nil | ||
223 | } | ||
224 | |||
225 | // Retrieve retrieves credentials from the MinIO service. | ||
226 | // Error will be returned if the request fails. | ||
227 | func (m *STSAssumeRole) Retrieve() (Value, error) { | ||
228 | a, err := getAssumeRoleCredentials(m.Client, m.STSEndpoint, m.Options) | ||
229 | if err != nil { | ||
230 | return Value{}, err | ||
231 | } | ||
232 | |||
233 | // Expiry window is set to 10secs. | ||
234 | m.SetExpiration(a.Result.Credentials.Expiration, DefaultExpiryWindow) | ||
235 | |||
236 | return Value{ | ||
237 | AccessKeyID: a.Result.Credentials.AccessKey, | ||
238 | SecretAccessKey: a.Result.Credentials.SecretKey, | ||
239 | SessionToken: a.Result.Credentials.SessionToken, | ||
240 | SignerType: SignatureV4, | ||
241 | }, nil | ||
242 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/pkg/credentials/chain.go b/vendor/github.com/minio/minio-go/v7/pkg/credentials/chain.go deleted file mode 100644 index ddccfb1..0000000 --- a/vendor/github.com/minio/minio-go/v7/pkg/credentials/chain.go +++ /dev/null | |||
@@ -1,88 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2017 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package credentials | ||
19 | |||
20 | // A Chain will search for a provider which returns credentials | ||
21 | // and cache that provider until Retrieve is called again. | ||
22 | // | ||
23 | // The Chain provides a way of chaining multiple providers together | ||
24 | // which will pick the first available using priority order of the | ||
25 | // Providers in the list. | ||
26 | // | ||
27 | // If none of the Providers retrieve valid credentials Value, ChainProvider's | ||
28 | // Retrieve() will return the no credentials value. | ||
29 | // | ||
30 | // If a Provider is found which returns valid credentials Value ChainProvider | ||
31 | // will cache that Provider for all calls to IsExpired(), until Retrieve is | ||
32 | // called again after IsExpired() is true. | ||
33 | // | ||
34 | // creds := credentials.NewChainCredentials( | ||
35 | // []credentials.Provider{ | ||
36 | // &credentials.EnvAWSS3{}, | ||
37 | // &credentials.EnvMinio{}, | ||
38 | // }) | ||
39 | // | ||
40 | // // Usage of ChainCredentials. | ||
41 | // mc, err := minio.NewWithCredentials(endpoint, creds, secure, "us-east-1") | ||
42 | // if err != nil { | ||
43 | // log.Fatalln(err) | ||
44 | // } | ||
45 | type Chain struct { | ||
46 | Providers []Provider | ||
47 | curr Provider | ||
48 | } | ||
49 | |||
50 | // NewChainCredentials returns a pointer to a new Credentials object | ||
51 | // wrapping a chain of providers. | ||
52 | func NewChainCredentials(providers []Provider) *Credentials { | ||
53 | return New(&Chain{ | ||
54 | Providers: append([]Provider{}, providers...), | ||
55 | }) | ||
56 | } | ||
57 | |||
58 | // Retrieve returns the credentials value, returns no credentials(anonymous) | ||
59 | // if no credentials provider returned any value. | ||
60 | // | ||
61 | // If a provider is found with credentials, it will be cached and any calls | ||
62 | // to IsExpired() will return the expired state of the cached provider. | ||
63 | func (c *Chain) Retrieve() (Value, error) { | ||
64 | for _, p := range c.Providers { | ||
65 | creds, _ := p.Retrieve() | ||
66 | // Always prioritize non-anonymous providers, if any. | ||
67 | if creds.AccessKeyID == "" && creds.SecretAccessKey == "" { | ||
68 | continue | ||
69 | } | ||
70 | c.curr = p | ||
71 | return creds, nil | ||
72 | } | ||
73 | // At this point we have exhausted all the providers and | ||
74 | // are left without any credentials return anonymous. | ||
75 | return Value{ | ||
76 | SignerType: SignatureAnonymous, | ||
77 | }, nil | ||
78 | } | ||
79 | |||
80 | // IsExpired will returned the expired state of the currently cached provider | ||
81 | // if there is one. If there is no current provider, true will be returned. | ||
82 | func (c *Chain) IsExpired() bool { | ||
83 | if c.curr != nil { | ||
84 | return c.curr.IsExpired() | ||
85 | } | ||
86 | |||
87 | return true | ||
88 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/pkg/credentials/config.json.sample b/vendor/github.com/minio/minio-go/v7/pkg/credentials/config.json.sample deleted file mode 100644 index d793c9e..0000000 --- a/vendor/github.com/minio/minio-go/v7/pkg/credentials/config.json.sample +++ /dev/null | |||
@@ -1,17 +0,0 @@ | |||
1 | { | ||
2 | "version": "8", | ||
3 | "hosts": { | ||
4 | "play": { | ||
5 | "url": "https://play.min.io", | ||
6 | "accessKey": "Q3AM3UQ867SPQQA43P2F", | ||
7 | "secretKey": "zuf+tfteSlswRu7BJ86wekitnifILbZam1KYY3TG", | ||
8 | "api": "S3v2" | ||
9 | }, | ||
10 | "s3": { | ||
11 | "url": "https://s3.amazonaws.com", | ||
12 | "accessKey": "accessKey", | ||
13 | "secretKey": "secret", | ||
14 | "api": "S3v4" | ||
15 | } | ||
16 | } | ||
17 | } \ No newline at end of file | ||
diff --git a/vendor/github.com/minio/minio-go/v7/pkg/credentials/credentials.go b/vendor/github.com/minio/minio-go/v7/pkg/credentials/credentials.go deleted file mode 100644 index af61049..0000000 --- a/vendor/github.com/minio/minio-go/v7/pkg/credentials/credentials.go +++ /dev/null | |||
@@ -1,193 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2017 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package credentials | ||
19 | |||
20 | import ( | ||
21 | "sync" | ||
22 | "time" | ||
23 | ) | ||
24 | |||
25 | const ( | ||
26 | // STSVersion sts version string | ||
27 | STSVersion = "2011-06-15" | ||
28 | |||
29 | // How much duration to slash from the given expiration duration | ||
30 | defaultExpiryWindow = 0.8 | ||
31 | ) | ||
32 | |||
33 | // A Value is the AWS credentials value for individual credential fields. | ||
34 | type Value struct { | ||
35 | // AWS Access key ID | ||
36 | AccessKeyID string | ||
37 | |||
38 | // AWS Secret Access Key | ||
39 | SecretAccessKey string | ||
40 | |||
41 | // AWS Session Token | ||
42 | SessionToken string | ||
43 | |||
44 | // Signature Type. | ||
45 | SignerType SignatureType | ||
46 | } | ||
47 | |||
48 | // A Provider is the interface for any component which will provide credentials | ||
49 | // Value. A provider is required to manage its own Expired state, and what to | ||
50 | // be expired means. | ||
51 | type Provider interface { | ||
52 | // Retrieve returns nil if it successfully retrieved the value. | ||
53 | // Error is returned if the value were not obtainable, or empty. | ||
54 | Retrieve() (Value, error) | ||
55 | |||
56 | // IsExpired returns if the credentials are no longer valid, and need | ||
57 | // to be retrieved. | ||
58 | IsExpired() bool | ||
59 | } | ||
60 | |||
61 | // A Expiry provides shared expiration logic to be used by credentials | ||
62 | // providers to implement expiry functionality. | ||
63 | // | ||
64 | // The best method to use this struct is as an anonymous field within the | ||
65 | // provider's struct. | ||
66 | // | ||
67 | // Example: | ||
68 | // | ||
69 | // type IAMCredentialProvider struct { | ||
70 | // Expiry | ||
71 | // ... | ||
72 | // } | ||
73 | type Expiry struct { | ||
74 | // The date/time when to expire on | ||
75 | expiration time.Time | ||
76 | |||
77 | // If set will be used by IsExpired to determine the current time. | ||
78 | // Defaults to time.Now if CurrentTime is not set. | ||
79 | CurrentTime func() time.Time | ||
80 | } | ||
81 | |||
82 | // SetExpiration sets the expiration IsExpired will check when called. | ||
83 | // | ||
84 | // If window is greater than 0 the expiration time will be reduced by the | ||
85 | // window value. | ||
86 | // | ||
87 | // Using a window is helpful to trigger credentials to expire sooner than | ||
88 | // the expiration time given to ensure no requests are made with expired | ||
89 | // tokens. | ||
90 | func (e *Expiry) SetExpiration(expiration time.Time, window time.Duration) { | ||
91 | if e.CurrentTime == nil { | ||
92 | e.CurrentTime = time.Now | ||
93 | } | ||
94 | cut := window | ||
95 | if cut < 0 { | ||
96 | expireIn := expiration.Sub(e.CurrentTime()) | ||
97 | cut = time.Duration(float64(expireIn) * (1 - defaultExpiryWindow)) | ||
98 | } | ||
99 | e.expiration = expiration.Add(-cut) | ||
100 | } | ||
101 | |||
102 | // IsExpired returns if the credentials are expired. | ||
103 | func (e *Expiry) IsExpired() bool { | ||
104 | if e.CurrentTime == nil { | ||
105 | e.CurrentTime = time.Now | ||
106 | } | ||
107 | return e.expiration.Before(e.CurrentTime()) | ||
108 | } | ||
109 | |||
110 | // Credentials - A container for synchronous safe retrieval of credentials Value. | ||
111 | // Credentials will cache the credentials value until they expire. Once the value | ||
112 | // expires the next Get will attempt to retrieve valid credentials. | ||
113 | // | ||
114 | // Credentials is safe to use across multiple goroutines and will manage the | ||
115 | // synchronous state so the Providers do not need to implement their own | ||
116 | // synchronization. | ||
117 | // | ||
118 | // The first Credentials.Get() will always call Provider.Retrieve() to get the | ||
119 | // first instance of the credentials Value. All calls to Get() after that | ||
120 | // will return the cached credentials Value until IsExpired() returns true. | ||
121 | type Credentials struct { | ||
122 | sync.Mutex | ||
123 | |||
124 | creds Value | ||
125 | forceRefresh bool | ||
126 | provider Provider | ||
127 | } | ||
128 | |||
129 | // New returns a pointer to a new Credentials with the provider set. | ||
130 | func New(provider Provider) *Credentials { | ||
131 | return &Credentials{ | ||
132 | provider: provider, | ||
133 | forceRefresh: true, | ||
134 | } | ||
135 | } | ||
136 | |||
137 | // Get returns the credentials value, or error if the credentials Value failed | ||
138 | // to be retrieved. | ||
139 | // | ||
140 | // Will return the cached credentials Value if it has not expired. If the | ||
141 | // credentials Value has expired the Provider's Retrieve() will be called | ||
142 | // to refresh the credentials. | ||
143 | // | ||
144 | // If Credentials.Expire() was called the credentials Value will be force | ||
145 | // expired, and the next call to Get() will cause them to be refreshed. | ||
146 | func (c *Credentials) Get() (Value, error) { | ||
147 | if c == nil { | ||
148 | return Value{}, nil | ||
149 | } | ||
150 | |||
151 | c.Lock() | ||
152 | defer c.Unlock() | ||
153 | |||
154 | if c.isExpired() { | ||
155 | creds, err := c.provider.Retrieve() | ||
156 | if err != nil { | ||
157 | return Value{}, err | ||
158 | } | ||
159 | c.creds = creds | ||
160 | c.forceRefresh = false | ||
161 | } | ||
162 | |||
163 | return c.creds, nil | ||
164 | } | ||
165 | |||
166 | // Expire expires the credentials and forces them to be retrieved on the | ||
167 | // next call to Get(). | ||
168 | // | ||
169 | // This will override the Provider's expired state, and force Credentials | ||
170 | // to call the Provider's Retrieve(). | ||
171 | func (c *Credentials) Expire() { | ||
172 | c.Lock() | ||
173 | defer c.Unlock() | ||
174 | |||
175 | c.forceRefresh = true | ||
176 | } | ||
177 | |||
178 | // IsExpired returns if the credentials are no longer valid, and need | ||
179 | // to be refreshed. | ||
180 | // | ||
181 | // If the Credentials were forced to be expired with Expire() this will | ||
182 | // reflect that override. | ||
183 | func (c *Credentials) IsExpired() bool { | ||
184 | c.Lock() | ||
185 | defer c.Unlock() | ||
186 | |||
187 | return c.isExpired() | ||
188 | } | ||
189 | |||
190 | // isExpired helper method wrapping the definition of expired credentials. | ||
191 | func (c *Credentials) isExpired() bool { | ||
192 | return c.forceRefresh || c.provider.IsExpired() | ||
193 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/pkg/credentials/credentials.json b/vendor/github.com/minio/minio-go/v7/pkg/credentials/credentials.json deleted file mode 100644 index afbfad5..0000000 --- a/vendor/github.com/minio/minio-go/v7/pkg/credentials/credentials.json +++ /dev/null | |||
@@ -1,7 +0,0 @@ | |||
1 | { | ||
2 | "Version": 1, | ||
3 | "SessionToken": "token", | ||
4 | "AccessKeyId": "accessKey", | ||
5 | "SecretAccessKey": "secret", | ||
6 | "Expiration": "9999-04-27T16:02:25.000Z" | ||
7 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/pkg/credentials/credentials.sample b/vendor/github.com/minio/minio-go/v7/pkg/credentials/credentials.sample deleted file mode 100644 index e2dc1bf..0000000 --- a/vendor/github.com/minio/minio-go/v7/pkg/credentials/credentials.sample +++ /dev/null | |||
@@ -1,15 +0,0 @@ | |||
1 | [default] | ||
2 | aws_access_key_id = accessKey | ||
3 | aws_secret_access_key = secret | ||
4 | aws_session_token = token | ||
5 | |||
6 | [no_token] | ||
7 | aws_access_key_id = accessKey | ||
8 | aws_secret_access_key = secret | ||
9 | |||
10 | [with_colon] | ||
11 | aws_access_key_id: accessKey | ||
12 | aws_secret_access_key: secret | ||
13 | |||
14 | [with_process] | ||
15 | credential_process = /bin/cat credentials.json | ||
diff --git a/vendor/github.com/minio/minio-go/v7/pkg/credentials/doc.go b/vendor/github.com/minio/minio-go/v7/pkg/credentials/doc.go deleted file mode 100644 index fbfb105..0000000 --- a/vendor/github.com/minio/minio-go/v7/pkg/credentials/doc.go +++ /dev/null | |||
@@ -1,60 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2017 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | // Package credentials provides credential retrieval and management | ||
19 | // for S3 compatible object storage. | ||
20 | // | ||
21 | // By default the Credentials.Get() will cache the successful result of a | ||
22 | // Provider's Retrieve() until Provider.IsExpired() returns true. At which | ||
23 | // point Credentials will call Provider's Retrieve() to get new credential Value. | ||
24 | // | ||
25 | // The Provider is responsible for determining when credentials have expired. | ||
26 | // It is also important to note that Credentials will always call Retrieve the | ||
27 | // first time Credentials.Get() is called. | ||
28 | // | ||
29 | // Example of using the environment variable credentials. | ||
30 | // | ||
31 | // creds := NewFromEnv() | ||
32 | // // Retrieve the credentials value | ||
33 | // credValue, err := creds.Get() | ||
34 | // if err != nil { | ||
35 | // // handle error | ||
36 | // } | ||
37 | // | ||
38 | // Example of forcing credentials to expire and be refreshed on the next Get(). | ||
39 | // This may be helpful to proactively expire credentials and refresh them sooner | ||
40 | // than they would naturally expire on their own. | ||
41 | // | ||
42 | // creds := NewFromIAM("") | ||
43 | // creds.Expire() | ||
44 | // credsValue, err := creds.Get() | ||
45 | // // New credentials will be retrieved instead of from cache. | ||
46 | // | ||
47 | // # Custom Provider | ||
48 | // | ||
49 | // Each Provider built into this package also provides a helper method to generate | ||
50 | // a Credentials pointer setup with the provider. To use a custom Provider just | ||
51 | // create a type which satisfies the Provider interface and pass it to the | ||
52 | // NewCredentials method. | ||
53 | // | ||
54 | // type MyProvider struct{} | ||
55 | // func (m *MyProvider) Retrieve() (Value, error) {...} | ||
56 | // func (m *MyProvider) IsExpired() bool {...} | ||
57 | // | ||
58 | // creds := NewCredentials(&MyProvider{}) | ||
59 | // credValue, err := creds.Get() | ||
60 | package credentials | ||
diff --git a/vendor/github.com/minio/minio-go/v7/pkg/credentials/env_aws.go b/vendor/github.com/minio/minio-go/v7/pkg/credentials/env_aws.go deleted file mode 100644 index b6e60d0..0000000 --- a/vendor/github.com/minio/minio-go/v7/pkg/credentials/env_aws.go +++ /dev/null | |||
@@ -1,71 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2017 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package credentials | ||
19 | |||
20 | import "os" | ||
21 | |||
22 | // A EnvAWS retrieves credentials from the environment variables of the | ||
23 | // running process. EnvAWSironment credentials never expire. | ||
24 | // | ||
25 | // EnvAWSironment variables used: | ||
26 | // | ||
27 | // * Access Key ID: AWS_ACCESS_KEY_ID or AWS_ACCESS_KEY. | ||
28 | // * Secret Access Key: AWS_SECRET_ACCESS_KEY or AWS_SECRET_KEY. | ||
29 | // * Secret Token: AWS_SESSION_TOKEN. | ||
30 | type EnvAWS struct { | ||
31 | retrieved bool | ||
32 | } | ||
33 | |||
34 | // NewEnvAWS returns a pointer to a new Credentials object | ||
35 | // wrapping the environment variable provider. | ||
36 | func NewEnvAWS() *Credentials { | ||
37 | return New(&EnvAWS{}) | ||
38 | } | ||
39 | |||
40 | // Retrieve retrieves the keys from the environment. | ||
41 | func (e *EnvAWS) Retrieve() (Value, error) { | ||
42 | e.retrieved = false | ||
43 | |||
44 | id := os.Getenv("AWS_ACCESS_KEY_ID") | ||
45 | if id == "" { | ||
46 | id = os.Getenv("AWS_ACCESS_KEY") | ||
47 | } | ||
48 | |||
49 | secret := os.Getenv("AWS_SECRET_ACCESS_KEY") | ||
50 | if secret == "" { | ||
51 | secret = os.Getenv("AWS_SECRET_KEY") | ||
52 | } | ||
53 | |||
54 | signerType := SignatureV4 | ||
55 | if id == "" || secret == "" { | ||
56 | signerType = SignatureAnonymous | ||
57 | } | ||
58 | |||
59 | e.retrieved = true | ||
60 | return Value{ | ||
61 | AccessKeyID: id, | ||
62 | SecretAccessKey: secret, | ||
63 | SessionToken: os.Getenv("AWS_SESSION_TOKEN"), | ||
64 | SignerType: signerType, | ||
65 | }, nil | ||
66 | } | ||
67 | |||
68 | // IsExpired returns if the credentials have been retrieved. | ||
69 | func (e *EnvAWS) IsExpired() bool { | ||
70 | return !e.retrieved | ||
71 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/pkg/credentials/env_minio.go b/vendor/github.com/minio/minio-go/v7/pkg/credentials/env_minio.go deleted file mode 100644 index 5bfeab1..0000000 --- a/vendor/github.com/minio/minio-go/v7/pkg/credentials/env_minio.go +++ /dev/null | |||
@@ -1,68 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2017 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package credentials | ||
19 | |||
20 | import "os" | ||
21 | |||
22 | // A EnvMinio retrieves credentials from the environment variables of the | ||
23 | // running process. EnvMinioironment credentials never expire. | ||
24 | // | ||
25 | // Environment variables used: | ||
26 | // | ||
27 | // * Access Key ID: MINIO_ACCESS_KEY. | ||
28 | // * Secret Access Key: MINIO_SECRET_KEY. | ||
29 | // * Access Key ID: MINIO_ROOT_USER. | ||
30 | // * Secret Access Key: MINIO_ROOT_PASSWORD. | ||
31 | type EnvMinio struct { | ||
32 | retrieved bool | ||
33 | } | ||
34 | |||
35 | // NewEnvMinio returns a pointer to a new Credentials object | ||
36 | // wrapping the environment variable provider. | ||
37 | func NewEnvMinio() *Credentials { | ||
38 | return New(&EnvMinio{}) | ||
39 | } | ||
40 | |||
41 | // Retrieve retrieves the keys from the environment. | ||
42 | func (e *EnvMinio) Retrieve() (Value, error) { | ||
43 | e.retrieved = false | ||
44 | |||
45 | id := os.Getenv("MINIO_ROOT_USER") | ||
46 | secret := os.Getenv("MINIO_ROOT_PASSWORD") | ||
47 | |||
48 | signerType := SignatureV4 | ||
49 | if id == "" || secret == "" { | ||
50 | id = os.Getenv("MINIO_ACCESS_KEY") | ||
51 | secret = os.Getenv("MINIO_SECRET_KEY") | ||
52 | if id == "" || secret == "" { | ||
53 | signerType = SignatureAnonymous | ||
54 | } | ||
55 | } | ||
56 | |||
57 | e.retrieved = true | ||
58 | return Value{ | ||
59 | AccessKeyID: id, | ||
60 | SecretAccessKey: secret, | ||
61 | SignerType: signerType, | ||
62 | }, nil | ||
63 | } | ||
64 | |||
65 | // IsExpired returns if the credentials have been retrieved. | ||
66 | func (e *EnvMinio) IsExpired() bool { | ||
67 | return !e.retrieved | ||
68 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/pkg/credentials/error_response.go b/vendor/github.com/minio/minio-go/v7/pkg/credentials/error_response.go deleted file mode 100644 index 07a9c2f..0000000 --- a/vendor/github.com/minio/minio-go/v7/pkg/credentials/error_response.go +++ /dev/null | |||
@@ -1,95 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2021 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package credentials | ||
19 | |||
20 | import ( | ||
21 | "bytes" | ||
22 | "encoding/xml" | ||
23 | "fmt" | ||
24 | "io" | ||
25 | ) | ||
26 | |||
27 | // ErrorResponse - Is the typed error returned. | ||
28 | // ErrorResponse struct should be comparable since it is compared inside | ||
29 | // golang http API (https://github.com/golang/go/issues/29768) | ||
30 | type ErrorResponse struct { | ||
31 | XMLName xml.Name `xml:"https://sts.amazonaws.com/doc/2011-06-15/ ErrorResponse" json:"-"` | ||
32 | STSError struct { | ||
33 | Type string `xml:"Type"` | ||
34 | Code string `xml:"Code"` | ||
35 | Message string `xml:"Message"` | ||
36 | } `xml:"Error"` | ||
37 | RequestID string `xml:"RequestId"` | ||
38 | } | ||
39 | |||
40 | // Error - Is the typed error returned by all API operations. | ||
41 | type Error struct { | ||
42 | XMLName xml.Name `xml:"Error" json:"-"` | ||
43 | Code string | ||
44 | Message string | ||
45 | BucketName string | ||
46 | Key string | ||
47 | Resource string | ||
48 | RequestID string `xml:"RequestId"` | ||
49 | HostID string `xml:"HostId"` | ||
50 | |||
51 | // Region where the bucket is located. This header is returned | ||
52 | // only in HEAD bucket and ListObjects response. | ||
53 | Region string | ||
54 | |||
55 | // Captures the server string returned in response header. | ||
56 | Server string | ||
57 | |||
58 | // Underlying HTTP status code for the returned error | ||
59 | StatusCode int `xml:"-" json:"-"` | ||
60 | } | ||
61 | |||
62 | // Error - Returns S3 error string. | ||
63 | func (e Error) Error() string { | ||
64 | if e.Message == "" { | ||
65 | return fmt.Sprintf("Error response code %s.", e.Code) | ||
66 | } | ||
67 | return e.Message | ||
68 | } | ||
69 | |||
70 | // Error - Returns STS error string. | ||
71 | func (e ErrorResponse) Error() string { | ||
72 | if e.STSError.Message == "" { | ||
73 | return fmt.Sprintf("Error response code %s.", e.STSError.Code) | ||
74 | } | ||
75 | return e.STSError.Message | ||
76 | } | ||
77 | |||
78 | // xmlDecoder provide decoded value in xml. | ||
79 | func xmlDecoder(body io.Reader, v interface{}) error { | ||
80 | d := xml.NewDecoder(body) | ||
81 | return d.Decode(v) | ||
82 | } | ||
83 | |||
84 | // xmlDecodeAndBody reads the whole body up to 1MB and | ||
85 | // tries to XML decode it into v. | ||
86 | // The body that was read and any error from reading or decoding is returned. | ||
87 | func xmlDecodeAndBody(bodyReader io.Reader, v interface{}) ([]byte, error) { | ||
88 | // read the whole body (up to 1MB) | ||
89 | const maxBodyLength = 1 << 20 | ||
90 | body, err := io.ReadAll(io.LimitReader(bodyReader, maxBodyLength)) | ||
91 | if err != nil { | ||
92 | return nil, err | ||
93 | } | ||
94 | return bytes.TrimSpace(body), xmlDecoder(bytes.NewReader(body), v) | ||
95 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/pkg/credentials/file_aws_credentials.go b/vendor/github.com/minio/minio-go/v7/pkg/credentials/file_aws_credentials.go deleted file mode 100644 index 5b07376..0000000 --- a/vendor/github.com/minio/minio-go/v7/pkg/credentials/file_aws_credentials.go +++ /dev/null | |||
@@ -1,157 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2017 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package credentials | ||
19 | |||
20 | import ( | ||
21 | "encoding/json" | ||
22 | "errors" | ||
23 | "os" | ||
24 | "os/exec" | ||
25 | "path/filepath" | ||
26 | "strings" | ||
27 | "time" | ||
28 | |||
29 | ini "gopkg.in/ini.v1" | ||
30 | ) | ||
31 | |||
32 | // A externalProcessCredentials stores the output of a credential_process | ||
33 | type externalProcessCredentials struct { | ||
34 | Version int | ||
35 | SessionToken string | ||
36 | AccessKeyID string `json:"AccessKeyId"` | ||
37 | SecretAccessKey string | ||
38 | Expiration time.Time | ||
39 | } | ||
40 | |||
41 | // A FileAWSCredentials retrieves credentials from the current user's home | ||
42 | // directory, and keeps track if those credentials are expired. | ||
43 | // | ||
44 | // Profile ini file example: $HOME/.aws/credentials | ||
45 | type FileAWSCredentials struct { | ||
46 | Expiry | ||
47 | |||
48 | // Path to the shared credentials file. | ||
49 | // | ||
50 | // If empty will look for "AWS_SHARED_CREDENTIALS_FILE" env variable. If the | ||
51 | // env value is empty will default to current user's home directory. | ||
52 | // Linux/OSX: "$HOME/.aws/credentials" | ||
53 | // Windows: "%USERPROFILE%\.aws\credentials" | ||
54 | Filename string | ||
55 | |||
56 | // AWS Profile to extract credentials from the shared credentials file. If empty | ||
57 | // will default to environment variable "AWS_PROFILE" or "default" if | ||
58 | // environment variable is also not set. | ||
59 | Profile string | ||
60 | |||
61 | // retrieved states if the credentials have been successfully retrieved. | ||
62 | retrieved bool | ||
63 | } | ||
64 | |||
65 | // NewFileAWSCredentials returns a pointer to a new Credentials object | ||
66 | // wrapping the Profile file provider. | ||
67 | func NewFileAWSCredentials(filename, profile string) *Credentials { | ||
68 | return New(&FileAWSCredentials{ | ||
69 | Filename: filename, | ||
70 | Profile: profile, | ||
71 | }) | ||
72 | } | ||
73 | |||
74 | // Retrieve reads and extracts the shared credentials from the current | ||
75 | // users home directory. | ||
76 | func (p *FileAWSCredentials) Retrieve() (Value, error) { | ||
77 | if p.Filename == "" { | ||
78 | p.Filename = os.Getenv("AWS_SHARED_CREDENTIALS_FILE") | ||
79 | if p.Filename == "" { | ||
80 | homeDir, err := os.UserHomeDir() | ||
81 | if err != nil { | ||
82 | return Value{}, err | ||
83 | } | ||
84 | p.Filename = filepath.Join(homeDir, ".aws", "credentials") | ||
85 | } | ||
86 | } | ||
87 | if p.Profile == "" { | ||
88 | p.Profile = os.Getenv("AWS_PROFILE") | ||
89 | if p.Profile == "" { | ||
90 | p.Profile = "default" | ||
91 | } | ||
92 | } | ||
93 | |||
94 | p.retrieved = false | ||
95 | |||
96 | iniProfile, err := loadProfile(p.Filename, p.Profile) | ||
97 | if err != nil { | ||
98 | return Value{}, err | ||
99 | } | ||
100 | |||
101 | // Default to empty string if not found. | ||
102 | id := iniProfile.Key("aws_access_key_id") | ||
103 | // Default to empty string if not found. | ||
104 | secret := iniProfile.Key("aws_secret_access_key") | ||
105 | // Default to empty string if not found. | ||
106 | token := iniProfile.Key("aws_session_token") | ||
107 | |||
108 | // If credential_process is defined, obtain credentials by executing | ||
109 | // the external process | ||
110 | credentialProcess := strings.TrimSpace(iniProfile.Key("credential_process").String()) | ||
111 | if credentialProcess != "" { | ||
112 | args := strings.Fields(credentialProcess) | ||
113 | if len(args) <= 1 { | ||
114 | return Value{}, errors.New("invalid credential process args") | ||
115 | } | ||
116 | cmd := exec.Command(args[0], args[1:]...) | ||
117 | out, err := cmd.Output() | ||
118 | if err != nil { | ||
119 | return Value{}, err | ||
120 | } | ||
121 | var externalProcessCredentials externalProcessCredentials | ||
122 | err = json.Unmarshal([]byte(out), &externalProcessCredentials) | ||
123 | if err != nil { | ||
124 | return Value{}, err | ||
125 | } | ||
126 | p.retrieved = true | ||
127 | p.SetExpiration(externalProcessCredentials.Expiration, DefaultExpiryWindow) | ||
128 | return Value{ | ||
129 | AccessKeyID: externalProcessCredentials.AccessKeyID, | ||
130 | SecretAccessKey: externalProcessCredentials.SecretAccessKey, | ||
131 | SessionToken: externalProcessCredentials.SessionToken, | ||
132 | SignerType: SignatureV4, | ||
133 | }, nil | ||
134 | } | ||
135 | p.retrieved = true | ||
136 | return Value{ | ||
137 | AccessKeyID: id.String(), | ||
138 | SecretAccessKey: secret.String(), | ||
139 | SessionToken: token.String(), | ||
140 | SignerType: SignatureV4, | ||
141 | }, nil | ||
142 | } | ||
143 | |||
144 | // loadProfiles loads from the file pointed to by shared credentials filename for profile. | ||
145 | // The credentials retrieved from the profile will be returned or error. Error will be | ||
146 | // returned if it fails to read from the file, or the data is invalid. | ||
147 | func loadProfile(filename, profile string) (*ini.Section, error) { | ||
148 | config, err := ini.Load(filename) | ||
149 | if err != nil { | ||
150 | return nil, err | ||
151 | } | ||
152 | iniProfile, err := config.GetSection(profile) | ||
153 | if err != nil { | ||
154 | return nil, err | ||
155 | } | ||
156 | return iniProfile, nil | ||
157 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/pkg/credentials/file_minio_client.go b/vendor/github.com/minio/minio-go/v7/pkg/credentials/file_minio_client.go deleted file mode 100644 index eb77767..0000000 --- a/vendor/github.com/minio/minio-go/v7/pkg/credentials/file_minio_client.go +++ /dev/null | |||
@@ -1,139 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2017 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package credentials | ||
19 | |||
20 | import ( | ||
21 | "os" | ||
22 | "path/filepath" | ||
23 | "runtime" | ||
24 | |||
25 | jsoniter "github.com/json-iterator/go" | ||
26 | ) | ||
27 | |||
28 | // A FileMinioClient retrieves credentials from the current user's home | ||
29 | // directory, and keeps track if those credentials are expired. | ||
30 | // | ||
31 | // Configuration file example: $HOME/.mc/config.json | ||
32 | type FileMinioClient struct { | ||
33 | // Path to the shared credentials file. | ||
34 | // | ||
35 | // If empty will look for "MINIO_SHARED_CREDENTIALS_FILE" env variable. If the | ||
36 | // env value is empty will default to current user's home directory. | ||
37 | // Linux/OSX: "$HOME/.mc/config.json" | ||
38 | // Windows: "%USERALIAS%\mc\config.json" | ||
39 | Filename string | ||
40 | |||
41 | // MinIO Alias to extract credentials from the shared credentials file. If empty | ||
42 | // will default to environment variable "MINIO_ALIAS" or "default" if | ||
43 | // environment variable is also not set. | ||
44 | Alias string | ||
45 | |||
46 | // retrieved states if the credentials have been successfully retrieved. | ||
47 | retrieved bool | ||
48 | } | ||
49 | |||
50 | // NewFileMinioClient returns a pointer to a new Credentials object | ||
51 | // wrapping the Alias file provider. | ||
52 | func NewFileMinioClient(filename, alias string) *Credentials { | ||
53 | return New(&FileMinioClient{ | ||
54 | Filename: filename, | ||
55 | Alias: alias, | ||
56 | }) | ||
57 | } | ||
58 | |||
59 | // Retrieve reads and extracts the shared credentials from the current | ||
60 | // users home directory. | ||
61 | func (p *FileMinioClient) Retrieve() (Value, error) { | ||
62 | if p.Filename == "" { | ||
63 | if value, ok := os.LookupEnv("MINIO_SHARED_CREDENTIALS_FILE"); ok { | ||
64 | p.Filename = value | ||
65 | } else { | ||
66 | homeDir, err := os.UserHomeDir() | ||
67 | if err != nil { | ||
68 | return Value{}, err | ||
69 | } | ||
70 | p.Filename = filepath.Join(homeDir, ".mc", "config.json") | ||
71 | if runtime.GOOS == "windows" { | ||
72 | p.Filename = filepath.Join(homeDir, "mc", "config.json") | ||
73 | } | ||
74 | } | ||
75 | } | ||
76 | |||
77 | if p.Alias == "" { | ||
78 | p.Alias = os.Getenv("MINIO_ALIAS") | ||
79 | if p.Alias == "" { | ||
80 | p.Alias = "s3" | ||
81 | } | ||
82 | } | ||
83 | |||
84 | p.retrieved = false | ||
85 | |||
86 | hostCfg, err := loadAlias(p.Filename, p.Alias) | ||
87 | if err != nil { | ||
88 | return Value{}, err | ||
89 | } | ||
90 | |||
91 | p.retrieved = true | ||
92 | return Value{ | ||
93 | AccessKeyID: hostCfg.AccessKey, | ||
94 | SecretAccessKey: hostCfg.SecretKey, | ||
95 | SignerType: parseSignatureType(hostCfg.API), | ||
96 | }, nil | ||
97 | } | ||
98 | |||
99 | // IsExpired returns if the shared credentials have expired. | ||
100 | func (p *FileMinioClient) IsExpired() bool { | ||
101 | return !p.retrieved | ||
102 | } | ||
103 | |||
104 | // hostConfig configuration of a host. | ||
105 | type hostConfig struct { | ||
106 | URL string `json:"url"` | ||
107 | AccessKey string `json:"accessKey"` | ||
108 | SecretKey string `json:"secretKey"` | ||
109 | API string `json:"api"` | ||
110 | } | ||
111 | |||
112 | // config config version. | ||
113 | type config struct { | ||
114 | Version string `json:"version"` | ||
115 | Hosts map[string]hostConfig `json:"hosts"` | ||
116 | Aliases map[string]hostConfig `json:"aliases"` | ||
117 | } | ||
118 | |||
119 | // loadAliass loads from the file pointed to by shared credentials filename for alias. | ||
120 | // The credentials retrieved from the alias will be returned or error. Error will be | ||
121 | // returned if it fails to read from the file. | ||
122 | func loadAlias(filename, alias string) (hostConfig, error) { | ||
123 | cfg := &config{} | ||
124 | json := jsoniter.ConfigCompatibleWithStandardLibrary | ||
125 | |||
126 | configBytes, err := os.ReadFile(filename) | ||
127 | if err != nil { | ||
128 | return hostConfig{}, err | ||
129 | } | ||
130 | if err = json.Unmarshal(configBytes, cfg); err != nil { | ||
131 | return hostConfig{}, err | ||
132 | } | ||
133 | |||
134 | if cfg.Version == "10" { | ||
135 | return cfg.Aliases[alias], nil | ||
136 | } | ||
137 | |||
138 | return cfg.Hosts[alias], nil | ||
139 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/pkg/credentials/iam_aws.go b/vendor/github.com/minio/minio-go/v7/pkg/credentials/iam_aws.go deleted file mode 100644 index c5153c4..0000000 --- a/vendor/github.com/minio/minio-go/v7/pkg/credentials/iam_aws.go +++ /dev/null | |||
@@ -1,433 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2017 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package credentials | ||
19 | |||
20 | import ( | ||
21 | "bufio" | ||
22 | "context" | ||
23 | "errors" | ||
24 | "fmt" | ||
25 | "io" | ||
26 | "net" | ||
27 | "net/http" | ||
28 | "net/url" | ||
29 | "os" | ||
30 | "path" | ||
31 | "strings" | ||
32 | "time" | ||
33 | |||
34 | jsoniter "github.com/json-iterator/go" | ||
35 | ) | ||
36 | |||
37 | // DefaultExpiryWindow - Default expiry window. | ||
38 | // ExpiryWindow will allow the credentials to trigger refreshing | ||
39 | // prior to the credentials actually expiring. This is beneficial | ||
40 | // so race conditions with expiring credentials do not cause | ||
41 | // request to fail unexpectedly due to ExpiredTokenException exceptions. | ||
42 | // DefaultExpiryWindow can be used as parameter to (*Expiry).SetExpiration. | ||
43 | // When used the tokens refresh will be triggered when 80% of the elapsed | ||
44 | // time until the actual expiration time is passed. | ||
45 | const DefaultExpiryWindow = -1 | ||
46 | |||
47 | // A IAM retrieves credentials from the EC2 service, and keeps track if | ||
48 | // those credentials are expired. | ||
49 | type IAM struct { | ||
50 | Expiry | ||
51 | |||
52 | // Required http Client to use when connecting to IAM metadata service. | ||
53 | Client *http.Client | ||
54 | |||
55 | // Custom endpoint to fetch IAM role credentials. | ||
56 | Endpoint string | ||
57 | |||
58 | // Region configurable custom region for STS | ||
59 | Region string | ||
60 | |||
61 | // Support for container authorization token https://docs.aws.amazon.com/sdkref/latest/guide/feature-container-credentials.html | ||
62 | Container struct { | ||
63 | AuthorizationToken string | ||
64 | CredentialsFullURI string | ||
65 | CredentialsRelativeURI string | ||
66 | } | ||
67 | |||
68 | // EKS based k8s RBAC authorization - https://docs.aws.amazon.com/eks/latest/userguide/pod-configuration.html | ||
69 | EKSIdentity struct { | ||
70 | TokenFile string | ||
71 | RoleARN string | ||
72 | RoleSessionName string | ||
73 | } | ||
74 | } | ||
75 | |||
76 | // IAM Roles for Amazon EC2 | ||
77 | // http://docs.aws.amazon.com/AWSEC2/latest/UserGuide/iam-roles-for-amazon-ec2.html | ||
78 | const ( | ||
79 | DefaultIAMRoleEndpoint = "http://169.254.169.254" | ||
80 | DefaultECSRoleEndpoint = "http://169.254.170.2" | ||
81 | DefaultSTSRoleEndpoint = "https://sts.amazonaws.com" | ||
82 | DefaultIAMSecurityCredsPath = "/latest/meta-data/iam/security-credentials/" | ||
83 | TokenRequestTTLHeader = "X-aws-ec2-metadata-token-ttl-seconds" | ||
84 | TokenPath = "/latest/api/token" | ||
85 | TokenTTL = "21600" | ||
86 | TokenRequestHeader = "X-aws-ec2-metadata-token" | ||
87 | ) | ||
88 | |||
89 | // NewIAM returns a pointer to a new Credentials object wrapping the IAM. | ||
90 | func NewIAM(endpoint string) *Credentials { | ||
91 | return New(&IAM{ | ||
92 | Client: &http.Client{ | ||
93 | Transport: http.DefaultTransport, | ||
94 | }, | ||
95 | Endpoint: endpoint, | ||
96 | }) | ||
97 | } | ||
98 | |||
99 | // Retrieve retrieves credentials from the EC2 service. | ||
100 | // Error will be returned if the request fails, or unable to extract | ||
101 | // the desired | ||
102 | func (m *IAM) Retrieve() (Value, error) { | ||
103 | token := os.Getenv("AWS_CONTAINER_AUTHORIZATION_TOKEN") | ||
104 | if token == "" { | ||
105 | token = m.Container.AuthorizationToken | ||
106 | } | ||
107 | |||
108 | relativeURI := os.Getenv("AWS_CONTAINER_CREDENTIALS_RELATIVE_URI") | ||
109 | if relativeURI == "" { | ||
110 | relativeURI = m.Container.CredentialsRelativeURI | ||
111 | } | ||
112 | |||
113 | fullURI := os.Getenv("AWS_CONTAINER_CREDENTIALS_FULL_URI") | ||
114 | if fullURI == "" { | ||
115 | fullURI = m.Container.CredentialsFullURI | ||
116 | } | ||
117 | |||
118 | identityFile := os.Getenv("AWS_WEB_IDENTITY_TOKEN_FILE") | ||
119 | if identityFile == "" { | ||
120 | identityFile = m.EKSIdentity.TokenFile | ||
121 | } | ||
122 | |||
123 | roleArn := os.Getenv("AWS_ROLE_ARN") | ||
124 | if roleArn == "" { | ||
125 | roleArn = m.EKSIdentity.RoleARN | ||
126 | } | ||
127 | |||
128 | roleSessionName := os.Getenv("AWS_ROLE_SESSION_NAME") | ||
129 | if roleSessionName == "" { | ||
130 | roleSessionName = m.EKSIdentity.RoleSessionName | ||
131 | } | ||
132 | |||
133 | region := os.Getenv("AWS_REGION") | ||
134 | if region == "" { | ||
135 | region = m.Region | ||
136 | } | ||
137 | |||
138 | var roleCreds ec2RoleCredRespBody | ||
139 | var err error | ||
140 | |||
141 | endpoint := m.Endpoint | ||
142 | switch { | ||
143 | case identityFile != "": | ||
144 | if len(endpoint) == 0 { | ||
145 | if region != "" { | ||
146 | if strings.HasPrefix(region, "cn-") { | ||
147 | endpoint = "https://sts." + region + ".amazonaws.com.cn" | ||
148 | } else { | ||
149 | endpoint = "https://sts." + region + ".amazonaws.com" | ||
150 | } | ||
151 | } else { | ||
152 | endpoint = DefaultSTSRoleEndpoint | ||
153 | } | ||
154 | } | ||
155 | |||
156 | creds := &STSWebIdentity{ | ||
157 | Client: m.Client, | ||
158 | STSEndpoint: endpoint, | ||
159 | GetWebIDTokenExpiry: func() (*WebIdentityToken, error) { | ||
160 | token, err := os.ReadFile(identityFile) | ||
161 | if err != nil { | ||
162 | return nil, err | ||
163 | } | ||
164 | |||
165 | return &WebIdentityToken{Token: string(token)}, nil | ||
166 | }, | ||
167 | RoleARN: roleArn, | ||
168 | roleSessionName: roleSessionName, | ||
169 | } | ||
170 | |||
171 | stsWebIdentityCreds, err := creds.Retrieve() | ||
172 | if err == nil { | ||
173 | m.SetExpiration(creds.Expiration(), DefaultExpiryWindow) | ||
174 | } | ||
175 | return stsWebIdentityCreds, err | ||
176 | |||
177 | case relativeURI != "": | ||
178 | if len(endpoint) == 0 { | ||
179 | endpoint = fmt.Sprintf("%s%s", DefaultECSRoleEndpoint, relativeURI) | ||
180 | } | ||
181 | |||
182 | roleCreds, err = getEcsTaskCredentials(m.Client, endpoint, token) | ||
183 | |||
184 | case fullURI != "": | ||
185 | if len(endpoint) == 0 { | ||
186 | endpoint = fullURI | ||
187 | var ok bool | ||
188 | if ok, err = isLoopback(endpoint); !ok { | ||
189 | if err == nil { | ||
190 | err = fmt.Errorf("uri host is not a loopback address: %s", endpoint) | ||
191 | } | ||
192 | break | ||
193 | } | ||
194 | } | ||
195 | |||
196 | roleCreds, err = getEcsTaskCredentials(m.Client, endpoint, token) | ||
197 | |||
198 | default: | ||
199 | roleCreds, err = getCredentials(m.Client, endpoint) | ||
200 | } | ||
201 | |||
202 | if err != nil { | ||
203 | return Value{}, err | ||
204 | } | ||
205 | // Expiry window is set to 10secs. | ||
206 | m.SetExpiration(roleCreds.Expiration, DefaultExpiryWindow) | ||
207 | |||
208 | return Value{ | ||
209 | AccessKeyID: roleCreds.AccessKeyID, | ||
210 | SecretAccessKey: roleCreds.SecretAccessKey, | ||
211 | SessionToken: roleCreds.Token, | ||
212 | SignerType: SignatureV4, | ||
213 | }, nil | ||
214 | } | ||
215 | |||
216 | // A ec2RoleCredRespBody provides the shape for unmarshaling credential | ||
217 | // request responses. | ||
218 | type ec2RoleCredRespBody struct { | ||
219 | // Success State | ||
220 | Expiration time.Time | ||
221 | AccessKeyID string | ||
222 | SecretAccessKey string | ||
223 | Token string | ||
224 | |||
225 | // Error state | ||
226 | Code string | ||
227 | Message string | ||
228 | |||
229 | // Unused params. | ||
230 | LastUpdated time.Time | ||
231 | Type string | ||
232 | } | ||
233 | |||
234 | // Get the final IAM role URL where the request will | ||
235 | // be sent to fetch the rolling access credentials. | ||
236 | // http://docs.aws.amazon.com/AWSEC2/latest/UserGuide/iam-roles-for-amazon-ec2.html | ||
237 | func getIAMRoleURL(endpoint string) (*url.URL, error) { | ||
238 | u, err := url.Parse(endpoint) | ||
239 | if err != nil { | ||
240 | return nil, err | ||
241 | } | ||
242 | u.Path = DefaultIAMSecurityCredsPath | ||
243 | return u, nil | ||
244 | } | ||
245 | |||
246 | // listRoleNames lists of credential role names associated | ||
247 | // with the current EC2 service. If there are no credentials, | ||
248 | // or there is an error making or receiving the request. | ||
249 | // http://docs.aws.amazon.com/AWSEC2/latest/UserGuide/iam-roles-for-amazon-ec2.html | ||
250 | func listRoleNames(client *http.Client, u *url.URL, token string) ([]string, error) { | ||
251 | req, err := http.NewRequest(http.MethodGet, u.String(), nil) | ||
252 | if err != nil { | ||
253 | return nil, err | ||
254 | } | ||
255 | if token != "" { | ||
256 | req.Header.Add(TokenRequestHeader, token) | ||
257 | } | ||
258 | resp, err := client.Do(req) | ||
259 | if err != nil { | ||
260 | return nil, err | ||
261 | } | ||
262 | defer resp.Body.Close() | ||
263 | if resp.StatusCode != http.StatusOK { | ||
264 | return nil, errors.New(resp.Status) | ||
265 | } | ||
266 | |||
267 | credsList := []string{} | ||
268 | s := bufio.NewScanner(resp.Body) | ||
269 | for s.Scan() { | ||
270 | credsList = append(credsList, s.Text()) | ||
271 | } | ||
272 | |||
273 | if err := s.Err(); err != nil { | ||
274 | return nil, err | ||
275 | } | ||
276 | |||
277 | return credsList, nil | ||
278 | } | ||
279 | |||
280 | func getEcsTaskCredentials(client *http.Client, endpoint, token string) (ec2RoleCredRespBody, error) { | ||
281 | req, err := http.NewRequest(http.MethodGet, endpoint, nil) | ||
282 | if err != nil { | ||
283 | return ec2RoleCredRespBody{}, err | ||
284 | } | ||
285 | |||
286 | if token != "" { | ||
287 | req.Header.Set("Authorization", token) | ||
288 | } | ||
289 | |||
290 | resp, err := client.Do(req) | ||
291 | if err != nil { | ||
292 | return ec2RoleCredRespBody{}, err | ||
293 | } | ||
294 | defer resp.Body.Close() | ||
295 | if resp.StatusCode != http.StatusOK { | ||
296 | return ec2RoleCredRespBody{}, errors.New(resp.Status) | ||
297 | } | ||
298 | |||
299 | respCreds := ec2RoleCredRespBody{} | ||
300 | if err := jsoniter.NewDecoder(resp.Body).Decode(&respCreds); err != nil { | ||
301 | return ec2RoleCredRespBody{}, err | ||
302 | } | ||
303 | |||
304 | return respCreds, nil | ||
305 | } | ||
306 | |||
307 | func fetchIMDSToken(client *http.Client, endpoint string) (string, error) { | ||
308 | ctx, cancel := context.WithTimeout(context.Background(), time.Second) | ||
309 | defer cancel() | ||
310 | |||
311 | req, err := http.NewRequestWithContext(ctx, http.MethodPut, endpoint+TokenPath, nil) | ||
312 | if err != nil { | ||
313 | return "", err | ||
314 | } | ||
315 | req.Header.Add(TokenRequestTTLHeader, TokenTTL) | ||
316 | resp, err := client.Do(req) | ||
317 | if err != nil { | ||
318 | return "", err | ||
319 | } | ||
320 | defer resp.Body.Close() | ||
321 | data, err := io.ReadAll(resp.Body) | ||
322 | if err != nil { | ||
323 | return "", err | ||
324 | } | ||
325 | if resp.StatusCode != http.StatusOK { | ||
326 | return "", errors.New(resp.Status) | ||
327 | } | ||
328 | return string(data), nil | ||
329 | } | ||
330 | |||
331 | // getCredentials - obtains the credentials from the IAM role name associated with | ||
332 | // the current EC2 service. | ||
333 | // | ||
334 | // If the credentials cannot be found, or there is an error | ||
335 | // reading the response an error will be returned. | ||
336 | func getCredentials(client *http.Client, endpoint string) (ec2RoleCredRespBody, error) { | ||
337 | if endpoint == "" { | ||
338 | endpoint = DefaultIAMRoleEndpoint | ||
339 | } | ||
340 | |||
341 | // https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/configuring-instance-metadata-service.html | ||
342 | token, err := fetchIMDSToken(client, endpoint) | ||
343 | if err != nil { | ||
344 | // Return only errors for valid situations, if the IMDSv2 is not enabled | ||
345 | // we will not be able to get the token, in such a situation we have | ||
346 | // to rely on IMDSv1 behavior as a fallback, this check ensures that. | ||
347 | // Refer https://github.com/minio/minio-go/issues/1866 | ||
348 | if !errors.Is(err, context.DeadlineExceeded) && !errors.Is(err, context.Canceled) { | ||
349 | return ec2RoleCredRespBody{}, err | ||
350 | } | ||
351 | } | ||
352 | |||
353 | // http://docs.aws.amazon.com/AWSEC2/latest/UserGuide/iam-roles-for-amazon-ec2.html | ||
354 | u, err := getIAMRoleURL(endpoint) | ||
355 | if err != nil { | ||
356 | return ec2RoleCredRespBody{}, err | ||
357 | } | ||
358 | |||
359 | // http://docs.aws.amazon.com/AWSEC2/latest/UserGuide/iam-roles-for-amazon-ec2.html | ||
360 | roleNames, err := listRoleNames(client, u, token) | ||
361 | if err != nil { | ||
362 | return ec2RoleCredRespBody{}, err | ||
363 | } | ||
364 | |||
365 | if len(roleNames) == 0 { | ||
366 | return ec2RoleCredRespBody{}, errors.New("No IAM roles attached to this EC2 service") | ||
367 | } | ||
368 | |||
369 | // http://docs.aws.amazon.com/AWSEC2/latest/UserGuide/iam-roles-for-amazon-ec2.html | ||
370 | // - An instance profile can contain only one IAM role. This limit cannot be increased. | ||
371 | roleName := roleNames[0] | ||
372 | |||
373 | // http://docs.aws.amazon.com/AWSEC2/latest/UserGuide/iam-roles-for-amazon-ec2.html | ||
374 | // The following command retrieves the security credentials for an | ||
375 | // IAM role named `s3access`. | ||
376 | // | ||
377 | // $ curl http://169.254.169.254/latest/meta-data/iam/security-credentials/s3access | ||
378 | // | ||
379 | u.Path = path.Join(u.Path, roleName) | ||
380 | req, err := http.NewRequest(http.MethodGet, u.String(), nil) | ||
381 | if err != nil { | ||
382 | return ec2RoleCredRespBody{}, err | ||
383 | } | ||
384 | if token != "" { | ||
385 | req.Header.Add(TokenRequestHeader, token) | ||
386 | } | ||
387 | |||
388 | resp, err := client.Do(req) | ||
389 | if err != nil { | ||
390 | return ec2RoleCredRespBody{}, err | ||
391 | } | ||
392 | defer resp.Body.Close() | ||
393 | if resp.StatusCode != http.StatusOK { | ||
394 | return ec2RoleCredRespBody{}, errors.New(resp.Status) | ||
395 | } | ||
396 | |||
397 | respCreds := ec2RoleCredRespBody{} | ||
398 | if err := jsoniter.NewDecoder(resp.Body).Decode(&respCreds); err != nil { | ||
399 | return ec2RoleCredRespBody{}, err | ||
400 | } | ||
401 | |||
402 | if respCreds.Code != "Success" { | ||
403 | // If an error code was returned something failed requesting the role. | ||
404 | return ec2RoleCredRespBody{}, errors.New(respCreds.Message) | ||
405 | } | ||
406 | |||
407 | return respCreds, nil | ||
408 | } | ||
409 | |||
410 | // isLoopback identifies if a uri's host is on a loopback address | ||
411 | func isLoopback(uri string) (bool, error) { | ||
412 | u, err := url.Parse(uri) | ||
413 | if err != nil { | ||
414 | return false, err | ||
415 | } | ||
416 | |||
417 | host := u.Hostname() | ||
418 | if len(host) == 0 { | ||
419 | return false, fmt.Errorf("can't parse host from uri: %s", uri) | ||
420 | } | ||
421 | |||
422 | ips, err := net.LookupHost(host) | ||
423 | if err != nil { | ||
424 | return false, err | ||
425 | } | ||
426 | for _, ip := range ips { | ||
427 | if !net.ParseIP(ip).IsLoopback() { | ||
428 | return false, nil | ||
429 | } | ||
430 | } | ||
431 | |||
432 | return true, nil | ||
433 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/pkg/credentials/signature_type.go b/vendor/github.com/minio/minio-go/v7/pkg/credentials/signature_type.go deleted file mode 100644 index b794333..0000000 --- a/vendor/github.com/minio/minio-go/v7/pkg/credentials/signature_type.go +++ /dev/null | |||
@@ -1,77 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2017 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package credentials | ||
19 | |||
20 | import "strings" | ||
21 | |||
22 | // SignatureType is type of Authorization requested for a given HTTP request. | ||
23 | type SignatureType int | ||
24 | |||
25 | // Different types of supported signatures - default is SignatureV4 or SignatureDefault. | ||
26 | const ( | ||
27 | // SignatureDefault is always set to v4. | ||
28 | SignatureDefault SignatureType = iota | ||
29 | SignatureV4 | ||
30 | SignatureV2 | ||
31 | SignatureV4Streaming | ||
32 | SignatureAnonymous // Anonymous signature signifies, no signature. | ||
33 | ) | ||
34 | |||
35 | // IsV2 - is signature SignatureV2? | ||
36 | func (s SignatureType) IsV2() bool { | ||
37 | return s == SignatureV2 | ||
38 | } | ||
39 | |||
40 | // IsV4 - is signature SignatureV4? | ||
41 | func (s SignatureType) IsV4() bool { | ||
42 | return s == SignatureV4 || s == SignatureDefault | ||
43 | } | ||
44 | |||
45 | // IsStreamingV4 - is signature SignatureV4Streaming? | ||
46 | func (s SignatureType) IsStreamingV4() bool { | ||
47 | return s == SignatureV4Streaming | ||
48 | } | ||
49 | |||
50 | // IsAnonymous - is signature empty? | ||
51 | func (s SignatureType) IsAnonymous() bool { | ||
52 | return s == SignatureAnonymous | ||
53 | } | ||
54 | |||
55 | // Stringer humanized version of signature type, | ||
56 | // strings returned here are case insensitive. | ||
57 | func (s SignatureType) String() string { | ||
58 | if s.IsV2() { | ||
59 | return "S3v2" | ||
60 | } else if s.IsV4() { | ||
61 | return "S3v4" | ||
62 | } else if s.IsStreamingV4() { | ||
63 | return "S3v4Streaming" | ||
64 | } | ||
65 | return "Anonymous" | ||
66 | } | ||
67 | |||
68 | func parseSignatureType(str string) SignatureType { | ||
69 | if strings.EqualFold(str, "S3v4") { | ||
70 | return SignatureV4 | ||
71 | } else if strings.EqualFold(str, "S3v2") { | ||
72 | return SignatureV2 | ||
73 | } else if strings.EqualFold(str, "S3v4Streaming") { | ||
74 | return SignatureV4Streaming | ||
75 | } | ||
76 | return SignatureAnonymous | ||
77 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/pkg/credentials/static.go b/vendor/github.com/minio/minio-go/v7/pkg/credentials/static.go deleted file mode 100644 index 7dde00b..0000000 --- a/vendor/github.com/minio/minio-go/v7/pkg/credentials/static.go +++ /dev/null | |||
@@ -1,67 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2017 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package credentials | ||
19 | |||
20 | // A Static is a set of credentials which are set programmatically, | ||
21 | // and will never expire. | ||
22 | type Static struct { | ||
23 | Value | ||
24 | } | ||
25 | |||
26 | // NewStaticV2 returns a pointer to a new Credentials object | ||
27 | // wrapping a static credentials value provider, signature is | ||
28 | // set to v2. If access and secret are not specified then | ||
29 | // regardless of signature type set it Value will return | ||
30 | // as anonymous. | ||
31 | func NewStaticV2(id, secret, token string) *Credentials { | ||
32 | return NewStatic(id, secret, token, SignatureV2) | ||
33 | } | ||
34 | |||
35 | // NewStaticV4 is similar to NewStaticV2 with similar considerations. | ||
36 | func NewStaticV4(id, secret, token string) *Credentials { | ||
37 | return NewStatic(id, secret, token, SignatureV4) | ||
38 | } | ||
39 | |||
40 | // NewStatic returns a pointer to a new Credentials object | ||
41 | // wrapping a static credentials value provider. | ||
42 | func NewStatic(id, secret, token string, signerType SignatureType) *Credentials { | ||
43 | return New(&Static{ | ||
44 | Value: Value{ | ||
45 | AccessKeyID: id, | ||
46 | SecretAccessKey: secret, | ||
47 | SessionToken: token, | ||
48 | SignerType: signerType, | ||
49 | }, | ||
50 | }) | ||
51 | } | ||
52 | |||
53 | // Retrieve returns the static credentials. | ||
54 | func (s *Static) Retrieve() (Value, error) { | ||
55 | if s.AccessKeyID == "" || s.SecretAccessKey == "" { | ||
56 | // Anonymous is not an error | ||
57 | return Value{SignerType: SignatureAnonymous}, nil | ||
58 | } | ||
59 | return s.Value, nil | ||
60 | } | ||
61 | |||
62 | // IsExpired returns if the credentials are expired. | ||
63 | // | ||
64 | // For Static, the credentials never expired. | ||
65 | func (s *Static) IsExpired() bool { | ||
66 | return false | ||
67 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/pkg/credentials/sts_client_grants.go b/vendor/github.com/minio/minio-go/v7/pkg/credentials/sts_client_grants.go deleted file mode 100644 index 9e92c1e..0000000 --- a/vendor/github.com/minio/minio-go/v7/pkg/credentials/sts_client_grants.go +++ /dev/null | |||
@@ -1,182 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2019-2022 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package credentials | ||
19 | |||
20 | import ( | ||
21 | "bytes" | ||
22 | "encoding/xml" | ||
23 | "errors" | ||
24 | "fmt" | ||
25 | "io" | ||
26 | "net/http" | ||
27 | "net/url" | ||
28 | "strings" | ||
29 | "time" | ||
30 | ) | ||
31 | |||
32 | // AssumedRoleUser - The identifiers for the temporary security credentials that | ||
33 | // the operation returns. Please also see https://docs.aws.amazon.com/goto/WebAPI/sts-2011-06-15/AssumedRoleUser | ||
34 | type AssumedRoleUser struct { | ||
35 | Arn string | ||
36 | AssumedRoleID string `xml:"AssumeRoleId"` | ||
37 | } | ||
38 | |||
39 | // AssumeRoleWithClientGrantsResponse contains the result of successful AssumeRoleWithClientGrants request. | ||
40 | type AssumeRoleWithClientGrantsResponse struct { | ||
41 | XMLName xml.Name `xml:"https://sts.amazonaws.com/doc/2011-06-15/ AssumeRoleWithClientGrantsResponse" json:"-"` | ||
42 | Result ClientGrantsResult `xml:"AssumeRoleWithClientGrantsResult"` | ||
43 | ResponseMetadata struct { | ||
44 | RequestID string `xml:"RequestId,omitempty"` | ||
45 | } `xml:"ResponseMetadata,omitempty"` | ||
46 | } | ||
47 | |||
48 | // ClientGrantsResult - Contains the response to a successful AssumeRoleWithClientGrants | ||
49 | // request, including temporary credentials that can be used to make MinIO API requests. | ||
50 | type ClientGrantsResult struct { | ||
51 | AssumedRoleUser AssumedRoleUser `xml:",omitempty"` | ||
52 | Audience string `xml:",omitempty"` | ||
53 | Credentials struct { | ||
54 | AccessKey string `xml:"AccessKeyId" json:"accessKey,omitempty"` | ||
55 | SecretKey string `xml:"SecretAccessKey" json:"secretKey,omitempty"` | ||
56 | Expiration time.Time `xml:"Expiration" json:"expiration,omitempty"` | ||
57 | SessionToken string `xml:"SessionToken" json:"sessionToken,omitempty"` | ||
58 | } `xml:",omitempty"` | ||
59 | PackedPolicySize int `xml:",omitempty"` | ||
60 | Provider string `xml:",omitempty"` | ||
61 | SubjectFromClientGrantsToken string `xml:",omitempty"` | ||
62 | } | ||
63 | |||
64 | // ClientGrantsToken - client grants token with expiry. | ||
65 | type ClientGrantsToken struct { | ||
66 | Token string | ||
67 | Expiry int | ||
68 | } | ||
69 | |||
70 | // A STSClientGrants retrieves credentials from MinIO service, and keeps track if | ||
71 | // those credentials are expired. | ||
72 | type STSClientGrants struct { | ||
73 | Expiry | ||
74 | |||
75 | // Required http Client to use when connecting to MinIO STS service. | ||
76 | Client *http.Client | ||
77 | |||
78 | // MinIO endpoint to fetch STS credentials. | ||
79 | STSEndpoint string | ||
80 | |||
81 | // getClientGrantsTokenExpiry function to retrieve tokens | ||
82 | // from IDP This function should return two values one is | ||
83 | // accessToken which is a self contained access token (JWT) | ||
84 | // and second return value is the expiry associated with | ||
85 | // this token. This is a customer provided function and | ||
86 | // is mandatory. | ||
87 | GetClientGrantsTokenExpiry func() (*ClientGrantsToken, error) | ||
88 | } | ||
89 | |||
90 | // NewSTSClientGrants returns a pointer to a new | ||
91 | // Credentials object wrapping the STSClientGrants. | ||
92 | func NewSTSClientGrants(stsEndpoint string, getClientGrantsTokenExpiry func() (*ClientGrantsToken, error)) (*Credentials, error) { | ||
93 | if stsEndpoint == "" { | ||
94 | return nil, errors.New("STS endpoint cannot be empty") | ||
95 | } | ||
96 | if getClientGrantsTokenExpiry == nil { | ||
97 | return nil, errors.New("Client grants access token and expiry retrieval function should be defined") | ||
98 | } | ||
99 | return New(&STSClientGrants{ | ||
100 | Client: &http.Client{ | ||
101 | Transport: http.DefaultTransport, | ||
102 | }, | ||
103 | STSEndpoint: stsEndpoint, | ||
104 | GetClientGrantsTokenExpiry: getClientGrantsTokenExpiry, | ||
105 | }), nil | ||
106 | } | ||
107 | |||
108 | func getClientGrantsCredentials(clnt *http.Client, endpoint string, | ||
109 | getClientGrantsTokenExpiry func() (*ClientGrantsToken, error), | ||
110 | ) (AssumeRoleWithClientGrantsResponse, error) { | ||
111 | accessToken, err := getClientGrantsTokenExpiry() | ||
112 | if err != nil { | ||
113 | return AssumeRoleWithClientGrantsResponse{}, err | ||
114 | } | ||
115 | |||
116 | v := url.Values{} | ||
117 | v.Set("Action", "AssumeRoleWithClientGrants") | ||
118 | v.Set("Token", accessToken.Token) | ||
119 | v.Set("DurationSeconds", fmt.Sprintf("%d", accessToken.Expiry)) | ||
120 | v.Set("Version", STSVersion) | ||
121 | |||
122 | u, err := url.Parse(endpoint) | ||
123 | if err != nil { | ||
124 | return AssumeRoleWithClientGrantsResponse{}, err | ||
125 | } | ||
126 | |||
127 | req, err := http.NewRequest(http.MethodPost, u.String(), strings.NewReader(v.Encode())) | ||
128 | if err != nil { | ||
129 | return AssumeRoleWithClientGrantsResponse{}, err | ||
130 | } | ||
131 | |||
132 | req.Header.Set("Content-Type", "application/x-www-form-urlencoded") | ||
133 | |||
134 | resp, err := clnt.Do(req) | ||
135 | if err != nil { | ||
136 | return AssumeRoleWithClientGrantsResponse{}, err | ||
137 | } | ||
138 | defer resp.Body.Close() | ||
139 | if resp.StatusCode != http.StatusOK { | ||
140 | var errResp ErrorResponse | ||
141 | buf, err := io.ReadAll(resp.Body) | ||
142 | if err != nil { | ||
143 | return AssumeRoleWithClientGrantsResponse{}, err | ||
144 | } | ||
145 | _, err = xmlDecodeAndBody(bytes.NewReader(buf), &errResp) | ||
146 | if err != nil { | ||
147 | var s3Err Error | ||
148 | if _, err = xmlDecodeAndBody(bytes.NewReader(buf), &s3Err); err != nil { | ||
149 | return AssumeRoleWithClientGrantsResponse{}, err | ||
150 | } | ||
151 | errResp.RequestID = s3Err.RequestID | ||
152 | errResp.STSError.Code = s3Err.Code | ||
153 | errResp.STSError.Message = s3Err.Message | ||
154 | } | ||
155 | return AssumeRoleWithClientGrantsResponse{}, errResp | ||
156 | } | ||
157 | |||
158 | a := AssumeRoleWithClientGrantsResponse{} | ||
159 | if err = xml.NewDecoder(resp.Body).Decode(&a); err != nil { | ||
160 | return AssumeRoleWithClientGrantsResponse{}, err | ||
161 | } | ||
162 | return a, nil | ||
163 | } | ||
164 | |||
165 | // Retrieve retrieves credentials from the MinIO service. | ||
166 | // Error will be returned if the request fails. | ||
167 | func (m *STSClientGrants) Retrieve() (Value, error) { | ||
168 | a, err := getClientGrantsCredentials(m.Client, m.STSEndpoint, m.GetClientGrantsTokenExpiry) | ||
169 | if err != nil { | ||
170 | return Value{}, err | ||
171 | } | ||
172 | |||
173 | // Expiry window is set to 10secs. | ||
174 | m.SetExpiration(a.Result.Credentials.Expiration, DefaultExpiryWindow) | ||
175 | |||
176 | return Value{ | ||
177 | AccessKeyID: a.Result.Credentials.AccessKey, | ||
178 | SecretAccessKey: a.Result.Credentials.SecretKey, | ||
179 | SessionToken: a.Result.Credentials.SessionToken, | ||
180 | SignerType: SignatureV4, | ||
181 | }, nil | ||
182 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/pkg/credentials/sts_custom_identity.go b/vendor/github.com/minio/minio-go/v7/pkg/credentials/sts_custom_identity.go deleted file mode 100644 index e1f9ce4..0000000 --- a/vendor/github.com/minio/minio-go/v7/pkg/credentials/sts_custom_identity.go +++ /dev/null | |||
@@ -1,146 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2015-2022 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package credentials | ||
19 | |||
20 | import ( | ||
21 | "encoding/xml" | ||
22 | "errors" | ||
23 | "fmt" | ||
24 | "net/http" | ||
25 | "net/url" | ||
26 | "time" | ||
27 | ) | ||
28 | |||
29 | // CustomTokenResult - Contains temporary creds and user metadata. | ||
30 | type CustomTokenResult struct { | ||
31 | Credentials struct { | ||
32 | AccessKey string `xml:"AccessKeyId"` | ||
33 | SecretKey string `xml:"SecretAccessKey"` | ||
34 | Expiration time.Time `xml:"Expiration"` | ||
35 | SessionToken string `xml:"SessionToken"` | ||
36 | } `xml:",omitempty"` | ||
37 | |||
38 | AssumedUser string `xml:",omitempty"` | ||
39 | } | ||
40 | |||
41 | // AssumeRoleWithCustomTokenResponse contains the result of a successful | ||
42 | // AssumeRoleWithCustomToken request. | ||
43 | type AssumeRoleWithCustomTokenResponse struct { | ||
44 | XMLName xml.Name `xml:"https://sts.amazonaws.com/doc/2011-06-15/ AssumeRoleWithCustomTokenResponse" json:"-"` | ||
45 | Result CustomTokenResult `xml:"AssumeRoleWithCustomTokenResult"` | ||
46 | Metadata struct { | ||
47 | RequestID string `xml:"RequestId,omitempty"` | ||
48 | } `xml:"ResponseMetadata,omitempty"` | ||
49 | } | ||
50 | |||
51 | // CustomTokenIdentity - satisfies the Provider interface, and retrieves | ||
52 | // credentials from MinIO using the AssumeRoleWithCustomToken STS API. | ||
53 | type CustomTokenIdentity struct { | ||
54 | Expiry | ||
55 | |||
56 | Client *http.Client | ||
57 | |||
58 | // MinIO server STS endpoint to fetch STS credentials. | ||
59 | STSEndpoint string | ||
60 | |||
61 | // The custom token to use with the request. | ||
62 | Token string | ||
63 | |||
64 | // RoleArn associated with the identity | ||
65 | RoleArn string | ||
66 | |||
67 | // RequestedExpiry is to set the validity of the generated credentials | ||
68 | // (this value bounded by server). | ||
69 | RequestedExpiry time.Duration | ||
70 | } | ||
71 | |||
72 | // Retrieve - to satisfy Provider interface; fetches credentials from MinIO. | ||
73 | func (c *CustomTokenIdentity) Retrieve() (value Value, err error) { | ||
74 | u, err := url.Parse(c.STSEndpoint) | ||
75 | if err != nil { | ||
76 | return value, err | ||
77 | } | ||
78 | |||
79 | v := url.Values{} | ||
80 | v.Set("Action", "AssumeRoleWithCustomToken") | ||
81 | v.Set("Version", STSVersion) | ||
82 | v.Set("RoleArn", c.RoleArn) | ||
83 | v.Set("Token", c.Token) | ||
84 | if c.RequestedExpiry != 0 { | ||
85 | v.Set("DurationSeconds", fmt.Sprintf("%d", int(c.RequestedExpiry.Seconds()))) | ||
86 | } | ||
87 | |||
88 | u.RawQuery = v.Encode() | ||
89 | |||
90 | req, err := http.NewRequest(http.MethodPost, u.String(), nil) | ||
91 | if err != nil { | ||
92 | return value, err | ||
93 | } | ||
94 | |||
95 | resp, err := c.Client.Do(req) | ||
96 | if err != nil { | ||
97 | return value, err | ||
98 | } | ||
99 | |||
100 | defer resp.Body.Close() | ||
101 | if resp.StatusCode != http.StatusOK { | ||
102 | return value, errors.New(resp.Status) | ||
103 | } | ||
104 | |||
105 | r := AssumeRoleWithCustomTokenResponse{} | ||
106 | if err = xml.NewDecoder(resp.Body).Decode(&r); err != nil { | ||
107 | return | ||
108 | } | ||
109 | |||
110 | cr := r.Result.Credentials | ||
111 | c.SetExpiration(cr.Expiration, DefaultExpiryWindow) | ||
112 | return Value{ | ||
113 | AccessKeyID: cr.AccessKey, | ||
114 | SecretAccessKey: cr.SecretKey, | ||
115 | SessionToken: cr.SessionToken, | ||
116 | SignerType: SignatureV4, | ||
117 | }, nil | ||
118 | } | ||
119 | |||
120 | // NewCustomTokenCredentials - returns credentials using the | ||
121 | // AssumeRoleWithCustomToken STS API. | ||
122 | func NewCustomTokenCredentials(stsEndpoint, token, roleArn string, optFuncs ...CustomTokenOpt) (*Credentials, error) { | ||
123 | c := CustomTokenIdentity{ | ||
124 | Client: &http.Client{Transport: http.DefaultTransport}, | ||
125 | STSEndpoint: stsEndpoint, | ||
126 | Token: token, | ||
127 | RoleArn: roleArn, | ||
128 | } | ||
129 | for _, optFunc := range optFuncs { | ||
130 | optFunc(&c) | ||
131 | } | ||
132 | return New(&c), nil | ||
133 | } | ||
134 | |||
135 | // CustomTokenOpt is a function type to configure the custom-token based | ||
136 | // credentials using NewCustomTokenCredentials. | ||
137 | type CustomTokenOpt func(*CustomTokenIdentity) | ||
138 | |||
139 | // CustomTokenValidityOpt sets the validity duration of the requested | ||
140 | // credentials. This value is ignored if the server enforces a lower validity | ||
141 | // period. | ||
142 | func CustomTokenValidityOpt(d time.Duration) CustomTokenOpt { | ||
143 | return func(c *CustomTokenIdentity) { | ||
144 | c.RequestedExpiry = d | ||
145 | } | ||
146 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/pkg/credentials/sts_ldap_identity.go b/vendor/github.com/minio/minio-go/v7/pkg/credentials/sts_ldap_identity.go deleted file mode 100644 index ec5f3f0..0000000 --- a/vendor/github.com/minio/minio-go/v7/pkg/credentials/sts_ldap_identity.go +++ /dev/null | |||
@@ -1,189 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2019-2022 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package credentials | ||
19 | |||
20 | import ( | ||
21 | "bytes" | ||
22 | "encoding/xml" | ||
23 | "fmt" | ||
24 | "io" | ||
25 | "net/http" | ||
26 | "net/url" | ||
27 | "strings" | ||
28 | "time" | ||
29 | ) | ||
30 | |||
31 | // AssumeRoleWithLDAPResponse contains the result of successful | ||
32 | // AssumeRoleWithLDAPIdentity request | ||
33 | type AssumeRoleWithLDAPResponse struct { | ||
34 | XMLName xml.Name `xml:"https://sts.amazonaws.com/doc/2011-06-15/ AssumeRoleWithLDAPIdentityResponse" json:"-"` | ||
35 | Result LDAPIdentityResult `xml:"AssumeRoleWithLDAPIdentityResult"` | ||
36 | ResponseMetadata struct { | ||
37 | RequestID string `xml:"RequestId,omitempty"` | ||
38 | } `xml:"ResponseMetadata,omitempty"` | ||
39 | } | ||
40 | |||
41 | // LDAPIdentityResult - contains credentials for a successful | ||
42 | // AssumeRoleWithLDAPIdentity request. | ||
43 | type LDAPIdentityResult struct { | ||
44 | Credentials struct { | ||
45 | AccessKey string `xml:"AccessKeyId" json:"accessKey,omitempty"` | ||
46 | SecretKey string `xml:"SecretAccessKey" json:"secretKey,omitempty"` | ||
47 | Expiration time.Time `xml:"Expiration" json:"expiration,omitempty"` | ||
48 | SessionToken string `xml:"SessionToken" json:"sessionToken,omitempty"` | ||
49 | } `xml:",omitempty"` | ||
50 | |||
51 | SubjectFromToken string `xml:",omitempty"` | ||
52 | } | ||
53 | |||
54 | // LDAPIdentity retrieves credentials from MinIO | ||
55 | type LDAPIdentity struct { | ||
56 | Expiry | ||
57 | |||
58 | // Required http Client to use when connecting to MinIO STS service. | ||
59 | Client *http.Client | ||
60 | |||
61 | // Exported STS endpoint to fetch STS credentials. | ||
62 | STSEndpoint string | ||
63 | |||
64 | // LDAP username/password used to fetch LDAP STS credentials. | ||
65 | LDAPUsername, LDAPPassword string | ||
66 | |||
67 | // Session policy to apply to the generated credentials. Leave empty to | ||
68 | // use the full access policy available to the user. | ||
69 | Policy string | ||
70 | |||
71 | // RequestedExpiry is the configured expiry duration for credentials | ||
72 | // requested from LDAP. | ||
73 | RequestedExpiry time.Duration | ||
74 | } | ||
75 | |||
76 | // NewLDAPIdentity returns new credentials object that uses LDAP | ||
77 | // Identity. | ||
78 | func NewLDAPIdentity(stsEndpoint, ldapUsername, ldapPassword string, optFuncs ...LDAPIdentityOpt) (*Credentials, error) { | ||
79 | l := LDAPIdentity{ | ||
80 | Client: &http.Client{Transport: http.DefaultTransport}, | ||
81 | STSEndpoint: stsEndpoint, | ||
82 | LDAPUsername: ldapUsername, | ||
83 | LDAPPassword: ldapPassword, | ||
84 | } | ||
85 | for _, optFunc := range optFuncs { | ||
86 | optFunc(&l) | ||
87 | } | ||
88 | return New(&l), nil | ||
89 | } | ||
90 | |||
91 | // LDAPIdentityOpt is a function type used to configured the LDAPIdentity | ||
92 | // instance. | ||
93 | type LDAPIdentityOpt func(*LDAPIdentity) | ||
94 | |||
95 | // LDAPIdentityPolicyOpt sets the session policy for requested credentials. | ||
96 | func LDAPIdentityPolicyOpt(policy string) LDAPIdentityOpt { | ||
97 | return func(k *LDAPIdentity) { | ||
98 | k.Policy = policy | ||
99 | } | ||
100 | } | ||
101 | |||
102 | // LDAPIdentityExpiryOpt sets the expiry duration for requested credentials. | ||
103 | func LDAPIdentityExpiryOpt(d time.Duration) LDAPIdentityOpt { | ||
104 | return func(k *LDAPIdentity) { | ||
105 | k.RequestedExpiry = d | ||
106 | } | ||
107 | } | ||
108 | |||
109 | // NewLDAPIdentityWithSessionPolicy returns new credentials object that uses | ||
110 | // LDAP Identity with a specified session policy. The `policy` parameter must be | ||
111 | // a JSON string specifying the policy document. | ||
112 | // | ||
113 | // Deprecated: Use the `LDAPIdentityPolicyOpt` with `NewLDAPIdentity` instead. | ||
114 | func NewLDAPIdentityWithSessionPolicy(stsEndpoint, ldapUsername, ldapPassword, policy string) (*Credentials, error) { | ||
115 | return New(&LDAPIdentity{ | ||
116 | Client: &http.Client{Transport: http.DefaultTransport}, | ||
117 | STSEndpoint: stsEndpoint, | ||
118 | LDAPUsername: ldapUsername, | ||
119 | LDAPPassword: ldapPassword, | ||
120 | Policy: policy, | ||
121 | }), nil | ||
122 | } | ||
123 | |||
124 | // Retrieve gets the credential by calling the MinIO STS API for | ||
125 | // LDAP on the configured stsEndpoint. | ||
126 | func (k *LDAPIdentity) Retrieve() (value Value, err error) { | ||
127 | u, err := url.Parse(k.STSEndpoint) | ||
128 | if err != nil { | ||
129 | return value, err | ||
130 | } | ||
131 | |||
132 | v := url.Values{} | ||
133 | v.Set("Action", "AssumeRoleWithLDAPIdentity") | ||
134 | v.Set("Version", STSVersion) | ||
135 | v.Set("LDAPUsername", k.LDAPUsername) | ||
136 | v.Set("LDAPPassword", k.LDAPPassword) | ||
137 | if k.Policy != "" { | ||
138 | v.Set("Policy", k.Policy) | ||
139 | } | ||
140 | if k.RequestedExpiry != 0 { | ||
141 | v.Set("DurationSeconds", fmt.Sprintf("%d", int(k.RequestedExpiry.Seconds()))) | ||
142 | } | ||
143 | |||
144 | req, err := http.NewRequest(http.MethodPost, u.String(), strings.NewReader(v.Encode())) | ||
145 | if err != nil { | ||
146 | return value, err | ||
147 | } | ||
148 | |||
149 | req.Header.Set("Content-Type", "application/x-www-form-urlencoded") | ||
150 | |||
151 | resp, err := k.Client.Do(req) | ||
152 | if err != nil { | ||
153 | return value, err | ||
154 | } | ||
155 | |||
156 | defer resp.Body.Close() | ||
157 | if resp.StatusCode != http.StatusOK { | ||
158 | var errResp ErrorResponse | ||
159 | buf, err := io.ReadAll(resp.Body) | ||
160 | if err != nil { | ||
161 | return value, err | ||
162 | } | ||
163 | _, err = xmlDecodeAndBody(bytes.NewReader(buf), &errResp) | ||
164 | if err != nil { | ||
165 | var s3Err Error | ||
166 | if _, err = xmlDecodeAndBody(bytes.NewReader(buf), &s3Err); err != nil { | ||
167 | return value, err | ||
168 | } | ||
169 | errResp.RequestID = s3Err.RequestID | ||
170 | errResp.STSError.Code = s3Err.Code | ||
171 | errResp.STSError.Message = s3Err.Message | ||
172 | } | ||
173 | return value, errResp | ||
174 | } | ||
175 | |||
176 | r := AssumeRoleWithLDAPResponse{} | ||
177 | if err = xml.NewDecoder(resp.Body).Decode(&r); err != nil { | ||
178 | return | ||
179 | } | ||
180 | |||
181 | cr := r.Result.Credentials | ||
182 | k.SetExpiration(cr.Expiration, DefaultExpiryWindow) | ||
183 | return Value{ | ||
184 | AccessKeyID: cr.AccessKey, | ||
185 | SecretAccessKey: cr.SecretKey, | ||
186 | SessionToken: cr.SessionToken, | ||
187 | SignerType: SignatureV4, | ||
188 | }, nil | ||
189 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/pkg/credentials/sts_tls_identity.go b/vendor/github.com/minio/minio-go/v7/pkg/credentials/sts_tls_identity.go deleted file mode 100644 index dee0a8c..0000000 --- a/vendor/github.com/minio/minio-go/v7/pkg/credentials/sts_tls_identity.go +++ /dev/null | |||
@@ -1,211 +0,0 @@ | |||
1 | // MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
2 | // Copyright 2021 MinIO, Inc. | ||
3 | // | ||
4 | // Licensed under the Apache License, Version 2.0 (the "License"); | ||
5 | // you may not use this file except in compliance with the License. | ||
6 | // You may obtain a copy of the License at | ||
7 | // | ||
8 | // http://www.apache.org/licenses/LICENSE-2.0 | ||
9 | // | ||
10 | // Unless required by applicable law or agreed to in writing, software | ||
11 | // distributed under the License is distributed on an "AS IS" BASIS, | ||
12 | // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
13 | // See the License for the specific language governing permissions and | ||
14 | // limitations under the License. | ||
15 | |||
16 | package credentials | ||
17 | |||
18 | import ( | ||
19 | "bytes" | ||
20 | "crypto/tls" | ||
21 | "encoding/xml" | ||
22 | "errors" | ||
23 | "io" | ||
24 | "net" | ||
25 | "net/http" | ||
26 | "net/url" | ||
27 | "strconv" | ||
28 | "time" | ||
29 | ) | ||
30 | |||
31 | // CertificateIdentityOption is an optional AssumeRoleWithCertificate | ||
32 | // parameter - e.g. a custom HTTP transport configuration or S3 credental | ||
33 | // livetime. | ||
34 | type CertificateIdentityOption func(*STSCertificateIdentity) | ||
35 | |||
36 | // CertificateIdentityWithTransport returns a CertificateIdentityOption that | ||
37 | // customizes the STSCertificateIdentity with the given http.RoundTripper. | ||
38 | func CertificateIdentityWithTransport(t http.RoundTripper) CertificateIdentityOption { | ||
39 | return CertificateIdentityOption(func(i *STSCertificateIdentity) { i.Client.Transport = t }) | ||
40 | } | ||
41 | |||
42 | // CertificateIdentityWithExpiry returns a CertificateIdentityOption that | ||
43 | // customizes the STSCertificateIdentity with the given livetime. | ||
44 | // | ||
45 | // Fetched S3 credentials will have the given livetime if the STS server | ||
46 | // allows such credentials. | ||
47 | func CertificateIdentityWithExpiry(livetime time.Duration) CertificateIdentityOption { | ||
48 | return CertificateIdentityOption(func(i *STSCertificateIdentity) { i.S3CredentialLivetime = livetime }) | ||
49 | } | ||
50 | |||
51 | // A STSCertificateIdentity retrieves S3 credentials from the MinIO STS API and | ||
52 | // rotates those credentials once they expire. | ||
53 | type STSCertificateIdentity struct { | ||
54 | Expiry | ||
55 | |||
56 | // STSEndpoint is the base URL endpoint of the STS API. | ||
57 | // For example, https://minio.local:9000 | ||
58 | STSEndpoint string | ||
59 | |||
60 | // S3CredentialLivetime is the duration temp. S3 access | ||
61 | // credentials should be valid. | ||
62 | // | ||
63 | // It represents the access credential livetime requested | ||
64 | // by the client. The STS server may choose to issue | ||
65 | // temp. S3 credentials that have a different - usually | ||
66 | // shorter - livetime. | ||
67 | // | ||
68 | // The default livetime is one hour. | ||
69 | S3CredentialLivetime time.Duration | ||
70 | |||
71 | // Client is the HTTP client used to authenticate and fetch | ||
72 | // S3 credentials. | ||
73 | // | ||
74 | // A custom TLS client configuration can be specified by | ||
75 | // using a custom http.Transport: | ||
76 | // Client: http.Client { | ||
77 | // Transport: &http.Transport{ | ||
78 | // TLSClientConfig: &tls.Config{}, | ||
79 | // }, | ||
80 | // } | ||
81 | Client http.Client | ||
82 | } | ||
83 | |||
84 | var _ Provider = (*STSWebIdentity)(nil) // compiler check | ||
85 | |||
86 | // NewSTSCertificateIdentity returns a STSCertificateIdentity that authenticates | ||
87 | // to the given STS endpoint with the given TLS certificate and retrieves and | ||
88 | // rotates S3 credentials. | ||
89 | func NewSTSCertificateIdentity(endpoint string, certificate tls.Certificate, options ...CertificateIdentityOption) (*Credentials, error) { | ||
90 | if endpoint == "" { | ||
91 | return nil, errors.New("STS endpoint cannot be empty") | ||
92 | } | ||
93 | if _, err := url.Parse(endpoint); err != nil { | ||
94 | return nil, err | ||
95 | } | ||
96 | identity := &STSCertificateIdentity{ | ||
97 | STSEndpoint: endpoint, | ||
98 | Client: http.Client{ | ||
99 | Transport: &http.Transport{ | ||
100 | Proxy: http.ProxyFromEnvironment, | ||
101 | DialContext: (&net.Dialer{ | ||
102 | Timeout: 30 * time.Second, | ||
103 | KeepAlive: 30 * time.Second, | ||
104 | }).DialContext, | ||
105 | ForceAttemptHTTP2: true, | ||
106 | MaxIdleConns: 100, | ||
107 | IdleConnTimeout: 90 * time.Second, | ||
108 | TLSHandshakeTimeout: 10 * time.Second, | ||
109 | ExpectContinueTimeout: 5 * time.Second, | ||
110 | TLSClientConfig: &tls.Config{ | ||
111 | Certificates: []tls.Certificate{certificate}, | ||
112 | }, | ||
113 | }, | ||
114 | }, | ||
115 | } | ||
116 | for _, option := range options { | ||
117 | option(identity) | ||
118 | } | ||
119 | return New(identity), nil | ||
120 | } | ||
121 | |||
122 | // Retrieve fetches a new set of S3 credentials from the configured | ||
123 | // STS API endpoint. | ||
124 | func (i *STSCertificateIdentity) Retrieve() (Value, error) { | ||
125 | endpointURL, err := url.Parse(i.STSEndpoint) | ||
126 | if err != nil { | ||
127 | return Value{}, err | ||
128 | } | ||
129 | livetime := i.S3CredentialLivetime | ||
130 | if livetime == 0 { | ||
131 | livetime = 1 * time.Hour | ||
132 | } | ||
133 | |||
134 | queryValues := url.Values{} | ||
135 | queryValues.Set("Action", "AssumeRoleWithCertificate") | ||
136 | queryValues.Set("Version", STSVersion) | ||
137 | endpointURL.RawQuery = queryValues.Encode() | ||
138 | |||
139 | req, err := http.NewRequest(http.MethodPost, endpointURL.String(), nil) | ||
140 | if err != nil { | ||
141 | return Value{}, err | ||
142 | } | ||
143 | if req.Form == nil { | ||
144 | req.Form = url.Values{} | ||
145 | } | ||
146 | req.Form.Add("DurationSeconds", strconv.FormatUint(uint64(livetime.Seconds()), 10)) | ||
147 | |||
148 | resp, err := i.Client.Do(req) | ||
149 | if err != nil { | ||
150 | return Value{}, err | ||
151 | } | ||
152 | if resp.Body != nil { | ||
153 | defer resp.Body.Close() | ||
154 | } | ||
155 | if resp.StatusCode != http.StatusOK { | ||
156 | var errResp ErrorResponse | ||
157 | buf, err := io.ReadAll(resp.Body) | ||
158 | if err != nil { | ||
159 | return Value{}, err | ||
160 | } | ||
161 | _, err = xmlDecodeAndBody(bytes.NewReader(buf), &errResp) | ||
162 | if err != nil { | ||
163 | var s3Err Error | ||
164 | if _, err = xmlDecodeAndBody(bytes.NewReader(buf), &s3Err); err != nil { | ||
165 | return Value{}, err | ||
166 | } | ||
167 | errResp.RequestID = s3Err.RequestID | ||
168 | errResp.STSError.Code = s3Err.Code | ||
169 | errResp.STSError.Message = s3Err.Message | ||
170 | } | ||
171 | return Value{}, errResp | ||
172 | } | ||
173 | |||
174 | const MaxSize = 10 * 1 << 20 | ||
175 | var body io.Reader = resp.Body | ||
176 | if resp.ContentLength > 0 && resp.ContentLength < MaxSize { | ||
177 | body = io.LimitReader(body, resp.ContentLength) | ||
178 | } else { | ||
179 | body = io.LimitReader(body, MaxSize) | ||
180 | } | ||
181 | |||
182 | var response assumeRoleWithCertificateResponse | ||
183 | if err = xml.NewDecoder(body).Decode(&response); err != nil { | ||
184 | return Value{}, err | ||
185 | } | ||
186 | i.SetExpiration(response.Result.Credentials.Expiration, DefaultExpiryWindow) | ||
187 | return Value{ | ||
188 | AccessKeyID: response.Result.Credentials.AccessKey, | ||
189 | SecretAccessKey: response.Result.Credentials.SecretKey, | ||
190 | SessionToken: response.Result.Credentials.SessionToken, | ||
191 | SignerType: SignatureDefault, | ||
192 | }, nil | ||
193 | } | ||
194 | |||
195 | // Expiration returns the expiration time of the current S3 credentials. | ||
196 | func (i *STSCertificateIdentity) Expiration() time.Time { return i.expiration } | ||
197 | |||
198 | type assumeRoleWithCertificateResponse struct { | ||
199 | XMLName xml.Name `xml:"https://sts.amazonaws.com/doc/2011-06-15/ AssumeRoleWithCertificateResponse" json:"-"` | ||
200 | Result struct { | ||
201 | Credentials struct { | ||
202 | AccessKey string `xml:"AccessKeyId" json:"accessKey,omitempty"` | ||
203 | SecretKey string `xml:"SecretAccessKey" json:"secretKey,omitempty"` | ||
204 | Expiration time.Time `xml:"Expiration" json:"expiration,omitempty"` | ||
205 | SessionToken string `xml:"SessionToken" json:"sessionToken,omitempty"` | ||
206 | } `xml:"Credentials" json:"credentials,omitempty"` | ||
207 | } `xml:"AssumeRoleWithCertificateResult"` | ||
208 | ResponseMetadata struct { | ||
209 | RequestID string `xml:"RequestId,omitempty"` | ||
210 | } `xml:"ResponseMetadata,omitempty"` | ||
211 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/pkg/credentials/sts_web_identity.go b/vendor/github.com/minio/minio-go/v7/pkg/credentials/sts_web_identity.go deleted file mode 100644 index 2e2af50..0000000 --- a/vendor/github.com/minio/minio-go/v7/pkg/credentials/sts_web_identity.go +++ /dev/null | |||
@@ -1,205 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2019-2022 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package credentials | ||
19 | |||
20 | import ( | ||
21 | "bytes" | ||
22 | "encoding/xml" | ||
23 | "errors" | ||
24 | "fmt" | ||
25 | "io" | ||
26 | "net/http" | ||
27 | "net/url" | ||
28 | "strconv" | ||
29 | "strings" | ||
30 | "time" | ||
31 | ) | ||
32 | |||
33 | // AssumeRoleWithWebIdentityResponse contains the result of successful AssumeRoleWithWebIdentity request. | ||
34 | type AssumeRoleWithWebIdentityResponse struct { | ||
35 | XMLName xml.Name `xml:"https://sts.amazonaws.com/doc/2011-06-15/ AssumeRoleWithWebIdentityResponse" json:"-"` | ||
36 | Result WebIdentityResult `xml:"AssumeRoleWithWebIdentityResult"` | ||
37 | ResponseMetadata struct { | ||
38 | RequestID string `xml:"RequestId,omitempty"` | ||
39 | } `xml:"ResponseMetadata,omitempty"` | ||
40 | } | ||
41 | |||
42 | // WebIdentityResult - Contains the response to a successful AssumeRoleWithWebIdentity | ||
43 | // request, including temporary credentials that can be used to make MinIO API requests. | ||
44 | type WebIdentityResult struct { | ||
45 | AssumedRoleUser AssumedRoleUser `xml:",omitempty"` | ||
46 | Audience string `xml:",omitempty"` | ||
47 | Credentials struct { | ||
48 | AccessKey string `xml:"AccessKeyId" json:"accessKey,omitempty"` | ||
49 | SecretKey string `xml:"SecretAccessKey" json:"secretKey,omitempty"` | ||
50 | Expiration time.Time `xml:"Expiration" json:"expiration,omitempty"` | ||
51 | SessionToken string `xml:"SessionToken" json:"sessionToken,omitempty"` | ||
52 | } `xml:",omitempty"` | ||
53 | PackedPolicySize int `xml:",omitempty"` | ||
54 | Provider string `xml:",omitempty"` | ||
55 | SubjectFromWebIdentityToken string `xml:",omitempty"` | ||
56 | } | ||
57 | |||
58 | // WebIdentityToken - web identity token with expiry. | ||
59 | type WebIdentityToken struct { | ||
60 | Token string | ||
61 | AccessToken string | ||
62 | Expiry int | ||
63 | } | ||
64 | |||
65 | // A STSWebIdentity retrieves credentials from MinIO service, and keeps track if | ||
66 | // those credentials are expired. | ||
67 | type STSWebIdentity struct { | ||
68 | Expiry | ||
69 | |||
70 | // Required http Client to use when connecting to MinIO STS service. | ||
71 | Client *http.Client | ||
72 | |||
73 | // Exported STS endpoint to fetch STS credentials. | ||
74 | STSEndpoint string | ||
75 | |||
76 | // Exported GetWebIDTokenExpiry function which returns ID | ||
77 | // tokens from IDP. This function should return two values | ||
78 | // one is ID token which is a self contained ID token (JWT) | ||
79 | // and second return value is the expiry associated with | ||
80 | // this token. | ||
81 | // This is a customer provided function and is mandatory. | ||
82 | GetWebIDTokenExpiry func() (*WebIdentityToken, error) | ||
83 | |||
84 | // RoleARN is the Amazon Resource Name (ARN) of the role that the caller is | ||
85 | // assuming. | ||
86 | RoleARN string | ||
87 | |||
88 | // roleSessionName is the identifier for the assumed role session. | ||
89 | roleSessionName string | ||
90 | } | ||
91 | |||
92 | // NewSTSWebIdentity returns a pointer to a new | ||
93 | // Credentials object wrapping the STSWebIdentity. | ||
94 | func NewSTSWebIdentity(stsEndpoint string, getWebIDTokenExpiry func() (*WebIdentityToken, error)) (*Credentials, error) { | ||
95 | if stsEndpoint == "" { | ||
96 | return nil, errors.New("STS endpoint cannot be empty") | ||
97 | } | ||
98 | if getWebIDTokenExpiry == nil { | ||
99 | return nil, errors.New("Web ID token and expiry retrieval function should be defined") | ||
100 | } | ||
101 | return New(&STSWebIdentity{ | ||
102 | Client: &http.Client{ | ||
103 | Transport: http.DefaultTransport, | ||
104 | }, | ||
105 | STSEndpoint: stsEndpoint, | ||
106 | GetWebIDTokenExpiry: getWebIDTokenExpiry, | ||
107 | }), nil | ||
108 | } | ||
109 | |||
110 | func getWebIdentityCredentials(clnt *http.Client, endpoint, roleARN, roleSessionName string, | ||
111 | getWebIDTokenExpiry func() (*WebIdentityToken, error), | ||
112 | ) (AssumeRoleWithWebIdentityResponse, error) { | ||
113 | idToken, err := getWebIDTokenExpiry() | ||
114 | if err != nil { | ||
115 | return AssumeRoleWithWebIdentityResponse{}, err | ||
116 | } | ||
117 | |||
118 | v := url.Values{} | ||
119 | v.Set("Action", "AssumeRoleWithWebIdentity") | ||
120 | if len(roleARN) > 0 { | ||
121 | v.Set("RoleArn", roleARN) | ||
122 | |||
123 | if len(roleSessionName) == 0 { | ||
124 | roleSessionName = strconv.FormatInt(time.Now().UnixNano(), 10) | ||
125 | } | ||
126 | v.Set("RoleSessionName", roleSessionName) | ||
127 | } | ||
128 | v.Set("WebIdentityToken", idToken.Token) | ||
129 | if idToken.AccessToken != "" { | ||
130 | // Usually set when server is using extended userInfo endpoint. | ||
131 | v.Set("WebIdentityAccessToken", idToken.AccessToken) | ||
132 | } | ||
133 | if idToken.Expiry > 0 { | ||
134 | v.Set("DurationSeconds", fmt.Sprintf("%d", idToken.Expiry)) | ||
135 | } | ||
136 | v.Set("Version", STSVersion) | ||
137 | |||
138 | u, err := url.Parse(endpoint) | ||
139 | if err != nil { | ||
140 | return AssumeRoleWithWebIdentityResponse{}, err | ||
141 | } | ||
142 | |||
143 | req, err := http.NewRequest(http.MethodPost, u.String(), strings.NewReader(v.Encode())) | ||
144 | if err != nil { | ||
145 | return AssumeRoleWithWebIdentityResponse{}, err | ||
146 | } | ||
147 | |||
148 | req.Header.Set("Content-Type", "application/x-www-form-urlencoded") | ||
149 | |||
150 | resp, err := clnt.Do(req) | ||
151 | if err != nil { | ||
152 | return AssumeRoleWithWebIdentityResponse{}, err | ||
153 | } | ||
154 | |||
155 | defer resp.Body.Close() | ||
156 | if resp.StatusCode != http.StatusOK { | ||
157 | var errResp ErrorResponse | ||
158 | buf, err := io.ReadAll(resp.Body) | ||
159 | if err != nil { | ||
160 | return AssumeRoleWithWebIdentityResponse{}, err | ||
161 | } | ||
162 | _, err = xmlDecodeAndBody(bytes.NewReader(buf), &errResp) | ||
163 | if err != nil { | ||
164 | var s3Err Error | ||
165 | if _, err = xmlDecodeAndBody(bytes.NewReader(buf), &s3Err); err != nil { | ||
166 | return AssumeRoleWithWebIdentityResponse{}, err | ||
167 | } | ||
168 | errResp.RequestID = s3Err.RequestID | ||
169 | errResp.STSError.Code = s3Err.Code | ||
170 | errResp.STSError.Message = s3Err.Message | ||
171 | } | ||
172 | return AssumeRoleWithWebIdentityResponse{}, errResp | ||
173 | } | ||
174 | |||
175 | a := AssumeRoleWithWebIdentityResponse{} | ||
176 | if err = xml.NewDecoder(resp.Body).Decode(&a); err != nil { | ||
177 | return AssumeRoleWithWebIdentityResponse{}, err | ||
178 | } | ||
179 | |||
180 | return a, nil | ||
181 | } | ||
182 | |||
183 | // Retrieve retrieves credentials from the MinIO service. | ||
184 | // Error will be returned if the request fails. | ||
185 | func (m *STSWebIdentity) Retrieve() (Value, error) { | ||
186 | a, err := getWebIdentityCredentials(m.Client, m.STSEndpoint, m.RoleARN, m.roleSessionName, m.GetWebIDTokenExpiry) | ||
187 | if err != nil { | ||
188 | return Value{}, err | ||
189 | } | ||
190 | |||
191 | // Expiry window is set to 10secs. | ||
192 | m.SetExpiration(a.Result.Credentials.Expiration, DefaultExpiryWindow) | ||
193 | |||
194 | return Value{ | ||
195 | AccessKeyID: a.Result.Credentials.AccessKey, | ||
196 | SecretAccessKey: a.Result.Credentials.SecretKey, | ||
197 | SessionToken: a.Result.Credentials.SessionToken, | ||
198 | SignerType: SignatureV4, | ||
199 | }, nil | ||
200 | } | ||
201 | |||
202 | // Expiration returns the expiration time of the credentials | ||
203 | func (m *STSWebIdentity) Expiration() time.Time { | ||
204 | return m.expiration | ||
205 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/pkg/encrypt/fips_disabled.go b/vendor/github.com/minio/minio-go/v7/pkg/encrypt/fips_disabled.go deleted file mode 100644 index 6db26c0..0000000 --- a/vendor/github.com/minio/minio-go/v7/pkg/encrypt/fips_disabled.go +++ /dev/null | |||
@@ -1,24 +0,0 @@ | |||
1 | //go:build !fips | ||
2 | // +build !fips | ||
3 | |||
4 | /* | ||
5 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
6 | * Copyright 2022 MinIO, Inc. | ||
7 | * | ||
8 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
9 | * you may not use this file except in compliance with the License. | ||
10 | * You may obtain a copy of the License at | ||
11 | * | ||
12 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
13 | * | ||
14 | * Unless required by applicable law or agreed to in writing, software | ||
15 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
16 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
17 | * See the License for the specific language governing permissions and | ||
18 | * limitations under the License. | ||
19 | */ | ||
20 | |||
21 | package encrypt | ||
22 | |||
23 | // FIPS is true if 'fips' build tag was specified. | ||
24 | const FIPS = false | ||
diff --git a/vendor/github.com/minio/minio-go/v7/pkg/encrypt/fips_enabled.go b/vendor/github.com/minio/minio-go/v7/pkg/encrypt/fips_enabled.go deleted file mode 100644 index 6402582..0000000 --- a/vendor/github.com/minio/minio-go/v7/pkg/encrypt/fips_enabled.go +++ /dev/null | |||
@@ -1,24 +0,0 @@ | |||
1 | //go:build fips | ||
2 | // +build fips | ||
3 | |||
4 | /* | ||
5 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
6 | * Copyright 2022 MinIO, Inc. | ||
7 | * | ||
8 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
9 | * you may not use this file except in compliance with the License. | ||
10 | * You may obtain a copy of the License at | ||
11 | * | ||
12 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
13 | * | ||
14 | * Unless required by applicable law or agreed to in writing, software | ||
15 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
16 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
17 | * See the License for the specific language governing permissions and | ||
18 | * limitations under the License. | ||
19 | */ | ||
20 | |||
21 | package encrypt | ||
22 | |||
23 | // FIPS is true if 'fips' build tag was specified. | ||
24 | const FIPS = true | ||
diff --git a/vendor/github.com/minio/minio-go/v7/pkg/encrypt/server-side.go b/vendor/github.com/minio/minio-go/v7/pkg/encrypt/server-side.go deleted file mode 100644 index a7081c5..0000000 --- a/vendor/github.com/minio/minio-go/v7/pkg/encrypt/server-side.go +++ /dev/null | |||
@@ -1,198 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2018 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package encrypt | ||
19 | |||
20 | import ( | ||
21 | "crypto/md5" | ||
22 | "encoding/base64" | ||
23 | "errors" | ||
24 | "net/http" | ||
25 | |||
26 | jsoniter "github.com/json-iterator/go" | ||
27 | "golang.org/x/crypto/argon2" | ||
28 | ) | ||
29 | |||
30 | const ( | ||
31 | // SseGenericHeader is the AWS SSE header used for SSE-S3 and SSE-KMS. | ||
32 | SseGenericHeader = "X-Amz-Server-Side-Encryption" | ||
33 | |||
34 | // SseKmsKeyID is the AWS SSE-KMS key id. | ||
35 | SseKmsKeyID = SseGenericHeader + "-Aws-Kms-Key-Id" | ||
36 | // SseEncryptionContext is the AWS SSE-KMS Encryption Context data. | ||
37 | SseEncryptionContext = SseGenericHeader + "-Context" | ||
38 | |||
39 | // SseCustomerAlgorithm is the AWS SSE-C algorithm HTTP header key. | ||
40 | SseCustomerAlgorithm = SseGenericHeader + "-Customer-Algorithm" | ||
41 | // SseCustomerKey is the AWS SSE-C encryption key HTTP header key. | ||
42 | SseCustomerKey = SseGenericHeader + "-Customer-Key" | ||
43 | // SseCustomerKeyMD5 is the AWS SSE-C encryption key MD5 HTTP header key. | ||
44 | SseCustomerKeyMD5 = SseGenericHeader + "-Customer-Key-MD5" | ||
45 | |||
46 | // SseCopyCustomerAlgorithm is the AWS SSE-C algorithm HTTP header key for CopyObject API. | ||
47 | SseCopyCustomerAlgorithm = "X-Amz-Copy-Source-Server-Side-Encryption-Customer-Algorithm" | ||
48 | // SseCopyCustomerKey is the AWS SSE-C encryption key HTTP header key for CopyObject API. | ||
49 | SseCopyCustomerKey = "X-Amz-Copy-Source-Server-Side-Encryption-Customer-Key" | ||
50 | // SseCopyCustomerKeyMD5 is the AWS SSE-C encryption key MD5 HTTP header key for CopyObject API. | ||
51 | SseCopyCustomerKeyMD5 = "X-Amz-Copy-Source-Server-Side-Encryption-Customer-Key-MD5" | ||
52 | ) | ||
53 | |||
54 | // PBKDF creates a SSE-C key from the provided password and salt. | ||
55 | // PBKDF is a password-based key derivation function | ||
56 | // which can be used to derive a high-entropy cryptographic | ||
57 | // key from a low-entropy password and a salt. | ||
58 | type PBKDF func(password, salt []byte) ServerSide | ||
59 | |||
60 | // DefaultPBKDF is the default PBKDF. It uses Argon2id with the | ||
61 | // recommended parameters from the RFC draft (1 pass, 64 MB memory, 4 threads). | ||
62 | var DefaultPBKDF PBKDF = func(password, salt []byte) ServerSide { | ||
63 | sse := ssec{} | ||
64 | copy(sse[:], argon2.IDKey(password, salt, 1, 64*1024, 4, 32)) | ||
65 | return sse | ||
66 | } | ||
67 | |||
68 | // Type is the server-side-encryption method. It represents one of | ||
69 | // the following encryption methods: | ||
70 | // - SSE-C: server-side-encryption with customer provided keys | ||
71 | // - KMS: server-side-encryption with managed keys | ||
72 | // - S3: server-side-encryption using S3 storage encryption | ||
73 | type Type string | ||
74 | |||
75 | const ( | ||
76 | // SSEC represents server-side-encryption with customer provided keys | ||
77 | SSEC Type = "SSE-C" | ||
78 | // KMS represents server-side-encryption with managed keys | ||
79 | KMS Type = "KMS" | ||
80 | // S3 represents server-side-encryption using S3 storage encryption | ||
81 | S3 Type = "S3" | ||
82 | ) | ||
83 | |||
84 | // ServerSide is a form of S3 server-side-encryption. | ||
85 | type ServerSide interface { | ||
86 | // Type returns the server-side-encryption method. | ||
87 | Type() Type | ||
88 | |||
89 | // Marshal adds encryption headers to the provided HTTP headers. | ||
90 | // It marks an HTTP request as server-side-encryption request | ||
91 | // and inserts the required data into the headers. | ||
92 | Marshal(h http.Header) | ||
93 | } | ||
94 | |||
95 | // NewSSE returns a server-side-encryption using S3 storage encryption. | ||
96 | // Using SSE-S3 the server will encrypt the object with server-managed keys. | ||
97 | func NewSSE() ServerSide { return s3{} } | ||
98 | |||
99 | // NewSSEKMS returns a new server-side-encryption using SSE-KMS and the provided Key Id and context. | ||
100 | func NewSSEKMS(keyID string, context interface{}) (ServerSide, error) { | ||
101 | if context == nil { | ||
102 | return kms{key: keyID, hasContext: false}, nil | ||
103 | } | ||
104 | json := jsoniter.ConfigCompatibleWithStandardLibrary | ||
105 | serializedContext, err := json.Marshal(context) | ||
106 | if err != nil { | ||
107 | return nil, err | ||
108 | } | ||
109 | return kms{key: keyID, context: serializedContext, hasContext: true}, nil | ||
110 | } | ||
111 | |||
112 | // NewSSEC returns a new server-side-encryption using SSE-C and the provided key. | ||
113 | // The key must be 32 bytes long. | ||
114 | func NewSSEC(key []byte) (ServerSide, error) { | ||
115 | if len(key) != 32 { | ||
116 | return nil, errors.New("encrypt: SSE-C key must be 256 bit long") | ||
117 | } | ||
118 | sse := ssec{} | ||
119 | copy(sse[:], key) | ||
120 | return sse, nil | ||
121 | } | ||
122 | |||
123 | // SSE transforms a SSE-C copy encryption into a SSE-C encryption. | ||
124 | // It is the inverse of SSECopy(...). | ||
125 | // | ||
126 | // If the provided sse is no SSE-C copy encryption SSE returns | ||
127 | // sse unmodified. | ||
128 | func SSE(sse ServerSide) ServerSide { | ||
129 | if sse == nil || sse.Type() != SSEC { | ||
130 | return sse | ||
131 | } | ||
132 | if sse, ok := sse.(ssecCopy); ok { | ||
133 | return ssec(sse) | ||
134 | } | ||
135 | return sse | ||
136 | } | ||
137 | |||
138 | // SSECopy transforms a SSE-C encryption into a SSE-C copy | ||
139 | // encryption. This is required for SSE-C key rotation or a SSE-C | ||
140 | // copy where the source and the destination should be encrypted. | ||
141 | // | ||
142 | // If the provided sse is no SSE-C encryption SSECopy returns | ||
143 | // sse unmodified. | ||
144 | func SSECopy(sse ServerSide) ServerSide { | ||
145 | if sse == nil || sse.Type() != SSEC { | ||
146 | return sse | ||
147 | } | ||
148 | if sse, ok := sse.(ssec); ok { | ||
149 | return ssecCopy(sse) | ||
150 | } | ||
151 | return sse | ||
152 | } | ||
153 | |||
154 | type ssec [32]byte | ||
155 | |||
156 | func (s ssec) Type() Type { return SSEC } | ||
157 | |||
158 | func (s ssec) Marshal(h http.Header) { | ||
159 | keyMD5 := md5.Sum(s[:]) | ||
160 | h.Set(SseCustomerAlgorithm, "AES256") | ||
161 | h.Set(SseCustomerKey, base64.StdEncoding.EncodeToString(s[:])) | ||
162 | h.Set(SseCustomerKeyMD5, base64.StdEncoding.EncodeToString(keyMD5[:])) | ||
163 | } | ||
164 | |||
165 | type ssecCopy [32]byte | ||
166 | |||
167 | func (s ssecCopy) Type() Type { return SSEC } | ||
168 | |||
169 | func (s ssecCopy) Marshal(h http.Header) { | ||
170 | keyMD5 := md5.Sum(s[:]) | ||
171 | h.Set(SseCopyCustomerAlgorithm, "AES256") | ||
172 | h.Set(SseCopyCustomerKey, base64.StdEncoding.EncodeToString(s[:])) | ||
173 | h.Set(SseCopyCustomerKeyMD5, base64.StdEncoding.EncodeToString(keyMD5[:])) | ||
174 | } | ||
175 | |||
176 | type s3 struct{} | ||
177 | |||
178 | func (s s3) Type() Type { return S3 } | ||
179 | |||
180 | func (s s3) Marshal(h http.Header) { h.Set(SseGenericHeader, "AES256") } | ||
181 | |||
182 | type kms struct { | ||
183 | key string | ||
184 | context []byte | ||
185 | hasContext bool | ||
186 | } | ||
187 | |||
188 | func (s kms) Type() Type { return KMS } | ||
189 | |||
190 | func (s kms) Marshal(h http.Header) { | ||
191 | h.Set(SseGenericHeader, "aws:kms") | ||
192 | if s.key != "" { | ||
193 | h.Set(SseKmsKeyID, s.key) | ||
194 | } | ||
195 | if s.hasContext { | ||
196 | h.Set(SseEncryptionContext, base64.StdEncoding.EncodeToString(s.context)) | ||
197 | } | ||
198 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/pkg/lifecycle/lifecycle.go b/vendor/github.com/minio/minio-go/v7/pkg/lifecycle/lifecycle.go deleted file mode 100644 index c52f78c..0000000 --- a/vendor/github.com/minio/minio-go/v7/pkg/lifecycle/lifecycle.go +++ /dev/null | |||
@@ -1,491 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2020 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | // Package lifecycle contains all the lifecycle related data types and marshallers. | ||
19 | package lifecycle | ||
20 | |||
21 | import ( | ||
22 | "encoding/json" | ||
23 | "encoding/xml" | ||
24 | "errors" | ||
25 | "time" | ||
26 | ) | ||
27 | |||
28 | var errMissingStorageClass = errors.New("storage-class cannot be empty") | ||
29 | |||
30 | // AbortIncompleteMultipartUpload structure, not supported yet on MinIO | ||
31 | type AbortIncompleteMultipartUpload struct { | ||
32 | XMLName xml.Name `xml:"AbortIncompleteMultipartUpload,omitempty" json:"-"` | ||
33 | DaysAfterInitiation ExpirationDays `xml:"DaysAfterInitiation,omitempty" json:"DaysAfterInitiation,omitempty"` | ||
34 | } | ||
35 | |||
36 | // IsDaysNull returns true if days field is null | ||
37 | func (n AbortIncompleteMultipartUpload) IsDaysNull() bool { | ||
38 | return n.DaysAfterInitiation == ExpirationDays(0) | ||
39 | } | ||
40 | |||
41 | // MarshalXML if days after initiation is set to non-zero value | ||
42 | func (n AbortIncompleteMultipartUpload) MarshalXML(e *xml.Encoder, start xml.StartElement) error { | ||
43 | if n.IsDaysNull() { | ||
44 | return nil | ||
45 | } | ||
46 | type abortIncompleteMultipartUploadWrapper AbortIncompleteMultipartUpload | ||
47 | return e.EncodeElement(abortIncompleteMultipartUploadWrapper(n), start) | ||
48 | } | ||
49 | |||
50 | // NoncurrentVersionExpiration - Specifies when noncurrent object versions expire. | ||
51 | // Upon expiration, server permanently deletes the noncurrent object versions. | ||
52 | // Set this lifecycle configuration action on a bucket that has versioning enabled | ||
53 | // (or suspended) to request server delete noncurrent object versions at a | ||
54 | // specific period in the object's lifetime. | ||
55 | type NoncurrentVersionExpiration struct { | ||
56 | XMLName xml.Name `xml:"NoncurrentVersionExpiration" json:"-"` | ||
57 | NoncurrentDays ExpirationDays `xml:"NoncurrentDays,omitempty" json:"NoncurrentDays,omitempty"` | ||
58 | NewerNoncurrentVersions int `xml:"NewerNoncurrentVersions,omitempty" json:"NewerNoncurrentVersions,omitempty"` | ||
59 | } | ||
60 | |||
61 | // MarshalXML if n is non-empty, i.e has a non-zero NoncurrentDays or NewerNoncurrentVersions. | ||
62 | func (n NoncurrentVersionExpiration) MarshalXML(e *xml.Encoder, start xml.StartElement) error { | ||
63 | if n.isNull() { | ||
64 | return nil | ||
65 | } | ||
66 | type noncurrentVersionExpirationWrapper NoncurrentVersionExpiration | ||
67 | return e.EncodeElement(noncurrentVersionExpirationWrapper(n), start) | ||
68 | } | ||
69 | |||
70 | // IsDaysNull returns true if days field is null | ||
71 | func (n NoncurrentVersionExpiration) IsDaysNull() bool { | ||
72 | return n.NoncurrentDays == ExpirationDays(0) | ||
73 | } | ||
74 | |||
75 | func (n NoncurrentVersionExpiration) isNull() bool { | ||
76 | return n.IsDaysNull() && n.NewerNoncurrentVersions == 0 | ||
77 | } | ||
78 | |||
79 | // NoncurrentVersionTransition structure, set this action to request server to | ||
80 | // transition noncurrent object versions to different set storage classes | ||
81 | // at a specific period in the object's lifetime. | ||
82 | type NoncurrentVersionTransition struct { | ||
83 | XMLName xml.Name `xml:"NoncurrentVersionTransition,omitempty" json:"-"` | ||
84 | StorageClass string `xml:"StorageClass,omitempty" json:"StorageClass,omitempty"` | ||
85 | NoncurrentDays ExpirationDays `xml:"NoncurrentDays" json:"NoncurrentDays"` | ||
86 | NewerNoncurrentVersions int `xml:"NewerNoncurrentVersions,omitempty" json:"NewerNoncurrentVersions,omitempty"` | ||
87 | } | ||
88 | |||
89 | // IsDaysNull returns true if days field is null | ||
90 | func (n NoncurrentVersionTransition) IsDaysNull() bool { | ||
91 | return n.NoncurrentDays == ExpirationDays(0) | ||
92 | } | ||
93 | |||
94 | // IsStorageClassEmpty returns true if storage class field is empty | ||
95 | func (n NoncurrentVersionTransition) IsStorageClassEmpty() bool { | ||
96 | return n.StorageClass == "" | ||
97 | } | ||
98 | |||
99 | func (n NoncurrentVersionTransition) isNull() bool { | ||
100 | return n.StorageClass == "" | ||
101 | } | ||
102 | |||
103 | // UnmarshalJSON implements NoncurrentVersionTransition JSONify | ||
104 | func (n *NoncurrentVersionTransition) UnmarshalJSON(b []byte) error { | ||
105 | type noncurrentVersionTransition NoncurrentVersionTransition | ||
106 | var nt noncurrentVersionTransition | ||
107 | err := json.Unmarshal(b, &nt) | ||
108 | if err != nil { | ||
109 | return err | ||
110 | } | ||
111 | |||
112 | if nt.StorageClass == "" { | ||
113 | return errMissingStorageClass | ||
114 | } | ||
115 | *n = NoncurrentVersionTransition(nt) | ||
116 | return nil | ||
117 | } | ||
118 | |||
119 | // MarshalXML is extended to leave out | ||
120 | // <NoncurrentVersionTransition></NoncurrentVersionTransition> tags | ||
121 | func (n NoncurrentVersionTransition) MarshalXML(e *xml.Encoder, start xml.StartElement) error { | ||
122 | if n.isNull() { | ||
123 | return nil | ||
124 | } | ||
125 | type noncurrentVersionTransitionWrapper NoncurrentVersionTransition | ||
126 | return e.EncodeElement(noncurrentVersionTransitionWrapper(n), start) | ||
127 | } | ||
128 | |||
129 | // Tag structure key/value pair representing an object tag to apply lifecycle configuration | ||
130 | type Tag struct { | ||
131 | XMLName xml.Name `xml:"Tag,omitempty" json:"-"` | ||
132 | Key string `xml:"Key,omitempty" json:"Key,omitempty"` | ||
133 | Value string `xml:"Value,omitempty" json:"Value,omitempty"` | ||
134 | } | ||
135 | |||
136 | // IsEmpty returns whether this tag is empty or not. | ||
137 | func (tag Tag) IsEmpty() bool { | ||
138 | return tag.Key == "" | ||
139 | } | ||
140 | |||
141 | // Transition structure - transition details of lifecycle configuration | ||
142 | type Transition struct { | ||
143 | XMLName xml.Name `xml:"Transition" json:"-"` | ||
144 | Date ExpirationDate `xml:"Date,omitempty" json:"Date,omitempty"` | ||
145 | StorageClass string `xml:"StorageClass,omitempty" json:"StorageClass,omitempty"` | ||
146 | Days ExpirationDays `xml:"Days" json:"Days"` | ||
147 | } | ||
148 | |||
149 | // UnmarshalJSON returns an error if storage-class is empty. | ||
150 | func (t *Transition) UnmarshalJSON(b []byte) error { | ||
151 | type transition Transition | ||
152 | var tr transition | ||
153 | err := json.Unmarshal(b, &tr) | ||
154 | if err != nil { | ||
155 | return err | ||
156 | } | ||
157 | |||
158 | if tr.StorageClass == "" { | ||
159 | return errMissingStorageClass | ||
160 | } | ||
161 | *t = Transition(tr) | ||
162 | return nil | ||
163 | } | ||
164 | |||
165 | // MarshalJSON customizes json encoding by omitting empty values | ||
166 | func (t Transition) MarshalJSON() ([]byte, error) { | ||
167 | if t.IsNull() { | ||
168 | return nil, nil | ||
169 | } | ||
170 | type transition struct { | ||
171 | Date *ExpirationDate `json:"Date,omitempty"` | ||
172 | StorageClass string `json:"StorageClass,omitempty"` | ||
173 | Days *ExpirationDays `json:"Days"` | ||
174 | } | ||
175 | |||
176 | newt := transition{ | ||
177 | StorageClass: t.StorageClass, | ||
178 | } | ||
179 | |||
180 | if !t.IsDateNull() { | ||
181 | newt.Date = &t.Date | ||
182 | } else { | ||
183 | newt.Days = &t.Days | ||
184 | } | ||
185 | return json.Marshal(newt) | ||
186 | } | ||
187 | |||
188 | // IsDaysNull returns true if days field is null | ||
189 | func (t Transition) IsDaysNull() bool { | ||
190 | return t.Days == ExpirationDays(0) | ||
191 | } | ||
192 | |||
193 | // IsDateNull returns true if date field is null | ||
194 | func (t Transition) IsDateNull() bool { | ||
195 | return t.Date.Time.IsZero() | ||
196 | } | ||
197 | |||
198 | // IsNull returns true if no storage-class is set. | ||
199 | func (t Transition) IsNull() bool { | ||
200 | return t.StorageClass == "" | ||
201 | } | ||
202 | |||
203 | // MarshalXML is transition is non null | ||
204 | func (t Transition) MarshalXML(en *xml.Encoder, startElement xml.StartElement) error { | ||
205 | if t.IsNull() { | ||
206 | return nil | ||
207 | } | ||
208 | type transitionWrapper Transition | ||
209 | return en.EncodeElement(transitionWrapper(t), startElement) | ||
210 | } | ||
211 | |||
212 | // And And Rule for LifecycleTag, to be used in LifecycleRuleFilter | ||
213 | type And struct { | ||
214 | XMLName xml.Name `xml:"And" json:"-"` | ||
215 | Prefix string `xml:"Prefix" json:"Prefix,omitempty"` | ||
216 | Tags []Tag `xml:"Tag" json:"Tags,omitempty"` | ||
217 | ObjectSizeLessThan int64 `xml:"ObjectSizeLessThan,omitempty" json:"ObjectSizeLessThan,omitempty"` | ||
218 | ObjectSizeGreaterThan int64 `xml:"ObjectSizeGreaterThan,omitempty" json:"ObjectSizeGreaterThan,omitempty"` | ||
219 | } | ||
220 | |||
221 | // IsEmpty returns true if Tags field is null | ||
222 | func (a And) IsEmpty() bool { | ||
223 | return len(a.Tags) == 0 && a.Prefix == "" && | ||
224 | a.ObjectSizeLessThan == 0 && a.ObjectSizeGreaterThan == 0 | ||
225 | } | ||
226 | |||
227 | // Filter will be used in selecting rule(s) for lifecycle configuration | ||
228 | type Filter struct { | ||
229 | XMLName xml.Name `xml:"Filter" json:"-"` | ||
230 | And And `xml:"And,omitempty" json:"And,omitempty"` | ||
231 | Prefix string `xml:"Prefix,omitempty" json:"Prefix,omitempty"` | ||
232 | Tag Tag `xml:"Tag,omitempty" json:"Tag,omitempty"` | ||
233 | ObjectSizeLessThan int64 `xml:"ObjectSizeLessThan,omitempty" json:"ObjectSizeLessThan,omitempty"` | ||
234 | ObjectSizeGreaterThan int64 `xml:"ObjectSizeGreaterThan,omitempty" json:"ObjectSizeGreaterThan,omitempty"` | ||
235 | } | ||
236 | |||
237 | // IsNull returns true if all Filter fields are empty. | ||
238 | func (f Filter) IsNull() bool { | ||
239 | return f.Tag.IsEmpty() && f.And.IsEmpty() && f.Prefix == "" && | ||
240 | f.ObjectSizeLessThan == 0 && f.ObjectSizeGreaterThan == 0 | ||
241 | } | ||
242 | |||
243 | // MarshalJSON customizes json encoding by removing empty values. | ||
244 | func (f Filter) MarshalJSON() ([]byte, error) { | ||
245 | type filter struct { | ||
246 | And *And `json:"And,omitempty"` | ||
247 | Prefix string `json:"Prefix,omitempty"` | ||
248 | Tag *Tag `json:"Tag,omitempty"` | ||
249 | ObjectSizeLessThan int64 `json:"ObjectSizeLessThan,omitempty"` | ||
250 | ObjectSizeGreaterThan int64 `json:"ObjectSizeGreaterThan,omitempty"` | ||
251 | } | ||
252 | |||
253 | newf := filter{ | ||
254 | Prefix: f.Prefix, | ||
255 | } | ||
256 | if !f.Tag.IsEmpty() { | ||
257 | newf.Tag = &f.Tag | ||
258 | } | ||
259 | if !f.And.IsEmpty() { | ||
260 | newf.And = &f.And | ||
261 | } | ||
262 | newf.ObjectSizeLessThan = f.ObjectSizeLessThan | ||
263 | newf.ObjectSizeGreaterThan = f.ObjectSizeGreaterThan | ||
264 | return json.Marshal(newf) | ||
265 | } | ||
266 | |||
267 | // MarshalXML - produces the xml representation of the Filter struct | ||
268 | // only one of Prefix, And and Tag should be present in the output. | ||
269 | func (f Filter) MarshalXML(e *xml.Encoder, start xml.StartElement) error { | ||
270 | if err := e.EncodeToken(start); err != nil { | ||
271 | return err | ||
272 | } | ||
273 | |||
274 | switch { | ||
275 | case !f.And.IsEmpty(): | ||
276 | if err := e.EncodeElement(f.And, xml.StartElement{Name: xml.Name{Local: "And"}}); err != nil { | ||
277 | return err | ||
278 | } | ||
279 | case !f.Tag.IsEmpty(): | ||
280 | if err := e.EncodeElement(f.Tag, xml.StartElement{Name: xml.Name{Local: "Tag"}}); err != nil { | ||
281 | return err | ||
282 | } | ||
283 | default: | ||
284 | if f.ObjectSizeLessThan > 0 { | ||
285 | if err := e.EncodeElement(f.ObjectSizeLessThan, xml.StartElement{Name: xml.Name{Local: "ObjectSizeLessThan"}}); err != nil { | ||
286 | return err | ||
287 | } | ||
288 | break | ||
289 | } | ||
290 | if f.ObjectSizeGreaterThan > 0 { | ||
291 | if err := e.EncodeElement(f.ObjectSizeGreaterThan, xml.StartElement{Name: xml.Name{Local: "ObjectSizeGreaterThan"}}); err != nil { | ||
292 | return err | ||
293 | } | ||
294 | break | ||
295 | } | ||
296 | // Print empty Prefix field only when everything else is empty | ||
297 | if err := e.EncodeElement(f.Prefix, xml.StartElement{Name: xml.Name{Local: "Prefix"}}); err != nil { | ||
298 | return err | ||
299 | } | ||
300 | } | ||
301 | |||
302 | return e.EncodeToken(xml.EndElement{Name: start.Name}) | ||
303 | } | ||
304 | |||
305 | // ExpirationDays is a type alias to unmarshal Days in Expiration | ||
306 | type ExpirationDays int | ||
307 | |||
308 | // MarshalXML encodes number of days to expire if it is non-zero and | ||
309 | // encodes empty string otherwise | ||
310 | func (eDays ExpirationDays) MarshalXML(e *xml.Encoder, startElement xml.StartElement) error { | ||
311 | if eDays == 0 { | ||
312 | return nil | ||
313 | } | ||
314 | return e.EncodeElement(int(eDays), startElement) | ||
315 | } | ||
316 | |||
317 | // ExpirationDate is a embedded type containing time.Time to unmarshal | ||
318 | // Date in Expiration | ||
319 | type ExpirationDate struct { | ||
320 | time.Time | ||
321 | } | ||
322 | |||
323 | // MarshalXML encodes expiration date if it is non-zero and encodes | ||
324 | // empty string otherwise | ||
325 | func (eDate ExpirationDate) MarshalXML(e *xml.Encoder, startElement xml.StartElement) error { | ||
326 | if eDate.Time.IsZero() { | ||
327 | return nil | ||
328 | } | ||
329 | return e.EncodeElement(eDate.Format(time.RFC3339), startElement) | ||
330 | } | ||
331 | |||
332 | // ExpireDeleteMarker represents value of ExpiredObjectDeleteMarker field in Expiration XML element. | ||
333 | type ExpireDeleteMarker ExpirationBoolean | ||
334 | |||
335 | // IsEnabled returns true if the auto delete-marker expiration is enabled | ||
336 | func (e ExpireDeleteMarker) IsEnabled() bool { | ||
337 | return bool(e) | ||
338 | } | ||
339 | |||
340 | // ExpirationBoolean represents an XML version of 'bool' type | ||
341 | type ExpirationBoolean bool | ||
342 | |||
343 | // MarshalXML encodes delete marker boolean into an XML form. | ||
344 | func (b ExpirationBoolean) MarshalXML(e *xml.Encoder, startElement xml.StartElement) error { | ||
345 | if !b { | ||
346 | return nil | ||
347 | } | ||
348 | type booleanWrapper ExpirationBoolean | ||
349 | return e.EncodeElement(booleanWrapper(b), startElement) | ||
350 | } | ||
351 | |||
352 | // IsEnabled returns true if the expiration boolean is enabled | ||
353 | func (b ExpirationBoolean) IsEnabled() bool { | ||
354 | return bool(b) | ||
355 | } | ||
356 | |||
357 | // Expiration structure - expiration details of lifecycle configuration | ||
358 | type Expiration struct { | ||
359 | XMLName xml.Name `xml:"Expiration,omitempty" json:"-"` | ||
360 | Date ExpirationDate `xml:"Date,omitempty" json:"Date,omitempty"` | ||
361 | Days ExpirationDays `xml:"Days,omitempty" json:"Days,omitempty"` | ||
362 | DeleteMarker ExpireDeleteMarker `xml:"ExpiredObjectDeleteMarker,omitempty" json:"ExpiredObjectDeleteMarker,omitempty"` | ||
363 | DeleteAll ExpirationBoolean `xml:"ExpiredObjectAllVersions,omitempty" json:"ExpiredObjectAllVersions,omitempty"` | ||
364 | } | ||
365 | |||
366 | // MarshalJSON customizes json encoding by removing empty day/date specification. | ||
367 | func (e Expiration) MarshalJSON() ([]byte, error) { | ||
368 | type expiration struct { | ||
369 | Date *ExpirationDate `json:"Date,omitempty"` | ||
370 | Days *ExpirationDays `json:"Days,omitempty"` | ||
371 | DeleteMarker ExpireDeleteMarker `json:"ExpiredObjectDeleteMarker,omitempty"` | ||
372 | DeleteAll ExpirationBoolean `json:"ExpiredObjectAllVersions,omitempty"` | ||
373 | } | ||
374 | |||
375 | newexp := expiration{ | ||
376 | DeleteMarker: e.DeleteMarker, | ||
377 | DeleteAll: e.DeleteAll, | ||
378 | } | ||
379 | if !e.IsDaysNull() { | ||
380 | newexp.Days = &e.Days | ||
381 | } | ||
382 | if !e.IsDateNull() { | ||
383 | newexp.Date = &e.Date | ||
384 | } | ||
385 | return json.Marshal(newexp) | ||
386 | } | ||
387 | |||
388 | // IsDaysNull returns true if days field is null | ||
389 | func (e Expiration) IsDaysNull() bool { | ||
390 | return e.Days == ExpirationDays(0) | ||
391 | } | ||
392 | |||
393 | // IsDateNull returns true if date field is null | ||
394 | func (e Expiration) IsDateNull() bool { | ||
395 | return e.Date.Time.IsZero() | ||
396 | } | ||
397 | |||
398 | // IsDeleteMarkerExpirationEnabled returns true if the auto-expiration of delete marker is enabled | ||
399 | func (e Expiration) IsDeleteMarkerExpirationEnabled() bool { | ||
400 | return e.DeleteMarker.IsEnabled() | ||
401 | } | ||
402 | |||
403 | // IsNull returns true if both date and days fields are null | ||
404 | func (e Expiration) IsNull() bool { | ||
405 | return e.IsDaysNull() && e.IsDateNull() && !e.IsDeleteMarkerExpirationEnabled() | ||
406 | } | ||
407 | |||
408 | // MarshalXML is expiration is non null | ||
409 | func (e Expiration) MarshalXML(en *xml.Encoder, startElement xml.StartElement) error { | ||
410 | if e.IsNull() { | ||
411 | return nil | ||
412 | } | ||
413 | type expirationWrapper Expiration | ||
414 | return en.EncodeElement(expirationWrapper(e), startElement) | ||
415 | } | ||
416 | |||
417 | // MarshalJSON customizes json encoding by omitting empty values | ||
418 | func (r Rule) MarshalJSON() ([]byte, error) { | ||
419 | type rule struct { | ||
420 | AbortIncompleteMultipartUpload *AbortIncompleteMultipartUpload `json:"AbortIncompleteMultipartUpload,omitempty"` | ||
421 | Expiration *Expiration `json:"Expiration,omitempty"` | ||
422 | ID string `json:"ID"` | ||
423 | RuleFilter *Filter `json:"Filter,omitempty"` | ||
424 | NoncurrentVersionExpiration *NoncurrentVersionExpiration `json:"NoncurrentVersionExpiration,omitempty"` | ||
425 | NoncurrentVersionTransition *NoncurrentVersionTransition `json:"NoncurrentVersionTransition,omitempty"` | ||
426 | Prefix string `json:"Prefix,omitempty"` | ||
427 | Status string `json:"Status"` | ||
428 | Transition *Transition `json:"Transition,omitempty"` | ||
429 | } | ||
430 | newr := rule{ | ||
431 | Prefix: r.Prefix, | ||
432 | Status: r.Status, | ||
433 | ID: r.ID, | ||
434 | } | ||
435 | |||
436 | if !r.RuleFilter.IsNull() { | ||
437 | newr.RuleFilter = &r.RuleFilter | ||
438 | } | ||
439 | if !r.AbortIncompleteMultipartUpload.IsDaysNull() { | ||
440 | newr.AbortIncompleteMultipartUpload = &r.AbortIncompleteMultipartUpload | ||
441 | } | ||
442 | if !r.Expiration.IsNull() { | ||
443 | newr.Expiration = &r.Expiration | ||
444 | } | ||
445 | if !r.Transition.IsNull() { | ||
446 | newr.Transition = &r.Transition | ||
447 | } | ||
448 | if !r.NoncurrentVersionExpiration.isNull() { | ||
449 | newr.NoncurrentVersionExpiration = &r.NoncurrentVersionExpiration | ||
450 | } | ||
451 | if !r.NoncurrentVersionTransition.isNull() { | ||
452 | newr.NoncurrentVersionTransition = &r.NoncurrentVersionTransition | ||
453 | } | ||
454 | |||
455 | return json.Marshal(newr) | ||
456 | } | ||
457 | |||
458 | // Rule represents a single rule in lifecycle configuration | ||
459 | type Rule struct { | ||
460 | XMLName xml.Name `xml:"Rule,omitempty" json:"-"` | ||
461 | AbortIncompleteMultipartUpload AbortIncompleteMultipartUpload `xml:"AbortIncompleteMultipartUpload,omitempty" json:"AbortIncompleteMultipartUpload,omitempty"` | ||
462 | Expiration Expiration `xml:"Expiration,omitempty" json:"Expiration,omitempty"` | ||
463 | ID string `xml:"ID" json:"ID"` | ||
464 | RuleFilter Filter `xml:"Filter,omitempty" json:"Filter,omitempty"` | ||
465 | NoncurrentVersionExpiration NoncurrentVersionExpiration `xml:"NoncurrentVersionExpiration,omitempty" json:"NoncurrentVersionExpiration,omitempty"` | ||
466 | NoncurrentVersionTransition NoncurrentVersionTransition `xml:"NoncurrentVersionTransition,omitempty" json:"NoncurrentVersionTransition,omitempty"` | ||
467 | Prefix string `xml:"Prefix,omitempty" json:"Prefix,omitempty"` | ||
468 | Status string `xml:"Status" json:"Status"` | ||
469 | Transition Transition `xml:"Transition,omitempty" json:"Transition,omitempty"` | ||
470 | } | ||
471 | |||
472 | // Configuration is a collection of Rule objects. | ||
473 | type Configuration struct { | ||
474 | XMLName xml.Name `xml:"LifecycleConfiguration,omitempty" json:"-"` | ||
475 | Rules []Rule `xml:"Rule"` | ||
476 | } | ||
477 | |||
478 | // Empty check if lifecycle configuration is empty | ||
479 | func (c *Configuration) Empty() bool { | ||
480 | if c == nil { | ||
481 | return true | ||
482 | } | ||
483 | return len(c.Rules) == 0 | ||
484 | } | ||
485 | |||
486 | // NewConfiguration initializes a fresh lifecycle configuration | ||
487 | // for manipulation, such as setting and removing lifecycle rules | ||
488 | // and filters. | ||
489 | func NewConfiguration() *Configuration { | ||
490 | return &Configuration{} | ||
491 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/pkg/notification/info.go b/vendor/github.com/minio/minio-go/v7/pkg/notification/info.go deleted file mode 100644 index 126661a..0000000 --- a/vendor/github.com/minio/minio-go/v7/pkg/notification/info.go +++ /dev/null | |||
@@ -1,78 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2017-2020 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package notification | ||
19 | |||
20 | // Indentity represents the user id, this is a compliance field. | ||
21 | type identity struct { | ||
22 | PrincipalID string `json:"principalId"` | ||
23 | } | ||
24 | |||
25 | // event bucket metadata. | ||
26 | type bucketMeta struct { | ||
27 | Name string `json:"name"` | ||
28 | OwnerIdentity identity `json:"ownerIdentity"` | ||
29 | ARN string `json:"arn"` | ||
30 | } | ||
31 | |||
32 | // event object metadata. | ||
33 | type objectMeta struct { | ||
34 | Key string `json:"key"` | ||
35 | Size int64 `json:"size,omitempty"` | ||
36 | ETag string `json:"eTag,omitempty"` | ||
37 | ContentType string `json:"contentType,omitempty"` | ||
38 | UserMetadata map[string]string `json:"userMetadata,omitempty"` | ||
39 | VersionID string `json:"versionId,omitempty"` | ||
40 | Sequencer string `json:"sequencer"` | ||
41 | } | ||
42 | |||
43 | // event server specific metadata. | ||
44 | type eventMeta struct { | ||
45 | SchemaVersion string `json:"s3SchemaVersion"` | ||
46 | ConfigurationID string `json:"configurationId"` | ||
47 | Bucket bucketMeta `json:"bucket"` | ||
48 | Object objectMeta `json:"object"` | ||
49 | } | ||
50 | |||
51 | // sourceInfo represents information on the client that | ||
52 | // triggered the event notification. | ||
53 | type sourceInfo struct { | ||
54 | Host string `json:"host"` | ||
55 | Port string `json:"port"` | ||
56 | UserAgent string `json:"userAgent"` | ||
57 | } | ||
58 | |||
59 | // Event represents an Amazon an S3 bucket notification event. | ||
60 | type Event struct { | ||
61 | EventVersion string `json:"eventVersion"` | ||
62 | EventSource string `json:"eventSource"` | ||
63 | AwsRegion string `json:"awsRegion"` | ||
64 | EventTime string `json:"eventTime"` | ||
65 | EventName string `json:"eventName"` | ||
66 | UserIdentity identity `json:"userIdentity"` | ||
67 | RequestParameters map[string]string `json:"requestParameters"` | ||
68 | ResponseElements map[string]string `json:"responseElements"` | ||
69 | S3 eventMeta `json:"s3"` | ||
70 | Source sourceInfo `json:"source"` | ||
71 | } | ||
72 | |||
73 | // Info - represents the collection of notification events, additionally | ||
74 | // also reports errors if any while listening on bucket notifications. | ||
75 | type Info struct { | ||
76 | Records []Event | ||
77 | Err error | ||
78 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/pkg/notification/notification.go b/vendor/github.com/minio/minio-go/v7/pkg/notification/notification.go deleted file mode 100644 index a44799d..0000000 --- a/vendor/github.com/minio/minio-go/v7/pkg/notification/notification.go +++ /dev/null | |||
@@ -1,440 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2020 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package notification | ||
19 | |||
20 | import ( | ||
21 | "encoding/xml" | ||
22 | "errors" | ||
23 | "fmt" | ||
24 | "strings" | ||
25 | |||
26 | "github.com/minio/minio-go/v7/pkg/set" | ||
27 | ) | ||
28 | |||
29 | // EventType is a S3 notification event associated to the bucket notification configuration | ||
30 | type EventType string | ||
31 | |||
32 | // The role of all event types are described in : | ||
33 | // | ||
34 | // http://docs.aws.amazon.com/AmazonS3/latest/dev/NotificationHowTo.html#notification-how-to-event-types-and-destinations | ||
35 | const ( | ||
36 | ObjectCreatedAll EventType = "s3:ObjectCreated:*" | ||
37 | ObjectCreatedPut EventType = "s3:ObjectCreated:Put" | ||
38 | ObjectCreatedPost EventType = "s3:ObjectCreated:Post" | ||
39 | ObjectCreatedCopy EventType = "s3:ObjectCreated:Copy" | ||
40 | ObjectCreatedDeleteTagging EventType = "s3:ObjectCreated:DeleteTagging" | ||
41 | ObjectCreatedCompleteMultipartUpload EventType = "s3:ObjectCreated:CompleteMultipartUpload" | ||
42 | ObjectCreatedPutLegalHold EventType = "s3:ObjectCreated:PutLegalHold" | ||
43 | ObjectCreatedPutRetention EventType = "s3:ObjectCreated:PutRetention" | ||
44 | ObjectCreatedPutTagging EventType = "s3:ObjectCreated:PutTagging" | ||
45 | ObjectAccessedGet EventType = "s3:ObjectAccessed:Get" | ||
46 | ObjectAccessedHead EventType = "s3:ObjectAccessed:Head" | ||
47 | ObjectAccessedGetRetention EventType = "s3:ObjectAccessed:GetRetention" | ||
48 | ObjectAccessedGetLegalHold EventType = "s3:ObjectAccessed:GetLegalHold" | ||
49 | ObjectAccessedAll EventType = "s3:ObjectAccessed:*" | ||
50 | ObjectRemovedAll EventType = "s3:ObjectRemoved:*" | ||
51 | ObjectRemovedDelete EventType = "s3:ObjectRemoved:Delete" | ||
52 | ObjectRemovedDeleteMarkerCreated EventType = "s3:ObjectRemoved:DeleteMarkerCreated" | ||
53 | ObjectReducedRedundancyLostObject EventType = "s3:ReducedRedundancyLostObject" | ||
54 | ObjectTransitionAll EventType = "s3:ObjectTransition:*" | ||
55 | ObjectTransitionFailed EventType = "s3:ObjectTransition:Failed" | ||
56 | ObjectTransitionComplete EventType = "s3:ObjectTransition:Complete" | ||
57 | ObjectTransitionPost EventType = "s3:ObjectRestore:Post" | ||
58 | ObjectTransitionCompleted EventType = "s3:ObjectRestore:Completed" | ||
59 | ObjectReplicationAll EventType = "s3:Replication:*" | ||
60 | ObjectReplicationOperationCompletedReplication EventType = "s3:Replication:OperationCompletedReplication" | ||
61 | ObjectReplicationOperationFailedReplication EventType = "s3:Replication:OperationFailedReplication" | ||
62 | ObjectReplicationOperationMissedThreshold EventType = "s3:Replication:OperationMissedThreshold" | ||
63 | ObjectReplicationOperationNotTracked EventType = "s3:Replication:OperationNotTracked" | ||
64 | ObjectReplicationOperationReplicatedAfterThreshold EventType = "s3:Replication:OperationReplicatedAfterThreshold" | ||
65 | ObjectScannerManyVersions EventType = "s3:Scanner:ManyVersions" | ||
66 | ObjectScannerBigPrefix EventType = "s3:Scanner:BigPrefix" | ||
67 | ObjectScannerAll EventType = "s3:Scanner:*" | ||
68 | BucketCreatedAll EventType = "s3:BucketCreated:*" | ||
69 | BucketRemovedAll EventType = "s3:BucketRemoved:*" | ||
70 | ) | ||
71 | |||
72 | // FilterRule - child of S3Key, a tag in the notification xml which | ||
73 | // carries suffix/prefix filters | ||
74 | type FilterRule struct { | ||
75 | Name string `xml:"Name"` | ||
76 | Value string `xml:"Value"` | ||
77 | } | ||
78 | |||
79 | // S3Key - child of Filter, a tag in the notification xml which | ||
80 | // carries suffix/prefix filters | ||
81 | type S3Key struct { | ||
82 | FilterRules []FilterRule `xml:"FilterRule,omitempty"` | ||
83 | } | ||
84 | |||
85 | // Filter - a tag in the notification xml structure which carries | ||
86 | // suffix/prefix filters | ||
87 | type Filter struct { | ||
88 | S3Key S3Key `xml:"S3Key,omitempty"` | ||
89 | } | ||
90 | |||
91 | // Arn - holds ARN information that will be sent to the web service, | ||
92 | // ARN desciption can be found in http://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html | ||
93 | type Arn struct { | ||
94 | Partition string | ||
95 | Service string | ||
96 | Region string | ||
97 | AccountID string | ||
98 | Resource string | ||
99 | } | ||
100 | |||
101 | // NewArn creates new ARN based on the given partition, service, region, account id and resource | ||
102 | func NewArn(partition, service, region, accountID, resource string) Arn { | ||
103 | return Arn{ | ||
104 | Partition: partition, | ||
105 | Service: service, | ||
106 | Region: region, | ||
107 | AccountID: accountID, | ||
108 | Resource: resource, | ||
109 | } | ||
110 | } | ||
111 | |||
112 | var ( | ||
113 | // ErrInvalidArnPrefix is returned when ARN string format does not start with 'arn' | ||
114 | ErrInvalidArnPrefix = errors.New("invalid ARN format, must start with 'arn:'") | ||
115 | // ErrInvalidArnFormat is returned when ARN string format is not valid | ||
116 | ErrInvalidArnFormat = errors.New("invalid ARN format, must be 'arn:<partition>:<service>:<region>:<accountID>:<resource>'") | ||
117 | ) | ||
118 | |||
119 | // NewArnFromString parses string representation of ARN into Arn object. | ||
120 | // Returns an error if the string format is incorrect. | ||
121 | func NewArnFromString(arn string) (Arn, error) { | ||
122 | parts := strings.Split(arn, ":") | ||
123 | if len(parts) != 6 { | ||
124 | return Arn{}, ErrInvalidArnFormat | ||
125 | } | ||
126 | if parts[0] != "arn" { | ||
127 | return Arn{}, ErrInvalidArnPrefix | ||
128 | } | ||
129 | |||
130 | return NewArn(parts[1], parts[2], parts[3], parts[4], parts[5]), nil | ||
131 | } | ||
132 | |||
133 | // String returns the string format of the ARN | ||
134 | func (arn Arn) String() string { | ||
135 | return "arn:" + arn.Partition + ":" + arn.Service + ":" + arn.Region + ":" + arn.AccountID + ":" + arn.Resource | ||
136 | } | ||
137 | |||
138 | // Config - represents one single notification configuration | ||
139 | // such as topic, queue or lambda configuration. | ||
140 | type Config struct { | ||
141 | ID string `xml:"Id,omitempty"` | ||
142 | Arn Arn `xml:"-"` | ||
143 | Events []EventType `xml:"Event"` | ||
144 | Filter *Filter `xml:"Filter,omitempty"` | ||
145 | } | ||
146 | |||
147 | // NewConfig creates one notification config and sets the given ARN | ||
148 | func NewConfig(arn Arn) Config { | ||
149 | return Config{Arn: arn, Filter: &Filter{}} | ||
150 | } | ||
151 | |||
152 | // AddEvents adds one event to the current notification config | ||
153 | func (t *Config) AddEvents(events ...EventType) { | ||
154 | t.Events = append(t.Events, events...) | ||
155 | } | ||
156 | |||
157 | // AddFilterSuffix sets the suffix configuration to the current notification config | ||
158 | func (t *Config) AddFilterSuffix(suffix string) { | ||
159 | if t.Filter == nil { | ||
160 | t.Filter = &Filter{} | ||
161 | } | ||
162 | newFilterRule := FilterRule{Name: "suffix", Value: suffix} | ||
163 | // Replace any suffix rule if existing and add to the list otherwise | ||
164 | for index := range t.Filter.S3Key.FilterRules { | ||
165 | if t.Filter.S3Key.FilterRules[index].Name == "suffix" { | ||
166 | t.Filter.S3Key.FilterRules[index] = newFilterRule | ||
167 | return | ||
168 | } | ||
169 | } | ||
170 | t.Filter.S3Key.FilterRules = append(t.Filter.S3Key.FilterRules, newFilterRule) | ||
171 | } | ||
172 | |||
173 | // AddFilterPrefix sets the prefix configuration to the current notification config | ||
174 | func (t *Config) AddFilterPrefix(prefix string) { | ||
175 | if t.Filter == nil { | ||
176 | t.Filter = &Filter{} | ||
177 | } | ||
178 | newFilterRule := FilterRule{Name: "prefix", Value: prefix} | ||
179 | // Replace any prefix rule if existing and add to the list otherwise | ||
180 | for index := range t.Filter.S3Key.FilterRules { | ||
181 | if t.Filter.S3Key.FilterRules[index].Name == "prefix" { | ||
182 | t.Filter.S3Key.FilterRules[index] = newFilterRule | ||
183 | return | ||
184 | } | ||
185 | } | ||
186 | t.Filter.S3Key.FilterRules = append(t.Filter.S3Key.FilterRules, newFilterRule) | ||
187 | } | ||
188 | |||
189 | // EqualEventTypeList tells whether a and b contain the same events | ||
190 | func EqualEventTypeList(a, b []EventType) bool { | ||
191 | if len(a) != len(b) { | ||
192 | return false | ||
193 | } | ||
194 | setA := set.NewStringSet() | ||
195 | for _, i := range a { | ||
196 | setA.Add(string(i)) | ||
197 | } | ||
198 | |||
199 | setB := set.NewStringSet() | ||
200 | for _, i := range b { | ||
201 | setB.Add(string(i)) | ||
202 | } | ||
203 | |||
204 | return setA.Difference(setB).IsEmpty() | ||
205 | } | ||
206 | |||
207 | // EqualFilterRuleList tells whether a and b contain the same filters | ||
208 | func EqualFilterRuleList(a, b []FilterRule) bool { | ||
209 | if len(a) != len(b) { | ||
210 | return false | ||
211 | } | ||
212 | |||
213 | setA := set.NewStringSet() | ||
214 | for _, i := range a { | ||
215 | setA.Add(fmt.Sprintf("%s-%s", i.Name, i.Value)) | ||
216 | } | ||
217 | |||
218 | setB := set.NewStringSet() | ||
219 | for _, i := range b { | ||
220 | setB.Add(fmt.Sprintf("%s-%s", i.Name, i.Value)) | ||
221 | } | ||
222 | |||
223 | return setA.Difference(setB).IsEmpty() | ||
224 | } | ||
225 | |||
226 | // Equal returns whether this `Config` is equal to another defined by the passed parameters | ||
227 | func (t *Config) Equal(events []EventType, prefix, suffix string) bool { | ||
228 | if t == nil { | ||
229 | return false | ||
230 | } | ||
231 | |||
232 | // Compare events | ||
233 | passEvents := EqualEventTypeList(t.Events, events) | ||
234 | |||
235 | // Compare filters | ||
236 | var newFilterRules []FilterRule | ||
237 | if prefix != "" { | ||
238 | newFilterRules = append(newFilterRules, FilterRule{Name: "prefix", Value: prefix}) | ||
239 | } | ||
240 | if suffix != "" { | ||
241 | newFilterRules = append(newFilterRules, FilterRule{Name: "suffix", Value: suffix}) | ||
242 | } | ||
243 | |||
244 | var currentFilterRules []FilterRule | ||
245 | if t.Filter != nil { | ||
246 | currentFilterRules = t.Filter.S3Key.FilterRules | ||
247 | } | ||
248 | |||
249 | passFilters := EqualFilterRuleList(currentFilterRules, newFilterRules) | ||
250 | return passEvents && passFilters | ||
251 | } | ||
252 | |||
253 | // TopicConfig carries one single topic notification configuration | ||
254 | type TopicConfig struct { | ||
255 | Config | ||
256 | Topic string `xml:"Topic"` | ||
257 | } | ||
258 | |||
259 | // QueueConfig carries one single queue notification configuration | ||
260 | type QueueConfig struct { | ||
261 | Config | ||
262 | Queue string `xml:"Queue"` | ||
263 | } | ||
264 | |||
265 | // LambdaConfig carries one single cloudfunction notification configuration | ||
266 | type LambdaConfig struct { | ||
267 | Config | ||
268 | Lambda string `xml:"CloudFunction"` | ||
269 | } | ||
270 | |||
271 | // Configuration - the struct that represents the whole XML to be sent to the web service | ||
272 | type Configuration struct { | ||
273 | XMLName xml.Name `xml:"NotificationConfiguration"` | ||
274 | LambdaConfigs []LambdaConfig `xml:"CloudFunctionConfiguration"` | ||
275 | TopicConfigs []TopicConfig `xml:"TopicConfiguration"` | ||
276 | QueueConfigs []QueueConfig `xml:"QueueConfiguration"` | ||
277 | } | ||
278 | |||
279 | // AddTopic adds a given topic config to the general bucket notification config | ||
280 | func (b *Configuration) AddTopic(topicConfig Config) bool { | ||
281 | newTopicConfig := TopicConfig{Config: topicConfig, Topic: topicConfig.Arn.String()} | ||
282 | for _, n := range b.TopicConfigs { | ||
283 | // If new config matches existing one | ||
284 | if n.Topic == newTopicConfig.Arn.String() && newTopicConfig.Filter == n.Filter { | ||
285 | |||
286 | existingConfig := set.NewStringSet() | ||
287 | for _, v := range n.Events { | ||
288 | existingConfig.Add(string(v)) | ||
289 | } | ||
290 | |||
291 | newConfig := set.NewStringSet() | ||
292 | for _, v := range topicConfig.Events { | ||
293 | newConfig.Add(string(v)) | ||
294 | } | ||
295 | |||
296 | if !newConfig.Intersection(existingConfig).IsEmpty() { | ||
297 | return false | ||
298 | } | ||
299 | } | ||
300 | } | ||
301 | b.TopicConfigs = append(b.TopicConfigs, newTopicConfig) | ||
302 | return true | ||
303 | } | ||
304 | |||
305 | // AddQueue adds a given queue config to the general bucket notification config | ||
306 | func (b *Configuration) AddQueue(queueConfig Config) bool { | ||
307 | newQueueConfig := QueueConfig{Config: queueConfig, Queue: queueConfig.Arn.String()} | ||
308 | for _, n := range b.QueueConfigs { | ||
309 | if n.Queue == newQueueConfig.Arn.String() && newQueueConfig.Filter == n.Filter { | ||
310 | |||
311 | existingConfig := set.NewStringSet() | ||
312 | for _, v := range n.Events { | ||
313 | existingConfig.Add(string(v)) | ||
314 | } | ||
315 | |||
316 | newConfig := set.NewStringSet() | ||
317 | for _, v := range queueConfig.Events { | ||
318 | newConfig.Add(string(v)) | ||
319 | } | ||
320 | |||
321 | if !newConfig.Intersection(existingConfig).IsEmpty() { | ||
322 | return false | ||
323 | } | ||
324 | } | ||
325 | } | ||
326 | b.QueueConfigs = append(b.QueueConfigs, newQueueConfig) | ||
327 | return true | ||
328 | } | ||
329 | |||
330 | // AddLambda adds a given lambda config to the general bucket notification config | ||
331 | func (b *Configuration) AddLambda(lambdaConfig Config) bool { | ||
332 | newLambdaConfig := LambdaConfig{Config: lambdaConfig, Lambda: lambdaConfig.Arn.String()} | ||
333 | for _, n := range b.LambdaConfigs { | ||
334 | if n.Lambda == newLambdaConfig.Arn.String() && newLambdaConfig.Filter == n.Filter { | ||
335 | |||
336 | existingConfig := set.NewStringSet() | ||
337 | for _, v := range n.Events { | ||
338 | existingConfig.Add(string(v)) | ||
339 | } | ||
340 | |||
341 | newConfig := set.NewStringSet() | ||
342 | for _, v := range lambdaConfig.Events { | ||
343 | newConfig.Add(string(v)) | ||
344 | } | ||
345 | |||
346 | if !newConfig.Intersection(existingConfig).IsEmpty() { | ||
347 | return false | ||
348 | } | ||
349 | } | ||
350 | } | ||
351 | b.LambdaConfigs = append(b.LambdaConfigs, newLambdaConfig) | ||
352 | return true | ||
353 | } | ||
354 | |||
355 | // RemoveTopicByArn removes all topic configurations that match the exact specified ARN | ||
356 | func (b *Configuration) RemoveTopicByArn(arn Arn) { | ||
357 | var topics []TopicConfig | ||
358 | for _, topic := range b.TopicConfigs { | ||
359 | if topic.Topic != arn.String() { | ||
360 | topics = append(topics, topic) | ||
361 | } | ||
362 | } | ||
363 | b.TopicConfigs = topics | ||
364 | } | ||
365 | |||
366 | // ErrNoConfigMatch is returned when a notification configuration (sqs,sns,lambda) is not found when trying to delete | ||
367 | var ErrNoConfigMatch = errors.New("no notification configuration matched") | ||
368 | |||
369 | // RemoveTopicByArnEventsPrefixSuffix removes a topic configuration that match the exact specified ARN, events, prefix and suffix | ||
370 | func (b *Configuration) RemoveTopicByArnEventsPrefixSuffix(arn Arn, events []EventType, prefix, suffix string) error { | ||
371 | removeIndex := -1 | ||
372 | for i, v := range b.TopicConfigs { | ||
373 | // if it matches events and filters, mark the index for deletion | ||
374 | if v.Topic == arn.String() && v.Config.Equal(events, prefix, suffix) { | ||
375 | removeIndex = i | ||
376 | break // since we have at most one matching config | ||
377 | } | ||
378 | } | ||
379 | if removeIndex >= 0 { | ||
380 | b.TopicConfigs = append(b.TopicConfigs[:removeIndex], b.TopicConfigs[removeIndex+1:]...) | ||
381 | return nil | ||
382 | } | ||
383 | return ErrNoConfigMatch | ||
384 | } | ||
385 | |||
386 | // RemoveQueueByArn removes all queue configurations that match the exact specified ARN | ||
387 | func (b *Configuration) RemoveQueueByArn(arn Arn) { | ||
388 | var queues []QueueConfig | ||
389 | for _, queue := range b.QueueConfigs { | ||
390 | if queue.Queue != arn.String() { | ||
391 | queues = append(queues, queue) | ||
392 | } | ||
393 | } | ||
394 | b.QueueConfigs = queues | ||
395 | } | ||
396 | |||
397 | // RemoveQueueByArnEventsPrefixSuffix removes a queue configuration that match the exact specified ARN, events, prefix and suffix | ||
398 | func (b *Configuration) RemoveQueueByArnEventsPrefixSuffix(arn Arn, events []EventType, prefix, suffix string) error { | ||
399 | removeIndex := -1 | ||
400 | for i, v := range b.QueueConfigs { | ||
401 | // if it matches events and filters, mark the index for deletion | ||
402 | if v.Queue == arn.String() && v.Config.Equal(events, prefix, suffix) { | ||
403 | removeIndex = i | ||
404 | break // since we have at most one matching config | ||
405 | } | ||
406 | } | ||
407 | if removeIndex >= 0 { | ||
408 | b.QueueConfigs = append(b.QueueConfigs[:removeIndex], b.QueueConfigs[removeIndex+1:]...) | ||
409 | return nil | ||
410 | } | ||
411 | return ErrNoConfigMatch | ||
412 | } | ||
413 | |||
414 | // RemoveLambdaByArn removes all lambda configurations that match the exact specified ARN | ||
415 | func (b *Configuration) RemoveLambdaByArn(arn Arn) { | ||
416 | var lambdas []LambdaConfig | ||
417 | for _, lambda := range b.LambdaConfigs { | ||
418 | if lambda.Lambda != arn.String() { | ||
419 | lambdas = append(lambdas, lambda) | ||
420 | } | ||
421 | } | ||
422 | b.LambdaConfigs = lambdas | ||
423 | } | ||
424 | |||
425 | // RemoveLambdaByArnEventsPrefixSuffix removes a topic configuration that match the exact specified ARN, events, prefix and suffix | ||
426 | func (b *Configuration) RemoveLambdaByArnEventsPrefixSuffix(arn Arn, events []EventType, prefix, suffix string) error { | ||
427 | removeIndex := -1 | ||
428 | for i, v := range b.LambdaConfigs { | ||
429 | // if it matches events and filters, mark the index for deletion | ||
430 | if v.Lambda == arn.String() && v.Config.Equal(events, prefix, suffix) { | ||
431 | removeIndex = i | ||
432 | break // since we have at most one matching config | ||
433 | } | ||
434 | } | ||
435 | if removeIndex >= 0 { | ||
436 | b.LambdaConfigs = append(b.LambdaConfigs[:removeIndex], b.LambdaConfigs[removeIndex+1:]...) | ||
437 | return nil | ||
438 | } | ||
439 | return ErrNoConfigMatch | ||
440 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/pkg/replication/replication.go b/vendor/github.com/minio/minio-go/v7/pkg/replication/replication.go deleted file mode 100644 index 0abbf6e..0000000 --- a/vendor/github.com/minio/minio-go/v7/pkg/replication/replication.go +++ /dev/null | |||
@@ -1,971 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Client (C) 2020 MinIO, Inc. | ||
3 | * | ||
4 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
5 | * you may not use this file except in compliance with the License. | ||
6 | * You may obtain a copy of the License at | ||
7 | * | ||
8 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
9 | * | ||
10 | * Unless required by applicable law or agreed to in writing, software | ||
11 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
12 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
13 | * See the License for the specific language governing permissions and | ||
14 | * limitations under the License. | ||
15 | */ | ||
16 | |||
17 | package replication | ||
18 | |||
19 | import ( | ||
20 | "bytes" | ||
21 | "encoding/xml" | ||
22 | "fmt" | ||
23 | "math" | ||
24 | "strconv" | ||
25 | "strings" | ||
26 | "time" | ||
27 | "unicode/utf8" | ||
28 | |||
29 | "github.com/rs/xid" | ||
30 | ) | ||
31 | |||
32 | var errInvalidFilter = fmt.Errorf("invalid filter") | ||
33 | |||
34 | // OptionType specifies operation to be performed on config | ||
35 | type OptionType string | ||
36 | |||
37 | const ( | ||
38 | // AddOption specifies addition of rule to config | ||
39 | AddOption OptionType = "Add" | ||
40 | // SetOption specifies modification of existing rule to config | ||
41 | SetOption OptionType = "Set" | ||
42 | |||
43 | // RemoveOption specifies rule options are for removing a rule | ||
44 | RemoveOption OptionType = "Remove" | ||
45 | // ImportOption is for getting current config | ||
46 | ImportOption OptionType = "Import" | ||
47 | ) | ||
48 | |||
49 | // Options represents options to set a replication configuration rule | ||
50 | type Options struct { | ||
51 | Op OptionType | ||
52 | RoleArn string | ||
53 | ID string | ||
54 | Prefix string | ||
55 | RuleStatus string | ||
56 | Priority string | ||
57 | TagString string | ||
58 | StorageClass string | ||
59 | DestBucket string | ||
60 | IsTagSet bool | ||
61 | IsSCSet bool | ||
62 | ReplicateDeletes string // replicate versioned deletes | ||
63 | ReplicateDeleteMarkers string // replicate soft deletes | ||
64 | ReplicaSync string // replicate replica metadata modifications | ||
65 | ExistingObjectReplicate string | ||
66 | } | ||
67 | |||
68 | // Tags returns a slice of tags for a rule | ||
69 | func (opts Options) Tags() ([]Tag, error) { | ||
70 | var tagList []Tag | ||
71 | tagTokens := strings.Split(opts.TagString, "&") | ||
72 | for _, tok := range tagTokens { | ||
73 | if tok == "" { | ||
74 | break | ||
75 | } | ||
76 | kv := strings.SplitN(tok, "=", 2) | ||
77 | if len(kv) != 2 { | ||
78 | return []Tag{}, fmt.Errorf("tags should be entered as comma separated k=v pairs") | ||
79 | } | ||
80 | tagList = append(tagList, Tag{ | ||
81 | Key: kv[0], | ||
82 | Value: kv[1], | ||
83 | }) | ||
84 | } | ||
85 | return tagList, nil | ||
86 | } | ||
87 | |||
88 | // Config - replication configuration specified in | ||
89 | // https://docs.aws.amazon.com/AmazonS3/latest/dev/replication-add-config.html | ||
90 | type Config struct { | ||
91 | XMLName xml.Name `xml:"ReplicationConfiguration" json:"-"` | ||
92 | Rules []Rule `xml:"Rule" json:"Rules"` | ||
93 | Role string `xml:"Role" json:"Role"` | ||
94 | } | ||
95 | |||
96 | // Empty returns true if config is not set | ||
97 | func (c *Config) Empty() bool { | ||
98 | return len(c.Rules) == 0 | ||
99 | } | ||
100 | |||
101 | // AddRule adds a new rule to existing replication config. If a rule exists with the | ||
102 | // same ID, then the rule is replaced. | ||
103 | func (c *Config) AddRule(opts Options) error { | ||
104 | priority, err := strconv.Atoi(opts.Priority) | ||
105 | if err != nil { | ||
106 | return err | ||
107 | } | ||
108 | var compatSw bool // true if RoleArn is used with new mc client and older minio version prior to multisite | ||
109 | if opts.RoleArn != "" { | ||
110 | tokens := strings.Split(opts.RoleArn, ":") | ||
111 | if len(tokens) != 6 { | ||
112 | return fmt.Errorf("invalid format for replication Role Arn: %v", opts.RoleArn) | ||
113 | } | ||
114 | switch { | ||
115 | case strings.HasPrefix(opts.RoleArn, "arn:minio:replication") && len(c.Rules) == 0: | ||
116 | c.Role = opts.RoleArn | ||
117 | compatSw = true | ||
118 | case strings.HasPrefix(opts.RoleArn, "arn:aws:iam"): | ||
119 | c.Role = opts.RoleArn | ||
120 | default: | ||
121 | return fmt.Errorf("RoleArn invalid for AWS replication configuration: %v", opts.RoleArn) | ||
122 | } | ||
123 | } | ||
124 | |||
125 | var status Status | ||
126 | // toggle rule status for edit option | ||
127 | switch opts.RuleStatus { | ||
128 | case "enable": | ||
129 | status = Enabled | ||
130 | case "disable": | ||
131 | status = Disabled | ||
132 | default: | ||
133 | return fmt.Errorf("rule state should be either [enable|disable]") | ||
134 | } | ||
135 | |||
136 | tags, err := opts.Tags() | ||
137 | if err != nil { | ||
138 | return err | ||
139 | } | ||
140 | andVal := And{ | ||
141 | Tags: tags, | ||
142 | } | ||
143 | filter := Filter{Prefix: opts.Prefix} | ||
144 | // only a single tag is set. | ||
145 | if opts.Prefix == "" && len(tags) == 1 { | ||
146 | filter.Tag = tags[0] | ||
147 | } | ||
148 | // both prefix and tag are present | ||
149 | if len(andVal.Tags) > 1 || opts.Prefix != "" { | ||
150 | filter.And = andVal | ||
151 | filter.And.Prefix = opts.Prefix | ||
152 | filter.Prefix = "" | ||
153 | filter.Tag = Tag{} | ||
154 | } | ||
155 | if opts.ID == "" { | ||
156 | opts.ID = xid.New().String() | ||
157 | } | ||
158 | |||
159 | destBucket := opts.DestBucket | ||
160 | // ref https://docs.aws.amazon.com/AmazonS3/latest/dev/s3-arn-format.html | ||
161 | if btokens := strings.Split(destBucket, ":"); len(btokens) != 6 { | ||
162 | if len(btokens) == 1 && compatSw { | ||
163 | destBucket = fmt.Sprintf("arn:aws:s3:::%s", destBucket) | ||
164 | } else { | ||
165 | return fmt.Errorf("destination bucket needs to be in Arn format") | ||
166 | } | ||
167 | } | ||
168 | dmStatus := Disabled | ||
169 | if opts.ReplicateDeleteMarkers != "" { | ||
170 | switch opts.ReplicateDeleteMarkers { | ||
171 | case "enable": | ||
172 | dmStatus = Enabled | ||
173 | case "disable": | ||
174 | dmStatus = Disabled | ||
175 | default: | ||
176 | return fmt.Errorf("ReplicateDeleteMarkers should be either enable|disable") | ||
177 | } | ||
178 | } | ||
179 | |||
180 | vDeleteStatus := Disabled | ||
181 | if opts.ReplicateDeletes != "" { | ||
182 | switch opts.ReplicateDeletes { | ||
183 | case "enable": | ||
184 | vDeleteStatus = Enabled | ||
185 | case "disable": | ||
186 | vDeleteStatus = Disabled | ||
187 | default: | ||
188 | return fmt.Errorf("ReplicateDeletes should be either enable|disable") | ||
189 | } | ||
190 | } | ||
191 | var replicaSync Status | ||
192 | // replica sync is by default Enabled, unless specified. | ||
193 | switch opts.ReplicaSync { | ||
194 | case "enable", "": | ||
195 | replicaSync = Enabled | ||
196 | case "disable": | ||
197 | replicaSync = Disabled | ||
198 | default: | ||
199 | return fmt.Errorf("replica metadata sync should be either [enable|disable]") | ||
200 | } | ||
201 | |||
202 | var existingStatus Status | ||
203 | if opts.ExistingObjectReplicate != "" { | ||
204 | switch opts.ExistingObjectReplicate { | ||
205 | case "enable": | ||
206 | existingStatus = Enabled | ||
207 | case "disable", "": | ||
208 | existingStatus = Disabled | ||
209 | default: | ||
210 | return fmt.Errorf("existingObjectReplicate should be either enable|disable") | ||
211 | } | ||
212 | } | ||
213 | newRule := Rule{ | ||
214 | ID: opts.ID, | ||
215 | Priority: priority, | ||
216 | Status: status, | ||
217 | Filter: filter, | ||
218 | Destination: Destination{ | ||
219 | Bucket: destBucket, | ||
220 | StorageClass: opts.StorageClass, | ||
221 | }, | ||
222 | DeleteMarkerReplication: DeleteMarkerReplication{Status: dmStatus}, | ||
223 | DeleteReplication: DeleteReplication{Status: vDeleteStatus}, | ||
224 | // MinIO enables replica metadata syncing by default in the case of bi-directional replication to allow | ||
225 | // automatic failover as the expectation in this case is that replica and source should be identical. | ||
226 | // However AWS leaves this configurable https://docs.aws.amazon.com/AmazonS3/latest/dev/replication-for-metadata-changes.html | ||
227 | SourceSelectionCriteria: SourceSelectionCriteria{ | ||
228 | ReplicaModifications: ReplicaModifications{ | ||
229 | Status: replicaSync, | ||
230 | }, | ||
231 | }, | ||
232 | // By default disable existing object replication unless selected | ||
233 | ExistingObjectReplication: ExistingObjectReplication{ | ||
234 | Status: existingStatus, | ||
235 | }, | ||
236 | } | ||
237 | |||
238 | // validate rule after overlaying priority for pre-existing rule being disabled. | ||
239 | if err := newRule.Validate(); err != nil { | ||
240 | return err | ||
241 | } | ||
242 | // if replication config uses RoleArn, migrate this to the destination element as target ARN for remote bucket for MinIO configuration | ||
243 | if c.Role != "" && !strings.HasPrefix(c.Role, "arn:aws:iam") && !compatSw { | ||
244 | for i := range c.Rules { | ||
245 | c.Rules[i].Destination.Bucket = c.Role | ||
246 | } | ||
247 | c.Role = "" | ||
248 | } | ||
249 | |||
250 | for _, rule := range c.Rules { | ||
251 | if rule.Priority == newRule.Priority { | ||
252 | return fmt.Errorf("priority must be unique. Replication configuration already has a rule with this priority") | ||
253 | } | ||
254 | if rule.ID == newRule.ID { | ||
255 | return fmt.Errorf("a rule exists with this ID") | ||
256 | } | ||
257 | } | ||
258 | |||
259 | c.Rules = append(c.Rules, newRule) | ||
260 | return nil | ||
261 | } | ||
262 | |||
263 | // EditRule modifies an existing rule in replication config | ||
264 | func (c *Config) EditRule(opts Options) error { | ||
265 | if opts.ID == "" { | ||
266 | return fmt.Errorf("rule ID missing") | ||
267 | } | ||
268 | // if replication config uses RoleArn, migrate this to the destination element as target ARN for remote bucket for non AWS. | ||
269 | if c.Role != "" && !strings.HasPrefix(c.Role, "arn:aws:iam") && len(c.Rules) > 1 { | ||
270 | for i := range c.Rules { | ||
271 | c.Rules[i].Destination.Bucket = c.Role | ||
272 | } | ||
273 | c.Role = "" | ||
274 | } | ||
275 | |||
276 | rIdx := -1 | ||
277 | var newRule Rule | ||
278 | for i, rule := range c.Rules { | ||
279 | if rule.ID == opts.ID { | ||
280 | rIdx = i | ||
281 | newRule = rule | ||
282 | break | ||
283 | } | ||
284 | } | ||
285 | if rIdx < 0 { | ||
286 | return fmt.Errorf("rule with ID %s not found in replication configuration", opts.ID) | ||
287 | } | ||
288 | prefixChg := opts.Prefix != newRule.Prefix() | ||
289 | if opts.IsTagSet || prefixChg { | ||
290 | prefix := newRule.Prefix() | ||
291 | if prefix != opts.Prefix { | ||
292 | prefix = opts.Prefix | ||
293 | } | ||
294 | tags := []Tag{newRule.Filter.Tag} | ||
295 | if len(newRule.Filter.And.Tags) != 0 { | ||
296 | tags = newRule.Filter.And.Tags | ||
297 | } | ||
298 | var err error | ||
299 | if opts.IsTagSet { | ||
300 | tags, err = opts.Tags() | ||
301 | if err != nil { | ||
302 | return err | ||
303 | } | ||
304 | } | ||
305 | andVal := And{ | ||
306 | Tags: tags, | ||
307 | } | ||
308 | |||
309 | filter := Filter{Prefix: prefix} | ||
310 | // only a single tag is set. | ||
311 | if prefix == "" && len(tags) == 1 { | ||
312 | filter.Tag = tags[0] | ||
313 | } | ||
314 | // both prefix and tag are present | ||
315 | if len(andVal.Tags) > 1 || prefix != "" { | ||
316 | filter.And = andVal | ||
317 | filter.And.Prefix = prefix | ||
318 | filter.Prefix = "" | ||
319 | filter.Tag = Tag{} | ||
320 | } | ||
321 | newRule.Filter = filter | ||
322 | } | ||
323 | |||
324 | // toggle rule status for edit option | ||
325 | if opts.RuleStatus != "" { | ||
326 | switch opts.RuleStatus { | ||
327 | case "enable": | ||
328 | newRule.Status = Enabled | ||
329 | case "disable": | ||
330 | newRule.Status = Disabled | ||
331 | default: | ||
332 | return fmt.Errorf("rule state should be either [enable|disable]") | ||
333 | } | ||
334 | } | ||
335 | // set DeleteMarkerReplication rule status for edit option | ||
336 | if opts.ReplicateDeleteMarkers != "" { | ||
337 | switch opts.ReplicateDeleteMarkers { | ||
338 | case "enable": | ||
339 | newRule.DeleteMarkerReplication.Status = Enabled | ||
340 | case "disable": | ||
341 | newRule.DeleteMarkerReplication.Status = Disabled | ||
342 | default: | ||
343 | return fmt.Errorf("ReplicateDeleteMarkers state should be either [enable|disable]") | ||
344 | } | ||
345 | } | ||
346 | |||
347 | // set DeleteReplication rule status for edit option. This is a MinIO specific | ||
348 | // option to replicate versioned deletes | ||
349 | if opts.ReplicateDeletes != "" { | ||
350 | switch opts.ReplicateDeletes { | ||
351 | case "enable": | ||
352 | newRule.DeleteReplication.Status = Enabled | ||
353 | case "disable": | ||
354 | newRule.DeleteReplication.Status = Disabled | ||
355 | default: | ||
356 | return fmt.Errorf("ReplicateDeletes state should be either [enable|disable]") | ||
357 | } | ||
358 | } | ||
359 | |||
360 | if opts.ReplicaSync != "" { | ||
361 | switch opts.ReplicaSync { | ||
362 | case "enable", "": | ||
363 | newRule.SourceSelectionCriteria.ReplicaModifications.Status = Enabled | ||
364 | case "disable": | ||
365 | newRule.SourceSelectionCriteria.ReplicaModifications.Status = Disabled | ||
366 | default: | ||
367 | return fmt.Errorf("replica metadata sync should be either [enable|disable]") | ||
368 | } | ||
369 | } | ||
370 | |||
371 | if opts.ExistingObjectReplicate != "" { | ||
372 | switch opts.ExistingObjectReplicate { | ||
373 | case "enable": | ||
374 | newRule.ExistingObjectReplication.Status = Enabled | ||
375 | case "disable": | ||
376 | newRule.ExistingObjectReplication.Status = Disabled | ||
377 | default: | ||
378 | return fmt.Errorf("existingObjectsReplication state should be either [enable|disable]") | ||
379 | } | ||
380 | } | ||
381 | if opts.IsSCSet { | ||
382 | newRule.Destination.StorageClass = opts.StorageClass | ||
383 | } | ||
384 | if opts.Priority != "" { | ||
385 | priority, err := strconv.Atoi(opts.Priority) | ||
386 | if err != nil { | ||
387 | return err | ||
388 | } | ||
389 | newRule.Priority = priority | ||
390 | } | ||
391 | if opts.DestBucket != "" { | ||
392 | destBucket := opts.DestBucket | ||
393 | // ref https://docs.aws.amazon.com/AmazonS3/latest/dev/s3-arn-format.html | ||
394 | if btokens := strings.Split(opts.DestBucket, ":"); len(btokens) != 6 { | ||
395 | return fmt.Errorf("destination bucket needs to be in Arn format") | ||
396 | } | ||
397 | newRule.Destination.Bucket = destBucket | ||
398 | } | ||
399 | // validate rule | ||
400 | if err := newRule.Validate(); err != nil { | ||
401 | return err | ||
402 | } | ||
403 | // ensure priority and destination bucket restrictions are not violated | ||
404 | for idx, rule := range c.Rules { | ||
405 | if rule.Priority == newRule.Priority && rIdx != idx { | ||
406 | return fmt.Errorf("priority must be unique. Replication configuration already has a rule with this priority") | ||
407 | } | ||
408 | if rule.Destination.Bucket != newRule.Destination.Bucket && rule.ID == newRule.ID { | ||
409 | return fmt.Errorf("invalid destination bucket for this rule") | ||
410 | } | ||
411 | } | ||
412 | |||
413 | c.Rules[rIdx] = newRule | ||
414 | return nil | ||
415 | } | ||
416 | |||
417 | // RemoveRule removes a rule from replication config. | ||
418 | func (c *Config) RemoveRule(opts Options) error { | ||
419 | var newRules []Rule | ||
420 | ruleFound := false | ||
421 | for _, rule := range c.Rules { | ||
422 | if rule.ID != opts.ID { | ||
423 | newRules = append(newRules, rule) | ||
424 | continue | ||
425 | } | ||
426 | ruleFound = true | ||
427 | } | ||
428 | if !ruleFound { | ||
429 | return fmt.Errorf("Rule with ID %s not found", opts.ID) | ||
430 | } | ||
431 | if len(newRules) == 0 { | ||
432 | return fmt.Errorf("replication configuration should have at least one rule") | ||
433 | } | ||
434 | c.Rules = newRules | ||
435 | return nil | ||
436 | } | ||
437 | |||
438 | // Rule - a rule for replication configuration. | ||
439 | type Rule struct { | ||
440 | XMLName xml.Name `xml:"Rule" json:"-"` | ||
441 | ID string `xml:"ID,omitempty"` | ||
442 | Status Status `xml:"Status"` | ||
443 | Priority int `xml:"Priority"` | ||
444 | DeleteMarkerReplication DeleteMarkerReplication `xml:"DeleteMarkerReplication"` | ||
445 | DeleteReplication DeleteReplication `xml:"DeleteReplication"` | ||
446 | Destination Destination `xml:"Destination"` | ||
447 | Filter Filter `xml:"Filter" json:"Filter"` | ||
448 | SourceSelectionCriteria SourceSelectionCriteria `xml:"SourceSelectionCriteria" json:"SourceSelectionCriteria"` | ||
449 | ExistingObjectReplication ExistingObjectReplication `xml:"ExistingObjectReplication,omitempty" json:"ExistingObjectReplication,omitempty"` | ||
450 | } | ||
451 | |||
452 | // Validate validates the rule for correctness | ||
453 | func (r Rule) Validate() error { | ||
454 | if err := r.validateID(); err != nil { | ||
455 | return err | ||
456 | } | ||
457 | if err := r.validateStatus(); err != nil { | ||
458 | return err | ||
459 | } | ||
460 | if err := r.validateFilter(); err != nil { | ||
461 | return err | ||
462 | } | ||
463 | |||
464 | if r.Priority < 0 && r.Status == Enabled { | ||
465 | return fmt.Errorf("priority must be set for the rule") | ||
466 | } | ||
467 | |||
468 | if err := r.validateStatus(); err != nil { | ||
469 | return err | ||
470 | } | ||
471 | return r.ExistingObjectReplication.Validate() | ||
472 | } | ||
473 | |||
474 | // validateID - checks if ID is valid or not. | ||
475 | func (r Rule) validateID() error { | ||
476 | // cannot be longer than 255 characters | ||
477 | if len(r.ID) > 255 { | ||
478 | return fmt.Errorf("ID must be less than 255 characters") | ||
479 | } | ||
480 | return nil | ||
481 | } | ||
482 | |||
483 | // validateStatus - checks if status is valid or not. | ||
484 | func (r Rule) validateStatus() error { | ||
485 | // Status can't be empty | ||
486 | if len(r.Status) == 0 { | ||
487 | return fmt.Errorf("status cannot be empty") | ||
488 | } | ||
489 | |||
490 | // Status must be one of Enabled or Disabled | ||
491 | if r.Status != Enabled && r.Status != Disabled { | ||
492 | return fmt.Errorf("status must be set to either Enabled or Disabled") | ||
493 | } | ||
494 | return nil | ||
495 | } | ||
496 | |||
497 | func (r Rule) validateFilter() error { | ||
498 | return r.Filter.Validate() | ||
499 | } | ||
500 | |||
501 | // Prefix - a rule can either have prefix under <filter></filter> or under | ||
502 | // <filter><and></and></filter>. This method returns the prefix from the | ||
503 | // location where it is available | ||
504 | func (r Rule) Prefix() string { | ||
505 | if r.Filter.Prefix != "" { | ||
506 | return r.Filter.Prefix | ||
507 | } | ||
508 | return r.Filter.And.Prefix | ||
509 | } | ||
510 | |||
511 | // Tags - a rule can either have tag under <filter></filter> or under | ||
512 | // <filter><and></and></filter>. This method returns all the tags from the | ||
513 | // rule in the format tag1=value1&tag2=value2 | ||
514 | func (r Rule) Tags() string { | ||
515 | ts := []Tag{r.Filter.Tag} | ||
516 | if len(r.Filter.And.Tags) != 0 { | ||
517 | ts = r.Filter.And.Tags | ||
518 | } | ||
519 | |||
520 | var buf bytes.Buffer | ||
521 | for _, t := range ts { | ||
522 | if buf.Len() > 0 { | ||
523 | buf.WriteString("&") | ||
524 | } | ||
525 | buf.WriteString(t.String()) | ||
526 | } | ||
527 | return buf.String() | ||
528 | } | ||
529 | |||
530 | // Filter - a filter for a replication configuration Rule. | ||
531 | type Filter struct { | ||
532 | XMLName xml.Name `xml:"Filter" json:"-"` | ||
533 | Prefix string `json:"Prefix,omitempty"` | ||
534 | And And `xml:"And,omitempty" json:"And,omitempty"` | ||
535 | Tag Tag `xml:"Tag,omitempty" json:"Tag,omitempty"` | ||
536 | } | ||
537 | |||
538 | // Validate - validates the filter element | ||
539 | func (f Filter) Validate() error { | ||
540 | // A Filter must have exactly one of Prefix, Tag, or And specified. | ||
541 | if !f.And.isEmpty() { | ||
542 | if f.Prefix != "" { | ||
543 | return errInvalidFilter | ||
544 | } | ||
545 | if !f.Tag.IsEmpty() { | ||
546 | return errInvalidFilter | ||
547 | } | ||
548 | } | ||
549 | if f.Prefix != "" { | ||
550 | if !f.Tag.IsEmpty() { | ||
551 | return errInvalidFilter | ||
552 | } | ||
553 | } | ||
554 | if !f.Tag.IsEmpty() { | ||
555 | if err := f.Tag.Validate(); err != nil { | ||
556 | return err | ||
557 | } | ||
558 | } | ||
559 | return nil | ||
560 | } | ||
561 | |||
562 | // Tag - a tag for a replication configuration Rule filter. | ||
563 | type Tag struct { | ||
564 | XMLName xml.Name `json:"-"` | ||
565 | Key string `xml:"Key,omitempty" json:"Key,omitempty"` | ||
566 | Value string `xml:"Value,omitempty" json:"Value,omitempty"` | ||
567 | } | ||
568 | |||
569 | func (tag Tag) String() string { | ||
570 | if tag.IsEmpty() { | ||
571 | return "" | ||
572 | } | ||
573 | return tag.Key + "=" + tag.Value | ||
574 | } | ||
575 | |||
576 | // IsEmpty returns whether this tag is empty or not. | ||
577 | func (tag Tag) IsEmpty() bool { | ||
578 | return tag.Key == "" | ||
579 | } | ||
580 | |||
581 | // Validate checks this tag. | ||
582 | func (tag Tag) Validate() error { | ||
583 | if len(tag.Key) == 0 || utf8.RuneCountInString(tag.Key) > 128 { | ||
584 | return fmt.Errorf("invalid Tag Key") | ||
585 | } | ||
586 | |||
587 | if utf8.RuneCountInString(tag.Value) > 256 { | ||
588 | return fmt.Errorf("invalid Tag Value") | ||
589 | } | ||
590 | return nil | ||
591 | } | ||
592 | |||
593 | // Destination - destination in ReplicationConfiguration. | ||
594 | type Destination struct { | ||
595 | XMLName xml.Name `xml:"Destination" json:"-"` | ||
596 | Bucket string `xml:"Bucket" json:"Bucket"` | ||
597 | StorageClass string `xml:"StorageClass,omitempty" json:"StorageClass,omitempty"` | ||
598 | } | ||
599 | |||
600 | // And - a tag to combine a prefix and multiple tags for replication configuration rule. | ||
601 | type And struct { | ||
602 | XMLName xml.Name `xml:"And,omitempty" json:"-"` | ||
603 | Prefix string `xml:"Prefix,omitempty" json:"Prefix,omitempty"` | ||
604 | Tags []Tag `xml:"Tag,omitempty" json:"Tag,omitempty"` | ||
605 | } | ||
606 | |||
607 | // isEmpty returns true if Tags field is null | ||
608 | func (a And) isEmpty() bool { | ||
609 | return len(a.Tags) == 0 && a.Prefix == "" | ||
610 | } | ||
611 | |||
612 | // Status represents Enabled/Disabled status | ||
613 | type Status string | ||
614 | |||
615 | // Supported status types | ||
616 | const ( | ||
617 | Enabled Status = "Enabled" | ||
618 | Disabled Status = "Disabled" | ||
619 | ) | ||
620 | |||
621 | // DeleteMarkerReplication - whether delete markers are replicated - https://docs.aws.amazon.com/AmazonS3/latest/dev/replication-add-config.html | ||
622 | type DeleteMarkerReplication struct { | ||
623 | Status Status `xml:"Status" json:"Status"` // should be set to "Disabled" by default | ||
624 | } | ||
625 | |||
626 | // IsEmpty returns true if DeleteMarkerReplication is not set | ||
627 | func (d DeleteMarkerReplication) IsEmpty() bool { | ||
628 | return len(d.Status) == 0 | ||
629 | } | ||
630 | |||
631 | // DeleteReplication - whether versioned deletes are replicated - this | ||
632 | // is a MinIO specific extension | ||
633 | type DeleteReplication struct { | ||
634 | Status Status `xml:"Status" json:"Status"` // should be set to "Disabled" by default | ||
635 | } | ||
636 | |||
637 | // IsEmpty returns true if DeleteReplication is not set | ||
638 | func (d DeleteReplication) IsEmpty() bool { | ||
639 | return len(d.Status) == 0 | ||
640 | } | ||
641 | |||
642 | // ReplicaModifications specifies if replica modification sync is enabled | ||
643 | type ReplicaModifications struct { | ||
644 | Status Status `xml:"Status" json:"Status"` // should be set to "Enabled" by default | ||
645 | } | ||
646 | |||
647 | // SourceSelectionCriteria - specifies additional source selection criteria in ReplicationConfiguration. | ||
648 | type SourceSelectionCriteria struct { | ||
649 | ReplicaModifications ReplicaModifications `xml:"ReplicaModifications" json:"ReplicaModifications"` | ||
650 | } | ||
651 | |||
652 | // IsValid - checks whether SourceSelectionCriteria is valid or not. | ||
653 | func (s SourceSelectionCriteria) IsValid() bool { | ||
654 | return s.ReplicaModifications.Status == Enabled || s.ReplicaModifications.Status == Disabled | ||
655 | } | ||
656 | |||
657 | // Validate source selection criteria | ||
658 | func (s SourceSelectionCriteria) Validate() error { | ||
659 | if (s == SourceSelectionCriteria{}) { | ||
660 | return nil | ||
661 | } | ||
662 | if !s.IsValid() { | ||
663 | return fmt.Errorf("invalid ReplicaModification status") | ||
664 | } | ||
665 | return nil | ||
666 | } | ||
667 | |||
668 | // ExistingObjectReplication - whether existing object replication is enabled | ||
669 | type ExistingObjectReplication struct { | ||
670 | Status Status `xml:"Status"` // should be set to "Disabled" by default | ||
671 | } | ||
672 | |||
673 | // IsEmpty returns true if DeleteMarkerReplication is not set | ||
674 | func (e ExistingObjectReplication) IsEmpty() bool { | ||
675 | return len(e.Status) == 0 | ||
676 | } | ||
677 | |||
678 | // Validate validates whether the status is disabled. | ||
679 | func (e ExistingObjectReplication) Validate() error { | ||
680 | if e.IsEmpty() { | ||
681 | return nil | ||
682 | } | ||
683 | if e.Status != Disabled && e.Status != Enabled { | ||
684 | return fmt.Errorf("invalid ExistingObjectReplication status") | ||
685 | } | ||
686 | return nil | ||
687 | } | ||
688 | |||
689 | // TargetMetrics represents inline replication metrics | ||
690 | // such as pending, failed and completed bytes in total for a bucket remote target | ||
691 | type TargetMetrics struct { | ||
692 | // Completed count | ||
693 | ReplicatedCount uint64 `json:"replicationCount,omitempty"` | ||
694 | // Completed size in bytes | ||
695 | ReplicatedSize uint64 `json:"completedReplicationSize,omitempty"` | ||
696 | // Bandwidth limit in bytes/sec for this target | ||
697 | BandWidthLimitInBytesPerSecond int64 `json:"limitInBits,omitempty"` | ||
698 | // Current bandwidth used in bytes/sec for this target | ||
699 | CurrentBandwidthInBytesPerSecond float64 `json:"currentBandwidth,omitempty"` | ||
700 | // errors seen in replication in last minute, hour and total | ||
701 | Failed TimedErrStats `json:"failed,omitempty"` | ||
702 | // Deprecated fields | ||
703 | // Pending size in bytes | ||
704 | PendingSize uint64 `json:"pendingReplicationSize,omitempty"` | ||
705 | // Total Replica size in bytes | ||
706 | ReplicaSize uint64 `json:"replicaSize,omitempty"` | ||
707 | // Failed size in bytes | ||
708 | FailedSize uint64 `json:"failedReplicationSize,omitempty"` | ||
709 | // Total number of pending operations including metadata updates | ||
710 | PendingCount uint64 `json:"pendingReplicationCount,omitempty"` | ||
711 | // Total number of failed operations including metadata updates | ||
712 | FailedCount uint64 `json:"failedReplicationCount,omitempty"` | ||
713 | } | ||
714 | |||
715 | // Metrics represents inline replication metrics for a bucket. | ||
716 | type Metrics struct { | ||
717 | Stats map[string]TargetMetrics | ||
718 | // Completed size in bytes across targets | ||
719 | ReplicatedSize uint64 `json:"completedReplicationSize,omitempty"` | ||
720 | // Total Replica size in bytes across targets | ||
721 | ReplicaSize uint64 `json:"replicaSize,omitempty"` | ||
722 | // Total Replica counts | ||
723 | ReplicaCount int64 `json:"replicaCount,omitempty"` | ||
724 | // Total Replicated count | ||
725 | ReplicatedCount int64 `json:"replicationCount,omitempty"` | ||
726 | // errors seen in replication in last minute, hour and total | ||
727 | Errors TimedErrStats `json:"failed,omitempty"` | ||
728 | // Total number of entries that are queued for replication | ||
729 | QStats InQueueMetric `json:"queued"` | ||
730 | // Deprecated fields | ||
731 | // Total Pending size in bytes across targets | ||
732 | PendingSize uint64 `json:"pendingReplicationSize,omitempty"` | ||
733 | // Failed size in bytes across targets | ||
734 | FailedSize uint64 `json:"failedReplicationSize,omitempty"` | ||
735 | // Total number of pending operations including metadata updates across targets | ||
736 | PendingCount uint64 `json:"pendingReplicationCount,omitempty"` | ||
737 | // Total number of failed operations including metadata updates across targets | ||
738 | FailedCount uint64 `json:"failedReplicationCount,omitempty"` | ||
739 | } | ||
740 | |||
741 | // RStat - has count and bytes for replication metrics | ||
742 | type RStat struct { | ||
743 | Count float64 `json:"count"` | ||
744 | Bytes int64 `json:"bytes"` | ||
745 | } | ||
746 | |||
747 | // Add two RStat | ||
748 | func (r RStat) Add(r1 RStat) RStat { | ||
749 | return RStat{ | ||
750 | Count: r.Count + r1.Count, | ||
751 | Bytes: r.Bytes + r1.Bytes, | ||
752 | } | ||
753 | } | ||
754 | |||
755 | // TimedErrStats holds error stats for a time period | ||
756 | type TimedErrStats struct { | ||
757 | LastMinute RStat `json:"lastMinute"` | ||
758 | LastHour RStat `json:"lastHour"` | ||
759 | Totals RStat `json:"totals"` | ||
760 | } | ||
761 | |||
762 | // Add two TimedErrStats | ||
763 | func (te TimedErrStats) Add(o TimedErrStats) TimedErrStats { | ||
764 | return TimedErrStats{ | ||
765 | LastMinute: te.LastMinute.Add(o.LastMinute), | ||
766 | LastHour: te.LastHour.Add(o.LastHour), | ||
767 | Totals: te.Totals.Add(o.Totals), | ||
768 | } | ||
769 | } | ||
770 | |||
771 | // ResyncTargetsInfo provides replication target information to resync replicated data. | ||
772 | type ResyncTargetsInfo struct { | ||
773 | Targets []ResyncTarget `json:"target,omitempty"` | ||
774 | } | ||
775 | |||
776 | // ResyncTarget provides the replica resources and resetID to initiate resync replication. | ||
777 | type ResyncTarget struct { | ||
778 | Arn string `json:"arn"` | ||
779 | ResetID string `json:"resetid"` | ||
780 | StartTime time.Time `json:"startTime,omitempty"` | ||
781 | EndTime time.Time `json:"endTime,omitempty"` | ||
782 | // Status of resync operation | ||
783 | ResyncStatus string `json:"resyncStatus,omitempty"` | ||
784 | // Completed size in bytes | ||
785 | ReplicatedSize int64 `json:"completedReplicationSize,omitempty"` | ||
786 | // Failed size in bytes | ||
787 | FailedSize int64 `json:"failedReplicationSize,omitempty"` | ||
788 | // Total number of failed operations | ||
789 | FailedCount int64 `json:"failedReplicationCount,omitempty"` | ||
790 | // Total number of completed operations | ||
791 | ReplicatedCount int64 `json:"replicationCount,omitempty"` | ||
792 | // Last bucket/object replicated. | ||
793 | Bucket string `json:"bucket,omitempty"` | ||
794 | Object string `json:"object,omitempty"` | ||
795 | } | ||
796 | |||
797 | // XferStats holds transfer rate info for uploads/sec | ||
798 | type XferStats struct { | ||
799 | AvgRate float64 `json:"avgRate"` | ||
800 | PeakRate float64 `json:"peakRate"` | ||
801 | CurrRate float64 `json:"currRate"` | ||
802 | } | ||
803 | |||
804 | // Merge two XferStats | ||
805 | func (x *XferStats) Merge(x1 XferStats) { | ||
806 | x.AvgRate += x1.AvgRate | ||
807 | x.PeakRate += x1.PeakRate | ||
808 | x.CurrRate += x1.CurrRate | ||
809 | } | ||
810 | |||
811 | // QStat holds count and bytes for objects in replication queue | ||
812 | type QStat struct { | ||
813 | Count float64 `json:"count"` | ||
814 | Bytes float64 `json:"bytes"` | ||
815 | } | ||
816 | |||
817 | // Add 2 QStat entries | ||
818 | func (q *QStat) Add(q1 QStat) { | ||
819 | q.Count += q1.Count | ||
820 | q.Bytes += q1.Bytes | ||
821 | } | ||
822 | |||
823 | // InQueueMetric holds stats for objects in replication queue | ||
824 | type InQueueMetric struct { | ||
825 | Curr QStat `json:"curr" msg:"cq"` | ||
826 | Avg QStat `json:"avg" msg:"aq"` | ||
827 | Max QStat `json:"peak" msg:"pq"` | ||
828 | } | ||
829 | |||
830 | // MetricName name of replication metric | ||
831 | type MetricName string | ||
832 | |||
833 | const ( | ||
834 | // Large is a metric name for large objects >=128MiB | ||
835 | Large MetricName = "Large" | ||
836 | // Small is a metric name for objects <128MiB size | ||
837 | Small MetricName = "Small" | ||
838 | // Total is a metric name for total objects | ||
839 | Total MetricName = "Total" | ||
840 | ) | ||
841 | |||
842 | // WorkerStat has stats on number of replication workers | ||
843 | type WorkerStat struct { | ||
844 | Curr int32 `json:"curr"` | ||
845 | Avg float32 `json:"avg"` | ||
846 | Max int32 `json:"max"` | ||
847 | } | ||
848 | |||
849 | // ReplMRFStats holds stats of MRF backlog saved to disk in the last 5 minutes | ||
850 | // and number of entries that failed replication after 3 retries | ||
851 | type ReplMRFStats struct { | ||
852 | LastFailedCount uint64 `json:"failedCount_last5min"` | ||
853 | // Count of unreplicated entries that were dropped after MRF retry limit reached since cluster start. | ||
854 | TotalDroppedCount uint64 `json:"droppedCount_since_uptime"` | ||
855 | // Bytes of unreplicated entries that were dropped after MRF retry limit reached since cluster start. | ||
856 | TotalDroppedBytes uint64 `json:"droppedBytes_since_uptime"` | ||
857 | } | ||
858 | |||
859 | // ReplQNodeStats holds stats for a node in replication queue | ||
860 | type ReplQNodeStats struct { | ||
861 | NodeName string `json:"nodeName"` | ||
862 | Uptime int64 `json:"uptime"` | ||
863 | Workers WorkerStat `json:"activeWorkers"` | ||
864 | |||
865 | XferStats map[MetricName]XferStats `json:"transferSummary"` | ||
866 | TgtXferStats map[string]map[MetricName]XferStats `json:"tgtTransferStats"` | ||
867 | |||
868 | QStats InQueueMetric `json:"queueStats"` | ||
869 | MRFStats ReplMRFStats `json:"mrfStats"` | ||
870 | } | ||
871 | |||
872 | // ReplQueueStats holds stats for replication queue across nodes | ||
873 | type ReplQueueStats struct { | ||
874 | Nodes []ReplQNodeStats `json:"nodes"` | ||
875 | } | ||
876 | |||
877 | // Workers returns number of workers across all nodes | ||
878 | func (q ReplQueueStats) Workers() (tot WorkerStat) { | ||
879 | for _, node := range q.Nodes { | ||
880 | tot.Avg += node.Workers.Avg | ||
881 | tot.Curr += node.Workers.Curr | ||
882 | if tot.Max < node.Workers.Max { | ||
883 | tot.Max = node.Workers.Max | ||
884 | } | ||
885 | } | ||
886 | if len(q.Nodes) > 0 { | ||
887 | tot.Avg /= float32(len(q.Nodes)) | ||
888 | tot.Curr /= int32(len(q.Nodes)) | ||
889 | } | ||
890 | return tot | ||
891 | } | ||
892 | |||
893 | // qStatSummary returns cluster level stats for objects in replication queue | ||
894 | func (q ReplQueueStats) qStatSummary() InQueueMetric { | ||
895 | m := InQueueMetric{} | ||
896 | for _, v := range q.Nodes { | ||
897 | m.Avg.Add(v.QStats.Avg) | ||
898 | m.Curr.Add(v.QStats.Curr) | ||
899 | if m.Max.Count < v.QStats.Max.Count { | ||
900 | m.Max.Add(v.QStats.Max) | ||
901 | } | ||
902 | } | ||
903 | return m | ||
904 | } | ||
905 | |||
906 | // ReplQStats holds stats for objects in replication queue | ||
907 | type ReplQStats struct { | ||
908 | Uptime int64 `json:"uptime"` | ||
909 | Workers WorkerStat `json:"workers"` | ||
910 | |||
911 | XferStats map[MetricName]XferStats `json:"xferStats"` | ||
912 | TgtXferStats map[string]map[MetricName]XferStats `json:"tgtXferStats"` | ||
913 | |||
914 | QStats InQueueMetric `json:"qStats"` | ||
915 | MRFStats ReplMRFStats `json:"mrfStats"` | ||
916 | } | ||
917 | |||
918 | // QStats returns cluster level stats for objects in replication queue | ||
919 | func (q ReplQueueStats) QStats() (r ReplQStats) { | ||
920 | r.QStats = q.qStatSummary() | ||
921 | r.XferStats = make(map[MetricName]XferStats) | ||
922 | r.TgtXferStats = make(map[string]map[MetricName]XferStats) | ||
923 | r.Workers = q.Workers() | ||
924 | |||
925 | for _, node := range q.Nodes { | ||
926 | for arn := range node.TgtXferStats { | ||
927 | xmap, ok := node.TgtXferStats[arn] | ||
928 | if !ok { | ||
929 | xmap = make(map[MetricName]XferStats) | ||
930 | } | ||
931 | for m, v := range xmap { | ||
932 | st, ok := r.XferStats[m] | ||
933 | if !ok { | ||
934 | st = XferStats{} | ||
935 | } | ||
936 | st.AvgRate += v.AvgRate | ||
937 | st.CurrRate += v.CurrRate | ||
938 | st.PeakRate = math.Max(st.PeakRate, v.PeakRate) | ||
939 | if _, ok := r.TgtXferStats[arn]; !ok { | ||
940 | r.TgtXferStats[arn] = make(map[MetricName]XferStats) | ||
941 | } | ||
942 | r.TgtXferStats[arn][m] = st | ||
943 | } | ||
944 | } | ||
945 | for k, v := range node.XferStats { | ||
946 | st, ok := r.XferStats[k] | ||
947 | if !ok { | ||
948 | st = XferStats{} | ||
949 | } | ||
950 | st.AvgRate += v.AvgRate | ||
951 | st.CurrRate += v.CurrRate | ||
952 | st.PeakRate = math.Max(st.PeakRate, v.PeakRate) | ||
953 | r.XferStats[k] = st | ||
954 | } | ||
955 | r.MRFStats.LastFailedCount += node.MRFStats.LastFailedCount | ||
956 | r.MRFStats.TotalDroppedCount += node.MRFStats.TotalDroppedCount | ||
957 | r.MRFStats.TotalDroppedBytes += node.MRFStats.TotalDroppedBytes | ||
958 | r.Uptime += node.Uptime | ||
959 | } | ||
960 | if len(q.Nodes) > 0 { | ||
961 | r.Uptime /= int64(len(q.Nodes)) // average uptime | ||
962 | } | ||
963 | return | ||
964 | } | ||
965 | |||
966 | // MetricsV2 represents replication metrics for a bucket. | ||
967 | type MetricsV2 struct { | ||
968 | Uptime int64 `json:"uptime"` | ||
969 | CurrentStats Metrics `json:"currStats"` | ||
970 | QueueStats ReplQueueStats `json:"queueStats"` | ||
971 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/pkg/s3utils/utils.go b/vendor/github.com/minio/minio-go/v7/pkg/s3utils/utils.go deleted file mode 100644 index 056e78a..0000000 --- a/vendor/github.com/minio/minio-go/v7/pkg/s3utils/utils.go +++ /dev/null | |||
@@ -1,411 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2015-2020 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package s3utils | ||
19 | |||
20 | import ( | ||
21 | "bytes" | ||
22 | "encoding/hex" | ||
23 | "errors" | ||
24 | "net" | ||
25 | "net/url" | ||
26 | "regexp" | ||
27 | "sort" | ||
28 | "strings" | ||
29 | "unicode/utf8" | ||
30 | ) | ||
31 | |||
32 | // Sentinel URL is the default url value which is invalid. | ||
33 | var sentinelURL = url.URL{} | ||
34 | |||
35 | // IsValidDomain validates if input string is a valid domain name. | ||
36 | func IsValidDomain(host string) bool { | ||
37 | // See RFC 1035, RFC 3696. | ||
38 | host = strings.TrimSpace(host) | ||
39 | if len(host) == 0 || len(host) > 255 { | ||
40 | return false | ||
41 | } | ||
42 | // host cannot start or end with "-" | ||
43 | if host[len(host)-1:] == "-" || host[:1] == "-" { | ||
44 | return false | ||
45 | } | ||
46 | // host cannot start or end with "_" | ||
47 | if host[len(host)-1:] == "_" || host[:1] == "_" { | ||
48 | return false | ||
49 | } | ||
50 | // host cannot start with a "." | ||
51 | if host[:1] == "." { | ||
52 | return false | ||
53 | } | ||
54 | // All non alphanumeric characters are invalid. | ||
55 | if strings.ContainsAny(host, "`~!@#$%^&*()+={}[]|\\\"';:><?/") { | ||
56 | return false | ||
57 | } | ||
58 | // No need to regexp match, since the list is non-exhaustive. | ||
59 | // We let it valid and fail later. | ||
60 | return true | ||
61 | } | ||
62 | |||
63 | // IsValidIP parses input string for ip address validity. | ||
64 | func IsValidIP(ip string) bool { | ||
65 | return net.ParseIP(ip) != nil | ||
66 | } | ||
67 | |||
68 | // IsVirtualHostSupported - verifies if bucketName can be part of | ||
69 | // virtual host. Currently only Amazon S3 and Google Cloud Storage | ||
70 | // would support this. | ||
71 | func IsVirtualHostSupported(endpointURL url.URL, bucketName string) bool { | ||
72 | if endpointURL == sentinelURL { | ||
73 | return false | ||
74 | } | ||
75 | // bucketName can be valid but '.' in the hostname will fail SSL | ||
76 | // certificate validation. So do not use host-style for such buckets. | ||
77 | if endpointURL.Scheme == "https" && strings.Contains(bucketName, ".") { | ||
78 | return false | ||
79 | } | ||
80 | // Return true for all other cases | ||
81 | return IsAmazonEndpoint(endpointURL) || IsGoogleEndpoint(endpointURL) || IsAliyunOSSEndpoint(endpointURL) | ||
82 | } | ||
83 | |||
84 | // Refer for region styles - https://docs.aws.amazon.com/general/latest/gr/rande.html#s3_region | ||
85 | |||
86 | // amazonS3HostHyphen - regular expression used to determine if an arg is s3 host in hyphenated style. | ||
87 | var amazonS3HostHyphen = regexp.MustCompile(`^s3-(.*?).amazonaws.com$`) | ||
88 | |||
89 | // amazonS3HostDualStack - regular expression used to determine if an arg is s3 host dualstack. | ||
90 | var amazonS3HostDualStack = regexp.MustCompile(`^s3.dualstack.(.*?).amazonaws.com$`) | ||
91 | |||
92 | // amazonS3HostFIPS - regular expression used to determine if an arg is s3 FIPS host. | ||
93 | var amazonS3HostFIPS = regexp.MustCompile(`^s3-fips.(.*?).amazonaws.com$`) | ||
94 | |||
95 | // amazonS3HostFIPSDualStack - regular expression used to determine if an arg is s3 FIPS host dualstack. | ||
96 | var amazonS3HostFIPSDualStack = regexp.MustCompile(`^s3-fips.dualstack.(.*?).amazonaws.com$`) | ||
97 | |||
98 | // amazonS3HostDot - regular expression used to determine if an arg is s3 host in . style. | ||
99 | var amazonS3HostDot = regexp.MustCompile(`^s3.(.*?).amazonaws.com$`) | ||
100 | |||
101 | // amazonS3ChinaHost - regular expression used to determine if the arg is s3 china host. | ||
102 | var amazonS3ChinaHost = regexp.MustCompile(`^s3.(cn.*?).amazonaws.com.cn$`) | ||
103 | |||
104 | // amazonS3ChinaHostDualStack - regular expression used to determine if the arg is s3 china host dualstack. | ||
105 | var amazonS3ChinaHostDualStack = regexp.MustCompile(`^s3.dualstack.(cn.*?).amazonaws.com.cn$`) | ||
106 | |||
107 | // Regular expression used to determine if the arg is elb host. | ||
108 | var elbAmazonRegex = regexp.MustCompile(`elb(.*?).amazonaws.com$`) | ||
109 | |||
110 | // Regular expression used to determine if the arg is elb host in china. | ||
111 | var elbAmazonCnRegex = regexp.MustCompile(`elb(.*?).amazonaws.com.cn$`) | ||
112 | |||
113 | // amazonS3HostPrivateLink - regular expression used to determine if an arg is s3 host in AWS PrivateLink interface endpoints style | ||
114 | var amazonS3HostPrivateLink = regexp.MustCompile(`^(?:bucket|accesspoint).vpce-.*?.s3.(.*?).vpce.amazonaws.com$`) | ||
115 | |||
116 | // GetRegionFromURL - returns a region from url host. | ||
117 | func GetRegionFromURL(endpointURL url.URL) string { | ||
118 | if endpointURL == sentinelURL { | ||
119 | return "" | ||
120 | } | ||
121 | if endpointURL.Host == "s3-external-1.amazonaws.com" { | ||
122 | return "" | ||
123 | } | ||
124 | |||
125 | // if elb's are used we cannot calculate which region it may be, just return empty. | ||
126 | if elbAmazonRegex.MatchString(endpointURL.Host) || elbAmazonCnRegex.MatchString(endpointURL.Host) { | ||
127 | return "" | ||
128 | } | ||
129 | |||
130 | // We check for FIPS dualstack matching first to avoid the non-greedy | ||
131 | // regex for FIPS non-dualstack matching a dualstack URL | ||
132 | parts := amazonS3HostFIPSDualStack.FindStringSubmatch(endpointURL.Host) | ||
133 | if len(parts) > 1 { | ||
134 | return parts[1] | ||
135 | } | ||
136 | |||
137 | parts = amazonS3HostFIPS.FindStringSubmatch(endpointURL.Host) | ||
138 | if len(parts) > 1 { | ||
139 | return parts[1] | ||
140 | } | ||
141 | |||
142 | parts = amazonS3HostDualStack.FindStringSubmatch(endpointURL.Host) | ||
143 | if len(parts) > 1 { | ||
144 | return parts[1] | ||
145 | } | ||
146 | |||
147 | parts = amazonS3HostHyphen.FindStringSubmatch(endpointURL.Host) | ||
148 | if len(parts) > 1 { | ||
149 | return parts[1] | ||
150 | } | ||
151 | |||
152 | parts = amazonS3ChinaHost.FindStringSubmatch(endpointURL.Host) | ||
153 | if len(parts) > 1 { | ||
154 | return parts[1] | ||
155 | } | ||
156 | |||
157 | parts = amazonS3ChinaHostDualStack.FindStringSubmatch(endpointURL.Host) | ||
158 | if len(parts) > 1 { | ||
159 | return parts[1] | ||
160 | } | ||
161 | |||
162 | parts = amazonS3HostDot.FindStringSubmatch(endpointURL.Host) | ||
163 | if len(parts) > 1 { | ||
164 | return parts[1] | ||
165 | } | ||
166 | |||
167 | parts = amazonS3HostPrivateLink.FindStringSubmatch(endpointURL.Host) | ||
168 | if len(parts) > 1 { | ||
169 | return parts[1] | ||
170 | } | ||
171 | |||
172 | return "" | ||
173 | } | ||
174 | |||
175 | // IsAliyunOSSEndpoint - Match if it is exactly Aliyun OSS endpoint. | ||
176 | func IsAliyunOSSEndpoint(endpointURL url.URL) bool { | ||
177 | return strings.HasSuffix(endpointURL.Host, "aliyuncs.com") | ||
178 | } | ||
179 | |||
180 | // IsAmazonEndpoint - Match if it is exactly Amazon S3 endpoint. | ||
181 | func IsAmazonEndpoint(endpointURL url.URL) bool { | ||
182 | if endpointURL.Host == "s3-external-1.amazonaws.com" || endpointURL.Host == "s3.amazonaws.com" { | ||
183 | return true | ||
184 | } | ||
185 | return GetRegionFromURL(endpointURL) != "" | ||
186 | } | ||
187 | |||
188 | // IsAmazonGovCloudEndpoint - Match if it is exactly Amazon S3 GovCloud endpoint. | ||
189 | func IsAmazonGovCloudEndpoint(endpointURL url.URL) bool { | ||
190 | if endpointURL == sentinelURL { | ||
191 | return false | ||
192 | } | ||
193 | return (endpointURL.Host == "s3-us-gov-west-1.amazonaws.com" || | ||
194 | endpointURL.Host == "s3-us-gov-east-1.amazonaws.com" || | ||
195 | IsAmazonFIPSGovCloudEndpoint(endpointURL)) | ||
196 | } | ||
197 | |||
198 | // IsAmazonFIPSGovCloudEndpoint - match if the endpoint is FIPS and GovCloud. | ||
199 | func IsAmazonFIPSGovCloudEndpoint(endpointURL url.URL) bool { | ||
200 | if endpointURL == sentinelURL { | ||
201 | return false | ||
202 | } | ||
203 | return IsAmazonFIPSEndpoint(endpointURL) && strings.Contains(endpointURL.Host, "us-gov-") | ||
204 | } | ||
205 | |||
206 | // IsAmazonFIPSEndpoint - Match if it is exactly Amazon S3 FIPS endpoint. | ||
207 | // See https://aws.amazon.com/compliance/fips. | ||
208 | func IsAmazonFIPSEndpoint(endpointURL url.URL) bool { | ||
209 | if endpointURL == sentinelURL { | ||
210 | return false | ||
211 | } | ||
212 | return strings.HasPrefix(endpointURL.Host, "s3-fips") && strings.HasSuffix(endpointURL.Host, ".amazonaws.com") | ||
213 | } | ||
214 | |||
215 | // IsAmazonPrivateLinkEndpoint - Match if it is exactly Amazon S3 PrivateLink interface endpoint | ||
216 | // See https://docs.aws.amazon.com/AmazonS3/latest/userguide/privatelink-interface-endpoints.html. | ||
217 | func IsAmazonPrivateLinkEndpoint(endpointURL url.URL) bool { | ||
218 | if endpointURL == sentinelURL { | ||
219 | return false | ||
220 | } | ||
221 | return amazonS3HostPrivateLink.MatchString(endpointURL.Host) | ||
222 | } | ||
223 | |||
224 | // IsGoogleEndpoint - Match if it is exactly Google cloud storage endpoint. | ||
225 | func IsGoogleEndpoint(endpointURL url.URL) bool { | ||
226 | if endpointURL == sentinelURL { | ||
227 | return false | ||
228 | } | ||
229 | return endpointURL.Host == "storage.googleapis.com" | ||
230 | } | ||
231 | |||
232 | // Expects ascii encoded strings - from output of urlEncodePath | ||
233 | func percentEncodeSlash(s string) string { | ||
234 | return strings.ReplaceAll(s, "/", "%2F") | ||
235 | } | ||
236 | |||
237 | // QueryEncode - encodes query values in their URL encoded form. In | ||
238 | // addition to the percent encoding performed by urlEncodePath() used | ||
239 | // here, it also percent encodes '/' (forward slash) | ||
240 | func QueryEncode(v url.Values) string { | ||
241 | if v == nil { | ||
242 | return "" | ||
243 | } | ||
244 | var buf bytes.Buffer | ||
245 | keys := make([]string, 0, len(v)) | ||
246 | for k := range v { | ||
247 | keys = append(keys, k) | ||
248 | } | ||
249 | sort.Strings(keys) | ||
250 | for _, k := range keys { | ||
251 | vs := v[k] | ||
252 | prefix := percentEncodeSlash(EncodePath(k)) + "=" | ||
253 | for _, v := range vs { | ||
254 | if buf.Len() > 0 { | ||
255 | buf.WriteByte('&') | ||
256 | } | ||
257 | buf.WriteString(prefix) | ||
258 | buf.WriteString(percentEncodeSlash(EncodePath(v))) | ||
259 | } | ||
260 | } | ||
261 | return buf.String() | ||
262 | } | ||
263 | |||
264 | // TagDecode - decodes canonical tag into map of key and value. | ||
265 | func TagDecode(ctag string) map[string]string { | ||
266 | if ctag == "" { | ||
267 | return map[string]string{} | ||
268 | } | ||
269 | tags := strings.Split(ctag, "&") | ||
270 | tagMap := make(map[string]string, len(tags)) | ||
271 | var err error | ||
272 | for _, tag := range tags { | ||
273 | kvs := strings.SplitN(tag, "=", 2) | ||
274 | if len(kvs) == 0 { | ||
275 | return map[string]string{} | ||
276 | } | ||
277 | if len(kvs) == 1 { | ||
278 | return map[string]string{} | ||
279 | } | ||
280 | tagMap[kvs[0]], err = url.PathUnescape(kvs[1]) | ||
281 | if err != nil { | ||
282 | continue | ||
283 | } | ||
284 | } | ||
285 | return tagMap | ||
286 | } | ||
287 | |||
288 | // TagEncode - encodes tag values in their URL encoded form. In | ||
289 | // addition to the percent encoding performed by urlEncodePath() used | ||
290 | // here, it also percent encodes '/' (forward slash) | ||
291 | func TagEncode(tags map[string]string) string { | ||
292 | if tags == nil { | ||
293 | return "" | ||
294 | } | ||
295 | values := url.Values{} | ||
296 | for k, v := range tags { | ||
297 | values[k] = []string{v} | ||
298 | } | ||
299 | return QueryEncode(values) | ||
300 | } | ||
301 | |||
302 | // if object matches reserved string, no need to encode them | ||
303 | var reservedObjectNames = regexp.MustCompile("^[a-zA-Z0-9-_.~/]+$") | ||
304 | |||
305 | // EncodePath encode the strings from UTF-8 byte representations to HTML hex escape sequences | ||
306 | // | ||
307 | // This is necessary since regular url.Parse() and url.Encode() functions do not support UTF-8 | ||
308 | // non english characters cannot be parsed due to the nature in which url.Encode() is written | ||
309 | // | ||
310 | // This function on the other hand is a direct replacement for url.Encode() technique to support | ||
311 | // pretty much every UTF-8 character. | ||
312 | func EncodePath(pathName string) string { | ||
313 | if reservedObjectNames.MatchString(pathName) { | ||
314 | return pathName | ||
315 | } | ||
316 | var encodedPathname strings.Builder | ||
317 | for _, s := range pathName { | ||
318 | if 'A' <= s && s <= 'Z' || 'a' <= s && s <= 'z' || '0' <= s && s <= '9' { // §2.3 Unreserved characters (mark) | ||
319 | encodedPathname.WriteRune(s) | ||
320 | continue | ||
321 | } | ||
322 | switch s { | ||
323 | case '-', '_', '.', '~', '/': // §2.3 Unreserved characters (mark) | ||
324 | encodedPathname.WriteRune(s) | ||
325 | continue | ||
326 | default: | ||
327 | l := utf8.RuneLen(s) | ||
328 | if l < 0 { | ||
329 | // if utf8 cannot convert return the same string as is | ||
330 | return pathName | ||
331 | } | ||
332 | u := make([]byte, l) | ||
333 | utf8.EncodeRune(u, s) | ||
334 | for _, r := range u { | ||
335 | hex := hex.EncodeToString([]byte{r}) | ||
336 | encodedPathname.WriteString("%" + strings.ToUpper(hex)) | ||
337 | } | ||
338 | } | ||
339 | } | ||
340 | return encodedPathname.String() | ||
341 | } | ||
342 | |||
343 | // We support '.' with bucket names but we fallback to using path | ||
344 | // style requests instead for such buckets. | ||
345 | var ( | ||
346 | validBucketName = regexp.MustCompile(`^[A-Za-z0-9][A-Za-z0-9\.\-\_\:]{1,61}[A-Za-z0-9]$`) | ||
347 | validBucketNameStrict = regexp.MustCompile(`^[a-z0-9][a-z0-9\.\-]{1,61}[a-z0-9]$`) | ||
348 | ipAddress = regexp.MustCompile(`^(\d+\.){3}\d+$`) | ||
349 | ) | ||
350 | |||
351 | // Common checker for both stricter and basic validation. | ||
352 | func checkBucketNameCommon(bucketName string, strict bool) (err error) { | ||
353 | if strings.TrimSpace(bucketName) == "" { | ||
354 | return errors.New("Bucket name cannot be empty") | ||
355 | } | ||
356 | if len(bucketName) < 3 { | ||
357 | return errors.New("Bucket name cannot be shorter than 3 characters") | ||
358 | } | ||
359 | if len(bucketName) > 63 { | ||
360 | return errors.New("Bucket name cannot be longer than 63 characters") | ||
361 | } | ||
362 | if ipAddress.MatchString(bucketName) { | ||
363 | return errors.New("Bucket name cannot be an ip address") | ||
364 | } | ||
365 | if strings.Contains(bucketName, "..") || strings.Contains(bucketName, ".-") || strings.Contains(bucketName, "-.") { | ||
366 | return errors.New("Bucket name contains invalid characters") | ||
367 | } | ||
368 | if strict { | ||
369 | if !validBucketNameStrict.MatchString(bucketName) { | ||
370 | err = errors.New("Bucket name contains invalid characters") | ||
371 | } | ||
372 | return err | ||
373 | } | ||
374 | if !validBucketName.MatchString(bucketName) { | ||
375 | err = errors.New("Bucket name contains invalid characters") | ||
376 | } | ||
377 | return err | ||
378 | } | ||
379 | |||
380 | // CheckValidBucketName - checks if we have a valid input bucket name. | ||
381 | func CheckValidBucketName(bucketName string) (err error) { | ||
382 | return checkBucketNameCommon(bucketName, false) | ||
383 | } | ||
384 | |||
385 | // CheckValidBucketNameStrict - checks if we have a valid input bucket name. | ||
386 | // This is a stricter version. | ||
387 | // - http://docs.aws.amazon.com/AmazonS3/latest/dev/UsingBucket.html | ||
388 | func CheckValidBucketNameStrict(bucketName string) (err error) { | ||
389 | return checkBucketNameCommon(bucketName, true) | ||
390 | } | ||
391 | |||
392 | // CheckValidObjectNamePrefix - checks if we have a valid input object name prefix. | ||
393 | // - http://docs.aws.amazon.com/AmazonS3/latest/dev/UsingMetadata.html | ||
394 | func CheckValidObjectNamePrefix(objectName string) error { | ||
395 | if len(objectName) > 1024 { | ||
396 | return errors.New("Object name cannot be longer than 1024 characters") | ||
397 | } | ||
398 | if !utf8.ValidString(objectName) { | ||
399 | return errors.New("Object name with non UTF-8 strings are not supported") | ||
400 | } | ||
401 | return nil | ||
402 | } | ||
403 | |||
404 | // CheckValidObjectName - checks if we have a valid input object name. | ||
405 | // - http://docs.aws.amazon.com/AmazonS3/latest/dev/UsingMetadata.html | ||
406 | func CheckValidObjectName(objectName string) error { | ||
407 | if strings.TrimSpace(objectName) == "" { | ||
408 | return errors.New("Object name cannot be empty") | ||
409 | } | ||
410 | return CheckValidObjectNamePrefix(objectName) | ||
411 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/pkg/set/stringset.go b/vendor/github.com/minio/minio-go/v7/pkg/set/stringset.go deleted file mode 100644 index c35e58e..0000000 --- a/vendor/github.com/minio/minio-go/v7/pkg/set/stringset.go +++ /dev/null | |||
@@ -1,200 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2015-2017 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package set | ||
19 | |||
20 | import ( | ||
21 | "fmt" | ||
22 | "sort" | ||
23 | |||
24 | jsoniter "github.com/json-iterator/go" | ||
25 | ) | ||
26 | |||
27 | // StringSet - uses map as set of strings. | ||
28 | type StringSet map[string]struct{} | ||
29 | |||
30 | var json = jsoniter.ConfigCompatibleWithStandardLibrary | ||
31 | |||
32 | // ToSlice - returns StringSet as string slice. | ||
33 | func (set StringSet) ToSlice() []string { | ||
34 | keys := make([]string, 0, len(set)) | ||
35 | for k := range set { | ||
36 | keys = append(keys, k) | ||
37 | } | ||
38 | sort.Strings(keys) | ||
39 | return keys | ||
40 | } | ||
41 | |||
42 | // IsEmpty - returns whether the set is empty or not. | ||
43 | func (set StringSet) IsEmpty() bool { | ||
44 | return len(set) == 0 | ||
45 | } | ||
46 | |||
47 | // Add - adds string to the set. | ||
48 | func (set StringSet) Add(s string) { | ||
49 | set[s] = struct{}{} | ||
50 | } | ||
51 | |||
52 | // Remove - removes string in the set. It does nothing if string does not exist in the set. | ||
53 | func (set StringSet) Remove(s string) { | ||
54 | delete(set, s) | ||
55 | } | ||
56 | |||
57 | // Contains - checks if string is in the set. | ||
58 | func (set StringSet) Contains(s string) bool { | ||
59 | _, ok := set[s] | ||
60 | return ok | ||
61 | } | ||
62 | |||
63 | // FuncMatch - returns new set containing each value who passes match function. | ||
64 | // A 'matchFn' should accept element in a set as first argument and | ||
65 | // 'matchString' as second argument. The function can do any logic to | ||
66 | // compare both the arguments and should return true to accept element in | ||
67 | // a set to include in output set else the element is ignored. | ||
68 | func (set StringSet) FuncMatch(matchFn func(string, string) bool, matchString string) StringSet { | ||
69 | nset := NewStringSet() | ||
70 | for k := range set { | ||
71 | if matchFn(k, matchString) { | ||
72 | nset.Add(k) | ||
73 | } | ||
74 | } | ||
75 | return nset | ||
76 | } | ||
77 | |||
78 | // ApplyFunc - returns new set containing each value processed by 'applyFn'. | ||
79 | // A 'applyFn' should accept element in a set as a argument and return | ||
80 | // a processed string. The function can do any logic to return a processed | ||
81 | // string. | ||
82 | func (set StringSet) ApplyFunc(applyFn func(string) string) StringSet { | ||
83 | nset := NewStringSet() | ||
84 | for k := range set { | ||
85 | nset.Add(applyFn(k)) | ||
86 | } | ||
87 | return nset | ||
88 | } | ||
89 | |||
90 | // Equals - checks whether given set is equal to current set or not. | ||
91 | func (set StringSet) Equals(sset StringSet) bool { | ||
92 | // If length of set is not equal to length of given set, the | ||
93 | // set is not equal to given set. | ||
94 | if len(set) != len(sset) { | ||
95 | return false | ||
96 | } | ||
97 | |||
98 | // As both sets are equal in length, check each elements are equal. | ||
99 | for k := range set { | ||
100 | if _, ok := sset[k]; !ok { | ||
101 | return false | ||
102 | } | ||
103 | } | ||
104 | |||
105 | return true | ||
106 | } | ||
107 | |||
108 | // Intersection - returns the intersection with given set as new set. | ||
109 | func (set StringSet) Intersection(sset StringSet) StringSet { | ||
110 | nset := NewStringSet() | ||
111 | for k := range set { | ||
112 | if _, ok := sset[k]; ok { | ||
113 | nset.Add(k) | ||
114 | } | ||
115 | } | ||
116 | |||
117 | return nset | ||
118 | } | ||
119 | |||
120 | // Difference - returns the difference with given set as new set. | ||
121 | func (set StringSet) Difference(sset StringSet) StringSet { | ||
122 | nset := NewStringSet() | ||
123 | for k := range set { | ||
124 | if _, ok := sset[k]; !ok { | ||
125 | nset.Add(k) | ||
126 | } | ||
127 | } | ||
128 | |||
129 | return nset | ||
130 | } | ||
131 | |||
132 | // Union - returns the union with given set as new set. | ||
133 | func (set StringSet) Union(sset StringSet) StringSet { | ||
134 | nset := NewStringSet() | ||
135 | for k := range set { | ||
136 | nset.Add(k) | ||
137 | } | ||
138 | |||
139 | for k := range sset { | ||
140 | nset.Add(k) | ||
141 | } | ||
142 | |||
143 | return nset | ||
144 | } | ||
145 | |||
146 | // MarshalJSON - converts to JSON data. | ||
147 | func (set StringSet) MarshalJSON() ([]byte, error) { | ||
148 | return json.Marshal(set.ToSlice()) | ||
149 | } | ||
150 | |||
151 | // UnmarshalJSON - parses JSON data and creates new set with it. | ||
152 | // If 'data' contains JSON string array, the set contains each string. | ||
153 | // If 'data' contains JSON string, the set contains the string as one element. | ||
154 | // If 'data' contains Other JSON types, JSON parse error is returned. | ||
155 | func (set *StringSet) UnmarshalJSON(data []byte) error { | ||
156 | sl := []string{} | ||
157 | var err error | ||
158 | if err = json.Unmarshal(data, &sl); err == nil { | ||
159 | *set = make(StringSet) | ||
160 | for _, s := range sl { | ||
161 | set.Add(s) | ||
162 | } | ||
163 | } else { | ||
164 | var s string | ||
165 | if err = json.Unmarshal(data, &s); err == nil { | ||
166 | *set = make(StringSet) | ||
167 | set.Add(s) | ||
168 | } | ||
169 | } | ||
170 | |||
171 | return err | ||
172 | } | ||
173 | |||
174 | // String - returns printable string of the set. | ||
175 | func (set StringSet) String() string { | ||
176 | return fmt.Sprintf("%s", set.ToSlice()) | ||
177 | } | ||
178 | |||
179 | // NewStringSet - creates new string set. | ||
180 | func NewStringSet() StringSet { | ||
181 | return make(StringSet) | ||
182 | } | ||
183 | |||
184 | // CreateStringSet - creates new string set with given string values. | ||
185 | func CreateStringSet(sl ...string) StringSet { | ||
186 | set := make(StringSet) | ||
187 | for _, k := range sl { | ||
188 | set.Add(k) | ||
189 | } | ||
190 | return set | ||
191 | } | ||
192 | |||
193 | // CopyStringSet - returns copy of given set. | ||
194 | func CopyStringSet(set StringSet) StringSet { | ||
195 | nset := NewStringSet() | ||
196 | for k, v := range set { | ||
197 | nset[k] = v | ||
198 | } | ||
199 | return nset | ||
200 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/pkg/signer/request-signature-streaming-unsigned-trailer.go b/vendor/github.com/minio/minio-go/v7/pkg/signer/request-signature-streaming-unsigned-trailer.go deleted file mode 100644 index 77540e2..0000000 --- a/vendor/github.com/minio/minio-go/v7/pkg/signer/request-signature-streaming-unsigned-trailer.go +++ /dev/null | |||
@@ -1,224 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2022 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package signer | ||
19 | |||
20 | import ( | ||
21 | "bytes" | ||
22 | "fmt" | ||
23 | "io" | ||
24 | "net/http" | ||
25 | "strconv" | ||
26 | "strings" | ||
27 | "time" | ||
28 | ) | ||
29 | |||
30 | // getUnsignedChunkLength - calculates the length of chunk metadata | ||
31 | func getUnsignedChunkLength(chunkDataSize int64) int64 { | ||
32 | return int64(len(fmt.Sprintf("%x", chunkDataSize))) + | ||
33 | crlfLen + | ||
34 | chunkDataSize + | ||
35 | crlfLen | ||
36 | } | ||
37 | |||
38 | // getUSStreamLength - calculates the length of the overall stream (data + metadata) | ||
39 | func getUSStreamLength(dataLen, chunkSize int64, trailers http.Header) int64 { | ||
40 | if dataLen <= 0 { | ||
41 | return 0 | ||
42 | } | ||
43 | |||
44 | chunksCount := int64(dataLen / chunkSize) | ||
45 | remainingBytes := int64(dataLen % chunkSize) | ||
46 | streamLen := int64(0) | ||
47 | streamLen += chunksCount * getUnsignedChunkLength(chunkSize) | ||
48 | if remainingBytes > 0 { | ||
49 | streamLen += getUnsignedChunkLength(remainingBytes) | ||
50 | } | ||
51 | streamLen += getUnsignedChunkLength(0) | ||
52 | if len(trailers) > 0 { | ||
53 | for name, placeholder := range trailers { | ||
54 | if len(placeholder) > 0 { | ||
55 | streamLen += int64(len(name) + len(trailerKVSeparator) + len(placeholder[0]) + 1) | ||
56 | } | ||
57 | } | ||
58 | streamLen += crlfLen | ||
59 | } | ||
60 | |||
61 | return streamLen | ||
62 | } | ||
63 | |||
64 | // prepareStreamingRequest - prepares a request with appropriate | ||
65 | // headers before computing the seed signature. | ||
66 | func prepareUSStreamingRequest(req *http.Request, sessionToken string, dataLen int64, timestamp time.Time) { | ||
67 | req.TransferEncoding = []string{"aws-chunked"} | ||
68 | if sessionToken != "" { | ||
69 | req.Header.Set("X-Amz-Security-Token", sessionToken) | ||
70 | } | ||
71 | |||
72 | req.Header.Set("X-Amz-Date", timestamp.Format(iso8601DateFormat)) | ||
73 | // Set content length with streaming signature for each chunk included. | ||
74 | req.ContentLength = getUSStreamLength(dataLen, int64(payloadChunkSize), req.Trailer) | ||
75 | } | ||
76 | |||
77 | // StreamingUSReader implements chunked upload signature as a reader on | ||
78 | // top of req.Body's ReaderCloser chunk header;data;... repeat | ||
79 | type StreamingUSReader struct { | ||
80 | contentLen int64 // Content-Length from req header | ||
81 | baseReadCloser io.ReadCloser // underlying io.Reader | ||
82 | bytesRead int64 // bytes read from underlying io.Reader | ||
83 | buf bytes.Buffer // holds signed chunk | ||
84 | chunkBuf []byte // holds raw data read from req Body | ||
85 | chunkBufLen int // no. of bytes read so far into chunkBuf | ||
86 | done bool // done reading the underlying reader to EOF | ||
87 | chunkNum int | ||
88 | totalChunks int | ||
89 | lastChunkSize int | ||
90 | trailer http.Header | ||
91 | } | ||
92 | |||
93 | // writeChunk - signs a chunk read from s.baseReader of chunkLen size. | ||
94 | func (s *StreamingUSReader) writeChunk(chunkLen int, addCrLf bool) { | ||
95 | s.buf.WriteString(strconv.FormatInt(int64(chunkLen), 16) + "\r\n") | ||
96 | |||
97 | // Write chunk data into streaming buffer | ||
98 | s.buf.Write(s.chunkBuf[:chunkLen]) | ||
99 | |||
100 | // Write the chunk trailer. | ||
101 | if addCrLf { | ||
102 | s.buf.Write([]byte("\r\n")) | ||
103 | } | ||
104 | |||
105 | // Reset chunkBufLen for next chunk read. | ||
106 | s.chunkBufLen = 0 | ||
107 | s.chunkNum++ | ||
108 | } | ||
109 | |||
110 | // addSignedTrailer - adds a trailer with the provided headers, | ||
111 | // then signs a chunk and adds it to output. | ||
112 | func (s *StreamingUSReader) addTrailer(h http.Header) { | ||
113 | olen := len(s.chunkBuf) | ||
114 | s.chunkBuf = s.chunkBuf[:0] | ||
115 | for k, v := range h { | ||
116 | s.chunkBuf = append(s.chunkBuf, []byte(strings.ToLower(k)+trailerKVSeparator+v[0]+"\n")...) | ||
117 | } | ||
118 | |||
119 | s.buf.Write(s.chunkBuf) | ||
120 | s.buf.WriteString("\r\n\r\n") | ||
121 | |||
122 | // Reset chunkBufLen for next chunk read. | ||
123 | s.chunkBuf = s.chunkBuf[:olen] | ||
124 | s.chunkBufLen = 0 | ||
125 | s.chunkNum++ | ||
126 | } | ||
127 | |||
128 | // StreamingUnsignedV4 - provides chunked upload | ||
129 | func StreamingUnsignedV4(req *http.Request, sessionToken string, dataLen int64, reqTime time.Time) *http.Request { | ||
130 | // Set headers needed for streaming signature. | ||
131 | prepareUSStreamingRequest(req, sessionToken, dataLen, reqTime) | ||
132 | |||
133 | if req.Body == nil { | ||
134 | req.Body = io.NopCloser(bytes.NewReader([]byte(""))) | ||
135 | } | ||
136 | |||
137 | stReader := &StreamingUSReader{ | ||
138 | baseReadCloser: req.Body, | ||
139 | chunkBuf: make([]byte, payloadChunkSize), | ||
140 | contentLen: dataLen, | ||
141 | chunkNum: 1, | ||
142 | totalChunks: int((dataLen+payloadChunkSize-1)/payloadChunkSize) + 1, | ||
143 | lastChunkSize: int(dataLen % payloadChunkSize), | ||
144 | } | ||
145 | if len(req.Trailer) > 0 { | ||
146 | stReader.trailer = req.Trailer | ||
147 | // Remove... | ||
148 | req.Trailer = nil | ||
149 | } | ||
150 | |||
151 | req.Body = stReader | ||
152 | |||
153 | return req | ||
154 | } | ||
155 | |||
156 | // Read - this method performs chunk upload signature providing a | ||
157 | // io.Reader interface. | ||
158 | func (s *StreamingUSReader) Read(buf []byte) (int, error) { | ||
159 | switch { | ||
160 | // After the last chunk is read from underlying reader, we | ||
161 | // never re-fill s.buf. | ||
162 | case s.done: | ||
163 | |||
164 | // s.buf will be (re-)filled with next chunk when has lesser | ||
165 | // bytes than asked for. | ||
166 | case s.buf.Len() < len(buf): | ||
167 | s.chunkBufLen = 0 | ||
168 | for { | ||
169 | n1, err := s.baseReadCloser.Read(s.chunkBuf[s.chunkBufLen:]) | ||
170 | // Usually we validate `err` first, but in this case | ||
171 | // we are validating n > 0 for the following reasons. | ||
172 | // | ||
173 | // 1. n > 0, err is one of io.EOF, nil (near end of stream) | ||
174 | // A Reader returning a non-zero number of bytes at the end | ||
175 | // of the input stream may return either err == EOF or err == nil | ||
176 | // | ||
177 | // 2. n == 0, err is io.EOF (actual end of stream) | ||
178 | // | ||
179 | // Callers should always process the n > 0 bytes returned | ||
180 | // before considering the error err. | ||
181 | if n1 > 0 { | ||
182 | s.chunkBufLen += n1 | ||
183 | s.bytesRead += int64(n1) | ||
184 | |||
185 | if s.chunkBufLen == payloadChunkSize || | ||
186 | (s.chunkNum == s.totalChunks-1 && | ||
187 | s.chunkBufLen == s.lastChunkSize) { | ||
188 | // Sign the chunk and write it to s.buf. | ||
189 | s.writeChunk(s.chunkBufLen, true) | ||
190 | break | ||
191 | } | ||
192 | } | ||
193 | if err != nil { | ||
194 | if err == io.EOF { | ||
195 | // No more data left in baseReader - last chunk. | ||
196 | // Done reading the last chunk from baseReader. | ||
197 | s.done = true | ||
198 | |||
199 | // bytes read from baseReader different than | ||
200 | // content length provided. | ||
201 | if s.bytesRead != s.contentLen { | ||
202 | return 0, fmt.Errorf("http: ContentLength=%d with Body length %d", s.contentLen, s.bytesRead) | ||
203 | } | ||
204 | |||
205 | // Sign the chunk and write it to s.buf. | ||
206 | s.writeChunk(0, len(s.trailer) == 0) | ||
207 | if len(s.trailer) > 0 { | ||
208 | // Trailer must be set now. | ||
209 | s.addTrailer(s.trailer) | ||
210 | } | ||
211 | break | ||
212 | } | ||
213 | return 0, err | ||
214 | } | ||
215 | |||
216 | } | ||
217 | } | ||
218 | return s.buf.Read(buf) | ||
219 | } | ||
220 | |||
221 | // Close - this method makes underlying io.ReadCloser's Close method available. | ||
222 | func (s *StreamingUSReader) Close() error { | ||
223 | return s.baseReadCloser.Close() | ||
224 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/pkg/signer/request-signature-streaming.go b/vendor/github.com/minio/minio-go/v7/pkg/signer/request-signature-streaming.go deleted file mode 100644 index 1c2f1dc..0000000 --- a/vendor/github.com/minio/minio-go/v7/pkg/signer/request-signature-streaming.go +++ /dev/null | |||
@@ -1,403 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2017 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package signer | ||
19 | |||
20 | import ( | ||
21 | "bytes" | ||
22 | "encoding/hex" | ||
23 | "fmt" | ||
24 | "io" | ||
25 | "net/http" | ||
26 | "strconv" | ||
27 | "strings" | ||
28 | "time" | ||
29 | |||
30 | md5simd "github.com/minio/md5-simd" | ||
31 | ) | ||
32 | |||
33 | // Reference for constants used below - | ||
34 | // http://docs.aws.amazon.com/AmazonS3/latest/API/sigv4-streaming.html#example-signature-calculations-streaming | ||
35 | const ( | ||
36 | streamingSignAlgorithm = "STREAMING-AWS4-HMAC-SHA256-PAYLOAD" | ||
37 | streamingSignTrailerAlgorithm = "STREAMING-AWS4-HMAC-SHA256-PAYLOAD-TRAILER" | ||
38 | streamingPayloadHdr = "AWS4-HMAC-SHA256-PAYLOAD" | ||
39 | streamingTrailerHdr = "AWS4-HMAC-SHA256-TRAILER" | ||
40 | emptySHA256 = "e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855" | ||
41 | payloadChunkSize = 64 * 1024 | ||
42 | chunkSigConstLen = 17 // ";chunk-signature=" | ||
43 | signatureStrLen = 64 // e.g. "f2ca1bb6c7e907d06dafe4687e579fce76b37e4e93b7605022da52e6ccc26fd2" | ||
44 | crlfLen = 2 // CRLF | ||
45 | trailerKVSeparator = ":" | ||
46 | trailerSignature = "x-amz-trailer-signature" | ||
47 | ) | ||
48 | |||
49 | // Request headers to be ignored while calculating seed signature for | ||
50 | // a request. | ||
51 | var ignoredStreamingHeaders = map[string]bool{ | ||
52 | "Authorization": true, | ||
53 | "User-Agent": true, | ||
54 | "Content-Type": true, | ||
55 | } | ||
56 | |||
57 | // getSignedChunkLength - calculates the length of chunk metadata | ||
58 | func getSignedChunkLength(chunkDataSize int64) int64 { | ||
59 | return int64(len(fmt.Sprintf("%x", chunkDataSize))) + | ||
60 | chunkSigConstLen + | ||
61 | signatureStrLen + | ||
62 | crlfLen + | ||
63 | chunkDataSize + | ||
64 | crlfLen | ||
65 | } | ||
66 | |||
67 | // getStreamLength - calculates the length of the overall stream (data + metadata) | ||
68 | func getStreamLength(dataLen, chunkSize int64, trailers http.Header) int64 { | ||
69 | if dataLen <= 0 { | ||
70 | return 0 | ||
71 | } | ||
72 | |||
73 | chunksCount := int64(dataLen / chunkSize) | ||
74 | remainingBytes := int64(dataLen % chunkSize) | ||
75 | streamLen := int64(0) | ||
76 | streamLen += chunksCount * getSignedChunkLength(chunkSize) | ||
77 | if remainingBytes > 0 { | ||
78 | streamLen += getSignedChunkLength(remainingBytes) | ||
79 | } | ||
80 | streamLen += getSignedChunkLength(0) | ||
81 | if len(trailers) > 0 { | ||
82 | for name, placeholder := range trailers { | ||
83 | if len(placeholder) > 0 { | ||
84 | streamLen += int64(len(name) + len(trailerKVSeparator) + len(placeholder[0]) + 1) | ||
85 | } | ||
86 | } | ||
87 | streamLen += int64(len(trailerSignature)+len(trailerKVSeparator)) + signatureStrLen + crlfLen + crlfLen | ||
88 | } | ||
89 | |||
90 | return streamLen | ||
91 | } | ||
92 | |||
93 | // buildChunkStringToSign - returns the string to sign given chunk data | ||
94 | // and previous signature. | ||
95 | func buildChunkStringToSign(t time.Time, region, previousSig, chunkChecksum string) string { | ||
96 | stringToSignParts := []string{ | ||
97 | streamingPayloadHdr, | ||
98 | t.Format(iso8601DateFormat), | ||
99 | getScope(region, t, ServiceTypeS3), | ||
100 | previousSig, | ||
101 | emptySHA256, | ||
102 | chunkChecksum, | ||
103 | } | ||
104 | |||
105 | return strings.Join(stringToSignParts, "\n") | ||
106 | } | ||
107 | |||
108 | // buildTrailerChunkStringToSign - returns the string to sign given chunk data | ||
109 | // and previous signature. | ||
110 | func buildTrailerChunkStringToSign(t time.Time, region, previousSig, chunkChecksum string) string { | ||
111 | stringToSignParts := []string{ | ||
112 | streamingTrailerHdr, | ||
113 | t.Format(iso8601DateFormat), | ||
114 | getScope(region, t, ServiceTypeS3), | ||
115 | previousSig, | ||
116 | chunkChecksum, | ||
117 | } | ||
118 | |||
119 | return strings.Join(stringToSignParts, "\n") | ||
120 | } | ||
121 | |||
122 | // prepareStreamingRequest - prepares a request with appropriate | ||
123 | // headers before computing the seed signature. | ||
124 | func prepareStreamingRequest(req *http.Request, sessionToken string, dataLen int64, timestamp time.Time) { | ||
125 | // Set x-amz-content-sha256 header. | ||
126 | if len(req.Trailer) == 0 { | ||
127 | req.Header.Set("X-Amz-Content-Sha256", streamingSignAlgorithm) | ||
128 | } else { | ||
129 | req.Header.Set("X-Amz-Content-Sha256", streamingSignTrailerAlgorithm) | ||
130 | for k := range req.Trailer { | ||
131 | req.Header.Add("X-Amz-Trailer", strings.ToLower(k)) | ||
132 | } | ||
133 | req.TransferEncoding = []string{"aws-chunked"} | ||
134 | } | ||
135 | |||
136 | if sessionToken != "" { | ||
137 | req.Header.Set("X-Amz-Security-Token", sessionToken) | ||
138 | } | ||
139 | |||
140 | req.Header.Set("X-Amz-Date", timestamp.Format(iso8601DateFormat)) | ||
141 | // Set content length with streaming signature for each chunk included. | ||
142 | req.ContentLength = getStreamLength(dataLen, int64(payloadChunkSize), req.Trailer) | ||
143 | req.Header.Set("x-amz-decoded-content-length", strconv.FormatInt(dataLen, 10)) | ||
144 | } | ||
145 | |||
146 | // buildChunkHeader - returns the chunk header. | ||
147 | // e.g string(IntHexBase(chunk-size)) + ";chunk-signature=" + signature + \r\n + chunk-data + \r\n | ||
148 | func buildChunkHeader(chunkLen int64, signature string) []byte { | ||
149 | return []byte(strconv.FormatInt(chunkLen, 16) + ";chunk-signature=" + signature + "\r\n") | ||
150 | } | ||
151 | |||
152 | // buildChunkSignature - returns chunk signature for a given chunk and previous signature. | ||
153 | func buildChunkSignature(chunkCheckSum string, reqTime time.Time, region, | ||
154 | previousSignature, secretAccessKey string, | ||
155 | ) string { | ||
156 | chunkStringToSign := buildChunkStringToSign(reqTime, region, | ||
157 | previousSignature, chunkCheckSum) | ||
158 | signingKey := getSigningKey(secretAccessKey, region, reqTime, ServiceTypeS3) | ||
159 | return getSignature(signingKey, chunkStringToSign) | ||
160 | } | ||
161 | |||
162 | // buildChunkSignature - returns chunk signature for a given chunk and previous signature. | ||
163 | func buildTrailerChunkSignature(chunkChecksum string, reqTime time.Time, region, | ||
164 | previousSignature, secretAccessKey string, | ||
165 | ) string { | ||
166 | chunkStringToSign := buildTrailerChunkStringToSign(reqTime, region, | ||
167 | previousSignature, chunkChecksum) | ||
168 | signingKey := getSigningKey(secretAccessKey, region, reqTime, ServiceTypeS3) | ||
169 | return getSignature(signingKey, chunkStringToSign) | ||
170 | } | ||
171 | |||
172 | // getSeedSignature - returns the seed signature for a given request. | ||
173 | func (s *StreamingReader) setSeedSignature(req *http.Request) { | ||
174 | // Get canonical request | ||
175 | canonicalRequest := getCanonicalRequest(*req, ignoredStreamingHeaders, getHashedPayload(*req)) | ||
176 | |||
177 | // Get string to sign from canonical request. | ||
178 | stringToSign := getStringToSignV4(s.reqTime, s.region, canonicalRequest, ServiceTypeS3) | ||
179 | |||
180 | signingKey := getSigningKey(s.secretAccessKey, s.region, s.reqTime, ServiceTypeS3) | ||
181 | |||
182 | // Calculate signature. | ||
183 | s.seedSignature = getSignature(signingKey, stringToSign) | ||
184 | } | ||
185 | |||
186 | // StreamingReader implements chunked upload signature as a reader on | ||
187 | // top of req.Body's ReaderCloser chunk header;data;... repeat | ||
188 | type StreamingReader struct { | ||
189 | accessKeyID string | ||
190 | secretAccessKey string | ||
191 | sessionToken string | ||
192 | region string | ||
193 | prevSignature string | ||
194 | seedSignature string | ||
195 | contentLen int64 // Content-Length from req header | ||
196 | baseReadCloser io.ReadCloser // underlying io.Reader | ||
197 | bytesRead int64 // bytes read from underlying io.Reader | ||
198 | buf bytes.Buffer // holds signed chunk | ||
199 | chunkBuf []byte // holds raw data read from req Body | ||
200 | chunkBufLen int // no. of bytes read so far into chunkBuf | ||
201 | done bool // done reading the underlying reader to EOF | ||
202 | reqTime time.Time | ||
203 | chunkNum int | ||
204 | totalChunks int | ||
205 | lastChunkSize int | ||
206 | trailer http.Header | ||
207 | sh256 md5simd.Hasher | ||
208 | } | ||
209 | |||
210 | // signChunk - signs a chunk read from s.baseReader of chunkLen size. | ||
211 | func (s *StreamingReader) signChunk(chunkLen int, addCrLf bool) { | ||
212 | // Compute chunk signature for next header | ||
213 | s.sh256.Reset() | ||
214 | s.sh256.Write(s.chunkBuf[:chunkLen]) | ||
215 | chunckChecksum := hex.EncodeToString(s.sh256.Sum(nil)) | ||
216 | |||
217 | signature := buildChunkSignature(chunckChecksum, s.reqTime, | ||
218 | s.region, s.prevSignature, s.secretAccessKey) | ||
219 | |||
220 | // For next chunk signature computation | ||
221 | s.prevSignature = signature | ||
222 | |||
223 | // Write chunk header into streaming buffer | ||
224 | chunkHdr := buildChunkHeader(int64(chunkLen), signature) | ||
225 | s.buf.Write(chunkHdr) | ||
226 | |||
227 | // Write chunk data into streaming buffer | ||
228 | s.buf.Write(s.chunkBuf[:chunkLen]) | ||
229 | |||
230 | // Write the chunk trailer. | ||
231 | if addCrLf { | ||
232 | s.buf.Write([]byte("\r\n")) | ||
233 | } | ||
234 | |||
235 | // Reset chunkBufLen for next chunk read. | ||
236 | s.chunkBufLen = 0 | ||
237 | s.chunkNum++ | ||
238 | } | ||
239 | |||
240 | // addSignedTrailer - adds a trailer with the provided headers, | ||
241 | // then signs a chunk and adds it to output. | ||
242 | func (s *StreamingReader) addSignedTrailer(h http.Header) { | ||
243 | olen := len(s.chunkBuf) | ||
244 | s.chunkBuf = s.chunkBuf[:0] | ||
245 | for k, v := range h { | ||
246 | s.chunkBuf = append(s.chunkBuf, []byte(strings.ToLower(k)+trailerKVSeparator+v[0]+"\n")...) | ||
247 | } | ||
248 | |||
249 | s.sh256.Reset() | ||
250 | s.sh256.Write(s.chunkBuf) | ||
251 | chunkChecksum := hex.EncodeToString(s.sh256.Sum(nil)) | ||
252 | // Compute chunk signature | ||
253 | signature := buildTrailerChunkSignature(chunkChecksum, s.reqTime, | ||
254 | s.region, s.prevSignature, s.secretAccessKey) | ||
255 | |||
256 | // For next chunk signature computation | ||
257 | s.prevSignature = signature | ||
258 | |||
259 | s.buf.Write(s.chunkBuf) | ||
260 | s.buf.WriteString("\r\n" + trailerSignature + trailerKVSeparator + signature + "\r\n\r\n") | ||
261 | |||
262 | // Reset chunkBufLen for next chunk read. | ||
263 | s.chunkBuf = s.chunkBuf[:olen] | ||
264 | s.chunkBufLen = 0 | ||
265 | s.chunkNum++ | ||
266 | } | ||
267 | |||
268 | // setStreamingAuthHeader - builds and sets authorization header value | ||
269 | // for streaming signature. | ||
270 | func (s *StreamingReader) setStreamingAuthHeader(req *http.Request) { | ||
271 | credential := GetCredential(s.accessKeyID, s.region, s.reqTime, ServiceTypeS3) | ||
272 | authParts := []string{ | ||
273 | signV4Algorithm + " Credential=" + credential, | ||
274 | "SignedHeaders=" + getSignedHeaders(*req, ignoredStreamingHeaders), | ||
275 | "Signature=" + s.seedSignature, | ||
276 | } | ||
277 | |||
278 | // Set authorization header. | ||
279 | auth := strings.Join(authParts, ",") | ||
280 | req.Header.Set("Authorization", auth) | ||
281 | } | ||
282 | |||
283 | // StreamingSignV4 - provides chunked upload signatureV4 support by | ||
284 | // implementing io.Reader. | ||
285 | func StreamingSignV4(req *http.Request, accessKeyID, secretAccessKey, sessionToken, | ||
286 | region string, dataLen int64, reqTime time.Time, sh256 md5simd.Hasher, | ||
287 | ) *http.Request { | ||
288 | // Set headers needed for streaming signature. | ||
289 | prepareStreamingRequest(req, sessionToken, dataLen, reqTime) | ||
290 | |||
291 | if req.Body == nil { | ||
292 | req.Body = io.NopCloser(bytes.NewReader([]byte(""))) | ||
293 | } | ||
294 | |||
295 | stReader := &StreamingReader{ | ||
296 | baseReadCloser: req.Body, | ||
297 | accessKeyID: accessKeyID, | ||
298 | secretAccessKey: secretAccessKey, | ||
299 | sessionToken: sessionToken, | ||
300 | region: region, | ||
301 | reqTime: reqTime, | ||
302 | chunkBuf: make([]byte, payloadChunkSize), | ||
303 | contentLen: dataLen, | ||
304 | chunkNum: 1, | ||
305 | totalChunks: int((dataLen+payloadChunkSize-1)/payloadChunkSize) + 1, | ||
306 | lastChunkSize: int(dataLen % payloadChunkSize), | ||
307 | sh256: sh256, | ||
308 | } | ||
309 | if len(req.Trailer) > 0 { | ||
310 | stReader.trailer = req.Trailer | ||
311 | // Remove... | ||
312 | req.Trailer = nil | ||
313 | } | ||
314 | |||
315 | // Add the request headers required for chunk upload signing. | ||
316 | |||
317 | // Compute the seed signature. | ||
318 | stReader.setSeedSignature(req) | ||
319 | |||
320 | // Set the authorization header with the seed signature. | ||
321 | stReader.setStreamingAuthHeader(req) | ||
322 | |||
323 | // Set seed signature as prevSignature for subsequent | ||
324 | // streaming signing process. | ||
325 | stReader.prevSignature = stReader.seedSignature | ||
326 | req.Body = stReader | ||
327 | |||
328 | return req | ||
329 | } | ||
330 | |||
331 | // Read - this method performs chunk upload signature providing a | ||
332 | // io.Reader interface. | ||
333 | func (s *StreamingReader) Read(buf []byte) (int, error) { | ||
334 | switch { | ||
335 | // After the last chunk is read from underlying reader, we | ||
336 | // never re-fill s.buf. | ||
337 | case s.done: | ||
338 | |||
339 | // s.buf will be (re-)filled with next chunk when has lesser | ||
340 | // bytes than asked for. | ||
341 | case s.buf.Len() < len(buf): | ||
342 | s.chunkBufLen = 0 | ||
343 | for { | ||
344 | n1, err := s.baseReadCloser.Read(s.chunkBuf[s.chunkBufLen:]) | ||
345 | // Usually we validate `err` first, but in this case | ||
346 | // we are validating n > 0 for the following reasons. | ||
347 | // | ||
348 | // 1. n > 0, err is one of io.EOF, nil (near end of stream) | ||
349 | // A Reader returning a non-zero number of bytes at the end | ||
350 | // of the input stream may return either err == EOF or err == nil | ||
351 | // | ||
352 | // 2. n == 0, err is io.EOF (actual end of stream) | ||
353 | // | ||
354 | // Callers should always process the n > 0 bytes returned | ||
355 | // before considering the error err. | ||
356 | if n1 > 0 { | ||
357 | s.chunkBufLen += n1 | ||
358 | s.bytesRead += int64(n1) | ||
359 | |||
360 | if s.chunkBufLen == payloadChunkSize || | ||
361 | (s.chunkNum == s.totalChunks-1 && | ||
362 | s.chunkBufLen == s.lastChunkSize) { | ||
363 | // Sign the chunk and write it to s.buf. | ||
364 | s.signChunk(s.chunkBufLen, true) | ||
365 | break | ||
366 | } | ||
367 | } | ||
368 | if err != nil { | ||
369 | if err == io.EOF { | ||
370 | // No more data left in baseReader - last chunk. | ||
371 | // Done reading the last chunk from baseReader. | ||
372 | s.done = true | ||
373 | |||
374 | // bytes read from baseReader different than | ||
375 | // content length provided. | ||
376 | if s.bytesRead != s.contentLen { | ||
377 | return 0, fmt.Errorf("http: ContentLength=%d with Body length %d", s.contentLen, s.bytesRead) | ||
378 | } | ||
379 | |||
380 | // Sign the chunk and write it to s.buf. | ||
381 | s.signChunk(0, len(s.trailer) == 0) | ||
382 | if len(s.trailer) > 0 { | ||
383 | // Trailer must be set now. | ||
384 | s.addSignedTrailer(s.trailer) | ||
385 | } | ||
386 | break | ||
387 | } | ||
388 | return 0, err | ||
389 | } | ||
390 | |||
391 | } | ||
392 | } | ||
393 | return s.buf.Read(buf) | ||
394 | } | ||
395 | |||
396 | // Close - this method makes underlying io.ReadCloser's Close method available. | ||
397 | func (s *StreamingReader) Close() error { | ||
398 | if s.sh256 != nil { | ||
399 | s.sh256.Close() | ||
400 | s.sh256 = nil | ||
401 | } | ||
402 | return s.baseReadCloser.Close() | ||
403 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/pkg/signer/request-signature-v2.go b/vendor/github.com/minio/minio-go/v7/pkg/signer/request-signature-v2.go deleted file mode 100644 index fa4f8c9..0000000 --- a/vendor/github.com/minio/minio-go/v7/pkg/signer/request-signature-v2.go +++ /dev/null | |||
@@ -1,319 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2015-2017 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package signer | ||
19 | |||
20 | import ( | ||
21 | "bytes" | ||
22 | "crypto/hmac" | ||
23 | "crypto/sha1" | ||
24 | "encoding/base64" | ||
25 | "fmt" | ||
26 | "net/http" | ||
27 | "net/url" | ||
28 | "sort" | ||
29 | "strconv" | ||
30 | "strings" | ||
31 | "time" | ||
32 | |||
33 | "github.com/minio/minio-go/v7/pkg/s3utils" | ||
34 | ) | ||
35 | |||
36 | // Signature and API related constants. | ||
37 | const ( | ||
38 | signV2Algorithm = "AWS" | ||
39 | ) | ||
40 | |||
41 | // Encode input URL path to URL encoded path. | ||
42 | func encodeURL2Path(req *http.Request, virtualHost bool) (path string) { | ||
43 | if virtualHost { | ||
44 | reqHost := getHostAddr(req) | ||
45 | dotPos := strings.Index(reqHost, ".") | ||
46 | if dotPos > -1 { | ||
47 | bucketName := reqHost[:dotPos] | ||
48 | path = "/" + bucketName | ||
49 | path += req.URL.Path | ||
50 | path = s3utils.EncodePath(path) | ||
51 | return | ||
52 | } | ||
53 | } | ||
54 | path = s3utils.EncodePath(req.URL.Path) | ||
55 | return | ||
56 | } | ||
57 | |||
58 | // PreSignV2 - presign the request in following style. | ||
59 | // https://${S3_BUCKET}.s3.amazonaws.com/${S3_OBJECT}?AWSAccessKeyId=${S3_ACCESS_KEY}&Expires=${TIMESTAMP}&Signature=${SIGNATURE}. | ||
60 | func PreSignV2(req http.Request, accessKeyID, secretAccessKey string, expires int64, virtualHost bool) *http.Request { | ||
61 | // Presign is not needed for anonymous credentials. | ||
62 | if accessKeyID == "" || secretAccessKey == "" { | ||
63 | return &req | ||
64 | } | ||
65 | |||
66 | d := time.Now().UTC() | ||
67 | // Find epoch expires when the request will expire. | ||
68 | epochExpires := d.Unix() + expires | ||
69 | |||
70 | // Add expires header if not present. | ||
71 | if expiresStr := req.Header.Get("Expires"); expiresStr == "" { | ||
72 | req.Header.Set("Expires", strconv.FormatInt(epochExpires, 10)) | ||
73 | } | ||
74 | |||
75 | // Get presigned string to sign. | ||
76 | stringToSign := preStringToSignV2(req, virtualHost) | ||
77 | hm := hmac.New(sha1.New, []byte(secretAccessKey)) | ||
78 | hm.Write([]byte(stringToSign)) | ||
79 | |||
80 | // Calculate signature. | ||
81 | signature := base64.StdEncoding.EncodeToString(hm.Sum(nil)) | ||
82 | |||
83 | query := req.URL.Query() | ||
84 | // Handle specially for Google Cloud Storage. | ||
85 | if strings.Contains(getHostAddr(&req), ".storage.googleapis.com") { | ||
86 | query.Set("GoogleAccessId", accessKeyID) | ||
87 | } else { | ||
88 | query.Set("AWSAccessKeyId", accessKeyID) | ||
89 | } | ||
90 | |||
91 | // Fill in Expires for presigned query. | ||
92 | query.Set("Expires", strconv.FormatInt(epochExpires, 10)) | ||
93 | |||
94 | // Encode query and save. | ||
95 | req.URL.RawQuery = s3utils.QueryEncode(query) | ||
96 | |||
97 | // Save signature finally. | ||
98 | req.URL.RawQuery += "&Signature=" + s3utils.EncodePath(signature) | ||
99 | |||
100 | // Return. | ||
101 | return &req | ||
102 | } | ||
103 | |||
104 | // PostPresignSignatureV2 - presigned signature for PostPolicy | ||
105 | // request. | ||
106 | func PostPresignSignatureV2(policyBase64, secretAccessKey string) string { | ||
107 | hm := hmac.New(sha1.New, []byte(secretAccessKey)) | ||
108 | hm.Write([]byte(policyBase64)) | ||
109 | signature := base64.StdEncoding.EncodeToString(hm.Sum(nil)) | ||
110 | return signature | ||
111 | } | ||
112 | |||
113 | // Authorization = "AWS" + " " + AWSAccessKeyId + ":" + Signature; | ||
114 | // Signature = Base64( HMAC-SHA1( YourSecretAccessKeyID, UTF-8-Encoding-Of( StringToSign ) ) ); | ||
115 | // | ||
116 | // StringToSign = HTTP-Verb + "\n" + | ||
117 | // Content-Md5 + "\n" + | ||
118 | // Content-Type + "\n" + | ||
119 | // Date + "\n" + | ||
120 | // CanonicalizedProtocolHeaders + | ||
121 | // CanonicalizedResource; | ||
122 | // | ||
123 | // CanonicalizedResource = [ "/" + Bucket ] + | ||
124 | // <HTTP-Request-URI, from the protocol name up to the query string> + | ||
125 | // [ subresource, if present. For example "?acl", "?location", "?logging", or "?torrent"]; | ||
126 | // | ||
127 | // CanonicalizedProtocolHeaders = <described below> | ||
128 | |||
129 | // SignV2 sign the request before Do() (AWS Signature Version 2). | ||
130 | func SignV2(req http.Request, accessKeyID, secretAccessKey string, virtualHost bool) *http.Request { | ||
131 | // Signature calculation is not needed for anonymous credentials. | ||
132 | if accessKeyID == "" || secretAccessKey == "" { | ||
133 | return &req | ||
134 | } | ||
135 | |||
136 | // Initial time. | ||
137 | d := time.Now().UTC() | ||
138 | |||
139 | // Add date if not present. | ||
140 | if date := req.Header.Get("Date"); date == "" { | ||
141 | req.Header.Set("Date", d.Format(http.TimeFormat)) | ||
142 | } | ||
143 | |||
144 | // Calculate HMAC for secretAccessKey. | ||
145 | stringToSign := stringToSignV2(req, virtualHost) | ||
146 | hm := hmac.New(sha1.New, []byte(secretAccessKey)) | ||
147 | hm.Write([]byte(stringToSign)) | ||
148 | |||
149 | // Prepare auth header. | ||
150 | authHeader := new(bytes.Buffer) | ||
151 | authHeader.WriteString(fmt.Sprintf("%s %s:", signV2Algorithm, accessKeyID)) | ||
152 | encoder := base64.NewEncoder(base64.StdEncoding, authHeader) | ||
153 | encoder.Write(hm.Sum(nil)) | ||
154 | encoder.Close() | ||
155 | |||
156 | // Set Authorization header. | ||
157 | req.Header.Set("Authorization", authHeader.String()) | ||
158 | |||
159 | return &req | ||
160 | } | ||
161 | |||
162 | // From the Amazon docs: | ||
163 | // | ||
164 | // StringToSign = HTTP-Verb + "\n" + | ||
165 | // | ||
166 | // Content-Md5 + "\n" + | ||
167 | // Content-Type + "\n" + | ||
168 | // Expires + "\n" + | ||
169 | // CanonicalizedProtocolHeaders + | ||
170 | // CanonicalizedResource; | ||
171 | func preStringToSignV2(req http.Request, virtualHost bool) string { | ||
172 | buf := new(bytes.Buffer) | ||
173 | // Write standard headers. | ||
174 | writePreSignV2Headers(buf, req) | ||
175 | // Write canonicalized protocol headers if any. | ||
176 | writeCanonicalizedHeaders(buf, req) | ||
177 | // Write canonicalized Query resources if any. | ||
178 | writeCanonicalizedResource(buf, req, virtualHost) | ||
179 | return buf.String() | ||
180 | } | ||
181 | |||
182 | // writePreSignV2Headers - write preSign v2 required headers. | ||
183 | func writePreSignV2Headers(buf *bytes.Buffer, req http.Request) { | ||
184 | buf.WriteString(req.Method + "\n") | ||
185 | buf.WriteString(req.Header.Get("Content-Md5") + "\n") | ||
186 | buf.WriteString(req.Header.Get("Content-Type") + "\n") | ||
187 | buf.WriteString(req.Header.Get("Expires") + "\n") | ||
188 | } | ||
189 | |||
190 | // From the Amazon docs: | ||
191 | // | ||
192 | // StringToSign = HTTP-Verb + "\n" + | ||
193 | // | ||
194 | // Content-Md5 + "\n" + | ||
195 | // Content-Type + "\n" + | ||
196 | // Date + "\n" + | ||
197 | // CanonicalizedProtocolHeaders + | ||
198 | // CanonicalizedResource; | ||
199 | func stringToSignV2(req http.Request, virtualHost bool) string { | ||
200 | buf := new(bytes.Buffer) | ||
201 | // Write standard headers. | ||
202 | writeSignV2Headers(buf, req) | ||
203 | // Write canonicalized protocol headers if any. | ||
204 | writeCanonicalizedHeaders(buf, req) | ||
205 | // Write canonicalized Query resources if any. | ||
206 | writeCanonicalizedResource(buf, req, virtualHost) | ||
207 | return buf.String() | ||
208 | } | ||
209 | |||
210 | // writeSignV2Headers - write signV2 required headers. | ||
211 | func writeSignV2Headers(buf *bytes.Buffer, req http.Request) { | ||
212 | buf.WriteString(req.Method + "\n") | ||
213 | buf.WriteString(req.Header.Get("Content-Md5") + "\n") | ||
214 | buf.WriteString(req.Header.Get("Content-Type") + "\n") | ||
215 | buf.WriteString(req.Header.Get("Date") + "\n") | ||
216 | } | ||
217 | |||
218 | // writeCanonicalizedHeaders - write canonicalized headers. | ||
219 | func writeCanonicalizedHeaders(buf *bytes.Buffer, req http.Request) { | ||
220 | var protoHeaders []string | ||
221 | vals := make(map[string][]string) | ||
222 | for k, vv := range req.Header { | ||
223 | // All the AMZ headers should be lowercase | ||
224 | lk := strings.ToLower(k) | ||
225 | if strings.HasPrefix(lk, "x-amz") { | ||
226 | protoHeaders = append(protoHeaders, lk) | ||
227 | vals[lk] = vv | ||
228 | } | ||
229 | } | ||
230 | sort.Strings(protoHeaders) | ||
231 | for _, k := range protoHeaders { | ||
232 | buf.WriteString(k) | ||
233 | buf.WriteByte(':') | ||
234 | for idx, v := range vals[k] { | ||
235 | if idx > 0 { | ||
236 | buf.WriteByte(',') | ||
237 | } | ||
238 | buf.WriteString(v) | ||
239 | } | ||
240 | buf.WriteByte('\n') | ||
241 | } | ||
242 | } | ||
243 | |||
244 | // AWS S3 Signature V2 calculation rule is give here: | ||
245 | // http://docs.aws.amazon.com/AmazonS3/latest/dev/RESTAuthentication.html#RESTAuthenticationStringToSign | ||
246 | |||
247 | // Whitelist resource list that will be used in query string for signature-V2 calculation. | ||
248 | // | ||
249 | // This list should be kept alphabetically sorted, do not hastily edit. | ||
250 | var resourceList = []string{ | ||
251 | "acl", | ||
252 | "cors", | ||
253 | "delete", | ||
254 | "encryption", | ||
255 | "legal-hold", | ||
256 | "lifecycle", | ||
257 | "location", | ||
258 | "logging", | ||
259 | "notification", | ||
260 | "partNumber", | ||
261 | "policy", | ||
262 | "replication", | ||
263 | "requestPayment", | ||
264 | "response-cache-control", | ||
265 | "response-content-disposition", | ||
266 | "response-content-encoding", | ||
267 | "response-content-language", | ||
268 | "response-content-type", | ||
269 | "response-expires", | ||
270 | "retention", | ||
271 | "select", | ||
272 | "select-type", | ||
273 | "tagging", | ||
274 | "torrent", | ||
275 | "uploadId", | ||
276 | "uploads", | ||
277 | "versionId", | ||
278 | "versioning", | ||
279 | "versions", | ||
280 | "website", | ||
281 | } | ||
282 | |||
283 | // From the Amazon docs: | ||
284 | // | ||
285 | // CanonicalizedResource = [ "/" + Bucket ] + | ||
286 | // | ||
287 | // <HTTP-Request-URI, from the protocol name up to the query string> + | ||
288 | // [ sub-resource, if present. For example "?acl", "?location", "?logging", or "?torrent"]; | ||
289 | func writeCanonicalizedResource(buf *bytes.Buffer, req http.Request, virtualHost bool) { | ||
290 | // Save request URL. | ||
291 | requestURL := req.URL | ||
292 | // Get encoded URL path. | ||
293 | buf.WriteString(encodeURL2Path(&req, virtualHost)) | ||
294 | if requestURL.RawQuery != "" { | ||
295 | var n int | ||
296 | vals, _ := url.ParseQuery(requestURL.RawQuery) | ||
297 | // Verify if any sub resource queries are present, if yes | ||
298 | // canonicallize them. | ||
299 | for _, resource := range resourceList { | ||
300 | if vv, ok := vals[resource]; ok && len(vv) > 0 { | ||
301 | n++ | ||
302 | // First element | ||
303 | switch n { | ||
304 | case 1: | ||
305 | buf.WriteByte('?') | ||
306 | // The rest | ||
307 | default: | ||
308 | buf.WriteByte('&') | ||
309 | } | ||
310 | buf.WriteString(resource) | ||
311 | // Request parameters | ||
312 | if len(vv[0]) > 0 { | ||
313 | buf.WriteByte('=') | ||
314 | buf.WriteString(vv[0]) | ||
315 | } | ||
316 | } | ||
317 | } | ||
318 | } | ||
319 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/pkg/signer/request-signature-v4.go b/vendor/github.com/minio/minio-go/v7/pkg/signer/request-signature-v4.go deleted file mode 100644 index ffd2514..0000000 --- a/vendor/github.com/minio/minio-go/v7/pkg/signer/request-signature-v4.go +++ /dev/null | |||
@@ -1,351 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2015-2017 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package signer | ||
19 | |||
20 | import ( | ||
21 | "bytes" | ||
22 | "encoding/hex" | ||
23 | "net/http" | ||
24 | "sort" | ||
25 | "strconv" | ||
26 | "strings" | ||
27 | "time" | ||
28 | |||
29 | "github.com/minio/minio-go/v7/pkg/s3utils" | ||
30 | ) | ||
31 | |||
32 | // Signature and API related constants. | ||
33 | const ( | ||
34 | signV4Algorithm = "AWS4-HMAC-SHA256" | ||
35 | iso8601DateFormat = "20060102T150405Z" | ||
36 | yyyymmdd = "20060102" | ||
37 | ) | ||
38 | |||
39 | // Different service types | ||
40 | const ( | ||
41 | ServiceTypeS3 = "s3" | ||
42 | ServiceTypeSTS = "sts" | ||
43 | ) | ||
44 | |||
45 | // Excerpts from @lsegal - | ||
46 | // https:/github.com/aws/aws-sdk-js/issues/659#issuecomment-120477258. | ||
47 | // | ||
48 | // * User-Agent | ||
49 | // This is ignored from signing because signing this causes problems with generating pre-signed | ||
50 | // URLs (that are executed by other agents) or when customers pass requests through proxies, which | ||
51 | // may modify the user-agent. | ||
52 | // | ||
53 | // * Authorization | ||
54 | // Is skipped for obvious reasons. | ||
55 | // | ||
56 | // * Accept-Encoding | ||
57 | // Some S3 servers like Hitachi Content Platform do not honor this header for signature | ||
58 | // calculation. | ||
59 | var v4IgnoredHeaders = map[string]bool{ | ||
60 | "Accept-Encoding": true, | ||
61 | "Authorization": true, | ||
62 | "User-Agent": true, | ||
63 | } | ||
64 | |||
65 | // getSigningKey hmac seed to calculate final signature. | ||
66 | func getSigningKey(secret, loc string, t time.Time, serviceType string) []byte { | ||
67 | date := sumHMAC([]byte("AWS4"+secret), []byte(t.Format(yyyymmdd))) | ||
68 | location := sumHMAC(date, []byte(loc)) | ||
69 | service := sumHMAC(location, []byte(serviceType)) | ||
70 | signingKey := sumHMAC(service, []byte("aws4_request")) | ||
71 | return signingKey | ||
72 | } | ||
73 | |||
74 | // getSignature final signature in hexadecimal form. | ||
75 | func getSignature(signingKey []byte, stringToSign string) string { | ||
76 | return hex.EncodeToString(sumHMAC(signingKey, []byte(stringToSign))) | ||
77 | } | ||
78 | |||
79 | // getScope generate a string of a specific date, an AWS region, and a | ||
80 | // service. | ||
81 | func getScope(location string, t time.Time, serviceType string) string { | ||
82 | scope := strings.Join([]string{ | ||
83 | t.Format(yyyymmdd), | ||
84 | location, | ||
85 | serviceType, | ||
86 | "aws4_request", | ||
87 | }, "/") | ||
88 | return scope | ||
89 | } | ||
90 | |||
91 | // GetCredential generate a credential string. | ||
92 | func GetCredential(accessKeyID, location string, t time.Time, serviceType string) string { | ||
93 | scope := getScope(location, t, serviceType) | ||
94 | return accessKeyID + "/" + scope | ||
95 | } | ||
96 | |||
97 | // getHashedPayload get the hexadecimal value of the SHA256 hash of | ||
98 | // the request payload. | ||
99 | func getHashedPayload(req http.Request) string { | ||
100 | hashedPayload := req.Header.Get("X-Amz-Content-Sha256") | ||
101 | if hashedPayload == "" { | ||
102 | // Presign does not have a payload, use S3 recommended value. | ||
103 | hashedPayload = unsignedPayload | ||
104 | } | ||
105 | return hashedPayload | ||
106 | } | ||
107 | |||
108 | // getCanonicalHeaders generate a list of request headers for | ||
109 | // signature. | ||
110 | func getCanonicalHeaders(req http.Request, ignoredHeaders map[string]bool) string { | ||
111 | var headers []string | ||
112 | vals := make(map[string][]string) | ||
113 | for k, vv := range req.Header { | ||
114 | if _, ok := ignoredHeaders[http.CanonicalHeaderKey(k)]; ok { | ||
115 | continue // ignored header | ||
116 | } | ||
117 | headers = append(headers, strings.ToLower(k)) | ||
118 | vals[strings.ToLower(k)] = vv | ||
119 | } | ||
120 | if !headerExists("host", headers) { | ||
121 | headers = append(headers, "host") | ||
122 | } | ||
123 | sort.Strings(headers) | ||
124 | |||
125 | var buf bytes.Buffer | ||
126 | // Save all the headers in canonical form <header>:<value> newline | ||
127 | // separated for each header. | ||
128 | for _, k := range headers { | ||
129 | buf.WriteString(k) | ||
130 | buf.WriteByte(':') | ||
131 | switch { | ||
132 | case k == "host": | ||
133 | buf.WriteString(getHostAddr(&req)) | ||
134 | buf.WriteByte('\n') | ||
135 | default: | ||
136 | for idx, v := range vals[k] { | ||
137 | if idx > 0 { | ||
138 | buf.WriteByte(',') | ||
139 | } | ||
140 | buf.WriteString(signV4TrimAll(v)) | ||
141 | } | ||
142 | buf.WriteByte('\n') | ||
143 | } | ||
144 | } | ||
145 | return buf.String() | ||
146 | } | ||
147 | |||
148 | func headerExists(key string, headers []string) bool { | ||
149 | for _, k := range headers { | ||
150 | if k == key { | ||
151 | return true | ||
152 | } | ||
153 | } | ||
154 | return false | ||
155 | } | ||
156 | |||
157 | // getSignedHeaders generate all signed request headers. | ||
158 | // i.e lexically sorted, semicolon-separated list of lowercase | ||
159 | // request header names. | ||
160 | func getSignedHeaders(req http.Request, ignoredHeaders map[string]bool) string { | ||
161 | var headers []string | ||
162 | for k := range req.Header { | ||
163 | if _, ok := ignoredHeaders[http.CanonicalHeaderKey(k)]; ok { | ||
164 | continue // Ignored header found continue. | ||
165 | } | ||
166 | headers = append(headers, strings.ToLower(k)) | ||
167 | } | ||
168 | if !headerExists("host", headers) { | ||
169 | headers = append(headers, "host") | ||
170 | } | ||
171 | sort.Strings(headers) | ||
172 | return strings.Join(headers, ";") | ||
173 | } | ||
174 | |||
175 | // getCanonicalRequest generate a canonical request of style. | ||
176 | // | ||
177 | // canonicalRequest = | ||
178 | // | ||
179 | // <HTTPMethod>\n | ||
180 | // <CanonicalURI>\n | ||
181 | // <CanonicalQueryString>\n | ||
182 | // <CanonicalHeaders>\n | ||
183 | // <SignedHeaders>\n | ||
184 | // <HashedPayload> | ||
185 | func getCanonicalRequest(req http.Request, ignoredHeaders map[string]bool, hashedPayload string) string { | ||
186 | req.URL.RawQuery = strings.ReplaceAll(req.URL.Query().Encode(), "+", "%20") | ||
187 | canonicalRequest := strings.Join([]string{ | ||
188 | req.Method, | ||
189 | s3utils.EncodePath(req.URL.Path), | ||
190 | req.URL.RawQuery, | ||
191 | getCanonicalHeaders(req, ignoredHeaders), | ||
192 | getSignedHeaders(req, ignoredHeaders), | ||
193 | hashedPayload, | ||
194 | }, "\n") | ||
195 | return canonicalRequest | ||
196 | } | ||
197 | |||
198 | // getStringToSign a string based on selected query values. | ||
199 | func getStringToSignV4(t time.Time, location, canonicalRequest, serviceType string) string { | ||
200 | stringToSign := signV4Algorithm + "\n" + t.Format(iso8601DateFormat) + "\n" | ||
201 | stringToSign = stringToSign + getScope(location, t, serviceType) + "\n" | ||
202 | stringToSign += hex.EncodeToString(sum256([]byte(canonicalRequest))) | ||
203 | return stringToSign | ||
204 | } | ||
205 | |||
206 | // PreSignV4 presign the request, in accordance with | ||
207 | // http://docs.aws.amazon.com/AmazonS3/latest/API/sigv4-query-string-auth.html. | ||
208 | func PreSignV4(req http.Request, accessKeyID, secretAccessKey, sessionToken, location string, expires int64) *http.Request { | ||
209 | // Presign is not needed for anonymous credentials. | ||
210 | if accessKeyID == "" || secretAccessKey == "" { | ||
211 | return &req | ||
212 | } | ||
213 | |||
214 | // Initial time. | ||
215 | t := time.Now().UTC() | ||
216 | |||
217 | // Get credential string. | ||
218 | credential := GetCredential(accessKeyID, location, t, ServiceTypeS3) | ||
219 | |||
220 | // Get all signed headers. | ||
221 | signedHeaders := getSignedHeaders(req, v4IgnoredHeaders) | ||
222 | |||
223 | // Set URL query. | ||
224 | query := req.URL.Query() | ||
225 | query.Set("X-Amz-Algorithm", signV4Algorithm) | ||
226 | query.Set("X-Amz-Date", t.Format(iso8601DateFormat)) | ||
227 | query.Set("X-Amz-Expires", strconv.FormatInt(expires, 10)) | ||
228 | query.Set("X-Amz-SignedHeaders", signedHeaders) | ||
229 | query.Set("X-Amz-Credential", credential) | ||
230 | // Set session token if available. | ||
231 | if sessionToken != "" { | ||
232 | query.Set("X-Amz-Security-Token", sessionToken) | ||
233 | } | ||
234 | req.URL.RawQuery = query.Encode() | ||
235 | |||
236 | // Get canonical request. | ||
237 | canonicalRequest := getCanonicalRequest(req, v4IgnoredHeaders, getHashedPayload(req)) | ||
238 | |||
239 | // Get string to sign from canonical request. | ||
240 | stringToSign := getStringToSignV4(t, location, canonicalRequest, ServiceTypeS3) | ||
241 | |||
242 | // Gext hmac signing key. | ||
243 | signingKey := getSigningKey(secretAccessKey, location, t, ServiceTypeS3) | ||
244 | |||
245 | // Calculate signature. | ||
246 | signature := getSignature(signingKey, stringToSign) | ||
247 | |||
248 | // Add signature header to RawQuery. | ||
249 | req.URL.RawQuery += "&X-Amz-Signature=" + signature | ||
250 | |||
251 | return &req | ||
252 | } | ||
253 | |||
254 | // PostPresignSignatureV4 - presigned signature for PostPolicy | ||
255 | // requests. | ||
256 | func PostPresignSignatureV4(policyBase64 string, t time.Time, secretAccessKey, location string) string { | ||
257 | // Get signining key. | ||
258 | signingkey := getSigningKey(secretAccessKey, location, t, ServiceTypeS3) | ||
259 | // Calculate signature. | ||
260 | signature := getSignature(signingkey, policyBase64) | ||
261 | return signature | ||
262 | } | ||
263 | |||
264 | // SignV4STS - signature v4 for STS request. | ||
265 | func SignV4STS(req http.Request, accessKeyID, secretAccessKey, location string) *http.Request { | ||
266 | return signV4(req, accessKeyID, secretAccessKey, "", location, ServiceTypeSTS, nil) | ||
267 | } | ||
268 | |||
269 | // Internal function called for different service types. | ||
270 | func signV4(req http.Request, accessKeyID, secretAccessKey, sessionToken, location, serviceType string, trailer http.Header) *http.Request { | ||
271 | // Signature calculation is not needed for anonymous credentials. | ||
272 | if accessKeyID == "" || secretAccessKey == "" { | ||
273 | return &req | ||
274 | } | ||
275 | |||
276 | // Initial time. | ||
277 | t := time.Now().UTC() | ||
278 | |||
279 | // Set x-amz-date. | ||
280 | req.Header.Set("X-Amz-Date", t.Format(iso8601DateFormat)) | ||
281 | |||
282 | // Set session token if available. | ||
283 | if sessionToken != "" { | ||
284 | req.Header.Set("X-Amz-Security-Token", sessionToken) | ||
285 | } | ||
286 | |||
287 | if len(trailer) > 0 { | ||
288 | for k := range trailer { | ||
289 | req.Header.Add("X-Amz-Trailer", strings.ToLower(k)) | ||
290 | } | ||
291 | |||
292 | req.Header.Set("Content-Encoding", "aws-chunked") | ||
293 | req.Header.Set("x-amz-decoded-content-length", strconv.FormatInt(req.ContentLength, 10)) | ||
294 | } | ||
295 | |||
296 | hashedPayload := getHashedPayload(req) | ||
297 | if serviceType == ServiceTypeSTS { | ||
298 | // Content sha256 header is not sent with the request | ||
299 | // but it is expected to have sha256 of payload for signature | ||
300 | // in STS service type request. | ||
301 | req.Header.Del("X-Amz-Content-Sha256") | ||
302 | } | ||
303 | |||
304 | // Get canonical request. | ||
305 | canonicalRequest := getCanonicalRequest(req, v4IgnoredHeaders, hashedPayload) | ||
306 | |||
307 | // Get string to sign from canonical request. | ||
308 | stringToSign := getStringToSignV4(t, location, canonicalRequest, serviceType) | ||
309 | |||
310 | // Get hmac signing key. | ||
311 | signingKey := getSigningKey(secretAccessKey, location, t, serviceType) | ||
312 | |||
313 | // Get credential string. | ||
314 | credential := GetCredential(accessKeyID, location, t, serviceType) | ||
315 | |||
316 | // Get all signed headers. | ||
317 | signedHeaders := getSignedHeaders(req, v4IgnoredHeaders) | ||
318 | |||
319 | // Calculate signature. | ||
320 | signature := getSignature(signingKey, stringToSign) | ||
321 | |||
322 | // If regular request, construct the final authorization header. | ||
323 | parts := []string{ | ||
324 | signV4Algorithm + " Credential=" + credential, | ||
325 | "SignedHeaders=" + signedHeaders, | ||
326 | "Signature=" + signature, | ||
327 | } | ||
328 | |||
329 | // Set authorization header. | ||
330 | auth := strings.Join(parts, ", ") | ||
331 | req.Header.Set("Authorization", auth) | ||
332 | |||
333 | if len(trailer) > 0 { | ||
334 | // Use custom chunked encoding. | ||
335 | req.Trailer = trailer | ||
336 | return StreamingUnsignedV4(&req, sessionToken, req.ContentLength, time.Now().UTC()) | ||
337 | } | ||
338 | return &req | ||
339 | } | ||
340 | |||
341 | // SignV4 sign the request before Do(), in accordance with | ||
342 | // http://docs.aws.amazon.com/AmazonS3/latest/API/sig-v4-authenticating-requests.html. | ||
343 | func SignV4(req http.Request, accessKeyID, secretAccessKey, sessionToken, location string) *http.Request { | ||
344 | return signV4(req, accessKeyID, secretAccessKey, sessionToken, location, ServiceTypeS3, nil) | ||
345 | } | ||
346 | |||
347 | // SignV4Trailer sign the request before Do(), in accordance with | ||
348 | // http://docs.aws.amazon.com/AmazonS3/latest/API/sig-v4-authenticating-requests.html | ||
349 | func SignV4Trailer(req http.Request, accessKeyID, secretAccessKey, sessionToken, location string, trailer http.Header) *http.Request { | ||
350 | return signV4(req, accessKeyID, secretAccessKey, sessionToken, location, ServiceTypeS3, trailer) | ||
351 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/pkg/signer/utils.go b/vendor/github.com/minio/minio-go/v7/pkg/signer/utils.go deleted file mode 100644 index 87c9939..0000000 --- a/vendor/github.com/minio/minio-go/v7/pkg/signer/utils.go +++ /dev/null | |||
@@ -1,62 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2015-2017 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package signer | ||
19 | |||
20 | import ( | ||
21 | "crypto/hmac" | ||
22 | "crypto/sha256" | ||
23 | "net/http" | ||
24 | "strings" | ||
25 | ) | ||
26 | |||
27 | // unsignedPayload - value to be set to X-Amz-Content-Sha256 header when | ||
28 | const unsignedPayload = "UNSIGNED-PAYLOAD" | ||
29 | |||
30 | // sum256 calculate sha256 sum for an input byte array. | ||
31 | func sum256(data []byte) []byte { | ||
32 | hash := sha256.New() | ||
33 | hash.Write(data) | ||
34 | return hash.Sum(nil) | ||
35 | } | ||
36 | |||
37 | // sumHMAC calculate hmac between two input byte array. | ||
38 | func sumHMAC(key, data []byte) []byte { | ||
39 | hash := hmac.New(sha256.New, key) | ||
40 | hash.Write(data) | ||
41 | return hash.Sum(nil) | ||
42 | } | ||
43 | |||
44 | // getHostAddr returns host header if available, otherwise returns host from URL | ||
45 | func getHostAddr(req *http.Request) string { | ||
46 | host := req.Header.Get("host") | ||
47 | if host != "" && req.Host != host { | ||
48 | return host | ||
49 | } | ||
50 | if req.Host != "" { | ||
51 | return req.Host | ||
52 | } | ||
53 | return req.URL.Host | ||
54 | } | ||
55 | |||
56 | // Trim leading and trailing spaces and replace sequential spaces with one space, following Trimall() | ||
57 | // in http://docs.aws.amazon.com/general/latest/gr/sigv4-create-canonical-request.html | ||
58 | func signV4TrimAll(input string) string { | ||
59 | // Compress adjacent spaces (a space is determined by | ||
60 | // unicode.IsSpace() internally here) to one space and return | ||
61 | return strings.Join(strings.Fields(input), " ") | ||
62 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/pkg/sse/sse.go b/vendor/github.com/minio/minio-go/v7/pkg/sse/sse.go deleted file mode 100644 index b5fb956..0000000 --- a/vendor/github.com/minio/minio-go/v7/pkg/sse/sse.go +++ /dev/null | |||
@@ -1,66 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2020 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package sse | ||
19 | |||
20 | import "encoding/xml" | ||
21 | |||
22 | // ApplySSEByDefault defines default encryption configuration, KMS or SSE. To activate | ||
23 | // KMS, SSEAlgoritm needs to be set to "aws:kms" | ||
24 | // Minio currently does not support Kms. | ||
25 | type ApplySSEByDefault struct { | ||
26 | KmsMasterKeyID string `xml:"KMSMasterKeyID,omitempty"` | ||
27 | SSEAlgorithm string `xml:"SSEAlgorithm"` | ||
28 | } | ||
29 | |||
30 | // Rule layer encapsulates default encryption configuration | ||
31 | type Rule struct { | ||
32 | Apply ApplySSEByDefault `xml:"ApplyServerSideEncryptionByDefault"` | ||
33 | } | ||
34 | |||
35 | // Configuration is the default encryption configuration structure | ||
36 | type Configuration struct { | ||
37 | XMLName xml.Name `xml:"ServerSideEncryptionConfiguration"` | ||
38 | Rules []Rule `xml:"Rule"` | ||
39 | } | ||
40 | |||
41 | // NewConfigurationSSES3 initializes a new SSE-S3 configuration | ||
42 | func NewConfigurationSSES3() *Configuration { | ||
43 | return &Configuration{ | ||
44 | Rules: []Rule{ | ||
45 | { | ||
46 | Apply: ApplySSEByDefault{ | ||
47 | SSEAlgorithm: "AES256", | ||
48 | }, | ||
49 | }, | ||
50 | }, | ||
51 | } | ||
52 | } | ||
53 | |||
54 | // NewConfigurationSSEKMS initializes a new SSE-KMS configuration | ||
55 | func NewConfigurationSSEKMS(kmsMasterKey string) *Configuration { | ||
56 | return &Configuration{ | ||
57 | Rules: []Rule{ | ||
58 | { | ||
59 | Apply: ApplySSEByDefault{ | ||
60 | KmsMasterKeyID: kmsMasterKey, | ||
61 | SSEAlgorithm: "aws:kms", | ||
62 | }, | ||
63 | }, | ||
64 | }, | ||
65 | } | ||
66 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/pkg/tags/tags.go b/vendor/github.com/minio/minio-go/v7/pkg/tags/tags.go deleted file mode 100644 index 7a84a6f..0000000 --- a/vendor/github.com/minio/minio-go/v7/pkg/tags/tags.go +++ /dev/null | |||
@@ -1,413 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2020-2022 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package tags | ||
19 | |||
20 | import ( | ||
21 | "encoding/xml" | ||
22 | "io" | ||
23 | "net/url" | ||
24 | "regexp" | ||
25 | "sort" | ||
26 | "strings" | ||
27 | "unicode/utf8" | ||
28 | ) | ||
29 | |||
30 | // Error contains tag specific error. | ||
31 | type Error interface { | ||
32 | error | ||
33 | Code() string | ||
34 | } | ||
35 | |||
36 | type errTag struct { | ||
37 | code string | ||
38 | message string | ||
39 | } | ||
40 | |||
41 | // Code contains error code. | ||
42 | func (err errTag) Code() string { | ||
43 | return err.code | ||
44 | } | ||
45 | |||
46 | // Error contains error message. | ||
47 | func (err errTag) Error() string { | ||
48 | return err.message | ||
49 | } | ||
50 | |||
51 | var ( | ||
52 | errTooManyObjectTags = &errTag{"BadRequest", "Tags cannot be more than 10"} | ||
53 | errTooManyTags = &errTag{"BadRequest", "Tags cannot be more than 50"} | ||
54 | errInvalidTagKey = &errTag{"InvalidTag", "The TagKey you have provided is invalid"} | ||
55 | errInvalidTagValue = &errTag{"InvalidTag", "The TagValue you have provided is invalid"} | ||
56 | errDuplicateTagKey = &errTag{"InvalidTag", "Cannot provide multiple Tags with the same key"} | ||
57 | ) | ||
58 | |||
59 | // Tag comes with limitation as per | ||
60 | // https://docs.aws.amazon.com/AmazonS3/latest/dev/object-tagging.html amd | ||
61 | // https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/Using_Tags.html#tag-restrictions | ||
62 | const ( | ||
63 | maxKeyLength = 128 | ||
64 | maxValueLength = 256 | ||
65 | maxObjectTagCount = 10 | ||
66 | maxTagCount = 50 | ||
67 | ) | ||
68 | |||
69 | // https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/Using_Tags.html#tag-restrictions | ||
70 | // borrowed from this article and also testing various ASCII characters following regex | ||
71 | // is supported by AWS S3 for both tags and values. | ||
72 | var validTagKeyValue = regexp.MustCompile(`^[a-zA-Z0-9-+\-._:/@ ]+$`) | ||
73 | |||
74 | func checkKey(key string) error { | ||
75 | if len(key) == 0 { | ||
76 | return errInvalidTagKey | ||
77 | } | ||
78 | |||
79 | if utf8.RuneCountInString(key) > maxKeyLength || !validTagKeyValue.MatchString(key) { | ||
80 | return errInvalidTagKey | ||
81 | } | ||
82 | |||
83 | return nil | ||
84 | } | ||
85 | |||
86 | func checkValue(value string) error { | ||
87 | if value != "" { | ||
88 | if utf8.RuneCountInString(value) > maxValueLength || !validTagKeyValue.MatchString(value) { | ||
89 | return errInvalidTagValue | ||
90 | } | ||
91 | } | ||
92 | |||
93 | return nil | ||
94 | } | ||
95 | |||
96 | // Tag denotes key and value. | ||
97 | type Tag struct { | ||
98 | Key string `xml:"Key"` | ||
99 | Value string `xml:"Value"` | ||
100 | } | ||
101 | |||
102 | func (tag Tag) String() string { | ||
103 | return tag.Key + "=" + tag.Value | ||
104 | } | ||
105 | |||
106 | // IsEmpty returns whether this tag is empty or not. | ||
107 | func (tag Tag) IsEmpty() bool { | ||
108 | return tag.Key == "" | ||
109 | } | ||
110 | |||
111 | // Validate checks this tag. | ||
112 | func (tag Tag) Validate() error { | ||
113 | if err := checkKey(tag.Key); err != nil { | ||
114 | return err | ||
115 | } | ||
116 | |||
117 | return checkValue(tag.Value) | ||
118 | } | ||
119 | |||
120 | // MarshalXML encodes to XML data. | ||
121 | func (tag Tag) MarshalXML(e *xml.Encoder, start xml.StartElement) error { | ||
122 | if err := tag.Validate(); err != nil { | ||
123 | return err | ||
124 | } | ||
125 | |||
126 | type subTag Tag // to avoid recursively calling MarshalXML() | ||
127 | return e.EncodeElement(subTag(tag), start) | ||
128 | } | ||
129 | |||
130 | // UnmarshalXML decodes XML data to tag. | ||
131 | func (tag *Tag) UnmarshalXML(d *xml.Decoder, start xml.StartElement) error { | ||
132 | type subTag Tag // to avoid recursively calling UnmarshalXML() | ||
133 | var st subTag | ||
134 | if err := d.DecodeElement(&st, &start); err != nil { | ||
135 | return err | ||
136 | } | ||
137 | |||
138 | if err := Tag(st).Validate(); err != nil { | ||
139 | return err | ||
140 | } | ||
141 | |||
142 | *tag = Tag(st) | ||
143 | return nil | ||
144 | } | ||
145 | |||
146 | // tagSet represents list of unique tags. | ||
147 | type tagSet struct { | ||
148 | tagMap map[string]string | ||
149 | isObject bool | ||
150 | } | ||
151 | |||
152 | func (tags tagSet) String() string { | ||
153 | if len(tags.tagMap) == 0 { | ||
154 | return "" | ||
155 | } | ||
156 | var buf strings.Builder | ||
157 | keys := make([]string, 0, len(tags.tagMap)) | ||
158 | for k := range tags.tagMap { | ||
159 | keys = append(keys, k) | ||
160 | } | ||
161 | sort.Strings(keys) | ||
162 | for _, k := range keys { | ||
163 | keyEscaped := url.QueryEscape(k) | ||
164 | valueEscaped := url.QueryEscape(tags.tagMap[k]) | ||
165 | if buf.Len() > 0 { | ||
166 | buf.WriteByte('&') | ||
167 | } | ||
168 | buf.WriteString(keyEscaped) | ||
169 | buf.WriteByte('=') | ||
170 | buf.WriteString(valueEscaped) | ||
171 | } | ||
172 | return buf.String() | ||
173 | } | ||
174 | |||
175 | func (tags *tagSet) remove(key string) { | ||
176 | delete(tags.tagMap, key) | ||
177 | } | ||
178 | |||
179 | func (tags *tagSet) set(key, value string, failOnExist bool) error { | ||
180 | if failOnExist { | ||
181 | if _, found := tags.tagMap[key]; found { | ||
182 | return errDuplicateTagKey | ||
183 | } | ||
184 | } | ||
185 | |||
186 | if err := checkKey(key); err != nil { | ||
187 | return err | ||
188 | } | ||
189 | |||
190 | if err := checkValue(value); err != nil { | ||
191 | return err | ||
192 | } | ||
193 | |||
194 | if tags.isObject { | ||
195 | if len(tags.tagMap) == maxObjectTagCount { | ||
196 | return errTooManyObjectTags | ||
197 | } | ||
198 | } else if len(tags.tagMap) == maxTagCount { | ||
199 | return errTooManyTags | ||
200 | } | ||
201 | |||
202 | tags.tagMap[key] = value | ||
203 | return nil | ||
204 | } | ||
205 | |||
206 | func (tags tagSet) count() int { | ||
207 | return len(tags.tagMap) | ||
208 | } | ||
209 | |||
210 | func (tags tagSet) toMap() map[string]string { | ||
211 | m := make(map[string]string, len(tags.tagMap)) | ||
212 | for key, value := range tags.tagMap { | ||
213 | m[key] = value | ||
214 | } | ||
215 | return m | ||
216 | } | ||
217 | |||
218 | // MarshalXML encodes to XML data. | ||
219 | func (tags tagSet) MarshalXML(e *xml.Encoder, start xml.StartElement) error { | ||
220 | tagList := struct { | ||
221 | Tags []Tag `xml:"Tag"` | ||
222 | }{} | ||
223 | |||
224 | tagList.Tags = make([]Tag, 0, len(tags.tagMap)) | ||
225 | for key, value := range tags.tagMap { | ||
226 | tagList.Tags = append(tagList.Tags, Tag{key, value}) | ||
227 | } | ||
228 | |||
229 | return e.EncodeElement(tagList, start) | ||
230 | } | ||
231 | |||
232 | // UnmarshalXML decodes XML data to tag list. | ||
233 | func (tags *tagSet) UnmarshalXML(d *xml.Decoder, start xml.StartElement) error { | ||
234 | tagList := struct { | ||
235 | Tags []Tag `xml:"Tag"` | ||
236 | }{} | ||
237 | |||
238 | if err := d.DecodeElement(&tagList, &start); err != nil { | ||
239 | return err | ||
240 | } | ||
241 | |||
242 | if tags.isObject { | ||
243 | if len(tagList.Tags) > maxObjectTagCount { | ||
244 | return errTooManyObjectTags | ||
245 | } | ||
246 | } else if len(tagList.Tags) > maxTagCount { | ||
247 | return errTooManyTags | ||
248 | } | ||
249 | |||
250 | m := make(map[string]string, len(tagList.Tags)) | ||
251 | for _, tag := range tagList.Tags { | ||
252 | if _, found := m[tag.Key]; found { | ||
253 | return errDuplicateTagKey | ||
254 | } | ||
255 | |||
256 | m[tag.Key] = tag.Value | ||
257 | } | ||
258 | |||
259 | tags.tagMap = m | ||
260 | return nil | ||
261 | } | ||
262 | |||
263 | type tagging struct { | ||
264 | XMLName xml.Name `xml:"Tagging"` | ||
265 | TagSet *tagSet `xml:"TagSet"` | ||
266 | } | ||
267 | |||
268 | // Tags is list of tags of XML request/response as per | ||
269 | // https://docs.aws.amazon.com/AmazonS3/latest/API/API_GetBucketTagging.html#API_GetBucketTagging_RequestBody | ||
270 | type Tags tagging | ||
271 | |||
272 | func (tags Tags) String() string { | ||
273 | return tags.TagSet.String() | ||
274 | } | ||
275 | |||
276 | // Remove removes a tag by its key. | ||
277 | func (tags *Tags) Remove(key string) { | ||
278 | tags.TagSet.remove(key) | ||
279 | } | ||
280 | |||
281 | // Set sets new tag. | ||
282 | func (tags *Tags) Set(key, value string) error { | ||
283 | return tags.TagSet.set(key, value, false) | ||
284 | } | ||
285 | |||
286 | // Count - return number of tags accounted for | ||
287 | func (tags Tags) Count() int { | ||
288 | return tags.TagSet.count() | ||
289 | } | ||
290 | |||
291 | // ToMap returns copy of tags. | ||
292 | func (tags Tags) ToMap() map[string]string { | ||
293 | return tags.TagSet.toMap() | ||
294 | } | ||
295 | |||
296 | // MapToObjectTags converts an input map of key and value into | ||
297 | // *Tags data structure with validation. | ||
298 | func MapToObjectTags(tagMap map[string]string) (*Tags, error) { | ||
299 | return NewTags(tagMap, true) | ||
300 | } | ||
301 | |||
302 | // MapToBucketTags converts an input map of key and value into | ||
303 | // *Tags data structure with validation. | ||
304 | func MapToBucketTags(tagMap map[string]string) (*Tags, error) { | ||
305 | return NewTags(tagMap, false) | ||
306 | } | ||
307 | |||
308 | // NewTags creates Tags from tagMap, If isObject is set, it validates for object tags. | ||
309 | func NewTags(tagMap map[string]string, isObject bool) (*Tags, error) { | ||
310 | tagging := &Tags{ | ||
311 | TagSet: &tagSet{ | ||
312 | tagMap: make(map[string]string), | ||
313 | isObject: isObject, | ||
314 | }, | ||
315 | } | ||
316 | |||
317 | for key, value := range tagMap { | ||
318 | if err := tagging.TagSet.set(key, value, true); err != nil { | ||
319 | return nil, err | ||
320 | } | ||
321 | } | ||
322 | |||
323 | return tagging, nil | ||
324 | } | ||
325 | |||
326 | func unmarshalXML(reader io.Reader, isObject bool) (*Tags, error) { | ||
327 | tagging := &Tags{ | ||
328 | TagSet: &tagSet{ | ||
329 | tagMap: make(map[string]string), | ||
330 | isObject: isObject, | ||
331 | }, | ||
332 | } | ||
333 | |||
334 | if err := xml.NewDecoder(reader).Decode(tagging); err != nil { | ||
335 | return nil, err | ||
336 | } | ||
337 | |||
338 | return tagging, nil | ||
339 | } | ||
340 | |||
341 | // ParseBucketXML decodes XML data of tags in reader specified in | ||
342 | // https://docs.aws.amazon.com/AmazonS3/latest/API/API_PutBucketTagging.html#API_PutBucketTagging_RequestSyntax. | ||
343 | func ParseBucketXML(reader io.Reader) (*Tags, error) { | ||
344 | return unmarshalXML(reader, false) | ||
345 | } | ||
346 | |||
347 | // ParseObjectXML decodes XML data of tags in reader specified in | ||
348 | // https://docs.aws.amazon.com/AmazonS3/latest/API/API_PutObjectTagging.html#API_PutObjectTagging_RequestSyntax | ||
349 | func ParseObjectXML(reader io.Reader) (*Tags, error) { | ||
350 | return unmarshalXML(reader, true) | ||
351 | } | ||
352 | |||
353 | // stringsCut slices s around the first instance of sep, | ||
354 | // returning the text before and after sep. | ||
355 | // The found result reports whether sep appears in s. | ||
356 | // If sep does not appear in s, cut returns s, "", false. | ||
357 | func stringsCut(s, sep string) (before, after string, found bool) { | ||
358 | if i := strings.Index(s, sep); i >= 0 { | ||
359 | return s[:i], s[i+len(sep):], true | ||
360 | } | ||
361 | return s, "", false | ||
362 | } | ||
363 | |||
364 | func (tags *tagSet) parseTags(tgs string) (err error) { | ||
365 | for tgs != "" { | ||
366 | var key string | ||
367 | key, tgs, _ = stringsCut(tgs, "&") | ||
368 | if key == "" { | ||
369 | continue | ||
370 | } | ||
371 | key, value, _ := stringsCut(key, "=") | ||
372 | key, err1 := url.QueryUnescape(key) | ||
373 | if err1 != nil { | ||
374 | if err == nil { | ||
375 | err = err1 | ||
376 | } | ||
377 | continue | ||
378 | } | ||
379 | value, err1 = url.QueryUnescape(value) | ||
380 | if err1 != nil { | ||
381 | if err == nil { | ||
382 | err = err1 | ||
383 | } | ||
384 | continue | ||
385 | } | ||
386 | if err = tags.set(key, value, true); err != nil { | ||
387 | return err | ||
388 | } | ||
389 | } | ||
390 | return err | ||
391 | } | ||
392 | |||
393 | // Parse decodes HTTP query formatted string into tags which is limited by isObject. | ||
394 | // A query formatted string is like "key1=value1&key2=value2". | ||
395 | func Parse(s string, isObject bool) (*Tags, error) { | ||
396 | tagging := &Tags{ | ||
397 | TagSet: &tagSet{ | ||
398 | tagMap: make(map[string]string), | ||
399 | isObject: isObject, | ||
400 | }, | ||
401 | } | ||
402 | |||
403 | if err := tagging.TagSet.parseTags(s); err != nil { | ||
404 | return nil, err | ||
405 | } | ||
406 | |||
407 | return tagging, nil | ||
408 | } | ||
409 | |||
410 | // ParseObjectTags decodes HTTP query formatted string into tags. A query formatted string is like "key1=value1&key2=value2". | ||
411 | func ParseObjectTags(s string) (*Tags, error) { | ||
412 | return Parse(s, true) | ||
413 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/post-policy.go b/vendor/github.com/minio/minio-go/v7/post-policy.go deleted file mode 100644 index 3f4881e..0000000 --- a/vendor/github.com/minio/minio-go/v7/post-policy.go +++ /dev/null | |||
@@ -1,349 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2015-2023 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package minio | ||
19 | |||
20 | import ( | ||
21 | "encoding/base64" | ||
22 | "fmt" | ||
23 | "net/http" | ||
24 | "strings" | ||
25 | "time" | ||
26 | |||
27 | "github.com/minio/minio-go/v7/pkg/encrypt" | ||
28 | ) | ||
29 | |||
30 | // expirationDateFormat date format for expiration key in json policy. | ||
31 | const expirationDateFormat = "2006-01-02T15:04:05.000Z" | ||
32 | |||
33 | // policyCondition explanation: | ||
34 | // http://docs.aws.amazon.com/AmazonS3/latest/API/sigv4-HTTPPOSTConstructPolicy.html | ||
35 | // | ||
36 | // Example: | ||
37 | // | ||
38 | // policyCondition { | ||
39 | // matchType: "$eq", | ||
40 | // key: "$Content-Type", | ||
41 | // value: "image/png", | ||
42 | // } | ||
43 | type policyCondition struct { | ||
44 | matchType string | ||
45 | condition string | ||
46 | value string | ||
47 | } | ||
48 | |||
49 | // PostPolicy - Provides strict static type conversion and validation | ||
50 | // for Amazon S3's POST policy JSON string. | ||
51 | type PostPolicy struct { | ||
52 | // Expiration date and time of the POST policy. | ||
53 | expiration time.Time | ||
54 | // Collection of different policy conditions. | ||
55 | conditions []policyCondition | ||
56 | // ContentLengthRange minimum and maximum allowable size for the | ||
57 | // uploaded content. | ||
58 | contentLengthRange struct { | ||
59 | min int64 | ||
60 | max int64 | ||
61 | } | ||
62 | |||
63 | // Post form data. | ||
64 | formData map[string]string | ||
65 | } | ||
66 | |||
67 | // NewPostPolicy - Instantiate new post policy. | ||
68 | func NewPostPolicy() *PostPolicy { | ||
69 | p := &PostPolicy{} | ||
70 | p.conditions = make([]policyCondition, 0) | ||
71 | p.formData = make(map[string]string) | ||
72 | return p | ||
73 | } | ||
74 | |||
75 | // SetExpires - Sets expiration time for the new policy. | ||
76 | func (p *PostPolicy) SetExpires(t time.Time) error { | ||
77 | if t.IsZero() { | ||
78 | return errInvalidArgument("No expiry time set.") | ||
79 | } | ||
80 | p.expiration = t | ||
81 | return nil | ||
82 | } | ||
83 | |||
84 | // SetKey - Sets an object name for the policy based upload. | ||
85 | func (p *PostPolicy) SetKey(key string) error { | ||
86 | if strings.TrimSpace(key) == "" || key == "" { | ||
87 | return errInvalidArgument("Object name is empty.") | ||
88 | } | ||
89 | policyCond := policyCondition{ | ||
90 | matchType: "eq", | ||
91 | condition: "$key", | ||
92 | value: key, | ||
93 | } | ||
94 | if err := p.addNewPolicy(policyCond); err != nil { | ||
95 | return err | ||
96 | } | ||
97 | p.formData["key"] = key | ||
98 | return nil | ||
99 | } | ||
100 | |||
101 | // SetKeyStartsWith - Sets an object name that an policy based upload | ||
102 | // can start with. | ||
103 | // Can use an empty value ("") to allow any key. | ||
104 | func (p *PostPolicy) SetKeyStartsWith(keyStartsWith string) error { | ||
105 | policyCond := policyCondition{ | ||
106 | matchType: "starts-with", | ||
107 | condition: "$key", | ||
108 | value: keyStartsWith, | ||
109 | } | ||
110 | if err := p.addNewPolicy(policyCond); err != nil { | ||
111 | return err | ||
112 | } | ||
113 | p.formData["key"] = keyStartsWith | ||
114 | return nil | ||
115 | } | ||
116 | |||
117 | // SetBucket - Sets bucket at which objects will be uploaded to. | ||
118 | func (p *PostPolicy) SetBucket(bucketName string) error { | ||
119 | if strings.TrimSpace(bucketName) == "" || bucketName == "" { | ||
120 | return errInvalidArgument("Bucket name is empty.") | ||
121 | } | ||
122 | policyCond := policyCondition{ | ||
123 | matchType: "eq", | ||
124 | condition: "$bucket", | ||
125 | value: bucketName, | ||
126 | } | ||
127 | if err := p.addNewPolicy(policyCond); err != nil { | ||
128 | return err | ||
129 | } | ||
130 | p.formData["bucket"] = bucketName | ||
131 | return nil | ||
132 | } | ||
133 | |||
134 | // SetCondition - Sets condition for credentials, date and algorithm | ||
135 | func (p *PostPolicy) SetCondition(matchType, condition, value string) error { | ||
136 | if strings.TrimSpace(value) == "" || value == "" { | ||
137 | return errInvalidArgument("No value specified for condition") | ||
138 | } | ||
139 | |||
140 | policyCond := policyCondition{ | ||
141 | matchType: matchType, | ||
142 | condition: "$" + condition, | ||
143 | value: value, | ||
144 | } | ||
145 | if condition == "X-Amz-Credential" || condition == "X-Amz-Date" || condition == "X-Amz-Algorithm" { | ||
146 | if err := p.addNewPolicy(policyCond); err != nil { | ||
147 | return err | ||
148 | } | ||
149 | p.formData[condition] = value | ||
150 | return nil | ||
151 | } | ||
152 | return errInvalidArgument("Invalid condition in policy") | ||
153 | } | ||
154 | |||
155 | // SetContentType - Sets content-type of the object for this policy | ||
156 | // based upload. | ||
157 | func (p *PostPolicy) SetContentType(contentType string) error { | ||
158 | if strings.TrimSpace(contentType) == "" || contentType == "" { | ||
159 | return errInvalidArgument("No content type specified.") | ||
160 | } | ||
161 | policyCond := policyCondition{ | ||
162 | matchType: "eq", | ||
163 | condition: "$Content-Type", | ||
164 | value: contentType, | ||
165 | } | ||
166 | if err := p.addNewPolicy(policyCond); err != nil { | ||
167 | return err | ||
168 | } | ||
169 | p.formData["Content-Type"] = contentType | ||
170 | return nil | ||
171 | } | ||
172 | |||
173 | // SetContentTypeStartsWith - Sets what content-type of the object for this policy | ||
174 | // based upload can start with. | ||
175 | // Can use an empty value ("") to allow any content-type. | ||
176 | func (p *PostPolicy) SetContentTypeStartsWith(contentTypeStartsWith string) error { | ||
177 | policyCond := policyCondition{ | ||
178 | matchType: "starts-with", | ||
179 | condition: "$Content-Type", | ||
180 | value: contentTypeStartsWith, | ||
181 | } | ||
182 | if err := p.addNewPolicy(policyCond); err != nil { | ||
183 | return err | ||
184 | } | ||
185 | p.formData["Content-Type"] = contentTypeStartsWith | ||
186 | return nil | ||
187 | } | ||
188 | |||
189 | // SetContentLengthRange - Set new min and max content length | ||
190 | // condition for all incoming uploads. | ||
191 | func (p *PostPolicy) SetContentLengthRange(min, max int64) error { | ||
192 | if min > max { | ||
193 | return errInvalidArgument("Minimum limit is larger than maximum limit.") | ||
194 | } | ||
195 | if min < 0 { | ||
196 | return errInvalidArgument("Minimum limit cannot be negative.") | ||
197 | } | ||
198 | if max <= 0 { | ||
199 | return errInvalidArgument("Maximum limit cannot be non-positive.") | ||
200 | } | ||
201 | p.contentLengthRange.min = min | ||
202 | p.contentLengthRange.max = max | ||
203 | return nil | ||
204 | } | ||
205 | |||
206 | // SetSuccessActionRedirect - Sets the redirect success url of the object for this policy | ||
207 | // based upload. | ||
208 | func (p *PostPolicy) SetSuccessActionRedirect(redirect string) error { | ||
209 | if strings.TrimSpace(redirect) == "" || redirect == "" { | ||
210 | return errInvalidArgument("Redirect is empty") | ||
211 | } | ||
212 | policyCond := policyCondition{ | ||
213 | matchType: "eq", | ||
214 | condition: "$success_action_redirect", | ||
215 | value: redirect, | ||
216 | } | ||
217 | if err := p.addNewPolicy(policyCond); err != nil { | ||
218 | return err | ||
219 | } | ||
220 | p.formData["success_action_redirect"] = redirect | ||
221 | return nil | ||
222 | } | ||
223 | |||
224 | // SetSuccessStatusAction - Sets the status success code of the object for this policy | ||
225 | // based upload. | ||
226 | func (p *PostPolicy) SetSuccessStatusAction(status string) error { | ||
227 | if strings.TrimSpace(status) == "" || status == "" { | ||
228 | return errInvalidArgument("Status is empty") | ||
229 | } | ||
230 | policyCond := policyCondition{ | ||
231 | matchType: "eq", | ||
232 | condition: "$success_action_status", | ||
233 | value: status, | ||
234 | } | ||
235 | if err := p.addNewPolicy(policyCond); err != nil { | ||
236 | return err | ||
237 | } | ||
238 | p.formData["success_action_status"] = status | ||
239 | return nil | ||
240 | } | ||
241 | |||
242 | // SetUserMetadata - Set user metadata as a key/value couple. | ||
243 | // Can be retrieved through a HEAD request or an event. | ||
244 | func (p *PostPolicy) SetUserMetadata(key, value string) error { | ||
245 | if strings.TrimSpace(key) == "" || key == "" { | ||
246 | return errInvalidArgument("Key is empty") | ||
247 | } | ||
248 | if strings.TrimSpace(value) == "" || value == "" { | ||
249 | return errInvalidArgument("Value is empty") | ||
250 | } | ||
251 | headerName := fmt.Sprintf("x-amz-meta-%s", key) | ||
252 | policyCond := policyCondition{ | ||
253 | matchType: "eq", | ||
254 | condition: fmt.Sprintf("$%s", headerName), | ||
255 | value: value, | ||
256 | } | ||
257 | if err := p.addNewPolicy(policyCond); err != nil { | ||
258 | return err | ||
259 | } | ||
260 | p.formData[headerName] = value | ||
261 | return nil | ||
262 | } | ||
263 | |||
264 | // SetChecksum sets the checksum of the request. | ||
265 | func (p *PostPolicy) SetChecksum(c Checksum) { | ||
266 | if c.IsSet() { | ||
267 | p.formData[amzChecksumAlgo] = c.Type.String() | ||
268 | p.formData[c.Type.Key()] = c.Encoded() | ||
269 | } | ||
270 | } | ||
271 | |||
272 | // SetEncryption - sets encryption headers for POST API | ||
273 | func (p *PostPolicy) SetEncryption(sse encrypt.ServerSide) { | ||
274 | if sse == nil { | ||
275 | return | ||
276 | } | ||
277 | h := http.Header{} | ||
278 | sse.Marshal(h) | ||
279 | for k, v := range h { | ||
280 | p.formData[k] = v[0] | ||
281 | } | ||
282 | } | ||
283 | |||
284 | // SetUserData - Set user data as a key/value couple. | ||
285 | // Can be retrieved through a HEAD request or an event. | ||
286 | func (p *PostPolicy) SetUserData(key, value string) error { | ||
287 | if key == "" { | ||
288 | return errInvalidArgument("Key is empty") | ||
289 | } | ||
290 | if value == "" { | ||
291 | return errInvalidArgument("Value is empty") | ||
292 | } | ||
293 | headerName := fmt.Sprintf("x-amz-%s", key) | ||
294 | policyCond := policyCondition{ | ||
295 | matchType: "eq", | ||
296 | condition: fmt.Sprintf("$%s", headerName), | ||
297 | value: value, | ||
298 | } | ||
299 | if err := p.addNewPolicy(policyCond); err != nil { | ||
300 | return err | ||
301 | } | ||
302 | p.formData[headerName] = value | ||
303 | return nil | ||
304 | } | ||
305 | |||
306 | // addNewPolicy - internal helper to validate adding new policies. | ||
307 | // Can use starts-with with an empty value ("") to allow any content within a form field. | ||
308 | func (p *PostPolicy) addNewPolicy(policyCond policyCondition) error { | ||
309 | if policyCond.matchType == "" || policyCond.condition == "" { | ||
310 | return errInvalidArgument("Policy fields are empty.") | ||
311 | } | ||
312 | if policyCond.matchType != "starts-with" && policyCond.value == "" { | ||
313 | return errInvalidArgument("Policy value is empty.") | ||
314 | } | ||
315 | p.conditions = append(p.conditions, policyCond) | ||
316 | return nil | ||
317 | } | ||
318 | |||
319 | // String function for printing policy in json formatted string. | ||
320 | func (p PostPolicy) String() string { | ||
321 | return string(p.marshalJSON()) | ||
322 | } | ||
323 | |||
324 | // marshalJSON - Provides Marshaled JSON in bytes. | ||
325 | func (p PostPolicy) marshalJSON() []byte { | ||
326 | expirationStr := `"expiration":"` + p.expiration.Format(expirationDateFormat) + `"` | ||
327 | var conditionsStr string | ||
328 | conditions := []string{} | ||
329 | for _, po := range p.conditions { | ||
330 | conditions = append(conditions, fmt.Sprintf("[\"%s\",\"%s\",\"%s\"]", po.matchType, po.condition, po.value)) | ||
331 | } | ||
332 | if p.contentLengthRange.min != 0 || p.contentLengthRange.max != 0 { | ||
333 | conditions = append(conditions, fmt.Sprintf("[\"content-length-range\", %d, %d]", | ||
334 | p.contentLengthRange.min, p.contentLengthRange.max)) | ||
335 | } | ||
336 | if len(conditions) > 0 { | ||
337 | conditionsStr = `"conditions":[` + strings.Join(conditions, ",") + "]" | ||
338 | } | ||
339 | retStr := "{" | ||
340 | retStr = retStr + expirationStr + "," | ||
341 | retStr += conditionsStr | ||
342 | retStr += "}" | ||
343 | return []byte(retStr) | ||
344 | } | ||
345 | |||
346 | // base64 - Produces base64 of PostPolicy's Marshaled json. | ||
347 | func (p PostPolicy) base64() string { | ||
348 | return base64.StdEncoding.EncodeToString(p.marshalJSON()) | ||
349 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/retry-continous.go b/vendor/github.com/minio/minio-go/v7/retry-continous.go deleted file mode 100644 index bfeea95..0000000 --- a/vendor/github.com/minio/minio-go/v7/retry-continous.go +++ /dev/null | |||
@@ -1,69 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2015-2017 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package minio | ||
19 | |||
20 | import "time" | ||
21 | |||
22 | // newRetryTimerContinous creates a timer with exponentially increasing delays forever. | ||
23 | func (c *Client) newRetryTimerContinous(unit, cap time.Duration, jitter float64, doneCh chan struct{}) <-chan int { | ||
24 | attemptCh := make(chan int) | ||
25 | |||
26 | // normalize jitter to the range [0, 1.0] | ||
27 | if jitter < NoJitter { | ||
28 | jitter = NoJitter | ||
29 | } | ||
30 | if jitter > MaxJitter { | ||
31 | jitter = MaxJitter | ||
32 | } | ||
33 | |||
34 | // computes the exponential backoff duration according to | ||
35 | // https://www.awsarchitectureblog.com/2015/03/backoff.html | ||
36 | exponentialBackoffWait := func(attempt int) time.Duration { | ||
37 | // 1<<uint(attempt) below could overflow, so limit the value of attempt | ||
38 | maxAttempt := 30 | ||
39 | if attempt > maxAttempt { | ||
40 | attempt = maxAttempt | ||
41 | } | ||
42 | // sleep = random_between(0, min(cap, base * 2 ** attempt)) | ||
43 | sleep := unit * time.Duration(1<<uint(attempt)) | ||
44 | if sleep > cap { | ||
45 | sleep = cap | ||
46 | } | ||
47 | if jitter != NoJitter { | ||
48 | sleep -= time.Duration(c.random.Float64() * float64(sleep) * jitter) | ||
49 | } | ||
50 | return sleep | ||
51 | } | ||
52 | |||
53 | go func() { | ||
54 | defer close(attemptCh) | ||
55 | var nextBackoff int | ||
56 | for { | ||
57 | select { | ||
58 | // Attempts starts. | ||
59 | case attemptCh <- nextBackoff: | ||
60 | nextBackoff++ | ||
61 | case <-doneCh: | ||
62 | // Stop the routine. | ||
63 | return | ||
64 | } | ||
65 | time.Sleep(exponentialBackoffWait(nextBackoff)) | ||
66 | } | ||
67 | }() | ||
68 | return attemptCh | ||
69 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/retry.go b/vendor/github.com/minio/minio-go/v7/retry.go deleted file mode 100644 index 1c6105e..0000000 --- a/vendor/github.com/minio/minio-go/v7/retry.go +++ /dev/null | |||
@@ -1,148 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2015-2017 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package minio | ||
19 | |||
20 | import ( | ||
21 | "context" | ||
22 | "crypto/x509" | ||
23 | "errors" | ||
24 | "net/http" | ||
25 | "net/url" | ||
26 | "time" | ||
27 | ) | ||
28 | |||
29 | // MaxRetry is the maximum number of retries before stopping. | ||
30 | var MaxRetry = 10 | ||
31 | |||
32 | // MaxJitter will randomize over the full exponential backoff time | ||
33 | const MaxJitter = 1.0 | ||
34 | |||
35 | // NoJitter disables the use of jitter for randomizing the exponential backoff time | ||
36 | const NoJitter = 0.0 | ||
37 | |||
38 | // DefaultRetryUnit - default unit multiplicative per retry. | ||
39 | // defaults to 200 * time.Millisecond | ||
40 | var DefaultRetryUnit = 200 * time.Millisecond | ||
41 | |||
42 | // DefaultRetryCap - Each retry attempt never waits no longer than | ||
43 | // this maximum time duration. | ||
44 | var DefaultRetryCap = time.Second | ||
45 | |||
46 | // newRetryTimer creates a timer with exponentially increasing | ||
47 | // delays until the maximum retry attempts are reached. | ||
48 | func (c *Client) newRetryTimer(ctx context.Context, maxRetry int, unit, cap time.Duration, jitter float64) <-chan int { | ||
49 | attemptCh := make(chan int) | ||
50 | |||
51 | // computes the exponential backoff duration according to | ||
52 | // https://www.awsarchitectureblog.com/2015/03/backoff.html | ||
53 | exponentialBackoffWait := func(attempt int) time.Duration { | ||
54 | // normalize jitter to the range [0, 1.0] | ||
55 | if jitter < NoJitter { | ||
56 | jitter = NoJitter | ||
57 | } | ||
58 | if jitter > MaxJitter { | ||
59 | jitter = MaxJitter | ||
60 | } | ||
61 | |||
62 | // sleep = random_between(0, min(cap, base * 2 ** attempt)) | ||
63 | sleep := unit * time.Duration(1<<uint(attempt)) | ||
64 | if sleep > cap { | ||
65 | sleep = cap | ||
66 | } | ||
67 | if jitter != NoJitter { | ||
68 | sleep -= time.Duration(c.random.Float64() * float64(sleep) * jitter) | ||
69 | } | ||
70 | return sleep | ||
71 | } | ||
72 | |||
73 | go func() { | ||
74 | defer close(attemptCh) | ||
75 | for i := 0; i < maxRetry; i++ { | ||
76 | select { | ||
77 | case attemptCh <- i + 1: | ||
78 | case <-ctx.Done(): | ||
79 | return | ||
80 | } | ||
81 | |||
82 | select { | ||
83 | case <-time.After(exponentialBackoffWait(i)): | ||
84 | case <-ctx.Done(): | ||
85 | return | ||
86 | } | ||
87 | } | ||
88 | }() | ||
89 | return attemptCh | ||
90 | } | ||
91 | |||
92 | // List of AWS S3 error codes which are retryable. | ||
93 | var retryableS3Codes = map[string]struct{}{ | ||
94 | "RequestError": {}, | ||
95 | "RequestTimeout": {}, | ||
96 | "Throttling": {}, | ||
97 | "ThrottlingException": {}, | ||
98 | "RequestLimitExceeded": {}, | ||
99 | "RequestThrottled": {}, | ||
100 | "InternalError": {}, | ||
101 | "ExpiredToken": {}, | ||
102 | "ExpiredTokenException": {}, | ||
103 | "SlowDown": {}, | ||
104 | // Add more AWS S3 codes here. | ||
105 | } | ||
106 | |||
107 | // isS3CodeRetryable - is s3 error code retryable. | ||
108 | func isS3CodeRetryable(s3Code string) (ok bool) { | ||
109 | _, ok = retryableS3Codes[s3Code] | ||
110 | return ok | ||
111 | } | ||
112 | |||
113 | // List of HTTP status codes which are retryable. | ||
114 | var retryableHTTPStatusCodes = map[int]struct{}{ | ||
115 | 429: {}, // http.StatusTooManyRequests is not part of the Go 1.5 library, yet | ||
116 | 499: {}, // client closed request, retry. A non-standard status code introduced by nginx. | ||
117 | http.StatusInternalServerError: {}, | ||
118 | http.StatusBadGateway: {}, | ||
119 | http.StatusServiceUnavailable: {}, | ||
120 | http.StatusGatewayTimeout: {}, | ||
121 | // Add more HTTP status codes here. | ||
122 | } | ||
123 | |||
124 | // isHTTPStatusRetryable - is HTTP error code retryable. | ||
125 | func isHTTPStatusRetryable(httpStatusCode int) (ok bool) { | ||
126 | _, ok = retryableHTTPStatusCodes[httpStatusCode] | ||
127 | return ok | ||
128 | } | ||
129 | |||
130 | // For now, all http Do() requests are retriable except some well defined errors | ||
131 | func isRequestErrorRetryable(err error) bool { | ||
132 | if errors.Is(err, context.Canceled) || errors.Is(err, context.DeadlineExceeded) { | ||
133 | return false | ||
134 | } | ||
135 | if ue, ok := err.(*url.Error); ok { | ||
136 | e := ue.Unwrap() | ||
137 | switch e.(type) { | ||
138 | // x509: certificate signed by unknown authority | ||
139 | case x509.UnknownAuthorityError: | ||
140 | return false | ||
141 | } | ||
142 | switch e.Error() { | ||
143 | case "http: server gave HTTP response to HTTPS client": | ||
144 | return false | ||
145 | } | ||
146 | } | ||
147 | return true | ||
148 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/s3-endpoints.go b/vendor/github.com/minio/minio-go/v7/s3-endpoints.go deleted file mode 100644 index b1de7b6..0000000 --- a/vendor/github.com/minio/minio-go/v7/s3-endpoints.go +++ /dev/null | |||
@@ -1,64 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2015-2017 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package minio | ||
19 | |||
20 | // awsS3EndpointMap Amazon S3 endpoint map. | ||
21 | var awsS3EndpointMap = map[string]string{ | ||
22 | "us-east-1": "s3.dualstack.us-east-1.amazonaws.com", | ||
23 | "us-east-2": "s3.dualstack.us-east-2.amazonaws.com", | ||
24 | "us-west-2": "s3.dualstack.us-west-2.amazonaws.com", | ||
25 | "us-west-1": "s3.dualstack.us-west-1.amazonaws.com", | ||
26 | "ca-central-1": "s3.dualstack.ca-central-1.amazonaws.com", | ||
27 | "eu-west-1": "s3.dualstack.eu-west-1.amazonaws.com", | ||
28 | "eu-west-2": "s3.dualstack.eu-west-2.amazonaws.com", | ||
29 | "eu-west-3": "s3.dualstack.eu-west-3.amazonaws.com", | ||
30 | "eu-central-1": "s3.dualstack.eu-central-1.amazonaws.com", | ||
31 | "eu-central-2": "s3.dualstack.eu-central-2.amazonaws.com", | ||
32 | "eu-north-1": "s3.dualstack.eu-north-1.amazonaws.com", | ||
33 | "eu-south-1": "s3.dualstack.eu-south-1.amazonaws.com", | ||
34 | "eu-south-2": "s3.dualstack.eu-south-2.amazonaws.com", | ||
35 | "ap-east-1": "s3.dualstack.ap-east-1.amazonaws.com", | ||
36 | "ap-south-1": "s3.dualstack.ap-south-1.amazonaws.com", | ||
37 | "ap-south-2": "s3.dualstack.ap-south-2.amazonaws.com", | ||
38 | "ap-southeast-1": "s3.dualstack.ap-southeast-1.amazonaws.com", | ||
39 | "ap-southeast-2": "s3.dualstack.ap-southeast-2.amazonaws.com", | ||
40 | "ap-northeast-1": "s3.dualstack.ap-northeast-1.amazonaws.com", | ||
41 | "ap-northeast-2": "s3.dualstack.ap-northeast-2.amazonaws.com", | ||
42 | "ap-northeast-3": "s3.dualstack.ap-northeast-3.amazonaws.com", | ||
43 | "af-south-1": "s3.dualstack.af-south-1.amazonaws.com", | ||
44 | "me-central-1": "s3.dualstack.me-central-1.amazonaws.com", | ||
45 | "me-south-1": "s3.dualstack.me-south-1.amazonaws.com", | ||
46 | "sa-east-1": "s3.dualstack.sa-east-1.amazonaws.com", | ||
47 | "us-gov-west-1": "s3.dualstack.us-gov-west-1.amazonaws.com", | ||
48 | "us-gov-east-1": "s3.dualstack.us-gov-east-1.amazonaws.com", | ||
49 | "cn-north-1": "s3.dualstack.cn-north-1.amazonaws.com.cn", | ||
50 | "cn-northwest-1": "s3.dualstack.cn-northwest-1.amazonaws.com.cn", | ||
51 | "ap-southeast-3": "s3.dualstack.ap-southeast-3.amazonaws.com", | ||
52 | "ap-southeast-4": "s3.dualstack.ap-southeast-4.amazonaws.com", | ||
53 | "il-central-1": "s3.dualstack.il-central-1.amazonaws.com", | ||
54 | } | ||
55 | |||
56 | // getS3Endpoint get Amazon S3 endpoint based on the bucket location. | ||
57 | func getS3Endpoint(bucketLocation string) (s3Endpoint string) { | ||
58 | s3Endpoint, ok := awsS3EndpointMap[bucketLocation] | ||
59 | if !ok { | ||
60 | // Default to 's3.dualstack.us-east-1.amazonaws.com' endpoint. | ||
61 | s3Endpoint = "s3.dualstack.us-east-1.amazonaws.com" | ||
62 | } | ||
63 | return s3Endpoint | ||
64 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/s3-error.go b/vendor/github.com/minio/minio-go/v7/s3-error.go deleted file mode 100644 index f365157..0000000 --- a/vendor/github.com/minio/minio-go/v7/s3-error.go +++ /dev/null | |||
@@ -1,61 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2015-2017 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package minio | ||
19 | |||
20 | // Non exhaustive list of AWS S3 standard error responses - | ||
21 | // http://docs.aws.amazon.com/AmazonS3/latest/API/ErrorResponses.html | ||
22 | var s3ErrorResponseMap = map[string]string{ | ||
23 | "AccessDenied": "Access Denied.", | ||
24 | "BadDigest": "The Content-Md5 you specified did not match what we received.", | ||
25 | "EntityTooSmall": "Your proposed upload is smaller than the minimum allowed object size.", | ||
26 | "EntityTooLarge": "Your proposed upload exceeds the maximum allowed object size.", | ||
27 | "IncompleteBody": "You did not provide the number of bytes specified by the Content-Length HTTP header.", | ||
28 | "InternalError": "We encountered an internal error, please try again.", | ||
29 | "InvalidAccessKeyId": "The access key ID you provided does not exist in our records.", | ||
30 | "InvalidBucketName": "The specified bucket is not valid.", | ||
31 | "InvalidDigest": "The Content-Md5 you specified is not valid.", | ||
32 | "InvalidRange": "The requested range is not satisfiable", | ||
33 | "MalformedXML": "The XML you provided was not well-formed or did not validate against our published schema.", | ||
34 | "MissingContentLength": "You must provide the Content-Length HTTP header.", | ||
35 | "MissingContentMD5": "Missing required header for this request: Content-Md5.", | ||
36 | "MissingRequestBodyError": "Request body is empty.", | ||
37 | "NoSuchBucket": "The specified bucket does not exist.", | ||
38 | "NoSuchBucketPolicy": "The bucket policy does not exist", | ||
39 | "NoSuchKey": "The specified key does not exist.", | ||
40 | "NoSuchUpload": "The specified multipart upload does not exist. The upload ID may be invalid, or the upload may have been aborted or completed.", | ||
41 | "NotImplemented": "A header you provided implies functionality that is not implemented", | ||
42 | "PreconditionFailed": "At least one of the pre-conditions you specified did not hold", | ||
43 | "RequestTimeTooSkewed": "The difference between the request time and the server's time is too large.", | ||
44 | "SignatureDoesNotMatch": "The request signature we calculated does not match the signature you provided. Check your key and signing method.", | ||
45 | "MethodNotAllowed": "The specified method is not allowed against this resource.", | ||
46 | "InvalidPart": "One or more of the specified parts could not be found.", | ||
47 | "InvalidPartOrder": "The list of parts was not in ascending order. The parts list must be specified in order by part number.", | ||
48 | "InvalidObjectState": "The operation is not valid for the current state of the object.", | ||
49 | "AuthorizationHeaderMalformed": "The authorization header is malformed; the region is wrong.", | ||
50 | "MalformedPOSTRequest": "The body of your POST request is not well-formed multipart/form-data.", | ||
51 | "BucketNotEmpty": "The bucket you tried to delete is not empty", | ||
52 | "AllAccessDisabled": "All access to this bucket has been disabled.", | ||
53 | "MalformedPolicy": "Policy has invalid resource.", | ||
54 | "MissingFields": "Missing fields in request.", | ||
55 | "AuthorizationQueryParametersError": "Error parsing the X-Amz-Credential parameter; the Credential is mal-formed; expecting \"<YOUR-AKID>/YYYYMMDD/REGION/SERVICE/aws4_request\".", | ||
56 | "MalformedDate": "Invalid date format header, expected to be in ISO8601, RFC1123 or RFC1123Z time format.", | ||
57 | "BucketAlreadyOwnedByYou": "Your previous request to create the named bucket succeeded and you already own it.", | ||
58 | "InvalidDuration": "Duration provided in the request is invalid.", | ||
59 | "XAmzContentSHA256Mismatch": "The provided 'x-amz-content-sha256' header does not match what was computed.", | ||
60 | // Add new API errors here. | ||
61 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/transport.go b/vendor/github.com/minio/minio-go/v7/transport.go deleted file mode 100644 index 1bff664..0000000 --- a/vendor/github.com/minio/minio-go/v7/transport.go +++ /dev/null | |||
@@ -1,83 +0,0 @@ | |||
1 | //go:build go1.7 || go1.8 | ||
2 | // +build go1.7 go1.8 | ||
3 | |||
4 | /* | ||
5 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
6 | * Copyright 2017-2018 MinIO, Inc. | ||
7 | * | ||
8 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
9 | * you may not use this file except in compliance with the License. | ||
10 | * You may obtain a copy of the License at | ||
11 | * | ||
12 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
13 | * | ||
14 | * Unless required by applicable law or agreed to in writing, software | ||
15 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
16 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
17 | * See the License for the specific language governing permissions and | ||
18 | * limitations under the License. | ||
19 | */ | ||
20 | |||
21 | package minio | ||
22 | |||
23 | import ( | ||
24 | "crypto/tls" | ||
25 | "crypto/x509" | ||
26 | "net" | ||
27 | "net/http" | ||
28 | "os" | ||
29 | "time" | ||
30 | ) | ||
31 | |||
32 | // mustGetSystemCertPool - return system CAs or empty pool in case of error (or windows) | ||
33 | func mustGetSystemCertPool() *x509.CertPool { | ||
34 | pool, err := x509.SystemCertPool() | ||
35 | if err != nil { | ||
36 | return x509.NewCertPool() | ||
37 | } | ||
38 | return pool | ||
39 | } | ||
40 | |||
41 | // DefaultTransport - this default transport is similar to | ||
42 | // http.DefaultTransport but with additional param DisableCompression | ||
43 | // is set to true to avoid decompressing content with 'gzip' encoding. | ||
44 | var DefaultTransport = func(secure bool) (*http.Transport, error) { | ||
45 | tr := &http.Transport{ | ||
46 | Proxy: http.ProxyFromEnvironment, | ||
47 | DialContext: (&net.Dialer{ | ||
48 | Timeout: 30 * time.Second, | ||
49 | KeepAlive: 30 * time.Second, | ||
50 | }).DialContext, | ||
51 | MaxIdleConns: 256, | ||
52 | MaxIdleConnsPerHost: 16, | ||
53 | ResponseHeaderTimeout: time.Minute, | ||
54 | IdleConnTimeout: time.Minute, | ||
55 | TLSHandshakeTimeout: 10 * time.Second, | ||
56 | ExpectContinueTimeout: 10 * time.Second, | ||
57 | // Set this value so that the underlying transport round-tripper | ||
58 | // doesn't try to auto decode the body of objects with | ||
59 | // content-encoding set to `gzip`. | ||
60 | // | ||
61 | // Refer: | ||
62 | // https://golang.org/src/net/http/transport.go?h=roundTrip#L1843 | ||
63 | DisableCompression: true, | ||
64 | } | ||
65 | |||
66 | if secure { | ||
67 | tr.TLSClientConfig = &tls.Config{ | ||
68 | // Can't use SSLv3 because of POODLE and BEAST | ||
69 | // Can't use TLSv1.0 because of POODLE and BEAST using CBC cipher | ||
70 | // Can't use TLSv1.1 because of RC4 cipher usage | ||
71 | MinVersion: tls.VersionTLS12, | ||
72 | } | ||
73 | if f := os.Getenv("SSL_CERT_FILE"); f != "" { | ||
74 | rootCAs := mustGetSystemCertPool() | ||
75 | data, err := os.ReadFile(f) | ||
76 | if err == nil { | ||
77 | rootCAs.AppendCertsFromPEM(data) | ||
78 | } | ||
79 | tr.TLSClientConfig.RootCAs = rootCAs | ||
80 | } | ||
81 | } | ||
82 | return tr, nil | ||
83 | } | ||
diff --git a/vendor/github.com/minio/minio-go/v7/utils.go b/vendor/github.com/minio/minio-go/v7/utils.go deleted file mode 100644 index e39eba0..0000000 --- a/vendor/github.com/minio/minio-go/v7/utils.go +++ /dev/null | |||
@@ -1,693 +0,0 @@ | |||
1 | /* | ||
2 | * MinIO Go Library for Amazon S3 Compatible Cloud Storage | ||
3 | * Copyright 2015-2017 MinIO, Inc. | ||
4 | * | ||
5 | * Licensed under the Apache License, Version 2.0 (the "License"); | ||
6 | * you may not use this file except in compliance with the License. | ||
7 | * You may obtain a copy of the License at | ||
8 | * | ||
9 | * http://www.apache.org/licenses/LICENSE-2.0 | ||
10 | * | ||
11 | * Unless required by applicable law or agreed to in writing, software | ||
12 | * distributed under the License is distributed on an "AS IS" BASIS, | ||
13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
14 | * See the License for the specific language governing permissions and | ||
15 | * limitations under the License. | ||
16 | */ | ||
17 | |||
18 | package minio | ||
19 | |||
20 | import ( | ||
21 | "context" | ||
22 | "crypto/md5" | ||
23 | fipssha256 "crypto/sha256" | ||
24 | "encoding/base64" | ||
25 | "encoding/hex" | ||
26 | "encoding/xml" | ||
27 | "errors" | ||
28 | "fmt" | ||
29 | "hash" | ||
30 | "io" | ||
31 | "math/rand" | ||
32 | "net" | ||
33 | "net/http" | ||
34 | "net/url" | ||
35 | "regexp" | ||
36 | "strconv" | ||
37 | "strings" | ||
38 | "sync" | ||
39 | "time" | ||
40 | |||
41 | md5simd "github.com/minio/md5-simd" | ||
42 | "github.com/minio/minio-go/v7/pkg/encrypt" | ||
43 | "github.com/minio/minio-go/v7/pkg/s3utils" | ||
44 | "github.com/minio/sha256-simd" | ||
45 | ) | ||
46 | |||
47 | func trimEtag(etag string) string { | ||
48 | etag = strings.TrimPrefix(etag, "\"") | ||
49 | return strings.TrimSuffix(etag, "\"") | ||
50 | } | ||
51 | |||
52 | var expirationRegex = regexp.MustCompile(`expiry-date="(.*?)", rule-id="(.*?)"`) | ||
53 | |||
54 | func amzExpirationToExpiryDateRuleID(expiration string) (time.Time, string) { | ||
55 | if matches := expirationRegex.FindStringSubmatch(expiration); len(matches) == 3 { | ||
56 | expTime, err := parseRFC7231Time(matches[1]) | ||
57 | if err != nil { | ||
58 | return time.Time{}, "" | ||
59 | } | ||
60 | return expTime, matches[2] | ||
61 | } | ||
62 | return time.Time{}, "" | ||
63 | } | ||
64 | |||
65 | var restoreRegex = regexp.MustCompile(`ongoing-request="(.*?)"(, expiry-date="(.*?)")?`) | ||
66 | |||
67 | func amzRestoreToStruct(restore string) (ongoing bool, expTime time.Time, err error) { | ||
68 | matches := restoreRegex.FindStringSubmatch(restore) | ||
69 | if len(matches) != 4 { | ||
70 | return false, time.Time{}, errors.New("unexpected restore header") | ||
71 | } | ||
72 | ongoing, err = strconv.ParseBool(matches[1]) | ||
73 | if err != nil { | ||
74 | return false, time.Time{}, err | ||
75 | } | ||
76 | if matches[3] != "" { | ||
77 | expTime, err = parseRFC7231Time(matches[3]) | ||
78 | if err != nil { | ||
79 | return false, time.Time{}, err | ||
80 | } | ||
81 | } | ||
82 | return | ||
83 | } | ||
84 | |||
85 | // xmlDecoder provide decoded value in xml. | ||
86 | func xmlDecoder(body io.Reader, v interface{}) error { | ||
87 | d := xml.NewDecoder(body) | ||
88 | return d.Decode(v) | ||
89 | } | ||
90 | |||
91 | // sum256 calculate sha256sum for an input byte array, returns hex encoded. | ||
92 | func sum256Hex(data []byte) string { | ||
93 | hash := newSHA256Hasher() | ||
94 | defer hash.Close() | ||
95 | hash.Write(data) | ||
96 | return hex.EncodeToString(hash.Sum(nil)) | ||
97 | } | ||
98 | |||
99 | // sumMD5Base64 calculate md5sum for an input byte array, returns base64 encoded. | ||
100 | func sumMD5Base64(data []byte) string { | ||
101 | hash := newMd5Hasher() | ||
102 | defer hash.Close() | ||
103 | hash.Write(data) | ||
104 | return base64.StdEncoding.EncodeToString(hash.Sum(nil)) | ||
105 | } | ||
106 | |||
107 | // getEndpointURL - construct a new endpoint. | ||
108 | func getEndpointURL(endpoint string, secure bool) (*url.URL, error) { | ||
109 | // If secure is false, use 'http' scheme. | ||
110 | scheme := "https" | ||
111 | if !secure { | ||
112 | scheme = "http" | ||
113 | } | ||
114 | |||
115 | // Construct a secured endpoint URL. | ||
116 | endpointURLStr := scheme + "://" + endpoint | ||
117 | endpointURL, err := url.Parse(endpointURLStr) | ||
118 | if err != nil { | ||
119 | return nil, err | ||
120 | } | ||
121 | |||
122 | // Validate incoming endpoint URL. | ||
123 | if err := isValidEndpointURL(*endpointURL); err != nil { | ||
124 | return nil, err | ||
125 | } | ||
126 | return endpointURL, nil | ||
127 | } | ||
128 | |||
129 | // closeResponse close non nil response with any response Body. | ||
130 | // convenient wrapper to drain any remaining data on response body. | ||
131 | // | ||
132 | // Subsequently this allows golang http RoundTripper | ||
133 | // to re-use the same connection for future requests. | ||
134 | func closeResponse(resp *http.Response) { | ||
135 | // Callers should close resp.Body when done reading from it. | ||
136 | // If resp.Body is not closed, the Client's underlying RoundTripper | ||
137 | // (typically Transport) may not be able to re-use a persistent TCP | ||
138 | // connection to the server for a subsequent "keep-alive" request. | ||
139 | if resp != nil && resp.Body != nil { | ||
140 | // Drain any remaining Body and then close the connection. | ||
141 | // Without this closing connection would disallow re-using | ||
142 | // the same connection for future uses. | ||
143 | // - http://stackoverflow.com/a/17961593/4465767 | ||
144 | io.Copy(io.Discard, resp.Body) | ||
145 | resp.Body.Close() | ||
146 | } | ||
147 | } | ||
148 | |||
149 | var ( | ||
150 | // Hex encoded string of nil sha256sum bytes. | ||
151 | emptySHA256Hex = "e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855" | ||
152 | |||
153 | // Sentinel URL is the default url value which is invalid. | ||
154 | sentinelURL = url.URL{} | ||
155 | ) | ||
156 | |||
157 | // Verify if input endpoint URL is valid. | ||
158 | func isValidEndpointURL(endpointURL url.URL) error { | ||
159 | if endpointURL == sentinelURL { | ||
160 | return errInvalidArgument("Endpoint url cannot be empty.") | ||
161 | } | ||
162 | if endpointURL.Path != "/" && endpointURL.Path != "" { | ||
163 | return errInvalidArgument("Endpoint url cannot have fully qualified paths.") | ||
164 | } | ||
165 | host := endpointURL.Hostname() | ||
166 | if !s3utils.IsValidIP(host) && !s3utils.IsValidDomain(host) { | ||
167 | msg := "Endpoint: " + endpointURL.Host + " does not follow ip address or domain name standards." | ||
168 | return errInvalidArgument(msg) | ||
169 | } | ||
170 | |||
171 | if strings.Contains(host, ".s3.amazonaws.com") { | ||
172 | if !s3utils.IsAmazonEndpoint(endpointURL) { | ||
173 | return errInvalidArgument("Amazon S3 endpoint should be 's3.amazonaws.com'.") | ||
174 | } | ||
175 | } | ||
176 | if strings.Contains(host, ".googleapis.com") { | ||
177 | if !s3utils.IsGoogleEndpoint(endpointURL) { | ||
178 | return errInvalidArgument("Google Cloud Storage endpoint should be 'storage.googleapis.com'.") | ||
179 | } | ||
180 | } | ||
181 | return nil | ||
182 | } | ||
183 | |||
184 | // Verify if input expires value is valid. | ||
185 | func isValidExpiry(expires time.Duration) error { | ||
186 | expireSeconds := int64(expires / time.Second) | ||
187 | if expireSeconds < 1 { | ||
188 | return errInvalidArgument("Expires cannot be lesser than 1 second.") | ||
189 | } | ||
190 | if expireSeconds > 604800 { | ||
191 | return errInvalidArgument("Expires cannot be greater than 7 days.") | ||
192 | } | ||
193 | return nil | ||
194 | } | ||
195 | |||
196 | // Extract only necessary metadata header key/values by | ||
197 | // filtering them out with a list of custom header keys. | ||
198 | func extractObjMetadata(header http.Header) http.Header { | ||
199 | preserveKeys := []string{ | ||
200 | "Content-Type", | ||
201 | "Cache-Control", | ||
202 | "Content-Encoding", | ||
203 | "Content-Language", | ||
204 | "Content-Disposition", | ||
205 | "X-Amz-Storage-Class", | ||
206 | "X-Amz-Object-Lock-Mode", | ||
207 | "X-Amz-Object-Lock-Retain-Until-Date", | ||
208 | "X-Amz-Object-Lock-Legal-Hold", | ||
209 | "X-Amz-Website-Redirect-Location", | ||
210 | "X-Amz-Server-Side-Encryption", | ||
211 | "X-Amz-Tagging-Count", | ||
212 | "X-Amz-Meta-", | ||
213 | // Add new headers to be preserved. | ||
214 | // if you add new headers here, please extend | ||
215 | // PutObjectOptions{} to preserve them | ||
216 | // upon upload as well. | ||
217 | } | ||
218 | filteredHeader := make(http.Header) | ||
219 | for k, v := range header { | ||
220 | var found bool | ||
221 | for _, prefix := range preserveKeys { | ||
222 | if !strings.HasPrefix(k, prefix) { | ||
223 | continue | ||
224 | } | ||
225 | found = true | ||
226 | break | ||
227 | } | ||
228 | if found { | ||
229 | filteredHeader[k] = v | ||
230 | } | ||
231 | } | ||
232 | return filteredHeader | ||
233 | } | ||
234 | |||
235 | const ( | ||
236 | // RFC 7231#section-7.1.1.1 timetamp format. e.g Tue, 29 Apr 2014 18:30:38 GMT | ||
237 | rfc822TimeFormat = "Mon, 2 Jan 2006 15:04:05 GMT" | ||
238 | rfc822TimeFormatSingleDigitDay = "Mon, _2 Jan 2006 15:04:05 GMT" | ||
239 | rfc822TimeFormatSingleDigitDayTwoDigitYear = "Mon, _2 Jan 06 15:04:05 GMT" | ||
240 | ) | ||
241 | |||
242 | func parseTime(t string, formats ...string) (time.Time, error) { | ||
243 | for _, format := range formats { | ||
244 | tt, err := time.Parse(format, t) | ||
245 | if err == nil { | ||
246 | return tt, nil | ||
247 | } | ||
248 | } | ||
249 | return time.Time{}, fmt.Errorf("unable to parse %s in any of the input formats: %s", t, formats) | ||
250 | } | ||
251 | |||
252 | func parseRFC7231Time(lastModified string) (time.Time, error) { | ||
253 | return parseTime(lastModified, rfc822TimeFormat, rfc822TimeFormatSingleDigitDay, rfc822TimeFormatSingleDigitDayTwoDigitYear) | ||
254 | } | ||
255 | |||
256 | // ToObjectInfo converts http header values into ObjectInfo type, | ||
257 | // extracts metadata and fills in all the necessary fields in ObjectInfo. | ||
258 | func ToObjectInfo(bucketName, objectName string, h http.Header) (ObjectInfo, error) { | ||
259 | var err error | ||
260 | // Trim off the odd double quotes from ETag in the beginning and end. | ||
261 | etag := trimEtag(h.Get("ETag")) | ||
262 | |||
263 | // Parse content length is exists | ||
264 | var size int64 = -1 | ||
265 | contentLengthStr := h.Get("Content-Length") | ||
266 | if contentLengthStr != "" { | ||
267 | size, err = strconv.ParseInt(contentLengthStr, 10, 64) | ||
268 | if err != nil { | ||
269 | // Content-Length is not valid | ||
270 | return ObjectInfo{}, ErrorResponse{ | ||
271 | Code: "InternalError", | ||
272 | Message: fmt.Sprintf("Content-Length is not an integer, failed with %v", err), | ||
273 | BucketName: bucketName, | ||
274 | Key: objectName, | ||
275 | RequestID: h.Get("x-amz-request-id"), | ||
276 | HostID: h.Get("x-amz-id-2"), | ||
277 | Region: h.Get("x-amz-bucket-region"), | ||
278 | } | ||
279 | } | ||
280 | } | ||
281 | |||
282 | // Parse Last-Modified has http time format. | ||
283 | mtime, err := parseRFC7231Time(h.Get("Last-Modified")) | ||
284 | if err != nil { | ||
285 | return ObjectInfo{}, ErrorResponse{ | ||
286 | Code: "InternalError", | ||
287 | Message: fmt.Sprintf("Last-Modified time format is invalid, failed with %v", err), | ||
288 | BucketName: bucketName, | ||
289 | Key: objectName, | ||
290 | RequestID: h.Get("x-amz-request-id"), | ||
291 | HostID: h.Get("x-amz-id-2"), | ||
292 | Region: h.Get("x-amz-bucket-region"), | ||
293 | } | ||
294 | } | ||
295 | |||
296 | // Fetch content type if any present. | ||
297 | contentType := strings.TrimSpace(h.Get("Content-Type")) | ||
298 | if contentType == "" { | ||
299 | contentType = "application/octet-stream" | ||
300 | } | ||
301 | |||
302 | expiryStr := h.Get("Expires") | ||
303 | var expiry time.Time | ||
304 | if expiryStr != "" { | ||
305 | expiry, err = parseRFC7231Time(expiryStr) | ||
306 | if err != nil { | ||
307 | return ObjectInfo{}, ErrorResponse{ | ||
308 | Code: "InternalError", | ||
309 | Message: fmt.Sprintf("'Expiry' is not in supported format: %v", err), | ||
310 | BucketName: bucketName, | ||
311 | Key: objectName, | ||
312 | RequestID: h.Get("x-amz-request-id"), | ||
313 | HostID: h.Get("x-amz-id-2"), | ||
314 | Region: h.Get("x-amz-bucket-region"), | ||
315 | } | ||
316 | } | ||
317 | } | ||
318 | |||
319 | metadata := extractObjMetadata(h) | ||
320 | userMetadata := make(map[string]string) | ||
321 | for k, v := range metadata { | ||
322 | if strings.HasPrefix(k, "X-Amz-Meta-") { | ||
323 | userMetadata[strings.TrimPrefix(k, "X-Amz-Meta-")] = v[0] | ||
324 | } | ||
325 | } | ||
326 | userTags := s3utils.TagDecode(h.Get(amzTaggingHeader)) | ||
327 | |||
328 | var tagCount int | ||
329 | if count := h.Get(amzTaggingCount); count != "" { | ||
330 | tagCount, err = strconv.Atoi(count) | ||
331 | if err != nil { | ||
332 | return ObjectInfo{}, ErrorResponse{ | ||
333 | Code: "InternalError", | ||
334 | Message: fmt.Sprintf("x-amz-tagging-count is not an integer, failed with %v", err), | ||
335 | BucketName: bucketName, | ||
336 | Key: objectName, | ||
337 | RequestID: h.Get("x-amz-request-id"), | ||
338 | HostID: h.Get("x-amz-id-2"), | ||
339 | Region: h.Get("x-amz-bucket-region"), | ||
340 | } | ||
341 | } | ||
342 | } | ||
343 | |||
344 | // Nil if not found | ||
345 | var restore *RestoreInfo | ||
346 | if restoreHdr := h.Get(amzRestore); restoreHdr != "" { | ||
347 | ongoing, expTime, err := amzRestoreToStruct(restoreHdr) | ||
348 | if err != nil { | ||
349 | return ObjectInfo{}, err | ||
350 | } | ||
351 | restore = &RestoreInfo{OngoingRestore: ongoing, ExpiryTime: expTime} | ||
352 | } | ||
353 | |||
354 | // extract lifecycle expiry date and rule ID | ||
355 | expTime, ruleID := amzExpirationToExpiryDateRuleID(h.Get(amzExpiration)) | ||
356 | |||
357 | deleteMarker := h.Get(amzDeleteMarker) == "true" | ||
358 | |||
359 | // Save object metadata info. | ||
360 | return ObjectInfo{ | ||
361 | ETag: etag, | ||
362 | Key: objectName, | ||
363 | Size: size, | ||
364 | LastModified: mtime, | ||
365 | ContentType: contentType, | ||
366 | Expires: expiry, | ||
367 | VersionID: h.Get(amzVersionID), | ||
368 | IsDeleteMarker: deleteMarker, | ||
369 | ReplicationStatus: h.Get(amzReplicationStatus), | ||
370 | Expiration: expTime, | ||
371 | ExpirationRuleID: ruleID, | ||
372 | // Extract only the relevant header keys describing the object. | ||
373 | // following function filters out a list of standard set of keys | ||
374 | // which are not part of object metadata. | ||
375 | Metadata: metadata, | ||
376 | UserMetadata: userMetadata, | ||
377 | UserTags: userTags, | ||
378 | UserTagCount: tagCount, | ||
379 | Restore: restore, | ||
380 | |||
381 | // Checksum values | ||
382 | ChecksumCRC32: h.Get("x-amz-checksum-crc32"), | ||
383 | ChecksumCRC32C: h.Get("x-amz-checksum-crc32c"), | ||
384 | ChecksumSHA1: h.Get("x-amz-checksum-sha1"), | ||
385 | ChecksumSHA256: h.Get("x-amz-checksum-sha256"), | ||
386 | }, nil | ||
387 | } | ||
388 | |||
389 | var readFull = func(r io.Reader, buf []byte) (n int, err error) { | ||
390 | // ReadFull reads exactly len(buf) bytes from r into buf. | ||
391 | // It returns the number of bytes copied and an error if | ||
392 | // fewer bytes were read. The error is EOF only if no bytes | ||
393 | // were read. If an EOF happens after reading some but not | ||
394 | // all the bytes, ReadFull returns ErrUnexpectedEOF. | ||
395 | // On return, n == len(buf) if and only if err == nil. | ||
396 | // If r returns an error having read at least len(buf) bytes, | ||
397 | // the error is dropped. | ||
398 | for n < len(buf) && err == nil { | ||
399 | var nn int | ||
400 | nn, err = r.Read(buf[n:]) | ||
401 | // Some spurious io.Reader's return | ||
402 | // io.ErrUnexpectedEOF when nn == 0 | ||
403 | // this behavior is undocumented | ||
404 | // so we are on purpose not using io.ReadFull | ||
405 | // implementation because this can lead | ||
406 | // to custom handling, to avoid that | ||
407 | // we simply modify the original io.ReadFull | ||
408 | // implementation to avoid this issue. | ||
409 | // io.ErrUnexpectedEOF with nn == 0 really | ||
410 | // means that io.EOF | ||
411 | if err == io.ErrUnexpectedEOF && nn == 0 { | ||
412 | err = io.EOF | ||
413 | } | ||
414 | n += nn | ||
415 | } | ||
416 | if n >= len(buf) { | ||
417 | err = nil | ||
418 | } else if n > 0 && err == io.EOF { | ||
419 | err = io.ErrUnexpectedEOF | ||
420 | } | ||
421 | return | ||
422 | } | ||
423 | |||
424 | // regCred matches credential string in HTTP header | ||
425 | var regCred = regexp.MustCompile("Credential=([A-Z0-9]+)/") | ||
426 | |||
427 | // regCred matches signature string in HTTP header | ||
428 | var regSign = regexp.MustCompile("Signature=([[0-9a-f]+)") | ||
429 | |||
430 | // Redact out signature value from authorization string. | ||
431 | func redactSignature(origAuth string) string { | ||
432 | if !strings.HasPrefix(origAuth, signV4Algorithm) { | ||
433 | // Set a temporary redacted auth | ||
434 | return "AWS **REDACTED**:**REDACTED**" | ||
435 | } | ||
436 | |||
437 | // Signature V4 authorization header. | ||
438 | |||
439 | // Strip out accessKeyID from: | ||
440 | // Credential=<access-key-id>/<date>/<aws-region>/<aws-service>/aws4_request | ||
441 | newAuth := regCred.ReplaceAllString(origAuth, "Credential=**REDACTED**/") | ||
442 | |||
443 | // Strip out 256-bit signature from: Signature=<256-bit signature> | ||
444 | return regSign.ReplaceAllString(newAuth, "Signature=**REDACTED**") | ||
445 | } | ||
446 | |||
447 | // Get default location returns the location based on the input | ||
448 | // URL `u`, if region override is provided then all location | ||
449 | // defaults to regionOverride. | ||
450 | // | ||
451 | // If no other cases match then the location is set to `us-east-1` | ||
452 | // as a last resort. | ||
453 | func getDefaultLocation(u url.URL, regionOverride string) (location string) { | ||
454 | if regionOverride != "" { | ||
455 | return regionOverride | ||
456 | } | ||
457 | region := s3utils.GetRegionFromURL(u) | ||
458 | if region == "" { | ||
459 | region = "us-east-1" | ||
460 | } | ||
461 | return region | ||
462 | } | ||
463 | |||
464 | var supportedHeaders = map[string]bool{ | ||
465 | "content-type": true, | ||
466 | "cache-control": true, | ||
467 | "content-encoding": true, | ||
468 | "content-disposition": true, | ||
469 | "content-language": true, | ||
470 | "x-amz-website-redirect-location": true, | ||
471 | "x-amz-object-lock-mode": true, | ||
472 | "x-amz-metadata-directive": true, | ||
473 | "x-amz-object-lock-retain-until-date": true, | ||
474 | "expires": true, | ||
475 | "x-amz-replication-status": true, | ||
476 | // Add more supported headers here. | ||
477 | // Must be lower case. | ||
478 | } | ||
479 | |||
480 | // isStorageClassHeader returns true if the header is a supported storage class header | ||
481 | func isStorageClassHeader(headerKey string) bool { | ||
482 | return strings.EqualFold(amzStorageClass, headerKey) | ||
483 | } | ||
484 | |||
485 | // isStandardHeader returns true if header is a supported header and not a custom header | ||
486 | func isStandardHeader(headerKey string) bool { | ||
487 | return supportedHeaders[strings.ToLower(headerKey)] | ||
488 | } | ||
489 | |||
490 | // sseHeaders is list of server side encryption headers | ||
491 | var sseHeaders = map[string]bool{ | ||
492 | "x-amz-server-side-encryption": true, | ||
493 | "x-amz-server-side-encryption-aws-kms-key-id": true, | ||
494 | "x-amz-server-side-encryption-context": true, | ||
495 | "x-amz-server-side-encryption-customer-algorithm": true, | ||
496 | "x-amz-server-side-encryption-customer-key": true, | ||
497 | "x-amz-server-side-encryption-customer-key-md5": true, | ||
498 | // Add more supported headers here. | ||
499 | // Must be lower case. | ||
500 | } | ||
501 | |||
502 | // isSSEHeader returns true if header is a server side encryption header. | ||
503 | func isSSEHeader(headerKey string) bool { | ||
504 | return sseHeaders[strings.ToLower(headerKey)] | ||
505 | } | ||
506 | |||
507 | // isAmzHeader returns true if header is a x-amz-meta-* or x-amz-acl header. | ||
508 | func isAmzHeader(headerKey string) bool { | ||
509 | key := strings.ToLower(headerKey) | ||
510 | |||
511 | return strings.HasPrefix(key, "x-amz-meta-") || strings.HasPrefix(key, "x-amz-grant-") || key == "x-amz-acl" || isSSEHeader(headerKey) || strings.HasPrefix(key, "x-amz-checksum-") | ||
512 | } | ||
513 | |||
514 | // supportedQueryValues is a list of query strings that can be passed in when using GetObject. | ||
515 | var supportedQueryValues = map[string]bool{ | ||
516 | "partNumber": true, | ||
517 | "versionId": true, | ||
518 | "response-cache-control": true, | ||
519 | "response-content-disposition": true, | ||
520 | "response-content-encoding": true, | ||
521 | "response-content-language": true, | ||
522 | "response-content-type": true, | ||
523 | "response-expires": true, | ||
524 | } | ||
525 | |||
526 | // isStandardQueryValue will return true when the passed in query string parameter is supported rather than customized. | ||
527 | func isStandardQueryValue(qsKey string) bool { | ||
528 | return supportedQueryValues[qsKey] | ||
529 | } | ||
530 | |||
531 | // Per documentation at https://docs.aws.amazon.com/AmazonS3/latest/userguide/LogFormat.html#LogFormatCustom, the | ||
532 | // set of query params starting with "x-" are ignored by S3. | ||
533 | const allowedCustomQueryPrefix = "x-" | ||
534 | |||
535 | func isCustomQueryValue(qsKey string) bool { | ||
536 | return strings.HasPrefix(qsKey, allowedCustomQueryPrefix) | ||
537 | } | ||
538 | |||
539 | var ( | ||
540 | md5Pool = sync.Pool{New: func() interface{} { return md5.New() }} | ||
541 | sha256Pool = sync.Pool{New: func() interface{} { return sha256.New() }} | ||
542 | ) | ||
543 | |||
544 | func newMd5Hasher() md5simd.Hasher { | ||
545 | return &hashWrapper{Hash: md5Pool.Get().(hash.Hash), isMD5: true} | ||
546 | } | ||
547 | |||
548 | func newSHA256Hasher() md5simd.Hasher { | ||
549 | if encrypt.FIPS { | ||
550 | return &hashWrapper{Hash: fipssha256.New(), isSHA256: true} | ||
551 | } | ||
552 | return &hashWrapper{Hash: sha256Pool.Get().(hash.Hash), isSHA256: true} | ||
553 | } | ||
554 | |||
555 | // hashWrapper implements the md5simd.Hasher interface. | ||
556 | type hashWrapper struct { | ||
557 | hash.Hash | ||
558 | isMD5 bool | ||
559 | isSHA256 bool | ||
560 | } | ||
561 | |||
562 | // Close will put the hasher back into the pool. | ||
563 | func (m *hashWrapper) Close() { | ||
564 | if m.isMD5 && m.Hash != nil { | ||
565 | m.Reset() | ||
566 | md5Pool.Put(m.Hash) | ||
567 | } | ||
568 | if m.isSHA256 && m.Hash != nil { | ||
569 | m.Reset() | ||
570 | sha256Pool.Put(m.Hash) | ||
571 | } | ||
572 | m.Hash = nil | ||
573 | } | ||
574 | |||
575 | const letterBytes = "abcdefghijklmnopqrstuvwxyz01234569" | ||
576 | const ( | ||
577 | letterIdxBits = 6 // 6 bits to represent a letter index | ||
578 | letterIdxMask = 1<<letterIdxBits - 1 // All 1-bits, as many as letterIdxBits | ||
579 | letterIdxMax = 63 / letterIdxBits // # of letter indices fitting in 63 bits | ||
580 | ) | ||
581 | |||
582 | // randString generates random names and prepends them with a known prefix. | ||
583 | func randString(n int, src rand.Source, prefix string) string { | ||
584 | b := make([]byte, n) | ||
585 | // A rand.Int63() generates 63 random bits, enough for letterIdxMax letters! | ||
586 | for i, cache, remain := n-1, src.Int63(), letterIdxMax; i >= 0; { | ||
587 | if remain == 0 { | ||
588 | cache, remain = src.Int63(), letterIdxMax | ||
589 | } | ||
590 | if idx := int(cache & letterIdxMask); idx < len(letterBytes) { | ||
591 | b[i] = letterBytes[idx] | ||
592 | i-- | ||
593 | } | ||
594 | cache >>= letterIdxBits | ||
595 | remain-- | ||
596 | } | ||
597 | return prefix + string(b[0:30-len(prefix)]) | ||
598 | } | ||
599 | |||
600 | // IsNetworkOrHostDown - if there was a network error or if the host is down. | ||
601 | // expectTimeouts indicates that *context* timeouts are expected and does not | ||
602 | // indicate a downed host. Other timeouts still returns down. | ||
603 | func IsNetworkOrHostDown(err error, expectTimeouts bool) bool { | ||
604 | if err == nil { | ||
605 | return false | ||
606 | } | ||
607 | |||
608 | if errors.Is(err, context.Canceled) { | ||
609 | return false | ||
610 | } | ||
611 | |||
612 | if expectTimeouts && errors.Is(err, context.DeadlineExceeded) { | ||
613 | return false | ||
614 | } | ||
615 | |||
616 | if errors.Is(err, context.DeadlineExceeded) { | ||
617 | return true | ||
618 | } | ||
619 | |||
620 | // We need to figure if the error either a timeout | ||
621 | // or a non-temporary error. | ||
622 | urlErr := &url.Error{} | ||
623 | if errors.As(err, &urlErr) { | ||
624 | switch urlErr.Err.(type) { | ||
625 | case *net.DNSError, *net.OpError, net.UnknownNetworkError: | ||
626 | return true | ||
627 | } | ||
628 | } | ||
629 | var e net.Error | ||
630 | if errors.As(err, &e) { | ||
631 | if e.Timeout() { | ||
632 | return true | ||
633 | } | ||
634 | } | ||
635 | |||
636 | // Fallback to other mechanisms. | ||
637 | switch { | ||
638 | case strings.Contains(err.Error(), "Connection closed by foreign host"): | ||
639 | return true | ||
640 | case strings.Contains(err.Error(), "TLS handshake timeout"): | ||
641 | // If error is - tlsHandshakeTimeoutError. | ||
642 | return true | ||
643 | case strings.Contains(err.Error(), "i/o timeout"): | ||
644 | // If error is - tcp timeoutError. | ||
645 | return true | ||
646 | case strings.Contains(err.Error(), "connection timed out"): | ||
647 | // If err is a net.Dial timeout. | ||
648 | return true | ||
649 | case strings.Contains(err.Error(), "connection refused"): | ||
650 | // If err is connection refused | ||
651 | return true | ||
652 | |||
653 | case strings.Contains(strings.ToLower(err.Error()), "503 service unavailable"): | ||
654 | // Denial errors | ||
655 | return true | ||
656 | } | ||
657 | return false | ||
658 | } | ||
659 | |||
660 | // newHashReaderWrapper will hash all reads done through r. | ||
661 | // When r returns io.EOF the done function will be called with the sum. | ||
662 | func newHashReaderWrapper(r io.Reader, h hash.Hash, done func(hash []byte)) *hashReaderWrapper { | ||
663 | return &hashReaderWrapper{ | ||
664 | r: r, | ||
665 | h: h, | ||
666 | done: done, | ||
667 | } | ||
668 | } | ||
669 | |||
670 | type hashReaderWrapper struct { | ||
671 | r io.Reader | ||
672 | h hash.Hash | ||
673 | done func(hash []byte) | ||
674 | } | ||
675 | |||
676 | // Read implements the io.Reader interface. | ||
677 | func (h *hashReaderWrapper) Read(p []byte) (n int, err error) { | ||
678 | n, err = h.r.Read(p) | ||
679 | if n > 0 { | ||
680 | n2, err := h.h.Write(p[:n]) | ||
681 | if err != nil { | ||
682 | return 0, err | ||
683 | } | ||
684 | if n2 != n { | ||
685 | return 0, io.ErrShortWrite | ||
686 | } | ||
687 | } | ||
688 | if err == io.EOF { | ||
689 | // Call back | ||
690 | h.done(h.h.Sum(nil)) | ||
691 | } | ||
692 | return n, err | ||
693 | } | ||