2
2
mirror of https://github.com/octoleo/restic.git synced 2024-11-30 00:33:57 +00:00
restic/vendor/github.com/minio/minio-go/api-put-object.go

268 lines
8.4 KiB
Go
Raw Normal View History

2017-07-23 12:24:45 +00:00
/*
2017-12-08 19:45:59 +00:00
* Minio Go Library for Amazon S3 Compatible Cloud Storage
* Copyright 2015-2017 Minio, Inc.
2017-07-23 12:24:45 +00:00
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package minio
import (
2017-09-13 12:09:48 +00:00
"bytes"
2017-12-08 19:45:59 +00:00
"context"
2017-09-13 12:09:48 +00:00
"fmt"
2017-07-23 12:24:45 +00:00
"io"
2017-12-08 19:45:59 +00:00
"net/http"
2017-09-13 12:09:48 +00:00
"runtime/debug"
"sort"
2017-07-23 12:24:45 +00:00
2017-12-08 19:45:59 +00:00
"github.com/minio/minio-go/pkg/encrypt"
2017-07-23 12:24:45 +00:00
"github.com/minio/minio-go/pkg/s3utils"
"golang.org/x/net/http/httpguts"
2017-07-23 12:24:45 +00:00
)
2017-12-08 19:45:59 +00:00
// PutObjectOptions represents options specified by user for PutObject call
type PutObjectOptions struct {
UserMetadata map[string]string
Progress io.Reader
ContentType string
ContentEncoding string
ContentDisposition string
ContentLanguage string
CacheControl string
ServerSideEncryption encrypt.ServerSide
NumThreads uint
StorageClass string
WebsiteRedirectLocation string
2017-12-08 19:45:59 +00:00
}
// getNumThreads - gets the number of threads to be used in the multipart
// put object operation
func (opts PutObjectOptions) getNumThreads() (numThreads int) {
if opts.NumThreads > 0 {
numThreads = int(opts.NumThreads)
} else {
numThreads = totalWorkers
2017-07-23 12:24:45 +00:00
}
2017-12-08 19:45:59 +00:00
return
2017-07-23 12:24:45 +00:00
}
2017-12-08 19:45:59 +00:00
// Header - constructs the headers from metadata entered by user in
// PutObjectOptions struct
func (opts PutObjectOptions) Header() (header http.Header) {
header = make(http.Header)
if opts.ContentType != "" {
header["Content-Type"] = []string{opts.ContentType}
} else {
header["Content-Type"] = []string{"application/octet-stream"}
2017-07-23 12:24:45 +00:00
}
2017-12-08 19:45:59 +00:00
if opts.ContentEncoding != "" {
header["Content-Encoding"] = []string{opts.ContentEncoding}
}
if opts.ContentDisposition != "" {
header["Content-Disposition"] = []string{opts.ContentDisposition}
}
if opts.ContentLanguage != "" {
header["Content-Language"] = []string{opts.ContentLanguage}
}
2017-12-08 19:45:59 +00:00
if opts.CacheControl != "" {
header["Cache-Control"] = []string{opts.CacheControl}
}
if opts.ServerSideEncryption != nil {
opts.ServerSideEncryption.Marshal(header)
2017-12-08 19:45:59 +00:00
}
if opts.StorageClass != "" {
header[amzStorageClass] = []string{opts.StorageClass}
}
if opts.WebsiteRedirectLocation != "" {
header[amzWebsiteRedirectLocation] = []string{opts.WebsiteRedirectLocation}
}
2017-12-08 19:45:59 +00:00
for k, v := range opts.UserMetadata {
if !isAmzHeader(k) && !isStandardHeader(k) && !isStorageClassHeader(k) {
2017-12-08 19:45:59 +00:00
header["X-Amz-Meta-"+k] = []string{v}
} else {
header[k] = []string{v}
2017-07-23 12:24:45 +00:00
}
2017-12-08 19:45:59 +00:00
}
return
}
// validate() checks if the UserMetadata map has standard headers or and raises an error if so.
2017-12-08 19:45:59 +00:00
func (opts PutObjectOptions) validate() (err error) {
for k, v := range opts.UserMetadata {
if !httpguts.ValidHeaderFieldName(k) || isStandardHeader(k) || isSSEHeader(k) || isStorageClassHeader(k) {
return ErrInvalidArgument(k + " unsupported user defined metadata name")
}
if !httpguts.ValidHeaderFieldValue(v) {
return ErrInvalidArgument(v + " unsupported user defined metadata value")
2017-07-23 12:24:45 +00:00
}
}
2017-12-08 19:45:59 +00:00
return nil
2017-07-23 12:24:45 +00:00
}
// completedParts is a collection of parts sortable by their part numbers.
// used for sorting the uploaded parts before completing the multipart request.
type completedParts []CompletePart
func (a completedParts) Len() int { return len(a) }
func (a completedParts) Swap(i, j int) { a[i], a[j] = a[j], a[i] }
func (a completedParts) Less(i, j int) bool { return a[i].PartNumber < a[j].PartNumber }
// PutObject creates an object in a bucket.
//
// You must have WRITE permissions on a bucket to create an object.
//
// - For size smaller than 64MiB PutObject automatically does a
// single atomic Put operation.
// - For size larger than 64MiB PutObject automatically does a
// multipart Put operation.
// - For size input as -1 PutObject does a multipart Put operation
// until input stream reaches EOF. Maximum object size that can
// be uploaded through this operation will be 5TiB.
2017-12-08 19:45:59 +00:00
func (c Client) PutObject(bucketName, objectName string, reader io.Reader, objectSize int64,
opts PutObjectOptions) (n int64, err error) {
return c.PutObjectWithContext(context.Background(), bucketName, objectName, reader, objectSize, opts)
2017-07-23 12:24:45 +00:00
}
2017-12-08 19:45:59 +00:00
func (c Client) putObjectCommon(ctx context.Context, bucketName, objectName string, reader io.Reader, size int64, opts PutObjectOptions) (n int64, err error) {
2017-07-23 12:24:45 +00:00
// Check for largest object size allowed.
if size > int64(maxMultipartPutObjectSize) {
return 0, ErrEntityTooLarge(size, maxMultipartPutObjectSize, bucketName, objectName)
}
// NOTE: Streaming signature is not supported by GCS.
if s3utils.IsGoogleEndpoint(*c.endpointURL) {
2017-07-23 12:24:45 +00:00
// Do not compute MD5 for Google Cloud Storage.
2017-12-08 19:45:59 +00:00
return c.putObjectNoChecksum(ctx, bucketName, objectName, reader, size, opts)
2017-07-23 12:24:45 +00:00
}
if c.overrideSignerType.IsV2() {
2017-09-13 12:09:48 +00:00
if size >= 0 && size < minPartSize {
2017-12-08 19:45:59 +00:00
return c.putObjectNoChecksum(ctx, bucketName, objectName, reader, size, opts)
2017-07-23 12:24:45 +00:00
}
2017-12-08 19:45:59 +00:00
return c.putObjectMultipart(ctx, bucketName, objectName, reader, size, opts)
2017-07-23 12:24:45 +00:00
}
if size < 0 {
2017-12-08 19:45:59 +00:00
return c.putObjectMultipartStreamNoLength(ctx, bucketName, objectName, reader, opts)
2017-07-23 12:24:45 +00:00
}
if size < minPartSize {
2017-12-08 19:45:59 +00:00
return c.putObjectNoChecksum(ctx, bucketName, objectName, reader, size, opts)
2017-07-23 12:24:45 +00:00
}
// For all sizes greater than 64MiB do multipart.
2017-12-08 19:45:59 +00:00
return c.putObjectMultipartStream(ctx, bucketName, objectName, reader, size, opts)
2017-07-23 12:24:45 +00:00
}
2017-09-13 12:09:48 +00:00
2017-12-08 19:45:59 +00:00
func (c Client) putObjectMultipartStreamNoLength(ctx context.Context, bucketName, objectName string, reader io.Reader, opts PutObjectOptions) (n int64, err error) {
2017-09-13 12:09:48 +00:00
// Input validation.
if err = s3utils.CheckValidBucketName(bucketName); err != nil {
return 0, err
}
if err = s3utils.CheckValidObjectName(objectName); err != nil {
return 0, err
}
// Total data read and written to server. should be equal to
// 'size' at the end of the call.
var totalUploadedSize int64
// Complete multipart upload.
var complMultipartUpload completeMultipartUpload
// Calculate the optimal parts info for a given size.
totalPartsCount, partSize, _, err := optimalPartInfo(-1)
if err != nil {
return 0, err
}
// Initiate a new multipart upload.
2017-12-08 19:45:59 +00:00
uploadID, err := c.newUploadID(ctx, bucketName, objectName, opts)
2017-09-13 12:09:48 +00:00
if err != nil {
return 0, err
}
defer func() {
if err != nil {
2017-12-08 19:45:59 +00:00
c.abortMultipartUpload(ctx, bucketName, objectName, uploadID)
2017-09-13 12:09:48 +00:00
}
}()
// Part number always starts with '1'.
partNumber := 1
// Initialize parts uploaded map.
partsInfo := make(map[int]ObjectPart)
// Create a buffer.
buf := make([]byte, partSize)
defer debug.FreeOSMemory()
for partNumber <= totalPartsCount {
length, rErr := io.ReadFull(reader, buf)
2017-12-08 19:45:59 +00:00
if rErr == io.EOF && partNumber > 1 {
2017-09-13 12:09:48 +00:00
break
}
if rErr != nil && rErr != io.ErrUnexpectedEOF && rErr != io.EOF {
2017-09-13 12:09:48 +00:00
return 0, rErr
}
// Update progress reader appropriately to the latest offset
// as we read from the source.
2017-12-08 19:45:59 +00:00
rd := newHook(bytes.NewReader(buf[:length]), opts.Progress)
2017-09-13 12:09:48 +00:00
// Proceed to upload the part.
var objPart ObjectPart
2017-12-08 19:45:59 +00:00
objPart, err = c.uploadPart(ctx, bucketName, objectName, uploadID, rd, partNumber,
"", "", int64(length), opts.ServerSideEncryption)
2017-09-13 12:09:48 +00:00
if err != nil {
return totalUploadedSize, err
}
// Save successfully uploaded part metadata.
partsInfo[partNumber] = objPart
// Save successfully uploaded size.
totalUploadedSize += int64(length)
// Increment part number.
partNumber++
// For unknown size, Read EOF we break away.
// We do not have to upload till totalPartsCount.
if rErr == io.EOF {
break
}
}
// Loop over total uploaded parts to save them in
// Parts array before completing the multipart request.
for i := 1; i < partNumber; i++ {
part, ok := partsInfo[i]
if !ok {
return 0, ErrInvalidArgument(fmt.Sprintf("Missing part number %d", i))
}
complMultipartUpload.Parts = append(complMultipartUpload.Parts, CompletePart{
ETag: part.ETag,
PartNumber: part.PartNumber,
})
}
// Sort all completed parts.
sort.Sort(completedParts(complMultipartUpload.Parts))
2017-12-08 19:45:59 +00:00
if _, err = c.completeMultipartUpload(ctx, bucketName, objectName, uploadID, complMultipartUpload); err != nil {
2017-09-13 12:09:48 +00:00
return totalUploadedSize, err
}
// Return final size.
return totalUploadedSize, nil
}