mirror of https://github.com/minio/minio.git
				
				
				
			
		
			
				
	
	
		
			379 lines
		
	
	
		
			11 KiB
		
	
	
	
		
			Go
		
	
	
	
			
		
		
	
	
			379 lines
		
	
	
		
			11 KiB
		
	
	
	
		
			Go
		
	
	
	
/*
 | 
						|
 * Minio Cloud Storage, (C) 2015, 2016, 2017 Minio, Inc.
 | 
						|
 *
 | 
						|
 * Licensed under the Apache License, Version 2.0 (the "License");
 | 
						|
 * you may not use this file except in compliance with the License.
 | 
						|
 * You may obtain a copy of the License at
 | 
						|
 *
 | 
						|
 *     http://www.apache.org/licenses/LICENSE-2.0
 | 
						|
 *
 | 
						|
 * Unless required by applicable law or agreed to in writing, software
 | 
						|
 * distributed under the License is distributed on an "AS IS" BASIS,
 | 
						|
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | 
						|
 * See the License for the specific language governing permissions and
 | 
						|
 * limitations under the License.
 | 
						|
 */
 | 
						|
 | 
						|
package cmd
 | 
						|
 | 
						|
import (
 | 
						|
	"bytes"
 | 
						|
	"context"
 | 
						|
	"io"
 | 
						|
	"io/ioutil"
 | 
						|
	"mime/multipart"
 | 
						|
	"net"
 | 
						|
	"net/http"
 | 
						|
	"net/url"
 | 
						|
	"strings"
 | 
						|
 | 
						|
	"github.com/minio/minio/cmd/logger"
 | 
						|
	"github.com/minio/minio/pkg/auth"
 | 
						|
	"github.com/minio/minio/pkg/handlers"
 | 
						|
	httptracer "github.com/minio/minio/pkg/handlers"
 | 
						|
)
 | 
						|
 | 
						|
// Parses location constraint from the incoming reader.
 | 
						|
func parseLocationConstraint(r *http.Request) (location string, s3Error APIErrorCode) {
 | 
						|
	// If the request has no body with content-length set to 0,
 | 
						|
	// we do not have to validate location constraint. Bucket will
 | 
						|
	// be created at default region.
 | 
						|
	locationConstraint := createBucketLocationConfiguration{}
 | 
						|
	err := xmlDecoder(r.Body, &locationConstraint, r.ContentLength)
 | 
						|
	if err != nil && r.ContentLength != 0 {
 | 
						|
		logger.LogIf(context.Background(), err)
 | 
						|
		// Treat all other failures as XML parsing errors.
 | 
						|
		return "", ErrMalformedXML
 | 
						|
	} // else for both err as nil or io.EOF
 | 
						|
	location = locationConstraint.Location
 | 
						|
	if location == "" {
 | 
						|
		location = globalServerConfig.GetRegion()
 | 
						|
	}
 | 
						|
	return location, ErrNone
 | 
						|
}
 | 
						|
 | 
						|
// Validates input location is same as configured region
 | 
						|
// of Minio server.
 | 
						|
func isValidLocation(location string) bool {
 | 
						|
	return globalServerConfig.GetRegion() == "" || globalServerConfig.GetRegion() == location
 | 
						|
}
 | 
						|
 | 
						|
// Supported headers that needs to be extracted.
 | 
						|
var supportedHeaders = []string{
 | 
						|
	"content-type",
 | 
						|
	"cache-control",
 | 
						|
	"content-language",
 | 
						|
	"content-encoding",
 | 
						|
	"content-disposition",
 | 
						|
	amzStorageClass,
 | 
						|
	"expires",
 | 
						|
	// Add more supported headers here.
 | 
						|
}
 | 
						|
 | 
						|
// isMetadataDirectiveValid - check if metadata-directive is valid.
 | 
						|
func isMetadataDirectiveValid(h http.Header) bool {
 | 
						|
	_, ok := h[http.CanonicalHeaderKey("X-Amz-Metadata-Directive")]
 | 
						|
	if ok {
 | 
						|
		// Check atleast set metadata-directive is valid.
 | 
						|
		return (isMetadataCopy(h) || isMetadataReplace(h))
 | 
						|
	}
 | 
						|
	// By default if x-amz-metadata-directive is not we
 | 
						|
	// treat it as 'COPY' this function returns true.
 | 
						|
	return true
 | 
						|
}
 | 
						|
 | 
						|
// Check if the metadata COPY is requested.
 | 
						|
func isMetadataCopy(h http.Header) bool {
 | 
						|
	return h.Get("X-Amz-Metadata-Directive") == "COPY"
 | 
						|
}
 | 
						|
 | 
						|
// Check if the metadata REPLACE is requested.
 | 
						|
func isMetadataReplace(h http.Header) bool {
 | 
						|
	return h.Get("X-Amz-Metadata-Directive") == "REPLACE"
 | 
						|
}
 | 
						|
 | 
						|
// Splits an incoming path into bucket and object components.
 | 
						|
func path2BucketAndObject(path string) (bucket, object string) {
 | 
						|
	// Skip the first element if it is '/', split the rest.
 | 
						|
	path = strings.TrimPrefix(path, "/")
 | 
						|
	pathComponents := strings.SplitN(path, "/", 2)
 | 
						|
 | 
						|
	// Save the bucket and object extracted from path.
 | 
						|
	switch len(pathComponents) {
 | 
						|
	case 1:
 | 
						|
		bucket = pathComponents[0]
 | 
						|
	case 2:
 | 
						|
		bucket = pathComponents[0]
 | 
						|
		object = pathComponents[1]
 | 
						|
	}
 | 
						|
	return bucket, object
 | 
						|
}
 | 
						|
 | 
						|
// userMetadataKeyPrefixes contains the prefixes of used-defined metadata keys.
 | 
						|
// All values stored with a key starting with one of the following prefixes
 | 
						|
// must be extracted from the header.
 | 
						|
var userMetadataKeyPrefixes = []string{
 | 
						|
	"X-Amz-Meta-",
 | 
						|
	"X-Minio-Meta-",
 | 
						|
}
 | 
						|
 | 
						|
// extractMetadata extracts metadata from HTTP header and HTTP queryString.
 | 
						|
func extractMetadata(ctx context.Context, r *http.Request) (metadata map[string]string, err error) {
 | 
						|
	query := r.URL.Query()
 | 
						|
	header := r.Header
 | 
						|
	metadata = make(map[string]string)
 | 
						|
	// Extract all query values.
 | 
						|
	err = extractMetadataFromMap(ctx, query, metadata)
 | 
						|
	if err != nil {
 | 
						|
		return nil, err
 | 
						|
	}
 | 
						|
 | 
						|
	// Extract all header values.
 | 
						|
	err = extractMetadataFromMap(ctx, header, metadata)
 | 
						|
	if err != nil {
 | 
						|
		return nil, err
 | 
						|
	}
 | 
						|
 | 
						|
	// Set content-type to default value if it is not set.
 | 
						|
	if _, ok := metadata["content-type"]; !ok {
 | 
						|
		metadata["content-type"] = "application/octet-stream"
 | 
						|
	}
 | 
						|
 | 
						|
	// Success.
 | 
						|
	return metadata, nil
 | 
						|
}
 | 
						|
 | 
						|
// extractMetadata extracts metadata from map values.
 | 
						|
func extractMetadataFromMap(ctx context.Context, v map[string][]string, m map[string]string) error {
 | 
						|
	if v == nil {
 | 
						|
		logger.LogIf(ctx, errInvalidArgument)
 | 
						|
		return errInvalidArgument
 | 
						|
	}
 | 
						|
	// Save all supported headers.
 | 
						|
	for _, supportedHeader := range supportedHeaders {
 | 
						|
		if value, ok := v[http.CanonicalHeaderKey(supportedHeader)]; ok {
 | 
						|
			m[supportedHeader] = value[0]
 | 
						|
		} else if value, ok := v[supportedHeader]; ok {
 | 
						|
			m[supportedHeader] = value[0]
 | 
						|
		}
 | 
						|
	}
 | 
						|
	for key := range v {
 | 
						|
		for _, prefix := range userMetadataKeyPrefixes {
 | 
						|
			if !strings.HasPrefix(strings.ToLower(key), strings.ToLower(prefix)) {
 | 
						|
				continue
 | 
						|
			}
 | 
						|
			value, ok := v[key]
 | 
						|
			if ok {
 | 
						|
				m[key] = strings.Join(value, ",")
 | 
						|
				break
 | 
						|
			}
 | 
						|
		}
 | 
						|
	}
 | 
						|
	return nil
 | 
						|
}
 | 
						|
 | 
						|
// The Query string for the redirect URL the client is
 | 
						|
// redirected on successful upload.
 | 
						|
func getRedirectPostRawQuery(objInfo ObjectInfo) string {
 | 
						|
	redirectValues := make(url.Values)
 | 
						|
	redirectValues.Set("bucket", objInfo.Bucket)
 | 
						|
	redirectValues.Set("key", objInfo.Name)
 | 
						|
	redirectValues.Set("etag", "\""+objInfo.ETag+"\"")
 | 
						|
	return redirectValues.Encode()
 | 
						|
}
 | 
						|
 | 
						|
// Returns access credentials in the request Authorization header.
 | 
						|
func getReqAccessCred(r *http.Request, region string) (cred auth.Credentials) {
 | 
						|
	cred, _, _ = getReqAccessKeyV4(r, region, serviceS3)
 | 
						|
	if cred.AccessKey == "" {
 | 
						|
		cred, _, _ = getReqAccessKeyV2(r)
 | 
						|
	}
 | 
						|
	if cred.AccessKey == "" {
 | 
						|
		claims, owner, _ := webRequestAuthenticate(r)
 | 
						|
		if owner {
 | 
						|
			return globalServerConfig.GetCredential()
 | 
						|
		}
 | 
						|
		cred, _ = globalIAMSys.GetUser(claims.Subject)
 | 
						|
	}
 | 
						|
	return cred
 | 
						|
}
 | 
						|
 | 
						|
// Extract request params to be sent with event notifiation.
 | 
						|
func extractReqParams(r *http.Request) map[string]string {
 | 
						|
	if r == nil {
 | 
						|
		return nil
 | 
						|
	}
 | 
						|
 | 
						|
	region := globalServerConfig.GetRegion()
 | 
						|
	cred := getReqAccessCred(r, region)
 | 
						|
 | 
						|
	// Success.
 | 
						|
	return map[string]string{
 | 
						|
		"region":          region,
 | 
						|
		"accessKey":       cred.AccessKey,
 | 
						|
		"sourceIPAddress": handlers.GetSourceIP(r),
 | 
						|
		// Add more fields here.
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
// Extract response elements to be sent with event notifiation.
 | 
						|
func extractRespElements(w http.ResponseWriter) map[string]string {
 | 
						|
 | 
						|
	return map[string]string{
 | 
						|
		"requestId":      w.Header().Get(responseRequestIDKey),
 | 
						|
		"content-length": w.Header().Get("Content-Length"),
 | 
						|
		// Add more fields here.
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
// Trims away `aws-chunked` from the content-encoding header if present.
 | 
						|
// Streaming signature clients can have custom content-encoding such as
 | 
						|
// `aws-chunked,gzip` here we need to only save `gzip`.
 | 
						|
// For more refer http://docs.aws.amazon.com/AmazonS3/latest/API/sigv4-streaming.html
 | 
						|
func trimAwsChunkedContentEncoding(contentEnc string) (trimmedContentEnc string) {
 | 
						|
	if contentEnc == "" {
 | 
						|
		return contentEnc
 | 
						|
	}
 | 
						|
	var newEncs []string
 | 
						|
	for _, enc := range strings.Split(contentEnc, ",") {
 | 
						|
		if enc != streamingContentEncoding {
 | 
						|
			newEncs = append(newEncs, enc)
 | 
						|
		}
 | 
						|
	}
 | 
						|
	return strings.Join(newEncs, ",")
 | 
						|
}
 | 
						|
 | 
						|
// Validate form field size for s3 specification requirement.
 | 
						|
func validateFormFieldSize(ctx context.Context, formValues http.Header) error {
 | 
						|
	// Iterate over form values
 | 
						|
	for k := range formValues {
 | 
						|
		// Check if value's field exceeds S3 limit
 | 
						|
		if int64(len(formValues.Get(k))) > maxFormFieldSize {
 | 
						|
			logger.LogIf(ctx, errSizeUnexpected)
 | 
						|
			return errSizeUnexpected
 | 
						|
		}
 | 
						|
	}
 | 
						|
 | 
						|
	// Success.
 | 
						|
	return nil
 | 
						|
}
 | 
						|
 | 
						|
// Extract form fields and file data from a HTTP POST Policy
 | 
						|
func extractPostPolicyFormValues(ctx context.Context, form *multipart.Form) (filePart io.ReadCloser, fileName string, fileSize int64, formValues http.Header, err error) {
 | 
						|
	/// HTML Form values
 | 
						|
	fileName = ""
 | 
						|
 | 
						|
	// Canonicalize the form values into http.Header.
 | 
						|
	formValues = make(http.Header)
 | 
						|
	for k, v := range form.Value {
 | 
						|
		formValues[http.CanonicalHeaderKey(k)] = v
 | 
						|
	}
 | 
						|
 | 
						|
	// Validate form values.
 | 
						|
	if err = validateFormFieldSize(ctx, formValues); err != nil {
 | 
						|
		return nil, "", 0, nil, err
 | 
						|
	}
 | 
						|
 | 
						|
	// this means that filename="" was not specified for file key and Go has
 | 
						|
	// an ugly way of handling this situation. Refer here
 | 
						|
	// https://golang.org/src/mime/multipart/formdata.go#L61
 | 
						|
	if len(form.File) == 0 {
 | 
						|
		var b = &bytes.Buffer{}
 | 
						|
		for _, v := range formValues["File"] {
 | 
						|
			b.WriteString(v)
 | 
						|
		}
 | 
						|
		fileSize = int64(b.Len())
 | 
						|
		filePart = ioutil.NopCloser(b)
 | 
						|
		return filePart, fileName, fileSize, formValues, nil
 | 
						|
	}
 | 
						|
 | 
						|
	// Iterator until we find a valid File field and break
 | 
						|
	for k, v := range form.File {
 | 
						|
		canonicalFormName := http.CanonicalHeaderKey(k)
 | 
						|
		if canonicalFormName == "File" {
 | 
						|
			if len(v) == 0 {
 | 
						|
				logger.LogIf(ctx, errInvalidArgument)
 | 
						|
				return nil, "", 0, nil, errInvalidArgument
 | 
						|
			}
 | 
						|
			// Fetch fileHeader which has the uploaded file information
 | 
						|
			fileHeader := v[0]
 | 
						|
			// Set filename
 | 
						|
			fileName = fileHeader.Filename
 | 
						|
			// Open the uploaded part
 | 
						|
			filePart, err = fileHeader.Open()
 | 
						|
			if err != nil {
 | 
						|
				logger.LogIf(ctx, err)
 | 
						|
				return nil, "", 0, nil, err
 | 
						|
			}
 | 
						|
			// Compute file size
 | 
						|
			fileSize, err = filePart.(io.Seeker).Seek(0, 2)
 | 
						|
			if err != nil {
 | 
						|
				logger.LogIf(ctx, err)
 | 
						|
				return nil, "", 0, nil, err
 | 
						|
			}
 | 
						|
			// Reset Seek to the beginning
 | 
						|
			_, err = filePart.(io.Seeker).Seek(0, 0)
 | 
						|
			if err != nil {
 | 
						|
				logger.LogIf(ctx, err)
 | 
						|
				return nil, "", 0, nil, err
 | 
						|
			}
 | 
						|
			// File found and ready for reading
 | 
						|
			break
 | 
						|
		}
 | 
						|
	}
 | 
						|
	return filePart, fileName, fileSize, formValues, nil
 | 
						|
}
 | 
						|
 | 
						|
// Log headers and body.
 | 
						|
func httpTraceAll(f http.HandlerFunc) http.HandlerFunc {
 | 
						|
	if globalHTTPTraceFile == nil {
 | 
						|
		return f
 | 
						|
	}
 | 
						|
	return httptracer.TraceReqHandlerFunc(f, globalHTTPTraceFile, true)
 | 
						|
}
 | 
						|
 | 
						|
// Log only the headers.
 | 
						|
func httpTraceHdrs(f http.HandlerFunc) http.HandlerFunc {
 | 
						|
	if globalHTTPTraceFile == nil {
 | 
						|
		return f
 | 
						|
	}
 | 
						|
	return httptracer.TraceReqHandlerFunc(f, globalHTTPTraceFile, false)
 | 
						|
}
 | 
						|
 | 
						|
// Returns "/bucketName/objectName" for path-style or virtual-host-style requests.
 | 
						|
func getResource(path string, host string, domains []string) (string, error) {
 | 
						|
	if len(domains) == 0 {
 | 
						|
		return path, nil
 | 
						|
	}
 | 
						|
	// If virtual-host-style is enabled construct the "resource" properly.
 | 
						|
	if strings.Contains(host, ":") {
 | 
						|
		// In bucket.mydomain.com:9000, strip out :9000
 | 
						|
		var err error
 | 
						|
		if host, _, err = net.SplitHostPort(host); err != nil {
 | 
						|
			reqInfo := (&logger.ReqInfo{}).AppendTags("host", host)
 | 
						|
			reqInfo.AppendTags("path", path)
 | 
						|
			ctx := logger.SetReqInfo(context.Background(), reqInfo)
 | 
						|
			logger.LogIf(ctx, err)
 | 
						|
			return "", err
 | 
						|
		}
 | 
						|
	}
 | 
						|
	for _, domain := range domains {
 | 
						|
		if !strings.HasSuffix(host, "."+domain) {
 | 
						|
			continue
 | 
						|
		}
 | 
						|
		bucket := strings.TrimSuffix(host, "."+domain)
 | 
						|
		return slashSeparator + pathJoin(bucket, path), nil
 | 
						|
	}
 | 
						|
	return path, nil
 | 
						|
}
 | 
						|
 | 
						|
// If none of the http routes match respond with MethodNotAllowed, in JSON
 | 
						|
func notFoundHandlerJSON(w http.ResponseWriter, r *http.Request) {
 | 
						|
	writeErrorResponseJSON(context.Background(), w, errorCodes.ToAPIErr(ErrMethodNotAllowed), r.URL)
 | 
						|
}
 | 
						|
 | 
						|
// If none of the http routes match respond with MethodNotAllowed
 | 
						|
func notFoundHandler(w http.ResponseWriter, r *http.Request) {
 | 
						|
	writeErrorResponse(context.Background(), w, errorCodes.ToAPIErr(ErrMethodNotAllowed), r.URL, guessIsBrowserReq(r))
 | 
						|
}
 |