mirror of https://github.com/minio/minio.git
				
				
				
			
		
			
	
	
		
			1022 lines
		
	
	
		
			34 KiB
		
	
	
	
		
			Go
		
	
	
	
		
		
			
		
	
	
			1022 lines
		
	
	
		
			34 KiB
		
	
	
	
		
			Go
		
	
	
	
|  | /* | ||
|  |  * Minio Cloud Storage, (C) 2018 Minio, Inc. | ||
|  |  * | ||
|  |  * Licensed under the Apache License, Version 2.0 (the "License"); | ||
|  |  * you may not use this file except in compliance with the License. | ||
|  |  * You may obtain a copy of the License at | ||
|  |  * | ||
|  |  *     http://www.apache.org/licenses/LICENSE-2.0
 | ||
|  |  * | ||
|  |  * Unless required by applicable law or agreed to in writing, software | ||
|  |  * distributed under the License is distributed on an "AS IS" BASIS, | ||
|  |  * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
|  |  * See the License for the specific language governing permissions and | ||
|  |  * limitations under the License. | ||
|  |  */ | ||
|  | 
 | ||
|  | package cmd | ||
|  | 
 | ||
|  | import ( | ||
|  | 	"context" | ||
|  | 	"crypto/sha256" | ||
|  | 	"errors" | ||
|  | 	"fmt" | ||
|  | 	"hash/crc32" | ||
|  | 	"io" | ||
|  | 	"io/ioutil" | ||
|  | 	"net/http" | ||
|  | 	"os" | ||
|  | 	"path" | ||
|  | 	"sort" | ||
|  | 	"strconv" | ||
|  | 	"strings" | ||
|  | 	"time" | ||
|  | 
 | ||
|  | 	"github.com/djherbis/atime" | ||
|  | 
 | ||
|  | 	errors2 "github.com/minio/minio/pkg/errors" | ||
|  | 	"github.com/minio/minio/pkg/wildcard" | ||
|  | 
 | ||
|  | 	"github.com/minio/minio/pkg/hash" | ||
|  | ) | ||
|  | 
 | ||
|  | // list of all errors that can be ignored in tree walk operation in disk cache
 | ||
|  | var cacheTreeWalkIgnoredErrs = append(baseIgnoredErrs, errDiskAccessDenied, errVolumeNotFound, errFileNotFound) | ||
|  | 
 | ||
|  | const ( | ||
|  | 	// disk cache needs to have cacheSizeMultiplier * object size space free for a cache entry to be created.
 | ||
|  | 	cacheSizeMultiplier  = 100 | ||
|  | 	cacheTrashDir        = "trash" | ||
|  | 	cacheMaxDiskUsagePct = 80 // in %
 | ||
|  | 	cacheCleanupInterval = 10 // in minutes
 | ||
|  | ) | ||
|  | 
 | ||
|  | // abstract slice of cache drives backed by FS.
 | ||
|  | type diskCache struct { | ||
|  | 	cfs []*cacheFSObjects | ||
|  | } | ||
|  | 
 | ||
|  | // Abstracts disk caching - used by the S3 layer
 | ||
|  | type cacheObjects struct { | ||
|  | 	// pointer to disk cache
 | ||
|  | 	cache *diskCache | ||
|  | 	// ListObjects pool management.
 | ||
|  | 	listPool *treeWalkPool | ||
|  | 	// file path patterns to exclude from cache
 | ||
|  | 	exclude []string | ||
|  | 	// Object functions pointing to the corresponding functions of backend implementation.
 | ||
|  | 	GetObjectFn               func(ctx context.Context, bucket, object string, startOffset int64, length int64, writer io.Writer, etag string) (err error) | ||
|  | 	GetObjectInfoFn           func(ctx context.Context, bucket, object string) (objInfo ObjectInfo, err error) | ||
|  | 	PutObjectFn               func(ctx context.Context, bucket, object string, data *hash.Reader, metadata map[string]string) (objInfo ObjectInfo, err error) | ||
|  | 	DeleteObjectFn            func(ctx context.Context, bucket, object string) error | ||
|  | 	ListObjectsFn             func(ctx context.Context, bucket, prefix, marker, delimiter string, maxKeys int) (result ListObjectsInfo, err error) | ||
|  | 	ListObjectsV2Fn           func(ctx context.Context, bucket, prefix, continuationToken, delimiter string, maxKeys int, fetchOwner bool, startAfter string) (result ListObjectsV2Info, err error) | ||
|  | 	ListBucketsFn             func(ctx context.Context) (buckets []BucketInfo, err error) | ||
|  | 	GetBucketInfoFn           func(ctx context.Context, bucket string) (bucketInfo BucketInfo, err error) | ||
|  | 	NewMultipartUploadFn      func(ctx context.Context, bucket, object string, metadata map[string]string) (uploadID string, err error) | ||
|  | 	PutObjectPartFn           func(ctx context.Context, bucket, object, uploadID string, partID int, data *hash.Reader) (info PartInfo, err error) | ||
|  | 	AbortMultipartUploadFn    func(ctx context.Context, bucket, object, uploadID string) error | ||
|  | 	CompleteMultipartUploadFn func(ctx context.Context, bucket, object, uploadID string, uploadedParts []CompletePart) (objInfo ObjectInfo, err error) | ||
|  | 	DeleteBucketFn            func(ctx context.Context, bucket string) error | ||
|  | } | ||
|  | 
 | ||
|  | // CacheObjectLayer implements primitives for cache object API layer.
 | ||
|  | type CacheObjectLayer interface { | ||
|  | 	// Bucket operations.
 | ||
|  | 	ListObjects(ctx context.Context, bucket, prefix, marker, delimiter string, maxKeys int) (result ListObjectsInfo, err error) | ||
|  | 	ListObjectsV2(ctx context.Context, bucket, prefix, continuationToken, delimiter string, maxKeys int, fetchOwner bool, startAfter string) (result ListObjectsV2Info, err error) | ||
|  | 	GetBucketInfo(ctx context.Context, bucket string) (bucketInfo BucketInfo, err error) | ||
|  | 	ListBuckets(ctx context.Context) (buckets []BucketInfo, err error) | ||
|  | 	DeleteBucket(ctx context.Context, bucket string) error | ||
|  | 	// Object operations.
 | ||
|  | 	GetObject(ctx context.Context, bucket, object string, startOffset int64, length int64, writer io.Writer, etag string) (err error) | ||
|  | 	GetObjectInfo(ctx context.Context, bucket, object string) (objInfo ObjectInfo, err error) | ||
|  | 	PutObject(ctx context.Context, bucket, object string, data *hash.Reader, metadata map[string]string) (objInfo ObjectInfo, err error) | ||
|  | 	DeleteObject(ctx context.Context, bucket, object string) error | ||
|  | 
 | ||
|  | 	// Multipart operations.
 | ||
|  | 	NewMultipartUpload(ctx context.Context, bucket, object string, metadata map[string]string) (uploadID string, err error) | ||
|  | 	PutObjectPart(ctx context.Context, bucket, object, uploadID string, partID int, data *hash.Reader) (info PartInfo, err error) | ||
|  | 	AbortMultipartUpload(ctx context.Context, bucket, object, uploadID string) error | ||
|  | 	CompleteMultipartUpload(ctx context.Context, bucket, object, uploadID string, uploadedParts []CompletePart) (objInfo ObjectInfo, err error) | ||
|  | 
 | ||
|  | 	// Storage operations.
 | ||
|  | 	StorageInfo(ctx context.Context) StorageInfo | ||
|  | } | ||
|  | 
 | ||
|  | // backendDownError returns true if err is due to backend failure or faulty disk if in server mode
 | ||
|  | func backendDownError(err error) bool { | ||
|  | 	_, backendDown := errors2.Cause(err).(BackendDown) | ||
|  | 	return backendDown || errors2.IsErr(err, baseErrs...) | ||
|  | } | ||
|  | 
 | ||
|  | // get cache disk where object is currently cached for a GET operation. If object does not exist at that location,
 | ||
|  | // treat the list of cache drives as a circular buffer and walk through them starting at hash index
 | ||
|  | // until an online drive is found.If object is not found, fall back to the first online cache drive
 | ||
|  | // closest to the hash index, so that object can be recached.
 | ||
|  | func (c diskCache) getCachedFSLoc(ctx context.Context, bucket, object string) (*cacheFSObjects, error) { | ||
|  | 
 | ||
|  | 	index := c.hashIndex(bucket, object) | ||
|  | 	numDisks := len(c.cfs) | ||
|  | 	// save first online cache disk closest to the hint index
 | ||
|  | 	var firstOnlineDisk *cacheFSObjects | ||
|  | 	for k := 0; k < numDisks; k++ { | ||
|  | 		i := (index + k) % numDisks | ||
|  | 		if c.cfs[i] == nil { | ||
|  | 			continue | ||
|  | 		} | ||
|  | 		if c.cfs[i].IsOnline() { | ||
|  | 			if firstOnlineDisk == nil { | ||
|  | 				firstOnlineDisk = c.cfs[i] | ||
|  | 			} | ||
|  | 			if c.cfs[i].Exists(ctx, bucket, object) { | ||
|  | 				return c.cfs[i], nil | ||
|  | 			} | ||
|  | 		} | ||
|  | 	} | ||
|  | 
 | ||
|  | 	if firstOnlineDisk != nil { | ||
|  | 		return firstOnlineDisk, nil | ||
|  | 	} | ||
|  | 	return nil, errDiskNotFound | ||
|  | } | ||
|  | 
 | ||
|  | // choose a cache deterministically based on hash of bucket,object. The hash index is treated as
 | ||
|  | // a hint. In the event that the cache drive at hash index is offline, treat the list of cache drives
 | ||
|  | // as a circular buffer and walk through them starting at hash index until an online drive is found.
 | ||
|  | func (c diskCache) getCacheFS(ctx context.Context, bucket, object string) (*cacheFSObjects, error) { | ||
|  | 
 | ||
|  | 	index := c.hashIndex(bucket, object) | ||
|  | 	numDisks := len(c.cfs) | ||
|  | 	for k := 0; k < numDisks; k++ { | ||
|  | 		i := (index + k) % numDisks | ||
|  | 		if c.cfs[i] == nil { | ||
|  | 			continue | ||
|  | 		} | ||
|  | 		if c.cfs[i].IsOnline() { | ||
|  | 			return c.cfs[i], nil | ||
|  | 		} | ||
|  | 	} | ||
|  | 	return nil, errDiskNotFound | ||
|  | } | ||
|  | 
 | ||
|  | // Compute a unique hash sum for bucket and object
 | ||
|  | func (c diskCache) hashIndex(bucket, object string) int { | ||
|  | 	key := fmt.Sprintf("%x", sha256.Sum256([]byte(path.Join(bucket, object)))) | ||
|  | 	return int(crc32.Checksum([]byte(key), crc32.IEEETable)) % len(c.cfs) | ||
|  | } | ||
|  | 
 | ||
|  | // construct a metadata k-v map
 | ||
|  | func (c cacheObjects) getMetadata(objInfo ObjectInfo) map[string]string { | ||
|  | 	metadata := make(map[string]string) | ||
|  | 	metadata["etag"] = objInfo.ETag | ||
|  | 	metadata["content-type"] = objInfo.ContentType | ||
|  | 	metadata["content-encoding"] = objInfo.ContentEncoding | ||
|  | 
 | ||
|  | 	for key, val := range objInfo.UserDefined { | ||
|  | 		metadata[key] = val | ||
|  | 	} | ||
|  | 	return metadata | ||
|  | } | ||
|  | 
 | ||
|  | // Uses cached-object to serve the request. If object is not cached it serves the request from the backend and also
 | ||
|  | // stores it in the cache for serving subsequent requests.
 | ||
|  | func (c cacheObjects) GetObject(ctx context.Context, bucket, object string, startOffset int64, length int64, writer io.Writer, etag string) (err error) { | ||
|  | 	GetObjectFn := c.GetObjectFn | ||
|  | 	GetObjectInfoFn := c.GetObjectInfoFn | ||
|  | 
 | ||
|  | 	if c.isCacheExclude(bucket, object) { | ||
|  | 		return GetObjectFn(ctx, bucket, object, startOffset, length, writer, etag) | ||
|  | 	} | ||
|  | 	// fetch cacheFSObjects if object is currently cached or nearest available cache drive
 | ||
|  | 	dcache, err := c.cache.getCachedFSLoc(ctx, bucket, object) | ||
|  | 	if err != nil { | ||
|  | 		return GetObjectFn(ctx, bucket, object, startOffset, length, writer, etag) | ||
|  | 	} | ||
|  | 	// stat object on backend
 | ||
|  | 	objInfo, err := GetObjectInfoFn(ctx, bucket, object) | ||
|  | 	backendDown := backendDownError(err) | ||
|  | 	if err != nil && !backendDown { | ||
|  | 		if _, ok := errors2.Cause(err).(ObjectNotFound); ok { | ||
|  | 			// Delete the cached entry if backend object was deleted.
 | ||
|  | 			dcache.Delete(ctx, bucket, object) | ||
|  | 		} | ||
|  | 		return err | ||
|  | 	} | ||
|  | 
 | ||
|  | 	if !backendDown && filterFromCache(objInfo.UserDefined) { | ||
|  | 		return GetObjectFn(ctx, bucket, object, startOffset, length, writer, etag) | ||
|  | 	} | ||
|  | 
 | ||
|  | 	cachedObjInfo, err := dcache.GetObjectInfo(ctx, bucket, object) | ||
|  | 	if err == nil { | ||
|  | 		if backendDown { | ||
|  | 			// If the backend is down, serve the request from cache.
 | ||
|  | 			return dcache.Get(ctx, bucket, object, startOffset, length, writer, etag) | ||
|  | 		} | ||
|  | 		if cachedObjInfo.ETag == objInfo.ETag && !isStaleCache(objInfo) { | ||
|  | 			return dcache.Get(ctx, bucket, object, startOffset, length, writer, etag) | ||
|  | 		} | ||
|  | 		dcache.Delete(ctx, bucket, object) | ||
|  | 	} | ||
|  | 	if startOffset != 0 || length != objInfo.Size { | ||
|  | 		// We don't cache partial objects.
 | ||
|  | 		return GetObjectFn(ctx, bucket, object, startOffset, length, writer, etag) | ||
|  | 	} | ||
|  | 	if !dcache.diskAvailable(objInfo.Size * cacheSizeMultiplier) { | ||
|  | 		// cache only objects < 1/100th of disk capacity
 | ||
|  | 		return GetObjectFn(ctx, bucket, object, startOffset, length, writer, etag) | ||
|  | 	} | ||
|  | 	// Initialize pipe.
 | ||
|  | 	pipeReader, pipeWriter := io.Pipe() | ||
|  | 	hashReader, err := hash.NewReader(pipeReader, objInfo.Size, "", "") | ||
|  | 	if err != nil { | ||
|  | 		return err | ||
|  | 	} | ||
|  | 	go func() { | ||
|  | 		if err = GetObjectFn(ctx, bucket, object, 0, objInfo.Size, io.MultiWriter(writer, pipeWriter), etag); err != nil { | ||
|  | 			pipeWriter.CloseWithError(err) | ||
|  | 			return | ||
|  | 		} | ||
|  | 		pipeWriter.Close() // Close writer explicitly signalling we wrote all data.
 | ||
|  | 	}() | ||
|  | 	err = dcache.Put(ctx, bucket, object, hashReader, c.getMetadata(objInfo)) | ||
|  | 	if err != nil { | ||
|  | 		return err | ||
|  | 	} | ||
|  | 	pipeReader.Close() | ||
|  | 	return | ||
|  | } | ||
|  | 
 | ||
|  | // Returns ObjectInfo from cache if available.
 | ||
|  | func (c cacheObjects) GetObjectInfo(ctx context.Context, bucket, object string) (ObjectInfo, error) { | ||
|  | 	getObjectInfoFn := c.GetObjectInfoFn | ||
|  | 	if c.isCacheExclude(bucket, object) { | ||
|  | 		return getObjectInfoFn(ctx, bucket, object) | ||
|  | 	} | ||
|  | 	// fetch cacheFSObjects if object is currently cached or nearest available cache drive
 | ||
|  | 	dcache, err := c.cache.getCachedFSLoc(ctx, bucket, object) | ||
|  | 	if err != nil { | ||
|  | 		return getObjectInfoFn(ctx, bucket, object) | ||
|  | 	} | ||
|  | 	objInfo, err := getObjectInfoFn(ctx, bucket, object) | ||
|  | 	if err != nil { | ||
|  | 		if _, ok := errors2.Cause(err).(ObjectNotFound); ok { | ||
|  | 			// Delete the cached entry if backend object was deleted.
 | ||
|  | 			dcache.Delete(ctx, bucket, object) | ||
|  | 			return ObjectInfo{}, err | ||
|  | 		} | ||
|  | 		if !backendDownError(err) { | ||
|  | 			return ObjectInfo{}, err | ||
|  | 		} | ||
|  | 		// when backend is down, serve from cache.
 | ||
|  | 		cachedObjInfo, cerr := dcache.GetObjectInfo(ctx, bucket, object) | ||
|  | 		if cerr == nil { | ||
|  | 			return cachedObjInfo, nil | ||
|  | 		} | ||
|  | 		return ObjectInfo{}, BackendDown{} | ||
|  | 	} | ||
|  | 	// when backend is up, do a sanity check on cached object
 | ||
|  | 	cachedObjInfo, err := dcache.GetObjectInfo(ctx, bucket, object) | ||
|  | 	if err != nil { | ||
|  | 		return objInfo, nil | ||
|  | 	} | ||
|  | 	if cachedObjInfo.ETag != objInfo.ETag { | ||
|  | 		// Delete the cached entry if the backend object was replaced.
 | ||
|  | 		dcache.Delete(ctx, bucket, object) | ||
|  | 	} | ||
|  | 	return objInfo, nil | ||
|  | } | ||
|  | 
 | ||
|  | // Returns function "listDir" of the type listDirFunc.
 | ||
|  | // isLeaf - is used by listDir function to check if an entry is a leaf or non-leaf entry.
 | ||
|  | // disks - list of fsObjects
 | ||
|  | func listDirCacheFactory(isLeaf isLeafFunc, treeWalkIgnoredErrs []error, disks []*cacheFSObjects) listDirFunc { | ||
|  | 	listCacheDirs := func(bucket, prefixDir, prefixEntry string) (dirs []string, err error) { | ||
|  | 		var entries []string | ||
|  | 		for _, disk := range disks { | ||
|  | 			fs := disk.FSObjects | ||
|  | 			entries, err = readDir(pathJoin(fs.fsPath, bucket, prefixDir)) | ||
|  | 			if err != nil { | ||
|  | 				// For any reason disk was deleted or goes offline, continue
 | ||
|  | 				// and list from other disks if possible.
 | ||
|  | 				continue | ||
|  | 			} | ||
|  | 
 | ||
|  | 			// Filter entries that have the prefix prefixEntry.
 | ||
|  | 			entries = filterMatchingPrefix(entries, prefixEntry) | ||
|  | 			dirs = append(dirs, entries...) | ||
|  | 		} | ||
|  | 		return dirs, nil | ||
|  | 	} | ||
|  | 
 | ||
|  | 	// listDir - lists all the entries at a given prefix and given entry in the prefix.
 | ||
|  | 	listDir := func(bucket, prefixDir, prefixEntry string) (mergedEntries []string, delayIsLeaf bool, err error) { | ||
|  | 		var cacheEntries []string | ||
|  | 		cacheEntries, err = listCacheDirs(bucket, prefixDir, prefixEntry) | ||
|  | 		if err != nil { | ||
|  | 			return nil, false, err | ||
|  | 		} | ||
|  | 		for _, entry := range cacheEntries { | ||
|  | 			// Find elements in entries which are not in mergedEntries
 | ||
|  | 			idx := sort.SearchStrings(mergedEntries, entry) | ||
|  | 			// if entry is already present in mergedEntries don't add.
 | ||
|  | 			if idx < len(mergedEntries) && mergedEntries[idx] == entry { | ||
|  | 				continue | ||
|  | 			} | ||
|  | 			mergedEntries = append(mergedEntries, entry) | ||
|  | 			sort.Strings(mergedEntries) | ||
|  | 		} | ||
|  | 		return mergedEntries, false, nil | ||
|  | 	} | ||
|  | 	return listDir | ||
|  | } | ||
|  | 
 | ||
|  | // List all objects at prefix upto maxKeys, optionally delimited by '/' from the cache. Maintains the list pool
 | ||
|  | // state for future re-entrant list requests.
 | ||
|  | func (c cacheObjects) listCacheObjects(ctx context.Context, bucket, prefix, marker, delimiter string, maxKeys int) (result ListObjectsInfo, err error) { | ||
|  | 	var objInfos []ObjectInfo | ||
|  | 	var eof bool | ||
|  | 	var nextMarker string | ||
|  | 
 | ||
|  | 	recursive := true | ||
|  | 	if delimiter == slashSeparator { | ||
|  | 		recursive = false | ||
|  | 	} | ||
|  | 	walkResultCh, endWalkCh := c.listPool.Release(listParams{bucket, recursive, marker, prefix, false}) | ||
|  | 	if walkResultCh == nil { | ||
|  | 		endWalkCh = make(chan struct{}) | ||
|  | 		isLeaf := func(bucket, object string) bool { | ||
|  | 			fs, err := c.cache.getCacheFS(ctx, bucket, object) | ||
|  | 			if err != nil { | ||
|  | 				return false | ||
|  | 			} | ||
|  | 			_, err = fs.getObjectInfo(bucket, object) | ||
|  | 			return err == nil | ||
|  | 		} | ||
|  | 
 | ||
|  | 		listDir := listDirCacheFactory(isLeaf, cacheTreeWalkIgnoredErrs, c.cache.cfs) | ||
|  | 		walkResultCh = startTreeWalk(bucket, prefix, marker, recursive, listDir, isLeaf, endWalkCh) | ||
|  | 	} | ||
|  | 
 | ||
|  | 	for i := 0; i < maxKeys; { | ||
|  | 		walkResult, ok := <-walkResultCh | ||
|  | 		if !ok { | ||
|  | 			// Closed channel.
 | ||
|  | 			eof = true | ||
|  | 			break | ||
|  | 		} | ||
|  | 		// For any walk error return right away.
 | ||
|  | 		if walkResult.err != nil { | ||
|  | 			return result, toObjectErr(walkResult.err, bucket, prefix) | ||
|  | 		} | ||
|  | 
 | ||
|  | 		entry := walkResult.entry | ||
|  | 		var objInfo ObjectInfo | ||
|  | 		if hasSuffix(entry, slashSeparator) { | ||
|  | 			// Object name needs to be full path.
 | ||
|  | 			objInfo.Bucket = bucket | ||
|  | 			objInfo.Name = entry | ||
|  | 			objInfo.IsDir = true | ||
|  | 		} else { | ||
|  | 			// Set the Mode to a "regular" file.
 | ||
|  | 			var err error | ||
|  | 			fs, err := c.cache.getCacheFS(ctx, bucket, entry) | ||
|  | 			if err != nil { | ||
|  | 				// Ignore errFileNotFound
 | ||
|  | 				if errors2.Cause(err) == errFileNotFound { | ||
|  | 					continue | ||
|  | 				} | ||
|  | 				return result, toObjectErr(err, bucket, prefix) | ||
|  | 			} | ||
|  | 			objInfo, err = fs.getObjectInfo(bucket, entry) | ||
|  | 			if err != nil { | ||
|  | 				// Ignore errFileNotFound
 | ||
|  | 				if errors2.Cause(err) == errFileNotFound { | ||
|  | 					continue | ||
|  | 				} | ||
|  | 				return result, toObjectErr(err, bucket, prefix) | ||
|  | 			} | ||
|  | 		} | ||
|  | 		nextMarker = objInfo.Name | ||
|  | 		objInfos = append(objInfos, objInfo) | ||
|  | 		i++ | ||
|  | 		if walkResult.end { | ||
|  | 			eof = true | ||
|  | 			break | ||
|  | 		} | ||
|  | 	} | ||
|  | 
 | ||
|  | 	params := listParams{bucket, recursive, nextMarker, prefix, false} | ||
|  | 	if !eof { | ||
|  | 		c.listPool.Set(params, walkResultCh, endWalkCh) | ||
|  | 	} | ||
|  | 
 | ||
|  | 	result = ListObjectsInfo{IsTruncated: !eof} | ||
|  | 	for _, objInfo := range objInfos { | ||
|  | 		result.NextMarker = objInfo.Name | ||
|  | 		if objInfo.IsDir { | ||
|  | 			result.Prefixes = append(result.Prefixes, objInfo.Name) | ||
|  | 			continue | ||
|  | 		} | ||
|  | 		result.Objects = append(result.Objects, objInfo) | ||
|  | 	} | ||
|  | 	return result, nil | ||
|  | } | ||
|  | 
 | ||
|  | // listCacheV2Objects lists all blobs in bucket filtered by prefix from the cache
 | ||
|  | func (c cacheObjects) listCacheV2Objects(ctx context.Context, bucket, prefix, continuationToken, delimiter string, maxKeys int, fetchOwner bool, startAfter string) (result ListObjectsV2Info, err error) { | ||
|  | 	loi, err := c.listCacheObjects(ctx, bucket, prefix, continuationToken, delimiter, maxKeys) | ||
|  | 	if err != nil { | ||
|  | 		return result, err | ||
|  | 	} | ||
|  | 
 | ||
|  | 	listObjectsV2Info := ListObjectsV2Info{ | ||
|  | 		IsTruncated:           loi.IsTruncated, | ||
|  | 		ContinuationToken:     continuationToken, | ||
|  | 		NextContinuationToken: loi.NextMarker, | ||
|  | 		Objects:               loi.Objects, | ||
|  | 		Prefixes:              loi.Prefixes, | ||
|  | 	} | ||
|  | 	return listObjectsV2Info, err | ||
|  | } | ||
|  | 
 | ||
|  | // List all objects at prefix upto maxKeys., optionally delimited by '/'. Maintains the list pool
 | ||
|  | // state for future re-entrant list requests. Retrieve from cache if backend is down
 | ||
|  | func (c cacheObjects) ListObjects(ctx context.Context, bucket, prefix, marker, delimiter string, maxKeys int) (result ListObjectsInfo, err error) { | ||
|  | 
 | ||
|  | 	listObjectsFn := c.ListObjectsFn | ||
|  | 
 | ||
|  | 	result, err = listObjectsFn(ctx, bucket, prefix, marker, delimiter, maxKeys) | ||
|  | 	if err != nil { | ||
|  | 		if backendDownError(err) { | ||
|  | 			return c.listCacheObjects(ctx, bucket, prefix, marker, delimiter, maxKeys) | ||
|  | 		} | ||
|  | 		return | ||
|  | 	} | ||
|  | 	return | ||
|  | } | ||
|  | 
 | ||
|  | // ListObjectsV2 lists all blobs in bucket filtered by prefix
 | ||
|  | func (c cacheObjects) ListObjectsV2(ctx context.Context, bucket, prefix, continuationToken, delimiter string, maxKeys int, fetchOwner bool, startAfter string) (result ListObjectsV2Info, err error) { | ||
|  | 	listObjectsV2Fn := c.ListObjectsV2Fn | ||
|  | 
 | ||
|  | 	result, err = listObjectsV2Fn(ctx, bucket, prefix, continuationToken, delimiter, maxKeys, fetchOwner, startAfter) | ||
|  | 	if err != nil { | ||
|  | 		if backendDownError(err) { | ||
|  | 			return c.listCacheV2Objects(ctx, bucket, prefix, continuationToken, delimiter, maxKeys, fetchOwner, startAfter) | ||
|  | 		} | ||
|  | 		return | ||
|  | 	} | ||
|  | 	return | ||
|  | } | ||
|  | 
 | ||
|  | // Lists all the buckets in the cache
 | ||
|  | func (c cacheObjects) listBuckets(ctx context.Context) (buckets []BucketInfo, err error) { | ||
|  | 	m := make(map[string]string) | ||
|  | 	for _, cache := range c.cache.cfs { | ||
|  | 		entries, err := cache.ListBuckets(ctx) | ||
|  | 
 | ||
|  | 		if err != nil { | ||
|  | 			return nil, err | ||
|  | 		} | ||
|  | 		for _, entry := range entries { | ||
|  | 			_, ok := m[entry.Name] | ||
|  | 			if !ok { | ||
|  | 				m[entry.Name] = entry.Name | ||
|  | 				buckets = append(buckets, entry) | ||
|  | 			} | ||
|  | 		} | ||
|  | 	} | ||
|  | 	// Sort bucket infos by bucket name.
 | ||
|  | 	sort.Sort(byBucketName(buckets)) | ||
|  | 	return | ||
|  | } | ||
|  | 
 | ||
|  | // Returns list of buckets from cache or the backend. If the backend is down, buckets
 | ||
|  | // available on cache are served.
 | ||
|  | func (c cacheObjects) ListBuckets(ctx context.Context) (buckets []BucketInfo, err error) { | ||
|  | 
 | ||
|  | 	listBucketsFn := c.ListBucketsFn | ||
|  | 	buckets, err = listBucketsFn(ctx) | ||
|  | 	if err != nil { | ||
|  | 		if backendDownError(err) { | ||
|  | 			return c.listBuckets(ctx) | ||
|  | 		} | ||
|  | 		return []BucketInfo{}, err | ||
|  | 	} | ||
|  | 	return | ||
|  | } | ||
|  | 
 | ||
|  | // Returns bucket info from cache if backend is down.
 | ||
|  | func (c cacheObjects) GetBucketInfo(ctx context.Context, bucket string) (bucketInfo BucketInfo, err error) { | ||
|  | 
 | ||
|  | 	getBucketInfoFn := c.GetBucketInfoFn | ||
|  | 	bucketInfo, err = getBucketInfoFn(ctx, bucket) | ||
|  | 	if backendDownError(err) { | ||
|  | 		for _, cache := range c.cache.cfs { | ||
|  | 			if bucketInfo, err = cache.GetBucketInfo(ctx, bucket); err == nil { | ||
|  | 				return | ||
|  | 			} | ||
|  | 		} | ||
|  | 	} | ||
|  | 	return | ||
|  | } | ||
|  | 
 | ||
|  | // Delete Object deletes from cache as well if backend operation succeeds
 | ||
|  | func (c cacheObjects) DeleteObject(ctx context.Context, bucket, object string) (err error) { | ||
|  | 
 | ||
|  | 	if err = c.DeleteObjectFn(ctx, bucket, object); err != nil { | ||
|  | 		return | ||
|  | 	} | ||
|  | 	if c.isCacheExclude(bucket, object) { | ||
|  | 		return | ||
|  | 	} | ||
|  | 	dcache, cerr := c.cache.getCachedFSLoc(ctx, bucket, object) | ||
|  | 	if cerr == nil { | ||
|  | 		_ = dcache.DeleteObject(ctx, bucket, object) | ||
|  | 	} | ||
|  | 	return | ||
|  | } | ||
|  | 
 | ||
|  | // Returns true if object should be excluded from cache
 | ||
|  | func (c cacheObjects) isCacheExclude(bucket, object string) bool { | ||
|  | 	for _, pattern := range c.exclude { | ||
|  | 		matchStr := fmt.Sprintf("%s/%s", bucket, object) | ||
|  | 		if ok := wildcard.MatchSimple(pattern, matchStr); ok { | ||
|  | 			return true | ||
|  | 		} | ||
|  | 	} | ||
|  | 	return false | ||
|  | } | ||
|  | 
 | ||
|  | // PutObject - caches the uploaded object for single Put operations
 | ||
|  | func (c cacheObjects) PutObject(ctx context.Context, bucket, object string, r *hash.Reader, metadata map[string]string) (objInfo ObjectInfo, err error) { | ||
|  | 	putObjectFn := c.PutObjectFn | ||
|  | 	dcache, err := c.cache.getCacheFS(ctx, bucket, object) | ||
|  | 	if err != nil { | ||
|  | 		// disk cache could not be located,execute backend call.
 | ||
|  | 		return putObjectFn(ctx, bucket, object, r, metadata) | ||
|  | 	} | ||
|  | 	size := r.Size() | ||
|  | 
 | ||
|  | 	// fetch from backend if there is no space on cache drive
 | ||
|  | 	if !dcache.diskAvailable(size * cacheSizeMultiplier) { | ||
|  | 		return putObjectFn(ctx, bucket, object, r, metadata) | ||
|  | 	} | ||
|  | 	// fetch from backend if cache exclude pattern or cache-control
 | ||
|  | 	// directive set to exclude
 | ||
|  | 	if c.isCacheExclude(bucket, object) || filterFromCache(metadata) { | ||
|  | 		dcache.Delete(ctx, bucket, object) | ||
|  | 		return putObjectFn(ctx, bucket, object, r, metadata) | ||
|  | 	} | ||
|  | 	objInfo = ObjectInfo{} | ||
|  | 	// Initialize pipe to stream data to backend
 | ||
|  | 	pipeReader, pipeWriter := io.Pipe() | ||
|  | 	hashReader, err := hash.NewReader(pipeReader, size, r.MD5HexString(), r.SHA256HexString()) | ||
|  | 	if err != nil { | ||
|  | 		return ObjectInfo{}, err | ||
|  | 	} | ||
|  | 	// Initialize pipe to stream data to cache
 | ||
|  | 	rPipe, wPipe := io.Pipe() | ||
|  | 	cHashReader, err := hash.NewReader(rPipe, size, r.MD5HexString(), r.SHA256HexString()) | ||
|  | 	if err != nil { | ||
|  | 		return ObjectInfo{}, err | ||
|  | 	} | ||
|  | 	oinfoCh := make(chan ObjectInfo) | ||
|  | 	errCh := make(chan error) | ||
|  | 	go func() { | ||
|  | 		oinfo, perr := putObjectFn(ctx, bucket, object, hashReader, metadata) | ||
|  | 		if perr != nil { | ||
|  | 			pipeWriter.CloseWithError(perr) | ||
|  | 			wPipe.CloseWithError(perr) | ||
|  | 			close(oinfoCh) | ||
|  | 			errCh <- perr | ||
|  | 			return | ||
|  | 		} | ||
|  | 		close(errCh) | ||
|  | 		oinfoCh <- oinfo | ||
|  | 	}() | ||
|  | 
 | ||
|  | 	go func() { | ||
|  | 		if err = dcache.Put(ctx, bucket, object, cHashReader, metadata); err != nil { | ||
|  | 			wPipe.CloseWithError(err) | ||
|  | 			return | ||
|  | 		} | ||
|  | 	}() | ||
|  | 
 | ||
|  | 	mwriter := io.MultiWriter(pipeWriter, wPipe) | ||
|  | 	_, err = io.Copy(mwriter, r) | ||
|  | 	if err != nil { | ||
|  | 		err = <-errCh | ||
|  | 		return objInfo, err | ||
|  | 	} | ||
|  | 	pipeWriter.Close() | ||
|  | 	wPipe.Close() | ||
|  | 	objInfo = <-oinfoCh | ||
|  | 	return objInfo, err | ||
|  | } | ||
|  | 
 | ||
|  | // NewMultipartUpload - Starts a new multipart upload operation to backend and cache.
 | ||
|  | func (c cacheObjects) NewMultipartUpload(ctx context.Context, bucket, object string, metadata map[string]string) (uploadID string, err error) { | ||
|  | 
 | ||
|  | 	newMultipartUploadFn := c.NewMultipartUploadFn | ||
|  | 
 | ||
|  | 	if c.isCacheExclude(bucket, object) || filterFromCache(metadata) { | ||
|  | 		return newMultipartUploadFn(ctx, bucket, object, metadata) | ||
|  | 	} | ||
|  | 
 | ||
|  | 	dcache, err := c.cache.getCacheFS(ctx, bucket, object) | ||
|  | 	if err != nil { | ||
|  | 		// disk cache could not be located,execute backend call.
 | ||
|  | 		return newMultipartUploadFn(ctx, bucket, object, metadata) | ||
|  | 	} | ||
|  | 
 | ||
|  | 	uploadID, err = newMultipartUploadFn(ctx, bucket, object, metadata) | ||
|  | 	if err != nil { | ||
|  | 		return | ||
|  | 	} | ||
|  | 	// create new multipart upload in cache with same uploadID
 | ||
|  | 	dcache.NewMultipartUpload(ctx, bucket, object, metadata, uploadID) | ||
|  | 	return uploadID, err | ||
|  | } | ||
|  | 
 | ||
|  | // PutObjectPart - uploads part to backend and cache simultaneously.
 | ||
|  | func (c cacheObjects) PutObjectPart(ctx context.Context, bucket, object, uploadID string, partID int, data *hash.Reader) (info PartInfo, err error) { | ||
|  | 
 | ||
|  | 	putObjectPartFn := c.PutObjectPartFn | ||
|  | 	dcache, err := c.cache.getCacheFS(ctx, bucket, object) | ||
|  | 	if err != nil { | ||
|  | 		// disk cache could not be located,execute backend call.
 | ||
|  | 		return putObjectPartFn(ctx, bucket, object, uploadID, partID, data) | ||
|  | 	} | ||
|  | 
 | ||
|  | 	if c.isCacheExclude(bucket, object) { | ||
|  | 		return putObjectPartFn(ctx, bucket, object, uploadID, partID, data) | ||
|  | 	} | ||
|  | 
 | ||
|  | 	// make sure cache has at least cacheSizeMultiplier * size available
 | ||
|  | 	size := data.Size() | ||
|  | 	if !dcache.diskAvailable(size * cacheSizeMultiplier) { | ||
|  | 		select { | ||
|  | 		case dcache.purgeChan <- struct{}{}: | ||
|  | 		default: | ||
|  | 		} | ||
|  | 		return putObjectPartFn(ctx, bucket, object, uploadID, partID, data) | ||
|  | 	} | ||
|  | 
 | ||
|  | 	info = PartInfo{} | ||
|  | 	// Initialize pipe to stream data to backend
 | ||
|  | 	pipeReader, pipeWriter := io.Pipe() | ||
|  | 	hashReader, err := hash.NewReader(pipeReader, size, data.MD5HexString(), data.SHA256HexString()) | ||
|  | 	if err != nil { | ||
|  | 		return | ||
|  | 	} | ||
|  | 	// Initialize pipe to stream data to cache
 | ||
|  | 	rPipe, wPipe := io.Pipe() | ||
|  | 	cHashReader, err := hash.NewReader(rPipe, size, data.MD5HexString(), data.SHA256HexString()) | ||
|  | 	if err != nil { | ||
|  | 		return | ||
|  | 	} | ||
|  | 	pinfoCh := make(chan PartInfo) | ||
|  | 	errorCh := make(chan error) | ||
|  | 	go func() { | ||
|  | 		info, err = putObjectPartFn(ctx, bucket, object, uploadID, partID, hashReader) | ||
|  | 		if err != nil { | ||
|  | 			close(pinfoCh) | ||
|  | 			pipeWriter.CloseWithError(err) | ||
|  | 			wPipe.CloseWithError(err) | ||
|  | 			errorCh <- err | ||
|  | 			return | ||
|  | 		} | ||
|  | 		close(errorCh) | ||
|  | 		pinfoCh <- info | ||
|  | 	}() | ||
|  | 	go func() { | ||
|  | 		if _, perr := dcache.PutObjectPart(ctx, bucket, object, uploadID, partID, cHashReader); perr != nil { | ||
|  | 			wPipe.CloseWithError(perr) | ||
|  | 			return | ||
|  | 		} | ||
|  | 	}() | ||
|  | 
 | ||
|  | 	mwriter := io.MultiWriter(pipeWriter, wPipe) | ||
|  | 	_, err = io.Copy(mwriter, data) | ||
|  | 	if err != nil { | ||
|  | 		err = <-errorCh | ||
|  | 		return PartInfo{}, err | ||
|  | 	} | ||
|  | 	pipeWriter.Close() | ||
|  | 	wPipe.Close() | ||
|  | 	info = <-pinfoCh | ||
|  | 	return info, err | ||
|  | } | ||
|  | 
 | ||
|  | // AbortMultipartUpload - aborts multipart upload on backend and cache.
 | ||
|  | func (c cacheObjects) AbortMultipartUpload(ctx context.Context, bucket, object, uploadID string) error { | ||
|  | 
 | ||
|  | 	abortMultipartUploadFn := c.AbortMultipartUploadFn | ||
|  | 
 | ||
|  | 	if c.isCacheExclude(bucket, object) { | ||
|  | 		return abortMultipartUploadFn(ctx, bucket, object, uploadID) | ||
|  | 	} | ||
|  | 
 | ||
|  | 	dcache, err := c.cache.getCacheFS(ctx, bucket, object) | ||
|  | 	if err != nil { | ||
|  | 		// disk cache could not be located,execute backend call.
 | ||
|  | 		return abortMultipartUploadFn(ctx, bucket, object, uploadID) | ||
|  | 	} | ||
|  | 	// execute backend operation
 | ||
|  | 	err = abortMultipartUploadFn(ctx, bucket, object, uploadID) | ||
|  | 	if err != nil { | ||
|  | 		return err | ||
|  | 	} | ||
|  | 	// abort multipart upload on cache
 | ||
|  | 	dcache.AbortMultipartUpload(ctx, bucket, object, uploadID) | ||
|  | 	return nil | ||
|  | } | ||
|  | 
 | ||
|  | // CompleteMultipartUpload - completes multipart upload operation on backend and cache.
 | ||
|  | func (c cacheObjects) CompleteMultipartUpload(ctx context.Context, bucket, object, uploadID string, uploadedParts []CompletePart) (objInfo ObjectInfo, err error) { | ||
|  | 
 | ||
|  | 	completeMultipartUploadFn := c.CompleteMultipartUploadFn | ||
|  | 
 | ||
|  | 	if c.isCacheExclude(bucket, object) { | ||
|  | 		return completeMultipartUploadFn(ctx, bucket, object, uploadID, uploadedParts) | ||
|  | 	} | ||
|  | 
 | ||
|  | 	dcache, err := c.cache.getCacheFS(ctx, bucket, object) | ||
|  | 	if err != nil { | ||
|  | 		// disk cache could not be located,execute backend call.
 | ||
|  | 		return completeMultipartUploadFn(ctx, bucket, object, uploadID, uploadedParts) | ||
|  | 	} | ||
|  | 	// perform backend operation
 | ||
|  | 	objInfo, err = completeMultipartUploadFn(ctx, bucket, object, uploadID, uploadedParts) | ||
|  | 	if err != nil { | ||
|  | 		return | ||
|  | 	} | ||
|  | 	// create new multipart upload in cache with same uploadID
 | ||
|  | 	dcache.CompleteMultipartUpload(ctx, bucket, object, uploadID, uploadedParts) | ||
|  | 	return | ||
|  | } | ||
|  | 
 | ||
|  | // StorageInfo - returns underlying storage statistics.
 | ||
|  | func (c cacheObjects) StorageInfo(ctx context.Context) (storageInfo StorageInfo) { | ||
|  | 	var total, free uint64 | ||
|  | 	for _, cfs := range c.cache.cfs { | ||
|  | 		if cfs == nil { | ||
|  | 			continue | ||
|  | 		} | ||
|  | 		info, err := getDiskInfo((cfs.fsPath)) | ||
|  | 		errorIf(err, "Unable to get disk info %#v", cfs.fsPath) | ||
|  | 		total += info.Total | ||
|  | 		free += info.Free | ||
|  | 	} | ||
|  | 	storageInfo = StorageInfo{ | ||
|  | 		Total: total, | ||
|  | 		Free:  free, | ||
|  | 	} | ||
|  | 	storageInfo.Backend.Type = FS | ||
|  | 	return storageInfo | ||
|  | } | ||
|  | 
 | ||
|  | // DeleteBucket - marks bucket to be deleted from cache if bucket is deleted from backend.
 | ||
|  | func (c cacheObjects) DeleteBucket(ctx context.Context, bucket string) (err error) { | ||
|  | 	deleteBucketFn := c.DeleteBucketFn | ||
|  | 	var toDel []*cacheFSObjects | ||
|  | 	for _, cfs := range c.cache.cfs { | ||
|  | 		if _, cerr := cfs.GetBucketInfo(ctx, bucket); cerr == nil { | ||
|  | 			toDel = append(toDel, cfs) | ||
|  | 		} | ||
|  | 	} | ||
|  | 	// perform backend operation
 | ||
|  | 	err = deleteBucketFn(ctx, bucket) | ||
|  | 	if err != nil { | ||
|  | 		return | ||
|  | 	} | ||
|  | 	// move bucket metadata and content to cache's trash dir
 | ||
|  | 	for _, d := range toDel { | ||
|  | 		d.moveBucketToTrash(ctx, bucket) | ||
|  | 	} | ||
|  | 	return | ||
|  | } | ||
|  | 
 | ||
|  | // newCache initializes the cacheFSObjects for the "drives" specified in config.json
 | ||
|  | // or the global env overrides.
 | ||
|  | func newCache(c CacheConfig) (*diskCache, error) { | ||
|  | 	var cfsObjects []*cacheFSObjects | ||
|  | 	formats, err := loadAndValidateCacheFormat(c.Drives) | ||
|  | 	if err != nil { | ||
|  | 		errorIf(err, "Cache drives validation error") | ||
|  | 	} | ||
|  | 	if len(formats) == 0 { | ||
|  | 		return nil, errors.New("Cache drives validation error") | ||
|  | 	} | ||
|  | 	for i, dir := range c.Drives { | ||
|  | 		// skip cacheFSObjects creation for cache drives missing a format.json
 | ||
|  | 		if formats[i] == nil { | ||
|  | 			cfsObjects = append(cfsObjects, nil) | ||
|  | 			continue | ||
|  | 		} | ||
|  | 		c, err := newCacheFSObjects(dir, c.Expiry, cacheMaxDiskUsagePct) | ||
|  | 		if err != nil { | ||
|  | 			return nil, err | ||
|  | 		} | ||
|  | 		if err := checkAtimeSupport(dir); err != nil { | ||
|  | 			return nil, errors.New("Atime support required for disk caching") | ||
|  | 		} | ||
|  | 		// Start the purging go-routine for entries that have expired
 | ||
|  | 		go c.purge() | ||
|  | 		// Start trash purge routine for deleted buckets.
 | ||
|  | 		go c.purgeTrash() | ||
|  | 		cfsObjects = append(cfsObjects, c) | ||
|  | 	} | ||
|  | 	return &diskCache{cfs: cfsObjects}, nil | ||
|  | } | ||
|  | 
 | ||
|  | // Return error if Atime is disabled on the O/S
 | ||
|  | func checkAtimeSupport(dir string) (err error) { | ||
|  | 	file, err := ioutil.TempFile(dir, "prefix") | ||
|  | 	if err != nil { | ||
|  | 		return | ||
|  | 	} | ||
|  | 	defer os.Remove(file.Name()) | ||
|  | 	finfo1, err := os.Stat(file.Name()) | ||
|  | 	if err != nil { | ||
|  | 		return | ||
|  | 	} | ||
|  | 	if _, err = io.Copy(ioutil.Discard, file); err != io.EOF { | ||
|  | 		return | ||
|  | 	} | ||
|  | 
 | ||
|  | 	finfo2, err := os.Stat(file.Name()) | ||
|  | 
 | ||
|  | 	if atime.Get(finfo2).Equal(atime.Get(finfo1)) { | ||
|  | 		return errors.New("Atime not supported") | ||
|  | 	} | ||
|  | 	return | ||
|  | } | ||
|  | 
 | ||
|  | // Returns cacheObjects for use by Server.
 | ||
|  | func newServerCacheObjects(c CacheConfig) (CacheObjectLayer, error) { | ||
|  | 
 | ||
|  | 	// list of disk caches for cache "drives" specified in config.json or MINIO_CACHE_DRIVES env var.
 | ||
|  | 	dcache, err := newCache(c) | ||
|  | 	if err != nil { | ||
|  | 		return nil, err | ||
|  | 	} | ||
|  | 
 | ||
|  | 	return &cacheObjects{ | ||
|  | 		cache:    dcache, | ||
|  | 		exclude:  c.Exclude, | ||
|  | 		listPool: newTreeWalkPool(globalLookupTimeout), | ||
|  | 		GetObjectFn: func(ctx context.Context, bucket, object string, startOffset int64, length int64, writer io.Writer, etag string) error { | ||
|  | 			return newObjectLayerFn().GetObject(ctx, bucket, object, startOffset, length, writer, etag) | ||
|  | 		}, | ||
|  | 		GetObjectInfoFn: func(ctx context.Context, bucket, object string) (ObjectInfo, error) { | ||
|  | 			return newObjectLayerFn().GetObjectInfo(ctx, bucket, object) | ||
|  | 		}, | ||
|  | 		PutObjectFn: func(ctx context.Context, bucket, object string, data *hash.Reader, metadata map[string]string) (objInfo ObjectInfo, err error) { | ||
|  | 			return newObjectLayerFn().PutObject(ctx, bucket, object, data, metadata) | ||
|  | 		}, | ||
|  | 		DeleteObjectFn: func(ctx context.Context, bucket, object string) error { | ||
|  | 			return newObjectLayerFn().DeleteObject(ctx, bucket, object) | ||
|  | 		}, | ||
|  | 		ListObjectsFn: func(ctx context.Context, bucket, prefix, marker, delimiter string, maxKeys int) (result ListObjectsInfo, err error) { | ||
|  | 			return newObjectLayerFn().ListObjects(ctx, bucket, prefix, marker, delimiter, maxKeys) | ||
|  | 		}, | ||
|  | 		ListObjectsV2Fn: func(ctx context.Context, bucket, prefix, continuationToken, delimiter string, maxKeys int, fetchOwner bool, startAfter string) (result ListObjectsV2Info, err error) { | ||
|  | 			return newObjectLayerFn().ListObjectsV2(ctx, bucket, prefix, continuationToken, delimiter, maxKeys, fetchOwner, startAfter) | ||
|  | 		}, | ||
|  | 		ListBucketsFn: func(ctx context.Context) (buckets []BucketInfo, err error) { | ||
|  | 			return newObjectLayerFn().ListBuckets(ctx) | ||
|  | 		}, | ||
|  | 		GetBucketInfoFn: func(ctx context.Context, bucket string) (bucketInfo BucketInfo, err error) { | ||
|  | 			return newObjectLayerFn().GetBucketInfo(ctx, bucket) | ||
|  | 		}, | ||
|  | 		NewMultipartUploadFn: func(ctx context.Context, bucket, object string, metadata map[string]string) (uploadID string, err error) { | ||
|  | 			return newObjectLayerFn().NewMultipartUpload(ctx, bucket, object, metadata) | ||
|  | 		}, | ||
|  | 		PutObjectPartFn: func(ctx context.Context, bucket, object, uploadID string, partID int, data *hash.Reader) (info PartInfo, err error) { | ||
|  | 			return newObjectLayerFn().PutObjectPart(ctx, bucket, object, uploadID, partID, data) | ||
|  | 		}, | ||
|  | 		AbortMultipartUploadFn: func(ctx context.Context, bucket, object, uploadID string) error { | ||
|  | 			return newObjectLayerFn().AbortMultipartUpload(ctx, bucket, object, uploadID) | ||
|  | 		}, | ||
|  | 		CompleteMultipartUploadFn: func(ctx context.Context, bucket, object, uploadID string, uploadedParts []CompletePart) (objInfo ObjectInfo, err error) { | ||
|  | 			return newObjectLayerFn().CompleteMultipartUpload(ctx, bucket, object, uploadID, uploadedParts) | ||
|  | 		}, | ||
|  | 		DeleteBucketFn: func(ctx context.Context, bucket string) error { | ||
|  | 			return newObjectLayerFn().DeleteBucket(ctx, bucket) | ||
|  | 		}, | ||
|  | 	}, nil | ||
|  | } | ||
|  | 
 | ||
|  | type cacheControl struct { | ||
|  | 	exclude  bool | ||
|  | 	expiry   time.Time | ||
|  | 	maxAge   int | ||
|  | 	sMaxAge  int | ||
|  | 	minFresh int | ||
|  | } | ||
|  | 
 | ||
|  | // cache exclude directives in cache-control header
 | ||
|  | var cacheExcludeDirectives = []string{ | ||
|  | 	"no-cache", | ||
|  | 	"no-store", | ||
|  | 	"must-revalidate", | ||
|  | } | ||
|  | 
 | ||
|  | // returns true if cache exclude directives are set.
 | ||
|  | func isCacheExcludeDirective(s string) bool { | ||
|  | 	for _, directive := range cacheExcludeDirectives { | ||
|  | 		if s == directive { | ||
|  | 			return true | ||
|  | 		} | ||
|  | 	} | ||
|  | 	return false | ||
|  | } | ||
|  | 
 | ||
|  | // returns struct with cache-control settings from user metadata.
 | ||
|  | func getCacheControlOpts(m map[string]string) (c cacheControl, err error) { | ||
|  | 	var headerVal string | ||
|  | 	for k, v := range m { | ||
|  | 		if k == "cache-control" { | ||
|  | 			headerVal = v | ||
|  | 		} | ||
|  | 		if k == "expires" { | ||
|  | 			if e, err := http.ParseTime(v); err == nil { | ||
|  | 				c.expiry = e | ||
|  | 			} | ||
|  | 		} | ||
|  | 	} | ||
|  | 	if headerVal == "" { | ||
|  | 		return | ||
|  | 	} | ||
|  | 	headerVal = strings.ToLower(headerVal) | ||
|  | 	headerVal = strings.TrimSpace(headerVal) | ||
|  | 
 | ||
|  | 	vals := strings.Split(headerVal, ",") | ||
|  | 	for _, val := range vals { | ||
|  | 		val = strings.TrimSpace(val) | ||
|  | 		p := strings.Split(val, "=") | ||
|  | 		if isCacheExcludeDirective(p[0]) { | ||
|  | 			c.exclude = true | ||
|  | 			continue | ||
|  | 		} | ||
|  | 
 | ||
|  | 		if len(p) != 2 { | ||
|  | 			continue | ||
|  | 		} | ||
|  | 		if p[0] == "max-age" || | ||
|  | 			p[0] == "s-maxage" || | ||
|  | 			p[0] == "min-fresh" { | ||
|  | 			i, err := strconv.Atoi(p[1]) | ||
|  | 			if err != nil { | ||
|  | 				return c, err | ||
|  | 			} | ||
|  | 			if p[0] == "max-age" { | ||
|  | 				c.maxAge = i | ||
|  | 			} | ||
|  | 			if p[0] == "s-maxage" { | ||
|  | 				c.sMaxAge = i | ||
|  | 			} | ||
|  | 			if p[0] == "min-fresh" { | ||
|  | 				c.minFresh = i | ||
|  | 			} | ||
|  | 		} | ||
|  | 	} | ||
|  | 	return c, nil | ||
|  | } | ||
|  | 
 | ||
|  | // return true if metadata has a cache-control header
 | ||
|  | // directive to exclude object from cache.
 | ||
|  | func filterFromCache(m map[string]string) bool { | ||
|  | 	c, err := getCacheControlOpts(m) | ||
|  | 	if err != nil { | ||
|  | 		return false | ||
|  | 	} | ||
|  | 	return c.exclude | ||
|  | } | ||
|  | 
 | ||
|  | // returns true if cache expiry conditions met in cache-control/expiry metadata.
 | ||
|  | func isStaleCache(objInfo ObjectInfo) bool { | ||
|  | 	c, err := getCacheControlOpts(objInfo.UserDefined) | ||
|  | 	if err != nil { | ||
|  | 		return false | ||
|  | 	} | ||
|  | 	now := time.Now() | ||
|  | 	if c.sMaxAge > 0 && c.sMaxAge > int(now.Sub(objInfo.ModTime).Seconds()) { | ||
|  | 		return true | ||
|  | 	} | ||
|  | 	if c.maxAge > 0 && c.maxAge > int(now.Sub(objInfo.ModTime).Seconds()) { | ||
|  | 		return true | ||
|  | 	} | ||
|  | 	if !c.expiry.Equal(time.Time{}) && c.expiry.Before(time.Now()) { | ||
|  | 		return true | ||
|  | 	} | ||
|  | 	if c.minFresh > 0 && c.minFresh <= int(now.Sub(objInfo.ModTime).Seconds()) { | ||
|  | 		return true | ||
|  | 	} | ||
|  | 	return false | ||
|  | } |