mirror of https://github.com/minio/minio.git
				
				
				
			
		
			
				
	
	
		
			2385 lines
		
	
	
		
			76 KiB
		
	
	
	
		
			Go
		
	
	
	
			
		
		
	
	
			2385 lines
		
	
	
		
			76 KiB
		
	
	
	
		
			Go
		
	
	
	
| // Copyright (c) 2015-2021 MinIO, Inc.
 | |
| //
 | |
| // This file is part of MinIO Object Storage stack
 | |
| //
 | |
| // This program is free software: you can redistribute it and/or modify
 | |
| // it under the terms of the GNU Affero General Public License as published by
 | |
| // the Free Software Foundation, either version 3 of the License, or
 | |
| // (at your option) any later version.
 | |
| //
 | |
| // This program is distributed in the hope that it will be useful
 | |
| // but WITHOUT ANY WARRANTY; without even the implied warranty of
 | |
| // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 | |
| // GNU Affero General Public License for more details.
 | |
| //
 | |
| // You should have received a copy of the GNU Affero General Public License
 | |
| // along with this program.  If not, see <http://www.gnu.org/licenses/>.
 | |
| 
 | |
| package cmd
 | |
| 
 | |
| import (
 | |
| 	"archive/zip"
 | |
| 	"bufio"
 | |
| 	"bytes"
 | |
| 	"context"
 | |
| 	"crypto/ecdsa"
 | |
| 	"crypto/hmac"
 | |
| 	crand "crypto/rand"
 | |
| 	"crypto/rsa"
 | |
| 	"crypto/sha1"
 | |
| 	"crypto/tls"
 | |
| 	"crypto/x509"
 | |
| 	"crypto/x509/pkix"
 | |
| 	"encoding/base64"
 | |
| 	"encoding/hex"
 | |
| 	"encoding/pem"
 | |
| 	"encoding/xml"
 | |
| 	"errors"
 | |
| 	"flag"
 | |
| 	"fmt"
 | |
| 	"io"
 | |
| 	"math/big"
 | |
| 	"math/rand"
 | |
| 	"net"
 | |
| 	"net/http"
 | |
| 	"net/http/httptest"
 | |
| 	"net/url"
 | |
| 	"os"
 | |
| 	"path"
 | |
| 	"path/filepath"
 | |
| 	"reflect"
 | |
| 	"sort"
 | |
| 	"strconv"
 | |
| 	"strings"
 | |
| 	"sync"
 | |
| 	"testing"
 | |
| 	"time"
 | |
| 	"unsafe"
 | |
| 
 | |
| 	"github.com/fatih/color"
 | |
| 
 | |
| 	"github.com/gorilla/mux"
 | |
| 	"github.com/minio/minio-go/v7/pkg/s3utils"
 | |
| 	"github.com/minio/minio-go/v7/pkg/signer"
 | |
| 	"github.com/minio/minio/internal/auth"
 | |
| 	"github.com/minio/minio/internal/config"
 | |
| 	"github.com/minio/minio/internal/crypto"
 | |
| 	"github.com/minio/minio/internal/hash"
 | |
| 	"github.com/minio/minio/internal/logger"
 | |
| 	"github.com/minio/minio/internal/rest"
 | |
| 	"github.com/minio/pkg/bucket/policy"
 | |
| )
 | |
| 
 | |
| // TestMain to set up global env.
 | |
| func TestMain(m *testing.M) {
 | |
| 	flag.Parse()
 | |
| 
 | |
| 	// set to 'true' when testing is invoked
 | |
| 	globalIsTesting = true
 | |
| 
 | |
| 	globalIsCICD = globalIsTesting
 | |
| 
 | |
| 	globalActiveCred = auth.Credentials{
 | |
| 		AccessKey: auth.DefaultAccessKey,
 | |
| 		SecretKey: auth.DefaultSecretKey,
 | |
| 	}
 | |
| 
 | |
| 	// disable ENVs which interfere with tests.
 | |
| 	for _, env := range []string{
 | |
| 		crypto.EnvKMSAutoEncryption,
 | |
| 		config.EnvAccessKey,
 | |
| 		config.EnvSecretKey,
 | |
| 		config.EnvRootUser,
 | |
| 		config.EnvRootPassword,
 | |
| 	} {
 | |
| 		os.Unsetenv(env)
 | |
| 	}
 | |
| 
 | |
| 	// Set as non-distributed.
 | |
| 	globalIsDistErasure = false
 | |
| 
 | |
| 	// Disable printing console messages during tests.
 | |
| 	color.Output = io.Discard
 | |
| 	// Minimum is error logs for testing
 | |
| 	logger.MinimumLogLevel = logger.ErrorLvl
 | |
| 
 | |
| 	// Uncomment the following line to see trace logs during unit tests.
 | |
| 	// logger.AddTarget(console.New())
 | |
| 
 | |
| 	// Set system resources to maximum.
 | |
| 	setMaxResources()
 | |
| 
 | |
| 	// Initialize globalConsoleSys system
 | |
| 	globalConsoleSys = NewConsoleLogger(context.Background())
 | |
| 
 | |
| 	globalInternodeTransport = newInternodeHTTPTransport(nil, rest.DefaultTimeout)()
 | |
| 
 | |
| 	initHelp()
 | |
| 
 | |
| 	resetTestGlobals()
 | |
| 
 | |
| 	globalIsCICD = true
 | |
| 
 | |
| 	os.Exit(m.Run())
 | |
| }
 | |
| 
 | |
| // concurrency level for certain parallel tests.
 | |
| const testConcurrencyLevel = 10
 | |
| 
 | |
| // Excerpts from @lsegal - https://github.com/aws/aws-sdk-js/issues/659#issuecomment-120477258
 | |
| //
 | |
| //	User-Agent:
 | |
| //
 | |
| //	    This is ignored from signing because signing this causes problems with generating pre-signed URLs
 | |
| //	    (that are executed by other agents) or when customers pass requests through proxies, which may
 | |
| //	    modify the user-agent.
 | |
| //
 | |
| //	Authorization:
 | |
| //
 | |
| //	    Is skipped for obvious reasons
 | |
| var ignoredHeaders = map[string]bool{
 | |
| 	"Authorization": true,
 | |
| 	"User-Agent":    true,
 | |
| }
 | |
| 
 | |
| // Headers to ignore in streaming v4
 | |
| var ignoredStreamingHeaders = map[string]bool{
 | |
| 	"Authorization": true,
 | |
| 	"Content-Type":  true,
 | |
| 	"Content-Md5":   true,
 | |
| 	"User-Agent":    true,
 | |
| }
 | |
| 
 | |
| // calculateSignedChunkLength - calculates the length of chunk metadata
 | |
| func calculateSignedChunkLength(chunkDataSize int64) int64 {
 | |
| 	return int64(len(fmt.Sprintf("%x", chunkDataSize))) +
 | |
| 		17 + // ";chunk-signature="
 | |
| 		64 + // e.g. "f2ca1bb6c7e907d06dafe4687e579fce76b37e4e93b7605022da52e6ccc26fd2"
 | |
| 		2 + // CRLF
 | |
| 		chunkDataSize +
 | |
| 		2 // CRLF
 | |
| }
 | |
| 
 | |
| func mustGetPutObjReader(t TestErrHandler, data io.Reader, size int64, md5hex, sha256hex string) *PutObjReader {
 | |
| 	hr, err := hash.NewReader(data, size, md5hex, sha256hex, size)
 | |
| 	if err != nil {
 | |
| 		t.Fatal(err)
 | |
| 	}
 | |
| 	return NewPutObjReader(hr)
 | |
| }
 | |
| 
 | |
| // calculateSignedChunkLength - calculates the length of the overall stream (data + metadata)
 | |
| func calculateStreamContentLength(dataLen, chunkSize int64) int64 {
 | |
| 	if dataLen <= 0 {
 | |
| 		return 0
 | |
| 	}
 | |
| 	chunksCount := dataLen / chunkSize
 | |
| 	remainingBytes := dataLen % chunkSize
 | |
| 	var streamLen int64
 | |
| 	streamLen += chunksCount * calculateSignedChunkLength(chunkSize)
 | |
| 	if remainingBytes > 0 {
 | |
| 		streamLen += calculateSignedChunkLength(remainingBytes)
 | |
| 	}
 | |
| 	streamLen += calculateSignedChunkLength(0)
 | |
| 	return streamLen
 | |
| }
 | |
| 
 | |
| func prepareFS(ctx context.Context) (ObjectLayer, string, error) {
 | |
| 	nDisks := 1
 | |
| 	fsDirs, err := getRandomDisks(nDisks)
 | |
| 	if err != nil {
 | |
| 		return nil, "", err
 | |
| 	}
 | |
| 	obj, _, err := initObjectLayer(context.Background(), mustGetPoolEndpoints(fsDirs...))
 | |
| 	if err != nil {
 | |
| 		return nil, "", err
 | |
| 	}
 | |
| 
 | |
| 	initAllSubsystems(ctx)
 | |
| 
 | |
| 	globalIAMSys.Init(ctx, obj, globalEtcdClient, 2*time.Second)
 | |
| 	return obj, fsDirs[0], nil
 | |
| }
 | |
| 
 | |
| func prepareErasureSets32(ctx context.Context) (ObjectLayer, []string, error) {
 | |
| 	return prepareErasure(ctx, 32)
 | |
| }
 | |
| 
 | |
| func prepareErasure(ctx context.Context, nDisks int) (ObjectLayer, []string, error) {
 | |
| 	fsDirs, err := getRandomDisks(nDisks)
 | |
| 	if err != nil {
 | |
| 		return nil, nil, err
 | |
| 	}
 | |
| 	obj, _, err := initObjectLayer(ctx, mustGetPoolEndpoints(fsDirs...))
 | |
| 	if err != nil {
 | |
| 		removeRoots(fsDirs)
 | |
| 		return nil, nil, err
 | |
| 	}
 | |
| 
 | |
| 	return obj, fsDirs, nil
 | |
| }
 | |
| 
 | |
| func prepareErasure16(ctx context.Context) (ObjectLayer, []string, error) {
 | |
| 	return prepareErasure(ctx, 16)
 | |
| }
 | |
| 
 | |
| // Initialize FS objects.
 | |
| func initFSObjects(disk string, t *testing.T) (obj ObjectLayer) {
 | |
| 	obj, _, err := initObjectLayer(context.Background(), mustGetPoolEndpoints(disk))
 | |
| 	if err != nil {
 | |
| 		t.Fatal(err)
 | |
| 	}
 | |
| 
 | |
| 	newTestConfig(globalMinioDefaultRegion, obj)
 | |
| 
 | |
| 	initAllSubsystems(GlobalContext)
 | |
| 	return obj
 | |
| }
 | |
| 
 | |
| // TestErrHandler - Go testing.T satisfy this interface.
 | |
| // This makes it easy to run the TestServer from any of the tests.
 | |
| // Using this interface, functionalities to be used in tests can be
 | |
| // made generalized, and can be integrated in benchmarks/unit tests/go check suite tests.
 | |
| type TestErrHandler interface {
 | |
| 	testing.TB
 | |
| }
 | |
| 
 | |
| const (
 | |
| 	// ErasureSDStr is the string which is used as notation for Single node ObjectLayer in the unit tests.
 | |
| 	ErasureSDStr string = "ErasureSD"
 | |
| 
 | |
| 	// ErasureTestStr is the string which is used as notation for Erasure ObjectLayer in the unit tests.
 | |
| 	ErasureTestStr string = "Erasure"
 | |
| 
 | |
| 	// ErasureSetsTestStr is the string which is used as notation for Erasure sets object layer in the unit tests.
 | |
| 	ErasureSetsTestStr string = "ErasureSet"
 | |
| )
 | |
| 
 | |
| const letterBytes = "abcdefghijklmnopqrstuvwxyz01234569"
 | |
| const (
 | |
| 	letterIdxBits = 6                    // 6 bits to represent a letter index
 | |
| 	letterIdxMask = 1<<letterIdxBits - 1 // All 1-bits, as many as letterIdxBits
 | |
| 	letterIdxMax  = 63 / letterIdxBits   // # of letter indices fitting in 63 bits
 | |
| )
 | |
| 
 | |
| // Random number state.
 | |
| // We generate random temporary file names so that there's a good
 | |
| // chance the file doesn't exist yet.
 | |
| var (
 | |
| 	randN  uint32
 | |
| 	randmu sync.Mutex
 | |
| )
 | |
| 
 | |
| // Temp files created in default Tmp dir
 | |
| var globalTestTmpDir = os.TempDir()
 | |
| 
 | |
| // reseed - returns a new seed every time the function is called.
 | |
| func reseed() uint32 {
 | |
| 	return uint32(time.Now().UnixNano() + int64(os.Getpid()))
 | |
| }
 | |
| 
 | |
| // nextSuffix - provides a new unique suffix every time the function is called.
 | |
| func nextSuffix() string {
 | |
| 	randmu.Lock()
 | |
| 	r := randN
 | |
| 	// Initial seed required, generate one.
 | |
| 	if r == 0 {
 | |
| 		r = reseed()
 | |
| 	}
 | |
| 	// constants from Numerical Recipes
 | |
| 	r = r*1664525 + 1013904223
 | |
| 	randN = r
 | |
| 	randmu.Unlock()
 | |
| 	return strconv.Itoa(int(1e9 + r%1e9))[1:]
 | |
| }
 | |
| 
 | |
| // isSameType - compares two object types via reflect.TypeOf
 | |
| func isSameType(obj1, obj2 interface{}) bool {
 | |
| 	return reflect.TypeOf(obj1) == reflect.TypeOf(obj2)
 | |
| }
 | |
| 
 | |
| // TestServer encapsulates an instantiation of a MinIO instance with a temporary backend.
 | |
| // Example usage:
 | |
| //
 | |
| //	s := StartTestServer(t,"Erasure")
 | |
| //	defer s.Stop()
 | |
| type TestServer struct {
 | |
| 	Root         string
 | |
| 	Disks        EndpointServerPools
 | |
| 	AccessKey    string
 | |
| 	SecretKey    string
 | |
| 	Server       *httptest.Server
 | |
| 	Obj          ObjectLayer
 | |
| 	cancel       context.CancelFunc
 | |
| 	rawDiskPaths []string
 | |
| }
 | |
| 
 | |
| // UnstartedTestServer - Configures a temp FS/Erasure backend,
 | |
| // initializes the endpoints and configures the test server.
 | |
| // The server should be started using the Start() method.
 | |
| func UnstartedTestServer(t TestErrHandler, instanceType string) TestServer {
 | |
| 	ctx, cancel := context.WithCancel(context.Background())
 | |
| 	// create an instance of TestServer.
 | |
| 	testServer := TestServer{cancel: cancel}
 | |
| 	// return FS/Erasure object layer and temp backend.
 | |
| 	objLayer, disks, err := prepareTestBackend(ctx, instanceType)
 | |
| 	if err != nil {
 | |
| 		t.Fatal(err)
 | |
| 	}
 | |
| 
 | |
| 	// set new server configuration.
 | |
| 	if err = newTestConfig(globalMinioDefaultRegion, objLayer); err != nil {
 | |
| 		t.Fatalf("%s", err)
 | |
| 	}
 | |
| 
 | |
| 	return initTestServerWithBackend(ctx, t, testServer, objLayer, disks)
 | |
| }
 | |
| 
 | |
| // initializes a test server with the given object layer and disks.
 | |
| func initTestServerWithBackend(ctx context.Context, t TestErrHandler, testServer TestServer, objLayer ObjectLayer, disks []string) TestServer {
 | |
| 	// Test Server needs to start before formatting of disks.
 | |
| 	// Get credential.
 | |
| 	credentials := globalActiveCred
 | |
| 
 | |
| 	testServer.Obj = objLayer
 | |
| 	testServer.rawDiskPaths = disks
 | |
| 	testServer.Disks = mustGetPoolEndpoints(disks...)
 | |
| 	testServer.AccessKey = credentials.AccessKey
 | |
| 	testServer.SecretKey = credentials.SecretKey
 | |
| 
 | |
| 	httpHandler, err := configureServerHandler(testServer.Disks)
 | |
| 	if err != nil {
 | |
| 		t.Fatalf("Failed to configure one of the RPC services <ERROR> %s", err)
 | |
| 	}
 | |
| 
 | |
| 	// Run TestServer.
 | |
| 	testServer.Server = httptest.NewUnstartedServer(setCriticalErrorHandler(corsHandler(httpHandler)))
 | |
| 
 | |
| 	globalObjLayerMutex.Lock()
 | |
| 	globalObjectAPI = objLayer
 | |
| 	globalObjLayerMutex.Unlock()
 | |
| 
 | |
| 	// initialize peer rpc
 | |
| 	host, port := mustSplitHostPort(testServer.Server.Listener.Addr().String())
 | |
| 	globalMinioHost = host
 | |
| 	globalMinioPort = port
 | |
| 	globalMinioAddr = getEndpointsLocalAddr(testServer.Disks)
 | |
| 
 | |
| 	initAllSubsystems(ctx)
 | |
| 
 | |
| 	globalEtcdClient = nil
 | |
| 
 | |
| 	initConfigSubsystem(ctx, objLayer)
 | |
| 
 | |
| 	globalIAMSys.Init(ctx, objLayer, globalEtcdClient, 2*time.Second)
 | |
| 
 | |
| 	globalEventNotifier.InitBucketTargets(ctx, objLayer)
 | |
| 
 | |
| 	return testServer
 | |
| }
 | |
| 
 | |
| // testServerCertPEM and testServerKeyPEM are generated by
 | |
| //  https://golang.org/src/crypto/tls/generate_cert.go
 | |
| //    $ go run generate_cert.go -ca --host 127.0.0.1
 | |
| // The generated certificate contains IP SAN, that way we don't need
 | |
| // to enable InsecureSkipVerify in TLS config
 | |
| 
 | |
| // Starts the test server and returns the TestServer with TLS configured instance.
 | |
| func StartTestTLSServer(t TestErrHandler, instanceType string, cert, key []byte) TestServer {
 | |
| 	// Fetch TLS key and pem files from test-data/ directory.
 | |
| 	//	dir, _ := os.Getwd()
 | |
| 	//	testDataDir := filepath.Join(filepath.Dir(dir), "test-data")
 | |
| 	//
 | |
| 	//	pemFile := filepath.Join(testDataDir, "server.pem")
 | |
| 	//	keyFile := filepath.Join(testDataDir, "server.key")
 | |
| 	cer, err := tls.X509KeyPair(cert, key)
 | |
| 	if err != nil {
 | |
| 		t.Fatalf("Failed to load certificate: %v", err)
 | |
| 	}
 | |
| 	config := &tls.Config{Certificates: []tls.Certificate{cer}}
 | |
| 
 | |
| 	testServer := UnstartedTestServer(t, instanceType)
 | |
| 	testServer.Server.TLS = config
 | |
| 	testServer.Server.StartTLS()
 | |
| 	return testServer
 | |
| }
 | |
| 
 | |
| // Starts the test server and returns the TestServer instance.
 | |
| func StartTestServer(t TestErrHandler, instanceType string) TestServer {
 | |
| 	// create an instance of TestServer.
 | |
| 	testServer := UnstartedTestServer(t, instanceType)
 | |
| 	testServer.Server.Start()
 | |
| 	return testServer
 | |
| }
 | |
| 
 | |
| // Sets the global config path to empty string.
 | |
| func resetGlobalConfigPath() {
 | |
| 	globalConfigDir = &ConfigDir{path: ""}
 | |
| }
 | |
| 
 | |
| // sets globalObjectAPI to `nil`.
 | |
| func resetGlobalObjectAPI() {
 | |
| 	globalObjLayerMutex.Lock()
 | |
| 	globalObjectAPI = nil
 | |
| 	globalObjLayerMutex.Unlock()
 | |
| }
 | |
| 
 | |
| // reset the value of the Global server config.
 | |
| // set it to `nil`.
 | |
| func resetGlobalConfig() {
 | |
| 	// hold the mutex lock before a new config is assigned.
 | |
| 	globalServerConfigMu.Lock()
 | |
| 	// Save the loaded config globally.
 | |
| 	globalServerConfig = nil
 | |
| 	globalServerConfigMu.Unlock()
 | |
| }
 | |
| 
 | |
| func resetGlobalEndpoints() {
 | |
| 	globalEndpoints = EndpointServerPools{}
 | |
| }
 | |
| 
 | |
| func resetGlobalIsErasure() {
 | |
| 	globalIsErasure = false
 | |
| }
 | |
| 
 | |
| // reset global heal state
 | |
| func resetGlobalHealState() {
 | |
| 	// Init global heal state
 | |
| 	if globalAllHealState == nil {
 | |
| 		globalAllHealState = newHealState(GlobalContext, false)
 | |
| 	} else {
 | |
| 		globalAllHealState.Lock()
 | |
| 		for _, v := range globalAllHealState.healSeqMap {
 | |
| 			if !v.hasEnded() {
 | |
| 				v.stop()
 | |
| 			}
 | |
| 		}
 | |
| 		globalAllHealState.Unlock()
 | |
| 	}
 | |
| 
 | |
| 	// Init background heal state
 | |
| 	if globalBackgroundHealState == nil {
 | |
| 		globalBackgroundHealState = newHealState(GlobalContext, false)
 | |
| 	} else {
 | |
| 		globalBackgroundHealState.Lock()
 | |
| 		for _, v := range globalBackgroundHealState.healSeqMap {
 | |
| 			if !v.hasEnded() {
 | |
| 				v.stop()
 | |
| 			}
 | |
| 		}
 | |
| 		globalBackgroundHealState.Unlock()
 | |
| 	}
 | |
| }
 | |
| 
 | |
| // sets globalIAMSys to `nil`.
 | |
| func resetGlobalIAMSys() {
 | |
| 	globalIAMSys = nil
 | |
| }
 | |
| 
 | |
| // Resets all the globals used modified in tests.
 | |
| // Resetting ensures that the changes made to globals by one test doesn't affect others.
 | |
| func resetTestGlobals() {
 | |
| 	// set globalObjectAPI to `nil`.
 | |
| 	resetGlobalObjectAPI()
 | |
| 	// Reset config path set.
 | |
| 	resetGlobalConfigPath()
 | |
| 	// Reset Global server config.
 | |
| 	resetGlobalConfig()
 | |
| 	// Reset global endpoints.
 | |
| 	resetGlobalEndpoints()
 | |
| 	// Reset global isErasure flag.
 | |
| 	resetGlobalIsErasure()
 | |
| 	// Reset global heal state
 | |
| 	resetGlobalHealState()
 | |
| 	// Reset globalIAMSys to `nil`
 | |
| 	resetGlobalIAMSys()
 | |
| }
 | |
| 
 | |
| // Configure the server for the test run.
 | |
| func newTestConfig(bucketLocation string, obj ObjectLayer) (err error) {
 | |
| 	// Initialize server config.
 | |
| 	if err = newSrvConfig(obj); err != nil {
 | |
| 		return err
 | |
| 	}
 | |
| 
 | |
| 	// Set a default region.
 | |
| 	config.SetRegion(globalServerConfig, bucketLocation)
 | |
| 
 | |
| 	applyDynamicConfigForSubSys(context.Background(), obj, globalServerConfig, config.StorageClassSubSys)
 | |
| 
 | |
| 	// Save config.
 | |
| 	return saveServerConfig(context.Background(), obj, globalServerConfig)
 | |
| }
 | |
| 
 | |
| // Deleting the temporary backend and stopping the server.
 | |
| func (testServer TestServer) Stop() {
 | |
| 	testServer.cancel()
 | |
| 	testServer.Server.Close()
 | |
| 	testServer.Obj.Shutdown(context.Background())
 | |
| 	os.RemoveAll(testServer.Root)
 | |
| 	for _, ep := range testServer.Disks {
 | |
| 		for _, disk := range ep.Endpoints {
 | |
| 			os.RemoveAll(disk.Path)
 | |
| 		}
 | |
| 	}
 | |
| }
 | |
| 
 | |
| // Truncate request to simulate unexpected EOF for a request signed using streaming signature v4.
 | |
| func truncateChunkByHalfSigv4(req *http.Request) (*http.Request, error) {
 | |
| 	bufReader := bufio.NewReader(req.Body)
 | |
| 	hexChunkSize, chunkSignature, err := readChunkLine(bufReader)
 | |
| 	if err != nil {
 | |
| 		return nil, err
 | |
| 	}
 | |
| 
 | |
| 	newChunkHdr := []byte(fmt.Sprintf("%s"+s3ChunkSignatureStr+"%s\r\n",
 | |
| 		hexChunkSize, chunkSignature))
 | |
| 	newChunk, err := io.ReadAll(bufReader)
 | |
| 	if err != nil {
 | |
| 		return nil, err
 | |
| 	}
 | |
| 	newReq := req
 | |
| 	newReq.Body = io.NopCloser(
 | |
| 		bytes.NewReader(bytes.Join([][]byte{newChunkHdr, newChunk[:len(newChunk)/2]},
 | |
| 			[]byte(""))),
 | |
| 	)
 | |
| 	return newReq, nil
 | |
| }
 | |
| 
 | |
| // Malform data given a request signed using streaming signature V4.
 | |
| func malformDataSigV4(req *http.Request, newByte byte) (*http.Request, error) {
 | |
| 	bufReader := bufio.NewReader(req.Body)
 | |
| 	hexChunkSize, chunkSignature, err := readChunkLine(bufReader)
 | |
| 	if err != nil {
 | |
| 		return nil, err
 | |
| 	}
 | |
| 
 | |
| 	newChunkHdr := []byte(fmt.Sprintf("%s"+s3ChunkSignatureStr+"%s\r\n",
 | |
| 		hexChunkSize, chunkSignature))
 | |
| 	newChunk, err := io.ReadAll(bufReader)
 | |
| 	if err != nil {
 | |
| 		return nil, err
 | |
| 	}
 | |
| 
 | |
| 	newChunk[0] = newByte
 | |
| 	newReq := req
 | |
| 	newReq.Body = io.NopCloser(
 | |
| 		bytes.NewReader(bytes.Join([][]byte{newChunkHdr, newChunk},
 | |
| 			[]byte(""))),
 | |
| 	)
 | |
| 
 | |
| 	return newReq, nil
 | |
| }
 | |
| 
 | |
| // Malform chunk size given a request signed using streaming signatureV4.
 | |
| func malformChunkSizeSigV4(req *http.Request, badSize int64) (*http.Request, error) {
 | |
| 	bufReader := bufio.NewReader(req.Body)
 | |
| 	_, chunkSignature, err := readChunkLine(bufReader)
 | |
| 	if err != nil {
 | |
| 		return nil, err
 | |
| 	}
 | |
| 
 | |
| 	n := badSize
 | |
| 	newHexChunkSize := []byte(fmt.Sprintf("%x", n))
 | |
| 	newChunkHdr := []byte(fmt.Sprintf("%s"+s3ChunkSignatureStr+"%s\r\n",
 | |
| 		newHexChunkSize, chunkSignature))
 | |
| 	newChunk, err := io.ReadAll(bufReader)
 | |
| 	if err != nil {
 | |
| 		return nil, err
 | |
| 	}
 | |
| 
 | |
| 	newReq := req
 | |
| 	newReq.Body = io.NopCloser(
 | |
| 		bytes.NewReader(bytes.Join([][]byte{newChunkHdr, newChunk},
 | |
| 			[]byte(""))),
 | |
| 	)
 | |
| 
 | |
| 	return newReq, nil
 | |
| }
 | |
| 
 | |
| // Sign given request using Signature V4.
 | |
| func signStreamingRequest(req *http.Request, accessKey, secretKey string, currTime time.Time) (string, error) {
 | |
| 	// Get hashed payload.
 | |
| 	hashedPayload := req.Header.Get("x-amz-content-sha256")
 | |
| 	if hashedPayload == "" {
 | |
| 		return "", fmt.Errorf("Invalid hashed payload")
 | |
| 	}
 | |
| 
 | |
| 	// Set x-amz-date.
 | |
| 	req.Header.Set("x-amz-date", currTime.Format(iso8601Format))
 | |
| 
 | |
| 	// Get header map.
 | |
| 	headerMap := make(map[string][]string)
 | |
| 	for k, vv := range req.Header {
 | |
| 		// If request header key is not in ignored headers, then add it.
 | |
| 		if _, ok := ignoredStreamingHeaders[http.CanonicalHeaderKey(k)]; !ok {
 | |
| 			headerMap[strings.ToLower(k)] = vv
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	// Get header keys.
 | |
| 	headers := []string{"host"}
 | |
| 	for k := range headerMap {
 | |
| 		headers = append(headers, k)
 | |
| 	}
 | |
| 	sort.Strings(headers)
 | |
| 
 | |
| 	// Get canonical headers.
 | |
| 	var buf bytes.Buffer
 | |
| 	for _, k := range headers {
 | |
| 		buf.WriteString(k)
 | |
| 		buf.WriteByte(':')
 | |
| 		switch {
 | |
| 		case k == "host":
 | |
| 			buf.WriteString(req.URL.Host)
 | |
| 			fallthrough
 | |
| 		default:
 | |
| 			for idx, v := range headerMap[k] {
 | |
| 				if idx > 0 {
 | |
| 					buf.WriteByte(',')
 | |
| 				}
 | |
| 				buf.WriteString(v)
 | |
| 			}
 | |
| 			buf.WriteByte('\n')
 | |
| 		}
 | |
| 	}
 | |
| 	canonicalHeaders := buf.String()
 | |
| 
 | |
| 	// Get signed headers.
 | |
| 	signedHeaders := strings.Join(headers, ";")
 | |
| 
 | |
| 	// Get canonical query string.
 | |
| 	req.URL.RawQuery = strings.ReplaceAll(req.URL.Query().Encode(), "+", "%20")
 | |
| 
 | |
| 	// Get canonical URI.
 | |
| 	canonicalURI := s3utils.EncodePath(req.URL.Path)
 | |
| 
 | |
| 	// Get canonical request.
 | |
| 	// canonicalRequest =
 | |
| 	//  <HTTPMethod>\n
 | |
| 	//  <CanonicalURI>\n
 | |
| 	//  <CanonicalQueryString>\n
 | |
| 	//  <CanonicalHeaders>\n
 | |
| 	//  <SignedHeaders>\n
 | |
| 	//  <HashedPayload>
 | |
| 	//
 | |
| 	canonicalRequest := strings.Join([]string{
 | |
| 		req.Method,
 | |
| 		canonicalURI,
 | |
| 		req.URL.RawQuery,
 | |
| 		canonicalHeaders,
 | |
| 		signedHeaders,
 | |
| 		hashedPayload,
 | |
| 	}, "\n")
 | |
| 
 | |
| 	// Get scope.
 | |
| 	scope := strings.Join([]string{
 | |
| 		currTime.Format(yyyymmdd),
 | |
| 		globalMinioDefaultRegion,
 | |
| 		string(serviceS3),
 | |
| 		"aws4_request",
 | |
| 	}, SlashSeparator)
 | |
| 
 | |
| 	stringToSign := "AWS4-HMAC-SHA256" + "\n" + currTime.Format(iso8601Format) + "\n"
 | |
| 	stringToSign += scope + "\n"
 | |
| 	stringToSign += getSHA256Hash([]byte(canonicalRequest))
 | |
| 
 | |
| 	date := sumHMAC([]byte("AWS4"+secretKey), []byte(currTime.Format(yyyymmdd)))
 | |
| 	region := sumHMAC(date, []byte(globalMinioDefaultRegion))
 | |
| 	service := sumHMAC(region, []byte(string(serviceS3)))
 | |
| 	signingKey := sumHMAC(service, []byte("aws4_request"))
 | |
| 
 | |
| 	signature := hex.EncodeToString(sumHMAC(signingKey, []byte(stringToSign)))
 | |
| 
 | |
| 	// final Authorization header
 | |
| 	parts := []string{
 | |
| 		"AWS4-HMAC-SHA256" + " Credential=" + accessKey + SlashSeparator + scope,
 | |
| 		"SignedHeaders=" + signedHeaders,
 | |
| 		"Signature=" + signature,
 | |
| 	}
 | |
| 	auth := strings.Join(parts, ", ")
 | |
| 	req.Header.Set("Authorization", auth)
 | |
| 
 | |
| 	return signature, nil
 | |
| }
 | |
| 
 | |
| // Returns new HTTP request object.
 | |
| func newTestStreamingRequest(method, urlStr string, dataLength, chunkSize int64, body io.ReadSeeker) (*http.Request, error) {
 | |
| 	if method == "" {
 | |
| 		method = http.MethodPost
 | |
| 	}
 | |
| 
 | |
| 	req, err := http.NewRequest(method, urlStr, nil)
 | |
| 	if err != nil {
 | |
| 		return nil, err
 | |
| 	}
 | |
| 
 | |
| 	if body == nil {
 | |
| 		// this is added to avoid panic during io.ReadAll(req.Body).
 | |
| 		// th stack trace can be found here  https://github.com/minio/minio/pull/2074 .
 | |
| 		// This is very similar to https://github.com/golang/go/issues/7527.
 | |
| 		req.Body = io.NopCloser(bytes.NewReader([]byte("")))
 | |
| 	}
 | |
| 
 | |
| 	contentLength := calculateStreamContentLength(dataLength, chunkSize)
 | |
| 
 | |
| 	req.Header.Set("x-amz-content-sha256", "STREAMING-AWS4-HMAC-SHA256-PAYLOAD")
 | |
| 	req.Header.Set("content-encoding", "aws-chunked")
 | |
| 	req.Header.Set("x-amz-decoded-content-length", strconv.FormatInt(dataLength, 10))
 | |
| 	req.Header.Set("content-length", strconv.FormatInt(contentLength, 10))
 | |
| 
 | |
| 	// Seek back to beginning.
 | |
| 	body.Seek(0, 0)
 | |
| 
 | |
| 	// Add body
 | |
| 	req.Body = io.NopCloser(body)
 | |
| 	req.ContentLength = contentLength
 | |
| 
 | |
| 	return req, nil
 | |
| }
 | |
| 
 | |
| func assembleStreamingChunks(req *http.Request, body io.ReadSeeker, chunkSize int64,
 | |
| 	secretKey, signature string, currTime time.Time) (*http.Request, error,
 | |
| ) {
 | |
| 	regionStr := globalSite.Region
 | |
| 	var stream []byte
 | |
| 	var buffer []byte
 | |
| 	body.Seek(0, 0)
 | |
| 	for {
 | |
| 		buffer = make([]byte, chunkSize)
 | |
| 		n, err := body.Read(buffer)
 | |
| 		if err != nil && err != io.EOF {
 | |
| 			return nil, err
 | |
| 		}
 | |
| 
 | |
| 		// Get scope.
 | |
| 		scope := strings.Join([]string{
 | |
| 			currTime.Format(yyyymmdd),
 | |
| 			regionStr,
 | |
| 			string(serviceS3),
 | |
| 			"aws4_request",
 | |
| 		}, SlashSeparator)
 | |
| 
 | |
| 		stringToSign := "AWS4-HMAC-SHA256-PAYLOAD" + "\n"
 | |
| 		stringToSign = stringToSign + currTime.Format(iso8601Format) + "\n"
 | |
| 		stringToSign = stringToSign + scope + "\n"
 | |
| 		stringToSign = stringToSign + signature + "\n"
 | |
| 		stringToSign = stringToSign + "e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855" + "\n" // hex(sum256(""))
 | |
| 		stringToSign += getSHA256Hash(buffer[:n])
 | |
| 
 | |
| 		date := sumHMAC([]byte("AWS4"+secretKey), []byte(currTime.Format(yyyymmdd)))
 | |
| 		region := sumHMAC(date, []byte(regionStr))
 | |
| 		service := sumHMAC(region, []byte(serviceS3))
 | |
| 		signingKey := sumHMAC(service, []byte("aws4_request"))
 | |
| 
 | |
| 		signature = hex.EncodeToString(sumHMAC(signingKey, []byte(stringToSign)))
 | |
| 
 | |
| 		stream = append(stream, []byte(fmt.Sprintf("%x", n)+";chunk-signature="+signature+"\r\n")...)
 | |
| 		stream = append(stream, buffer[:n]...)
 | |
| 		stream = append(stream, []byte("\r\n")...)
 | |
| 
 | |
| 		if n <= 0 {
 | |
| 			break
 | |
| 		}
 | |
| 
 | |
| 	}
 | |
| 	req.Body = io.NopCloser(bytes.NewReader(stream))
 | |
| 	return req, nil
 | |
| }
 | |
| 
 | |
| func newTestStreamingSignedBadChunkDateRequest(method, urlStr string, contentLength, chunkSize int64, body io.ReadSeeker, accessKey, secretKey string) (*http.Request, error) {
 | |
| 	req, err := newTestStreamingRequest(method, urlStr, contentLength, chunkSize, body)
 | |
| 	if err != nil {
 | |
| 		return nil, err
 | |
| 	}
 | |
| 
 | |
| 	currTime := UTCNow()
 | |
| 	signature, err := signStreamingRequest(req, accessKey, secretKey, currTime)
 | |
| 	if err != nil {
 | |
| 		return nil, err
 | |
| 	}
 | |
| 
 | |
| 	// skew the time between the chunk signature calculation and seed signature.
 | |
| 	currTime = currTime.Add(1 * time.Second)
 | |
| 	req, err = assembleStreamingChunks(req, body, chunkSize, secretKey, signature, currTime)
 | |
| 	return req, err
 | |
| }
 | |
| 
 | |
| func newTestStreamingSignedCustomEncodingRequest(method, urlStr string, contentLength, chunkSize int64, body io.ReadSeeker, accessKey, secretKey, contentEncoding string) (*http.Request, error) {
 | |
| 	req, err := newTestStreamingRequest(method, urlStr, contentLength, chunkSize, body)
 | |
| 	if err != nil {
 | |
| 		return nil, err
 | |
| 	}
 | |
| 
 | |
| 	// Set custom encoding.
 | |
| 	req.Header.Set("content-encoding", contentEncoding)
 | |
| 
 | |
| 	currTime := UTCNow()
 | |
| 	signature, err := signStreamingRequest(req, accessKey, secretKey, currTime)
 | |
| 	if err != nil {
 | |
| 		return nil, err
 | |
| 	}
 | |
| 
 | |
| 	req, err = assembleStreamingChunks(req, body, chunkSize, secretKey, signature, currTime)
 | |
| 	return req, err
 | |
| }
 | |
| 
 | |
| // Returns new HTTP request object signed with streaming signature v4.
 | |
| func newTestStreamingSignedRequest(method, urlStr string, contentLength, chunkSize int64, body io.ReadSeeker, accessKey, secretKey string) (*http.Request, error) {
 | |
| 	req, err := newTestStreamingRequest(method, urlStr, contentLength, chunkSize, body)
 | |
| 	if err != nil {
 | |
| 		return nil, err
 | |
| 	}
 | |
| 
 | |
| 	currTime := UTCNow()
 | |
| 	signature, err := signStreamingRequest(req, accessKey, secretKey, currTime)
 | |
| 	if err != nil {
 | |
| 		return nil, err
 | |
| 	}
 | |
| 
 | |
| 	req, err = assembleStreamingChunks(req, body, chunkSize, secretKey, signature, currTime)
 | |
| 	return req, err
 | |
| }
 | |
| 
 | |
| // preSignV4 presign the request, in accordance with
 | |
| // http://docs.aws.amazon.com/AmazonS3/latest/API/sigv4-query-string-auth.html.
 | |
| func preSignV4(req *http.Request, accessKeyID, secretAccessKey string, expires int64) error {
 | |
| 	// Presign is not needed for anonymous credentials.
 | |
| 	if accessKeyID == "" || secretAccessKey == "" {
 | |
| 		return errors.New("Presign cannot be generated without access and secret keys")
 | |
| 	}
 | |
| 
 | |
| 	region := globalSite.Region
 | |
| 	date := UTCNow()
 | |
| 	scope := getScope(date, region)
 | |
| 	credential := fmt.Sprintf("%s/%s", accessKeyID, scope)
 | |
| 
 | |
| 	// Set URL query.
 | |
| 	query := req.URL.Query()
 | |
| 	query.Set("X-Amz-Algorithm", signV4Algorithm)
 | |
| 	query.Set("X-Amz-Date", date.Format(iso8601Format))
 | |
| 	query.Set("X-Amz-Expires", strconv.FormatInt(expires, 10))
 | |
| 	query.Set("X-Amz-SignedHeaders", "host")
 | |
| 	query.Set("X-Amz-Credential", credential)
 | |
| 	query.Set("X-Amz-Content-Sha256", unsignedPayload)
 | |
| 
 | |
| 	// "host" is the only header required to be signed for Presigned URLs.
 | |
| 	extractedSignedHeaders := make(http.Header)
 | |
| 	extractedSignedHeaders.Set("host", req.Host)
 | |
| 
 | |
| 	queryStr := strings.ReplaceAll(query.Encode(), "+", "%20")
 | |
| 	canonicalRequest := getCanonicalRequest(extractedSignedHeaders, unsignedPayload, queryStr, req.URL.Path, req.Method)
 | |
| 	stringToSign := getStringToSign(canonicalRequest, date, scope)
 | |
| 	signingKey := getSigningKey(secretAccessKey, date, region, serviceS3)
 | |
| 	signature := getSignature(signingKey, stringToSign)
 | |
| 
 | |
| 	req.URL.RawQuery = query.Encode()
 | |
| 
 | |
| 	// Add signature header to RawQuery.
 | |
| 	req.URL.RawQuery += "&X-Amz-Signature=" + url.QueryEscape(signature)
 | |
| 
 | |
| 	// Construct the final presigned URL.
 | |
| 	return nil
 | |
| }
 | |
| 
 | |
| // preSignV2 - presign the request in following style.
 | |
| // https://${S3_BUCKET}.s3.amazonaws.com/${S3_OBJECT}?AWSAccessKeyId=${S3_ACCESS_KEY}&Expires=${TIMESTAMP}&Signature=${SIGNATURE}.
 | |
| func preSignV2(req *http.Request, accessKeyID, secretAccessKey string, expires int64) error {
 | |
| 	// Presign is not needed for anonymous credentials.
 | |
| 	if accessKeyID == "" || secretAccessKey == "" {
 | |
| 		return errors.New("Presign cannot be generated without access and secret keys")
 | |
| 	}
 | |
| 
 | |
| 	// FIXME: Remove following portion of code after fixing a bug in minio-go preSignV2.
 | |
| 
 | |
| 	d := UTCNow()
 | |
| 	// Find epoch expires when the request will expire.
 | |
| 	epochExpires := d.Unix() + expires
 | |
| 
 | |
| 	// Add expires header if not present.
 | |
| 	expiresStr := req.Header.Get("Expires")
 | |
| 	if expiresStr == "" {
 | |
| 		expiresStr = strconv.FormatInt(epochExpires, 10)
 | |
| 		req.Header.Set("Expires", expiresStr)
 | |
| 	}
 | |
| 
 | |
| 	// url.RawPath will be valid if path has any encoded characters, if not it will
 | |
| 	// be empty - in which case we need to consider url.Path (bug in net/http?)
 | |
| 	encodedResource := req.URL.RawPath
 | |
| 	encodedQuery := req.URL.RawQuery
 | |
| 	if encodedResource == "" {
 | |
| 		splits := strings.SplitN(req.URL.Path, "?", 2)
 | |
| 		encodedResource = splits[0]
 | |
| 		if len(splits) == 2 {
 | |
| 			encodedQuery = splits[1]
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	unescapedQueries, err := unescapeQueries(encodedQuery)
 | |
| 	if err != nil {
 | |
| 		return err
 | |
| 	}
 | |
| 
 | |
| 	// Get presigned string to sign.
 | |
| 	stringToSign := getStringToSignV2(req.Method, encodedResource, strings.Join(unescapedQueries, "&"), req.Header, expiresStr)
 | |
| 	hm := hmac.New(sha1.New, []byte(secretAccessKey))
 | |
| 	hm.Write([]byte(stringToSign))
 | |
| 
 | |
| 	// Calculate signature.
 | |
| 	signature := base64.StdEncoding.EncodeToString(hm.Sum(nil))
 | |
| 
 | |
| 	query := req.URL.Query()
 | |
| 	// Handle specially for Google Cloud Storage.
 | |
| 	query.Set("AWSAccessKeyId", accessKeyID)
 | |
| 	// Fill in Expires for presigned query.
 | |
| 	query.Set("Expires", strconv.FormatInt(epochExpires, 10))
 | |
| 
 | |
| 	// Encode query and save.
 | |
| 	req.URL.RawQuery = query.Encode()
 | |
| 
 | |
| 	// Save signature finally.
 | |
| 	req.URL.RawQuery += "&Signature=" + url.QueryEscape(signature)
 | |
| 	return nil
 | |
| }
 | |
| 
 | |
| // Sign given request using Signature V2.
 | |
| func signRequestV2(req *http.Request, accessKey, secretKey string) error {
 | |
| 	signer.SignV2(*req, accessKey, secretKey, false)
 | |
| 	return nil
 | |
| }
 | |
| 
 | |
| // Sign given request using Signature V4.
 | |
| func signRequestV4(req *http.Request, accessKey, secretKey string) error {
 | |
| 	// Get hashed payload.
 | |
| 	hashedPayload := req.Header.Get("x-amz-content-sha256")
 | |
| 	if hashedPayload == "" {
 | |
| 		return fmt.Errorf("Invalid hashed payload")
 | |
| 	}
 | |
| 
 | |
| 	currTime := UTCNow()
 | |
| 
 | |
| 	// Set x-amz-date.
 | |
| 	req.Header.Set("x-amz-date", currTime.Format(iso8601Format))
 | |
| 
 | |
| 	// Get header map.
 | |
| 	headerMap := make(map[string][]string)
 | |
| 	for k, vv := range req.Header {
 | |
| 		// If request header key is not in ignored headers, then add it.
 | |
| 		if _, ok := ignoredHeaders[http.CanonicalHeaderKey(k)]; !ok {
 | |
| 			headerMap[strings.ToLower(k)] = vv
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	// Get header keys.
 | |
| 	headers := []string{"host"}
 | |
| 	for k := range headerMap {
 | |
| 		headers = append(headers, k)
 | |
| 	}
 | |
| 	sort.Strings(headers)
 | |
| 
 | |
| 	region := globalSite.Region
 | |
| 
 | |
| 	// Get canonical headers.
 | |
| 	var buf bytes.Buffer
 | |
| 	for _, k := range headers {
 | |
| 		buf.WriteString(k)
 | |
| 		buf.WriteByte(':')
 | |
| 		switch {
 | |
| 		case k == "host":
 | |
| 			buf.WriteString(req.URL.Host)
 | |
| 			fallthrough
 | |
| 		default:
 | |
| 			for idx, v := range headerMap[k] {
 | |
| 				if idx > 0 {
 | |
| 					buf.WriteByte(',')
 | |
| 				}
 | |
| 				buf.WriteString(v)
 | |
| 			}
 | |
| 			buf.WriteByte('\n')
 | |
| 		}
 | |
| 	}
 | |
| 	canonicalHeaders := buf.String()
 | |
| 
 | |
| 	// Get signed headers.
 | |
| 	signedHeaders := strings.Join(headers, ";")
 | |
| 
 | |
| 	// Get canonical query string.
 | |
| 	req.URL.RawQuery = strings.ReplaceAll(req.URL.Query().Encode(), "+", "%20")
 | |
| 
 | |
| 	// Get canonical URI.
 | |
| 	canonicalURI := s3utils.EncodePath(req.URL.Path)
 | |
| 
 | |
| 	// Get canonical request.
 | |
| 	// canonicalRequest =
 | |
| 	//  <HTTPMethod>\n
 | |
| 	//  <CanonicalURI>\n
 | |
| 	//  <CanonicalQueryString>\n
 | |
| 	//  <CanonicalHeaders>\n
 | |
| 	//  <SignedHeaders>\n
 | |
| 	//  <HashedPayload>
 | |
| 	//
 | |
| 	canonicalRequest := strings.Join([]string{
 | |
| 		req.Method,
 | |
| 		canonicalURI,
 | |
| 		req.URL.RawQuery,
 | |
| 		canonicalHeaders,
 | |
| 		signedHeaders,
 | |
| 		hashedPayload,
 | |
| 	}, "\n")
 | |
| 
 | |
| 	// Get scope.
 | |
| 	scope := strings.Join([]string{
 | |
| 		currTime.Format(yyyymmdd),
 | |
| 		region,
 | |
| 		string(serviceS3),
 | |
| 		"aws4_request",
 | |
| 	}, SlashSeparator)
 | |
| 
 | |
| 	stringToSign := "AWS4-HMAC-SHA256" + "\n" + currTime.Format(iso8601Format) + "\n"
 | |
| 	stringToSign = stringToSign + scope + "\n"
 | |
| 	stringToSign += getSHA256Hash([]byte(canonicalRequest))
 | |
| 
 | |
| 	date := sumHMAC([]byte("AWS4"+secretKey), []byte(currTime.Format(yyyymmdd)))
 | |
| 	regionHMAC := sumHMAC(date, []byte(region))
 | |
| 	service := sumHMAC(regionHMAC, []byte(serviceS3))
 | |
| 	signingKey := sumHMAC(service, []byte("aws4_request"))
 | |
| 
 | |
| 	signature := hex.EncodeToString(sumHMAC(signingKey, []byte(stringToSign)))
 | |
| 
 | |
| 	// final Authorization header
 | |
| 	parts := []string{
 | |
| 		"AWS4-HMAC-SHA256" + " Credential=" + accessKey + SlashSeparator + scope,
 | |
| 		"SignedHeaders=" + signedHeaders,
 | |
| 		"Signature=" + signature,
 | |
| 	}
 | |
| 	auth := strings.Join(parts, ", ")
 | |
| 	req.Header.Set("Authorization", auth)
 | |
| 
 | |
| 	return nil
 | |
| }
 | |
| 
 | |
| // getCredentialString generate a credential string.
 | |
| func getCredentialString(accessKeyID, location string, t time.Time) string {
 | |
| 	return accessKeyID + SlashSeparator + getScope(t, location)
 | |
| }
 | |
| 
 | |
| // getMD5HashBase64 returns MD5 hash in base64 encoding of given data.
 | |
| func getMD5HashBase64(data []byte) string {
 | |
| 	return base64.StdEncoding.EncodeToString(getMD5Sum(data))
 | |
| }
 | |
| 
 | |
| // Returns new HTTP request object.
 | |
| func newTestRequest(method, urlStr string, contentLength int64, body io.ReadSeeker) (*http.Request, error) {
 | |
| 	if method == "" {
 | |
| 		method = http.MethodPost
 | |
| 	}
 | |
| 
 | |
| 	// Save for subsequent use
 | |
| 	var hashedPayload string
 | |
| 	var md5Base64 string
 | |
| 	switch {
 | |
| 	case body == nil:
 | |
| 		hashedPayload = getSHA256Hash([]byte{})
 | |
| 	default:
 | |
| 		payloadBytes, err := io.ReadAll(body)
 | |
| 		if err != nil {
 | |
| 			return nil, err
 | |
| 		}
 | |
| 		hashedPayload = getSHA256Hash(payloadBytes)
 | |
| 		md5Base64 = getMD5HashBase64(payloadBytes)
 | |
| 	}
 | |
| 	// Seek back to beginning.
 | |
| 	if body != nil {
 | |
| 		body.Seek(0, 0)
 | |
| 	} else {
 | |
| 		body = bytes.NewReader([]byte(""))
 | |
| 	}
 | |
| 	req, err := http.NewRequest(method, urlStr, body)
 | |
| 	if err != nil {
 | |
| 		return nil, err
 | |
| 	}
 | |
| 	if md5Base64 != "" {
 | |
| 		req.Header.Set("Content-Md5", md5Base64)
 | |
| 	}
 | |
| 	req.Header.Set("x-amz-content-sha256", hashedPayload)
 | |
| 
 | |
| 	// Add Content-Length
 | |
| 	req.ContentLength = contentLength
 | |
| 
 | |
| 	return req, nil
 | |
| }
 | |
| 
 | |
| // Various signature types we are supporting, currently
 | |
| // two main signature types.
 | |
| type signerType int
 | |
| 
 | |
| const (
 | |
| 	signerV2 signerType = iota
 | |
| 	signerV4
 | |
| )
 | |
| 
 | |
| func newTestSignedRequest(method, urlStr string, contentLength int64, body io.ReadSeeker, accessKey, secretKey string, signer signerType) (*http.Request, error) {
 | |
| 	if signer == signerV2 {
 | |
| 		return newTestSignedRequestV2(method, urlStr, contentLength, body, accessKey, secretKey, nil)
 | |
| 	}
 | |
| 	return newTestSignedRequestV4(method, urlStr, contentLength, body, accessKey, secretKey, nil)
 | |
| }
 | |
| 
 | |
| // Returns request with correct signature but with incorrect SHA256.
 | |
| func newTestSignedBadSHARequest(method, urlStr string, contentLength int64, body io.ReadSeeker, accessKey, secretKey string, signer signerType) (*http.Request, error) {
 | |
| 	req, err := newTestRequest(method, urlStr, contentLength, body)
 | |
| 	if err != nil {
 | |
| 		return nil, err
 | |
| 	}
 | |
| 
 | |
| 	// Anonymous request return early.
 | |
| 	if accessKey == "" || secretKey == "" {
 | |
| 		return req, nil
 | |
| 	}
 | |
| 
 | |
| 	if signer == signerV2 {
 | |
| 		err = signRequestV2(req, accessKey, secretKey)
 | |
| 		req.Header.Del("x-amz-content-sha256")
 | |
| 	} else {
 | |
| 		req.Header.Set("x-amz-content-sha256", "92b165232fbd011da355eca0b033db22b934ba9af0145a437a832d27310b89f9")
 | |
| 		err = signRequestV4(req, accessKey, secretKey)
 | |
| 	}
 | |
| 
 | |
| 	return req, err
 | |
| }
 | |
| 
 | |
| // Returns new HTTP request object signed with signature v2.
 | |
| func newTestSignedRequestV2(method, urlStr string, contentLength int64, body io.ReadSeeker, accessKey, secretKey string, headers map[string]string) (*http.Request, error) {
 | |
| 	req, err := newTestRequest(method, urlStr, contentLength, body)
 | |
| 	if err != nil {
 | |
| 		return nil, err
 | |
| 	}
 | |
| 	req.Header.Del("x-amz-content-sha256")
 | |
| 
 | |
| 	// Anonymous request return quickly.
 | |
| 	if accessKey == "" || secretKey == "" {
 | |
| 		return req, nil
 | |
| 	}
 | |
| 
 | |
| 	for k, v := range headers {
 | |
| 		req.Header.Set(k, v)
 | |
| 	}
 | |
| 
 | |
| 	err = signRequestV2(req, accessKey, secretKey)
 | |
| 	if err != nil {
 | |
| 		return nil, err
 | |
| 	}
 | |
| 
 | |
| 	return req, nil
 | |
| }
 | |
| 
 | |
| // Returns new HTTP request object signed with signature v4.
 | |
| func newTestSignedRequestV4(method, urlStr string, contentLength int64, body io.ReadSeeker, accessKey, secretKey string, headers map[string]string) (*http.Request, error) {
 | |
| 	req, err := newTestRequest(method, urlStr, contentLength, body)
 | |
| 	if err != nil {
 | |
| 		return nil, err
 | |
| 	}
 | |
| 
 | |
| 	// Anonymous request return quickly.
 | |
| 	if accessKey == "" || secretKey == "" {
 | |
| 		return req, nil
 | |
| 	}
 | |
| 
 | |
| 	for k, v := range headers {
 | |
| 		req.Header.Set(k, v)
 | |
| 	}
 | |
| 
 | |
| 	err = signRequestV4(req, accessKey, secretKey)
 | |
| 	if err != nil {
 | |
| 		return nil, err
 | |
| 	}
 | |
| 
 | |
| 	return req, nil
 | |
| }
 | |
| 
 | |
| var src = rand.NewSource(time.Now().UnixNano())
 | |
| 
 | |
| func randString(n int) string {
 | |
| 	b := make([]byte, n)
 | |
| 	// A src.Int63() generates 63 random bits, enough for letterIdxMax characters!
 | |
| 	for i, cache, remain := n-1, src.Int63(), letterIdxMax; i >= 0; {
 | |
| 		if remain == 0 {
 | |
| 			cache, remain = src.Int63(), letterIdxMax
 | |
| 		}
 | |
| 		if idx := int(cache & letterIdxMask); idx < len(letterBytes) {
 | |
| 			b[i] = letterBytes[idx]
 | |
| 			i--
 | |
| 		}
 | |
| 		cache >>= letterIdxBits
 | |
| 		remain--
 | |
| 	}
 | |
| 
 | |
| 	return *(*string)(unsafe.Pointer(&b))
 | |
| }
 | |
| 
 | |
| // generate random object name.
 | |
| func getRandomObjectName() string {
 | |
| 	return randString(16)
 | |
| }
 | |
| 
 | |
| // generate random bucket name.
 | |
| func getRandomBucketName() string {
 | |
| 	return randString(60)
 | |
| }
 | |
| 
 | |
| // construct URL for http requests for bucket operations.
 | |
| func makeTestTargetURL(endPoint, bucketName, objectName string, queryValues url.Values) string {
 | |
| 	urlStr := endPoint + SlashSeparator
 | |
| 	if bucketName != "" {
 | |
| 		urlStr = urlStr + bucketName + SlashSeparator
 | |
| 	}
 | |
| 	if objectName != "" {
 | |
| 		urlStr += s3utils.EncodePath(objectName)
 | |
| 	}
 | |
| 	if len(queryValues) > 0 {
 | |
| 		urlStr = urlStr + "?" + queryValues.Encode()
 | |
| 	}
 | |
| 	return urlStr
 | |
| }
 | |
| 
 | |
| // return URL for uploading object into the bucket.
 | |
| func getPutObjectURL(endPoint, bucketName, objectName string) string {
 | |
| 	return makeTestTargetURL(endPoint, bucketName, objectName, url.Values{})
 | |
| }
 | |
| 
 | |
| func getPutObjectPartURL(endPoint, bucketName, objectName, uploadID, partNumber string) string {
 | |
| 	queryValues := url.Values{}
 | |
| 	queryValues.Set("uploadId", uploadID)
 | |
| 	queryValues.Set("partNumber", partNumber)
 | |
| 	return makeTestTargetURL(endPoint, bucketName, objectName, queryValues)
 | |
| }
 | |
| 
 | |
| func getCopyObjectPartURL(endPoint, bucketName, objectName, uploadID, partNumber string) string {
 | |
| 	queryValues := url.Values{}
 | |
| 	queryValues.Set("uploadId", uploadID)
 | |
| 	queryValues.Set("partNumber", partNumber)
 | |
| 	return makeTestTargetURL(endPoint, bucketName, objectName, queryValues)
 | |
| }
 | |
| 
 | |
| // return URL for fetching object from the bucket.
 | |
| func getGetObjectURL(endPoint, bucketName, objectName string) string {
 | |
| 	return makeTestTargetURL(endPoint, bucketName, objectName, url.Values{})
 | |
| }
 | |
| 
 | |
| // return URL for deleting the object from the bucket.
 | |
| func getDeleteObjectURL(endPoint, bucketName, objectName string) string {
 | |
| 	return makeTestTargetURL(endPoint, bucketName, objectName, url.Values{})
 | |
| }
 | |
| 
 | |
| // return URL for deleting multiple objects from a bucket.
 | |
| func getMultiDeleteObjectURL(endPoint, bucketName string) string {
 | |
| 	queryValue := url.Values{}
 | |
| 	queryValue.Set("delete", "")
 | |
| 	return makeTestTargetURL(endPoint, bucketName, "", queryValue)
 | |
| }
 | |
| 
 | |
| // return URL for HEAD on the object.
 | |
| func getHeadObjectURL(endPoint, bucketName, objectName string) string {
 | |
| 	return makeTestTargetURL(endPoint, bucketName, objectName, url.Values{})
 | |
| }
 | |
| 
 | |
| // return url to be used while copying the object.
 | |
| func getCopyObjectURL(endPoint, bucketName, objectName string) string {
 | |
| 	return makeTestTargetURL(endPoint, bucketName, objectName, url.Values{})
 | |
| }
 | |
| 
 | |
| // return URL for inserting bucket notification.
 | |
| func getPutNotificationURL(endPoint, bucketName string) string {
 | |
| 	queryValue := url.Values{}
 | |
| 	queryValue.Set("notification", "")
 | |
| 	return makeTestTargetURL(endPoint, bucketName, "", queryValue)
 | |
| }
 | |
| 
 | |
| // return URL for inserting bucket policy.
 | |
| func getPutPolicyURL(endPoint, bucketName string) string {
 | |
| 	queryValue := url.Values{}
 | |
| 	queryValue.Set("policy", "")
 | |
| 	return makeTestTargetURL(endPoint, bucketName, "", queryValue)
 | |
| }
 | |
| 
 | |
| // return URL for fetching bucket policy.
 | |
| func getGetPolicyURL(endPoint, bucketName string) string {
 | |
| 	queryValue := url.Values{}
 | |
| 	queryValue.Set("policy", "")
 | |
| 	return makeTestTargetURL(endPoint, bucketName, "", queryValue)
 | |
| }
 | |
| 
 | |
| // return URL for deleting bucket policy.
 | |
| func getDeletePolicyURL(endPoint, bucketName string) string {
 | |
| 	queryValue := url.Values{}
 | |
| 	queryValue.Set("policy", "")
 | |
| 	return makeTestTargetURL(endPoint, bucketName, "", queryValue)
 | |
| }
 | |
| 
 | |
| // return URL for creating the bucket.
 | |
| func getMakeBucketURL(endPoint, bucketName string) string {
 | |
| 	return makeTestTargetURL(endPoint, bucketName, "", url.Values{})
 | |
| }
 | |
| 
 | |
| // return URL for creating the bucket.
 | |
| func getBucketVersioningConfigURL(endPoint, bucketName string) string {
 | |
| 	vals := make(url.Values)
 | |
| 	vals.Set("versioning", "")
 | |
| 	return makeTestTargetURL(endPoint, bucketName, "", vals)
 | |
| }
 | |
| 
 | |
| // return URL for listing buckets.
 | |
| func getListBucketURL(endPoint string) string {
 | |
| 	return makeTestTargetURL(endPoint, "", "", url.Values{})
 | |
| }
 | |
| 
 | |
| // return URL for HEAD on the bucket.
 | |
| func getHEADBucketURL(endPoint, bucketName string) string {
 | |
| 	return makeTestTargetURL(endPoint, bucketName, "", url.Values{})
 | |
| }
 | |
| 
 | |
| // return URL for deleting the bucket.
 | |
| func getDeleteBucketURL(endPoint, bucketName string) string {
 | |
| 	return makeTestTargetURL(endPoint, bucketName, "", url.Values{})
 | |
| }
 | |
| 
 | |
| // return URL for deleting the bucket.
 | |
| func getDeleteMultipleObjectsURL(endPoint, bucketName string) string {
 | |
| 	queryValue := url.Values{}
 | |
| 	queryValue.Set("delete", "")
 | |
| 	return makeTestTargetURL(endPoint, bucketName, "", queryValue)
 | |
| }
 | |
| 
 | |
| // return URL For fetching location of the bucket.
 | |
| func getBucketLocationURL(endPoint, bucketName string) string {
 | |
| 	queryValue := url.Values{}
 | |
| 	queryValue.Set("location", "")
 | |
| 	return makeTestTargetURL(endPoint, bucketName, "", queryValue)
 | |
| }
 | |
| 
 | |
| // return URL For set/get lifecycle of the bucket.
 | |
| func getBucketLifecycleURL(endPoint, bucketName string) (ret string) {
 | |
| 	queryValue := url.Values{}
 | |
| 	queryValue.Set("lifecycle", "")
 | |
| 	return makeTestTargetURL(endPoint, bucketName, "", queryValue)
 | |
| }
 | |
| 
 | |
| // return URL for listing objects in the bucket with V1 legacy API.
 | |
| func getListObjectsV1URL(endPoint, bucketName, prefix, maxKeys, encodingType string) string {
 | |
| 	queryValue := url.Values{}
 | |
| 	if maxKeys != "" {
 | |
| 		queryValue.Set("max-keys", maxKeys)
 | |
| 	}
 | |
| 	if encodingType != "" {
 | |
| 		queryValue.Set("encoding-type", encodingType)
 | |
| 	}
 | |
| 	return makeTestTargetURL(endPoint, bucketName, prefix, queryValue)
 | |
| }
 | |
| 
 | |
| // return URL for listing objects in the bucket with V1 legacy API.
 | |
| func getListObjectVersionsURL(endPoint, bucketName, prefix, maxKeys, encodingType string) string {
 | |
| 	queryValue := url.Values{}
 | |
| 	if maxKeys != "" {
 | |
| 		queryValue.Set("max-keys", maxKeys)
 | |
| 	}
 | |
| 	if encodingType != "" {
 | |
| 		queryValue.Set("encoding-type", encodingType)
 | |
| 	}
 | |
| 	queryValue.Set("versions", "")
 | |
| 	return makeTestTargetURL(endPoint, bucketName, prefix, queryValue)
 | |
| }
 | |
| 
 | |
| // return URL for listing objects in the bucket with V2 API.
 | |
| func getListObjectsV2URL(endPoint, bucketName, prefix, maxKeys, fetchOwner, encodingType string) string {
 | |
| 	queryValue := url.Values{}
 | |
| 	queryValue.Set("list-type", "2") // Enables list objects V2 URL.
 | |
| 	if maxKeys != "" {
 | |
| 		queryValue.Set("max-keys", maxKeys)
 | |
| 	}
 | |
| 	if fetchOwner != "" {
 | |
| 		queryValue.Set("fetch-owner", fetchOwner)
 | |
| 	}
 | |
| 	if encodingType != "" {
 | |
| 		queryValue.Set("encoding-type", encodingType)
 | |
| 	}
 | |
| 	return makeTestTargetURL(endPoint, bucketName, prefix, queryValue)
 | |
| }
 | |
| 
 | |
| // return URL for a new multipart upload.
 | |
| func getNewMultipartURL(endPoint, bucketName, objectName string) string {
 | |
| 	queryValue := url.Values{}
 | |
| 	queryValue.Set("uploads", "")
 | |
| 	return makeTestTargetURL(endPoint, bucketName, objectName, queryValue)
 | |
| }
 | |
| 
 | |
| // return URL for a new multipart upload.
 | |
| func getPartUploadURL(endPoint, bucketName, objectName, uploadID, partNumber string) string {
 | |
| 	queryValues := url.Values{}
 | |
| 	queryValues.Set("uploadId", uploadID)
 | |
| 	queryValues.Set("partNumber", partNumber)
 | |
| 	return makeTestTargetURL(endPoint, bucketName, objectName, queryValues)
 | |
| }
 | |
| 
 | |
| // return URL for aborting multipart upload.
 | |
| func getAbortMultipartUploadURL(endPoint, bucketName, objectName, uploadID string) string {
 | |
| 	queryValue := url.Values{}
 | |
| 	queryValue.Set("uploadId", uploadID)
 | |
| 	return makeTestTargetURL(endPoint, bucketName, objectName, queryValue)
 | |
| }
 | |
| 
 | |
| // return URL for a listing pending multipart uploads.
 | |
| func getListMultipartURL(endPoint, bucketName string) string {
 | |
| 	queryValue := url.Values{}
 | |
| 	queryValue.Set("uploads", "")
 | |
| 	return makeTestTargetURL(endPoint, bucketName, "", queryValue)
 | |
| }
 | |
| 
 | |
| // return URL for listing pending multipart uploads with parameters.
 | |
| func getListMultipartUploadsURLWithParams(endPoint, bucketName, prefix, keyMarker, uploadIDMarker, delimiter, maxUploads string) string {
 | |
| 	queryValue := url.Values{}
 | |
| 	queryValue.Set("uploads", "")
 | |
| 	queryValue.Set("prefix", prefix)
 | |
| 	queryValue.Set("delimiter", delimiter)
 | |
| 	queryValue.Set("key-marker", keyMarker)
 | |
| 	queryValue.Set("upload-id-marker", uploadIDMarker)
 | |
| 	queryValue.Set("max-uploads", maxUploads)
 | |
| 	return makeTestTargetURL(endPoint, bucketName, "", queryValue)
 | |
| }
 | |
| 
 | |
| // return URL for a listing parts on a given upload id.
 | |
| func getListMultipartURLWithParams(endPoint, bucketName, objectName, uploadID, maxParts, partNumberMarker, encoding string) string {
 | |
| 	queryValues := url.Values{}
 | |
| 	queryValues.Set("uploadId", uploadID)
 | |
| 	queryValues.Set("max-parts", maxParts)
 | |
| 	if partNumberMarker != "" {
 | |
| 		queryValues.Set("part-number-marker", partNumberMarker)
 | |
| 	}
 | |
| 	return makeTestTargetURL(endPoint, bucketName, objectName, queryValues)
 | |
| }
 | |
| 
 | |
| // return URL for completing multipart upload.
 | |
| // complete multipart upload request is sent after all parts are uploaded.
 | |
| func getCompleteMultipartUploadURL(endPoint, bucketName, objectName, uploadID string) string {
 | |
| 	queryValue := url.Values{}
 | |
| 	queryValue.Set("uploadId", uploadID)
 | |
| 	return makeTestTargetURL(endPoint, bucketName, objectName, queryValue)
 | |
| }
 | |
| 
 | |
| // return URL for listen bucket notification.
 | |
| func getListenNotificationURL(endPoint, bucketName string, prefixes, suffixes, events []string) string {
 | |
| 	queryValue := url.Values{}
 | |
| 
 | |
| 	queryValue["prefix"] = prefixes
 | |
| 	queryValue["suffix"] = suffixes
 | |
| 	queryValue["events"] = events
 | |
| 	return makeTestTargetURL(endPoint, bucketName, "", queryValue)
 | |
| }
 | |
| 
 | |
| // getRandomDisks - Creates a slice of N random disks, each of the form - minio-XXX
 | |
| func getRandomDisks(N int) ([]string, error) {
 | |
| 	var erasureDisks []string
 | |
| 	for i := 0; i < N; i++ {
 | |
| 		path, err := os.MkdirTemp(globalTestTmpDir, "minio-")
 | |
| 		if err != nil {
 | |
| 			// Remove directories created so far.
 | |
| 			removeRoots(erasureDisks)
 | |
| 			return nil, err
 | |
| 		}
 | |
| 		erasureDisks = append(erasureDisks, path)
 | |
| 	}
 | |
| 	return erasureDisks, nil
 | |
| }
 | |
| 
 | |
| // Initialize object layer with the supplied disks, objectLayer is nil upon any error.
 | |
| func newTestObjectLayer(ctx context.Context, endpointServerPools EndpointServerPools) (newObject ObjectLayer, err error) {
 | |
| 	initAllSubsystems(ctx)
 | |
| 
 | |
| 	return newErasureServerPools(ctx, endpointServerPools)
 | |
| }
 | |
| 
 | |
| // initObjectLayer - Instantiates object layer and returns it.
 | |
| func initObjectLayer(ctx context.Context, endpointServerPools EndpointServerPools) (ObjectLayer, []StorageAPI, error) {
 | |
| 	objLayer, err := newTestObjectLayer(ctx, endpointServerPools)
 | |
| 	if err != nil {
 | |
| 		return nil, nil, err
 | |
| 	}
 | |
| 
 | |
| 	var formattedDisks []StorageAPI
 | |
| 	// Should use the object layer tests for validating cache.
 | |
| 	if z, ok := objLayer.(*erasureServerPools); ok {
 | |
| 		formattedDisks = z.serverPools[0].GetDisks(0)()
 | |
| 	}
 | |
| 
 | |
| 	// Success.
 | |
| 	return objLayer, formattedDisks, nil
 | |
| }
 | |
| 
 | |
| // removeRoots - Cleans up initialized directories during tests.
 | |
| func removeRoots(roots []string) {
 | |
| 	for _, root := range roots {
 | |
| 		os.RemoveAll(root)
 | |
| 	}
 | |
| }
 | |
| 
 | |
| // removeDiskN - removes N disks from supplied disk slice.
 | |
| func removeDiskN(disks []string, n int) {
 | |
| 	if n > len(disks) {
 | |
| 		n = len(disks)
 | |
| 	}
 | |
| 	for _, disk := range disks[:n] {
 | |
| 		os.RemoveAll(disk)
 | |
| 	}
 | |
| }
 | |
| 
 | |
| // creates a bucket for the tests and returns the bucket name.
 | |
| // initializes the specified API endpoints for the tests.
 | |
| // initialies the root and returns its path.
 | |
| // return credentials.
 | |
| func initAPIHandlerTest(ctx context.Context, obj ObjectLayer, endpoints []string) (string, http.Handler, error) {
 | |
| 	initAllSubsystems(ctx)
 | |
| 
 | |
| 	initConfigSubsystem(ctx, obj)
 | |
| 
 | |
| 	globalIAMSys.Init(ctx, obj, globalEtcdClient, 2*time.Second)
 | |
| 
 | |
| 	// get random bucket name.
 | |
| 	bucketName := getRandomBucketName()
 | |
| 
 | |
| 	// Create bucket.
 | |
| 	err := obj.MakeBucketWithLocation(context.Background(), bucketName, MakeBucketOptions{})
 | |
| 	if err != nil {
 | |
| 		// failed to create newbucket, return err.
 | |
| 		return "", nil, err
 | |
| 	}
 | |
| 	// Register the API end points with Erasure object layer.
 | |
| 	// Registering only the GetObject handler.
 | |
| 	apiRouter := initTestAPIEndPoints(obj, endpoints)
 | |
| 	f := http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {
 | |
| 		r.RequestURI = r.URL.RequestURI()
 | |
| 		apiRouter.ServeHTTP(w, r)
 | |
| 	})
 | |
| 	return bucketName, f, nil
 | |
| }
 | |
| 
 | |
| // prepare test backend.
 | |
| // create FS/Erasure/ErasureSet backend.
 | |
| // return object layer, backend disks.
 | |
| func prepareTestBackend(ctx context.Context, instanceType string) (ObjectLayer, []string, error) {
 | |
| 	switch instanceType {
 | |
| 	// Total number of disks for Erasure sets backend is set to 32.
 | |
| 	case ErasureSetsTestStr:
 | |
| 		return prepareErasureSets32(ctx)
 | |
| 	// Total number of disks for Erasure backend is set to 16.
 | |
| 	case ErasureTestStr:
 | |
| 		return prepareErasure16(ctx)
 | |
| 	default:
 | |
| 		// return FS backend by default.
 | |
| 		obj, disk, err := prepareFS(ctx)
 | |
| 		if err != nil {
 | |
| 			return nil, nil, err
 | |
| 		}
 | |
| 		return obj, []string{disk}, nil
 | |
| 	}
 | |
| }
 | |
| 
 | |
| // ExecObjectLayerAPIAnonTest - Helper function to validate object Layer API handler
 | |
| // response for anonymous/unsigned and unknown signature type HTTP request.
 | |
| 
 | |
| // Here is the brief description of some of the arguments to the function below.
 | |
| //
 | |
| //	apiRouter - http.Handler with the relevant API endPoint (API endPoint under test) registered.
 | |
| //	anonReq   - unsigned *http.Request to invoke the handler's response for anonymous requests.
 | |
| //	policyFunc    - function to return bucketPolicy statement which would permit the anonymous request to be served.
 | |
| //
 | |
| // The test works in 2 steps, here is the description of the steps.
 | |
| //
 | |
| //	STEP 1: Call the handler with the unsigned HTTP request (anonReq), assert for the `ErrAccessDenied` error response.
 | |
| func ExecObjectLayerAPIAnonTest(t *testing.T, obj ObjectLayer, testName, bucketName, objectName, instanceType string, apiRouter http.Handler,
 | |
| 	anonReq *http.Request, bucketPolicy *policy.Policy,
 | |
| ) {
 | |
| 	anonTestStr := "Anonymous HTTP request test"
 | |
| 	unknownSignTestStr := "Unknown HTTP signature test"
 | |
| 
 | |
| 	// simple function which returns a message which gives the context of the test
 | |
| 	// and then followed by the actual error message.
 | |
| 	failTestStr := func(testType, failMsg string) string {
 | |
| 		return fmt.Sprintf("MinIO %s: %s fail for \"%s\": \n<Error> %s", instanceType, testType, testName, failMsg)
 | |
| 	}
 | |
| 
 | |
| 	// httptest Recorder to capture all the response by the http handler.
 | |
| 	rec := httptest.NewRecorder()
 | |
| 	// reading the body to preserve it so that it can be used again for second attempt of sending unsigned HTTP request.
 | |
| 	// If the body is read in the handler the same request cannot be made use of.
 | |
| 	buf, err := io.ReadAll(anonReq.Body)
 | |
| 	if err != nil {
 | |
| 		t.Fatal(failTestStr(anonTestStr, err.Error()))
 | |
| 	}
 | |
| 
 | |
| 	// creating 2 read closer (to set as request body) from the body content.
 | |
| 	readerOne := io.NopCloser(bytes.NewBuffer(buf))
 | |
| 	readerTwo := io.NopCloser(bytes.NewBuffer(buf))
 | |
| 
 | |
| 	anonReq.Body = readerOne
 | |
| 
 | |
| 	// call the HTTP handler.
 | |
| 	apiRouter.ServeHTTP(rec, anonReq)
 | |
| 
 | |
| 	// expected error response when the unsigned HTTP request is not permitted.
 | |
| 	accessDenied := getAPIError(ErrAccessDenied).HTTPStatusCode
 | |
| 	if rec.Code != accessDenied {
 | |
| 		t.Fatal(failTestStr(anonTestStr, fmt.Sprintf("Object API Nil Test expected to fail with %d, but failed with %d", accessDenied, rec.Code)))
 | |
| 	}
 | |
| 
 | |
| 	// HEAD HTTTP request doesn't contain response body.
 | |
| 	if anonReq.Method != http.MethodHead {
 | |
| 		// read the response body.
 | |
| 		var actualContent []byte
 | |
| 		actualContent, err = io.ReadAll(rec.Body)
 | |
| 		if err != nil {
 | |
| 			t.Fatal(failTestStr(anonTestStr, fmt.Sprintf("Failed parsing response body: <ERROR> %v", err)))
 | |
| 		}
 | |
| 
 | |
| 		actualError := &APIErrorResponse{}
 | |
| 		if err = xml.Unmarshal(actualContent, actualError); err != nil {
 | |
| 			t.Fatal(failTestStr(anonTestStr, "error response failed to parse error XML"))
 | |
| 		}
 | |
| 
 | |
| 		if actualError.BucketName != bucketName {
 | |
| 			t.Fatal(failTestStr(anonTestStr, "error response bucket name differs from expected value"))
 | |
| 		}
 | |
| 
 | |
| 		if actualError.Key != objectName {
 | |
| 			t.Fatal(failTestStr(anonTestStr, "error response object name differs from expected value"))
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	// test for unknown auth case.
 | |
| 	anonReq.Body = readerTwo
 | |
| 	// Setting the `Authorization` header to a random value so that the signature falls into unknown auth case.
 | |
| 	anonReq.Header.Set("Authorization", "nothingElse")
 | |
| 	// initialize new response recorder.
 | |
| 	rec = httptest.NewRecorder()
 | |
| 	// call the handler using the HTTP Request.
 | |
| 	apiRouter.ServeHTTP(rec, anonReq)
 | |
| 	// verify the response body for `ErrAccessDenied` message =.
 | |
| 	if anonReq.Method != http.MethodHead {
 | |
| 		// read the response body.
 | |
| 		actualContent, err := io.ReadAll(rec.Body)
 | |
| 		if err != nil {
 | |
| 			t.Fatal(failTestStr(unknownSignTestStr, fmt.Sprintf("Failed parsing response body: <ERROR> %v", err)))
 | |
| 		}
 | |
| 
 | |
| 		actualError := &APIErrorResponse{}
 | |
| 		if err = xml.Unmarshal(actualContent, actualError); err != nil {
 | |
| 			t.Fatal(failTestStr(unknownSignTestStr, "error response failed to parse error XML"))
 | |
| 		}
 | |
| 
 | |
| 		if path.Clean(actualError.Resource) != pathJoin(SlashSeparator, bucketName, SlashSeparator, objectName) {
 | |
| 			t.Fatal(failTestStr(unknownSignTestStr, "error response resource differs from expected value"))
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	// expected error response when the unsigned HTTP request is not permitted.
 | |
| 	unsupportedSignature := getAPIError(ErrSignatureVersionNotSupported).HTTPStatusCode
 | |
| 	if rec.Code != unsupportedSignature {
 | |
| 		t.Fatal(failTestStr(unknownSignTestStr, fmt.Sprintf("Object API Unknow auth test for \"%s\", expected to fail with %d, but failed with %d", testName, unsupportedSignature, rec.Code)))
 | |
| 	}
 | |
| }
 | |
| 
 | |
| // ExecObjectLayerAPINilTest - Sets the object layer to `nil`, and calls rhe registered object layer API endpoint,
 | |
| // and assert the error response. The purpose is to validate the API handlers response when the object layer is uninitialized.
 | |
| // Usage hint: Should be used at the end of the API end points tests (ex: check the last few lines of `testAPIListObjectPartsHandler`),
 | |
| // need a sample HTTP request to be sent as argument so that the relevant handler is called, the handler registration is expected
 | |
| // to be done since its called from within the API handler tests, the reference to the registered HTTP handler has to be sent
 | |
| // as an argument.
 | |
| func ExecObjectLayerAPINilTest(t TestErrHandler, bucketName, objectName, instanceType string, apiRouter http.Handler, req *http.Request) {
 | |
| 	// httptest Recorder to capture all the response by the http handler.
 | |
| 	rec := httptest.NewRecorder()
 | |
| 
 | |
| 	// The  API handler gets the referece to the object layer via the global object Layer,
 | |
| 	// setting it to `nil` in order test for handlers response for uninitialized object layer.
 | |
| 	globalObjLayerMutex.Lock()
 | |
| 	globalObjectAPI = nil
 | |
| 	globalObjLayerMutex.Unlock()
 | |
| 
 | |
| 	// call the HTTP handler.
 | |
| 	apiRouter.ServeHTTP(rec, req)
 | |
| 
 | |
| 	// expected error response when the API handler is called before the object layer is initialized,
 | |
| 	// or when objectLayer is `nil`.
 | |
| 	serverNotInitializedErr := getAPIError(ErrServerNotInitialized).HTTPStatusCode
 | |
| 	if rec.Code != serverNotInitializedErr {
 | |
| 		t.Errorf("Object API Nil Test expected to fail with %d, but failed with %d", serverNotInitializedErr, rec.Code)
 | |
| 	}
 | |
| 
 | |
| 	// HEAD HTTP Request doesn't contain body in its response,
 | |
| 	// for other type of HTTP requests compare the response body content with the expected one.
 | |
| 	if req.Method != http.MethodHead {
 | |
| 		// read the response body.
 | |
| 		actualContent, err := io.ReadAll(rec.Body)
 | |
| 		if err != nil {
 | |
| 			t.Fatalf("MinIO %s: Failed parsing response body: <ERROR> %v", instanceType, err)
 | |
| 		}
 | |
| 
 | |
| 		actualError := &APIErrorResponse{}
 | |
| 		if err = xml.Unmarshal(actualContent, actualError); err != nil {
 | |
| 			t.Errorf("MinIO %s: error response failed to parse error XML", instanceType)
 | |
| 		}
 | |
| 
 | |
| 		if actualError.BucketName != bucketName {
 | |
| 			t.Errorf("MinIO %s: error response bucket name differs from expected value", instanceType)
 | |
| 		}
 | |
| 
 | |
| 		if actualError.Key != objectName {
 | |
| 			t.Errorf("MinIO %s: error response object name differs from expected value", instanceType)
 | |
| 		}
 | |
| 	}
 | |
| }
 | |
| 
 | |
| // ExecObjectLayerAPITest - executes object layer API tests.
 | |
| // Creates single node and Erasure ObjectLayer instance, registers the specified API end points and runs test for both the layers.
 | |
| func ExecObjectLayerAPITest(t *testing.T, objAPITest objAPITestType, endpoints []string) {
 | |
| 	ctx, cancel := context.WithCancel(context.Background())
 | |
| 	defer cancel()
 | |
| 
 | |
| 	// reset globals.
 | |
| 	// this is to make sure that the tests are not affected by modified value.
 | |
| 	resetTestGlobals()
 | |
| 
 | |
| 	objLayer, fsDir, err := prepareFS(ctx)
 | |
| 	if err != nil {
 | |
| 		t.Fatalf("Initialization of object layer failed for single node setup: %s", err)
 | |
| 	}
 | |
| 
 | |
| 	bucketFS, fsAPIRouter, err := initAPIHandlerTest(ctx, objLayer, endpoints)
 | |
| 	if err != nil {
 | |
| 		t.Fatalf("Initialization of API handler tests failed: <ERROR> %s", err)
 | |
| 	}
 | |
| 
 | |
| 	// initialize the server and obtain the credentials and root.
 | |
| 	// credentials are necessary to sign the HTTP request.
 | |
| 	if err = newTestConfig(globalMinioDefaultRegion, objLayer); err != nil {
 | |
| 		t.Fatalf("Unable to initialize server config. %s", err)
 | |
| 	}
 | |
| 
 | |
| 	credentials := globalActiveCred
 | |
| 
 | |
| 	// Executing the object layer tests for single node setup.
 | |
| 	objAPITest(objLayer, ErasureSDStr, bucketFS, fsAPIRouter, credentials, t)
 | |
| 
 | |
| 	// reset globals.
 | |
| 	// this is to make sure that the tests are not affected by modified value.
 | |
| 	resetTestGlobals()
 | |
| 
 | |
| 	objLayer, erasureDisks, err := prepareErasure16(ctx)
 | |
| 	if err != nil {
 | |
| 		t.Fatalf("Initialization of object layer failed for Erasure setup: %s", err)
 | |
| 	}
 | |
| 	defer objLayer.Shutdown(ctx)
 | |
| 
 | |
| 	bucketErasure, erAPIRouter, err := initAPIHandlerTest(ctx, objLayer, endpoints)
 | |
| 	if err != nil {
 | |
| 		t.Fatalf("Initialzation of API handler tests failed: <ERROR> %s", err)
 | |
| 	}
 | |
| 
 | |
| 	// initialize the server and obtain the credentials and root.
 | |
| 	// credentials are necessary to sign the HTTP request.
 | |
| 	if err = newTestConfig(globalMinioDefaultRegion, objLayer); err != nil {
 | |
| 		t.Fatalf("Unable to initialize server config. %s", err)
 | |
| 	}
 | |
| 
 | |
| 	// Executing the object layer tests for Erasure.
 | |
| 	objAPITest(objLayer, ErasureTestStr, bucketErasure, erAPIRouter, credentials, t)
 | |
| 
 | |
| 	// clean up the temporary test backend.
 | |
| 	removeRoots(append(erasureDisks, fsDir))
 | |
| }
 | |
| 
 | |
| // ExecExtendedObjectLayerTest will execute the tests with combinations of encrypted & compressed.
 | |
| // This can be used to test functionality when reading and writing data.
 | |
| func ExecExtendedObjectLayerAPITest(t *testing.T, objAPITest objAPITestType, endpoints []string) {
 | |
| 	execExtended(t, func(t *testing.T) {
 | |
| 		ExecObjectLayerAPITest(t, objAPITest, endpoints)
 | |
| 	})
 | |
| }
 | |
| 
 | |
| // function to be passed to ExecObjectLayerAPITest, for executing object layr API handler tests.
 | |
| type objAPITestType func(obj ObjectLayer, instanceType string, bucketName string,
 | |
| 	apiRouter http.Handler, credentials auth.Credentials, t *testing.T)
 | |
| 
 | |
| // Regular object test type.
 | |
| type objTestType func(obj ObjectLayer, instanceType string, t TestErrHandler)
 | |
| 
 | |
| // Special test type for test with directories
 | |
| type objTestTypeWithDirs func(obj ObjectLayer, instanceType string, dirs []string, t TestErrHandler)
 | |
| 
 | |
| // Special object test type for disk not found situations.
 | |
| type objTestDiskNotFoundType func(obj ObjectLayer, instanceType string, dirs []string, t *testing.T)
 | |
| 
 | |
| // ExecObjectLayerTest - executes object layer tests.
 | |
| // Creates single node and Erasure ObjectLayer instance and runs test for both the layers.
 | |
| func ExecObjectLayerTest(t TestErrHandler, objTest objTestType) {
 | |
| 	{
 | |
| 		ctx, cancel := context.WithCancel(context.Background())
 | |
| 		if localMetacacheMgr != nil {
 | |
| 			localMetacacheMgr.deleteAll()
 | |
| 		}
 | |
| 
 | |
| 		objLayer, fsDir, err := prepareFS(ctx)
 | |
| 		if err != nil {
 | |
| 			t.Fatalf("Initialization of object layer failed for single node setup: %s", err)
 | |
| 		}
 | |
| 		setObjectLayer(objLayer)
 | |
| 		initAllSubsystems(ctx)
 | |
| 
 | |
| 		// initialize the server and obtain the credentials and root.
 | |
| 		// credentials are necessary to sign the HTTP request.
 | |
| 		if err = newTestConfig(globalMinioDefaultRegion, objLayer); err != nil {
 | |
| 			t.Fatal("Unexpected error", err)
 | |
| 		}
 | |
| 		initConfigSubsystem(ctx, objLayer)
 | |
| 		globalIAMSys.Init(ctx, objLayer, globalEtcdClient, 2*time.Second)
 | |
| 
 | |
| 		// Executing the object layer tests for single node setup.
 | |
| 		objTest(objLayer, ErasureSDStr, t)
 | |
| 
 | |
| 		// Call clean up functions
 | |
| 		cancel()
 | |
| 		setObjectLayer(newObjectLayerFn())
 | |
| 		removeRoots([]string{fsDir})
 | |
| 	}
 | |
| 
 | |
| 	{
 | |
| 		ctx, cancel := context.WithCancel(context.Background())
 | |
| 
 | |
| 		if localMetacacheMgr != nil {
 | |
| 			localMetacacheMgr.deleteAll()
 | |
| 		}
 | |
| 
 | |
| 		initAllSubsystems(ctx)
 | |
| 		objLayer, fsDirs, err := prepareErasureSets32(ctx)
 | |
| 		if err != nil {
 | |
| 			t.Fatalf("Initialization of object layer failed for Erasure setup: %s", err)
 | |
| 		}
 | |
| 		setObjectLayer(objLayer)
 | |
| 		initConfigSubsystem(ctx, objLayer)
 | |
| 		globalIAMSys.Init(ctx, objLayer, globalEtcdClient, 2*time.Second)
 | |
| 
 | |
| 		// Executing the object layer tests for Erasure.
 | |
| 		objTest(objLayer, ErasureTestStr, t)
 | |
| 
 | |
| 		objLayer.Shutdown(context.Background())
 | |
| 		if localMetacacheMgr != nil {
 | |
| 			localMetacacheMgr.deleteAll()
 | |
| 		}
 | |
| 		setObjectLayer(newObjectLayerFn())
 | |
| 		cancel()
 | |
| 		removeRoots(fsDirs)
 | |
| 	}
 | |
| }
 | |
| 
 | |
| // ExecObjectLayerTestWithDirs - executes object layer tests.
 | |
| // Creates single node and Erasure ObjectLayer instance and runs test for both the layers.
 | |
| func ExecObjectLayerTestWithDirs(t TestErrHandler, objTest objTestTypeWithDirs) {
 | |
| 	ctx, cancel := context.WithCancel(context.Background())
 | |
| 	defer cancel()
 | |
| 
 | |
| 	objLayer, fsDirs, err := prepareErasure16(ctx)
 | |
| 	if err != nil {
 | |
| 		t.Fatalf("Initialization of object layer failed for Erasure setup: %s", err)
 | |
| 	}
 | |
| 	defer objLayer.Shutdown(ctx)
 | |
| 
 | |
| 	// initialize the server and obtain the credentials and root.
 | |
| 	// credentials are necessary to sign the HTTP request.
 | |
| 	if err = newTestConfig(globalMinioDefaultRegion, objLayer); err != nil {
 | |
| 		t.Fatal("Unexpected error", err)
 | |
| 	}
 | |
| 
 | |
| 	// Executing the object layer tests for Erasure.
 | |
| 	objTest(objLayer, ErasureTestStr, fsDirs, t)
 | |
| 	defer removeRoots(fsDirs)
 | |
| }
 | |
| 
 | |
| // ExecObjectLayerDiskAlteredTest - executes object layer tests while altering
 | |
| // disks in between tests. Creates Erasure ObjectLayer instance and runs test for Erasure layer.
 | |
| func ExecObjectLayerDiskAlteredTest(t *testing.T, objTest objTestDiskNotFoundType) {
 | |
| 	ctx, cancel := context.WithCancel(context.Background())
 | |
| 	defer cancel()
 | |
| 
 | |
| 	objLayer, fsDirs, err := prepareErasure16(ctx)
 | |
| 	if err != nil {
 | |
| 		t.Fatalf("Initialization of object layer failed for Erasure setup: %s", err)
 | |
| 	}
 | |
| 	defer objLayer.Shutdown(ctx)
 | |
| 
 | |
| 	if err = newTestConfig(globalMinioDefaultRegion, objLayer); err != nil {
 | |
| 		t.Fatal("Failed to create config directory", err)
 | |
| 	}
 | |
| 
 | |
| 	// Executing the object layer tests for Erasure.
 | |
| 	objTest(objLayer, ErasureTestStr, fsDirs, t)
 | |
| 	defer removeRoots(fsDirs)
 | |
| }
 | |
| 
 | |
| // Special object test type for stale files situations.
 | |
| type objTestStaleFilesType func(obj ObjectLayer, instanceType string, dirs []string, t *testing.T)
 | |
| 
 | |
| // ExecObjectLayerStaleFilesTest - executes object layer tests those leaves stale
 | |
| // files/directories under .minio/tmp.  Creates Erasure ObjectLayer instance and runs test for Erasure layer.
 | |
| func ExecObjectLayerStaleFilesTest(t *testing.T, objTest objTestStaleFilesType) {
 | |
| 	ctx, cancel := context.WithCancel(context.Background())
 | |
| 	defer cancel()
 | |
| 
 | |
| 	nDisks := 16
 | |
| 	erasureDisks, err := getRandomDisks(nDisks)
 | |
| 	if err != nil {
 | |
| 		t.Fatalf("Initialization of drives for Erasure setup: %s", err)
 | |
| 	}
 | |
| 	objLayer, _, err := initObjectLayer(ctx, mustGetPoolEndpoints(erasureDisks...))
 | |
| 	if err != nil {
 | |
| 		t.Fatalf("Initialization of object layer failed for Erasure setup: %s", err)
 | |
| 	}
 | |
| 	if err = newTestConfig(globalMinioDefaultRegion, objLayer); err != nil {
 | |
| 		t.Fatal("Failed to create config directory", err)
 | |
| 	}
 | |
| 
 | |
| 	// Executing the object layer tests for Erasure.
 | |
| 	objTest(objLayer, ErasureTestStr, erasureDisks, t)
 | |
| 	defer removeRoots(erasureDisks)
 | |
| }
 | |
| 
 | |
| func registerBucketLevelFunc(bucket *mux.Router, api objectAPIHandlers, apiFunctions ...string) {
 | |
| 	for _, apiFunction := range apiFunctions {
 | |
| 		switch apiFunction {
 | |
| 		case "PostPolicy":
 | |
| 			// Register PostPolicy handler.
 | |
| 			bucket.Methods(http.MethodPost).HeadersRegexp("Content-Type", "multipart/form-data*").HandlerFunc(api.PostPolicyBucketHandler)
 | |
| 		case "HeadObject":
 | |
| 			// Register HeadObject handler.
 | |
| 			bucket.Methods("Head").Path("/{object:.+}").HandlerFunc(api.HeadObjectHandler)
 | |
| 		case "GetObject":
 | |
| 			// Register GetObject handler.
 | |
| 			bucket.Methods(http.MethodGet).Path("/{object:.+}").HandlerFunc(api.GetObjectHandler)
 | |
| 		case "PutObject":
 | |
| 			// Register PutObject handler.
 | |
| 			bucket.Methods(http.MethodPut).Path("/{object:.+}").HandlerFunc(api.PutObjectHandler)
 | |
| 		case "DeleteObject":
 | |
| 			// Register Delete Object handler.
 | |
| 			bucket.Methods(http.MethodDelete).Path("/{object:.+}").HandlerFunc(api.DeleteObjectHandler)
 | |
| 		case "CopyObject":
 | |
| 			// Register Copy Object  handler.
 | |
| 			bucket.Methods(http.MethodPut).Path("/{object:.+}").HeadersRegexp("X-Amz-Copy-Source", ".*?(\\/|%2F).*?").HandlerFunc(api.CopyObjectHandler)
 | |
| 		case "PutBucketPolicy":
 | |
| 			// Register PutBucket Policy handler.
 | |
| 			bucket.Methods(http.MethodPut).HandlerFunc(api.PutBucketPolicyHandler).Queries("policy", "")
 | |
| 		case "DeleteBucketPolicy":
 | |
| 			// Register Delete bucket HTTP policy handler.
 | |
| 			bucket.Methods(http.MethodDelete).HandlerFunc(api.DeleteBucketPolicyHandler).Queries("policy", "")
 | |
| 		case "GetBucketPolicy":
 | |
| 			// Register Get Bucket policy HTTP Handler.
 | |
| 			bucket.Methods(http.MethodGet).HandlerFunc(api.GetBucketPolicyHandler).Queries("policy", "")
 | |
| 		case "GetBucketLifecycle":
 | |
| 			bucket.Methods(http.MethodGet).HandlerFunc(api.GetBucketLifecycleHandler).Queries("lifecycle", "")
 | |
| 		case "PutBucketLifecycle":
 | |
| 			bucket.Methods(http.MethodPut).HandlerFunc(api.PutBucketLifecycleHandler).Queries("lifecycle", "")
 | |
| 		case "DeleteBucketLifecycle":
 | |
| 			bucket.Methods(http.MethodDelete).HandlerFunc(api.DeleteBucketLifecycleHandler).Queries("lifecycle", "")
 | |
| 		case "GetBucketLocation":
 | |
| 			// Register GetBucketLocation handler.
 | |
| 			bucket.Methods(http.MethodGet).HandlerFunc(api.GetBucketLocationHandler).Queries("location", "")
 | |
| 		case "HeadBucket":
 | |
| 			// Register HeadBucket handler.
 | |
| 			bucket.Methods(http.MethodHead).HandlerFunc(api.HeadBucketHandler)
 | |
| 		case "DeleteMultipleObjects":
 | |
| 			// Register DeleteMultipleObjects handler.
 | |
| 			bucket.Methods(http.MethodPost).HandlerFunc(api.DeleteMultipleObjectsHandler).Queries("delete", "")
 | |
| 		case "NewMultipart":
 | |
| 			// Register New Multipart upload handler.
 | |
| 			bucket.Methods(http.MethodPost).Path("/{object:.+}").HandlerFunc(api.NewMultipartUploadHandler).Queries("uploads", "")
 | |
| 		case "CopyObjectPart":
 | |
| 			// Register CopyObjectPart handler.
 | |
| 			bucket.Methods(http.MethodPut).Path("/{object:.+}").HeadersRegexp("X-Amz-Copy-Source", ".*?(\\/|%2F).*?").HandlerFunc(api.CopyObjectPartHandler).Queries("partNumber", "{partNumber:[0-9]+}", "uploadId", "{uploadId:.*}")
 | |
| 		case "PutObjectPart":
 | |
| 			// Register PutObjectPart handler.
 | |
| 			bucket.Methods(http.MethodPut).Path("/{object:.+}").HandlerFunc(api.PutObjectPartHandler).Queries("partNumber", "{partNumber:[0-9]+}", "uploadId", "{uploadId:.*}")
 | |
| 		case "ListObjectParts":
 | |
| 			// Register ListObjectParts handler.
 | |
| 			bucket.Methods(http.MethodGet).Path("/{object:.+}").HandlerFunc(api.ListObjectPartsHandler).Queries("uploadId", "{uploadId:.*}")
 | |
| 		case "ListMultipartUploads":
 | |
| 			// Register ListMultipartUploads handler.
 | |
| 			bucket.Methods(http.MethodGet).HandlerFunc(api.ListMultipartUploadsHandler).Queries("uploads", "")
 | |
| 		case "CompleteMultipart":
 | |
| 			// Register Complete Multipart Upload handler.
 | |
| 			bucket.Methods(http.MethodPost).Path("/{object:.+}").HandlerFunc(api.CompleteMultipartUploadHandler).Queries("uploadId", "{uploadId:.*}")
 | |
| 		case "AbortMultipart":
 | |
| 			// Register AbortMultipart Handler.
 | |
| 			bucket.Methods(http.MethodDelete).Path("/{object:.+}").HandlerFunc(api.AbortMultipartUploadHandler).Queries("uploadId", "{uploadId:.*}")
 | |
| 		case "GetBucketNotification":
 | |
| 			// Register GetBucketNotification Handler.
 | |
| 			bucket.Methods(http.MethodGet).HandlerFunc(api.GetBucketNotificationHandler).Queries("notification", "")
 | |
| 		case "PutBucketNotification":
 | |
| 			// Register PutBucketNotification Handler.
 | |
| 			bucket.Methods(http.MethodPut).HandlerFunc(api.PutBucketNotificationHandler).Queries("notification", "")
 | |
| 		case "ListenNotification":
 | |
| 			// Register ListenNotification Handler.
 | |
| 			bucket.Methods(http.MethodGet).HandlerFunc(api.ListenNotificationHandler).Queries("events", "{events:.*}")
 | |
| 		}
 | |
| 	}
 | |
| }
 | |
| 
 | |
| // registerAPIFunctions helper function to add API functions identified by name to the routers.
 | |
| func registerAPIFunctions(muxRouter *mux.Router, objLayer ObjectLayer, apiFunctions ...string) {
 | |
| 	if len(apiFunctions) == 0 {
 | |
| 		// Register all api endpoints by default.
 | |
| 		registerAPIRouter(muxRouter)
 | |
| 		return
 | |
| 	}
 | |
| 	// API Router.
 | |
| 	apiRouter := muxRouter.PathPrefix(SlashSeparator).Subrouter()
 | |
| 	// Bucket router.
 | |
| 	bucketRouter := apiRouter.PathPrefix("/{bucket}").Subrouter()
 | |
| 
 | |
| 	// All object storage operations are registered as HTTP handlers on `objectAPIHandlers`.
 | |
| 	// When the handlers get a HTTP request they use the underlying ObjectLayer to perform operations.
 | |
| 	globalObjLayerMutex.Lock()
 | |
| 	globalObjectAPI = objLayer
 | |
| 	globalObjLayerMutex.Unlock()
 | |
| 
 | |
| 	// When cache is enabled, Put and Get operations are passed
 | |
| 	// to underlying cache layer to manage object layer operation and disk caching
 | |
| 	// operation
 | |
| 	api := objectAPIHandlers{
 | |
| 		ObjectAPI: func() ObjectLayer {
 | |
| 			return globalObjectAPI
 | |
| 		},
 | |
| 		CacheAPI: func() CacheObjectLayer {
 | |
| 			return globalCacheObjectAPI
 | |
| 		},
 | |
| 	}
 | |
| 
 | |
| 	// Register ListBuckets	handler.
 | |
| 	apiRouter.Methods(http.MethodGet).HandlerFunc(api.ListBucketsHandler)
 | |
| 	// Register all bucket level handlers.
 | |
| 	registerBucketLevelFunc(bucketRouter, api, apiFunctions...)
 | |
| }
 | |
| 
 | |
| // Takes in Erasure object layer, and the list of API end points to be tested/required, registers the API end points and returns the HTTP handler.
 | |
| // Need isolated registration of API end points while writing unit tests for end points.
 | |
| // All the API end points are registered only for the default case.
 | |
| func initTestAPIEndPoints(objLayer ObjectLayer, apiFunctions []string) http.Handler {
 | |
| 	// initialize a new mux router.
 | |
| 	// goriilla/mux is the library used to register all the routes and handle them.
 | |
| 	muxRouter := mux.NewRouter().SkipClean(true).UseEncodedPath()
 | |
| 	if len(apiFunctions) > 0 {
 | |
| 		// Iterate the list of API functions requested for and register them in mux HTTP handler.
 | |
| 		registerAPIFunctions(muxRouter, objLayer, apiFunctions...)
 | |
| 		muxRouter.Use(globalHandlers...)
 | |
| 		return muxRouter
 | |
| 	}
 | |
| 	registerAPIRouter(muxRouter)
 | |
| 	muxRouter.Use(globalHandlers...)
 | |
| 	return muxRouter
 | |
| }
 | |
| 
 | |
| // generateTLSCertKey creates valid key/cert with registered DNS or IP address
 | |
| // depending on the passed parameter. That way, we can use tls config without
 | |
| // passing InsecureSkipVerify flag.  This code is a simplified version of
 | |
| // https://golang.org/src/crypto/tls/generate_cert.go
 | |
| func generateTLSCertKey(host string) ([]byte, []byte, error) {
 | |
| 	validFor := 365 * 24 * time.Hour
 | |
| 	rsaBits := 2048
 | |
| 
 | |
| 	if len(host) == 0 {
 | |
| 		return nil, nil, fmt.Errorf("Missing host parameter")
 | |
| 	}
 | |
| 
 | |
| 	publicKey := func(priv interface{}) interface{} {
 | |
| 		switch k := priv.(type) {
 | |
| 		case *rsa.PrivateKey:
 | |
| 			return &k.PublicKey
 | |
| 		case *ecdsa.PrivateKey:
 | |
| 			return &k.PublicKey
 | |
| 		default:
 | |
| 			return nil
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	pemBlockForKey := func(priv interface{}) *pem.Block {
 | |
| 		switch k := priv.(type) {
 | |
| 		case *rsa.PrivateKey:
 | |
| 			return &pem.Block{Type: "RSA PRIVATE KEY", Bytes: x509.MarshalPKCS1PrivateKey(k)}
 | |
| 		case *ecdsa.PrivateKey:
 | |
| 			b, err := x509.MarshalECPrivateKey(k)
 | |
| 			if err != nil {
 | |
| 				fmt.Fprintf(os.Stderr, "Unable to marshal ECDSA private key: %v", err)
 | |
| 				os.Exit(2)
 | |
| 			}
 | |
| 			return &pem.Block{Type: "EC PRIVATE KEY", Bytes: b}
 | |
| 		default:
 | |
| 			return nil
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	var priv interface{}
 | |
| 	var err error
 | |
| 	priv, err = rsa.GenerateKey(crand.Reader, rsaBits)
 | |
| 	if err != nil {
 | |
| 		return nil, nil, fmt.Errorf("failed to generate private key: %w", err)
 | |
| 	}
 | |
| 
 | |
| 	notBefore := time.Now()
 | |
| 	notAfter := notBefore.Add(validFor)
 | |
| 
 | |
| 	serialNumberLimit := new(big.Int).Lsh(big.NewInt(1), 128)
 | |
| 	serialNumber, err := crand.Int(crand.Reader, serialNumberLimit)
 | |
| 	if err != nil {
 | |
| 		return nil, nil, fmt.Errorf("failed to generate serial number: %w", err)
 | |
| 	}
 | |
| 
 | |
| 	template := x509.Certificate{
 | |
| 		SerialNumber: serialNumber,
 | |
| 		Subject: pkix.Name{
 | |
| 			Organization: []string{"Acme Co"},
 | |
| 		},
 | |
| 		NotBefore: notBefore,
 | |
| 		NotAfter:  notAfter,
 | |
| 
 | |
| 		KeyUsage:              x509.KeyUsageKeyEncipherment | x509.KeyUsageDigitalSignature,
 | |
| 		ExtKeyUsage:           []x509.ExtKeyUsage{x509.ExtKeyUsageServerAuth},
 | |
| 		BasicConstraintsValid: true,
 | |
| 	}
 | |
| 
 | |
| 	hosts := strings.Split(host, ",")
 | |
| 	for _, h := range hosts {
 | |
| 		if ip := net.ParseIP(h); ip != nil {
 | |
| 			template.IPAddresses = append(template.IPAddresses, ip)
 | |
| 		} else {
 | |
| 			template.DNSNames = append(template.DNSNames, h)
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	template.IsCA = true
 | |
| 	template.KeyUsage |= x509.KeyUsageCertSign
 | |
| 
 | |
| 	derBytes, err := x509.CreateCertificate(crand.Reader, &template, &template, publicKey(priv), priv)
 | |
| 	if err != nil {
 | |
| 		return nil, nil, fmt.Errorf("Failed to create certificate: %w", err)
 | |
| 	}
 | |
| 
 | |
| 	certOut := bytes.NewBuffer([]byte{})
 | |
| 	pem.Encode(certOut, &pem.Block{Type: "CERTIFICATE", Bytes: derBytes})
 | |
| 
 | |
| 	keyOut := bytes.NewBuffer([]byte{})
 | |
| 	pem.Encode(keyOut, pemBlockForKey(priv))
 | |
| 
 | |
| 	return certOut.Bytes(), keyOut.Bytes(), nil
 | |
| }
 | |
| 
 | |
| func mustGetPoolEndpoints(args ...string) EndpointServerPools {
 | |
| 	endpoints := mustGetNewEndpoints(args...)
 | |
| 	drivesPerSet := len(args)
 | |
| 	setCount := 1
 | |
| 	if len(args) >= 16 {
 | |
| 		drivesPerSet = 16
 | |
| 		setCount = len(args) / 16
 | |
| 	}
 | |
| 	return []PoolEndpoints{{
 | |
| 		SetCount:     setCount,
 | |
| 		DrivesPerSet: drivesPerSet,
 | |
| 		Endpoints:    endpoints,
 | |
| 		CmdLine:      strings.Join(args, " "),
 | |
| 	}}
 | |
| }
 | |
| 
 | |
| func mustGetNewEndpoints(args ...string) (endpoints Endpoints) {
 | |
| 	endpoints, err := NewEndpoints(args...)
 | |
| 	logger.FatalIf(err, "unable to create new endpoint list")
 | |
| 	return endpoints
 | |
| }
 | |
| 
 | |
| func getEndpointsLocalAddr(endpointServerPools EndpointServerPools) string {
 | |
| 	for _, endpoints := range endpointServerPools {
 | |
| 		for _, endpoint := range endpoints.Endpoints {
 | |
| 			if endpoint.IsLocal && endpoint.Type() == URLEndpointType {
 | |
| 				return endpoint.Host
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return net.JoinHostPort(globalMinioHost, globalMinioPort)
 | |
| }
 | |
| 
 | |
| // fetches a random number between range min-max.
 | |
| func getRandomRange(min, max int, seed int64) int {
 | |
| 	// special value -1 means no explicit seeding.
 | |
| 	if seed != -1 {
 | |
| 		rand.Seed(seed)
 | |
| 	}
 | |
| 	return rand.Intn(max-min) + min
 | |
| }
 | |
| 
 | |
| // Randomizes the order of bytes in the byte array
 | |
| // using Knuth Fisher-Yates shuffle algorithm.
 | |
| func randomizeBytes(s []byte, seed int64) []byte {
 | |
| 	// special value -1 means no explicit seeding.
 | |
| 	if seed != -1 {
 | |
| 		rand.Seed(seed)
 | |
| 	}
 | |
| 	n := len(s)
 | |
| 	var j int
 | |
| 	for i := 0; i < n-1; i++ {
 | |
| 		j = i + rand.Intn(n-i)
 | |
| 		s[i], s[j] = s[j], s[i]
 | |
| 	}
 | |
| 	return s
 | |
| }
 | |
| 
 | |
| func TestToErrIsNil(t *testing.T) {
 | |
| 	if toObjectErr(nil) != nil {
 | |
| 		t.Errorf("Test expected to return nil, failed instead got a non-nil value %s", toObjectErr(nil))
 | |
| 	}
 | |
| 	if toStorageErr(nil) != nil {
 | |
| 		t.Errorf("Test expected to return nil, failed instead got a non-nil value %s", toStorageErr(nil))
 | |
| 	}
 | |
| 	ctx := context.Background()
 | |
| 	if toAPIError(ctx, nil) != noError {
 | |
| 		t.Errorf("Test expected error code to be ErrNone, failed instead provided %s", toAPIError(ctx, nil).Code)
 | |
| 	}
 | |
| }
 | |
| 
 | |
| // Uploads an object using DummyDataGen directly via the http
 | |
| // handler. Each part in a multipart object is a new DummyDataGen
 | |
| // instance (so the part sizes are needed to reconstruct the whole
 | |
| // object). When `len(partSizes) == 1`, asMultipart is used to upload
 | |
| // the object as multipart with 1 part or as a regular single object.
 | |
| //
 | |
| // All upload failures are considered test errors - this function is
 | |
| // intended as a helper for other tests.
 | |
| func uploadTestObject(t *testing.T, apiRouter http.Handler, creds auth.Credentials, bucketName, objectName string,
 | |
| 	partSizes []int64, metadata map[string]string, asMultipart bool,
 | |
| ) {
 | |
| 	if len(partSizes) == 0 {
 | |
| 		t.Fatalf("Cannot upload an object without part sizes")
 | |
| 	}
 | |
| 	if len(partSizes) > 1 {
 | |
| 		asMultipart = true
 | |
| 	}
 | |
| 
 | |
| 	checkRespErr := func(rec *httptest.ResponseRecorder, exp int) {
 | |
| 		if rec.Code != exp {
 | |
| 			b, err := io.ReadAll(rec.Body)
 | |
| 			t.Fatalf("Expected: %v, Got: %v, Body: %s, err: %v", exp, rec.Code, string(b), err)
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	if !asMultipart {
 | |
| 		srcData := NewDummyDataGen(partSizes[0], 0)
 | |
| 		req, err := newTestSignedRequestV4(http.MethodPut, getPutObjectURL("", bucketName, objectName),
 | |
| 			partSizes[0], srcData, creds.AccessKey, creds.SecretKey, metadata)
 | |
| 		if err != nil {
 | |
| 			t.Fatalf("Unexpected err: %#v", err)
 | |
| 		}
 | |
| 		rec := httptest.NewRecorder()
 | |
| 		apiRouter.ServeHTTP(rec, req)
 | |
| 		checkRespErr(rec, http.StatusOK)
 | |
| 	} else {
 | |
| 		// Multipart upload - each part is a new DummyDataGen
 | |
| 		// (so the part lengths are required to verify the
 | |
| 		// object when reading).
 | |
| 
 | |
| 		// Initiate mp upload
 | |
| 		reqI, err := newTestSignedRequestV4(http.MethodPost, getNewMultipartURL("", bucketName, objectName),
 | |
| 			0, nil, creds.AccessKey, creds.SecretKey, metadata)
 | |
| 		if err != nil {
 | |
| 			t.Fatalf("Unexpected err: %#v", err)
 | |
| 		}
 | |
| 		rec := httptest.NewRecorder()
 | |
| 		apiRouter.ServeHTTP(rec, reqI)
 | |
| 		checkRespErr(rec, http.StatusOK)
 | |
| 		decoder := xml.NewDecoder(rec.Body)
 | |
| 		multipartResponse := &InitiateMultipartUploadResponse{}
 | |
| 		err = decoder.Decode(multipartResponse)
 | |
| 		if err != nil {
 | |
| 			t.Fatalf("Error decoding the recorded response Body")
 | |
| 		}
 | |
| 		upID := multipartResponse.UploadID
 | |
| 
 | |
| 		// Upload each part
 | |
| 		var cp []CompletePart
 | |
| 		cumulativeSum := int64(0)
 | |
| 		for i, partLen := range partSizes {
 | |
| 			partID := i + 1
 | |
| 			partSrc := NewDummyDataGen(partLen, cumulativeSum)
 | |
| 			cumulativeSum += partLen
 | |
| 			req, errP := newTestSignedRequestV4(http.MethodPut,
 | |
| 				getPutObjectPartURL("", bucketName, objectName, upID, fmt.Sprintf("%d", partID)),
 | |
| 				partLen, partSrc, creds.AccessKey, creds.SecretKey, metadata)
 | |
| 			if errP != nil {
 | |
| 				t.Fatalf("Unexpected err: %#v", errP)
 | |
| 			}
 | |
| 			rec = httptest.NewRecorder()
 | |
| 			apiRouter.ServeHTTP(rec, req)
 | |
| 			checkRespErr(rec, http.StatusOK)
 | |
| 			header := rec.Header()
 | |
| 			if v, ok := header["ETag"]; ok {
 | |
| 				etag := v[0]
 | |
| 				if etag == "" {
 | |
| 					t.Fatalf("Unexpected empty etag")
 | |
| 				}
 | |
| 				cp = append(cp, CompletePart{PartNumber: partID, ETag: etag[1 : len(etag)-1]})
 | |
| 			} else {
 | |
| 				t.Fatalf("Missing etag header")
 | |
| 			}
 | |
| 		}
 | |
| 
 | |
| 		// Call CompleteMultipart API
 | |
| 		compMpBody, err := xml.Marshal(CompleteMultipartUpload{Parts: cp})
 | |
| 		if err != nil {
 | |
| 			t.Fatalf("Unexpected err: %#v", err)
 | |
| 		}
 | |
| 		reqC, errP := newTestSignedRequestV4(http.MethodPost,
 | |
| 			getCompleteMultipartUploadURL("", bucketName, objectName, upID),
 | |
| 			int64(len(compMpBody)), bytes.NewReader(compMpBody),
 | |
| 			creds.AccessKey, creds.SecretKey, metadata)
 | |
| 		if errP != nil {
 | |
| 			t.Fatalf("Unexpected err: %#v", errP)
 | |
| 		}
 | |
| 		rec = httptest.NewRecorder()
 | |
| 		apiRouter.ServeHTTP(rec, reqC)
 | |
| 		checkRespErr(rec, http.StatusOK)
 | |
| 	}
 | |
| }
 | |
| 
 | |
| // unzip a file into a specific target dir - used to unzip sample data in cmd/testdata/
 | |
| func unzipArchive(zipFilePath, targetDir string) error {
 | |
| 	zipReader, err := zip.OpenReader(zipFilePath)
 | |
| 	if err != nil {
 | |
| 		return err
 | |
| 	}
 | |
| 	for _, file := range zipReader.Reader.File {
 | |
| 		zippedFile, err := file.Open()
 | |
| 		if err != nil {
 | |
| 			return err
 | |
| 		}
 | |
| 		err = func() (err error) {
 | |
| 			defer zippedFile.Close()
 | |
| 			extractedFilePath := filepath.Join(targetDir, file.Name)
 | |
| 			if file.FileInfo().IsDir() {
 | |
| 				return os.MkdirAll(extractedFilePath, file.Mode())
 | |
| 			}
 | |
| 			outputFile, err := os.OpenFile(extractedFilePath, os.O_WRONLY|os.O_CREATE|os.O_TRUNC, file.Mode())
 | |
| 			if err != nil {
 | |
| 				return err
 | |
| 			}
 | |
| 			defer outputFile.Close()
 | |
| 			_, err = io.Copy(outputFile, zippedFile)
 | |
| 			return err
 | |
| 		}()
 | |
| 		if err != nil {
 | |
| 			return err
 | |
| 		}
 | |
| 	}
 | |
| 	return nil
 | |
| }
 |