mirror of https://github.com/minio/minio.git
				
				
				
			
		
			
				
	
	
		
			589 lines
		
	
	
		
			19 KiB
		
	
	
	
		
			Go
		
	
	
	
			
		
		
	
	
			589 lines
		
	
	
		
			19 KiB
		
	
	
	
		
			Go
		
	
	
	
| // Copyright (c) 2015-2021 MinIO, Inc.
 | |
| //
 | |
| // This file is part of MinIO Object Storage stack
 | |
| //
 | |
| // This program is free software: you can redistribute it and/or modify
 | |
| // it under the terms of the GNU Affero General Public License as published by
 | |
| // the Free Software Foundation, either version 3 of the License, or
 | |
| // (at your option) any later version.
 | |
| //
 | |
| // This program is distributed in the hope that it will be useful
 | |
| // but WITHOUT ANY WARRANTY; without even the implied warranty of
 | |
| // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 | |
| // GNU Affero General Public License for more details.
 | |
| //
 | |
| // You should have received a copy of the GNU Affero General Public License
 | |
| // along with this program.  If not, see <http://www.gnu.org/licenses/>.
 | |
| 
 | |
| package cmd
 | |
| 
 | |
| import (
 | |
| 	"context"
 | |
| 	"crypto/tls"
 | |
| 	"errors"
 | |
| 	"fmt"
 | |
| 	"math/rand"
 | |
| 	"net"
 | |
| 	"os"
 | |
| 	"os/signal"
 | |
| 	"strings"
 | |
| 	"sync"
 | |
| 	"syscall"
 | |
| 	"time"
 | |
| 
 | |
| 	"github.com/minio/cli"
 | |
| 	"github.com/minio/minio/cmd/config"
 | |
| 	xhttp "github.com/minio/minio/cmd/http"
 | |
| 	"github.com/minio/minio/cmd/logger"
 | |
| 	"github.com/minio/minio/cmd/rest"
 | |
| 	"github.com/minio/minio/pkg/auth"
 | |
| 	"github.com/minio/minio/pkg/bucket/bandwidth"
 | |
| 	"github.com/minio/minio/pkg/certs"
 | |
| 	"github.com/minio/minio/pkg/color"
 | |
| 	"github.com/minio/minio/pkg/env"
 | |
| 	"github.com/minio/minio/pkg/fips"
 | |
| 	"github.com/minio/minio/pkg/madmin"
 | |
| 	"github.com/minio/minio/pkg/sync/errgroup"
 | |
| )
 | |
| 
 | |
| // ServerFlags - server command specific flags
 | |
| var ServerFlags = []cli.Flag{
 | |
| 	cli.StringFlag{
 | |
| 		Name:  "address",
 | |
| 		Value: ":" + GlobalMinioDefaultPort,
 | |
| 		Usage: "bind to a specific ADDRESS:PORT, ADDRESS can be an IP or hostname",
 | |
| 	},
 | |
| }
 | |
| 
 | |
| var serverCmd = cli.Command{
 | |
| 	Name:   "server",
 | |
| 	Usage:  "start object storage server",
 | |
| 	Flags:  append(ServerFlags, GlobalFlags...),
 | |
| 	Action: serverMain,
 | |
| 	CustomHelpTemplate: `NAME:
 | |
|   {{.HelpName}} - {{.Usage}}
 | |
| 
 | |
| USAGE:
 | |
|   {{.HelpName}} {{if .VisibleFlags}}[FLAGS] {{end}}DIR1 [DIR2..]
 | |
|   {{.HelpName}} {{if .VisibleFlags}}[FLAGS] {{end}}DIR{1...64}
 | |
|   {{.HelpName}} {{if .VisibleFlags}}[FLAGS] {{end}}DIR{1...64} DIR{65...128}
 | |
| 
 | |
| DIR:
 | |
|   DIR points to a directory on a filesystem. When you want to combine
 | |
|   multiple drives into a single large system, pass one directory per
 | |
|   filesystem separated by space. You may also use a '...' convention
 | |
|   to abbreviate the directory arguments. Remote directories in a
 | |
|   distributed setup are encoded as HTTP(s) URIs.
 | |
| {{if .VisibleFlags}}
 | |
| FLAGS:
 | |
|   {{range .VisibleFlags}}{{.}}
 | |
|   {{end}}{{end}}
 | |
| EXAMPLES:
 | |
|   1. Start minio server on "/home/shared" directory.
 | |
|      {{.Prompt}} {{.HelpName}} /home/shared
 | |
| 
 | |
|   2. Start single node server with 64 local drives "/mnt/data1" to "/mnt/data64".
 | |
|      {{.Prompt}} {{.HelpName}} /mnt/data{1...64}
 | |
| 
 | |
|   3. Start distributed minio server on an 32 node setup with 32 drives each, run following command on all the nodes
 | |
|      {{.Prompt}} {{.EnvVarSetCommand}} MINIO_ROOT_USER{{.AssignmentOperator}}minio
 | |
|      {{.Prompt}} {{.EnvVarSetCommand}} MINIO_ROOT_PASSWORD{{.AssignmentOperator}}miniostorage
 | |
|      {{.Prompt}} {{.HelpName}} http://node{1...32}.example.com/mnt/export{1...32}
 | |
| 
 | |
|   4. Start distributed minio server in an expanded setup, run the following command on all the nodes
 | |
|      {{.Prompt}} {{.EnvVarSetCommand}} MINIO_ROOT_USER{{.AssignmentOperator}}minio
 | |
|      {{.Prompt}} {{.EnvVarSetCommand}} MINIO_ROOT_PASSWORD{{.AssignmentOperator}}miniostorage
 | |
|      {{.Prompt}} {{.HelpName}} http://node{1...16}.example.com/mnt/export{1...32} \
 | |
|             http://node{17...64}.example.com/mnt/export{1...64}
 | |
| `,
 | |
| }
 | |
| 
 | |
| func serverCmdArgs(ctx *cli.Context) []string {
 | |
| 	v := env.Get(config.EnvArgs, "")
 | |
| 	if v == "" {
 | |
| 		// Fall back to older ENV MINIO_ENDPOINTS
 | |
| 		v = env.Get(config.EnvEndpoints, "")
 | |
| 	}
 | |
| 	if v == "" {
 | |
| 		if !ctx.Args().Present() || ctx.Args().First() == "help" {
 | |
| 			cli.ShowCommandHelpAndExit(ctx, ctx.Command.Name, 1)
 | |
| 		}
 | |
| 		return ctx.Args()
 | |
| 	}
 | |
| 	return strings.Fields(v)
 | |
| }
 | |
| 
 | |
| func serverHandleCmdArgs(ctx *cli.Context) {
 | |
| 	// Handle common command args.
 | |
| 	handleCommonCmdArgs(ctx)
 | |
| 
 | |
| 	logger.FatalIf(CheckLocalServerAddr(globalCLIContext.Addr), "Unable to validate passed arguments")
 | |
| 
 | |
| 	var err error
 | |
| 	var setupType SetupType
 | |
| 
 | |
| 	// Check and load TLS certificates.
 | |
| 	globalPublicCerts, globalTLSCerts, globalIsTLS, err = getTLSConfig()
 | |
| 	logger.FatalIf(err, "Unable to load the TLS configuration")
 | |
| 
 | |
| 	// Check and load Root CAs.
 | |
| 	globalRootCAs, err = certs.GetRootCAs(globalCertsCADir.Get())
 | |
| 	logger.FatalIf(err, "Failed to read root CAs (%v)", err)
 | |
| 
 | |
| 	// Add the global public crts as part of global root CAs
 | |
| 	for _, publicCrt := range globalPublicCerts {
 | |
| 		globalRootCAs.AddCert(publicCrt)
 | |
| 	}
 | |
| 
 | |
| 	// Register root CAs for remote ENVs
 | |
| 	env.RegisterGlobalCAs(globalRootCAs)
 | |
| 
 | |
| 	globalMinioAddr = globalCLIContext.Addr
 | |
| 
 | |
| 	globalMinioHost, globalMinioPort = mustSplitHostPort(globalMinioAddr)
 | |
| 	globalEndpoints, setupType, err = createServerEndpoints(globalCLIContext.Addr, serverCmdArgs(ctx)...)
 | |
| 	logger.FatalIf(err, "Invalid command line arguments")
 | |
| 
 | |
| 	globalLocalNodeName = GetLocalPeer(globalEndpoints, globalMinioHost, globalMinioPort)
 | |
| 
 | |
| 	globalRemoteEndpoints = make(map[string]Endpoint)
 | |
| 	for _, z := range globalEndpoints {
 | |
| 		for _, ep := range z.Endpoints {
 | |
| 			if ep.IsLocal {
 | |
| 				globalRemoteEndpoints[globalLocalNodeName] = ep
 | |
| 			} else {
 | |
| 				globalRemoteEndpoints[ep.Host] = ep
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	// allow transport to be HTTP/1.1 for proxying.
 | |
| 	globalProxyTransport = newCustomHTTPProxyTransport(&tls.Config{
 | |
| 		RootCAs:          globalRootCAs,
 | |
| 		CipherSuites:     fips.CipherSuitesTLS(),
 | |
| 		CurvePreferences: fips.EllipticCurvesTLS(),
 | |
| 	}, rest.DefaultTimeout)()
 | |
| 	globalProxyEndpoints = GetProxyEndpoints(globalEndpoints)
 | |
| 	globalInternodeTransport = newInternodeHTTPTransport(&tls.Config{
 | |
| 		RootCAs:          globalRootCAs,
 | |
| 		CipherSuites:     fips.CipherSuitesTLS(),
 | |
| 		CurvePreferences: fips.EllipticCurvesTLS(),
 | |
| 	}, rest.DefaultTimeout)()
 | |
| 
 | |
| 	// On macOS, if a process already listens on LOCALIPADDR:PORT, net.Listen() falls back
 | |
| 	// to IPv6 address ie minio will start listening on IPv6 address whereas another
 | |
| 	// (non-)minio process is listening on IPv4 of given port.
 | |
| 	// To avoid this error situation we check for port availability.
 | |
| 	logger.FatalIf(checkPortAvailability(globalMinioHost, globalMinioPort), "Unable to start the server")
 | |
| 
 | |
| 	globalIsErasure = (setupType == ErasureSetupType)
 | |
| 	globalIsDistErasure = (setupType == DistErasureSetupType)
 | |
| 	if globalIsDistErasure {
 | |
| 		globalIsErasure = true
 | |
| 	}
 | |
| }
 | |
| 
 | |
| func serverHandleEnvVars() {
 | |
| 	// Handle common environment variables.
 | |
| 	handleCommonEnvVars()
 | |
| }
 | |
| 
 | |
| var globalHealStateLK sync.RWMutex
 | |
| 
 | |
| func newAllSubsystems() {
 | |
| 	if globalIsErasure {
 | |
| 		globalHealStateLK.Lock()
 | |
| 		// New global heal state
 | |
| 		globalAllHealState = newHealState(true)
 | |
| 		globalBackgroundHealState = newHealState(false)
 | |
| 		globalHealStateLK.Unlock()
 | |
| 	}
 | |
| 
 | |
| 	// Create new notification system and initialize notification targets
 | |
| 	globalNotificationSys = NewNotificationSys(globalEndpoints)
 | |
| 
 | |
| 	// Create new bucket metadata system.
 | |
| 	if globalBucketMetadataSys == nil {
 | |
| 		globalBucketMetadataSys = NewBucketMetadataSys()
 | |
| 	} else {
 | |
| 		// Reinitialize safely when testing.
 | |
| 		globalBucketMetadataSys.Reset()
 | |
| 	}
 | |
| 
 | |
| 	// Create the bucket bandwidth monitor
 | |
| 	globalBucketMonitor = bandwidth.NewMonitor(GlobalServiceDoneCh)
 | |
| 
 | |
| 	// Create a new config system.
 | |
| 	globalConfigSys = NewConfigSys()
 | |
| 
 | |
| 	// Create new IAM system.
 | |
| 	globalIAMSys = NewIAMSys()
 | |
| 
 | |
| 	// Create new policy system.
 | |
| 	globalPolicySys = NewPolicySys()
 | |
| 
 | |
| 	// Create new lifecycle system.
 | |
| 	globalLifecycleSys = NewLifecycleSys()
 | |
| 
 | |
| 	// Create new bucket encryption subsystem
 | |
| 	globalBucketSSEConfigSys = NewBucketSSEConfigSys()
 | |
| 
 | |
| 	// Create new bucket object lock subsystem
 | |
| 	globalBucketObjectLockSys = NewBucketObjectLockSys()
 | |
| 
 | |
| 	// Create new bucket quota subsystem
 | |
| 	globalBucketQuotaSys = NewBucketQuotaSys()
 | |
| 
 | |
| 	// Create new bucket versioning subsystem
 | |
| 	if globalBucketVersioningSys == nil {
 | |
| 		globalBucketVersioningSys = NewBucketVersioningSys()
 | |
| 	} else {
 | |
| 		globalBucketVersioningSys.Reset()
 | |
| 	}
 | |
| 
 | |
| 	// Create new bucket replication subsytem
 | |
| 	globalBucketTargetSys = NewBucketTargetSys()
 | |
| 
 | |
| 	// Create new ILM tier configuration subsystem
 | |
| 	globalTierConfigMgr = NewTierConfigMgr()
 | |
| }
 | |
| 
 | |
| func configRetriableErrors(err error) bool {
 | |
| 	// Initializing sub-systems needs a retry mechanism for
 | |
| 	// the following reasons:
 | |
| 	//  - Read quorum is lost just after the initialization
 | |
| 	//    of the object layer.
 | |
| 	//  - Write quorum not met when upgrading configuration
 | |
| 	//    version is needed, migration is needed etc.
 | |
| 	rquorum := InsufficientReadQuorum{}
 | |
| 	wquorum := InsufficientWriteQuorum{}
 | |
| 
 | |
| 	// One of these retriable errors shall be retried.
 | |
| 	return errors.Is(err, errDiskNotFound) ||
 | |
| 		errors.Is(err, errConfigNotFound) ||
 | |
| 		errors.Is(err, context.DeadlineExceeded) ||
 | |
| 		errors.Is(err, errErasureWriteQuorum) ||
 | |
| 		errors.Is(err, errErasureReadQuorum) ||
 | |
| 		errors.As(err, &rquorum) ||
 | |
| 		errors.As(err, &wquorum) ||
 | |
| 		isErrBucketNotFound(err) ||
 | |
| 		errors.Is(err, os.ErrDeadlineExceeded)
 | |
| }
 | |
| 
 | |
| func initServer(ctx context.Context, newObject ObjectLayer) error {
 | |
| 	// Once the config is fully loaded, initialize the new object layer.
 | |
| 	setObjectLayer(newObject)
 | |
| 
 | |
| 	// Make sure to hold lock for entire migration to avoid
 | |
| 	// such that only one server should migrate the entire config
 | |
| 	// at a given time, this big transaction lock ensures this
 | |
| 	// appropriately. This is also true for rotation of encrypted
 | |
| 	// content.
 | |
| 	txnLk := newObject.NewNSLock(minioMetaBucket, minioConfigPrefix+"/transaction.lock")
 | |
| 
 | |
| 	// ****  WARNING ****
 | |
| 	// Migrating to encrypted backend should happen before initialization of any
 | |
| 	// sub-systems, make sure that we do not move the above codeblock elsewhere.
 | |
| 
 | |
| 	r := rand.New(rand.NewSource(time.Now().UnixNano()))
 | |
| 
 | |
| 	lockTimeout := newDynamicTimeout(5*time.Second, 3*time.Second)
 | |
| 
 | |
| 	for {
 | |
| 		select {
 | |
| 		case <-ctx.Done():
 | |
| 			// Retry was canceled successfully.
 | |
| 			return fmt.Errorf("Initializing sub-systems stopped gracefully %w", ctx.Err())
 | |
| 		default:
 | |
| 		}
 | |
| 
 | |
| 		// let one of the server acquire the lock, if not let them timeout.
 | |
| 		// which shall be retried again by this loop.
 | |
| 		_, cancel, err := txnLk.GetLock(ctx, lockTimeout)
 | |
| 		if err != nil {
 | |
| 			logger.Info("Waiting for all MinIO sub-systems to be initialized.. trying to acquire lock")
 | |
| 
 | |
| 			time.Sleep(time.Duration(r.Float64() * float64(5*time.Second)))
 | |
| 			continue
 | |
| 		}
 | |
| 
 | |
| 		// These messages only meant primarily for distributed setup, so only log during distributed setup.
 | |
| 		if globalIsDistErasure {
 | |
| 			logger.Info("Waiting for all MinIO sub-systems to be initialized.. lock acquired")
 | |
| 		}
 | |
| 
 | |
| 		// Migrate all backend configs to encrypted backend configs, optionally
 | |
| 		// handles rotating keys for encryption, if there is any retriable failure
 | |
| 		// that shall be retried if there is an error.
 | |
| 		if err = handleEncryptedConfigBackend(newObject); err == nil {
 | |
| 			// Upon success migrating the config, initialize all sub-systems
 | |
| 			// if all sub-systems initialized successfully return right away
 | |
| 			if err = initAllSubsystems(ctx, newObject); err == nil {
 | |
| 				txnLk.Unlock(cancel)
 | |
| 				// All successful return.
 | |
| 				if globalIsDistErasure {
 | |
| 					// These messages only meant primarily for distributed setup, so only log during distributed setup.
 | |
| 					logger.Info("All MinIO sub-systems initialized successfully")
 | |
| 				}
 | |
| 				return nil
 | |
| 			}
 | |
| 		}
 | |
| 
 | |
| 		txnLk.Unlock(cancel) // Unlock the transaction lock and allow other nodes to acquire the lock if possible.
 | |
| 
 | |
| 		if configRetriableErrors(err) {
 | |
| 			logger.Info("Waiting for all MinIO sub-systems to be initialized.. possible cause (%v)", err)
 | |
| 			time.Sleep(time.Duration(r.Float64() * float64(5*time.Second)))
 | |
| 			continue
 | |
| 		}
 | |
| 
 | |
| 		// Any other unhandled return right here.
 | |
| 		return fmt.Errorf("Unable to initialize sub-systems: %w", err)
 | |
| 	}
 | |
| }
 | |
| 
 | |
| func initAllSubsystems(ctx context.Context, newObject ObjectLayer) (err error) {
 | |
| 	// %w is used by all error returns here to make sure
 | |
| 	// we wrap the underlying error, make sure when you
 | |
| 	// are modifying this code that you do so, if and when
 | |
| 	// you want to add extra context to your error. This
 | |
| 	// ensures top level retry works accordingly.
 | |
| 	// List buckets to heal, and be re-used for loading configs.
 | |
| 
 | |
| 	buckets, err := newObject.ListBuckets(ctx)
 | |
| 	if err != nil {
 | |
| 		return fmt.Errorf("Unable to list buckets to heal: %w", err)
 | |
| 	}
 | |
| 
 | |
| 	if globalIsErasure {
 | |
| 		if len(buckets) > 0 {
 | |
| 			if len(buckets) == 1 {
 | |
| 				logger.Info(fmt.Sprintf("Verifying if %d bucket is consistent across drives...", len(buckets)))
 | |
| 			} else {
 | |
| 				logger.Info(fmt.Sprintf("Verifying if %d buckets are consistent across drives...", len(buckets)))
 | |
| 			}
 | |
| 		}
 | |
| 
 | |
| 		// Limit to no more than 50 concurrent buckets.
 | |
| 		g := errgroup.WithNErrs(len(buckets)).WithConcurrency(50)
 | |
| 		ctx, cancel := g.WithCancelOnError(ctx)
 | |
| 		defer cancel()
 | |
| 		for index := range buckets {
 | |
| 			index := index
 | |
| 			g.Go(func() error {
 | |
| 				_, berr := newObject.HealBucket(ctx, buckets[index].Name, madmin.HealOpts{Recreate: true})
 | |
| 				return berr
 | |
| 			}, index)
 | |
| 		}
 | |
| 		if err := g.WaitErr(); err != nil {
 | |
| 			return fmt.Errorf("Unable to list buckets to heal: %w", err)
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	// Initialize config system.
 | |
| 	if err = globalConfigSys.Init(newObject); err != nil {
 | |
| 		if configRetriableErrors(err) {
 | |
| 			return fmt.Errorf("Unable to initialize config system: %w", err)
 | |
| 		}
 | |
| 		// Any other config errors we simply print a message and proceed forward.
 | |
| 		logger.LogIf(ctx, fmt.Errorf("Unable to initialize config, some features may be missing %w", err))
 | |
| 	}
 | |
| 
 | |
| 	// Populate existing buckets to the etcd backend
 | |
| 	if globalDNSConfig != nil {
 | |
| 		// Background this operation.
 | |
| 		go initFederatorBackend(buckets, newObject)
 | |
| 	}
 | |
| 
 | |
| 	// Initialize bucket metadata sub-system.
 | |
| 	globalBucketMetadataSys.Init(ctx, buckets, newObject)
 | |
| 
 | |
| 	// Initialize notification system.
 | |
| 	globalNotificationSys.Init(ctx, buckets, newObject)
 | |
| 
 | |
| 	// Initialize bucket targets sub-system.
 | |
| 	globalBucketTargetSys.Init(ctx, buckets, newObject)
 | |
| 
 | |
| 	if globalIsErasure {
 | |
| 		// Initialize transition tier configuration manager
 | |
| 		err = globalTierConfigMgr.Init(ctx, newObject)
 | |
| 		if err != nil {
 | |
| 			return err
 | |
| 		}
 | |
| 	}
 | |
| 	return nil
 | |
| }
 | |
| 
 | |
| // serverMain handler called for 'minio server' command.
 | |
| func serverMain(ctx *cli.Context) {
 | |
| 	defer globalDNSCache.Stop()
 | |
| 
 | |
| 	signal.Notify(globalOSSignalCh, os.Interrupt, syscall.SIGTERM, syscall.SIGQUIT)
 | |
| 
 | |
| 	go handleSignals()
 | |
| 
 | |
| 	setDefaultProfilerRates()
 | |
| 
 | |
| 	// Initialize globalConsoleSys system
 | |
| 	globalConsoleSys = NewConsoleLogger(GlobalContext)
 | |
| 	logger.AddTarget(globalConsoleSys)
 | |
| 
 | |
| 	// Perform any self-tests
 | |
| 	bitrotSelfTest()
 | |
| 	erasureSelfTest()
 | |
| 	compressSelfTest()
 | |
| 
 | |
| 	// Handle all server command args.
 | |
| 	serverHandleCmdArgs(ctx)
 | |
| 
 | |
| 	// Handle all server environment vars.
 | |
| 	serverHandleEnvVars()
 | |
| 
 | |
| 	// Set node name, only set for distributed setup.
 | |
| 	globalConsoleSys.SetNodeName(globalLocalNodeName)
 | |
| 
 | |
| 	// Initialize all help
 | |
| 	initHelp()
 | |
| 
 | |
| 	// Initialize all sub-systems
 | |
| 	newAllSubsystems()
 | |
| 
 | |
| 	globalMinioEndpoint = func() string {
 | |
| 		host := globalMinioHost
 | |
| 		if host == "" {
 | |
| 			host = sortIPs(localIP4.ToSlice())[0]
 | |
| 		}
 | |
| 		return fmt.Sprintf("%s://%s", getURLScheme(globalIsTLS), net.JoinHostPort(host, globalMinioPort))
 | |
| 	}()
 | |
| 
 | |
| 	// Is distributed setup, error out if no certificates are found for HTTPS endpoints.
 | |
| 	if globalIsDistErasure {
 | |
| 		if globalEndpoints.HTTPS() && !globalIsTLS {
 | |
| 			logger.Fatal(config.ErrNoCertsAndHTTPSEndpoints(nil), "Unable to start the server")
 | |
| 		}
 | |
| 		if !globalEndpoints.HTTPS() && globalIsTLS {
 | |
| 			logger.Fatal(config.ErrCertsAndHTTPEndpoints(nil), "Unable to start the server")
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	if !globalCLIContext.Quiet && !globalInplaceUpdateDisabled {
 | |
| 		// Check for new updates from dl.min.io.
 | |
| 		checkUpdate(getMinioMode())
 | |
| 	}
 | |
| 
 | |
| 	if !globalActiveCred.IsValid() && globalIsDistErasure {
 | |
| 		logger.Fatal(config.ErrEnvCredentialsMissingDistributed(nil),
 | |
| 			"Unable to initialize the server in distributed mode")
 | |
| 	}
 | |
| 
 | |
| 	// Set system resources to maximum.
 | |
| 	setMaxResources()
 | |
| 
 | |
| 	// Configure server.
 | |
| 	handler, err := configureServerHandler(globalEndpoints)
 | |
| 	if err != nil {
 | |
| 		logger.Fatal(config.ErrUnexpectedError(err), "Unable to configure one of server's RPC services")
 | |
| 	}
 | |
| 
 | |
| 	var getCert certs.GetCertificateFunc
 | |
| 	if globalTLSCerts != nil {
 | |
| 		getCert = globalTLSCerts.GetCertificate
 | |
| 	}
 | |
| 
 | |
| 	httpServer := xhttp.NewServer([]string{globalMinioAddr}, criticalErrorHandler{corsHandler(handler)}, getCert)
 | |
| 	httpServer.BaseContext = func(listener net.Listener) context.Context {
 | |
| 		return GlobalContext
 | |
| 	}
 | |
| 	go func() {
 | |
| 		globalHTTPServerErrorCh <- httpServer.Start()
 | |
| 	}()
 | |
| 
 | |
| 	setHTTPServer(httpServer)
 | |
| 
 | |
| 	if globalIsDistErasure && globalEndpoints.FirstLocal() {
 | |
| 		for {
 | |
| 			// Additionally in distributed setup, validate the setup and configuration.
 | |
| 			err := verifyServerSystemConfig(GlobalContext, globalEndpoints)
 | |
| 			if err == nil || errors.Is(err, context.Canceled) {
 | |
| 				break
 | |
| 			}
 | |
| 			logger.LogIf(GlobalContext, err, "Unable to initialize distributed setup, retrying.. after 5 seconds")
 | |
| 			select {
 | |
| 			case <-GlobalContext.Done():
 | |
| 				return
 | |
| 			case <-time.After(500 * time.Millisecond):
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	newObject, err := newObjectLayer(GlobalContext, globalEndpoints)
 | |
| 	if err != nil {
 | |
| 		logFatalErrs(err, Endpoint{}, true)
 | |
| 	}
 | |
| 	logger.SetDeploymentID(globalDeploymentID)
 | |
| 
 | |
| 	// Enable background operations for erasure coding
 | |
| 	if globalIsErasure {
 | |
| 		initAutoHeal(GlobalContext, newObject)
 | |
| 		initBackgroundTransition(GlobalContext, newObject)
 | |
| 	}
 | |
| 
 | |
| 	initBackgroundExpiry(GlobalContext, newObject)
 | |
| 	initDataScanner(GlobalContext, newObject)
 | |
| 
 | |
| 	if err = initServer(GlobalContext, newObject); err != nil {
 | |
| 		var cerr config.Err
 | |
| 		// For any config error, we don't need to drop into safe-mode
 | |
| 		// instead its a user error and should be fixed by user.
 | |
| 		if errors.As(err, &cerr) {
 | |
| 			logger.FatalIf(err, "Unable to initialize the server")
 | |
| 		}
 | |
| 
 | |
| 		// If context was canceled
 | |
| 		if errors.Is(err, context.Canceled) {
 | |
| 			logger.FatalIf(err, "Server startup canceled upon user request")
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	if globalIsErasure { // to be done after config init
 | |
| 		initBackgroundReplication(GlobalContext, newObject)
 | |
| 		globalTierJournal, err = initTierDeletionJournal(GlobalContext.Done())
 | |
| 		if err != nil {
 | |
| 			logger.FatalIf(err, "Unable to initialize remote tier pending deletes journal")
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	if globalCacheConfig.Enabled {
 | |
| 		// initialize the new disk cache objects.
 | |
| 		var cacheAPI CacheObjectLayer
 | |
| 		cacheAPI, err = newServerCacheObjects(GlobalContext, globalCacheConfig)
 | |
| 		logger.FatalIf(err, "Unable to initialize disk caching")
 | |
| 
 | |
| 		setCacheObjectLayer(cacheAPI)
 | |
| 	}
 | |
| 
 | |
| 	// Initialize users credentials and policies in background right after config has initialized.
 | |
| 	go globalIAMSys.Init(GlobalContext, newObject)
 | |
| 
 | |
| 	// Prints the formatted startup message, if err is not nil then it prints additional information as well.
 | |
| 	printStartupMessage(getAPIEndpoints(), err)
 | |
| 
 | |
| 	if globalActiveCred.Equal(auth.DefaultCredentials) {
 | |
| 		msg := fmt.Sprintf("Detected default credentials '%s', please change the credentials immediately using 'MINIO_ROOT_USER' and 'MINIO_ROOT_PASSWORD'", globalActiveCred)
 | |
| 		logger.StartupMessage(color.RedBold(msg))
 | |
| 	}
 | |
| 
 | |
| 	<-globalOSSignalCh
 | |
| }
 | |
| 
 | |
| // Initialize object layer with the supplied disks, objectLayer is nil upon any error.
 | |
| func newObjectLayer(ctx context.Context, endpointServerPools EndpointServerPools) (newObject ObjectLayer, err error) {
 | |
| 	// For FS only, directly use the disk.
 | |
| 	if endpointServerPools.NEndpoints() == 1 {
 | |
| 		// Initialize new FS object layer.
 | |
| 		return NewFSObjectLayer(endpointServerPools[0].Endpoints[0].Path)
 | |
| 	}
 | |
| 
 | |
| 	return newErasureServerPools(ctx, endpointServerPools)
 | |
| }
 |