package cluster import ( "context" "net/url" "strings" "github.com/k3s-io/kine/pkg/endpoint" "github.com/pkg/errors" "github.com/rancher/k3s/pkg/clientaccess" "github.com/rancher/k3s/pkg/cluster/managed" "github.com/rancher/k3s/pkg/daemons/config" "github.com/rancher/k3s/pkg/etcd" "github.com/sirupsen/logrus" ) type Cluster struct { clientAccessInfo *clientaccess.Info config *config.Control runtime *config.ControlRuntime managedDB managed.Driver shouldBootstrap bool storageStarted bool etcdConfig endpoint.ETCDConfig joining bool saveBootstrap bool } // Start creates the dynamic tls listener, http request handler, // handles starting and writing/reading bootstrap data, and returns a channel // that will be closed when datastore is ready. If embedded etcd is in use, // a secondary call to Cluster.save is made. func (c *Cluster) Start(ctx context.Context) (<-chan struct{}, error) { // Set up the dynamiclistener and http request handlers if err := c.initClusterAndHTTPS(ctx); err != nil { return nil, errors.Wrap(err, "init cluster datastore and https") } if c.config.DisableETCD { ready := make(chan struct{}) defer close(ready) // try to get /db/info urls first before attempting to use join url clientURLs, _, err := etcd.ClientURLs(ctx, c.clientAccessInfo, c.config.PrivateIP) if err != nil { return nil, err } if len(clientURLs) < 1 { clientURL, err := url.Parse(c.config.JoinURL) if err != nil { return nil, err } clientURL.Host = clientURL.Hostname() + ":2379" clientURLs = append(clientURLs, clientURL.String()) } etcdProxy, err := etcd.NewETCDProxy(ctx, true, c.config.DataDir, clientURLs[0]) if err != nil { return nil, err } c.setupEtcdProxy(ctx, etcdProxy) // remove etcd member if it exists if err := c.managedDB.RemoveSelf(ctx); err != nil { logrus.Warnf("Failed to remove this node from etcd members") } return ready, nil } // start managed database (if necessary) if err := c.start(ctx); err != nil { return nil, errors.Wrap(err, "start managed database") } // get the wait channel for testing managed database readiness ready, err := c.testClusterDB(ctx) if err != nil { return nil, err } // if necessary, store bootstrap data to datastore if c.saveBootstrap { if err := c.save(ctx); err != nil { return nil, err } } // if necessary, record successful bootstrap if c.shouldBootstrap { if err := c.bootstrapped(); err != nil { return nil, err } } if err := c.startStorage(ctx); err != nil { return nil, err } // at this point, if etcd is in use, it's bootstrapping is complete // so save the bootstrap data. We will need for etcd to be up. If // the save call returns an error, we panic since subsequent etcd // snapshots will be empty. if c.managedDB != nil { go func() { for range ready { if err := c.save(ctx); err != nil { panic(err) } } if err := c.managedDB.StoreSnapshotData(ctx); err != nil { logrus.Errorf("Failed to record snapshots for cluster: %v", err) } }() } return ready, nil } // startStorage starts the kine listener and configures the endpoints, if necessary. // This calls into the kine endpoint code, which sets up the database client // and unix domain socket listener if using an external database. In the case of an etcd // backend it just returns the user-provided etcd endpoints and tls config. func (c *Cluster) startStorage(ctx context.Context) error { if c.storageStarted { return nil } c.storageStarted = true // start listening on the kine socket as an etcd endpoint, or return the external etcd endpoints etcdConfig, err := endpoint.Listen(ctx, c.config.Datastore) if err != nil { return errors.Wrap(err, "creating storage endpoint") } // Persist the returned etcd configuration. We decide if we're doing leader election for embedded controllers // based on what the kine wrapper tells us about the datastore. Single-node datastores like sqlite don't require // leader election, while basically all others (etcd, external database, etc) do since they allow multiple servers. c.etcdConfig = etcdConfig c.config.Datastore.Config = etcdConfig.TLSConfig c.config.Datastore.Endpoint = strings.Join(etcdConfig.Endpoints, ",") c.config.NoLeaderElect = !etcdConfig.LeaderElect return nil } // New creates an initial cluster using the provided configuration func New(config *config.Control) *Cluster { return &Cluster{ config: config, runtime: config.Runtime, } }