/* Copyright The Helm Authors. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. */ package main import ( "bytes" "encoding/json" "errors" "fmt" "io" "os" "time" "github.com/spf13/cobra" apierrors "k8s.io/apimachinery/pkg/api/errors" "k8s.io/client-go/kubernetes" "k8s.io/apimachinery/pkg/util/yaml" "k8s.io/helm/cmd/helm/installer" "k8s.io/helm/pkg/helm" "k8s.io/helm/pkg/helm/helmpath" "k8s.io/helm/pkg/helm/portforwarder" "k8s.io/helm/pkg/version" ) const initDesc = ` This command installs Tiller (the Helm server-side component) onto your Kubernetes Cluster and sets up local configuration in $HELM_HOME (default ~/.helm/). As with the rest of the Helm commands, 'helm init' discovers Kubernetes clusters by reading $KUBECONFIG (default '~/.kube/config') and using the default context. To set up just a local environment, use '--client-only'. That will configure $HELM_HOME, but not attempt to connect to a Kubernetes cluster and install the Tiller deployment. When installing Tiller, 'helm init' will attempt to install the latest released version. You can specify an alternative image with '--tiller-image'. For those frequently working on the latest code, the flag '--canary-image' will install the latest pre-release version of Tiller (e.g. the HEAD commit in the GitHub repository on the master branch). To dump a manifest containing the Tiller deployment YAML, combine the '--dry-run' and '--debug' flags. ` var ( stableRepositoryURL = "https://charts.helm.sh/stable" oldStableRepositoryURL = "https://kubernetes-charts.storage.googleapis.com" // This is the IPv4 loopback, not localhost, because we have to force IPv4 // for Dockerized Helm: https://github.com/kubernetes/helm/issues/1410 localRepositoryURL = "http://127.0.0.1:8879/charts" tlsServerName string // overrides the server name used to verify the hostname on the returned certificates from the server. tlsCaCertFile string // path to TLS CA certificate file tlsCertFile string // path to TLS certificate file tlsKeyFile string // path to TLS key file tlsVerify bool // enable TLS and verify remote certificates tlsEnable bool // enable TLS ) type initCmd struct { image string clientOnly bool canary bool upgrade bool namespace string dryRun bool forceUpgrade bool skipRefresh bool skipRepos bool out io.Writer client helm.Interface home helmpath.Home opts installer.Options kubeClient kubernetes.Interface serviceAccount string maxHistory int replicas int wait bool useDeprecatedRepo bool } func newInitCmd(out io.Writer) *cobra.Command { i := &initCmd{out: out} cmd := &cobra.Command{ Use: "init", Short: "Initialize Helm on both client and server", Long: initDesc, RunE: func(cmd *cobra.Command, args []string) error { if len(args) != 0 { return errors.New("This command does not accept arguments") } i.namespace = settings.TillerNamespace i.home = settings.Home i.client = ensureHelmClient(i.client) return i.run() }, } f := cmd.Flags() f.StringVarP(&i.image, "tiller-image", "i", "", "Override Tiller image") f.BoolVar(&i.canary, "canary-image", false, "Use the canary Tiller image") f.BoolVar(&i.upgrade, "upgrade", false, "Upgrade if Tiller is already installed") f.BoolVar(&i.forceUpgrade, "force-upgrade", false, "Force upgrade of Tiller to the current helm version") f.BoolVarP(&i.clientOnly, "client-only", "c", false, "If set does not install Tiller") f.BoolVar(&i.dryRun, "dry-run", false, "Do not install local or remote") f.BoolVar(&i.skipRefresh, "skip-refresh", false, "Do not refresh (download) the local repository cache") f.BoolVar(&i.skipRepos, "skip-repos", false, "Skip adding the stable and local repositories") f.BoolVar(&i.wait, "wait", false, "Block until Tiller is running and ready to receive requests") f.BoolVar(&i.useDeprecatedRepo, "use-deprecated-stable-repository", false, "Use the old (googleapis) repository URL even though that URL is being shutdown.") // TODO: replace TLS flags with pkg/helm/environment.AddFlagsTLS() in Helm 3 // // NOTE (bacongobbler): we can't do this in Helm 2 because the flag names differ, and `helm init --tls-ca-cert` // doesn't conform with the rest of the TLS flag names (should be --tiller-tls-ca-cert in Helm 3) f.BoolVar(&tlsEnable, "tiller-tls", false, "Install Tiller with TLS enabled") f.BoolVar(&tlsVerify, "tiller-tls-verify", false, "Install Tiller with TLS enabled and to verify remote certificates") f.StringVar(&tlsKeyFile, "tiller-tls-key", "", "Path to TLS key file to install with Tiller") f.StringVar(&tlsCertFile, "tiller-tls-cert", "", "Path to TLS certificate file to install with Tiller") f.StringVar(&tlsCaCertFile, "tls-ca-cert", "", "Path to CA root certificate") f.StringVar(&tlsServerName, "tiller-tls-hostname", settings.TillerHost, "The server name used to verify the hostname on the returned certificates from Tiller") f.StringVar(&stableRepositoryURL, "stable-repo-url", stableRepositoryURL, "URL for stable repository") f.StringVar(&localRepositoryURL, "local-repo-url", localRepositoryURL, "URL for local repository") f.BoolVar(&i.opts.EnableHostNetwork, "net-host", false, "Install Tiller with net=host") f.StringVar(&i.serviceAccount, "service-account", "", "Name of service account") f.IntVar(&i.maxHistory, "history-max", 0, "Limit the maximum number of revisions saved per release. Use 0 for no limit.") f.IntVar(&i.replicas, "replicas", 1, "Amount of tiller instances to run on the cluster") f.StringVar(&i.opts.NodeSelectors, "node-selectors", "", "Labels to specify the node on which Tiller is installed (app=tiller,helm=rocks)") f.VarP(&i.opts.Output, "output", "o", "Skip installation and output Tiller's manifest in specified format (json or yaml)") f.StringArrayVar(&i.opts.Values, "override", []string{}, "Override values for the Tiller Deployment manifest (can specify multiple or separate values with commas: key1=val1,key2=val2)") f.BoolVar(&i.opts.AutoMountServiceAccountToken, "automount-service-account-token", true, "Auto-mount the given service account to tiller") return cmd } // tlsOptions sanitizes the tls flags as well as checks for the existence of required // tls files indicated by those flags, if any. func (i *initCmd) tlsOptions() error { i.opts.EnableTLS = tlsEnable || tlsVerify i.opts.VerifyTLS = tlsVerify if i.opts.EnableTLS { missing := func(file string) bool { _, err := os.Stat(file) return os.IsNotExist(err) } if i.opts.TLSKeyFile = tlsKeyFile; i.opts.TLSKeyFile == "" || missing(i.opts.TLSKeyFile) { return errors.New("missing required TLS key file") } if i.opts.TLSCertFile = tlsCertFile; i.opts.TLSCertFile == "" || missing(i.opts.TLSCertFile) { return errors.New("missing required TLS certificate file") } if i.opts.VerifyTLS { if i.opts.TLSCaCertFile = tlsCaCertFile; i.opts.TLSCaCertFile == "" || missing(i.opts.TLSCaCertFile) { return errors.New("missing required TLS CA file") } } // FIXME: remove once we use pkg/helm/environment.AddFlagsTLS() in Helm 3 settings.TLSEnable = tlsEnable settings.TLSVerify = tlsVerify settings.TLSServerName = tlsServerName settings.TLSCaCertFile = tlsCaCertFile settings.TLSCertFile = tlsCertFile settings.TLSKeyFile = tlsKeyFile } return nil } // run initializes local config and installs Tiller to Kubernetes cluster. func (i *initCmd) run() error { if err := i.tlsOptions(); err != nil { return err } i.opts.Namespace = i.namespace i.opts.UseCanary = i.canary i.opts.ImageSpec = i.image i.opts.ForceUpgrade = i.forceUpgrade i.opts.ServiceAccount = i.serviceAccount i.opts.MaxHistory = i.maxHistory i.opts.Replicas = i.replicas writeYAMLManifests := func(manifests []string) error { w := i.out for _, manifest := range manifests { if _, err := fmt.Fprintln(w, "---"); err != nil { return err } if _, err := fmt.Fprintln(w, manifest); err != nil { return err } } // YAML ending document boundary marker _, err := fmt.Fprintln(w, "...") return err } if len(i.opts.Output) > 0 { var manifests []string var err error if manifests, err = installer.TillerManifests(&i.opts); err != nil { return err } switch i.opts.Output.String() { case "json": for _, manifest := range manifests { var out bytes.Buffer jsonb, err := yaml.ToJSON([]byte(manifest)) if err != nil { return err } buf := bytes.NewBuffer(jsonb) if err := json.Indent(&out, buf.Bytes(), "", " "); err != nil { return err } if _, err = i.out.Write(out.Bytes()); err != nil { return err } fmt.Fprint(i.out, "\n") } return nil case "yaml": return writeYAMLManifests(manifests) default: return fmt.Errorf("unknown output format: %q", i.opts.Output) } } if settings.Debug { var manifests []string var err error // write Tiller manifests if manifests, err = installer.TillerManifests(&i.opts); err != nil { return err } if err = writeYAMLManifests(manifests); err != nil { return err } } if i.dryRun { return nil } if i.skipRepos { if err := installer.InitializeWithoutRepos(i.home, i.out); err != nil { return fmt.Errorf("error initializing: %s", err) } } else { // If this is set, override user config, default config, and set it to the old // URL. if i.useDeprecatedRepo { stableRepositoryURL = oldStableRepositoryURL } if err := installer.Initialize(i.home, i.out, i.skipRefresh, settings, stableRepositoryURL, localRepositoryURL); err != nil { return fmt.Errorf("error initializing: %s", err) } } fmt.Fprintf(i.out, "$HELM_HOME has been configured at %s.\n", settings.Home) if !i.clientOnly { if i.kubeClient == nil { _, c, err := getKubeClient(settings.KubeContext, settings.KubeConfig) if err != nil { return fmt.Errorf("could not get kubernetes client: %s", err) } i.kubeClient = c } if err := installer.Install(i.kubeClient, &i.opts); err != nil { if !apierrors.IsAlreadyExists(err) { return fmt.Errorf("error installing: %s", err) } if i.upgrade { if err := installer.Upgrade(i.kubeClient, &i.opts); err != nil { return fmt.Errorf("error when upgrading: %s", err) } if err := i.ping(i.opts.SelectImage()); err != nil { return err } fmt.Fprintln(i.out, "\nTiller (the Helm server-side component) has been updated to", i.opts.SelectImage(), ".") } else { debug("The error received while trying to init: %s", err) fmt.Fprintln(i.out, "Warning: Tiller is already installed in the cluster.\n"+ "(Use --client-only to suppress this message, or --upgrade to upgrade Tiller to the current version.)") } } else { fmt.Fprintln(i.out, "\nTiller (the Helm server-side component) has been installed into your Kubernetes Cluster.") if !tlsVerify { fmt.Fprintln(i.out, "\nPlease note: by default, Tiller is deployed with an insecure 'allow unauthenticated users' policy.\n"+ "To prevent this, run `helm init` with the --tiller-tls-verify flag.\n"+ "For more information on securing your installation see: https://v2.helm.sh/docs/securing_installation/") } } if err := i.ping(i.opts.SelectImage()); err != nil { return err } } else { fmt.Fprintln(i.out, "Not installing Tiller due to 'client-only' flag having been set") } needsDefaultImage := !i.clientOnly && !i.opts.UseCanary && len(i.opts.ImageSpec) == 0 && version.BuildMetadata == "unreleased" if needsDefaultImage { fmt.Fprintf(i.out, "\nWarning: You appear to be using an unreleased version of Helm. Please either use the\n"+ "--canary-image flag, or specify your desired tiller version with --tiller-image.\n\n"+ "Ex:\n"+ "$ helm init --tiller-image ghcr.io/helm/tiller:v2.17.0\n\n") } return nil } func (i *initCmd) ping(image string) error { if i.wait { _, kubeClient, err := getKubeClient(settings.KubeContext, settings.KubeConfig) if err != nil { return err } if !watchTillerUntilReady(settings.TillerNamespace, kubeClient, settings.TillerConnectionTimeout, image) { return fmt.Errorf("tiller was not found. polling deadline exceeded") } // establish a connection to Tiller now that we've effectively guaranteed it's available if err := setupConnection(); err != nil { return err } i.client = newClient() if err := i.client.PingTiller(); err != nil { return fmt.Errorf("could not ping Tiller: %s", err) } } return nil } // watchTillerUntilReady waits for the tiller pod to become available. This is useful in situations where we // want to wait before we call New(). // // Returns true if it exists. If the timeout was reached and it could not find the pod, it returns false. func watchTillerUntilReady(namespace string, client kubernetes.Interface, timeout int64, newImage string) bool { deadlinePollingChan := time.NewTimer(time.Duration(timeout) * time.Second).C checkTillerPodTicker := time.NewTicker(500 * time.Millisecond) doneChan := make(chan bool) defer checkTillerPodTicker.Stop() go func() { for range checkTillerPodTicker.C { image, err := portforwarder.GetTillerPodImage(client.CoreV1(), namespace) if err == nil && image == newImage { doneChan <- true break } } }() for { select { case <-deadlinePollingChan: return false case <-doneChan: return true } } }