package commands import ( "context" "encoding/json" "errors" "fmt" "io/ioutil" "net/url" "os" "path/filepath" "strconv" "strings" "time" v2 "github.com/porter-dev/porter/cli/cmd/v2" "github.com/cli/cli/git" "github.com/fatih/color" "github.com/mitchellh/mapstructure" api "github.com/porter-dev/porter/api/client" "github.com/porter-dev/porter/api/types" "github.com/porter-dev/porter/cli/cmd/config" "github.com/porter-dev/porter/cli/cmd/deploy" "github.com/porter-dev/porter/cli/cmd/deploy/wait" porter_app "github.com/porter-dev/porter/cli/cmd/porter_app" "github.com/porter-dev/porter/cli/cmd/preview" previewV2Beta1 "github.com/porter-dev/porter/cli/cmd/preview/v2beta1" cliUtils "github.com/porter-dev/porter/cli/cmd/utils" previewInt "github.com/porter-dev/porter/internal/integrations/preview" "github.com/porter-dev/porter/internal/templater/utils" "github.com/porter-dev/switchboard/pkg/drivers" switchboardModels "github.com/porter-dev/switchboard/pkg/models" "github.com/porter-dev/switchboard/pkg/parser" switchboardTypes "github.com/porter-dev/switchboard/pkg/types" switchboardWorker "github.com/porter-dev/switchboard/pkg/worker" "github.com/rs/zerolog" "github.com/spf13/cobra" "gopkg.in/yaml.v2" ) var ( porterYAML string previewApply bool ) func registerCommand_Apply(cliConf config.CLIConfig) *cobra.Command { applyCmd := &cobra.Command{ Use: "apply", Short: "Applies a configuration to an application", Long: fmt.Sprintf(` %s Applies a configuration to an application by either creating a new one or updating an existing one. For example: %s This command will apply the configuration contained in porter.yaml to the requested project and cluster either provided inside the porter.yaml file or through environment variables. Note that environment variables will always take precendence over values specified in the porter.yaml file. By default, this command expects to be run from a local git repository. The following are the environment variables that can be used to set certain values while applying a configuration: PORTER_CLUSTER Cluster ID that contains the project PORTER_PROJECT Project ID that contains the application PORTER_NAMESPACE The Kubernetes namespace that the application belongs to PORTER_SOURCE_NAME Name of the source Helm chart PORTER_SOURCE_REPO The URL of the Helm charts registry PORTER_SOURCE_VERSION The version of the Helm chart to use PORTER_TAG The Docker image tag to use (like the git commit hash) `, color.New(color.FgBlue, color.Bold).Sprintf("Help for \"porter apply\":"), color.New(color.FgGreen, color.Bold).Sprintf("porter apply -f porter.yaml"), ), Run: func(cmd *cobra.Command, args []string) { err := checkLoginAndRunWithConfig(cmd, cliConf, args, apply) if err != nil { if strings.Contains(err.Error(), "Forbidden") { _, _ = color.New(color.FgRed).Fprintf(os.Stderr, "You may have to update your GitHub secret token") } os.Exit(1) } }, } // applyValidateCmd represents the "porter apply validate" command when called // with a porter.yaml file as an argument applyValidateCmd := &cobra.Command{ Use: "validate", Short: "Validates a porter.yaml", Run: func(*cobra.Command, []string) { err := applyValidate() if err != nil { _, _ = color.New(color.FgRed).Fprintf(os.Stderr, "Error: %s\n", err.Error()) os.Exit(1) } else { _, _ = color.New(color.FgGreen).Printf("The porter.yaml file is valid!\n") } }, } applyCmd.AddCommand(applyValidateCmd) applyCmd.PersistentFlags().StringVarP(&porterYAML, "file", "f", "", "path to porter.yaml") applyCmd.PersistentFlags().BoolVarP(&previewApply, "preview", "p", false, "apply as preview environment based on current git branch") applyCmd.MarkFlagRequired("file") return applyCmd } func appNameFromEnvironmentVariable() string { if os.Getenv("PORTER_APP_NAME") != "" { return os.Getenv("PORTER_APP_NAME") } if os.Getenv("PORTER_STACK_NAME") != "" { return os.Getenv("PORTER_STACK_NAME") } return "" } func apply(ctx context.Context, _ *types.GetAuthenticatedUserResponse, client api.Client, cliConfig config.CLIConfig, _ config.FeatureFlags, _ *cobra.Command, _ []string) (err error) { project, err := client.GetProject(ctx, cliConfig.Project) if err != nil { return fmt.Errorf("could not retrieve project from Porter API. Please contact support@porter.run") } appName := appNameFromEnvironmentVariable() if project.ValidateApplyV2 { if previewApply && !project.PreviewEnvsEnabled { return fmt.Errorf("preview environments are not enabled for this project. Please contact support@porter.run") } inp := v2.ApplyInput{ CLIConfig: cliConfig, Client: client, PorterYamlPath: porterYAML, AppName: appName, PreviewApply: previewApply, } err = v2.Apply(ctx, inp) if err != nil { return err } return nil } fileBytes, err := os.ReadFile(porterYAML) //nolint:errcheck,gosec // do not want to change logic of CLI. New linter error if err != nil && appName == "" { return fmt.Errorf("a valid porter.yaml file must be specified. Run porter apply --help for more information") } var previewVersion struct { Version string `json:"version"` } err = yaml.Unmarshal(fileBytes, &previewVersion) if err != nil { return fmt.Errorf("error unmarshaling porter.yaml: %w", err) } var resGroup *switchboardTypes.ResourceGroup worker := switchboardWorker.NewWorker() if previewVersion.Version == "v2beta1" { ns := os.Getenv("PORTER_NAMESPACE") applier, err := previewV2Beta1.NewApplier(client, cliConfig, fileBytes, ns) if err != nil { return err } resGroup, err = applier.DowngradeToV1() if err != nil { return err } } else if previewVersion.Version == "v1" { if _, ok := os.LookupEnv("PORTER_VALIDATE_YAML"); ok { err := applyValidate() if err != nil { return err } } resGroup, err = parser.ParseRawBytes(fileBytes) if err != nil { return fmt.Errorf("error parsing porter.yaml: %w", err) } } else if previewVersion.Version == "v1stack" || previewVersion.Version == "" { parsed, err := porter_app.ValidateAndMarshal(fileBytes) if err != nil { return fmt.Errorf("error parsing porter.yaml: %w", err) } resGroup = &switchboardTypes.ResourceGroup{ Version: "v1", Resources: []*switchboardTypes.Resource{ { Name: "get-env", Driver: "os-env", }, }, } if parsed.Applications != nil { for name, app := range parsed.Applications { resources, err := porter_app.CreateApplicationDeploy(ctx, client, worker, app, name, cliConfig) if err != nil { return fmt.Errorf("error parsing porter.yaml for build resources: %w", err) } resGroup.Resources = append(resGroup.Resources, resources...) } } else { if appName == "" { return fmt.Errorf("environment variable PORTER_STACK_NAME must be set") } if parsed.Apps != nil && parsed.Services != nil { return fmt.Errorf("'apps' and 'services' are synonymous but both were defined") } var services map[string]*porter_app.Service if parsed.Apps != nil { services = parsed.Apps } if parsed.Services != nil { services = parsed.Services } app := &porter_app.Application{ Env: parsed.Env, Services: services, Build: parsed.Build, Release: parsed.Release, } if err != nil { return fmt.Errorf("error parsing porter.yaml for build resources: %w", err) } resources, err := porter_app.CreateApplicationDeploy(ctx, client, worker, app, appName, cliConfig) if err != nil { return fmt.Errorf("error parsing porter.yaml for build resources: %w", err) } resGroup.Resources = append(resGroup.Resources, resources...) } } else if previewVersion.Version == "v2" { return errors.New("porter.yaml v2 is not enabled for this project") } else { return fmt.Errorf("unknown porter.yaml version: %s", previewVersion.Version) } basePath, err := os.Getwd() if err != nil { err = fmt.Errorf("error getting working directory: %w", err) return } drivers := []struct { name string funcName func(resource *switchboardModels.Resource, opts *drivers.SharedDriverOpts) (drivers.Driver, error) }{ {"deploy", NewDeployDriver(ctx, client, cliConfig)}, {"build-image", preview.NewBuildDriver(ctx, client, cliConfig)}, {"push-image", preview.NewPushDriver(ctx, client, cliConfig)}, {"update-config", preview.NewUpdateConfigDriver(ctx, client, cliConfig)}, {"random-string", preview.NewRandomStringDriver}, {"env-group", preview.NewEnvGroupDriver(ctx, client, cliConfig)}, {"os-env", preview.NewOSEnvDriver}, } for _, driver := range drivers { err = worker.RegisterDriver(driver.name, driver.funcName) if err != nil { err = fmt.Errorf("error registering driver %s: %w", driver.name, err) return } } worker.SetDefaultDriver("deploy") if hasDeploymentHookEnvVars() { deplNamespace := os.Getenv("PORTER_NAMESPACE") if deplNamespace == "" { err = fmt.Errorf("namespace must be set by PORTER_NAMESPACE") return } deploymentHook, err := NewDeploymentHook(cliConfig, client, resGroup, deplNamespace) if err != nil { err = fmt.Errorf("error creating deployment hook: %w", err) return err } err = worker.RegisterHook("deployment", deploymentHook) if err != nil { err = fmt.Errorf("error registering deployment hook: %w", err) return err } } errorEmitterHook := NewErrorEmitterHook(client, resGroup) err = worker.RegisterHook("erroremitter", errorEmitterHook) if err != nil { err = fmt.Errorf("error registering error emitter hook: %w", err) return err } cloneEnvGroupHook := NewCloneEnvGroupHook(client, cliConfig, resGroup) err = worker.RegisterHook("cloneenvgroup", cloneEnvGroupHook) if err != nil { err = fmt.Errorf("error registering clone env group hook: %w", err) return err } err = worker.Apply(resGroup, &switchboardTypes.ApplyOpts{ BasePath: basePath, }) return } func applyValidate() error { fileBytes, err := ioutil.ReadFile(porterYAML) if err != nil { return fmt.Errorf("error reading porter.yaml: %w", err) } validationErrors := previewInt.Validate(string(fileBytes)) if len(validationErrors) > 0 { errString := "the following error(s) were found while validating the porter.yaml file:" for _, err := range validationErrors { errString += "\n- " + strings.ReplaceAll(err.Error(), "\n\n*", "\n *") } return fmt.Errorf(errString) } return nil } func hasDeploymentHookEnvVars() bool { if ghIDStr := os.Getenv("PORTER_GIT_INSTALLATION_ID"); ghIDStr == "" { return false } if prIDStr := os.Getenv("PORTER_PULL_REQUEST_ID"); prIDStr == "" { return false } if branchFrom := os.Getenv("PORTER_BRANCH_FROM"); branchFrom == "" { return false } if branchInto := os.Getenv("PORTER_BRANCH_INTO"); branchInto == "" { return false } if actionIDStr := os.Getenv("PORTER_ACTION_ID"); actionIDStr == "" { return false } if repoName := os.Getenv("PORTER_REPO_NAME"); repoName == "" { return false } if repoOwner := os.Getenv("PORTER_REPO_OWNER"); repoOwner == "" { return false } if prName := os.Getenv("PORTER_PR_NAME"); prName == "" { return false } return true } // DeployDriver contains all information needed for deploying with switchboard type DeployDriver struct { source *previewInt.Source target *previewInt.Target output map[string]interface{} lookupTable *map[string]drivers.Driver logger *zerolog.Logger cliConfig config.CLIConfig apiClient api.Client } // NewDeployDriver creates a deployment driver for use with switchboard func NewDeployDriver(ctx context.Context, apiClient api.Client, cliConfig config.CLIConfig) func(resource *switchboardModels.Resource, opts *drivers.SharedDriverOpts) (drivers.Driver, error) { return func(resource *switchboardModels.Resource, opts *drivers.SharedDriverOpts) (drivers.Driver, error) { driver := &DeployDriver{ lookupTable: opts.DriverLookupTable, logger: opts.Logger, output: make(map[string]interface{}), cliConfig: cliConfig, apiClient: apiClient, } target, err := preview.GetTarget(ctx, resource.Name, resource.Target, apiClient, cliConfig) if err != nil { return nil, err } driver.target = target source, err := preview.GetSource(ctx, target.Project, resource.Name, resource.Source, apiClient) if err != nil { return nil, err } driver.source = source return driver, nil } } // ShouldApply extends switchboard func (d *DeployDriver) ShouldApply(_ *switchboardModels.Resource) bool { return true } // Apply extends switchboard func (d *DeployDriver) Apply(resource *switchboardModels.Resource) (*switchboardModels.Resource, error) { ctx := context.TODO() // blocked from switchboard for now _, err := d.apiClient.GetRelease( ctx, d.target.Project, d.target.Cluster, d.target.Namespace, resource.Name, ) shouldCreate := err != nil if err != nil { color.New(color.FgYellow).Printf("Could not read release %s/%s (%s): attempting creation\n", d.target.Namespace, resource.Name, err.Error()) } if d.source.IsApplication { return d.applyApplication(ctx, resource, d.apiClient, shouldCreate) } return d.applyAddon(ctx, resource, d.apiClient, shouldCreate) } // Simple apply for addons func (d *DeployDriver) applyAddon(ctx context.Context, resource *switchboardModels.Resource, client api.Client, shouldCreate bool) (*switchboardModels.Resource, error) { addonConfig, err := d.getAddonConfig(resource) if err != nil { return nil, fmt.Errorf("error getting addon config for resource %s: %w", resource.Name, err) } if shouldCreate { err := client.DeployAddon( ctx, d.target.Project, d.target.Cluster, d.target.Namespace, &types.CreateAddonRequest{ CreateReleaseBaseRequest: &types.CreateReleaseBaseRequest{ RepoURL: d.source.Repo, TemplateName: d.source.Name, TemplateVersion: d.source.Version, Values: addonConfig, Name: resource.Name, }, }, ) if err != nil { return nil, fmt.Errorf("error creating addon from resource %s: %w", resource.Name, err) } } else { bytes, err := json.Marshal(addonConfig) if err != nil { return nil, fmt.Errorf("error marshalling addon config from resource %s: %w", resource.Name, err) } err = client.UpgradeRelease( ctx, d.target.Project, d.target.Cluster, d.target.Namespace, resource.Name, &types.UpgradeReleaseRequest{ Values: string(bytes), }, ) if err != nil { return nil, fmt.Errorf("error updating addon from resource %s: %w", resource.Name, err) } } if err = d.assignOutput(ctx, resource, client); err != nil { return nil, err } return resource, nil } func (d *DeployDriver) applyApplication(ctx context.Context, resource *switchboardModels.Resource, client api.Client, shouldCreate bool) (*switchboardModels.Resource, error) { if resource == nil { return nil, fmt.Errorf("nil resource") } resourceName := resource.Name appConfig, err := d.getApplicationConfig(resource) if err != nil { return nil, err } fullPath, err := filepath.Abs(appConfig.Build.Context) if err != nil { return nil, fmt.Errorf("for resource %s, error getting absolute path for config.build.context: %w", resourceName, err) } tag := os.Getenv("PORTER_TAG") if tag == "" { color.New(color.FgYellow).Printf("for resource %s, since PORTER_TAG is not set, the Docker image tag will default to"+ " the git repo SHA\n", resourceName) commit, err := git.LastCommit() if err != nil { return nil, fmt.Errorf("for resource %s, error getting last git commit: %w", resourceName, err) } tag = commit.Sha[:7] color.New(color.FgYellow).Printf("for resource %s, using tag %s\n", resourceName, tag) } // if the method is registry and a tag is defined, we use the provided tag if appConfig.Build.Method == "registry" { imageSpl := strings.Split(appConfig.Build.Image, ":") if len(imageSpl) == 2 { tag = imageSpl[1] } if tag == "" { tag = "latest" } } sharedOpts := &deploy.SharedOpts{ ProjectID: d.target.Project, ClusterID: d.target.Cluster, Namespace: d.target.Namespace, LocalPath: fullPath, LocalDockerfile: appConfig.Build.Dockerfile, OverrideTag: tag, Method: deploy.DeployBuildType(appConfig.Build.Method), EnvGroups: appConfig.EnvGroups, UseCache: appConfig.Build.UseCache, } if appConfig.Build.UseCache { // set the docker config so that pack caching can use the repo credentials err := config.SetDockerConfig(ctx, client, d.target.Project) if err != nil { return nil, err } } if shouldCreate { resource, err = d.createApplication(ctx, resource, client, sharedOpts, appConfig) if err != nil { return nil, fmt.Errorf("error creating app from resource %s: %w", resourceName, err) } } else if !appConfig.OnlyCreate { resource, err = d.updateApplication(ctx, resource, client, sharedOpts, appConfig) if err != nil { return nil, fmt.Errorf("error updating application from resource %s: %w", resourceName, err) } } else { color.New(color.FgYellow).Printf("Skipping creation for resource %s as onlyCreate is set to true\n", resourceName) } if err = d.assignOutput(ctx, resource, client); err != nil { return nil, err } if d.source.Name == "job" && appConfig.WaitForJob && (shouldCreate || !appConfig.OnlyCreate) { color.New(color.FgYellow).Printf("Waiting for job '%s' to finish\n", resourceName) var predeployEventResponseID string stackNameWithoutRelease := strings.TrimSuffix(d.target.AppName, "-r") if strings.Contains(d.target.Namespace, "porter-stack-") { eventRequest := types.CreateOrUpdatePorterAppEventRequest{ Status: "PROGRESSING", Type: types.PorterAppEventType_PreDeploy, Metadata: map[string]any{ "start_time": time.Now().UTC(), }, } eventResponse, err := client.CreateOrUpdatePorterAppEvent(ctx, d.target.Project, d.target.Cluster, stackNameWithoutRelease, &eventRequest) if err != nil { return nil, fmt.Errorf("error creating porter app event for pre-deploy job: %s", err.Error()) } predeployEventResponseID = eventResponse.ID } err = wait.WaitForJob(ctx, client, &wait.WaitOpts{ ProjectID: d.target.Project, ClusterID: d.target.Cluster, Namespace: d.target.Namespace, Name: resourceName, }) if err != nil { if strings.Contains(d.target.Namespace, "porter-stack-") { if predeployEventResponseID == "" { return nil, errors.New("unable to find pre-deploy event response ID for failed pre-deploy event") } eventRequest := types.CreateOrUpdatePorterAppEventRequest{ ID: predeployEventResponseID, Status: "FAILED", Type: types.PorterAppEventType_PreDeploy, Metadata: map[string]any{ "end_time": time.Now().UTC(), }, } _, err := client.CreateOrUpdatePorterAppEvent(ctx, d.target.Project, d.target.Cluster, stackNameWithoutRelease, &eventRequest) if err != nil { return nil, fmt.Errorf("error updating failed porter app event for pre-deploy job: %s", err.Error()) } } if appConfig.OnlyCreate { deleteJobErr := client.DeleteRelease( ctx, d.target.Project, d.target.Cluster, d.target.Namespace, resourceName, ) if deleteJobErr != nil { return nil, fmt.Errorf("error deleting job %s with waitForJob and onlyCreate set to true: %w", resourceName, deleteJobErr) } } return nil, fmt.Errorf("error waiting for job %s: %w", resourceName, err) } if strings.Contains(d.target.Namespace, "porter-stack-") { stackNameWithoutRelease := strings.TrimSuffix(d.target.AppName, "-r") if predeployEventResponseID == "" { return nil, errors.New("unable to find pre-deploy event response ID for successful pre-deploy event") } eventRequest := types.CreateOrUpdatePorterAppEventRequest{ ID: predeployEventResponseID, Status: "SUCCESS", Type: types.PorterAppEventType_PreDeploy, Metadata: map[string]any{ "end_time": time.Now().UTC(), }, } _, err := client.CreateOrUpdatePorterAppEvent(ctx, d.target.Project, d.target.Cluster, stackNameWithoutRelease, &eventRequest) if err != nil { return nil, fmt.Errorf("error updating successful porter app event for pre-deploy job: %s", err.Error()) } } } return resource, err } func (d *DeployDriver) createApplication(ctx context.Context, resource *switchboardModels.Resource, client api.Client, sharedOpts *deploy.SharedOpts, appConf *previewInt.ApplicationConfig) (*switchboardModels.Resource, error) { // create new release color.New(color.FgGreen).Printf("Creating %s release: %s\n", d.source.Name, resource.Name) color.New(color.FgBlue).Printf("for resource %s, using registry %s\n", resource.Name, d.target.RegistryURL) // attempt to get repo suffix from environment variables var repoSuffix string if repoName := os.Getenv("PORTER_REPO_NAME"); repoName != "" { if repoOwner := os.Getenv("PORTER_REPO_OWNER"); repoOwner != "" { repoSuffix = cliUtils.SlugifyRepoSuffix(repoOwner, repoName) } } createAgent := &deploy.CreateAgent{ Client: client, CreateOpts: &deploy.CreateOpts{ SharedOpts: sharedOpts, Kind: d.source.Name, ReleaseName: resource.Name, RegistryURL: registryURL, RepoSuffix: repoSuffix, }, } var buildConfig *types.BuildConfig if appConf.Build.Builder != "" { buildConfig = &types.BuildConfig{ Builder: appConf.Build.Builder, Buildpacks: appConf.Build.Buildpacks, } } var subdomain string var err error if appConf.Build.Method == "registry" { subdomain, err = createAgent.CreateFromRegistry(ctx, appConf.Build.Image, appConf.Values) } else { // if useCache is set, create the image repository first if appConf.Build.UseCache { regID, imageURL, err := createAgent.GetImageRepoURL(ctx, resource.Name, sharedOpts.Namespace) if err != nil { return nil, err } err = client.CreateRepository( ctx, sharedOpts.ProjectID, regID, &types.CreateRegistryRepositoryRequest{ ImageRepoURI: imageURL, }, ) if err != nil { return nil, err } } subdomain, err = createAgent.CreateFromDocker(ctx, appConf.Values, sharedOpts.OverrideTag, buildConfig) } if err != nil { return nil, err } return resource, handleSubdomainCreate(subdomain, err) } func (d *DeployDriver) updateApplication(ctx context.Context, resource *switchboardModels.Resource, client api.Client, sharedOpts *deploy.SharedOpts, appConf *previewInt.ApplicationConfig) (*switchboardModels.Resource, error) { color.New(color.FgGreen).Println("Updating existing release:", resource.Name) if len(appConf.Build.Env) > 0 { sharedOpts.AdditionalEnv = appConf.Build.Env } updateAgent, err := deploy.NewDeployAgent(ctx, client, resource.Name, &deploy.DeployOpts{ SharedOpts: sharedOpts, Local: appConf.Build.Method != "registry", }) if err != nil { return nil, err } // if the build method is registry, we do not trigger a build if appConf.Build.Method != "registry" { buildEnv, err := updateAgent.GetBuildEnv(ctx, &deploy.GetBuildEnvOpts{ UseNewConfig: true, NewConfig: appConf.Values, }) if err != nil { return nil, err } err = updateAgent.SetBuildEnv(buildEnv) if err != nil { return nil, err } var buildConfig *types.BuildConfig if appConf.Build.Builder != "" { buildConfig = &types.BuildConfig{ Builder: appConf.Build.Builder, Buildpacks: appConf.Build.Buildpacks, } } err = updateAgent.Build(ctx, buildConfig) if err != nil { return nil, err } if !appConf.Build.UseCache { err = updateAgent.Push(ctx) if err != nil { return nil, err } } } if appConf.InjectBuild { // use the built image in the values if it is set // if it contains a $, then the query did not resolve if appConf.Build.Image != "" && !strings.Contains(appConf.Build.Image, "$") { imageSpl := strings.Split(appConf.Build.Image, ":") if len(imageSpl) == 2 { appConf.Values["image"] = map[string]interface{}{ "repository": imageSpl[0], "tag": imageSpl[1], } } else { return nil, fmt.Errorf("could not parse image info %s", appConf.Build.Image) } } } err = updateAgent.UpdateImageAndValues(ctx, appConf.Values) if err != nil { return nil, err } return resource, nil } func (d *DeployDriver) assignOutput(ctx context.Context, resource *switchboardModels.Resource, client api.Client) error { release, err := client.GetRelease( ctx, d.target.Project, d.target.Cluster, d.target.Namespace, resource.Name, ) if err != nil { return err } d.output = utils.CoalesceValues(d.source.SourceValues, release.Config) return nil } // Output extends switchboard func (d *DeployDriver) Output() (map[string]interface{}, error) { return d.output, nil } func (d *DeployDriver) getApplicationConfig(resource *switchboardModels.Resource) (*previewInt.ApplicationConfig, error) { populatedConf, err := drivers.ConstructConfig(&drivers.ConstructConfigOpts{ RawConf: resource.Config, LookupTable: *d.lookupTable, Dependencies: resource.Dependencies, }) if err != nil { return nil, err } appConf := &previewInt.ApplicationConfig{} err = mapstructure.Decode(populatedConf, appConf) if err != nil { return nil, err } if _, ok := resource.Config["waitForJob"]; !ok && d.source.Name == "job" { // default to true and wait for the job to finish appConf.WaitForJob = true } return appConf, nil } func (d *DeployDriver) getAddonConfig(resource *switchboardModels.Resource) (map[string]interface{}, error) { return drivers.ConstructConfig(&drivers.ConstructConfigOpts{ RawConf: resource.Config, LookupTable: *d.lookupTable, Dependencies: resource.Dependencies, }) } // DeploymentHook contains all information needed for deploying with switchboard type DeploymentHook struct { client api.Client resourceGroup *switchboardTypes.ResourceGroup gitInstallationID, projectID, clusterID, prID, actionID, envID uint branchFrom, branchInto, namespace, repoName, repoOwner, prName, commitSHA string cliConfig config.CLIConfig } // NewDeploymentHook creates a new deployment using switchboard func NewDeploymentHook(cliConfig config.CLIConfig, client api.Client, resourceGroup *switchboardTypes.ResourceGroup, namespace string) (*DeploymentHook, error) { res := &DeploymentHook{ client: client, resourceGroup: resourceGroup, namespace: namespace, cliConfig: cliConfig, } ghIDStr := os.Getenv("PORTER_GIT_INSTALLATION_ID") ghID, err := strconv.Atoi(ghIDStr) if err != nil { return nil, err } res.gitInstallationID = uint(ghID) prIDStr := os.Getenv("PORTER_PULL_REQUEST_ID") prID, err := strconv.Atoi(prIDStr) if err != nil { return nil, err } res.prID = uint(prID) res.projectID = cliConfig.Project if res.projectID == 0 { return nil, fmt.Errorf("project id must be set") } res.clusterID = cliConfig.Cluster if res.clusterID == 0 { return nil, fmt.Errorf("cluster id must be set") } branchFrom := os.Getenv("PORTER_BRANCH_FROM") res.branchFrom = branchFrom branchInto := os.Getenv("PORTER_BRANCH_INTO") res.branchInto = branchInto actionIDStr := os.Getenv("PORTER_ACTION_ID") actionID, err := strconv.Atoi(actionIDStr) if err != nil { return nil, err } res.actionID = uint(actionID) repoName := os.Getenv("PORTER_REPO_NAME") res.repoName = repoName repoOwner := os.Getenv("PORTER_REPO_OWNER") res.repoOwner = repoOwner prName := os.Getenv("PORTER_PR_NAME") res.prName = prName commit, err := git.LastCommit() if err != nil { return nil, fmt.Errorf(err.Error()) } res.commitSHA = commit.Sha[:7] return res, nil } func (t *DeploymentHook) isBranchDeploy() bool { return t.branchFrom != "" && t.branchInto != "" && t.branchFrom == t.branchInto } // PreApply extends switchboard func (t *DeploymentHook) PreApply() error { ctx := context.TODO() // switchboard blocks changing this for now if isSystemNamespace(t.namespace) { color.New(color.FgYellow).Printf("attempting to deploy to system namespace '%s'\n", t.namespace) } envList, err := t.client.ListEnvironments( ctx, t.projectID, t.clusterID, ) if err != nil { return err } envs := *envList var deplEnv *types.Environment for _, env := range envs { if strings.EqualFold(env.GitRepoOwner, t.repoOwner) && strings.EqualFold(env.GitRepoName, t.repoName) && env.GitInstallationID == t.gitInstallationID { t.envID = env.ID deplEnv = env break } } if t.envID == 0 { return fmt.Errorf("could not find environment for deployment") } nsList, err := t.client.GetK8sNamespaces( ctx, t.projectID, t.clusterID, ) if err != nil { return fmt.Errorf("error fetching namespaces: %w", err) } found := false for _, ns := range *nsList { if ns.Name == t.namespace { found = true break } } if !found { if isSystemNamespace(t.namespace) { return fmt.Errorf("attempting to deploy to system namespace '%s' which does not exist, please create it "+ "to continue", t.namespace) } createNS := &types.CreateNamespaceRequest{ Name: t.namespace, } if len(deplEnv.NamespaceLabels) > 0 { createNS.Labels = deplEnv.NamespaceLabels } // create the new namespace _, err := t.client.CreateNewK8sNamespace(ctx, t.projectID, t.clusterID, createNS) if err != nil && !strings.Contains(err.Error(), "namespace already exists") { // ignore the error if the namespace already exists // // this might happen if someone creates the namespace in between this operation return fmt.Errorf("error creating namespace: %w", err) } } var deplErr error if t.isBranchDeploy() { _, deplErr = t.client.GetDeployment( ctx, t.projectID, t.clusterID, t.envID, &types.GetDeploymentRequest{ Branch: t.branchFrom, }, ) } else { _, deplErr = t.client.GetDeployment( ctx, t.projectID, t.clusterID, t.envID, &types.GetDeploymentRequest{ PRNumber: t.prID, }, ) } if deplErr != nil && strings.Contains(deplErr.Error(), "not found") { // in this case, create the deployment createReq := &types.CreateDeploymentRequest{ Namespace: t.namespace, PullRequestID: t.prID, CreateGHDeploymentRequest: &types.CreateGHDeploymentRequest{ ActionID: t.actionID, }, GitHubMetadata: &types.GitHubMetadata{ PRName: t.prName, RepoName: t.repoName, RepoOwner: t.repoOwner, CommitSHA: t.commitSHA, PRBranchFrom: t.branchFrom, PRBranchInto: t.branchInto, }, } if t.isBranchDeploy() { createReq.PullRequestID = 0 } _, err = t.client.CreateDeployment( ctx, t.projectID, t.clusterID, createReq, ) } else if err == nil { updateReq := &types.UpdateDeploymentByClusterRequest{ RepoOwner: t.repoOwner, RepoName: t.repoName, Namespace: t.namespace, PRNumber: t.prID, CreateGHDeploymentRequest: &types.CreateGHDeploymentRequest{ ActionID: t.actionID, }, PRBranchFrom: t.branchFrom, CommitSHA: t.commitSHA, } if t.isBranchDeploy() { updateReq.PRNumber = 0 } _, err = t.client.UpdateDeployment(ctx, t.projectID, t.clusterID, updateReq) } return err } // DataQueries extends switchboard func (t *DeploymentHook) DataQueries() map[string]interface{} { res := make(map[string]interface{}) // use the resource group to find all web applications that can have an exposed subdomain // that we can query for for _, resource := range t.resourceGroup.Resources { isWeb := false if sourceNameInter, exists := resource.Source["name"]; exists { if sourceName, ok := sourceNameInter.(string); ok { if sourceName == "web" { isWeb = true } } } if isWeb { // determine if we should query for porter_hosts or just hosts isCustomDomain := false ingressMap, err := deploy.GetNestedMap(resource.Config, "values", "ingress") if err == nil { enabledVal, enabledExists := ingressMap["enabled"] customDomVal, customDomExists := ingressMap["custom_domain"] if enabledExists && customDomExists { enabled, eOK := enabledVal.(bool) customDomain, cOK := customDomVal.(bool) if eOK && cOK && enabled { if customDomain { // return the first custom domain when one exists hostsArr, hostsExists := ingressMap["hosts"] if hostsExists { hostsArrVal, hostsArrOk := hostsArr.([]interface{}) if hostsArrOk && len(hostsArrVal) > 0 { if _, ok := hostsArrVal[0].(string); ok { res[resource.Name] = fmt.Sprintf("{ .%s.ingress.hosts[0] }", resource.Name) isCustomDomain = true } } } } } } } if !isCustomDomain { res[resource.Name] = fmt.Sprintf("{ .%s.ingress.porter_hosts[0] }", resource.Name) } } } return res } // PostApply extends switchboard func (t *DeploymentHook) PostApply(populatedData map[string]interface{}) error { ctx := context.TODO() // switchboard blocks changing this for now subdomains := make([]string, 0) for _, data := range populatedData { domain, ok := data.(string) if !ok { continue } if _, err := url.Parse("https://" + domain); err == nil { subdomains = append(subdomains, "https://"+domain) } } req := &types.FinalizeDeploymentByClusterRequest{ RepoOwner: t.repoOwner, RepoName: t.repoName, Subdomain: strings.Join(subdomains, ", "), } if t.isBranchDeploy() { req.Namespace = t.namespace } else { req.PRNumber = t.prID } for _, res := range t.resourceGroup.Resources { releaseType := getReleaseType(ctx, t.projectID, res, t.client) releaseName := getReleaseName(ctx, res, t.client, t.cliConfig) if releaseType != "" && releaseName != "" { req.SuccessfulResources = append(req.SuccessfulResources, &types.SuccessfullyDeployedResource{ ReleaseName: releaseName, ReleaseType: releaseType, }) } } // finalize the deployment _, err := t.client.FinalizeDeployment(ctx, t.projectID, t.clusterID, req) return err } // OnError extends switchboard func (t *DeploymentHook) OnError(error) { ctx := context.TODO() // switchboard blocks changing this for now var deplErr error if t.isBranchDeploy() { _, deplErr = t.client.GetDeployment( ctx, t.projectID, t.clusterID, t.envID, &types.GetDeploymentRequest{ Branch: t.branchFrom, }, ) } else { _, deplErr = t.client.GetDeployment( ctx, t.projectID, t.clusterID, t.envID, &types.GetDeploymentRequest{ PRNumber: t.prID, }, ) } // if the deployment exists, throw an error for that deployment if deplErr == nil { req := &types.UpdateDeploymentStatusByClusterRequest{ RepoOwner: t.repoOwner, RepoName: t.repoName, CreateGHDeploymentRequest: &types.CreateGHDeploymentRequest{ ActionID: t.actionID, }, PRBranchFrom: t.branchFrom, Status: string(types.DeploymentStatusFailed), } if t.isBranchDeploy() { req.Namespace = t.namespace } else { req.PRNumber = t.prID } // FIXME: try to use the error with a custom logger t.client.UpdateDeploymentStatus(ctx, t.projectID, t.clusterID, req) //nolint:errcheck,gosec // do not want to change logic of CLI. New linter error } } // OnConsolidatedErrors extends switchboard func (t *DeploymentHook) OnConsolidatedErrors(allErrors map[string]error) { ctx := context.TODO() // switchboard blocks changing this for now var deplErr error if t.isBranchDeploy() { _, deplErr = t.client.GetDeployment( ctx, t.projectID, t.clusterID, t.envID, &types.GetDeploymentRequest{ Branch: t.branchFrom, }, ) } else { _, deplErr = t.client.GetDeployment( ctx, t.projectID, t.clusterID, t.envID, &types.GetDeploymentRequest{ PRNumber: t.prID, }, ) } // if the deployment exists, throw an error for that deployment if deplErr == nil { req := &types.FinalizeDeploymentWithErrorsByClusterRequest{ RepoOwner: t.repoOwner, RepoName: t.repoName, Errors: make(map[string]string), } if t.isBranchDeploy() { req.Namespace = t.namespace } else { req.PRNumber = t.prID } for _, res := range t.resourceGroup.Resources { if _, ok := allErrors[res.Name]; !ok { req.SuccessfulResources = append(req.SuccessfulResources, &types.SuccessfullyDeployedResource{ ReleaseName: getReleaseName(ctx, res, t.client, t.cliConfig), ReleaseType: getReleaseType(ctx, t.projectID, res, t.client), }) } } for res, err := range allErrors { req.Errors[res] = err.Error() } // FIXME: handle the error t.client.FinalizeDeploymentWithErrors(ctx, t.projectID, t.clusterID, req) //nolint:errcheck,gosec // do not want to change logic of CLI. New linter error } } // CloneEnvGroupHook contains all information needed to clone an env group type CloneEnvGroupHook struct { client api.Client resGroup *switchboardTypes.ResourceGroup cliConfig config.CLIConfig } // NewCloneEnvGroupHook wraps switchboard for cloning env groups func NewCloneEnvGroupHook(client api.Client, cliConfig config.CLIConfig, resourceGroup *switchboardTypes.ResourceGroup) *CloneEnvGroupHook { return &CloneEnvGroupHook{ client: client, cliConfig: cliConfig, resGroup: resourceGroup, } } func (t *CloneEnvGroupHook) PreApply() error { ctx := context.TODO() // switchboard blocks changing this for now for _, res := range t.resGroup.Resources { if res.Driver == "env-group" { continue } appConf := &previewInt.ApplicationConfig{} err := mapstructure.Decode(res.Config, &appConf) if err != nil { continue } if appConf != nil && len(appConf.EnvGroups) > 0 { target, err := preview.GetTarget(ctx, res.Name, res.Target, t.client, t.cliConfig) if err != nil { return err } for _, group := range appConf.EnvGroups { if group.Name == "" { return fmt.Errorf("env group name cannot be empty") } _, err := t.client.GetEnvGroup( ctx, target.Project, target.Cluster, target.Namespace, &types.GetEnvGroupRequest{ Name: group.Name, Version: group.Version, }, ) if err != nil && err.Error() == "env group not found" { if group.Namespace == "" { return fmt.Errorf("env group namespace cannot be empty") } color.New(color.FgBlue, color.Bold). Printf("Env group '%s' does not exist in the target namespace '%s'\n", group.Name, target.Namespace) color.New(color.FgBlue, color.Bold). Printf("Cloning env group '%s' from namespace '%s' to target namespace '%s'\n", group.Name, group.Namespace, target.Namespace) _, err = t.client.CloneEnvGroup( ctx, target.Project, target.Cluster, group.Namespace, &types.CloneEnvGroupRequest{ SourceName: group.Name, TargetNamespace: target.Namespace, }, ) if err != nil { return err } } else if err != nil { return err } } } } return nil } func (t *CloneEnvGroupHook) DataQueries() map[string]interface{} { return nil } func (t *CloneEnvGroupHook) PostApply(map[string]interface{}) error { return nil } func (t *CloneEnvGroupHook) OnError(error) {} func (t *CloneEnvGroupHook) OnConsolidatedErrors(map[string]error) {} func getReleaseName(ctx context.Context, res *switchboardTypes.Resource, apiClient api.Client, cliConfig config.CLIConfig) string { // can ignore the error because this method is called once // GetTarget has alrealy been called and validated previously target, _ := preview.GetTarget(ctx, res.Name, res.Target, apiClient, cliConfig) if target.AppName != "" { return target.AppName } return res.Name } func getReleaseType(ctx context.Context, projectID uint, res *switchboardTypes.Resource, apiClient api.Client) string { // can ignore the error because this method is called once // GetSource has alrealy been called and validated previously source, _ := preview.GetSource(ctx, projectID, res.Name, res.Source, apiClient) if source != nil && source.Name != "" { return source.Name } return "" } func isSystemNamespace(namespace string) bool { systemNamespaces := map[string]bool{ "ack-system": true, "cert-manager": true, "default": true, "ingress-nginx": true, "ingress-nginx-private": true, "kube-node-lease": true, "kube-public": true, "kube-system": true, "monitoring": true, "porter-agent-system": true, } return systemNamespaces[namespace] } type ErrorEmitterHook struct{} // NewErrorEmitterHook handles switchboard errors func NewErrorEmitterHook(api.Client, *switchboardTypes.ResourceGroup) *ErrorEmitterHook { return &ErrorEmitterHook{} } func (t *ErrorEmitterHook) PreApply() error { return nil } func (t *ErrorEmitterHook) DataQueries() map[string]interface{} { return nil } func (t *ErrorEmitterHook) PostApply(map[string]interface{}) error { return nil } func (t *ErrorEmitterHook) OnError(err error) { color.New(color.FgRed).Fprintf(os.Stderr, "Errors while building: %s\n", err.Error()) } func (t *ErrorEmitterHook) OnConsolidatedErrors(errMap map[string]error) { color.New(color.FgRed).Fprintf(os.Stderr, "Errors while building:\n") for resName, err := range errMap { color.New(color.FgRed).Fprintf(os.Stderr, " - %s: %s\n", resName, err.Error()) } }