diff --git a/Gopkg.lock b/Gopkg.lock index 3595589ac4..bd5d27c2b5 100644 --- a/Gopkg.lock +++ b/Gopkg.lock @@ -322,7 +322,7 @@ [[projects]] branch = "master" - digest = "1:80634274ffb798752b241ccd930a2f371e8c82aa0a10a50d723b40fe35f5769e" + digest = "1:5c49f94452161da04aeff1fc01e503050e6de02a704de579cfcd4cb643ef41bb" name = "github.com/openshift/api" packages = [ "build/v1", @@ -333,7 +333,7 @@ "osin/v1", ] pruneopts = "NT" - revision = "fb5c7e9aaf18be4888c03b8a4c9be78c613e2ef9" + revision = "13b403bfb6ce84ddc053bd3b401b5d67bf175efa" [[projects]] branch = "master" diff --git a/pkg/controller/clusterconfig/clusterconfig_controller.go b/pkg/controller/clusterconfig/clusterconfig_controller.go index 667160839e..30b9d3b139 100644 --- a/pkg/controller/clusterconfig/clusterconfig_controller.go +++ b/pkg/controller/clusterconfig/clusterconfig_controller.go @@ -91,7 +91,7 @@ func (r *ReconcileClusterConfig) Reconcile(request reconcile.Request) (reconcile // Validate the cluster config if err := network.ValidateClusterConfig(clusterConfig.Spec); err != nil { log.Printf("Failed to validate Network.Spec: %v", err) - r.status.SetFailing(statusmanager.ClusterConfig, "InvalidClusterConfig", + r.status.SetDegraded(statusmanager.ClusterConfig, "InvalidClusterConfig", fmt.Sprintf("The cluster configuration is invalid (%v). Use 'oc edit network.config.openshift.io cluster' to fix.", err)) return reconcile.Result{}, err } @@ -99,7 +99,7 @@ func (r *ReconcileClusterConfig) Reconcile(request reconcile.Request) (reconcile operatorConfig, err := r.UpdateOperatorConfig(context.TODO(), *clusterConfig) if err != nil { log.Printf("Failed to generate NetworkConfig CRD: %v", err) - r.status.SetFailing(statusmanager.ClusterConfig, "UpdateOperatorConfig", + r.status.SetDegraded(statusmanager.ClusterConfig, "UpdateOperatorConfig", fmt.Sprintf("Internal error while converting cluster configuration: %v", err)) return reconcile.Result{}, err } @@ -107,12 +107,12 @@ func (r *ReconcileClusterConfig) Reconcile(request reconcile.Request) (reconcile if operatorConfig != nil { if err := apply.ApplyObject(context.TODO(), r.client, operatorConfig); err != nil { log.Printf("Could not apply operator config: %v", err) - r.status.SetFailing(statusmanager.ClusterConfig, "ApplyOperatorConfig", + r.status.SetDegraded(statusmanager.ClusterConfig, "ApplyOperatorConfig", fmt.Sprintf("Error while trying to update operator configuration: %v", err)) return reconcile.Result{}, err } } - r.status.SetNotFailing(statusmanager.ClusterConfig) + r.status.SetNotDegraded(statusmanager.ClusterConfig) return reconcile.Result{}, nil } diff --git a/pkg/controller/operconfig/operconfig_controller.go b/pkg/controller/operconfig/operconfig_controller.go index 73bbd34c4d..3aff13c799 100644 --- a/pkg/controller/operconfig/operconfig_controller.go +++ b/pkg/controller/operconfig/operconfig_controller.go @@ -118,7 +118,7 @@ func (r *ReconcileOperConfig) Reconcile(request reconcile.Request) (reconcile.Re err := r.client.Get(context.TODO(), request.NamespacedName, operConfig) if err != nil { if apierrors.IsNotFound(err) { - r.status.SetFailing(statusmanager.OperatorConfig, "NoOperatorConfig", + r.status.SetDegraded(statusmanager.OperatorConfig, "NoOperatorConfig", fmt.Sprintf("Operator configuration %s was deleted", request.NamespacedName.String())) // Request object not found, could have been deleted after reconcile request. // Owned objects are automatically garbage collected, since we set @@ -136,7 +136,7 @@ func (r *ReconcileOperConfig) Reconcile(request reconcile.Request) (reconcile.Re // This will also commit the change back to the apiserver. if err := r.MergeClusterConfig(context.TODO(), operConfig); err != nil { log.Printf("Failed to merge the cluster configuration: %v", err) - r.status.SetFailing(statusmanager.OperatorConfig, "MergeClusterConfig", + r.status.SetDegraded(statusmanager.OperatorConfig, "MergeClusterConfig", fmt.Sprintf("Internal error while merging cluster configuration and operator configuration: %v", err)) return reconcile.Result{}, err } @@ -147,7 +147,7 @@ func (r *ReconcileOperConfig) Reconcile(request reconcile.Request) (reconcile.Re // Validate the configuration if err := network.Validate(&operConfig.Spec); err != nil { log.Printf("Failed to validate Network.operator.openshift.io.Spec: %v", err) - r.status.SetFailing(statusmanager.OperatorConfig, "InvalidOperatorConfig", + r.status.SetDegraded(statusmanager.OperatorConfig, "InvalidOperatorConfig", fmt.Sprintf("The operator configuration is invalid (%v). Use 'oc edit network.operator.openshift.io cluster' to fix.", err)) return reconcile.Result{}, err } @@ -171,7 +171,7 @@ func (r *ReconcileOperConfig) Reconcile(request reconcile.Request) (reconcile.Re err = network.IsChangeSafe(prev, &operConfig.Spec) if err != nil { log.Printf("Not applying unsafe change: %v", err) - r.status.SetFailing(statusmanager.OperatorConfig, "InvalidOperatorConfig", + r.status.SetDegraded(statusmanager.OperatorConfig, "InvalidOperatorConfig", fmt.Sprintf("Not applying unsafe configuration change: %v. Use 'oc edit network.operator.openshift.io cluster' to undo the change.", err)) return reconcile.Result{}, err } @@ -181,7 +181,7 @@ func (r *ReconcileOperConfig) Reconcile(request reconcile.Request) (reconcile.Re objs, err := network.Render(&operConfig.Spec, ManifestPath) if err != nil { log.Printf("Failed to render: %v", err) - r.status.SetFailing(statusmanager.OperatorConfig, "RenderError", + r.status.SetDegraded(statusmanager.OperatorConfig, "RenderError", fmt.Sprintf("Internal error while rendering operator configuration: %v", err)) return reconcile.Result{}, err } @@ -190,7 +190,7 @@ func (r *ReconcileOperConfig) Reconcile(request reconcile.Request) (reconcile.Re app, err := AppliedConfiguration(operConfig) if err != nil { log.Printf("Failed to render applied: %v", err) - r.status.SetFailing(statusmanager.OperatorConfig, "RenderError", + r.status.SetDegraded(statusmanager.OperatorConfig, "RenderError", fmt.Sprintf("Internal error while recording new operator configuration: %v", err)) return reconcile.Result{}, err } @@ -220,7 +220,7 @@ func (r *ReconcileOperConfig) Reconcile(request reconcile.Request) (reconcile.Re if err := controllerutil.SetControllerReference(operConfig, obj, r.scheme); err != nil { err = errors.Wrapf(err, "could not set reference for (%s) %s/%s", obj.GroupVersionKind(), obj.GetNamespace(), obj.GetName()) log.Println(err) - r.status.SetFailing(statusmanager.OperatorConfig, "InternalError", + r.status.SetDegraded(statusmanager.OperatorConfig, "InternalError", fmt.Sprintf("Internal error while updating operator configuration: %v", err)) return reconcile.Result{}, err } @@ -238,7 +238,7 @@ func (r *ReconcileOperConfig) Reconcile(request reconcile.Request) (reconcile.Re continue } } - r.status.SetFailing(statusmanager.OperatorConfig, "ApplyOperatorConfig", + r.status.SetDegraded(statusmanager.OperatorConfig, "ApplyOperatorConfig", fmt.Sprintf("Error while updating operator configuration: %v", err)) return reconcile.Result{}, err } @@ -248,7 +248,7 @@ func (r *ReconcileOperConfig) Reconcile(request reconcile.Request) (reconcile.Re status, err := r.ClusterNetworkStatus(context.TODO(), operConfig) if err != nil { log.Printf("Could not generate network status: %v", err) - r.status.SetFailing(statusmanager.OperatorConfig, "StatusError", + r.status.SetDegraded(statusmanager.OperatorConfig, "StatusError", fmt.Sprintf("Could not update cluster configuration status: %v", err)) return reconcile.Result{}, err } @@ -258,13 +258,13 @@ func (r *ReconcileOperConfig) Reconcile(request reconcile.Request) (reconcile.Re if err := apply.ApplyObject(context.TODO(), r.client, status); err != nil { err = errors.Wrapf(err, "could not apply (%s) %s/%s", status.GroupVersionKind(), status.GetNamespace(), status.GetName()) log.Println(err) - r.status.SetFailing(statusmanager.OperatorConfig, "StatusError", + r.status.SetDegraded(statusmanager.OperatorConfig, "StatusError", fmt.Sprintf("Could not update cluster configuration status: %v", err)) return reconcile.Result{}, err } } - r.status.SetNotFailing(statusmanager.OperatorConfig) + r.status.SetNotDegraded(statusmanager.OperatorConfig) // All was successful. Request that this be re-triggered after ResyncPeriod, // so we can reconcile state again. diff --git a/pkg/controller/statusmanager/status_manager.go b/pkg/controller/statusmanager/status_manager.go index be4936145e..83f2708045 100644 --- a/pkg/controller/statusmanager/status_manager.go +++ b/pkg/controller/statusmanager/status_manager.go @@ -109,8 +109,8 @@ func (status *StatusManager) Set(reachedAvailableLevel bool, conditions ...confi } } -// syncFailing syncs the current Failing status -func (status *StatusManager) syncFailing() { +// syncDegraded syncs the current Degraded status +func (status *StatusManager) syncDegraded() { for _, c := range status.failing { if c != nil { status.Set(false, *c) @@ -120,32 +120,32 @@ func (status *StatusManager) syncFailing() { status.Set( false, configv1.ClusterOperatorStatusCondition{ - Type: configv1.OperatorFailing, + Type: configv1.OperatorDegraded, Status: configv1.ConditionFalse, }, ) } -// SetFailing marks the operator as Failing with the given reason and message. If it +// SetDegraded marks the operator as Degraded with the given reason and message. If it // is not already failing for a lower-level reason, the operator's status will be updated. -func (status *StatusManager) SetFailing(level StatusLevel, reason, message string) { +func (status *StatusManager) SetDegraded(level StatusLevel, reason, message string) { status.failing[level] = &configv1.ClusterOperatorStatusCondition{ - Type: configv1.OperatorFailing, + Type: configv1.OperatorDegraded, Status: configv1.ConditionTrue, Reason: reason, Message: message, } - status.syncFailing() + status.syncDegraded() } -// SetNotFailing marks the operator as not Failing at the given level. If the operator +// SetNotDegraded marks the operator as not Degraded at the given level. If the operator // status previously indicated failure at this level, it will updated to show the next // higher-level failure, or else to show that the operator is no longer failing. -func (status *StatusManager) SetNotFailing(level StatusLevel) { +func (status *StatusManager) SetNotDegraded(level StatusLevel) { if status.failing[level] != nil { status.failing[level] = nil } - status.syncFailing() + status.syncDegraded() } func (status *StatusManager) SetDaemonSets(daemonSets []types.NamespacedName) { @@ -156,9 +156,8 @@ func (status *StatusManager) SetDeployments(deployments []types.NamespacedName) status.deployments = deployments } -// SetFromPods sets the operator status to Failing, Progressing, or Available, based on -// the current status of the manager's DaemonSets and Deployments. However, this is a -// no-op if the StatusManager is currently marked as failing due to a configuration error. +// SetFromPods sets the operator Degraded/Progressing/Available status, based on +// the current status of the manager's DaemonSets and Deployments. func (status *StatusManager) SetFromPods() { targetLevel := os.Getenv("RELEASE_VERSION") @@ -170,10 +169,10 @@ func (status *StatusManager) SetFromPods() { ns := &corev1.Namespace{} if err := status.client.Get(context.TODO(), types.NamespacedName{Name: dsName.Namespace}, ns); err != nil { if errors.IsNotFound(err) { - status.SetFailing(PodDeployment, "NoNamespace", + status.SetDegraded(PodDeployment, "NoNamespace", fmt.Sprintf("Namespace %q does not exist", dsName.Namespace)) } else { - status.SetFailing(PodDeployment, "InternalError", + status.SetDegraded(PodDeployment, "InternalError", fmt.Sprintf("Internal error deploying pods: %v", err)) } return @@ -182,10 +181,10 @@ func (status *StatusManager) SetFromPods() { ds := &appsv1.DaemonSet{} if err := status.client.Get(context.TODO(), dsName, ds); err != nil { if errors.IsNotFound(err) { - status.SetFailing(PodDeployment, "NoDaemonSet", + status.SetDegraded(PodDeployment, "NoDaemonSet", fmt.Sprintf("Expected DaemonSet %q does not exist", dsName.String())) } else { - status.SetFailing(PodDeployment, "InternalError", + status.SetDegraded(PodDeployment, "InternalError", fmt.Sprintf("Internal error deploying pods: %v", err)) } return @@ -210,10 +209,10 @@ func (status *StatusManager) SetFromPods() { ns := &corev1.Namespace{} if err := status.client.Get(context.TODO(), types.NamespacedName{Name: depName.Namespace}, ns); err != nil { if errors.IsNotFound(err) { - status.SetFailing(PodDeployment, "NoNamespace", + status.SetDegraded(PodDeployment, "NoNamespace", fmt.Sprintf("Namespace %q does not exist", depName.Namespace)) } else { - status.SetFailing(PodDeployment, "InternalError", + status.SetDegraded(PodDeployment, "InternalError", fmt.Sprintf("Internal error deploying pods: %v", err)) } return @@ -222,10 +221,10 @@ func (status *StatusManager) SetFromPods() { dep := &appsv1.Deployment{} if err := status.client.Get(context.TODO(), depName, dep); err != nil { if errors.IsNotFound(err) { - status.SetFailing(PodDeployment, "NoDeployment", + status.SetDegraded(PodDeployment, "NoDeployment", fmt.Sprintf("Expected Deployment %q does not exist", depName.String())) } else { - status.SetFailing(PodDeployment, "InternalError", + status.SetDegraded(PodDeployment, "InternalError", fmt.Sprintf("Internal error deploying pods: %v", err)) } return @@ -250,7 +249,7 @@ func (status *StatusManager) SetFromPods() { } } - status.SetNotFailing(PodDeployment) + status.SetNotDegraded(PodDeployment) if len(progressing) > 0 { status.Set( diff --git a/pkg/controller/statusmanager/status_manager_test.go b/pkg/controller/statusmanager/status_manager_test.go index 403eb78e9b..337418e6fa 100644 --- a/pkg/controller/statusmanager/status_manager_test.go +++ b/pkg/controller/statusmanager/status_manager_test.go @@ -72,7 +72,7 @@ func TestStatusManager_set(t *testing.T) { } condFail := configv1.ClusterOperatorStatusCondition{ - Type: configv1.OperatorFailing, + Type: configv1.OperatorDegraded, Status: configv1.ConditionTrue, Reason: "Reason", Message: "Message", @@ -102,7 +102,7 @@ func TestStatusManager_set(t *testing.T) { } condNoFail := configv1.ClusterOperatorStatusCondition{ - Type: configv1.OperatorFailing, + Type: configv1.OperatorDegraded, Status: configv1.ConditionFalse, } status.Set(false, condNoFail) @@ -134,7 +134,7 @@ func TestStatusManager_set(t *testing.T) { } } -func TestStatusManagerSetFailing(t *testing.T) { +func TestStatusManagerSetDegraded(t *testing.T) { client := fake.NewFakeClient() status := New(client, "testing", "1.2.3") @@ -144,23 +144,23 @@ func TestStatusManagerSetFailing(t *testing.T) { } condFailCluster := configv1.ClusterOperatorStatusCondition{ - Type: configv1.OperatorFailing, + Type: configv1.OperatorDegraded, Status: configv1.ConditionTrue, Reason: "Cluster", } condFailOperator := configv1.ClusterOperatorStatusCondition{ - Type: configv1.OperatorFailing, + Type: configv1.OperatorDegraded, Status: configv1.ConditionTrue, Reason: "Operator", } condFailPods := configv1.ClusterOperatorStatusCondition{ - Type: configv1.OperatorFailing, + Type: configv1.OperatorDegraded, Status: configv1.ConditionTrue, Reason: "Pods", } // Initial failure status - status.SetFailing(OperatorConfig, "Operator", "") + status.SetDegraded(OperatorConfig, "Operator", "") co, err = getCO(client, "testing") if err != nil { t.Fatalf("error getting ClusterOperator: %v", err) @@ -170,7 +170,7 @@ func TestStatusManagerSetFailing(t *testing.T) { } // Setting a higher-level status should override it - status.SetFailing(ClusterConfig, "Cluster", "") + status.SetDegraded(ClusterConfig, "Cluster", "") co, err = getCO(client, "testing") if err != nil { t.Fatalf("error getting ClusterOperator: %v", err) @@ -180,7 +180,7 @@ func TestStatusManagerSetFailing(t *testing.T) { } // Setting a lower-level status should be ignored - status.SetFailing(PodDeployment, "Pods", "") + status.SetDegraded(PodDeployment, "Pods", "") co, err = getCO(client, "testing") if err != nil { t.Fatalf("error getting ClusterOperator: %v", err) @@ -190,7 +190,7 @@ func TestStatusManagerSetFailing(t *testing.T) { } // Clearing an unseen status should have no effect - status.SetNotFailing(OperatorConfig) + status.SetNotDegraded(OperatorConfig) co, err = getCO(client, "testing") if err != nil { t.Fatalf("error getting ClusterOperator: %v", err) @@ -200,7 +200,7 @@ func TestStatusManagerSetFailing(t *testing.T) { } // Clearing the currently-seen status should reveal the higher-level status - status.SetNotFailing(ClusterConfig) + status.SetNotDegraded(ClusterConfig) co, err = getCO(client, "testing") if err != nil { t.Fatalf("error getting ClusterOperator: %v", err) @@ -210,12 +210,12 @@ func TestStatusManagerSetFailing(t *testing.T) { } // Clearing all failing statuses should result in not failing - status.SetNotFailing(PodDeployment) + status.SetNotDegraded(PodDeployment) co, err = getCO(client, "testing") if err != nil { t.Fatalf("error getting ClusterOperator: %v", err) } - if !v1helpers.IsStatusConditionFalse(co.Status.Conditions, configv1.OperatorFailing) { + if !v1helpers.IsStatusConditionFalse(co.Status.Conditions, configv1.OperatorDegraded) { t.Fatalf("unexpected Status.Conditions: %#v", co.Status.Conditions) } } @@ -236,7 +236,7 @@ func TestStatusManagerSetFromDaemonSets(t *testing.T) { } if !conditionsInclude(co.Status.Conditions, []configv1.ClusterOperatorStatusCondition{ { - Type: configv1.OperatorFailing, + Type: configv1.OperatorDegraded, Status: configv1.ConditionTrue, Reason: "NoNamespace", }, @@ -263,7 +263,7 @@ func TestStatusManagerSetFromDaemonSets(t *testing.T) { } if !conditionsInclude(co.Status.Conditions, []configv1.ClusterOperatorStatusCondition{ { - Type: configv1.OperatorFailing, + Type: configv1.OperatorDegraded, Status: configv1.ConditionTrue, Reason: "NoDaemonSet", }, @@ -291,7 +291,7 @@ func TestStatusManagerSetFromDaemonSets(t *testing.T) { } if !conditionsInclude(co.Status.Conditions, []configv1.ClusterOperatorStatusCondition{ { - Type: configv1.OperatorFailing, + Type: configv1.OperatorDegraded, Status: configv1.ConditionFalse, }, { @@ -396,7 +396,7 @@ func TestStatusManagerSetFromDaemonSets(t *testing.T) { } if !conditionsInclude(co.Status.Conditions, []configv1.ClusterOperatorStatusCondition{ { - Type: configv1.OperatorFailing, + Type: configv1.OperatorDegraded, Status: configv1.ConditionFalse, }, { @@ -438,7 +438,7 @@ func TestStatusManagerSetFromPods(t *testing.T) { } if !conditionsInclude(co.Status.Conditions, []configv1.ClusterOperatorStatusCondition{ { - Type: configv1.OperatorFailing, + Type: configv1.OperatorDegraded, Status: configv1.ConditionTrue, Reason: "NoNamespace", }, @@ -461,7 +461,7 @@ func TestStatusManagerSetFromPods(t *testing.T) { } if !conditionsInclude(co.Status.Conditions, []configv1.ClusterOperatorStatusCondition{ { - Type: configv1.OperatorFailing, + Type: configv1.OperatorDegraded, Status: configv1.ConditionTrue, Reason: "NoDeployment", }, @@ -488,7 +488,7 @@ func TestStatusManagerSetFromPods(t *testing.T) { } if !conditionsInclude(co.Status.Conditions, []configv1.ClusterOperatorStatusCondition{ { - Type: configv1.OperatorFailing, + Type: configv1.OperatorDegraded, Status: configv1.ConditionTrue, Reason: "NoDeployment", }, @@ -510,7 +510,7 @@ func TestStatusManagerSetFromPods(t *testing.T) { } if !conditionsInclude(co.Status.Conditions, []configv1.ClusterOperatorStatusCondition{ { - Type: configv1.OperatorFailing, + Type: configv1.OperatorDegraded, Status: configv1.ConditionFalse, }, { @@ -542,7 +542,7 @@ func TestStatusManagerSetFromPods(t *testing.T) { } if !conditionsInclude(co.Status.Conditions, []configv1.ClusterOperatorStatusCondition{ { - Type: configv1.OperatorFailing, + Type: configv1.OperatorDegraded, Status: configv1.ConditionFalse, }, { @@ -574,7 +574,7 @@ func TestStatusManagerSetFromPods(t *testing.T) { } if !conditionsInclude(co.Status.Conditions, []configv1.ClusterOperatorStatusCondition{ { - Type: configv1.OperatorFailing, + Type: configv1.OperatorDegraded, Status: configv1.ConditionTrue, Reason: "NoDaemonSet", }, @@ -604,7 +604,7 @@ func TestStatusManagerSetFromPods(t *testing.T) { // Available before if !conditionsInclude(co.Status.Conditions, []configv1.ClusterOperatorStatusCondition{ { - Type: configv1.OperatorFailing, + Type: configv1.OperatorDegraded, Status: configv1.ConditionFalse, }, { @@ -633,7 +633,7 @@ func TestStatusManagerSetFromPods(t *testing.T) { } if !conditionsInclude(co.Status.Conditions, []configv1.ClusterOperatorStatusCondition{ { - Type: configv1.OperatorFailing, + Type: configv1.OperatorDegraded, Status: configv1.ConditionFalse, }, { diff --git a/vendor/github.com/openshift/api/apps/v1/generated.pb.go b/vendor/github.com/openshift/api/apps/v1/generated.pb.go index 59e0b05516..ae30800935 100644 --- a/vendor/github.com/openshift/api/apps/v1/generated.pb.go +++ b/vendor/github.com/openshift/api/apps/v1/generated.pb.go @@ -1,6 +1,5 @@ -// Code generated by protoc-gen-gogo. +// Code generated by protoc-gen-gogo. DO NOT EDIT. // source: github.com/openshift/api/apps/v1/generated.proto -// DO NOT EDIT! /* Package v1 is a generated protocol buffer package. @@ -1380,24 +1379,6 @@ func (m *TagImageHook) MarshalTo(dAtA []byte) (int, error) { return i, nil } -func encodeFixed64Generated(dAtA []byte, offset int, v uint64) int { - dAtA[offset] = uint8(v) - dAtA[offset+1] = uint8(v >> 8) - dAtA[offset+2] = uint8(v >> 16) - dAtA[offset+3] = uint8(v >> 24) - dAtA[offset+4] = uint8(v >> 32) - dAtA[offset+5] = uint8(v >> 40) - dAtA[offset+6] = uint8(v >> 48) - dAtA[offset+7] = uint8(v >> 56) - return offset + 8 -} -func encodeFixed32Generated(dAtA []byte, offset int, v uint32) int { - dAtA[offset] = uint8(v) - dAtA[offset+1] = uint8(v >> 8) - dAtA[offset+2] = uint8(v >> 16) - dAtA[offset+3] = uint8(v >> 24) - return offset + 4 -} func encodeVarintGenerated(dAtA []byte, offset int, v uint64) int { for v >= 1<<7 { dAtA[offset] = uint8(v&0x7f | 0x80) @@ -3073,51 +3054,14 @@ func (m *DeploymentConfigRollback) Unmarshal(dAtA []byte) error { if postIndex > l { return io.ErrUnexpectedEOF } - var keykey uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - keykey |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - var stringLenmapkey uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - stringLenmapkey |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - intStringLenmapkey := int(stringLenmapkey) - if intStringLenmapkey < 0 { - return ErrInvalidLengthGenerated - } - postStringIndexmapkey := iNdEx + intStringLenmapkey - if postStringIndexmapkey > l { - return io.ErrUnexpectedEOF - } - mapkey := string(dAtA[iNdEx:postStringIndexmapkey]) - iNdEx = postStringIndexmapkey if m.UpdatedAnnotations == nil { m.UpdatedAnnotations = make(map[string]string) } - if iNdEx < postIndex { - var valuekey uint64 + var mapkey string + var mapvalue string + for iNdEx < postIndex { + entryPreIndex := iNdEx + var wire uint64 for shift := uint(0); ; shift += 7 { if shift >= 64 { return ErrIntOverflowGenerated @@ -3127,41 +3071,80 @@ func (m *DeploymentConfigRollback) Unmarshal(dAtA []byte) error { } b := dAtA[iNdEx] iNdEx++ - valuekey |= (uint64(b) & 0x7F) << shift + wire |= (uint64(b) & 0x7F) << shift if b < 0x80 { break } } - var stringLenmapvalue uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated + fieldNum := int32(wire >> 3) + if fieldNum == 1 { + var stringLenmapkey uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLenmapkey |= (uint64(b) & 0x7F) << shift + if b < 0x80 { + break + } } - if iNdEx >= l { + intStringLenmapkey := int(stringLenmapkey) + if intStringLenmapkey < 0 { + return ErrInvalidLengthGenerated + } + postStringIndexmapkey := iNdEx + intStringLenmapkey + if postStringIndexmapkey > l { return io.ErrUnexpectedEOF } - b := dAtA[iNdEx] - iNdEx++ - stringLenmapvalue |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break + mapkey = string(dAtA[iNdEx:postStringIndexmapkey]) + iNdEx = postStringIndexmapkey + } else if fieldNum == 2 { + var stringLenmapvalue uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLenmapvalue |= (uint64(b) & 0x7F) << shift + if b < 0x80 { + break + } } + intStringLenmapvalue := int(stringLenmapvalue) + if intStringLenmapvalue < 0 { + return ErrInvalidLengthGenerated + } + postStringIndexmapvalue := iNdEx + intStringLenmapvalue + if postStringIndexmapvalue > l { + return io.ErrUnexpectedEOF + } + mapvalue = string(dAtA[iNdEx:postStringIndexmapvalue]) + iNdEx = postStringIndexmapvalue + } else { + iNdEx = entryPreIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > postIndex { + return io.ErrUnexpectedEOF + } + iNdEx += skippy } - intStringLenmapvalue := int(stringLenmapvalue) - if intStringLenmapvalue < 0 { - return ErrInvalidLengthGenerated - } - postStringIndexmapvalue := iNdEx + intStringLenmapvalue - if postStringIndexmapvalue > l { - return io.ErrUnexpectedEOF - } - mapvalue := string(dAtA[iNdEx:postStringIndexmapvalue]) - iNdEx = postStringIndexmapvalue - m.UpdatedAnnotations[mapkey] = mapvalue - } else { - var mapvalue string - m.UpdatedAnnotations[mapkey] = mapvalue } + m.UpdatedAnnotations[mapkey] = mapvalue iNdEx = postIndex case 3: if wireType != 2 { @@ -3590,51 +3573,14 @@ func (m *DeploymentConfigSpec) Unmarshal(dAtA []byte) error { if postIndex > l { return io.ErrUnexpectedEOF } - var keykey uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - keykey |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - var stringLenmapkey uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - stringLenmapkey |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - intStringLenmapkey := int(stringLenmapkey) - if intStringLenmapkey < 0 { - return ErrInvalidLengthGenerated - } - postStringIndexmapkey := iNdEx + intStringLenmapkey - if postStringIndexmapkey > l { - return io.ErrUnexpectedEOF - } - mapkey := string(dAtA[iNdEx:postStringIndexmapkey]) - iNdEx = postStringIndexmapkey if m.Selector == nil { m.Selector = make(map[string]string) } - if iNdEx < postIndex { - var valuekey uint64 + var mapkey string + var mapvalue string + for iNdEx < postIndex { + entryPreIndex := iNdEx + var wire uint64 for shift := uint(0); ; shift += 7 { if shift >= 64 { return ErrIntOverflowGenerated @@ -3644,41 +3590,80 @@ func (m *DeploymentConfigSpec) Unmarshal(dAtA []byte) error { } b := dAtA[iNdEx] iNdEx++ - valuekey |= (uint64(b) & 0x7F) << shift + wire |= (uint64(b) & 0x7F) << shift if b < 0x80 { break } } - var stringLenmapvalue uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated + fieldNum := int32(wire >> 3) + if fieldNum == 1 { + var stringLenmapkey uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLenmapkey |= (uint64(b) & 0x7F) << shift + if b < 0x80 { + break + } } - if iNdEx >= l { + intStringLenmapkey := int(stringLenmapkey) + if intStringLenmapkey < 0 { + return ErrInvalidLengthGenerated + } + postStringIndexmapkey := iNdEx + intStringLenmapkey + if postStringIndexmapkey > l { return io.ErrUnexpectedEOF } - b := dAtA[iNdEx] - iNdEx++ - stringLenmapvalue |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break + mapkey = string(dAtA[iNdEx:postStringIndexmapkey]) + iNdEx = postStringIndexmapkey + } else if fieldNum == 2 { + var stringLenmapvalue uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLenmapvalue |= (uint64(b) & 0x7F) << shift + if b < 0x80 { + break + } } + intStringLenmapvalue := int(stringLenmapvalue) + if intStringLenmapvalue < 0 { + return ErrInvalidLengthGenerated + } + postStringIndexmapvalue := iNdEx + intStringLenmapvalue + if postStringIndexmapvalue > l { + return io.ErrUnexpectedEOF + } + mapvalue = string(dAtA[iNdEx:postStringIndexmapvalue]) + iNdEx = postStringIndexmapvalue + } else { + iNdEx = entryPreIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > postIndex { + return io.ErrUnexpectedEOF + } + iNdEx += skippy } - intStringLenmapvalue := int(stringLenmapvalue) - if intStringLenmapvalue < 0 { - return ErrInvalidLengthGenerated - } - postStringIndexmapvalue := iNdEx + intStringLenmapvalue - if postStringIndexmapvalue > l { - return io.ErrUnexpectedEOF - } - mapvalue := string(dAtA[iNdEx:postStringIndexmapvalue]) - iNdEx = postStringIndexmapvalue - m.Selector[mapkey] = mapvalue - } else { - var mapvalue string - m.Selector[mapkey] = mapvalue } + m.Selector[mapkey] = mapvalue iNdEx = postIndex case 8: if wireType != 2 { @@ -4793,51 +4778,14 @@ func (m *DeploymentStrategy) Unmarshal(dAtA []byte) error { if postIndex > l { return io.ErrUnexpectedEOF } - var keykey uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - keykey |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - var stringLenmapkey uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - stringLenmapkey |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - intStringLenmapkey := int(stringLenmapkey) - if intStringLenmapkey < 0 { - return ErrInvalidLengthGenerated - } - postStringIndexmapkey := iNdEx + intStringLenmapkey - if postStringIndexmapkey > l { - return io.ErrUnexpectedEOF - } - mapkey := string(dAtA[iNdEx:postStringIndexmapkey]) - iNdEx = postStringIndexmapkey if m.Labels == nil { m.Labels = make(map[string]string) } - if iNdEx < postIndex { - var valuekey uint64 + var mapkey string + var mapvalue string + for iNdEx < postIndex { + entryPreIndex := iNdEx + var wire uint64 for shift := uint(0); ; shift += 7 { if shift >= 64 { return ErrIntOverflowGenerated @@ -4847,41 +4795,80 @@ func (m *DeploymentStrategy) Unmarshal(dAtA []byte) error { } b := dAtA[iNdEx] iNdEx++ - valuekey |= (uint64(b) & 0x7F) << shift + wire |= (uint64(b) & 0x7F) << shift if b < 0x80 { break } } - var stringLenmapvalue uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated + fieldNum := int32(wire >> 3) + if fieldNum == 1 { + var stringLenmapkey uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLenmapkey |= (uint64(b) & 0x7F) << shift + if b < 0x80 { + break + } } - if iNdEx >= l { + intStringLenmapkey := int(stringLenmapkey) + if intStringLenmapkey < 0 { + return ErrInvalidLengthGenerated + } + postStringIndexmapkey := iNdEx + intStringLenmapkey + if postStringIndexmapkey > l { return io.ErrUnexpectedEOF } - b := dAtA[iNdEx] - iNdEx++ - stringLenmapvalue |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break + mapkey = string(dAtA[iNdEx:postStringIndexmapkey]) + iNdEx = postStringIndexmapkey + } else if fieldNum == 2 { + var stringLenmapvalue uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLenmapvalue |= (uint64(b) & 0x7F) << shift + if b < 0x80 { + break + } } + intStringLenmapvalue := int(stringLenmapvalue) + if intStringLenmapvalue < 0 { + return ErrInvalidLengthGenerated + } + postStringIndexmapvalue := iNdEx + intStringLenmapvalue + if postStringIndexmapvalue > l { + return io.ErrUnexpectedEOF + } + mapvalue = string(dAtA[iNdEx:postStringIndexmapvalue]) + iNdEx = postStringIndexmapvalue + } else { + iNdEx = entryPreIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > postIndex { + return io.ErrUnexpectedEOF + } + iNdEx += skippy } - intStringLenmapvalue := int(stringLenmapvalue) - if intStringLenmapvalue < 0 { - return ErrInvalidLengthGenerated - } - postStringIndexmapvalue := iNdEx + intStringLenmapvalue - if postStringIndexmapvalue > l { - return io.ErrUnexpectedEOF - } - mapvalue := string(dAtA[iNdEx:postStringIndexmapvalue]) - iNdEx = postStringIndexmapvalue - m.Labels[mapkey] = mapvalue - } else { - var mapvalue string - m.Labels[mapkey] = mapvalue } + m.Labels[mapkey] = mapvalue iNdEx = postIndex case 7: if wireType != 2 { @@ -4909,51 +4896,14 @@ func (m *DeploymentStrategy) Unmarshal(dAtA []byte) error { if postIndex > l { return io.ErrUnexpectedEOF } - var keykey uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - keykey |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - var stringLenmapkey uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - stringLenmapkey |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - intStringLenmapkey := int(stringLenmapkey) - if intStringLenmapkey < 0 { - return ErrInvalidLengthGenerated - } - postStringIndexmapkey := iNdEx + intStringLenmapkey - if postStringIndexmapkey > l { - return io.ErrUnexpectedEOF - } - mapkey := string(dAtA[iNdEx:postStringIndexmapkey]) - iNdEx = postStringIndexmapkey if m.Annotations == nil { m.Annotations = make(map[string]string) } - if iNdEx < postIndex { - var valuekey uint64 + var mapkey string + var mapvalue string + for iNdEx < postIndex { + entryPreIndex := iNdEx + var wire uint64 for shift := uint(0); ; shift += 7 { if shift >= 64 { return ErrIntOverflowGenerated @@ -4963,41 +4913,80 @@ func (m *DeploymentStrategy) Unmarshal(dAtA []byte) error { } b := dAtA[iNdEx] iNdEx++ - valuekey |= (uint64(b) & 0x7F) << shift + wire |= (uint64(b) & 0x7F) << shift if b < 0x80 { break } } - var stringLenmapvalue uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated + fieldNum := int32(wire >> 3) + if fieldNum == 1 { + var stringLenmapkey uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLenmapkey |= (uint64(b) & 0x7F) << shift + if b < 0x80 { + break + } } - if iNdEx >= l { + intStringLenmapkey := int(stringLenmapkey) + if intStringLenmapkey < 0 { + return ErrInvalidLengthGenerated + } + postStringIndexmapkey := iNdEx + intStringLenmapkey + if postStringIndexmapkey > l { return io.ErrUnexpectedEOF } - b := dAtA[iNdEx] - iNdEx++ - stringLenmapvalue |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break + mapkey = string(dAtA[iNdEx:postStringIndexmapkey]) + iNdEx = postStringIndexmapkey + } else if fieldNum == 2 { + var stringLenmapvalue uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLenmapvalue |= (uint64(b) & 0x7F) << shift + if b < 0x80 { + break + } } + intStringLenmapvalue := int(stringLenmapvalue) + if intStringLenmapvalue < 0 { + return ErrInvalidLengthGenerated + } + postStringIndexmapvalue := iNdEx + intStringLenmapvalue + if postStringIndexmapvalue > l { + return io.ErrUnexpectedEOF + } + mapvalue = string(dAtA[iNdEx:postStringIndexmapvalue]) + iNdEx = postStringIndexmapvalue + } else { + iNdEx = entryPreIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > postIndex { + return io.ErrUnexpectedEOF + } + iNdEx += skippy } - intStringLenmapvalue := int(stringLenmapvalue) - if intStringLenmapvalue < 0 { - return ErrInvalidLengthGenerated - } - postStringIndexmapvalue := iNdEx + intStringLenmapvalue - if postStringIndexmapvalue > l { - return io.ErrUnexpectedEOF - } - mapvalue := string(dAtA[iNdEx:postStringIndexmapvalue]) - iNdEx = postStringIndexmapvalue - m.Annotations[mapkey] = mapvalue - } else { - var mapvalue string - m.Annotations[mapkey] = mapvalue } + m.Annotations[mapkey] = mapvalue iNdEx = postIndex case 8: if wireType != 0 { diff --git a/vendor/github.com/openshift/api/apps/v1/zz_generated.deepcopy.go b/vendor/github.com/openshift/api/apps/v1/zz_generated.deepcopy.go index 6d76bea56a..f58fc08673 100644 --- a/vendor/github.com/openshift/api/apps/v1/zz_generated.deepcopy.go +++ b/vendor/github.com/openshift/api/apps/v1/zz_generated.deepcopy.go @@ -5,7 +5,7 @@ package v1 import ( - core_v1 "k8s.io/api/core/v1" + corev1 "k8s.io/api/core/v1" runtime "k8s.io/apimachinery/pkg/runtime" intstr "k8s.io/apimachinery/pkg/util/intstr" ) @@ -15,7 +15,7 @@ func (in *CustomDeploymentStrategyParams) DeepCopyInto(out *CustomDeploymentStra *out = *in if in.Environment != nil { in, out := &in.Environment, &out.Environment - *out = make([]core_v1.EnvVar, len(*in)) + *out = make([]corev1.EnvVar, len(*in)) for i := range *in { (*in)[i].DeepCopyInto(&(*out)[i]) } @@ -43,12 +43,8 @@ func (in *DeploymentCause) DeepCopyInto(out *DeploymentCause) { *out = *in if in.ImageTrigger != nil { in, out := &in.ImageTrigger, &out.ImageTrigger - if *in == nil { - *out = nil - } else { - *out = new(DeploymentCauseImageTrigger) - **out = **in - } + *out = new(DeploymentCauseImageTrigger) + **out = **in } return } @@ -222,12 +218,8 @@ func (in *DeploymentConfigSpec) DeepCopyInto(out *DeploymentConfigSpec) { } if in.RevisionHistoryLimit != nil { in, out := &in.RevisionHistoryLimit, &out.RevisionHistoryLimit - if *in == nil { - *out = nil - } else { - *out = new(int32) - **out = **in - } + *out = new(int32) + **out = **in } if in.Selector != nil { in, out := &in.Selector, &out.Selector @@ -238,12 +230,8 @@ func (in *DeploymentConfigSpec) DeepCopyInto(out *DeploymentConfigSpec) { } if in.Template != nil { in, out := &in.Template, &out.Template - if *in == nil { - *out = nil - } else { - *out = new(core_v1.PodTemplateSpec) - (*in).DeepCopyInto(*out) - } + *out = new(corev1.PodTemplateSpec) + (*in).DeepCopyInto(*out) } return } @@ -263,12 +251,8 @@ func (in *DeploymentConfigStatus) DeepCopyInto(out *DeploymentConfigStatus) { *out = *in if in.Details != nil { in, out := &in.Details, &out.Details - if *in == nil { - *out = nil - } else { - *out = new(DeploymentDetails) - (*in).DeepCopyInto(*out) - } + *out = new(DeploymentDetails) + (*in).DeepCopyInto(*out) } if in.Conditions != nil { in, out := &in.Conditions, &out.Conditions @@ -344,47 +328,27 @@ func (in *DeploymentLogOptions) DeepCopyInto(out *DeploymentLogOptions) { out.TypeMeta = in.TypeMeta if in.SinceSeconds != nil { in, out := &in.SinceSeconds, &out.SinceSeconds - if *in == nil { - *out = nil - } else { - *out = new(int64) - **out = **in - } + *out = new(int64) + **out = **in } if in.SinceTime != nil { in, out := &in.SinceTime, &out.SinceTime - if *in == nil { - *out = nil - } else { - *out = (*in).DeepCopy() - } + *out = (*in).DeepCopy() } if in.TailLines != nil { in, out := &in.TailLines, &out.TailLines - if *in == nil { - *out = nil - } else { - *out = new(int64) - **out = **in - } + *out = new(int64) + **out = **in } if in.LimitBytes != nil { in, out := &in.LimitBytes, &out.LimitBytes - if *in == nil { - *out = nil - } else { - *out = new(int64) - **out = **in - } + *out = new(int64) + **out = **in } if in.Version != nil { in, out := &in.Version, &out.Version - if *in == nil { - *out = nil - } else { - *out = new(int64) - **out = **in - } + *out = new(int64) + **out = **in } return } @@ -442,30 +406,18 @@ func (in *DeploymentStrategy) DeepCopyInto(out *DeploymentStrategy) { *out = *in if in.CustomParams != nil { in, out := &in.CustomParams, &out.CustomParams - if *in == nil { - *out = nil - } else { - *out = new(CustomDeploymentStrategyParams) - (*in).DeepCopyInto(*out) - } + *out = new(CustomDeploymentStrategyParams) + (*in).DeepCopyInto(*out) } if in.RecreateParams != nil { in, out := &in.RecreateParams, &out.RecreateParams - if *in == nil { - *out = nil - } else { - *out = new(RecreateDeploymentStrategyParams) - (*in).DeepCopyInto(*out) - } + *out = new(RecreateDeploymentStrategyParams) + (*in).DeepCopyInto(*out) } if in.RollingParams != nil { in, out := &in.RollingParams, &out.RollingParams - if *in == nil { - *out = nil - } else { - *out = new(RollingDeploymentStrategyParams) - (*in).DeepCopyInto(*out) - } + *out = new(RollingDeploymentStrategyParams) + (*in).DeepCopyInto(*out) } in.Resources.DeepCopyInto(&out.Resources) if in.Labels != nil { @@ -484,12 +436,8 @@ func (in *DeploymentStrategy) DeepCopyInto(out *DeploymentStrategy) { } if in.ActiveDeadlineSeconds != nil { in, out := &in.ActiveDeadlineSeconds, &out.ActiveDeadlineSeconds - if *in == nil { - *out = nil - } else { - *out = new(int64) - **out = **in - } + *out = new(int64) + **out = **in } return } @@ -553,12 +501,8 @@ func (in *DeploymentTriggerPolicy) DeepCopyInto(out *DeploymentTriggerPolicy) { *out = *in if in.ImageChangeParams != nil { in, out := &in.ImageChangeParams, &out.ImageChangeParams - if *in == nil { - *out = nil - } else { - *out = new(DeploymentTriggerImageChangeParams) - (*in).DeepCopyInto(*out) - } + *out = new(DeploymentTriggerImageChangeParams) + (*in).DeepCopyInto(*out) } return } @@ -583,7 +527,7 @@ func (in *ExecNewPodHook) DeepCopyInto(out *ExecNewPodHook) { } if in.Env != nil { in, out := &in.Env, &out.Env - *out = make([]core_v1.EnvVar, len(*in)) + *out = make([]corev1.EnvVar, len(*in)) for i := range *in { (*in)[i].DeepCopyInto(&(*out)[i]) } @@ -611,12 +555,8 @@ func (in *LifecycleHook) DeepCopyInto(out *LifecycleHook) { *out = *in if in.ExecNewPod != nil { in, out := &in.ExecNewPod, &out.ExecNewPod - if *in == nil { - *out = nil - } else { - *out = new(ExecNewPodHook) - (*in).DeepCopyInto(*out) - } + *out = new(ExecNewPodHook) + (*in).DeepCopyInto(*out) } if in.TagImages != nil { in, out := &in.TagImages, &out.TagImages @@ -641,39 +581,23 @@ func (in *RecreateDeploymentStrategyParams) DeepCopyInto(out *RecreateDeployment *out = *in if in.TimeoutSeconds != nil { in, out := &in.TimeoutSeconds, &out.TimeoutSeconds - if *in == nil { - *out = nil - } else { - *out = new(int64) - **out = **in - } + *out = new(int64) + **out = **in } if in.Pre != nil { in, out := &in.Pre, &out.Pre - if *in == nil { - *out = nil - } else { - *out = new(LifecycleHook) - (*in).DeepCopyInto(*out) - } + *out = new(LifecycleHook) + (*in).DeepCopyInto(*out) } if in.Mid != nil { in, out := &in.Mid, &out.Mid - if *in == nil { - *out = nil - } else { - *out = new(LifecycleHook) - (*in).DeepCopyInto(*out) - } + *out = new(LifecycleHook) + (*in).DeepCopyInto(*out) } if in.Post != nil { in, out := &in.Post, &out.Post - if *in == nil { - *out = nil - } else { - *out = new(LifecycleHook) - (*in).DeepCopyInto(*out) - } + *out = new(LifecycleHook) + (*in).DeepCopyInto(*out) } return } @@ -693,66 +617,38 @@ func (in *RollingDeploymentStrategyParams) DeepCopyInto(out *RollingDeploymentSt *out = *in if in.UpdatePeriodSeconds != nil { in, out := &in.UpdatePeriodSeconds, &out.UpdatePeriodSeconds - if *in == nil { - *out = nil - } else { - *out = new(int64) - **out = **in - } + *out = new(int64) + **out = **in } if in.IntervalSeconds != nil { in, out := &in.IntervalSeconds, &out.IntervalSeconds - if *in == nil { - *out = nil - } else { - *out = new(int64) - **out = **in - } + *out = new(int64) + **out = **in } if in.TimeoutSeconds != nil { in, out := &in.TimeoutSeconds, &out.TimeoutSeconds - if *in == nil { - *out = nil - } else { - *out = new(int64) - **out = **in - } + *out = new(int64) + **out = **in } if in.MaxUnavailable != nil { in, out := &in.MaxUnavailable, &out.MaxUnavailable - if *in == nil { - *out = nil - } else { - *out = new(intstr.IntOrString) - **out = **in - } + *out = new(intstr.IntOrString) + **out = **in } if in.MaxSurge != nil { in, out := &in.MaxSurge, &out.MaxSurge - if *in == nil { - *out = nil - } else { - *out = new(intstr.IntOrString) - **out = **in - } + *out = new(intstr.IntOrString) + **out = **in } if in.Pre != nil { in, out := &in.Pre, &out.Pre - if *in == nil { - *out = nil - } else { - *out = new(LifecycleHook) - (*in).DeepCopyInto(*out) - } + *out = new(LifecycleHook) + (*in).DeepCopyInto(*out) } if in.Post != nil { in, out := &in.Post, &out.Post - if *in == nil { - *out = nil - } else { - *out = new(LifecycleHook) - (*in).DeepCopyInto(*out) - } + *out = new(LifecycleHook) + (*in).DeepCopyInto(*out) } return } diff --git a/vendor/github.com/openshift/api/apps/v1/zz_generated.swagger_doc_generated.go b/vendor/github.com/openshift/api/apps/v1/zz_generated.swagger_doc_generated.go index a7870003aa..f8ec2bd1dd 100644 --- a/vendor/github.com/openshift/api/apps/v1/zz_generated.swagger_doc_generated.go +++ b/vendor/github.com/openshift/api/apps/v1/zz_generated.swagger_doc_generated.go @@ -257,7 +257,7 @@ func (RecreateDeploymentStrategyParams) SwaggerDoc() map[string]string { } var map_RollingDeploymentStrategyParams = map[string]string{ - "": "RollingDeploymentStrategyParams are the input to the Rolling deployment strategy.", + "": "RollingDeploymentStrategyParams are the input to the Rolling deployment strategy.", "updatePeriodSeconds": "UpdatePeriodSeconds is the time to wait between individual pod updates. If the value is nil, a default will be used.", "intervalSeconds": "IntervalSeconds is the time to wait between polling deployment status after update. If the value is nil, a default will be used.", "timeoutSeconds": "TimeoutSeconds is the time to wait for updates before giving up. If the value is nil, a default will be used.", diff --git a/vendor/github.com/openshift/api/authorization/v1/generated.pb.go b/vendor/github.com/openshift/api/authorization/v1/generated.pb.go index 24f7fa52ee..c8e008826a 100644 --- a/vendor/github.com/openshift/api/authorization/v1/generated.pb.go +++ b/vendor/github.com/openshift/api/authorization/v1/generated.pb.go @@ -1,6 +1,5 @@ -// Code generated by protoc-gen-gogo. +// Code generated by protoc-gen-gogo. DO NOT EDIT. // source: github.com/openshift/api/authorization/v1/generated.proto -// DO NOT EDIT! /* Package v1 is a generated protocol buffer package. @@ -1738,24 +1737,6 @@ func (m *UserRestriction) MarshalTo(dAtA []byte) (int, error) { return i, nil } -func encodeFixed64Generated(dAtA []byte, offset int, v uint64) int { - dAtA[offset] = uint8(v) - dAtA[offset+1] = uint8(v >> 8) - dAtA[offset+2] = uint8(v >> 16) - dAtA[offset+3] = uint8(v >> 24) - dAtA[offset+4] = uint8(v >> 32) - dAtA[offset+5] = uint8(v >> 40) - dAtA[offset+6] = uint8(v >> 48) - dAtA[offset+7] = uint8(v >> 56) - return offset + 8 -} -func encodeFixed32Generated(dAtA []byte, offset int, v uint32) int { - dAtA[offset] = uint8(v) - dAtA[offset+1] = uint8(v >> 8) - dAtA[offset+2] = uint8(v >> 16) - dAtA[offset+3] = uint8(v >> 24) - return offset + 4 -} func encodeVarintGenerated(dAtA []byte, offset int, v uint64) int { for v >= 1<<7 { dAtA[offset] = uint8(v&0x7f | 0x80) diff --git a/vendor/github.com/openshift/api/authorization/v1/types.go b/vendor/github.com/openshift/api/authorization/v1/types.go index df165f6cdc..3e619df0a7 100644 --- a/vendor/github.com/openshift/api/authorization/v1/types.go +++ b/vendor/github.com/openshift/api/authorization/v1/types.go @@ -460,12 +460,15 @@ type RoleBindingRestriction struct { // field must be non-nil. type RoleBindingRestrictionSpec struct { // UserRestriction matches against user subjects. + // +nullable UserRestriction *UserRestriction `json:"userrestriction" protobuf:"bytes,1,opt,name=userrestriction"` // GroupRestriction matches against group subjects. + // +nullable GroupRestriction *GroupRestriction `json:"grouprestriction" protobuf:"bytes,2,opt,name=grouprestriction"` // ServiceAccountRestriction matches against service-account subjects. + // +nullable ServiceAccountRestriction *ServiceAccountRestriction `json:"serviceaccountrestriction" protobuf:"bytes,3,opt,name=serviceaccountrestriction"` } diff --git a/vendor/github.com/openshift/api/authorization/v1/zz_generated.deepcopy.go b/vendor/github.com/openshift/api/authorization/v1/zz_generated.deepcopy.go index e09b9ab6fa..930d8cf3db 100644 --- a/vendor/github.com/openshift/api/authorization/v1/zz_generated.deepcopy.go +++ b/vendor/github.com/openshift/api/authorization/v1/zz_generated.deepcopy.go @@ -5,9 +5,9 @@ package v1 import ( - core_v1 "k8s.io/api/core/v1" - rbac_v1 "k8s.io/api/rbac/v1" - meta_v1 "k8s.io/apimachinery/pkg/apis/meta/v1" + corev1 "k8s.io/api/core/v1" + rbacv1 "k8s.io/api/rbac/v1" + metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" runtime "k8s.io/apimachinery/pkg/runtime" ) @@ -42,12 +42,8 @@ func (in *ClusterRole) DeepCopyInto(out *ClusterRole) { } if in.AggregationRule != nil { in, out := &in.AggregationRule, &out.AggregationRule - if *in == nil { - *out = nil - } else { - *out = new(rbac_v1.AggregationRule) - (*in).DeepCopyInto(*out) - } + *out = new(rbacv1.AggregationRule) + (*in).DeepCopyInto(*out) } return } @@ -87,7 +83,7 @@ func (in *ClusterRoleBinding) DeepCopyInto(out *ClusterRoleBinding) { } if in.Subjects != nil { in, out := &in.Subjects, &out.Subjects - *out = make([]core_v1.ObjectReference, len(*in)) + *out = make([]corev1.ObjectReference, len(*in)) copy(*out, *in) } out.RoleRef = in.RoleRef @@ -188,7 +184,7 @@ func (in *GroupRestriction) DeepCopyInto(out *GroupRestriction) { } if in.Selectors != nil { in, out := &in.Selectors, &out.Selectors - *out = make([]meta_v1.LabelSelector, len(*in)) + *out = make([]metav1.LabelSelector, len(*in)) for i := range *in { (*in)[i].DeepCopyInto(&(*out)[i]) } @@ -554,7 +550,7 @@ func (in *RoleBinding) DeepCopyInto(out *RoleBinding) { } if in.Subjects != nil { in, out := &in.Subjects, &out.Subjects - *out = make([]core_v1.ObjectReference, len(*in)) + *out = make([]corev1.ObjectReference, len(*in)) copy(*out, *in) } out.RoleRef = in.RoleRef @@ -677,30 +673,18 @@ func (in *RoleBindingRestrictionSpec) DeepCopyInto(out *RoleBindingRestrictionSp *out = *in if in.UserRestriction != nil { in, out := &in.UserRestriction, &out.UserRestriction - if *in == nil { - *out = nil - } else { - *out = new(UserRestriction) - (*in).DeepCopyInto(*out) - } + *out = new(UserRestriction) + (*in).DeepCopyInto(*out) } if in.GroupRestriction != nil { in, out := &in.GroupRestriction, &out.GroupRestriction - if *in == nil { - *out = nil - } else { - *out = new(GroupRestriction) - (*in).DeepCopyInto(*out) - } + *out = new(GroupRestriction) + (*in).DeepCopyInto(*out) } if in.ServiceAccountRestriction != nil { in, out := &in.ServiceAccountRestriction, &out.ServiceAccountRestriction - if *in == nil { - *out = nil - } else { - *out = new(ServiceAccountRestriction) - (*in).DeepCopyInto(*out) - } + *out = new(ServiceAccountRestriction) + (*in).DeepCopyInto(*out) } return } @@ -990,7 +974,7 @@ func (in *UserRestriction) DeepCopyInto(out *UserRestriction) { } if in.Selectors != nil { in, out := &in.Selectors, &out.Selectors - *out = make([]meta_v1.LabelSelector, len(*in)) + *out = make([]metav1.LabelSelector, len(*in)) for i := range *in { (*in)[i].DeepCopyInto(&(*out)[i]) } diff --git a/vendor/github.com/openshift/api/build/v1/consts.go b/vendor/github.com/openshift/api/build/v1/consts.go index 685cebc123..96917232d4 100644 --- a/vendor/github.com/openshift/api/build/v1/consts.go +++ b/vendor/github.com/openshift/api/build/v1/consts.go @@ -87,7 +87,11 @@ const ( // StatusReasonOutOfMemoryKilled indicates that the build pod was killed for its memory consumption StatusReasonOutOfMemoryKilled StatusReason = "OutOfMemoryKilled" - // StatusCannotRetrieveServiceAccount is the reason associated with a failure + // StatusReasonCannotRetrieveServiceAccount is the reason associated with a failure // to look up the service account associated with the BuildConfig. StatusReasonCannotRetrieveServiceAccount StatusReason = "CannotRetrieveServiceAccount" + + // StatusReasonBuildPodEvicted is the reason a build fails due to the build pod being evicted + // from its node + StatusReasonBuildPodEvicted StatusReason = "BuildPodEvicted" ) diff --git a/vendor/github.com/openshift/api/build/v1/generated.pb.go b/vendor/github.com/openshift/api/build/v1/generated.pb.go index 76cc3daaba..bcec53e134 100644 --- a/vendor/github.com/openshift/api/build/v1/generated.pb.go +++ b/vendor/github.com/openshift/api/build/v1/generated.pb.go @@ -1,6 +1,5 @@ -// Code generated by protoc-gen-gogo. +// Code generated by protoc-gen-gogo. DO NOT EDIT. // source: github.com/openshift/api/build/v1/generated.proto -// DO NOT EDIT! /* Package v1 is a generated protocol buffer package. @@ -2829,24 +2828,6 @@ func (m *WebHookTrigger) MarshalTo(dAtA []byte) (int, error) { return i, nil } -func encodeFixed64Generated(dAtA []byte, offset int, v uint64) int { - dAtA[offset] = uint8(v) - dAtA[offset+1] = uint8(v >> 8) - dAtA[offset+2] = uint8(v >> 16) - dAtA[offset+3] = uint8(v >> 24) - dAtA[offset+4] = uint8(v >> 32) - dAtA[offset+5] = uint8(v >> 40) - dAtA[offset+6] = uint8(v >> 48) - dAtA[offset+7] = uint8(v >> 56) - return offset + 8 -} -func encodeFixed32Generated(dAtA []byte, offset int, v uint32) int { - dAtA[offset] = uint8(v) - dAtA[offset+1] = uint8(v >> 8) - dAtA[offset+2] = uint8(v >> 16) - dAtA[offset+3] = uint8(v >> 24) - return offset + 4 -} func encodeVarintGenerated(dAtA []byte, offset int, v uint64) int { for v >= 1<<7 { dAtA[offset] = uint8(v&0x7f | 0x80) @@ -11331,51 +11312,14 @@ func (m *OptionalNodeSelector) Unmarshal(dAtA []byte) error { if postIndex > l { return io.ErrUnexpectedEOF } - var keykey uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - keykey |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - var stringLenmapkey uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - stringLenmapkey |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - intStringLenmapkey := int(stringLenmapkey) - if intStringLenmapkey < 0 { - return ErrInvalidLengthGenerated - } - postStringIndexmapkey := iNdEx + intStringLenmapkey - if postStringIndexmapkey > l { - return io.ErrUnexpectedEOF - } - mapkey := string(dAtA[iNdEx:postStringIndexmapkey]) - iNdEx = postStringIndexmapkey if *m == nil { *m = make(map[string]string) } - if iNdEx < postIndex { - var valuekey uint64 + var mapkey string + var mapvalue string + for iNdEx < postIndex { + entryPreIndex := iNdEx + var wire uint64 for shift := uint(0); ; shift += 7 { if shift >= 64 { return ErrIntOverflowGenerated @@ -11385,41 +11329,80 @@ func (m *OptionalNodeSelector) Unmarshal(dAtA []byte) error { } b := dAtA[iNdEx] iNdEx++ - valuekey |= (uint64(b) & 0x7F) << shift + wire |= (uint64(b) & 0x7F) << shift if b < 0x80 { break } } - var stringLenmapvalue uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated + fieldNum := int32(wire >> 3) + if fieldNum == 1 { + var stringLenmapkey uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLenmapkey |= (uint64(b) & 0x7F) << shift + if b < 0x80 { + break + } } - if iNdEx >= l { + intStringLenmapkey := int(stringLenmapkey) + if intStringLenmapkey < 0 { + return ErrInvalidLengthGenerated + } + postStringIndexmapkey := iNdEx + intStringLenmapkey + if postStringIndexmapkey > l { return io.ErrUnexpectedEOF } - b := dAtA[iNdEx] - iNdEx++ - stringLenmapvalue |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break + mapkey = string(dAtA[iNdEx:postStringIndexmapkey]) + iNdEx = postStringIndexmapkey + } else if fieldNum == 2 { + var stringLenmapvalue uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLenmapvalue |= (uint64(b) & 0x7F) << shift + if b < 0x80 { + break + } } + intStringLenmapvalue := int(stringLenmapvalue) + if intStringLenmapvalue < 0 { + return ErrInvalidLengthGenerated + } + postStringIndexmapvalue := iNdEx + intStringLenmapvalue + if postStringIndexmapvalue > l { + return io.ErrUnexpectedEOF + } + mapvalue = string(dAtA[iNdEx:postStringIndexmapvalue]) + iNdEx = postStringIndexmapvalue + } else { + iNdEx = entryPreIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > postIndex { + return io.ErrUnexpectedEOF + } + iNdEx += skippy } - intStringLenmapvalue := int(stringLenmapvalue) - if intStringLenmapvalue < 0 { - return ErrInvalidLengthGenerated - } - postStringIndexmapvalue := iNdEx + intStringLenmapvalue - if postStringIndexmapvalue > l { - return io.ErrUnexpectedEOF - } - mapvalue := string(dAtA[iNdEx:postStringIndexmapvalue]) - iNdEx = postStringIndexmapvalue - (*m)[mapkey] = mapvalue - } else { - var mapvalue string - (*m)[mapkey] = mapvalue } + (*m)[mapkey] = mapvalue iNdEx = postIndex default: iNdEx = preIndex diff --git a/vendor/github.com/openshift/api/build/v1/zz_generated.deepcopy.go b/vendor/github.com/openshift/api/build/v1/zz_generated.deepcopy.go index 68f987dc9e..96a895606d 100644 --- a/vendor/github.com/openshift/api/build/v1/zz_generated.deepcopy.go +++ b/vendor/github.com/openshift/api/build/v1/zz_generated.deepcopy.go @@ -5,7 +5,7 @@ package v1 import ( - core_v1 "k8s.io/api/core/v1" + corev1 "k8s.io/api/core/v1" runtime "k8s.io/apimachinery/pkg/runtime" ) @@ -170,21 +170,13 @@ func (in *BuildConfigSpec) DeepCopyInto(out *BuildConfigSpec) { in.CommonSpec.DeepCopyInto(&out.CommonSpec) if in.SuccessfulBuildsHistoryLimit != nil { in, out := &in.SuccessfulBuildsHistoryLimit, &out.SuccessfulBuildsHistoryLimit - if *in == nil { - *out = nil - } else { - *out = new(int32) - **out = **in - } + *out = new(int32) + **out = **in } if in.FailedBuildsHistoryLimit != nil { in, out := &in.FailedBuildsHistoryLimit, &out.FailedBuildsHistoryLimit - if *in == nil { - *out = nil - } else { - *out = new(int32) - **out = **in - } + *out = new(int32) + **out = **in } return } @@ -279,47 +271,27 @@ func (in *BuildLogOptions) DeepCopyInto(out *BuildLogOptions) { out.TypeMeta = in.TypeMeta if in.SinceSeconds != nil { in, out := &in.SinceSeconds, &out.SinceSeconds - if *in == nil { - *out = nil - } else { - *out = new(int64) - **out = **in - } + *out = new(int64) + **out = **in } if in.SinceTime != nil { in, out := &in.SinceTime, &out.SinceTime - if *in == nil { - *out = nil - } else { - *out = (*in).DeepCopy() - } + *out = (*in).DeepCopy() } if in.TailLines != nil { in, out := &in.TailLines, &out.TailLines - if *in == nil { - *out = nil - } else { - *out = new(int64) - **out = **in - } + *out = new(int64) + **out = **in } if in.LimitBytes != nil { in, out := &in.LimitBytes, &out.LimitBytes - if *in == nil { - *out = nil - } else { - *out = new(int64) - **out = **in - } + *out = new(int64) + **out = **in } if in.Version != nil { in, out := &in.Version, &out.Version - if *in == nil { - *out = nil - } else { - *out = new(int64) - **out = **in - } + *out = new(int64) + **out = **in } return } @@ -347,21 +319,13 @@ func (in *BuildOutput) DeepCopyInto(out *BuildOutput) { *out = *in if in.To != nil { in, out := &in.To, &out.To - if *in == nil { - *out = nil - } else { - *out = new(core_v1.ObjectReference) - **out = **in - } + *out = new(corev1.ObjectReference) + **out = **in } if in.PushSecret != nil { in, out := &in.PushSecret, &out.PushSecret - if *in == nil { - *out = nil - } else { - *out = new(core_v1.LocalObjectReference) - **out = **in - } + *out = new(corev1.LocalObjectReference) + **out = **in } if in.ImageLabels != nil { in, out := &in.ImageLabels, &out.ImageLabels @@ -414,52 +378,32 @@ func (in *BuildRequest) DeepCopyInto(out *BuildRequest) { in.ObjectMeta.DeepCopyInto(&out.ObjectMeta) if in.Revision != nil { in, out := &in.Revision, &out.Revision - if *in == nil { - *out = nil - } else { - *out = new(SourceRevision) - (*in).DeepCopyInto(*out) - } + *out = new(SourceRevision) + (*in).DeepCopyInto(*out) } if in.TriggeredByImage != nil { in, out := &in.TriggeredByImage, &out.TriggeredByImage - if *in == nil { - *out = nil - } else { - *out = new(core_v1.ObjectReference) - **out = **in - } + *out = new(corev1.ObjectReference) + **out = **in } if in.From != nil { in, out := &in.From, &out.From - if *in == nil { - *out = nil - } else { - *out = new(core_v1.ObjectReference) - **out = **in - } + *out = new(corev1.ObjectReference) + **out = **in } if in.Binary != nil { in, out := &in.Binary, &out.Binary - if *in == nil { - *out = nil - } else { - *out = new(BinaryBuildSource) - **out = **in - } + *out = new(BinaryBuildSource) + **out = **in } if in.LastVersion != nil { in, out := &in.LastVersion, &out.LastVersion - if *in == nil { - *out = nil - } else { - *out = new(int64) - **out = **in - } + *out = new(int64) + **out = **in } if in.Env != nil { in, out := &in.Env, &out.Env - *out = make([]core_v1.EnvVar, len(*in)) + *out = make([]corev1.EnvVar, len(*in)) for i := range *in { (*in)[i].DeepCopyInto(&(*out)[i]) } @@ -473,21 +417,13 @@ func (in *BuildRequest) DeepCopyInto(out *BuildRequest) { } if in.DockerStrategyOptions != nil { in, out := &in.DockerStrategyOptions, &out.DockerStrategyOptions - if *in == nil { - *out = nil - } else { - *out = new(DockerStrategyOptions) - (*in).DeepCopyInto(*out) - } + *out = new(DockerStrategyOptions) + (*in).DeepCopyInto(*out) } if in.SourceStrategyOptions != nil { in, out := &in.SourceStrategyOptions, &out.SourceStrategyOptions - if *in == nil { - *out = nil - } else { - *out = new(SourceStrategyOptions) - (*in).DeepCopyInto(*out) - } + *out = new(SourceStrategyOptions) + (*in).DeepCopyInto(*out) } return } @@ -515,30 +451,18 @@ func (in *BuildSource) DeepCopyInto(out *BuildSource) { *out = *in if in.Binary != nil { in, out := &in.Binary, &out.Binary - if *in == nil { - *out = nil - } else { - *out = new(BinaryBuildSource) - **out = **in - } + *out = new(BinaryBuildSource) + **out = **in } if in.Dockerfile != nil { in, out := &in.Dockerfile, &out.Dockerfile - if *in == nil { - *out = nil - } else { - *out = new(string) - **out = **in - } + *out = new(string) + **out = **in } if in.Git != nil { in, out := &in.Git, &out.Git - if *in == nil { - *out = nil - } else { - *out = new(GitBuildSource) - (*in).DeepCopyInto(*out) - } + *out = new(GitBuildSource) + (*in).DeepCopyInto(*out) } if in.Images != nil { in, out := &in.Images, &out.Images @@ -549,12 +473,8 @@ func (in *BuildSource) DeepCopyInto(out *BuildSource) { } if in.SourceSecret != nil { in, out := &in.SourceSecret, &out.SourceSecret - if *in == nil { - *out = nil - } else { - *out = new(core_v1.LocalObjectReference) - **out = **in - } + *out = new(corev1.LocalObjectReference) + **out = **in } if in.Secrets != nil { in, out := &in.Secrets, &out.Secrets @@ -608,28 +528,16 @@ func (in *BuildStatus) DeepCopyInto(out *BuildStatus) { *out = *in if in.StartTimestamp != nil { in, out := &in.StartTimestamp, &out.StartTimestamp - if *in == nil { - *out = nil - } else { - *out = (*in).DeepCopy() - } + *out = (*in).DeepCopy() } if in.CompletionTimestamp != nil { in, out := &in.CompletionTimestamp, &out.CompletionTimestamp - if *in == nil { - *out = nil - } else { - *out = (*in).DeepCopy() - } + *out = (*in).DeepCopy() } if in.Config != nil { in, out := &in.Config, &out.Config - if *in == nil { - *out = nil - } else { - *out = new(core_v1.ObjectReference) - **out = **in - } + *out = new(corev1.ObjectReference) + **out = **in } in.Output.DeepCopyInto(&out.Output) if in.Stages != nil { @@ -657,12 +565,8 @@ func (in *BuildStatusOutput) DeepCopyInto(out *BuildStatusOutput) { *out = *in if in.To != nil { in, out := &in.To, &out.To - if *in == nil { - *out = nil - } else { - *out = new(BuildStatusOutputTo) - **out = **in - } + *out = new(BuildStatusOutputTo) + **out = **in } return } @@ -698,39 +602,23 @@ func (in *BuildStrategy) DeepCopyInto(out *BuildStrategy) { *out = *in if in.DockerStrategy != nil { in, out := &in.DockerStrategy, &out.DockerStrategy - if *in == nil { - *out = nil - } else { - *out = new(DockerBuildStrategy) - (*in).DeepCopyInto(*out) - } + *out = new(DockerBuildStrategy) + (*in).DeepCopyInto(*out) } if in.SourceStrategy != nil { in, out := &in.SourceStrategy, &out.SourceStrategy - if *in == nil { - *out = nil - } else { - *out = new(SourceBuildStrategy) - (*in).DeepCopyInto(*out) - } + *out = new(SourceBuildStrategy) + (*in).DeepCopyInto(*out) } if in.CustomStrategy != nil { in, out := &in.CustomStrategy, &out.CustomStrategy - if *in == nil { - *out = nil - } else { - *out = new(CustomBuildStrategy) - (*in).DeepCopyInto(*out) - } + *out = new(CustomBuildStrategy) + (*in).DeepCopyInto(*out) } if in.JenkinsPipelineStrategy != nil { in, out := &in.JenkinsPipelineStrategy, &out.JenkinsPipelineStrategy - if *in == nil { - *out = nil - } else { - *out = new(JenkinsPipelineBuildStrategy) - (*in).DeepCopyInto(*out) - } + *out = new(JenkinsPipelineBuildStrategy) + (*in).DeepCopyInto(*out) } return } @@ -750,48 +638,28 @@ func (in *BuildTriggerCause) DeepCopyInto(out *BuildTriggerCause) { *out = *in if in.GenericWebHook != nil { in, out := &in.GenericWebHook, &out.GenericWebHook - if *in == nil { - *out = nil - } else { - *out = new(GenericWebHookCause) - (*in).DeepCopyInto(*out) - } + *out = new(GenericWebHookCause) + (*in).DeepCopyInto(*out) } if in.GitHubWebHook != nil { in, out := &in.GitHubWebHook, &out.GitHubWebHook - if *in == nil { - *out = nil - } else { - *out = new(GitHubWebHookCause) - (*in).DeepCopyInto(*out) - } + *out = new(GitHubWebHookCause) + (*in).DeepCopyInto(*out) } if in.ImageChangeBuild != nil { in, out := &in.ImageChangeBuild, &out.ImageChangeBuild - if *in == nil { - *out = nil - } else { - *out = new(ImageChangeCause) - (*in).DeepCopyInto(*out) - } + *out = new(ImageChangeCause) + (*in).DeepCopyInto(*out) } if in.GitLabWebHook != nil { in, out := &in.GitLabWebHook, &out.GitLabWebHook - if *in == nil { - *out = nil - } else { - *out = new(GitLabWebHookCause) - (*in).DeepCopyInto(*out) - } + *out = new(GitLabWebHookCause) + (*in).DeepCopyInto(*out) } if in.BitbucketWebHook != nil { in, out := &in.BitbucketWebHook, &out.BitbucketWebHook - if *in == nil { - *out = nil - } else { - *out = new(BitbucketWebHookCause) - (*in).DeepCopyInto(*out) - } + *out = new(BitbucketWebHookCause) + (*in).DeepCopyInto(*out) } return } @@ -811,48 +679,28 @@ func (in *BuildTriggerPolicy) DeepCopyInto(out *BuildTriggerPolicy) { *out = *in if in.GitHubWebHook != nil { in, out := &in.GitHubWebHook, &out.GitHubWebHook - if *in == nil { - *out = nil - } else { - *out = new(WebHookTrigger) - (*in).DeepCopyInto(*out) - } + *out = new(WebHookTrigger) + (*in).DeepCopyInto(*out) } if in.GenericWebHook != nil { in, out := &in.GenericWebHook, &out.GenericWebHook - if *in == nil { - *out = nil - } else { - *out = new(WebHookTrigger) - (*in).DeepCopyInto(*out) - } + *out = new(WebHookTrigger) + (*in).DeepCopyInto(*out) } if in.ImageChange != nil { in, out := &in.ImageChange, &out.ImageChange - if *in == nil { - *out = nil - } else { - *out = new(ImageChangeTrigger) - (*in).DeepCopyInto(*out) - } + *out = new(ImageChangeTrigger) + (*in).DeepCopyInto(*out) } if in.GitLabWebHook != nil { in, out := &in.GitLabWebHook, &out.GitLabWebHook - if *in == nil { - *out = nil - } else { - *out = new(WebHookTrigger) - (*in).DeepCopyInto(*out) - } + *out = new(WebHookTrigger) + (*in).DeepCopyInto(*out) } if in.BitbucketWebHook != nil { in, out := &in.BitbucketWebHook, &out.BitbucketWebHook - if *in == nil { - *out = nil - } else { - *out = new(WebHookTrigger) - (*in).DeepCopyInto(*out) - } + *out = new(WebHookTrigger) + (*in).DeepCopyInto(*out) } return } @@ -873,12 +721,8 @@ func (in *CommonSpec) DeepCopyInto(out *CommonSpec) { in.Source.DeepCopyInto(&out.Source) if in.Revision != nil { in, out := &in.Revision, &out.Revision - if *in == nil { - *out = nil - } else { - *out = new(SourceRevision) - (*in).DeepCopyInto(*out) - } + *out = new(SourceRevision) + (*in).DeepCopyInto(*out) } in.Strategy.DeepCopyInto(&out.Strategy) in.Output.DeepCopyInto(&out.Output) @@ -886,12 +730,8 @@ func (in *CommonSpec) DeepCopyInto(out *CommonSpec) { in.PostCommit.DeepCopyInto(&out.PostCommit) if in.CompletionDeadlineSeconds != nil { in, out := &in.CompletionDeadlineSeconds, &out.CompletionDeadlineSeconds - if *in == nil { - *out = nil - } else { - *out = new(int64) - **out = **in - } + *out = new(int64) + **out = **in } if in.NodeSelector != nil { in, out := &in.NodeSelector, &out.NodeSelector @@ -918,12 +758,8 @@ func (in *CommonWebHookCause) DeepCopyInto(out *CommonWebHookCause) { *out = *in if in.Revision != nil { in, out := &in.Revision, &out.Revision - if *in == nil { - *out = nil - } else { - *out = new(SourceRevision) - (*in).DeepCopyInto(*out) - } + *out = new(SourceRevision) + (*in).DeepCopyInto(*out) } return } @@ -961,16 +797,12 @@ func (in *CustomBuildStrategy) DeepCopyInto(out *CustomBuildStrategy) { out.From = in.From if in.PullSecret != nil { in, out := &in.PullSecret, &out.PullSecret - if *in == nil { - *out = nil - } else { - *out = new(core_v1.LocalObjectReference) - **out = **in - } + *out = new(corev1.LocalObjectReference) + **out = **in } if in.Env != nil { in, out := &in.Env, &out.Env - *out = make([]core_v1.EnvVar, len(*in)) + *out = make([]corev1.EnvVar, len(*in)) for i := range *in { (*in)[i].DeepCopyInto(&(*out)[i]) } @@ -998,44 +830,32 @@ func (in *DockerBuildStrategy) DeepCopyInto(out *DockerBuildStrategy) { *out = *in if in.From != nil { in, out := &in.From, &out.From - if *in == nil { - *out = nil - } else { - *out = new(core_v1.ObjectReference) - **out = **in - } + *out = new(corev1.ObjectReference) + **out = **in } if in.PullSecret != nil { in, out := &in.PullSecret, &out.PullSecret - if *in == nil { - *out = nil - } else { - *out = new(core_v1.LocalObjectReference) - **out = **in - } + *out = new(corev1.LocalObjectReference) + **out = **in } if in.Env != nil { in, out := &in.Env, &out.Env - *out = make([]core_v1.EnvVar, len(*in)) + *out = make([]corev1.EnvVar, len(*in)) for i := range *in { (*in)[i].DeepCopyInto(&(*out)[i]) } } if in.BuildArgs != nil { in, out := &in.BuildArgs, &out.BuildArgs - *out = make([]core_v1.EnvVar, len(*in)) + *out = make([]corev1.EnvVar, len(*in)) for i := range *in { (*in)[i].DeepCopyInto(&(*out)[i]) } } if in.ImageOptimizationPolicy != nil { in, out := &in.ImageOptimizationPolicy, &out.ImageOptimizationPolicy - if *in == nil { - *out = nil - } else { - *out = new(ImageOptimizationPolicy) - **out = **in - } + *out = new(ImageOptimizationPolicy) + **out = **in } return } @@ -1055,19 +875,15 @@ func (in *DockerStrategyOptions) DeepCopyInto(out *DockerStrategyOptions) { *out = *in if in.BuildArgs != nil { in, out := &in.BuildArgs, &out.BuildArgs - *out = make([]core_v1.EnvVar, len(*in)) + *out = make([]corev1.EnvVar, len(*in)) for i := range *in { (*in)[i].DeepCopyInto(&(*out)[i]) } } if in.NoCache != nil { in, out := &in.NoCache, &out.NoCache - if *in == nil { - *out = nil - } else { - *out = new(bool) - **out = **in - } + *out = new(bool) + **out = **in } return } @@ -1087,12 +903,8 @@ func (in *GenericWebHookCause) DeepCopyInto(out *GenericWebHookCause) { *out = *in if in.Revision != nil { in, out := &in.Revision, &out.Revision - if *in == nil { - *out = nil - } else { - *out = new(SourceRevision) - (*in).DeepCopyInto(*out) - } + *out = new(SourceRevision) + (*in).DeepCopyInto(*out) } return } @@ -1112,28 +924,20 @@ func (in *GenericWebHookEvent) DeepCopyInto(out *GenericWebHookEvent) { *out = *in if in.Git != nil { in, out := &in.Git, &out.Git - if *in == nil { - *out = nil - } else { - *out = new(GitInfo) - (*in).DeepCopyInto(*out) - } + *out = new(GitInfo) + (*in).DeepCopyInto(*out) } if in.Env != nil { in, out := &in.Env, &out.Env - *out = make([]core_v1.EnvVar, len(*in)) + *out = make([]corev1.EnvVar, len(*in)) for i := range *in { (*in)[i].DeepCopyInto(&(*out)[i]) } } if in.DockerStrategyOptions != nil { in, out := &in.DockerStrategyOptions, &out.DockerStrategyOptions - if *in == nil { - *out = nil - } else { - *out = new(DockerStrategyOptions) - (*in).DeepCopyInto(*out) - } + *out = new(DockerStrategyOptions) + (*in).DeepCopyInto(*out) } return } @@ -1170,12 +974,8 @@ func (in *GitHubWebHookCause) DeepCopyInto(out *GitHubWebHookCause) { *out = *in if in.Revision != nil { in, out := &in.Revision, &out.Revision - if *in == nil { - *out = nil - } else { - *out = new(SourceRevision) - (*in).DeepCopyInto(*out) - } + *out = new(SourceRevision) + (*in).DeepCopyInto(*out) } return } @@ -1273,12 +1073,8 @@ func (in *ImageChangeCause) DeepCopyInto(out *ImageChangeCause) { *out = *in if in.FromRef != nil { in, out := &in.FromRef, &out.FromRef - if *in == nil { - *out = nil - } else { - *out = new(core_v1.ObjectReference) - **out = **in - } + *out = new(corev1.ObjectReference) + **out = **in } return } @@ -1298,12 +1094,8 @@ func (in *ImageChangeTrigger) DeepCopyInto(out *ImageChangeTrigger) { *out = *in if in.From != nil { in, out := &in.From, &out.From - if *in == nil { - *out = nil - } else { - *out = new(core_v1.ObjectReference) - **out = **in - } + *out = new(corev1.ObjectReference) + **out = **in } return } @@ -1350,12 +1142,8 @@ func (in *ImageSource) DeepCopyInto(out *ImageSource) { } if in.PullSecret != nil { in, out := &in.PullSecret, &out.PullSecret - if *in == nil { - *out = nil - } else { - *out = new(core_v1.LocalObjectReference) - **out = **in - } + *out = new(corev1.LocalObjectReference) + **out = **in } return } @@ -1391,7 +1179,7 @@ func (in *JenkinsPipelineBuildStrategy) DeepCopyInto(out *JenkinsPipelineBuildSt *out = *in if in.Env != nil { in, out := &in.Env, &out.Env - *out = make([]core_v1.EnvVar, len(*in)) + *out = make([]corev1.EnvVar, len(*in)) for i := range *in { (*in)[i].DeepCopyInto(&(*out)[i]) } @@ -1436,30 +1224,18 @@ func (in *ProxyConfig) DeepCopyInto(out *ProxyConfig) { *out = *in if in.HTTPProxy != nil { in, out := &in.HTTPProxy, &out.HTTPProxy - if *in == nil { - *out = nil - } else { - *out = new(string) - **out = **in - } + *out = new(string) + **out = **in } if in.HTTPSProxy != nil { in, out := &in.HTTPSProxy, &out.HTTPSProxy - if *in == nil { - *out = nil - } else { - *out = new(string) - **out = **in - } + *out = new(string) + **out = **in } if in.NoProxy != nil { in, out := &in.NoProxy, &out.NoProxy - if *in == nil { - *out = nil - } else { - *out = new(string) - **out = **in - } + *out = new(string) + **out = **in } return } @@ -1530,28 +1306,20 @@ func (in *SourceBuildStrategy) DeepCopyInto(out *SourceBuildStrategy) { out.From = in.From if in.PullSecret != nil { in, out := &in.PullSecret, &out.PullSecret - if *in == nil { - *out = nil - } else { - *out = new(core_v1.LocalObjectReference) - **out = **in - } + *out = new(corev1.LocalObjectReference) + **out = **in } if in.Env != nil { in, out := &in.Env, &out.Env - *out = make([]core_v1.EnvVar, len(*in)) + *out = make([]corev1.EnvVar, len(*in)) for i := range *in { (*in)[i].DeepCopyInto(&(*out)[i]) } } if in.Incremental != nil { in, out := &in.Incremental, &out.Incremental - if *in == nil { - *out = nil - } else { - *out = new(bool) - **out = **in - } + *out = new(bool) + **out = **in } return } @@ -1587,12 +1355,8 @@ func (in *SourceRevision) DeepCopyInto(out *SourceRevision) { *out = *in if in.Git != nil { in, out := &in.Git, &out.Git - if *in == nil { - *out = nil - } else { - *out = new(GitSourceRevision) - **out = **in - } + *out = new(GitSourceRevision) + **out = **in } return } @@ -1612,12 +1376,8 @@ func (in *SourceStrategyOptions) DeepCopyInto(out *SourceStrategyOptions) { *out = *in if in.Incremental != nil { in, out := &in.Incremental, &out.Incremental - if *in == nil { - *out = nil - } else { - *out = new(bool) - **out = **in - } + *out = new(bool) + **out = **in } return } @@ -1678,12 +1438,8 @@ func (in *WebHookTrigger) DeepCopyInto(out *WebHookTrigger) { *out = *in if in.SecretReference != nil { in, out := &in.SecretReference, &out.SecretReference - if *in == nil { - *out = nil - } else { - *out = new(SecretLocalReference) - **out = **in - } + *out = new(SecretLocalReference) + **out = **in } return } diff --git a/vendor/github.com/openshift/api/build/v1/zz_generated.swagger_doc_generated.go b/vendor/github.com/openshift/api/build/v1/zz_generated.swagger_doc_generated.go index 84b0c1c1dd..cc926d4431 100644 --- a/vendor/github.com/openshift/api/build/v1/zz_generated.swagger_doc_generated.go +++ b/vendor/github.com/openshift/api/build/v1/zz_generated.swagger_doc_generated.go @@ -367,10 +367,10 @@ func (GenericWebHookCause) SwaggerDoc() map[string]string { } var map_GenericWebHookEvent = map[string]string{ - "": "GenericWebHookEvent is the payload expected for a generic webhook post", - "type": "type is the type of source repository", - "git": "git is the git information if the Type is BuildSourceGit", - "env": "env contains additional environment variables you want to pass into a builder container. ValueFrom is not supported.", + "": "GenericWebHookEvent is the payload expected for a generic webhook post", + "type": "type is the type of source repository", + "git": "git is the git information if the Type is BuildSourceGit", + "env": "env contains additional environment variables you want to pass into a builder container. ValueFrom is not supported.", "dockerStrategyOptions": "DockerStrategyOptions contains additional docker-strategy specific options for the build", } @@ -446,7 +446,7 @@ func (ImageChangeCause) SwaggerDoc() map[string]string { } var map_ImageChangeTrigger = map[string]string{ - "": "ImageChangeTrigger allows builds to be triggered when an ImageStream changes", + "": "ImageChangeTrigger allows builds to be triggered when an ImageStream changes", "lastTriggeredImageID": "lastTriggeredImageID is used internally by the ImageChangeController to save last used image ID for build", "from": "from is a reference to an ImageStreamTag that will trigger a build when updated It is optional. If no From is specified, the From image from the build strategy will be used. Only one ImageChangeTrigger with an empty From reference is allowed in a build configuration.", "paused": "paused is true if this trigger is temporarily disabled. Optional.", diff --git a/vendor/github.com/openshift/api/config/v1/register.go b/vendor/github.com/openshift/api/config/v1/register.go index e096d9e313..66c342569a 100644 --- a/vendor/github.com/openshift/api/config/v1/register.go +++ b/vendor/github.com/openshift/api/config/v1/register.go @@ -44,8 +44,8 @@ func addKnownTypes(scheme *runtime.Scheme) error { &ConsoleList{}, &DNS{}, &DNSList{}, - &Features{}, - &FeaturesList{}, + &FeatureGate{}, + &FeatureGateList{}, &Image{}, &ImageList{}, &Infrastructure{}, diff --git a/vendor/github.com/openshift/api/config/v1/types.go b/vendor/github.com/openshift/api/config/v1/types.go index 13d008d420..7cca094710 100644 --- a/vendor/github.com/openshift/api/config/v1/types.go +++ b/vendor/github.com/openshift/api/config/v1/types.go @@ -254,9 +254,6 @@ type GenericAPIServerConfig struct { // admissionConfig holds information about how to configure admission. AdmissionConfig AdmissionConfig `json:"admission"` - // TODO remove this. We need a cut-over or we'll have a gap. - AdmissionPluginConfig map[string]AdmissionPluginConfig `json:"admissionPluginConfig,omitempty"` - KubeClientConfig KubeClientConfig `json:"kubeClientConfig"` } diff --git a/vendor/github.com/openshift/api/config/v1/types_cluster_operator.go b/vendor/github.com/openshift/api/config/v1/types_cluster_operator.go index b5cbd222ef..8508b5cd07 100644 --- a/vendor/github.com/openshift/api/config/v1/types_cluster_operator.go +++ b/vendor/github.com/openshift/api/config/v1/types_cluster_operator.go @@ -127,10 +127,10 @@ const ( // operator (eg: openshift-apiserver for the openshift-apiserver-operator). OperatorProgressing ClusterStatusConditionType = "Progressing" - // Failing indicates that the operator has encountered an error that is preventing it from working properly. - // The binary maintained by the operator (eg: openshift-apiserver for the openshift-apiserver-operator) may still be - // available, but the user intent cannot be fulfilled. - OperatorFailing ClusterStatusConditionType = "Failing" + // Degraded indicates that the operand is not functioning completely. An example of a degraded state + // would be if there should be 5 copies of the operand running but only 4 are running. It may still be available, + // but it is degraded + OperatorDegraded ClusterStatusConditionType = "Degraded" // Upgradeable indicates whether the operator is in a state that is safe to upgrade. When status is `False` // administrators should not upgrade their cluster and the message field should contain a human readable description diff --git a/vendor/github.com/openshift/api/config/v1/types_cluster_version.go b/vendor/github.com/openshift/api/config/v1/types_cluster_version.go index 8d4cb7776e..d650816d26 100644 --- a/vendor/github.com/openshift/api/config/v1/types_cluster_version.go +++ b/vendor/github.com/openshift/api/config/v1/types_cluster_version.go @@ -104,7 +104,7 @@ type ClusterVersionStatus struct { // conditions provides information about the cluster version. The condition // "Available" is set to true if the desiredUpdate has been reached. The // condition "Progressing" is set to true if an update is being applied. - // The condition "Failing" is set to true if an update is currently blocked + // The condition "Degraded" is set to true if an update is currently blocked // by a temporary or permanent error. Conditions are only valid for the // current desiredUpdate when metadata.generation is equal to // status.generation. diff --git a/vendor/github.com/openshift/api/config/v1/types_console.go b/vendor/github.com/openshift/api/config/v1/types_console.go index b1dea9cdf4..c8b5b482f5 100644 --- a/vendor/github.com/openshift/api/config/v1/types_console.go +++ b/vendor/github.com/openshift/api/config/v1/types_console.go @@ -52,5 +52,6 @@ type ConsoleAuthentication struct { // provides the user the option to perform single logout (SLO) through the identity // provider to destroy their single sign-on session. // +optional + // +kubebuilder:validation:Pattern=^$|^((https):\/\/?)[^\s()<>]+(?:\([\w\d]+\)|([^[:punct:]\s]|\/?))$ LogoutRedirect string `json:"logoutRedirect,omitempty"` } diff --git a/vendor/github.com/openshift/api/config/v1/types_features.go b/vendor/github.com/openshift/api/config/v1/types_feature.go similarity index 75% rename from vendor/github.com/openshift/api/config/v1/types_features.go rename to vendor/github.com/openshift/api/config/v1/types_feature.go index 0cc54890cf..a072f13824 100644 --- a/vendor/github.com/openshift/api/config/v1/types_features.go +++ b/vendor/github.com/openshift/api/config/v1/types_feature.go @@ -6,18 +6,18 @@ import metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" // +genclient:nonNamespaced // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object -// Features holds cluster-wide information about feature gates. The canonical name is `cluster` -type Features struct { +// Feature holds cluster-wide information about feature gates. The canonical name is `cluster` +type FeatureGate struct { metav1.TypeMeta `json:",inline"` // Standard object's metadata. metav1.ObjectMeta `json:"metadata,omitempty"` // spec holds user settable values for configuration // +required - Spec FeaturesSpec `json:"spec"` + Spec FeatureGateSpec `json:"spec"` // status holds observed values from the cluster. They may not be overridden. // +optional - Status FeaturesStatus `json:"status"` + Status FeatureGateStatus `json:"status"` } type FeatureSet string @@ -31,30 +31,30 @@ var ( TechPreviewNoUpgrade FeatureSet = "TechPreviewNoUpgrade" ) -type FeaturesSpec struct { +type FeatureGateSpec struct { // featureSet changes the list of features in the cluster. The default is empty. Be very careful adjusting this setting. // Turning on or off features may cause irreversible changes in your cluster which cannot be undone. FeatureSet FeatureSet `json:"featureSet,omitempty"` } -type FeaturesStatus struct { +type FeatureGateStatus struct { } // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object -type FeaturesList struct { +type FeatureGateList struct { metav1.TypeMeta `json:",inline"` // Standard object's metadata. metav1.ListMeta `json:"metadata"` - Items []Features `json:"items"` + Items []FeatureGate `json:"items"` } -type FeatureEnabledDisabled struct { +type FeatureGateEnabledDisabled struct { Enabled []string Disabled []string } -// FeatureSets Contains a map of Feature names to Enabled/Disabled Features. +// FeatureSets Contains a map of Feature names to Enabled/Disabled Feature. // // NOTE: The caller needs to make sure to check for the existence of the value // using golang's existence field. A possible scenario is an upgrade where new @@ -62,28 +62,29 @@ type FeatureEnabledDisabled struct { // version of this file. In this upgrade scenario the map could return nil. // // example: -// if featureSet, ok := FeaturesSets["SomeNewFeature"]; ok { } +// if featureSet, ok := FeatureSets["SomeNewFeature"]; ok { } // // If you put an item in either of these lists, put your area and name on it so we can find owners. -var FeatureSets = map[FeatureSet]*FeatureEnabledDisabled{ - Default: &FeatureEnabledDisabled{ +var FeatureSets = map[FeatureSet]*FeatureGateEnabledDisabled{ + Default: { Enabled: []string{ "ExperimentalCriticalPodAnnotation", // sig-pod, sjenning "RotateKubeletServerCertificate", // sig-pod, sjenning + "SupportPodPidsLimit", // sig-pod, sjenning }, Disabled: []string{ "LocalStorageCapacityIsolation", // sig-pod, sjenning - "PersistentLocalVolumes", // sig-storage, hekumar@redhat.com }, }, - TechPreviewNoUpgrade: &FeatureEnabledDisabled{ + TechPreviewNoUpgrade: { Enabled: []string{ "ExperimentalCriticalPodAnnotation", // sig-pod, sjenning "RotateKubeletServerCertificate", // sig-pod, sjenning + "SupportPodPidsLimit", // sig-pod, sjenning + "CSIBlockVolume", // sig-storage, j-griffith }, Disabled: []string{ "LocalStorageCapacityIsolation", // sig-pod, sjenning - "PersistentLocalVolumes", // sig-storage, hekumar@redhat.com }, }, } diff --git a/vendor/github.com/openshift/api/config/v1/types_infrastructure.go b/vendor/github.com/openshift/api/config/v1/types_infrastructure.go index 7993f57cfe..40e3f2c279 100644 --- a/vendor/github.com/openshift/api/config/v1/types_infrastructure.go +++ b/vendor/github.com/openshift/api/config/v1/types_infrastructure.go @@ -22,17 +22,26 @@ type Infrastructure struct { // InfrastructureSpec contains settings that apply to the cluster infrastructure. type InfrastructureSpec struct { - // secret reference? - // configmap reference to file? + // cloudConfig is a reference to a ConfigMap containing the cloud provider configuration file. + // This configuration file is used to configure the Kubernetes cloud provider integration + // when using the built-in cloud provider integration or the external cloud controller manager. + // The namespace for this config map is openshift-config. + // +optional + CloudConfig ConfigMapFileReference `json:"cloudConfig"` } // InfrastructureStatus describes the infrastructure the cluster is leveraging. type InfrastructureStatus struct { + // infrastructureName uniquely identifies a cluster with a human friendly name. + // Once set it should not be changed. Must be of max length 27 and must have only + // alphanumeric or hyphen characters. + InfrastructureName string `json:"infrastructureName"` + // platform is the underlying infrastructure provider for the cluster. This // value controls whether infrastructure automation such as service load // balancers, dynamic volume provisioning, machine creation and deletion, and // other integrations are enabled. If None, no infrastructure automation is - // enabled. Allowed values are "AWS", "Azure", "GCP", "Libvirt", + // enabled. Allowed values are "AWS", "Azure", "BareMetal", "GCP", "Libvirt", // "OpenStack", "VSphere", and "None". Individual components may not support // all platforms, and must handle unrecognized platforms as None if they do // not support that platform. @@ -53,26 +62,29 @@ type InfrastructureStatus struct { type PlatformType string const ( - // AWSPlatform represents Amazon Web Services infrastructure. - AWSPlatform PlatformType = "AWS" + // AWSPlatformType represents Amazon Web Services infrastructure. + AWSPlatformType PlatformType = "AWS" + + // AzurePlatformType represents Microsoft Azure infrastructure. + AzurePlatformType PlatformType = "Azure" - // AzurePlatform represents Microsoft Azure infrastructure. - AzurePlatform PlatformType = "Azure" + // BareMetalPlatformType represents managed bare metal infrastructure. + BareMetalPlatformType PlatformType = "BareMetal" - // GCPPlatform represents Google Cloud Platform infrastructure. - GCPPlatform PlatformType = "GCP" + // GCPPlatformType represents Google Cloud Platform infrastructure. + GCPPlatformType PlatformType = "GCP" - // LibvirtPlatform represents libvirt infrastructure. - LibvirtPlatform PlatformType = "Libvirt" + // LibvirtPlatformType represents libvirt infrastructure. + LibvirtPlatformType PlatformType = "Libvirt" - // OpenStackPlatform represents OpenStack infrastructure. - OpenStackPlatform PlatformType = "OpenStack" + // OpenStackPlatformType represents OpenStack infrastructure. + OpenStackPlatformType PlatformType = "OpenStack" - // NonePlatform means there is no infrastructure provider. - NonePlatform PlatformType = "None" + // NonePlatformType means there is no infrastructure provider. + NonePlatformType PlatformType = "None" - // VSpherePlatform represents VMWare vSphere infrastructure. - VSpherePlatform PlatformType = "VSphere" + // VSpherePlatformType represents VMWare vSphere infrastructure. + VSpherePlatformType PlatformType = "VSphere" ) // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object diff --git a/vendor/github.com/openshift/api/config/v1/types_oauth.go b/vendor/github.com/openshift/api/config/v1/types_oauth.go index 6e4ed1ca1a..7aa0133e73 100644 --- a/vendor/github.com/openshift/api/config/v1/types_oauth.go +++ b/vendor/github.com/openshift/api/config/v1/types_oauth.go @@ -121,12 +121,6 @@ type IdentityProvider struct { // Ref: https://godoc.org/github.com/openshift/origin/pkg/user/apis/user/validation#ValidateIdentityProviderName Name string `json:"name"` - // challenge indicates whether to issue WWW-Authenticate challenges for this provider - UseAsChallenger bool `json:"challenge"` - - // login indicates whether to use this identity provider for unauthenticated browsers to login against - UseAsLogin bool `json:"login"` - // mappingMethod determines how identities from this provider are mapped to users // Defaults to "claim" // +optional @@ -517,28 +511,14 @@ type OpenIDIdentityProvider struct { // +optional ExtraAuthorizeParameters map[string]string `json:"extraAuthorizeParameters,omitempty"` - // urls to use to authenticate - URLs OpenIDURLs `json:"urls"` + // issuer is the URL that the OpenID Provider asserts as its Issuer Identifier. + // It must use the https scheme with no query or fragment component. + Issuer string `json:"issuer"` // claims mappings Claims OpenIDClaims `json:"claims"` } -// OpenIDURLs are URLs to use when authenticating with an OpenID identity provider -type OpenIDURLs struct { - // authorize is the oauth authorization URL - Authorize string `json:"authorize"` - - // token is the oauth token granting URL - Token string `json:"token"` - - // userInfo is the optional userinfo URL. - // If present, a granted access_token is used to request claims - // If empty, a granted id_token is parsed for claims - // +optional - UserInfo string `json:"userInfo"` -} - // UserIDClaim is the claim used to provide a stable identifier for OIDC identities. // Per http://openid.net/specs/openid-connect-core-1_0.html#ClaimStability // "The sub (subject) and iss (issuer) Claims, used together, are the only Claims that an RP can diff --git a/vendor/github.com/openshift/api/config/v1/types_scheduling.go b/vendor/github.com/openshift/api/config/v1/types_scheduling.go index f07e7b083e..48195e8b2f 100644 --- a/vendor/github.com/openshift/api/config/v1/types_scheduling.go +++ b/vendor/github.com/openshift/api/config/v1/types_scheduling.go @@ -27,6 +27,27 @@ type SchedulerSpec struct { // The namespace for this configmap is openshift-config. // +optional Policy ConfigMapNameReference `json:"policy"` + // defaultNodeSelector helps set the cluster-wide default node selector to + // restrict pod placement to specific nodes. This is applied to the pods + // created in all namespaces without a specified nodeSelector value. + // For example, + // defaultNodeSelector: "type=user-node,region=east" would set nodeSelector + // field in pod spec to "type=user-node,region=east" to all pods created + // in all namespaces. Namespaces having project-wide node selectors won't be + // impacted even if this field is set. This adds an annotation section to + // the namespace. + // For example, if a new namespace is created with + // node-selector='type=user-node,region=east', + // the annotation openshift.io/node-selector: type=user-node,region=east + // gets added to the project. When the openshift.io/node-selector annotation + // is set on the project the value is used in preference to the value we are setting + // for defaultNodeSelector field. + // For instance, + // openshift.io/node-selector: "type=user-node,region=west" means + // that the default of "type=user-node,region=east" set in defaultNodeSelector + // would not be applied. + // +optional + DefaultNodeSelector string `json:"defaultNodeSelector,omitempty"` } type SchedulerStatus struct { diff --git a/vendor/github.com/openshift/api/config/v1/zz_generated.deepcopy.go b/vendor/github.com/openshift/api/config/v1/zz_generated.deepcopy.go index a3be0e654f..3ad614c5e3 100644 --- a/vendor/github.com/openshift/api/config/v1/zz_generated.deepcopy.go +++ b/vendor/github.com/openshift/api/config/v1/zz_generated.deepcopy.go @@ -5,7 +5,7 @@ package v1 import ( - core_v1 "k8s.io/api/core/v1" + corev1 "k8s.io/api/core/v1" runtime "k8s.io/apimachinery/pkg/runtime" ) @@ -157,9 +157,7 @@ func (in *AdmissionConfig) DeepCopyInto(out *AdmissionConfig) { in, out := &in.PluginConfig, &out.PluginConfig *out = make(map[string]AdmissionPluginConfig, len(*in)) for key, val := range *in { - newVal := new(AdmissionPluginConfig) - val.DeepCopyInto(newVal) - (*out)[key] = *newVal + (*out)[key] = *val.DeepCopy() } } if in.EnabledAdmissionPlugins != nil { @@ -368,25 +366,17 @@ func (in *BuildDefaults) DeepCopyInto(out *BuildDefaults) { *out = *in if in.DefaultProxy != nil { in, out := &in.DefaultProxy, &out.DefaultProxy - if *in == nil { - *out = nil - } else { - *out = new(ProxySpec) - **out = **in - } + *out = new(ProxySpec) + **out = **in } if in.GitProxy != nil { in, out := &in.GitProxy, &out.GitProxy - if *in == nil { - *out = nil - } else { - *out = new(ProxySpec) - **out = **in - } + *out = new(ProxySpec) + **out = **in } if in.Env != nil { in, out := &in.Env, &out.Env - *out = make([]core_v1.EnvVar, len(*in)) + *out = make([]corev1.EnvVar, len(*in)) for i := range *in { (*in)[i].DeepCopyInto(&(*out)[i]) } @@ -460,7 +450,7 @@ func (in *BuildOverrides) DeepCopyInto(out *BuildOverrides) { } if in.Tolerations != nil { in, out := &in.Tolerations, &out.Tolerations - *out = make([]core_v1.Toleration, len(*in)) + *out = make([]corev1.Toleration, len(*in)) for i := range *in { (*in)[i].DeepCopyInto(&(*out)[i]) } @@ -739,12 +729,8 @@ func (in *ClusterVersionSpec) DeepCopyInto(out *ClusterVersionSpec) { *out = *in if in.DesiredUpdate != nil { in, out := &in.DesiredUpdate, &out.DesiredUpdate - if *in == nil { - *out = nil - } else { - *out = new(Update) - **out = **in - } + *out = new(Update) + **out = **in } if in.Overrides != nil { in, out := &in.Overrides, &out.Overrides @@ -1024,21 +1010,13 @@ func (in *DNSSpec) DeepCopyInto(out *DNSSpec) { *out = *in if in.PublicZone != nil { in, out := &in.PublicZone, &out.PublicZone - if *in == nil { - *out = nil - } else { - *out = new(DNSZone) - (*in).DeepCopyInto(*out) - } + *out = new(DNSZone) + (*in).DeepCopyInto(*out) } if in.PrivateZone != nil { in, out := &in.PrivateZone, &out.PrivateZone - if *in == nil { - *out = nil - } else { - *out = new(DNSZone) - (*in).DeepCopyInto(*out) - } + *out = new(DNSZone) + (*in).DeepCopyInto(*out) } return } @@ -1164,67 +1142,67 @@ func (in *EtcdStorageConfig) DeepCopy() *EtcdStorageConfig { } // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. -func (in *FeatureEnabledDisabled) DeepCopyInto(out *FeatureEnabledDisabled) { +func (in *FeatureGate) DeepCopyInto(out *FeatureGate) { *out = *in - if in.Enabled != nil { - in, out := &in.Enabled, &out.Enabled - *out = make([]string, len(*in)) - copy(*out, *in) - } - if in.Disabled != nil { - in, out := &in.Disabled, &out.Disabled - *out = make([]string, len(*in)) - copy(*out, *in) - } + out.TypeMeta = in.TypeMeta + in.ObjectMeta.DeepCopyInto(&out.ObjectMeta) + out.Spec = in.Spec + out.Status = in.Status return } -// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new FeatureEnabledDisabled. -func (in *FeatureEnabledDisabled) DeepCopy() *FeatureEnabledDisabled { +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new FeatureGate. +func (in *FeatureGate) DeepCopy() *FeatureGate { if in == nil { return nil } - out := new(FeatureEnabledDisabled) + out := new(FeatureGate) in.DeepCopyInto(out) return out } +// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object. +func (in *FeatureGate) DeepCopyObject() runtime.Object { + if c := in.DeepCopy(); c != nil { + return c + } + return nil +} + // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. -func (in *Features) DeepCopyInto(out *Features) { +func (in *FeatureGateEnabledDisabled) DeepCopyInto(out *FeatureGateEnabledDisabled) { *out = *in - out.TypeMeta = in.TypeMeta - in.ObjectMeta.DeepCopyInto(&out.ObjectMeta) - out.Spec = in.Spec - out.Status = in.Status + if in.Enabled != nil { + in, out := &in.Enabled, &out.Enabled + *out = make([]string, len(*in)) + copy(*out, *in) + } + if in.Disabled != nil { + in, out := &in.Disabled, &out.Disabled + *out = make([]string, len(*in)) + copy(*out, *in) + } return } -// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new Features. -func (in *Features) DeepCopy() *Features { +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new FeatureGateEnabledDisabled. +func (in *FeatureGateEnabledDisabled) DeepCopy() *FeatureGateEnabledDisabled { if in == nil { return nil } - out := new(Features) + out := new(FeatureGateEnabledDisabled) in.DeepCopyInto(out) return out } -// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object. -func (in *Features) DeepCopyObject() runtime.Object { - if c := in.DeepCopy(); c != nil { - return c - } - return nil -} - // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. -func (in *FeaturesList) DeepCopyInto(out *FeaturesList) { +func (in *FeatureGateList) DeepCopyInto(out *FeatureGateList) { *out = *in out.TypeMeta = in.TypeMeta out.ListMeta = in.ListMeta if in.Items != nil { in, out := &in.Items, &out.Items - *out = make([]Features, len(*in)) + *out = make([]FeatureGate, len(*in)) for i := range *in { (*in)[i].DeepCopyInto(&(*out)[i]) } @@ -1232,18 +1210,18 @@ func (in *FeaturesList) DeepCopyInto(out *FeaturesList) { return } -// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new FeaturesList. -func (in *FeaturesList) DeepCopy() *FeaturesList { +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new FeatureGateList. +func (in *FeatureGateList) DeepCopy() *FeatureGateList { if in == nil { return nil } - out := new(FeaturesList) + out := new(FeatureGateList) in.DeepCopyInto(out) return out } // DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object. -func (in *FeaturesList) DeepCopyObject() runtime.Object { +func (in *FeatureGateList) DeepCopyObject() runtime.Object { if c := in.DeepCopy(); c != nil { return c } @@ -1251,33 +1229,33 @@ func (in *FeaturesList) DeepCopyObject() runtime.Object { } // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. -func (in *FeaturesSpec) DeepCopyInto(out *FeaturesSpec) { +func (in *FeatureGateSpec) DeepCopyInto(out *FeatureGateSpec) { *out = *in return } -// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new FeaturesSpec. -func (in *FeaturesSpec) DeepCopy() *FeaturesSpec { +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new FeatureGateSpec. +func (in *FeatureGateSpec) DeepCopy() *FeatureGateSpec { if in == nil { return nil } - out := new(FeaturesSpec) + out := new(FeatureGateSpec) in.DeepCopyInto(out) return out } // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. -func (in *FeaturesStatus) DeepCopyInto(out *FeaturesStatus) { +func (in *FeatureGateStatus) DeepCopyInto(out *FeatureGateStatus) { *out = *in return } -// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new FeaturesStatus. -func (in *FeaturesStatus) DeepCopy() *FeaturesStatus { +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new FeatureGateStatus. +func (in *FeatureGateStatus) DeepCopy() *FeatureGateStatus { if in == nil { return nil } - out := new(FeaturesStatus) + out := new(FeatureGateStatus) in.DeepCopyInto(out) return out } @@ -1294,15 +1272,6 @@ func (in *GenericAPIServerConfig) DeepCopyInto(out *GenericAPIServerConfig) { in.AuditConfig.DeepCopyInto(&out.AuditConfig) in.StorageConfig.DeepCopyInto(&out.StorageConfig) in.AdmissionConfig.DeepCopyInto(&out.AdmissionConfig) - if in.AdmissionPluginConfig != nil { - in, out := &in.AdmissionPluginConfig, &out.AdmissionPluginConfig - *out = make(map[string]AdmissionPluginConfig, len(*in)) - for key, val := range *in { - newVal := new(AdmissionPluginConfig) - val.DeepCopyInto(newVal) - (*out)[key] = *newVal - } - } out.KubeClientConfig = in.KubeClientConfig return } @@ -1456,84 +1425,48 @@ func (in *IdentityProviderConfig) DeepCopyInto(out *IdentityProviderConfig) { *out = *in if in.BasicAuth != nil { in, out := &in.BasicAuth, &out.BasicAuth - if *in == nil { - *out = nil - } else { - *out = new(BasicAuthIdentityProvider) - **out = **in - } + *out = new(BasicAuthIdentityProvider) + **out = **in } if in.GitHub != nil { in, out := &in.GitHub, &out.GitHub - if *in == nil { - *out = nil - } else { - *out = new(GitHubIdentityProvider) - (*in).DeepCopyInto(*out) - } + *out = new(GitHubIdentityProvider) + (*in).DeepCopyInto(*out) } if in.GitLab != nil { in, out := &in.GitLab, &out.GitLab - if *in == nil { - *out = nil - } else { - *out = new(GitLabIdentityProvider) - **out = **in - } + *out = new(GitLabIdentityProvider) + **out = **in } if in.Google != nil { in, out := &in.Google, &out.Google - if *in == nil { - *out = nil - } else { - *out = new(GoogleIdentityProvider) - **out = **in - } + *out = new(GoogleIdentityProvider) + **out = **in } if in.HTPasswd != nil { in, out := &in.HTPasswd, &out.HTPasswd - if *in == nil { - *out = nil - } else { - *out = new(HTPasswdIdentityProvider) - **out = **in - } + *out = new(HTPasswdIdentityProvider) + **out = **in } if in.Keystone != nil { in, out := &in.Keystone, &out.Keystone - if *in == nil { - *out = nil - } else { - *out = new(KeystoneIdentityProvider) - **out = **in - } + *out = new(KeystoneIdentityProvider) + **out = **in } if in.LDAP != nil { in, out := &in.LDAP, &out.LDAP - if *in == nil { - *out = nil - } else { - *out = new(LDAPIdentityProvider) - (*in).DeepCopyInto(*out) - } + *out = new(LDAPIdentityProvider) + (*in).DeepCopyInto(*out) } if in.OpenID != nil { in, out := &in.OpenID, &out.OpenID - if *in == nil { - *out = nil - } else { - *out = new(OpenIDIdentityProvider) - (*in).DeepCopyInto(*out) - } + *out = new(OpenIDIdentityProvider) + (*in).DeepCopyInto(*out) } if in.RequestHeader != nil { in, out := &in.RequestHeader, &out.RequestHeader - if *in == nil { - *out = nil - } else { - *out = new(RequestHeaderIdentityProvider) - (*in).DeepCopyInto(*out) - } + *out = new(RequestHeaderIdentityProvider) + (*in).DeepCopyInto(*out) } return } @@ -1738,6 +1671,7 @@ func (in *InfrastructureList) DeepCopyObject() runtime.Object { // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *InfrastructureSpec) DeepCopyInto(out *InfrastructureSpec) { *out = *in + out.CloudConfig = in.CloudConfig return } @@ -2307,7 +2241,6 @@ func (in *OpenIDIdentityProvider) DeepCopyInto(out *OpenIDIdentityProvider) { (*out)[key] = val } } - out.URLs = in.URLs in.Claims.DeepCopyInto(&out.Claims) return } @@ -2322,22 +2255,6 @@ func (in *OpenIDIdentityProvider) DeepCopy() *OpenIDIdentityProvider { return out } -// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. -func (in *OpenIDURLs) DeepCopyInto(out *OpenIDURLs) { - *out = *in - return -} - -// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new OpenIDURLs. -func (in *OpenIDURLs) DeepCopy() *OpenIDURLs { - if in == nil { - return nil - } - out := new(OpenIDURLs) - in.DeepCopyInto(out) - return out -} - // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *OperandVersion) DeepCopyInto(out *OperandVersion) { *out = *in @@ -2856,11 +2773,7 @@ func (in *UpdateHistory) DeepCopyInto(out *UpdateHistory) { in.StartedTime.DeepCopyInto(&out.StartedTime) if in.CompletionTime != nil { in, out := &in.CompletionTime, &out.CompletionTime - if *in == nil { - *out = nil - } else { - *out = (*in).DeepCopy() - } + *out = (*in).DeepCopy() } return } diff --git a/vendor/github.com/openshift/api/config/v1/zz_generated.swagger_doc_generated.go b/vendor/github.com/openshift/api/config/v1/zz_generated.swagger_doc_generated.go index 1545e4e2b4..91f3a7353f 100644 --- a/vendor/github.com/openshift/api/config/v1/zz_generated.swagger_doc_generated.go +++ b/vendor/github.com/openshift/api/config/v1/zz_generated.swagger_doc_generated.go @@ -149,7 +149,7 @@ func (GenericControllerConfig) SwaggerDoc() map[string]string { } var map_HTTPServingInfo = map[string]string{ - "": "HTTPServingInfo holds configuration for serving HTTP", + "": "HTTPServingInfo holds configuration for serving HTTP", "maxRequestsInFlight": "MaxRequestsInFlight is the number of concurrent requests allowed to the server. If zero, no limit.", "requestTimeoutSeconds": "RequestTimeoutSeconds is the number of seconds before requests are timed out. The default is 60 minutes, if -1 there is no limit on requests.", } @@ -492,7 +492,7 @@ var map_ClusterVersionStatus = map[string]string{ "history": "history contains a list of the most recent versions applied to the cluster. This value may be empty during cluster startup, and then will be updated when a new update is being applied. The newest update is first in the list and it is ordered by recency. Updates in the history have state Completed if the rollout completed - if an update was failing or halfway applied the state will be Partial. Only a limited amount of update history is preserved.", "observedGeneration": "observedGeneration reports which version of the spec is being synced. If this value is not equal to metadata.generation, then the desired and conditions fields may represent from a previous version.", "versionHash": "versionHash is a fingerprint of the content that the cluster will be updated with. It is used by the operator to avoid unnecessary work and is for internal use only.", - "conditions": "conditions provides information about the cluster version. The condition \"Available\" is set to true if the desiredUpdate has been reached. The condition \"Progressing\" is set to true if an update is being applied. The condition \"Failing\" is set to true if an update is currently blocked by a temporary or permanent error. Conditions are only valid for the current desiredUpdate when metadata.generation is equal to status.generation.", + "conditions": "conditions provides information about the cluster version. The condition \"Available\" is set to true if the desiredUpdate has been reached. The condition \"Progressing\" is set to true if an update is being applied. The condition \"Degraded\" is set to true if an update is currently blocked by a temporary or permanent error. Conditions are only valid for the current desiredUpdate when metadata.generation is equal to status.generation.", "availableUpdates": "availableUpdates contains the list of updates that are appropriate for this cluster. This list may be empty if no updates are recommended, if the update service is unavailable, or if an invalid channel has been specified.", } @@ -610,31 +610,31 @@ func (DNSZone) SwaggerDoc() map[string]string { return map_DNSZone } -var map_Features = map[string]string{ - "": "Features holds cluster-wide information about feature gates. The canonical name is `cluster`", +var map_FeatureGate = map[string]string{ + "": "Feature holds cluster-wide information about feature gates. The canonical name is `cluster`", "metadata": "Standard object's metadata.", "spec": "spec holds user settable values for configuration", "status": "status holds observed values from the cluster. They may not be overridden.", } -func (Features) SwaggerDoc() map[string]string { - return map_Features +func (FeatureGate) SwaggerDoc() map[string]string { + return map_FeatureGate } -var map_FeaturesList = map[string]string{ +var map_FeatureGateList = map[string]string{ "metadata": "Standard object's metadata.", } -func (FeaturesList) SwaggerDoc() map[string]string { - return map_FeaturesList +func (FeatureGateList) SwaggerDoc() map[string]string { + return map_FeatureGateList } -var map_FeaturesSpec = map[string]string{ +var map_FeatureGateSpec = map[string]string{ "featureSet": "featureSet changes the list of features in the cluster. The default is empty. Be very careful adjusting this setting. Turning on or off features may cause irreversible changes in your cluster which cannot be undone.", } -func (FeaturesSpec) SwaggerDoc() map[string]string { - return map_FeaturesSpec +func (FeatureGateSpec) SwaggerDoc() map[string]string { + return map_FeatureGateSpec } var map_Image = map[string]string{ @@ -718,7 +718,8 @@ func (InfrastructureList) SwaggerDoc() map[string]string { } var map_InfrastructureSpec = map[string]string{ - "": "InfrastructureSpec contains settings that apply to the cluster infrastructure.", + "": "InfrastructureSpec contains settings that apply to the cluster infrastructure.", + "cloudConfig": "cloudConfig is a reference to a ConfigMap containing the cloud provider configuration file. This configuration file is used to configure the Kubernetes cloud provider integration when using the built-in cloud provider integration or the external cloud controller manager. The namespace for this config map is openshift-config.", } func (InfrastructureSpec) SwaggerDoc() map[string]string { @@ -727,7 +728,8 @@ func (InfrastructureSpec) SwaggerDoc() map[string]string { var map_InfrastructureStatus = map[string]string{ "": "InfrastructureStatus describes the infrastructure the cluster is leveraging.", - "platform": "platform is the underlying infrastructure provider for the cluster. This value controls whether infrastructure automation such as service load balancers, dynamic volume provisioning, machine creation and deletion, and other integrations are enabled. If None, no infrastructure automation is enabled. Allowed values are \"AWS\", \"Azure\", \"GCP\", \"Libvirt\", \"OpenStack\", \"VSphere\", and \"None\". Individual components may not support all platforms, and must handle unrecognized platforms as None if they do not support that platform.", + "infrastructureName": "infrastructureName uniquely identifies a cluster with a human friendly name. Once set it should not be changed. Must be of max length 27 and must have only alphanumeric or hyphen characters.", + "platform": "platform is the underlying infrastructure provider for the cluster. This value controls whether infrastructure automation such as service load balancers, dynamic volume provisioning, machine creation and deletion, and other integrations are enabled. If None, no infrastructure automation is enabled. Allowed values are \"AWS\", \"Azure\", \"BareMetal\", \"GCP\", \"Libvirt\", \"OpenStack\", \"VSphere\", and \"None\". Individual components may not support all platforms, and must handle unrecognized platforms as None if they do not support that platform.", "etcdDiscoveryDomain": "etcdDiscoveryDomain is the domain used to fetch the SRV records for discovering etcd servers and clients. For more info: https://github.com/etcd-io/etcd/blob/329be66e8b3f9e2e6af83c123ff89297e49ebd15/Documentation/op-guide/clustering.md#dns-discovery", "apiServerURL": "apiServerURL is a valid URL with scheme(http/https), address and port. apiServerURL can be used by components like kubelet on machines, to contact the `apisever` using the infrastructure provider rather than the kubernetes networking.", } @@ -872,8 +874,6 @@ func (HTPasswdIdentityProvider) SwaggerDoc() map[string]string { var map_IdentityProvider = map[string]string{ "": "IdentityProvider provides identities for users authenticating using credentials", "name": "name is used to qualify the identities returned by this provider. - It MUST be unique and not shared by any other identity provider used - It MUST be a valid path segment: name cannot equal \".\" or \"..\" or contain \"/\" or \"%\" or \":\"\n Ref: https://godoc.org/github.com/openshift/origin/pkg/user/apis/user/validation#ValidateIdentityProviderName", - "challenge": "challenge indicates whether to issue WWW-Authenticate challenges for this provider", - "login": "login indicates whether to use this identity provider for unauthenticated browsers to login against", "mappingMethod": "mappingMethod determines how identities from this provider are mapped to users Defaults to \"claim\"", } @@ -1002,25 +1002,14 @@ var map_OpenIDIdentityProvider = map[string]string{ "ca": "ca is an optional reference to a config map by name containing the PEM-encoded CA bundle. It is used as a trust anchor to validate the TLS certificate presented by the remote server. The key \"ca.crt\" is used to locate the data. If specified and the config map or expected key is not found, the identity provider is not honored. If the specified ca data is not valid, the identity provider is not honored. If empty, the default system roots are used. The namespace for this config map is openshift-config.", "extraScopes": "extraScopes are any scopes to request in addition to the standard \"openid\" scope.", "extraAuthorizeParameters": "extraAuthorizeParameters are any custom parameters to add to the authorize request.", - "urls": "urls to use to authenticate", - "claims": "claims mappings", + "issuer": "issuer is the URL that the OpenID Provider asserts as its Issuer Identifier. It must use the https scheme with no query or fragment component.", + "claims": "claims mappings", } func (OpenIDIdentityProvider) SwaggerDoc() map[string]string { return map_OpenIDIdentityProvider } -var map_OpenIDURLs = map[string]string{ - "": "OpenIDURLs are URLs to use when authenticating with an OpenID identity provider", - "authorize": "authorize is the oauth authorization URL", - "token": "token is the oauth token granting URL", - "userInfo": "userInfo is the optional userinfo URL. If present, a granted access_token is used to request claims If empty, a granted id_token is parsed for claims", -} - -func (OpenIDURLs) SwaggerDoc() map[string]string { - return map_OpenIDURLs -} - var map_RequestHeaderIdentityProvider = map[string]string{ "": "RequestHeaderIdentityProvider provides identities for users authenticating using request header credentials", "loginURL": "loginURL is a URL to redirect unauthenticated /authorize requests to Unauthenticated requests from OAuth clients which expect interactive logins will be redirected here ${url} is replaced with the current URL, escaped to be safe in a query parameter\n https://www.example.com/sso-login?then=${url}\n${query} is replaced with the current query string\n https://www.example.com/auth-proxy/oauth/authorize?${query}\nRequired when login is set to true.", @@ -1038,7 +1027,7 @@ func (RequestHeaderIdentityProvider) SwaggerDoc() map[string]string { } var map_TokenConfig = map[string]string{ - "": "TokenConfig holds the necessary configuration options for authorization and access tokens", + "": "TokenConfig holds the necessary configuration options for authorization and access tokens", "accessTokenMaxAgeSeconds": "accessTokenMaxAgeSeconds defines the maximum age of access tokens", "accessTokenInactivityTimeoutSeconds": "accessTokenInactivityTimeoutSeconds defines the default token inactivity timeout for tokens granted by any client. The value represents the maximum amount of time that can occur between consecutive uses of the token. Tokens become invalid if they are not used within this temporal window. The user will need to acquire a new token to regain access once a token times out. Valid values are integer values:\n x < 0 Tokens time out is enabled but tokens never timeout unless configured per client (e.g. `-1`)\n x = 0 Tokens time out is disabled (default)\n x > 0 Tokens time out if there is no activity for x seconds\nThe current minimum allowed value for X is 300 (5 minutes)", } @@ -1067,7 +1056,7 @@ func (ProjectList) SwaggerDoc() map[string]string { } var map_ProjectSpec = map[string]string{ - "": "ProjectSpec holds the project creation configuration.", + "": "ProjectSpec holds the project creation configuration.", "projectRequestMessage": "projectRequestMessage is the string presented to a user if they are unable to request a project via the projectrequest api endpoint", "projectRequestTemplate": "projectRequestTemplate is the template to use for creating projects in response to projectrequest. This must point to a template in 'openshift-config' namespace. It is optional. If it is not specified, a default template is used.", } @@ -1132,7 +1121,8 @@ func (SchedulerList) SwaggerDoc() map[string]string { } var map_SchedulerSpec = map[string]string{ - "policy": "policy is a reference to a ConfigMap containing scheduler policy which has user specified predicates and priorities. If this ConfigMap is not available scheduler will default to use DefaultAlgorithmProvider. The namespace for this configmap is openshift-config.", + "policy": "policy is a reference to a ConfigMap containing scheduler policy which has user specified predicates and priorities. If this ConfigMap is not available scheduler will default to use DefaultAlgorithmProvider. The namespace for this configmap is openshift-config.", + "defaultNodeSelector": "defaultNodeSelector helps set the cluster-wide default node selector to restrict pod placement to specific nodes. This is applied to the pods created in all namespaces without a specified nodeSelector value. For example, defaultNodeSelector: \"type=user-node,region=east\" would set nodeSelector field in pod spec to \"type=user-node,region=east\" to all pods created in all namespaces. Namespaces having project-wide node selectors won't be impacted even if this field is set. This adds an annotation section to the namespace. For example, if a new namespace is created with node-selector='type=user-node,region=east', the annotation openshift.io/node-selector: type=user-node,region=east gets added to the project. When the openshift.io/node-selector annotation is set on the project the value is used in preference to the value we are setting for defaultNodeSelector field. For instance, openshift.io/node-selector: \"type=user-node,region=west\" means that the default of \"type=user-node,region=east\" set in defaultNodeSelector would not be applied.", } func (SchedulerSpec) SwaggerDoc() map[string]string { diff --git a/vendor/github.com/openshift/api/image/docker10/zz_generated.deepcopy.go b/vendor/github.com/openshift/api/image/docker10/zz_generated.deepcopy.go index b9301dcf20..b59f75ac21 100644 --- a/vendor/github.com/openshift/api/image/docker10/zz_generated.deepcopy.go +++ b/vendor/github.com/openshift/api/image/docker10/zz_generated.deepcopy.go @@ -19,8 +19,8 @@ func (in *DockerConfig) DeepCopyInto(out *DockerConfig) { if in.ExposedPorts != nil { in, out := &in.ExposedPorts, &out.ExposedPorts *out = make(map[string]struct{}, len(*in)) - for key := range *in { - (*out)[key] = struct{}{} + for key, val := range *in { + (*out)[key] = val } } if in.Env != nil { @@ -41,8 +41,8 @@ func (in *DockerConfig) DeepCopyInto(out *DockerConfig) { if in.Volumes != nil { in, out := &in.Volumes, &out.Volumes *out = make(map[string]struct{}, len(*in)) - for key := range *in { - (*out)[key] = struct{}{} + for key, val := range *in { + (*out)[key] = val } } if in.Entrypoint != nil { @@ -88,12 +88,8 @@ func (in *DockerImage) DeepCopyInto(out *DockerImage) { in.ContainerConfig.DeepCopyInto(&out.ContainerConfig) if in.Config != nil { in, out := &in.Config, &out.Config - if *in == nil { - *out = nil - } else { - *out = new(DockerConfig) - (*in).DeepCopyInto(*out) - } + *out = new(DockerConfig) + (*in).DeepCopyInto(*out) } return } diff --git a/vendor/github.com/openshift/api/image/dockerpre012/zz_generated.deepcopy.go b/vendor/github.com/openshift/api/image/dockerpre012/zz_generated.deepcopy.go index 14353b7056..d9042704ad 100644 --- a/vendor/github.com/openshift/api/image/dockerpre012/zz_generated.deepcopy.go +++ b/vendor/github.com/openshift/api/image/dockerpre012/zz_generated.deepcopy.go @@ -19,8 +19,8 @@ func (in *Config) DeepCopyInto(out *Config) { if in.ExposedPorts != nil { in, out := &in.ExposedPorts, &out.ExposedPorts *out = make(map[Port]struct{}, len(*in)) - for key := range *in { - (*out)[key] = struct{}{} + for key, val := range *in { + (*out)[key] = val } } if in.Env != nil { @@ -41,8 +41,8 @@ func (in *Config) DeepCopyInto(out *Config) { if in.Volumes != nil { in, out := &in.Volumes, &out.Volumes *out = make(map[string]struct{}, len(*in)) - for key := range *in { - (*out)[key] = struct{}{} + for key, val := range *in { + (*out)[key] = val } } if in.Entrypoint != nil { @@ -96,8 +96,8 @@ func (in *DockerConfig) DeepCopyInto(out *DockerConfig) { if in.ExposedPorts != nil { in, out := &in.ExposedPorts, &out.ExposedPorts *out = make(map[string]struct{}, len(*in)) - for key := range *in { - (*out)[key] = struct{}{} + for key, val := range *in { + (*out)[key] = val } } if in.Env != nil { @@ -118,8 +118,8 @@ func (in *DockerConfig) DeepCopyInto(out *DockerConfig) { if in.Volumes != nil { in, out := &in.Volumes, &out.Volumes *out = make(map[string]struct{}, len(*in)) - for key := range *in { - (*out)[key] = struct{}{} + for key, val := range *in { + (*out)[key] = val } } if in.Entrypoint != nil { @@ -165,12 +165,8 @@ func (in *DockerImage) DeepCopyInto(out *DockerImage) { in.ContainerConfig.DeepCopyInto(&out.ContainerConfig) if in.Config != nil { in, out := &in.Config, &out.Config - if *in == nil { - *out = nil - } else { - *out = new(DockerConfig) - (*in).DeepCopyInto(*out) - } + *out = new(DockerConfig) + (*in).DeepCopyInto(*out) } return } diff --git a/vendor/github.com/openshift/api/image/v1/generated.pb.go b/vendor/github.com/openshift/api/image/v1/generated.pb.go index 10b77e5c81..ae113b7d35 100644 --- a/vendor/github.com/openshift/api/image/v1/generated.pb.go +++ b/vendor/github.com/openshift/api/image/v1/generated.pb.go @@ -1,6 +1,5 @@ -// Code generated by protoc-gen-gogo. +// Code generated by protoc-gen-gogo. DO NOT EDIT. // source: github.com/openshift/api/image/v1/generated.proto -// DO NOT EDIT! /* Package v1 is a generated protocol buffer package. @@ -1807,24 +1806,6 @@ func (m *TagReferencePolicy) MarshalTo(dAtA []byte) (int, error) { return i, nil } -func encodeFixed64Generated(dAtA []byte, offset int, v uint64) int { - dAtA[offset] = uint8(v) - dAtA[offset+1] = uint8(v >> 8) - dAtA[offset+2] = uint8(v >> 16) - dAtA[offset+3] = uint8(v >> 24) - dAtA[offset+4] = uint8(v >> 32) - dAtA[offset+5] = uint8(v >> 40) - dAtA[offset+6] = uint8(v >> 48) - dAtA[offset+7] = uint8(v >> 56) - return offset + 8 -} -func encodeFixed32Generated(dAtA []byte, offset int, v uint32) int { - dAtA[offset] = uint8(v) - dAtA[offset+1] = uint8(v >> 8) - dAtA[offset+2] = uint8(v >> 16) - dAtA[offset+3] = uint8(v >> 24) - return offset + 4 -} func encodeVarintGenerated(dAtA []byte, offset int, v uint64) int { for v >= 1<<7 { dAtA[offset] = uint8(v&0x7f | 0x80) @@ -4490,51 +4471,14 @@ func (m *ImageSignature) Unmarshal(dAtA []byte) error { if postIndex > l { return io.ErrUnexpectedEOF } - var keykey uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - keykey |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - var stringLenmapkey uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - stringLenmapkey |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - intStringLenmapkey := int(stringLenmapkey) - if intStringLenmapkey < 0 { - return ErrInvalidLengthGenerated - } - postStringIndexmapkey := iNdEx + intStringLenmapkey - if postStringIndexmapkey > l { - return io.ErrUnexpectedEOF - } - mapkey := string(dAtA[iNdEx:postStringIndexmapkey]) - iNdEx = postStringIndexmapkey if m.SignedClaims == nil { m.SignedClaims = make(map[string]string) } - if iNdEx < postIndex { - var valuekey uint64 + var mapkey string + var mapvalue string + for iNdEx < postIndex { + entryPreIndex := iNdEx + var wire uint64 for shift := uint(0); ; shift += 7 { if shift >= 64 { return ErrIntOverflowGenerated @@ -4544,41 +4488,80 @@ func (m *ImageSignature) Unmarshal(dAtA []byte) error { } b := dAtA[iNdEx] iNdEx++ - valuekey |= (uint64(b) & 0x7F) << shift + wire |= (uint64(b) & 0x7F) << shift if b < 0x80 { break } } - var stringLenmapvalue uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated + fieldNum := int32(wire >> 3) + if fieldNum == 1 { + var stringLenmapkey uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLenmapkey |= (uint64(b) & 0x7F) << shift + if b < 0x80 { + break + } } - if iNdEx >= l { + intStringLenmapkey := int(stringLenmapkey) + if intStringLenmapkey < 0 { + return ErrInvalidLengthGenerated + } + postStringIndexmapkey := iNdEx + intStringLenmapkey + if postStringIndexmapkey > l { return io.ErrUnexpectedEOF } - b := dAtA[iNdEx] - iNdEx++ - stringLenmapvalue |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break + mapkey = string(dAtA[iNdEx:postStringIndexmapkey]) + iNdEx = postStringIndexmapkey + } else if fieldNum == 2 { + var stringLenmapvalue uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLenmapvalue |= (uint64(b) & 0x7F) << shift + if b < 0x80 { + break + } } + intStringLenmapvalue := int(stringLenmapvalue) + if intStringLenmapvalue < 0 { + return ErrInvalidLengthGenerated + } + postStringIndexmapvalue := iNdEx + intStringLenmapvalue + if postStringIndexmapvalue > l { + return io.ErrUnexpectedEOF + } + mapvalue = string(dAtA[iNdEx:postStringIndexmapvalue]) + iNdEx = postStringIndexmapvalue + } else { + iNdEx = entryPreIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > postIndex { + return io.ErrUnexpectedEOF + } + iNdEx += skippy } - intStringLenmapvalue := int(stringLenmapvalue) - if intStringLenmapvalue < 0 { - return ErrInvalidLengthGenerated - } - postStringIndexmapvalue := iNdEx + intStringLenmapvalue - if postStringIndexmapvalue > l { - return io.ErrUnexpectedEOF - } - mapvalue := string(dAtA[iNdEx:postStringIndexmapvalue]) - iNdEx = postStringIndexmapvalue - m.SignedClaims[mapkey] = mapvalue - } else { - var mapvalue string - m.SignedClaims[mapkey] = mapvalue } + m.SignedClaims[mapkey] = mapvalue iNdEx = postIndex case 7: if wireType != 2 { @@ -5456,51 +5439,14 @@ func (m *ImageStreamLayers) Unmarshal(dAtA []byte) error { if postIndex > l { return io.ErrUnexpectedEOF } - var keykey uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - keykey |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - var stringLenmapkey uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - stringLenmapkey |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - intStringLenmapkey := int(stringLenmapkey) - if intStringLenmapkey < 0 { - return ErrInvalidLengthGenerated - } - postStringIndexmapkey := iNdEx + intStringLenmapkey - if postStringIndexmapkey > l { - return io.ErrUnexpectedEOF - } - mapkey := string(dAtA[iNdEx:postStringIndexmapkey]) - iNdEx = postStringIndexmapkey if m.Blobs == nil { m.Blobs = make(map[string]ImageLayerData) } - if iNdEx < postIndex { - var valuekey uint64 + var mapkey string + mapvalue := &ImageLayerData{} + for iNdEx < postIndex { + entryPreIndex := iNdEx + var wire uint64 for shift := uint(0); ; shift += 7 { if shift >= 64 { return ErrIntOverflowGenerated @@ -5510,46 +5456,85 @@ func (m *ImageStreamLayers) Unmarshal(dAtA []byte) error { } b := dAtA[iNdEx] iNdEx++ - valuekey |= (uint64(b) & 0x7F) << shift + wire |= (uint64(b) & 0x7F) << shift if b < 0x80 { break } } - var mapmsglen int - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated + fieldNum := int32(wire >> 3) + if fieldNum == 1 { + var stringLenmapkey uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLenmapkey |= (uint64(b) & 0x7F) << shift + if b < 0x80 { + break + } } - if iNdEx >= l { + intStringLenmapkey := int(stringLenmapkey) + if intStringLenmapkey < 0 { + return ErrInvalidLengthGenerated + } + postStringIndexmapkey := iNdEx + intStringLenmapkey + if postStringIndexmapkey > l { return io.ErrUnexpectedEOF } - b := dAtA[iNdEx] - iNdEx++ - mapmsglen |= (int(b) & 0x7F) << shift - if b < 0x80 { - break + mapkey = string(dAtA[iNdEx:postStringIndexmapkey]) + iNdEx = postStringIndexmapkey + } else if fieldNum == 2 { + var mapmsglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + mapmsglen |= (int(b) & 0x7F) << shift + if b < 0x80 { + break + } } + if mapmsglen < 0 { + return ErrInvalidLengthGenerated + } + postmsgIndex := iNdEx + mapmsglen + if mapmsglen < 0 { + return ErrInvalidLengthGenerated + } + if postmsgIndex > l { + return io.ErrUnexpectedEOF + } + mapvalue = &ImageLayerData{} + if err := mapvalue.Unmarshal(dAtA[iNdEx:postmsgIndex]); err != nil { + return err + } + iNdEx = postmsgIndex + } else { + iNdEx = entryPreIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > postIndex { + return io.ErrUnexpectedEOF + } + iNdEx += skippy } - if mapmsglen < 0 { - return ErrInvalidLengthGenerated - } - postmsgIndex := iNdEx + mapmsglen - if mapmsglen < 0 { - return ErrInvalidLengthGenerated - } - if postmsgIndex > l { - return io.ErrUnexpectedEOF - } - mapvalue := &ImageLayerData{} - if err := mapvalue.Unmarshal(dAtA[iNdEx:postmsgIndex]); err != nil { - return err - } - iNdEx = postmsgIndex - m.Blobs[mapkey] = *mapvalue - } else { - var mapvalue ImageLayerData - m.Blobs[mapkey] = mapvalue } + m.Blobs[mapkey] = *mapvalue iNdEx = postIndex case 3: if wireType != 2 { @@ -5577,51 +5562,14 @@ func (m *ImageStreamLayers) Unmarshal(dAtA []byte) error { if postIndex > l { return io.ErrUnexpectedEOF } - var keykey uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - keykey |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - var stringLenmapkey uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - stringLenmapkey |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - intStringLenmapkey := int(stringLenmapkey) - if intStringLenmapkey < 0 { - return ErrInvalidLengthGenerated - } - postStringIndexmapkey := iNdEx + intStringLenmapkey - if postStringIndexmapkey > l { - return io.ErrUnexpectedEOF - } - mapkey := string(dAtA[iNdEx:postStringIndexmapkey]) - iNdEx = postStringIndexmapkey if m.Images == nil { m.Images = make(map[string]ImageBlobReferences) } - if iNdEx < postIndex { - var valuekey uint64 + var mapkey string + mapvalue := &ImageBlobReferences{} + for iNdEx < postIndex { + entryPreIndex := iNdEx + var wire uint64 for shift := uint(0); ; shift += 7 { if shift >= 64 { return ErrIntOverflowGenerated @@ -5631,46 +5579,85 @@ func (m *ImageStreamLayers) Unmarshal(dAtA []byte) error { } b := dAtA[iNdEx] iNdEx++ - valuekey |= (uint64(b) & 0x7F) << shift + wire |= (uint64(b) & 0x7F) << shift if b < 0x80 { break } } - var mapmsglen int - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated + fieldNum := int32(wire >> 3) + if fieldNum == 1 { + var stringLenmapkey uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLenmapkey |= (uint64(b) & 0x7F) << shift + if b < 0x80 { + break + } } - if iNdEx >= l { + intStringLenmapkey := int(stringLenmapkey) + if intStringLenmapkey < 0 { + return ErrInvalidLengthGenerated + } + postStringIndexmapkey := iNdEx + intStringLenmapkey + if postStringIndexmapkey > l { return io.ErrUnexpectedEOF } - b := dAtA[iNdEx] - iNdEx++ - mapmsglen |= (int(b) & 0x7F) << shift - if b < 0x80 { - break + mapkey = string(dAtA[iNdEx:postStringIndexmapkey]) + iNdEx = postStringIndexmapkey + } else if fieldNum == 2 { + var mapmsglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + mapmsglen |= (int(b) & 0x7F) << shift + if b < 0x80 { + break + } } + if mapmsglen < 0 { + return ErrInvalidLengthGenerated + } + postmsgIndex := iNdEx + mapmsglen + if mapmsglen < 0 { + return ErrInvalidLengthGenerated + } + if postmsgIndex > l { + return io.ErrUnexpectedEOF + } + mapvalue = &ImageBlobReferences{} + if err := mapvalue.Unmarshal(dAtA[iNdEx:postmsgIndex]); err != nil { + return err + } + iNdEx = postmsgIndex + } else { + iNdEx = entryPreIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > postIndex { + return io.ErrUnexpectedEOF + } + iNdEx += skippy } - if mapmsglen < 0 { - return ErrInvalidLengthGenerated - } - postmsgIndex := iNdEx + mapmsglen - if mapmsglen < 0 { - return ErrInvalidLengthGenerated - } - if postmsgIndex > l { - return io.ErrUnexpectedEOF - } - mapvalue := &ImageBlobReferences{} - if err := mapvalue.Unmarshal(dAtA[iNdEx:postmsgIndex]); err != nil { - return err - } - iNdEx = postmsgIndex - m.Images[mapkey] = *mapvalue - } else { - var mapvalue ImageBlobReferences - m.Images[mapkey] = mapvalue } + m.Images[mapkey] = *mapvalue iNdEx = postIndex default: iNdEx = preIndex @@ -8066,51 +8053,14 @@ func (m *TagReference) Unmarshal(dAtA []byte) error { if postIndex > l { return io.ErrUnexpectedEOF } - var keykey uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - keykey |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - var stringLenmapkey uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - stringLenmapkey |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - intStringLenmapkey := int(stringLenmapkey) - if intStringLenmapkey < 0 { - return ErrInvalidLengthGenerated - } - postStringIndexmapkey := iNdEx + intStringLenmapkey - if postStringIndexmapkey > l { - return io.ErrUnexpectedEOF - } - mapkey := string(dAtA[iNdEx:postStringIndexmapkey]) - iNdEx = postStringIndexmapkey if m.Annotations == nil { m.Annotations = make(map[string]string) } - if iNdEx < postIndex { - var valuekey uint64 + var mapkey string + var mapvalue string + for iNdEx < postIndex { + entryPreIndex := iNdEx + var wire uint64 for shift := uint(0); ; shift += 7 { if shift >= 64 { return ErrIntOverflowGenerated @@ -8120,41 +8070,80 @@ func (m *TagReference) Unmarshal(dAtA []byte) error { } b := dAtA[iNdEx] iNdEx++ - valuekey |= (uint64(b) & 0x7F) << shift + wire |= (uint64(b) & 0x7F) << shift if b < 0x80 { break } } - var stringLenmapvalue uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated + fieldNum := int32(wire >> 3) + if fieldNum == 1 { + var stringLenmapkey uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLenmapkey |= (uint64(b) & 0x7F) << shift + if b < 0x80 { + break + } } - if iNdEx >= l { + intStringLenmapkey := int(stringLenmapkey) + if intStringLenmapkey < 0 { + return ErrInvalidLengthGenerated + } + postStringIndexmapkey := iNdEx + intStringLenmapkey + if postStringIndexmapkey > l { return io.ErrUnexpectedEOF } - b := dAtA[iNdEx] - iNdEx++ - stringLenmapvalue |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break + mapkey = string(dAtA[iNdEx:postStringIndexmapkey]) + iNdEx = postStringIndexmapkey + } else if fieldNum == 2 { + var stringLenmapvalue uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLenmapvalue |= (uint64(b) & 0x7F) << shift + if b < 0x80 { + break + } } + intStringLenmapvalue := int(stringLenmapvalue) + if intStringLenmapvalue < 0 { + return ErrInvalidLengthGenerated + } + postStringIndexmapvalue := iNdEx + intStringLenmapvalue + if postStringIndexmapvalue > l { + return io.ErrUnexpectedEOF + } + mapvalue = string(dAtA[iNdEx:postStringIndexmapvalue]) + iNdEx = postStringIndexmapvalue + } else { + iNdEx = entryPreIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > postIndex { + return io.ErrUnexpectedEOF + } + iNdEx += skippy } - intStringLenmapvalue := int(stringLenmapvalue) - if intStringLenmapvalue < 0 { - return ErrInvalidLengthGenerated - } - postStringIndexmapvalue := iNdEx + intStringLenmapvalue - if postStringIndexmapvalue > l { - return io.ErrUnexpectedEOF - } - mapvalue := string(dAtA[iNdEx:postStringIndexmapvalue]) - iNdEx = postStringIndexmapvalue - m.Annotations[mapkey] = mapvalue - } else { - var mapvalue string - m.Annotations[mapkey] = mapvalue } + m.Annotations[mapkey] = mapvalue iNdEx = postIndex case 3: if wireType != 2 { diff --git a/vendor/github.com/openshift/api/image/v1/zz_generated.deepcopy.go b/vendor/github.com/openshift/api/image/v1/zz_generated.deepcopy.go index 93525f773d..cbde157c64 100644 --- a/vendor/github.com/openshift/api/image/v1/zz_generated.deepcopy.go +++ b/vendor/github.com/openshift/api/image/v1/zz_generated.deepcopy.go @@ -5,7 +5,7 @@ package v1 import ( - core_v1 "k8s.io/api/core/v1" + corev1 "k8s.io/api/core/v1" runtime "k8s.io/apimachinery/pkg/runtime" ) @@ -85,12 +85,8 @@ func (in *ImageBlobReferences) DeepCopyInto(out *ImageBlobReferences) { } if in.Config != nil { in, out := &in.Config, &out.Config - if *in == nil { - *out = nil - } else { - *out = new(string) - **out = **in - } + *out = new(string) + **out = **in } return } @@ -111,12 +107,8 @@ func (in *ImageImportSpec) DeepCopyInto(out *ImageImportSpec) { out.From = in.From if in.To != nil { in, out := &in.To, &out.To - if *in == nil { - *out = nil - } else { - *out = new(core_v1.LocalObjectReference) - **out = **in - } + *out = new(corev1.LocalObjectReference) + **out = **in } out.ImportPolicy = in.ImportPolicy out.ReferencePolicy = in.ReferencePolicy @@ -139,12 +131,8 @@ func (in *ImageImportStatus) DeepCopyInto(out *ImageImportStatus) { in.Status.DeepCopyInto(&out.Status) if in.Image != nil { in, out := &in.Image, &out.Image - if *in == nil { - *out = nil - } else { - *out = new(Image) - (*in).DeepCopyInto(*out) - } + *out = new(Image) + (*in).DeepCopyInto(*out) } return } @@ -180,12 +168,8 @@ func (in *ImageLayerData) DeepCopyInto(out *ImageLayerData) { *out = *in if in.LayerSize != nil { in, out := &in.LayerSize, &out.LayerSize - if *in == nil { - *out = nil - } else { - *out = new(int64) - **out = **in - } + *out = new(int64) + **out = **in } return } @@ -275,29 +259,17 @@ func (in *ImageSignature) DeepCopyInto(out *ImageSignature) { } if in.Created != nil { in, out := &in.Created, &out.Created - if *in == nil { - *out = nil - } else { - *out = (*in).DeepCopy() - } + *out = (*in).DeepCopy() } if in.IssuedBy != nil { in, out := &in.IssuedBy, &out.IssuedBy - if *in == nil { - *out = nil - } else { - *out = new(SignatureIssuer) - **out = **in - } + *out = new(SignatureIssuer) + **out = **in } if in.IssuedTo != nil { in, out := &in.IssuedTo, &out.IssuedTo - if *in == nil { - *out = nil - } else { - *out = new(SignatureSubject) - **out = **in - } + *out = new(SignatureSubject) + **out = **in } return } @@ -408,12 +380,8 @@ func (in *ImageStreamImportSpec) DeepCopyInto(out *ImageStreamImportSpec) { *out = *in if in.Repository != nil { in, out := &in.Repository, &out.Repository - if *in == nil { - *out = nil - } else { - *out = new(RepositoryImportSpec) - **out = **in - } + *out = new(RepositoryImportSpec) + **out = **in } if in.Images != nil { in, out := &in.Images, &out.Images @@ -440,21 +408,13 @@ func (in *ImageStreamImportStatus) DeepCopyInto(out *ImageStreamImportStatus) { *out = *in if in.Import != nil { in, out := &in.Import, &out.Import - if *in == nil { - *out = nil - } else { - *out = new(ImageStream) - (*in).DeepCopyInto(*out) - } + *out = new(ImageStream) + (*in).DeepCopyInto(*out) } if in.Repository != nil { in, out := &in.Repository, &out.Repository - if *in == nil { - *out = nil - } else { - *out = new(RepositoryImportStatus) - (*in).DeepCopyInto(*out) - } + *out = new(RepositoryImportStatus) + (*in).DeepCopyInto(*out) } if in.Images != nil { in, out := &in.Images, &out.Images @@ -485,18 +445,14 @@ func (in *ImageStreamLayers) DeepCopyInto(out *ImageStreamLayers) { in, out := &in.Blobs, &out.Blobs *out = make(map[string]ImageLayerData, len(*in)) for key, val := range *in { - newVal := new(ImageLayerData) - val.DeepCopyInto(newVal) - (*out)[key] = *newVal + (*out)[key] = *val.DeepCopy() } } if in.Images != nil { in, out := &in.Images, &out.Images *out = make(map[string]ImageBlobReferences, len(*in)) for key, val := range *in { - newVal := new(ImageBlobReferences) - val.DeepCopyInto(newVal) - (*out)[key] = *newVal + (*out)[key] = *val.DeepCopy() } } return @@ -634,12 +590,8 @@ func (in *ImageStreamTag) DeepCopyInto(out *ImageStreamTag) { in.ObjectMeta.DeepCopyInto(&out.ObjectMeta) if in.Tag != nil { in, out := &in.Tag, &out.Tag - if *in == nil { - *out = nil - } else { - *out = new(TagReference) - (*in).DeepCopyInto(*out) - } + *out = new(TagReference) + (*in).DeepCopyInto(*out) } out.LookupPolicy = in.LookupPolicy if in.Conditions != nil { @@ -912,21 +864,13 @@ func (in *TagReference) DeepCopyInto(out *TagReference) { } if in.From != nil { in, out := &in.From, &out.From - if *in == nil { - *out = nil - } else { - *out = new(core_v1.ObjectReference) - **out = **in - } + *out = new(corev1.ObjectReference) + **out = **in } if in.Generation != nil { in, out := &in.Generation, &out.Generation - if *in == nil { - *out = nil - } else { - *out = new(int64) - **out = **in - } + *out = new(int64) + **out = **in } out.ImportPolicy = in.ImportPolicy out.ReferencePolicy = in.ReferencePolicy diff --git a/vendor/github.com/openshift/api/image/v1/zz_generated.swagger_doc_generated.go b/vendor/github.com/openshift/api/image/v1/zz_generated.swagger_doc_generated.go index 7815909872..a31b6316cb 100644 --- a/vendor/github.com/openshift/api/image/v1/zz_generated.swagger_doc_generated.go +++ b/vendor/github.com/openshift/api/image/v1/zz_generated.swagger_doc_generated.go @@ -232,10 +232,10 @@ func (ImageStreamSpec) SwaggerDoc() map[string]string { } var map_ImageStreamStatus = map[string]string{ - "": "ImageStreamStatus contains information about the state of this image stream.", + "": "ImageStreamStatus contains information about the state of this image stream.", "dockerImageRepository": "DockerImageRepository represents the effective location this stream may be accessed at. May be empty until the server determines where the repository is located", "publicDockerImageRepository": "PublicDockerImageRepository represents the public location from where the image can be pulled outside the cluster. This field may be empty if the administrator has not exposed the integrated registry externally.", - "tags": "Tags are a historical record of images associated with each tag. The first entry in the TagEvent array is the currently tagged image.", + "tags": "Tags are a historical record of images associated with each tag. The first entry in the TagEvent array is the currently tagged image.", } func (ImageStreamStatus) SwaggerDoc() map[string]string { diff --git a/vendor/github.com/openshift/api/kubecontrolplane/v1/zz_generated.deepcopy.go b/vendor/github.com/openshift/api/kubecontrolplane/v1/zz_generated.deepcopy.go index d244856a24..3e4b5730f2 100644 --- a/vendor/github.com/openshift/api/kubecontrolplane/v1/zz_generated.deepcopy.go +++ b/vendor/github.com/openshift/api/kubecontrolplane/v1/zz_generated.deepcopy.go @@ -5,7 +5,7 @@ package v1 import ( - osin_v1 "github.com/openshift/api/osin/v1" + osinv1 "github.com/openshift/api/osin/v1" runtime "k8s.io/apimachinery/pkg/runtime" ) @@ -64,23 +64,22 @@ func (in *KubeAPIServerConfig) DeepCopyInto(out *KubeAPIServerConfig) { } if in.OAuthConfig != nil { in, out := &in.OAuthConfig, &out.OAuthConfig - if *in == nil { - *out = nil - } else { - *out = new(osin_v1.OAuthConfig) - (*in).DeepCopyInto(*out) - } + *out = new(osinv1.OAuthConfig) + (*in).DeepCopyInto(*out) } if in.APIServerArguments != nil { in, out := &in.APIServerArguments, &out.APIServerArguments *out = make(map[string]Arguments, len(*in)) for key, val := range *in { + var outVal []string if val == nil { (*out)[key] = nil } else { - (*out)[key] = make([]string, len(val)) - copy((*out)[key], val) + in, out := &val, &outVal + *out = make(Arguments, len(*in)) + copy(*out, *in) } + (*out)[key] = outVal } } return @@ -151,12 +150,15 @@ func (in *KubeControllerManagerConfig) DeepCopyInto(out *KubeControllerManagerCo in, out := &in.ExtendedArguments, &out.ExtendedArguments *out = make(map[string]Arguments, len(*in)) for key, val := range *in { + var outVal []string if val == nil { (*out)[key] = nil } else { - (*out)[key] = make([]string, len(val)) - copy((*out)[key], val) + in, out := &val, &outVal + *out = make(Arguments, len(*in)) + copy(*out, *in) } + (*out)[key] = outVal } } return @@ -218,12 +220,8 @@ func (in *MasterAuthConfig) DeepCopyInto(out *MasterAuthConfig) { *out = *in if in.RequestHeader != nil { in, out := &in.RequestHeader, &out.RequestHeader - if *in == nil { - *out = nil - } else { - *out = new(RequestHeaderAuthenticationOptions) - (*in).DeepCopyInto(*out) - } + *out = new(RequestHeaderAuthenticationOptions) + (*in).DeepCopyInto(*out) } if in.WebhookTokenAuthenticators != nil { in, out := &in.WebhookTokenAuthenticators, &out.WebhookTokenAuthenticators diff --git a/vendor/github.com/openshift/api/legacyconfig/v1/zz_generated.deepcopy.go b/vendor/github.com/openshift/api/legacyconfig/v1/zz_generated.deepcopy.go index e33a4c5cfd..b65c4beb7f 100644 --- a/vendor/github.com/openshift/api/legacyconfig/v1/zz_generated.deepcopy.go +++ b/vendor/github.com/openshift/api/legacyconfig/v1/zz_generated.deepcopy.go @@ -5,8 +5,8 @@ package v1 import ( - build_v1 "github.com/openshift/api/build/v1" - core_v1 "k8s.io/api/core/v1" + buildv1 "github.com/openshift/api/build/v1" + corev1 "k8s.io/api/core/v1" runtime "k8s.io/apimachinery/pkg/runtime" ) @@ -44,12 +44,15 @@ func (in *AdmissionConfig) DeepCopyInto(out *AdmissionConfig) { in, out := &in.PluginConfig, &out.PluginConfig *out = make(map[string]*AdmissionPluginConfig, len(*in)) for key, val := range *in { + var outVal *AdmissionPluginConfig if val == nil { (*out)[key] = nil } else { - (*out)[key] = new(AdmissionPluginConfig) - val.DeepCopyInto((*out)[key]) + in, out := &val, &outVal + *out = new(AdmissionPluginConfig) + (*in).DeepCopyInto(*out) } + (*out)[key] = outVal } } if in.PluginOrderOverride != nil { @@ -231,23 +234,19 @@ func (in *BuildDefaultsConfig) DeepCopyInto(out *BuildDefaultsConfig) { out.TypeMeta = in.TypeMeta if in.Env != nil { in, out := &in.Env, &out.Env - *out = make([]core_v1.EnvVar, len(*in)) + *out = make([]corev1.EnvVar, len(*in)) for i := range *in { (*in)[i].DeepCopyInto(&(*out)[i]) } } if in.SourceStrategyDefaults != nil { in, out := &in.SourceStrategyDefaults, &out.SourceStrategyDefaults - if *in == nil { - *out = nil - } else { - *out = new(SourceStrategyDefaultsConfig) - (*in).DeepCopyInto(*out) - } + *out = new(SourceStrategyDefaultsConfig) + (*in).DeepCopyInto(*out) } if in.ImageLabels != nil { in, out := &in.ImageLabels, &out.ImageLabels - *out = make([]build_v1.ImageLabel, len(*in)) + *out = make([]buildv1.ImageLabel, len(*in)) copy(*out, *in) } if in.NodeSelector != nil { @@ -292,7 +291,7 @@ func (in *BuildOverridesConfig) DeepCopyInto(out *BuildOverridesConfig) { out.TypeMeta = in.TypeMeta if in.ImageLabels != nil { in, out := &in.ImageLabels, &out.ImageLabels - *out = make([]build_v1.ImageLabel, len(*in)) + *out = make([]buildv1.ImageLabel, len(*in)) copy(*out, *in) } if in.NodeSelector != nil { @@ -311,7 +310,7 @@ func (in *BuildOverridesConfig) DeepCopyInto(out *BuildOverridesConfig) { } if in.Tolerations != nil { in, out := &in.Tolerations, &out.Tolerations - *out = make([]core_v1.Toleration, len(*in)) + *out = make([]corev1.Toleration, len(*in)) for i := range *in { (*in)[i].DeepCopyInto(&(*out)[i]) } @@ -395,12 +394,8 @@ func (in *ControllerConfig) DeepCopyInto(out *ControllerConfig) { } if in.Election != nil { in, out := &in.Election, &out.Election - if *in == nil { - *out = nil - } else { - *out = new(ControllerElectionConfig) - **out = **in - } + *out = new(ControllerElectionConfig) + **out = **in } in.ServiceServingCert.DeepCopyInto(&out.ServiceServingCert) return @@ -577,12 +572,15 @@ func (in ExtendedArguments) DeepCopyInto(out *ExtendedArguments) { in := &in *out = make(ExtendedArguments, len(*in)) for key, val := range *in { + var outVal []string if val == nil { (*out)[key] = nil } else { - (*out)[key] = make([]string, len(val)) - copy((*out)[key], val) + in, out := &val, &outVal + *out = make([]string, len(*in)) + copy(*out, *in) } + (*out)[key] = outVal } return } @@ -661,12 +659,8 @@ func (in *GitLabIdentityProvider) DeepCopyInto(out *GitLabIdentityProvider) { out.ClientSecret = in.ClientSecret if in.Legacy != nil { in, out := &in.Legacy, &out.Legacy - if *in == nil { - *out = nil - } else { - *out = new(bool) - **out = **in - } + *out = new(bool) + **out = **in } return } @@ -827,15 +821,11 @@ func (in *ImagePolicyConfig) DeepCopyInto(out *ImagePolicyConfig) { *out = *in if in.AllowedRegistriesForImport != nil { in, out := &in.AllowedRegistriesForImport, &out.AllowedRegistriesForImport - if *in == nil { - *out = nil - } else { - *out = new(AllowedRegistries) - if **in != nil { - in, out := *in, *out - *out = make([]RegistryLocation, len(*in)) - copy(*out, *in) - } + *out = new(AllowedRegistries) + if **in != nil { + in, out := *in, *out + *out = make([]RegistryLocation, len(*in)) + copy(*out, *in) } } return @@ -856,12 +846,8 @@ func (in *JenkinsPipelineConfig) DeepCopyInto(out *JenkinsPipelineConfig) { *out = *in if in.AutoProvisionEnabled != nil { in, out := &in.AutoProvisionEnabled, &out.AutoProvisionEnabled - if *in == nil { - *out = nil - } else { - *out = new(bool) - **out = **in - } + *out = new(bool) + **out = **in } if in.Parameters != nil { in, out := &in.Parameters, &out.Parameters @@ -938,12 +924,15 @@ func (in *KubernetesMasterConfig) DeepCopyInto(out *KubernetesMasterConfig) { in, out := &in.DisabledAPIGroupVersions, &out.DisabledAPIGroupVersions *out = make(map[string][]string, len(*in)) for key, val := range *in { + var outVal []string if val == nil { (*out)[key] = nil } else { - (*out)[key] = make([]string, len(val)) - copy((*out)[key], val) + in, out := &val, &outVal + *out = make([]string, len(*in)) + copy(*out, *in) } + (*out)[key] = outVal } } out.ProxyClientInfo = in.ProxyClientInfo @@ -951,36 +940,45 @@ func (in *KubernetesMasterConfig) DeepCopyInto(out *KubernetesMasterConfig) { in, out := &in.APIServerArguments, &out.APIServerArguments *out = make(ExtendedArguments, len(*in)) for key, val := range *in { + var outVal []string if val == nil { (*out)[key] = nil } else { - (*out)[key] = make([]string, len(val)) - copy((*out)[key], val) + in, out := &val, &outVal + *out = make([]string, len(*in)) + copy(*out, *in) } + (*out)[key] = outVal } } if in.ControllerArguments != nil { in, out := &in.ControllerArguments, &out.ControllerArguments *out = make(ExtendedArguments, len(*in)) for key, val := range *in { + var outVal []string if val == nil { (*out)[key] = nil } else { - (*out)[key] = make([]string, len(val)) - copy((*out)[key], val) + in, out := &val, &outVal + *out = make([]string, len(*in)) + copy(*out, *in) } + (*out)[key] = outVal } } if in.SchedulerArguments != nil { in, out := &in.SchedulerArguments, &out.SchedulerArguments *out = make(ExtendedArguments, len(*in)) for key, val := range *in { + var outVal []string if val == nil { (*out)[key] = nil } else { - (*out)[key] = make([]string, len(val)) - copy((*out)[key], val) + in, out := &val, &outVal + *out = make([]string, len(*in)) + copy(*out, *in) } + (*out)[key] = outVal } } return @@ -1089,30 +1087,18 @@ func (in *LDAPSyncConfig) DeepCopyInto(out *LDAPSyncConfig) { } if in.RFC2307Config != nil { in, out := &in.RFC2307Config, &out.RFC2307Config - if *in == nil { - *out = nil - } else { - *out = new(RFC2307Config) - (*in).DeepCopyInto(*out) - } + *out = new(RFC2307Config) + (*in).DeepCopyInto(*out) } if in.ActiveDirectoryConfig != nil { in, out := &in.ActiveDirectoryConfig, &out.ActiveDirectoryConfig - if *in == nil { - *out = nil - } else { - *out = new(ActiveDirectoryConfig) - (*in).DeepCopyInto(*out) - } + *out = new(ActiveDirectoryConfig) + (*in).DeepCopyInto(*out) } if in.AugmentedActiveDirectoryConfig != nil { in, out := &in.AugmentedActiveDirectoryConfig, &out.AugmentedActiveDirectoryConfig - if *in == nil { - *out = nil - } else { - *out = new(AugmentedActiveDirectoryConfig) - (*in).DeepCopyInto(*out) - } + *out = new(AugmentedActiveDirectoryConfig) + (*in).DeepCopyInto(*out) } return } @@ -1140,12 +1126,8 @@ func (in *LocalQuota) DeepCopyInto(out *LocalQuota) { *out = *in if in.PerFSGroup != nil { in, out := &in.PerFSGroup, &out.PerFSGroup - if *in == nil { - *out = nil - } else { - x := (*in).DeepCopy() - *out = &x - } + x := (*in).DeepCopy() + *out = &x } return } @@ -1165,12 +1147,8 @@ func (in *MasterAuthConfig) DeepCopyInto(out *MasterAuthConfig) { *out = *in if in.RequestHeader != nil { in, out := &in.RequestHeader, &out.RequestHeader - if *in == nil { - *out = nil - } else { - *out = new(RequestHeaderAuthenticationOptions) - (*in).DeepCopyInto(*out) - } + *out = new(RequestHeaderAuthenticationOptions) + (*in).DeepCopyInto(*out) } if in.WebhookTokenAuthenticators != nil { in, out := &in.WebhookTokenAuthenticators, &out.WebhookTokenAuthenticators @@ -1195,12 +1173,8 @@ func (in *MasterClients) DeepCopyInto(out *MasterClients) { *out = *in if in.OpenShiftLoopbackClientConnectionOverrides != nil { in, out := &in.OpenShiftLoopbackClientConnectionOverrides, &out.OpenShiftLoopbackClientConnectionOverrides - if *in == nil { - *out = nil - } else { - *out = new(ClientConnectionOverrides) - **out = **in - } + *out = new(ClientConnectionOverrides) + **out = **in } return } @@ -1240,30 +1214,18 @@ func (in *MasterConfig) DeepCopyInto(out *MasterConfig) { in.KubernetesMasterConfig.DeepCopyInto(&out.KubernetesMasterConfig) if in.EtcdConfig != nil { in, out := &in.EtcdConfig, &out.EtcdConfig - if *in == nil { - *out = nil - } else { - *out = new(EtcdConfig) - (*in).DeepCopyInto(*out) - } + *out = new(EtcdConfig) + (*in).DeepCopyInto(*out) } if in.OAuthConfig != nil { in, out := &in.OAuthConfig, &out.OAuthConfig - if *in == nil { - *out = nil - } else { - *out = new(OAuthConfig) - (*in).DeepCopyInto(*out) - } + *out = new(OAuthConfig) + (*in).DeepCopyInto(*out) } if in.DNSConfig != nil { in, out := &in.DNSConfig, &out.DNSConfig - if *in == nil { - *out = nil - } else { - *out = new(DNSConfig) - **out = **in - } + *out = new(DNSConfig) + **out = **in } in.ServiceAccountConfig.DeepCopyInto(&out.ServiceAccountConfig) in.MasterClients.DeepCopyInto(&out.MasterClients) @@ -1328,12 +1290,8 @@ func (in *MasterVolumeConfig) DeepCopyInto(out *MasterVolumeConfig) { *out = *in if in.DynamicProvisioningEnabled != nil { in, out := &in.DynamicProvisioningEnabled, &out.DynamicProvisioningEnabled - if *in == nil { - *out = nil - } else { - *out = new(bool) - **out = **in - } + *out = new(bool) + **out = **in } return } @@ -1393,12 +1351,8 @@ func (in *NodeConfig) DeepCopyInto(out *NodeConfig) { in.ServingInfo.DeepCopyInto(&out.ServingInfo) if in.MasterClientConnectionOverrides != nil { in, out := &in.MasterClientConnectionOverrides, &out.MasterClientConnectionOverrides - if *in == nil { - *out = nil - } else { - *out = new(ClientConnectionOverrides) - **out = **in - } + *out = new(ClientConnectionOverrides) + **out = **in } if in.DNSNameservers != nil { in, out := &in.DNSNameservers, &out.DNSNameservers @@ -1409,12 +1363,8 @@ func (in *NodeConfig) DeepCopyInto(out *NodeConfig) { out.ImageConfig = in.ImageConfig if in.PodManifestConfig != nil { in, out := &in.PodManifestConfig, &out.PodManifestConfig - if *in == nil { - *out = nil - } else { - *out = new(PodManifestConfig) - **out = **in - } + *out = new(PodManifestConfig) + **out = **in } out.AuthConfig = in.AuthConfig out.DockerConfig = in.DockerConfig @@ -1422,34 +1372,36 @@ func (in *NodeConfig) DeepCopyInto(out *NodeConfig) { in, out := &in.KubeletArguments, &out.KubeletArguments *out = make(ExtendedArguments, len(*in)) for key, val := range *in { + var outVal []string if val == nil { (*out)[key] = nil } else { - (*out)[key] = make([]string, len(val)) - copy((*out)[key], val) + in, out := &val, &outVal + *out = make([]string, len(*in)) + copy(*out, *in) } + (*out)[key] = outVal } } if in.ProxyArguments != nil { in, out := &in.ProxyArguments, &out.ProxyArguments *out = make(ExtendedArguments, len(*in)) for key, val := range *in { + var outVal []string if val == nil { (*out)[key] = nil } else { - (*out)[key] = make([]string, len(val)) - copy((*out)[key], val) + in, out := &val, &outVal + *out = make([]string, len(*in)) + copy(*out, *in) } + (*out)[key] = outVal } } if in.EnableUnidling != nil { in, out := &in.EnableUnidling, &out.EnableUnidling - if *in == nil { - *out = nil - } else { - *out = new(bool) - **out = **in - } + *out = new(bool) + **out = **in } in.VolumeConfig.DeepCopyInto(&out.VolumeConfig) return @@ -1511,12 +1463,8 @@ func (in *OAuthConfig) DeepCopyInto(out *OAuthConfig) { *out = *in if in.MasterCA != nil { in, out := &in.MasterCA, &out.MasterCA - if *in == nil { - *out = nil - } else { - *out = new(string) - **out = **in - } + *out = new(string) + **out = **in } if in.IdentityProviders != nil { in, out := &in.IdentityProviders, &out.IdentityProviders @@ -1528,22 +1476,14 @@ func (in *OAuthConfig) DeepCopyInto(out *OAuthConfig) { out.GrantConfig = in.GrantConfig if in.SessionConfig != nil { in, out := &in.SessionConfig, &out.SessionConfig - if *in == nil { - *out = nil - } else { - *out = new(SessionConfig) - **out = **in - } + *out = new(SessionConfig) + **out = **in } in.TokenConfig.DeepCopyInto(&out.TokenConfig) if in.Templates != nil { in, out := &in.Templates, &out.Templates - if *in == nil { - *out = nil - } else { - *out = new(OAuthTemplates) - **out = **in - } + *out = new(OAuthTemplates) + **out = **in } return } @@ -1704,12 +1644,8 @@ func (in *ProjectConfig) DeepCopyInto(out *ProjectConfig) { *out = *in if in.SecurityAllocator != nil { in, out := &in.SecurityAllocator, &out.SecurityAllocator - if *in == nil { - *out = nil - } else { - *out = new(SecurityAllocator) - **out = **in - } + *out = new(SecurityAllocator) + **out = **in } return } @@ -1939,12 +1875,8 @@ func (in *ServiceServingCert) DeepCopyInto(out *ServiceServingCert) { *out = *in if in.Signer != nil { in, out := &in.Signer, &out.Signer - if *in == nil { - *out = nil - } else { - *out = new(CertInfo) - **out = **in - } + *out = new(CertInfo) + **out = **in } return } @@ -2055,12 +1987,8 @@ func (in *SourceStrategyDefaultsConfig) DeepCopyInto(out *SourceStrategyDefaults *out = *in if in.Incremental != nil { in, out := &in.Incremental, &out.Incremental - if *in == nil { - *out = nil - } else { - *out = new(bool) - **out = **in - } + *out = new(bool) + **out = **in } return } @@ -2113,12 +2041,8 @@ func (in *TokenConfig) DeepCopyInto(out *TokenConfig) { *out = *in if in.AccessTokenInactivityTimeoutSeconds != nil { in, out := &in.AccessTokenInactivityTimeoutSeconds, &out.AccessTokenInactivityTimeoutSeconds - if *in == nil { - *out = nil - } else { - *out = new(int32) - **out = **in - } + *out = new(int32) + **out = **in } return } diff --git a/vendor/github.com/openshift/api/legacyconfig/v1/zz_generated.swagger_doc_generated.go b/vendor/github.com/openshift/api/legacyconfig/v1/zz_generated.swagger_doc_generated.go index 8c6488156c..75ee2a42b1 100644 --- a/vendor/github.com/openshift/api/legacyconfig/v1/zz_generated.swagger_doc_generated.go +++ b/vendor/github.com/openshift/api/legacyconfig/v1/zz_generated.swagger_doc_generated.go @@ -100,11 +100,11 @@ func (BasicAuthPasswordIdentityProvider) SwaggerDoc() map[string]string { } var map_BuildDefaultsConfig = map[string]string{ - "": "BuildDefaultsConfig controls the default information for Builds", - "gitHTTPProxy": "gitHTTPProxy is the location of the HTTPProxy for Git source", - "gitHTTPSProxy": "gitHTTPSProxy is the location of the HTTPSProxy for Git source", - "gitNoProxy": "gitNoProxy is the list of domains for which the proxy should not be used", - "env": "env is a set of default environment variables that will be applied to the build if the specified variables do not exist on the build", + "": "BuildDefaultsConfig controls the default information for Builds", + "gitHTTPProxy": "gitHTTPProxy is the location of the HTTPProxy for Git source", + "gitHTTPSProxy": "gitHTTPSProxy is the location of the HTTPSProxy for Git source", + "gitNoProxy": "gitNoProxy is the list of domains for which the proxy should not be used", + "env": "env is a set of default environment variables that will be applied to the build if the specified variables do not exist on the build", "sourceStrategyDefaults": "sourceStrategyDefaults are default values that apply to builds using the source strategy.", "imageLabels": "imageLabels is a list of labels that are applied to the resulting image. User can override a default label by providing a label with the same name in their Build/BuildConfig.", "nodeSelector": "nodeSelector is a selector which must be true for the build pod to fit on a node", @@ -246,7 +246,7 @@ func (EtcdConnectionInfo) SwaggerDoc() map[string]string { } var map_EtcdStorageConfig = map[string]string{ - "": "EtcdStorageConfig holds the necessary configuration options for the etcd storage underlying OpenShift and Kubernetes", + "": "EtcdStorageConfig holds the necessary configuration options for the etcd storage underlying OpenShift and Kubernetes", "kubernetesStorageVersion": "KubernetesStorageVersion is the API version that Kube resources in etcd should be serialized to. This value should *not* be advanced until all clients in the cluster that read from etcd have code that allows them to read the new version.", "kubernetesStoragePrefix": "KubernetesStoragePrefix is the path within etcd that the Kubernetes resources will be rooted under. This value, if changed, will mean existing objects in etcd will no longer be located. The default value is 'kubernetes.io'.", "openShiftStorageVersion": "OpenShiftStorageVersion is the API version that OS resources in etcd should be serialized to. This value should *not* be advanced until all clients in the cluster that read from etcd have code that allows them to read the new version.", @@ -325,7 +325,7 @@ func (HTPasswdPasswordIdentityProvider) SwaggerDoc() map[string]string { } var map_HTTPServingInfo = map[string]string{ - "": "HTTPServingInfo holds configuration for serving HTTP", + "": "HTTPServingInfo holds configuration for serving HTTP", "maxRequestsInFlight": "MaxRequestsInFlight is the number of concurrent requests allowed to the server. If zero, no limit.", "requestTimeoutSeconds": "RequestTimeoutSeconds is the number of seconds before requests are timed out. The default is 60 minutes, if -1 there is no limit on requests.", } @@ -358,9 +358,9 @@ func (ImageConfig) SwaggerDoc() map[string]string { } var map_ImagePolicyConfig = map[string]string{ - "": "ImagePolicyConfig holds the necessary configuration options for limits and behavior for importing images", - "maxImagesBulkImportedPerRepository": "MaxImagesBulkImportedPerRepository controls the number of images that are imported when a user does a bulk import of a container repository. This number defaults to 50 to prevent users from importing large numbers of images accidentally. Set -1 for no limit.", - "disableScheduledImport": "DisableScheduledImport allows scheduled background import of images to be disabled.", + "": "ImagePolicyConfig holds the necessary configuration options for limits and behavior for importing images", + "maxImagesBulkImportedPerRepository": "MaxImagesBulkImportedPerRepository controls the number of images that are imported when a user does a bulk import of a container repository. This number defaults to 50 to prevent users from importing large numbers of images accidentally. Set -1 for no limit.", + "disableScheduledImport": "DisableScheduledImport allows scheduled background import of images to be disabled.", "scheduledImageImportMinimumIntervalSeconds": "ScheduledImageImportMinimumIntervalSeconds is the minimum number of seconds that can elapse between when image streams scheduled for background import are checked against the upstream repository. The default value is 15 minutes.", "maxScheduledImageImportsPerMinute": "MaxScheduledImageImportsPerMinute is the maximum number of scheduled image streams that will be imported in the background per minute. The default value is 60. Set to -1 for unlimited.", "allowedRegistriesForImport": "AllowedRegistriesForImport limits the container image registries that normal users may import images from. Set this list to the registries that you trust to contain valid Docker images and that you want applications to be able to import from. Users with permission to create Images or ImageStreamMappings via the API are not affected by this policy - typically only administrators or system integrations will have those permissions.", @@ -374,7 +374,7 @@ func (ImagePolicyConfig) SwaggerDoc() map[string]string { } var map_JenkinsPipelineConfig = map[string]string{ - "": "JenkinsPipelineConfig holds configuration for the Jenkins pipeline strategy", + "": "JenkinsPipelineConfig holds configuration for the Jenkins pipeline strategy", "autoProvisionEnabled": "AutoProvisionEnabled determines whether a Jenkins server will be spawned from the provided template when the first build config in the project with type JenkinsPipeline is created. When not specified this option defaults to true.", "templateNamespace": "TemplateNamespace contains the namespace name where the Jenkins template is stored", "templateName": "TemplateName is the name of the default Jenkins template", @@ -467,12 +467,12 @@ func (LDAPQuery) SwaggerDoc() map[string]string { } var map_LDAPSyncConfig = map[string]string{ - "": "LDAPSyncConfig holds the necessary configuration options to define an LDAP group sync", - "url": "Host is the scheme, host and port of the LDAP server to connect to: scheme://host:port", - "bindDN": "BindDN is an optional DN to bind to the LDAP server with", - "bindPassword": "BindPassword is an optional password to bind with during the search phase.", - "insecure": "Insecure, if true, indicates the connection should not use TLS. Cannot be set to true with a URL scheme of \"ldaps://\" If false, \"ldaps://\" URLs connect using TLS, and \"ldap://\" URLs are upgraded to a TLS connection using StartTLS as specified in https://tools.ietf.org/html/rfc2830", - "ca": "CA is the optional trusted certificate authority bundle to use when making requests to the server If empty, the default system roots are used", + "": "LDAPSyncConfig holds the necessary configuration options to define an LDAP group sync", + "url": "Host is the scheme, host and port of the LDAP server to connect to: scheme://host:port", + "bindDN": "BindDN is an optional DN to bind to the LDAP server with", + "bindPassword": "BindPassword is an optional password to bind with during the search phase.", + "insecure": "Insecure, if true, indicates the connection should not use TLS. Cannot be set to true with a URL scheme of \"ldaps://\" If false, \"ldaps://\" URLs connect using TLS, and \"ldap://\" URLs are upgraded to a TLS connection using StartTLS as specified in https://tools.ietf.org/html/rfc2830", + "ca": "CA is the optional trusted certificate authority bundle to use when making requests to the server If empty, the default system roots are used", "groupUIDNameMapping": "LDAPGroupUIDToOpenShiftGroupNameMapping is an optional direct mapping of LDAP group UIDs to OpenShift Group names", "rfc2307": "RFC2307Config holds the configuration for extracting data from an LDAP server set up in a fashion similar to RFC2307: first-class group and user entries, with group membership determined by a multi-valued attribute on the group entry listing its members", "activeDirectory": "ActiveDirectoryConfig holds the configuration for extracting data from an LDAP server set up in a fashion similar to that used in Active Directory: first-class user entries, with group membership determined by a multi-valued attribute on members listing groups they are a member of", @@ -504,8 +504,8 @@ func (MasterAuthConfig) SwaggerDoc() map[string]string { } var map_MasterClients = map[string]string{ - "": "MasterClients holds references to `.kubeconfig` files that qualify master clients for OpenShift and Kubernetes", - "openshiftLoopbackKubeConfig": "OpenShiftLoopbackKubeConfig is a .kubeconfig filename for system components to loopback to this master", + "": "MasterClients holds references to `.kubeconfig` files that qualify master clients for OpenShift and Kubernetes", + "openshiftLoopbackKubeConfig": "OpenShiftLoopbackKubeConfig is a .kubeconfig filename for system components to loopback to this master", "openshiftLoopbackClientConnectionOverrides": "OpenShiftLoopbackClientConnectionOverrides specifies client overrides for system components to loop back to this master.", } @@ -565,7 +565,7 @@ func (MasterNetworkConfig) SwaggerDoc() map[string]string { } var map_MasterVolumeConfig = map[string]string{ - "": "MasterVolumeConfig contains options for configuring volume plugins in the master node.", + "": "MasterVolumeConfig contains options for configuring volume plugins in the master node.", "dynamicProvisioningEnabled": "DynamicProvisioningEnabled is a boolean that toggles dynamic provisioning off when false, defaults to true", } @@ -583,7 +583,7 @@ func (NamedCertificate) SwaggerDoc() map[string]string { } var map_NodeAuthConfig = map[string]string{ - "": "NodeAuthConfig holds authn/authz configuration options", + "": "NodeAuthConfig holds authn/authz configuration options", "authenticationCacheTTL": "AuthenticationCacheTTL indicates how long an authentication result should be cached. It takes a valid time duration string (e.g. \"5m\"). If empty, you get the default timeout. If zero (e.g. \"0m\"), caching is disabled", "authenticationCacheSize": "AuthenticationCacheSize indicates how many authentication results should be cached. If 0, the default cache size is used.", "authorizationCacheTTL": "AuthorizationCacheTTL indicates how long an authorization result should be cached. It takes a valid time duration string (e.g. \"5m\"). If empty, you get the default timeout. If zero (e.g. \"0m\"), caching is disabled", @@ -692,8 +692,8 @@ var map_OpenIDIdentityProvider = map[string]string{ "clientSecret": "ClientSecret is the oauth client secret", "extraScopes": "ExtraScopes are any scopes to request in addition to the standard \"openid\" scope.", "extraAuthorizeParameters": "ExtraAuthorizeParameters are any custom parameters to add to the authorize request.", - "urls": "URLs to use to authenticate", - "claims": "Claims mappings", + "urls": "URLs to use to authenticate", + "claims": "Claims mappings", } func (OpenIDIdentityProvider) SwaggerDoc() map[string]string { @@ -712,8 +712,8 @@ func (OpenIDURLs) SwaggerDoc() map[string]string { } var map_PodManifestConfig = map[string]string{ - "": "PodManifestConfig holds the necessary configuration options for using pod manifests", - "path": "Path specifies the path for the pod manifest file or directory If its a directory, its expected to contain on or more manifest files This is used by the Kubelet to create pods on the node", + "": "PodManifestConfig holds the necessary configuration options for using pod manifests", + "path": "Path specifies the path for the pod manifest file or directory If its a directory, its expected to contain on or more manifest files This is used by the Kubelet to create pods on the node", "fileCheckIntervalSeconds": "FileCheckIntervalSeconds is the interval in seconds for checking the manifest file(s) for new data The interval needs to be a positive value", } @@ -722,7 +722,7 @@ func (PodManifestConfig) SwaggerDoc() map[string]string { } var map_PolicyConfig = map[string]string{ - "": "\n holds the necessary configuration options for", + "": "\n holds the necessary configuration options for", "userAgentMatchingConfig": "UserAgentMatchingConfig controls how API calls from *voluntarily* identifying clients will be handled. THIS DOES NOT DEFEND AGAINST MALICIOUS CLIENTS!", } @@ -731,7 +731,7 @@ func (PolicyConfig) SwaggerDoc() map[string]string { } var map_ProjectConfig = map[string]string{ - "": "\n holds the necessary configuration options for", + "": "\n holds the necessary configuration options for", "defaultNodeSelector": "DefaultNodeSelector holds default project node label selector", "projectRequestMessage": "ProjectRequestMessage is the string presented to a user if they are unable to request a project via the projectrequest api endpoint", "projectRequestTemplate": "ProjectRequestTemplate is the template to use for creating projects in response to projectrequest. It is in the format namespace/template and it is optional. If it is not specified, a default template is used.", @@ -924,7 +924,7 @@ func (StringSourceSpec) SwaggerDoc() map[string]string { } var map_TokenConfig = map[string]string{ - "": "TokenConfig holds the necessary configuration options for authorization and access tokens", + "": "TokenConfig holds the necessary configuration options for authorization and access tokens", "authorizeTokenMaxAgeSeconds": "AuthorizeTokenMaxAgeSeconds defines the maximum age of authorize tokens", "accessTokenMaxAgeSeconds": "AccessTokenMaxAgeSeconds defines the maximum age of access tokens", "accessTokenInactivityTimeoutSeconds": "AccessTokenInactivityTimeoutSeconds defined the default token inactivity timeout for tokens granted by any client. Setting it to nil means the feature is completely disabled (default) The default setting can be overriden on OAuthClient basis. The value represents the maximum amount of time that can occur between consecutive uses of the token. Tokens become invalid if they are not used within this temporal window. The user will need to acquire a new token to regain access once a token times out. Valid values are: - 0: Tokens never time out - X: Tokens time out if there is no activity for X seconds The current minimum allowed value for X is 300 (5 minutes)", diff --git a/vendor/github.com/openshift/api/network/v1/generated.pb.go b/vendor/github.com/openshift/api/network/v1/generated.pb.go index 942dec6f16..4350b10a56 100644 --- a/vendor/github.com/openshift/api/network/v1/generated.pb.go +++ b/vendor/github.com/openshift/api/network/v1/generated.pb.go @@ -1,6 +1,5 @@ -// Code generated by protoc-gen-gogo. +// Code generated by protoc-gen-gogo. DO NOT EDIT. // source: github.com/openshift/api/network/v1/generated.proto -// DO NOT EDIT! /* Package v1 is a generated protocol buffer package. @@ -577,24 +576,6 @@ func (m *NetNamespaceList) MarshalTo(dAtA []byte) (int, error) { return i, nil } -func encodeFixed64Generated(dAtA []byte, offset int, v uint64) int { - dAtA[offset] = uint8(v) - dAtA[offset+1] = uint8(v >> 8) - dAtA[offset+2] = uint8(v >> 16) - dAtA[offset+3] = uint8(v >> 24) - dAtA[offset+4] = uint8(v >> 32) - dAtA[offset+5] = uint8(v >> 40) - dAtA[offset+6] = uint8(v >> 48) - dAtA[offset+7] = uint8(v >> 56) - return offset + 8 -} -func encodeFixed32Generated(dAtA []byte, offset int, v uint32) int { - dAtA[offset] = uint8(v) - dAtA[offset+1] = uint8(v >> 8) - dAtA[offset+2] = uint8(v >> 16) - dAtA[offset+3] = uint8(v >> 24) - return offset + 4 -} func encodeVarintGenerated(dAtA []byte, offset int, v uint64) int { for v >= 1<<7 { dAtA[offset] = uint8(v&0x7f | 0x80) diff --git a/vendor/github.com/openshift/api/network/v1/zz_generated.deepcopy.go b/vendor/github.com/openshift/api/network/v1/zz_generated.deepcopy.go index e445797967..1f55d84627 100644 --- a/vendor/github.com/openshift/api/network/v1/zz_generated.deepcopy.go +++ b/vendor/github.com/openshift/api/network/v1/zz_generated.deepcopy.go @@ -20,12 +20,8 @@ func (in *ClusterNetwork) DeepCopyInto(out *ClusterNetwork) { } if in.VXLANPort != nil { in, out := &in.VXLANPort, &out.VXLANPort - if *in == nil { - *out = nil - } else { - *out = new(uint32) - **out = **in - } + *out = new(uint32) + **out = **in } return } diff --git a/vendor/github.com/openshift/api/oauth/v1/generated.pb.go b/vendor/github.com/openshift/api/oauth/v1/generated.pb.go index 0964309c08..0a4f507f12 100644 --- a/vendor/github.com/openshift/api/oauth/v1/generated.pb.go +++ b/vendor/github.com/openshift/api/oauth/v1/generated.pb.go @@ -1,6 +1,5 @@ -// Code generated by protoc-gen-gogo. +// Code generated by protoc-gen-gogo. DO NOT EDIT. // source: github.com/openshift/api/oauth/v1/generated.proto -// DO NOT EDIT! /* Package v1 is a generated protocol buffer package. @@ -717,24 +716,6 @@ func (m *ScopeRestriction) MarshalTo(dAtA []byte) (int, error) { return i, nil } -func encodeFixed64Generated(dAtA []byte, offset int, v uint64) int { - dAtA[offset] = uint8(v) - dAtA[offset+1] = uint8(v >> 8) - dAtA[offset+2] = uint8(v >> 16) - dAtA[offset+3] = uint8(v >> 24) - dAtA[offset+4] = uint8(v >> 32) - dAtA[offset+5] = uint8(v >> 40) - dAtA[offset+6] = uint8(v >> 48) - dAtA[offset+7] = uint8(v >> 56) - return offset + 8 -} -func encodeFixed32Generated(dAtA []byte, offset int, v uint32) int { - dAtA[offset] = uint8(v) - dAtA[offset+1] = uint8(v >> 8) - dAtA[offset+2] = uint8(v >> 16) - dAtA[offset+3] = uint8(v >> 24) - return offset + 4 -} func encodeVarintGenerated(dAtA []byte, offset int, v uint64) int { for v >= 1<<7 { dAtA[offset] = uint8(v&0x7f | 0x80) diff --git a/vendor/github.com/openshift/api/oauth/v1/zz_generated.deepcopy.go b/vendor/github.com/openshift/api/oauth/v1/zz_generated.deepcopy.go index d4eed458b3..ccafa19467 100644 --- a/vendor/github.com/openshift/api/oauth/v1/zz_generated.deepcopy.go +++ b/vendor/github.com/openshift/api/oauth/v1/zz_generated.deepcopy.go @@ -186,21 +186,13 @@ func (in *OAuthClient) DeepCopyInto(out *OAuthClient) { } if in.AccessTokenMaxAgeSeconds != nil { in, out := &in.AccessTokenMaxAgeSeconds, &out.AccessTokenMaxAgeSeconds - if *in == nil { - *out = nil - } else { - *out = new(int32) - **out = **in - } + *out = new(int32) + **out = **in } if in.AccessTokenInactivityTimeoutSeconds != nil { in, out := &in.AccessTokenInactivityTimeoutSeconds, &out.AccessTokenInactivityTimeoutSeconds - if *in == nil { - *out = nil - } else { - *out = new(int32) - **out = **in - } + *out = new(int32) + **out = **in } return } @@ -373,12 +365,8 @@ func (in *ScopeRestriction) DeepCopyInto(out *ScopeRestriction) { } if in.ClusterRole != nil { in, out := &in.ClusterRole, &out.ClusterRole - if *in == nil { - *out = nil - } else { - *out = new(ClusterRoleScopeRestriction) - (*in).DeepCopyInto(*out) - } + *out = new(ClusterRoleScopeRestriction) + (*in).DeepCopyInto(*out) } return } diff --git a/vendor/github.com/openshift/api/openshiftcontrolplane/v1/zz_generated.deepcopy.go b/vendor/github.com/openshift/api/openshiftcontrolplane/v1/zz_generated.deepcopy.go index a6dad208d5..9e8e2011e5 100644 --- a/vendor/github.com/openshift/api/openshiftcontrolplane/v1/zz_generated.deepcopy.go +++ b/vendor/github.com/openshift/api/openshiftcontrolplane/v1/zz_generated.deepcopy.go @@ -5,9 +5,9 @@ package v1 import ( - build_v1 "github.com/openshift/api/build/v1" - config_v1 "github.com/openshift/api/config/v1" - core_v1 "k8s.io/api/core/v1" + buildv1 "github.com/openshift/api/build/v1" + configv1 "github.com/openshift/api/config/v1" + corev1 "k8s.io/api/core/v1" runtime "k8s.io/apimachinery/pkg/runtime" ) @@ -37,21 +37,13 @@ func (in *BuildControllerConfig) DeepCopyInto(out *BuildControllerConfig) { out.ImageTemplateFormat = in.ImageTemplateFormat if in.BuildDefaults != nil { in, out := &in.BuildDefaults, &out.BuildDefaults - if *in == nil { - *out = nil - } else { - *out = new(BuildDefaultsConfig) - (*in).DeepCopyInto(*out) - } + *out = new(BuildDefaultsConfig) + (*in).DeepCopyInto(*out) } if in.BuildOverrides != nil { in, out := &in.BuildOverrides, &out.BuildOverrides - if *in == nil { - *out = nil - } else { - *out = new(BuildOverridesConfig) - (*in).DeepCopyInto(*out) - } + *out = new(BuildOverridesConfig) + (*in).DeepCopyInto(*out) } return } @@ -72,23 +64,19 @@ func (in *BuildDefaultsConfig) DeepCopyInto(out *BuildDefaultsConfig) { out.TypeMeta = in.TypeMeta if in.Env != nil { in, out := &in.Env, &out.Env - *out = make([]core_v1.EnvVar, len(*in)) + *out = make([]corev1.EnvVar, len(*in)) for i := range *in { (*in)[i].DeepCopyInto(&(*out)[i]) } } if in.SourceStrategyDefaults != nil { in, out := &in.SourceStrategyDefaults, &out.SourceStrategyDefaults - if *in == nil { - *out = nil - } else { - *out = new(SourceStrategyDefaultsConfig) - (*in).DeepCopyInto(*out) - } + *out = new(SourceStrategyDefaultsConfig) + (*in).DeepCopyInto(*out) } if in.ImageLabels != nil { in, out := &in.ImageLabels, &out.ImageLabels - *out = make([]build_v1.ImageLabel, len(*in)) + *out = make([]buildv1.ImageLabel, len(*in)) copy(*out, *in) } if in.NodeSelector != nil { @@ -133,7 +121,7 @@ func (in *BuildOverridesConfig) DeepCopyInto(out *BuildOverridesConfig) { out.TypeMeta = in.TypeMeta if in.ImageLabels != nil { in, out := &in.ImageLabels, &out.ImageLabels - *out = make([]build_v1.ImageLabel, len(*in)) + *out = make([]buildv1.ImageLabel, len(*in)) copy(*out, *in) } if in.NodeSelector != nil { @@ -152,7 +140,7 @@ func (in *BuildOverridesConfig) DeepCopyInto(out *BuildOverridesConfig) { } if in.Tolerations != nil { in, out := &in.Tolerations, &out.Tolerations - *out = make([]core_v1.Toleration, len(*in)) + *out = make([]corev1.Toleration, len(*in)) for i := range *in { (*in)[i].DeepCopyInto(&(*out)[i]) } @@ -347,12 +335,8 @@ func (in *JenkinsPipelineConfig) DeepCopyInto(out *JenkinsPipelineConfig) { *out = *in if in.AutoProvisionEnabled != nil { in, out := &in.AutoProvisionEnabled, &out.AutoProvisionEnabled - if *in == nil { - *out = nil - } else { - *out = new(bool) - **out = **in - } + *out = new(bool) + **out = **in } if in.Parameters != nil { in, out := &in.Parameters, &out.Parameters @@ -409,12 +393,15 @@ func (in *OpenShiftAPIServerConfig) DeepCopyInto(out *OpenShiftAPIServerConfig) in, out := &in.APIServerArguments, &out.APIServerArguments *out = make(map[string][]string, len(*in)) for key, val := range *in { + var outVal []string if val == nil { (*out)[key] = nil } else { - (*out)[key] = make([]string, len(val)) - copy((*out)[key], val) + in, out := &val, &outVal + *out = make([]string, len(*in)) + copy(*out, *in) } + (*out)[key] = outVal } } return @@ -445,12 +432,8 @@ func (in *OpenShiftControllerManagerConfig) DeepCopyInto(out *OpenShiftControlle out.KubeClientConfig = in.KubeClientConfig if in.ServingInfo != nil { in, out := &in.ServingInfo, &out.ServingInfo - if *in == nil { - *out = nil - } else { - *out = new(config_v1.HTTPServingInfo) - (*in).DeepCopyInto(*out) - } + *out = new(configv1.HTTPServingInfo) + (*in).DeepCopyInto(*out) } out.LeaderElection = in.LeaderElection if in.Controllers != nil { @@ -597,12 +580,8 @@ func (in *ServiceServingCert) DeepCopyInto(out *ServiceServingCert) { *out = *in if in.Signer != nil { in, out := &in.Signer, &out.Signer - if *in == nil { - *out = nil - } else { - *out = new(config_v1.CertInfo) - **out = **in - } + *out = new(configv1.CertInfo) + **out = **in } return } @@ -622,12 +601,8 @@ func (in *SourceStrategyDefaultsConfig) DeepCopyInto(out *SourceStrategyDefaults *out = *in if in.Incremental != nil { in, out := &in.Incremental, &out.Incremental - if *in == nil { - *out = nil - } else { - *out = new(bool) - **out = **in - } + *out = new(bool) + **out = **in } return } diff --git a/vendor/github.com/openshift/api/openshiftcontrolplane/v1/zz_generated.swagger_doc_generated.go b/vendor/github.com/openshift/api/openshiftcontrolplane/v1/zz_generated.swagger_doc_generated.go index 0cd9f1dade..9547ae2bea 100644 --- a/vendor/github.com/openshift/api/openshiftcontrolplane/v1/zz_generated.swagger_doc_generated.go +++ b/vendor/github.com/openshift/api/openshiftcontrolplane/v1/zz_generated.swagger_doc_generated.go @@ -20,11 +20,11 @@ func (BuildControllerConfig) SwaggerDoc() map[string]string { } var map_BuildDefaultsConfig = map[string]string{ - "": "BuildDefaultsConfig controls the default information for Builds", - "gitHTTPProxy": "gitHTTPProxy is the location of the HTTPProxy for Git source", - "gitHTTPSProxy": "gitHTTPSProxy is the location of the HTTPSProxy for Git source", - "gitNoProxy": "gitNoProxy is the list of domains for which the proxy should not be used", - "env": "env is a set of default environment variables that will be applied to the build if the specified variables do not exist on the build", + "": "BuildDefaultsConfig controls the default information for Builds", + "gitHTTPProxy": "gitHTTPProxy is the location of the HTTPProxy for Git source", + "gitHTTPSProxy": "gitHTTPSProxy is the location of the HTTPSProxy for Git source", + "gitNoProxy": "gitNoProxy is the list of domains for which the proxy should not be used", + "env": "env is a set of default environment variables that will be applied to the build if the specified variables do not exist on the build", "sourceStrategyDefaults": "sourceStrategyDefaults are default values that apply to builds using the source strategy.", "imageLabels": "imageLabels is a list of labels that are applied to the resulting image. User can override a default label by providing a label with the same name in their Build/BuildConfig.", "nodeSelector": "nodeSelector is a selector which must be true for the build pod to fit on a node", @@ -121,7 +121,7 @@ func (IngressControllerConfig) SwaggerDoc() map[string]string { } var map_JenkinsPipelineConfig = map[string]string{ - "": "JenkinsPipelineConfig holds configuration for the Jenkins pipeline strategy", + "": "JenkinsPipelineConfig holds configuration for the Jenkins pipeline strategy", "autoProvisionEnabled": "autoProvisionEnabled determines whether a Jenkins server will be spawned from the provided template when the first build config in the project with type JenkinsPipeline is created. When not specified this option defaults to true.", "templateNamespace": "templateNamespace contains the namespace name where the Jenkins template is stored", "templateName": "templateName is the name of the default Jenkins template", diff --git a/vendor/github.com/openshift/api/operator/v1/register.go b/vendor/github.com/openshift/api/operator/v1/register.go index 1d4a93399d..423a41a9af 100644 --- a/vendor/github.com/openshift/api/operator/v1/register.go +++ b/vendor/github.com/openshift/api/operator/v1/register.go @@ -34,6 +34,8 @@ func addKnownTypes(scheme *runtime.Scheme) error { scheme.AddKnownTypes(GroupVersion, &Authentication{}, &AuthenticationList{}, + &DNS{}, + &DNSList{}, &Console{}, &ConsoleList{}, &Etcd{}, diff --git a/vendor/github.com/openshift/api/operator/v1/types.go b/vendor/github.com/openshift/api/operator/v1/types.go index 1e7509fca7..2a4fe24d84 100644 --- a/vendor/github.com/openshift/api/operator/v1/types.go +++ b/vendor/github.com/openshift/api/operator/v1/types.go @@ -1,9 +1,8 @@ package v1 import ( - corev1 "k8s.io/api/core/v1" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" - runtime "k8s.io/apimachinery/pkg/runtime" + "k8s.io/apimachinery/pkg/runtime" ) // MyOperatorResource is an example operator configuration type @@ -46,7 +45,7 @@ var ( // inside of the Spec struct for your particular operator. type OperatorSpec struct { // managementState indicates whether and how the operator should manage the component - // +kubebuilder:validation:Pattern=^Managed|Unmanaged|Force|Removed$ + // +kubebuilder:validation:Pattern=^(Managed|Unmanaged|Force|Removed)$ ManagementState ManagementState `json:"managementState"` // logLevel is an intent based logging for an overall component. It does not give fine grained control, but it is a @@ -54,9 +53,10 @@ type OperatorSpec struct { // +optional LogLevel LogLevel `json:"logLevel"` - // operandSpecs provide customization for functional units within the component + // operatorLogLevel is an intent based logging for the operator itself. It does not give fine grained control, but it is a + // simple way to manage coarse grained logging choices that operators have to interpret for themselves. // +optional - OperandSpecs []OperandSpec `json:"operandSpecs,omitempty"` + OperatorLogLevel LogLevel `json:"operatorLogLevel"` // unsupportedConfigOverrides holds a sparse config that will override any previously set options. It only needs to be the fields to override // it will end up overlaying in the following order: @@ -91,54 +91,25 @@ var ( TraceAll LogLevel = "TraceAll" ) -// ResourcePatch is a way to represent the patch you would issue to `kubectl patch` in the API -type ResourcePatch struct { - // type is the type of patch to apply: jsonmerge, strategicmerge - Type string `json:"type"` - // patch the patch itself - Patch string `json:"patch"` -} - -// OperandSpec holds information for customization of a particular functional unit - logically maps to a workload -type OperandSpec struct { - // name is the name of this unit. The operator must be aware of it. - Name string `json:"name"` - - // operandContainerSpecs are per-container options - // +optional - OperandContainerSpecs []OperandContainerSpec `json:"operandContainerSpecs,omitempty"` - - // unsupportedResourcePatches are applied to the workload resource for this unit. This is an unsupported - // workaround if anything needs to be modified on the workload that is not otherwise configurable. - // TODO Decide: alternatively, we could simply include a RawExtension which is used in place of the "normal" default manifest - // +optional - UnsupportedResourcePatches []ResourcePatch `json:"unsupportedResourcePatches,omitempty"` -} - -type OperandContainerSpec struct { - // name is the name of the container to modify - Name string `json:"name"` - - // resources are the requests and limits to place in the container. Nil means to accept the defaults. - Resources *corev1.ResourceRequirements `json:"resources,omitempty"` -} - type OperatorStatus struct { // observedGeneration is the last generation change you've dealt with // +optional ObservedGeneration int64 `json:"observedGeneration,omitempty"` // conditions is a list of conditions and their status + // +optional Conditions []OperatorCondition `json:"conditions,omitempty"` // version is the level this availability applies to - Version string `json:"version"` + // +optional + Version string `json:"version,omitempty"` // readyReplicas indicates how many replicas are ready and at the desired state ReadyReplicas int32 `json:"readyReplicas"` // generations are used to determine when an item needs to be reconciled or has changed in a way that needs a reaction. - Generations []GenerationStatus `json:"generations"` + // +optional + Generations []GenerationStatus `json:"generations,omitempty"` } // GenerationStatus keeps track of the generation for a given resource so that decisions about forced updates can be made. @@ -164,6 +135,8 @@ var ( OperatorStatusTypeProgressing = "Progressing" // Failing indicates that the operator (not the operand) is unable to fulfill the user intent OperatorStatusTypeFailing = "Failing" + // Degraded indicates that the operator (not the operand) is unable to fulfill the user intent + OperatorStatusTypeDegraded = "Degraded" // PrereqsSatisfied indicates that the things this operator depends on are present and at levels compatible with the // current and desired states. OperatorStatusTypePrereqsSatisfied = "PrereqsSatisfied" @@ -192,6 +165,11 @@ const ( type StaticPodOperatorSpec struct { OperatorSpec `json:",inline"` + // forceRedeploymentReason can be used to force the redeployment of the operand by providing a unique string. + // This provides a mechanism to kick a previously failed deployment and provide a reason why you think it will work + // this time instead of failing again on the same config. + ForceRedeploymentReason string `json:"forceRedeploymentReason"` + // failedRevisionLimit is the number of failed static pod installer revisions to keep on disk and in the api // -1 = unlimited, 0 or unset = 5 (default) FailedRevisionLimit int32 `json:"failedRevisionLimit,omitempty"` @@ -206,9 +184,11 @@ type StaticPodOperatorStatus struct { OperatorStatus `json:",inline"` // latestAvailableRevision is the deploymentID of the most recent deployment - LatestAvailableRevision int32 `json:"latestAvailableRevision"` + // +optional + LatestAvailableRevision int32 `json:"latestAvailableRevision,omitEmpty"` // nodeStatuses track the deployment values and errors across individual nodes + // +optional NodeStatuses []NodeStatus `json:"nodeStatuses,omitempty"` } diff --git a/vendor/github.com/openshift/api/operator/v1/types_authentication.go b/vendor/github.com/openshift/api/operator/v1/types_authentication.go index d697cf8f13..1dee7ca27a 100644 --- a/vendor/github.com/openshift/api/operator/v1/types_authentication.go +++ b/vendor/github.com/openshift/api/operator/v1/types_authentication.go @@ -14,7 +14,7 @@ type Authentication struct { metav1.ObjectMeta `json:"metadata,omitempty"` // +required - Spec AuthenticationSpec `json:"spec,omitempty"` + Spec AuthenticationSpec `json:"spec,omitempty"` // +optional Status AuthenticationStatus `json:"status,omitempty"` } diff --git a/vendor/github.com/openshift/api/operator/v1/types_console.go b/vendor/github.com/openshift/api/operator/v1/types_console.go index 3f3ee26a3a..92e74e5644 100644 --- a/vendor/github.com/openshift/api/operator/v1/types_console.go +++ b/vendor/github.com/openshift/api/operator/v1/types_console.go @@ -11,7 +11,7 @@ type Console struct { metav1.ObjectMeta `json:"metadata,omitempty"` // +required - Spec ConsoleSpec `json:"spec,omitempty"` + Spec ConsoleSpec `json:"spec,omitempty"` // +optional Status ConsoleStatus `json:"status,omitempty"` } @@ -48,13 +48,15 @@ const ( // Branding for OpenShift BrandOpenShift Brand = "openshift" // Branding for The Origin Community Distribution of Kubernetes - BrandOKD Brand = "okd" + BrandOKD Brand = "okd" // Branding for OpenShift Online - BrandOnline Brand = "online" + BrandOnline Brand = "online" // Branding for OpenShift Container Platform - BrandOCP Brand = "ocp" + BrandOCP Brand = "ocp" // Branding for OpenShift Dedicated BrandDedicated Brand = "dedicated" + // Branding for Azure Red Hat OpenShift + BrandAzure Brand = "azure" ) // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object diff --git a/vendor/github.com/openshift/api/operator/v1/types_dns.go b/vendor/github.com/openshift/api/operator/v1/types_dns.go new file mode 100644 index 0000000000..1638090b4c --- /dev/null +++ b/vendor/github.com/openshift/api/operator/v1/types_dns.go @@ -0,0 +1,82 @@ +package v1 + +import ( + metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" +) + +// +genclient +// +genclient:nonNamespaced +// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object + +// DNS manages the CoreDNS component to provide a name resolution service +// for pods and services in the cluster. +// +// This supports the DNS-based service discovery specification: +// https://github.com/kubernetes/dns/blob/master/docs/specification.md +// +// More details: https://kubernetes.io/docs/tasks/administer-cluster/coredns +type DNS struct { + metav1.TypeMeta `json:",inline"` + metav1.ObjectMeta `json:"metadata,omitempty"` + + // spec is the specification of the desired behavior of the DNS. + Spec DNSSpec `json:"spec,omitempty"` + // status is the most recently observed status of the DNS. + Status DNSStatus `json:"status,omitempty"` +} + +// DNSSpec is the specification of the desired behavior of the DNS. +type DNSSpec struct { +} + +const ( + // Available indicates the DNS controller daemonset is available. + DNSAvailable = "Available" +) + +// DNSStatus defines the observed status of the DNS. +type DNSStatus struct { + // clusterIP is the service IP through which this DNS is made available. + // + // In the case of the default DNS, this will be a well known IP that is used + // as the default nameserver for pods that are using the default ClusterFirst DNS policy. + // + // In general, this IP can be specified in a pod's spec.dnsConfig.nameservers list + // or used explicitly when performing name resolution from within the cluster. + // Example: dig foo.com @ + // + // More info: https://kubernetes.io/docs/concepts/services-networking/service/#virtual-ips-and-service-proxies + ClusterIP string `json:"clusterIP"` + + // clusterDomain is the local cluster DNS domain suffix for DNS services. + // This will be a subdomain as defined in RFC 1034, + // section 3.5: https://tools.ietf.org/html/rfc1034#section-3.5 + // Example: "cluster.local" + // + // More info: https://kubernetes.io/docs/concepts/services-networking/dns-pod-service + ClusterDomain string `json:"clusterDomain"` + + // conditions provide information about the state of the DNS on the cluster. + // + // These are the supported DNS conditions: + // + // * Available + // - True if the following conditions are met: + // * DNS controller daemonset is available. + // - False if any of those conditions are unsatisfied. + // + // +patchMergeKey=type + // +patchStrategy=merge + // +optional + Conditions []OperatorCondition `json:"conditions,omitempty" patchStrategy:"merge" patchMergeKey:"type"` +} + +// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object + +// DNSList contains a list of DNS +type DNSList struct { + metav1.TypeMeta `json:",inline"` + metav1.ListMeta `json:"metadata,omitempty"` + + Items []DNS `json:"items"` +} diff --git a/vendor/github.com/openshift/api/operator/v1/types_etcd.go b/vendor/github.com/openshift/api/operator/v1/types_etcd.go index a3135120fa..b1e1adda68 100644 --- a/vendor/github.com/openshift/api/operator/v1/types_etcd.go +++ b/vendor/github.com/openshift/api/operator/v1/types_etcd.go @@ -14,7 +14,7 @@ type Etcd struct { metav1.ObjectMeta `json:"metadata"` // +required - Spec EtcdSpec `json:"spec"` + Spec EtcdSpec `json:"spec"` // +optional Status EtcdStatus `json:"status"` } diff --git a/vendor/github.com/openshift/api/operator/v1/types_ingress.go b/vendor/github.com/openshift/api/operator/v1/types_ingress.go index 5562d20a7d..4e44b0116a 100644 --- a/vendor/github.com/openshift/api/operator/v1/types_ingress.go +++ b/vendor/github.com/openshift/api/operator/v1/types_ingress.go @@ -67,8 +67,10 @@ type IngressControllerSpec struct { // If unset, the default is based on // infrastructure.config.openshift.io/cluster .status.platform: // - // AWS: LoadBalancerService - // All other platform types: Private + // AWS: LoadBalancerService + // Libvirt: HostNetwork + // + // Any other platform types (including None) default to HostNetwork. // // endpointPublishingStrategy cannot be updated. // @@ -135,6 +137,16 @@ type NodePlacement struct { // // +optional NodeSelector *metav1.LabelSelector `json:"nodeSelector,omitempty"` + + // tolerations is a list of tolerations applied to ingress controller + // deployments. + // + // The default is an empty list. + // + // See https://kubernetes.io/docs/concepts/configuration/taint-and-toleration/ + // + // +optional + Tolerations []corev1.Toleration `json:"tolerations,omitempty"` } // EndpointPublishingStrategyType is a way to publish ingress controller endpoints. diff --git a/vendor/github.com/openshift/api/operator/v1/types_kubeapiserver.go b/vendor/github.com/openshift/api/operator/v1/types_kubeapiserver.go index d0d14f3cc5..6b79a8e0c6 100644 --- a/vendor/github.com/openshift/api/operator/v1/types_kubeapiserver.go +++ b/vendor/github.com/openshift/api/operator/v1/types_kubeapiserver.go @@ -14,18 +14,13 @@ type KubeAPIServer struct { metav1.ObjectMeta `json:"metadata"` // +required - Spec KubeAPIServerSpec `json:"spec"` + Spec KubeAPIServerSpec `json:"spec"` // +optional Status KubeAPIServerStatus `json:"status"` } type KubeAPIServerSpec struct { StaticPodOperatorSpec `json:",inline"` - - // forceRedeploymentReason can be used to force the redeployment of the kube-apiserver by providing a unique string. - // This provides a mechanism to kick a previously failed deployment and provide a reason why you think it will work - // this time instead of failing again on the same config. - ForceRedeploymentReason string `json:"forceRedeploymentReason"` } type KubeAPIServerStatus struct { diff --git a/vendor/github.com/openshift/api/operator/v1/types_kubecontrollermanager.go b/vendor/github.com/openshift/api/operator/v1/types_kubecontrollermanager.go index 564be1ac09..36ecc5edad 100644 --- a/vendor/github.com/openshift/api/operator/v1/types_kubecontrollermanager.go +++ b/vendor/github.com/openshift/api/operator/v1/types_kubecontrollermanager.go @@ -14,18 +14,13 @@ type KubeControllerManager struct { metav1.ObjectMeta `json:"metadata"` // +required - Spec KubeControllerManagerSpec `json:"spec"` + Spec KubeControllerManagerSpec `json:"spec"` // +optional Status KubeControllerManagerStatus `json:"status"` } type KubeControllerManagerSpec struct { StaticPodOperatorSpec `json:",inline"` - - // forceRedeploymentReason can be used to force the redeployment of the kube-controller-manager by providing a unique string. - // This provides a mechanism to kick a previously failed deployment and provide a reason why you think it will work - // this time instead of failing again on the same config. - ForceRedeploymentReason string `json:"forceRedeploymentReason"` } type KubeControllerManagerStatus struct { diff --git a/vendor/github.com/openshift/api/operator/v1/types_openshiftapiserver.go b/vendor/github.com/openshift/api/operator/v1/types_openshiftapiserver.go index e26ee9c13e..ec0487ccfc 100644 --- a/vendor/github.com/openshift/api/operator/v1/types_openshiftapiserver.go +++ b/vendor/github.com/openshift/api/operator/v1/types_openshiftapiserver.go @@ -14,7 +14,7 @@ type OpenShiftAPIServer struct { metav1.ObjectMeta `json:"metadata"` // +required - Spec OpenShiftAPIServerSpec `json:"spec"` + Spec OpenShiftAPIServerSpec `json:"spec"` // +optional Status OpenShiftAPIServerStatus `json:"status"` } diff --git a/vendor/github.com/openshift/api/operator/v1/types_openshiftcontrollermanager.go b/vendor/github.com/openshift/api/operator/v1/types_openshiftcontrollermanager.go index bad14acbe3..57ebe8e3a6 100644 --- a/vendor/github.com/openshift/api/operator/v1/types_openshiftcontrollermanager.go +++ b/vendor/github.com/openshift/api/operator/v1/types_openshiftcontrollermanager.go @@ -14,7 +14,7 @@ type OpenShiftControllerManager struct { metav1.ObjectMeta `json:"metadata"` // +required - Spec OpenShiftControllerManagerSpec `json:"spec"` + Spec OpenShiftControllerManagerSpec `json:"spec"` // +optional Status OpenShiftControllerManagerStatus `json:"status"` } diff --git a/vendor/github.com/openshift/api/operator/v1/types_scheduler.go b/vendor/github.com/openshift/api/operator/v1/types_scheduler.go index 0d08b80444..69fa5fbe4c 100644 --- a/vendor/github.com/openshift/api/operator/v1/types_scheduler.go +++ b/vendor/github.com/openshift/api/operator/v1/types_scheduler.go @@ -14,18 +14,13 @@ type KubeScheduler struct { metav1.ObjectMeta `json:"metadata"` // +required - Spec KubeSchedulerSpec `json:"spec"` + Spec KubeSchedulerSpec `json:"spec"` // +optional Status KubeSchedulerStatus `json:"status"` } type KubeSchedulerSpec struct { StaticPodOperatorSpec `json:",inline"` - - // forceRedeploymentReason can be used to force the redeployment of the kube-scheduler by providing a unique string. - // This provides a mechanism to kick a previously failed deployment and provide a reason why you think it will work - // this time instead of failing again on the same config. - ForceRedeploymentReason string `json:"forceRedeploymentReason"` } type KubeSchedulerStatus struct { diff --git a/vendor/github.com/openshift/api/operator/v1/types_serviceca.go b/vendor/github.com/openshift/api/operator/v1/types_serviceca.go index 42a51dbc9a..e6bb242aa2 100644 --- a/vendor/github.com/openshift/api/operator/v1/types_serviceca.go +++ b/vendor/github.com/openshift/api/operator/v1/types_serviceca.go @@ -14,8 +14,10 @@ type ServiceCA struct { metav1.ObjectMeta `json:"metadata"` // +required - Spec ServiceCASpec `json:"spec"` + //spec holds user settable values for configuration + Spec ServiceCASpec `json:"spec"` // +optional + // status holds observed values from the cluster. They may not be overridden. Status ServiceCAStatus `json:"status"` } diff --git a/vendor/github.com/openshift/api/operator/v1/types_servicecatalogapiserver.go b/vendor/github.com/openshift/api/operator/v1/types_servicecatalogapiserver.go index 92e0098546..aa480d7322 100644 --- a/vendor/github.com/openshift/api/operator/v1/types_servicecatalogapiserver.go +++ b/vendor/github.com/openshift/api/operator/v1/types_servicecatalogapiserver.go @@ -14,7 +14,7 @@ type ServiceCatalogAPIServer struct { metav1.ObjectMeta `json:"metadata,omitempty"` // +required - Spec ServiceCatalogAPIServerSpec `json:"spec"` + Spec ServiceCatalogAPIServerSpec `json:"spec"` // +optional Status ServiceCatalogAPIServerStatus `json:"status"` } diff --git a/vendor/github.com/openshift/api/operator/v1/zz_generated.deepcopy.go b/vendor/github.com/openshift/api/operator/v1/zz_generated.deepcopy.go index 105519ea8d..8974078f61 100644 --- a/vendor/github.com/openshift/api/operator/v1/zz_generated.deepcopy.go +++ b/vendor/github.com/openshift/api/operator/v1/zz_generated.deepcopy.go @@ -5,8 +5,8 @@ package v1 import ( - core_v1 "k8s.io/api/core/v1" - meta_v1 "k8s.io/apimachinery/pkg/apis/meta/v1" + corev1 "k8s.io/api/core/v1" + metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" runtime "k8s.io/apimachinery/pkg/runtime" ) @@ -249,26 +249,118 @@ func (in *ConsoleStatus) DeepCopy() *ConsoleStatus { return out } +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *DNS) DeepCopyInto(out *DNS) { + *out = *in + out.TypeMeta = in.TypeMeta + in.ObjectMeta.DeepCopyInto(&out.ObjectMeta) + out.Spec = in.Spec + in.Status.DeepCopyInto(&out.Status) + return +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new DNS. +func (in *DNS) DeepCopy() *DNS { + if in == nil { + return nil + } + out := new(DNS) + in.DeepCopyInto(out) + return out +} + +// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object. +func (in *DNS) DeepCopyObject() runtime.Object { + if c := in.DeepCopy(); c != nil { + return c + } + return nil +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *DNSList) DeepCopyInto(out *DNSList) { + *out = *in + out.TypeMeta = in.TypeMeta + out.ListMeta = in.ListMeta + if in.Items != nil { + in, out := &in.Items, &out.Items + *out = make([]DNS, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } + return +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new DNSList. +func (in *DNSList) DeepCopy() *DNSList { + if in == nil { + return nil + } + out := new(DNSList) + in.DeepCopyInto(out) + return out +} + +// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object. +func (in *DNSList) DeepCopyObject() runtime.Object { + if c := in.DeepCopy(); c != nil { + return c + } + return nil +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *DNSSpec) DeepCopyInto(out *DNSSpec) { + *out = *in + return +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new DNSSpec. +func (in *DNSSpec) DeepCopy() *DNSSpec { + if in == nil { + return nil + } + out := new(DNSSpec) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *DNSStatus) DeepCopyInto(out *DNSStatus) { + *out = *in + if in.Conditions != nil { + in, out := &in.Conditions, &out.Conditions + *out = make([]OperatorCondition, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } + return +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new DNSStatus. +func (in *DNSStatus) DeepCopy() *DNSStatus { + if in == nil { + return nil + } + out := new(DNSStatus) + in.DeepCopyInto(out) + return out +} + // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *DefaultNetworkDefinition) DeepCopyInto(out *DefaultNetworkDefinition) { *out = *in if in.OpenShiftSDNConfig != nil { in, out := &in.OpenShiftSDNConfig, &out.OpenShiftSDNConfig - if *in == nil { - *out = nil - } else { - *out = new(OpenShiftSDNConfig) - (*in).DeepCopyInto(*out) - } + *out = new(OpenShiftSDNConfig) + (*in).DeepCopyInto(*out) } if in.OVNKubernetesConfig != nil { in, out := &in.OVNKubernetesConfig, &out.OVNKubernetesConfig - if *in == nil { - *out = nil - } else { - *out = new(OVNKubernetesConfig) - (*in).DeepCopyInto(*out) - } + *out = new(OVNKubernetesConfig) + (*in).DeepCopyInto(*out) } return } @@ -476,57 +568,33 @@ func (in *IngressControllerSpec) DeepCopyInto(out *IngressControllerSpec) { *out = *in if in.Replicas != nil { in, out := &in.Replicas, &out.Replicas - if *in == nil { - *out = nil - } else { - *out = new(int32) - **out = **in - } + *out = new(int32) + **out = **in } if in.EndpointPublishingStrategy != nil { in, out := &in.EndpointPublishingStrategy, &out.EndpointPublishingStrategy - if *in == nil { - *out = nil - } else { - *out = new(EndpointPublishingStrategy) - **out = **in - } + *out = new(EndpointPublishingStrategy) + **out = **in } if in.DefaultCertificate != nil { in, out := &in.DefaultCertificate, &out.DefaultCertificate - if *in == nil { - *out = nil - } else { - *out = new(core_v1.LocalObjectReference) - **out = **in - } + *out = new(corev1.LocalObjectReference) + **out = **in } if in.NamespaceSelector != nil { in, out := &in.NamespaceSelector, &out.NamespaceSelector - if *in == nil { - *out = nil - } else { - *out = new(meta_v1.LabelSelector) - (*in).DeepCopyInto(*out) - } + *out = new(metav1.LabelSelector) + (*in).DeepCopyInto(*out) } if in.RouteSelector != nil { in, out := &in.RouteSelector, &out.RouteSelector - if *in == nil { - *out = nil - } else { - *out = new(meta_v1.LabelSelector) - (*in).DeepCopyInto(*out) - } + *out = new(metav1.LabelSelector) + (*in).DeepCopyInto(*out) } if in.NodePlacement != nil { in, out := &in.NodePlacement, &out.NodePlacement - if *in == nil { - *out = nil - } else { - *out = new(NodePlacement) - (*in).DeepCopyInto(*out) - } + *out = new(NodePlacement) + (*in).DeepCopyInto(*out) } return } @@ -546,12 +614,8 @@ func (in *IngressControllerStatus) DeepCopyInto(out *IngressControllerStatus) { *out = *in if in.EndpointPublishingStrategy != nil { in, out := &in.EndpointPublishingStrategy, &out.EndpointPublishingStrategy - if *in == nil { - *out = nil - } else { - *out = new(EndpointPublishingStrategy) - **out = **in - } + *out = new(EndpointPublishingStrategy) + **out = **in } if in.Conditions != nil { in, out := &in.Conditions, &out.Conditions @@ -994,30 +1058,18 @@ func (in *NetworkSpec) DeepCopyInto(out *NetworkSpec) { } if in.DisableMultiNetwork != nil { in, out := &in.DisableMultiNetwork, &out.DisableMultiNetwork - if *in == nil { - *out = nil - } else { - *out = new(bool) - **out = **in - } + *out = new(bool) + **out = **in } if in.DeployKubeProxy != nil { in, out := &in.DeployKubeProxy, &out.DeployKubeProxy - if *in == nil { - *out = nil - } else { - *out = new(bool) - **out = **in - } + *out = new(bool) + **out = **in } if in.KubeProxyConfig != nil { in, out := &in.KubeProxyConfig, &out.KubeProxyConfig - if *in == nil { - *out = nil - } else { - *out = new(ProxyConfig) - (*in).DeepCopyInto(*out) - } + *out = new(ProxyConfig) + (*in).DeepCopyInto(*out) } return } @@ -1053,11 +1105,14 @@ func (in *NodePlacement) DeepCopyInto(out *NodePlacement) { *out = *in if in.NodeSelector != nil { in, out := &in.NodeSelector, &out.NodeSelector - if *in == nil { - *out = nil - } else { - *out = new(meta_v1.LabelSelector) - (*in).DeepCopyInto(*out) + *out = new(metav1.LabelSelector) + (*in).DeepCopyInto(*out) + } + if in.Tolerations != nil { + in, out := &in.Tolerations, &out.Tolerations + *out = make([]corev1.Toleration, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) } } return @@ -1099,12 +1154,8 @@ func (in *OVNKubernetesConfig) DeepCopyInto(out *OVNKubernetesConfig) { *out = *in if in.MTU != nil { in, out := &in.MTU, &out.MTU - if *in == nil { - *out = nil - } else { - *out = new(uint32) - **out = **in - } + *out = new(uint32) + **out = **in } return } @@ -1314,30 +1365,18 @@ func (in *OpenShiftSDNConfig) DeepCopyInto(out *OpenShiftSDNConfig) { *out = *in if in.VXLANPort != nil { in, out := &in.VXLANPort, &out.VXLANPort - if *in == nil { - *out = nil - } else { - *out = new(uint32) - **out = **in - } + *out = new(uint32) + **out = **in } if in.MTU != nil { in, out := &in.MTU, &out.MTU - if *in == nil { - *out = nil - } else { - *out = new(uint32) - **out = **in - } + *out = new(uint32) + **out = **in } if in.UseExternalOpenvswitch != nil { in, out := &in.UseExternalOpenvswitch, &out.UseExternalOpenvswitch - if *in == nil { - *out = nil - } else { - *out = new(bool) - **out = **in - } + *out = new(bool) + **out = **in } return } @@ -1352,59 +1391,6 @@ func (in *OpenShiftSDNConfig) DeepCopy() *OpenShiftSDNConfig { return out } -// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. -func (in *OperandContainerSpec) DeepCopyInto(out *OperandContainerSpec) { - *out = *in - if in.Resources != nil { - in, out := &in.Resources, &out.Resources - if *in == nil { - *out = nil - } else { - *out = new(core_v1.ResourceRequirements) - (*in).DeepCopyInto(*out) - } - } - return -} - -// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new OperandContainerSpec. -func (in *OperandContainerSpec) DeepCopy() *OperandContainerSpec { - if in == nil { - return nil - } - out := new(OperandContainerSpec) - in.DeepCopyInto(out) - return out -} - -// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. -func (in *OperandSpec) DeepCopyInto(out *OperandSpec) { - *out = *in - if in.OperandContainerSpecs != nil { - in, out := &in.OperandContainerSpecs, &out.OperandContainerSpecs - *out = make([]OperandContainerSpec, len(*in)) - for i := range *in { - (*in)[i].DeepCopyInto(&(*out)[i]) - } - } - if in.UnsupportedResourcePatches != nil { - in, out := &in.UnsupportedResourcePatches, &out.UnsupportedResourcePatches - *out = make([]ResourcePatch, len(*in)) - copy(*out, *in) - } - return -} - -// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new OperandSpec. -func (in *OperandSpec) DeepCopy() *OperandSpec { - if in == nil { - return nil - } - out := new(OperandSpec) - in.DeepCopyInto(out) - return out -} - // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *OperatorCondition) DeepCopyInto(out *OperatorCondition) { *out = *in @@ -1425,13 +1411,6 @@ func (in *OperatorCondition) DeepCopy() *OperatorCondition { // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *OperatorSpec) DeepCopyInto(out *OperatorSpec) { *out = *in - if in.OperandSpecs != nil { - in, out := &in.OperandSpecs, &out.OperandSpecs - *out = make([]OperandSpec, len(*in)) - for i := range *in { - (*in)[i].DeepCopyInto(&(*out)[i]) - } - } in.UnsupportedConfigOverrides.DeepCopyInto(&out.UnsupportedConfigOverrides) in.ObservedConfig.DeepCopyInto(&out.ObservedConfig) return @@ -1482,12 +1461,15 @@ func (in *ProxyConfig) DeepCopyInto(out *ProxyConfig) { in, out := &in.ProxyArguments, &out.ProxyArguments *out = make(map[string][]string, len(*in)) for key, val := range *in { + var outVal []string if val == nil { (*out)[key] = nil } else { - (*out)[key] = make([]string, len(val)) - copy((*out)[key], val) + in, out := &val, &outVal + *out = make([]string, len(*in)) + copy(*out, *in) } + (*out)[key] = outVal } } return @@ -1503,22 +1485,6 @@ func (in *ProxyConfig) DeepCopy() *ProxyConfig { return out } -// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. -func (in *ResourcePatch) DeepCopyInto(out *ResourcePatch) { - *out = *in - return -} - -// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ResourcePatch. -func (in *ResourcePatch) DeepCopy() *ResourcePatch { - if in == nil { - return nil - } - out := new(ResourcePatch) - in.DeepCopyInto(out) - return out -} - // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *ServiceCA) DeepCopyInto(out *ServiceCA) { *out = *in diff --git a/vendor/github.com/openshift/api/operator/v1/zz_generated.swagger_doc_generated.go b/vendor/github.com/openshift/api/operator/v1/zz_generated.swagger_doc_generated.go index d106095e34..da167ba6d2 100644 --- a/vendor/github.com/openshift/api/operator/v1/zz_generated.swagger_doc_generated.go +++ b/vendor/github.com/openshift/api/operator/v1/zz_generated.swagger_doc_generated.go @@ -46,26 +46,6 @@ func (NodeStatus) SwaggerDoc() map[string]string { return map_NodeStatus } -var map_OperandContainerSpec = map[string]string{ - "name": "name is the name of the container to modify", - "resources": "resources are the requests and limits to place in the container. Nil means to accept the defaults.", -} - -func (OperandContainerSpec) SwaggerDoc() map[string]string { - return map_OperandContainerSpec -} - -var map_OperandSpec = map[string]string{ - "": "OperandSpec holds information for customization of a particular functional unit - logically maps to a workload", - "name": "name is the name of this unit. The operator must be aware of it.", - "operandContainerSpecs": "operandContainerSpecs are per-container options", - "unsupportedResourcePatches": "unsupportedResourcePatches are applied to the workload resource for this unit. This is an unsupported workaround if anything needs to be modified on the workload that is not otherwise configurable.", -} - -func (OperandSpec) SwaggerDoc() map[string]string { - return map_OperandSpec -} - var map_OperatorCondition = map[string]string{ "": "OperatorCondition is just the standard condition fields.", } @@ -78,7 +58,7 @@ var map_OperatorSpec = map[string]string{ "": "OperatorSpec contains common fields operators need. It is intended to be anonymous included inside of the Spec struct for your particular operator.", "managementState": "managementState indicates whether and how the operator should manage the component", "logLevel": "logLevel is an intent based logging for an overall component. It does not give fine grained control, but it is a simple way to manage coarse grained logging choices that operators have to interpret for their operands.", - "operandSpecs": "operandSpecs provide customization for functional units within the component", + "operatorLogLevel": "operatorLogLevel is an intent based logging for the operator itself. It does not give fine grained control, but it is a simple way to manage coarse grained logging choices that operators have to interpret for themselves.", "unsupportedConfigOverrides": "unsupportedConfigOverrides holds a sparse config that will override any previously set options. It only needs to be the fields to override it will end up overlaying in the following order: 1. hardcoded defaults 2. observedConfig 3. unsupportedConfigOverrides", "observedConfig": "observedConfig holds a sparse config that controller has observed from the cluster state. It exists in spec because it is an input to the level for the operator", } @@ -99,20 +79,11 @@ func (OperatorStatus) SwaggerDoc() map[string]string { return map_OperatorStatus } -var map_ResourcePatch = map[string]string{ - "": "ResourcePatch is a way to represent the patch you would issue to `kubectl patch` in the API", - "type": "type is the type of patch to apply: jsonmerge, strategicmerge", - "patch": "patch the patch itself", -} - -func (ResourcePatch) SwaggerDoc() map[string]string { - return map_ResourcePatch -} - var map_StaticPodOperatorSpec = map[string]string{ - "": "StaticPodOperatorSpec is spec for controllers that manage static pods.", - "failedRevisionLimit": "failedRevisionLimit is the number of failed static pod installer revisions to keep on disk and in the api -1 = unlimited, 0 or unset = 5 (default)", - "succeededRevisionLimit": "succeededRevisionLimit is the number of successful static pod installer revisions to keep on disk and in the api -1 = unlimited, 0 or unset = 5 (default)", + "": "StaticPodOperatorSpec is spec for controllers that manage static pods.", + "forceRedeploymentReason": "forceRedeploymentReason can be used to force the redeployment of the operand by providing a unique string. This provides a mechanism to kick a previously failed deployment and provide a reason why you think it will work this time instead of failing again on the same config.", + "failedRevisionLimit": "failedRevisionLimit is the number of failed static pod installer revisions to keep on disk and in the api -1 = unlimited, 0 or unset = 5 (default)", + "succeededRevisionLimit": "succeededRevisionLimit is the number of successful static pod installer revisions to keep on disk and in the api -1 = unlimited, 0 or unset = 5 (default)", } func (StaticPodOperatorSpec) SwaggerDoc() map[string]string { @@ -120,7 +91,7 @@ func (StaticPodOperatorSpec) SwaggerDoc() map[string]string { } var map_StaticPodOperatorStatus = map[string]string{ - "": "StaticPodOperatorStatus is status for controllers that manage static pods. There are different needs because individual node status must be tracked.", + "": "StaticPodOperatorStatus is status for controllers that manage static pods. There are different needs because individual node status must be tracked.", "latestAvailableRevision": "latestAvailableRevision is the deploymentID of the most recent deployment", "nodeStatuses": "nodeStatuses track the deployment values and errors across individual nodes", } @@ -162,6 +133,43 @@ func (ConsoleSpec) SwaggerDoc() map[string]string { return map_ConsoleSpec } +var map_DNS = map[string]string{ + "": "DNS manages the CoreDNS component to provide a name resolution service for pods and services in the cluster.\n\nThis supports the DNS-based service discovery specification: https://github.com/kubernetes/dns/blob/master/docs/specification.md\n\nMore details: https://kubernetes.io/docs/tasks/administer-cluster/coredns", + "spec": "spec is the specification of the desired behavior of the DNS.", + "status": "status is the most recently observed status of the DNS.", +} + +func (DNS) SwaggerDoc() map[string]string { + return map_DNS +} + +var map_DNSList = map[string]string{ + "": "DNSList contains a list of DNS", +} + +func (DNSList) SwaggerDoc() map[string]string { + return map_DNSList +} + +var map_DNSSpec = map[string]string{ + "": "DNSSpec is the specification of the desired behavior of the DNS.", +} + +func (DNSSpec) SwaggerDoc() map[string]string { + return map_DNSSpec +} + +var map_DNSStatus = map[string]string{ + "": "DNSStatus defines the observed status of the DNS.", + "clusterIP": "clusterIP is the service IP through which this DNS is made available.\n\nIn the case of the default DNS, this will be a well known IP that is used as the default nameserver for pods that are using the default ClusterFirst DNS policy.\n\nIn general, this IP can be specified in a pod's spec.dnsConfig.nameservers list or used explicitly when performing name resolution from within the cluster. Example: dig foo.com @\n\nMore info: https://kubernetes.io/docs/concepts/services-networking/service/#virtual-ips-and-service-proxies", + "clusterDomain": "clusterDomain is the local cluster DNS domain suffix for DNS services. This will be a subdomain as defined in RFC 1034, section 3.5: https://tools.ietf.org/html/rfc1034#section-3.5 Example: \"cluster.local\"\n\nMore info: https://kubernetes.io/docs/concepts/services-networking/dns-pod-service", + "conditions": "conditions provide information about the state of the DNS on the cluster.\n\nThese are the supported DNS conditions:\n\n * Available\n - True if the following conditions are met:\n * DNS controller daemonset is available.\n - False if any of those conditions are unsatisfied.", +} + +func (DNSStatus) SwaggerDoc() map[string]string { + return map_DNSStatus +} + var map_Etcd = map[string]string{ "": "Etcd provides information to configure an operator to manage kube-apiserver.", } @@ -219,7 +227,7 @@ var map_IngressControllerSpec = map[string]string{ "": "IngressControllerSpec is the specification of the desired behavior of the IngressController.", "domain": "domain is a DNS name serviced by the ingress controller and is used to configure multiple features:\n\n* For the LoadBalancerService endpoint publishing strategy, domain is\n used to configure DNS records. See endpointPublishingStrategy.\n\n* When using a generated default certificate, the certificate will be valid\n for domain and its subdomains. See defaultCertificate.\n\n* The value is published to individual Route statuses so that end-users\n know where to target external DNS records.\n\ndomain must be unique among all IngressControllers, and cannot be updated.\n\nIf empty, defaults to ingress.config.openshift.io/cluster .spec.domain.", "replicas": "replicas is the desired number of ingress controller replicas. If unset, defaults to 2.", - "endpointPublishingStrategy": "endpointPublishingStrategy is used to publish the ingress controller endpoints to other networks, enable load balancer integrations, etc.\n\nIf unset, the default is based on infrastructure.config.openshift.io/cluster .status.platform:\n\n AWS: LoadBalancerService\n All other platform types: Private\n\nendpointPublishingStrategy cannot be updated.", + "endpointPublishingStrategy": "endpointPublishingStrategy is used to publish the ingress controller endpoints to other networks, enable load balancer integrations, etc.\n\nIf unset, the default is based on infrastructure.config.openshift.io/cluster .status.platform:\n\n AWS: LoadBalancerService\n Libvirt: HostNetwork\n\nAny other platform types (including None) default to HostNetwork.\n\nendpointPublishingStrategy cannot be updated.", "defaultCertificate": "defaultCertificate is a reference to a secret containing the default certificate served by the ingress controller. When Routes don't specify their own certificate, defaultCertificate is used.\n\nThe secret must contain the following keys and data:\n\n tls.crt: certificate file contents\n tls.key: key file contents\n\nIf unset, a wildcard certificate is automatically generated and used. The certificate is valid for the ingress controller domain (and subdomains) and the generated certificate's CA will be automatically integrated with the cluster's trust store.\n\nThe in-use certificate (whether generated or user-specified) will be automatically integrated with OpenShift's built-in OAuth server.", "namespaceSelector": "namespaceSelector is used to filter the set of namespaces serviced by the ingress controller. This is useful for implementing shards.\n\nIf unset, the default is no filtering.", "routeSelector": "routeSelector is used to filter the set of Routes serviced by the ingress controller. This is useful for implementing shards.\n\nIf unset, the default is no filtering.", @@ -246,6 +254,7 @@ func (IngressControllerStatus) SwaggerDoc() map[string]string { var map_NodePlacement = map[string]string{ "": "NodePlacement describes node scheduling configuration for an ingress controller.", "nodeSelector": "nodeSelector is the node selector applied to ingress controller deployments.\n\nIf unset, the default is:\n\n beta.kubernetes.io/os: linux\n node-role.kubernetes.io/worker: ''\n\nIf set, the specified selector is used and replaces the default.", + "tolerations": "tolerations is a list of tolerations applied to ingress controller deployments.\n\nThe default is an empty list.\n\nSee https://kubernetes.io/docs/concepts/configuration/taint-and-toleration/", } func (NodePlacement) SwaggerDoc() map[string]string { @@ -270,14 +279,6 @@ func (KubeAPIServerList) SwaggerDoc() map[string]string { return map_KubeAPIServerList } -var map_KubeAPIServerSpec = map[string]string{ - "forceRedeploymentReason": "forceRedeploymentReason can be used to force the redeployment of the kube-apiserver by providing a unique string. This provides a mechanism to kick a previously failed deployment and provide a reason why you think it will work this time instead of failing again on the same config.", -} - -func (KubeAPIServerSpec) SwaggerDoc() map[string]string { - return map_KubeAPIServerSpec -} - var map_KubeControllerManager = map[string]string{ "": "KubeControllerManager provides information to configure an operator to manage kube-controller-manager.", } @@ -296,14 +297,6 @@ func (KubeControllerManagerList) SwaggerDoc() map[string]string { return map_KubeControllerManagerList } -var map_KubeControllerManagerSpec = map[string]string{ - "forceRedeploymentReason": "forceRedeploymentReason can be used to force the redeployment of the kube-controller-manager by providing a unique string. This provides a mechanism to kick a previously failed deployment and provide a reason why you think it will work this time instead of failing again on the same config.", -} - -func (KubeControllerManagerSpec) SwaggerDoc() map[string]string { - return map_KubeControllerManagerSpec -} - var map_AdditionalNetworkDefinition = map[string]string{ "": "AdditionalNetworkDefinition configures an extra network that is available but not created by default. Instead, pods must request them by name. type must be specified, along with exactly one \"Config\" that matches the type.", "type": "type is the type of network The only supported value is NetworkTypeRaw", @@ -383,10 +376,10 @@ func (OVNKubernetesConfig) SwaggerDoc() map[string]string { } var map_OpenShiftSDNConfig = map[string]string{ - "": "OpenShiftSDNConfig configures the three openshift-sdn plugins", - "mode": "mode is one of \"Multitenant\", \"Subnet\", or \"NetworkPolicy\"", - "vxlanPort": "vxlanPort is the port to use for all vxlan packets. The default is 4789.", - "mtu": "mtu is the mtu to use for the tunnel interface. Defaults to 1450 if unset. This must be 50 bytes smaller than the machine's uplink.", + "": "OpenShiftSDNConfig configures the three openshift-sdn plugins", + "mode": "mode is one of \"Multitenant\", \"Subnet\", or \"NetworkPolicy\"", + "vxlanPort": "vxlanPort is the port to use for all vxlan packets. The default is 4789.", + "mtu": "mtu is the mtu to use for the tunnel interface. Defaults to 1450 if unset. This must be 50 bytes smaller than the machine's uplink.", "useExternalOpenvswitch": "useExternalOpenvswitch tells the operator not to install openvswitch, because it will be provided separately. If set, you must provide it yourself.", } @@ -459,16 +452,10 @@ func (KubeSchedulerList) SwaggerDoc() map[string]string { return map_KubeSchedulerList } -var map_KubeSchedulerSpec = map[string]string{ - "forceRedeploymentReason": "forceRedeploymentReason can be used to force the redeployment of the kube-scheduler by providing a unique string. This provides a mechanism to kick a previously failed deployment and provide a reason why you think it will work this time instead of failing again on the same config.", -} - -func (KubeSchedulerSpec) SwaggerDoc() map[string]string { - return map_KubeSchedulerSpec -} - var map_ServiceCA = map[string]string{ - "": "ServiceCA provides information to configure an operator to manage the service cert controllers", + "": "ServiceCA provides information to configure an operator to manage the service cert controllers", + "spec": "spec holds user settable values for configuration", + "status": "status holds observed values from the cluster. They may not be overridden.", } func (ServiceCA) SwaggerDoc() map[string]string { diff --git a/vendor/github.com/openshift/api/operator/v1alpha1/zz_generated.deepcopy.go b/vendor/github.com/openshift/api/operator/v1alpha1/zz_generated.deepcopy.go index 0004598b5b..72ae4b3db0 100644 --- a/vendor/github.com/openshift/api/operator/v1alpha1/zz_generated.deepcopy.go +++ b/vendor/github.com/openshift/api/operator/v1alpha1/zz_generated.deepcopy.go @@ -168,21 +168,13 @@ func (in *OperatorStatus) DeepCopyInto(out *OperatorStatus) { } if in.CurrentAvailability != nil { in, out := &in.CurrentAvailability, &out.CurrentAvailability - if *in == nil { - *out = nil - } else { - *out = new(VersionAvailability) - (*in).DeepCopyInto(*out) - } + *out = new(VersionAvailability) + (*in).DeepCopyInto(*out) } if in.TargetAvailability != nil { in, out := &in.TargetAvailability, &out.TargetAvailability - if *in == nil { - *out = nil - } else { - *out = new(VersionAvailability) - (*in).DeepCopyInto(*out) - } + *out = new(VersionAvailability) + (*in).DeepCopyInto(*out) } return } diff --git a/vendor/github.com/openshift/api/operator/v1alpha1/zz_generated.swagger_doc_generated.go b/vendor/github.com/openshift/api/operator/v1alpha1/zz_generated.swagger_doc_generated.go index 76b45b0568..59a145211e 100644 --- a/vendor/github.com/openshift/api/operator/v1alpha1/zz_generated.swagger_doc_generated.go +++ b/vendor/github.com/openshift/api/operator/v1alpha1/zz_generated.swagger_doc_generated.go @@ -113,7 +113,7 @@ func (OperatorStatus) SwaggerDoc() map[string]string { } var map_StaticPodOperatorStatus = map[string]string{ - "": "StaticPodOperatorStatus is status for controllers that manage static pods. There are different needs because individual node status must be tracked.", + "": "StaticPodOperatorStatus is status for controllers that manage static pods. There are different needs because individual node status must be tracked.", "latestAvailableDeploymentGeneration": "latestAvailableDeploymentGeneration is the deploymentID of the most recent deployment", "nodeStatuses": "nodeStatuses track the deployment values and errors across individual nodes", } diff --git a/vendor/github.com/openshift/api/osin/v1/zz_generated.deepcopy.go b/vendor/github.com/openshift/api/osin/v1/zz_generated.deepcopy.go index 59964e1000..fb73989345 100644 --- a/vendor/github.com/openshift/api/osin/v1/zz_generated.deepcopy.go +++ b/vendor/github.com/openshift/api/osin/v1/zz_generated.deepcopy.go @@ -127,12 +127,8 @@ func (in *GitLabIdentityProvider) DeepCopyInto(out *GitLabIdentityProvider) { out.ClientSecret = in.ClientSecret if in.Legacy != nil { in, out := &in.Legacy, &out.Legacy - if *in == nil { - *out = nil - } else { - *out = new(bool) - **out = **in - } + *out = new(bool) + **out = **in } return } @@ -333,12 +329,8 @@ func (in *OAuthConfig) DeepCopyInto(out *OAuthConfig) { *out = *in if in.MasterCA != nil { in, out := &in.MasterCA, &out.MasterCA - if *in == nil { - *out = nil - } else { - *out = new(string) - **out = **in - } + *out = new(string) + **out = **in } if in.IdentityProviders != nil { in, out := &in.IdentityProviders, &out.IdentityProviders @@ -350,22 +342,14 @@ func (in *OAuthConfig) DeepCopyInto(out *OAuthConfig) { out.GrantConfig = in.GrantConfig if in.SessionConfig != nil { in, out := &in.SessionConfig, &out.SessionConfig - if *in == nil { - *out = nil - } else { - *out = new(SessionConfig) - **out = **in - } + *out = new(SessionConfig) + **out = **in } in.TokenConfig.DeepCopyInto(&out.TokenConfig) if in.Templates != nil { in, out := &in.Templates, &out.Templates - if *in == nil { - *out = nil - } else { - *out = new(OAuthTemplates) - **out = **in - } + *out = new(OAuthTemplates) + **out = **in } return } @@ -586,12 +570,8 @@ func (in *TokenConfig) DeepCopyInto(out *TokenConfig) { *out = *in if in.AccessTokenInactivityTimeoutSeconds != nil { in, out := &in.AccessTokenInactivityTimeoutSeconds, &out.AccessTokenInactivityTimeoutSeconds - if *in == nil { - *out = nil - } else { - *out = new(int32) - **out = **in - } + *out = new(int32) + **out = **in } return } diff --git a/vendor/github.com/openshift/api/project/v1/generated.pb.go b/vendor/github.com/openshift/api/project/v1/generated.pb.go index 126f0b1e02..018c6acd1f 100644 --- a/vendor/github.com/openshift/api/project/v1/generated.pb.go +++ b/vendor/github.com/openshift/api/project/v1/generated.pb.go @@ -1,6 +1,5 @@ -// Code generated by protoc-gen-gogo. +// Code generated by protoc-gen-gogo. DO NOT EDIT. // source: github.com/openshift/api/project/v1/generated.proto -// DO NOT EDIT! /* Package v1 is a generated protocol buffer package. @@ -235,24 +234,6 @@ func (m *ProjectStatus) MarshalTo(dAtA []byte) (int, error) { return i, nil } -func encodeFixed64Generated(dAtA []byte, offset int, v uint64) int { - dAtA[offset] = uint8(v) - dAtA[offset+1] = uint8(v >> 8) - dAtA[offset+2] = uint8(v >> 16) - dAtA[offset+3] = uint8(v >> 24) - dAtA[offset+4] = uint8(v >> 32) - dAtA[offset+5] = uint8(v >> 40) - dAtA[offset+6] = uint8(v >> 48) - dAtA[offset+7] = uint8(v >> 56) - return offset + 8 -} -func encodeFixed32Generated(dAtA []byte, offset int, v uint32) int { - dAtA[offset] = uint8(v) - dAtA[offset+1] = uint8(v >> 8) - dAtA[offset+2] = uint8(v >> 16) - dAtA[offset+3] = uint8(v >> 24) - return offset + 4 -} func encodeVarintGenerated(dAtA []byte, offset int, v uint64) int { for v >= 1<<7 { dAtA[offset] = uint8(v&0x7f | 0x80) diff --git a/vendor/github.com/openshift/api/project/v1/zz_generated.deepcopy.go b/vendor/github.com/openshift/api/project/v1/zz_generated.deepcopy.go index 61e15a5924..552dc51b1f 100644 --- a/vendor/github.com/openshift/api/project/v1/zz_generated.deepcopy.go +++ b/vendor/github.com/openshift/api/project/v1/zz_generated.deepcopy.go @@ -5,7 +5,7 @@ package v1 import ( - core_v1 "k8s.io/api/core/v1" + corev1 "k8s.io/api/core/v1" runtime "k8s.io/apimachinery/pkg/runtime" ) @@ -101,7 +101,7 @@ func (in *ProjectSpec) DeepCopyInto(out *ProjectSpec) { *out = *in if in.Finalizers != nil { in, out := &in.Finalizers, &out.Finalizers - *out = make([]core_v1.FinalizerName, len(*in)) + *out = make([]corev1.FinalizerName, len(*in)) copy(*out, *in) } return diff --git a/vendor/github.com/openshift/api/quota/v1/generated.pb.go b/vendor/github.com/openshift/api/quota/v1/generated.pb.go index b1fd54bd3b..94ff472379 100644 --- a/vendor/github.com/openshift/api/quota/v1/generated.pb.go +++ b/vendor/github.com/openshift/api/quota/v1/generated.pb.go @@ -1,6 +1,5 @@ -// Code generated by protoc-gen-gogo. +// Code generated by protoc-gen-gogo. DO NOT EDIT. // source: github.com/openshift/api/quota/v1/generated.proto -// DO NOT EDIT! /* Package v1 is a generated protocol buffer package. @@ -412,24 +411,6 @@ func (m *ResourceQuotaStatusByNamespace) MarshalTo(dAtA []byte) (int, error) { return i, nil } -func encodeFixed64Generated(dAtA []byte, offset int, v uint64) int { - dAtA[offset] = uint8(v) - dAtA[offset+1] = uint8(v >> 8) - dAtA[offset+2] = uint8(v >> 16) - dAtA[offset+3] = uint8(v >> 24) - dAtA[offset+4] = uint8(v >> 32) - dAtA[offset+5] = uint8(v >> 40) - dAtA[offset+6] = uint8(v >> 48) - dAtA[offset+7] = uint8(v >> 56) - return offset + 8 -} -func encodeFixed32Generated(dAtA []byte, offset int, v uint32) int { - dAtA[offset] = uint8(v) - dAtA[offset+1] = uint8(v >> 8) - dAtA[offset+2] = uint8(v >> 16) - dAtA[offset+3] = uint8(v >> 24) - return offset + 4 -} func encodeVarintGenerated(dAtA []byte, offset int, v uint64) int { for v >= 1<<7 { dAtA[offset] = uint8(v&0x7f | 0x80) @@ -1254,51 +1235,14 @@ func (m *ClusterResourceQuotaSelector) Unmarshal(dAtA []byte) error { if postIndex > l { return io.ErrUnexpectedEOF } - var keykey uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - keykey |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - var stringLenmapkey uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - stringLenmapkey |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - intStringLenmapkey := int(stringLenmapkey) - if intStringLenmapkey < 0 { - return ErrInvalidLengthGenerated - } - postStringIndexmapkey := iNdEx + intStringLenmapkey - if postStringIndexmapkey > l { - return io.ErrUnexpectedEOF - } - mapkey := string(dAtA[iNdEx:postStringIndexmapkey]) - iNdEx = postStringIndexmapkey if m.AnnotationSelector == nil { m.AnnotationSelector = make(map[string]string) } - if iNdEx < postIndex { - var valuekey uint64 + var mapkey string + var mapvalue string + for iNdEx < postIndex { + entryPreIndex := iNdEx + var wire uint64 for shift := uint(0); ; shift += 7 { if shift >= 64 { return ErrIntOverflowGenerated @@ -1308,41 +1252,80 @@ func (m *ClusterResourceQuotaSelector) Unmarshal(dAtA []byte) error { } b := dAtA[iNdEx] iNdEx++ - valuekey |= (uint64(b) & 0x7F) << shift + wire |= (uint64(b) & 0x7F) << shift if b < 0x80 { break } } - var stringLenmapvalue uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated + fieldNum := int32(wire >> 3) + if fieldNum == 1 { + var stringLenmapkey uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLenmapkey |= (uint64(b) & 0x7F) << shift + if b < 0x80 { + break + } } - if iNdEx >= l { + intStringLenmapkey := int(stringLenmapkey) + if intStringLenmapkey < 0 { + return ErrInvalidLengthGenerated + } + postStringIndexmapkey := iNdEx + intStringLenmapkey + if postStringIndexmapkey > l { return io.ErrUnexpectedEOF } - b := dAtA[iNdEx] - iNdEx++ - stringLenmapvalue |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break + mapkey = string(dAtA[iNdEx:postStringIndexmapkey]) + iNdEx = postStringIndexmapkey + } else if fieldNum == 2 { + var stringLenmapvalue uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLenmapvalue |= (uint64(b) & 0x7F) << shift + if b < 0x80 { + break + } } + intStringLenmapvalue := int(stringLenmapvalue) + if intStringLenmapvalue < 0 { + return ErrInvalidLengthGenerated + } + postStringIndexmapvalue := iNdEx + intStringLenmapvalue + if postStringIndexmapvalue > l { + return io.ErrUnexpectedEOF + } + mapvalue = string(dAtA[iNdEx:postStringIndexmapvalue]) + iNdEx = postStringIndexmapvalue + } else { + iNdEx = entryPreIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > postIndex { + return io.ErrUnexpectedEOF + } + iNdEx += skippy } - intStringLenmapvalue := int(stringLenmapvalue) - if intStringLenmapvalue < 0 { - return ErrInvalidLengthGenerated - } - postStringIndexmapvalue := iNdEx + intStringLenmapvalue - if postStringIndexmapvalue > l { - return io.ErrUnexpectedEOF - } - mapvalue := string(dAtA[iNdEx:postStringIndexmapvalue]) - iNdEx = postStringIndexmapvalue - m.AnnotationSelector[mapkey] = mapvalue - } else { - var mapvalue string - m.AnnotationSelector[mapkey] = mapvalue } + m.AnnotationSelector[mapkey] = mapvalue iNdEx = postIndex default: iNdEx = preIndex diff --git a/vendor/github.com/openshift/api/quota/v1/types.go b/vendor/github.com/openshift/api/quota/v1/types.go index 50236e0682..e34e815162 100644 --- a/vendor/github.com/openshift/api/quota/v1/types.go +++ b/vendor/github.com/openshift/api/quota/v1/types.go @@ -40,9 +40,13 @@ type ClusterResourceQuotaSpec struct { // the project must match both restrictions. type ClusterResourceQuotaSelector struct { // LabelSelector is used to select projects by label. + // +optional + // +nullable LabelSelector *metav1.LabelSelector `json:"labels" protobuf:"bytes,1,opt,name=labels"` // AnnotationSelector is used to select projects by annotation. + // +optional + // +nullable AnnotationSelector map[string]string `json:"annotations" protobuf:"bytes,2,rep,name=annotations"` } @@ -54,6 +58,8 @@ type ClusterResourceQuotaStatus struct { // Namespaces slices the usage by project. This division allows for quick resolution of // deletion reconciliation inside of a single project without requiring a recalculation // across all projects. This can be used to pull the deltas for a given project. + // +optional + // +nullable Namespaces ResourceQuotasStatusByNamespace `json:"namespaces" protobuf:"bytes,2,rep,name=namespaces"` } diff --git a/vendor/github.com/openshift/api/quota/v1/zz_generated.deepcopy.go b/vendor/github.com/openshift/api/quota/v1/zz_generated.deepcopy.go index e6d6e8708b..a7170a1556 100644 --- a/vendor/github.com/openshift/api/quota/v1/zz_generated.deepcopy.go +++ b/vendor/github.com/openshift/api/quota/v1/zz_generated.deepcopy.go @@ -5,7 +5,7 @@ package v1 import ( - meta_v1 "k8s.io/apimachinery/pkg/apis/meta/v1" + metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" runtime "k8s.io/apimachinery/pkg/runtime" ) @@ -136,12 +136,8 @@ func (in *ClusterResourceQuotaSelector) DeepCopyInto(out *ClusterResourceQuotaSe *out = *in if in.LabelSelector != nil { in, out := &in.LabelSelector, &out.LabelSelector - if *in == nil { - *out = nil - } else { - *out = new(meta_v1.LabelSelector) - (*in).DeepCopyInto(*out) - } + *out = new(metav1.LabelSelector) + (*in).DeepCopyInto(*out) } if in.AnnotationSelector != nil { in, out := &in.AnnotationSelector, &out.AnnotationSelector diff --git a/vendor/github.com/openshift/api/route/v1/generated.pb.go b/vendor/github.com/openshift/api/route/v1/generated.pb.go index 8575bccbbf..b0ad233717 100644 --- a/vendor/github.com/openshift/api/route/v1/generated.pb.go +++ b/vendor/github.com/openshift/api/route/v1/generated.pb.go @@ -1,6 +1,5 @@ -// Code generated by protoc-gen-gogo. +// Code generated by protoc-gen-gogo. DO NOT EDIT. // source: github.com/openshift/api/route/v1/generated.proto -// DO NOT EDIT! /* Package v1 is a generated protocol buffer package. @@ -493,24 +492,6 @@ func (m *TLSConfig) MarshalTo(dAtA []byte) (int, error) { return i, nil } -func encodeFixed64Generated(dAtA []byte, offset int, v uint64) int { - dAtA[offset] = uint8(v) - dAtA[offset+1] = uint8(v >> 8) - dAtA[offset+2] = uint8(v >> 16) - dAtA[offset+3] = uint8(v >> 24) - dAtA[offset+4] = uint8(v >> 32) - dAtA[offset+5] = uint8(v >> 40) - dAtA[offset+6] = uint8(v >> 48) - dAtA[offset+7] = uint8(v >> 56) - return offset + 8 -} -func encodeFixed32Generated(dAtA []byte, offset int, v uint32) int { - dAtA[offset] = uint8(v) - dAtA[offset+1] = uint8(v >> 8) - dAtA[offset+2] = uint8(v >> 16) - dAtA[offset+3] = uint8(v >> 24) - return offset + 4 -} func encodeVarintGenerated(dAtA []byte, offset int, v uint64) int { for v >= 1<<7 { dAtA[offset] = uint8(v&0x7f | 0x80) diff --git a/vendor/github.com/openshift/api/route/v1/zz_generated.deepcopy.go b/vendor/github.com/openshift/api/route/v1/zz_generated.deepcopy.go index c88cd03839..f3de8e47a0 100644 --- a/vendor/github.com/openshift/api/route/v1/zz_generated.deepcopy.go +++ b/vendor/github.com/openshift/api/route/v1/zz_generated.deepcopy.go @@ -64,11 +64,7 @@ func (in *RouteIngressCondition) DeepCopyInto(out *RouteIngressCondition) { *out = *in if in.LastTransitionTime != nil { in, out := &in.LastTransitionTime, &out.LastTransitionTime - if *in == nil { - *out = nil - } else { - *out = (*in).DeepCopy() - } + *out = (*in).DeepCopy() } return } @@ -146,21 +142,13 @@ func (in *RouteSpec) DeepCopyInto(out *RouteSpec) { } if in.Port != nil { in, out := &in.Port, &out.Port - if *in == nil { - *out = nil - } else { - *out = new(RoutePort) - **out = **in - } + *out = new(RoutePort) + **out = **in } if in.TLS != nil { in, out := &in.TLS, &out.TLS - if *in == nil { - *out = nil - } else { - *out = new(TLSConfig) - **out = **in - } + *out = new(TLSConfig) + **out = **in } return } @@ -203,12 +191,8 @@ func (in *RouteTargetReference) DeepCopyInto(out *RouteTargetReference) { *out = *in if in.Weight != nil { in, out := &in.Weight, &out.Weight - if *in == nil { - *out = nil - } else { - *out = new(int32) - **out = **in - } + *out = new(int32) + **out = **in } return } diff --git a/vendor/github.com/openshift/api/security/v1/generated.pb.go b/vendor/github.com/openshift/api/security/v1/generated.pb.go index 64228a4099..02fbad99be 100644 --- a/vendor/github.com/openshift/api/security/v1/generated.pb.go +++ b/vendor/github.com/openshift/api/security/v1/generated.pb.go @@ -1,6 +1,5 @@ -// Code generated by protoc-gen-gogo. +// Code generated by protoc-gen-gogo. DO NOT EDIT. // source: github.com/openshift/api/security/v1/generated.proto -// DO NOT EDIT! /* Package v1 is a generated protocol buffer package. @@ -1095,24 +1094,6 @@ func (m *SupplementalGroupsStrategyOptions) MarshalTo(dAtA []byte) (int, error) return i, nil } -func encodeFixed64Generated(dAtA []byte, offset int, v uint64) int { - dAtA[offset] = uint8(v) - dAtA[offset+1] = uint8(v >> 8) - dAtA[offset+2] = uint8(v >> 16) - dAtA[offset+3] = uint8(v >> 24) - dAtA[offset+4] = uint8(v >> 32) - dAtA[offset+5] = uint8(v >> 40) - dAtA[offset+6] = uint8(v >> 48) - dAtA[offset+7] = uint8(v >> 56) - return offset + 8 -} -func encodeFixed32Generated(dAtA []byte, offset int, v uint32) int { - dAtA[offset] = uint8(v) - dAtA[offset+1] = uint8(v >> 8) - dAtA[offset+2] = uint8(v >> 16) - dAtA[offset+3] = uint8(v >> 24) - return offset + 4 -} func encodeVarintGenerated(dAtA []byte, offset int, v uint64) int { for v >= 1<<7 { dAtA[offset] = uint8(v&0x7f | 0x80) diff --git a/vendor/github.com/openshift/api/security/v1/zz_generated.deepcopy.go b/vendor/github.com/openshift/api/security/v1/zz_generated.deepcopy.go index 892ecc4d35..70eef789e7 100644 --- a/vendor/github.com/openshift/api/security/v1/zz_generated.deepcopy.go +++ b/vendor/github.com/openshift/api/security/v1/zz_generated.deepcopy.go @@ -5,7 +5,7 @@ package v1 import ( - core_v1 "k8s.io/api/core/v1" + corev1 "k8s.io/api/core/v1" runtime "k8s.io/apimachinery/pkg/runtime" ) @@ -232,12 +232,8 @@ func (in *PodSecurityPolicySubjectReviewStatus) DeepCopyInto(out *PodSecurityPol *out = *in if in.AllowedBy != nil { in, out := &in.AllowedBy, &out.AllowedBy - if *in == nil { - *out = nil - } else { - *out = new(core_v1.ObjectReference) - **out = **in - } + *out = new(corev1.ObjectReference) + **out = **in } in.Template.DeepCopyInto(&out.Template) return @@ -322,30 +318,18 @@ func (in *RunAsUserStrategyOptions) DeepCopyInto(out *RunAsUserStrategyOptions) *out = *in if in.UID != nil { in, out := &in.UID, &out.UID - if *in == nil { - *out = nil - } else { - *out = new(int64) - **out = **in - } + *out = new(int64) + **out = **in } if in.UIDRangeMin != nil { in, out := &in.UIDRangeMin, &out.UIDRangeMin - if *in == nil { - *out = nil - } else { - *out = new(int64) - **out = **in - } + *out = new(int64) + **out = **in } if in.UIDRangeMax != nil { in, out := &in.UIDRangeMax, &out.UIDRangeMax - if *in == nil { - *out = nil - } else { - *out = new(int64) - **out = **in - } + *out = new(int64) + **out = **in } return } @@ -365,12 +349,8 @@ func (in *SELinuxContextStrategyOptions) DeepCopyInto(out *SELinuxContextStrateg *out = *in if in.SELinuxOptions != nil { in, out := &in.SELinuxOptions, &out.SELinuxOptions - if *in == nil { - *out = nil - } else { - *out = new(core_v1.SELinuxOptions) - **out = **in - } + *out = new(corev1.SELinuxOptions) + **out = **in } return } @@ -392,26 +372,22 @@ func (in *SecurityContextConstraints) DeepCopyInto(out *SecurityContextConstrain in.ObjectMeta.DeepCopyInto(&out.ObjectMeta) if in.Priority != nil { in, out := &in.Priority, &out.Priority - if *in == nil { - *out = nil - } else { - *out = new(int32) - **out = **in - } + *out = new(int32) + **out = **in } if in.DefaultAddCapabilities != nil { in, out := &in.DefaultAddCapabilities, &out.DefaultAddCapabilities - *out = make([]core_v1.Capability, len(*in)) + *out = make([]corev1.Capability, len(*in)) copy(*out, *in) } if in.RequiredDropCapabilities != nil { in, out := &in.RequiredDropCapabilities, &out.RequiredDropCapabilities - *out = make([]core_v1.Capability, len(*in)) + *out = make([]corev1.Capability, len(*in)) copy(*out, *in) } if in.AllowedCapabilities != nil { in, out := &in.AllowedCapabilities, &out.AllowedCapabilities - *out = make([]core_v1.Capability, len(*in)) + *out = make([]corev1.Capability, len(*in)) copy(*out, *in) } if in.Volumes != nil { @@ -426,21 +402,13 @@ func (in *SecurityContextConstraints) DeepCopyInto(out *SecurityContextConstrain } if in.DefaultAllowPrivilegeEscalation != nil { in, out := &in.DefaultAllowPrivilegeEscalation, &out.DefaultAllowPrivilegeEscalation - if *in == nil { - *out = nil - } else { - *out = new(bool) - **out = **in - } + *out = new(bool) + **out = **in } if in.AllowPrivilegeEscalation != nil { in, out := &in.AllowPrivilegeEscalation, &out.AllowPrivilegeEscalation - if *in == nil { - *out = nil - } else { - *out = new(bool) - **out = **in - } + *out = new(bool) + **out = **in } in.SELinuxContext.DeepCopyInto(&out.SELinuxContext) in.RunAsUser.DeepCopyInto(&out.RunAsUser) diff --git a/vendor/github.com/openshift/api/security/v1/zz_generated.swagger_doc_generated.go b/vendor/github.com/openshift/api/security/v1/zz_generated.swagger_doc_generated.go index 3a6db5284c..60a167915e 100644 --- a/vendor/github.com/openshift/api/security/v1/zz_generated.swagger_doc_generated.go +++ b/vendor/github.com/openshift/api/security/v1/zz_generated.swagger_doc_generated.go @@ -61,7 +61,7 @@ func (PodSecurityPolicyReviewSpec) SwaggerDoc() map[string]string { } var map_PodSecurityPolicyReviewStatus = map[string]string{ - "": "PodSecurityPolicyReviewStatus represents the status of PodSecurityPolicyReview.", + "": "PodSecurityPolicyReviewStatus represents the status of PodSecurityPolicyReview.", "allowedServiceAccounts": "allowedServiceAccounts returns the list of service accounts in *this* namespace that have the power to create the PodTemplateSpec.", } diff --git a/vendor/github.com/openshift/api/template/v1/generated.pb.go b/vendor/github.com/openshift/api/template/v1/generated.pb.go index 03380d8b6a..1de3d14173 100644 --- a/vendor/github.com/openshift/api/template/v1/generated.pb.go +++ b/vendor/github.com/openshift/api/template/v1/generated.pb.go @@ -1,6 +1,5 @@ -// Code generated by protoc-gen-gogo. +// Code generated by protoc-gen-gogo. DO NOT EDIT. // source: github.com/openshift/api/template/v1/generated.proto -// DO NOT EDIT! /* Package v1 is a generated protocol buffer package. @@ -758,24 +757,6 @@ func (m *TemplateList) MarshalTo(dAtA []byte) (int, error) { return i, nil } -func encodeFixed64Generated(dAtA []byte, offset int, v uint64) int { - dAtA[offset] = uint8(v) - dAtA[offset+1] = uint8(v >> 8) - dAtA[offset+2] = uint8(v >> 16) - dAtA[offset+3] = uint8(v >> 24) - dAtA[offset+4] = uint8(v >> 32) - dAtA[offset+5] = uint8(v >> 40) - dAtA[offset+6] = uint8(v >> 48) - dAtA[offset+7] = uint8(v >> 56) - return offset + 8 -} -func encodeFixed32Generated(dAtA []byte, offset int, v uint32) int { - dAtA[offset] = uint8(v) - dAtA[offset+1] = uint8(v >> 8) - dAtA[offset+2] = uint8(v >> 16) - dAtA[offset+3] = uint8(v >> 24) - return offset + 4 -} func encodeVarintGenerated(dAtA []byte, offset int, v uint64) int { for v >= 1<<7 { dAtA[offset] = uint8(v&0x7f | 0x80) @@ -2067,51 +2048,14 @@ func (m *Template) Unmarshal(dAtA []byte) error { if postIndex > l { return io.ErrUnexpectedEOF } - var keykey uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - keykey |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - var stringLenmapkey uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - stringLenmapkey |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - intStringLenmapkey := int(stringLenmapkey) - if intStringLenmapkey < 0 { - return ErrInvalidLengthGenerated - } - postStringIndexmapkey := iNdEx + intStringLenmapkey - if postStringIndexmapkey > l { - return io.ErrUnexpectedEOF - } - mapkey := string(dAtA[iNdEx:postStringIndexmapkey]) - iNdEx = postStringIndexmapkey if m.ObjectLabels == nil { m.ObjectLabels = make(map[string]string) } - if iNdEx < postIndex { - var valuekey uint64 + var mapkey string + var mapvalue string + for iNdEx < postIndex { + entryPreIndex := iNdEx + var wire uint64 for shift := uint(0); ; shift += 7 { if shift >= 64 { return ErrIntOverflowGenerated @@ -2121,41 +2065,80 @@ func (m *Template) Unmarshal(dAtA []byte) error { } b := dAtA[iNdEx] iNdEx++ - valuekey |= (uint64(b) & 0x7F) << shift + wire |= (uint64(b) & 0x7F) << shift if b < 0x80 { break } } - var stringLenmapvalue uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated + fieldNum := int32(wire >> 3) + if fieldNum == 1 { + var stringLenmapkey uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLenmapkey |= (uint64(b) & 0x7F) << shift + if b < 0x80 { + break + } } - if iNdEx >= l { + intStringLenmapkey := int(stringLenmapkey) + if intStringLenmapkey < 0 { + return ErrInvalidLengthGenerated + } + postStringIndexmapkey := iNdEx + intStringLenmapkey + if postStringIndexmapkey > l { return io.ErrUnexpectedEOF } - b := dAtA[iNdEx] - iNdEx++ - stringLenmapvalue |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break + mapkey = string(dAtA[iNdEx:postStringIndexmapkey]) + iNdEx = postStringIndexmapkey + } else if fieldNum == 2 { + var stringLenmapvalue uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLenmapvalue |= (uint64(b) & 0x7F) << shift + if b < 0x80 { + break + } } + intStringLenmapvalue := int(stringLenmapvalue) + if intStringLenmapvalue < 0 { + return ErrInvalidLengthGenerated + } + postStringIndexmapvalue := iNdEx + intStringLenmapvalue + if postStringIndexmapvalue > l { + return io.ErrUnexpectedEOF + } + mapvalue = string(dAtA[iNdEx:postStringIndexmapvalue]) + iNdEx = postStringIndexmapvalue + } else { + iNdEx = entryPreIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > postIndex { + return io.ErrUnexpectedEOF + } + iNdEx += skippy } - intStringLenmapvalue := int(stringLenmapvalue) - if intStringLenmapvalue < 0 { - return ErrInvalidLengthGenerated - } - postStringIndexmapvalue := iNdEx + intStringLenmapvalue - if postStringIndexmapvalue > l { - return io.ErrUnexpectedEOF - } - mapvalue := string(dAtA[iNdEx:postStringIndexmapvalue]) - iNdEx = postStringIndexmapvalue - m.ObjectLabels[mapkey] = mapvalue - } else { - var mapvalue string - m.ObjectLabels[mapkey] = mapvalue } + m.ObjectLabels[mapkey] = mapvalue iNdEx = postIndex default: iNdEx = preIndex @@ -2847,51 +2830,14 @@ func (m *TemplateInstanceRequester) Unmarshal(dAtA []byte) error { if postIndex > l { return io.ErrUnexpectedEOF } - var keykey uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - keykey |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - var stringLenmapkey uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - stringLenmapkey |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - intStringLenmapkey := int(stringLenmapkey) - if intStringLenmapkey < 0 { - return ErrInvalidLengthGenerated - } - postStringIndexmapkey := iNdEx + intStringLenmapkey - if postStringIndexmapkey > l { - return io.ErrUnexpectedEOF - } - mapkey := string(dAtA[iNdEx:postStringIndexmapkey]) - iNdEx = postStringIndexmapkey if m.Extra == nil { m.Extra = make(map[string]ExtraValue) } - if iNdEx < postIndex { - var valuekey uint64 + var mapkey string + mapvalue := &ExtraValue{} + for iNdEx < postIndex { + entryPreIndex := iNdEx + var wire uint64 for shift := uint(0); ; shift += 7 { if shift >= 64 { return ErrIntOverflowGenerated @@ -2901,46 +2847,85 @@ func (m *TemplateInstanceRequester) Unmarshal(dAtA []byte) error { } b := dAtA[iNdEx] iNdEx++ - valuekey |= (uint64(b) & 0x7F) << shift + wire |= (uint64(b) & 0x7F) << shift if b < 0x80 { break } } - var mapmsglen int - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated + fieldNum := int32(wire >> 3) + if fieldNum == 1 { + var stringLenmapkey uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLenmapkey |= (uint64(b) & 0x7F) << shift + if b < 0x80 { + break + } } - if iNdEx >= l { + intStringLenmapkey := int(stringLenmapkey) + if intStringLenmapkey < 0 { + return ErrInvalidLengthGenerated + } + postStringIndexmapkey := iNdEx + intStringLenmapkey + if postStringIndexmapkey > l { return io.ErrUnexpectedEOF } - b := dAtA[iNdEx] - iNdEx++ - mapmsglen |= (int(b) & 0x7F) << shift - if b < 0x80 { - break + mapkey = string(dAtA[iNdEx:postStringIndexmapkey]) + iNdEx = postStringIndexmapkey + } else if fieldNum == 2 { + var mapmsglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + mapmsglen |= (int(b) & 0x7F) << shift + if b < 0x80 { + break + } } + if mapmsglen < 0 { + return ErrInvalidLengthGenerated + } + postmsgIndex := iNdEx + mapmsglen + if mapmsglen < 0 { + return ErrInvalidLengthGenerated + } + if postmsgIndex > l { + return io.ErrUnexpectedEOF + } + mapvalue = &ExtraValue{} + if err := mapvalue.Unmarshal(dAtA[iNdEx:postmsgIndex]); err != nil { + return err + } + iNdEx = postmsgIndex + } else { + iNdEx = entryPreIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > postIndex { + return io.ErrUnexpectedEOF + } + iNdEx += skippy } - if mapmsglen < 0 { - return ErrInvalidLengthGenerated - } - postmsgIndex := iNdEx + mapmsglen - if mapmsglen < 0 { - return ErrInvalidLengthGenerated - } - if postmsgIndex > l { - return io.ErrUnexpectedEOF - } - mapvalue := &ExtraValue{} - if err := mapvalue.Unmarshal(dAtA[iNdEx:postmsgIndex]); err != nil { - return err - } - iNdEx = postmsgIndex - m.Extra[mapkey] = *mapvalue - } else { - var mapvalue ExtraValue - m.Extra[mapkey] = mapvalue } + m.Extra[mapkey] = *mapvalue iNdEx = postIndex default: iNdEx = preIndex diff --git a/vendor/github.com/openshift/api/template/v1/zz_generated.deepcopy.go b/vendor/github.com/openshift/api/template/v1/zz_generated.deepcopy.go index 7549c4ad78..3e76a1b4cb 100644 --- a/vendor/github.com/openshift/api/template/v1/zz_generated.deepcopy.go +++ b/vendor/github.com/openshift/api/template/v1/zz_generated.deepcopy.go @@ -5,7 +5,7 @@ package v1 import ( - core_v1 "k8s.io/api/core/v1" + corev1 "k8s.io/api/core/v1" runtime "k8s.io/apimachinery/pkg/runtime" ) @@ -280,12 +280,15 @@ func (in *TemplateInstanceRequester) DeepCopyInto(out *TemplateInstanceRequester in, out := &in.Extra, &out.Extra *out = make(map[string]ExtraValue, len(*in)) for key, val := range *in { + var outVal []string if val == nil { (*out)[key] = nil } else { - (*out)[key] = make([]string, len(val)) - copy((*out)[key], val) + in, out := &val, &outVal + *out = make(ExtraValue, len(*in)) + copy(*out, *in) } + (*out)[key] = outVal } } return @@ -307,21 +310,13 @@ func (in *TemplateInstanceSpec) DeepCopyInto(out *TemplateInstanceSpec) { in.Template.DeepCopyInto(&out.Template) if in.Secret != nil { in, out := &in.Secret, &out.Secret - if *in == nil { - *out = nil - } else { - *out = new(core_v1.LocalObjectReference) - **out = **in - } + *out = new(corev1.LocalObjectReference) + **out = **in } if in.Requester != nil { in, out := &in.Requester, &out.Requester - if *in == nil { - *out = nil - } else { - *out = new(TemplateInstanceRequester) - (*in).DeepCopyInto(*out) - } + *out = new(TemplateInstanceRequester) + (*in).DeepCopyInto(*out) } return } diff --git a/vendor/github.com/openshift/api/tools/genswaggertypedocs/swagger_type_docs.go b/vendor/github.com/openshift/api/tools/genswaggertypedocs/swagger_type_docs.go index 1d3e33d1f8..8aa77585b8 100644 --- a/vendor/github.com/openshift/api/tools/genswaggertypedocs/swagger_type_docs.go +++ b/vendor/github.com/openshift/api/tools/genswaggertypedocs/swagger_type_docs.go @@ -23,10 +23,10 @@ import ( "os" "path/filepath" - kruntime "k8s.io/apimachinery/pkg/runtime" - - "github.com/golang/glog" flag "github.com/spf13/pflag" + + kruntime "k8s.io/apimachinery/pkg/runtime" + "k8s.io/klog" ) var ( @@ -39,7 +39,7 @@ func main() { flag.Parse() if *typeSrc == "" { - glog.Fatalf("Please define -s flag as it is the source file") + klog.Fatalf("Please define -s flag as it is the source file") } var funcOut io.Writer @@ -48,7 +48,7 @@ func main() { } else { file, err := os.Create(*functionDest) if err != nil { - glog.Fatalf("Couldn't open %v: %v", *functionDest, err) + klog.Fatalf("Couldn't open %v: %v", *functionDest, err) } defer file.Close() funcOut = file @@ -58,14 +58,14 @@ func main() { if fi, err := os.Stat(*typeSrc); err == nil && !fi.IsDir() { docsForTypes = kruntime.ParseDocumentationFrom(*typeSrc) } else if err == nil && fi.IsDir() { - glog.Fatalf("-s must be a valid file or file glob pattern, not a directory") + klog.Fatalf("-s must be a valid file or file glob pattern, not a directory") } else { m, err := filepath.Glob(*typeSrc) if err != nil { - glog.Fatalf("Couldn't search for files matching -s: %v", err) + klog.Fatalf("Couldn't search for files matching -s: %v", err) } if len(m) == 0 { - glog.Fatalf("-s must be a valid file or file glob pattern") + klog.Fatalf("-s must be a valid file or file glob pattern") } for _, file := range m { docsForTypes = append(docsForTypes, kruntime.ParseDocumentationFrom(file)...) diff --git a/vendor/github.com/openshift/api/user/v1/generated.pb.go b/vendor/github.com/openshift/api/user/v1/generated.pb.go index 2c8d16e94d..172e4a7cb7 100644 --- a/vendor/github.com/openshift/api/user/v1/generated.pb.go +++ b/vendor/github.com/openshift/api/user/v1/generated.pb.go @@ -1,6 +1,5 @@ -// Code generated by protoc-gen-gogo. +// Code generated by protoc-gen-gogo. DO NOT EDIT. // source: github.com/openshift/api/user/v1/generated.proto -// DO NOT EDIT! /* Package v1 is a generated protocol buffer package. @@ -433,24 +432,6 @@ func (m *UserList) MarshalTo(dAtA []byte) (int, error) { return i, nil } -func encodeFixed64Generated(dAtA []byte, offset int, v uint64) int { - dAtA[offset] = uint8(v) - dAtA[offset+1] = uint8(v >> 8) - dAtA[offset+2] = uint8(v >> 16) - dAtA[offset+3] = uint8(v >> 24) - dAtA[offset+4] = uint8(v >> 32) - dAtA[offset+5] = uint8(v >> 40) - dAtA[offset+6] = uint8(v >> 48) - dAtA[offset+7] = uint8(v >> 56) - return offset + 8 -} -func encodeFixed32Generated(dAtA []byte, offset int, v uint32) int { - dAtA[offset] = uint8(v) - dAtA[offset+1] = uint8(v >> 8) - dAtA[offset+2] = uint8(v >> 16) - dAtA[offset+3] = uint8(v >> 24) - return offset + 4 -} func encodeVarintGenerated(dAtA []byte, offset int, v uint64) int { for v >= 1<<7 { dAtA[offset] = uint8(v&0x7f | 0x80) @@ -1093,51 +1074,14 @@ func (m *Identity) Unmarshal(dAtA []byte) error { if postIndex > l { return io.ErrUnexpectedEOF } - var keykey uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - keykey |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - var stringLenmapkey uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - stringLenmapkey |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - intStringLenmapkey := int(stringLenmapkey) - if intStringLenmapkey < 0 { - return ErrInvalidLengthGenerated - } - postStringIndexmapkey := iNdEx + intStringLenmapkey - if postStringIndexmapkey > l { - return io.ErrUnexpectedEOF - } - mapkey := string(dAtA[iNdEx:postStringIndexmapkey]) - iNdEx = postStringIndexmapkey if m.Extra == nil { m.Extra = make(map[string]string) } - if iNdEx < postIndex { - var valuekey uint64 + var mapkey string + var mapvalue string + for iNdEx < postIndex { + entryPreIndex := iNdEx + var wire uint64 for shift := uint(0); ; shift += 7 { if shift >= 64 { return ErrIntOverflowGenerated @@ -1147,41 +1091,80 @@ func (m *Identity) Unmarshal(dAtA []byte) error { } b := dAtA[iNdEx] iNdEx++ - valuekey |= (uint64(b) & 0x7F) << shift + wire |= (uint64(b) & 0x7F) << shift if b < 0x80 { break } } - var stringLenmapvalue uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated + fieldNum := int32(wire >> 3) + if fieldNum == 1 { + var stringLenmapkey uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLenmapkey |= (uint64(b) & 0x7F) << shift + if b < 0x80 { + break + } } - if iNdEx >= l { + intStringLenmapkey := int(stringLenmapkey) + if intStringLenmapkey < 0 { + return ErrInvalidLengthGenerated + } + postStringIndexmapkey := iNdEx + intStringLenmapkey + if postStringIndexmapkey > l { return io.ErrUnexpectedEOF } - b := dAtA[iNdEx] - iNdEx++ - stringLenmapvalue |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break + mapkey = string(dAtA[iNdEx:postStringIndexmapkey]) + iNdEx = postStringIndexmapkey + } else if fieldNum == 2 { + var stringLenmapvalue uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLenmapvalue |= (uint64(b) & 0x7F) << shift + if b < 0x80 { + break + } } + intStringLenmapvalue := int(stringLenmapvalue) + if intStringLenmapvalue < 0 { + return ErrInvalidLengthGenerated + } + postStringIndexmapvalue := iNdEx + intStringLenmapvalue + if postStringIndexmapvalue > l { + return io.ErrUnexpectedEOF + } + mapvalue = string(dAtA[iNdEx:postStringIndexmapvalue]) + iNdEx = postStringIndexmapvalue + } else { + iNdEx = entryPreIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > postIndex { + return io.ErrUnexpectedEOF + } + iNdEx += skippy } - intStringLenmapvalue := int(stringLenmapvalue) - if intStringLenmapvalue < 0 { - return ErrInvalidLengthGenerated - } - postStringIndexmapvalue := iNdEx + intStringLenmapvalue - if postStringIndexmapvalue > l { - return io.ErrUnexpectedEOF - } - mapvalue := string(dAtA[iNdEx:postStringIndexmapvalue]) - iNdEx = postStringIndexmapvalue - m.Extra[mapkey] = mapvalue - } else { - var mapvalue string - m.Extra[mapkey] = mapvalue } + m.Extra[mapkey] = mapvalue iNdEx = postIndex default: iNdEx = preIndex diff --git a/vendor/github.com/openshift/api/webconsole/v1/zz_generated.swagger_doc_generated.go b/vendor/github.com/openshift/api/webconsole/v1/zz_generated.swagger_doc_generated.go index 850d7c34d6..9e96d0774c 100644 --- a/vendor/github.com/openshift/api/webconsole/v1/zz_generated.swagger_doc_generated.go +++ b/vendor/github.com/openshift/api/webconsole/v1/zz_generated.swagger_doc_generated.go @@ -37,7 +37,7 @@ func (ExtensionsConfiguration) SwaggerDoc() map[string]string { } var map_FeaturesConfiguration = map[string]string{ - "": "FeaturesConfiguration defines various feature gates for the web console", + "": "FeaturesConfiguration defines various feature gates for the web console", "inactivityTimeoutMinutes": "InactivityTimeoutMinutes is the number of minutes of inactivity before you are automatically logged out of the web console (optional). If set to 0, inactivity timeout is disabled.", "clusterResourceOverridesEnabled": "ClusterResourceOverridesEnabled indicates that the cluster is configured for overcommit. When set to true, the web console will hide the CPU request, CPU limit, and memory request fields in its editors and skip validation on those fields. The memory limit field will still be displayed.", }