diff --git a/welcome/index.adoc b/welcome/index.adoc index 0aff3b364026..d57eeb7f0b65 100644 --- a/welcome/index.adoc +++ b/welcome/index.adoc @@ -10,7 +10,7 @@ Welcome to the official {product-title} {product-version} documentation, where y endif::openshift-rosa[] ifdef::openshift-rosa[] Welcome to the official {product-title} (ROSA) documentation, where you can learn about ROSA and start exploring its features. -To learn about ROSA, interacting with ROSA using {cluster-manager-first} and command-line interface (CLI) tools, consumption experience, and integration with Amazon Web Services (AWS) services, start with xref:../rosa_architecture/rosa-understanding.adoc#rosa-understanding[the Introduction to ROSA documentation]. +To learn about ROSA, interacting with ROSA by using {cluster-manager-first} and command-line interface (CLI) tools, consumption experience, and integration with Amazon Web Services (AWS) services, start with xref:../rosa_architecture/rosa-understanding.adoc#rosa-understanding[the Introduction to ROSA documentation]. image::291_OpenShift_on_AWS_Intro_1122_docs.png[{product-title}] endif::[] @@ -18,7 +18,7 @@ endif::[] ifdef::openshift-rosa[] To navigate the ROSA documentation, use the left navigation bar. -For documentation that is not ROSA-specific, see the link:https://docs.openshift.com/container-platform/latest/welcome/index.html[OpenShift Container Platform documentation]. +For documentation that is not ROSA-specific, see the link:https://docs.openshift.com/container-platform/latest/welcome/index.html[{product-title} documentation]. endif::[] ifndef::openshift-rosa[] @@ -40,31 +40,31 @@ endif::microshift[] ifdef::openshift-dpu[] To navigate the {product-title} data processing unit (DPU) documentation, use the left navigation bar. -For documentation that is not DPU-specific, see the link:https://docs.openshift.com/container-platform/latest/welcome/index.html[OpenShift Container Platform documentation]. +For documentation that is not DPU-specific, see the link:https://docs.openshift.com/container-platform/latest/welcome/index.html[{product-title} documentation]. endif::[] ifdef::openshift-dedicated[] To navigate the {product-title} documentation, use the left navigation bar. -For documentation that is not specific to {product-title}, see the link:https://docs.openshift.com/container-platform/latest/welcome/index.html[OpenShift Container Platform documentation]. +For documentation that is not specific to {product-title}, see the link:https://docs.openshift.com/container-platform/latest/welcome/index.html[{product-title} documentation]. endif::[] ifdef::microshift[] To navigate the {product-title} documentation, use the navigation bars and links. Start with xref:../microshift_getting_started/microshift-understanding.adoc#microshift-understanding[Understanding {product-title}] and xref:../microshift_install/microshift-install-rpm.adoc#microshift-install-rpm[Installing]. -Then, see the xref:../microshift_release_notes/microshift-4-13-release-notes.adoc#microshift-4-13-release-notes[release notes]. +Next, view the xref:../microshift_release_notes/microshift-4-13-release-notes.adoc#microshift-4-13-release-notes[release notes]. * For information about Red Hat Device Edge, read the link:https://access.redhat.com/documentation/en-us/red_hat_device_edge/4/html/overview/device-edge-overview[Red Hat Device Edge overview]. * For information about Red Hat Enterprise Linux for Edge, read link:https://access.redhat.com/documentation/en-us/red_hat_enterprise_linux/8/html/composing_installing_and_managing_rhel_for_edge_images/index[RHEL for Edge documentation]. -* For documentation that is not specific to {product-title}, read link:https://docs.openshift.com/container-platform/latest/welcome/index.html[OpenShift Container Platform documentation]. +* For documentation that is not specific to {product-title}, read link:https://docs.openshift.com/container-platform/latest/welcome/index.html[{product-title} documentation]. endif::[] ifdef::openshift-enterprise,openshift-webscale,openshift-origin[] Start with xref:../architecture/architecture.adoc#architecture-overview-architecture[Architecture] and xref:../security/container_security/security-understanding.adoc#understanding-security[Security and compliance]. ifdef::openshift-enterprise,openshift-webscale[] -Then, see the +Next, view the xref:../release_notes/ocp-4-13-release-notes.adoc#ocp-4-13-release-notes[release notes]. endif::[] endif::[] @@ -86,7 +86,7 @@ endif::[] ifdef::openshift-enterprise,openshift-webscale,openshift-origin[] == Cluster installer activities -Explore these {product-title} installation tasks. +Explore the following {product-title} installation tasks: - **xref:../installing/index.adoc#ocp-installation-overview[{product-title} installation overview]**: You can install {product-title} on installer-provisioned or user-provisioned infrastructure. The {product-title} installation program provides the flexibility to deploy {product-title} on a range of different platforms. @@ -109,7 +109,7 @@ endif::openshift-origin[] - **xref:../installing/installing_vsphere/installing-vsphere-installer-provisioned.adoc#installing-vsphere-installer-provisioned[Install a cluster on VMware vSphere]**: You can install {product-title} on supported versions of vSphere. -- **xref:../installing/installing_vmc/preparing-to-install-on-vmc.adoc#preparing-to-install-on-vmc[Install a cluster on VMware Cloud]**: You can install OpenShift Container Platform on supported versions of VMware Cloud (VMC) on AWS. +- **xref:../installing/installing_vmc/preparing-to-install-on-vmc.adoc#preparing-to-install-on-vmc[Install a cluster on VMware Cloud]**: You can install {product-title} on supported versions of VMware Cloud (VMC) on AWS. ifndef::openshift-origin[] - **xref:../installing/installing_ibm_z/installing-ibm-z.adoc#installing-ibm-z[Install a cluster with z/VM on {ibmzProductName} and {linuxoneProductName}]**: You can install {product-title} with z/VM on {ibmzProductName} and {linuxoneProductName} on user-provisioned infrastructure. @@ -119,7 +119,7 @@ endif::openshift-origin[] - **xref:../installing/installing_bare_metal_ipi/ipi-install-overview.adoc#ipi-install-overview[Install an installer-provisioned cluster on bare metal]**: You can install {product-title} on bare metal with an installer-provisioned architecture. -- **xref:../installing/installing_bare_metal/installing-bare-metal.adoc#installing-bare-metal[Install a user-provisioned cluster on bare metal]**: If none of the available platform and cloud provider deployment options meet your needs, you can install {product-title} on user-provisioned bare metal infrastructure. +- **xref:../installing/installing_bare_metal/installing-bare-metal.adoc#installing-bare-metal[Install a user-provisioned cluster on bare metal]**: If none of the available platform and cloud provider deployment options meet your needs, you can install {product-title} on user-provisioned, bare-metal infrastructure. - **Install a cluster on {rh-openstack-first}**: You can install a cluster on xref:../installing/installing_openstack/installing-openstack-installer-custom.adoc#installing-openstack-installer-custom[{rh-openstack} with customizations], with xref:../installing/installing_openstack/installing-openstack-installer-kuryr.adoc#installing-openstack-installer-kuryr[network customizations], or xref:../installing/installing_openstack/installing-openstack-installer-restricted.adoc#installing-openstack-installer-restricted[on a restricted network] on installer-provisioned infrastructure. @@ -146,9 +146,9 @@ xref:../installing/installing_vsphere/installing-restricted-networks-vsphere.ado xref:../installing/installing_ibm_z/installing-restricted-networks-ibm-z.adoc#installing-restricted-networks-ibm-z[{ibmzProductName} and {linuxoneProductName} with z/VM], xref:../installing/installing_ibm_z/installing-restricted-networks-ibm-z-kvm.adoc#installing-restricted-networks-ibm-z-kvm[{ibmzProductName} and {linuxoneProductName} with RHEL KVM], xref:../installing/installing_ibm_power/installing-restricted-networks-ibm-power.adoc#installing-restricted-networks-ibm-power[{ibmpowerProductName}], or xref:../installing/installing_bare_metal/installing-restricted-networks-bare-metal.adoc#installing-restricted-networks-bare-metal[bare metal] -does not have full access to the internet, then mirror the OpenShift Container Platform installation images using one of the following methods and install a cluster in a restricted network. +does not have full access to the internet, then mirror the {product-title} installation images by using one of the following methods and install a cluster in a restricted network. *** xref:../installing/disconnected_install/installing-mirroring-installation-images.adoc#installing-mirroring-installation-images[Mirroring images for a disconnected installation] -*** xref:../installing/disconnected_install/installing-mirroring-disconnected.adoc#installing-mirroring-disconnected[Mirroring images for a disconnected installation using the oc-mirror plug-in] +*** xref:../installing/disconnected_install/installing-mirroring-disconnected.adoc#installing-mirroring-disconnected[Mirroring images for a disconnected installation by using the oc-mirror plug-in] endif::openshift-origin[] @@ -168,15 +168,14 @@ endif::openshift-origin[] xref:../installing/installing_aws/installing-aws-vpc.adoc#installing-aws-vpc[AWS] or xref:../installing/installing_gcp/installing-gcp-vpc.adoc#installing-gcp-vpc[GCP] or an existing xref:../installing/installing_azure/installing-azure-vnet.adoc#installing-azure-vnet[VNet] -on Azure, you can install a cluster. +on Azure, you can install a cluster. You can install a cluster in xref:../installing/installing_gcp/installing-gcp-shared-vpc.adoc#installation-gcp-shared-vpc-prerequisites_installing-gcp-shared-vpc[Installing a cluster on GCP into a shared VPC] - **Install a private cluster**: If your cluster does not require external internet access, you can install a private cluster on xref:../installing/installing_aws/installing-aws-private.adoc#installing-aws-private[AWS], -xref:../installing/installing_azure/installing-azure-private.adoc#installing-aws-private[Azure], +xref:../installing/installing_azure/installing-azure-private.adoc#installing-aws-private[Azure], xref:../installing/installing_gcp/installing-gcp-private.adoc#installing-gcp-private[GCP], or -xref:../installing/installing_ibm_cloud_public/preparing-to-install-on-ibm-cloud.adoc#preparing-to-install-on-ibm-cloud[IBM Cloud VPC] -Internet access is still required to access the cloud APIs and installation media. +xref:../installing/installing_ibm_cloud_public/preparing-to-install-on-ibm-cloud.adoc#preparing-to-install-on-ibm-cloud[IBM Cloud VPC]. Internet access is still required to access the cloud APIs and installation media. - **xref:../installing/installing-troubleshooting.adoc#installing-troubleshooting[Check installation logs]**: Access installation logs to evaluate issues that occur during {product-title} installation. @@ -184,9 +183,9 @@ Internet access is still required to access the cloud APIs and installation medi - **xref:../storage/persistent_storage/persistent-storage-ocs.adoc#red-hat-openshift-data-foundation[Install Red Hat OpenShift Data Foundation]**: You can install {rh-storage-first} as an Operator to provide highly integrated and simplified persistent storage management for containers. -- **xref:../installing/installing_nutanix/preparing-to-install-on-nutanix.adoc#preparing-to-install-nutanix[Install a cluster on Nutanix]**: You can install a cluster on your Nutanix instance that uses installer-provisioned infrastructure. This type of installation lets you use the installation program to deploy a cluster on infrastructure that the installation program provisions and the cluster maintains. +- **xref:../installing/installing_nutanix/preparing-to-install-on-nutanix.adoc#preparing-to-install-nutanix[Install a cluster on Nutanix]**: You can install a cluster on your Nutanix instance that uses installer-provisioned infrastructure. With this type of installation, you can use the installation program to deploy a cluster on infrastructure that the installation program provisions and the cluster maintains. -- **xref:../post_installation_configuration/coreos-layering.adoc#coreos-layering[Red Hat Enterprise Linux CoreOS (RHCOS) image layering]** allows you to add new images on top of the base RHCOS image. This layering does not modify the base RHCOS image. Instead, it creates a custom layered image that includes all RHCOS functionality and adds additional functionality to specific nodes in the cluster. +- **xref:../post_installation_configuration/coreos-layering.adoc#coreos-layering[Red Hat Enterprise Linux CoreOS (RHCOS) image layering]** provides a way for you to add new images on top of the base RHCOS image. This layering does not modify the base RHCOS image. Instead, the layering creates a custom layered image that includes all RHCOS functionality and adds additional functionality to specific nodes in the cluster. endif::[] @@ -198,7 +197,7 @@ Develop and deploy containerized applications with {product-title}. {product-tit - **xref:../applications/projects/working-with-projects.adoc#working-with-projects[Work with projects]**: Create projects from the {product-title} web console or OpenShift CLI (`oc`) to organize and share the software you develop. -- **xref:../applications/creating_applications/odc-creating-applications-using-developer-perspective.adoc#odc-creating-applications-using-developer-perspective[Work with applications]**: +- **xref:../applications/creating_applications/odc-creating-applications-using-developer-perspective.adoc#odc-creating-applications-using-developer-perspective[Work with applications]**. - **Use the xref:../web_console/web-console-overview.adoc#about-developer-perspective_web-console-overview[*Developer perspective*]** in the {product-title} web console to xref:../applications/creating_applications/odc-creating-applications-using-developer-perspective.adoc#odc-creating-applications-using-developer-perspective[create and deploy applications]. @@ -206,34 +205,31 @@ xref:../applications/creating_applications/odc-creating-applications-using-devel - **Use the xref:../applications/odc-viewing-application-composition-using-topology-view.adoc#odc-viewing-application-composition-using-topology-view[*Topology*] view** to see your applications, monitor status, connect and group components, and modify your code base. -- **xref:../applications/connecting_applications_to_services/understanding-service-binding-operator.adoc#understanding-service-binding-operator[Connect your workloads to backing services]**: The Service Binding Operator enables application developers to easily bind workloads with Operator-managed backing services by automatically collecting and sharing binding data with the workloads. The Service Binding Operator improves the development lifecycle with a consistent and declarative service binding method that prevents discrepancies in cluster environments. - -- ** xref:../cli_reference/odo-important-update.adoc#odo-important_update[Use the developer CLI tool (`odo`)] **: -The `odo` CLI tool lets developers create single or multi-component applications easily and automates deployment, build, and service route configurations. It abstracts complex Kubernetes and {product-title} concepts, allowing you to focus on developing your applications. +- **xref:../applications/connecting_applications_to_services/understanding-service-binding-operator.adoc#understanding-service-binding-operator[Connect your workloads to backing services]**: With the Service Binding Operator, an application developer can bind workloads with Operator-managed backing services by automatically collecting and sharing binding data with the workloads. The Service Binding Operator improves the development lifecycle with a consistent and declarative service binding method that prevents discrepancies in cluster environments. - **xref:../cicd/pipelines/understanding-openshift-pipelines.adoc#op-key-features[Create CI/CD Pipelines]**: Pipelines are serverless, cloud-native, continuous integration and continuous deployment systems that run in isolated containers. -They use standard Tekton custom resources to automate deployments and are designed for decentralized teams that work on microservice-based architecture. +Pipelines use standard Tekton custom resources to automate deployments and are designed for decentralized teams that work on microservice-based architecture. ifdef::openshift-enterprise,openshift-webscale,openshift-origin[] -- **xref:../cicd/gitops/understanding-openshift-gitops.adoc#understanding-openshift-gitops[Manage your infrastructure and application configurations]**: GitOps is a declarative way to implement continuous deployment for cloud native applications. GitOps defines infrastructure and application definitions as code. Then, it uses this code to manage multiple workspaces and clusters to simplify the creation of infrastructure and application configurations. GitOps also handles and automates complex deployments at a fast pace, saving time during deployment and release cycles. +- **xref:../cicd/gitops/understanding-openshift-gitops.adoc#understanding-openshift-gitops[Manage your infrastructure and application configurations]**: GitOps is a declarative way to implement continuous deployment for cloud native applications. GitOps defines infrastructure and application definitions as code. GitOps uses this code to manage multiple workspaces and clusters to simplify the creation of infrastructure and application configurations. GitOps also handles and automates complex deployments at a fast pace, which saves time during deployment and release cycles. - **xref:../applications/working_with_helm_charts/configuring-custom-helm-chart-repositories.adoc#installing-a-helm-chart-on-an-openshift-cluster_configuring-custom-helm-chart-repositories[Deploy Helm charts]**: -xref:../applications/working_with_helm_charts/understanding-helm.adoc#understanding-helm[Helm] is a software package manager that simplifies deployment of applications and services to OpenShift Container Platform clusters. Helm uses a packaging format called charts. A Helm chart is a collection of files that describes the OpenShift Container Platform resources. +xref:../applications/working_with_helm_charts/understanding-helm.adoc#understanding-helm[Helm] is a software package manager that simplifies deployment of applications and services to {product-title} clusters. Helm uses a packaging format called _charts_. A Helm chart is a collection of files that describes the {product-title} resources. -- **xref:../cicd/builds/understanding-image-builds.adoc#understanding-image-builds[Understand image builds]**: Choose from different build strategies (Docker, S2I, custom, and pipeline) that can include different kinds of source materials (from places like Git repositories, local binary inputs, and external artifacts). Then, follow examples of build types from basic builds to advanced builds. +- **xref:../cicd/builds/understanding-image-builds.adoc#understanding-image-builds[Understand image builds]**: Choose from different build strategies (Docker, S2I, custom, and pipeline) that can include different kinds of source materials, such as Git repositories, local binary inputs, and external artifacts. You can follow examples of build types from basic builds to advanced builds. -- **xref:../openshift_images/index.adoc#overview-of-images[Create container images]**: A container image is the most basic building block in {product-title} (and Kubernetes) applications. Defining image streams lets you gather multiple versions of an image in one place as you continue its development. S2I containers let you insert your source code into a base container that is set up to run code of a particular type, such as Ruby, Node.js, or Python. +- **xref:../openshift_images/index.adoc#overview-of-images[Create container images]**: A container image is the most basic building block in {product-title} and Kubernetes applications. By defining image streams, you can gather multiple versions of an image in one place as you continue to development the image stream. With S2I containers, you can insert your source code into a base container. The base container is configured to run code of a particular type, such as Ruby, Node.js, or Python. - **xref:../applications/deployments/what-deployments-are.adoc#what-deployments-are[Create deployments]**: Use `Deployment` and `DeploymentConfig` objects to exert fine-grained management over applications. - xref:../applications/deployments/managing-deployment-processes.adoc#deployment-operations[Manage deployments] using the *Workloads* page or OpenShift CLI (`oc`). Learn xref:../applications/deployments/deployment-strategies.adoc#deployment-strategies[rolling, recreate, and custom] deployment strategies. + xref:../applications/deployments/managing-deployment-processes.adoc#deployment-operations[Manage deployments] by using the *Workloads* page or OpenShift CLI (`oc`). Learn xref:../applications/deployments/deployment-strategies.adoc#deployment-strategies[rolling, recreate, and custom] deployment strategies. - **xref:../openshift_images/using-templates.adoc#using-templates[Create templates]**: Use existing templates or create your own templates that describe how an application is built or deployed. A template can combine images with descriptions, parameters, replicas, exposed ports and other content that defines how an application can be run or built. -- **xref:../operators/understanding/olm-what-operators-are.adoc#olm-what-operators-are[Understand Operators]**: Operators are the preferred method for creating on-cluster applications for {product-title} {product-version}. Learn about the Operator Framework and how to deploy applications using installed Operators into your projects. +- **xref:../operators/understanding/olm-what-operators-are.adoc#olm-what-operators-are[Understand Operators]**: Operators are the preferred method for creating on-cluster applications for {product-title} {product-version}. Learn about the Operator Framework and how to deploy applications by using installed Operators into your projects. -- **xref:../operators/operator_sdk/osdk-about.adoc#osdk-about[Develop Operators]**: Operators are the preferred method for creating on-cluster applications for {product-title} {product-version}. Learn the workflow for building, testing, and deploying Operators. Then, create your own Operators based on xref:../operators/operator_sdk/ansible/osdk-ansible-support.adoc#osdk-ansible-support[Ansible] or -xref:../operators/operator_sdk/helm/osdk-helm-support.adoc#osdk-helm-support[Helm], or configure xref:../operators/operator_sdk/osdk-monitoring-prometheus.adoc#osdk-monitoring-prometheus[built-in Prometheus monitoring] using the Operator SDK. +- **xref:../operators/operator_sdk/osdk-about.adoc#osdk-about[Develop Operators]**: Operators are the preferred method for creating on-cluster applications for {product-title} {product-version}. Learn the workflow for building, testing, and deploying Operators. You can then create your own Operators based on xref:../operators/operator_sdk/ansible/osdk-ansible-support.adoc#osdk-ansible-support[Ansible] or +xref:../operators/operator_sdk/helm/osdk-helm-support.adoc#osdk-helm-support[Helm], or configure xref:../operators/operator_sdk/osdk-monitoring-prometheus.adoc#osdk-monitoring-prometheus[built-in Prometheus monitoring] by using the Operator SDK. - **xref:../rest_api/index.adoc#api-index[REST API reference]**: Learn about {product-title} application programming interface endpoints. endif::[] @@ -241,27 +237,23 @@ endif::[] ifdef::openshift-dedicated[] == Developer activities -Ultimately, {product-title} is a platform for developing and deploying containerized applications. As an application developer, this documentation helps you: +{product-title} is a platform for developing and deploying containerized applications. As an application developer, you can use this documentation to complete the following tasks: - *Understand {product-title} development*: Learn the different types of containerized applications, from simple containers to advanced Kubernetes deployments and Operators. - *Work with projects*: Create projects from the web console or CLI to organize and share the software you develop. -- *Work with applications*: Use the Developer perspective in the {product-title} web console to easily create and deploy applications. Use the Topology view to visually interact with your applications, monitor status, connect and group components, and modify your code base. - -- *Use the developer CLI tool (odo)*: The odo CLI tool lets developers create single or multi-component applications easily and automates deployment, build, and service route configurations. -It abstracts complex Kubernetes and {product-title} concepts, allowing developers to focus on developing their applications. +- *Work with applications*: Use the Developer perspective in the {product-title} web console to easily create and deploy applications. Use the *Topology* view to visually interact with your applications, monitor status, connect and group components, and modify your code base. -- *Create CI/CD Pipelines*: Pipelines are serverless, cloud-native, continuous integration and continuous deployment systems that run in isolated containers. They use standard Tekton custom resources to automate deployments and are designed for decentralized teams that work on microservices-based architecture. +- *Create CI/CD Pipelines*: Pipelines are serverless, cloud-native, continuous integration and continuous deployment systems that run in isolated containers. Pipelines use standard Tekton custom resources to automate deployments and are designed for decentralized teams that work on microservices-based architecture. -- *Understand Operators*: Operators are the preferred method for creating on-cluster applications for {product-title} {product-version}. Learn about the Operator Framework and how to deploy applications using installed Operators into your projects. +- *Understand Operators*: Operators are the preferred method for creating on-cluster applications for {product-title} {product-version}. Learn about the Operator Framework and how to deploy applications by using installed Operators into your projects. -- *Understand image builds*: Choose from different build strategies (Docker, S2I, custom, and pipeline) that can include different kinds of source materials (from places like Git repositories, local binary inputs, and external artifacts). Then, follow examples of build types from basic builds to advanced builds. +- *Understand image builds*: Choose from different build strategies (Docker, S2I, custom, and pipeline) that can include different kinds of source materials, such as Git repositories, local binary inputs, and external artifacts. You can follow examples of build types from basic builds to advanced builds. -- *Create container images*: A container image is the most basic building block in {product-title} (and Kubernetes) applications. Defining image streams lets you gather multiple versions of an image in one place as you continue its development. S2I containers let you insert your source code into a base container that is set up to run code of a particular type (such as Ruby, Node.js, or Python). +- *Create container images*: A container image is the most basic building block in {product-title} (and Kubernetes) applications. By defining image streams, you can gather multiple versions of an image in one place as you continue its development. With S2I containers, you can insert your source code into a base container that is set up to run code of a particular type (such as Ruby, Node.js, or Python). -- *Create deployments*: Use `Deployment` and `DeploymentConfig` objects to exert fine-grained management over applications. -Use the Workloads page or `oc` CLI to manage deployments. Learn rolling, recreate, and custom deployment strategies. +- *Create deployments*: Use `Deployment` and `DeploymentConfig` objects to exert fine-grained management over applications. Use the Workloads page or `oc` CLI to manage deployments. Learn rolling, recreate, and custom deployment strategies. - *Create templates*: Use existing templates or create your own templates that describe how an application is built or deployed. A template can combine images with descriptions, parameters, replicas, exposed ports and other content that defines how an application can be run or built. endif::[] @@ -269,12 +261,12 @@ endif::[] ifdef::openshift-enterprise,openshift-webscale,openshift-origin[] == Cluster administrator activities -As a cluster administrator for {product-title}, this documentation helps you: +Manage machines, provide services to users, and follow monitoring and logging reports. This documentation helps you: - **xref:../architecture/architecture.adoc#architecture-overview-architecture[Understand {product-title} management]**: Learn about components of the {product-title} {product-version} control plane. See how {product-title} control plane and worker nodes are managed and updated through the xref:../machine_management/creating_machinesets/creating-machineset-aws.adoc#machine-api-overview_creating-machineset-aws[Machine API] and xref:../architecture/control-plane.adoc#operators-overview_control-plane[Operators]. -- **Enable cluster capabilities that were disabled prior to installation** Cluster administrators can enable cluster capabilities that were disabled prior to installation. For more information, see xref:../post_installation_configuration/enabling-cluster-capabilities.adoc#enabling-cluster-capabilities[Enabling cluster capabilities]. +- **Enable cluster capabilities that were disabled prior to installation** Cluster administrators can enable cluster capabilities that were disabled prior to installation. For more information, see xref:../post_installation_configuration/enabling-cluster-capabilities.adoc#enabling-cluster-capabilities[Enabling cluster capabilities]. === Manage cluster components @@ -285,21 +277,18 @@ or xref:../machine_management/creating_machinesets/creating-machineset-gcp.adoc#creating-machineset-gcp[GCP] by xref:../machine_management/deploying-machine-health-checks.adoc#deploying-machine-health-checks[deploying health checks] and xref:../machine_management/applying-autoscaling.adoc#applying-autoscaling[applying autoscaling to machines]. -- **xref:../registry/index.adoc#registry-overview[Manage container registries]**: Each {product-title} cluster includes a built-in container registry for storing its images. You can also configure a separate link:https://access.redhat.com/documentation/en-us/red_hat_quay/[Red Hat Quay] registry to use with {product-title}. The link:https://quay.io[Quay.io] web site provides a public container registry that stores {product-title} containers and Operators. +- **xref:../registry/index.adoc#registry-overview[Manage container registries]**: Each {product-title} cluster includes a built-in container registry for storing its images. You can also configure a separate link:https://access.redhat.com/documentation/en-us/red_hat_quay/[Red Hat Quay] registry to use with {product-title}. The link:https://quay.io[Quay.io] website provides a public container registry that stores {product-title} containers and Operators. - **xref:../authentication/understanding-authentication.adoc#understanding-authentication[Manage users and groups]**: Add users and groups with different levels of permissions to use or modify clusters. - **xref:../authentication/understanding-authentication.adoc#understanding-authentication[Manage authentication]**: Learn how user, group, and API authentication works in {product-title}. {product-title} supports xref:../authentication/understanding-identity-provider.adoc#supported-identity-providers[multiple identity providers]. -- **Manage xref:../security/certificates/replacing-default-ingress-certificate.adoc#replacing-default-ingress[ingress], xref:../security/certificates/api-server.adoc#api-server-certificates[API server], and xref:../security/certificates/service-serving-certificate.adoc#add-service-serving[service] certificates**: {product-title} creates certificates by default for the Ingress Operator, the API server, and for -services needed by complex middleware applications that require encryption. You might need to change, add, or -rotate these certificates. +- **Manage xref:../security/certificates/replacing-default-ingress-certificate.adoc#replacing-default-ingress[ingress], xref:../security/certificates/api-server.adoc#api-server-certificates[API server], and xref:../security/certificates/service-serving-certificate.adoc#add-service-serving[service] certificates**: {product-title} creates certificates by default for the Ingress Operator, the API server, and for services needed by complex middleware applications that require encryption. You might need to change, add, or rotate these certificates. -- **xref:../networking/understanding-networking.adoc#understanding-networking[Manage networking]**: The cluster network in {product-title} is managed by the xref:../networking/cluster-network-operator.adoc#cluster-network-operator[Cluster Network Operator] (CNO). The CNO uses iptables rules in xref:../networking/openshift_sdn/configuring-kube-proxy.adoc#configuring-kube-proxy[kube-proxy] to direct traffic between nodes and pods running on those nodes. The Multus Container Network Interface adds the capability to attach xref:../networking/multiple_networks/understanding-multiple-networks.adoc#understanding-multiple-networks[multiple network interfaces] to a pod. Using +- **xref:../networking/understanding-networking.adoc#understanding-networking[Manage networking]**: The cluster network in {product-title} is managed by the xref:../networking/cluster-network-operator.adoc#cluster-network-operator[Cluster Network Operator] (CNO). The CNO uses `iptables` rules in xref:../networking/openshift_sdn/configuring-kube-proxy.adoc#configuring-kube-proxy[kube-proxy] to direct traffic between nodes and pods running on those nodes. The Multus Container Network Interface adds the capability to attach xref:../networking/multiple_networks/understanding-multiple-networks.adoc#understanding-multiple-networks[multiple network interfaces] to a pod. By using xref:../networking/network_policy/about-network-policy.adoc#about-network-policy[network policy] features, you can isolate your pods or permit selected traffic. -- **xref:../storage/understanding-persistent-storage.adoc#understanding-persistent-storage[Manage storage]**: {product-title} allows cluster administrators to -configure persistent storage using +- **xref:../storage/understanding-persistent-storage.adoc#understanding-persistent-storage[Manage storage]**: With {product-title}, a cluster administrator can configure persistent storage by using xref:../storage/persistent_storage/persistent-storage-ocs.adoc#red-hat-openshift-data-foundation[Red Hat OpenShift Data Foundation], xref:../storage/persistent_storage/persistent-storage-aws.adoc#persistent-storage-using-aws-ebs[AWS Elastic Block Store], xref:../storage/persistent_storage/persistent-storage-nfs.adoc#persistent-storage-using-nfs[NFS], @@ -308,8 +297,12 @@ xref:../storage/container_storage_interface/persistent-storage-csi.adoc#persiste and more. You can xref:../storage/expanding-persistent-volumes.adoc#expanding-persistent-volumes[expand persistent volumes], configure xref:../storage/dynamic-provisioning.adoc#dynamic-provisioning[dynamic provisioning], and use CSI to xref:../storage/container_storage_interface/persistent-storage-csi.adoc#persistent-storage-using-csi[configure], xref:../storage/container_storage_interface/persistent-storage-csi-cloning.adoc#persistent-storage-csi-cloning[clone], and use xref:../storage/container_storage_interface/persistent-storage-csi-snapshots.adoc#persistent-storage-csi-snapshots[snapshots] of persistent storage. -- **xref:../operators/understanding/olm-understanding-operatorhub.adoc#olm-understanding-operatorhub[Manage Operators]**: Lists of Red Hat, ISV, and community Operators can -be reviewed by cluster administrators and xref:../operators/admin/olm-adding-operators-to-cluster.adoc#olm-adding-operators-to-a-cluster[installed on their clusters]. After you install them, you can xref:../operators/user/olm-creating-apps-from-installed-operators.adoc#olm-creating-apps-from-installed-operators[run], xref:../operators/admin/olm-upgrading-operators.adoc#olm-upgrading-operators[upgrade], back up, or otherwise manage the Operator on your cluster. +- **xref:../operators/understanding/olm-understanding-operatorhub.adoc#olm-understanding-operatorhub[Manage Operators]**: Lists of Red Hat, ISV, and community Operators can be reviewed by cluster administrators and xref:../operators/admin/olm-adding-operators-to-cluster.adoc#olm-adding-operators-to-a-cluster[installed on their clusters]. After you install them, you can xref:../operators/user/olm-creating-apps-from-installed-operators.adoc#olm-creating-apps-from-installed-operators[run], xref:../operators/admin/olm-upgrading-operators.adoc#olm-upgrading-operators[upgrade], back up, or otherwise manage the Operator on your cluster. + +// Check in with M.Burke on May 10, 2023 +//// +- **xref:../windows_containers/understanding-windows-container-workloads.html#understanding-windows-container-workloads_understanding-windows-container-workloads[Understanding Windows container workloads]**. You can use the {productwinc} feature to run Windows compute nodes in an {product-title} cluster. This is possible by using the Red Hat Windows Machine Config Operator (WMCO) to install and manage Windows nodes. +//// === Change cluster components @@ -333,7 +326,7 @@ xref:../updating/updating-disconnected-cluster.adoc#updating-disconnected-cluste - **xref:../updating/updating-restricted-network-cluster/restricted-network-update-osus.adoc#update-service-overview_updating-restricted-network-cluster-osus[Understanding the OpenShift Update Service]**: Learn about installing and managing a local OpenShift Update Service for recommending {product-title} updates in disconnected environments. -- **xref:../nodes/clusters/nodes-cluster-worker-latency-profiles.adoc#nodes-cluster-worker-latency-profiles[Improving cluster stability in high latency environments using worker latency profiles]**: If your network has latency issues, you can use one of three _worker latency profiles_ to help ensure that your control plane does not accidentally evict pods in case it cannot reach a worker node. You can configure or modify the profile at any time during the life of the cluster. +- **xref:../nodes/clusters/nodes-cluster-worker-latency-profiles.adoc#nodes-cluster-worker-latency-profiles[Improving cluster stability in high latency environments by using worker latency profiles]**: If your network has latency issues, you can use one of three _worker latency profiles_ to help ensure that your control plane does not accidentally evict pods in case it cannot reach a worker node. You can configure or modify the profile at any time during the life of the cluster. === Monitor the cluster @@ -343,7 +336,7 @@ xref:../updating/updating-disconnected-cluster.adoc#updating-disconnected-cluste Learn to xref:../monitoring/configuring-the-monitoring-stack.adoc#configuring-the-monitoring-stack[configure the monitoring stack]. After configuring monitoring, use the web console to access xref:../monitoring/reviewing-monitoring-dashboards.adoc#reviewing-monitoring-dashboards[monitoring dashboards]. In addition to infrastructure metrics, you can also scrape and view metrics for your own services. -- **xref:../support/remote_health_monitoring/about-remote-health-monitoring.adoc#about-remote-health-monitoring_about-remote-health-monitoring[Remote health monitoring]**: {product-title} collects anonymized aggregated information about your cluster. Using Telemetry and the Insights Operator, this data is received by Red Hat and used to improve {product-title}. You can view the xref:../support/remote_health_monitoring/showing-data-collected-by-remote-health-monitoring.adoc#showing-data-collected-by-remote-health-monitoring_showing-data-collected-by-remote-health-monitoring[data collected by remote health monitoring]. +- **xref:../support/remote_health_monitoring/about-remote-health-monitoring.adoc#about-remote-health-monitoring_about-remote-health-monitoring[Remote health monitoring]**: {product-title} collects anonymized aggregated information about your cluster. By using Telemetry and the Insights Operator, this data is received by Red Hat and used to improve {product-title}. You can view the xref:../support/remote_health_monitoring/showing-data-collected-by-remote-health-monitoring.adoc#showing-data-collected-by-remote-health-monitoring_showing-data-collected-by-remote-health-monitoring[data collected by remote health monitoring]. endif::[] ifdef::openshift-dedicated[]