$ cat <install_dir>/.openshift_install.log
You can check the status of an OKD cluster after an installation by following the procedures in this document.
You can review a summary of an installation in the OKD installation log. If an installation succeeds, the information required to access the cluster is included in the log.
You have access to the installation host.
Review the .openshift_install.log
log file in the installation directory on your installation host:
$ cat <install_dir>/.openshift_install.log
Cluster credentials are included at the end of the log if the installation is successful, as outlined in the following example:
...
time="2020-12-03T09:50:47Z" level=info msg="Install complete!"
time="2020-12-03T09:50:47Z" level=info msg="To access the cluster as the system:admin user when using 'oc', run 'export KUBECONFIG=/home/myuser/install_dir/auth/kubeconfig'"
time="2020-12-03T09:50:47Z" level=info msg="Access the OpenShift web-console here: https://console-openshift-console.apps.mycluster.example.com"
time="2020-12-03T09:50:47Z" level=info msg="Login to the console with user: \"kubeadmin\", and password: \"password\""
time="2020-12-03T09:50:47Z" level=debug msg="Time elapsed per stage:"
time="2020-12-03T09:50:47Z" level=debug msg=" Infrastructure: 6m45s"
time="2020-12-03T09:50:47Z" level=debug msg="Bootstrap Complete: 11m30s"
time="2020-12-03T09:50:47Z" level=debug msg=" Bootstrap Destroy: 1m5s"
time="2020-12-03T09:50:47Z" level=debug msg=" Cluster Operators: 17m31s"
time="2020-12-03T09:50:47Z" level=info msg="Time elapsed: 37m26s"
For clusters with unrestricted network connectivity, you can view the source of your pulled images by using a command on a node, such as crictl images
.
However, for disconnected installations, to view the source of pulled images, you must review the CRI-O logs to locate the Trying to access
log entry, as shown in the following procedure. Other methods to view the image pull source, such as the crictl images
command, show the non-mirrored image name, even though the image is pulled from the mirrored location.
You have access to the cluster as a user with the cluster-admin
role.
Review the CRI-O logs for a master or worker node:
$ oc adm node-logs <node_name> -u crio
The Trying to access
log entry indicates where the image is being pulled from.
...
Mar 17 02:52:50 ip-10-0-138-140.ec2.internal crio[1366]: time="2021-08-05 10:33:21.594930907Z" level=info msg="Pulling image: quay.io/openshift-release-dev/ocp-release:4.10.0-ppc64le" id=abcd713b-d0e1-4844-ac1c-474c5b60c07c name=/runtime.v1alpha2.ImageService/PullImage
Mar 17 02:52:50 ip-10-0-138-140.ec2.internal crio[1484]: time="2021-03-17 02:52:50.194341109Z" level=info msg="Trying to access \"li0317gcp1.mirror-registry.qe.gcp.devcluster.openshift.com:5000/ocp/release@sha256:1926eae7cacb9c00f142ec98b00628970e974284b6ddaf9a6a086cb9af7a6c31\""
Mar 17 02:52:50 ip-10-0-138-140.ec2.internal crio[1484]: time="2021-03-17 02:52:50.226788351Z" level=info msg="Trying to access \"li0317gcp1.mirror-registry.qe.gcp.devcluster.openshift.com:5000/ocp/release@sha256:1926eae7cacb9c00f142ec98b00628970e974284b6ddaf9a6a086cb9af7a6c31\""
...
The log might show the image pull source twice, as shown in the preceding example.
If your ImageContentSourcePolicy
object lists multiple mirrors, OKD attempts to pull the images in the order listed in the configuration, for example:
Trying to access \"li0317gcp1.mirror-registry.qe.gcp.devcluster.openshift.com:5000/ocp/release@sha256:1926eae7cacb9c00f142ec98b00628970e974284b6ddaf9a6a086cb9af7a6c31\" Trying to access \"li0317gcp2.mirror-registry.qe.gcp.devcluster.openshift.com:5000/ocp/release@sha256:1926eae7cacb9c00f142ec98b00628970e974284b6ddaf9a6a086cb9af7a6c31\"
You can view the cluster version and status by running the oc get clusterversion
command. If the status shows that the installation is still progressing, you can review the status of the Operators for more information.
You can also list the current update channel and review the available cluster updates.
You have access to the cluster as a user with the cluster-admin
role.
You have installed the OpenShift CLI (oc
).
Obtain the cluster version and overall status:
$ oc get clusterversion
NAME VERSION AVAILABLE PROGRESSING SINCE STATUS
version 4.6.4 True False 6m25s Cluster version is 4.6.4
The example output indicates that the cluster has been installed successfully.
If the cluster status indicates that the installation is still progressing, you can obtain more detailed progress information by checking the status of the Operators:
$ oc get clusteroperators.config.openshift.io
View a detailed summary of cluster specifications, update availability, and update history:
$ oc describe clusterversion
List the current update channel:
$ oc get clusterversion -o jsonpath='{.items[0].spec}{"\n"}'
{"channel":"stable-4.6","clusterID":"245539c1-72a3-41aa-9cec-72ed8cf25c5c"}
Review the available cluster updates:
$ oc adm upgrade
Cluster version is 4.6.4
Updates:
VERSION IMAGE
4.6.6 quay.io/openshift-release-dev/ocp-release@sha256:c7e8f18e8116356701bd23ae3a23fb9892dd5ea66c8300662ef30563d7104f39
See Querying Operator status after installation for more information about querying Operator status if your installation is still progressing.
See Troubleshooting Operator issues for information about investigating issues with Operators.
See Updating a cluster using the web console for more information on updating your cluster.
See Understanding update channels and releases for an overview about update release channels.
You can verify that a cluster uses short-term security credentials for individual components by checking the Cloud Credential Operator (CCO) configuration and other values in the cluster.
You deployed an OKD cluster using the Cloud Credential Operator utility (ccoctl
) to implement short-term credentials.
You installed the OpenShift CLI (oc
).
You are logged in as a user with cluster-admin
privileges.
Verify that the CCO is configured to operate in manual mode by running the following command:
$ oc get cloudcredentials cluster \
-o=jsonpath={.spec.credentialsMode}
The following output confirms that the CCO is operating in manual mode:
Manual
Verify that the cluster does not have root
credentials by running the following command:
$ oc get secrets \
-n kube-system <secret_name>
where <secret_name>
is the name of the root secret for your cloud provider.
Platform | Secret name |
---|---|
Amazon Web Services (AWS) |
|
Microsoft Azure |
|
Google Cloud Platform (GCP) |
|
An error confirms that the root secret is not present on the cluster.
Error from server (NotFound): secrets "aws-creds" not found
Verify that the components are using short-term security credentials for individual components by running the following command:
$ oc get authentication cluster \
-o jsonpath \
--template='{ .spec.serviceAccountIssuer }'
This command displays the value of the .spec.serviceAccountIssuer
parameter in the cluster Authentication
object.
An output of a URL that is associated with your cloud provider indicates that the cluster is using manual mode with short-term credentials that are created and managed from outside of the cluster.
Azure clusters: Verify that the components are assuming the Azure client ID that is specified in the secret manifests by running the following command:
$ oc get secrets \
-n openshift-image-registry installer-cloud-credentials \
-o jsonpath='{.data}'
An output that contains the azure_client_id
and azure_federated_token_file
felids confirms that the components are assuming the Azure client ID.
Azure clusters: Verify that the pod identity webhook is running by running the following command:
$ oc get pods \
-n openshift-cloud-credential-operator
NAME READY STATUS RESTARTS AGE
cloud-credential-operator-59cf744f78-r8pbq 2/2 Running 2 71m
pod-identity-webhook-548f977b4c-859lz 1/1 Running 1 70m
You can verify the status of the cluster nodes after an installation.
You have access to the cluster as a user with the cluster-admin
role.
You have installed the OpenShift CLI (oc
).
List the status of the cluster nodes. Verify that the output lists all of the expected control plane and compute nodes and that each node has a Ready
status:
$ oc get nodes
NAME STATUS ROLES AGE VERSION
compute-1.example.com Ready worker 33m v1.30.3
control-plane-1.example.com Ready master 41m v1.30.3
control-plane-2.example.com Ready master 45m v1.30.3
compute-2.example.com Ready worker 38m v1.30.3
compute-3.example.com Ready worker 33m v1.30.3
control-plane-3.example.com Ready master 41m v1.30.3
Review CPU and memory resource availability for each cluster node:
$ oc adm top nodes
NAME CPU(cores) CPU% MEMORY(bytes) MEMORY%
compute-1.example.com 128m 8% 1132Mi 16%
control-plane-1.example.com 801m 22% 3471Mi 23%
control-plane-2.example.com 1718m 49% 6085Mi 40%
compute-2.example.com 935m 62% 5178Mi 75%
compute-3.example.com 111m 7% 1131Mi 16%
control-plane-3.example.com 942m 26% 4100Mi 27%
See Verifying node health for more details about reviewing node health and investigating node issues.
You can review the following information in the Overview page in the OKD web console:
The general status of your cluster
The status of the control plane, cluster Operators, and storage
CPU, memory, file system, network transfer, and pod availability
The API address of the cluster, the cluster ID, and the name of the provider
Cluster version information
Cluster update status, including details of the current update channel and available updates
A cluster inventory detailing node, pod, storage class, and persistent volume claim (PVC) information
A list of ongoing cluster activities and recent events
You have access to the cluster as a user with the cluster-admin
role.
In the Administrator perspective, navigate to Home → Overview.
From the OKD web console, you can review detailed information about the status of your cluster on OpenShift Cluster Manager.
You are logged in to OpenShift Cluster Manager.
You have access to the cluster as a user with the cluster-admin
role.
Go to the Cluster List list in OpenShift Cluster Manager and locate your OKD cluster.
Click the Overview tab for your cluster.
Review the following information about your cluster:
vCPU and memory availability and resource usage
The cluster ID, status, type, region, and the provider name
Node counts by node type
Cluster version details, the creation date of the cluster, and the name of the cluster owner
The life cycle support status of the cluster
Subscription information, including the service level agreement (SLA) status, the subscription unit type, the production status of the cluster, the subscription obligation, and the service level
To view the history for your cluster, click the Cluster history tab. |
Navigate to the Monitoring page to review the following information:
A list of any issues that have been detected
A list of alerts that are firing
The cluster Operator status and version
The cluster’s resource usage
Optional: You can view information about your cluster that Red Hat Insights collects by navigating to the Overview menu. From this menu you can view the following information:
Potential issues that your cluster might be exposed to, categorized by risk level
Health-check status by category
See Using Insights to identify issues with your cluster for more information about reviewing potential issues with your cluster.
OKD provides a comprehensive set of monitoring dashboards that help you understand the state of cluster components.
In the Administrator perspective, you can access dashboards for core OKD components, including:
etcd
Kubernetes compute resources
Kubernetes network resources
Prometheus
Dashboards relating to cluster and node performance
You have access to the cluster as a user with the cluster-admin
role.
In the Administrator perspective in the OKD web console, navigate to Observe → Dashboards.
Choose a dashboard in the Dashboard list. Some dashboards, such as the etcd dashboard, produce additional sub-menus when selected.
Optional: Select a time range for the graphs in the Time Range list.
Select a pre-defined time period.
Set a custom time range by selecting Custom time range in the Time Range list.
Input or select the From and To dates and times.
Click Save to save the custom time range.
Optional: Select a Refresh Interval.
Hover over each of the graphs within a dashboard to display detailed information about specific items.
See Monitoring overview for more information about the OKD monitoring stack.
Alerts provide notifications when a set of defined conditions are true in an OKD cluster. You can review the alerts that are firing in your cluster by using the Alerting UI in the OKD web console.
You have access to the cluster as a user with the cluster-admin
role.
In the Administrator perspective, navigate to the Observe → Alerting → Alerts page.
Review the alerts that are firing, including their Severity, State, and Source.
Select an alert to view more detailed information in the Alert Details page.
See Managing alerts for further details about alerting in OKD.
See Troubleshooting installations if you experience issues when installing your cluster.
After installing OKD, you can further expand and customize your cluster.