AdmissionConfig
The openshift start
command is used to launch OpenShift Container Platform servers. The command
and its subcommands (master
to launch a
master
server and node
to launch a
node
server) all take a limited set of arguments that are sufficient for launching
servers in a development or experimental environment.
However, these arguments are insufficient to describe and control the full set of configuration and security options that are necessary in a production environment. To provide those options, it is necessary to use the dedicated master and node configuration files.
Master configuration files and
node configuration files are fully specified
with no default values. Therefore, any empty value indicates that you want to
start up with an empty value for that parameter. This makes it easy to reason
about exactly what your configuration is, but it also makes it difficult to
remember all of the options to specify. To make this easier, the configuration
files can be created with the --write-config
option and then used with the
--config
option.
This section reviews parameters mentioned in the master-config.yaml file.
You can create a new master configuration file to see the valid options for your installed version of OpenShift Container Platform.
Whenever you modify the master-config.yaml file, you must restart the master for the changes to take effect. See Restarting OpenShift Container Platform services. |
Parameter Name | Description |
---|---|
|
Contains admission control plug-in configuration. |
|
Key-value pairs that will be passed directly to the Kube API server that match
the API servers' command line arguments. These are not migrated, but if you
reference a value that does not exist the server will not start. These values
may override other settings in |
|
Key-value pairs that will be passed directly to the Kube controller manager
that match the controller manager’s command line arguments. These are not
migrated, but if you reference a value that does not exist the server will not
start. These values may override other settings in |
|
Used to enable or disable various admission plug-ins. When this type is present
as the configuration object under |
|
Allows specifying a configuration file per admission control plug-in. |
|
A list of admission control plug-in names that will be installed on the master. Order is significant. If empty, a default list of plug-ins is used. |
|
Key-value pairs that will be passed directly to the Kube scheduler that match
the scheduler’s command line arguments. These are not migrated, but if you
reference a value that does not exist the server will not start. These values
may override other settings in |
Parameter Name | Description |
---|---|
|
Holds the necessary configuration options for serving assets. |
|
To access the API server from a web application using a different host name, you
must whitelist that host name by specifying |
|
A list of features that should not be started. You will likely want to set this as null. It is very unlikely that anyone will want to manually disable features and that is not encouraged. |
|
Files to serve from the asset server file system under a subcontext. |
|
When set to true, tells the asset server to reload extension scripts and stylesheets for every request rather than only at startup. It lets you develop extensions without having to restart the server for every change. |
|
Key- (string) and value- (string) pairs that will be injected into the console under
the global variable |
|
File paths on the asset server files to load as scripts when the web console loads. |
|
File paths on the asset server files to load as style sheets when the web console loads. |
|
The public endpoint for logging (optional). |
|
An optional, absolute URL to redirect web browsers to after logging out of the web console. If not specified, the built-in logout page is shown. |
|
How the web console can access the OpenShift Container Platform server. |
|
The public endpoint for metrics (optional). |
|
URL of the the asset server. |
Parameter Name | Description |
---|---|
|
Holds authentication and authorization configuration options. |
|
Indicates how many authentication results should be cached. If 0, the default cache size is used. |
|
Indicates how long an authorization result should be cached. It takes a valid time duration string (e.g. "5m"). If empty, you get the default timeout. If zero (e.g. "0m"), caching is disabled. |
Parameter Name | Description |
---|---|
|
List of the controllers that should be started. If set to none, no
controllers will start automatically. The default value is * which will start
all controllers. When using *, you may exclude controllers by prepending a |
|
enables controller election, instructing the master to attempt to acquire a
lease before controllers start and renewing it within a number of seconds
defined by this value. Setting this value non-negative forces
|
|
Instructs the master to not automatically start controllers, but instead to wait until a notification to the server is received before launching them. |
Parameter Name | Description |
---|---|
|
The advertised host:port for client connections to etcd. |
|
Contains information about how to connect to etcd. |
|
Holds the necessary configuration options for connecting with an etcd database. |
|
Contains information about how API resources are stored in etcd. These values are only relevant when etcd is the backing store for the cluster. |
|
The path within etcd that the Kubernetes resources will be rooted under. This value, if changed, will mean existing objects in etcd will no longer be located. The default value is kubernetes.io. |
|
The API version that Kubernetes resources in etcd should be serialized to. This value should not be advanced until all clients in the cluster that read from etcd have code that allows them to read the new version. |
|
The path within etcd that the OpenShift Container Platform resources will be rooted under. This value, if changed, will mean existing objects in etcd will no longer be located. The default value is openshift.io. |
|
API version that OS resources in etcd should be serialized to. This value should not be advanced until all clients in the cluster that read from etcd have code that allows them to read the new version. |
|
The advertised host:port for peer connections to etcd. |
|
Describes how to start serving the etcd peer. |
|
Describes how to start serving the etcd master. |
|
The path to the etcd storage directory. |
Parameter Name | Description |
---|---|
|
Describes how to handle grants. |
|
Auto-approves client authorization grant requests. |
|
Auto-denies client authorization grant requests. |
|
Prompts the user to approve new client authorization grant requests. |
|
Determines the default strategy to use when an OAuth client requests a grant.This method will be used only if the specific OAuth client does not provide a strategy of their own. Valid grant handling methods are:
|
Parameter Name | Description |
---|---|
|
Allows scheduled background import of images to be disabled. |
|
The format of the name to be built for the system component. |
|
Holds options that describe how to build image names for system components. |
|
Controls limits and behavior for importing images. |
|
Determines if the latest tag will be pulled from the registry. |
|
Controls the number of images that are imported when a user does a bulk import of a Docker repository. This number defaults to 5 to prevent users from importing large numbers of images accidentally. Set -1 for no limit. |
|
The maximum number of scheduled image streams that will be imported in the background per minute. The default value is 60. |
|
The minimum number of seconds that can elapse between when image streams scheduled for background import are checked against the upstream repository. The default value is 15 minutes. |
Parameter Name | Description |
---|---|
|
A list of API levels that should be enabled on startup, v1 as examples. |
|
A map of groups to the versions (or |
|
Contains information about how to connect to kubelets. |
|
Holds the necessary configuration options for the Kubernetes master. |
|
The number of expected masters that should be running. This value defaults to 1 and may be set to a positive integer, or if set to -1, indicates this is part of a cluster. |
|
The public IP address of Kubernetes resources. If empty, the first result from
|
|
File name for the .kubeconfig file that describes how to connect this node to the master. |
|
The range to use for assigning service public ports on a host. Default 30000-32767. |
|
The subnet to use for assigning service IPs. |
|
The list of nodes that are statically known. |
Choose the CIDRs in the following parameters carefully, because the IPv4 address space is shared by all users of the nodes. OpenShift Container Platform reserves CIDRs from the IPv4 address space for its own use, and reserves CIDRs from the IPv4 address space for addresses that are shared between the external user and the cluster.
Parameter Name | Description |
---|---|
|
The CIDR string to specify the global overlay network’s L3 space. This is reserved for the internal use of the cluster networking. |
|
Controls what values are acceptable for the service external IP field. If
empty, no |
|
The number of bits to allocate to each host’s subnet. For example, 8 would mean a /24 network on the host. |
|
Controls the range to assign ingress IPs from for services of type
LoadBalancer on bare metal. It may contain a single CIDR that it will be
allocated from. By default |
|
The number of bits to allocate to each host’s subnet. For example, 8 would mean a /24 network on the host. |
|
Provides network options for the node. |
|
The name of the network plug-in to use. |
|
The CIDR string to specify the service networks. |
Parameter Name | Description |
---|---|
|
Forces the provider selection page to render even when there is only a single provider. |
|
Used for building valid client redirect URLs for external access. |
|
A path to a file containing a go template used to render error pages during the authentication or grant flow If unspecified, the default error page is used. |
|
Ordered list of ways for a user to identify themselves. |
|
A path to a file containing a go template used to render the login page. If unspecified, the default login page is used. |
|
CA for verifying the TLS connection back to the |
|
Used for building valid client redirect URLs for external access. |
|
Used for making server-to-server calls to exchange authorization codes for access tokens. |
|
Holds the necessary configuration options for OAuth authentication. |
|
Allows for customization of pages like the login page. |
|
A path to a file containing a go template used to render the provider selection page. If unspecified, the default provider selection page is used. |
|
Holds information about configuring sessions. |
|
Allows you to customize pages like the login page. |
|
Contains options for authorization and access tokens. |
Parameter Name | Description |
---|---|
|
Holds default project node label selector. |
|
Holds information about project creation and defaults. |
|
The string presented to a user if they are unable to request a project via the project request API endpoint. |
|
The template to use for creating projects in response to a projectrequest. It is in the format namespace/template and it is optional. If it is not specified, a default template is used. |
Parameter Name | Description |
---|---|
|
Points to a file that describes how to set up the scheduler. If empty, you get the default scheduling rules |
Parameter Name | Description |
---|---|
|
Defines the range of MCS categories that will be assigned to namespaces. The
format is |
|
Controls the automatic allocation of UIDs and MCS labels to a project. If nil, allocation is disabled. |
|
Defines the total set of Unix user IDs (UIDs) that will be allocated to projects automatically, and the size of the block that each namespace gets. For example, 1000-1999/10 will allocate ten UIDs per namespace, and will be able to allocate up to 100 blocks before running out of space. The default is to allocate from 1 billion to 2 billion in 10k blocks (which is the expected size of the ranges container images will use once user namespaces are started). |
Parameter Name | Description |
---|---|
|
Controls whether or not to allow a service account to reference any secret in a namespace without explicitly referencing them. |
|
A list of service account names that will be auto-created in every namespace.
If no names are specified, the |
|
The CA for verifying the TLS connection back to the master. The service account controller will automatically inject the contents of this file into pods so they can verify connections to the master. |
|
A file containing a PeM-encoded private RSA key, used to sign service account
tokens. If no private key is specified, the service account |
|
A list of files, each containing a PeM-encoded public RSA key. If any file contains a private key, the public portion of the key is used. The list of public keys is used to verify presented service account tokens. each key is tried in order until the list is exhausted or verification succeeds. If no keys are specified, no service account authentication will be available. |
|
Holds the necessary configuration options for a service account. |
Parameter Name | Description |
---|---|
|
Allows the DNS server on the master to answer queries recursively. Note that open resolvers can be used for DNS amplification attacks and the master DNS should not be made accessible to public networks. |
|
The ip:port to serve on. |
|
Controls limits and behavior for importing images. |
|
A file containing a PeM-encoded certificate. |
|
TLS cert information for serving secure traffic. |
|
The certificate bundle for all the signers that you recognize for incoming client certificates. |
|
Holds the necessary configuration options for DNS. |
|
Holds the domain suffix. |
|
Holds the IP. |
|
A file containing a PeM-encoded private key for the certificate specified by
|
|
Provides overrides to the client connection used to connect to the master. |
|
The number of concurrent requests allowed to the server. If zero, no limit. |
|
A list of certificates to use to secure requests to specific host names. |
|
The number of seconds before requests are timed out. The default is 60 minutes. If -1, there is no limit on requests. |
|
The HTTP serving information for the assets. |
Parameter Name | Description |
---|---|
|
A boolean to enable or disable dynamic provisioning. Default is true. |
FSGroup |
Can be specified to enable a quota on local storage use per unique FSGroup ID.
At present this is only implemented for emptyDir volumes, and if the underlying
|
|
Contains options for controlling local volume quota on the node. |
|
Contains options for configuring volume plug-ins in the master node. |
|
Contains options for configuring volumes on the node. |
|
Contains options for configuring volumes on the node. |
|
The directory that volumes are stored under. |
Audit provides a security-relevant chronological set of records documenting the sequence of activities that have affected system by individual users, administrators, or other components of the system.
Audit works at the API server level, logging all requests coming to the server. each audit log contains two entries:
The request line containing:
A Unique ID allowing to match the response line (see #2)
The source IP of the request
The HTTP method being invoked
The original user invoking the operation
The impersonated user for the operation (self
meaning himself)
The impersonated group for the operation (lookup
meaning user’s group)
The namespace of the request or <none>
The URI as requested
The response line containing:
The the unique ID from #1
The response code
example output for user admin asking for a list of pods:
AUDIT: id="5c3b8227-4af9-4322-8a71-542231c3887b" ip="127.0.0.1" method="GeT" user="admin" as="<self>" asgroups="<lookup>" namespace="default" uri="/api/v1/namespaces/default/pods" AUDIT: id="5c3b8227-4af9-4322-8a71-542231c3887b" response="200"
The openshift_master_audit_config
variable enables API service auditing. It
takes an array of the following options:
Parameter Name | Description |
---|---|
|
A boolean to enable or disable audit logs. Default is |
|
File path where the requests should be logged to. If not set, logs are printed to master logs. |
|
Specifies maximum number of days to retain old audit log files based on the time stamp encoded in their filename. |
|
Specifies the maximum number of old audit log files to retain. |
|
Specifies maximum size in megabytes of the log file before it gets rotated. Defaults to 100MB. |
auditConfig: auditFilePath: "/var/log/audit-ocp.log" enabled: true maximumFileRetentionDays: 10 maximumFileSizeMegabytes: 10 maximumRetainedFiles: 10
If you want more advanced setup for the audit log, you can use:
openshift_master_audit_config={"enabled": true}
The directory in auditFilePath
will be created if it does not exist.
openshift_master_audit_config={"enabled": true, "auditFilePath": "/var/log/openpaas-oscp-audit/openpaas-oscp-audit.log", "maximumFileRetentionDays": 14, "maximumFileSizeMegabytes": 500, "maximumRetainedFiles": 5}
The following node-config.yaml file is a sample node configuration file that was generated with the default values as of writing. You can create a new node configuration file to see the valid options for your installed version of OpenShift Container Platform.
allowDisabledDocker: false
apiVersion: v1
authConfig:
authenticationCacheSize: 1000
authenticationCacheTTL: 5m
authorizationCacheSize: 1000
authorizationCacheTTL: 5m
dnsDomain: cluster.local
dnsIP: 10.0.2.15 (1)
dockerConfig:
execHandlerName: native
imageConfig:
format: openshift/origin-${component}:${version}
latest: false
iptablesSyncPeriod: 5s
kind: NodeConfig
masterKubeConfig: node.kubeconfig
networkConfig:
mtu: 1450
networkPluginName: ""
nodeIP: ""
nodeName: node1.example.com
podManifestConfig: (2)
path: "/path/to/pod-manifest-file" (3)
fileCheckIntervalSeconds: 30 (4)
proxyArguments:
proxy-mode:
- iptables (5)
volumeConfig:
localQuota:
perFSGroup: null(6)
servingInfo:
bindAddress: 0.0.0.0:10250
bindNetwork: tcp4
certFile: server.crt
clientCA: node-client-ca.crt
keyFile: server.key
namedCertificates: null
volumeDirectory: /root/openshift.local.volumes
1 | Configures an IP address to be prepended to a pod’s /etc/resolv.conf by adding the address here. |
2 | Allows pods to be placed directly on certain set of nodes, or on all nodes without going through the scheduler. You can then use pods to perform the same administrative tasks and support the same services on each node. |
3 | Specifies the path for the pod manifest file or directory. If it is a directory, then it is expected to contain one or more manifest files. This is used by the Kubelet to create pods on the node. |
4 | This is the interval (in seconds) for checking the manifest file for new data. The interval must be a positive value. |
5 | The service proxy implementation to use. |
6 | Preliminary support for local emptyDir volume quotas, set this value to a resource quantity representing the desired quota per FSGroup, per node. (i.e. 1Gi, 512Mi, etc) Currently requires that the volumeDirectory be on an XFS filesystem mounted with the 'gquota' option, and the matching security context contraint’s fsGroup type set to 'MustRunAs'. |
Parameter Name | Description |
---|---|
|
The fully specified configuration starting an OpenShift Container Platform node. |
|
Node may have multiple IPs, so this specifies the IP to use for pod traffic routing. If not specified, network parse/lookup on the nodeName is performed and the first non-loopback address is used. |
|
The value used to identify this particular node in the cluster. If possible, this should be your fully qualified hostname. If you are describing a set of static nodes to the master, this value must match one of the values in the list. |
|
Controls grace period for deleting pods on failed nodes. It takes valid time duration string. If empty, you get the default pod eviction timeout. |
|
Specifies the client cert/key to use when proxying to pods. |
Parameter Name | Description |
---|---|
|
If true, the kubelet will ignore errors from Docker. This means that a node can start on a machine that does not have docker started. |
|
Holds Docker related configuration options |
|
The handler to use for executing commands in Docker containers. |
If you are using Docker 1.9+, you may want to consider enabling parallel image pulling, as the default is to pull images one at a time.
There is a potential issue with data corruption prior to Docker 1.9. However, starting with 1.9, the corruption issue is resolved and it is safe to switch to parallel pulls. |
kubeletArguments:
serialize-image-pulls:
- "false" (1)
1 | Change to true to disable parallel pulls. (This is the default config) |
For some authentication configurations,
an LDAP bindPassword
or OAuth clientSecret
value is required.
Instead of specifying these values directly in the master configuration file,
these values may be provided as environment variables, external files,
or in encrypted files.
...
bindPassword:
env: BIND_PASSWORD_eNV_VAR_NAMe
...
bindPassword:
file: bindPassword.txt
...
bindPassword:
file: bindPassword.encrypted
keyFile: bindPassword.key
To create the encrypted file and key file for the above example:
$ oc adm ca encrypt --genkey=bindPassword.key --out=bindPassword.encrypted > Data to encrypt: B1ndPass0rd!
Run oc adm
commands only from the first master listed in the Ansible host inventory file,
by default /etc/ansible/hosts.
encrypted data is only as secure as the decrypting key. Care should be taken to limit filesystem permissions and access to the key file. |
When defining an OpenShift Container Platform configuration from scratch, start by creating new configuration files.
For master host configuration files, use the openshift start
command with the
--write-config
option to write the configuration files. For node hosts, use
the oc adm create-node-config
command to write the configuration files.
The following commands write the relevant launch configuration file(s),
certificate files, and any other necessary files to the specified
--write-config
or --node-dir
directory.
Generated certificate files are valid for two years, while the certification
authority (CA) certificate is valid for five years. This can be altered with the
--expire-days
and --signer-expire-days
options, but for security reasons, it
is recommended to not make them greater than these values.
To create configuration files for an all-in-one server (a master and a node on the same host) in the specified directory:
$ openshift start --write-config=/openshift.local.config
To create a master configuration file and other required files in the specified directory:
$ openshift start master --write-config=/openshift.local.config/master
To create a node configuration file and other related files in the specified directory:
$ oc adm create-node-config \ --node-dir=/openshift.local.config/node-<node_hostname> \ --node=<node_hostname> \ --hostnames=<node_hostname>,<ip_address> \ --certificate-authority="/path/to/ca.crt" \ --signer-cert="/path/to/ca.crt" \ --signer-key="/path/to/ca.key" --signer-serial="/path/to/ca.serial.txt" --node-client-certificate-authority="/path/to/ca.crt"
When creating node configuration files, the --hostnames
option accepts a
comma-delimited list of every host name or IP address you want server
certificates to be valid for.
Once you have modified the master and/or node configuration files to your specifications, you can use them when launching servers by specifying them as an argument. Keep in mind that if you specify a configuration file, none of the other command line options you pass are respected.
To launch an all-in-one server using a master configuration and a node configuration file:
$ openshift start --master-config=/openshift.local.config/master/master-config.yaml --node-config=/openshift.local.config/node-<node_hostname>/node-config.yaml
To launch a master server using a master configuration file:
$ openshift start master --config=/openshift.local.config/master/master-config.yaml
To launch a node server using a node configuration file:
$ openshift start node --config=/openshift.local.config/node-<node_hostname>/node-config.yaml
OpenShift Container Platform uses the systemd-journald.service
to collect log messages for debugging, using five log message severities. The logging levels are based on Kubernetes logging conventions, as follows:
Option | Description |
---|---|
0 |
errors and warnings only |
2 |
Normal information |
4 |
Debugging-level information |
6 |
API-level debugging information (request / response) |
8 |
Body-level API debugging information |
You can control which INFO messages are logged by setting the loglevel option in the in /etc/sysconfig/atomic-openshift-node, the /etc/sysconfig/atomic-openshift-master-api file and the /etc/sysconfig/atomic-openshift-master-controllers file. Configuring the logs to collect all messages can lead to large logs that are difficult to interpret and can take up excessive space. Collecting all messages should only be used in debug situations.
Messages with FATAL, eRROR, WARNING and some INFO severities appear in the logs regardless of the log configuration. |
You can view logs for the master or the node system using the following command:
# journalctl -r -u <journal_name>
Use the -r
option to show the newest entries first.
For example:
# journalctl -r -u atomic-openshift-master-controllers # journalctl -r -u atomic-openshift-master-api # journalctl -r -u atomic-openshift-node.service
To change the logging level:
edit the /etc/sysconfig/atomic-openshift-master file for the master or /etc/sysconfig/atomic-openshift-node file for the nodes.
enter a value from the Log Level Options table above in the OPTIONS=--loglevel=
field.
For example:
OPTIONS=--loglevel=4
Restart the master or node host as appropriate. See Restarting OpenShift Container Platform services.
After the restart, all new log messages will conform to the new setting. Older messages do not change.
The default log level can be set using the Advanced Install. For more information, see Cluster Variables. |
The following examples are excerpts from a master journald log at various log levels. Timestamps and system information have been removed from these examples.
4897 plugins.go:77] Registered admission plugin "NamespaceLifecycle" 4897 start_master.go:290] Warning: assetConfig.loggingPublicURL: Invalid value: "": required to view aggregated container logs in the console, master start will continue. 4897 start_master.go:290] Warning: assetConfig.metricsPublicURL: Invalid value: "": required to view cluster metrics in the console, master start will continue. 4897 start_master.go:290] Warning: aggregatorConfig.proxyClientInfo: Invalid value: "": if no client certificate is specified, the aggregator will be unable to proxy to remote servers, 4897 start_master.go:412] Starting controllers on 0.0.0.0:8444 (v3.7.14) 4897 start_master.go:416] Using images from "openshift3/ose-<component>:v3.7.14" 4897 standalone_apiserver.go:106] Started health checks at 0.0.0.0:8444 4897 plugins.go:77] Registered admission plugin "NamespaceLifecycle" 4897 configgetter.go:53] Initializing cache sizes based on 0MB limit 4897 leaderelection.go:105] Attempting to acquire openshift-master-controllers lease as master-bkr-hv03-guest44.dsal.lab.eng.bos.redhat.com-10.19.41.74-xtz6lbqb, renewing every 3s, hold 4897 leaderelection.go:179] attempting to acquire leader lease... systemd[1]: Started Atomic OpenShift Master Controllers. 4897 leaderelection.go:189] successfully acquired lease kube-system/openshift-master-controllers 4897 event.go:218] event(v1.ObjectReference{Kind:"ConfigMap", Namespace:"kube-system", Name:"openshift-master-controllers", UID:"aca86731-ffbe-11e7-8d33-525400c845a8", APIVersion:"v1", 4897 start_master.go:627] Started serviceaccount-token controller 4897 factory.go:351] Creating scheduler from configuration: {{ } [{NoVolumeZoneConflict <nil>} {MaxeBSVolumeCount <nil>} {MaxGCePDVolumeCount <nil>} {MaxAzureDiskVolumeCount <nil>} {Mat 4897 factory.go:360] Registering predicate: NoVolumeZoneConflict 4897 plugins.go:145] Predicate type NoVolumeZoneConflict already registered, reusing. 4897 factory.go:360] Registering predicate: MaxeBSVolumeCount 4897 plugins.go:145] Predicate type MaxeBSVolumeCount already registered, reusing. 4897 factory.go:360] Registering predicate: MaxGCePDVolumeCount
4897 master.go:47] Initializing SDN master of type "redhat/openshift-ovs-subnet" 4897 master.go:107] Created ClusterNetwork default (network: "10.128.0.0/14", hostSubnetBits: 9, serviceNetwork: "172.30.0.0/16", pluginName: "redhat/openshift-ovs-subnet") 4897 start_master.go:690] Started "openshift.io/sdn" 4897 start_master.go:680] Starting "openshift.io/service-serving-cert" 4897 controllermanager.go:466] Started "namespace" 4897 controllermanager.go:456] Starting "daemonset" 4897 controller_utils.go:1025] Waiting for caches to sync for namespace controller 4897 shared_informer.go:298] resyncPeriod 120000000000 is smaller than resyncCheckPeriod 600000000000 and the informer has already started. Changing it to 600000000000 4897 start_master.go:690] Started "openshift.io/service-serving-cert" 4897 start_master.go:680] Starting "openshift.io/image-signature-import" 4897 start_master.go:690] Started "openshift.io/image-signature-import" 4897 start_master.go:680] Starting "openshift.io/templateinstance" 4897 controllermanager.go:466] Started "daemonset" 4897 controllermanager.go:456] Starting "statefulset" 4897 daemoncontroller.go:222] Starting daemon sets controller 4897 controller_utils.go:1025] Waiting for caches to sync for daemon sets controller 4897 controllermanager.go:466] Started "statefulset" 4897 controllermanager.go:456] Starting "cronjob" 4897 stateful_set.go:147] Starting stateful set controller 4897 controller_utils.go:1025] Waiting for caches to sync for stateful set controller 4897 start_master.go:690] Started "openshift.io/templateinstance" 4897 start_master.go:680] Starting "openshift.io/horizontalpodautoscaling
4897 factory.go:366] Registering priority: Zone 4897 factory.go:401] Creating scheduler with fit predicates 'map[GeneralPredicates:{} CheckNodeMemoryPressure:{} CheckNodeDiskPressure:{} NoVolumeNodeConflict:{} Region:{} NoVolumeZoneC 4897 controller_utils.go:1025] Waiting for caches to sync for tokens controller 4897 controllermanager.go:108] Version: v1.7.6+a08f5eeb62 4897 leaderelection.go:179] attempting to acquire leader lease... 4897 leaderelection.go:189] successfully acquired lease kube-system/kube-controller-manager 4897 event.go:218] event(v1.ObjectReference{Kind:"ConfigMap", Namespace:"kube-system", Name:"kube-controller-manager", UID:"acb3e9c6-ffbe-11e7-8d33-525400c845a8", APIVersion:"v1", Resou 4897 controller_utils.go:1032] Caches are synced for tokens controller 4897 plugins.go:101] No cloud provider specified. 4897 controllermanager.go:481] "serviceaccount-token" is disabled 4897 controllermanager.go:450] "bootstrapsigner" is disabled 4897 controllermanager.go:450] "tokencleaner" is disabled 4897 controllermanager.go:456] Starting "garbagecollector" 4897 start_master.go:680] Starting "openshift.io/build" 4897 controllermanager.go:466] Started "garbagecollector" 4897 controllermanager.go:456] Starting "deployment" 4897 garbagecollector.go:126] Starting garbage collector controller 4897 controller_utils.go:1025] Waiting for caches to sync for garbage collector controller 4897 controllermanager.go:466] Started "deployment" 4897 controllermanager.go:450] "horizontalpodautoscaling" is disabled 4897 controllermanager.go:456] Starting "disruption" 4897 deployment_controller.go:152] Starting deployment controller
4897 plugins.go:77] Registered admission plugin "NamespaceLifecycle" 4897 start_master.go:290] Warning: assetConfig.loggingPublicURL: Invalid value: "": required to view aggregated container logs in the console, master start will continue. 4897 start_master.go:290] Warning: assetConfig.metricsPublicURL: Invalid value: "": required to view cluster metrics in the console, master start will continue. 4897 start_master.go:290] Warning: aggregatorConfig.proxyClientInfo: Invalid value: "": if no client certificate is specified, the aggregator will be unable to proxy to remote serv 4897 start_master.go:412] Starting controllers on 0.0.0.0:8444 (v3.7.14) 4897 start_master.go:416] Using images from "openshift3/ose-<component>:v3.7.14" 4897 standalone_apiserver.go:106] Started health checks at 0.0.0.0:8444 4897 plugins.go:77] Registered admission plugin "NamespaceLifecycle" 4897 configgetter.go:53] Initializing cache sizes based on 0MB limit 4897 leaderelection.go:105] Attempting to acquire openshift-master-controllers lease as master-bkr-hv03-guest44.dsal.lab.eng.bos.redhat.com-10.19.41.74-xtz6lbqb, renewing every 3s, 4897 leaderelection.go:179] attempting to acquire leader lease... systemd[1]: Started Atomic OpenShift Master Controllers. 4897 leaderelection.go:189] successfully acquired lease kube-system/openshift-master-controllers 4897 event.go:218] event(v1.ObjectReference{Kind:"ConfigMap", Namespace:"kube-system", Name:"openshift-master-controllers", UID:"aca86731-ffbe-11e7-8d33-525400c845a8", APIVersion:" 4897 start_master.go:627] Started serviceaccount-token controller
4613 plugins.go:77] Registered admission plugin "NamespaceLifecycle" 4613 master.go:320] Starting Web Console https://bkr-hv03-guest44.dsal.lab.eng.bos.redhat.com:8443/console/ 4613 master.go:329] Starting OAuth2 API at /oauth 4613 master.go:320] Starting Web Console https://bkr-hv03-guest44.dsal.lab.eng.bos.redhat.com:8443/console/ 4613 master.go:329] Starting OAuth2 API at /oauth 4613 master.go:320] Starting Web Console https://bkr-hv03-guest44.dsal.lab.eng.bos.redhat.com:8443/console/ 4613 master.go:329] Starting OAuth2 API at /oauth 4613 swagger.go:38] No API exists for predefined swagger description /oapi/v1 4613 swagger.go:38] No API exists for predefined swagger description /api/v1 [restful] 2018/01/22 16:53:14 log.go:33: [restful/swagger] listing is available at https://bkr-hv03-guest44.dsal.lab.eng.bos.redhat.com:8443/swaggerapi [restful] 2018/01/22 16:53:14 log.go:33: [restful/swagger] https://bkr-hv03-guest44.dsal.lab.eng.bos.redhat.com:8443/swaggerui/ is mapped to folder /swagger-ui/ 4613 master.go:320] Starting Web Console https://bkr-hv03-guest44.dsal.lab.eng.bos.redhat.com:8443/console/ 4613 master.go:329] Starting OAuth2 API at /oauth 4613 swagger.go:38] No API exists for predefined swagger description /oapi/v1 4613 swagger.go:38] No API exists for predefined swagger description /api/v1 [restful] 2018/01/22 16:53:14 log.go:33: [restful/swagger] listing is available at https://bkr-hv03-guest44.dsal.lab.eng.bos.redhat.com:8443/swaggerapi [restful] 2018/01/22 16:53:14 log.go:33: [restful/swagger] https://bkr-hv03-guest44.dsal.lab.eng.bos.redhat.com:8443/swaggerui/ is mapped to folder /swagger-ui/ Starting Web Console https://bkr-hv03-guest44.dsal.lab.eng.bos.redhat.com:8443/console/ Starting OAuth2 API at /oauth No API exists for predefined swagger description /oapi/v1 No API exists for predefined swagger description /api/v1
To apply configuration changes, you must restart OpenShift Container Platform services.
To restart master, run the command:
# systemctl restart atomic-openshift-master-api atomic-openshift-master-controllers
To restart node hosts, on each node, run the command:
# systemctl restart atomic-openshift-node