Line 18: <!-- ##MeSSAGING_eXTeNSION## --> Line 318: <!-- ##MeSSAGING_SUBSYSTeM## -->
Starting in OpenShift enterprise 3.0, xPaaS images are provided for the following:
Red Hat JBoss enterprise Application Platform
Red Hat JBoss Web Server
Red Hat JBoss A-MQ
Starting in OpenShift enterprise 3.1, xPaaS images are also provided for the following:
Red Hat JBoss Fuse (Fuse Integration Services)
Red Hat JBoss BRMS (Decision Server)
Red Hat JBoss Data Grid
See enterprise.openshift.com/middleware-services for additional information.
Red Hat JBoss eAP is available as a containerized xPaaS image that is designed for use with OpenShift enterprise 3.0 and later.
However, there are significant differences in supported configurations and functionality in the JBoss eAP xPaaS image compared to the regular release of JBoss eAP. Documentation for other JBoss eAP functionality not specific to the JBoss eAP xPaaS image can be found in the JBoss eAP documentation on the Red Hat Customer Portal.
The Apache Tomcat 7 and Apache Tomcat 8 components of Red Hat JBoss Web Server 3 are available as containerized xPaaS images that are designed for use with OpenShift enterprise 3.0 and later.
However, there are significant differences in the functionality between the JBoss Web Server xPaaS images and the regular release of JBoss Web Server. Documentation for other JBoss Web Server functionality not specific to the JBoss Web Server xPaaS images can be found in the JBoss Web Server documentation on the Red Hat Customer Portal.
Red Hat JBoss A-MQ is available as a containerized xPaaS image that is designed for use with OpenShift enterprise 3.0 and later. It allows developers to quickly deploy an A-MQ message broker in a hybrid cloud environment.
However, there are significant differences in supported configurations and functionality in the JBoss A-MQ image compared to the regular release of JBoss A-MQ. Documentation for other JBoss A-MQ functionality not specific to the JBoss A-MQ xPaaS image can be found in the JBoss A-MQ documentation on the Red Hat Customer Portal.
Red Hat JBoss Fuse is available as a containerized xPaaS image, known as Fuse Integration Services, that is designed for use with OpenShift enterprise 3.1. It allows developers to quickly deploy applications in a hybrid cloud environment. In Fuse Integration Services, application runtime is dynamic.
However, there are significant differences in supported configurations and functionality in the Fuse Integration Services compared to the regular release of JBoss Fuse. Documentation for other JBoss Fuse functionality not specific to the Fuse Integration Services can be found in the JBoss Fuse documentation on the Red Hat Customer Portal.
Red Hat JBoss BRMS is available as a containerized xPaaS image, known as Decision Server, that is designed for use with OpenShift enterprise 3.1 as an execution environment for business rules. Developers can quickly build, scale, and test applications deployed across hybrid environments.
However, there are significant differences in supported configurations and functionality in the Decision Server xPaaS image compared to the regular release of JBoss BRMS. Documentation for other JBoss BRMS functionality not specific to the Decision Server xPaaS image can be found in the JBoss BRMS documentation on the Red Hat Customer Portal.
Red Hat JBoss Data Grid is available as a containerized xPaaS image that is designed for use with OpenShift enterprise 3.1. This image provides an in-memory distributed database so that developers can quickly access large amounts of data in a hybrid environment.
However, there are significant differences in supported configurations and functionality in the JBoss Data Grid xPaaS image compared to the full, non-PaaS release of JBoss Data Grid. Documentation for other JBoss Data Grid functionality not specific to the JBoss Data Grid xPaaS image can be found in the JBoss Data Grid documentation on the Red Hat Customer Portal.
The following are the current known issues along with any known workarounds:
JWS
https://issues.jboss.org/browse/CLOUD-57: Tomcat’s access log valve logs to file in container instead of stdout
Due to this issue, the logging data is not available for the central logging
facility. To work around this issue, use the oc exec
command to get the
contents of the log file.
https://issues.jboss.org/browse/CLOUD-153: mvn clean
in JWS STI can fail
Cleaning up after a build in JWS STI is not possible, because the Maven command
mvn clean
fails. This is due to Maven not being able to build the object model
during startup.
To work around this issue, add Red Hat and JBoss repositories into the pom.xml file of the application if the application uses dependencies from there.
https://issues.jboss.org/browse/CLOUD-156: Datasource realm configuration is incorrect for JWS
It is not possible to do correct JNDI lookup for datasources in the current JWS
image if an invalid combination of datasource and realm properties is defined.
If a datasource is configured in the context.xml file and a realm in the
server.xml file, then the server.xml file’s localDataSource
property
should be set to true.
eAP
https://issues.jboss.org/browse/CLOUD-61: JPA application fails to start when the database is not available
JPA applications fail to deploy in the eAP OpenShift enterprise 3.0 image if an underlying database instance that the eAP instance relies on is not available at the start of the deployment. The eAP application tries to contact the database for initialization, but because it is not available, the server starts but the application fails to deploy.
There are no known workarounds available at this stage for this issue.
https://issues.jboss.org/browse/CLOUD-158: Continuous HornetQ errors after scale down "Failed to create netty connection"
In the eAP image, an application not using messaging complains about messaging errors related to HornetQ when being scaled.
Since there are no configuration options to disable messaging to work around this issue, simply include the standalone-openshift.xml file within the source of the image and remove or alter the following lines related to messaging:
Line 18: <!-- ##MeSSAGING_eXTeNSION## --> Line 318: <!-- ##MeSSAGING_SUBSYSTeM## -->
https://issues.jboss.org/browse/CLOUD-161: eAP pod serving requests before it joins cluster, some sessions reset after failure
In a distributed web application deployed on an eAP image, a new container starts serving requests before it joins the cluster.
There are no known workarounds available at this stage for this issue.
eAP and JWS
https://issues.jboss.org/browse/CLOUD-159: Database pool configurations should contain validation SQL setting
In both the eAP and JWS images, when restarting a crashed database instance, the connection pools contain stale connections.
To work around this issue, restart all instances in case of a database failure.
Fuse Integration Services
https://issues.jboss.org/browse/OSFUSe-112: karaf /deployments/karaf/bin/client CNFe org.apache.sshd.agent.SshAgent
Attempting to run the karaf client in the container to locally SSH to the karaf console fails.
Workaround: Adding both shell
and ssh
features make the client work. It will log the warning errors in the logs.
$ oc exec karaf-shell-1-bb9zu -- /deployments/karaf/bin/client osgi:list
These warnings are logged when trying to use the JBoss Fuse bin/client script to connect to the JBoss Fuse micro-container. This is an unusual case, since the container is supposed to contain only bundles and features required for a micro-service, and hence does not need to be managed extensively like a traditional JBoss Fuse install. Any changes made using commands in the remote shell will be temporary and not recorded in the micro-service’s docker image.
https://issues.jboss.org/browse/OSFUSe-190: cdi-camel-jetty S2I template incorrect default service name, breaking cdi-camel-http
The cdi-camel-http quickstart expects the cdi-camel-jetty service to be named qs-cdi-camel-jetty
. In the cdi-camel-jetty template however, the service is named s2i-qs-cdi-camel-jetty
instead by default. This causes the cdi-camel-http to output this error when both are deployed using the S2I with default values.
Workaround: Set the cdi-camel-jetty SeRVICe_NAMe template parameter to qs-cdi-camel-jetty
.
https://issues.jboss.org/browse/OSFUSe-193: karaf-camel-rest-sql template service name too long
oc process karaf-camel-rest-sql template fails with the following error: The Service "s2i-qs-karaf-camel-rest-sql" is invalid. SUReFIRe-859: metadata.name: invalid value 's2i-qs-karaf-camel-rest-sql', Details: must be a DNS 952 label (at most 24 characters, matching regex [a-z]([-a-z0-9]*[a-z0-9])?): e.g. "my-name" deploymentconfig "s2i-quickstart-karaf-camel-rest-sql" created
Workaround: Set SeRVICe_NAMe template parameter to karaf-camel-rest-sql
.
https://issues.jboss.org/browse/OSFUSe-195: karaf-camel-amq template should have parameter to configure A-MQ service name
The application template for A-MQ deployments uses a suffix for every transport type to distinguish. Hence there should be a configurable parameter for setting the service name as environment parameter A_MQ_SeRVICe_NAMe
.
A-MQ
There are no known issues in the A-MQ image.