singleton.war: /META-INF/ /META-INF/MANIFEST.MF /WEB-INF/ /WEB-INF/classes/ /WEB-INF/classes/org/ /WEB-INF/classes/org/jboss/ /WEB-INF/classes/org/jboss/as/ /WEB-INF/classes/org/jboss/as/test/ /WEB-INF/classes/org/jboss/as/test/clustering/ /WEB-INF/classes/org/jboss/as/test/clustering/unmanaged/ /WEB-INF/classes/org/jboss/as/test/clustering/unmanaged/singleton/ /WEB-INF/classes/org/jboss/as/test/clustering/unmanaged/singleton/service/ /WEB-INF/classes/org/jboss/as/test/clustering/unmanaged/singleton/service/MyService.class /WEB-INF/classes/org/jboss/as/test/clustering/unmanaged/singleton/service/MyServiceContextListener.class /WEB-INF/classes/org/jboss/as/test/clustering/unmanaged/singleton/service/MyServiceServlet.class /WEB-INF/classes/org/jboss/as/test/clustering/unmanaged/singleton/service/MyServiceContextListener$NotifyingServiceListener.class /WEB-INF/classes/org/jboss/as/test/clustering/unmanaged/singleton/service/MyServiceContextListener$1.class singleton.war: /META-INF/ /META-INF/MANIFEST.MF /WEB-INF/ /WEB-INF/force-hashcode-change.txt /WEB-INF/classes/ /WEB-INF/classes/org/ /WEB-INF/classes/org/jboss/ /WEB-INF/classes/org/jboss/as/ /WEB-INF/classes/org/jboss/as/test/ /WEB-INF/classes/org/jboss/as/test/clustering/ /WEB-INF/classes/org/jboss/as/test/clustering/unmanaged/ /WEB-INF/classes/org/jboss/as/test/clustering/unmanaged/singleton/ /WEB-INF/classes/org/jboss/as/test/clustering/unmanaged/singleton/service/ /WEB-INF/classes/org/jboss/as/test/clustering/unmanaged/singleton/service/MyService.class /WEB-INF/classes/org/jboss/as/test/clustering/unmanaged/singleton/service/MyServiceContextListener.class /WEB-INF/classes/org/jboss/as/test/clustering/unmanaged/singleton/service/MyServiceServlet.class /WEB-INF/classes/org/jboss/as/test/clustering/unmanaged/singleton/service/MyServiceContextListener$NotifyingServiceListener.class /WEB-INF/classes/org/jboss/as/test/clustering/unmanaged/singleton/service/MyServiceContextListener$1.class System properties: {java.vendor=Apple Inc., localRepository=/Users/bstansberry/.m2/repository, sun.java.launcher=SUN_STANDARD, sun.management.compiler=HotSpot 64-Bit Tiered Compilers, os.name=Mac OS X, sun.boot.class.path=/System/Library/Java/JavaVirtualMachines/1.6.0.jdk/Contents/Classes/jsfd.jar:/System/Library/Java/JavaVirtualMachines/1.6.0.jdk/Contents/Classes/classes.jar:/System/Library/Frameworks/JavaVM.framework/Frameworks/JavaRuntimeSupport.framework/Resources/Java/JavaRuntimeSupport.jar:/System/Library/Java/JavaVirtualMachines/1.6.0.jdk/Contents/Classes/ui.jar:/System/Library/Java/JavaVirtualMachines/1.6.0.jdk/Contents/Classes/laf.jar:/System/Library/Java/JavaVirtualMachines/1.6.0.jdk/Contents/Classes/sunrsasign.jar:/System/Library/Java/JavaVirtualMachines/1.6.0.jdk/Contents/Classes/jsse.jar:/System/Library/Java/JavaVirtualMachines/1.6.0.jdk/Contents/Classes/jce.jar:/System/Library/Java/JavaVirtualMachines/1.6.0.jdk/Contents/Classes/charsets.jar, java.vm.specification.vendor=Sun Microsystems Inc., java.runtime.version=1.6.0_29-b11-402-10M3527, jbossas.ts.submodule.dir=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust, user.name=bstansberry, awt.nativeDoubleBuffering=true, user.language=en, ts.tr.memio=100, sun.boot.library.path=/System/Library/Java/JavaVirtualMachines/1.6.0.jdk/Contents/Libraries, java.util.logging.manager=org.jboss.logmanager.LogManager, jbossas.project.dir=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/../../.., java.version=1.6.0_29, server.jvm.args=-Xmx512m -XX:MaxPermSize=256m -Djboss.dist=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/../../../build/target/jboss-as-7.1.0.Final-SNAPSHOT -Djava.net.preferIPv4Stack=true -Djava.net.preferIPv6Addresses=false -Dts.tr.fsio=100 -Dts.tr.netio=100 -Dts.tr.memio=100 -Dts.tr.db=100 -Dnode0=127.0.0.1 -Dnode1=127.0.0.1 -DudpGroup=230.0.0.4 -Djbossas.ts.submodule.dir=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust -Djbossas.ts.integ.dir=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/.. -Djbossas.ts.dir=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/../.. -Djbossas.project.dir=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/../../.. -Djboss.dist=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/../../../build/target/jboss-as-7.1.0.Final-SNAPSHOT, user.timezone=America/Chicago, java.net.preferIPv4Stack=true, sun.arch.data.model=64, http.nonProxyHosts=local|*.local|169.254/16|*.169.254/16, java.endorsed.dirs=/System/Library/Java/JavaVirtualMachines/1.6.0.jdk/Contents/Home/lib/endorsed, sun.cpu.isalist=, sun.jnu.encoding=MacRoman, file.encoding.pkg=sun.io, jbossas.ts.dir=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/../.., file.separator=/, java.specification.name=Java Platform API Specification, java.class.version=50.0, user.country=US, node1=127.0.0.1, java.home=/System/Library/Java/JavaVirtualMachines/1.6.0.jdk/Contents/Home, node0=127.0.0.1, java.vm.info=mixed mode, arquillian.launch=clustering-udp-unmanaged, os.version=10.6.8, surefire.real.class.path=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/target/surefire/surefirebooter5665103425292242151.jar, path.separator=:, java.vm.version=20.4-b02-402, jboss.server.config.file.name=standalone-ha.xml, java.awt.printerjob=apple.awt.CPrinterJob, sun.io.unicode.encoding=UnicodeLittle, awt.toolkit=apple.awt.CToolkit, module.path=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/../../../build/target/jboss-as-7.1.0.Final-SNAPSHOT/modules, socksNonProxyHosts=local|*.local|169.254/16|*.169.254/16, ftp.nonProxyHosts=local|*.local|169.254/16|*.169.254/16, ts.tr.fsio=100, user.home=/Users/bstansberry, java.specification.vendor=Sun Microsystems Inc., ts.smoke=true, jboss.home=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/target/jbossas, java.library.path=.:/Library/Java/Extensions:/System/Library/Java/Extensions:/usr/lib/java, java.vendor.url=http://www.apple.com/, java.vm.vendor=Apple Inc., gopherProxySet=false, jboss.dist=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/../../../build/target/jboss-as-7.1.0.Final-SNAPSHOT, java.runtime.name=Java(TM) SE Runtime Environment, sun.java.command=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/target/surefire/surefirebooter5665103425292242151.jar /Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/target/surefire/surefire1950426917762447363tmp /Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/target/surefire/surefire4517623595175827033tmp, allTests=true, java.class.path=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/target/test-classes:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/target/classes:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/clustering/singleton/target/jboss-as-clustering-singleton-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/clustering/service/target/jboss-as-clustering-service-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/webservices/tests-integration/target/jboss-as-webservices-tests-integration-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/.m2/repository/org/jboss/ws/jbossws-spi/2.0.1.GA/jbossws-spi-2.0.1.GA.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/controller-client/target/jboss-as-controller-client-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/protocol/target/jboss-as-protocol-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/.m2/repository/org/jboss/xnio/xnio-nio/3.0.0.GA/xnio-nio-3.0.0.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/threads/jboss-threads/2.0.0.GA/jboss-threads-2.0.0.GA.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/security/target/jboss-as-security-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/.m2/repository/org/jboss/marshalling/jboss-marshalling-river/1.3.6.GA/jboss-marshalling-river-1.3.6.GA.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/subsystem-test/target/jboss-as-subsystem-test-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/.m2/repository/org/jboss/logging/jboss-logging/3.1.0.CR2/jboss-logging-3.1.0.CR2.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/controller/target/jboss-as-controller-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/server/target/jboss-as-server-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/domain-http/interface/target/jboss-as-domain-http-interface-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/deployment-repository/target/jboss-as-deployment-repository-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/embedded/target/jboss-as-embedded-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/platform-mbean/target/jboss-as-platform-mbean-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/process-controller/target/jboss-as-process-controller-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/.m2/repository/org/jboss/remoting3/jboss-remoting/3.2.0.CR9/jboss-remoting-3.2.0.CR9.jar:/Users/bstansberry/.m2/repository/org/jboss/stdio/jboss-stdio/1.0.1.GA/jboss-stdio-1.0.1.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/jboss-vfs/3.1.0.CR1/jboss-vfs-3.1.0.CR1.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/demos/legacy/target/jboss-as-demos-legacy-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/.m2/repository/org/jboss/spec/javax/ws/rs/jboss-jaxrs-api_1.1_spec/1.0.0.Final/jboss-jaxrs-api_1.1_spec-1.0.0.Final.jar:/Users/bstansberry/.m2/repository/org/hornetq/hornetq-core/2.2.10.Final/hornetq-core-2.2.10.Final.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/jmx/target/jboss-as-jmx-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/.m2/repository/org/jboss/jboss-common-core/2.2.17.GA/jboss-common-core-2.2.17.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/spec/javax/resource/jboss-connector-api_1.6_spec/1.0.0.Final/jboss-connector-api_1.6_spec-1.0.0.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/shrinkwrap/shrinkwrap-api/1.0.0-cr-1/shrinkwrap-api-1.0.0-cr-1.jar:/Users/bstansberry/.m2/repository/org/jboss/shrinkwrap/shrinkwrap-impl-base/1.0.0-cr-1/shrinkwrap-impl-base-1.0.0-cr-1.jar:/Users/bstansberry/.m2/repository/org/jboss/spec/javax/jms/jboss-jms-api_1.1_spec/1.0.0.Final/jboss-jms-api_1.1_spec-1.0.0.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/netty/netty/3.2.6.Final/netty-3.2.6.Final.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/connector/target/jboss-as-connector-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/logging/target/jboss-as-logging-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/messaging/target/jboss-as-messaging-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/network/target/jboss-as-network-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/naming/target/jboss-as-naming-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/osgi/service/target/jboss-as-osgi-service-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/remoting/target/jboss-as-remoting-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/domain-management/target/jboss-as-domain-management-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/.m2/repository/org/jboss/xnio/xnio-api/3.0.0.GA/xnio-api-3.0.0.GA.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/sar/target/jboss-as-sar-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/threads/target/jboss-as-threads-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/transactions/target/jboss-as-transactions-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/web/target/jboss-as-web-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/weld/target/jboss-as-weld-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/.m2/repository/org/jboss/jboss-transaction-spi/7.0.0.CR2/jboss-transaction-spi-7.0.0.CR2.jar:/Users/bstansberry/.m2/repository/org/jboss/spec/javax/xml/ws/jboss-jaxws-api_2.2_spec/2.0.0.Final/jboss-jaxws-api_2.2_spec-2.0.0.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/ws/cxf/jbossws-cxf-client/4.0.0.GA/jbossws-cxf-client-4.0.0.GA.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-rt-frontend-jaxws/2.4.4/cxf-rt-frontend-jaxws-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-rt-frontend-simple/2.4.4/cxf-rt-frontend-simple-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-common-utilities/2.4.4/cxf-common-utilities-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-rt-core/2.4.4/cxf-rt-core-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-rt-ws-addr/2.4.4/cxf-rt-ws-addr-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-rt-ws-policy/2.4.4/cxf-rt-ws-policy-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-rt-ws-rm/2.4.4/cxf-rt-ws-rm-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-rt-transports-http/2.4.4/cxf-rt-transports-http-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-rt-databinding-jaxb/2.4.4/cxf-rt-databinding-jaxb-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-tools-common/2.4.4/cxf-tools-common-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/ws/xmlschema/xmlschema-core/2.0/xmlschema-core-2.0.jar:/Users/bstansberry/.m2/repository/org/apache/neethi/neethi/3.0.1/neethi-3.0.1.jar:/Users/bstansberry/.m2/repository/wsdl4j/wsdl4j/1.6.2/wsdl4j-1.6.2.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/demos/spec/target/jboss-as-demos-spec-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/.m2/repository/org/jboss/spec/javax/ejb/jboss-ejb-api_3.1_spec/1.0.1.Final/jboss-ejb-api_3.1_spec-1.0.1.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/spec/javax/el/jboss-el-api_2.2_spec/1.0.0.Final/jboss-el-api_2.2_spec-1.0.0.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/spec/javax/enterprise/deploy/jboss-jad-api_1.2_spec/1.0.0.Final/jboss-jad-api_1.2_spec-1.0.0.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/spec/javax/faces/jboss-jsf-api_2.1_spec/2.0.0.Beta1/jboss-jsf-api_2.1_spec-2.0.0.Beta1.jar:/Users/bstansberry/.m2/repository/org/jboss/spec/javax/interceptor/jboss-interceptors-api_1.1_spec/1.0.0.Final/jboss-interceptors-api_1.1_spec-1.0.0.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/spec/javax/security/auth/message/jboss-jaspi-api_1.0_spec/1.0.0.Final/jboss-jaspi-api_1.0_spec-1.0.0.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/spec/javax/security/jacc/jboss-jacc-api_1.4_spec/1.0.1.Final/jboss-jacc-api_1.4_spec-1.0.1.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/spec/javax/servlet/jboss-servlet-api_3.0_spec/1.0.0.Final/jboss-servlet-api_3.0_spec-1.0.0.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/spec/javax/servlet/jsp/jboss-jsp-api_2.2_spec/1.0.0.Final/jboss-jsp-api_2.2_spec-1.0.0.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/spec/javax/servlet/jstl/jboss-jstl-api_1.2_spec/1.0.2.Final/jboss-jstl-api_1.2_spec-1.0.2.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/spec/javax/transaction/jboss-transaction-api_1.1_spec/1.0.0.Final/jboss-transaction-api_1.1_spec-1.0.0.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/spec/javax/xml/bind/jboss-jaxb-api_2.2_spec/1.0.3.Final/jboss-jaxb-api_2.2_spec-1.0.3.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/spec/javax/xml/registry/jboss-jaxr-api_1.0_spec/1.0.0.Final/jboss-jaxr-api_1.0_spec-1.0.0.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/spec/javax/xml/rpc/jboss-jaxrpc-api_1.1_spec/1.0.0.Final/jboss-jaxrpc-api_1.1_spec-1.0.0.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/spec/javax/xml/soap/jboss-saaj-api_1.3_spec/1.0.1.Final/jboss-saaj-api_1.3_spec-1.0.1.Final.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/shared/target/jboss-as-testsuite-shared-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/.m2/repository/log4j/log4j/1.2.16/log4j-1.2.16.jar:/Users/bstansberry/.m2/repository/org/jboss/jandex/1.0.3.Final/jandex-1.0.3.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/jboss-dmr/1.1.1.Final/jboss-dmr-1.1.1.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/staxmapper/1.0.0.Final/staxmapper-1.0.0.Final.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/domain-http/error-context/target/jboss-as-domain-http-error-context-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/.m2/repository/org/jboss/as/jbossweb-native/2.0.10.Beta1/jbossweb-native-2.0.10.Beta1.jar:/Users/bstansberry/.m2/repository/org/jboss/invocation/jboss-invocation/1.1.1.Final/jboss-invocation-1.1.1.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/logmanager/jboss-logmanager/1.2.0.GA/jboss-logmanager-1.2.0.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/logmanager/jboss-logmanager-log4j/1.0.0.GA/jboss-logmanager-log4j-1.0.0.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/marshalling/jboss-marshalling/1.3.6.GA/jboss-marshalling-1.3.6.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/modules/jboss-modules/1.1.0.CR8/jboss-modules-1.1.0.CR8.jar:/Users/bstansberry/.m2/repository/org/jboss/msc/jboss-msc/1.0.1.GA/jboss-msc-1.0.1.GA.jar:/Users/bstansberry/.m2/repository/antlr/antlr/2.7.7/antlr-2.7.7.jar:/Users/bstansberry/.m2/repository/asm/asm/3.3.1/asm-3.3.1.jar:/Users/bstansberry/.m2/repository/ch/qos/cal10n/cal10n-api/0.7.3/cal10n-api-0.7.3.jar:/Users/bstansberry/.m2/repository/com/h2database/h2/1.3.161/h2-1.3.161.jar:/Users/bstansberry/.m2/repository/com/google/guava/guava/10.0.1/guava-10.0.1.jar:/Users/bstansberry/.m2/repository/com/sun/faces/jsf-impl/2.1.5-jbossorg-1/jsf-impl-2.1.5-jbossorg-1.jar:/Users/bstansberry/.m2/repository/com/sun/xml/bind/jaxb-impl/2.2/jaxb-impl-2.2.jar:/Users/bstansberry/.m2/repository/com/sun/xml/bind/jaxb-xjc/2.2/jaxb-xjc-2.2.jar:/Users/bstansberry/.m2/repository/com/sun/xml/messaging/saaj/saaj-impl/1.3.16-jbossorg-1/saaj-impl-1.3.16-jbossorg-1.jar:/Users/bstansberry/.m2/repository/commons-beanutils/commons-beanutils/1.8.0/commons-beanutils-1.8.0.jar:/Users/bstansberry/.m2/repository/commons-cli/commons-cli/1.2/commons-cli-1.2.jar:/Users/bstansberry/.m2/repository/commons-codec/commons-codec/1.4/commons-codec-1.4.jar:/Users/bstansberry/.m2/repository/commons-collections/commons-collections/3.2.1/commons-collections-3.2.1.jar:/Users/bstansberry/.m2/repository/commons-io/commons-io/2.1/commons-io-2.1.jar:/Users/bstansberry/.m2/repository/commons-lang/commons-lang/2.4/commons-lang-2.4.jar:/Users/bstansberry/.m2/repository/commons-pool/commons-pool/1.5.6/commons-pool-1.5.6.jar:/Users/bstansberry/.m2/repository/dom4j/dom4j/1.6.1/dom4j-1.6.1.jar:/Users/bstansberry/.m2/repository/gnu-getopt/getopt/1.0.13/getopt-1.0.13.jar:/Users/bstansberry/.m2/repository/org/jacorb/jacorb/2.3.1.jbossorg-1/jacorb-2.3.1.jbossorg-1.jar:/Users/bstansberry/.m2/repository/javax/activation/activation/1.1.1/activation-1.1.1.jar:/Users/bstansberry/.m2/repository/javax/enterprise/cdi-api/1.0-SP4/cdi-api-1.0-SP4.jar:/Users/bstansberry/.m2/repository/javax/faces/jsf-api/1.2_15-jbossorg-2/jsf-api-1.2_15-jbossorg-2.jar:/Users/bstansberry/.m2/repository/javax/faces/jsf-impl/1.2_15-jbossorg-2/jsf-impl-1.2_15-jbossorg-2.jar:/Users/bstansberry/.m2/repository/javax/inject/javax.inject/1/javax.inject-1.jar:/Users/bstansberry/.m2/repository/javax/jws/jsr181-api/1.0-MR1/jsr181-api-1.0-MR1.jar:/Users/bstansberry/.m2/repository/javax/mail/mail/1.4.4/mail-1.4.4.jar:/Users/bstansberry/.m2/repository/javax/validation/validation-api/1.0.0.GA/validation-api-1.0.0.GA.jar:/Users/bstansberry/.m2/repository/jaxen/jaxen/1.1.3/jaxen-1.1.3.jar:/Users/bstansberry/.m2/repository/jboss/jaxbintros/jboss-jaxb-intros/1.0.2.GA/jboss-jaxb-intros-1.0.2.GA.jar:/Users/bstansberry/.m2/repository/jline/jline/0.9.94/jline-0.9.94.jar:/Users/bstansberry/.m2/repository/joda-time/joda-time/1.6.2/joda-time-1.6.2.jar:/Users/bstansberry/.m2/repository/net/sourceforge/cssparser/cssparser/0.9.5/cssparser-0.9.5.jar:/Users/bstansberry/.m2/repository/net/sourceforge/htmlunit/htmlunit/2.8/htmlunit-2.8.jar:/Users/bstansberry/.m2/repository/net/sourceforge/htmlunit/htmlunit-core-js/2.8/htmlunit-core-js-2.8.jar:/Users/bstansberry/.m2/repository/net/sourceforge/nekohtml/nekohtml/1.9.14/nekohtml-1.9.14.jar:/Users/bstansberry/.m2/repository/org/apache/ant/ant/1.8.2/ant-1.8.2.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-api/2.4.4/cxf-api-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-rt-bindings-coloc/2.4.4/cxf-rt-bindings-coloc-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-rt-bindings-http/2.4.4/cxf-rt-bindings-http-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-rt-bindings-object/2.4.4/cxf-rt-bindings-object-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-rt-bindings-soap/2.4.4/cxf-rt-bindings-soap-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-rt-bindings-xml/2.4.4/cxf-rt-bindings-xml-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-rt-databinding-aegis/2.4.4/cxf-rt-databinding-aegis-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-rt-management/2.4.4/cxf-rt-management-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-rt-transports-common/2.4.4/cxf-rt-transports-common-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-rt-transports-local/2.4.4/cxf-rt-transports-local-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-rt-transports-jms/2.4.4/cxf-rt-transports-jms-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-rt-ws-security/2.4.4/cxf-rt-ws-security-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-tools-java2ws/2.4.4/cxf-tools-java2ws-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-tools-validator/2.4.4/cxf-tools-validator-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-tools-wsdlto-core/2.4.4/cxf-tools-wsdlto-core-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-tools-wsdlto-databinding-jaxb/2.4.4/cxf-tools-wsdlto-databinding-jaxb-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-tools-wsdlto-frontend-jaxws/2.4.4/cxf-tools-wsdlto-frontend-jaxws-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/xjcplugins/cxf-xjc-boolean/2.4.0/cxf-xjc-boolean-2.4.0.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/xjcplugins/cxf-xjc-bug671/2.4.0/cxf-xjc-bug671-2.4.0.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/xjcplugins/cxf-xjc-dv/2.4.0/cxf-xjc-dv-2.4.0.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/xjcplugins/cxf-xjc-ts/2.4.0/cxf-xjc-ts-2.4.0.jar:/Users/bstansberry/.m2/repository/org/apache/felix/org.apache.felix.configadmin/1.2.8/org.apache.felix.configadmin-1.2.8.jar:/Users/bstansberry/.m2/repository/org/apache/httpcomponents/httpclient/4.1.2/httpclient-4.1.2.jar:/Users/bstansberry/.m2/repository/org/apache/httpcomponents/httpcore/4.1.2/httpcore-4.1.2.jar:/Users/bstansberry/.m2/repository/org/apache/httpcomponents/httpmime/4.1.2/httpmime-4.1.2.jar:/Users/bstansberry/.m2/repository/org/apache/james/apache-mime4j/0.6/apache-mime4j-0.6.jar:/Users/bstansberry/.m2/repository/org/apache/santuario/xmlsec/1.4.5/xmlsec-1.4.5.jar:/Users/bstansberry/.m2/repository/org/apache/ws/scout/scout/1.1.1/scout-1.1.1.jar:/Users/bstansberry/.m2/repository/org/apache/velocity/velocity/1.6.3/velocity-1.6.3.jar:/Users/bstansberry/.m2/repository/org/apache/ws/security/wss4j/1.6.3/wss4j-1.6.3.jar:/Users/bstansberry/.m2/repository/org/codehaus/jackson/jackson-core-asl/1.9.2/jackson-core-asl-1.9.2.jar:/Users/bstansberry/.m2/repository/org/codehaus/jackson/jackson-mapper-asl/1.9.2/jackson-mapper-asl-1.9.2.jar:/Users/bstansberry/.m2/repository/org/codehaus/jackson/jackson-jaxrs/1.9.2/jackson-jaxrs-1.9.2.jar:/Users/bstansberry/.m2/repository/org/codehaus/jackson/jackson-xc/1.9.2/jackson-xc-1.9.2.jar:/Users/bstansberry/.m2/repository/org/codehaus/jettison/jettison/1.3.1/jettison-1.3.1.jar:/Users/bstansberry/.m2/repository/stax/stax-api/1.0.1/stax-api-1.0.1.jar:/Users/bstansberry/.m2/repository/org/codehaus/woodstox/woodstox-core-asl/4.1.1/woodstox-core-asl-4.1.1.jar:/Users/bstansberry/.m2/repository/org/codehaus/woodstox/stax2-api/3.1.1/stax2-api-3.1.1.jar:/Users/bstansberry/.m2/repository/org/hibernate/hibernate-core/4.0.1.Final/hibernate-core-4.0.1.Final.jar:/Users/bstansberry/.m2/repository/org/hibernate/common/hibernate-commons-annotations/4.0.1.Final/hibernate-commons-annotations-4.0.1.Final.jar:/Users/bstansberry/.m2/repository/org/hibernate/hibernate-entitymanager/4.0.1.Final/hibernate-entitymanager-4.0.1.Final.jar:/Users/bstansberry/.m2/repository/org/hibernate/hibernate-infinispan/4.0.1.Final/hibernate-infinispan-4.0.1.Final.jar:/Users/bstansberry/.m2/repository/org/hibernate/hibernate-validator/4.2.0.Final/hibernate-validator-4.2.0.Final.jar:/Users/bstansberry/.m2/repository/org/hibernate/hibernate-envers/4.0.1.Final/hibernate-envers-4.0.1.Final.jar:/Users/bstansberry/.m2/repository/org/hibernate/javax/persistence/hibernate-jpa-2.0-api/1.0.1.Final/hibernate-jpa-2.0-api-1.0.1.Final.jar:/Users/bstansberry/.m2/repository/org/hornetq/hornetq-jms/2.2.10.Final/hornetq-jms-2.2.10.Final.jar:/Users/bstansberry/.m2/repository/org/hornetq/hornetq-ra/2.2.10.Final/hornetq-ra-2.2.10.Final.jar:/Users/bstansberry/.m2/repository/org/infinispan/infinispan-cachestore-jdbc/5.1.0.FINAL/infinispan-cachestore-jdbc-5.1.0.FINAL.jar:/Users/bstansberry/.m2/repository/org/rhq/helpers/rhq-pluginAnnotations/3.0.4/rhq-pluginAnnotations-3.0.4.jar:/Users/bstansberry/.m2/repository/org/infinispan/infinispan-cachestore-remote/5.1.0.FINAL/infinispan-cachestore-remote-5.1.0.FINAL.jar:/Users/bstansberry/.m2/repository/org/infinispan/infinispan-client-hotrod/5.1.0.FINAL/infinispan-client-hotrod-5.1.0.FINAL.jar:/Users/bstansberry/.m2/repository/org/infinispan/infinispan-core/5.1.0.FINAL/infinispan-core-5.1.0.FINAL.jar:/Users/bstansberry/.m2/repository/org/javassist/javassist/3.15.0-GA/javassist-3.15.0-GA.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/appclient/target/jboss-as-appclient-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/cli/target/jboss-as-cli-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/clustering/api/target/jboss-as-clustering-api-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/clustering/common/target/jboss-as-clustering-common-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/clustering/ejb3-infinispan/target/jboss-as-clustering-ejb3-infinispan-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/clustering/impl/target/jboss-as-clustering-impl-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/clustering/infinispan/target/jboss-as-clustering-infinispan-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/clustering/jgroups/target/jboss-as-clustering-jgroups-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/clustering/registry/target/jboss-as-clustering-registry-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/clustering/web-infinispan/target/jboss-as-clustering-web-infinispan-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/clustering/web-spi/target/jboss-as-clustering-web-spi-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/cmp/target/jboss-as-cmp-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/configadmin/target/jboss-as-configadmin-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/.m2/repository/org/jboss/as/jboss-as-console/1.0.0.RC1/jboss-as-console-1.0.0.RC1-resources.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/deployment-scanner/target/jboss-as-deployment-scanner-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/ee/target/jboss-as-ee-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/ee-deployment/target/jboss-as-ee-deployment-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/ejb3/target/jboss-as-ejb3-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/.m2/repository/org/jboss/jboss-iiop-client/1.0.0.Beta2/jboss-iiop-client-1.0.0.Beta2.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/jaxr/target/jboss-as-jaxr-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/jaxrs/target/jboss-as-jaxrs-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/jdr/jboss-as-jdr/target/jboss-as-jdr-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/.m2/repository/org/jboss/logging/jboss-logging-processor/1.0.0.CR5/jboss-logging-processor-1.0.0.CR5.jar:/Users/bstansberry/.m2/repository/org/jboss/logging/jboss-logging-generator/1.0.0.CR5/jboss-logging-generator-1.0.0.CR5.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/jsr77/target/jboss-as-jsr77-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/host-controller/target/jboss-as-host-controller-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/management-client-content/target/jboss-as-management-client-content-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/jacorb/target/jboss-as-jacorb-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/mail/target/jboss-as-mail-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/modcluster/target/jboss-as-modcluster-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/pojo/target/jboss-as-pojo-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/osgi/configadmin/target/jboss-as-osgi-configadmin-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/jdr/jboss-as-sos/target/jboss-as-sos-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/webservices/server-integration/target/jboss-as-webservices-server-integration-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/xts/target/jboss-as-xts-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/.m2/repository/org/jboss/classfilewriter/jboss-classfilewriter/1.0.0.Final/jboss-classfilewriter-1.0.0.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/com/sun/httpserver/httpserver/1.0.0.Beta3/httpserver-1.0.0.Beta3.jar:/Users/bstansberry/.m2/repository/org/jboss/jboss-ejb-client/1.0.0.Beta12/jboss-ejb-client-1.0.0.Beta12.jar:/Users/bstansberry/.m2/repository/org/jboss/ejb3/jboss-ejb3-ext-api/2.0.0-beta-3/jboss-ejb3-ext-api-2.0.0-beta-3.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/jpa/core/target/jboss-as-jpa-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/jpa/hibernate3/target/jboss-as-jpa-hibernate3-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/.m2/repository/org/hibernate/hibernate-commons-annotations/3.2.0.Final/hibernate-commons-annotations-3.2.0.Final.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/jpa/hibernate4/target/jboss-as-jpa-hibernate4-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/jpa/hibernate-infinispan/target/jboss-as-jpa-hibernate-infinispan-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/jpa/openjpa/target/jboss-as-jpa-openjpa-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/.m2/repository/org/apache/openjpa/openjpa-kernel/2.1.1/openjpa-kernel-2.1.1.jar:/Users/bstansberry/.m2/repository/org/apache/openjpa/openjpa-lib/2.1.1/openjpa-lib-2.1.1.jar:/Users/bstansberry/.m2/repository/org/apache/openjpa/openjpa-persistence/2.1.1/openjpa-persistence-2.1.1.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/jpa/spi/target/jboss-as-jpa-spi-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/jpa/util/target/jboss-as-jpa-util-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/.m2/repository/org/jboss/interceptor/jboss-interceptor-core/2.0.0.Alpha3/jboss-interceptor-core-2.0.0.Alpha3.jar:/Users/bstansberry/.m2/repository/org/jboss/interceptor/jboss-interceptor-spi/2.0.0.Alpha3/jboss-interceptor-spi-2.0.0.Alpha3.jar:/Users/bstansberry/.m2/repository/org/jboss/ironjacamar/ironjacamar-core-api/1.0.7.Final/ironjacamar-core-api-1.0.7.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/ironjacamar/ironjacamar-core-impl/1.0.7.Final/ironjacamar-core-impl-1.0.7.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/ironjacamar/ironjacamar-common-api/1.0.7.Final/ironjacamar-common-api-1.0.7.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/ironjacamar/ironjacamar-common-spi/1.0.7.Final/ironjacamar-common-spi-1.0.7.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/ironjacamar/ironjacamar-common-impl/1.0.7.Final/ironjacamar-common-impl-1.0.7.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/ironjacamar/ironjacamar-deployers-common/1.0.7.Final/ironjacamar-deployers-common-1.0.7.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/ironjacamar/ironjacamar-jdbc/1.0.7.Final/ironjacamar-jdbc-1.0.7.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/ironjacamar/ironjacamar-spec-api/1.0.7.Final/ironjacamar-spec-api-1.0.7.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/ironjacamar/ironjacamar-validator/1.0.7.Final/ironjacamar-validator-1.0.7.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/jbossts/jbossjts/4.16.1.Final/jbossjts-4.16.1.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/jbossts/jbossjts-integration/4.16.1.Final/jbossjts-integration-4.16.1.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/jbossts/jbosstxbridge/4.16.1.Final/jbosstxbridge-4.16.1.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/jbossts/jbossxts/4.16.1.Final/jbossxts-4.16.1.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/jbossts/jbossxts-api/4.16.1.Final/jbossxts-api-4.16.1.Final.jar:/Users/bstansberry/.m2/repository/jacorb/jacorb/2.3.1jboss.patch01-brew/jacorb-2.3.1jboss.patch01-brew.jar:/Users/bstansberry/.m2/repository/org/jboss/jsfunit/jboss-jsfunit-core/2.0.0.Beta1/jboss-jsfunit-core-2.0.0.Beta1.jar:/Users/bstansberry/.m2/repository/org/jboss/metadata/jboss-metadata-appclient/7.0.0.Beta33/jboss-metadata-appclient-7.0.0.Beta33.jar:/Users/bstansberry/.m2/repository/org/jboss/metadata/jboss-metadata-common/7.0.0.Beta33/jboss-metadata-common-7.0.0.Beta33.jar:/Users/bstansberry/.m2/repository/org/jboss/metadata/jboss-metadata-ear/7.0.0.Beta33/jboss-metadata-ear-7.0.0.Beta33.jar:/Users/bstansberry/.m2/repository/org/jboss/metadata/jboss-metadata-ejb/7.0.0.Beta33/jboss-metadata-ejb-7.0.0.Beta33.jar:/Users/bstansberry/.m2/repository/org/jboss/metadata/jboss-metadata-web/7.0.0.Beta33/jboss-metadata-web-7.0.0.Beta33.jar:/Users/bstansberry/.m2/repository/org/jboss/mod_cluster/mod_cluster-container-spi/1.2.0.Beta4/mod_cluster-container-spi-1.2.0.Beta4.jar:/Users/bstansberry/.m2/repository/org/jboss/mod_cluster/mod_cluster-core/1.2.0.Beta4/mod_cluster-core-1.2.0.Beta4.jar:/Users/bstansberry/.m2/repository/net/jcip/jcip-annotations/1.0/jcip-annotations-1.0.jar:/Users/bstansberry/.m2/repository/org/jboss/mod_cluster/mod_cluster-container-catalina/1.2.0.Beta4/mod_cluster-container-catalina-1.2.0.Beta4.jar:/Users/bstansberry/.m2/repository/org/jboss/mod_cluster/mod_cluster-container-jbossweb/1.2.0.Beta4/mod_cluster-container-jbossweb-1.2.0.Beta4.jar:/Users/bstansberry/.m2/repository/org/jboss/osgi/deployment/jbosgi-deployment/1.0.10/jbosgi-deployment-1.0.10.jar:/Users/bstansberry/.m2/repository/org/jboss/osgi/framework/jbosgi-framework-core/1.1.0/jbosgi-framework-core-1.1.0.jar:/Users/bstansberry/.m2/repository/org/jboss/osgi/resolver/jbosgi-resolver-api/1.0.10/jbosgi-resolver-api-1.0.10.jar:/Users/bstansberry/.m2/repository/org/jboss/osgi/resolver/jbosgi-resolver-felix/1.0.10/jbosgi-resolver-felix-1.0.10.jar:/Users/bstansberry/.m2/repository/org/jboss/osgi/resolver/jbosgi-resolver-metadata/1.0.10/jbosgi-resolver-metadata-1.0.10.jar:/Users/bstansberry/.m2/repository/org/jboss/osgi/resolver/jbosgi-resolver-spi/1.0.10/jbosgi-resolver-spi-1.0.10.jar:/Users/bstansberry/.m2/repository/org/jboss/osgi/spi/jbosgi-spi/1.0.28/jbosgi-spi-1.0.28.jar:/Users/bstansberry/.m2/repository/org/jboss/osgi/vfs/jbosgi-vfs/1.0.6/jbosgi-vfs-1.0.6.jar:/Users/bstansberry/.m2/repository/org/jboss/osgi/vfs/jbosgi-vfs30/1.0.6/jbosgi-vfs30-1.0.6.jar:/Users/bstansberry/.m2/repository/org/jboss/remoting3/remoting-jmx/1.0.0.Beta1/remoting-jmx-1.0.0.Beta1.jar:/Users/bstansberry/.m2/repository/org/jboss/jboss-remote-naming/1.0.0.Beta3/jboss-remote-naming-1.0.0.Beta3.jar:/Users/bstansberry/.m2/repository/org/jboss/resteasy/resteasy-jaxrs/2.3.1.GA/resteasy-jaxrs-2.3.1.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/resteasy/async-http-servlet-3.0/2.3.1.GA/async-http-servlet-3.0-2.3.1.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/resteasy/resteasy-atom-provider/2.3.1.GA/resteasy-atom-provider-2.3.1.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/resteasy/resteasy-cdi/2.3.1.GA/resteasy-cdi-2.3.1.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/resteasy/resteasy-jackson-provider/2.3.1.GA/resteasy-jackson-provider-2.3.1.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/resteasy/resteasy-jaxb-provider/2.3.1.GA/resteasy-jaxb-provider-2.3.1.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/resteasy/resteasy-jettison-provider/2.3.1.GA/resteasy-jettison-provider-2.3.1.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/resteasy/resteasy-jsapi/2.3.1.GA/resteasy-jsapi-2.3.1.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/resteasy/resteasy-multipart-provider/2.3.1.GA/resteasy-multipart-provider-2.3.1.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/resteasy/resteasy-yaml-provider/2.3.1.GA/resteasy-yaml-provider-2.3.1.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/seam/integration/jboss-seam-int-jbossas/6.0.0.GA/jboss-seam-int-jbossas-6.0.0.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/security/jboss-negotiation/2.2.0.Beta3/jboss-negotiation-2.2.0.Beta3.jar:/Users/bstansberry/.m2/repository/org/jboss/security/jbossxacml/2.0.6.Final/jbossxacml-2.0.6.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/shrinkwrap/shrinkwrap-spi/1.0.0-cr-1/shrinkwrap-spi-1.0.0-cr-1.jar:/Users/bstansberry/.m2/repository/org/jboss/slf4j/slf4j-jboss-logmanager/1.0.0.GA/slf4j-jboss-logmanager-1.0.0.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/spec/javax/management/j2ee/jboss-j2eemgmt-api_1.1_spec/1.0.0.Final/jboss-j2eemgmt-api_1.1_spec-1.0.0.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/spec/javax/rmi/jboss-rmi-api_1.0_spec/1.0.4.Final/jboss-rmi-api_1.0_spec-1.0.4.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/web/jasper-jdt/7.0.3.Final/jasper-jdt-7.0.3.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/web/jbossweb/7.0.9.Final/jbossweb-7.0.9.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/weld/weld-api/1.1.Final/weld-api-1.1.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/weld/weld-core/1.1.5.Final/weld-core-1.1.5.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/weld/weld-spi/1.1.Final/weld-spi-1.1.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/ws/jbossws-api/1.0.0.GA/jbossws-api-1.0.0.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/ws/jbossws-common/2.0.0.GA/jbossws-common-2.0.0.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/ws/jbossws-common-tools/1.0.0.GA/jbossws-common-tools-1.0.0.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/ws/projects/jaxws-jboss-httpserver-httpspi/1.0.0.GA/jaxws-jboss-httpserver-httpspi-1.0.0.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/ws/cxf/jbossws-cxf-factories/4.0.0.GA/jbossws-cxf-factories-4.0.0.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/ws/cxf/jbossws-cxf-resources/4.0.0.GA/jbossws-cxf-resources-4.0.0.GA-jboss700.jar:/Users/bstansberry/.m2/repository/org/jboss/ws/cxf/jbossws-cxf-server/4.0.0.GA/jbossws-cxf-server-4.0.0.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/ws/cxf/jbossws-cxf-transports-httpserver/4.0.0.GA/jbossws-cxf-transports-httpserver-4.0.0.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/ws/native/jbossws-native-core/4.0.0.GA/jbossws-native-core-4.0.0.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/ws/native/jbossws-native-factories/4.0.0.GA/jbossws-native-factories-4.0.0.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/ws/native/jbossws-native-services/4.0.0.GA/jbossws-native-services-4.0.0.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/jbossxb/2.0.3.GA/jbossxb-2.0.3.GA.jar:/Users/bstansberry/.m2/repository/org/jdom/jdom/1.1.2/jdom-1.1.2.jar:/Users/bstansberry/.m2/repository/org/jgroups/jgroups/3.0.3.Final/jgroups-3.0.3.Final.jar:/Users/bstansberry/.m2/repository/org/osgi/org.osgi.core/4.2.0/org.osgi.core-4.2.0.jar:/Users/bstansberry/.m2/repository/org/osgi/org.osgi.compendium/4.2.0/org.osgi.compendium-4.2.0.jar:/Users/bstansberry/.m2/repository/org/picketlink/picketlink-fed/2.0.1.final/picketlink-fed-2.0.1.final.jar:/Users/bstansberry/.m2/repository/org/picketlink/picketlink-bindings/2.0.1.final/picketlink-bindings-2.0.1.final.jar:/Users/bstansberry/.m2/repository/org/picketlink/picketlink-bindings-jboss/2.0.1.final/picketlink-bindings-jboss-2.0.1.final.jar:/Users/bstansberry/.m2/repository/org/picketlink/picketlink-trust-jbossws/2.0.1.final/picketlink-trust-jbossws-2.0.1.final.jar:/Users/bstansberry/.m2/repository/org/projectodd/stilts/stilts-stomplet-server-bundle/0.1.20/stilts-stomplet-server-bundle-0.1.20.jar:/Users/bstansberry/.m2/repository/org/scannotation/scannotation/1.0.2/scannotation-1.0.2.jar:/Users/bstansberry/.m2/repository/org/picketbox/picketbox/4.0.6.final/picketbox-4.0.6.final.jar:/Users/bstansberry/.m2/repository/org/picketbox/picketbox-commons/1.0.0.final/picketbox-commons-1.0.0.final.jar:/Users/bstansberry/.m2/repository/org/picketbox/picketbox-infinispan/4.0.6.final/picketbox-infinispan-4.0.6.final.jar:/Users/bstansberry/.m2/repository/org/python/jython-standalone/2.5.2/jython-standalone-2.5.2.jar:/Users/bstansberry/.m2/repository/org/slf4j/jcl-over-slf4j/1.6.1/jcl-over-slf4j-1.6.1.jar:/Users/bstansberry/.m2/repository/org/slf4j/slf4j-api/1.6.1/slf4j-api-1.6.1.jar:/Users/bstansberry/.m2/repository/org/slf4j/slf4j-ext/1.6.1/slf4j-ext-1.6.1.jar:/Users/bstansberry/.m2/repository/org/w3c/css/sac/1.3/sac-1.3.jar:/Users/bstansberry/.m2/repository/org/yaml/snakeyaml/1.8/snakeyaml-1.8.jar:/Users/bstansberry/.m2/repository/xalan/serializer/2.7.1.jbossorg-1/serializer-2.7.1.jbossorg-1.jar:/Users/bstansberry/.m2/repository/xalan/xalan/2.7.1.jbossorg-1/xalan-2.7.1.jbossorg-1.jar:/Users/bstansberry/.m2/repository/xerces/xercesImpl/2.9.1-jbossas-1/xercesImpl-2.9.1-jbossas-1.jar:/Users/bstansberry/.m2/repository/xml-resolver/xml-resolver/1.2/xml-resolver-1.2.jar:/Users/bstansberry/.m2/repository/xom/xom/1.2.5/xom-1.2.5.jar:/Users/bstansberry/.m2/repository/org/projectodd/stilts/stilts-stomp-client/0.1.20/stilts-stomp-client-0.1.20.jar:/Users/bstansberry/.m2/repository/org/projectodd/stilts/stilts-stomp-api/0.1.20/stilts-stomp-api-0.1.20.jar:/Users/bstansberry/.m2/repository/org/projectodd/stilts/stilts-stomp-common/0.1.20/stilts-stomp-common-0.1.20.jar:/Users/bstansberry/.m2/repository/org/projectodd/stilts/stilts-stomplet-server-core/0.1.20/stilts-stomplet-server-core-0.1.20.jar:/Users/bstansberry/.m2/repository/org/projectodd/stilts/stilts-conduit-server-core/0.1.20/stilts-conduit-server-core-0.1.20.jar:/Users/bstansberry/.m2/repository/org/projectodd/stilts/stilts-stomp-server-core/0.1.20/stilts-stomp-server-core-0.1.20.jar:/Users/bstansberry/.m2/repository/org/projectodd/stilts/stilts-conduit-server-spi/0.1.20/stilts-conduit-server-spi-0.1.20.jar:/Users/bstansberry/.m2/repository/org/projectodd/stilts/stilts-stomplet-api/0.1.20/stilts-stomplet-api-0.1.20.jar:/Users/bstansberry/.m2/repository/org/projectodd/stilts/stilts-stomp-server-spi/0.1.20/stilts-stomp-server-spi-0.1.20.jar:/Users/bstansberry/.m2/repository/org/jboss/spec/javax/annotation/jboss-annotations-api_1.1_spec/1.0.0.Final/jboss-annotations-api_1.1_spec-1.0.0.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/arquillian/junit/arquillian-junit-container/1.0.0.CR7/arquillian-junit-container-1.0.0.CR7.jar:/Users/bstansberry/.m2/repository/org/jboss/arquillian/junit/arquillian-junit-core/1.0.0.CR7/arquillian-junit-core-1.0.0.CR7.jar:/Users/bstansberry/.m2/repository/org/jboss/arquillian/test/arquillian-test-api/1.0.0.CR7/arquillian-test-api-1.0.0.CR7.jar:/Users/bstansberry/.m2/repository/org/jboss/arquillian/test/arquillian-test-spi/1.0.0.CR7/arquillian-test-spi-1.0.0.CR7.jar:/Users/bstansberry/.m2/repository/org/jboss/arquillian/core/arquillian-core-spi/1.0.0.CR7/arquillian-core-spi-1.0.0.CR7.jar:/Users/bstansberry/.m2/repository/org/jboss/arquillian/container/arquillian-container-test-api/1.0.0.CR7/arquillian-container-test-api-1.0.0.CR7.jar:/Users/bstansberry/.m2/repository/org/jboss/arquillian/container/arquillian-container-test-spi/1.0.0.CR7/arquillian-container-test-spi-1.0.0.CR7.jar:/Users/bstansberry/.m2/repository/org/jboss/arquillian/container/arquillian-container-spi/1.0.0.CR7/arquillian-container-spi-1.0.0.CR7.jar:/Users/bstansberry/.m2/repository/org/jboss/shrinkwrap/descriptors/shrinkwrap-descriptors-api/1.1.0-beta-1/shrinkwrap-descriptors-api-1.1.0-beta-1.jar:/Users/bstansberry/.m2/repository/org/jboss/arquillian/core/arquillian-core-impl-base/1.0.0.CR7/arquillian-core-impl-base-1.0.0.CR7.jar:/Users/bstansberry/.m2/repository/org/jboss/arquillian/test/arquillian-test-impl-base/1.0.0.CR7/arquillian-test-impl-base-1.0.0.CR7.jar:/Users/bstansberry/.m2/repository/org/jboss/arquillian/container/arquillian-container-impl-base/1.0.0.CR7/arquillian-container-impl-base-1.0.0.CR7.jar:/Users/bstansberry/.m2/repository/org/jboss/arquillian/config/arquillian-config-api/1.0.0.CR7/arquillian-config-api-1.0.0.CR7.jar:/Users/bstansberry/.m2/repository/org/jboss/arquillian/config/arquillian-config-impl-base/1.0.0.CR7/arquillian-config-impl-base-1.0.0.CR7.jar:/Users/bstansberry/.m2/repository/org/jboss/shrinkwrap/descriptors/shrinkwrap-descriptors-spi/1.1.0-beta-1/shrinkwrap-descriptors-spi-1.1.0-beta-1.jar:/Users/bstansberry/.m2/repository/org/jboss/arquillian/container/arquillian-container-test-impl-base/1.0.0.CR7/arquillian-container-test-impl-base-1.0.0.CR7.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/arquillian/container-managed/target/jboss-as-arquillian-container-managed-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/arquillian/common/target/jboss-as-arquillian-common-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/.m2/repository/org/jboss/arquillian/container/arquillian-container-osgi/1.0.0.CR4/arquillian-container-osgi-1.0.0.CR4.jar:/Users/bstansberry/.m2/repository/org/jboss/arquillian/testenricher/arquillian-testenricher-cdi/1.0.0.CR7/arquillian-testenricher-cdi-1.0.0.CR7.jar:/Users/bstansberry/.m2/repository/org/jboss/arquillian/testenricher/arquillian-testenricher-ejb/1.0.0.CR7/arquillian-testenricher-ejb-1.0.0.CR7.jar:/Users/bstansberry/.m2/repository/org/jboss/arquillian/testenricher/arquillian-testenricher-initialcontext/1.0.0.CR7/arquillian-testenricher-initialcontext-1.0.0.CR7.jar:/Users/bstansberry/.m2/repository/org/jboss/arquillian/testenricher/arquillian-testenricher-osgi/1.0.0.CR4/arquillian-testenricher-osgi-1.0.0.CR4.jar:/Users/bstansberry/.m2/repository/org/jboss/arquillian/testenricher/arquillian-testenricher-resource/1.0.0.CR7/arquillian-testenricher-resource-1.0.0.CR7.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/arquillian/testenricher-msc/target/jboss-as-arquillian-testenricher-msc-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/.m2/repository/org/jboss/shrinkwrap/resolver/shrinkwrap-resolver-impl-maven/1.0.0-beta-5/shrinkwrap-resolver-impl-maven-1.0.0-beta-5.jar:/Users/bstansberry/.m2/repository/org/jboss/shrinkwrap/resolver/shrinkwrap-resolver-api/1.0.0-beta-5/shrinkwrap-resolver-api-1.0.0-beta-5.jar:/Users/bstansberry/.m2/repository/org/jboss/shrinkwrap/resolver/shrinkwrap-resolver-api-maven/1.0.0-beta-5/shrinkwrap-resolver-api-maven-1.0.0-beta-5.jar:/Users/bstansberry/.m2/repository/org/sonatype/aether/aether-api/1.8/aether-api-1.8.jar:/Users/bstansberry/.m2/repository/org/sonatype/aether/aether-util/1.8/aether-util-1.8.jar:/Users/bstansberry/.m2/repository/org/sonatype/aether/aether-connector-wagon/1.8/aether-connector-wagon-1.8.jar:/Users/bstansberry/.m2/repository/org/sonatype/aether/aether-spi/1.8/aether-spi-1.8.jar:/Users/bstansberry/.m2/repository/org/codehaus/plexus/plexus-container-default/1.5.5/plexus-container-default-1.5.5.jar:/Users/bstansberry/.m2/repository/org/codehaus/plexus/plexus-classworlds/2.2.2/plexus-classworlds-2.2.2.jar:/Users/bstansberry/.m2/repository/org/apache/xbean/xbean-reflect/3.4/xbean-reflect-3.4.jar:/Users/bstansberry/.m2/repository/com/google/collections/google-collections/1.0/google-collections-1.0.jar:/Users/bstansberry/.m2/repository/org/apache/maven/maven-aether-provider/3.0.1/maven-aether-provider-3.0.1.jar:/Users/bstansberry/.m2/repository/org/apache/maven/maven-model/3.0.1/maven-model-3.0.1.jar:/Users/bstansberry/.m2/repository/org/apache/maven/maven-model-builder/3.0.1/maven-model-builder-3.0.1.jar:/Users/bstansberry/.m2/repository/org/apache/maven/maven-repository-metadata/3.0.1/maven-repository-metadata-3.0.1.jar:/Users/bstansberry/.m2/repository/org/sonatype/aether/aether-impl/1.8/aether-impl-1.8.jar:/Users/bstansberry/.m2/repository/org/codehaus/plexus/plexus-component-annotations/1.5.5/plexus-component-annotations-1.5.5.jar:/Users/bstansberry/.m2/repository/org/codehaus/plexus/plexus-utils/2.0.4/plexus-utils-2.0.4.jar:/Users/bstansberry/.m2/repository/org/apache/maven/maven-settings-builder/3.0.1/maven-settings-builder-3.0.1.jar:/Users/bstansberry/.m2/repository/org/codehaus/plexus/plexus-interpolation/1.14/plexus-interpolation-1.14.jar:/Users/bstansberry/.m2/repository/org/apache/maven/maven-settings/3.0.1/maven-settings-3.0.1.jar:/Users/bstansberry/.m2/repository/org/sonatype/plexus/plexus-sec-dispatcher/1.3/plexus-sec-dispatcher-1.3.jar:/Users/bstansberry/.m2/repository/org/sonatype/plexus/plexus-cipher/1.4/plexus-cipher-1.4.jar:/Users/bstansberry/.m2/repository/org/apache/maven/wagon/wagon-provider-api/1.0-beta-7/wagon-provider-api-1.0-beta-7.jar:/Users/bstansberry/.m2/repository/org/apache/maven/wagon/wagon-file/1.0-beta-7/wagon-file-1.0-beta-7.jar:/Users/bstansberry/.m2/repository/org/apache/maven/wagon/wagon-http-lightweight/1.0-beta-7/wagon-http-lightweight-1.0-beta-7.jar:/Users/bstansberry/.m2/repository/org/apache/maven/wagon/wagon-http-shared/1.0-beta-7/wagon-http-shared-1.0-beta-7.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/arquillian/protocol-jmx/target/jboss-as-arquillian-protocol-jmx-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/.m2/repository/org/jboss/arquillian/protocol/arquillian-protocol-jmx/1.0.0.CR7/arquillian-protocol-jmx-1.0.0.CR7.jar:/Users/bstansberry/.m2/repository/org/jboss/sasl/jboss-sasl/1.0.0.Beta9/jboss-sasl-1.0.0.Beta9.jar:/Users/bstansberry/.m2/repository/org/jboss/arquillian/core/arquillian-core-api/1.0.0.CR7/arquillian-core-api-1.0.0.CR7.jar:/Users/bstansberry/.m2/repository/junit/junit/4.10/junit-4.10.jar:/Users/bstansberry/.m2/repository/org/hamcrest/hamcrest-core/1.1/hamcrest-core-1.1.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/build-config/target/jboss-as-build-config-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/src/test/config/arq:, java.vm.specification.name=Java Virtual Machine Specification, java.vm.specification.version=1.0, sun.cpu.endian=little, sun.os.patch.level=unknown, ts.tr.db=100, surefire.test.class.path=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/target/test-classes:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/target/classes:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/clustering/singleton/target/jboss-as-clustering-singleton-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/clustering/service/target/jboss-as-clustering-service-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/webservices/tests-integration/target/jboss-as-webservices-tests-integration-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/.m2/repository/org/jboss/ws/jbossws-spi/2.0.1.GA/jbossws-spi-2.0.1.GA.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/controller-client/target/jboss-as-controller-client-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/protocol/target/jboss-as-protocol-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/.m2/repository/org/jboss/xnio/xnio-nio/3.0.0.GA/xnio-nio-3.0.0.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/threads/jboss-threads/2.0.0.GA/jboss-threads-2.0.0.GA.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/security/target/jboss-as-security-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/.m2/repository/org/jboss/marshalling/jboss-marshalling-river/1.3.6.GA/jboss-marshalling-river-1.3.6.GA.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/subsystem-test/target/jboss-as-subsystem-test-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/.m2/repository/org/jboss/logging/jboss-logging/3.1.0.CR2/jboss-logging-3.1.0.CR2.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/controller/target/jboss-as-controller-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/server/target/jboss-as-server-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/domain-http/interface/target/jboss-as-domain-http-interface-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/deployment-repository/target/jboss-as-deployment-repository-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/embedded/target/jboss-as-embedded-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/platform-mbean/target/jboss-as-platform-mbean-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/process-controller/target/jboss-as-process-controller-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/.m2/repository/org/jboss/remoting3/jboss-remoting/3.2.0.CR9/jboss-remoting-3.2.0.CR9.jar:/Users/bstansberry/.m2/repository/org/jboss/stdio/jboss-stdio/1.0.1.GA/jboss-stdio-1.0.1.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/jboss-vfs/3.1.0.CR1/jboss-vfs-3.1.0.CR1.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/demos/legacy/target/jboss-as-demos-legacy-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/.m2/repository/org/jboss/spec/javax/ws/rs/jboss-jaxrs-api_1.1_spec/1.0.0.Final/jboss-jaxrs-api_1.1_spec-1.0.0.Final.jar:/Users/bstansberry/.m2/repository/org/hornetq/hornetq-core/2.2.10.Final/hornetq-core-2.2.10.Final.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/jmx/target/jboss-as-jmx-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/.m2/repository/org/jboss/jboss-common-core/2.2.17.GA/jboss-common-core-2.2.17.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/spec/javax/resource/jboss-connector-api_1.6_spec/1.0.0.Final/jboss-connector-api_1.6_spec-1.0.0.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/shrinkwrap/shrinkwrap-api/1.0.0-cr-1/shrinkwrap-api-1.0.0-cr-1.jar:/Users/bstansberry/.m2/repository/org/jboss/shrinkwrap/shrinkwrap-impl-base/1.0.0-cr-1/shrinkwrap-impl-base-1.0.0-cr-1.jar:/Users/bstansberry/.m2/repository/org/jboss/spec/javax/jms/jboss-jms-api_1.1_spec/1.0.0.Final/jboss-jms-api_1.1_spec-1.0.0.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/netty/netty/3.2.6.Final/netty-3.2.6.Final.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/connector/target/jboss-as-connector-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/logging/target/jboss-as-logging-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/messaging/target/jboss-as-messaging-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/network/target/jboss-as-network-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/naming/target/jboss-as-naming-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/osgi/service/target/jboss-as-osgi-service-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/remoting/target/jboss-as-remoting-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/domain-management/target/jboss-as-domain-management-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/.m2/repository/org/jboss/xnio/xnio-api/3.0.0.GA/xnio-api-3.0.0.GA.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/sar/target/jboss-as-sar-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/threads/target/jboss-as-threads-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/transactions/target/jboss-as-transactions-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/web/target/jboss-as-web-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/weld/target/jboss-as-weld-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/.m2/repository/org/jboss/jboss-transaction-spi/7.0.0.CR2/jboss-transaction-spi-7.0.0.CR2.jar:/Users/bstansberry/.m2/repository/org/jboss/spec/javax/xml/ws/jboss-jaxws-api_2.2_spec/2.0.0.Final/jboss-jaxws-api_2.2_spec-2.0.0.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/ws/cxf/jbossws-cxf-client/4.0.0.GA/jbossws-cxf-client-4.0.0.GA.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-rt-frontend-jaxws/2.4.4/cxf-rt-frontend-jaxws-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-rt-frontend-simple/2.4.4/cxf-rt-frontend-simple-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-common-utilities/2.4.4/cxf-common-utilities-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-rt-core/2.4.4/cxf-rt-core-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-rt-ws-addr/2.4.4/cxf-rt-ws-addr-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-rt-ws-policy/2.4.4/cxf-rt-ws-policy-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-rt-ws-rm/2.4.4/cxf-rt-ws-rm-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-rt-transports-http/2.4.4/cxf-rt-transports-http-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-rt-databinding-jaxb/2.4.4/cxf-rt-databinding-jaxb-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-tools-common/2.4.4/cxf-tools-common-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/ws/xmlschema/xmlschema-core/2.0/xmlschema-core-2.0.jar:/Users/bstansberry/.m2/repository/org/apache/neethi/neethi/3.0.1/neethi-3.0.1.jar:/Users/bstansberry/.m2/repository/wsdl4j/wsdl4j/1.6.2/wsdl4j-1.6.2.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/demos/spec/target/jboss-as-demos-spec-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/.m2/repository/org/jboss/spec/javax/ejb/jboss-ejb-api_3.1_spec/1.0.1.Final/jboss-ejb-api_3.1_spec-1.0.1.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/spec/javax/el/jboss-el-api_2.2_spec/1.0.0.Final/jboss-el-api_2.2_spec-1.0.0.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/spec/javax/enterprise/deploy/jboss-jad-api_1.2_spec/1.0.0.Final/jboss-jad-api_1.2_spec-1.0.0.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/spec/javax/faces/jboss-jsf-api_2.1_spec/2.0.0.Beta1/jboss-jsf-api_2.1_spec-2.0.0.Beta1.jar:/Users/bstansberry/.m2/repository/org/jboss/spec/javax/interceptor/jboss-interceptors-api_1.1_spec/1.0.0.Final/jboss-interceptors-api_1.1_spec-1.0.0.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/spec/javax/security/auth/message/jboss-jaspi-api_1.0_spec/1.0.0.Final/jboss-jaspi-api_1.0_spec-1.0.0.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/spec/javax/security/jacc/jboss-jacc-api_1.4_spec/1.0.1.Final/jboss-jacc-api_1.4_spec-1.0.1.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/spec/javax/servlet/jboss-servlet-api_3.0_spec/1.0.0.Final/jboss-servlet-api_3.0_spec-1.0.0.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/spec/javax/servlet/jsp/jboss-jsp-api_2.2_spec/1.0.0.Final/jboss-jsp-api_2.2_spec-1.0.0.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/spec/javax/servlet/jstl/jboss-jstl-api_1.2_spec/1.0.2.Final/jboss-jstl-api_1.2_spec-1.0.2.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/spec/javax/transaction/jboss-transaction-api_1.1_spec/1.0.0.Final/jboss-transaction-api_1.1_spec-1.0.0.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/spec/javax/xml/bind/jboss-jaxb-api_2.2_spec/1.0.3.Final/jboss-jaxb-api_2.2_spec-1.0.3.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/spec/javax/xml/registry/jboss-jaxr-api_1.0_spec/1.0.0.Final/jboss-jaxr-api_1.0_spec-1.0.0.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/spec/javax/xml/rpc/jboss-jaxrpc-api_1.1_spec/1.0.0.Final/jboss-jaxrpc-api_1.1_spec-1.0.0.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/spec/javax/xml/soap/jboss-saaj-api_1.3_spec/1.0.1.Final/jboss-saaj-api_1.3_spec-1.0.1.Final.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/shared/target/jboss-as-testsuite-shared-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/.m2/repository/log4j/log4j/1.2.16/log4j-1.2.16.jar:/Users/bstansberry/.m2/repository/org/jboss/jandex/1.0.3.Final/jandex-1.0.3.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/jboss-dmr/1.1.1.Final/jboss-dmr-1.1.1.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/staxmapper/1.0.0.Final/staxmapper-1.0.0.Final.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/domain-http/error-context/target/jboss-as-domain-http-error-context-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/.m2/repository/org/jboss/as/jbossweb-native/2.0.10.Beta1/jbossweb-native-2.0.10.Beta1.jar:/Users/bstansberry/.m2/repository/org/jboss/invocation/jboss-invocation/1.1.1.Final/jboss-invocation-1.1.1.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/logmanager/jboss-logmanager/1.2.0.GA/jboss-logmanager-1.2.0.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/logmanager/jboss-logmanager-log4j/1.0.0.GA/jboss-logmanager-log4j-1.0.0.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/marshalling/jboss-marshalling/1.3.6.GA/jboss-marshalling-1.3.6.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/modules/jboss-modules/1.1.0.CR8/jboss-modules-1.1.0.CR8.jar:/Users/bstansberry/.m2/repository/org/jboss/msc/jboss-msc/1.0.1.GA/jboss-msc-1.0.1.GA.jar:/Users/bstansberry/.m2/repository/antlr/antlr/2.7.7/antlr-2.7.7.jar:/Users/bstansberry/.m2/repository/asm/asm/3.3.1/asm-3.3.1.jar:/Users/bstansberry/.m2/repository/ch/qos/cal10n/cal10n-api/0.7.3/cal10n-api-0.7.3.jar:/Users/bstansberry/.m2/repository/com/h2database/h2/1.3.161/h2-1.3.161.jar:/Users/bstansberry/.m2/repository/com/google/guava/guava/10.0.1/guava-10.0.1.jar:/Users/bstansberry/.m2/repository/com/sun/faces/jsf-impl/2.1.5-jbossorg-1/jsf-impl-2.1.5-jbossorg-1.jar:/Users/bstansberry/.m2/repository/com/sun/xml/bind/jaxb-impl/2.2/jaxb-impl-2.2.jar:/Users/bstansberry/.m2/repository/com/sun/xml/bind/jaxb-xjc/2.2/jaxb-xjc-2.2.jar:/Users/bstansberry/.m2/repository/com/sun/xml/messaging/saaj/saaj-impl/1.3.16-jbossorg-1/saaj-impl-1.3.16-jbossorg-1.jar:/Users/bstansberry/.m2/repository/commons-beanutils/commons-beanutils/1.8.0/commons-beanutils-1.8.0.jar:/Users/bstansberry/.m2/repository/commons-cli/commons-cli/1.2/commons-cli-1.2.jar:/Users/bstansberry/.m2/repository/commons-codec/commons-codec/1.4/commons-codec-1.4.jar:/Users/bstansberry/.m2/repository/commons-collections/commons-collections/3.2.1/commons-collections-3.2.1.jar:/Users/bstansberry/.m2/repository/commons-io/commons-io/2.1/commons-io-2.1.jar:/Users/bstansberry/.m2/repository/commons-lang/commons-lang/2.4/commons-lang-2.4.jar:/Users/bstansberry/.m2/repository/commons-pool/commons-pool/1.5.6/commons-pool-1.5.6.jar:/Users/bstansberry/.m2/repository/dom4j/dom4j/1.6.1/dom4j-1.6.1.jar:/Users/bstansberry/.m2/repository/gnu-getopt/getopt/1.0.13/getopt-1.0.13.jar:/Users/bstansberry/.m2/repository/org/jacorb/jacorb/2.3.1.jbossorg-1/jacorb-2.3.1.jbossorg-1.jar:/Users/bstansberry/.m2/repository/javax/activation/activation/1.1.1/activation-1.1.1.jar:/Users/bstansberry/.m2/repository/javax/enterprise/cdi-api/1.0-SP4/cdi-api-1.0-SP4.jar:/Users/bstansberry/.m2/repository/javax/faces/jsf-api/1.2_15-jbossorg-2/jsf-api-1.2_15-jbossorg-2.jar:/Users/bstansberry/.m2/repository/javax/faces/jsf-impl/1.2_15-jbossorg-2/jsf-impl-1.2_15-jbossorg-2.jar:/Users/bstansberry/.m2/repository/javax/inject/javax.inject/1/javax.inject-1.jar:/Users/bstansberry/.m2/repository/javax/jws/jsr181-api/1.0-MR1/jsr181-api-1.0-MR1.jar:/Users/bstansberry/.m2/repository/javax/mail/mail/1.4.4/mail-1.4.4.jar:/Users/bstansberry/.m2/repository/javax/validation/validation-api/1.0.0.GA/validation-api-1.0.0.GA.jar:/Users/bstansberry/.m2/repository/jaxen/jaxen/1.1.3/jaxen-1.1.3.jar:/Users/bstansberry/.m2/repository/jboss/jaxbintros/jboss-jaxb-intros/1.0.2.GA/jboss-jaxb-intros-1.0.2.GA.jar:/Users/bstansberry/.m2/repository/jline/jline/0.9.94/jline-0.9.94.jar:/Users/bstansberry/.m2/repository/joda-time/joda-time/1.6.2/joda-time-1.6.2.jar:/Users/bstansberry/.m2/repository/net/sourceforge/cssparser/cssparser/0.9.5/cssparser-0.9.5.jar:/Users/bstansberry/.m2/repository/net/sourceforge/htmlunit/htmlunit/2.8/htmlunit-2.8.jar:/Users/bstansberry/.m2/repository/net/sourceforge/htmlunit/htmlunit-core-js/2.8/htmlunit-core-js-2.8.jar:/Users/bstansberry/.m2/repository/net/sourceforge/nekohtml/nekohtml/1.9.14/nekohtml-1.9.14.jar:/Users/bstansberry/.m2/repository/org/apache/ant/ant/1.8.2/ant-1.8.2.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-api/2.4.4/cxf-api-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-rt-bindings-coloc/2.4.4/cxf-rt-bindings-coloc-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-rt-bindings-http/2.4.4/cxf-rt-bindings-http-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-rt-bindings-object/2.4.4/cxf-rt-bindings-object-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-rt-bindings-soap/2.4.4/cxf-rt-bindings-soap-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-rt-bindings-xml/2.4.4/cxf-rt-bindings-xml-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-rt-databinding-aegis/2.4.4/cxf-rt-databinding-aegis-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-rt-management/2.4.4/cxf-rt-management-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-rt-transports-common/2.4.4/cxf-rt-transports-common-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-rt-transports-local/2.4.4/cxf-rt-transports-local-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-rt-transports-jms/2.4.4/cxf-rt-transports-jms-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-rt-ws-security/2.4.4/cxf-rt-ws-security-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-tools-java2ws/2.4.4/cxf-tools-java2ws-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-tools-validator/2.4.4/cxf-tools-validator-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-tools-wsdlto-core/2.4.4/cxf-tools-wsdlto-core-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-tools-wsdlto-databinding-jaxb/2.4.4/cxf-tools-wsdlto-databinding-jaxb-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/cxf-tools-wsdlto-frontend-jaxws/2.4.4/cxf-tools-wsdlto-frontend-jaxws-2.4.4.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/xjcplugins/cxf-xjc-boolean/2.4.0/cxf-xjc-boolean-2.4.0.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/xjcplugins/cxf-xjc-bug671/2.4.0/cxf-xjc-bug671-2.4.0.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/xjcplugins/cxf-xjc-dv/2.4.0/cxf-xjc-dv-2.4.0.jar:/Users/bstansberry/.m2/repository/org/apache/cxf/xjcplugins/cxf-xjc-ts/2.4.0/cxf-xjc-ts-2.4.0.jar:/Users/bstansberry/.m2/repository/org/apache/felix/org.apache.felix.configadmin/1.2.8/org.apache.felix.configadmin-1.2.8.jar:/Users/bstansberry/.m2/repository/org/apache/httpcomponents/httpclient/4.1.2/httpclient-4.1.2.jar:/Users/bstansberry/.m2/repository/org/apache/httpcomponents/httpcore/4.1.2/httpcore-4.1.2.jar:/Users/bstansberry/.m2/repository/org/apache/httpcomponents/httpmime/4.1.2/httpmime-4.1.2.jar:/Users/bstansberry/.m2/repository/org/apache/james/apache-mime4j/0.6/apache-mime4j-0.6.jar:/Users/bstansberry/.m2/repository/org/apache/santuario/xmlsec/1.4.5/xmlsec-1.4.5.jar:/Users/bstansberry/.m2/repository/org/apache/ws/scout/scout/1.1.1/scout-1.1.1.jar:/Users/bstansberry/.m2/repository/org/apache/velocity/velocity/1.6.3/velocity-1.6.3.jar:/Users/bstansberry/.m2/repository/org/apache/ws/security/wss4j/1.6.3/wss4j-1.6.3.jar:/Users/bstansberry/.m2/repository/org/codehaus/jackson/jackson-core-asl/1.9.2/jackson-core-asl-1.9.2.jar:/Users/bstansberry/.m2/repository/org/codehaus/jackson/jackson-mapper-asl/1.9.2/jackson-mapper-asl-1.9.2.jar:/Users/bstansberry/.m2/repository/org/codehaus/jackson/jackson-jaxrs/1.9.2/jackson-jaxrs-1.9.2.jar:/Users/bstansberry/.m2/repository/org/codehaus/jackson/jackson-xc/1.9.2/jackson-xc-1.9.2.jar:/Users/bstansberry/.m2/repository/org/codehaus/jettison/jettison/1.3.1/jettison-1.3.1.jar:/Users/bstansberry/.m2/repository/stax/stax-api/1.0.1/stax-api-1.0.1.jar:/Users/bstansberry/.m2/repository/org/codehaus/woodstox/woodstox-core-asl/4.1.1/woodstox-core-asl-4.1.1.jar:/Users/bstansberry/.m2/repository/org/codehaus/woodstox/stax2-api/3.1.1/stax2-api-3.1.1.jar:/Users/bstansberry/.m2/repository/org/hibernate/hibernate-core/4.0.1.Final/hibernate-core-4.0.1.Final.jar:/Users/bstansberry/.m2/repository/org/hibernate/common/hibernate-commons-annotations/4.0.1.Final/hibernate-commons-annotations-4.0.1.Final.jar:/Users/bstansberry/.m2/repository/org/hibernate/hibernate-entitymanager/4.0.1.Final/hibernate-entitymanager-4.0.1.Final.jar:/Users/bstansberry/.m2/repository/org/hibernate/hibernate-infinispan/4.0.1.Final/hibernate-infinispan-4.0.1.Final.jar:/Users/bstansberry/.m2/repository/org/hibernate/hibernate-validator/4.2.0.Final/hibernate-validator-4.2.0.Final.jar:/Users/bstansberry/.m2/repository/org/hibernate/hibernate-envers/4.0.1.Final/hibernate-envers-4.0.1.Final.jar:/Users/bstansberry/.m2/repository/org/hibernate/javax/persistence/hibernate-jpa-2.0-api/1.0.1.Final/hibernate-jpa-2.0-api-1.0.1.Final.jar:/Users/bstansberry/.m2/repository/org/hornetq/hornetq-jms/2.2.10.Final/hornetq-jms-2.2.10.Final.jar:/Users/bstansberry/.m2/repository/org/hornetq/hornetq-ra/2.2.10.Final/hornetq-ra-2.2.10.Final.jar:/Users/bstansberry/.m2/repository/org/infinispan/infinispan-cachestore-jdbc/5.1.0.FINAL/infinispan-cachestore-jdbc-5.1.0.FINAL.jar:/Users/bstansberry/.m2/repository/org/rhq/helpers/rhq-pluginAnnotations/3.0.4/rhq-pluginAnnotations-3.0.4.jar:/Users/bstansberry/.m2/repository/org/infinispan/infinispan-cachestore-remote/5.1.0.FINAL/infinispan-cachestore-remote-5.1.0.FINAL.jar:/Users/bstansberry/.m2/repository/org/infinispan/infinispan-client-hotrod/5.1.0.FINAL/infinispan-client-hotrod-5.1.0.FINAL.jar:/Users/bstansberry/.m2/repository/org/infinispan/infinispan-core/5.1.0.FINAL/infinispan-core-5.1.0.FINAL.jar:/Users/bstansberry/.m2/repository/org/javassist/javassist/3.15.0-GA/javassist-3.15.0-GA.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/appclient/target/jboss-as-appclient-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/cli/target/jboss-as-cli-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/clustering/api/target/jboss-as-clustering-api-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/clustering/common/target/jboss-as-clustering-common-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/clustering/ejb3-infinispan/target/jboss-as-clustering-ejb3-infinispan-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/clustering/impl/target/jboss-as-clustering-impl-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/clustering/infinispan/target/jboss-as-clustering-infinispan-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/clustering/jgroups/target/jboss-as-clustering-jgroups-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/clustering/registry/target/jboss-as-clustering-registry-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/clustering/web-infinispan/target/jboss-as-clustering-web-infinispan-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/clustering/web-spi/target/jboss-as-clustering-web-spi-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/cmp/target/jboss-as-cmp-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/configadmin/target/jboss-as-configadmin-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/.m2/repository/org/jboss/as/jboss-as-console/1.0.0.RC1/jboss-as-console-1.0.0.RC1-resources.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/deployment-scanner/target/jboss-as-deployment-scanner-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/ee/target/jboss-as-ee-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/ee-deployment/target/jboss-as-ee-deployment-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/ejb3/target/jboss-as-ejb3-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/.m2/repository/org/jboss/jboss-iiop-client/1.0.0.Beta2/jboss-iiop-client-1.0.0.Beta2.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/jaxr/target/jboss-as-jaxr-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/jaxrs/target/jboss-as-jaxrs-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/jdr/jboss-as-jdr/target/jboss-as-jdr-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/.m2/repository/org/jboss/logging/jboss-logging-processor/1.0.0.CR5/jboss-logging-processor-1.0.0.CR5.jar:/Users/bstansberry/.m2/repository/org/jboss/logging/jboss-logging-generator/1.0.0.CR5/jboss-logging-generator-1.0.0.CR5.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/jsr77/target/jboss-as-jsr77-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/host-controller/target/jboss-as-host-controller-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/management-client-content/target/jboss-as-management-client-content-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/jacorb/target/jboss-as-jacorb-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/mail/target/jboss-as-mail-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/modcluster/target/jboss-as-modcluster-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/pojo/target/jboss-as-pojo-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/osgi/configadmin/target/jboss-as-osgi-configadmin-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/jdr/jboss-as-sos/target/jboss-as-sos-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/webservices/server-integration/target/jboss-as-webservices-server-integration-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/xts/target/jboss-as-xts-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/.m2/repository/org/jboss/classfilewriter/jboss-classfilewriter/1.0.0.Final/jboss-classfilewriter-1.0.0.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/com/sun/httpserver/httpserver/1.0.0.Beta3/httpserver-1.0.0.Beta3.jar:/Users/bstansberry/.m2/repository/org/jboss/jboss-ejb-client/1.0.0.Beta12/jboss-ejb-client-1.0.0.Beta12.jar:/Users/bstansberry/.m2/repository/org/jboss/ejb3/jboss-ejb3-ext-api/2.0.0-beta-3/jboss-ejb3-ext-api-2.0.0-beta-3.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/jpa/core/target/jboss-as-jpa-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/jpa/hibernate3/target/jboss-as-jpa-hibernate3-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/.m2/repository/org/hibernate/hibernate-commons-annotations/3.2.0.Final/hibernate-commons-annotations-3.2.0.Final.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/jpa/hibernate4/target/jboss-as-jpa-hibernate4-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/jpa/hibernate-infinispan/target/jboss-as-jpa-hibernate-infinispan-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/jpa/openjpa/target/jboss-as-jpa-openjpa-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/.m2/repository/org/apache/openjpa/openjpa-kernel/2.1.1/openjpa-kernel-2.1.1.jar:/Users/bstansberry/.m2/repository/org/apache/openjpa/openjpa-lib/2.1.1/openjpa-lib-2.1.1.jar:/Users/bstansberry/.m2/repository/org/apache/openjpa/openjpa-persistence/2.1.1/openjpa-persistence-2.1.1.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/jpa/spi/target/jboss-as-jpa-spi-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/jpa/util/target/jboss-as-jpa-util-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/.m2/repository/org/jboss/interceptor/jboss-interceptor-core/2.0.0.Alpha3/jboss-interceptor-core-2.0.0.Alpha3.jar:/Users/bstansberry/.m2/repository/org/jboss/interceptor/jboss-interceptor-spi/2.0.0.Alpha3/jboss-interceptor-spi-2.0.0.Alpha3.jar:/Users/bstansberry/.m2/repository/org/jboss/ironjacamar/ironjacamar-core-api/1.0.7.Final/ironjacamar-core-api-1.0.7.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/ironjacamar/ironjacamar-core-impl/1.0.7.Final/ironjacamar-core-impl-1.0.7.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/ironjacamar/ironjacamar-common-api/1.0.7.Final/ironjacamar-common-api-1.0.7.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/ironjacamar/ironjacamar-common-spi/1.0.7.Final/ironjacamar-common-spi-1.0.7.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/ironjacamar/ironjacamar-common-impl/1.0.7.Final/ironjacamar-common-impl-1.0.7.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/ironjacamar/ironjacamar-deployers-common/1.0.7.Final/ironjacamar-deployers-common-1.0.7.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/ironjacamar/ironjacamar-jdbc/1.0.7.Final/ironjacamar-jdbc-1.0.7.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/ironjacamar/ironjacamar-spec-api/1.0.7.Final/ironjacamar-spec-api-1.0.7.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/ironjacamar/ironjacamar-validator/1.0.7.Final/ironjacamar-validator-1.0.7.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/jbossts/jbossjts/4.16.1.Final/jbossjts-4.16.1.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/jbossts/jbossjts-integration/4.16.1.Final/jbossjts-integration-4.16.1.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/jbossts/jbosstxbridge/4.16.1.Final/jbosstxbridge-4.16.1.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/jbossts/jbossxts/4.16.1.Final/jbossxts-4.16.1.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/jbossts/jbossxts-api/4.16.1.Final/jbossxts-api-4.16.1.Final.jar:/Users/bstansberry/.m2/repository/jacorb/jacorb/2.3.1jboss.patch01-brew/jacorb-2.3.1jboss.patch01-brew.jar:/Users/bstansberry/.m2/repository/org/jboss/jsfunit/jboss-jsfunit-core/2.0.0.Beta1/jboss-jsfunit-core-2.0.0.Beta1.jar:/Users/bstansberry/.m2/repository/org/jboss/metadata/jboss-metadata-appclient/7.0.0.Beta33/jboss-metadata-appclient-7.0.0.Beta33.jar:/Users/bstansberry/.m2/repository/org/jboss/metadata/jboss-metadata-common/7.0.0.Beta33/jboss-metadata-common-7.0.0.Beta33.jar:/Users/bstansberry/.m2/repository/org/jboss/metadata/jboss-metadata-ear/7.0.0.Beta33/jboss-metadata-ear-7.0.0.Beta33.jar:/Users/bstansberry/.m2/repository/org/jboss/metadata/jboss-metadata-ejb/7.0.0.Beta33/jboss-metadata-ejb-7.0.0.Beta33.jar:/Users/bstansberry/.m2/repository/org/jboss/metadata/jboss-metadata-web/7.0.0.Beta33/jboss-metadata-web-7.0.0.Beta33.jar:/Users/bstansberry/.m2/repository/org/jboss/mod_cluster/mod_cluster-container-spi/1.2.0.Beta4/mod_cluster-container-spi-1.2.0.Beta4.jar:/Users/bstansberry/.m2/repository/org/jboss/mod_cluster/mod_cluster-core/1.2.0.Beta4/mod_cluster-core-1.2.0.Beta4.jar:/Users/bstansberry/.m2/repository/net/jcip/jcip-annotations/1.0/jcip-annotations-1.0.jar:/Users/bstansberry/.m2/repository/org/jboss/mod_cluster/mod_cluster-container-catalina/1.2.0.Beta4/mod_cluster-container-catalina-1.2.0.Beta4.jar:/Users/bstansberry/.m2/repository/org/jboss/mod_cluster/mod_cluster-container-jbossweb/1.2.0.Beta4/mod_cluster-container-jbossweb-1.2.0.Beta4.jar:/Users/bstansberry/.m2/repository/org/jboss/osgi/deployment/jbosgi-deployment/1.0.10/jbosgi-deployment-1.0.10.jar:/Users/bstansberry/.m2/repository/org/jboss/osgi/framework/jbosgi-framework-core/1.1.0/jbosgi-framework-core-1.1.0.jar:/Users/bstansberry/.m2/repository/org/jboss/osgi/resolver/jbosgi-resolver-api/1.0.10/jbosgi-resolver-api-1.0.10.jar:/Users/bstansberry/.m2/repository/org/jboss/osgi/resolver/jbosgi-resolver-felix/1.0.10/jbosgi-resolver-felix-1.0.10.jar:/Users/bstansberry/.m2/repository/org/jboss/osgi/resolver/jbosgi-resolver-metadata/1.0.10/jbosgi-resolver-metadata-1.0.10.jar:/Users/bstansberry/.m2/repository/org/jboss/osgi/resolver/jbosgi-resolver-spi/1.0.10/jbosgi-resolver-spi-1.0.10.jar:/Users/bstansberry/.m2/repository/org/jboss/osgi/spi/jbosgi-spi/1.0.28/jbosgi-spi-1.0.28.jar:/Users/bstansberry/.m2/repository/org/jboss/osgi/vfs/jbosgi-vfs/1.0.6/jbosgi-vfs-1.0.6.jar:/Users/bstansberry/.m2/repository/org/jboss/osgi/vfs/jbosgi-vfs30/1.0.6/jbosgi-vfs30-1.0.6.jar:/Users/bstansberry/.m2/repository/org/jboss/remoting3/remoting-jmx/1.0.0.Beta1/remoting-jmx-1.0.0.Beta1.jar:/Users/bstansberry/.m2/repository/org/jboss/jboss-remote-naming/1.0.0.Beta3/jboss-remote-naming-1.0.0.Beta3.jar:/Users/bstansberry/.m2/repository/org/jboss/resteasy/resteasy-jaxrs/2.3.1.GA/resteasy-jaxrs-2.3.1.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/resteasy/async-http-servlet-3.0/2.3.1.GA/async-http-servlet-3.0-2.3.1.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/resteasy/resteasy-atom-provider/2.3.1.GA/resteasy-atom-provider-2.3.1.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/resteasy/resteasy-cdi/2.3.1.GA/resteasy-cdi-2.3.1.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/resteasy/resteasy-jackson-provider/2.3.1.GA/resteasy-jackson-provider-2.3.1.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/resteasy/resteasy-jaxb-provider/2.3.1.GA/resteasy-jaxb-provider-2.3.1.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/resteasy/resteasy-jettison-provider/2.3.1.GA/resteasy-jettison-provider-2.3.1.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/resteasy/resteasy-jsapi/2.3.1.GA/resteasy-jsapi-2.3.1.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/resteasy/resteasy-multipart-provider/2.3.1.GA/resteasy-multipart-provider-2.3.1.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/resteasy/resteasy-yaml-provider/2.3.1.GA/resteasy-yaml-provider-2.3.1.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/seam/integration/jboss-seam-int-jbossas/6.0.0.GA/jboss-seam-int-jbossas-6.0.0.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/security/jboss-negotiation/2.2.0.Beta3/jboss-negotiation-2.2.0.Beta3.jar:/Users/bstansberry/.m2/repository/org/jboss/security/jbossxacml/2.0.6.Final/jbossxacml-2.0.6.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/shrinkwrap/shrinkwrap-spi/1.0.0-cr-1/shrinkwrap-spi-1.0.0-cr-1.jar:/Users/bstansberry/.m2/repository/org/jboss/slf4j/slf4j-jboss-logmanager/1.0.0.GA/slf4j-jboss-logmanager-1.0.0.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/spec/javax/management/j2ee/jboss-j2eemgmt-api_1.1_spec/1.0.0.Final/jboss-j2eemgmt-api_1.1_spec-1.0.0.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/spec/javax/rmi/jboss-rmi-api_1.0_spec/1.0.4.Final/jboss-rmi-api_1.0_spec-1.0.4.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/web/jasper-jdt/7.0.3.Final/jasper-jdt-7.0.3.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/web/jbossweb/7.0.9.Final/jbossweb-7.0.9.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/weld/weld-api/1.1.Final/weld-api-1.1.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/weld/weld-core/1.1.5.Final/weld-core-1.1.5.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/weld/weld-spi/1.1.Final/weld-spi-1.1.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/ws/jbossws-api/1.0.0.GA/jbossws-api-1.0.0.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/ws/jbossws-common/2.0.0.GA/jbossws-common-2.0.0.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/ws/jbossws-common-tools/1.0.0.GA/jbossws-common-tools-1.0.0.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/ws/projects/jaxws-jboss-httpserver-httpspi/1.0.0.GA/jaxws-jboss-httpserver-httpspi-1.0.0.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/ws/cxf/jbossws-cxf-factories/4.0.0.GA/jbossws-cxf-factories-4.0.0.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/ws/cxf/jbossws-cxf-resources/4.0.0.GA/jbossws-cxf-resources-4.0.0.GA-jboss700.jar:/Users/bstansberry/.m2/repository/org/jboss/ws/cxf/jbossws-cxf-server/4.0.0.GA/jbossws-cxf-server-4.0.0.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/ws/cxf/jbossws-cxf-transports-httpserver/4.0.0.GA/jbossws-cxf-transports-httpserver-4.0.0.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/ws/native/jbossws-native-core/4.0.0.GA/jbossws-native-core-4.0.0.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/ws/native/jbossws-native-factories/4.0.0.GA/jbossws-native-factories-4.0.0.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/ws/native/jbossws-native-services/4.0.0.GA/jbossws-native-services-4.0.0.GA.jar:/Users/bstansberry/.m2/repository/org/jboss/jbossxb/2.0.3.GA/jbossxb-2.0.3.GA.jar:/Users/bstansberry/.m2/repository/org/jdom/jdom/1.1.2/jdom-1.1.2.jar:/Users/bstansberry/.m2/repository/org/jgroups/jgroups/3.0.3.Final/jgroups-3.0.3.Final.jar:/Users/bstansberry/.m2/repository/org/osgi/org.osgi.core/4.2.0/org.osgi.core-4.2.0.jar:/Users/bstansberry/.m2/repository/org/osgi/org.osgi.compendium/4.2.0/org.osgi.compendium-4.2.0.jar:/Users/bstansberry/.m2/repository/org/picketlink/picketlink-fed/2.0.1.final/picketlink-fed-2.0.1.final.jar:/Users/bstansberry/.m2/repository/org/picketlink/picketlink-bindings/2.0.1.final/picketlink-bindings-2.0.1.final.jar:/Users/bstansberry/.m2/repository/org/picketlink/picketlink-bindings-jboss/2.0.1.final/picketlink-bindings-jboss-2.0.1.final.jar:/Users/bstansberry/.m2/repository/org/picketlink/picketlink-trust-jbossws/2.0.1.final/picketlink-trust-jbossws-2.0.1.final.jar:/Users/bstansberry/.m2/repository/org/projectodd/stilts/stilts-stomplet-server-bundle/0.1.20/stilts-stomplet-server-bundle-0.1.20.jar:/Users/bstansberry/.m2/repository/org/scannotation/scannotation/1.0.2/scannotation-1.0.2.jar:/Users/bstansberry/.m2/repository/org/picketbox/picketbox/4.0.6.final/picketbox-4.0.6.final.jar:/Users/bstansberry/.m2/repository/org/picketbox/picketbox-commons/1.0.0.final/picketbox-commons-1.0.0.final.jar:/Users/bstansberry/.m2/repository/org/picketbox/picketbox-infinispan/4.0.6.final/picketbox-infinispan-4.0.6.final.jar:/Users/bstansberry/.m2/repository/org/python/jython-standalone/2.5.2/jython-standalone-2.5.2.jar:/Users/bstansberry/.m2/repository/org/slf4j/jcl-over-slf4j/1.6.1/jcl-over-slf4j-1.6.1.jar:/Users/bstansberry/.m2/repository/org/slf4j/slf4j-api/1.6.1/slf4j-api-1.6.1.jar:/Users/bstansberry/.m2/repository/org/slf4j/slf4j-ext/1.6.1/slf4j-ext-1.6.1.jar:/Users/bstansberry/.m2/repository/org/w3c/css/sac/1.3/sac-1.3.jar:/Users/bstansberry/.m2/repository/org/yaml/snakeyaml/1.8/snakeyaml-1.8.jar:/Users/bstansberry/.m2/repository/xalan/serializer/2.7.1.jbossorg-1/serializer-2.7.1.jbossorg-1.jar:/Users/bstansberry/.m2/repository/xalan/xalan/2.7.1.jbossorg-1/xalan-2.7.1.jbossorg-1.jar:/Users/bstansberry/.m2/repository/xerces/xercesImpl/2.9.1-jbossas-1/xercesImpl-2.9.1-jbossas-1.jar:/Users/bstansberry/.m2/repository/xml-resolver/xml-resolver/1.2/xml-resolver-1.2.jar:/Users/bstansberry/.m2/repository/xom/xom/1.2.5/xom-1.2.5.jar:/Users/bstansberry/.m2/repository/org/projectodd/stilts/stilts-stomp-client/0.1.20/stilts-stomp-client-0.1.20.jar:/Users/bstansberry/.m2/repository/org/projectodd/stilts/stilts-stomp-api/0.1.20/stilts-stomp-api-0.1.20.jar:/Users/bstansberry/.m2/repository/org/projectodd/stilts/stilts-stomp-common/0.1.20/stilts-stomp-common-0.1.20.jar:/Users/bstansberry/.m2/repository/org/projectodd/stilts/stilts-stomplet-server-core/0.1.20/stilts-stomplet-server-core-0.1.20.jar:/Users/bstansberry/.m2/repository/org/projectodd/stilts/stilts-conduit-server-core/0.1.20/stilts-conduit-server-core-0.1.20.jar:/Users/bstansberry/.m2/repository/org/projectodd/stilts/stilts-stomp-server-core/0.1.20/stilts-stomp-server-core-0.1.20.jar:/Users/bstansberry/.m2/repository/org/projectodd/stilts/stilts-conduit-server-spi/0.1.20/stilts-conduit-server-spi-0.1.20.jar:/Users/bstansberry/.m2/repository/org/projectodd/stilts/stilts-stomplet-api/0.1.20/stilts-stomplet-api-0.1.20.jar:/Users/bstansberry/.m2/repository/org/projectodd/stilts/stilts-stomp-server-spi/0.1.20/stilts-stomp-server-spi-0.1.20.jar:/Users/bstansberry/.m2/repository/org/jboss/spec/javax/annotation/jboss-annotations-api_1.1_spec/1.0.0.Final/jboss-annotations-api_1.1_spec-1.0.0.Final.jar:/Users/bstansberry/.m2/repository/org/jboss/arquillian/junit/arquillian-junit-container/1.0.0.CR7/arquillian-junit-container-1.0.0.CR7.jar:/Users/bstansberry/.m2/repository/org/jboss/arquillian/junit/arquillian-junit-core/1.0.0.CR7/arquillian-junit-core-1.0.0.CR7.jar:/Users/bstansberry/.m2/repository/org/jboss/arquillian/test/arquillian-test-api/1.0.0.CR7/arquillian-test-api-1.0.0.CR7.jar:/Users/bstansberry/.m2/repository/org/jboss/arquillian/test/arquillian-test-spi/1.0.0.CR7/arquillian-test-spi-1.0.0.CR7.jar:/Users/bstansberry/.m2/repository/org/jboss/arquillian/core/arquillian-core-spi/1.0.0.CR7/arquillian-core-spi-1.0.0.CR7.jar:/Users/bstansberry/.m2/repository/org/jboss/arquillian/container/arquillian-container-test-api/1.0.0.CR7/arquillian-container-test-api-1.0.0.CR7.jar:/Users/bstansberry/.m2/repository/org/jboss/arquillian/container/arquillian-container-test-spi/1.0.0.CR7/arquillian-container-test-spi-1.0.0.CR7.jar:/Users/bstansberry/.m2/repository/org/jboss/arquillian/container/arquillian-container-spi/1.0.0.CR7/arquillian-container-spi-1.0.0.CR7.jar:/Users/bstansberry/.m2/repository/org/jboss/shrinkwrap/descriptors/shrinkwrap-descriptors-api/1.1.0-beta-1/shrinkwrap-descriptors-api-1.1.0-beta-1.jar:/Users/bstansberry/.m2/repository/org/jboss/arquillian/core/arquillian-core-impl-base/1.0.0.CR7/arquillian-core-impl-base-1.0.0.CR7.jar:/Users/bstansberry/.m2/repository/org/jboss/arquillian/test/arquillian-test-impl-base/1.0.0.CR7/arquillian-test-impl-base-1.0.0.CR7.jar:/Users/bstansberry/.m2/repository/org/jboss/arquillian/container/arquillian-container-impl-base/1.0.0.CR7/arquillian-container-impl-base-1.0.0.CR7.jar:/Users/bstansberry/.m2/repository/org/jboss/arquillian/config/arquillian-config-api/1.0.0.CR7/arquillian-config-api-1.0.0.CR7.jar:/Users/bstansberry/.m2/repository/org/jboss/arquillian/config/arquillian-config-impl-base/1.0.0.CR7/arquillian-config-impl-base-1.0.0.CR7.jar:/Users/bstansberry/.m2/repository/org/jboss/shrinkwrap/descriptors/shrinkwrap-descriptors-spi/1.1.0-beta-1/shrinkwrap-descriptors-spi-1.1.0-beta-1.jar:/Users/bstansberry/.m2/repository/org/jboss/arquillian/container/arquillian-container-test-impl-base/1.0.0.CR7/arquillian-container-test-impl-base-1.0.0.CR7.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/arquillian/container-managed/target/jboss-as-arquillian-container-managed-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/arquillian/common/target/jboss-as-arquillian-common-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/.m2/repository/org/jboss/arquillian/container/arquillian-container-osgi/1.0.0.CR4/arquillian-container-osgi-1.0.0.CR4.jar:/Users/bstansberry/.m2/repository/org/jboss/arquillian/testenricher/arquillian-testenricher-cdi/1.0.0.CR7/arquillian-testenricher-cdi-1.0.0.CR7.jar:/Users/bstansberry/.m2/repository/org/jboss/arquillian/testenricher/arquillian-testenricher-ejb/1.0.0.CR7/arquillian-testenricher-ejb-1.0.0.CR7.jar:/Users/bstansberry/.m2/repository/org/jboss/arquillian/testenricher/arquillian-testenricher-initialcontext/1.0.0.CR7/arquillian-testenricher-initialcontext-1.0.0.CR7.jar:/Users/bstansberry/.m2/repository/org/jboss/arquillian/testenricher/arquillian-testenricher-osgi/1.0.0.CR4/arquillian-testenricher-osgi-1.0.0.CR4.jar:/Users/bstansberry/.m2/repository/org/jboss/arquillian/testenricher/arquillian-testenricher-resource/1.0.0.CR7/arquillian-testenricher-resource-1.0.0.CR7.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/arquillian/testenricher-msc/target/jboss-as-arquillian-testenricher-msc-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/.m2/repository/org/jboss/shrinkwrap/resolver/shrinkwrap-resolver-impl-maven/1.0.0-beta-5/shrinkwrap-resolver-impl-maven-1.0.0-beta-5.jar:/Users/bstansberry/.m2/repository/org/jboss/shrinkwrap/resolver/shrinkwrap-resolver-api/1.0.0-beta-5/shrinkwrap-resolver-api-1.0.0-beta-5.jar:/Users/bstansberry/.m2/repository/org/jboss/shrinkwrap/resolver/shrinkwrap-resolver-api-maven/1.0.0-beta-5/shrinkwrap-resolver-api-maven-1.0.0-beta-5.jar:/Users/bstansberry/.m2/repository/org/sonatype/aether/aether-api/1.8/aether-api-1.8.jar:/Users/bstansberry/.m2/repository/org/sonatype/aether/aether-util/1.8/aether-util-1.8.jar:/Users/bstansberry/.m2/repository/org/sonatype/aether/aether-connector-wagon/1.8/aether-connector-wagon-1.8.jar:/Users/bstansberry/.m2/repository/org/sonatype/aether/aether-spi/1.8/aether-spi-1.8.jar:/Users/bstansberry/.m2/repository/org/codehaus/plexus/plexus-container-default/1.5.5/plexus-container-default-1.5.5.jar:/Users/bstansberry/.m2/repository/org/codehaus/plexus/plexus-classworlds/2.2.2/plexus-classworlds-2.2.2.jar:/Users/bstansberry/.m2/repository/org/apache/xbean/xbean-reflect/3.4/xbean-reflect-3.4.jar:/Users/bstansberry/.m2/repository/com/google/collections/google-collections/1.0/google-collections-1.0.jar:/Users/bstansberry/.m2/repository/org/apache/maven/maven-aether-provider/3.0.1/maven-aether-provider-3.0.1.jar:/Users/bstansberry/.m2/repository/org/apache/maven/maven-model/3.0.1/maven-model-3.0.1.jar:/Users/bstansberry/.m2/repository/org/apache/maven/maven-model-builder/3.0.1/maven-model-builder-3.0.1.jar:/Users/bstansberry/.m2/repository/org/apache/maven/maven-repository-metadata/3.0.1/maven-repository-metadata-3.0.1.jar:/Users/bstansberry/.m2/repository/org/sonatype/aether/aether-impl/1.8/aether-impl-1.8.jar:/Users/bstansberry/.m2/repository/org/codehaus/plexus/plexus-component-annotations/1.5.5/plexus-component-annotations-1.5.5.jar:/Users/bstansberry/.m2/repository/org/codehaus/plexus/plexus-utils/2.0.4/plexus-utils-2.0.4.jar:/Users/bstansberry/.m2/repository/org/apache/maven/maven-settings-builder/3.0.1/maven-settings-builder-3.0.1.jar:/Users/bstansberry/.m2/repository/org/codehaus/plexus/plexus-interpolation/1.14/plexus-interpolation-1.14.jar:/Users/bstansberry/.m2/repository/org/apache/maven/maven-settings/3.0.1/maven-settings-3.0.1.jar:/Users/bstansberry/.m2/repository/org/sonatype/plexus/plexus-sec-dispatcher/1.3/plexus-sec-dispatcher-1.3.jar:/Users/bstansberry/.m2/repository/org/sonatype/plexus/plexus-cipher/1.4/plexus-cipher-1.4.jar:/Users/bstansberry/.m2/repository/org/apache/maven/wagon/wagon-provider-api/1.0-beta-7/wagon-provider-api-1.0-beta-7.jar:/Users/bstansberry/.m2/repository/org/apache/maven/wagon/wagon-file/1.0-beta-7/wagon-file-1.0-beta-7.jar:/Users/bstansberry/.m2/repository/org/apache/maven/wagon/wagon-http-lightweight/1.0-beta-7/wagon-http-lightweight-1.0-beta-7.jar:/Users/bstansberry/.m2/repository/org/apache/maven/wagon/wagon-http-shared/1.0-beta-7/wagon-http-shared-1.0-beta-7.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/arquillian/protocol-jmx/target/jboss-as-arquillian-protocol-jmx-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/.m2/repository/org/jboss/arquillian/protocol/arquillian-protocol-jmx/1.0.0.CR7/arquillian-protocol-jmx-1.0.0.CR7.jar:/Users/bstansberry/.m2/repository/org/jboss/sasl/jboss-sasl/1.0.0.Beta9/jboss-sasl-1.0.0.Beta9.jar:/Users/bstansberry/.m2/repository/org/jboss/arquillian/core/arquillian-core-api/1.0.0.CR7/arquillian-core-api-1.0.0.CR7.jar:/Users/bstansberry/.m2/repository/junit/junit/4.10/junit-4.10.jar:/Users/bstansberry/.m2/repository/org/hamcrest/hamcrest-core/1.1/hamcrest-core-1.1.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/build-config/target/jboss-as-build-config-7.1.0.Final-SNAPSHOT.jar:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/src/test/config/arq:, java.net.preferIPv6Addresses=false, java.io.tmpdir=/var/folders/IO/IOJd+-FEG0uH-mrEv+tGRk+++TI/-Tmp-/, java.vendor.url.bug=http://bugreport.apple.com/, mrj.build=10M3527, os.arch=x86_64, java.awt.graphicsenv=apple.awt.CGraphicsEnvironment, udpGroup=230.0.0.4, java.ext.dirs=/Library/Java/Extensions:/System/Library/Java/Extensions:/System/Library/Java/JavaVirtualMachines/1.6.0.jdk/Contents/Home/lib/ext, mrj.version=1060.1.6.0_29-402, user.dir=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/target/workdir, jbossas.ts.integ.dir=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/.., ts.tr.netio=100, line.separator= , java.vm.name=Java HotSpot(TM) 64-Bit Server VM, basedir=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust, file.encoding=MacRoman, jboss.options=-Xmx512m -XX:MaxPermSize=256m -Djboss.dist=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/../../../build/target/jboss-as-7.1.0.Final-SNAPSHOT, java.specification.version=1.6} 20:22:15,137 INFO [org.jboss.arquillian.container.test.impl.client.container.ClientContainerController] (main) Manual starting of a server instance 20:22:15,139 INFO [org.jboss.as.arquillian.container.managed.ManagedDeployableContainer] (main) Starting container with: [/System/Library/Frameworks/JavaVM.framework/Versions/1.6.0/Home/bin/java, -Xmx512m, -XX:MaxPermSize=256m, -Djboss.dist=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/../../../build/target/jboss-as-7.1.0.Final-SNAPSHOT, -Djava.net.preferIPv4Stack=true, -Djava.net.preferIPv6Addresses=false, -Dts.tr.fsio=100, -Dts.tr.netio=100, -Dts.tr.memio=100, -Dts.tr.db=100, -Dnode0=127.0.0.1, -Dnode1=127.0.0.1, -DudpGroup=230.0.0.4, -Djbossas.ts.submodule.dir=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust, -Djbossas.ts.integ.dir=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/.., -Djbossas.ts.dir=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/../.., -Djbossas.project.dir=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/../../.., -Djboss.dist=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/../../../build/target/jboss-as-7.1.0.Final-SNAPSHOT, -Djboss.inst=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/target/jbossas-clustering-udp-0, -Djboss.node.name=node-udp-0, -ea, -Djboss.home.dir=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/target/jbossas-clustering-udp-0, -Dorg.jboss.boot.log.file=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/target/jbossas-clustering-udp-0/standalone/log/boot.log, -Dlogging.configuration=file:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/target/jbossas-clustering-udp-0/standalone/configuration/logging.properties, -Djboss.modules.dir=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/build/target/jboss-as-7.1.0.Final-SNAPSHOT/modules, -Djboss.bundles.dir=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/build/target/jboss-as-7.1.0.Final-SNAPSHOT/bundles, -jar, /Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/target/jbossas-clustering-udp-0/jboss-modules.jar, -mp, /Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/../../../build/target/jboss-as-7.1.0.Final-SNAPSHOT/modules, -jaxpmodule, javax.xml.jaxp-provider, org.jboss.as.standalone, -server-config, standalone-ha.xml] 20:22:16,002 INFO [org.jboss.modules] JBoss Modules version 1.1.0.CR8 20:22:16,319 INFO [org.jboss.msc] JBoss MSC version 1.0.1.GA 20:22:16,378 INFO [org.jboss.as] JBoss AS 7.1.0.Final-SNAPSHOT "Flux Capacitor" starting 20:22:17,608 INFO [org.xnio] XNIO Version 3.0.0.GA 20:22:17,619 INFO [org.xnio.nio] XNIO NIO Implementation Version 3.0.0.GA 20:22:17,621 INFO [org.jboss.as.server] JBAS015888: Creating http management service using socket-binding (management-http) 20:22:17,638 INFO [org.jboss.remoting] JBoss Remoting version 3.2.0.CR9 20:22:17,671 INFO [org.jboss.as.logging] JBAS011502: Removing bootstrap log handlers 20:22:17,672 INFO [org.jboss.as.configadmin] (ServerService Thread Pool -- 30) JBAS016200: Activating ConfigAdmin Subsystem 20:22:17,683 INFO [org.jboss.as.clustering.infinispan] (ServerService Thread Pool -- 35) JBAS010280: Activating Infinispan subsystem. 20:22:17,689 INFO [org.jboss.as.jacorb] (ServerService Thread Pool -- 36) Activating JacORB Subsystem 20:22:17,698 INFO [org.jboss.as.clustering.jgroups] (ServerService Thread Pool -- 40) JBAS010260: Activating JGroups subsystem. 20:22:17,724 INFO [org.jboss.as.osgi] (ServerService Thread Pool -- 47) JBAS011910: Activating OSGi Subsystem 20:22:17,750 INFO [org.jboss.as.security] (ServerService Thread Pool -- 52) JBAS013101: Activating Security Subsystem 20:22:17,766 INFO [org.jboss.as.naming] (ServerService Thread Pool -- 46) JBAS011800: Activating Naming Subsystem 20:22:17,789 INFO [org.jboss.as.security] (MSC service thread 1-1) JBAS013100: Current PicketBox version=4.0.6.final 20:22:17,846 INFO [org.jboss.as.connector.subsystems.datasources] (ServerService Thread Pool -- 31) JBAS010403: Deploying JDBC-compliant driver class org.h2.Driver (version 1.3) 20:22:17,859 INFO [org.jboss.as.webservices] (ServerService Thread Pool -- 56) JBAS015537: Activating WebServices Extension 20:22:17,872 INFO [org.jboss.as.remoting] (MSC service thread 1-4) Listening on /127.0.0.1:4447 20:22:17,885 INFO [org.jboss.as.connector] (MSC service thread 1-4) JBAS010408: Starting JCA Subsystem (JBoss IronJacamar 1.0.7.Final) 20:22:18,036 INFO [org.jboss.as.naming] (MSC service thread 1-2) JBAS011802: Starting Naming Service 20:22:18,057 INFO [org.jboss.as.mail.extension] (MSC service thread 1-1) JBAS015400: Bound mail session [java:jboss/mail/Default] 20:22:18,136 WARN [org.jboss.as.messaging] (MSC service thread 1-1) JBAS011600: AIO wasn't located on this platform, it will fall back to using pure Java NIO. If your platform is Linux, install LibAIO to enable the AIO journal 20:22:18,249 INFO [org.jboss.ws.common.management.AbstractServerConfig] (MSC service thread 1-2) JBoss Web Services - Stack CXF Server 4.0.0.GA 20:22:18,499 INFO [org.hornetq.core.server.impl.HornetQServerImpl] (MSC service thread 1-2) live server is starting with configuration HornetQ Configuration (clustered=false,backup=false,sharedStore=true,journalDirectory=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/target/jbossas-clustering-udp-0/standalone/data/messagingjournal,bindingsDirectory=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/target/jbossas-clustering-udp-0/standalone/data/messagingbindings,largeMessagesDirectory=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/target/jbossas-clustering-udp-0/standalone/data/messaginglargemessages,pagingDirectory=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/target/jbossas-clustering-udp-0/standalone/data/messagingpaging) 20:22:18,526 WARNING [org.hornetq.core.server.impl.HornetQServerImpl] (MSC service thread 1-2) Security risk! It has been detected that the cluster admin user and password have not been changed from the installation default. Please see the HornetQ user guide, cluster chapter, for instructions on how to do this. 20:22:18,614 WARN [org.infinispan.configuration.cache.EvictionConfigurationBuilder] (ServerService Thread Pool -- 35) ISPN000152: Passivation configured without a valid eviction policy. This could mean that the cache store will never get used unless code calls Cache.evict() manually. 20:22:18,618 WARN [org.infinispan.configuration.cache.LoaderConfigurationBuilder] (ServerService Thread Pool -- 35) ISPN000149: Fetch persistent state and purge on startup are both disabled, cache may contain stale entries on startup 20:22:18,618 INFO [org.apache.tomcat.util.net.AprEndpoint] (MSC service thread 1-4) Failed to create poller with specified size of 32768 20:22:18,626 WARN [jacorb.codeset] (MSC service thread 1-1) Warning - unknown codeset (MacRoman) - defaulting to ISO-8859-1 20:22:18,654 INFO [org.apache.tomcat.util.net.AprEndpoint] (MSC service thread 1-4) Failed to create poller with specified size of 32768 20:22:18,656 INFO [org.jboss.as.jacorb] (MSC service thread 1-1) CORBA ORB Service Started 20:22:18,663 INFO [org.apache.tomcat.util.net.AprEndpoint] (MSC service thread 1-4) Failed to create poller with specified size of 16384 20:22:18,664 INFO [org.jboss.as.modcluster] (MSC service thread 1-1) JBAS011704: Mod_cluster uses default load balancer provider 20:22:18,739 INFO [org.apache.coyote.http11.Http11AprProtocol] (MSC service thread 1-4) Starting Coyote HTTP/1.1 on http-localhost-127.0.0.1-8080 20:22:18,754 WARN [org.infinispan.configuration.cache.EvictionConfigurationBuilder] (ServerService Thread Pool -- 35) ISPN000152: Passivation configured without a valid eviction policy. This could mean that the cache store will never get used unless code calls Cache.evict() manually. 20:22:18,755 WARN [org.infinispan.configuration.cache.LoaderConfigurationBuilder] (ServerService Thread Pool -- 35) ISPN000149: Fetch persistent state and purge on startup are both disabled, cache may contain stale entries on startup 20:22:18,785 WARN [org.infinispan.configuration.cache.LoaderConfigurationBuilder] (ServerService Thread Pool -- 35) ISPN000149: Fetch persistent state and purge on startup are both disabled, cache may contain stale entries on startup 20:22:18,825 WARN [org.infinispan.configuration.cache.LoaderConfigurationBuilder] (ServerService Thread Pool -- 35) ISPN000149: Fetch persistent state and purge on startup are both disabled, cache may contain stale entries on startup 20:22:18,834 INFO [org.jboss.modcluster.ModClusterService] (MSC service thread 1-1) Initializing mod_cluster 1.2.0.Beta4 20:22:18,866 INFO [org.jboss.modcluster.advertise.impl.AdvertiseListenerImpl] (MSC service thread 1-1) Listening to proxy advertisements on 224.0.1.105:23364 20:22:19,098 INFO [org.jboss.as.jacorb] (MSC service thread 1-1) CORBA Naming Service Started 20:22:19,131 INFO [org.jboss.as.connector.subsystems.datasources] (MSC service thread 1-1) JBAS010400: Bound data source [java:jboss/datasources/ExampleDS] 20:22:19,148 INFO [org.jboss.as.deployment.connector] (MSC service thread 1-4) JBAS010406: Registered connection factory java:/JmsXA 20:22:19,155 INFO [org.hornetq.ra.HornetQResourceAdapter] (MSC service thread 1-4) HornetQ resource adaptor started 20:22:19,155 INFO [org.jboss.as.connector.services.ResourceAdapterActivatorService$ResourceAdapterActivator] (MSC service thread 1-4) IJ020002: Deployed: file://RaActivatorhornetq-ra 20:22:19,157 INFO [org.jboss.as.deployment.connector] (MSC service thread 1-4) JBAS010401: Bound JCA ConnectionFactory [java:/JmsXA] 20:22:19,165 INFO [org.hornetq.core.remoting.impl.netty.NettyAcceptor] (MSC service thread 1-2) Started Netty Acceptor version 3.2.5.Final-a96d88c localhost:5445 for CORE protocol 20:22:19,196 INFO [org.hornetq.core.remoting.impl.netty.NettyAcceptor] (MSC service thread 1-2) Started Netty Acceptor version 3.2.5.Final-a96d88c localhost:5455 for CORE protocol 20:22:19,206 INFO [org.hornetq.core.server.impl.HornetQServerImpl] (MSC service thread 1-2) Server is now live 20:22:19,207 INFO [org.hornetq.core.server.impl.HornetQServerImpl] (MSC service thread 1-2) HornetQ Server version 2.2.10.Final (HQ_2_2_10_FINAL_AS7, 122) [92c192b7-4956-11e1-9991-002608e4cba7]) started 20:22:19,209 INFO [org.hornetq.core.server.impl.HornetQServerImpl] (MSC service thread 1-2) trying to deploy queue jms.queue.testQueue 20:22:19,365 INFO [org.jboss.as.messaging] (MSC service thread 1-2) JBAS011601: Bound messaging object to jndi name java:/queue/test 20:22:19,370 INFO [org.jboss.as.messaging] (MSC service thread 1-4) JBAS011601: Bound messaging object to jndi name java:/RemoteConnectionFactory 20:22:19,371 INFO [org.jboss.as.messaging] (MSC service thread 1-1) JBAS011601: Bound messaging object to jndi name java:/ConnectionFactory 20:22:19,373 INFO [org.hornetq.core.server.impl.HornetQServerImpl] (MSC service thread 1-3) trying to deploy queue jms.topic.testTopic 20:22:19,387 INFO [org.jboss.as.messaging] (MSC service thread 1-3) JBAS011601: Bound messaging object to jndi name java:/topic/test 20:22:19,397 INFO [org.jboss.as.server.deployment.scanner] (MSC service thread 1-4) JBAS015012: Started FileSystemDeploymentService for directory /Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/target/jbossas-clustering-udp-0/standalone/deployments 20:22:19,398 INFO [org.jboss.as.remoting] (MSC service thread 1-4) Listening on /127.0.0.1:9999 20:22:19,594 INFO [org.jboss.as] (Controller Boot Thread) JBAS015874: JBoss AS 7.1.0.Final-SNAPSHOT "Flux Capacitor" started in 4087ms - Started 158 of 275 services (116 services are passive or on-demand) 20:22:19,652 INFO [org.jboss.as.repository] (management-handler-threads - 2) JBAS014900: Content added at location /Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/target/jbossas-clustering-udp-0/standalone/data/content/4c/43c60a527fe03fcc8847126d65d8d589384e1c/content 20:22:19,664 INFO [org.jboss.as.server.deployment] (MSC service thread 1-1) JBAS015876: Starting deployment of "singleton.war" 20:22:19,776 WARN [org.jboss.as.dependency.private] (MSC service thread 1-3) JBAS018567: Deployment "deployment.singleton.war" is using a private module ("org.jboss.iiop-client:main") which may be changed or removed in future versions without notice. 20:22:19,777 WARN [org.jboss.as.dependency.private] (MSC service thread 1-3) JBAS018567: Deployment "deployment.singleton.war" is using a private module ("org.jboss.as.ejb3:main") which may be changed or removed in future versions without notice. 20:22:19,777 WARN [org.jboss.as.dependency.private] (MSC service thread 1-3) JBAS018567: Deployment "deployment.singleton.war" is using a private module ("org.jboss.as.jacorb:main") which may be changed or removed in future versions without notice. 20:22:19,786 WARN [org.jboss.as.dependency.private] (MSC service thread 1-3) JBAS018567: Deployment "deployment.singleton.war" is using a private module ("org.jboss.as.ee:main") which may be changed or removed in future versions without notice. 20:22:19,787 WARN [org.jboss.as.dependency.private] (MSC service thread 1-3) JBAS018567: Deployment "deployment.singleton.war" is using a private module ("com.sun.jsf-impl:main") which may be changed or removed in future versions without notice. 20:22:19,787 WARN [org.jboss.as.dependency.private] (MSC service thread 1-3) JBAS018567: Deployment "deployment.singleton.war" is using a private module ("org.jboss.as.web:main") which may be changed or removed in future versions without notice. 20:22:19,787 WARN [org.jboss.as.dependency.private] (MSC service thread 1-3) JBAS018567: Deployment "deployment.singleton.war" is using a private module ("org.jboss.ws.spi:main") which may be changed or removed in future versions without notice. 20:22:19,788 WARN [org.jboss.as.dependency.private] (MSC service thread 1-3) JBAS018567: Deployment "deployment.singleton.war" is using a private module ("org.jboss.as.security:main") which may be changed or removed in future versions without notice. 20:22:19,794 WARN [org.jboss.as.dependency.private] (MSC service thread 1-3) JBAS018567: Deployment "deployment.singleton.war" is using a private module ("org.jacorb:main") which may be changed or removed in future versions without notice. 20:22:19,797 WARN [org.jboss.as.dependency.private] (MSC service thread 1-3) JBAS018567: Deployment "deployment.singleton.war" is using a private module ("org.jboss.as.clustering.singleton:main") which may be changed or removed in future versions without notice. 20:22:19,798 WARN [org.jboss.as.dependency.private] (MSC service thread 1-3) JBAS018567: Deployment "deployment.singleton.war" is using a private module ("org.jboss.as.server:main") which may be changed or removed in future versions without notice. 20:22:20,054 INFO [stdout] (MSC service thread 1-4) Service controller state is DOWN, waiting for transition to UP 20:22:20,336 WARNING [org.jgroups.protocols.UDP] (pool-5-thread-1) receive buffer of socket java.net.DatagramSocket@6495a1b6 was set to 20MB, but the OS only allocated 65.51KB. This might lead to performance problems. Please set your max receive buffer in the OS correctly (e.g. net.core.rmem_max on Linux) 20:22:20,338 WARNING [org.jgroups.protocols.UDP] (pool-5-thread-1) receive buffer of socket java.net.MulticastSocket@1babe751 was set to 25MB, but the OS only allocated 65.51KB. This might lead to performance problems. Please set your max receive buffer in the OS correctly (e.g. net.core.rmem_max on Linux) 20:22:20,348 INFO [stdout] (pool-5-thread-1) 20:22:20,348 INFO [stdout] (pool-5-thread-1) ------------------------------------------------------------------- 20:22:20,349 INFO [stdout] (pool-5-thread-1) GMS: address=node-udp-0/cluster, cluster=cluster, physical address=127.0.0.1:55200 20:22:20,349 INFO [stdout] (pool-5-thread-1) ------------------------------------------------------------------- 20:22:22,390 INFO [org.jboss.as.clustering.CoreGroupCommunicationService.cluster] (MSC service thread 1-2) JBAS010206: Number of cluster members: 1 20:22:22,747 INFO [org.infinispan.remoting.transport.jgroups.JGroupsTransport] (pool-9-thread-1) ISPN000078: Starting JGroups Channel 20:22:22,748 WARNING [org.infinispan.remoting.transport.jgroups.CommandAwareRpcDispatcher] (pool-9-thread-1) Channel Muxer already has a default up handler installed (org.jboss.as.clustering.jgroups.ClassLoaderAwareUpHandler@3cb5e07a) but now it is being overridden 20:22:22,749 INFO [org.infinispan.remoting.transport.jgroups.JGroupsTransport] (pool-9-thread-1) ISPN000094: Received new cluster view: [node-udp-0/cluster|0] [node-udp-0/cluster] 20:22:22,749 INFO [org.infinispan.remoting.transport.jgroups.JGroupsTransport] (pool-9-thread-1) ISPN000079: Cache local address is node-udp-0/cluster, physical addresses are [127.0.0.1:55200] 20:22:22,753 INFO [org.infinispan.factories.GlobalComponentRegistry] (pool-9-thread-1) ISPN000128: Infinispan version: Infinispan 'Brahma' 5.1.0.FINAL 20:22:22,764 INFO [org.infinispan.jmx.CacheJmxRegistration] (pool-9-thread-1) ISPN000031: MBeans were successfully registered to the platform mbean server. 20:22:22,789 INFO [org.jboss.as.clustering.infinispan] (pool-9-thread-1) JBAS010281: Started default cache from cluster container 20:22:22,804 INFO [stdout] (MSC service thread 1-4) Service controller state is STARTING, waiting for transition to UP 20:22:22,822 INFO [org.jboss.as.clustering.singleton] (pool-26-thread-1) JBAS010340: This node will now operate as the singleton provider of the jboss.test.myservice service 20:22:22,900 INFO [org.jboss.web] (MSC service thread 1-4) JBAS018210: Registering web context: /singleton 20:22:22,977 INFO [org.jboss.as.server] (management-handler-threads - 2) JBAS018559: Deployed "singleton.war" 20:22:23,007 INFO [stdout] (management-handler-threads - 2) java.lang.Exception 20:22:23,007 INFO [stdout] (management-handler-threads - 2) at org.jboss.as.controller.operations.common.ProcessEnvironment$ProcessNameReadAttributeHandler.execute(ProcessEnvironment.java:176) 20:22:23,007 INFO [stdout] (management-handler-threads - 2) at org.jboss.as.controller.AbstractOperationContext.executeStep(AbstractOperationContext.java:378) 20:22:23,008 INFO [stdout] (management-handler-threads - 2) at org.jboss.as.controller.AbstractOperationContext.doCompleteStep(AbstractOperationContext.java:265) 20:22:23,008 INFO [stdout] (management-handler-threads - 2) at org.jboss.as.controller.AbstractOperationContext.completeStep(AbstractOperationContext.java:193) 20:22:23,008 INFO [stdout] (management-handler-threads - 2) at org.jboss.as.controller.ModelControllerImpl$DefaultPrepareStepHandler.execute(ModelControllerImpl.java:461) 20:22:23,008 INFO [stdout] (management-handler-threads - 2) at org.jboss.as.controller.AbstractOperationContext.executeStep(AbstractOperationContext.java:378) 20:22:23,009 INFO [stdout] (management-handler-threads - 2) at org.jboss.as.controller.AbstractOperationContext.doCompleteStep(AbstractOperationContext.java:265) 20:22:23,009 INFO [stdout] (management-handler-threads - 2) at org.jboss.as.controller.AbstractOperationContext.completeStep(AbstractOperationContext.java:193) 20:22:23,009 INFO [stdout] (management-handler-threads - 2) at org.jboss.as.controller.ModelControllerImpl.execute(ModelControllerImpl.java:121) 20:22:23,009 INFO [stdout] (management-handler-threads - 2) at org.jboss.as.controller.remote.ModelControllerClientOperationHandler$ExecuteRequestHandler.doExecute(ModelControllerClientOperationHandler.java:121) 20:22:23,009 INFO [stdout] (management-handler-threads - 2) at org.jboss.as.controller.remote.ModelControllerClientOperationHandler$ExecuteRequestHandler$1.execute(ModelControllerClientOperationHandler.java:98) 20:22:23,013 INFO [stdout] (management-handler-threads - 2) at org.jboss.as.protocol.mgmt.AbstractMessageHandler$2$1.doExecute(AbstractMessageHandler.java:250) 20:22:23,013 INFO [stdout] (management-handler-threads - 2) at org.jboss.as.protocol.mgmt.AbstractMessageHandler$AsyncTaskRunner.run(AbstractMessageHandler.java:444) 20:22:23,013 INFO [stdout] (management-handler-threads - 2) at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) 20:22:23,014 INFO [stdout] (management-handler-threads - 2) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) 20:22:23,014 INFO [stdout] (management-handler-threads - 2) at java.lang.Thread.run(Thread.java:680) 20:22:23,014 INFO [stdout] (management-handler-threads - 2) at org.jboss.threads.JBossThread.run(JBossThread.java:122) 20:22:23,550 INFO [org.jboss.arquillian.container.test.impl.client.container.ClientContainerController] (main) Manual starting of a server instance 20:22:23,553 INFO [org.jboss.as.arquillian.container.managed.ManagedDeployableContainer] (main) Starting container with: [/System/Library/Frameworks/JavaVM.framework/Versions/1.6.0/Home/bin/java, -Xmx512m, -XX:MaxPermSize=256m, -Djboss.dist=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/../../../build/target/jboss-as-7.1.0.Final-SNAPSHOT, -Djava.net.preferIPv4Stack=true, -Djava.net.preferIPv6Addresses=false, -Dts.tr.fsio=100, -Dts.tr.netio=100, -Dts.tr.memio=100, -Dts.tr.db=100, -Dnode0=127.0.0.1, -Dnode1=127.0.0.1, -DudpGroup=230.0.0.4, -Djbossas.ts.submodule.dir=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust, -Djbossas.ts.integ.dir=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/.., -Djbossas.ts.dir=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/../.., -Djbossas.project.dir=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/../../.., -Djboss.dist=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/../../../build/target/jboss-as-7.1.0.Final-SNAPSHOT, -Djboss.inst=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/target/jbossas-clustering-udp-1, -Djboss.node.name=node-udp-1, -Djboss.port.offset=100, -ea, -Djboss.home.dir=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/target/jbossas-clustering-udp-1, -Dorg.jboss.boot.log.file=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/target/jbossas-clustering-udp-1/standalone/log/boot.log, -Dlogging.configuration=file:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/target/jbossas-clustering-udp-1/standalone/configuration/logging.properties, -Djboss.modules.dir=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/build/target/jboss-as-7.1.0.Final-SNAPSHOT/modules, -Djboss.bundles.dir=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/build/target/jboss-as-7.1.0.Final-SNAPSHOT/bundles, -jar, /Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/target/jbossas-clustering-udp-1/jboss-modules.jar, -mp, /Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/../../../build/target/jboss-as-7.1.0.Final-SNAPSHOT/modules, -jaxpmodule, javax.xml.jaxp-provider, org.jboss.as.standalone, -server-config, standalone-ha.xml] 20:22:24,267 INFO [org.jboss.modules] JBoss Modules version 1.1.0.CR8 20:22:24,593 INFO [org.jboss.msc] JBoss MSC version 1.0.1.GA 20:22:24,660 INFO [org.jboss.as] JBoss AS 7.1.0.Final-SNAPSHOT "Flux Capacitor" starting 20:22:26,554 INFO [org.xnio] XNIO Version 3.0.0.GA 20:22:26,560 INFO [org.jboss.as.server] JBAS015888: Creating http management service using socket-binding (management-http) 20:22:26,568 INFO [org.xnio.nio] XNIO NIO Implementation Version 3.0.0.GA 20:22:26,578 INFO [org.jboss.remoting] JBoss Remoting version 3.2.0.CR9 20:22:26,640 INFO [org.jboss.as.logging] JBAS011502: Removing bootstrap log handlers 20:22:26,644 INFO [org.jboss.as.configadmin] (ServerService Thread Pool -- 30) JBAS016200: Activating ConfigAdmin Subsystem 20:22:26,667 INFO [org.jboss.as.clustering.infinispan] (ServerService Thread Pool -- 35) JBAS010280: Activating Infinispan subsystem. 20:22:26,671 INFO [org.jboss.as.jacorb] (ServerService Thread Pool -- 36) Activating JacORB Subsystem 20:22:26,689 INFO [org.jboss.as.clustering.jgroups] (ServerService Thread Pool -- 40) JBAS010260: Activating JGroups subsystem. 20:22:26,706 INFO [org.jboss.as.connector] (MSC service thread 1-4) JBAS010408: Starting JCA Subsystem (JBoss IronJacamar 1.0.7.Final) 20:22:26,757 INFO [org.jboss.as.connector.subsystems.datasources] (ServerService Thread Pool -- 31) JBAS010403: Deploying JDBC-compliant driver class org.h2.Driver (version 1.3) 20:22:26,789 INFO [org.jboss.as.naming] (ServerService Thread Pool -- 46) JBAS011800: Activating Naming Subsystem 20:22:26,800 INFO [org.jboss.as.osgi] (ServerService Thread Pool -- 47) JBAS011910: Activating OSGi Subsystem 20:22:26,817 INFO [org.jboss.as.naming] (MSC service thread 1-4) JBAS011802: Starting Naming Service 20:22:26,833 INFO [org.jboss.as.mail.extension] (MSC service thread 1-4) JBAS015400: Bound mail session [java:jboss/mail/Default] 20:22:26,872 WARN [org.jboss.as.messaging] (MSC service thread 1-3) JBAS011600: AIO wasn't located on this platform, it will fall back to using pure Java NIO. If your platform is Linux, install LibAIO to enable the AIO journal 20:22:26,935 INFO [org.jboss.as.security] (ServerService Thread Pool -- 52) JBAS013101: Activating Security Subsystem 20:22:26,946 INFO [org.jboss.as.webservices] (ServerService Thread Pool -- 56) JBAS015537: Activating WebServices Extension 20:22:27,092 INFO [org.jboss.as.remoting] (MSC service thread 1-1) Listening on localhost/127.0.0.1:4547 20:22:27,104 INFO [org.jboss.as.security] (MSC service thread 1-1) JBAS013100: Current PicketBox version=4.0.6.final 20:22:27,180 WARN [jacorb.codeset] (MSC service thread 1-2) Warning - unknown codeset (MacRoman) - defaulting to ISO-8859-1 20:22:27,200 INFO [org.jboss.as.jacorb] (MSC service thread 1-2) CORBA ORB Service Started 20:22:27,252 INFO [org.jboss.ws.common.management.AbstractServerConfig] (MSC service thread 1-4) JBoss Web Services - Stack CXF Server 4.0.0.GA 20:22:27,296 INFO [org.hornetq.core.server.impl.HornetQServerImpl] (MSC service thread 1-1) live server is starting with configuration HornetQ Configuration (clustered=false,backup=false,sharedStore=true,journalDirectory=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/target/jbossas-clustering-udp-1/standalone/data/messagingjournal,bindingsDirectory=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/target/jbossas-clustering-udp-1/standalone/data/messagingbindings,largeMessagesDirectory=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/target/jbossas-clustering-udp-1/standalone/data/messaginglargemessages,pagingDirectory=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/target/jbossas-clustering-udp-1/standalone/data/messagingpaging) 20:22:27,417 WARNING [org.hornetq.core.server.impl.HornetQServerImpl] (MSC service thread 1-1) Security risk! It has been detected that the cluster admin user and password have not been changed from the installation default. Please see the HornetQ user guide, cluster chapter, for instructions on how to do this. 20:22:27,538 WARN [org.infinispan.configuration.cache.EvictionConfigurationBuilder] (ServerService Thread Pool -- 35) ISPN000152: Passivation configured without a valid eviction policy. This could mean that the cache store will never get used unless code calls Cache.evict() manually. 20:22:27,539 WARN [org.infinispan.configuration.cache.LoaderConfigurationBuilder] (ServerService Thread Pool -- 35) ISPN000149: Fetch persistent state and purge on startup are both disabled, cache may contain stale entries on startup 20:22:27,581 WARN [org.infinispan.configuration.cache.EvictionConfigurationBuilder] (ServerService Thread Pool -- 35) ISPN000152: Passivation configured without a valid eviction policy. This could mean that the cache store will never get used unless code calls Cache.evict() manually. 20:22:27,582 WARN [org.infinispan.configuration.cache.LoaderConfigurationBuilder] (ServerService Thread Pool -- 35) ISPN000149: Fetch persistent state and purge on startup are both disabled, cache may contain stale entries on startup 20:22:27,600 WARN [org.infinispan.configuration.cache.LoaderConfigurationBuilder] (ServerService Thread Pool -- 35) ISPN000149: Fetch persistent state and purge on startup are both disabled, cache may contain stale entries on startup 20:22:27,639 WARN [org.infinispan.configuration.cache.LoaderConfigurationBuilder] (ServerService Thread Pool -- 35) ISPN000149: Fetch persistent state and purge on startup are both disabled, cache may contain stale entries on startup 20:22:27,726 INFO [org.apache.tomcat.util.net.AprEndpoint] (MSC service thread 1-3) Failed to create poller with specified size of 32768 20:22:27,771 INFO [org.apache.tomcat.util.net.AprEndpoint] (MSC service thread 1-3) Failed to create poller with specified size of 32768 20:22:27,775 INFO [org.apache.tomcat.util.net.AprEndpoint] (MSC service thread 1-3) Failed to create poller with specified size of 16384 20:22:27,777 INFO [org.jboss.as.modcluster] (MSC service thread 1-2) JBAS011704: Mod_cluster uses default load balancer provider 20:22:27,781 INFO [org.apache.coyote.http11.Http11AprProtocol] (MSC service thread 1-3) Starting Coyote HTTP/1.1 on http-localhost-127.0.0.1-8180 20:22:27,876 INFO [org.jboss.modcluster.ModClusterService] (MSC service thread 1-2) Initializing mod_cluster 1.2.0.Beta4 20:22:27,905 INFO [org.jboss.modcluster.advertise.impl.AdvertiseListenerImpl] (MSC service thread 1-2) Listening to proxy advertisements on 224.0.1.105:23364 20:22:27,999 INFO [org.jboss.as.jacorb] (MSC service thread 1-3) CORBA Naming Service Started 20:22:28,099 INFO [org.hornetq.core.remoting.impl.netty.NettyAcceptor] (MSC service thread 1-1) Started Netty Acceptor version 3.2.5.Final-a96d88c localhost:5555 for CORE protocol 20:22:28,100 INFO [org.hornetq.core.remoting.impl.netty.NettyAcceptor] (MSC service thread 1-1) Started Netty Acceptor version 3.2.5.Final-a96d88c localhost:5545 for CORE protocol 20:22:28,116 INFO [org.hornetq.core.server.impl.HornetQServerImpl] (MSC service thread 1-1) Server is now live 20:22:28,116 INFO [org.hornetq.core.server.impl.HornetQServerImpl] (MSC service thread 1-1) HornetQ Server version 2.2.10.Final (HQ_2_2_10_FINAL_AS7, 122) [9989b7ab-4956-11e1-b947-002608e4cba7]) started 20:22:28,154 INFO [org.jboss.as.messaging] (MSC service thread 1-1) JBAS011601: Bound messaging object to jndi name java:/ConnectionFactory 20:22:28,157 INFO [org.hornetq.core.server.impl.HornetQServerImpl] (MSC service thread 1-4) trying to deploy queue jms.topic.testTopic 20:22:28,257 INFO [org.jboss.as.deployment.connector] (MSC service thread 1-1) JBAS010406: Registered connection factory java:/JmsXA 20:22:28,274 INFO [org.hornetq.ra.HornetQResourceAdapter] (MSC service thread 1-1) HornetQ resource adaptor started 20:22:28,275 INFO [org.jboss.as.connector.services.ResourceAdapterActivatorService$ResourceAdapterActivator] (MSC service thread 1-1) IJ020002: Deployed: file://RaActivatorhornetq-ra 20:22:28,278 INFO [org.jboss.as.connector.subsystems.datasources] (MSC service thread 1-1) JBAS010400: Bound data source [java:jboss/datasources/ExampleDS] 20:22:28,292 INFO [org.jboss.as.deployment.connector] (MSC service thread 1-1) JBAS010401: Bound JCA ConnectionFactory [java:/JmsXA] 20:22:28,312 INFO [org.jboss.as.messaging] (MSC service thread 1-4) JBAS011601: Bound messaging object to jndi name java:/topic/test 20:22:28,323 INFO [org.hornetq.core.server.impl.HornetQServerImpl] (MSC service thread 1-2) trying to deploy queue jms.queue.testQueue 20:22:28,345 INFO [org.jboss.as.messaging] (MSC service thread 1-2) JBAS011601: Bound messaging object to jndi name java:/queue/test 20:22:28,348 INFO [org.jboss.as.messaging] (MSC service thread 1-3) JBAS011601: Bound messaging object to jndi name java:/RemoteConnectionFactory 20:22:28,430 INFO [org.jboss.as.remoting] (MSC service thread 1-2) Listening on /127.0.0.1:10099 20:22:28,476 INFO [org.jboss.as.server.deployment.scanner] (MSC service thread 1-4) JBAS015012: Started FileSystemDeploymentService for directory /Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/target/jbossas-clustering-udp-1/standalone/deployments 20:22:28,560 INFO [org.jboss.as] (Controller Boot Thread) JBAS015874: JBoss AS 7.1.0.Final-SNAPSHOT "Flux Capacitor" started in 4706ms - Started 158 of 275 services (116 services are passive or on-demand) 20:22:28,618 INFO [org.jboss.as.repository] (management-handler-threads - 1) JBAS014900: Content added at location /Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/target/jbossas-clustering-udp-1/standalone/data/content/48/394782342bb5c8760aa276097ba19ffa989563/content 20:22:28,631 INFO [org.jboss.as.server.deployment] (MSC service thread 1-2) JBAS015876: Starting deployment of "singleton.war" 20:22:28,751 WARN [org.jboss.as.dependency.private] (MSC service thread 1-3) JBAS018567: Deployment "deployment.singleton.war" is using a private module ("org.jboss.iiop-client:main") which may be changed or removed in future versions without notice. 20:22:28,757 WARN [org.jboss.as.dependency.private] (MSC service thread 1-3) JBAS018567: Deployment "deployment.singleton.war" is using a private module ("org.jboss.as.ejb3:main") which may be changed or removed in future versions without notice. 20:22:28,757 WARN [org.jboss.as.dependency.private] (MSC service thread 1-3) JBAS018567: Deployment "deployment.singleton.war" is using a private module ("org.jboss.as.jacorb:main") which may be changed or removed in future versions without notice. 20:22:28,762 WARN [org.jboss.as.dependency.private] (MSC service thread 1-3) JBAS018567: Deployment "deployment.singleton.war" is using a private module ("org.jboss.as.ee:main") which may be changed or removed in future versions without notice. 20:22:28,763 WARN [org.jboss.as.dependency.private] (MSC service thread 1-3) JBAS018567: Deployment "deployment.singleton.war" is using a private module ("com.sun.jsf-impl:main") which may be changed or removed in future versions without notice. 20:22:28,764 WARN [org.jboss.as.dependency.private] (MSC service thread 1-3) JBAS018567: Deployment "deployment.singleton.war" is using a private module ("org.jboss.as.web:main") which may be changed or removed in future versions without notice. 20:22:28,764 WARN [org.jboss.as.dependency.private] (MSC service thread 1-3) JBAS018567: Deployment "deployment.singleton.war" is using a private module ("org.jboss.ws.spi:main") which may be changed or removed in future versions without notice. 20:22:28,764 WARN [org.jboss.as.dependency.private] (MSC service thread 1-3) JBAS018567: Deployment "deployment.singleton.war" is using a private module ("org.jboss.as.security:main") which may be changed or removed in future versions without notice. 20:22:28,766 WARN [org.jboss.as.dependency.private] (MSC service thread 1-3) JBAS018567: Deployment "deployment.singleton.war" is using a private module ("org.jacorb:main") which may be changed or removed in future versions without notice. 20:22:28,768 WARN [org.jboss.as.dependency.private] (MSC service thread 1-3) JBAS018567: Deployment "deployment.singleton.war" is using a private module ("org.jboss.as.clustering.singleton:main") which may be changed or removed in future versions without notice. 20:22:28,769 WARN [org.jboss.as.dependency.private] (MSC service thread 1-3) JBAS018567: Deployment "deployment.singleton.war" is using a private module ("org.jboss.as.server:main") which may be changed or removed in future versions without notice. 20:22:29,011 INFO [stdout] (MSC service thread 1-3) Service controller state is DOWN, waiting for transition to UP 20:22:29,338 WARNING [org.jgroups.protocols.UDP] (pool-5-thread-1) receive buffer of socket java.net.DatagramSocket@64fa8cc6 was set to 20MB, but the OS only allocated 65.51KB. This might lead to performance problems. Please set your max receive buffer in the OS correctly (e.g. net.core.rmem_max on Linux) 20:22:29,339 WARNING [org.jgroups.protocols.UDP] (pool-5-thread-1) receive buffer of socket java.net.MulticastSocket@7d304ef4 was set to 25MB, but the OS only allocated 65.51KB. This might lead to performance problems. Please set your max receive buffer in the OS correctly (e.g. net.core.rmem_max on Linux) 20:22:29,400 INFO [stdout] (pool-5-thread-1) 20:22:29,400 INFO [stdout] (pool-5-thread-1) ------------------------------------------------------------------- 20:22:29,400 INFO [stdout] (pool-5-thread-1) GMS: address=node-udp-1/cluster, cluster=cluster, physical address=127.0.0.1:55300 20:22:29,400 INFO [stdout] (pool-5-thread-1) ------------------------------------------------------------------- 20:22:29,494 INFO [org.jboss.as.clustering.CoreGroupCommunicationService.lifecycle.cluster] (Incoming-1,null) JBAS010247: New cluster view for partition cluster (id: 1, delta: 1, merge: false) : [node-udp-0/cluster, node-udp-1/cluster] 20:22:29,495 INFO [org.infinispan.remoting.transport.jgroups.JGroupsTransport] (Incoming-1,null) ISPN000094: Received new cluster view: [node-udp-0/cluster|1] [node-udp-0/cluster, node-udp-1/cluster] 20:22:29,553 INFO [org.jboss.as.clustering.CoreGroupCommunicationService.cluster] (MSC service thread 1-4) JBAS010206: Number of cluster members: 2 20:22:30,069 INFO [org.infinispan.remoting.transport.jgroups.JGroupsTransport] (pool-8-thread-1) ISPN000078: Starting JGroups Channel 20:22:30,070 WARNING [org.infinispan.remoting.transport.jgroups.CommandAwareRpcDispatcher] (pool-8-thread-1) Channel Muxer already has a default up handler installed (org.jboss.as.clustering.jgroups.ClassLoaderAwareUpHandler@1502c065) but now it is being overridden 20:22:30,071 INFO [org.infinispan.remoting.transport.jgroups.JGroupsTransport] (pool-8-thread-1) ISPN000094: Received new cluster view: [node-udp-0/cluster|1] [node-udp-0/cluster, node-udp-1/cluster] 20:22:30,071 INFO [org.infinispan.remoting.transport.jgroups.JGroupsTransport] (pool-8-thread-1) ISPN000079: Cache local address is node-udp-1/cluster, physical addresses are [127.0.0.1:55300] 20:22:30,075 INFO [org.infinispan.factories.GlobalComponentRegistry] (pool-8-thread-1) ISPN000128: Infinispan version: Infinispan 'Brahma' 5.1.0.FINAL 20:22:30,087 INFO [org.infinispan.jmx.CacheJmxRegistration] (pool-8-thread-1) ISPN000031: MBeans were successfully registered to the platform mbean server. 20:22:30,155 INFO [org.jboss.as.clustering.infinispan] (pool-8-thread-1) JBAS010281: Started default cache from cluster container 20:22:30,173 INFO [stdout] (MSC service thread 1-3) Service controller state is STARTING, waiting for transition to UP 20:22:30,270 INFO [org.jboss.as.clustering.singleton] (OOB-20,null) JBAS010341: This node will no longer operate as the singleton provider of the jboss.test.myservice service 20:22:30,277 INFO [org.jboss.as.clustering.singleton] (pool-26-thread-1) JBAS010340: This node will now operate as the singleton provider of the jboss.test.myservice service 20:22:30,335 INFO [org.jboss.web] (MSC service thread 1-3) JBAS018210: Registering web context: /singleton 20:22:30,385 INFO [org.jboss.as.server] (management-handler-threads - 1) JBAS018559: Deployed "singleton.war" 20:22:30,402 INFO [stdout] (management-handler-threads - 1) java.lang.Exception 20:22:30,402 INFO [stdout] (management-handler-threads - 1) at org.jboss.as.controller.operations.common.ProcessEnvironment$ProcessNameReadAttributeHandler.execute(ProcessEnvironment.java:176) 20:22:30,402 INFO [stdout] (management-handler-threads - 1) at org.jboss.as.controller.AbstractOperationContext.executeStep(AbstractOperationContext.java:378) 20:22:30,403 INFO [stdout] (management-handler-threads - 1) at org.jboss.as.controller.AbstractOperationContext.doCompleteStep(AbstractOperationContext.java:265) 20:22:30,403 INFO [stdout] (management-handler-threads - 1) at org.jboss.as.controller.AbstractOperationContext.completeStep(AbstractOperationContext.java:193) 20:22:30,403 INFO [stdout] (management-handler-threads - 1) at org.jboss.as.controller.ModelControllerImpl$DefaultPrepareStepHandler.execute(ModelControllerImpl.java:461) 20:22:30,403 INFO [stdout] (management-handler-threads - 1) at org.jboss.as.controller.AbstractOperationContext.executeStep(AbstractOperationContext.java:378) 20:22:30,403 INFO [stdout] (management-handler-threads - 1) at org.jboss.as.controller.AbstractOperationContext.doCompleteStep(AbstractOperationContext.java:265) 20:22:30,403 INFO [stdout] (management-handler-threads - 1) at org.jboss.as.controller.AbstractOperationContext.completeStep(AbstractOperationContext.java:193) 20:22:30,404 INFO [stdout] (management-handler-threads - 1) at org.jboss.as.controller.ModelControllerImpl.execute(ModelControllerImpl.java:121) 20:22:30,404 INFO [stdout] (management-handler-threads - 1) at org.jboss.as.controller.remote.ModelControllerClientOperationHandler$ExecuteRequestHandler.doExecute(ModelControllerClientOperationHandler.java:121) 20:22:30,404 INFO [stdout] (management-handler-threads - 1) at org.jboss.as.controller.remote.ModelControllerClientOperationHandler$ExecuteRequestHandler$1.execute(ModelControllerClientOperationHandler.java:98) 20:22:30,404 INFO [stdout] (management-handler-threads - 1) at org.jboss.as.protocol.mgmt.AbstractMessageHandler$2$1.doExecute(AbstractMessageHandler.java:250) 20:22:30,404 INFO [stdout] (management-handler-threads - 1) at org.jboss.as.protocol.mgmt.AbstractMessageHandler$AsyncTaskRunner.run(AbstractMessageHandler.java:444) 20:22:30,404 INFO [stdout] (management-handler-threads - 1) at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) 20:22:30,405 INFO [stdout] (management-handler-threads - 1) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) 20:22:30,405 INFO [stdout] (management-handler-threads - 1) at java.lang.Thread.run(Thread.java:680) 20:22:30,405 INFO [stdout] (management-handler-threads - 1) at org.jboss.threads.JBossThread.run(JBossThread.java:122) 20:22:35,865 INFO [org.jboss.arquillian.container.test.impl.client.container.ClientContainerController] (main) Manual stopping of a server instance 20:22:35,950 INFO [org.jboss.as.clustering.singleton] (OOB-20,null) JBAS010340: This node will now operate as the singleton provider of the jboss.test.myservice service 20:22:36,602 INFO [org.jboss.as.clustering.CoreGroupCommunicationService.lifecycle.cluster] (Incoming-9,null) JBAS010247: New cluster view for partition cluster (id: 2, delta: -1, merge: false) : [node-udp-0/cluster] 20:22:36,603 INFO [org.infinispan.remoting.transport.jgroups.JGroupsTransport] (Incoming-9,null) ISPN000094: Received new cluster view: [node-udp-0/cluster|2] [node-udp-0/cluster] 20:22:36,604 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=4, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=3, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: Suspected member: node-udp-1/cluster at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: Suspected member: node-udp-1/cluster at org.infinispan.remoting.transport.AbstractTransport.parseResponseAndAddToResponseList(AbstractTransport.java:93) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:461) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:22:36,659 INFO [org.jboss.as.clustering.CoreGroupCommunicationService.lifecycle.cluster] (Incoming-11,null) JBAS010247: New cluster view for partition cluster (id: 3, delta: 1, merge: false) : [node-udp-0/cluster, node-udp-1/cluster] 20:22:36,659 INFO [org.infinispan.remoting.transport.jgroups.JGroupsTransport] (Incoming-11,null) ISPN000094: Received new cluster view: [node-udp-0/cluster|3] [node-udp-0/cluster, node-udp-1/cluster] 20:22:41,664 WARNING [org.jgroups.protocols.pbcast.GMS] (ViewHandler,cluster,node-udp-0/cluster) node-udp-0/cluster: failed to collect all ACKs (expected=1) for unicast view [node-udp-0/cluster|3] [node-udp-0/cluster, node-udp-1/cluster] after 5000ms, missing ACKs from [node-udp-1/cluster] 20:22:41,666 INFO [org.jboss.arquillian.container.test.impl.client.container.ClientContainerController] (main) Manual starting of a server instance 20:22:41,668 INFO [org.jboss.as.arquillian.container.managed.ManagedDeployableContainer] (main) Starting container with: [/System/Library/Frameworks/JavaVM.framework/Versions/1.6.0/Home/bin/java, -Xmx512m, -XX:MaxPermSize=256m, -Djboss.dist=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/../../../build/target/jboss-as-7.1.0.Final-SNAPSHOT, -Djava.net.preferIPv4Stack=true, -Djava.net.preferIPv6Addresses=false, -Dts.tr.fsio=100, -Dts.tr.netio=100, -Dts.tr.memio=100, -Dts.tr.db=100, -Dnode0=127.0.0.1, -Dnode1=127.0.0.1, -DudpGroup=230.0.0.4, -Djbossas.ts.submodule.dir=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust, -Djbossas.ts.integ.dir=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/.., -Djbossas.ts.dir=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/../.., -Djbossas.project.dir=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/../../.., -Djboss.dist=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/../../../build/target/jboss-as-7.1.0.Final-SNAPSHOT, -Djboss.inst=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/target/jbossas-clustering-udp-1, -Djboss.node.name=node-udp-1, -Djboss.port.offset=100, -ea, -Djboss.home.dir=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/target/jbossas-clustering-udp-1, -Dorg.jboss.boot.log.file=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/target/jbossas-clustering-udp-1/standalone/log/boot.log, -Dlogging.configuration=file:/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/target/jbossas-clustering-udp-1/standalone/configuration/logging.properties, -Djboss.modules.dir=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/build/target/jboss-as-7.1.0.Final-SNAPSHOT/modules, -Djboss.bundles.dir=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/build/target/jboss-as-7.1.0.Final-SNAPSHOT/bundles, -jar, /Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/target/jbossas-clustering-udp-1/jboss-modules.jar, -mp, /Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/../../../build/target/jboss-as-7.1.0.Final-SNAPSHOT/modules, -jaxpmodule, javax.xml.jaxp-provider, org.jboss.as.standalone, -server-config, standalone-ha.xml] 20:22:42,557 INFO [org.jboss.modules] JBoss Modules version 1.1.0.CR8 20:22:42,892 INFO [org.jboss.msc] JBoss MSC version 1.0.1.GA 20:22:43,000 INFO [org.jboss.as] JBoss AS 7.1.0.Final-SNAPSHOT "Flux Capacitor" starting 20:22:45,594 INFO [org.xnio] XNIO Version 3.0.0.GA 20:22:45,610 INFO [org.xnio.nio] XNIO NIO Implementation Version 3.0.0.GA 20:22:45,623 INFO [org.jboss.remoting] JBoss Remoting version 3.2.0.CR9 20:22:45,635 INFO [org.jboss.as.server] JBAS015888: Creating http management service using socket-binding (management-http) 20:22:45,707 INFO [org.jboss.as.logging] JBAS011502: Removing bootstrap log handlers 20:22:45,717 INFO [org.jboss.as.configadmin] (ServerService Thread Pool -- 30) JBAS016200: Activating ConfigAdmin Subsystem 20:22:45,741 INFO [org.jboss.as.connector.subsystems.datasources] (ServerService Thread Pool -- 31) JBAS010403: Deploying JDBC-compliant driver class org.h2.Driver (version 1.3) 20:22:45,752 INFO [org.jboss.as.jacorb] (ServerService Thread Pool -- 36) Activating JacORB Subsystem 20:22:45,752 INFO [org.jboss.as.clustering.infinispan] (ServerService Thread Pool -- 35) JBAS010280: Activating Infinispan subsystem. 20:22:45,766 INFO [org.jboss.as.clustering.jgroups] (ServerService Thread Pool -- 40) JBAS010260: Activating JGroups subsystem. 20:22:45,834 INFO [org.jboss.as.naming] (ServerService Thread Pool -- 46) JBAS011800: Activating Naming Subsystem 20:22:45,836 INFO [org.jboss.as.osgi] (ServerService Thread Pool -- 47) JBAS011910: Activating OSGi Subsystem 20:22:45,852 INFO [org.jboss.as.connector] (MSC service thread 1-2) JBAS010408: Starting JCA Subsystem (JBoss IronJacamar 1.0.7.Final) 20:22:45,886 INFO [org.jboss.as.security] (ServerService Thread Pool -- 52) JBAS013101: Activating Security Subsystem 20:22:45,891 INFO [org.jboss.as.naming] (MSC service thread 1-2) JBAS011802: Starting Naming Service 20:22:45,900 INFO [org.jboss.as.mail.extension] (MSC service thread 1-2) JBAS015400: Bound mail session [java:jboss/mail/Default] 20:22:45,909 INFO [org.jboss.as.security] (MSC service thread 1-2) JBAS013100: Current PicketBox version=4.0.6.final 20:22:45,929 INFO [org.jboss.as.webservices] (ServerService Thread Pool -- 56) JBAS015537: Activating WebServices Extension 20:22:45,945 INFO [org.jboss.as.remoting] (MSC service thread 1-4) Listening on /127.0.0.1:4547 20:22:46,265 INFO [org.jboss.ws.common.management.AbstractServerConfig] (MSC service thread 1-4) JBoss Web Services - Stack CXF Server 4.0.0.GA 20:22:46,466 WARN [jacorb.codeset] (MSC service thread 1-3) Warning - unknown codeset (MacRoman) - defaulting to ISO-8859-1 20:22:46,574 INFO [org.jboss.as.jacorb] (MSC service thread 1-3) CORBA ORB Service Started 20:22:46,659 WARN [org.jboss.as.messaging] (MSC service thread 1-4) JBAS011600: AIO wasn't located on this platform, it will fall back to using pure Java NIO. If your platform is Linux, install LibAIO to enable the AIO journal 20:22:46,690 WARN [org.infinispan.configuration.cache.EvictionConfigurationBuilder] (ServerService Thread Pool -- 35) ISPN000152: Passivation configured without a valid eviction policy. This could mean that the cache store will never get used unless code calls Cache.evict() manually. 20:22:46,692 WARN [org.infinispan.configuration.cache.LoaderConfigurationBuilder] (ServerService Thread Pool -- 35) ISPN000149: Fetch persistent state and purge on startup are both disabled, cache may contain stale entries on startup 20:22:46,786 WARN [org.infinispan.configuration.cache.EvictionConfigurationBuilder] (ServerService Thread Pool -- 35) ISPN000152: Passivation configured without a valid eviction policy. This could mean that the cache store will never get used unless code calls Cache.evict() manually. 20:22:46,787 WARN [org.infinispan.configuration.cache.LoaderConfigurationBuilder] (ServerService Thread Pool -- 35) ISPN000149: Fetch persistent state and purge on startup are both disabled, cache may contain stale entries on startup 20:22:46,856 WARN [org.infinispan.configuration.cache.LoaderConfigurationBuilder] (ServerService Thread Pool -- 35) ISPN000149: Fetch persistent state and purge on startup are both disabled, cache may contain stale entries on startup 20:22:46,867 WARN [org.infinispan.configuration.cache.LoaderConfigurationBuilder] (ServerService Thread Pool -- 35) ISPN000149: Fetch persistent state and purge on startup are both disabled, cache may contain stale entries on startup 20:22:46,913 INFO [org.jboss.as.modcluster] (MSC service thread 1-1) JBAS011704: Mod_cluster uses default load balancer provider 20:22:46,988 INFO [org.apache.tomcat.util.net.AprEndpoint] (MSC service thread 1-4) Failed to create poller with specified size of 32768 20:22:47,036 INFO [org.apache.tomcat.util.net.AprEndpoint] (MSC service thread 1-4) Failed to create poller with specified size of 32768 20:22:47,050 INFO [org.apache.tomcat.util.net.AprEndpoint] (MSC service thread 1-4) Failed to create poller with specified size of 16384 20:22:47,072 INFO [org.apache.coyote.http11.Http11AprProtocol] (MSC service thread 1-4) Starting Coyote HTTP/1.1 on http-localhost-127.0.0.1-8180 20:22:47,147 INFO [org.hornetq.core.server.impl.HornetQServerImpl] (MSC service thread 1-4) live server is starting with configuration HornetQ Configuration (clustered=false,backup=false,sharedStore=true,journalDirectory=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/target/jbossas-clustering-udp-1/standalone/data/messagingjournal,bindingsDirectory=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/target/jbossas-clustering-udp-1/standalone/data/messagingbindings,largeMessagesDirectory=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/target/jbossas-clustering-udp-1/standalone/data/messaginglargemessages,pagingDirectory=/Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/target/jbossas-clustering-udp-1/standalone/data/messagingpaging) 20:22:47,157 INFO [org.jboss.modcluster.ModClusterService] (MSC service thread 1-1) Initializing mod_cluster 1.2.0.Beta4 20:22:47,164 WARNING [org.hornetq.core.server.impl.HornetQServerImpl] (MSC service thread 1-4) Security risk! It has been detected that the cluster admin user and password have not been changed from the installation default. Please see the HornetQ user guide, cluster chapter, for instructions on how to do this. 20:22:47,226 INFO [org.jboss.modcluster.advertise.impl.AdvertiseListenerImpl] (MSC service thread 1-1) Listening to proxy advertisements on 224.0.1.105:23364 20:22:47,275 INFO [org.jboss.as.jacorb] (MSC service thread 1-3) CORBA Naming Service Started 20:22:47,768 INFO [org.jboss.as.connector.subsystems.datasources] (MSC service thread 1-1) JBAS010400: Bound data source [java:jboss/datasources/ExampleDS] 20:22:47,794 INFO [org.jboss.as.deployment.connector] (MSC service thread 1-2) JBAS010406: Registered connection factory java:/JmsXA 20:22:47,812 INFO [org.hornetq.ra.HornetQResourceAdapter] (MSC service thread 1-2) HornetQ resource adaptor started 20:22:47,812 INFO [org.jboss.as.connector.services.ResourceAdapterActivatorService$ResourceAdapterActivator] (MSC service thread 1-2) IJ020002: Deployed: file://RaActivatorhornetq-ra 20:22:47,815 INFO [org.jboss.as.deployment.connector] (MSC service thread 1-2) JBAS010401: Bound JCA ConnectionFactory [java:/JmsXA] 20:22:47,830 INFO [org.jboss.as.server.deployment] (MSC service thread 1-1) JBAS015876: Starting deployment of "singleton.war" 20:22:47,847 INFO [org.jboss.as.server.deployment.scanner] (MSC service thread 1-3) JBAS015012: Started FileSystemDeploymentService for directory /Users/bstansberry/dev/jbossas/bootstrap/jboss-as/testsuite/integration/clust/target/jbossas-clustering-udp-1/standalone/deployments 20:22:47,848 INFO [org.jboss.as.remoting] (MSC service thread 1-1) Listening on /127.0.0.1:10099 20:22:47,902 INFO [org.hornetq.core.remoting.impl.netty.NettyAcceptor] (MSC service thread 1-4) Started Netty Acceptor version 3.2.5.Final-a96d88c localhost:5545 for CORE protocol 20:22:47,924 INFO [org.hornetq.core.remoting.impl.netty.NettyAcceptor] (MSC service thread 1-4) Started Netty Acceptor version 3.2.5.Final-a96d88c localhost:5555 for CORE protocol 20:22:47,926 INFO [org.hornetq.core.server.impl.HornetQServerImpl] (MSC service thread 1-4) Server is now live 20:22:47,927 INFO [org.hornetq.core.server.impl.HornetQServerImpl] (MSC service thread 1-4) HornetQ Server version 2.2.10.Final (HQ_2_2_10_FINAL_AS7, 122) [9989b7ab-4956-11e1-b947-002608e4cba7]) started 20:22:47,929 INFO [org.hornetq.core.server.impl.HornetQServerImpl] (MSC service thread 1-4) trying to deploy queue jms.queue.testQueue 20:22:48,207 INFO [org.jboss.as.messaging] (MSC service thread 1-4) JBAS011601: Bound messaging object to jndi name java:/queue/test 20:22:48,208 INFO [org.hornetq.core.server.impl.HornetQServerImpl] (MSC service thread 1-2) trying to deploy queue jms.topic.testTopic 20:22:48,263 INFO [org.jboss.as.messaging] (MSC service thread 1-2) JBAS011601: Bound messaging object to jndi name java:/topic/test 20:22:48,294 INFO [org.jboss.as.messaging] (MSC service thread 1-3) JBAS011601: Bound messaging object to jndi name java:/ConnectionFactory 20:22:48,295 INFO [org.jboss.as.messaging] (MSC service thread 1-1) JBAS011601: Bound messaging object to jndi name java:/RemoteConnectionFactory 20:22:48,464 WARN [org.jboss.as.dependency.private] (MSC service thread 1-3) JBAS018567: Deployment "deployment.singleton.war" is using a private module ("org.jboss.iiop-client:main") which may be changed or removed in future versions without notice. 20:22:48,464 WARN [org.jboss.as.dependency.private] (MSC service thread 1-3) JBAS018567: Deployment "deployment.singleton.war" is using a private module ("org.jboss.as.ejb3:main") which may be changed or removed in future versions without notice. 20:22:48,465 WARN [org.jboss.as.dependency.private] (MSC service thread 1-3) JBAS018567: Deployment "deployment.singleton.war" is using a private module ("org.jboss.as.jacorb:main") which may be changed or removed in future versions without notice. 20:22:48,470 WARN [org.jboss.as.dependency.private] (MSC service thread 1-3) JBAS018567: Deployment "deployment.singleton.war" is using a private module ("org.jboss.as.ee:main") which may be changed or removed in future versions without notice. 20:22:48,499 WARN [org.jboss.as.dependency.private] (MSC service thread 1-3) JBAS018567: Deployment "deployment.singleton.war" is using a private module ("com.sun.jsf-impl:main") which may be changed or removed in future versions without notice. 20:22:48,499 WARN [org.jboss.as.dependency.private] (MSC service thread 1-3) JBAS018567: Deployment "deployment.singleton.war" is using a private module ("org.jboss.as.web:main") which may be changed or removed in future versions without notice. 20:22:48,499 WARN [org.jboss.as.dependency.private] (MSC service thread 1-3) JBAS018567: Deployment "deployment.singleton.war" is using a private module ("org.jboss.ws.spi:main") which may be changed or removed in future versions without notice. 20:22:48,499 WARN [org.jboss.as.dependency.private] (MSC service thread 1-3) JBAS018567: Deployment "deployment.singleton.war" is using a private module ("org.jboss.as.security:main") which may be changed or removed in future versions without notice. 20:22:48,504 WARN [org.jboss.as.dependency.private] (MSC service thread 1-3) JBAS018567: Deployment "deployment.singleton.war" is using a private module ("org.jacorb:main") which may be changed or removed in future versions without notice. 20:22:48,506 WARN [org.jboss.as.dependency.private] (MSC service thread 1-3) JBAS018567: Deployment "deployment.singleton.war" is using a private module ("org.jboss.as.clustering.singleton:main") which may be changed or removed in future versions without notice. 20:22:48,506 WARN [org.jboss.as.dependency.private] (MSC service thread 1-3) JBAS018567: Deployment "deployment.singleton.war" is using a private module ("org.jboss.as.server:main") which may be changed or removed in future versions without notice. 20:22:48,859 INFO [stdout] (MSC service thread 1-2) Service controller state is DOWN, waiting for transition to UP 20:22:49,466 WARNING [org.jgroups.protocols.UDP] (pool-5-thread-1) receive buffer of socket java.net.DatagramSocket@21a024c was set to 20MB, but the OS only allocated 65.51KB. This might lead to performance problems. Please set your max receive buffer in the OS correctly (e.g. net.core.rmem_max on Linux) 20:22:49,466 WARNING [org.jgroups.protocols.UDP] (pool-5-thread-1) receive buffer of socket java.net.MulticastSocket@798ccca7 was set to 25MB, but the OS only allocated 65.51KB. This might lead to performance problems. Please set your max receive buffer in the OS correctly (e.g. net.core.rmem_max on Linux) 20:22:49,479 INFO [stdout] (pool-5-thread-1) 20:22:49,479 INFO [stdout] (pool-5-thread-1) ------------------------------------------------------------------- 20:22:49,479 INFO [stdout] (pool-5-thread-1) GMS: address=node-udp-1/cluster, cluster=cluster, physical address=127.0.0.1:55300 20:22:49,480 INFO [stdout] (pool-5-thread-1) ------------------------------------------------------------------- 20:22:49,549 INFO [org.jboss.as.clustering.CoreGroupCommunicationService.lifecycle.cluster] (Incoming-13,null) JBAS010247: New cluster view for partition cluster (id: 4, delta: 1, merge: false) : [node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster] 20:22:49,552 INFO [org.infinispan.remoting.transport.jgroups.JGroupsTransport] (Incoming-13,null) ISPN000094: Received new cluster view: [node-udp-0/cluster|4] [node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster] 20:22:52,497 WARNING [org.jgroups.protocols.pbcast.GMS] (pool-5-thread-1) JOIN(node-udp-1/cluster) sent to node-udp-0/cluster timed out (after 3000 ms), retrying 20:22:54,549 WARNING [org.jgroups.protocols.pbcast.GMS] (ViewHandler,cluster,node-udp-0/cluster) node-udp-0/cluster: failed to collect all ACKs (expected=2) for view [node-udp-0/cluster|4] after 5000ms, missing ACKs from [node-udp-1/cluster] 20:22:54,589 ERROR [org.jboss.as.clustering.CoreGroupCommunicationService.cluster] (MSC service thread 1-3) JBAS010245: ViewAccepted failed: java.lang.IllegalStateException: JBAS010280: Address 1903f28a-f7d1-1488-62a3-03a2df0f5b62 not registered in transport layer at org.jboss.as.clustering.CoreGroupCommunicationService$ClusterNodeFactoryImpl.getClusterNode(CoreGroupCommunicationService.java:1581) [jboss-as-clustering-impl-7.1.0.Final-SNAPSHOT.jar:7.1.0.Final-SNAPSHOT] at org.jboss.as.clustering.CoreGroupCommunicationService.translateAddresses(CoreGroupCommunicationService.java:1211) [jboss-as-clustering-impl-7.1.0.Final-SNAPSHOT.jar:7.1.0.Final-SNAPSHOT] at org.jboss.as.clustering.CoreGroupCommunicationService$GroupView.(CoreGroupCommunicationService.java:1302) [jboss-as-clustering-impl-7.1.0.Final-SNAPSHOT.jar:7.1.0.Final-SNAPSHOT] at org.jboss.as.clustering.CoreGroupCommunicationService.processViewChange(CoreGroupCommunicationService.java:1111) [jboss-as-clustering-impl-7.1.0.Final-SNAPSHOT.jar:7.1.0.Final-SNAPSHOT] at org.jboss.as.clustering.CoreGroupCommunicationService$MembershipListenerImpl.viewAccepted(CoreGroupCommunicationService.java:1660) [jboss-as-clustering-impl-7.1.0.Final-SNAPSHOT.jar:7.1.0.Final-SNAPSHOT] at org.jboss.as.clustering.CoreGroupCommunicationService.startService(CoreGroupCommunicationService.java:905) [jboss-as-clustering-impl-7.1.0.Final-SNAPSHOT.jar:7.1.0.Final-SNAPSHOT] at org.jboss.as.clustering.CoreGroupCommunicationService.start(CoreGroupCommunicationService.java:806) [jboss-as-clustering-impl-7.1.0.Final-SNAPSHOT.jar:7.1.0.Final-SNAPSHOT] at org.jboss.as.clustering.CoreGroupCommunicationServiceService.start(CoreGroupCommunicationServiceService.java:81) [jboss-as-clustering-impl-7.1.0.Final-SNAPSHOT.jar:7.1.0.Final-SNAPSHOT] at org.jboss.msc.service.ServiceControllerImpl$StartTask.startService(ServiceControllerImpl.java:1824) [jboss-msc-1.0.1.GA.jar:1.0.1.GA] at org.jboss.msc.service.ServiceControllerImpl$StartTask.run(ServiceControllerImpl.java:1759) [jboss-msc-1.0.1.GA.jar:1.0.1.GA] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] 20:22:54,616 WARNING [org.jgroups.protocols.pbcast.GMS] (ViewHandler,cluster,node-udp-0/cluster) node-udp-1/cluster already present; returning existing view [node-udp-0/cluster|4] [node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster] 20:22:54,833 INFO [org.infinispan.remoting.transport.jgroups.JGroupsTransport] (pool-9-thread-1) ISPN000078: Starting JGroups Channel 20:22:54,834 WARNING [org.infinispan.remoting.transport.jgroups.CommandAwareRpcDispatcher] (pool-9-thread-1) Channel Muxer already has a default up handler installed (org.jboss.as.clustering.jgroups.ClassLoaderAwareUpHandler@45c803ae) but now it is being overridden 20:22:54,835 INFO [org.infinispan.remoting.transport.jgroups.JGroupsTransport] (pool-9-thread-1) ISPN000094: Received new cluster view: [node-udp-0/cluster|4] [node-udp-0/cluster, 1903f28a-f7d1-1488-62a3-03a2df0f5b62, node-udp-1/cluster] 20:22:54,835 INFO [org.infinispan.remoting.transport.jgroups.JGroupsTransport] (pool-9-thread-1) ISPN000079: Cache local address is node-udp-1/cluster, physical addresses are [127.0.0.1:55300] 20:22:54,839 INFO [org.infinispan.factories.GlobalComponentRegistry] (pool-9-thread-1) ISPN000128: Infinispan version: Infinispan 'Brahma' 5.1.0.FINAL 20:22:54,850 INFO [org.infinispan.jmx.CacheJmxRegistration] (pool-9-thread-1) ISPN000031: MBeans were successfully registered to the platform mbean server. 20:23:18,673 WARNING [org.jgroups.protocols.UDP] (OOB-19,null) null: no physical address for 1903f28a-f7d1-1488-62a3-03a2df0f5b62, dropping message 20:23:20,170 INFO [org.jboss.as.clustering.CoreGroupCommunicationService.lifecycle.cluster] (VERIFY_SUSPECT.TimerThread,cluster,node-udp-0/cluster) JBAS010254: Suspected member: node-udp-1/cluster 20:23:20,170 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=6, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=5, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: Suspected member: node-udp-1/cluster at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: Suspected member: node-udp-1/cluster at org.infinispan.remoting.transport.AbstractTransport.parseResponseAndAddToResponseList(AbstractTransport.java:93) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:461) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:23:20,174 INFO [org.jboss.as.clustering.CoreGroupCommunicationService.cluster] (VERIFY_SUSPECT.TimerThread,cluster,node-udp-1/cluster) JBAS010254: Suspected member: 1903f28a-f7d1-1488-62a3-03a2df0f5b62 20:23:20,224 INFO [org.jboss.as.clustering.CoreGroupCommunicationService.lifecycle.cluster] (Incoming-15,null) JBAS010247: New cluster view for partition cluster (id: 5, delta: -1, merge: false) : [node-udp-0/cluster, node-udp-1/cluster] 20:23:20,224 INFO [org.infinispan.remoting.transport.jgroups.JGroupsTransport] (Incoming-15,null) ISPN000094: Received new cluster view: [node-udp-0/cluster|5] [node-udp-0/cluster, node-udp-1/cluster] 20:23:20,226 INFO [org.jboss.as.clustering.CoreGroupCommunicationService.cluster] (Incoming-6,null) JBAS010206: Number of cluster members: 2 20:23:20,226 INFO [org.infinispan.remoting.transport.jgroups.JGroupsTransport] (Incoming-6,null) ISPN000094: Received new cluster view: [node-udp-0/cluster|5] [node-udp-0/cluster, node-udp-1/cluster] 20:23:20,226 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=8, members=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=7, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=8, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=7, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=8, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=7, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:23:21,227 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=10, members=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=9, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=10, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=9, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=10, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=9, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:23:22,227 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=12, members=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=11, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=12, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=11, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=12, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=11, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:23:23,227 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=14, members=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=13, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=14, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=13, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=14, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=13, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:23:24,227 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=16, members=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=15, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=16, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=15, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=16, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=15, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:23:25,229 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=18, members=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=17, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=18, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=17, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=18, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=17, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:23:26,229 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=20, members=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=19, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=20, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=19, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=20, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=19, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:23:27,229 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=22, members=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=21, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=22, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=21, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=22, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=21, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:23:28,229 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=24, members=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=23, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=24, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=23, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=24, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=23, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:23:29,229 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=26, members=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=25, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=26, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=25, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=26, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=25, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:23:30,231 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=28, members=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=27, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=28, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=27, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=28, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=27, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:23:31,230 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=30, members=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=29, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=30, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=29, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=30, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=29, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:23:32,230 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=32, members=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=31, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=32, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=31, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=32, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=31, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:23:33,230 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=34, members=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=33, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=34, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=33, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=34, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=33, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:23:34,230 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=36, members=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=35, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=36, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=35, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=36, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=35, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:23:35,232 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=38, members=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=37, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=38, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=37, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=38, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=37, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:23:36,232 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=40, members=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=39, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=40, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=39, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=40, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=39, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:23:36,604 ERROR [org.infinispan.interceptors.InvocationContextInterceptor] (AsynchViewChangeHandler Thread) ISPN000136: Execution error: org.infinispan.statetransfer.StateTransferInProgressException: Timed out waiting for the state transfer lock, state transfer in progress for view 40 at org.infinispan.interceptors.StateTransferLockInterceptor.signalStateTransferInProgress(StateTransferLockInterceptor.java:199) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.interceptors.StateTransferLockInterceptor.visitPrepareCommand(StateTransferLockInterceptor.java:80) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.commands.tx.PrepareCommand.acceptVisitor(PrepareCommand.java:131) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.interceptors.base.CommandInterceptor.invokeNextInterceptor(CommandInterceptor.java:116) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.interceptors.base.CommandInterceptor.handleDefault(CommandInterceptor.java:130) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.commands.AbstractVisitor.visitPrepareCommand(AbstractVisitor.java:113) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.commands.tx.PrepareCommand.acceptVisitor(PrepareCommand.java:131) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.interceptors.base.CommandInterceptor.invokeNextInterceptor(CommandInterceptor.java:116) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.interceptors.InvocationContextInterceptor.handleAll(InvocationContextInterceptor.java:130) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.interceptors.InvocationContextInterceptor.handleDefault(InvocationContextInterceptor.java:89) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.commands.AbstractVisitor.visitPrepareCommand(AbstractVisitor.java:113) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.commands.tx.PrepareCommand.acceptVisitor(PrepareCommand.java:131) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.interceptors.base.CommandInterceptor.invokeNextInterceptor(CommandInterceptor.java:116) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.interceptors.base.CommandInterceptor.handleDefault(CommandInterceptor.java:130) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.commands.AbstractVisitor.visitPrepareCommand(AbstractVisitor.java:113) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.commands.tx.PrepareCommand.acceptVisitor(PrepareCommand.java:131) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.interceptors.base.CommandInterceptor.invokeNextInterceptor(CommandInterceptor.java:116) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.interceptors.BatchingInterceptor.handleDefault(BatchingInterceptor.java:86) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.commands.AbstractVisitor.visitPrepareCommand(AbstractVisitor.java:113) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.commands.tx.PrepareCommand.acceptVisitor(PrepareCommand.java:131) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.interceptors.base.CommandInterceptor.invokeNextInterceptor(CommandInterceptor.java:116) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.interceptors.base.CommandInterceptor.handleDefault(CommandInterceptor.java:130) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.jboss.as.clustering.infinispan.DefaultEmbeddedCacheManager$ClassLoaderAwareCommandInterceptor.handleDefault(DefaultEmbeddedCacheManager.java:410) [jboss-as-clustering-infinispan-7.1.0.Final-SNAPSHOT.jar:7.1.0.Final-SNAPSHOT] at org.infinispan.commands.AbstractVisitor.visitPrepareCommand(AbstractVisitor.java:113) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.commands.tx.PrepareCommand.acceptVisitor(PrepareCommand.java:131) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.interceptors.InterceptorChain.invoke(InterceptorChain.java:345) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.transaction.TransactionCoordinator.prepare(TransactionCoordinator.java:140) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.transaction.TransactionCoordinator.prepare(TransactionCoordinator.java:122) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.transaction.synchronization.SynchronizationAdapter.beforeCompletion(SynchronizationAdapter.java:70) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.transaction.tm.DummyTransaction.notifyBeforeCompletion(DummyTransaction.java:223) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.transaction.tm.DummyTransaction.runPrepare(DummyTransaction.java:234) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.transaction.tm.DummyTransaction.commit(DummyTransaction.java:85) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.transaction.tm.DummyBaseTransactionManager.commit(DummyBaseTransactionManager.java:100) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.batch.BatchContainer.resolveTransaction(BatchContainer.java:123) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.batch.BatchContainer.endBatch(BatchContainer.java:105) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.batch.BatchContainer.endBatch(BatchContainer.java:86) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.CacheImpl.endBatch(CacheImpl.java:577) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.AbstractDelegatingCache.endBatch(AbstractDelegatingCache.java:75) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.AbstractDelegatingCache.endBatch(AbstractDelegatingCache.java:75) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.AbstractDelegatingCache.endBatch(AbstractDelegatingCache.java:75) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.jboss.as.clustering.infinispan.invoker.BatchOperation.invoke(BatchOperation.java:26) [jboss-as-clustering-infinispan-7.1.0.Final-SNAPSHOT.jar:7.1.0.Final-SNAPSHOT] at org.jboss.as.clustering.service.ServiceProviderRegistryService.invoke(ServiceProviderRegistryService.java:187) [jboss-as-clustering-service-7.1.0.Final-SNAPSHOT.jar:7.1.0.Final-SNAPSHOT] at org.jboss.as.clustering.service.ServiceProviderRegistryService.purgeDeadMembers(ServiceProviderRegistryService.java:167) [jboss-as-clustering-service-7.1.0.Final-SNAPSHOT.jar:7.1.0.Final-SNAPSHOT] at org.jboss.as.clustering.service.ServiceProviderRegistryService.membershipChanged(ServiceProviderRegistryService.java:140) [jboss-as-clustering-service-7.1.0.Final-SNAPSHOT.jar:7.1.0.Final-SNAPSHOT] at org.jboss.as.clustering.CoreGroupCommunicationService.notifyListeners(CoreGroupCommunicationService.java:1263) [jboss-as-clustering-impl-7.1.0.Final-SNAPSHOT.jar:7.1.0.Final-SNAPSHOT] at org.jboss.as.clustering.CoreGroupCommunicationService$ViewChangeEventProcessor.processEvent(CoreGroupCommunicationService.java:1513) [jboss-as-clustering-impl-7.1.0.Final-SNAPSHOT.jar:7.1.0.Final-SNAPSHOT] at org.jboss.as.clustering.AsynchEventHandler.run(AsynchEventHandler.java:96) [jboss-as-clustering-impl-7.1.0.Final-SNAPSHOT.jar:7.1.0.Final-SNAPSHOT] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] 20:23:36,607 ERROR [org.infinispan.transaction.TransactionCoordinator] (AsynchViewChangeHandler Thread) Error while processing prepare: org.infinispan.statetransfer.StateTransferInProgressException: Timed out waiting for the state transfer lock, state transfer in progress for view 40 at org.infinispan.interceptors.StateTransferLockInterceptor.signalStateTransferInProgress(StateTransferLockInterceptor.java:199) at org.infinispan.interceptors.StateTransferLockInterceptor.visitPrepareCommand(StateTransferLockInterceptor.java:80) at org.infinispan.commands.tx.PrepareCommand.acceptVisitor(PrepareCommand.java:131) at org.infinispan.interceptors.base.CommandInterceptor.invokeNextInterceptor(CommandInterceptor.java:116) at org.infinispan.interceptors.base.CommandInterceptor.handleDefault(CommandInterceptor.java:130) at org.infinispan.commands.AbstractVisitor.visitPrepareCommand(AbstractVisitor.java:113) at org.infinispan.commands.tx.PrepareCommand.acceptVisitor(PrepareCommand.java:131) at org.infinispan.interceptors.base.CommandInterceptor.invokeNextInterceptor(CommandInterceptor.java:116) at org.infinispan.interceptors.InvocationContextInterceptor.handleAll(InvocationContextInterceptor.java:130) at org.infinispan.interceptors.InvocationContextInterceptor.handleDefault(InvocationContextInterceptor.java:89) at org.infinispan.commands.AbstractVisitor.visitPrepareCommand(AbstractVisitor.java:113) at org.infinispan.commands.tx.PrepareCommand.acceptVisitor(PrepareCommand.java:131) at org.infinispan.interceptors.base.CommandInterceptor.invokeNextInterceptor(CommandInterceptor.java:116) at org.infinispan.interceptors.base.CommandInterceptor.handleDefault(CommandInterceptor.java:130) at org.infinispan.commands.AbstractVisitor.visitPrepareCommand(AbstractVisitor.java:113) at org.infinispan.commands.tx.PrepareCommand.acceptVisitor(PrepareCommand.java:131) at org.infinispan.interceptors.base.CommandInterceptor.invokeNextInterceptor(CommandInterceptor.java:116) at org.infinispan.interceptors.BatchingInterceptor.handleDefault(BatchingInterceptor.java:86) at org.infinispan.commands.AbstractVisitor.visitPrepareCommand(AbstractVisitor.java:113) at org.infinispan.commands.tx.PrepareCommand.acceptVisitor(PrepareCommand.java:131) at org.infinispan.interceptors.base.CommandInterceptor.invokeNextInterceptor(CommandInterceptor.java:116) at org.infinispan.interceptors.base.CommandInterceptor.handleDefault(CommandInterceptor.java:130) at org.jboss.as.clustering.infinispan.DefaultEmbeddedCacheManager$ClassLoaderAwareCommandInterceptor.handleDefault(DefaultEmbeddedCacheManager.java:410) at org.infinispan.commands.AbstractVisitor.visitPrepareCommand(AbstractVisitor.java:113) at org.infinispan.commands.tx.PrepareCommand.acceptVisitor(PrepareCommand.java:131) at org.infinispan.interceptors.InterceptorChain.invoke(InterceptorChain.java:345) at org.infinispan.transaction.TransactionCoordinator.prepare(TransactionCoordinator.java:140) at org.infinispan.transaction.TransactionCoordinator.prepare(TransactionCoordinator.java:122) at org.infinispan.transaction.synchronization.SynchronizationAdapter.beforeCompletion(SynchronizationAdapter.java:70) at org.infinispan.transaction.tm.DummyTransaction.notifyBeforeCompletion(DummyTransaction.java:223) at org.infinispan.transaction.tm.DummyTransaction.runPrepare(DummyTransaction.java:234) at org.infinispan.transaction.tm.DummyTransaction.commit(DummyTransaction.java:85) at org.infinispan.transaction.tm.DummyBaseTransactionManager.commit(DummyBaseTransactionManager.java:100) at org.infinispan.batch.BatchContainer.resolveTransaction(BatchContainer.java:123) at org.infinispan.batch.BatchContainer.endBatch(BatchContainer.java:105) at org.infinispan.batch.BatchContainer.endBatch(BatchContainer.java:86) at org.infinispan.CacheImpl.endBatch(CacheImpl.java:577) at org.infinispan.AbstractDelegatingCache.endBatch(AbstractDelegatingCache.java:75) at org.infinispan.AbstractDelegatingCache.endBatch(AbstractDelegatingCache.java:75) at org.infinispan.AbstractDelegatingCache.endBatch(AbstractDelegatingCache.java:75) at org.jboss.as.clustering.infinispan.invoker.BatchOperation.invoke(BatchOperation.java:26) at org.jboss.as.clustering.service.ServiceProviderRegistryService.invoke(ServiceProviderRegistryService.java:187) at org.jboss.as.clustering.service.ServiceProviderRegistryService.purgeDeadMembers(ServiceProviderRegistryService.java:167) at org.jboss.as.clustering.service.ServiceProviderRegistryService.membershipChanged(ServiceProviderRegistryService.java:140) at org.jboss.as.clustering.CoreGroupCommunicationService.notifyListeners(CoreGroupCommunicationService.java:1263) [jboss-as-clustering-impl-7.1.0.Final-SNAPSHOT.jar:7.1.0.Final-SNAPSHOT] at org.jboss.as.clustering.CoreGroupCommunicationService$ViewChangeEventProcessor.processEvent(CoreGroupCommunicationService.java:1513) [jboss-as-clustering-impl-7.1.0.Final-SNAPSHOT.jar:7.1.0.Final-SNAPSHOT] at org.jboss.as.clustering.AsynchEventHandler.run(AsynchEventHandler.java:96) [jboss-as-clustering-impl-7.1.0.Final-SNAPSHOT.jar:7.1.0.Final-SNAPSHOT] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] 20:23:36,614 ERROR [org.infinispan.transaction.tm.DummyTransaction] (AsynchViewChangeHandler Thread) ISPN000109: beforeCompletion() failed for SynchronizationAdapter{localTransaction=LocalTransaction{remoteLockedNodes=null, isMarkedForRollback=false, transaction=DummyTransaction{xid=DummyXid{id=3}, status=1}, lockedKeys=null, backupKeyLocks=null, viewId=1} org.infinispan.transaction.synchronization.SyncLocalTransaction@3} org.infinispan.transaction.synchronization.SynchronizationAdapter@22: org.infinispan.CacheException: Could not prepare. at org.infinispan.transaction.synchronization.SynchronizationAdapter.beforeCompletion(SynchronizationAdapter.java:72) at org.infinispan.transaction.tm.DummyTransaction.notifyBeforeCompletion(DummyTransaction.java:223) at org.infinispan.transaction.tm.DummyTransaction.runPrepare(DummyTransaction.java:234) at org.infinispan.transaction.tm.DummyTransaction.commit(DummyTransaction.java:85) at org.infinispan.transaction.tm.DummyBaseTransactionManager.commit(DummyBaseTransactionManager.java:100) at org.infinispan.batch.BatchContainer.resolveTransaction(BatchContainer.java:123) at org.infinispan.batch.BatchContainer.endBatch(BatchContainer.java:105) at org.infinispan.batch.BatchContainer.endBatch(BatchContainer.java:86) at org.infinispan.CacheImpl.endBatch(CacheImpl.java:577) at org.infinispan.AbstractDelegatingCache.endBatch(AbstractDelegatingCache.java:75) at org.infinispan.AbstractDelegatingCache.endBatch(AbstractDelegatingCache.java:75) at org.infinispan.AbstractDelegatingCache.endBatch(AbstractDelegatingCache.java:75) at org.jboss.as.clustering.infinispan.invoker.BatchOperation.invoke(BatchOperation.java:26) at org.jboss.as.clustering.service.ServiceProviderRegistryService.invoke(ServiceProviderRegistryService.java:187) at org.jboss.as.clustering.service.ServiceProviderRegistryService.purgeDeadMembers(ServiceProviderRegistryService.java:167) at org.jboss.as.clustering.service.ServiceProviderRegistryService.membershipChanged(ServiceProviderRegistryService.java:140) at org.jboss.as.clustering.CoreGroupCommunicationService.notifyListeners(CoreGroupCommunicationService.java:1263) [jboss-as-clustering-impl-7.1.0.Final-SNAPSHOT.jar:7.1.0.Final-SNAPSHOT] at org.jboss.as.clustering.CoreGroupCommunicationService$ViewChangeEventProcessor.processEvent(CoreGroupCommunicationService.java:1513) [jboss-as-clustering-impl-7.1.0.Final-SNAPSHOT.jar:7.1.0.Final-SNAPSHOT] at org.jboss.as.clustering.AsynchEventHandler.run(AsynchEventHandler.java:96) [jboss-as-clustering-impl-7.1.0.Final-SNAPSHOT.jar:7.1.0.Final-SNAPSHOT] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: javax.transaction.xa.XAException at org.infinispan.transaction.TransactionCoordinator.prepare(TransactionCoordinator.java:160) at org.infinispan.transaction.TransactionCoordinator.prepare(TransactionCoordinator.java:122) at org.infinispan.transaction.synchronization.SynchronizationAdapter.beforeCompletion(SynchronizationAdapter.java:70) ... 19 more 20:23:37,231 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=42, members=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=41, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=42, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=41, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=42, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=41, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:23:38,232 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=44, members=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=43, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=44, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=43, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=44, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=43, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:23:39,232 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=46, members=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=45, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=46, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=45, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=46, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=45, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:23:40,232 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=48, members=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=47, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=48, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=47, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=48, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=47, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:23:41,232 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=50, members=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=49, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=50, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=49, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=50, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=49, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:23:42,233 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=52, members=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=51, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=52, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=51, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=52, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=51, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:23:43,233 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=54, members=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=53, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=54, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=53, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=54, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=53, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:23:44,234 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=56, members=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=55, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=56, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=55, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=56, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=55, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:23:45,233 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=58, members=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=57, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=58, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=57, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=58, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=57, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:23:46,234 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=60, members=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=59, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=60, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=59, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=60, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=59, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:23:47,234 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=62, members=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=61, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=62, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=61, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=62, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=61, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:23:48,234 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=64, members=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=63, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=64, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=63, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=64, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=63, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:23:49,235 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=66, members=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=65, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=66, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=65, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=66, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=65, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:23:50,235 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=68, members=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=67, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=68, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=67, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=68, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=67, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:23:51,237 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=70, members=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=69, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=70, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=69, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=70, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=69, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:23:52,236 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=72, members=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=71, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=72, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=71, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=72, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=71, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:23:53,236 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=74, members=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=73, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=74, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=73, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=74, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=73, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:23:54,236 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=76, members=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=75, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=76, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=75, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=76, newMembers=[node-udp-0/cluster, node-udp-1/cluster, node-udp-1/cluster], oldViewId=75, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:23:54,898 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=78, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=77, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=78, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=77, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=78, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=77, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:23:55,256 INFO [org.jboss.as.clustering.CoreGroupCommunicationService.lifecycle.cluster] (Incoming-18,null) JBAS010247: New cluster view for partition cluster (id: 6, delta: -1, merge: false) : [node-udp-0/cluster] 20:23:55,256 INFO [org.infinispan.remoting.transport.jgroups.JGroupsTransport] (Incoming-18,null) ISPN000094: Received new cluster view: [node-udp-0/cluster|6] [node-udp-0/cluster] 20:23:55,260 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=80, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=79, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=80, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=79, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=80, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=79, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:23:55,392 INFO [stdout] (MSC service thread 1-3) Service controller state is UP, waiting for transition to REMOVED 20:23:55,394 INFO [stdout] (MSC service thread 1-3) Service controller state is STOPPING, waiting for transition to REMOVED 20:23:56,262 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=82, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=81, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=82, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=81, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=82, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=81, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:23:57,261 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=84, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=83, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=84, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=83, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=84, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=83, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:23:58,262 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=86, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=85, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=86, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=85, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=86, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=85, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:23:59,262 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=88, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=87, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=88, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=87, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=88, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=87, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:00,262 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=90, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=89, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=90, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=89, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=90, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=89, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:01,263 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=92, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=91, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=92, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=91, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=92, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=91, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:02,262 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=94, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=93, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=94, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=93, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=94, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=93, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:03,263 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=96, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=95, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=96, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=95, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=96, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=95, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:04,263 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=98, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=97, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=98, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=97, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=98, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=97, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:05,263 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=100, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=99, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=100, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=99, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=100, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=99, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:06,263 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=102, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=101, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=102, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=101, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=102, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=101, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:07,264 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=104, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=103, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=104, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=103, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=104, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=103, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:08,264 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=106, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=105, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=106, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=105, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=106, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=105, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:09,264 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=108, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=107, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=108, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=107, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=108, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=107, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:10,264 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=110, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=109, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=110, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=109, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=110, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=109, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:11,264 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=112, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=111, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=112, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=111, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=112, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=111, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:12,264 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=114, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=113, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=114, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=113, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=114, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=113, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:13,265 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=116, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=115, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=116, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=115, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=116, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=115, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:14,265 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=118, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=117, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=118, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=117, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=118, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=117, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:15,265 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=120, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=119, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=120, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=119, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=120, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=119, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:16,265 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=122, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=121, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=122, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=121, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=122, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=121, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:17,266 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=124, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=123, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=124, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=123, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=124, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=123, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:18,266 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=126, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=125, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=126, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=125, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=126, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=125, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:19,266 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=128, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=127, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=128, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=127, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=128, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=127, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:20,266 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=130, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=129, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=130, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=129, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=130, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=129, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:21,267 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=132, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=131, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=132, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=131, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=132, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=131, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:22,267 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=134, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=133, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=134, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=133, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=134, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=133, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:23,267 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=136, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=135, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=136, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=135, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=136, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=135, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:24,268 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=138, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=137, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=138, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=137, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=138, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=137, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:25,268 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=140, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=139, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=140, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=139, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=140, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=139, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:26,269 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=142, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=141, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=142, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=141, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=142, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=141, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:27,268 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=144, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=143, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=144, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=143, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=144, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=143, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:28,268 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=146, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=145, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=146, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=145, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=146, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=145, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:29,269 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=148, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=147, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=148, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=147, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=148, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=147, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:30,269 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=150, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=149, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=150, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=149, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=150, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=149, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:31,269 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=152, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=151, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=152, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=151, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=152, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=151, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:32,270 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=154, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=153, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=154, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=153, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=154, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=153, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:33,270 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=156, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=155, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=156, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=155, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=156, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=155, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:34,270 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=158, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=157, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=158, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=157, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=158, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=157, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:35,270 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=160, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=159, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=160, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=159, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=160, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=159, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:36,271 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=162, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=161, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=162, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=161, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=162, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=161, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:37,271 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=164, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=163, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=164, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=163, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=164, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=163, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:38,271 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=166, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=165, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=166, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=165, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=166, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=165, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:39,271 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=168, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=167, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=168, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=167, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=168, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=167, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:40,272 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=170, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=169, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=170, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=169, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=170, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=169, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:41,272 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=172, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=171, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=172, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=171, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=172, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=171, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:42,272 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=174, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=173, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=174, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=173, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=174, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=173, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:43,272 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=176, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=175, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=176, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=175, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=176, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=175, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:44,272 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=178, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=177, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=178, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=177, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=178, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=177, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:45,272 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=180, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=179, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=180, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=179, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=180, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=179, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:46,278 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=182, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=181, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=182, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=181, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=182, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=181, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:47,279 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=184, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=183, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=184, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=183, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=184, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=183, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:48,277 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=186, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=185, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=186, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=185, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=186, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=185, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:49,277 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=188, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=187, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=188, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=187, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=188, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=187, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:50,278 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=190, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=189, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=190, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=189, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=190, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=189, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:51,278 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=192, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=191, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=192, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=191, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=192, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=191, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:52,279 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=194, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=193, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=194, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=193, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=194, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=193, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:53,279 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=196, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=195, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=196, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=195, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=196, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=195, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:54,279 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=198, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=197, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=198, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=197, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=198, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=197, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:55,279 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=200, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=199, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=200, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=199, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=200, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=199, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:55,396 ERROR [org.infinispan.interceptors.InvocationContextInterceptor] (pool-26-thread-2) ISPN000136: Execution error: org.infinispan.statetransfer.StateTransferInProgressException: Timed out waiting for the state transfer lock, state transfer in progress for view 200 at org.infinispan.interceptors.StateTransferLockInterceptor.signalStateTransferInProgress(StateTransferLockInterceptor.java:199) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.interceptors.StateTransferLockInterceptor.visitPrepareCommand(StateTransferLockInterceptor.java:80) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.commands.tx.PrepareCommand.acceptVisitor(PrepareCommand.java:131) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.interceptors.base.CommandInterceptor.invokeNextInterceptor(CommandInterceptor.java:116) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.interceptors.base.CommandInterceptor.handleDefault(CommandInterceptor.java:130) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.commands.AbstractVisitor.visitPrepareCommand(AbstractVisitor.java:113) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.commands.tx.PrepareCommand.acceptVisitor(PrepareCommand.java:131) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.interceptors.base.CommandInterceptor.invokeNextInterceptor(CommandInterceptor.java:116) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.interceptors.InvocationContextInterceptor.handleAll(InvocationContextInterceptor.java:130) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.interceptors.InvocationContextInterceptor.handleDefault(InvocationContextInterceptor.java:89) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.commands.AbstractVisitor.visitPrepareCommand(AbstractVisitor.java:113) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.commands.tx.PrepareCommand.acceptVisitor(PrepareCommand.java:131) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.interceptors.base.CommandInterceptor.invokeNextInterceptor(CommandInterceptor.java:116) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.interceptors.base.CommandInterceptor.handleDefault(CommandInterceptor.java:130) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.commands.AbstractVisitor.visitPrepareCommand(AbstractVisitor.java:113) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.commands.tx.PrepareCommand.acceptVisitor(PrepareCommand.java:131) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.interceptors.base.CommandInterceptor.invokeNextInterceptor(CommandInterceptor.java:116) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.interceptors.BatchingInterceptor.handleDefault(BatchingInterceptor.java:86) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.commands.AbstractVisitor.visitPrepareCommand(AbstractVisitor.java:113) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.commands.tx.PrepareCommand.acceptVisitor(PrepareCommand.java:131) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.interceptors.base.CommandInterceptor.invokeNextInterceptor(CommandInterceptor.java:116) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.interceptors.base.CommandInterceptor.handleDefault(CommandInterceptor.java:130) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.jboss.as.clustering.infinispan.DefaultEmbeddedCacheManager$ClassLoaderAwareCommandInterceptor.handleDefault(DefaultEmbeddedCacheManager.java:410) [jboss-as-clustering-infinispan-7.1.0.Final-SNAPSHOT.jar:7.1.0.Final-SNAPSHOT] at org.infinispan.commands.AbstractVisitor.visitPrepareCommand(AbstractVisitor.java:113) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.commands.tx.PrepareCommand.acceptVisitor(PrepareCommand.java:131) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.interceptors.InterceptorChain.invoke(InterceptorChain.java:345) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.transaction.TransactionCoordinator.prepare(TransactionCoordinator.java:140) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.transaction.TransactionCoordinator.prepare(TransactionCoordinator.java:122) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.transaction.synchronization.SynchronizationAdapter.beforeCompletion(SynchronizationAdapter.java:70) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.transaction.tm.DummyTransaction.notifyBeforeCompletion(DummyTransaction.java:223) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.transaction.tm.DummyTransaction.runPrepare(DummyTransaction.java:234) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.transaction.tm.DummyTransaction.commit(DummyTransaction.java:85) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.transaction.tm.DummyBaseTransactionManager.commit(DummyBaseTransactionManager.java:100) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.batch.BatchContainer.resolveTransaction(BatchContainer.java:123) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.batch.BatchContainer.endBatch(BatchContainer.java:105) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.batch.BatchContainer.endBatch(BatchContainer.java:86) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.CacheImpl.endBatch(CacheImpl.java:577) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.AbstractDelegatingCache.endBatch(AbstractDelegatingCache.java:75) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.AbstractDelegatingCache.endBatch(AbstractDelegatingCache.java:75) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.AbstractDelegatingCache.endBatch(AbstractDelegatingCache.java:75) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.jboss.as.clustering.infinispan.invoker.BatchOperation.invoke(BatchOperation.java:26) [jboss-as-clustering-infinispan-7.1.0.Final-SNAPSHOT.jar:7.1.0.Final-SNAPSHOT] at org.jboss.as.clustering.service.ServiceProviderRegistryService.invoke(ServiceProviderRegistryService.java:187) [jboss-as-clustering-service-7.1.0.Final-SNAPSHOT.jar:7.1.0.Final-SNAPSHOT] at org.jboss.as.clustering.service.ServiceProviderRegistryService.unregister(ServiceProviderRegistryService.java:108) [jboss-as-clustering-service-7.1.0.Final-SNAPSHOT.jar:7.1.0.Final-SNAPSHOT] at org.jboss.as.clustering.singleton.SingletonService.stop(SingletonService.java:109) at org.jboss.as.clustering.AsynchronousService$2.run(AsynchronousService.java:64) at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] 20:24:55,402 ERROR [org.infinispan.transaction.TransactionCoordinator] (pool-26-thread-2) Error while processing prepare: org.infinispan.statetransfer.StateTransferInProgressException: Timed out waiting for the state transfer lock, state transfer in progress for view 200 at org.infinispan.interceptors.StateTransferLockInterceptor.signalStateTransferInProgress(StateTransferLockInterceptor.java:199) at org.infinispan.interceptors.StateTransferLockInterceptor.visitPrepareCommand(StateTransferLockInterceptor.java:80) at org.infinispan.commands.tx.PrepareCommand.acceptVisitor(PrepareCommand.java:131) at org.infinispan.interceptors.base.CommandInterceptor.invokeNextInterceptor(CommandInterceptor.java:116) at org.infinispan.interceptors.base.CommandInterceptor.handleDefault(CommandInterceptor.java:130) at org.infinispan.commands.AbstractVisitor.visitPrepareCommand(AbstractVisitor.java:113) at org.infinispan.commands.tx.PrepareCommand.acceptVisitor(PrepareCommand.java:131) at org.infinispan.interceptors.base.CommandInterceptor.invokeNextInterceptor(CommandInterceptor.java:116) at org.infinispan.interceptors.InvocationContextInterceptor.handleAll(InvocationContextInterceptor.java:130) at org.infinispan.interceptors.InvocationContextInterceptor.handleDefault(InvocationContextInterceptor.java:89) at org.infinispan.commands.AbstractVisitor.visitPrepareCommand(AbstractVisitor.java:113) at org.infinispan.commands.tx.PrepareCommand.acceptVisitor(PrepareCommand.java:131) at org.infinispan.interceptors.base.CommandInterceptor.invokeNextInterceptor(CommandInterceptor.java:116) at org.infinispan.interceptors.base.CommandInterceptor.handleDefault(CommandInterceptor.java:130) at org.infinispan.commands.AbstractVisitor.visitPrepareCommand(AbstractVisitor.java:113) at org.infinispan.commands.tx.PrepareCommand.acceptVisitor(PrepareCommand.java:131) at org.infinispan.interceptors.base.CommandInterceptor.invokeNextInterceptor(CommandInterceptor.java:116) at org.infinispan.interceptors.BatchingInterceptor.handleDefault(BatchingInterceptor.java:86) at org.infinispan.commands.AbstractVisitor.visitPrepareCommand(AbstractVisitor.java:113) at org.infinispan.commands.tx.PrepareCommand.acceptVisitor(PrepareCommand.java:131) at org.infinispan.interceptors.base.CommandInterceptor.invokeNextInterceptor(CommandInterceptor.java:116) at org.infinispan.interceptors.base.CommandInterceptor.handleDefault(CommandInterceptor.java:130) at org.jboss.as.clustering.infinispan.DefaultEmbeddedCacheManager$ClassLoaderAwareCommandInterceptor.handleDefault(DefaultEmbeddedCacheManager.java:410) at org.infinispan.commands.AbstractVisitor.visitPrepareCommand(AbstractVisitor.java:113) at org.infinispan.commands.tx.PrepareCommand.acceptVisitor(PrepareCommand.java:131) at org.infinispan.interceptors.InterceptorChain.invoke(InterceptorChain.java:345) at org.infinispan.transaction.TransactionCoordinator.prepare(TransactionCoordinator.java:140) at org.infinispan.transaction.TransactionCoordinator.prepare(TransactionCoordinator.java:122) at org.infinispan.transaction.synchronization.SynchronizationAdapter.beforeCompletion(SynchronizationAdapter.java:70) at org.infinispan.transaction.tm.DummyTransaction.notifyBeforeCompletion(DummyTransaction.java:223) at org.infinispan.transaction.tm.DummyTransaction.runPrepare(DummyTransaction.java:234) at org.infinispan.transaction.tm.DummyTransaction.commit(DummyTransaction.java:85) at org.infinispan.transaction.tm.DummyBaseTransactionManager.commit(DummyBaseTransactionManager.java:100) at org.infinispan.batch.BatchContainer.resolveTransaction(BatchContainer.java:123) at org.infinispan.batch.BatchContainer.endBatch(BatchContainer.java:105) at org.infinispan.batch.BatchContainer.endBatch(BatchContainer.java:86) at org.infinispan.CacheImpl.endBatch(CacheImpl.java:577) at org.infinispan.AbstractDelegatingCache.endBatch(AbstractDelegatingCache.java:75) at org.infinispan.AbstractDelegatingCache.endBatch(AbstractDelegatingCache.java:75) at org.infinispan.AbstractDelegatingCache.endBatch(AbstractDelegatingCache.java:75) at org.jboss.as.clustering.infinispan.invoker.BatchOperation.invoke(BatchOperation.java:26) at org.jboss.as.clustering.service.ServiceProviderRegistryService.invoke(ServiceProviderRegistryService.java:187) [jboss-as-clustering-service-7.1.0.Final-SNAPSHOT.jar:7.1.0.Final-SNAPSHOT] at org.jboss.as.clustering.service.ServiceProviderRegistryService.unregister(ServiceProviderRegistryService.java:108) [jboss-as-clustering-service-7.1.0.Final-SNAPSHOT.jar:7.1.0.Final-SNAPSHOT] at org.jboss.as.clustering.singleton.SingletonService.stop(SingletonService.java:109) [jboss-as-clustering-singleton-7.1.0.Final-SNAPSHOT.jar:7.1.0.Final-SNAPSHOT] at org.jboss.as.clustering.AsynchronousService$2.run(AsynchronousService.java:64) [jboss-as-clustering-common-7.1.0.Final-SNAPSHOT.jar:7.1.0.Final-SNAPSHOT] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] 20:24:55,406 ERROR [org.infinispan.transaction.tm.DummyTransaction] (pool-26-thread-2) ISPN000109: beforeCompletion() failed for SynchronizationAdapter{localTransaction=LocalTransaction{remoteLockedNodes=null, isMarkedForRollback=false, transaction=DummyTransaction{xid=DummyXid{id=4}, status=1}, lockedKeys=null, backupKeyLocks=null, viewId=6} org.infinispan.transaction.synchronization.SyncLocalTransaction@4} org.infinispan.transaction.synchronization.SynchronizationAdapter@23: org.infinispan.CacheException: Could not prepare. at org.infinispan.transaction.synchronization.SynchronizationAdapter.beforeCompletion(SynchronizationAdapter.java:72) at org.infinispan.transaction.tm.DummyTransaction.notifyBeforeCompletion(DummyTransaction.java:223) at org.infinispan.transaction.tm.DummyTransaction.runPrepare(DummyTransaction.java:234) at org.infinispan.transaction.tm.DummyTransaction.commit(DummyTransaction.java:85) at org.infinispan.transaction.tm.DummyBaseTransactionManager.commit(DummyBaseTransactionManager.java:100) at org.infinispan.batch.BatchContainer.resolveTransaction(BatchContainer.java:123) at org.infinispan.batch.BatchContainer.endBatch(BatchContainer.java:105) at org.infinispan.batch.BatchContainer.endBatch(BatchContainer.java:86) at org.infinispan.CacheImpl.endBatch(CacheImpl.java:577) at org.infinispan.AbstractDelegatingCache.endBatch(AbstractDelegatingCache.java:75) at org.infinispan.AbstractDelegatingCache.endBatch(AbstractDelegatingCache.java:75) at org.infinispan.AbstractDelegatingCache.endBatch(AbstractDelegatingCache.java:75) at org.jboss.as.clustering.infinispan.invoker.BatchOperation.invoke(BatchOperation.java:26) at org.jboss.as.clustering.service.ServiceProviderRegistryService.invoke(ServiceProviderRegistryService.java:187) [jboss-as-clustering-service-7.1.0.Final-SNAPSHOT.jar:7.1.0.Final-SNAPSHOT] at org.jboss.as.clustering.service.ServiceProviderRegistryService.unregister(ServiceProviderRegistryService.java:108) [jboss-as-clustering-service-7.1.0.Final-SNAPSHOT.jar:7.1.0.Final-SNAPSHOT] at org.jboss.as.clustering.singleton.SingletonService.stop(SingletonService.java:109) [jboss-as-clustering-singleton-7.1.0.Final-SNAPSHOT.jar:7.1.0.Final-SNAPSHOT] at org.jboss.as.clustering.AsynchronousService$2.run(AsynchronousService.java:64) [jboss-as-clustering-common-7.1.0.Final-SNAPSHOT.jar:7.1.0.Final-SNAPSHOT] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: javax.transaction.xa.XAException at org.infinispan.transaction.TransactionCoordinator.prepare(TransactionCoordinator.java:160) at org.infinispan.transaction.TransactionCoordinator.prepare(TransactionCoordinator.java:122) at org.infinispan.transaction.synchronization.SynchronizationAdapter.beforeCompletion(SynchronizationAdapter.java:70) ... 19 more 20:24:56,280 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=202, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=201, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=202, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=201, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=202, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=201, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:57,280 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=204, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=203, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=204, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=203, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=204, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=203, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:58,280 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=206, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=205, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=206, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=205, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=206, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=205, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:24:59,280 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=208, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=207, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=208, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=207, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=208, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=207, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:00,280 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=210, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=209, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=210, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=209, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=210, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=209, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:01,281 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=212, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=211, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=212, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=211, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=212, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=211, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:02,281 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=214, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=213, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=214, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=213, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=214, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=213, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:03,281 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=216, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=215, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=216, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=215, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=216, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=215, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:04,281 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=218, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=217, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=218, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=217, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=218, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=217, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:05,281 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=220, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=219, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=220, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=219, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=220, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=219, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:06,281 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=222, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=221, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=222, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=221, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=222, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=221, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:07,282 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=224, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=223, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=224, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=223, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=224, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=223, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:08,282 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=226, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=225, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=226, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=225, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=226, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=225, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:09,282 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=228, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=227, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=228, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=227, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=228, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=227, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:10,283 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=230, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=229, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=230, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=229, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=230, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=229, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:11,283 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=232, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=231, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=232, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=231, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=232, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=231, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:12,284 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=234, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=233, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=234, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=233, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=234, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=233, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:13,284 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=236, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=235, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=236, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=235, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=236, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=235, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:14,284 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=238, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=237, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=238, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=237, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=238, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=237, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:15,284 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=240, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=239, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=240, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=239, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=240, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=239, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:16,284 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=242, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=241, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=242, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=241, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=242, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=241, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:17,285 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=244, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=243, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=244, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=243, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=244, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=243, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:18,285 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=246, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=245, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=246, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=245, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=246, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=245, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:19,288 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=248, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=247, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=248, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=247, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=248, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=247, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:20,285 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=250, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=249, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=250, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=249, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=250, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=249, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:21,286 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=252, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=251, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=252, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=251, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=252, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=251, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:22,286 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=254, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=253, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=254, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=253, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=254, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=253, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:23,286 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=256, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=255, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=256, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=255, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=256, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=255, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:24,288 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=258, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=257, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=258, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=257, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=258, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=257, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:25,287 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=260, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=259, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=260, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=259, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=260, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=259, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:26,288 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=262, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=261, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=262, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=261, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=262, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=261, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:27,288 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=264, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=263, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=264, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=263, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=264, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=263, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:28,289 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=266, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=265, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=266, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=265, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=266, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=265, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:29,289 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=268, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=267, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=268, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=267, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=268, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=267, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:30,289 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=270, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=269, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=270, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=269, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=270, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=269, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:31,292 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=272, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=271, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=272, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=271, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=272, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=271, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:32,290 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=274, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=273, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=274, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=273, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=274, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=273, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:33,290 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=276, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=275, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=276, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=275, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=276, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=275, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:34,289 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=278, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=277, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=278, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=277, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=278, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=277, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:35,290 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=280, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=279, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=280, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=279, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=280, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=279, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:36,290 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=282, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=281, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=282, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=281, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=282, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=281, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:37,290 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=284, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=283, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=284, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=283, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=284, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=283, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:38,290 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=286, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=285, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=286, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=285, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=286, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=285, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:39,291 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=288, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=287, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=288, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=287, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=288, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=287, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:40,291 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=290, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=289, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=290, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=289, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=290, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=289, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:41,291 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=292, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=291, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=292, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=291, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=292, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=291, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:42,291 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=294, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=293, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=294, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=293, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=294, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=293, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:43,292 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=296, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=295, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=296, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=295, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=296, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=295, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:44,292 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=298, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=297, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=298, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=297, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=298, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=297, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:45,292 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=300, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=299, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=300, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=299, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=300, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=299, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:46,294 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=302, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=301, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=302, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=301, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=302, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=301, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:47,293 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=304, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=303, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=304, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=303, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=304, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=303, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:48,294 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=306, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=305, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=306, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=305, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=306, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=305, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:49,294 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=308, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=307, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=308, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=307, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=308, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=307, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:50,294 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=310, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=309, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=310, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=309, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=310, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=309, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:51,295 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=312, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=311, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=312, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=311, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=312, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=311, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:52,295 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=314, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=313, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=314, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=313, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=314, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=313, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:53,295 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=316, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=315, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=316, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=315, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=316, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=315, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:54,295 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=318, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=317, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=318, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=317, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=318, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=317, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:55,296 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=320, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=319, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=320, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=319, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=320, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=319, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:56,296 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=322, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=321, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=322, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=321, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=322, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=321, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:57,297 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=324, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=323, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=324, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=323, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=324, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=323, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:58,297 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=326, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=325, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=326, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=325, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=326, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=325, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:25:59,297 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=328, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=327, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=328, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=327, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=328, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=327, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:00,297 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=330, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=329, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=330, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=329, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=330, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=329, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:01,297 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=332, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=331, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=332, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=331, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=332, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=331, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:02,299 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=334, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=333, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=334, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=333, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=334, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=333, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:03,300 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=336, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=335, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=336, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=335, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=336, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=335, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:04,299 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=338, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=337, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=338, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=337, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=338, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=337, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:05,299 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=340, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=339, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=340, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=339, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=340, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=339, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:06,300 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=342, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=341, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=342, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=341, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=342, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=341, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:07,301 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=344, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=343, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=344, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=343, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=344, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=343, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:08,301 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=346, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=345, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=346, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=345, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=346, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=345, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:09,301 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=348, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=347, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=348, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=347, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=348, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=347, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:10,301 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=350, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=349, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=350, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=349, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=350, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=349, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:11,301 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=352, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=351, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=352, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=351, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=352, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=351, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:12,302 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=354, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=353, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=354, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=353, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=354, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=353, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:13,302 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=356, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=355, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=356, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=355, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=356, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=355, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:14,302 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=358, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=357, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=358, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=357, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=358, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=357, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:15,302 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=360, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=359, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=360, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=359, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=360, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=359, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:16,302 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=362, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=361, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=362, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=361, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=362, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=361, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:17,303 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=364, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=363, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=364, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=363, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=364, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=363, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:18,302 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=366, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=365, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=366, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=365, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=366, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=365, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:19,303 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=368, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=367, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=368, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=367, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=368, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=367, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:20,303 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=370, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=369, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=370, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=369, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=370, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=369, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:21,303 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=372, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=371, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=372, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=371, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=372, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=371, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:22,303 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=374, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=373, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=374, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=373, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=374, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=373, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:23,305 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=376, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=375, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=376, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=375, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=376, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=375, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:24,305 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=378, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=377, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=378, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=377, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=378, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=377, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:25,305 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=380, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=379, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=380, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=379, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=380, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=379, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:26,305 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=382, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=381, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=382, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=381, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=382, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=381, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:27,305 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=384, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=383, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=384, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=383, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=384, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=383, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:28,305 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=386, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=385, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=386, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=385, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=386, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=385, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:29,308 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=388, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=387, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=388, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=387, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=388, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=387, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:30,309 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=390, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=389, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=390, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=389, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=390, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=389, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:31,309 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=392, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=391, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=392, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=391, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=392, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=391, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:32,309 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=394, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=393, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=394, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=393, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=394, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=393, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:33,310 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=396, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=395, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=396, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=395, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=396, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=395, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:34,320 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=398, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=397, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=398, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=397, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=398, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=397, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:35,319 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=400, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=399, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=400, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=399, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=400, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=399, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:36,320 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=402, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=401, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=402, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=401, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=402, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=401, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:37,320 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=404, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=403, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=404, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=403, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=404, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=403, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:38,322 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=406, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=405, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=406, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=405, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=406, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=405, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:39,322 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=408, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=407, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=408, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=407, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=408, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=407, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:40,323 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=410, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=409, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=410, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=409, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=410, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=409, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:41,323 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=412, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=411, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=412, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=411, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=412, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=411, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:42,323 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=414, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=413, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=414, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=413, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=414, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=413, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:43,323 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=416, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=415, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=416, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=415, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=416, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=415, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:44,323 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=418, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=417, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=418, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=417, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=418, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=417, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:45,324 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=420, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=419, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=420, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=419, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=420, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=419, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:46,324 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=422, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=421, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=422, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=421, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=422, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=421, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:47,324 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=424, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=423, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=424, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=423, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=424, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=423, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:48,324 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=426, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=425, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=426, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=425, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=426, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=425, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:49,324 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=428, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=427, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=428, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=427, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=428, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=427, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:50,325 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=430, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=429, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=430, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=429, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=430, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=429, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:51,325 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=432, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=431, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=432, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=431, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=432, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=431, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:52,325 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=434, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=433, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=434, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=433, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=434, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=433, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:53,325 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=436, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=435, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=436, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=435, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=436, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=435, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:54,325 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=438, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=437, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=438, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=437, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=438, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=437, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:55,326 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=440, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=439, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=440, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=439, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=440, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=439, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:56,326 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=442, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=441, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=442, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=441, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=442, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=441, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:57,327 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=444, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=443, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=444, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=443, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=444, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=443, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:58,327 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=446, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=445, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=446, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=445, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=446, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=445, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:26:59,329 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=448, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=447, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=448, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=447, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=448, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=447, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:00,328 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=450, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=449, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=450, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=449, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=450, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=449, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:01,329 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=452, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=451, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=452, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=451, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=452, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=451, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:02,329 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=454, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=453, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=454, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=453, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=454, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=453, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:03,329 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=456, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=455, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=456, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=455, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=456, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=455, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:04,330 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=458, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=457, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=458, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=457, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=458, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=457, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:05,329 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=460, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=459, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=460, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=459, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=460, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=459, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:06,330 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=462, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=461, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=462, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=461, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=462, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=461, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:07,331 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=464, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=463, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=464, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=463, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=464, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=463, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:08,334 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=466, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=465, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=466, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=465, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=466, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=465, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:09,337 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=468, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=467, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=468, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=467, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=468, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=467, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:10,337 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=470, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=469, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=470, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=469, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=470, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=469, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:11,337 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=472, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=471, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=472, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=471, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=472, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=471, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:12,337 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=474, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=473, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=474, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=473, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=474, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=473, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:13,337 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=476, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=475, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=476, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=475, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=476, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=475, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:14,338 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=478, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=477, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=478, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=477, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=478, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=477, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:15,337 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=480, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=479, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=480, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=479, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=480, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=479, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:16,338 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=482, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=481, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=482, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=481, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=482, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=481, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:17,338 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=484, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=483, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=484, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=483, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=484, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=483, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:18,339 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=486, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=485, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=486, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=485, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=486, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=485, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:19,338 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=488, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=487, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=488, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=487, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=488, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=487, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:20,339 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=490, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=489, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=490, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=489, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=490, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=489, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:21,339 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=492, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=491, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=492, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=491, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=492, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=491, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:22,340 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=494, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=493, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=494, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=493, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=494, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=493, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:23,339 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=496, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=495, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=496, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=495, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=496, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=495, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:24,340 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=498, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=497, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=498, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=497, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=498, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=497, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:25,340 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=500, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=499, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=500, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=499, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=500, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=499, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:26,340 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=502, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=501, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=502, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=501, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=502, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=501, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:27,341 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=504, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=503, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=504, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=503, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=504, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=503, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:28,341 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=506, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=505, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=506, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=505, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=506, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=505, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:29,341 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=508, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=507, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=508, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=507, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=508, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=507, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:30,341 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=510, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=509, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=510, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=509, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=510, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=509, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:31,341 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=512, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=511, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=512, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=511, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=512, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=511, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:32,341 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=514, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=513, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=514, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=513, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=514, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=513, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:33,342 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=516, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=515, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=516, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=515, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=516, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=515, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:34,342 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=518, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=517, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=518, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=517, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=518, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=517, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:35,343 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=520, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=519, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=520, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=519, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=520, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=519, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:36,343 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=522, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=521, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=522, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=521, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=522, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=521, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:37,343 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=524, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=523, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=524, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=523, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=524, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=523, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:38,343 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=526, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=525, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=526, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=525, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=526, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=525, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:39,343 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=528, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=527, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=528, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=527, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=528, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=527, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:40,344 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=530, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=529, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=530, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=529, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=530, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=529, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:41,344 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=532, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=531, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=532, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=531, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=532, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=531, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:42,346 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=534, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=533, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=534, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=533, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=534, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=533, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:43,344 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=536, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=535, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=536, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=535, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=536, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=535, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:44,344 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=538, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=537, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=538, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=537, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=538, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=537, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:45,345 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=540, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=539, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=540, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=539, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=540, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=539, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:46,345 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=542, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=541, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=542, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=541, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=542, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=541, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:47,346 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=544, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=543, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=544, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=543, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=544, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=543, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:48,346 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=546, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=545, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=546, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=545, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=546, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=545, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:49,346 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=548, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=547, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=548, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=547, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=548, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=547, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:50,347 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=550, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=549, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=550, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=549, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=550, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=549, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:51,347 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=552, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=551, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=552, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=551, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=552, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=551, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:52,347 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=554, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=553, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=554, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=553, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=554, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=553, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:53,347 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=556, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=555, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=556, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=555, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=556, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=555, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:54,347 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=558, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=557, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=558, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=557, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=558, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=557, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:55,348 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=560, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=559, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=560, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=559, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=560, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=559, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:56,347 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=562, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=561, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=562, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=561, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=562, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=561, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:57,348 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=564, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=563, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=564, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=563, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=564, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=563, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:58,348 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=566, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=565, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=566, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=565, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=566, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=565, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:27:59,349 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=568, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=567, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=568, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=567, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=568, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=567, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:00,348 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=570, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=569, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=570, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=569, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=570, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=569, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:01,349 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=572, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=571, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=572, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=571, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=572, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=571, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:02,349 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=574, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=573, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=574, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=573, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=574, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=573, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:03,349 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=576, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=575, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=576, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=575, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=576, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=575, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:04,349 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=578, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=577, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=578, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=577, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=578, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=577, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:05,350 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=580, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=579, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=580, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=579, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=580, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=579, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:06,352 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=582, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=581, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=582, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=581, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=582, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=581, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:07,352 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=584, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=583, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=584, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=583, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=584, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=583, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:08,352 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=586, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=585, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=586, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=585, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=586, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=585, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:09,352 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=588, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=587, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=588, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=587, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=588, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=587, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:10,352 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=590, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=589, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=590, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=589, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=590, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=589, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:11,352 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=592, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=591, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=592, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=591, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=592, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=591, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:12,353 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=594, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=593, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=594, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=593, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=594, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=593, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:13,353 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=596, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=595, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=596, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=595, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=596, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=595, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:14,354 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=598, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=597, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=598, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=597, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=598, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=597, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:15,353 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=600, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=599, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=600, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=599, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=600, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=599, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:16,354 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=602, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=601, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=602, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=601, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=602, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=601, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:17,354 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=604, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=603, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=604, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=603, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=604, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=603, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:18,355 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=606, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=605, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=606, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=605, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=606, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=605, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:19,355 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=608, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=607, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=608, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=607, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=608, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=607, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:20,355 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=610, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=609, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=610, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=609, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=610, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=609, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:21,356 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=612, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=611, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=612, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=611, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=612, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=611, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:22,355 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=614, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=613, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=614, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=613, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=614, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=613, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:23,355 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=616, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=615, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=616, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=615, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=616, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=615, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:24,356 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=618, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=617, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=618, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=617, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=618, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=617, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:25,356 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=620, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=619, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=620, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=619, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=620, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=619, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:26,357 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=622, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=621, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=622, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=621, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=622, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=621, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:27,357 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=624, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=623, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=624, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=623, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=624, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=623, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:28,357 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=626, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=625, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=626, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=625, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=626, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=625, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:29,357 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=628, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=627, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=628, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=627, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=628, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=627, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:30,357 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=630, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=629, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=630, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=629, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=630, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=629, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:31,357 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=632, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=631, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=632, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=631, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=632, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=631, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:32,358 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=634, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=633, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=634, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=633, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=634, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=633, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:33,358 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=636, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=635, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=636, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=635, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=636, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=635, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:34,358 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=638, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=637, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=638, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=637, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=638, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=637, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:35,358 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=640, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=639, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=640, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=639, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=640, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=639, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:36,359 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=642, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=641, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=642, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=641, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=642, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=641, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:37,359 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=644, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=643, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=644, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=643, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=644, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=643, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:38,359 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=646, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=645, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=646, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=645, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=646, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=645, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:39,359 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=648, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=647, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=648, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=647, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=648, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=647, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:40,360 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=650, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=649, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=650, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=649, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=650, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=649, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:41,360 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=652, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=651, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=652, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=651, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=652, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=651, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:42,360 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=654, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=653, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=654, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=653, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=654, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=653, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:43,360 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=656, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=655, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=656, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=655, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=656, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=655, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:44,360 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=658, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=657, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=658, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=657, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=658, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=657, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:45,361 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=660, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=659, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=660, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=659, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=660, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=659, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:46,361 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=662, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=661, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=662, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=661, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=662, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=661, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:47,361 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=664, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=663, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=664, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=663, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=664, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=663, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:48,361 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=666, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=665, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=666, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=665, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=666, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=665, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:49,362 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=668, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=667, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=668, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=667, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=668, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=667, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:50,363 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=670, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=669, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=670, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=669, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=670, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=669, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:51,362 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=672, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=671, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=672, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=671, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=672, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=671, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:52,362 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=674, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=673, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=674, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=673, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=674, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=673, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:53,363 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=676, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=675, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=676, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=675, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=676, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=675, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:54,363 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=678, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=677, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=678, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=677, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=678, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=677, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:55,363 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=680, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=679, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=680, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=679, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=680, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=679, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:56,364 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=682, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=681, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=682, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=681, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=682, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=681, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:57,364 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=684, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=683, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=684, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=683, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=684, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=683, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:58,364 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=686, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=685, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=686, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=685, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=686, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=685, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:28:59,364 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=688, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=687, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=688, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=687, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=688, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=687, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:00,364 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=690, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=689, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=690, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=689, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=690, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=689, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:01,365 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=692, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=691, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=692, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=691, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=692, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=691, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:02,365 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=694, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=693, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=694, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=693, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=694, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=693, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:03,365 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=696, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=695, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=696, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=695, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=696, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=695, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:04,366 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=698, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=697, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=698, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=697, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=698, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=697, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:05,366 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=700, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=699, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=700, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=699, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=700, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=699, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:06,366 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=702, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=701, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=702, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=701, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=702, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=701, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:07,366 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=704, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=703, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=704, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=703, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=704, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=703, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:08,366 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=706, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=705, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=706, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=705, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=706, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=705, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:09,367 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=708, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=707, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=708, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=707, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=708, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=707, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:10,367 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=710, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=709, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=710, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=709, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=710, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=709, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:11,367 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=712, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=711, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=712, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=711, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=712, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=711, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:12,368 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=714, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=713, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=714, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=713, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=714, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=713, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:13,368 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=716, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=715, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=716, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=715, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=716, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=715, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:14,368 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=718, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=717, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=718, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=717, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=718, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=717, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:15,368 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=720, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=719, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=720, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=719, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=720, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=719, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:16,369 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=722, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=721, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=722, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=721, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=722, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=721, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:17,368 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=724, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=723, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=724, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=723, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=724, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=723, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:18,369 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=726, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=725, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=726, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=725, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=726, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=725, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:19,369 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=728, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=727, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=728, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=727, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=728, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=727, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:20,369 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=730, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=729, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=730, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=729, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=730, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=729, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:21,369 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=732, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=731, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=732, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=731, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=732, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=731, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:22,369 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=734, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=733, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=734, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=733, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=734, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=733, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:23,370 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=736, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=735, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=736, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=735, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=736, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=735, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:24,370 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=738, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=737, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=738, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=737, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=738, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=737, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:25,370 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=740, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=739, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=740, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=739, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=740, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=739, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:26,370 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=742, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=741, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=742, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=741, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=742, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=741, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:27,373 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=744, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=743, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=744, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=743, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=744, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=743, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:28,371 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=746, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=745, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=746, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=745, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=746, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=745, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:29,371 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=748, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=747, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=748, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=747, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=748, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=747, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:30,371 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=750, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=749, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=750, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=749, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=750, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=749, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:31,371 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=752, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=751, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=752, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=751, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=752, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=751, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:32,372 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=754, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=753, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=754, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=753, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=754, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=753, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:33,372 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=756, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=755, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=756, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=755, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=756, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=755, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:34,373 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=758, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=757, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=758, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=757, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=758, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=757, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:35,372 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=760, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=759, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=760, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=759, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=760, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=759, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:36,372 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=762, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=761, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=762, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=761, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=762, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=761, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:37,372 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=764, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=763, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=764, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=763, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=764, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=763, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:38,373 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=766, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=765, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=766, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=765, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=766, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=765, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:39,373 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=768, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=767, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=768, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=767, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=768, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=767, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:40,373 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=770, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=769, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=770, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=769, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=770, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=769, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:41,373 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=772, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=771, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=772, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=771, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=772, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=771, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:42,373 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=774, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=773, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=774, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=773, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=774, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=773, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:43,374 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=776, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=775, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=776, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=775, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=776, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=775, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:44,374 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=778, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=777, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=778, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=777, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=778, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=777, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:45,374 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=780, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=779, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=780, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=779, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=780, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=779, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:46,377 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=782, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=781, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=782, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=781, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=782, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=781, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:47,377 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=784, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=783, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=784, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=783, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=784, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=783, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:48,377 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=786, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=785, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=786, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=785, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=786, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=785, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:49,377 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=788, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=787, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=788, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=787, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=788, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=787, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:50,377 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=790, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=789, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=790, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=789, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=790, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=789, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:51,378 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=792, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=791, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=792, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=791, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=792, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=791, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:52,378 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=794, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=793, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=794, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=793, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=794, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=793, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:53,378 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=796, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=795, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=796, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=795, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=796, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=795, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:54,379 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=798, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=797, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=798, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=797, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=798, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=797, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:55,378 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=800, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=799, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=800, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=799, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=800, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=799, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:56,399 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=802, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=801, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=802, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=801, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=802, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=801, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:57,394 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=804, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=803, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=804, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=803, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=804, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=803, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:58,394 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=806, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=805, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=806, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=805, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=806, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=805, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:29:59,395 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=808, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=807, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=808, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=807, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=808, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=807, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:00,395 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=810, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=809, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=810, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=809, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=810, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=809, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:01,395 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=812, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=811, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=812, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=811, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=812, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=811, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:02,395 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=814, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=813, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=814, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=813, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=814, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=813, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:03,395 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=816, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=815, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=816, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=815, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=816, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=815, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:04,396 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=818, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=817, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=818, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=817, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=818, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=817, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:05,396 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=820, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=819, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=820, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=819, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=820, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=819, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:06,396 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=822, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=821, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=822, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=821, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=822, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=821, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:07,396 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=824, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=823, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=824, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=823, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=824, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=823, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:08,397 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=826, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=825, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=826, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=825, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=826, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=825, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:09,397 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=828, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=827, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=828, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=827, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=828, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=827, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:10,397 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=830, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=829, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=830, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=829, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=830, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=829, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:11,397 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=832, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=831, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=832, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=831, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=832, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=831, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:12,398 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=834, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=833, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=834, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=833, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=834, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=833, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:13,398 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=836, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=835, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=836, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=835, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=836, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=835, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:14,398 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=838, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=837, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=838, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=837, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=838, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=837, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:15,398 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=840, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=839, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=840, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=839, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=840, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=839, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:16,398 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=842, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=841, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=842, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=841, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=842, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=841, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:17,399 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=844, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=843, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=844, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=843, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=844, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=843, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:18,399 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=846, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=845, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=846, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=845, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=846, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=845, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:19,401 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=848, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=847, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=848, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=847, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=848, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=847, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:20,401 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=850, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=849, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=850, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=849, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=850, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=849, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:21,401 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=852, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=851, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=852, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=851, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=852, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=851, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:22,401 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=854, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=853, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=854, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=853, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=854, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=853, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:23,402 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=856, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=855, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=856, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=855, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=856, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=855, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:24,402 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=858, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=857, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=858, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=857, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=858, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=857, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:25,402 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=860, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=859, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=860, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=859, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=860, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=859, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:26,402 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=862, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=861, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=862, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=861, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=862, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=861, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:27,402 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=864, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=863, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=864, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=863, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=864, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=863, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:28,403 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=866, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=865, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=866, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=865, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=866, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=865, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:29,403 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=868, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=867, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=868, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=867, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=868, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=867, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:30,403 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=870, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=869, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=870, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=869, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=870, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=869, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:31,403 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=872, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=871, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=872, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=871, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=872, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=871, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:32,404 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=874, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=873, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=874, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=873, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=874, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=873, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:33,404 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=876, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=875, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=876, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=875, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=876, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=875, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:34,404 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=878, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=877, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=878, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=877, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=878, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=877, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:35,404 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=880, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=879, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=880, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=879, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=880, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=879, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:36,405 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=882, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=881, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=882, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=881, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=882, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=881, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:37,405 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=884, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=883, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=884, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=883, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=884, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=883, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:38,405 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=886, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=885, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=886, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=885, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=886, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=885, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:39,405 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=888, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=887, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=888, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=887, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=888, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=887, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:40,406 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=890, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=889, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=890, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=889, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=890, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=889, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:41,406 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=892, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=891, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=892, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=891, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=892, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=891, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:42,406 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=894, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=893, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=894, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=893, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=894, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=893, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:43,406 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=896, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=895, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=896, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=895, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=896, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=895, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:44,406 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=898, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=897, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=898, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=897, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=898, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=897, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:45,407 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=900, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=899, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=900, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=899, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=900, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=899, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:46,407 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=902, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=901, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=902, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=901, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=902, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=901, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:47,407 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=904, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=903, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=904, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=903, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=904, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=903, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:48,407 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=906, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=905, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=906, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=905, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=906, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=905, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:49,408 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=908, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=907, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=908, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=907, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=908, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=907, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:50,408 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=910, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=909, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=910, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=909, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=910, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=909, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:51,408 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=912, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=911, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=912, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=911, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=912, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=911, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:52,408 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=914, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=913, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=914, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=913, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=914, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=913, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:53,409 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=916, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=915, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=916, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=915, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=916, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=915, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:54,409 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=918, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=917, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=918, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=917, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=918, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=917, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:55,409 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=920, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=919, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=920, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=919, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=920, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=919, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:56,410 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=922, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=921, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=922, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=921, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=922, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=921, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:57,410 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=924, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=923, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=924, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=923, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=924, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=923, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:58,410 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=926, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=925, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=926, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=925, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=926, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=925, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:30:59,410 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=928, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=927, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=928, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=927, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=928, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=927, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:00,410 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=930, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=929, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=930, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=929, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=930, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=929, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:01,410 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=932, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=931, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=932, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=931, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=932, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=931, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:02,410 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=934, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=933, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=934, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=933, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=934, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=933, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:03,411 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=936, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=935, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=936, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=935, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=936, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=935, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:04,411 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=938, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=937, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=938, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=937, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=938, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=937, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:05,411 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=940, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=939, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=940, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=939, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=940, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=939, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:06,411 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=942, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=941, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=942, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=941, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=942, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=941, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:07,412 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=944, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=943, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=944, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=943, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=944, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=943, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:08,412 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=946, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=945, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=946, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=945, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=946, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=945, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:09,412 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=948, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=947, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=948, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=947, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=948, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=947, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:10,412 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=950, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=949, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=950, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=949, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=950, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=949, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:11,412 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=952, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=951, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=952, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=951, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=952, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=951, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:12,412 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=954, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=953, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=954, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=953, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=954, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=953, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:13,412 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=956, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=955, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=956, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=955, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=956, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=955, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:14,413 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=958, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=957, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=958, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=957, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=958, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=957, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:15,413 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=960, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=959, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=960, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=959, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=960, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=959, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:16,413 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=962, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=961, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=962, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=961, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=962, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=961, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:17,414 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=964, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=963, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=964, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=963, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=964, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=963, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:18,414 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=966, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=965, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=966, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=965, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=966, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=965, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:19,414 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=968, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=967, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=968, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=967, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=968, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=967, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:20,414 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=970, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=969, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=970, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=969, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=970, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=969, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:21,414 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=972, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=971, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=972, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=971, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=972, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=971, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:22,414 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=974, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=973, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=974, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=973, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=974, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=973, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:23,415 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=976, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=975, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=976, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=975, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=976, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=975, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:24,415 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=978, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=977, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=978, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=977, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=978, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=977, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:25,415 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=980, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=979, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=980, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=979, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=980, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=979, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:26,415 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=982, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=981, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=982, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=981, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=982, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=981, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:27,416 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=984, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=983, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=984, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=983, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=984, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=983, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:28,416 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=986, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=985, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=986, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=985, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=986, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=985, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:29,416 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=988, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=987, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=988, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=987, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=988, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=987, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:30,416 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=990, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=989, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=990, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=989, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=990, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=989, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:31,416 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=992, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=991, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=992, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=991, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=992, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=991, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:32,417 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=994, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=993, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=994, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=993, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=994, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=993, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:33,417 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=996, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=995, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=996, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=995, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=996, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=995, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:34,417 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=998, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=997, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=998, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=997, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=998, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=997, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:35,417 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1000, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=999, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1000, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=999, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1000, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=999, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:36,418 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1002, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1001, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1002, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1001, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1002, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1001, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:37,419 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1004, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1003, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1004, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1003, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1004, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1003, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:38,419 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1006, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1005, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1006, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1005, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1006, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1005, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:39,419 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1008, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1007, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1008, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1007, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1008, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1007, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:40,420 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1010, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1009, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1010, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1009, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1010, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1009, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:41,420 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1012, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1011, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1012, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1011, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1012, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1011, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:42,420 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1014, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1013, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1014, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1013, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1014, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1013, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:43,421 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1016, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1015, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1016, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1015, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1016, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1015, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:44,420 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1018, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1017, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1018, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1017, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1018, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1017, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:45,420 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1020, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1019, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1020, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1019, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1020, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1019, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:46,421 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1022, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1021, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1022, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1021, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1022, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1021, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:47,421 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1024, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1023, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1024, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1023, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1024, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1023, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:48,421 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1026, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1025, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1026, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1025, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1026, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1025, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:49,421 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1028, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1027, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1028, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1027, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1028, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1027, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:50,421 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1030, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1029, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1030, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1029, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1030, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1029, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:51,422 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1032, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1031, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1032, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1031, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1032, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1031, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:52,422 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1034, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1033, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1034, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1033, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1034, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1033, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:53,422 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1036, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1035, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1036, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1035, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1036, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1035, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:54,422 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1038, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1037, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1038, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1037, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1038, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1037, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:55,422 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1040, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1039, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1040, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1039, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1040, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1039, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:56,423 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1042, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1041, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1042, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1041, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1042, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1041, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:57,423 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1044, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1043, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1044, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1043, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1044, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1043, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:58,423 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1046, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1045, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1046, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1045, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1046, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1045, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:31:59,423 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1048, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1047, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1048, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1047, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1048, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1047, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:00,423 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1050, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1049, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1050, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1049, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1050, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1049, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:01,424 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1052, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1051, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1052, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1051, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1052, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1051, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:02,424 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1054, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1053, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1054, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1053, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1054, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1053, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:03,424 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1056, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1055, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1056, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1055, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1056, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1055, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:04,424 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1058, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1057, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1058, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1057, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1058, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1057, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:05,424 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1060, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1059, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1060, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1059, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1060, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1059, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:06,425 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1062, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1061, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1062, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1061, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1062, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1061, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:07,426 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1064, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1063, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1064, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1063, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1064, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1063, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:08,425 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1066, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1065, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1066, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1065, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1066, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1065, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:09,425 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1068, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1067, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1068, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1067, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1068, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1067, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:10,426 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1070, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1069, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1070, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1069, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1070, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1069, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:11,426 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1072, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1071, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1072, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1071, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1072, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1071, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:12,426 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1074, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1073, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1074, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1073, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1074, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1073, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:13,427 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1076, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1075, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1076, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1075, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1076, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1075, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:14,427 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1078, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1077, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1078, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1077, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1078, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1077, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:15,427 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1080, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1079, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1080, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1079, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1080, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1079, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:16,428 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1082, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1081, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1082, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1081, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1082, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1081, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:17,427 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1084, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1083, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1084, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1083, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1084, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1083, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:18,427 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1086, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1085, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1086, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1085, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1086, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1085, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:19,428 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1088, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1087, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1088, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1087, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1088, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1087, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:20,428 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1090, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1089, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1090, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1089, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1090, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1089, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:21,428 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1092, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1091, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1092, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1091, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1092, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1091, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:22,428 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1094, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1093, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1094, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1093, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1094, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1093, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:23,429 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1096, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1095, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1096, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1095, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1096, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1095, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:24,429 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1098, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1097, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1098, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1097, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1098, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1097, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:25,429 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1100, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1099, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1100, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1099, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1100, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1099, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:26,429 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1102, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1101, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1102, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1101, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1102, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1101, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:27,430 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1104, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1103, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1104, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1103, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1104, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1103, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:28,430 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1106, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1105, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1106, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1105, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1106, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1105, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:29,430 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1108, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1107, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1108, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1107, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1108, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1107, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:30,431 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1110, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1109, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1110, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1109, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1110, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1109, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:31,430 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1112, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1111, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1112, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1111, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1112, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1111, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:32,430 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1114, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1113, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1114, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1113, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1114, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1113, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:33,431 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1116, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1115, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1116, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1115, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1116, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1115, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:34,431 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1118, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1117, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1118, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1117, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1118, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1117, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:35,431 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1120, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1119, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1120, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1119, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1120, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1119, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:36,431 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1122, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1121, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1122, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1121, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1122, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1121, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:37,432 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1124, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1123, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1124, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1123, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1124, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1123, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:38,432 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1126, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1125, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1126, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1125, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1126, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1125, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:39,436 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1128, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1127, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1128, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1127, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1128, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1127, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:40,432 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1130, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1129, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1130, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1129, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1130, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1129, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:41,432 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1132, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1131, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1132, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1131, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1132, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1131, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:42,432 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1134, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1133, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1134, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1133, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1134, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1133, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:43,433 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1136, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1135, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1136, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1135, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1136, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1135, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:44,432 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1138, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1137, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1138, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1137, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1138, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1137, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:45,432 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1140, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1139, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1140, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1139, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1140, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1139, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:46,433 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1142, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1141, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1142, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1141, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1142, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1141, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:47,435 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1144, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1143, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1144, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1143, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1144, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1143, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:48,434 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1146, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1145, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1146, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1145, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1146, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1145, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:49,434 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1148, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1147, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1148, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1147, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1148, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1147, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:50,435 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1150, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1149, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1150, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1149, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1150, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1149, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:51,434 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1152, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1151, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1152, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1151, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1152, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1151, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:52,435 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1154, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1153, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1154, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1153, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1154, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1153, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:53,435 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1156, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1155, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1156, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1155, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1156, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1155, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:54,435 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1158, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1157, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1158, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1157, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1158, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1157, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:55,436 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1160, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1159, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1160, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1159, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1160, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1159, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:56,436 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1162, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1161, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1162, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1161, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1162, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1161, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:57,437 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1164, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1163, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1164, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1163, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1164, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1163, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:58,437 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1166, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1165, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1166, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1165, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1166, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1165, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:32:59,439 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1168, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1167, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1168, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1167, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1168, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1167, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:33:00,438 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1170, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1169, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1170, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1169, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1170, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1169, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:33:01,438 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1172, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1171, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1172, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1171, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1172, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1171, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:33:02,438 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1174, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1173, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1174, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1173, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1174, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1173, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:33:03,439 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1176, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1175, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1176, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1175, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1176, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1175, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:33:04,439 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1178, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1177, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1178, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1177, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1178, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1177, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:33:05,440 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1180, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1179, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1180, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1179, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1180, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1179, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:33:06,439 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1182, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1181, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1182, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1181, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1182, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1181, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:33:07,442 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1184, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1183, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1184, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1183, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1184, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1183, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:33:08,440 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1186, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1185, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1186, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1185, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1186, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1185, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:33:09,440 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1188, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1187, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1188, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1187, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1188, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1187, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:33:10,440 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1190, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1189, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1190, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1189, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1190, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1189, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:33:11,441 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1192, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1191, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1192, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1191, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1192, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1191, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:33:12,441 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1194, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1193, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1194, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1193, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1194, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1193, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:33:13,442 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1196, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1195, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1196, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1195, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1196, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1195, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:33:14,446 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1198, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1197, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1198, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1197, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1198, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1197, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:33:15,442 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1200, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1199, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1200, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1199, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1200, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1199, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:33:16,442 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1202, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1201, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1202, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1201, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1202, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1201, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:33:17,443 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1204, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1203, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1204, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1203, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1204, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1203, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:33:18,442 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1206, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1205, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1206, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1205, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1206, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1205, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:33:19,443 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1208, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1207, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1208, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1207, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1208, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1207, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:33:20,443 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1210, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1209, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1210, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1209, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1210, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1209, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:33:21,444 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1212, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1211, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1212, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1211, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1212, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1211, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:33:22,443 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1214, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1213, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1214, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1213, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1214, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1213, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 20:33:23,444 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1216, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rolling back to view CacheView{viewId=1215, members=[node-udp-0/cluster]}: java.util.concurrent.ExecutionException: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1216, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1215, oldMembers=[node-udp-0/cluster]} at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) [:1.6.0_29] at java.util.concurrent.FutureTask.get(FutureTask.java:91) [:1.6.0_29] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterPrepareView(CacheViewsManagerImpl.java:319) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl.clusterInstallView(CacheViewsManagerImpl.java:250) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewInstallationTask.call(CacheViewsManagerImpl.java:877) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) [:1.6.0_29] at java.util.concurrent.FutureTask.run(FutureTask.java:138) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) [:1.6.0_29] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) [:1.6.0_29] at java.lang.Thread.run(Thread.java:680) [:1.6.0_29] Caused by: org.infinispan.remoting.transport.jgroups.SuspectException: One or more nodes have left the cluster while replicating command CacheViewControlCommand{cache=default, type=PREPARE_VIEW, sender=node-udp-0/cluster, newViewId=1216, newMembers=[node-udp-0/cluster, node-udp-1/cluster], oldViewId=1215, oldMembers=[node-udp-0/cluster]} at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:436) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:303) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] at org.infinispan.cacheviews.CacheViewsManagerImpl$2.call(CacheViewsManagerImpl.java:300) [infinispan-core-5.1.0.FINAL.jar:5.1.0.FINAL] ... 5 more 2012-01-27 20:33:23 Full thread dump Java HotSpot(TM) 64-Bit Server VM (20.4-b02-402 mixed mode): "Thread-1 (HornetQ-client-global-threads-1074009022)" daemon prio=5 tid=10b804800 nid=0x11859b000 waiting on condition [11859a000] java.lang.Thread.State: TIMED_WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d40705f0> (a java.util.concurrent.SynchronousQueue$TransferStack) at java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:196) at java.util.concurrent.SynchronousQueue$TransferStack.awaitFulfill(SynchronousQueue.java:424) at java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:323) at java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:874) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:945) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Thread-4 (HornetQ-client-global-scheduled-threads-1269579719)" daemon prio=5 tid=10bc85000 nid=0x118395000 waiting on condition [118394000] java.lang.Thread.State: TIMED_WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d40706b0> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:196) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2025) at java.util.concurrent.DelayQueue.take(DelayQueue.java:164) at java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:609) at java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:602) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Thread-29 (HornetQ-server-HornetQServerImpl::serverUUID=92c192b7-4956-11e1-9991-002608e4cba7-623174149)" prio=5 tid=1095a8000 nid=0x118292000 waiting on condition [118291000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d362e018> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Thread-28 (HornetQ-server-HornetQServerImpl::serverUUID=92c192b7-4956-11e1-9991-002608e4cba7-623174149)" prio=5 tid=1095a7800 nid=0x11818f000 waiting on condition [11818e000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d362e018> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Thread-27 (HornetQ-server-HornetQServerImpl::serverUUID=92c192b7-4956-11e1-9991-002608e4cba7-623174149)" prio=5 tid=1095a6000 nid=0x117f9d000 waiting on condition [117f9c000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d362e018> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Thread-26 (HornetQ-server-HornetQServerImpl::serverUUID=92c192b7-4956-11e1-9991-002608e4cba7-623174149)" prio=5 tid=1095a5800 nid=0x117e9a000 waiting on condition [117e99000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d362e018> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Thread-3 (HornetQ-client-global-scheduled-threads-1269579719)" daemon prio=5 tid=1095a4800 nid=0x117d97000 waiting on condition [117d96000] java.lang.Thread.State: TIMED_WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d40706b0> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:196) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2025) at java.util.concurrent.DelayQueue.take(DelayQueue.java:164) at java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:609) at java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:602) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Thread-25 (HornetQ-server-HornetQServerImpl::serverUUID=92c192b7-4956-11e1-9991-002608e4cba7-623174149)" prio=5 tid=1095a4000 nid=0x117c94000 waiting on condition [117c93000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d362e018> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Thread-24 (HornetQ-server-HornetQServerImpl::serverUUID=92c192b7-4956-11e1-9991-002608e4cba7-623174149)" prio=5 tid=1095a3000 nid=0x117b91000 waiting on condition [117b90000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d362e018> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Thread-23 (HornetQ-server-HornetQServerImpl::serverUUID=92c192b7-4956-11e1-9991-002608e4cba7-623174149)" prio=5 tid=1095a2800 nid=0x117a8e000 waiting on condition [117a8d000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d362e018> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Thread-22 (HornetQ-server-HornetQServerImpl::serverUUID=92c192b7-4956-11e1-9991-002608e4cba7-623174149)" prio=5 tid=10905b000 nid=0x11798b000 waiting on condition [11798a000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d362e018> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Thread-21 (HornetQ-server-HornetQServerImpl::serverUUID=92c192b7-4956-11e1-9991-002608e4cba7-623174149)" prio=5 tid=101d42800 nid=0x117888000 waiting on condition [117887000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d362e018> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Thread-20 (HornetQ-server-HornetQServerImpl::serverUUID=92c192b7-4956-11e1-9991-002608e4cba7-623174149)" prio=5 tid=101b62000 nid=0x117785000 waiting on condition [117784000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d362e018> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Thread-19 (HornetQ-server-HornetQServerImpl::serverUUID=92c192b7-4956-11e1-9991-002608e4cba7-623174149)" prio=5 tid=1095a1800 nid=0x117682000 waiting on condition [117681000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d362e018> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Thread-18 (HornetQ-server-HornetQServerImpl::serverUUID=92c192b7-4956-11e1-9991-002608e4cba7-623174149)" prio=5 tid=10938d800 nid=0x11757f000 waiting on condition [11757e000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d362e018> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Thread-17 (HornetQ-server-HornetQServerImpl::serverUUID=92c192b7-4956-11e1-9991-002608e4cba7-623174149)" prio=5 tid=10938a000 nid=0x11747c000 waiting on condition [11747b000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d362e018> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Thread-16 (HornetQ-server-HornetQServerImpl::serverUUID=92c192b7-4956-11e1-9991-002608e4cba7-623174149)" prio=5 tid=109389000 nid=0x117379000 waiting on condition [117378000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d362e018> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Thread-2 (HornetQ-client-global-scheduled-threads-1269579719)" daemon prio=5 tid=101b60800 nid=0x117276000 waiting on condition [117275000] java.lang.Thread.State: TIMED_WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d40706b0> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:196) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2025) at java.util.concurrent.DelayQueue.take(DelayQueue.java:164) at java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:609) at java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:602) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Thread-15 (HornetQ-server-HornetQServerImpl::serverUUID=92c192b7-4956-11e1-9991-002608e4cba7-623174149)" prio=5 tid=101b60000 nid=0x117173000 waiting on condition [117172000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d362e018> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Thread-14 (HornetQ-server-HornetQServerImpl::serverUUID=92c192b7-4956-11e1-9991-002608e4cba7-623174149)" prio=5 tid=101b5f000 nid=0x117070000 waiting on condition [11706f000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d362e018> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Thread-13 (HornetQ-server-HornetQServerImpl::serverUUID=92c192b7-4956-11e1-9991-002608e4cba7-623174149)" prio=5 tid=101b5d800 nid=0x116f6d000 waiting on condition [116f6c000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d362e018> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Thread-12 (HornetQ-server-HornetQServerImpl::serverUUID=92c192b7-4956-11e1-9991-002608e4cba7-623174149)" prio=5 tid=101b5d000 nid=0x116e6a000 waiting on condition [116e69000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d362e018> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Thread-11 (HornetQ-server-HornetQServerImpl::serverUUID=92c192b7-4956-11e1-9991-002608e4cba7-623174149)" prio=5 tid=101b5c000 nid=0x116d67000 waiting on condition [116d66000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d362e018> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Thread-10 (HornetQ-server-HornetQServerImpl::serverUUID=92c192b7-4956-11e1-9991-002608e4cba7-623174149)" prio=5 tid=101b5b800 nid=0x116c64000 waiting on condition [116c63000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d362e018> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Thread-9 (HornetQ-server-HornetQServerImpl::serverUUID=92c192b7-4956-11e1-9991-002608e4cba7-623174149)" prio=5 tid=109388800 nid=0x116b61000 waiting on condition [116b60000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d362e018> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Thread-8 (HornetQ-server-HornetQServerImpl::serverUUID=92c192b7-4956-11e1-9991-002608e4cba7-623174149)" prio=5 tid=109387800 nid=0x116a5e000 waiting on condition [116a5d000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d362e018> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Thread-1 (HornetQ-client-global-scheduled-threads-1269579719)" daemon prio=5 tid=1095ba800 nid=0x11695b000 waiting on condition [11695a000] java.lang.Thread.State: TIMED_WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d40706b0> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:196) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2025) at java.util.concurrent.DelayQueue.take(DelayQueue.java:164) at java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:609) at java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:602) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Thread-7 (HornetQ-server-HornetQServerImpl::serverUUID=92c192b7-4956-11e1-9991-002608e4cba7-623174149)" prio=5 tid=1093af000 nid=0x116858000 waiting on condition [116857000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d362e018> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Thread-6 (HornetQ-server-HornetQServerImpl::serverUUID=92c192b7-4956-11e1-9991-002608e4cba7-623174149)" prio=5 tid=10966f000 nid=0x11654f000 waiting on condition [11654e000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d362e018> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "pool-26-thread-2" prio=5 tid=101b76000 nid=0x110283000 in Object.wait() [110281000] java.lang.Thread.State: WAITING (on object monitor) at java.lang.Object.wait(Native Method) - waiting on <7d3f0bd10> (a java.lang.Object) at java.lang.Object.wait(Object.java:485) at org.infinispan.statetransfer.StateTransferLockImpl.acquireLockForCommitCommand(StateTransferLockImpl.java:369) - locked <7d3f0bd10> (a java.lang.Object) at org.infinispan.statetransfer.StateTransferLockImpl.acquireForCommand(StateTransferLockImpl.java:144) at org.infinispan.interceptors.StateTransferLockInterceptor.visitCommitCommand(StateTransferLockInterceptor.java:115) at org.infinispan.commands.tx.CommitCommand.acceptVisitor(CommitCommand.java:66) at org.infinispan.interceptors.base.CommandInterceptor.invokeNextInterceptor(CommandInterceptor.java:116) at org.infinispan.interceptors.base.CommandInterceptor.handleDefault(CommandInterceptor.java:130) at org.infinispan.commands.AbstractVisitor.visitCommitCommand(AbstractVisitor.java:121) at org.infinispan.commands.tx.CommitCommand.acceptVisitor(CommitCommand.java:66) at org.infinispan.interceptors.base.CommandInterceptor.invokeNextInterceptor(CommandInterceptor.java:116) at org.infinispan.interceptors.InvocationContextInterceptor.handleAll(InvocationContextInterceptor.java:130) at org.infinispan.interceptors.InvocationContextInterceptor.handleDefault(InvocationContextInterceptor.java:89) at org.infinispan.commands.AbstractVisitor.visitCommitCommand(AbstractVisitor.java:121) at org.infinispan.commands.tx.CommitCommand.acceptVisitor(CommitCommand.java:66) at org.infinispan.interceptors.base.CommandInterceptor.invokeNextInterceptor(CommandInterceptor.java:116) at org.infinispan.interceptors.base.CommandInterceptor.handleDefault(CommandInterceptor.java:130) at org.infinispan.commands.AbstractVisitor.visitCommitCommand(AbstractVisitor.java:121) at org.infinispan.commands.tx.CommitCommand.acceptVisitor(CommitCommand.java:66) at org.infinispan.interceptors.base.CommandInterceptor.invokeNextInterceptor(CommandInterceptor.java:116) at org.infinispan.interceptors.BatchingInterceptor.handleDefault(BatchingInterceptor.java:86) at org.infinispan.commands.AbstractVisitor.visitCommitCommand(AbstractVisitor.java:121) at org.infinispan.commands.tx.CommitCommand.acceptVisitor(CommitCommand.java:66) at org.infinispan.interceptors.base.CommandInterceptor.invokeNextInterceptor(CommandInterceptor.java:116) at org.infinispan.interceptors.base.CommandInterceptor.handleDefault(CommandInterceptor.java:130) at org.jboss.as.clustering.infinispan.DefaultEmbeddedCacheManager$ClassLoaderAwareCommandInterceptor.handleDefault(DefaultEmbeddedCacheManager.java:410) at org.infinispan.commands.AbstractVisitor.visitCommitCommand(AbstractVisitor.java:121) at org.infinispan.commands.tx.CommitCommand.acceptVisitor(CommitCommand.java:66) at org.infinispan.interceptors.InterceptorChain.invoke(InterceptorChain.java:345) at org.infinispan.transaction.TransactionCoordinator.commit(TransactionCoordinator.java:181) at org.infinispan.transaction.synchronization.SynchronizationAdapter.afterCompletion(SynchronizationAdapter.java:83) at org.infinispan.transaction.tm.DummyTransaction.notifyAfterCompletion(DummyTransaction.java:272) at org.infinispan.transaction.tm.DummyTransaction.runCommitTx(DummyTransaction.java:321) at org.infinispan.transaction.tm.DummyTransaction.commit(DummyTransaction.java:90) at org.infinispan.transaction.tm.DummyBaseTransactionManager.commit(DummyBaseTransactionManager.java:100) at org.infinispan.batch.BatchContainer.resolveTransaction(BatchContainer.java:123) at org.infinispan.batch.BatchContainer.endBatch(BatchContainer.java:105) at org.infinispan.batch.BatchContainer.endBatch(BatchContainer.java:86) at org.infinispan.CacheImpl.endBatch(CacheImpl.java:577) at org.infinispan.AbstractDelegatingCache.endBatch(AbstractDelegatingCache.java:75) at org.infinispan.AbstractDelegatingCache.endBatch(AbstractDelegatingCache.java:75) at org.infinispan.AbstractDelegatingCache.endBatch(AbstractDelegatingCache.java:75) at org.jboss.as.clustering.infinispan.invoker.BatchOperation.invoke(BatchOperation.java:26) at org.jboss.as.clustering.service.ServiceProviderRegistryService.invoke(ServiceProviderRegistryService.java:187) at org.jboss.as.clustering.service.ServiceProviderRegistryService.unregister(ServiceProviderRegistryService.java:108) at org.jboss.as.clustering.singleton.SingletonService.stop(SingletonService.java:109) at org.jboss.as.clustering.AsynchronousService$2.run(AsynchronousService.java:64) at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) at java.lang.Thread.run(Thread.java:680) "management-handler-threads - 3" prio=5 tid=101b71800 nid=0x10f865000 in Object.wait() [10f863000] java.lang.Thread.State: WAITING (on object monitor) at java.lang.Object.wait(Native Method) - waiting on <7d40b6360> (a java.util.HashMap) at java.lang.Object.wait(Object.java:485) at org.jboss.as.controller.OperationContextImpl.awaitModelControllerContainerMonitor(OperationContextImpl.java:137) - locked <7d40b6360> (a java.util.HashMap) at org.jboss.as.controller.AbstractOperationContext.doCompleteStep(AbstractOperationContext.java:242) at org.jboss.as.controller.AbstractOperationContext.completeStep(AbstractOperationContext.java:193) at org.jboss.as.server.deployment.DeploymentRemoveHandler$1.execute(DeploymentRemoveHandler.java:85) at org.jboss.as.controller.AbstractOperationContext.executeStep(AbstractOperationContext.java:378) at org.jboss.as.controller.AbstractOperationContext.doCompleteStep(AbstractOperationContext.java:265) at org.jboss.as.controller.AbstractOperationContext.completeStep(AbstractOperationContext.java:193) at org.jboss.as.server.deployment.DeploymentHandlerUtil$5.execute(DeploymentHandlerUtil.java:295) at org.jboss.as.controller.AbstractOperationContext.executeStep(AbstractOperationContext.java:378) at org.jboss.as.controller.AbstractOperationContext.doCompleteStep(AbstractOperationContext.java:265) at org.jboss.as.controller.AbstractOperationContext.completeStep(AbstractOperationContext.java:193) at org.jboss.as.server.deployment.DeploymentRemoveHandler.execute(DeploymentRemoveHandler.java:112) at org.jboss.as.controller.AbstractOperationContext.executeStep(AbstractOperationContext.java:378) at org.jboss.as.controller.AbstractOperationContext.doCompleteStep(AbstractOperationContext.java:265) at org.jboss.as.controller.AbstractOperationContext.completeStep(AbstractOperationContext.java:193) at org.jboss.as.server.deployment.DeploymentUndeployHandler.execute(DeploymentUndeployHandler.java:63) at org.jboss.as.controller.AbstractOperationContext.executeStep(AbstractOperationContext.java:378) at org.jboss.as.controller.AbstractOperationContext.doCompleteStep(AbstractOperationContext.java:265) at org.jboss.as.controller.AbstractOperationContext.completeStep(AbstractOperationContext.java:193) at org.jboss.as.controller.CompositeOperationHandler.execute(CompositeOperationHandler.java:84) at org.jboss.as.controller.AbstractOperationContext.executeStep(AbstractOperationContext.java:378) at org.jboss.as.controller.AbstractOperationContext.doCompleteStep(AbstractOperationContext.java:265) at org.jboss.as.controller.AbstractOperationContext.completeStep(AbstractOperationContext.java:193) at org.jboss.as.controller.ModelControllerImpl$DefaultPrepareStepHandler.execute(ModelControllerImpl.java:461) at org.jboss.as.controller.AbstractOperationContext.executeStep(AbstractOperationContext.java:378) at org.jboss.as.controller.AbstractOperationContext.doCompleteStep(AbstractOperationContext.java:265) at org.jboss.as.controller.AbstractOperationContext.completeStep(AbstractOperationContext.java:193) at org.jboss.as.controller.ModelControllerImpl.execute(ModelControllerImpl.java:121) at org.jboss.as.controller.remote.ModelControllerClientOperationHandler$ExecuteRequestHandler.doExecute(ModelControllerClientOperationHandler.java:121) at org.jboss.as.controller.remote.ModelControllerClientOperationHandler$ExecuteRequestHandler$1.execute(ModelControllerClientOperationHandler.java:98) at org.jboss.as.protocol.mgmt.AbstractMessageHandler$2$1.doExecute(AbstractMessageHandler.java:250) at org.jboss.as.protocol.mgmt.AbstractMessageHandler$AsyncTaskRunner.run(AbstractMessageHandler.java:444) at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) at java.lang.Thread.run(Thread.java:680) at org.jboss.threads.JBossThread.run(JBossThread.java:122) "Incoming-19,null" prio=5 tid=1095b2800 nid=0x116755000 waiting on condition [116754000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3d28e90> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Incoming-18,null" prio=5 tid=1095b2000 nid=0x116652000 waiting on condition [116651000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3d28e90> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Incoming-17,null" prio=5 tid=1095b1000 nid=0x11644c000 waiting on condition [11644b000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3d28e90> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Thread-5 (HornetQ-server-HornetQServerImpl::serverUUID=92c192b7-4956-11e1-9991-002608e4cba7-623174149)" prio=5 tid=1095b0800 nid=0x115a2e000 waiting on condition [115a2d000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d362e018> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Thread-4 (HornetQ-server-HornetQServerImpl::serverUUID=92c192b7-4956-11e1-9991-002608e4cba7-623174149)" prio=5 tid=1095af800 nid=0x114378000 waiting on condition [114377000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d362e018> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "transport-thread-25" daemon prio=5 tid=101b70000 nid=0x11541c000 waiting on condition [11541b000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3e7d838> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "transport-thread-24" daemon prio=5 tid=101b6f000 nid=0x115216000 waiting on condition [115215000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3e7d838> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "transport-thread-23" daemon prio=5 tid=101c69800 nid=0x11472f000 waiting on condition [11472e000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3e7d838> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "transport-thread-22" daemon prio=5 tid=1018aa800 nid=0x116349000 waiting on condition [116348000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3e7d838> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "transport-thread-21" daemon prio=5 tid=1018aa000 nid=0x116246000 waiting on condition [116245000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3e7d838> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "transport-thread-20" daemon prio=5 tid=101c94800 nid=0x10ff7a000 waiting on condition [10ff79000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3e7d838> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "transport-thread-19" daemon prio=5 tid=1018a6000 nid=0x116143000 waiting on condition [116142000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3e7d838> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "transport-thread-18" daemon prio=5 tid=1018a5800 nid=0x116040000 waiting on condition [11603f000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3e7d838> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "transport-thread-17" daemon prio=5 tid=1018a4800 nid=0x115f3d000 waiting on condition [115f3c000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3e7d838> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "transport-thread-16" daemon prio=5 tid=1097ea800 nid=0x115e3a000 waiting on condition [115e39000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3e7d838> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "transport-thread-15" daemon prio=5 tid=1018a4000 nid=0x115d37000 waiting on condition [115d36000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3e7d838> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "transport-thread-14" daemon prio=5 tid=101c97800 nid=0x115c34000 waiting on condition [115c33000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3e7d838> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Incoming-16,null" prio=5 tid=101c96800 nid=0x11462c000 waiting on condition [11462b000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3d28e90> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Incoming-15,null" prio=5 tid=1093b0800 nid=0x114529000 waiting on condition [114528000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3d28e90> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Thread-3 (HornetQ-server-HornetQServerImpl::serverUUID=92c192b7-4956-11e1-9991-002608e4cba7-623174149)" prio=5 tid=101c96000 nid=0x110489000 waiting on condition [110488000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d362e018> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Thread-2 (HornetQ-server-HornetQServerImpl::serverUUID=92c192b7-4956-11e1-9991-002608e4cba7-623174149)" prio=5 tid=101c95000 nid=0x110386000 waiting on condition [110385000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d362e018> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Incoming-14,null" prio=5 tid=1097e9000 nid=0x110180000 waiting on condition [11017f000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3d28e90> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Incoming-13,null" prio=5 tid=101c93800 nid=0x11007d000 waiting on condition [11007c000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3d28e90> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Thread-1 (HornetQ-server-HornetQServerImpl::serverUUID=92c192b7-4956-11e1-9991-002608e4cba7-623174149)" prio=5 tid=101c93000 nid=0x10fe77000 waiting on condition [10fe76000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d362e018> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Thread-0 (HornetQ-server-HornetQServerImpl::serverUUID=92c192b7-4956-11e1-9991-002608e4cba7-623174149)" prio=5 tid=101c80800 nid=0x10f03a000 waiting on condition [10f039000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d362e018> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "transport-thread-13" daemon prio=5 tid=101c92000 nid=0x10fd74000 waiting on condition [10fd73000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3e7d838> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "transport-thread-12" daemon prio=5 tid=101c84000 nid=0x10fa6b000 waiting on condition [10fa6a000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3e7d838> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "transport-thread-11" daemon prio=5 tid=1093af800 nid=0x10f968000 waiting on condition [10f967000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3e7d838> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Incoming-12,null" prio=5 tid=101c83800 nid=0x10f762000 waiting on condition [10f761000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3d28e90> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Incoming-11,null" prio=5 tid=1092c2800 nid=0x10f65f000 waiting on condition [10f65e000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3d28e90> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Timer-5," daemon prio=5 tid=101ceb800 nid=0x10d7e1000 waiting on condition [10d7e0000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3d29178> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "LockBreakingService,default,node-udp-0/cluster" daemon prio=5 tid=1093ae000 nid=0x10cc43000 waiting on condition [10cc42000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3fc7900> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingDeque.takeFirst(LinkedBlockingDeque.java:440) at java.util.concurrent.LinkedBlockingDeque.take(LinkedBlockingDeque.java:629) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "transport-thread-10" daemon prio=5 tid=101c82800 nid=0x10f55c000 waiting on condition [10f55b000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3e7d838> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "transport-thread-9" daemon prio=5 tid=101c82000 nid=0x10f455000 waiting on condition [10f454000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3e7d838> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "transport-thread-8" daemon prio=5 tid=101c81000 nid=0x10f343000 waiting on condition [10f342000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3e7d838> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Incoming-10,null" prio=5 tid=1093ad800 nid=0x10f240000 waiting on condition [10f23f000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3d28e90> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Incoming-9,null" prio=5 tid=1093ac800 nid=0x10f13d000 waiting on condition [10f13c000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3d28e90> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "transport-thread-7" daemon prio=5 tid=1093ac000 nid=0x10ef37000 waiting on condition [10ef36000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3e7d838> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "transport-thread-6" daemon prio=5 tid=1093ab000 nid=0x10ee34000 waiting on condition [10ee33000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3e7d838> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Incoming-8,null" prio=5 tid=1093aa800 nid=0x10ed31000 waiting on condition [10ed30000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3d28e90> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Incoming-7,null" prio=5 tid=101c7d800 nid=0x10ec2e000 waiting on condition [10ec2d000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3d28e90> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Incoming-6,null" prio=5 tid=1093a9800 nid=0x10eb2b000 waiting on condition [10eb2a000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3d28e90> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Incoming-5,null" prio=5 tid=1096e8000 nid=0x10ea28000 waiting on condition [10ea27000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3d28e90> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Incoming-4,null" prio=5 tid=101c7d000 nid=0x10e925000 waiting on condition [10e924000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3d28e90> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Timer-4," daemon prio=5 tid=109315000 nid=0x10ca3d000 waiting on condition [10ca3c000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3d29178> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Timer-3," daemon prio=5 tid=101c7c000 nid=0x10e822000 waiting on condition [10e821000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3d29178> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Timer-2," daemon prio=5 tid=101c7b800 nid=0x10e71f000 waiting on condition [10e71e000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3d29178> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Incoming-3,null" prio=5 tid=1096e7000 nid=0x10e407000 waiting on condition [10e406000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3d28e90> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "OOB-20,null" prio=5 tid=109538800 nid=0x10e304000 waiting on condition [10e303000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3d16d68> (a java.util.concurrent.SynchronousQueue$TransferStack) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.SynchronousQueue$TransferStack.awaitFulfill(SynchronousQueue.java:422) at java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:323) at java.util.concurrent.SynchronousQueue.take(SynchronousQueue.java:857) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "OOB-19,null" prio=5 tid=101c7a800 nid=0x10e201000 waiting on condition [10e200000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3d16d68> (a java.util.concurrent.SynchronousQueue$TransferStack) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.SynchronousQueue$TransferStack.awaitFulfill(SynchronousQueue.java:422) at java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:323) at java.util.concurrent.SynchronousQueue.take(SynchronousQueue.java:857) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "OOB-18,null" prio=5 tid=1019b2000 nid=0x10e0fe000 waiting on condition [10e0fd000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3d16d68> (a java.util.concurrent.SynchronousQueue$TransferStack) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.SynchronousQueue$TransferStack.awaitFulfill(SynchronousQueue.java:422) at java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:323) at java.util.concurrent.SynchronousQueue.take(SynchronousQueue.java:857) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "OOB-17,null" prio=5 tid=109537800 nid=0x10dffb000 waiting on condition [10dffa000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3d16d68> (a java.util.concurrent.SynchronousQueue$TransferStack) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.SynchronousQueue$TransferStack.awaitFulfill(SynchronousQueue.java:422) at java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:323) at java.util.concurrent.SynchronousQueue.take(SynchronousQueue.java:857) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "OOB-16,null" prio=5 tid=109759800 nid=0x10def8000 waiting on condition [10def7000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3d16d68> (a java.util.concurrent.SynchronousQueue$TransferStack) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.SynchronousQueue$TransferStack.awaitFulfill(SynchronousQueue.java:422) at java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:323) at java.util.concurrent.SynchronousQueue.take(SynchronousQueue.java:857) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "OOB-15,null" prio=5 tid=109758800 nid=0x10ddf5000 waiting on condition [10ddf4000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3d16d68> (a java.util.concurrent.SynchronousQueue$TransferStack) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.SynchronousQueue$TransferStack.awaitFulfill(SynchronousQueue.java:422) at java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:323) at java.util.concurrent.SynchronousQueue.take(SynchronousQueue.java:857) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "OOB-14,null" prio=5 tid=1096ee000 nid=0x10dcf2000 waiting on condition [10dcf1000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3d16d68> (a java.util.concurrent.SynchronousQueue$TransferStack) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.SynchronousQueue$TransferStack.awaitFulfill(SynchronousQueue.java:422) at java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:323) at java.util.concurrent.SynchronousQueue.take(SynchronousQueue.java:857) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "transport-thread-5" daemon prio=5 tid=1019b1800 nid=0x10dbef000 waiting on condition [10dbee000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3e7d838> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "transport-thread-4" daemon prio=5 tid=1019b0800 nid=0x10daea000 waiting on condition [10dae9000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3e7d838> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "transport-thread-3" daemon prio=5 tid=1019b0000 nid=0x10d9e7000 waiting on condition [10d9e6000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3e7d838> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "OOB-13,null" prio=5 tid=1096ed000 nid=0x10d8e4000 waiting on condition [10d8e3000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3d16d68> (a java.util.concurrent.SynchronousQueue$TransferStack) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.SynchronousQueue$TransferStack.awaitFulfill(SynchronousQueue.java:422) at java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:323) at java.util.concurrent.SynchronousQueue.take(SynchronousQueue.java:857) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "OOB-12,null" prio=5 tid=1092c1800 nid=0x10d6de000 waiting on condition [10d6dd000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3d16d68> (a java.util.concurrent.SynchronousQueue$TransferStack) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.SynchronousQueue$TransferStack.awaitFulfill(SynchronousQueue.java:422) at java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:323) at java.util.concurrent.SynchronousQueue.take(SynchronousQueue.java:857) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "OOB-11,null" prio=5 tid=1092e2000 nid=0x10d5db000 waiting on condition [10d5da000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3d16d68> (a java.util.concurrent.SynchronousQueue$TransferStack) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.SynchronousQueue$TransferStack.awaitFulfill(SynchronousQueue.java:422) at java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:323) at java.util.concurrent.SynchronousQueue.take(SynchronousQueue.java:857) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "OOB-10,null" prio=5 tid=109159800 nid=0x10d4d8000 waiting on condition [10d4d7000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3d16d68> (a java.util.concurrent.SynchronousQueue$TransferStack) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.SynchronousQueue$TransferStack.awaitFulfill(SynchronousQueue.java:422) at java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:323) at java.util.concurrent.SynchronousQueue.take(SynchronousQueue.java:857) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "OOB-9,null" prio=5 tid=101a4a800 nid=0x10d3d5000 waiting on condition [10d3d4000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3d16d68> (a java.util.concurrent.SynchronousQueue$TransferStack) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.SynchronousQueue$TransferStack.awaitFulfill(SynchronousQueue.java:422) at java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:323) at java.util.concurrent.SynchronousQueue.take(SynchronousQueue.java:857) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "notification-thread-2" daemon prio=1 tid=101a4a000 nid=0x10d2d2000 waiting on condition [10d2d1000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3f0cb20> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "OOB-8,null" prio=5 tid=101cee800 nid=0x10d1cf000 waiting on condition [10d1ce000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3d16d68> (a java.util.concurrent.SynchronousQueue$TransferStack) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.SynchronousQueue$TransferStack.awaitFulfill(SynchronousQueue.java:422) at java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:323) at java.util.concurrent.SynchronousQueue.take(SynchronousQueue.java:857) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "OOB-7,null" prio=5 tid=109158800 nid=0x10d003000 waiting on condition [10d002000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3d16d68> (a java.util.concurrent.SynchronousQueue$TransferStack) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.SynchronousQueue$TransferStack.awaitFulfill(SynchronousQueue.java:422) at java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:323) at java.util.concurrent.SynchronousQueue.take(SynchronousQueue.java:857) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Incoming-2,null" prio=5 tid=109791800 nid=0x10cf00000 waiting on condition [10ceff000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3d28e90> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "OOB-6,null" prio=5 tid=101c87000 nid=0x10cde4000 waiting on condition [10cde3000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3d16d68> (a java.util.concurrent.SynchronousQueue$TransferStack) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.SynchronousQueue$TransferStack.awaitFulfill(SynchronousQueue.java:422) at java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:323) at java.util.concurrent.SynchronousQueue.take(SynchronousQueue.java:857) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Incoming-1,null" prio=5 tid=101ce9000 nid=0x10cb40000 waiting on condition [10cb3f000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3d28e90> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "OOB-5,null" prio=5 tid=1093f0800 nid=0x10c93a000 waiting on condition [10c939000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3d16d68> (a java.util.concurrent.SynchronousQueue$TransferStack) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.SynchronousQueue$TransferStack.awaitFulfill(SynchronousQueue.java:422) at java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:323) at java.util.concurrent.SynchronousQueue.take(SynchronousQueue.java:857) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "OOB-4,null" prio=5 tid=101e34800 nid=0x10c837000 waiting on condition [10c836000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3d16d68> (a java.util.concurrent.SynchronousQueue$TransferStack) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.SynchronousQueue$TransferStack.awaitFulfill(SynchronousQueue.java:422) at java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:323) at java.util.concurrent.SynchronousQueue.take(SynchronousQueue.java:857) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Thread-2 (HornetQ-remoting-threads-HornetQServerImpl::serverUUID=92c192b7-4956-11e1-9991-002608e4cba7-1763015052-14617658)" prio=5 tid=101e34000 nid=0x10c631000 waiting on condition [10c630000] java.lang.Thread.State: TIMED_WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d394b8a0> (a java.util.concurrent.SynchronousQueue$TransferStack) at java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:196) at java.util.concurrent.SynchronousQueue$TransferStack.awaitFulfill(SynchronousQueue.java:424) at java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:323) at java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:874) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:945) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Thread-0 (HornetQ-client-global-scheduled-threads-1269579719)" daemon prio=5 tid=1093ee000 nid=0x10c437000 waiting on condition [10c436000] java.lang.Thread.State: TIMED_WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d40706b0> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:196) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2025) at java.util.concurrent.DelayQueue.take(DelayQueue.java:164) at java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:609) at java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:602) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "http-localhost-127.0.0.1-8080-1" daemon prio=5 tid=109312000 nid=0x115b31000 in Object.wait() [115b30000] java.lang.Thread.State: WAITING (on object monitor) at java.lang.Object.wait(Native Method) - waiting on <7d3fdf788> (a org.apache.tomcat.util.net.AprEndpoint$Worker) at java.lang.Object.wait(Object.java:485) at org.apache.tomcat.util.net.AprEndpoint$Worker.await(AprEndpoint.java:1986) - locked <7d3fdf788> (a org.apache.tomcat.util.net.AprEndpoint$Worker) at org.apache.tomcat.util.net.AprEndpoint$Worker.run(AprEndpoint.java:2011) at java.lang.Thread.run(Thread.java:680) "transport-thread-1" daemon prio=5 tid=101c68800 nid=0x11592b000 waiting on condition [11592a000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3e7d838> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "transport-thread-0" daemon prio=5 tid=101bee000 nid=0x115828000 waiting on condition [115827000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3e7d838> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "CacheViewInstaller-1,node-udp-0/cluster" daemon prio=5 tid=101b5a800 nid=0x115725000 waiting on condition [115724000] java.lang.Thread.State: TIMED_WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3e7d880> (a java.util.concurrent.SynchronousQueue$TransferStack) at java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:196) at java.util.concurrent.SynchronousQueue$TransferStack.awaitFulfill(SynchronousQueue.java:424) at java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:323) at java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:874) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:945) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Scheduled-eviction-thread-0" daemon prio=1 tid=113044000 nid=0x115622000 waiting on condition [115621000] java.lang.Thread.State: TIMED_WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3ecf688> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:196) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2025) at java.util.concurrent.DelayQueue.take(DelayQueue.java:164) at java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:609) at java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:602) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "CacheViewTrigger,node-udp-0/cluster" daemon prio=5 tid=109063000 nid=0x11551f000 waiting on condition [11551e000] java.lang.Thread.State: TIMED_WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3e7d898> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:196) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2116) at org.infinispan.cacheviews.CacheViewsManagerImpl$ViewTriggerThread.run(CacheViewsManagerImpl.java:817) "AsynchViewChangeHandler Thread" prio=5 tid=101c64800 nid=0x115319000 in Object.wait() [115317000] java.lang.Thread.State: WAITING (on object monitor) at java.lang.Object.wait(Native Method) - waiting on <7d3f0bd10> (a java.lang.Object) at java.lang.Object.wait(Object.java:485) at org.infinispan.statetransfer.StateTransferLockImpl.acquireLockForCommitCommand(StateTransferLockImpl.java:369) - locked <7d3f0bd10> (a java.lang.Object) at org.infinispan.statetransfer.StateTransferLockImpl.acquireForCommand(StateTransferLockImpl.java:144) at org.infinispan.interceptors.StateTransferLockInterceptor.visitCommitCommand(StateTransferLockInterceptor.java:115) at org.infinispan.commands.tx.CommitCommand.acceptVisitor(CommitCommand.java:66) at org.infinispan.interceptors.base.CommandInterceptor.invokeNextInterceptor(CommandInterceptor.java:116) at org.infinispan.interceptors.base.CommandInterceptor.handleDefault(CommandInterceptor.java:130) at org.infinispan.commands.AbstractVisitor.visitCommitCommand(AbstractVisitor.java:121) at org.infinispan.commands.tx.CommitCommand.acceptVisitor(CommitCommand.java:66) at org.infinispan.interceptors.base.CommandInterceptor.invokeNextInterceptor(CommandInterceptor.java:116) at org.infinispan.interceptors.InvocationContextInterceptor.handleAll(InvocationContextInterceptor.java:130) at org.infinispan.interceptors.InvocationContextInterceptor.handleDefault(InvocationContextInterceptor.java:89) at org.infinispan.commands.AbstractVisitor.visitCommitCommand(AbstractVisitor.java:121) at org.infinispan.commands.tx.CommitCommand.acceptVisitor(CommitCommand.java:66) at org.infinispan.interceptors.base.CommandInterceptor.invokeNextInterceptor(CommandInterceptor.java:116) at org.infinispan.interceptors.base.CommandInterceptor.handleDefault(CommandInterceptor.java:130) at org.infinispan.commands.AbstractVisitor.visitCommitCommand(AbstractVisitor.java:121) at org.infinispan.commands.tx.CommitCommand.acceptVisitor(CommitCommand.java:66) at org.infinispan.interceptors.base.CommandInterceptor.invokeNextInterceptor(CommandInterceptor.java:116) at org.infinispan.interceptors.BatchingInterceptor.handleDefault(BatchingInterceptor.java:86) at org.infinispan.commands.AbstractVisitor.visitCommitCommand(AbstractVisitor.java:121) at org.infinispan.commands.tx.CommitCommand.acceptVisitor(CommitCommand.java:66) at org.infinispan.interceptors.base.CommandInterceptor.invokeNextInterceptor(CommandInterceptor.java:116) at org.infinispan.interceptors.base.CommandInterceptor.handleDefault(CommandInterceptor.java:130) at org.jboss.as.clustering.infinispan.DefaultEmbeddedCacheManager$ClassLoaderAwareCommandInterceptor.handleDefault(DefaultEmbeddedCacheManager.java:410) at org.infinispan.commands.AbstractVisitor.visitCommitCommand(AbstractVisitor.java:121) at org.infinispan.commands.tx.CommitCommand.acceptVisitor(CommitCommand.java:66) at org.infinispan.interceptors.InterceptorChain.invoke(InterceptorChain.java:345) at org.infinispan.transaction.TransactionCoordinator.commit(TransactionCoordinator.java:181) at org.infinispan.transaction.synchronization.SynchronizationAdapter.afterCompletion(SynchronizationAdapter.java:83) at org.infinispan.transaction.tm.DummyTransaction.notifyAfterCompletion(DummyTransaction.java:272) at org.infinispan.transaction.tm.DummyTransaction.runCommitTx(DummyTransaction.java:321) at org.infinispan.transaction.tm.DummyTransaction.commit(DummyTransaction.java:90) at org.infinispan.transaction.tm.DummyBaseTransactionManager.commit(DummyBaseTransactionManager.java:100) at org.infinispan.batch.BatchContainer.resolveTransaction(BatchContainer.java:123) at org.infinispan.batch.BatchContainer.endBatch(BatchContainer.java:105) at org.infinispan.batch.BatchContainer.endBatch(BatchContainer.java:86) at org.infinispan.CacheImpl.endBatch(CacheImpl.java:577) at org.infinispan.AbstractDelegatingCache.endBatch(AbstractDelegatingCache.java:75) at org.infinispan.AbstractDelegatingCache.endBatch(AbstractDelegatingCache.java:75) at org.infinispan.AbstractDelegatingCache.endBatch(AbstractDelegatingCache.java:75) at org.jboss.as.clustering.infinispan.invoker.BatchOperation.invoke(BatchOperation.java:26) at org.jboss.as.clustering.service.ServiceProviderRegistryService.invoke(ServiceProviderRegistryService.java:187) at org.jboss.as.clustering.service.ServiceProviderRegistryService.purgeDeadMembers(ServiceProviderRegistryService.java:167) at org.jboss.as.clustering.service.ServiceProviderRegistryService.membershipChanged(ServiceProviderRegistryService.java:140) at org.jboss.as.clustering.CoreGroupCommunicationService.notifyListeners(CoreGroupCommunicationService.java:1263) at org.jboss.as.clustering.CoreGroupCommunicationService$ViewChangeEventProcessor.processEvent(CoreGroupCommunicationService.java:1513) at org.jboss.as.clustering.AsynchEventHandler.run(AsynchEventHandler.java:96) at java.lang.Thread.run(Thread.java:680) "OOB-3,null" prio=5 tid=1090cd000 nid=0x115113000 waiting on condition [115112000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3d16d68> (a java.util.concurrent.SynchronousQueue$TransferStack) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.SynchronousQueue$TransferStack.awaitFulfill(SynchronousQueue.java:422) at java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:323) at java.util.concurrent.SynchronousQueue.take(SynchronousQueue.java:857) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "OOB-2,null" prio=5 tid=1095bc800 nid=0x115010000 waiting on condition [11500f000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3d16d68> (a java.util.concurrent.SynchronousQueue$TransferStack) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.SynchronousQueue$TransferStack.awaitFulfill(SynchronousQueue.java:422) at java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:323) at java.util.concurrent.SynchronousQueue.take(SynchronousQueue.java:857) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "OOB-1,null" prio=5 tid=11308b800 nid=0x114f0d000 waiting on condition [114f0c000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3d16d68> (a java.util.concurrent.SynchronousQueue$TransferStack) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.SynchronousQueue$TransferStack.awaitFulfill(SynchronousQueue.java:422) at java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:323) at java.util.concurrent.SynchronousQueue.take(SynchronousQueue.java:857) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "FD_SOCK server socket acceptor,cluster,node-udp-0/cluster" daemon prio=5 tid=101921800 nid=0x114e0a000 runnable [114e09000] java.lang.Thread.State: RUNNABLE at java.net.PlainSocketImpl.socketAccept(Native Method) at java.net.PlainSocketImpl.accept(PlainSocketImpl.java:408) - locked <7d3bf3560> (a java.net.SocksSocketImpl) at java.net.ServerSocket.implAccept(ServerSocket.java:462) at java.net.ServerSocket.accept(ServerSocket.java:430) at org.jgroups.protocols.FD_SOCK$ServerSocketHandler.run(FD_SOCK.java:1034) at java.lang.Thread.run(Thread.java:680) "multicast receiver,null" prio=5 tid=1095bc000 nid=0x114d07000 runnable [114d06000] java.lang.Thread.State: RUNNABLE at java.net.PlainDatagramSocketImpl.receive0(Native Method) - locked <7d3d08ea8> (a java.net.PlainDatagramSocketImpl) at java.net.PlainDatagramSocketImpl.receive(PlainDatagramSocketImpl.java:145) - locked <7d3d08ea8> (a java.net.PlainDatagramSocketImpl) at java.net.DatagramSocket.receive(DatagramSocket.java:725) - locked <7d3c157f0> (a java.net.DatagramPacket) - locked <7d3b9f688> (a java.net.MulticastSocket) at org.jgroups.protocols.UDP$PacketReceiver.run(UDP.java:674) at java.lang.Thread.run(Thread.java:680) "unicast receiver,null" prio=5 tid=10b972000 nid=0x114c04000 runnable [114c03000] java.lang.Thread.State: RUNNABLE at java.net.PlainDatagramSocketImpl.receive0(Native Method) - locked <7d3d08ee8> (a java.net.PlainDatagramSocketImpl) at java.net.PlainDatagramSocketImpl.receive(PlainDatagramSocketImpl.java:145) - locked <7d3d08ee8> (a java.net.PlainDatagramSocketImpl) at java.net.DatagramSocket.receive(DatagramSocket.java:725) - locked <7d3c15818> (a java.net.DatagramPacket) - locked <7d3d071e8> (a java.net.DatagramSocket) at org.jgroups.protocols.UDP$PacketReceiver.run(UDP.java:674) at java.lang.Thread.run(Thread.java:680) "DiagnosticsHandler,null" daemon prio=5 tid=113088800 nid=0x114b01000 runnable [114b00000] java.lang.Thread.State: RUNNABLE at java.net.PlainDatagramSocketImpl.receive0(Native Method) - locked <7d3d091a8> (a java.net.PlainDatagramSocketImpl) at java.net.PlainDatagramSocketImpl.receive(PlainDatagramSocketImpl.java:145) - locked <7d3d091a8> (a java.net.PlainDatagramSocketImpl) at java.net.DatagramSocket.receive(DatagramSocket.java:725) - locked <7d3c15840> (a java.net.DatagramPacket) - locked <7d3b9fa78> (a java.net.MulticastSocket) at org.jgroups.stack.DiagnosticsHandler.run(DiagnosticsHandler.java:94) at java.lang.Thread.run(Thread.java:680) "Timer runner-1," daemon prio=5 tid=101920800 nid=0x114847000 waiting on condition [114846000] java.lang.Thread.State: TIMED_WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d3d17068> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:196) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2116) at org.jgroups.util.TimeScheduler2.waitUntilNextExecution(TimeScheduler2.java:340) at org.jgroups.util.TimeScheduler2._run(TimeScheduler2.java:319) at org.jgroups.util.TimeScheduler2.run(TimeScheduler2.java:279) at java.lang.Thread.run(Thread.java:680) "Remoting "node-udp-0:MANAGEMENT" task-4" prio=5 tid=10b8f1000 nid=0x114275000 waiting on condition [114274000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d32ad5e0> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at org.xnio.LimitedBlockingQueue.take(LimitedBlockingQueue.java:95) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Remoting "node-udp-0:MANAGEMENT" task-3" prio=5 tid=101b07800 nid=0x114172000 waiting on condition [114171000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d32ad5e0> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at org.xnio.LimitedBlockingQueue.take(LimitedBlockingQueue.java:95) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Remoting "node-udp-0:MANAGEMENT" task-2" prio=5 tid=10b8fb800 nid=0x11406f000 waiting on condition [11406e000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d32ad5e0> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at org.xnio.LimitedBlockingQueue.take(LimitedBlockingQueue.java:95) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Remoting "node-udp-0:MANAGEMENT" task-1" prio=5 tid=10b8ae800 nid=0x113f6c000 waiting on condition [113f6b000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d32ad5e0> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at org.xnio.LimitedBlockingQueue.take(LimitedBlockingQueue.java:95) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Thread-74" prio=5 tid=109322800 nid=0x113e69000 runnable [113e68000] java.lang.Thread.State: RUNNABLE at sun.nio.ch.KQueueArrayWrapper.kevent0(Native Method) at sun.nio.ch.KQueueArrayWrapper.poll(KQueueArrayWrapper.java:136) at sun.nio.ch.KQueueSelectorImpl.doSelect(KQueueSelectorImpl.java:69) at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:69) - locked <7d3bc0470> (a sun.nio.ch.Util$2) - locked <7d3bc0458> (a java.util.Collections$UnmodifiableSet) - locked <7d3bebd70> (a sun.nio.ch.KQueueSelectorImpl) at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:80) at org.jboss.sun.net.httpserver.ServerImpl$Dispatcher.run(ServerImpl.java:382) at java.lang.Thread.run(Thread.java:680) "server-timer1" daemon prio=5 tid=101d37800 nid=0x113d0d000 in Object.wait() [113d0c000] java.lang.Thread.State: TIMED_WAITING (on object monitor) at java.lang.Object.wait(Native Method) - waiting on <7d3bc0758> (a java.util.TaskQueue) at java.util.TimerThread.mainLoop(Timer.java:509) - locked <7d3bc0758> (a java.util.TaskQueue) at java.util.TimerThread.run(Timer.java:462) "server-timer" daemon prio=5 tid=101d37000 nid=0x113c0a000 in Object.wait() [113c09000] java.lang.Thread.State: TIMED_WAITING (on object monitor) at java.lang.Object.wait(Native Method) - waiting on <7d3bc07d0> (a java.util.TaskQueue) at java.util.TimerThread.mainLoop(Timer.java:509) - locked <7d3bc07d0> (a java.util.TaskQueue) at java.util.TimerThread.run(Timer.java:462) "DeploymentScanner-threads - 1" prio=5 tid=101f33800 nid=0x110d6a000 waiting on condition [110d69000] java.lang.Thread.State: TIMED_WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d326f2a8> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:196) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2025) at java.util.concurrent.DelayQueue.take(DelayQueue.java:164) at java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:609) at java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:602) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) at org.jboss.threads.JBossThread.run(JBossThread.java:122) "Thread-3 (HornetQ-scheduled-threads-360560224)" prio=5 tid=101b4a000 nid=0x113b07000 runnable [113b06000] java.lang.Thread.State: TIMED_WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d36292c0> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:196) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2025) at java.util.concurrent.DelayQueue.take(DelayQueue.java:164) at java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:609) at java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:602) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Thread-2 (HornetQ-scheduled-threads-360560224)" prio=5 tid=101949000 nid=0x113a04000 runnable [113a03000] java.lang.Thread.State: TIMED_WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d36292c0> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:196) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2025) at java.util.concurrent.DelayQueue.take(DelayQueue.java:164) at java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:609) at java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:602) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "hornetq-failure-check-thread" prio=5 tid=10b979000 nid=0x113901000 in Object.wait() [113900000] java.lang.Thread.State: TIMED_WAITING (on object monitor) at java.lang.Object.wait(Native Method) - waiting on <7d3a432a0> (a org.hornetq.core.remoting.server.impl.RemotingServiceImpl$FailureCheckAndFlushThread) at org.hornetq.core.remoting.server.impl.RemotingServiceImpl$FailureCheckAndFlushThread.run(RemotingServiceImpl.java:632) - locked <7d3a432a0> (a org.hornetq.core.remoting.server.impl.RemotingServiceImpl$FailureCheckAndFlushThread) "Thread-1 (HornetQ-scheduled-threads-360560224)" prio=5 tid=10b8a7800 nid=0x11133b000 runnable [11133a000] java.lang.Thread.State: TIMED_WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d36292c0> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:196) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2025) at java.util.concurrent.DelayQueue.take(DelayQueue.java:164) at java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:609) at java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:602) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Old I/O server boss ([id: 0x7065ec20, localhost/127.0.0.1:5455])" prio=5 tid=10bc71000 nid=0x111235000 runnable [111234000] java.lang.Thread.State: RUNNABLE at java.net.PlainSocketImpl.socketAccept(Native Method) at java.net.PlainSocketImpl.accept(PlainSocketImpl.java:408) - locked <7d39da178> (a java.net.SocksSocketImpl) at java.net.ServerSocket.implAccept(ServerSocket.java:462) at java.net.ServerSocket.accept(ServerSocket.java:430) at org.jboss.netty.channel.socket.oio.OioServerSocketPipelineSink$Boss.run(OioServerSocketPipelineSink.java:202) at org.jboss.netty.util.ThreadRenamingRunnable.run(ThreadRenamingRunnable.java:108) at org.jboss.netty.util.internal.DeadLockProofWorker$1.run(DeadLockProofWorker.java:44) at org.jboss.netty.util.VirtualExecutorService$ChildExecutorRunnable.run(VirtualExecutorService.java:181) at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) at java.lang.Thread.run(Thread.java:680) "Old I/O server boss ([id: 0x39bbf341, localhost/127.0.0.1:5445])" prio=5 tid=1092e9800 nid=0x111132000 runnable [111131000] java.lang.Thread.State: RUNNABLE at java.net.PlainSocketImpl.socketAccept(Native Method) at java.net.PlainSocketImpl.accept(PlainSocketImpl.java:408) - locked <7d39da268> (a java.net.SocksSocketImpl) at java.net.ServerSocket.implAccept(ServerSocket.java:462) at java.net.ServerSocket.accept(ServerSocket.java:430) at org.jboss.netty.channel.socket.oio.OioServerSocketPipelineSink$Boss.run(OioServerSocketPipelineSink.java:202) at org.jboss.netty.util.ThreadRenamingRunnable.run(ThreadRenamingRunnable.java:108) at org.jboss.netty.util.internal.DeadLockProofWorker$1.run(DeadLockProofWorker.java:44) at org.jboss.netty.util.VirtualExecutorService$ChildExecutorRunnable.run(VirtualExecutorService.java:181) at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) at java.lang.Thread.run(Thread.java:680) "RequestController-3" daemon prio=10 tid=101b77000 nid=0x112ff2000 in Object.wait() [112ff1000] java.lang.Thread.State: WAITING (on object monitor) at java.lang.Object.wait(Native Method) - waiting on <7d394cef0> (a java.lang.Object) at java.lang.Object.wait(Object.java:485) at org.jacorb.poa.RequestController.waitForQueue(RequestController.java:581) - locked <7d394cef0> (a java.lang.Object) at org.jacorb.poa.RequestController.run(RequestController.java:498) "AOMRemoval" daemon prio=5 tid=101c50800 nid=0x112eef000 waiting on condition [112eee000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d394cfb0> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at org.jacorb.poa.AOM$1.run(AOM.java:120) "RequestController-2" daemon prio=10 tid=101c50000 nid=0x112dec000 in Object.wait() [112deb000] java.lang.Thread.State: WAITING (on object monitor) at java.lang.Object.wait(Native Method) - waiting on <7d394d598> (a java.lang.Object) at java.lang.Object.wait(Object.java:485) at org.jacorb.poa.RequestController.waitForQueue(RequestController.java:581) - locked <7d394d598> (a java.lang.Object) at org.jacorb.poa.RequestController.run(RequestController.java:498) "AOMRemoval" daemon prio=5 tid=101a8f800 nid=0x112ce9000 waiting on condition [112ce8000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d394d658> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at org.jacorb.poa.AOM$1.run(AOM.java:120) "ServerSocketListener" daemon prio=5 tid=10bc83800 nid=0x112be6000 runnable [112be5000] java.lang.Thread.State: RUNNABLE at java.net.PlainSocketImpl.socketAccept(Native Method) at java.net.PlainSocketImpl.accept(PlainSocketImpl.java:408) - locked <7d39da4c0> (a java.net.SocksSocketImpl) at java.net.ServerSocket.implAccept(ServerSocket.java:462) at java.net.ServerSocket.accept(ServerSocket.java:430) at org.jacorb.orb.iiop.IIOPListener$Acceptor.run(IIOPListener.java:467) "Thread-0 (HornetQ-scheduled-threads-360560224)" prio=5 tid=10bc81000 nid=0x112ae3000 runnable [112ae2000] java.lang.Thread.State: TIMED_WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d36292c0> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:196) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2025) at java.util.concurrent.DelayQueue.take(DelayQueue.java:164) at java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:609) at java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:602) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "pool-14-thread-1" prio=5 tid=10bc80000 nid=0x1129e0000 runnable [1129df000] java.lang.Thread.State: RUNNABLE at java.net.PlainDatagramSocketImpl.receive0(Native Method) - locked <7d3940180> (a java.net.PlainDatagramSocketImpl) at java.net.PlainDatagramSocketImpl.receive(PlainDatagramSocketImpl.java:145) - locked <7d3940180> (a java.net.PlainDatagramSocketImpl) at java.net.DatagramSocket.receive(DatagramSocket.java:725) - locked <7d4023d28> (a java.net.DatagramPacket) - locked <7d3ab53f0> (a java.net.MulticastSocket) at org.jboss.modcluster.advertise.impl.AdvertiseListenerImpl$AdvertiseListenerWorker.run(AdvertiseListenerImpl.java:369) at java.lang.Thread.run(Thread.java:680) "RequestController-1" daemon prio=10 tid=10bccd800 nid=0x1128dd000 in Object.wait() [1128dc000] java.lang.Thread.State: WAITING (on object monitor) at java.lang.Object.wait(Native Method) - waiting on <7d394d310> (a java.lang.Object) at java.lang.Object.wait(Object.java:485) at org.jacorb.poa.RequestController.waitForQueue(RequestController.java:581) - locked <7d394d310> (a java.lang.Object) at org.jacorb.poa.RequestController.run(RequestController.java:498) "AOMRemoval" daemon prio=5 tid=101d3e000 nid=0x1127da000 waiting on condition [1127d9000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d394d3d0> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at org.jacorb.poa.AOM$1.run(AOM.java:120) "Transaction Reaper Worker 0" daemon prio=5 tid=101b65000 nid=0x1126d7000 in Object.wait() [1126d6000] java.lang.Thread.State: WAITING (on object monitor) at java.lang.Object.wait(Native Method) - waiting on <7d394def8> (a java.util.LinkedList) at java.lang.Object.wait(Object.java:485) at com.arjuna.ats.arjuna.coordinator.TransactionReaper.waitForCancellations(TransactionReaper.java:321) - locked <7d394def8> (a java.util.LinkedList) at com.arjuna.ats.internal.arjuna.coordinator.ReaperWorkerThread.run(ReaperWorkerThread.java:65) "hornetq-expiry-reaper-thread" prio=3 tid=10bccc800 nid=0x1125d4000 in Object.wait() [1125d3000] java.lang.Thread.State: TIMED_WAITING (on object monitor) at java.lang.Object.wait(Native Method) - waiting on <7d36297e8> (a org.hornetq.core.postoffice.impl.PostOfficeImpl$Reaper) at org.hornetq.core.postoffice.impl.PostOfficeImpl$Reaper.run(PostOfficeImpl.java:1319) - locked <7d36297e8> (a org.hornetq.core.postoffice.impl.PostOfficeImpl$Reaper) at java.lang.Thread.run(Thread.java:680) "Transaction Reaper" daemon prio=5 tid=101b64000 nid=0x1124d1000 in Object.wait() [1124d0000] java.lang.Thread.State: TIMED_WAITING (on object monitor) at java.lang.Object.wait(Native Method) - waiting on <7d3940240> (a com.arjuna.ats.arjuna.coordinator.TransactionReaper) at com.arjuna.ats.internal.arjuna.coordinator.ReaperThread.run(ReaperThread.java:90) - locked <7d3940240> (a com.arjuna.ats.arjuna.coordinator.TransactionReaper) "http-localhost-127.0.0.1-8080-Acceptor" daemon prio=5 tid=10b8e8000 nid=0x112307000 runnable [112306000] java.lang.Thread.State: RUNNABLE at org.apache.tomcat.jni.Socket.accept(Native Method) at org.apache.tomcat.util.net.AprEndpoint$Acceptor.run(AprEndpoint.java:1130) at java.lang.Thread.run(Thread.java:680) "http-localhost-127.0.0.1-8080-Sendfile" daemon prio=5 tid=101d3f000 nid=0x112204000 in Object.wait() [112203000] java.lang.Thread.State: WAITING (on object monitor) at java.lang.Object.wait(Native Method) - waiting on <7d3626bc0> (a org.apache.tomcat.util.net.AprEndpoint$Sendfile) at java.lang.Object.wait(Object.java:485) at org.apache.tomcat.util.net.AprEndpoint$Sendfile.run(AprEndpoint.java:2258) - locked <7d3626bc0> (a org.apache.tomcat.util.net.AprEndpoint$Sendfile) at java.lang.Thread.run(Thread.java:680) "http-localhost-127.0.0.1-8080-EventPoller" daemon prio=5 tid=10bc61000 nid=0x112101000 in Object.wait() [112100000] java.lang.Thread.State: TIMED_WAITING (on object monitor) at java.lang.Object.wait(Native Method) - waiting on <7d34d8e18> (a org.apache.tomcat.util.net.AprEndpoint$Poller) at org.apache.tomcat.util.net.AprEndpoint$Poller.run(AprEndpoint.java:1678) - locked <7d34d8e18> (a org.apache.tomcat.util.net.AprEndpoint$Poller) at java.lang.Thread.run(Thread.java:680) "http-localhost-127.0.0.1-8080-Poller" daemon prio=5 tid=101b68800 nid=0x111755000 in Object.wait() [111754000] java.lang.Thread.State: TIMED_WAITING (on object monitor) at java.lang.Object.wait(Native Method) - waiting on <7d34d8dc8> (a org.apache.tomcat.util.net.AprEndpoint$Poller) at org.apache.tomcat.util.net.AprEndpoint$Poller.run(AprEndpoint.java:1678) - locked <7d34d8dc8> (a org.apache.tomcat.util.net.AprEndpoint$Poller) at java.lang.Thread.run(Thread.java:680) "ORB Run Thread" prio=5 tid=101b67800 nid=0x111652000 in Object.wait() [111651000] java.lang.Thread.State: WAITING (on object monitor) at java.lang.Object.wait(Native Method) - waiting on <7d33f5658> (a java.lang.Object) at java.lang.Object.wait(Object.java:485) at org.jacorb.orb.ORB.run(ORB.java:1417) - locked <7d33f5658> (a java.lang.Object) at org.jboss.as.jacorb.service.CorbaORBService$ORBRunner.run(CorbaORBService.java:215) at java.lang.Thread.run(Thread.java:680) "Periodic Recovery" prio=5 tid=10b993000 nid=0x110c67000 in Object.wait() [110c66000] java.lang.Thread.State: TIMED_WAITING (on object monitor) at java.lang.Object.wait(Native Method) - waiting on <7d3599640> (a java.lang.Object) at com.arjuna.ats.internal.arjuna.recovery.PeriodicRecovery.doPeriodicWait(PeriodicRecovery.java:672) at com.arjuna.ats.internal.arjuna.recovery.PeriodicRecovery.run(PeriodicRecovery.java:392) - locked <7d3599640> (a java.lang.Object) "Transaction Expired Entry Monitor" daemon prio=5 tid=10b97c000 nid=0x1114ef000 in Object.wait() [1114ee000] java.lang.Thread.State: TIMED_WAITING (on object monitor) at java.lang.Object.wait(Native Method) - waiting on <7d34e9950> (a com.arjuna.ats.internal.arjuna.recovery.ExpiredEntryMonitor) at com.arjuna.ats.internal.arjuna.recovery.ExpiredEntryMonitor.run(ExpiredEntryMonitor.java:190) - locked <7d34e9950> (a com.arjuna.ats.internal.arjuna.recovery.ExpiredEntryMonitor) "Timer-0" prio=5 tid=10b941000 nid=0x111019000 in Object.wait() [111018000] java.lang.Thread.State: WAITING (on object monitor) at java.lang.Object.wait(Native Method) - waiting on <7d3629b18> (a java.util.TaskQueue) at java.lang.Object.wait(Object.java:485) at java.util.TimerThread.mainLoop(Timer.java:483) - locked <7d3629b18> (a java.util.TaskQueue) at java.util.TimerThread.run(Timer.java:462) "ContainerBackgroundProcessor[StandardEngine[jboss.web]]" daemon prio=5 tid=101b4b000 nid=0x110f16000 waiting on condition [110f15000] java.lang.Thread.State: TIMED_WAITING (sleeping) at java.lang.Thread.sleep(Native Method) at org.apache.catalina.core.ContainerBase$ContainerBackgroundProcessor.run(ContainerBase.java:1579) at java.lang.Thread.run(Thread.java:680) "ConnectionValidator" daemon prio=5 tid=10b907800 nid=0x1106a0000 waiting on condition [11069f000] java.lang.Thread.State: TIMED_WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d33f5670> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:196) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2116) at org.jboss.jca.core.connectionmanager.pool.validator.ConnectionValidator$ConnectionValidatorRunner.run(ConnectionValidator.java:263) at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) at java.lang.Thread.run(Thread.java:680) "IdleRemover" daemon prio=5 tid=10b904800 nid=0x11059d000 waiting on condition [11059c000] java.lang.Thread.State: TIMED_WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d33e59a0> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:196) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2116) at org.jboss.jca.core.connectionmanager.pool.idle.IdleRemover$IdleRemoverRunner.run(IdleRemover.java:261) at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) at java.lang.Thread.run(Thread.java:680) "Remoting "node-udp-0" write-1" prio=5 tid=101c91800 nid=0x10fc71000 runnable [10fc70000] java.lang.Thread.State: RUNNABLE at sun.nio.ch.KQueueArrayWrapper.kevent0(Native Method) at sun.nio.ch.KQueueArrayWrapper.poll(KQueueArrayWrapper.java:136) at sun.nio.ch.KQueueSelectorImpl.doSelect(KQueueSelectorImpl.java:69) at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:69) - locked <7d31cf670> (a sun.nio.ch.Util$2) - locked <7d31cf658> (a java.util.Collections$UnmodifiableSet) - locked <7d326bab0> (a sun.nio.ch.KQueueSelectorImpl) at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:80) at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:84) at org.xnio.nio.WorkerThread.run(WorkerThread.java:153) "Remoting "node-udp-0" read-1" prio=5 tid=101c90800 nid=0x10fb6e000 runnable [10fb6d000] java.lang.Thread.State: RUNNABLE at sun.nio.ch.KQueueArrayWrapper.kevent0(Native Method) at sun.nio.ch.KQueueArrayWrapper.poll(KQueueArrayWrapper.java:136) at sun.nio.ch.KQueueSelectorImpl.doSelect(KQueueSelectorImpl.java:69) at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:69) - locked <7d319a478> (a sun.nio.ch.Util$2) - locked <7d319a460> (a java.util.Collections$UnmodifiableSet) - locked <7d326ba20> (a sun.nio.ch.KQueueSelectorImpl) at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:80) at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:84) at org.xnio.nio.WorkerThread.run(WorkerThread.java:153) "Remoting "node-udp-0:MANAGEMENT" write-1" prio=5 tid=101e54800 nid=0x10e61c000 runnable [10e61b000] java.lang.Thread.State: RUNNABLE at sun.nio.ch.KQueueArrayWrapper.kevent0(Native Method) at sun.nio.ch.KQueueArrayWrapper.poll(KQueueArrayWrapper.java:136) at sun.nio.ch.KQueueSelectorImpl.doSelect(KQueueSelectorImpl.java:69) at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:69) - locked <7d32ab708> (a sun.nio.ch.Util$2) - locked <7d32ab6f0> (a java.util.Collections$UnmodifiableSet) - locked <7d326b6c0> (a sun.nio.ch.KQueueSelectorImpl) at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:80) at org.xnio.nio.WorkerThread.run(WorkerThread.java:157) "Remoting "node-udp-0:MANAGEMENT" read-1" prio=5 tid=101d6a000 nid=0x10e519000 runnable [10e518000] java.lang.Thread.State: RUNNABLE at sun.nio.ch.KQueueArrayWrapper.kevent0(Native Method) at sun.nio.ch.KQueueArrayWrapper.poll(KQueueArrayWrapper.java:136) at sun.nio.ch.KQueueSelectorImpl.doSelect(KQueueSelectorImpl.java:69) at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:69) - locked <7d32ad490> (a sun.nio.ch.Util$2) - locked <7d32ad478> (a java.util.Collections$UnmodifiableSet) - locked <7d326b750> (a sun.nio.ch.KQueueSelectorImpl) at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:80) at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:84) at org.xnio.nio.WorkerThread.run(WorkerThread.java:153) "DestroyJavaVM" prio=5 tid=101801800 nid=0x100601000 waiting on condition [00000000] java.lang.Thread.State: RUNNABLE "Poller SunPKCS11-Darwin" daemon prio=1 tid=1097e7000 nid=0x10c334000 waiting on condition [10c333000] java.lang.Thread.State: TIMED_WAITING (sleeping) at java.lang.Thread.sleep(Native Method) at sun.security.pkcs11.SunPKCS11$TokenPoller.run(SunPKCS11.java:692) at java.lang.Thread.run(Thread.java:680) "MSC service thread 1-4" prio=5 tid=109249800 nid=0x10c231000 waiting on condition [10c230000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d2ae2430> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "MSC service thread 1-3" prio=5 tid=109248800 nid=0x10c101000 in Object.wait() [10c100000] java.lang.Thread.State: WAITING (on object monitor) at java.lang.Object.wait(Native Method) - waiting on <7d3ce7710> (a org.jboss.msc.service.ServiceControllerImpl) at java.lang.Object.wait(Object.java:485) at org.jboss.as.test.clustering.unmanaged.singleton.service.MyServiceContextListener.wait(MyServiceContextListener.java:55) - locked <7d3ce7710> (a org.jboss.msc.service.ServiceControllerImpl) at org.jboss.as.test.clustering.unmanaged.singleton.service.MyServiceContextListener.contextDestroyed(MyServiceContextListener.java:44) at org.apache.catalina.core.StandardContext.listenerStop(StandardContext.java:3489) at org.apache.catalina.core.StandardContext.stop(StandardContext.java:3999) - locked <7d3bc7438> (a org.apache.catalina.core.StandardContext) at org.jboss.as.web.deployment.WebDeploymentService.stop(WebDeploymentService.java:96) - locked <7d3d7d360> (a org.jboss.as.web.deployment.WebDeploymentService) at org.jboss.msc.service.ServiceControllerImpl$StopTask.stopService(ServiceControllerImpl.java:1909) at org.jboss.msc.service.ServiceControllerImpl$StopTask.run(ServiceControllerImpl.java:1872) at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) at java.lang.Thread.run(Thread.java:680) "MSC service thread 1-2" prio=5 tid=101d22800 nid=0x10b477000 waiting on condition [10b476000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d2ae2430> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "MSC service thread 1-1" prio=5 tid=1097f7000 nid=0x10b374000 waiting on condition [10b373000] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <7d2ae2430> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907) at java.lang.Thread.run(Thread.java:680) "Reference Reaper" daemon prio=5 tid=101a0b800 nid=0x10b001000 in Object.wait() [10b000000] java.lang.Thread.State: WAITING (on object monitor) at java.lang.Object.wait(Native Method) - waiting on <7d2aaa0e8> (a java.lang.ref.ReferenceQueue$Lock) at java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:118) - locked <7d2aaa0e8> (a java.lang.ref.ReferenceQueue$Lock) at java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:134) at org.jboss.modules.ref.References$ReaperThread.run(References.java:68) "Low Memory Detector" daemon prio=5 tid=1018a2000 nid=0x10ac0a000 runnable [00000000] java.lang.Thread.State: RUNNABLE "C2 CompilerThread1" daemon prio=9 tid=1018a1800 nid=0x10ab07000 waiting on condition [00000000] java.lang.Thread.State: RUNNABLE "C2 CompilerThread0" daemon prio=9 tid=10901f000 nid=0x10aa04000 waiting on condition [00000000] java.lang.Thread.State: RUNNABLE "Signal Dispatcher" daemon prio=9 tid=10901e000 nid=0x10a901000 waiting on condition [00000000] java.lang.Thread.State: RUNNABLE "Surrogate Locker Thread (Concurrent GC)" daemon prio=5 tid=10901d800 nid=0x109f01000 waiting on condition [00000000] java.lang.Thread.State: RUNNABLE "Finalizer" daemon prio=8 tid=109009800 nid=0x109c07000 in Object.wait() [109c06000] java.lang.Thread.State: WAITING (on object monitor) at java.lang.Object.wait(Native Method) - waiting on <7d2af3cd0> (a java.lang.ref.ReferenceQueue$Lock) at java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:118) - locked <7d2af3cd0> (a java.lang.ref.ReferenceQueue$Lock) at java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:134) at java.lang.ref.Finalizer$FinalizerThread.run(Finalizer.java:159) "Reference Handler" daemon prio=10 tid=109009000 nid=0x109b04000 in Object.wait() [109b03000] java.lang.Thread.State: WAITING (on object monitor) at java.lang.Object.wait(Native Method) - waiting on <7d29b0848> (a java.lang.ref.Reference$Lock) at java.lang.Object.wait(Object.java:485) at java.lang.ref.Reference$ReferenceHandler.run(Reference.java:116) - locked <7d29b0848> (a java.lang.ref.Reference$Lock) "VM Thread" prio=9 tid=109004000 nid=0x109a01000 runnable "Gang worker#0 (Parallel GC Threads)" prio=9 tid=101802800 nid=0x1017f9000 runnable "Gang worker#1 (Parallel GC Threads)" prio=9 tid=101803800 nid=0x105102000 runnable "Concurrent Mark-Sweep GC Thread" prio=9 tid=10184e000 nid=0x108bcc000 runnable "VM Periodic Task Thread" prio=10 tid=1018ac000 nid=0x10ad0d000 waiting on condition "Exception Catcher Thread" prio=10 tid=101802000 nid=0x100704000 runnable JNI global references: 1408 Heap par new generation total 19136K, used 11963K [7d0000000, 7d14c0000, 7d2990000) eden space 17024K, 69% used [7d0000000, 7d0b92900, 7d10a0000) from space 2112K, 5% used [7d12b0000, 7d12cc3e8, 7d14c0000) to space 2112K, 0% used [7d10a0000, 7d10a0000, 7d12b0000) concurrent mark-sweep generation total 63872K, used 24311K [7d2990000, 7d67f0000, 7f0000000) concurrent-mark-sweep perm gen total 70400K, used 70294K [7f0000000, 7f44c0000, 800000000) 20:33:24,444 ERROR [org.infinispan.cacheviews.CacheViewsManagerImpl] (CacheViewInstaller-1,node-udp-0/cluster) ISPN000172: Failed to prepare view CacheView{viewId=1218, members=[node-udp-0/cluster, node-udp-1/cluster]} for cache default, rol