<<<<<<<<<<<<<<<<<<<< START jenkins-openshift-ansible-1-provision-jslave-717 >>>>>>>>>>>>>>>>>> Started by upstream project "openshift-ansible-0-multijob" build number 731 originally caused by: GitHub pull request #3043 of commit ae88c63dda7c433cd2b42ad7d9527b483579dc4e, no merge conflicts. [EnvInject] - Loading node environment variables. Building on master in workspace /jenkins-data/jenkins/jobs/openshift-ansible-1-provision-jslave/workspace [WS-CLEANUP] Deleting project workspace... [WS-CLEANUP] Done Wiping out workspace first. Cloning the remote Git repository Cloning repository XXXXXXXXXXXXXXXXXXXXXXXXXXX/gerrit/atomic-ci-jobs > git init /jenkins-data/jenkins/jobs/openshift-ansible-1-provision-jslave/workspace/atomic-ci-jobs # timeout=10 Fetching upstream changes from XXXXXXXXXXXXXXXXXXXXXXXXXXX/gerrit/atomic-ci-jobs > git --version # timeout=10 > git fetch --tags --progress XXXXXXXXXXXXXXXXXXXXXXXXXXX/gerrit/atomic-ci-jobs +refs/heads/*:refs/remotes/origin/* > git config remote.origin.url XXXXXXXXXXXXXXXXXXXXXXXXXXX/gerrit/atomic-ci-jobs # timeout=10 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10 > git config remote.origin.url XXXXXXXXXXXXXXXXXXXXXXXXXXX/gerrit/atomic-ci-jobs # timeout=10 Fetching upstream changes from XXXXXXXXXXXXXXXXXXXXXXXXXXX/gerrit/atomic-ci-jobs > git fetch --tags --progress XXXXXXXXXXXXXXXXXXXXXXXXXXX/gerrit/atomic-ci-jobs +refs/heads/*:refs/remotes/origin/* > git rev-parse refs/remotes/origin/master^{commit} # timeout=10 > git rev-parse refs/remotes/origin/origin/master^{commit} # timeout=10 Checking out Revision 758ad5b12c38a0af1f9f10895e999af27a3a1541 (refs/remotes/origin/master) > git config core.sparsecheckout # timeout=10 > git checkout -f 758ad5b12c38a0af1f9f10895e999af27a3a1541 > git rev-list 758ad5b12c38a0af1f9f10895e999af27a3a1541 # timeout=10 Wiping out workspace first. Cloning the remote Git repository Cloning repository XXXXXXXXXXXXXXXXXXXXXXXXXXX/gerrit/ci-factory > git init /jenkins-data/jenkins/jobs/openshift-ansible-1-provision-jslave/workspace/ci-factory # timeout=10 Fetching upstream changes from XXXXXXXXXXXXXXXXXXXXXXXXXXX/gerrit/ci-factory > git --version # timeout=10 > git fetch --tags --progress XXXXXXXXXXXXXXXXXXXXXXXXXXX/gerrit/ci-factory +refs/heads/*:refs/remotes/origin/* > git config remote.origin.url XXXXXXXXXXXXXXXXXXXXXXXXXXX/gerrit/ci-factory # timeout=10 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10 > git config remote.origin.url XXXXXXXXXXXXXXXXXXXXXXXXXXX/gerrit/ci-factory # timeout=10 Fetching upstream changes from XXXXXXXXXXXXXXXXXXXXXXXXXXX/gerrit/ci-factory > git fetch --tags --progress XXXXXXXXXXXXXXXXXXXXXXXXXXX/gerrit/ci-factory +refs/heads/*:refs/remotes/origin/* > git rev-parse refs/remotes/origin/master^{commit} # timeout=10 > git rev-parse refs/remotes/origin/origin/master^{commit} # timeout=10 Checking out Revision 0d870824ed1657c03b43a31d0f9099dbccb107cb (refs/remotes/origin/master) > git config core.sparsecheckout # timeout=10 > git checkout -f 0d870824ed1657c03b43a31d0f9099dbccb107cb > git rev-list 0d870824ed1657c03b43a31d0f9099dbccb107cb # timeout=10 Wiping out workspace first. Cloning the remote Git repository Cloning repository XXXXXXXXXXXXXXXXXXXXXXXXXXX/gerrit/job-runner > git init /jenkins-data/jenkins/jobs/openshift-ansible-1-provision-jslave/workspace/job-runner # timeout=10 Fetching upstream changes from XXXXXXXXXXXXXXXXXXXXXXXXXXX/gerrit/job-runner > git --version # timeout=10 > git fetch --tags --progress XXXXXXXXXXXXXXXXXXXXXXXXXXX/gerrit/job-runner +refs/heads/*:refs/remotes/origin/* > git config remote.origin.url XXXXXXXXXXXXXXXXXXXXXXXXXXX/gerrit/job-runner # timeout=10 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10 > git config remote.origin.url XXXXXXXXXXXXXXXXXXXXXXXXXXX/gerrit/job-runner # timeout=10 Fetching upstream changes from XXXXXXXXXXXXXXXXXXXXXXXXXXX/gerrit/job-runner > git fetch --tags --progress XXXXXXXXXXXXXXXXXXXXXXXXXXX/gerrit/job-runner +refs/heads/*:refs/remotes/origin/* > git rev-parse refs/remotes/origin/master^{commit} # timeout=10 > git rev-parse refs/remotes/origin/origin/master^{commit} # timeout=10 Checking out Revision 9823d84d2db1b15cd82f1b60a2d1fe594fc8ece9 (refs/remotes/origin/master) > git config core.sparsecheckout # timeout=10 > git checkout -f 9823d84d2db1b15cd82f1b60a2d1fe594fc8ece9 > git rev-list 9823d84d2db1b15cd82f1b60a2d1fe594fc8ece9 # timeout=10 Run condition [Current build status] enabling prebuild for step [[Groovy Postbuild]] Run condition [Current build status] enabling prebuild for step [[Groovy Postbuild, Execute a set of scripts, Execute a set of scripts]] [EnvInject] - Executing scripts and injecting environment variables after the SCM step. [EnvInject] - Injecting as environment variables the properties content ANSIBLE_HOST_KEY_CHECKING=False ANSIBLE_PRIVATE_KEY_FILE=$WORKSPACE/atomic-ci-jobs/project/config/keys/ci-factory ANSIBLE_INVENTORY=$WORKSPACE/ci-factory/utils/central_ci_dynamic_hosts.py [EnvInject] - Variables injected successfully. [EnvInject] - Injecting environment variables from a build step. [EnvInject] - Injecting as environment variables the properties content GH_STATE=pending GH_CONTEXT=aos-ci-jenkins/OS_unit_tests GH_DESC='Running unit tests' GH_URL=https://aos-ci.s3.amazonaws.com/openshift/openshift-ansible/jenkins-openshift-ansible-1-provision-jslave-717/ae88c63dda7c433cd2b42ad7d9527b483579dc4e.txt [EnvInject] - Variables injected successfully. [workspace] $ python /tmp/hudson8335684563895674776.py /usr/lib/python2.6/site-packages/requests/packages/urllib3/util/ssl_.py:315: SNIMissingWarning: An HTTPS request has been made, but the SNI (Subject Name Indication) extension to TLS is not available on this platform. This may cause the server to present an incorrect TLS certificate, which can cause validation failures. For more information, see https://urllib3.readthedocs.org/en/latest/security.html#snimissingwarning. SNIMissingWarning /usr/lib/python2.6/site-packages/requests/packages/urllib3/util/ssl_.py:120: InsecurePlatformWarning: A true SSLContext object is not available. This prevents urllib3 from configuring SSL appropriately and may cause certain SSL connections to fail. For more information, see https://urllib3.readthedocs.org/en/latest/security.html#insecureplatformwarning. InsecurePlatformWarning Updating status for ae88c63dda7c433cd2b42ad7d9527b483579dc4e [workspace] $ /bin/bash /tmp/hudson4119383533980690358.sh + /jenkins-data/jenkins/jobs/openshift-ansible-1-provision-jslave/workspace/ci-factory/bootstrap/provision_jslave.sh --site=ci-osp --project_defaults=atomic-ci-jobs/project/config/project_defaults_osp7_ose --topology=atomic-ci-jobs/project/config/openshift_jslave_RHEL7-1 --ssh_keyfile=atomic-ci-jobs/project/config/keys/ci-factory --jslave_execs=10 --jslavename=openshift-ansible-slave --jslavecreate --skip_cust --resources_file=openshift-ansible-slave.json Provisioning with the following environment ------------------------------------------- SITE: ci-osp PROJECT_DEFAULTS: /jenkins-data/jenkins/jobs/openshift-ansible-1-provision-jslave/workspace/atomic-ci-jobs/project/config/project_defaults_osp7_ose.json TOPOLOGY: /jenkins-data/jenkins/jobs/openshift-ansible-1-provision-jslave/workspace/atomic-ci-jobs/project/config/openshift_jslave_RHEL7-1.json SSH_KEYFILE: /jenkins-data/jenkins/jobs/openshift-ansible-1-provision-jslave/workspace/atomic-ci-jobs/project/config/keys/ci-factory WORKSPACE: /jenkins-data/jenkins/jobs/openshift-ansible-1-provision-jslave/workspace JSLAVENAME: openshift-ansible-slave JSLAVELABEL: openshift-ansible-slave LABEL: openshift-ansible-slave JSWARM_VER: 2.0 JSWARM_JAR_LOC: /root JSWARM_USE_JENKINS: False JSWARM_DISABLE_UNIQUE_ID: True JSLAVEIMAGE: JSLAVEFLAVOR: JSLAVE_EXECS: 10 JSLAVECREATE: True SKIP_EPEL: False SKIP_CUSTOMIZATION: True SKIP_ANSIBLE: False RESOURCES_FILE: /jenkins-data/jenkins/jobs/openshift-ansible-1-provision-jslave/workspace/openshift-ansible-slave.json CLEANUP: on_failure JENKINS_MASTER_URL: XXXXXXXXXXXXXXXXXXXXXXXXXXX JSLAVENAME: openshift-ansible-slave JSLAVELABEL: openshift-ansible-slave JENKINS_CLI: /var/cache/jenkins/war/WEB-INF/jenkins-cli.jar JSLAVEIP=10.8.168.47 JSLAVENAME=openshift-ansible-slave JSLAVELABEL=openshift-ansible-slave ++ ls /jenkins-data/jenkins/jobs/openshift-ansible-1-provision-jslave/workspace/openshift-ansible-slave.slave + files=/jenkins-data/jenkins/jobs/openshift-ansible-1-provision-jslave/workspace/openshift-ansible-slave.slave + '[' -e /jenkins-data/jenkins/jobs/openshift-ansible-1-provision-jslave/workspace/openshift-ansible-slave.slave ']' + cat /jenkins-data/jenkins/jobs/openshift-ansible-1-provision-jslave/workspace/openshift-ansible-slave.slave [EnvInject] - Injecting environment variables from a build step. [EnvInject] - Injecting as environment variables the properties file path '/jenkins-data/jenkins/jobs/openshift-ansible-1-provision-jslave/workspace/RESOURCES.txt' [EnvInject] - Variables injected successfully. [EnvInject] - Injecting as environment variables the properties content PROJECT_DEFAULTS=atomic-ci-jobs/project/config/project_defaults_osp7_ose SSH_KEYFILE=atomic-ci-jobs/project/config/keys/ci-factory JSLAVE_TOPOLOGY_PATH=atomic-ci-jobs/project/config JSLAVE_TOPOLOGY=openshift_jslave_RHEL7-1 JSLAVE_USERNAME=root JSLAVE_TEARDOWN=False [EnvInject] - Variables injected successfully. [PostBuildScript] - Execution post build scripts. [workspace] $ /bin/bash /tmp/hudson1556676967465602105.sh <<<<<<<<<<<<<<<<<<<< END jenkins-openshift-ansible-1-provision-jslave-717 >>>>>>>>>>>>>>>>>> <<<<<<<<<<<<<<<<<<<< START jenkins-openshift-ansible-3-test-matrix-CONTAINERIZED=_containerized,OSE_VER=3.3,PYTHON=System-CPython-2.7,TOPOLOGY=openshift-cluster-containerized,TargetBranch=master,nodes=openshift-ansible-slave-688 >>>>>>>>>>>>>>>>>> Started by upstream project "openshift-ansible-3-test-matrix" build number 688 originally caused by: Started by upstream project "openshift-ansible-0-multijob" build number 731 originally caused by: GitHub pull request #3043 of commit ae88c63dda7c433cd2b42ad7d9527b483579dc4e, no merge conflicts. [EnvInject] - Loading node environment variables. Building remotely on openshift-ansible-slave (swarm) in workspace /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave [WS-CLEANUP] Deleting project workspace... Wiping out workspace first. Cloning the remote Git repository Cloning repository XXXXXXXXXXXXXXXXXXXXXXXXXXX/gerrit/atomic-ci-jobs > git init /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs # timeout=10 Fetching upstream changes from XXXXXXXXXXXXXXXXXXXXXXXXXXX/gerrit/atomic-ci-jobs > git --version # timeout=10 > git -c core.askpass=true fetch --tags --progress XXXXXXXXXXXXXXXXXXXXXXXXXXX/gerrit/atomic-ci-jobs +refs/heads/*:refs/remotes/origin/* > git config remote.origin.url XXXXXXXXXXXXXXXXXXXXXXXXXXX/gerrit/atomic-ci-jobs # timeout=10 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10 > git config remote.origin.url XXXXXXXXXXXXXXXXXXXXXXXXXXX/gerrit/atomic-ci-jobs # timeout=10 Fetching upstream changes from XXXXXXXXXXXXXXXXXXXXXXXXXXX/gerrit/atomic-ci-jobs > git -c core.askpass=true fetch --tags --progress XXXXXXXXXXXXXXXXXXXXXXXXXXX/gerrit/atomic-ci-jobs +refs/heads/*:refs/remotes/origin/* Checking out Revision 758ad5b12c38a0af1f9f10895e999af27a3a1541 (refs/remotes/origin/master) > git config core.sparsecheckout # timeout=10 > git checkout -f 758ad5b12c38a0af1f9f10895e999af27a3a1541 > git rev-list 758ad5b12c38a0af1f9f10895e999af27a3a1541 # timeout=10 Wiping out workspace first. Cloning the remote Git repository Cloning repository XXXXXXXXXXXXXXXXXXXXXXXXXXX/gerrit/ci-factory > git init /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/ci-factory # timeout=10 Fetching upstream changes from XXXXXXXXXXXXXXXXXXXXXXXXXXX/gerrit/ci-factory > git --version # timeout=10 > git -c core.askpass=true fetch --tags --progress XXXXXXXXXXXXXXXXXXXXXXXXXXX/gerrit/ci-factory +refs/heads/*:refs/remotes/origin/* > git config remote.origin.url XXXXXXXXXXXXXXXXXXXXXXXXXXX/gerrit/ci-factory # timeout=10 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10 > git config remote.origin.url XXXXXXXXXXXXXXXXXXXXXXXXXXX/gerrit/ci-factory # timeout=10 Fetching upstream changes from XXXXXXXXXXXXXXXXXXXXXXXXXXX/gerrit/ci-factory > git -c core.askpass=true fetch --tags --progress XXXXXXXXXXXXXXXXXXXXXXXXXXX/gerrit/ci-factory +refs/heads/*:refs/remotes/origin/* Checking out Revision 0d870824ed1657c03b43a31d0f9099dbccb107cb (refs/remotes/origin/master) > git config core.sparsecheckout # timeout=10 > git checkout -f 0d870824ed1657c03b43a31d0f9099dbccb107cb > git rev-list 0d870824ed1657c03b43a31d0f9099dbccb107cb # timeout=10 Wiping out workspace first. Cloning the remote Git repository Cloning repository XXXXXXXXXXXXXXXXXXXXXXXXXXX/gerrit/job-runner > git init /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/job-runner # timeout=10 Fetching upstream changes from XXXXXXXXXXXXXXXXXXXXXXXXXXX/gerrit/job-runner > git --version # timeout=10 > git -c core.askpass=true fetch --tags --progress XXXXXXXXXXXXXXXXXXXXXXXXXXX/gerrit/job-runner +refs/heads/*:refs/remotes/origin/* > git config remote.origin.url XXXXXXXXXXXXXXXXXXXXXXXXXXX/gerrit/job-runner # timeout=10 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10 > git config remote.origin.url XXXXXXXXXXXXXXXXXXXXXXXXXXX/gerrit/job-runner # timeout=10 Fetching upstream changes from XXXXXXXXXXXXXXXXXXXXXXXXXXX/gerrit/job-runner > git -c core.askpass=true fetch --tags --progress XXXXXXXXXXXXXXXXXXXXXXXXXXX/gerrit/job-runner +refs/heads/*:refs/remotes/origin/* Checking out Revision 9823d84d2db1b15cd82f1b60a2d1fe594fc8ece9 (refs/remotes/origin/master) > git config core.sparsecheckout # timeout=10 > git checkout -f 9823d84d2db1b15cd82f1b60a2d1fe594fc8ece9 > git rev-list 9823d84d2db1b15cd82f1b60a2d1fe594fc8ece9 # timeout=10 Wiping out workspace first. Cloning the remote Git repository Cloning repository https://github.com/openshift/openshift-ansible.git > git init /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible # timeout=10 Fetching upstream changes from https://github.com/openshift/openshift-ansible.git > git --version # timeout=10 > git -c core.askpass=true fetch --tags --progress https://github.com/openshift/openshift-ansible.git +refs/heads/*:refs/remotes/origin/* > git config remote.origin.url https://github.com/openshift/openshift-ansible.git # timeout=10 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10 > git config remote.origin.url https://github.com/openshift/openshift-ansible.git # timeout=10 Fetching upstream changes from https://github.com/openshift/openshift-ansible.git > git -c core.askpass=true fetch --tags --progress https://github.com/openshift/openshift-ansible.git +refs/head/*:refs/remotes/origin/* +refs/pull/*:refs/remotes/origin/pr/* Checking out Revision 696d7e0f3bb85e9133f20a5b9fc58ada9231a780 (refs/remotes/origin/pr/3043/merge) > git config core.sparsecheckout # timeout=10 > git checkout -f 696d7e0f3bb85e9133f20a5b9fc58ada9231a780 First time build. Skipping changelog. Run condition [Current build status] enabling prebuild for step [[Groovy Postbuild]] Run condition [Current build status] enabling prebuild for step [[Groovy Postbuild, Execute a set of scripts]] [EnvInject] - Executing scripts and injecting environment variables after the SCM step. [EnvInject] - Injecting as environment variables the properties content ANSIBLE_HOST_KEY_CHECKING=False ANSIBLE_PRIVATE_KEY_FILE=$WORKSPACE/atomic-ci-jobs/project/config/keys/ci-factory ANSIBLE_INVENTORY=$WORKSPACE/ci-factory/utils/central_ci_dynamic_hosts.py [EnvInject] - Variables injected successfully. Copied 2 artifacts from "openshift-ansible-1-provision-jslave" build number 717 [EnvInject] - Injecting environment variables from a build step. [EnvInject] - Injecting as environment variables the properties file path '/home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/RESOURCES.txt' [EnvInject] - Variables injected successfully. [EnvInject] - Injecting as environment variables the properties content SITE=ci-osp PROJECT_NAME=openshift-ansible TOPOLOGY_PATH=atomic-ci-jobs/project/config PROJECT_DEFAULTS=atomic-ci-jobs/project/config/project_defaults_osp7_ose SSH_KEYFILE=atomic-ci-jobs/project/config/keys/ci-factory JSLAVE_TOPOLOGY_PATH=atomic-ci-jobs/project/config JSLAVE_TOPOLOGY=openshift_jslave_RHEL7-1 JSLAVE_USERNAME=root JSLAVE_TEARDOWN=False GH_STATE=pending GH_DESC="openshift-ansible install running" GH_CONTEXT=aos-ci-jenkins/OS_3.3_containerized ALL_GH_CONTEXT=aos-ci-jenkins/OS_3.3_containerized GH_URL=https://aos-ci.s3.amazonaws.com/openshift/openshift-ansible/jenkins-openshift-ansible-3-test-matrix-CONTAINERIZED=_containerized,OSE_VER=3.3,PYTHON=System-CPython-2.7,TOPOLOGY=openshift-cluster-containerized,TargetBranch=master,nodes=openshift-ansible-slave-688/ae88c63dda7c433cd2b42ad7d9527b483579dc4e.txt [EnvInject] - Variables injected successfully. [openshift-ansible-slave] $ python /tmp/hudson1383379777275469531.py Updating status for ae88c63dda7c433cd2b42ad7d9527b483579dc4e [EnvInject] - Injecting environment variables from a build step. [EnvInject] - Injecting as environment variables the properties file path '/home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/RESOURCES.txt' [EnvInject] - Variables injected successfully. [EnvInject] - Injecting as environment variables the properties content GH_STATE=error GH_DESC="Internal Jenkins error during provisioning an 3.3 cluster" GH_CONTEXT=aos-ci-jenkins/OS_3.3_containerized [EnvInject] - Variables injected successfully. [openshift-ansible-slave] $ /usr/local/bin/python2.7 /home/jenkins/shiningpanda/jobs/501d0aeb/virtualenv.py /home/jenkins/shiningpanda/jobs/501d0aeb/virtualenvs/d41d8cd9 New python executable in /home/jenkins/shiningpanda/jobs/501d0aeb/virtualenvs/d41d8cd9/bin/python2.7 Also creating executable in /home/jenkins/shiningpanda/jobs/501d0aeb/virtualenvs/d41d8cd9/bin/python Please make sure you remove any previous custom paths from your /root/.pydistutils.cfg file. Installing setuptools, pip, wheel...done. [openshift-ansible-slave] $ /bin/sh -xe /tmp/shiningpanda5631475361250209626.sh + set -xeuo pipefail + pip install ansible==1.9.6 DEPRECATION: --no-use-wheel is deprecated and will be removed in the future. Please use --no-binary :all: instead. Collecting ansible==1.9.6 Using cached ansible-1.9.6.tar.gz Collecting paramiko (from ansible==1.9.6) Using cached paramiko-2.1.1.tar.gz Collecting jinja2 (from ansible==1.9.6) Using cached Jinja2-2.9.4.tar.gz Collecting PyYAML (from ansible==1.9.6) Using cached PyYAML-3.12.tar.gz Requirement already satisfied: setuptools in /home/jenkins/shiningpanda/jobs/501d0aeb/virtualenvs/d41d8cd9/lib/python2.7/site-packages (from ansible==1.9.6) Collecting pycrypto>=2.6 (from ansible==1.9.6) Using cached pycrypto-2.6.1.tar.gz Collecting cryptography>=1.1 (from paramiko->ansible==1.9.6) Using cached cryptography-1.7.1.tar.gz Collecting pyasn1>=0.1.7 (from paramiko->ansible==1.9.6) Using cached pyasn1-0.1.9.tar.gz Collecting MarkupSafe>=0.23 (from jinja2->ansible==1.9.6) Using cached MarkupSafe-0.23.tar.gz Collecting idna>=2.0 (from cryptography>=1.1->paramiko->ansible==1.9.6) Using cached idna-2.2.tar.gz Collecting six>=1.4.1 (from cryptography>=1.1->paramiko->ansible==1.9.6) Using cached six-1.10.0.tar.gz Collecting enum34 (from cryptography>=1.1->paramiko->ansible==1.9.6) Using cached enum34-1.1.6.tar.gz Collecting ipaddress (from cryptography>=1.1->paramiko->ansible==1.9.6) Using cached ipaddress-1.0.17.tar.gz Collecting cffi>=1.4.1 (from cryptography>=1.1->paramiko->ansible==1.9.6) Using cached cffi-1.9.1.tar.gz Collecting pycparser (from cffi>=1.4.1->cryptography>=1.1->paramiko->ansible==1.9.6) Using cached pycparser-2.17.tar.gz Skipping bdist_wheel for ansible, due to binaries being disabled for it. Skipping bdist_wheel for paramiko, due to binaries being disabled for it. Skipping bdist_wheel for jinja2, due to binaries being disabled for it. Skipping bdist_wheel for PyYAML, due to binaries being disabled for it. Skipping bdist_wheel for pycrypto, due to binaries being disabled for it. Skipping bdist_wheel for cryptography, due to binaries being disabled for it. Skipping bdist_wheel for pyasn1, due to binaries being disabled for it. Skipping bdist_wheel for MarkupSafe, due to binaries being disabled for it. Skipping bdist_wheel for idna, due to binaries being disabled for it. Skipping bdist_wheel for six, due to binaries being disabled for it. Skipping bdist_wheel for enum34, due to binaries being disabled for it. Skipping bdist_wheel for ipaddress, due to binaries being disabled for it. Skipping bdist_wheel for cffi, due to binaries being disabled for it. Skipping bdist_wheel for pycparser, due to binaries being disabled for it. Installing collected packages: idna, pyasn1, six, enum34, ipaddress, pycparser, cffi, cryptography, paramiko, MarkupSafe, jinja2, PyYAML, pycrypto, ansible Running setup.py install for idna: started Running setup.py install for idna: finished with status 'done' Running setup.py install for pyasn1: started Running setup.py install for pyasn1: finished with status 'done' Running setup.py install for six: started Running setup.py install for six: finished with status 'done' Running setup.py install for enum34: started Running setup.py install for enum34: finished with status 'done' Running setup.py install for ipaddress: started Running setup.py install for ipaddress: finished with status 'done' Running setup.py install for pycparser: started Running setup.py install for pycparser: finished with status 'done' Running setup.py install for cffi: started Running setup.py install for cffi: finished with status 'done' Running setup.py install for cryptography: started Running setup.py install for cryptography: finished with status 'done' Running setup.py install for paramiko: started Running setup.py install for paramiko: finished with status 'done' Running setup.py install for MarkupSafe: started Running setup.py install for MarkupSafe: finished with status 'done' Running setup.py install for jinja2: started Running setup.py install for jinja2: finished with status 'done' Running setup.py install for PyYAML: started Running setup.py install for PyYAML: finished with status 'done' Running setup.py install for pycrypto: started Running setup.py install for pycrypto: finished with status 'done' Running setup.py install for ansible: started Running setup.py install for ansible: finished with status 'done' Successfully installed MarkupSafe-0.23 PyYAML-3.12 ansible-1.9.6 cffi-1.9.1 cryptography-1.7.1 enum34-1.1.6 idna-2.2 ipaddress-1.0.17 jinja2-2.9.4 paramiko-2.1.1 pyasn1-0.1.9 pycparser-2.17 pycrypto-2.6.1 six-1.10.0 + pip install taskrunner lxml configobj requests foreman python-foreman==0.4.5 python-glanceclient python-keystoneclient python-neutronclient DEPRECATION: --no-use-wheel is deprecated and will be removed in the future. Please use --no-binary :all: instead. Collecting taskrunner Using cached taskrunner-0.2.1.tar.gz Collecting lxml Using cached lxml-3.7.2.tar.gz Collecting configobj Using cached configobj-5.0.6.tar.gz Collecting requests Using cached requests-2.12.4.tar.gz Collecting foreman Using cached foreman-0.9.7.tar.gz Collecting python-foreman==0.4.5 Using cached python-foreman-0.4.5.tar.gz Collecting python-glanceclient Using cached python-glanceclient-2.5.0.tar.gz Collecting python-keystoneclient Using cached python-keystoneclient-3.8.0.tar.gz Collecting python-neutronclient Using cached python-neutronclient-6.0.0.tar.gz Requirement already satisfied: six in /home/jenkins/shiningpanda/jobs/501d0aeb/virtualenvs/d41d8cd9/lib/python2.7/site-packages (from configobj) Collecting requests-futures (from foreman) Using cached requests-futures-0.9.7.tar.gz Collecting pbr (from python-foreman==0.4.5) Using cached pbr-1.10.0.tar.gz Collecting Babel>=2.3.4 (from python-glanceclient) Using cached Babel-2.3.4.tar.gz Collecting PrettyTable<0.8,>=0.7 (from python-glanceclient) Using cached prettytable-0.7.2.zip Collecting warlock!=1.3.0,<2,>=1.0.1 (from python-glanceclient) Using cached warlock-1.2.0.tar.gz Collecting oslo.utils>=3.16.0 (from python-glanceclient) Using cached oslo.utils-3.21.0.tar.gz Collecting oslo.i18n>=2.1.0 (from python-glanceclient) Using cached oslo.i18n-3.11.0.tar.gz Collecting debtcollector>=1.2.0 (from python-keystoneclient) Using cached debtcollector-1.10.0.tar.gz Collecting keystoneauth1>=2.14.0 (from python-keystoneclient) Using cached keystoneauth1-2.17.0.tar.gz Collecting oslo.config!=3.18.0,>=3.14.0 (from python-keystoneclient) Using cached oslo.config-3.21.0.tar.gz Collecting oslo.serialization>=1.10.0 (from python-keystoneclient) Using cached oslo.serialization-2.15.0.tar.gz Collecting positional>=1.1.1 (from python-keystoneclient) Using cached positional-1.1.1.tar.gz Collecting stevedore>=1.17.1 (from python-keystoneclient) Using cached stevedore-1.19.1.tar.gz Collecting cliff!=1.16.0,!=1.17.0,>=1.15.0 (from python-neutronclient) Using cached cliff-2.3.0.tar.gz Collecting iso8601>=0.1.11 (from python-neutronclient) Using cached iso8601-0.1.11.tar.gz Collecting netaddr!=0.7.16,>=0.7.12 (from python-neutronclient) Using cached netaddr-0.7.18.tar.gz Collecting osc-lib>=1.0.2 (from python-neutronclient) Using cached osc-lib-1.2.0.tar.gz Collecting os-client-config!=1.19.0,!=1.19.1,!=1.20.0,!=1.20.1,!=1.21.0,>=1.13.1 (from python-neutronclient) Using cached os-client-config-1.24.0.tar.gz Collecting simplejson>=2.2.0 (from python-neutronclient) Using cached simplejson-3.10.0.tar.gz Collecting futures>=2.1.3 (from requests-futures->foreman) Using cached futures-3.0.5.tar.gz Collecting pytz>=0a (from Babel>=2.3.4->python-glanceclient) Using cached pytz-2016.10.tar.gz Collecting jsonschema<3,>=0.7 (from warlock!=1.3.0,<2,>=1.0.1->python-glanceclient) Using cached jsonschema-2.5.1.tar.gz Collecting jsonpatch<2,>=0.10 (from warlock!=1.3.0,<2,>=1.0.1->python-glanceclient) Using cached jsonpatch-1.14.tar.gz Collecting funcsigs>=0.4 (from oslo.utils>=3.16.0->python-glanceclient) Using cached funcsigs-1.0.2.tar.gz Collecting monotonic>=0.6 (from oslo.utils>=3.16.0->python-glanceclient) Using cached monotonic-1.2.tar.gz Collecting netifaces>=0.10.4 (from oslo.utils>=3.16.0->python-glanceclient) Using cached netifaces-0.10.5.tar.gz Collecting pyparsing>=2.0.7 (from oslo.utils>=3.16.0->python-glanceclient) Using cached pyparsing-2.1.10.zip Collecting wrapt>=1.7.0 (from debtcollector>=1.2.0->python-keystoneclient) Using cached wrapt-1.10.8.tar.gz Collecting rfc3986>=0.2.2 (from oslo.config!=3.18.0,>=3.14.0->python-keystoneclient) Using cached rfc3986-0.4.1.tar.gz Collecting msgpack-python>=0.4.0 (from oslo.serialization>=1.10.0->python-keystoneclient) Using cached msgpack-python-0.4.8.tar.gz Collecting cmd2>=0.6.7 (from cliff!=1.16.0,!=1.17.0,>=1.15.0->python-neutronclient) Using cached cmd2-0.6.9.tar.gz Collecting unicodecsv>=0.8.0 (from cliff!=1.16.0,!=1.17.0,>=1.15.0->python-neutronclient) Using cached unicodecsv-0.14.1.tar.gz Requirement already satisfied: PyYAML>=3.10.0 in /home/jenkins/shiningpanda/jobs/501d0aeb/virtualenvs/d41d8cd9/lib/python2.7/site-packages (from cliff!=1.16.0,!=1.17.0,>=1.15.0->python-neutronclient) Collecting appdirs>=1.3.0 (from os-client-config!=1.19.0,!=1.19.1,!=1.20.0,!=1.20.1,!=1.21.0,>=1.13.1->python-neutronclient) Using cached appdirs-1.4.0.tar.gz Collecting requestsexceptions>=1.1.1 (from os-client-config!=1.19.0,!=1.19.1,!=1.20.0,!=1.20.1,!=1.21.0,>=1.13.1->python-neutronclient) Using cached requestsexceptions-1.1.3.tar.gz Collecting functools32 (from jsonschema<3,>=0.7->warlock!=1.3.0,<2,>=1.0.1->python-glanceclient) Using cached functools32-3.2.3-2.zip Collecting jsonpointer>=1.9 (from jsonpatch<2,>=0.10->warlock!=1.3.0,<2,>=1.0.1->python-glanceclient) Using cached jsonpointer-1.10.tar.gz Skipping bdist_wheel for taskrunner, due to binaries being disabled for it. Skipping bdist_wheel for lxml, due to binaries being disabled for it. Skipping bdist_wheel for configobj, due to binaries being disabled for it. Skipping bdist_wheel for requests, due to binaries being disabled for it. Skipping bdist_wheel for foreman, due to binaries being disabled for it. Skipping bdist_wheel for python-foreman, due to binaries being disabled for it. Skipping bdist_wheel for python-glanceclient, due to binaries being disabled for it. Skipping bdist_wheel for python-keystoneclient, due to binaries being disabled for it. Skipping bdist_wheel for python-neutronclient, due to binaries being disabled for it. Skipping bdist_wheel for requests-futures, due to binaries being disabled for it. Skipping bdist_wheel for pbr, due to binaries being disabled for it. Skipping bdist_wheel for Babel, due to binaries being disabled for it. Skipping bdist_wheel for PrettyTable, due to binaries being disabled for it. Skipping bdist_wheel for warlock, due to binaries being disabled for it. Skipping bdist_wheel for oslo.utils, due to binaries being disabled for it. Skipping bdist_wheel for oslo.i18n, due to binaries being disabled for it. Skipping bdist_wheel for debtcollector, due to binaries being disabled for it. Skipping bdist_wheel for keystoneauth1, due to binaries being disabled for it. Skipping bdist_wheel for oslo.config, due to binaries being disabled for it. Skipping bdist_wheel for oslo.serialization, due to binaries being disabled for it. Skipping bdist_wheel for positional, due to binaries being disabled for it. Skipping bdist_wheel for stevedore, due to binaries being disabled for it. Skipping bdist_wheel for cliff, due to binaries being disabled for it. Skipping bdist_wheel for iso8601, due to binaries being disabled for it. Skipping bdist_wheel for netaddr, due to binaries being disabled for it. Skipping bdist_wheel for osc-lib, due to binaries being disabled for it. Skipping bdist_wheel for os-client-config, due to binaries being disabled for it. Skipping bdist_wheel for simplejson, due to binaries being disabled for it. Skipping bdist_wheel for futures, due to binaries being disabled for it. Skipping bdist_wheel for pytz, due to binaries being disabled for it. Skipping bdist_wheel for jsonschema, due to binaries being disabled for it. Skipping bdist_wheel for jsonpatch, due to binaries being disabled for it. Skipping bdist_wheel for funcsigs, due to binaries being disabled for it. Skipping bdist_wheel for monotonic, due to binaries being disabled for it. Skipping bdist_wheel for netifaces, due to binaries being disabled for it. Skipping bdist_wheel for pyparsing, due to binaries being disabled for it. Skipping bdist_wheel for wrapt, due to binaries being disabled for it. Skipping bdist_wheel for rfc3986, due to binaries being disabled for it. Skipping bdist_wheel for msgpack-python, due to binaries being disabled for it. Skipping bdist_wheel for cmd2, due to binaries being disabled for it. Skipping bdist_wheel for unicodecsv, due to binaries being disabled for it. Skipping bdist_wheel for appdirs, due to binaries being disabled for it. Skipping bdist_wheel for requestsexceptions, due to binaries being disabled for it. Skipping bdist_wheel for functools32, due to binaries being disabled for it. Skipping bdist_wheel for jsonpointer, due to binaries being disabled for it. Installing collected packages: taskrunner, lxml, configobj, requests, futures, requests-futures, foreman, pbr, python-foreman, pytz, Babel, PrettyTable, wrapt, funcsigs, debtcollector, iso8601, positional, stevedore, keystoneauth1, netaddr, oslo.i18n, rfc3986, oslo.config, msgpack-python, monotonic, netifaces, pyparsing, oslo.utils, oslo.serialization, python-keystoneclient, functools32, jsonschema, jsonpointer, jsonpatch, warlock, python-glanceclient, cmd2, unicodecsv, cliff, appdirs, requestsexceptions, os-client-config, simplejson, osc-lib, python-neutronclient Running setup.py install for taskrunner: started Running setup.py install for taskrunner: finished with status 'done' Running setup.py install for lxml: started Running setup.py install for lxml: still running... Running setup.py install for lxml: finished with status 'done' Running setup.py install for configobj: started Running setup.py install for configobj: finished with status 'done' Running setup.py install for requests: started Running setup.py install for requests: finished with status 'done' Running setup.py install for futures: started Running setup.py install for futures: finished with status 'done' Running setup.py install for requests-futures: started Running setup.py install for requests-futures: finished with status 'done' Running setup.py install for foreman: started Running setup.py install for foreman: finished with status 'done' Running setup.py install for pbr: started Running setup.py install for pbr: finished with status 'done' Running setup.py install for python-foreman: started Running setup.py install for python-foreman: finished with status 'done' Running setup.py install for pytz: started Running setup.py install for pytz: finished with status 'done' Running setup.py install for Babel: started Running setup.py install for Babel: finished with status 'done' Running setup.py install for PrettyTable: started Running setup.py install for PrettyTable: finished with status 'done' Running setup.py install for wrapt: started Running setup.py install for wrapt: finished with status 'done' Running setup.py install for funcsigs: started Running setup.py install for funcsigs: finished with status 'done' Running setup.py install for debtcollector: started Running setup.py install for debtcollector: finished with status 'done' Running setup.py install for iso8601: started Running setup.py install for iso8601: finished with status 'done' Running setup.py install for positional: started Running setup.py install for positional: finished with status 'done' Running setup.py install for stevedore: started Running setup.py install for stevedore: finished with status 'done' Running setup.py install for keystoneauth1: started Running setup.py install for keystoneauth1: finished with status 'done' Running setup.py install for netaddr: started Running setup.py install for netaddr: finished with status 'done' Running setup.py install for oslo.i18n: started Running setup.py install for oslo.i18n: finished with status 'done' Running setup.py install for rfc3986: started Running setup.py install for rfc3986: finished with status 'done' Running setup.py install for oslo.config: started Running setup.py install for oslo.config: finished with status 'done' Running setup.py install for msgpack-python: started Running setup.py install for msgpack-python: finished with status 'done' Running setup.py install for monotonic: started Running setup.py install for monotonic: finished with status 'done' Running setup.py install for netifaces: started Running setup.py install for netifaces: finished with status 'done' Running setup.py install for pyparsing: started Running setup.py install for pyparsing: finished with status 'done' Running setup.py install for oslo.utils: started Running setup.py install for oslo.utils: finished with status 'done' Running setup.py install for oslo.serialization: started Running setup.py install for oslo.serialization: finished with status 'done' Running setup.py install for python-keystoneclient: started Running setup.py install for python-keystoneclient: finished with status 'done' Running setup.py install for functools32: started Running setup.py install for functools32: finished with status 'done' Running setup.py install for jsonschema: started Running setup.py install for jsonschema: finished with status 'done' Running setup.py install for jsonpointer: started Running setup.py install for jsonpointer: finished with status 'done' Running setup.py install for jsonpatch: started Running setup.py install for jsonpatch: finished with status 'done' Running setup.py install for warlock: started Running setup.py install for warlock: finished with status 'done' Running setup.py install for python-glanceclient: started Running setup.py install for python-glanceclient: finished with status 'done' Running setup.py install for cmd2: started Running setup.py install for cmd2: finished with status 'done' Running setup.py install for unicodecsv: started Running setup.py install for unicodecsv: finished with status 'done' Running setup.py install for cliff: started Running setup.py install for cliff: finished with status 'done' Running setup.py install for appdirs: started Running setup.py install for appdirs: finished with status 'done' Running setup.py install for requestsexceptions: started Running setup.py install for requestsexceptions: finished with status 'done' Running setup.py install for os-client-config: started Running setup.py install for os-client-config: finished with status 'done' Running setup.py install for simplejson: started Running setup.py install for simplejson: finished with status 'done' Running setup.py install for osc-lib: started Running setup.py install for osc-lib: finished with status 'done' Running setup.py install for python-neutronclient: started Running setup.py install for python-neutronclient: finished with status 'done' Successfully installed Babel-2.3.4 PrettyTable-0.7.2 appdirs-1.4.0 cliff-2.3.0 cmd2-0.6.9 configobj-5.0.6 debtcollector-1.10.0 foreman-0.9.7 funcsigs-1.0.2 functools32-3.2.3.post2 futures-3.0.5 iso8601-0.1.11 jsonpatch-1.14 jsonpointer-1.10 jsonschema-2.5.1 keystoneauth1-2.17.0 lxml-3.7.2 monotonic-1.2 msgpack-python-0.4.8 netaddr-0.7.18 netifaces-0.10.5 os-client-config-1.24.0 osc-lib-1.2.0 oslo.config-3.21.0 oslo.i18n-3.11.0 oslo.serialization-2.15.0 oslo.utils-3.21.0 pbr-1.10.0 positional-1.1.1 pyparsing-2.1.10 python-foreman-0.4.5 python-glanceclient-2.5.0 python-keystoneclient-3.8.0 python-neutronclient-6.0.0 pytz-2016.10 requests-2.12.4 requests-futures-0.9.7 requestsexceptions-1.1.3 rfc3986-0.4.1 simplejson-3.10.0 stevedore-1.19.1 taskrunner-0.2.1 unicodecsv-0.14.1 warlock-1.2.0 wrapt-1.10.8 + pip install python-novaclient==2.17.0 DEPRECATION: --no-use-wheel is deprecated and will be removed in the future. Please use --no-binary :all: instead. Collecting python-novaclient==2.17.0 Using cached python-novaclient-2.17.0.tar.gz Collecting pbr<1.0,>=0.6 (from python-novaclient==2.17.0) Using cached pbr-0.11.1.tar.gz Collecting argparse (from python-novaclient==2.17.0) Using cached argparse-1.4.0.tar.gz Requirement already satisfied: iso8601>=0.1.8 in /home/jenkins/shiningpanda/jobs/501d0aeb/virtualenvs/d41d8cd9/lib/python2.7/site-packages (from python-novaclient==2.17.0) Requirement already satisfied: PrettyTable<0.8,>=0.7 in /home/jenkins/shiningpanda/jobs/501d0aeb/virtualenvs/d41d8cd9/lib/python2.7/site-packages (from python-novaclient==2.17.0) Requirement already satisfied: requests>=1.1 in /home/jenkins/shiningpanda/jobs/501d0aeb/virtualenvs/d41d8cd9/lib/python2.7/site-packages (from python-novaclient==2.17.0) Requirement already satisfied: simplejson>=2.0.9 in /home/jenkins/shiningpanda/jobs/501d0aeb/virtualenvs/d41d8cd9/lib/python2.7/site-packages (from python-novaclient==2.17.0) Requirement already satisfied: six>=1.5.2 in /home/jenkins/shiningpanda/jobs/501d0aeb/virtualenvs/d41d8cd9/lib/python2.7/site-packages (from python-novaclient==2.17.0) Requirement already satisfied: Babel>=1.3 in /home/jenkins/shiningpanda/jobs/501d0aeb/virtualenvs/d41d8cd9/lib/python2.7/site-packages (from python-novaclient==2.17.0) Requirement already satisfied: pip in /home/jenkins/shiningpanda/jobs/501d0aeb/virtualenvs/d41d8cd9/lib/python2.7/site-packages (from pbr<1.0,>=0.6->python-novaclient==2.17.0) Requirement already satisfied: pytz>=0a in /home/jenkins/shiningpanda/jobs/501d0aeb/virtualenvs/d41d8cd9/lib/python2.7/site-packages (from Babel>=1.3->python-novaclient==2.17.0) Skipping bdist_wheel for python-novaclient, due to binaries being disabled for it. Skipping bdist_wheel for pbr, due to binaries being disabled for it. Skipping bdist_wheel for argparse, due to binaries being disabled for it. Installing collected packages: pbr, argparse, python-novaclient Found existing installation: pbr 1.10.0 Uninstalling pbr-1.10.0: Successfully uninstalled pbr-1.10.0 Running setup.py install for pbr: started Running setup.py install for pbr: finished with status 'done' Running setup.py install for argparse: started Running setup.py install for argparse: finished with status 'done' Running setup.py install for python-novaclient: started Running setup.py install for python-novaclient: finished with status 'done' Successfully installed argparse-1.4.0 pbr-0.11.1 python-novaclient-2.17.0 + export PYTHONPATH=/usr/lib/python2.7/site-packages/bkr + PYTHONPATH=/usr/lib/python2.7/site-packages/bkr + /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/ci-factory/bootstrap/provision_resources.sh --site=ci-osp --project_defaults=atomic-ci-jobs/project/config/project_defaults_osp7_ose --topology=atomic-ci-jobs/project/config/openshift-cluster-containerized --ssh_keyfile=atomic-ci-jobs/project/config/keys/ci-factory --name=openshift-ansible Provisioning resources with the following environment ----------------------------------------------------- SITE: ci-osp PROJECT_DEFAULTS: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/config/project_defaults_osp7_ose.json TOPOLOGY: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/config/openshift-cluster-containerized.json SSH_KEYFILE: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/config/keys/ci-factory UUID: SKIP_UUID: False WORKSPACE: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave LABEL: openshift-ansible-4Y5zDh CLEANUP: on_failure /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/ci-factory /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave 2017-01-10 18:04:55,060 - taskrunner - INFO - ************* TaskRunner start [2017-01-10 23:04:55.060475 UTC] ************* 2017-01-10 18:04:55,687 - provision - INFO - Provisioning in Cloud-Openstack since FLAVOR and IMAGE are defined 2017-01-10 18:04:55,687 - provision - INFO - Provisioning in Cloud-Openstack since FLAVOR and IMAGE are defined 2017-01-10 18:04:55,687 - provision - INFO - Provisioning in Cloud-Openstack since FLAVOR and IMAGE are defined 2017-01-10 18:04:55,692 - taskrunner - INFO - =========== run RunCommand =========== 2017-01-10 18:04:55,692 - util - INFO - Running command on localhost: rm -f /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/RESOURCES.txt /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/resources.json /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/ansible_inventory.txt 2017-01-10 18:04:55,697 - taskrunner - INFO - =========== run GetNodes =========== 2017-01-10 18:04:55,697 - tasks.nodefactory - INFO - Using OpenStack controller XXXXXXXXXXXXXXXXXXXXXXXXXXX/dashboard/project/instances/ 2017-01-10 18:04:56,308 - tasks.nodefactory - INFO - Checking if image, flavor, and network exist 2017-01-10 18:04:58,715 - tasks.nodefactory - INFO - Found image rhel-7.2-server-x86_64-updated (d49a9b5f-9a6f-4e0c-89a9-b25e5c3191f2). 2017-01-10 18:04:58,886 - tasks.nodefactory - INFO - Booting node : openshift-ansible-4Y5zDh-r1 2017-01-10 18:04:58,886 - tasks.nodefactory - INFO - Image : rhel-7.2-server-x86_64-updated 2017-01-10 18:04:58,886 - tasks.nodefactory - INFO - Flavor : m1.medium 2017-01-10 18:05:01,117 - tasks.nodefactory - INFO - Found image rhel-7.2-server-x86_64-updated (d49a9b5f-9a6f-4e0c-89a9-b25e5c3191f2). 2017-01-10 18:05:01,956 - tasks.nodefactory - INFO - openshift-ansible-4Y5zDh-r1 - BUILD/scheduling - created 2017-01-10T23:05:01Z - - 6169cbdb-1df1-4aa9-ab41-7198f5e611db 2017-01-10 18:05:02,280 - tasks.common - INFO - Waiting for end of BUILD state of openshift-ansible-4Y5zDh-r1 - START 2017-01-10 18:05:24,124 - tasks.common - INFO - Waiting for end of BUILD state of openshift-ansible-4Y5zDh-r1 - DONE 2017-01-10 18:05:24,694 - tasks.nodefactory - INFO - Added IP 10.8.169.210 with Neutron 2017-01-10 18:05:24,815 - tasks.get_nodes_task - INFO - Going to use ssh keyfile instead of password. 2017-01-10 18:05:24,816 - tasks.common - INFO - Waiting for SSH access to nodes 2017-01-10 18:05:24,816 - tasks.common - INFO - node 10.8.169.210 - START 2017-01-10 18:05:48,338 - tasks.common - INFO - node 10.8.169.210 - DONE 2017-01-10 18:05:48,338 - util - INFO - Running command on 10.8.169.210: uname -a; uptime 2017-01-10 18:05:48,780 - util - INFO - Output: Linux openshift-ansible-4y5zdh-r1.localdomain 3.10.0-327.44.2.el7.x86_64 #1 SMP Thu Nov 24 05:49:35 EST 2016 x86_64 x86_64 x86_64 GNU/Linux 18:05:48 up 0 min, 0 users, load average: 0.15, 0.03, 0.01 2017-01-10 18:05:48,780 - tasks.get_nodes_task - INFO - Using nodes: NODE1: floating ip 10.8.169.210, ip 172.16.132.148 2017-01-10 18:05:48,780 - util - INFO - Running command on 10.8.169.210: [[ -f ~/.prepared-by-cic-getnodes ]]; echo $? 2017-01-10 18:05:49,076 - util - INFO - Output: 0 2017-01-10 18:05:49,076 - tasks.get_nodes_task - INFO - Skipping preparation as nodes should be already prepared by us. 2017-01-10 18:05:49,077 - taskrunner - INFO - =========== run GetNodes =========== 2017-01-10 18:05:49,077 - tasks.nodefactory - INFO - Using OpenStack controller XXXXXXXXXXXXXXXXXXXXXXXXXXX/dashboard/project/instances/ 2017-01-10 18:05:49,570 - tasks.nodefactory - INFO - Checking if image, flavor, and network exist 2017-01-10 18:05:51,247 - tasks.nodefactory - INFO - Found image rhel-7.2-server-x86_64-updated (d49a9b5f-9a6f-4e0c-89a9-b25e5c3191f2). 2017-01-10 18:05:51,405 - tasks.nodefactory - INFO - Booting node : openshift-ansible-4Y5zDh-r2 2017-01-10 18:05:51,405 - tasks.nodefactory - INFO - Image : rhel-7.2-server-x86_64-updated 2017-01-10 18:05:51,406 - tasks.nodefactory - INFO - Flavor : m1.medium 2017-01-10 18:05:52,609 - tasks.nodefactory - INFO - Found image rhel-7.2-server-x86_64-updated (d49a9b5f-9a6f-4e0c-89a9-b25e5c3191f2). 2017-01-10 18:05:53,336 - tasks.nodefactory - INFO - openshift-ansible-4Y5zDh-r2 - BUILD/scheduling - created 2017-01-10T23:05:53Z - - 4cb6c697-ab31-4742-ac49-8f91eccaa4a8 2017-01-10 18:05:53,648 - tasks.common - INFO - Waiting for end of BUILD state of openshift-ansible-4Y5zDh-r2 - START 2017-01-10 18:06:02,647 - tasks.common - INFO - Waiting for end of BUILD state of openshift-ansible-4Y5zDh-r2 - DONE 2017-01-10 18:06:03,172 - tasks.nodefactory - INFO - Added IP 10.8.169.215 with Neutron 2017-01-10 18:06:03,335 - tasks.get_nodes_task - INFO - Going to use ssh keyfile instead of password. 2017-01-10 18:06:03,337 - tasks.common - INFO - Waiting for SSH access to nodes 2017-01-10 18:06:03,338 - tasks.common - INFO - node 10.8.169.215 - START 2017-01-10 18:06:34,379 - tasks.common - INFO - node 10.8.169.215 - DONE 2017-01-10 18:06:34,379 - util - INFO - Running command on 10.8.169.215: uname -a; uptime 2017-01-10 18:06:35,026 - util - INFO - Output: Linux openshift-ansible-4y5zdh-r2.localdomain 3.10.0-327.44.2.el7.x86_64 #1 SMP Thu Nov 24 05:49:35 EST 2016 x86_64 x86_64 x86_64 GNU/Linux 18:06:34 up 0 min, 0 users, load average: 0.34, 0.08, 0.03 2017-01-10 18:06:35,027 - tasks.get_nodes_task - INFO - Using nodes: NODE1: floating ip 10.8.169.215, ip 172.16.132.150 2017-01-10 18:06:35,027 - util - INFO - Running command on 10.8.169.215: [[ -f ~/.prepared-by-cic-getnodes ]]; echo $? 2017-01-10 18:06:35,317 - util - INFO - Output: 0 2017-01-10 18:06:35,317 - tasks.get_nodes_task - INFO - Skipping preparation as nodes should be already prepared by us. 2017-01-10 18:06:35,317 - taskrunner - INFO - =========== run GetNodes =========== 2017-01-10 18:06:35,317 - tasks.nodefactory - INFO - Using OpenStack controller XXXXXXXXXXXXXXXXXXXXXXXXXXX/dashboard/project/instances/ 2017-01-10 18:06:35,878 - tasks.get_nodes_task - INFO - Running /usr/bin/write-mime-multipart command result: True 2017-01-10 18:06:35,878 - tasks.nodefactory - INFO - Checking if image, flavor, and network exist 2017-01-10 18:06:38,273 - tasks.nodefactory - INFO - Found image rhel-7.2-server-x86_64-updated (d49a9b5f-9a6f-4e0c-89a9-b25e5c3191f2). 2017-01-10 18:06:38,487 - tasks.nodefactory - INFO - Booting node : openshift-ansible-4Y5zDh-r3 2017-01-10 18:06:38,487 - tasks.nodefactory - INFO - Image : rhel-7.2-server-x86_64-updated 2017-01-10 18:06:38,487 - tasks.nodefactory - INFO - Flavor : m1.large 2017-01-10 18:06:40,700 - tasks.nodefactory - INFO - Found image rhel-7.2-server-x86_64-updated (d49a9b5f-9a6f-4e0c-89a9-b25e5c3191f2). 2017-01-10 18:06:41,526 - tasks.nodefactory - INFO - openshift-ansible-4Y5zDh-r3 - BUILD/scheduling - created 2017-01-10T23:06:41Z - - 322ad2bb-bbf4-4c06-a0d8-99468b676e5c 2017-01-10 18:06:41,968 - tasks.common - INFO - Waiting for end of BUILD state of openshift-ansible-4Y5zDh-r3 - START 2017-01-10 18:06:49,739 - tasks.common - INFO - Waiting for end of BUILD state of openshift-ansible-4Y5zDh-r3 - DONE 2017-01-10 18:06:50,324 - tasks.nodefactory - INFO - Added IP 10.8.169.22 with Neutron 2017-01-10 18:06:50,475 - tasks.get_nodes_task - INFO - Going to use ssh keyfile instead of password. 2017-01-10 18:06:50,476 - tasks.common - INFO - Waiting for SSH access to nodes 2017-01-10 18:06:50,476 - tasks.common - INFO - node 10.8.169.22 - START 2017-01-10 18:07:13,191 - tasks.common - INFO - node 10.8.169.22 - DONE 2017-01-10 18:07:13,191 - util - INFO - Running command on 10.8.169.22: uname -a; uptime 2017-01-10 18:07:13,537 - util - INFO - Output: Linux openshift-ansible-4y5zdh-r3.localdomain 3.10.0-327.44.2.el7.x86_64 #1 SMP Thu Nov 24 05:49:35 EST 2016 x86_64 x86_64 x86_64 GNU/Linux 18:07:13 up 0 min, 0 users, load average: 0.24, 0.05, 0.02 2017-01-10 18:07:13,537 - tasks.get_nodes_task - INFO - Using nodes: NODE1: floating ip 10.8.169.22, ip 172.16.132.155 2017-01-10 18:07:13,537 - util - INFO - Running command on 10.8.169.22: [[ -f ~/.prepared-by-cic-getnodes ]]; echo $? 2017-01-10 18:07:13,855 - util - INFO - Output: 0 2017-01-10 18:07:13,855 - tasks.get_nodes_task - INFO - Skipping preparation as nodes should be already prepared by us. 2017-01-10 18:07:13,855 - taskrunner - INFO - =========== run ExtractResources =========== 2017-01-10 18:07:13,856 - taskrunner - INFO - Skipping cleanup: cleanup=on_failure and failures=[] 2017-01-10 18:07:13,856 - taskrunner - INFO - ************* TaskRunner end [2017-01-10 23:07:13.856827 UTC] ************* EXISTING_NODES=10.8.169.210,10.8.169.215,10.8.169.22 PRIVATE_IPS=172.16.132.148,172.16.132.150,172.16.132.155 BKR_JOBID=J: SITE=ci-osp LABEL=openshift-ansible-4Y5zDh UUID= PROVISION_JOB=XXXXXXXXXXXXXXXXXXXXXXXXXXX/view/All/job/openshift-ansible-3-test-matrix/CONTAINERIZED=_containerized,OSE_VER=3.3,PYTHON=System-CPython-2.7,TOPOLOGY=openshift-cluster-containerized,TargetBranch=master,nodes=openshift-ansible-slave/688 /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave [EnvInject] - Injecting environment variables from a build step. [EnvInject] - Injecting as environment variables the properties file path '/home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/RESOURCES.txt' [EnvInject] - Variables injected successfully. [EnvInject] - Injecting as environment variables the properties content GH_STATE=error GH_DESC="Error prepping the 3.3 cluster" GH_CONTEXT=aos-ci-jenkins/OS_3.3_containerized [EnvInject] - Variables injected successfully. [openshift-ansible-slave] $ /bin/bash /tmp/hudson7279610125709894696.sh + chmod 600 /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/config/keys/ci-factory + ansible-playbook -i /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/ci-factory/utils/central_ci_dynamic_hosts.py --private-key=/home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/config/keys/ci-factory --extra-vars 'enable_puddle=true ose_ver=3.3' /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/playbooks/openshift/setup.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/playbooks/openshift/roles/common/tasks/subscription_manager.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/playbooks/openshift/roles/common/tasks/enable_pulp_repos.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/playbooks/openshift/roles/common/tasks/install_required_pkgs.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/playbooks/openshift/roles/nodes/tasks/enable_openshift.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/playbooks/openshift/roles/nodes/tasks/copy_sshkeys.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/playbooks/openshift/roles/nodes/tasks/add_to_auth_keys.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/playbooks/openshift/roles/nodes/tasks/setup_docker.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/playbooks/openshift/roles/nodes/tasks/remove_from_auth_keys.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/playbooks/openshift/roles/nodes/tasks/remove_tmp_key_files.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/playbooks/openshift/roles/dss/tasks/create-vg.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/playbooks/openshift/roles/dss/tasks/update-ds-config.yml PLAY [nodes] ******************************************************************* TASK [setup] ******************************************************************* ok: [10.8.169.210] ok: [10.8.169.215] ok: [10.8.169.22] TASK [common : Register with subscription Manager on stage] ******************** skipping: [10.8.169.210] skipping: [10.8.169.215] skipping: [10.8.169.22] TASK [common : Subscribe to OpenShift Enterprise subscription] ***************** skipping: [10.8.169.215] skipping: [10.8.169.22] skipping: [10.8.169.210] TASK [common : Disable all repositories] *************************************** skipping: [10.8.169.22] skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [common : Enable the operating system repos] ****************************** skipping: [10.8.169.210] skipping: [10.8.169.215] skipping: [10.8.169.22] TASK [common : Enable the ose repo] ******************************************** skipping: [10.8.169.22] skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [common : rhel7 pulp repo] ************************************************ changed: [10.8.169.210] changed: [10.8.169.22] changed: [10.8.169.215] TASK [common : update all packages] ******************************************** changed: [10.8.169.210] [WARNING]: Consider using yum module rather than running yum changed: [10.8.169.215] changed: [10.8.169.22] TASK [common : Install packages on masters and nodes] ************************** changed: [10.8.169.210] => (item=[u'wget', u'git', u'rsync', u'net-tools', u'bind-utils', u'iptables-services', u'bridge-utils', u'gcc', u'python-virtualenv', u'docker', u'yum-utils']) changed: [10.8.169.215] => (item=[u'wget', u'git', u'rsync', u'net-tools', u'bind-utils', u'iptables-services', u'bridge-utils', u'gcc', u'python-virtualenv', u'docker', u'yum-utils']) changed: [10.8.169.22] => (item=[u'wget', u'git', u'rsync', u'net-tools', u'bind-utils', u'iptables-services', u'bridge-utils', u'gcc', u'python-virtualenv', u'docker', u'yum-utils']) TASK [nodes : Openshift repo from a puddle] ************************************ ok: [10.8.169.210] ok: [10.8.169.22] ok: [10.8.169.215] TASK [nodes : Openshift repo from pulp] **************************************** skipping: [10.8.169.215] skipping: [10.8.169.22] skipping: [10.8.169.210] TASK [nodes : Setup openshift repo] ******************************************** changed: [10.8.169.215] changed: [10.8.169.210] changed: [10.8.169.22] TASK [nodes : Retrieve hostnames from nodes] *********************************** changed: [10.8.169.210] changed: [10.8.169.22] changed: [10.8.169.215] TASK [nodes : Register label on general part of the node hostname] ************* changed: [10.8.169.22 -> localhost] changed: [10.8.169.215 -> localhost] changed: [10.8.169.210 -> localhost] TASK [nodes : Create directory based on the host_label] ************************ changed: [10.8.169.210 -> localhost] [WARNING]: Consider using file module with state=directory rather than running mkdir TASK [nodes : Create ssh key dir] ********************************************** changed: [10.8.169.210] changed: [10.8.169.22] changed: [10.8.169.215] TASK [nodes : Copy ssh keys] *************************************************** changed: [10.8.169.215] changed: [10.8.169.22] changed: [10.8.169.210] TASK [nodes : Add to authorized_keys] ****************************************** TASK [nodes : Change Docker configuration] ************************************* changed: [10.8.169.215] changed: [10.8.169.22] changed: [10.8.169.210] TASK [nodes : Remove from authorized_keys] ************************************* TASK [nodes : Remove ssh key files from /tmp] ********************************** changed: [10.8.169.210 -> localhost] TASK [dss : stat] ************************************************************** skipping: [10.8.169.210] TASK [dss : Update create partition script with correct device] **************** skipping: [10.8.169.210] TASK [dss : Create partition for docker storage] ******************************* skipping: [10.8.169.210] TASK [dss : Create physical device with pvcreate] ****************************** skipping: [10.8.169.210] TASK [dss : Create docker-vg volume group] ************************************* skipping: [10.8.169.210] TASK [dss : Update docker storage config with docker-vg] *********************** skipping: [10.8.169.210] TASK [dss : Run docker-storage-setup] ****************************************** skipping: [10.8.169.210] PLAY RECAP ********************************************************************* 10.8.169.210 : ok=13 changed=11 unreachable=0 failed=0 10.8.169.215 : ok=11 changed=9 unreachable=0 failed=0 10.8.169.22 : ok=11 changed=9 unreachable=0 failed=0 [EnvInject] - Injecting environment variables from a build step. [EnvInject] - Injecting as environment variables the properties content GH_STATE=error GH_DESC="openshift-ansible install failed" GH_CONTEXT=aos-ci-jenkins/OS_3.3_containerized [EnvInject] - Variables injected successfully. [openshift-ansible-slave] $ /bin/bash /tmp/hudson2685254913603561712.sh + chmod 600 /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/config/keys/ci-factory + ansible-playbook -i /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/ci-factory/utils/central_ci_dynamic_hosts.py --private-key=/home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/config/keys/ci-factory --extra-vars openshift_image_tag=v3.3 /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/playbooks/byo/config.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/os_firewall/tasks/firewall/iptables.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/docker/tasks/udev_workaround.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_version/tasks/set_version_rpm.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_version/tasks/set_version_containerized.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/os_firewall/tasks/firewall/iptables.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/docker/tasks/udev_workaround.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_version/tasks/set_version_rpm.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_version/tasks/set_version_containerized.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/os_firewall/tasks/firewall/iptables.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/docker/tasks/udev_workaround.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_version/tasks/set_version_rpm.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_version/tasks/set_version_containerized.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/docker/tasks/udev_workaround.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/os_firewall/tasks/firewall/iptables.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/etcd/tasks/etcdctl.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/os_firewall/tasks/firewall/iptables.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/os_firewall/tasks/firewall/iptables.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/os_firewall/tasks/firewall/iptables.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/os_firewall/tasks/firewall/iptables.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/docker/tasks/udev_workaround.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_version/tasks/set_version_rpm.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_version/tasks/set_version_containerized.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/docker/tasks/udev_workaround.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_cloud_provider/tasks/openstack.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_cloud_provider/tasks/aws.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_cloud_provider/tasks/gce.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/os_firewall/tasks/firewall/iptables.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/os_firewall/tasks/firewall/iptables.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/os_firewall/tasks/firewall/iptables.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/nuage_master/tasks/serviceaccount.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/nuage_master/tasks/certificates.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_master/tasks/systemd_units.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_master/tasks/set_loopback_context.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_master_cluster/tasks/configure.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/os_firewall/tasks/firewall/iptables.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/docker/tasks/udev_workaround.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_version/tasks/set_version_rpm.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_version/tasks/set_version_containerized.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/os_firewall/tasks/firewall/iptables.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/os_firewall/tasks/firewall/iptables.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/docker/tasks/udev_workaround.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_version/tasks/set_version_rpm.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_version/tasks/set_version_containerized.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/docker/tasks/udev_workaround.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_cloud_provider/tasks/openstack.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_cloud_provider/tasks/aws.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_cloud_provider/tasks/gce.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_node_dnsmasq/tasks/./network-manager.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_node_dnsmasq/tasks/./no-network-manager.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/os_firewall/tasks/firewall/iptables.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_node/tasks/systemd_units.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_node/tasks/storage_plugins/nfs.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_node/tasks/storage_plugins/glusterfs.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_node/tasks/storage_plugins/ceph.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_node/tasks/storage_plugins/iscsi.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/os_firewall/tasks/firewall/iptables.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/docker/tasks/udev_workaround.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_version/tasks/set_version_rpm.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_version/tasks/set_version_containerized.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/docker/tasks/udev_workaround.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_cloud_provider/tasks/openstack.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_cloud_provider/tasks/aws.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_cloud_provider/tasks/gce.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_node_dnsmasq/tasks/./network-manager.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_node_dnsmasq/tasks/./no-network-manager.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/os_firewall/tasks/firewall/iptables.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_node/tasks/systemd_units.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_node/tasks/storage_plugins/nfs.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_node/tasks/storage_plugins/glusterfs.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_node/tasks/storage_plugins/ceph.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_node/tasks/storage_plugins/iscsi.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/os_firewall/tasks/firewall/iptables.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/nuage_node/tasks/certificates.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/nuage_node/tasks/iptables.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/os_firewall/tasks/firewall/iptables.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/docker/tasks/udev_workaround.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_version/tasks/set_version_rpm.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_version/tasks/set_version_containerized.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/os_firewall/tasks/firewall/iptables.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/docker/tasks/udev_workaround.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_version/tasks/set_version_rpm.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_version/tasks/set_version_containerized.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/docker/tasks/udev_workaround.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_serviceaccounts/tasks/legacy_add_scc_to_user.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_serviceaccounts/tasks/legacy_add_scc_to_user.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_hosted/tasks/router/router.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_hosted/tasks/registry/registry.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_metrics/tasks/install.yml PLAY [Create initial host groups for localhost] ******************************** TASK [include_vars] ************************************************************ ok: [localhost] TASK [add_host] **************************************************************** ok: [localhost] => (item=10.8.169.215) ok: [localhost] => (item=10.8.169.210) ok: [localhost] => (item=10.8.169.22) PLAY [Create initial host groups for all hosts] ******************************** TASK [include_vars] ************************************************************ ok: [10.8.169.210] ok: [10.8.169.22] ok: [10.8.169.215] PLAY [Populate config host groups] ********************************************* TASK [fail] ******************************************************************** skipping: [localhost] TASK [fail] ******************************************************************** skipping: [localhost] TASK [fail] ******************************************************************** skipping: [localhost] TASK [fail] ******************************************************************** skipping: [localhost] TASK [fail] ******************************************************************** skipping: [localhost] TASK [fail] ******************************************************************** skipping: [localhost] TASK [Evaluate oo_all_hosts] *************************************************** ok: [localhost] => (item=10.8.169.22) ok: [localhost] => (item=10.8.169.215) ok: [localhost] => (item=10.8.169.210) TASK [Evaluate oo_masters] ***************************************************** ok: [localhost] => (item=10.8.169.22) TASK [Evaluate oo_etcd_to_config] ********************************************** TASK [Evaluate oo_masters_to_config] ******************************************* ok: [localhost] => (item=10.8.169.22) TASK [Evaluate oo_nodes_to_config] ********************************************* ok: [localhost] => (item=10.8.169.22) ok: [localhost] => (item=10.8.169.215) ok: [localhost] => (item=10.8.169.210) TASK [Add master to oo_nodes_to_config] **************************************** skipping: [localhost] => (item=10.8.169.22) TASK [Evaluate oo_first_etcd] ************************************************** skipping: [localhost] TASK [Evaluate oo_first_master] ************************************************ ok: [localhost] TASK [Evaluate oo_lb_to_config] ************************************************ TASK [Evaluate oo_nfs_to_config] *********************************************** PLAY [Ensure that all non-node hosts are accessible] *************************** TASK [setup] ******************************************************************* ok: [10.8.169.22] PLAY [Initialize host facts] *************************************************** TASK [setup] ******************************************************************* ok: [10.8.169.22] ok: [10.8.169.210] ok: [10.8.169.215] TASK [openshift_facts : Detecting Operating System] **************************** ok: [10.8.169.22] ok: [10.8.169.215] ok: [10.8.169.210] TASK [openshift_facts : set_fact] ********************************************** ok: [10.8.169.22] ok: [10.8.169.215] ok: [10.8.169.210] TASK [openshift_facts : set_fact] ********************************************** ok: [10.8.169.215] ok: [10.8.169.22] ok: [10.8.169.210] TASK [openshift_facts : Ensure various deps are installed] ********************* ok: [10.8.169.210] => (item=[u'iproute', u'python-dbus', u'python-six', u'PyYAML', u'yum-utils']) ok: [10.8.169.215] => (item=[u'iproute', u'python-dbus', u'python-six', u'PyYAML', u'yum-utils']) ok: [10.8.169.22] => (item=[u'iproute', u'python-dbus', u'python-six', u'PyYAML', u'yum-utils']) TASK [openshift_facts : Gather Cluster facts and set is_containerized if needed] *** changed: [10.8.169.22] changed: [10.8.169.210] changed: [10.8.169.215] TASK [openshift_facts : Set repoquery command] ********************************* ok: [10.8.169.210] ok: [10.8.169.22] ok: [10.8.169.215] TASK [openshift_facts] ********************************************************* ok: [10.8.169.22] ok: [10.8.169.215] ok: [10.8.169.210] TASK [set_fact] **************************************************************** ok: [10.8.169.215] ok: [10.8.169.22] ok: [10.8.169.210] TASK [set_fact] **************************************************************** ok: [10.8.169.210] ok: [10.8.169.22] ok: [10.8.169.215] PLAY [Gather and set facts for node hosts] ************************************* TASK [setup] ******************************************************************* ok: [10.8.169.210] ok: [10.8.169.215] ok: [10.8.169.22] TASK [openshift_facts : Detecting Operating System] **************************** ok: [10.8.169.210] ok: [10.8.169.215] ok: [10.8.169.22] TASK [openshift_facts : set_fact] ********************************************** ok: [10.8.169.210] ok: [10.8.169.22] ok: [10.8.169.215] TASK [openshift_facts : set_fact] ********************************************** ok: [10.8.169.22] ok: [10.8.169.210] ok: [10.8.169.215] TASK [openshift_facts : Ensure various deps are installed] ********************* ok: [10.8.169.210] => (item=[u'iproute', u'python-dbus', u'python-six', u'PyYAML', u'yum-utils']) ok: [10.8.169.22] => (item=[u'iproute', u'python-dbus', u'python-six', u'PyYAML', u'yum-utils']) ok: [10.8.169.215] => (item=[u'iproute', u'python-dbus', u'python-six', u'PyYAML', u'yum-utils']) TASK [openshift_facts : Gather Cluster facts and set is_containerized if needed] *** ok: [10.8.169.215] ok: [10.8.169.22] ok: [10.8.169.210] TASK [openshift_facts : Set repoquery command] ********************************* ok: [10.8.169.22] ok: [10.8.169.215] ok: [10.8.169.210] TASK [command] ***************************************************************** ok: [10.8.169.210] ok: [10.8.169.22] ok: [10.8.169.215] TASK [Warn user about bad openshift_hostname values] *************************** skipping: [10.8.169.210] PLAY [l_oo_all_hosts] ********************************************************** TASK [Check for bad combinations of yum and subscription-manager] ************** ok: [10.8.169.215] ok: [10.8.169.22] ok: [10.8.169.210] TASK [fail] ******************************************************************** skipping: [10.8.169.210] skipping: [10.8.169.215] skipping: [10.8.169.22] PLAY [Determine openshift_version to configure on first master] **************** TASK [setup] ******************************************************************* ok: [10.8.169.22] TASK [openshift_facts : Detecting Operating System] **************************** ok: [10.8.169.22] TASK [openshift_facts : set_fact] ********************************************** ok: [10.8.169.22] TASK [openshift_facts : set_fact] ********************************************** ok: [10.8.169.22] TASK [openshift_facts : Ensure various deps are installed] ********************* ok: [10.8.169.22] => (item=[u'iproute', u'python-dbus', u'python-six', u'PyYAML', u'yum-utils']) TASK [openshift_facts : Gather Cluster facts and set is_containerized if needed] *** ok: [10.8.169.22] TASK [openshift_facts : Set repoquery command] ********************************* ok: [10.8.169.22] TASK [openshift_repos : assert] ************************************************ skipping: [10.8.169.22] TASK [openshift_repos : Ensure libselinux-python is installed] ***************** skipping: [10.8.169.22] TASK [openshift_repos : Create any additional repos that are defined] ********** skipping: [10.8.169.22] TASK [openshift_repos : Remove the additional repos if no longer defined] ****** skipping: [10.8.169.22] TASK [openshift_repos : Configure origin gpg keys if needed] ******************* skipping: [10.8.169.22] TASK [openshift_repos : Configure origin yum repositories RHEL/CentOS] ********* skipping: [10.8.169.22] TASK [openshift_docker_facts : Set docker facts] ******************************* changed: [10.8.169.22] => (item={u'local_facts': {u'blocked_registries': u'', u'additional_registries': u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888,registry.ops.openshift.com', u'log_driver': u'', u'disable_push_dockerhub': u'', u'hosted_registry_insecure': False, u'hosted_registry_network': u'172.30.0.0/16', u'log_options': u'', u'insecure_registries': u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888,registry.ops.openshift.com', u'options': u''}, u'role': u'docker'}) TASK [openshift_docker_facts : set_fact] *************************************** ok: [10.8.169.22] TASK [openshift_docker_facts : set_fact] *************************************** skipping: [10.8.169.22] TASK [openshift_docker_facts : set_fact] *************************************** ok: [10.8.169.22] TASK [os_firewall : Assert - Do not use firewalld on Atomic Host] ************** skipping: [10.8.169.22] TASK [os_firewall : Install firewalld packages] ******************************** skipping: [10.8.169.22] TASK [os_firewall : Ensure iptables services are not enabled] ****************** changed: [10.8.169.22] => (item=iptables) changed: [10.8.169.22] => (item=ip6tables) TASK [os_firewall : Start and enable firewalld service] ************************ ok: [10.8.169.22] TASK [os_firewall : need to pause here, otherwise the firewalld service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.22] TASK [os_firewall : Add firewalld allow rules] ********************************* TASK [os_firewall : Remove firewalld allow rules] ****************************** TASK [os_firewall : Ensure firewalld service is not enabled] ******************* skipping: [10.8.169.22] TASK [os_firewall : Install iptables packages] ********************************* skipping: [10.8.169.22] => (item=[]) TASK [os_firewall : Start and enable iptables service] ************************* skipping: [10.8.169.22] TASK [os_firewall : need to pause here, otherwise the iptables service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.22] TASK [os_firewall : Add iptables allow rules] ********************************** TASK [os_firewall : Remove iptables rules] ************************************* TASK [docker : Get current installed Docker version] *************************** ok: [10.8.169.22] TASK [docker : Error out if Docker pre-installed but too old] ****************** skipping: [10.8.169.22] TASK [docker : Error out if requested Docker is too old] *********************** skipping: [10.8.169.22] TASK [docker : Get latest available version of Docker] ************************* skipping: [10.8.169.22] TASK [docker : Fail if Docker version requested but downgrade is required] ***** skipping: [10.8.169.22] TASK [docker : Error out if attempting to upgrade Docker across the 1.10 boundary] *** skipping: [10.8.169.22] TASK [docker : Install Docker] ************************************************* ok: [10.8.169.22] TASK [docker : Ensure docker.service.d directory exists] *********************** skipping: [10.8.169.22] TASK [docker : Configure Docker service unit file] ***************************** skipping: [10.8.169.22] TASK [docker : Getting current systemd-udevd exec command] ********************* skipping: [10.8.169.22] TASK [docker : Assure systemd-udevd.service.d directory exists] **************** skipping: [10.8.169.22] TASK [docker : Create systemd-udevd override file] ***************************** skipping: [10.8.169.22] TASK [docker : stat] *********************************************************** ok: [10.8.169.22] TASK [docker : Set registry params] ******************************************** changed: [10.8.169.22] => (item={u'reg_conf_var': u'ADD_REGISTRY', u'reg_flag': u'--add-registry', u'reg_fact_val': [u'registry.ops.openshift.com', u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888', u'XXXXXXXXXXXXXXXXXXXXXXXXXXX']}) skipping: [10.8.169.22] => (item={u'reg_conf_var': u'BLOCK_REGISTRY', u'reg_flag': u'--block-registry', u'reg_fact_val': u''}) changed: [10.8.169.22] => (item={u'reg_conf_var': u'INSECURE_REGISTRY', u'reg_flag': u'--insecure-registry', u'reg_fact_val': [u'registry.ops.openshift.com', u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888']}) TASK [docker : Set Proxy Settings] ********************************************* ok: [10.8.169.22] => (item={u'reg_conf_var': u'HTTP_PROXY', u'reg_fact_val': u''}) ok: [10.8.169.22] => (item={u'reg_conf_var': u'HTTPS_PROXY', u'reg_fact_val': u''}) ok: [10.8.169.22] => (item={u'reg_conf_var': u'NO_PROXY', u'reg_fact_val': u''}) TASK [docker : Set various Docker options] ************************************* changed: [10.8.169.22] TASK [docker : Start the Docker service] *************************************** changed: [10.8.169.22] TASK [docker : set_fact] ******************************************************* ok: [10.8.169.22] RUNNING HANDLER [docker : restart docker] ************************************** changed: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** ok: [10.8.169.22] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : Use openshift.common.version fact as version to configure if already installed] *** skipping: [10.8.169.22] TASK [openshift_version : Set rpm version to configure if openshift_pkg_version specified] *** skipping: [10.8.169.22] TASK [openshift_version : Gather common package version] *********************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : Set containerized version to configure if openshift_image_tag specified] *** ok: [10.8.169.22] TASK [openshift_version : Set containerized version to configure if openshift_release specified] *** skipping: [10.8.169.22] TASK [openshift_version : Lookup latest containerized version if no version specified] *** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : Set precise containerized version to configure if openshift_release specified] *** changed: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** ok: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** ok: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** ok: [10.8.169.22] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.22] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.22] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.22] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.22] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.22] TASK [openshift_version : debug] *********************************************** ok: [10.8.169.22] => { "openshift_release": "VARIABLE IS NOT DEFINED!" } TASK [openshift_version : debug] *********************************************** ok: [10.8.169.22] => { "openshift_image_tag": "v3.3" } TASK [openshift_version : debug] *********************************************** ok: [10.8.169.22] => { "openshift_pkg_version": "-3.3.1.9" } PLAY [Set openshift_version for all hosts] ************************************* TASK [setup] ******************************************************************* ok: [10.8.169.210] ok: [10.8.169.215] TASK [openshift_facts : Detecting Operating System] **************************** ok: [10.8.169.210] ok: [10.8.169.215] TASK [openshift_facts : set_fact] ********************************************** ok: [10.8.169.215] ok: [10.8.169.210] TASK [openshift_facts : set_fact] ********************************************** ok: [10.8.169.215] ok: [10.8.169.210] TASK [openshift_facts : Ensure various deps are installed] ********************* ok: [10.8.169.210] => (item=[u'iproute', u'python-dbus', u'python-six', u'PyYAML', u'yum-utils']) ok: [10.8.169.215] => (item=[u'iproute', u'python-dbus', u'python-six', u'PyYAML', u'yum-utils']) TASK [openshift_facts : Gather Cluster facts and set is_containerized if needed] *** ok: [10.8.169.210] ok: [10.8.169.215] TASK [openshift_facts : Set repoquery command] ********************************* ok: [10.8.169.210] ok: [10.8.169.215] TASK [openshift_repos : assert] ************************************************ skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_repos : Ensure libselinux-python is installed] ***************** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_repos : Create any additional repos that are defined] ********** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_repos : Remove the additional repos if no longer defined] ****** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_repos : Configure origin gpg keys if needed] ******************* skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_repos : Configure origin yum repositories RHEL/CentOS] ********* skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_docker_facts : Set docker facts] ******************************* changed: [10.8.169.215] => (item={u'local_facts': {u'blocked_registries': u'', u'additional_registries': u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888,registry.ops.openshift.com', u'log_driver': u'', u'disable_push_dockerhub': u'', u'hosted_registry_insecure': False, u'hosted_registry_network': u'172.30.0.0/16', u'log_options': u'', u'insecure_registries': u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888,registry.ops.openshift.com', u'options': u''}, u'role': u'docker'}) changed: [10.8.169.210] => (item={u'local_facts': {u'blocked_registries': u'', u'additional_registries': u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888,registry.ops.openshift.com', u'log_driver': u'', u'disable_push_dockerhub': u'', u'hosted_registry_insecure': False, u'hosted_registry_network': u'172.30.0.0/16', u'log_options': u'', u'insecure_registries': u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888,registry.ops.openshift.com', u'options': u''}, u'role': u'docker'}) TASK [openshift_docker_facts : set_fact] *************************************** ok: [10.8.169.210] ok: [10.8.169.215] TASK [openshift_docker_facts : set_fact] *************************************** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_docker_facts : set_fact] *************************************** ok: [10.8.169.215] ok: [10.8.169.210] TASK [os_firewall : Assert - Do not use firewalld on Atomic Host] ************** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [os_firewall : Install firewalld packages] ******************************** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [os_firewall : Ensure iptables services are not enabled] ****************** changed: [10.8.169.210] => (item=iptables) changed: [10.8.169.215] => (item=iptables) changed: [10.8.169.210] => (item=ip6tables) changed: [10.8.169.215] => (item=ip6tables) TASK [os_firewall : Start and enable firewalld service] ************************ ok: [10.8.169.215] ok: [10.8.169.210] TASK [os_firewall : need to pause here, otherwise the firewalld service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.210] TASK [os_firewall : Add firewalld allow rules] ********************************* TASK [os_firewall : Remove firewalld allow rules] ****************************** TASK [os_firewall : Ensure firewalld service is not enabled] ******************* skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [os_firewall : Install iptables packages] ********************************* skipping: [10.8.169.210] => (item=[]) skipping: [10.8.169.215] => (item=[]) TASK [os_firewall : Start and enable iptables service] ************************* skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [os_firewall : need to pause here, otherwise the iptables service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.210] TASK [os_firewall : Add iptables allow rules] ********************************** TASK [os_firewall : Remove iptables rules] ************************************* TASK [docker : Get current installed Docker version] *************************** ok: [10.8.169.215] ok: [10.8.169.210] TASK [docker : Error out if Docker pre-installed but too old] ****************** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [docker : Error out if requested Docker is too old] *********************** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [docker : Get latest available version of Docker] ************************* skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [docker : Fail if Docker version requested but downgrade is required] ***** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [docker : Error out if attempting to upgrade Docker across the 1.10 boundary] *** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [docker : Install Docker] ************************************************* ok: [10.8.169.215] ok: [10.8.169.210] TASK [docker : Ensure docker.service.d directory exists] *********************** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [docker : Configure Docker service unit file] ***************************** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [docker : Getting current systemd-udevd exec command] ********************* skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [docker : Assure systemd-udevd.service.d directory exists] **************** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [docker : Create systemd-udevd override file] ***************************** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [docker : stat] *********************************************************** ok: [10.8.169.210] ok: [10.8.169.215] TASK [docker : Set registry params] ******************************************** skipping: [10.8.169.215] => (item={u'reg_conf_var': u'BLOCK_REGISTRY', u'reg_flag': u'--block-registry', u'reg_fact_val': u''}) changed: [10.8.169.215] => (item={u'reg_conf_var': u'ADD_REGISTRY', u'reg_flag': u'--add-registry', u'reg_fact_val': [u'registry.ops.openshift.com', u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888', u'XXXXXXXXXXXXXXXXXXXXXXXXXXX']}) skipping: [10.8.169.210] => (item={u'reg_conf_var': u'BLOCK_REGISTRY', u'reg_flag': u'--block-registry', u'reg_fact_val': u''}) changed: [10.8.169.210] => (item={u'reg_conf_var': u'ADD_REGISTRY', u'reg_flag': u'--add-registry', u'reg_fact_val': [u'registry.ops.openshift.com', u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888', u'XXXXXXXXXXXXXXXXXXXXXXXXXXX']}) changed: [10.8.169.215] => (item={u'reg_conf_var': u'INSECURE_REGISTRY', u'reg_flag': u'--insecure-registry', u'reg_fact_val': [u'registry.ops.openshift.com', u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888']}) changed: [10.8.169.210] => (item={u'reg_conf_var': u'INSECURE_REGISTRY', u'reg_flag': u'--insecure-registry', u'reg_fact_val': [u'registry.ops.openshift.com', u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888']}) TASK [docker : Set Proxy Settings] ********************************************* ok: [10.8.169.210] => (item={u'reg_conf_var': u'HTTP_PROXY', u'reg_fact_val': u''}) ok: [10.8.169.215] => (item={u'reg_conf_var': u'HTTP_PROXY', u'reg_fact_val': u''}) ok: [10.8.169.210] => (item={u'reg_conf_var': u'HTTPS_PROXY', u'reg_fact_val': u''}) ok: [10.8.169.215] => (item={u'reg_conf_var': u'HTTPS_PROXY', u'reg_fact_val': u''}) ok: [10.8.169.210] => (item={u'reg_conf_var': u'NO_PROXY', u'reg_fact_val': u''}) ok: [10.8.169.215] => (item={u'reg_conf_var': u'NO_PROXY', u'reg_fact_val': u''}) TASK [docker : Set various Docker options] ************************************* changed: [10.8.169.210] changed: [10.8.169.215] TASK [docker : Start the Docker service] *************************************** changed: [10.8.169.215] changed: [10.8.169.210] TASK [docker : set_fact] ******************************************************* ok: [10.8.169.215] ok: [10.8.169.210] RUNNING HANDLER [docker : restart docker] ************************************** changed: [10.8.169.210] changed: [10.8.169.215] TASK [openshift_version : set_fact] ******************************************** ok: [10.8.169.210] ok: [10.8.169.215] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_version : set_fact] ******************************************** ok: [10.8.169.215] ok: [10.8.169.210] TASK [openshift_version : Use openshift.common.version fact as version to configure if already installed] *** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_version : Set rpm version to configure if openshift_pkg_version specified] *** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_version : Gather common package version] *********************** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_version : Set containerized version to configure if openshift_image_tag specified] *** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_version : Set containerized version to configure if openshift_release specified] *** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_version : Lookup latest containerized version if no version specified] *** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_version : Set precise containerized version to configure if openshift_release specified] *** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_version : set_fact] ******************************************** ok: [10.8.169.215] ok: [10.8.169.210] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_version : set_fact] ******************************************** ok: [10.8.169.215] ok: [10.8.169.210] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_version : debug] *********************************************** ok: [10.8.169.215] => { "openshift_release": "VARIABLE IS NOT DEFINED!" } ok: [10.8.169.210] => { "openshift_release": "VARIABLE IS NOT DEFINED!" } TASK [openshift_version : debug] *********************************************** ok: [10.8.169.210] => { "openshift_image_tag": "v3.3" } ok: [10.8.169.215] => { "openshift_image_tag": "v3.3" } TASK [openshift_version : debug] *********************************************** ok: [10.8.169.210] => { "openshift_pkg_version": "-3.3.1.9" } ok: [10.8.169.215] => { "openshift_pkg_version": "-3.3.1.9" } PLAY [Set oo_option facts] ***************************************************** TASK [setup] ******************************************************************* ok: [10.8.169.210] ok: [10.8.169.22] ok: [10.8.169.215] TASK [set_fact] **************************************************************** skipping: [10.8.169.215] skipping: [10.8.169.22] skipping: [10.8.169.210] TASK [set_fact] **************************************************************** skipping: [10.8.169.210] skipping: [10.8.169.215] skipping: [10.8.169.22] TASK [set_fact] **************************************************************** ok: [10.8.169.215] ok: [10.8.169.22] ok: [10.8.169.210] TASK [set_fact] **************************************************************** ok: [10.8.169.215] ok: [10.8.169.210] ok: [10.8.169.22] TASK [set_fact] **************************************************************** ok: [10.8.169.210] ok: [10.8.169.22] ok: [10.8.169.215] TASK [set_fact] **************************************************************** ok: [10.8.169.210] ok: [10.8.169.22] ok: [10.8.169.215] PLAY [Gather and set facts for master hosts] *********************************** TASK [setup] ******************************************************************* ok: [10.8.169.22] TASK [Check for RPM generated config marker file .config_managed] ************** ok: [10.8.169.22] TASK [Remove RPM generated config files if present] **************************** skipping: [10.8.169.22] => (item=.config_managed) skipping: [10.8.169.22] => (item=node) skipping: [10.8.169.22] => (item=master) TASK [set_fact] **************************************************************** ok: [10.8.169.22] TASK [set_fact] **************************************************************** ok: [10.8.169.22] TASK [set_fact] **************************************************************** skipping: [10.8.169.22] TASK [set_fact] **************************************************************** ok: [10.8.169.22] TASK [set_fact] **************************************************************** ok: [10.8.169.22] TASK [set_fact] **************************************************************** ok: [10.8.169.22] TASK [set_fact] **************************************************************** ok: [10.8.169.22] TASK [set_fact] **************************************************************** ok: [10.8.169.22] TASK [set_fact] **************************************************************** ok: [10.8.169.22] TASK [openshift_facts : Detecting Operating System] **************************** ok: [10.8.169.22] TASK [openshift_facts : set_fact] ********************************************** ok: [10.8.169.22] TASK [openshift_facts : set_fact] ********************************************** ok: [10.8.169.22] TASK [openshift_facts : Ensure various deps are installed] ********************* ok: [10.8.169.22] => (item=[u'iproute', u'python-dbus', u'python-six', u'PyYAML', u'yum-utils']) TASK [openshift_facts : Gather Cluster facts and set is_containerized if needed] *** ok: [10.8.169.22] TASK [openshift_facts : Set repoquery command] ********************************* ok: [10.8.169.22] TASK [openshift_facts] ********************************************************* changed: [10.8.169.22] TASK [openshift_facts] ********************************************************* skipping: [10.8.169.22] PLAY [Create temp directory for syncing certs] ********************************* TASK [Create local temp directory for syncing certs] *************************** ok: [localhost -> localhost] PLAY [Determine if session secrets must be generated] ************************** TASK [setup] ******************************************************************* ok: [10.8.169.22] TASK [openshift_facts : Detecting Operating System] **************************** ok: [10.8.169.22] TASK [openshift_facts : set_fact] ********************************************** ok: [10.8.169.22] TASK [openshift_facts : set_fact] ********************************************** ok: [10.8.169.22] TASK [openshift_facts : Ensure various deps are installed] ********************* ok: [10.8.169.22] => (item=[u'iproute', u'python-dbus', u'python-six', u'PyYAML', u'yum-utils']) TASK [openshift_facts : Gather Cluster facts and set is_containerized if needed] *** ok: [10.8.169.22] TASK [openshift_facts : Set repoquery command] ********************************* ok: [10.8.169.22] TASK [openshift_facts] ********************************************************* ok: [10.8.169.22] PLAY [Generate master session secrets] ***************************************** TASK [setup] ******************************************************************* ok: [10.8.169.22] TASK [openshift_facts : Detecting Operating System] **************************** ok: [10.8.169.22] TASK [openshift_facts : set_fact] ********************************************** ok: [10.8.169.22] TASK [openshift_facts : set_fact] ********************************************** ok: [10.8.169.22] TASK [openshift_facts : Ensure various deps are installed] ********************* ok: [10.8.169.22] => (item=[u'iproute', u'python-dbus', u'python-six', u'PyYAML', u'yum-utils']) TASK [openshift_facts : Gather Cluster facts and set is_containerized if needed] *** ok: [10.8.169.22] TASK [openshift_facts : Set repoquery command] ********************************* ok: [10.8.169.22] TASK [openshift_facts] ********************************************************* changed: [10.8.169.22] PLAY [Configure masters] ******************************************************* TASK [setup] ******************************************************************* ok: [10.8.169.22] TASK [openshift_facts : Detecting Operating System] **************************** ok: [10.8.169.22] TASK [openshift_facts : set_fact] ********************************************** ok: [10.8.169.22] TASK [openshift_facts : set_fact] ********************************************** ok: [10.8.169.22] TASK [openshift_facts : Ensure various deps are installed] ********************* ok: [10.8.169.22] => (item=[u'iproute', u'python-dbus', u'python-six', u'PyYAML', u'yum-utils']) TASK [openshift_facts : Gather Cluster facts and set is_containerized if needed] *** changed: [10.8.169.22] TASK [openshift_facts : Set repoquery command] ********************************* ok: [10.8.169.22] TASK [openshift_master_facts : Migrate legacy osm_default_subdomain fact] ****** skipping: [10.8.169.22] TASK [openshift_master_facts : fail] ******************************************* skipping: [10.8.169.22] TASK [openshift_master_facts : set_fact] *************************************** skipping: [10.8.169.22] TASK [openshift_master_facts : set_fact] *************************************** skipping: [10.8.169.22] TASK [openshift_master_facts : Set master facts] ******************************* changed: [10.8.169.22] TASK [openshift_master_facts : Determine if scheduler config present] ********** ok: [10.8.169.22] TASK [openshift_master_facts : set_fact] *************************************** ok: [10.8.169.22] TASK [openshift_master_facts : Retrieve current scheduler config] ************** skipping: [10.8.169.22] TASK [openshift_master_facts : set_fact] *************************************** skipping: [10.8.169.22] TASK [openshift_master_facts : fail] ******************************************* skipping: [10.8.169.22] TASK [openshift_master_facts : set_fact] *************************************** skipping: [10.8.169.22] TASK [openshift_hosted_facts : set_fact] *************************************** skipping: [10.8.169.22] TASK [openshift_hosted_facts : set_fact] *************************************** skipping: [10.8.169.22] TASK [openshift_hosted_facts : Set hosted facts] ******************************* changed: [10.8.169.22] TASK [openshift_repos : assert] ************************************************ skipping: [10.8.169.22] TASK [openshift_repos : Ensure libselinux-python is installed] ***************** skipping: [10.8.169.22] TASK [openshift_repos : Create any additional repos that are defined] ********** skipping: [10.8.169.22] TASK [openshift_repos : Remove the additional repos if no longer defined] ****** skipping: [10.8.169.22] TASK [openshift_repos : Configure origin gpg keys if needed] ******************* skipping: [10.8.169.22] TASK [openshift_repos : Configure origin yum repositories RHEL/CentOS] ********* skipping: [10.8.169.22] TASK [openshift_repos : assert] ************************************************ skipping: [10.8.169.22] TASK [openshift_repos : Ensure libselinux-python is installed] ***************** skipping: [10.8.169.22] TASK [openshift_repos : Create any additional repos that are defined] ********** skipping: [10.8.169.22] TASK [openshift_repos : Remove the additional repos if no longer defined] ****** skipping: [10.8.169.22] TASK [openshift_repos : Configure origin gpg keys if needed] ******************* skipping: [10.8.169.22] TASK [openshift_repos : Configure origin yum repositories RHEL/CentOS] ********* skipping: [10.8.169.22] TASK [openshift_docker_facts : Set docker facts] ******************************* ok: [10.8.169.22] => (item={u'local_facts': {u'blocked_registries': u'', u'additional_registries': u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888,registry.ops.openshift.com', u'log_driver': u'', u'disable_push_dockerhub': u'', u'hosted_registry_insecure': False, u'hosted_registry_network': u'172.30.0.0/16', u'log_options': u'', u'insecure_registries': u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888,registry.ops.openshift.com', u'options': u''}, u'role': u'docker'}) TASK [openshift_docker_facts : set_fact] *************************************** ok: [10.8.169.22] TASK [openshift_docker_facts : set_fact] *************************************** skipping: [10.8.169.22] TASK [openshift_docker_facts : set_fact] *************************************** ok: [10.8.169.22] TASK [os_firewall : Assert - Do not use firewalld on Atomic Host] ************** skipping: [10.8.169.22] TASK [os_firewall : Install firewalld packages] ******************************** skipping: [10.8.169.22] TASK [os_firewall : Ensure iptables services are not enabled] ****************** ok: [10.8.169.22] => (item=iptables) ok: [10.8.169.22] => (item=ip6tables) TASK [os_firewall : Start and enable firewalld service] ************************ ok: [10.8.169.22] TASK [os_firewall : need to pause here, otherwise the firewalld service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.22] TASK [os_firewall : Add firewalld allow rules] ********************************* TASK [os_firewall : Remove firewalld allow rules] ****************************** TASK [os_firewall : Ensure firewalld service is not enabled] ******************* skipping: [10.8.169.22] TASK [os_firewall : Install iptables packages] ********************************* skipping: [10.8.169.22] => (item=[]) TASK [os_firewall : Start and enable iptables service] ************************* skipping: [10.8.169.22] TASK [os_firewall : need to pause here, otherwise the iptables service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.22] TASK [os_firewall : Add iptables allow rules] ********************************** TASK [os_firewall : Remove iptables rules] ************************************* TASK [docker : Get current installed Docker version] *************************** ok: [10.8.169.22] TASK [docker : Error out if Docker pre-installed but too old] ****************** skipping: [10.8.169.22] TASK [docker : Error out if requested Docker is too old] *********************** skipping: [10.8.169.22] TASK [docker : Get latest available version of Docker] ************************* skipping: [10.8.169.22] TASK [docker : Fail if Docker version requested but downgrade is required] ***** skipping: [10.8.169.22] TASK [docker : Error out if attempting to upgrade Docker across the 1.10 boundary] *** skipping: [10.8.169.22] TASK [docker : Install Docker] ************************************************* ok: [10.8.169.22] TASK [docker : Ensure docker.service.d directory exists] *********************** skipping: [10.8.169.22] TASK [docker : Configure Docker service unit file] ***************************** skipping: [10.8.169.22] TASK [docker : Getting current systemd-udevd exec command] ********************* skipping: [10.8.169.22] TASK [docker : Assure systemd-udevd.service.d directory exists] **************** skipping: [10.8.169.22] TASK [docker : Create systemd-udevd override file] ***************************** skipping: [10.8.169.22] TASK [docker : stat] *********************************************************** ok: [10.8.169.22] TASK [docker : Set registry params] ******************************************** skipping: [10.8.169.22] => (item={u'reg_conf_var': u'BLOCK_REGISTRY', u'reg_flag': u'--block-registry', u'reg_fact_val': u''}) ok: [10.8.169.22] => (item={u'reg_conf_var': u'ADD_REGISTRY', u'reg_flag': u'--add-registry', u'reg_fact_val': [u'registry.ops.openshift.com', u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888', u'XXXXXXXXXXXXXXXXXXXXXXXXXXX']}) ok: [10.8.169.22] => (item={u'reg_conf_var': u'INSECURE_REGISTRY', u'reg_flag': u'--insecure-registry', u'reg_fact_val': [u'registry.ops.openshift.com', u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888']}) TASK [docker : Set Proxy Settings] ********************************************* ok: [10.8.169.22] => (item={u'reg_conf_var': u'HTTP_PROXY', u'reg_fact_val': u''}) ok: [10.8.169.22] => (item={u'reg_conf_var': u'HTTPS_PROXY', u'reg_fact_val': u''}) ok: [10.8.169.22] => (item={u'reg_conf_var': u'NO_PROXY', u'reg_fact_val': u''}) TASK [docker : Set various Docker options] ************************************* ok: [10.8.169.22] TASK [docker : Start the Docker service] *************************************** ok: [10.8.169.22] TASK [docker : set_fact] ******************************************************* ok: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** ok: [10.8.169.22] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** ok: [10.8.169.22] TASK [openshift_version : Use openshift.common.version fact as version to configure if already installed] *** skipping: [10.8.169.22] TASK [openshift_version : Set rpm version to configure if openshift_pkg_version specified] *** skipping: [10.8.169.22] TASK [openshift_version : Gather common package version] *********************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : Set containerized version to configure if openshift_image_tag specified] *** skipping: [10.8.169.22] TASK [openshift_version : Set containerized version to configure if openshift_release specified] *** skipping: [10.8.169.22] TASK [openshift_version : Lookup latest containerized version if no version specified] *** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : Set precise containerized version to configure if openshift_release specified] *** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** ok: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.22] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.22] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.22] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.22] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.22] TASK [openshift_version : debug] *********************************************** ok: [10.8.169.22] => { "openshift_release": "VARIABLE IS NOT DEFINED!" } TASK [openshift_version : debug] *********************************************** ok: [10.8.169.22] => { "openshift_image_tag": "v3.3" } TASK [openshift_version : debug] *********************************************** ok: [10.8.169.22] => { "openshift_pkg_version": "-3.3.1.9" } TASK [os_firewall : Assert - Do not use firewalld on Atomic Host] ************** skipping: [10.8.169.22] TASK [os_firewall : Install firewalld packages] ******************************** skipping: [10.8.169.22] TASK [os_firewall : Ensure iptables services are not enabled] ****************** ok: [10.8.169.22] => (item=iptables) ok: [10.8.169.22] => (item=ip6tables) TASK [os_firewall : Start and enable firewalld service] ************************ ok: [10.8.169.22] TASK [os_firewall : need to pause here, otherwise the firewalld service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.22] TASK [os_firewall : Add firewalld allow rules] ********************************* TASK [os_firewall : Remove firewalld allow rules] ****************************** TASK [os_firewall : Ensure firewalld service is not enabled] ******************* skipping: [10.8.169.22] TASK [os_firewall : Install iptables packages] ********************************* skipping: [10.8.169.22] => (item=[]) TASK [os_firewall : Start and enable iptables service] ************************* skipping: [10.8.169.22] TASK [os_firewall : need to pause here, otherwise the iptables service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.22] TASK [os_firewall : Add iptables allow rules] ********************************** TASK [os_firewall : Remove iptables rules] ************************************* TASK [docker : Get current installed Docker version] *************************** ok: [10.8.169.22] TASK [docker : Error out if Docker pre-installed but too old] ****************** skipping: [10.8.169.22] TASK [docker : Error out if requested Docker is too old] *********************** skipping: [10.8.169.22] TASK [docker : Get latest available version of Docker] ************************* skipping: [10.8.169.22] TASK [docker : Fail if Docker version requested but downgrade is required] ***** skipping: [10.8.169.22] TASK [docker : Error out if attempting to upgrade Docker across the 1.10 boundary] *** skipping: [10.8.169.22] TASK [docker : Install Docker] ************************************************* ok: [10.8.169.22] TASK [docker : Ensure docker.service.d directory exists] *********************** skipping: [10.8.169.22] TASK [docker : Configure Docker service unit file] ***************************** skipping: [10.8.169.22] TASK [docker : Getting current systemd-udevd exec command] ********************* skipping: [10.8.169.22] TASK [docker : Assure systemd-udevd.service.d directory exists] **************** skipping: [10.8.169.22] TASK [docker : Create systemd-udevd override file] ***************************** skipping: [10.8.169.22] TASK [docker : stat] *********************************************************** ok: [10.8.169.22] TASK [docker : Set registry params] ******************************************** skipping: [10.8.169.22] => (item={u'reg_conf_var': u'BLOCK_REGISTRY', u'reg_flag': u'--block-registry', u'reg_fact_val': u''}) ok: [10.8.169.22] => (item={u'reg_conf_var': u'ADD_REGISTRY', u'reg_flag': u'--add-registry', u'reg_fact_val': [u'registry.ops.openshift.com', u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888', u'XXXXXXXXXXXXXXXXXXXXXXXXXXX']}) ok: [10.8.169.22] => (item={u'reg_conf_var': u'INSECURE_REGISTRY', u'reg_flag': u'--insecure-registry', u'reg_fact_val': [u'registry.ops.openshift.com', u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888']}) TASK [docker : Set Proxy Settings] ********************************************* ok: [10.8.169.22] => (item={u'reg_conf_var': u'HTTP_PROXY', u'reg_fact_val': u''}) ok: [10.8.169.22] => (item={u'reg_conf_var': u'HTTPS_PROXY', u'reg_fact_val': u''}) ok: [10.8.169.22] => (item={u'reg_conf_var': u'NO_PROXY', u'reg_fact_val': u''}) TASK [docker : Set various Docker options] ************************************* ok: [10.8.169.22] TASK [docker : Start the Docker service] *************************************** ok: [10.8.169.22] TASK [docker : set_fact] ******************************************************* ok: [10.8.169.22] TASK [openshift_repos : assert] ************************************************ skipping: [10.8.169.22] TASK [openshift_repos : Ensure libselinux-python is installed] ***************** skipping: [10.8.169.22] TASK [openshift_repos : Create any additional repos that are defined] ********** skipping: [10.8.169.22] TASK [openshift_repos : Remove the additional repos if no longer defined] ****** skipping: [10.8.169.22] TASK [openshift_repos : Configure origin gpg keys if needed] ******************* skipping: [10.8.169.22] TASK [openshift_repos : Configure origin yum repositories RHEL/CentOS] ********* skipping: [10.8.169.22] TASK [openshift_repos : assert] ************************************************ skipping: [10.8.169.22] TASK [openshift_repos : Ensure libselinux-python is installed] ***************** skipping: [10.8.169.22] TASK [openshift_repos : Create any additional repos that are defined] ********** skipping: [10.8.169.22] TASK [openshift_repos : Remove the additional repos if no longer defined] ****** skipping: [10.8.169.22] TASK [openshift_repos : Configure origin gpg keys if needed] ******************* skipping: [10.8.169.22] TASK [openshift_repos : Configure origin yum repositories RHEL/CentOS] ********* skipping: [10.8.169.22] TASK [os_firewall : Assert - Do not use firewalld on Atomic Host] ************** skipping: [10.8.169.22] TASK [os_firewall : Install firewalld packages] ******************************** skipping: [10.8.169.22] TASK [os_firewall : Ensure iptables services are not enabled] ****************** ok: [10.8.169.22] => (item=iptables) ok: [10.8.169.22] => (item=ip6tables) TASK [os_firewall : Start and enable firewalld service] ************************ ok: [10.8.169.22] TASK [os_firewall : need to pause here, otherwise the firewalld service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.22] TASK [os_firewall : Add firewalld allow rules] ********************************* TASK [os_firewall : Remove firewalld allow rules] ****************************** TASK [os_firewall : Ensure firewalld service is not enabled] ******************* skipping: [10.8.169.22] TASK [os_firewall : Install iptables packages] ********************************* skipping: [10.8.169.22] => (item=[]) TASK [os_firewall : Start and enable iptables service] ************************* skipping: [10.8.169.22] TASK [os_firewall : need to pause here, otherwise the iptables service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.22] TASK [os_firewall : Add iptables allow rules] ********************************** TASK [os_firewall : Remove iptables rules] ************************************* TASK [openshift_common : fail] ************************************************* skipping: [10.8.169.22] TASK [openshift_common : fail] ************************************************* skipping: [10.8.169.22] TASK [openshift_common : fail] ************************************************* skipping: [10.8.169.22] TASK [openshift_common : fail] ************************************************* skipping: [10.8.169.22] TASK [openshift_common : Set common Cluster facts] ***************************** changed: [10.8.169.22] TASK [openshift_common : Install the base package for versioning] ************** skipping: [10.8.169.22] TASK [openshift_common : Set version facts] ************************************ ok: [10.8.169.22] TASK [openshift_common : set_fact] ********************************************* ok: [10.8.169.22] TASK [openshift_common : Set hostname] ***************************************** changed: [10.8.169.22] TASK [openshift_cli_facts : openshift_facts] *********************************** ok: [10.8.169.22] TASK [openshift_cli : Install clients] ***************************************** skipping: [10.8.169.22] TASK [openshift_cli : Pull CLI Image] ****************************************** ok: [10.8.169.22] TASK [openshift_cli : Copy client binaries/symlinks out of CLI image for use on the host] *** changed: [10.8.169.22] TASK [openshift_cli : Reload facts to pick up installed OpenShift version] ***** ok: [10.8.169.22] TASK [openshift_cli : Install bash completion for oc tools] ******************** skipping: [10.8.169.22] TASK [openshift_named_certificates : set_fact] ********************************* skipping: [10.8.169.22] TASK [openshift_named_certificates : openshift_facts] ************************** changed: [10.8.169.22] TASK [openshift_named_certificates : Clear named certificates] ***************** skipping: [10.8.169.22] TASK [openshift_named_certificates : Ensure named certificate directory exists] changed: [10.8.169.22] TASK [openshift_named_certificates : Land named certificates] ****************** TASK [openshift_named_certificates : Land named certificate keys] ************** TASK [openshift_named_certificates : Land named CA certificates] *************** TASK [openshift_ca : fail] ***************************************************** skipping: [10.8.169.22] TASK [openshift_ca : fail] ***************************************************** skipping: [10.8.169.22] TASK [openshift_ca : Install the base package for admin tooling] *************** skipping: [10.8.169.22] TASK [openshift_ca : Reload generated facts] *********************************** skipping: [10.8.169.22] TASK [openshift_ca : Create openshift_ca_config_dir if it does not exist] ****** ok: [10.8.169.22 -> 10.8.169.22] TASK [openshift_ca : Determine if CA must be created] ************************** ok: [10.8.169.22 -> 10.8.169.22] => (item=ca-bundle.crt) ok: [10.8.169.22 -> 10.8.169.22] => (item=ca.crt) ok: [10.8.169.22 -> 10.8.169.22] => (item=ca.key) TASK [openshift_ca : set_fact] ************************************************* ok: [10.8.169.22] TASK [openshift_ca : Retain original serviceaccount keys] ********************** skipping: [10.8.169.22] => (item=/etc/origin/master/serviceaccounts.public.key) skipping: [10.8.169.22] => (item=/etc/origin/master/serviceaccounts.private.key) TASK [openshift_ca : Deploy master ca certificate] ***************************** skipping: [10.8.169.22] => (item={u'dest': u'ca.key', u'src': u''}) skipping: [10.8.169.22] => (item={u'dest': u'ca.crt', u'src': u''}) TASK [openshift_ca : Create ca serial] ***************************************** skipping: [10.8.169.22] TASK [openshift_ca : Create the master certificates if they do not already exist] *** changed: [10.8.169.22 -> 10.8.169.22] TASK [openshift_ca : Restore original serviceaccount keys] ********************* skipping: [10.8.169.22] => (item=/etc/origin/master/serviceaccounts.public.key) skipping: [10.8.169.22] => (item=/etc/origin/master/serviceaccounts.private.key) TASK [openshift_ca : Remove backup serviceaccount keys] ************************ skipping: [10.8.169.22] => (item=/etc/origin/master/serviceaccounts.public.key) skipping: [10.8.169.22] => (item=/etc/origin/master/serviceaccounts.private.key) TASK [openshift_master_certificates : set_fact] ******************************** ok: [10.8.169.22] TASK [openshift_master_certificates : set_fact] ******************************** ok: [10.8.169.22] TASK [openshift_master_certificates : Check status of master certificates] ***** ok: [10.8.169.22] => (item=admin.crt) ok: [10.8.169.22] => (item=master.kubelet-client.crt) ok: [10.8.169.22] => (item=master.proxy-client.crt) ok: [10.8.169.22] => (item=master.server.crt) ok: [10.8.169.22] => (item=openshift-master.crt) ok: [10.8.169.22] => (item=openshift-registry.crt) ok: [10.8.169.22] => (item=openshift-router.crt) ok: [10.8.169.22] => (item=etcd.server.crt) TASK [openshift_master_certificates : set_fact] ******************************** ok: [10.8.169.22] TASK [openshift_master_certificates : Ensure the generated_configs directory present] *** skipping: [10.8.169.22] TASK [openshift_master_certificates : file] ************************************ skipping: [10.8.169.22] => (item=ca.serial.txt) skipping: [10.8.169.22] => (item=ca.key) skipping: [10.8.169.22] => (item=ca.crt) TASK [openshift_master_certificates : Create the master certificates if they do not already exist] *** skipping: [10.8.169.22] TASK [openshift_master_certificates : file] ************************************ skipping: [10.8.169.22] => (item=admin.key) skipping: [10.8.169.22] => (item=admin.crt) skipping: [10.8.169.22] => (item=service-signer.key) skipping: [10.8.169.22] => (item=service-signer.crt) skipping: [10.8.169.22] => (item=master.proxy-client.key) skipping: [10.8.169.22] => (item=master.proxy-client.crt) skipping: [10.8.169.22] => (item=serviceaccounts.public.key) skipping: [10.8.169.22] => (item=serviceaccounts.private.key) skipping: [10.8.169.22] => (item=openshift-router.kubeconfig) skipping: [10.8.169.22] => (item=openshift-router.key) skipping: [10.8.169.22] => (item=openshift-router.crt) skipping: [10.8.169.22] => (item=openshift-registry.kubeconfig) skipping: [10.8.169.22] => (item=openshift-registry.key) skipping: [10.8.169.22] => (item=openshift-registry.crt) skipping: [10.8.169.22] => (item=master.kubelet-client.key) skipping: [10.8.169.22] => (item=master.kubelet-client.crt) skipping: [10.8.169.22] => (item=admin.kubeconfig) TASK [openshift_master_certificates : Remove generated etcd client certs when using external etcd] *** skipping: [10.8.169.22] => (item=master.etcd-client.key) skipping: [10.8.169.22] => (item=master.etcd-client.crt) TASK [openshift_master_certificates : Create local temp directory for syncing certs] *** skipping: [10.8.169.22] TASK [openshift_master_certificates : Create a tarball of the master certs] **** skipping: [10.8.169.22] TASK [openshift_master_certificates : Retrieve the master cert tarball from the master] *** skipping: [10.8.169.22] TASK [openshift_master_certificates : Ensure certificate directory exists] ***** skipping: [10.8.169.22] TASK [openshift_master_certificates : Unarchive the tarball on the master] ***** skipping: [10.8.169.22] TASK [openshift_master_certificates : file] ************************************ skipping: [10.8.169.22] TASK [openshift_master_certificates : Lookup default group for ansible_ssh_user] *** ok: [10.8.169.22] TASK [openshift_master_certificates : set_fact] ******************************** ok: [10.8.169.22] TASK [openshift_master_certificates : Create the client config dir(s)] ********* changed: [10.8.169.22] => (item=root) TASK [openshift_master_certificates : Copy the admin client config(s)] ********* changed: [10.8.169.22] => (item=root) TASK [openshift_master_certificates : Update the permissions on the admin client config(s)] *** changed: [10.8.169.22] => (item=root) TASK [openshift_master_certificates : Check for ca-bundle.crt] ***************** ok: [10.8.169.22] TASK [openshift_master_certificates : Check for ca.crt] ************************ ok: [10.8.169.22] TASK [openshift_master_certificates : Migrate ca.crt to ca-bundle.crt] ********* skipping: [10.8.169.22] TASK [openshift_master_certificates : Link ca.crt to ca-bundle.crt] ************ skipping: [10.8.169.22] TASK [openshift_etcd_facts : openshift_facts] ********************************** skipping: [10.8.169.22] TASK [openshift_repos : assert] ************************************************ skipping: [10.8.169.22] TASK [openshift_repos : Ensure libselinux-python is installed] ***************** skipping: [10.8.169.22] TASK [openshift_repos : Create any additional repos that are defined] ********** skipping: [10.8.169.22] TASK [openshift_repos : Remove the additional repos if no longer defined] ****** skipping: [10.8.169.22] TASK [openshift_repos : Configure origin gpg keys if needed] ******************* skipping: [10.8.169.22] TASK [openshift_repos : Configure origin yum repositories RHEL/CentOS] ********* skipping: [10.8.169.22] TASK [etcd_client_certificates : Ensure CA certificate exists on etcd_ca_host] * skipping: [10.8.169.22] TASK [etcd_client_certificates : fail] ***************************************** skipping: [10.8.169.22] TASK [etcd_client_certificates : Check status of external etcd certificatees] ** skipping: [10.8.169.22] => (item=master.etcd-ca.crt) skipping: [10.8.169.22] => (item=master.etcd-client.key) skipping: [10.8.169.22] => (item=master.etcd-client.crt) TASK [etcd_client_certificates : set_fact] ************************************* skipping: [10.8.169.22] TASK [etcd_client_certificates : Ensure generated_certs directory present] ***** skipping: [10.8.169.22] TASK [etcd_client_certificates : Create the client csr] ************************ skipping: [10.8.169.22] TASK [etcd_client_certificates : Sign and create the client crt] *************** skipping: [10.8.169.22] TASK [etcd_client_certificates : file] ***************************************** skipping: [10.8.169.22] TASK [etcd_client_certificates : Create local temp directory for syncing certs] skipping: [10.8.169.22] TASK [etcd_client_certificates : Create a tarball of the etcd certs] *********** skipping: [10.8.169.22] TASK [etcd_client_certificates : Retrieve the etcd cert tarballs] ************** skipping: [10.8.169.22] TASK [etcd_client_certificates : Ensure certificate directory exists] ********** skipping: [10.8.169.22] TASK [etcd_client_certificates : Unarchive etcd cert tarballs] ***************** skipping: [10.8.169.22] TASK [etcd_client_certificates : file] ***************************************** skipping: [10.8.169.22] => (item=master.etcd-ca.crt) skipping: [10.8.169.22] => (item=master.etcd-client.key) skipping: [10.8.169.22] => (item=master.etcd-client.crt) TASK [etcd_client_certificates : Delete temporary directory] ******************* skipping: [10.8.169.22] TASK [openshift_clock : Set clock facts] *************************************** changed: [10.8.169.22] TASK [openshift_clock : Install ntp package] *********************************** skipping: [10.8.169.22] TASK [openshift_clock : Start and enable ntpd/chronyd] ************************* changed: [10.8.169.22] TASK [openshift_cloud_provider : Set cloud provider facts] ********************* changed: [10.8.169.22] TASK [openshift_cloud_provider : Create cloudprovider config dir] ************** skipping: [10.8.169.22] TASK [openshift_cloud_provider : fail] ***************************************** skipping: [10.8.169.22] TASK [openshift_cloud_provider : Create cloud config] ************************** skipping: [10.8.169.22] TASK [openshift_cloud_provider : Create cloud config file] ********************* skipping: [10.8.169.22] TASK [openshift_cloud_provider : Configure AWS cloud provider] ***************** skipping: [10.8.169.22] TASK [openshift_cloud_provider : Create cloud config file] ********************* skipping: [10.8.169.22] TASK [openshift_cloud_provider : Configure GCE cloud provider] ***************** skipping: [10.8.169.22] TASK [openshift_builddefaults : Set builddefaults] ***************************** changed: [10.8.169.22] TASK [openshift_builddefaults : Set builddefaults config structure] ************ changed: [10.8.169.22] TASK [os_firewall : Assert - Do not use firewalld on Atomic Host] ************** skipping: [10.8.169.22] TASK [os_firewall : Install firewalld packages] ******************************** skipping: [10.8.169.22] TASK [os_firewall : Ensure iptables services are not enabled] ****************** ok: [10.8.169.22] => (item=iptables) ok: [10.8.169.22] => (item=ip6tables) TASK [os_firewall : Start and enable firewalld service] ************************ ok: [10.8.169.22] TASK [os_firewall : need to pause here, otherwise the firewalld service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.22] TASK [os_firewall : Add firewalld allow rules] ********************************* changed: [10.8.169.22] => (item={u'port': u'8443/tcp', u'service': u'api server https'}) changed: [10.8.169.22] => (item={u'port': u'8444/tcp', u'service': u'api controllers https'}) changed: [10.8.169.22] => (item={u'port': u'8053/tcp', u'service': u'skydns tcp'}) changed: [10.8.169.22] => (item={u'port': u'8053/udp', u'service': u'skydns udp'}) TASK [os_firewall : Remove firewalld allow rules] ****************************** TASK [os_firewall : Ensure firewalld service is not enabled] ******************* skipping: [10.8.169.22] TASK [os_firewall : Install iptables packages] ********************************* skipping: [10.8.169.22] => (item=[]) TASK [os_firewall : Start and enable iptables service] ************************* skipping: [10.8.169.22] TASK [os_firewall : need to pause here, otherwise the iptables service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.22] TASK [os_firewall : Add iptables allow rules] ********************************** skipping: [10.8.169.22] => (item={u'port': u'8053/udp', u'service': u'skydns udp'}) skipping: [10.8.169.22] => (item={u'port': u'8053/tcp', u'service': u'skydns tcp'}) skipping: [10.8.169.22] => (item={u'port': u'8444/tcp', u'service': u'api controllers https'}) skipping: [10.8.169.22] => (item={u'port': u'8443/tcp', u'service': u'api server https'}) TASK [os_firewall : Remove iptables rules] ************************************* TASK [os_firewall : Assert - Do not use firewalld on Atomic Host] ************** skipping: [10.8.169.22] TASK [os_firewall : Install firewalld packages] ******************************** skipping: [10.8.169.22] TASK [os_firewall : Ensure iptables services are not enabled] ****************** ok: [10.8.169.22] => (item=iptables) ok: [10.8.169.22] => (item=ip6tables) TASK [os_firewall : Start and enable firewalld service] ************************ ok: [10.8.169.22] TASK [os_firewall : need to pause here, otherwise the firewalld service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.22] TASK [os_firewall : Add firewalld allow rules] ********************************* changed: [10.8.169.22] => (item={u'port': u'4001/tcp', u'service': u'etcd embedded'}) TASK [os_firewall : Remove firewalld allow rules] ****************************** TASK [os_firewall : Ensure firewalld service is not enabled] ******************* skipping: [10.8.169.22] TASK [os_firewall : Install iptables packages] ********************************* skipping: [10.8.169.22] => (item=[]) TASK [os_firewall : Start and enable iptables service] ************************* skipping: [10.8.169.22] TASK [os_firewall : need to pause here, otherwise the iptables service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.22] TASK [os_firewall : Add iptables allow rules] ********************************** skipping: [10.8.169.22] => (item={u'port': u'4001/tcp', u'service': u'etcd embedded'}) TASK [os_firewall : Remove iptables rules] ************************************* TASK [nickhammond.logrotate : nickhammond.logrotate | Install logrotate] ******* ok: [10.8.169.22] TASK [nickhammond.logrotate : nickhammond.logrotate | Setup logrotate.d scripts] *** TASK [nuage_ca : Install openssl] ********************************************** skipping: [10.8.169.22] TASK [nuage_ca : Create CA directory] ****************************************** skipping: [10.8.169.22] TASK [nuage_ca : Create certificate directory] ********************************* skipping: [10.8.169.22] TASK [nuage_ca : Check if the CA key already exists] *************************** skipping: [10.8.169.22] TASK [nuage_ca : Create CA key] ************************************************ skipping: [10.8.169.22] TASK [nuage_ca : Check if the CA crt already exists] *************************** skipping: [10.8.169.22] TASK [nuage_ca : Create CA crt] ************************************************ skipping: [10.8.169.22] TASK [nuage_ca : Create the serial file] *************************************** skipping: [10.8.169.22] TASK [nuage_ca : Copy SSL config file] ***************************************** skipping: [10.8.169.22] TASK [openshift_etcd_facts : openshift_facts] ********************************** skipping: [10.8.169.22] TASK [openshift_repos : assert] ************************************************ skipping: [10.8.169.22] TASK [openshift_repos : Ensure libselinux-python is installed] ***************** skipping: [10.8.169.22] TASK [openshift_repos : Create any additional repos that are defined] ********** skipping: [10.8.169.22] TASK [openshift_repos : Remove the additional repos if no longer defined] ****** skipping: [10.8.169.22] TASK [openshift_repos : Configure origin gpg keys if needed] ******************* skipping: [10.8.169.22] TASK [openshift_repos : Configure origin yum repositories RHEL/CentOS] ********* skipping: [10.8.169.22] TASK [etcd_client_certificates : Ensure CA certificate exists on etcd_ca_host] * skipping: [10.8.169.22] TASK [etcd_client_certificates : fail] ***************************************** skipping: [10.8.169.22] TASK [etcd_client_certificates : Check status of external etcd certificatees] ** skipping: [10.8.169.22] => (item=master.etcd-ca.crt) skipping: [10.8.169.22] => (item=master.etcd-client.key) skipping: [10.8.169.22] => (item=master.etcd-client.crt) TASK [etcd_client_certificates : set_fact] ************************************* skipping: [10.8.169.22] TASK [etcd_client_certificates : Ensure generated_certs directory present] ***** skipping: [10.8.169.22] TASK [etcd_client_certificates : Create the client csr] ************************ skipping: [10.8.169.22] TASK [etcd_client_certificates : Sign and create the client crt] *************** skipping: [10.8.169.22] TASK [etcd_client_certificates : file] ***************************************** skipping: [10.8.169.22] TASK [etcd_client_certificates : Create local temp directory for syncing certs] skipping: [10.8.169.22] TASK [etcd_client_certificates : Create a tarball of the etcd certs] *********** skipping: [10.8.169.22] TASK [etcd_client_certificates : Retrieve the etcd cert tarballs] ************** skipping: [10.8.169.22] TASK [etcd_client_certificates : Ensure certificate directory exists] ********** skipping: [10.8.169.22] TASK [etcd_client_certificates : Unarchive etcd cert tarballs] ***************** skipping: [10.8.169.22] TASK [etcd_client_certificates : file] ***************************************** skipping: [10.8.169.22] => (item=master.etcd-ca.crt) skipping: [10.8.169.22] => (item=master.etcd-client.key) skipping: [10.8.169.22] => (item=master.etcd-client.crt) TASK [etcd_client_certificates : Delete temporary directory] ******************* skipping: [10.8.169.22] TASK [os_firewall : Assert - Do not use firewalld on Atomic Host] ************** skipping: [10.8.169.22] TASK [os_firewall : Install firewalld packages] ******************************** skipping: [10.8.169.22] TASK [os_firewall : Ensure iptables services are not enabled] ****************** skipping: [10.8.169.22] => (item=ip6tables) skipping: [10.8.169.22] => (item=iptables) TASK [os_firewall : Start and enable firewalld service] ************************ skipping: [10.8.169.22] TASK [os_firewall : need to pause here, otherwise the firewalld service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.22] TASK [os_firewall : Add firewalld allow rules] ********************************* skipping: [10.8.169.22] => (item={u'port': u'9443/tcp', u'service': u'openshift-monitor'}) TASK [os_firewall : Remove firewalld allow rules] ****************************** TASK [os_firewall : Ensure firewalld service is not enabled] ******************* skipping: [10.8.169.22] TASK [os_firewall : Install iptables packages] ********************************* skipping: [10.8.169.22] => (item=[]) TASK [os_firewall : Start and enable iptables service] ************************* skipping: [10.8.169.22] TASK [os_firewall : need to pause here, otherwise the iptables service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.22] TASK [os_firewall : Add iptables allow rules] ********************************** skipping: [10.8.169.22] => (item={u'port': u'9443/tcp', u'service': u'openshift-monitor'}) TASK [os_firewall : Remove iptables rules] ************************************* TASK [nuage_master : Create directory /usr/share/nuage-openshift-monitor] ****** skipping: [10.8.169.22] TASK [nuage_master : Create the log directory] ********************************* skipping: [10.8.169.22] TASK [nuage_master : Install Nuage Openshift Monitor] ************************** skipping: [10.8.169.22] TASK [nuage_master : Create temporary directory for admin kubeconfig] ********** skipping: [10.8.169.22] TASK [nuage_master : set_fact] ************************************************* skipping: [10.8.169.22] TASK [nuage_master : Copy Configuration to temporary conf] ********************* skipping: [10.8.169.22] TASK [nuage_master : Create Admin Service Account] ***************************** skipping: [10.8.169.22] TASK [nuage_master : Configure role/user permissions] ************************** skipping: [10.8.169.22] => (item=policy add-cluster-role-to-user cluster-reader system:serviceaccount:default:nuage) TASK [nuage_master : Generate the node client config] ************************** skipping: [10.8.169.22] TASK [nuage_master : Clean temporary configuration file] *********************** skipping: [10.8.169.22] TASK [nuage_master : Download the certs and keys] ****************************** skipping: [10.8.169.22] => (item=nuage.kubeconfig) skipping: [10.8.169.22] => (item=nuage.key) skipping: [10.8.169.22] => (item=nuage.crt) skipping: [10.8.169.22] => (item=ca.crt) TASK [nuage_master : Create a directory to hold the certificates] ************** skipping: [10.8.169.22] TASK [nuage_master : Create the key] ******************************************* skipping: [10.8.169.22] TASK [nuage_master : Create the req file] ************************************** skipping: [10.8.169.22] TASK [nuage_master : Generate the crt file] ************************************ skipping: [10.8.169.22] TASK [nuage_master : Remove the req file] ************************************** skipping: [10.8.169.22] TASK [nuage_master : Copy nuage CA crt] **************************************** skipping: [10.8.169.22] TASK [nuage_master : Archive the certificate dir] ****************************** skipping: [10.8.169.22] TASK [nuage_master : Create a temp directory for the certificates] ************* skipping: [10.8.169.22] TASK [nuage_master : Download the certificates] ******************************** skipping: [10.8.169.22] TASK [nuage_master : Extract the certificates] ********************************* skipping: [10.8.169.22] TASK [nuage_master : Delete the certificates after copy] *********************** skipping: [10.8.169.22] TASK [nuage_master : Delete the temp directory] ******************************** skipping: [10.8.169.22] TASK [nuage_master : Create nuage-openshift-monitor.yaml] ********************** skipping: [10.8.169.22] TASK [openshift_master : fail] ************************************************* skipping: [10.8.169.22] TASK [openshift_master : fail] ************************************************* skipping: [10.8.169.22] TASK [openshift_master : fail] ************************************************* skipping: [10.8.169.22] TASK [openshift_master : fail] ************************************************* skipping: [10.8.169.22] TASK [openshift_master : fail] ************************************************* skipping: [10.8.169.22] TASK [openshift_master : Install Master package] ******************************* skipping: [10.8.169.22] TASK [openshift_master : Create openshift.common.data_dir] ********************* changed: [10.8.169.22] TASK [openshift_master : Reload systemd units] ********************************* skipping: [10.8.169.22] TASK [openshift_master : Re-gather package dependent master facts] ************* ok: [10.8.169.22] TASK [openshift_master : Create config parent directory if it does not exist] ** ok: [10.8.169.22] TASK [openshift_master : Create the policy file if it does not already exist] ** changed: [10.8.169.22] TASK [openshift_master : Create the scheduler config] ************************** changed: [10.8.169.22] TASK [openshift_master : Install httpd-tools if needed] ************************ skipping: [10.8.169.22] => (item={u'login': True, u'challenge': True, u'kind': u'DenyAllPasswordIdentityProvider', u'name': u'deny_all'}) TASK [openshift_master : Ensure htpasswd directory exists] ********************* skipping: [10.8.169.22] => (item={u'login': True, u'challenge': True, u'kind': u'DenyAllPasswordIdentityProvider', u'name': u'deny_all'}) TASK [openshift_master : Create the htpasswd file if needed] ******************* skipping: [10.8.169.22] => (item={u'login': True, u'challenge': True, u'kind': u'DenyAllPasswordIdentityProvider', u'name': u'deny_all'}) TASK [openshift_master : Ensure htpasswd file exists] ************************** skipping: [10.8.169.22] => (item={u'login': True, u'challenge': True, u'kind': u'DenyAllPasswordIdentityProvider', u'name': u'deny_all'}) TASK [openshift_master : Create the ldap ca file if needed] ******************** skipping: [10.8.169.22] => (item={u'login': True, u'challenge': True, u'kind': u'DenyAllPasswordIdentityProvider', u'name': u'deny_all'}) TASK [openshift_master : Create the openid ca file if needed] ****************** skipping: [10.8.169.22] => (item={u'login': True, u'challenge': True, u'kind': u'DenyAllPasswordIdentityProvider', u'name': u'deny_all'}) TASK [openshift_master : Create the request header ca file if needed] ********** skipping: [10.8.169.22] => (item={u'login': True, u'challenge': True, u'kind': u'DenyAllPasswordIdentityProvider', u'name': u'deny_all'}) TASK [openshift_master : Init HA Service Info] ********************************* ok: [10.8.169.22] TASK [openshift_master : Set HA Service Info for containerized installs] ******* ok: [10.8.169.22] TASK [openshift_master : Pre-pull master image] ******************************** ok: [10.8.169.22] TASK [openshift_master : Create the systemd unit files] ************************ changed: [10.8.169.22] TASK [openshift_master : command] ********************************************** changed: [10.8.169.22] TASK [openshift_master : Create the ha systemd unit files] ********************* skipping: [10.8.169.22] => (item=controllers) skipping: [10.8.169.22] => (item=api) TASK [openshift_master : command] ********************************************** skipping: [10.8.169.22] TASK [openshift_master : Preserve Master API Proxy Config options] ************* skipping: [10.8.169.22] TASK [openshift_master : Preserve Master API AWS options] ********************** skipping: [10.8.169.22] TASK [openshift_master : Create the master api service env file] *************** skipping: [10.8.169.22] TASK [openshift_master : Restore Master API Proxy Config Options] ************** TASK [openshift_master : Restore Master API AWS Options] *********************** TASK [openshift_master : Preserve Master Controllers Proxy Config options] ***** skipping: [10.8.169.22] TASK [openshift_master : Preserve Master Controllers AWS options] ************** skipping: [10.8.169.22] TASK [openshift_master : Create the master controllers service env file] ******* skipping: [10.8.169.22] TASK [openshift_master : Restore Master Controllers Proxy Config Options] ****** TASK [openshift_master : Restore Master Controllers AWS Options] *************** TASK [openshift_master : Install Master docker service file] ******************* ok: [10.8.169.22] TASK [openshift_master : Preserve Master Proxy Config options] ***************** ok: [10.8.169.22] TASK [openshift_master : set_fact] ********************************************* ok: [10.8.169.22] TASK [openshift_master : Preserve Master AWS options] ************************** ok: [10.8.169.22] TASK [openshift_master : set_fact] ********************************************* ok: [10.8.169.22] TASK [openshift_master : Create the master service env file] ******************* changed: [10.8.169.22] TASK [openshift_master : Create session secrets file] ************************** changed: [10.8.169.22] TASK [openshift_master : set_fact] ********************************************* ok: [10.8.169.22] TASK [openshift_master : Create master config] ********************************* changed: [10.8.169.22] TASK [openshift_master : Test local loopback context] ************************** ok: [10.8.169.22] TASK [openshift_master : command] ********************************************** skipping: [10.8.169.22] TASK [openshift_master : command] ********************************************** skipping: [10.8.169.22] TASK [openshift_master : command] ********************************************** skipping: [10.8.169.22] TASK [openshift_master : Start and enable master] ****************************** changed: [10.8.169.22] TASK [openshift_master : Stop and disable non-HA master when running HA] ******* skipping: [10.8.169.22] TASK [openshift_master : set_fact] ********************************************* ok: [10.8.169.22] TASK [openshift_master : Mask master service] ********************************** skipping: [10.8.169.22] TASK [openshift_master : Start and enable master api on first master] ********** skipping: [10.8.169.22] TASK [openshift_master : set_fact] ********************************************* skipping: [10.8.169.22] TASK [openshift_master : pause] ************************************************ skipping: [10.8.169.22] TASK [openshift_master : Start and enable master api all masters] ************** skipping: [10.8.169.22] TASK [openshift_master : set_fact] ********************************************* skipping: [10.8.169.22] TASK [openshift_master : Wait for API to become available] ********************* skipping: [10.8.169.22] TASK [openshift_master : Start and enable master controller on first master] *** skipping: [10.8.169.22] TASK [openshift_master : Wait for master controller service to start on first master] *** skipping: [10.8.169.22] TASK [openshift_master : Start and enable master controller on all masters] **** skipping: [10.8.169.22] TASK [openshift_master : set_fact] ********************************************* skipping: [10.8.169.22] TASK [openshift_master : Install cluster packages] ***************************** skipping: [10.8.169.22] TASK [openshift_master : Start and enable cluster service] ********************* skipping: [10.8.169.22] TASK [openshift_master : Set the cluster user password] ************************ skipping: [10.8.169.22] RUNNING HANDLER [nuage_master : restart master] ******************************** skipping: [10.8.169.22] RUNNING HANDLER [nuage_master : restart master api] **************************** skipping: [10.8.169.22] RUNNING HANDLER [nuage_master : restart master controllers] ******************** skipping: [10.8.169.22] RUNNING HANDLER [openshift_master : Verify API Server] ************************* ok: [10.8.169.22] TASK [Create group for deployment type] **************************************** ok: [10.8.169.22] PLAY [Delete temporary directory on localhost] ********************************* TASK [file] ******************************************************************** ok: [localhost] PLAY [Additional master configuration] ***************************************** TASK [setup] ******************************************************************* ok: [10.8.169.22] TASK [openshift_facts : Detecting Operating System] **************************** skipping: [10.8.169.22] TASK [openshift_facts : set_fact] ********************************************** skipping: [10.8.169.22] TASK [openshift_facts : set_fact] ********************************************** skipping: [10.8.169.22] TASK [openshift_facts : Ensure various deps are installed] ********************* skipping: [10.8.169.22] => (item=[]) TASK [openshift_facts : Gather Cluster facts and set is_containerized if needed] *** skipping: [10.8.169.22] TASK [openshift_facts : Set repoquery command] ********************************* skipping: [10.8.169.22] TASK [openshift_repos : assert] ************************************************ skipping: [10.8.169.22] TASK [openshift_repos : Ensure libselinux-python is installed] ***************** skipping: [10.8.169.22] TASK [openshift_repos : Create any additional repos that are defined] ********** skipping: [10.8.169.22] TASK [openshift_repos : Remove the additional repos if no longer defined] ****** skipping: [10.8.169.22] TASK [openshift_repos : Configure origin gpg keys if needed] ******************* skipping: [10.8.169.22] TASK [openshift_repos : Configure origin yum repositories RHEL/CentOS] ********* skipping: [10.8.169.22] TASK [openshift_master_cluster : fail] ***************************************** skipping: [10.8.169.22] TASK [openshift_master_cluster : fail] ***************************************** skipping: [10.8.169.22] TASK [openshift_master_cluster : Test if cluster is already configured] ******** skipping: [10.8.169.22] TASK [openshift_master_cluster : fail] ***************************************** skipping: [10.8.169.22] TASK [openshift_master_cluster : fail] ***************************************** skipping: [10.8.169.22] TASK [openshift_master_cluster : Authenticate to the cluster] ****************** skipping: [10.8.169.22] TASK [openshift_master_cluster : Create the cluster] *************************** skipping: [10.8.169.22] TASK [openshift_master_cluster : Start the cluster] **************************** skipping: [10.8.169.22] TASK [openshift_master_cluster : Enable the cluster on all nodes] ************** skipping: [10.8.169.22] TASK [openshift_master_cluster : Set default resource stickiness] ************** skipping: [10.8.169.22] TASK [openshift_master_cluster : Add the cluster VIP resource] ***************** skipping: [10.8.169.22] TASK [openshift_master_cluster : Add the cluster public VIP resource] ********** skipping: [10.8.169.22] TASK [openshift_master_cluster : Add the cluster master service resource] ****** skipping: [10.8.169.22] TASK [openshift_master_cluster : Disable stonith] ****************************** skipping: [10.8.169.22] TASK [openshift_master_cluster : Wait for the clustered master service to be available] *** skipping: [10.8.169.22] TASK [openshift_facts : Detecting Operating System] **************************** ok: [10.8.169.22] TASK [openshift_facts : set_fact] ********************************************** ok: [10.8.169.22] TASK [openshift_facts : set_fact] ********************************************** ok: [10.8.169.22] TASK [openshift_facts : Ensure various deps are installed] ********************* ok: [10.8.169.22] => (item=[u'iproute', u'python-dbus', u'python-six', u'PyYAML', u'yum-utils']) TASK [openshift_facts : Gather Cluster facts and set is_containerized if needed] *** changed: [10.8.169.22] TASK [openshift_facts : Set repoquery command] ********************************* ok: [10.8.169.22] TASK [openshift_repos : assert] ************************************************ skipping: [10.8.169.22] TASK [openshift_repos : Ensure libselinux-python is installed] ***************** skipping: [10.8.169.22] TASK [openshift_repos : Create any additional repos that are defined] ********** skipping: [10.8.169.22] TASK [openshift_repos : Remove the additional repos if no longer defined] ****** skipping: [10.8.169.22] TASK [openshift_repos : Configure origin gpg keys if needed] ******************* skipping: [10.8.169.22] TASK [openshift_repos : Configure origin yum repositories RHEL/CentOS] ********* skipping: [10.8.169.22] TASK [openshift_repos : assert] ************************************************ skipping: [10.8.169.22] TASK [openshift_repos : Ensure libselinux-python is installed] ***************** skipping: [10.8.169.22] TASK [openshift_repos : Create any additional repos that are defined] ********** skipping: [10.8.169.22] TASK [openshift_repos : Remove the additional repos if no longer defined] ****** skipping: [10.8.169.22] TASK [openshift_repos : Configure origin gpg keys if needed] ******************* skipping: [10.8.169.22] TASK [openshift_repos : Configure origin yum repositories RHEL/CentOS] ********* skipping: [10.8.169.22] TASK [openshift_docker_facts : Set docker facts] ******************************* ok: [10.8.169.22] => (item={u'local_facts': {u'blocked_registries': u'', u'additional_registries': u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888,registry.ops.openshift.com', u'log_driver': u'', u'disable_push_dockerhub': u'', u'hosted_registry_insecure': False, u'hosted_registry_network': u'172.30.0.0/16', u'log_options': u'', u'insecure_registries': u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888,registry.ops.openshift.com', u'options': u''}, u'role': u'docker'}) TASK [openshift_docker_facts : set_fact] *************************************** ok: [10.8.169.22] TASK [openshift_docker_facts : set_fact] *************************************** skipping: [10.8.169.22] TASK [openshift_docker_facts : set_fact] *************************************** ok: [10.8.169.22] TASK [os_firewall : Assert - Do not use firewalld on Atomic Host] ************** skipping: [10.8.169.22] TASK [os_firewall : Install firewalld packages] ******************************** skipping: [10.8.169.22] TASK [os_firewall : Ensure iptables services are not enabled] ****************** ok: [10.8.169.22] => (item=iptables) ok: [10.8.169.22] => (item=ip6tables) TASK [os_firewall : Start and enable firewalld service] ************************ ok: [10.8.169.22] TASK [os_firewall : need to pause here, otherwise the firewalld service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.22] TASK [os_firewall : Add firewalld allow rules] ********************************* TASK [os_firewall : Remove firewalld allow rules] ****************************** TASK [os_firewall : Ensure firewalld service is not enabled] ******************* skipping: [10.8.169.22] TASK [os_firewall : Install iptables packages] ********************************* skipping: [10.8.169.22] => (item=[]) TASK [os_firewall : Start and enable iptables service] ************************* skipping: [10.8.169.22] TASK [os_firewall : need to pause here, otherwise the iptables service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.22] TASK [os_firewall : Add iptables allow rules] ********************************** TASK [os_firewall : Remove iptables rules] ************************************* TASK [docker : Get current installed Docker version] *************************** ok: [10.8.169.22] TASK [docker : Error out if Docker pre-installed but too old] ****************** skipping: [10.8.169.22] TASK [docker : Error out if requested Docker is too old] *********************** skipping: [10.8.169.22] TASK [docker : Get latest available version of Docker] ************************* skipping: [10.8.169.22] TASK [docker : Fail if Docker version requested but downgrade is required] ***** skipping: [10.8.169.22] TASK [docker : Error out if attempting to upgrade Docker across the 1.10 boundary] *** skipping: [10.8.169.22] TASK [docker : Install Docker] ************************************************* ok: [10.8.169.22] TASK [docker : Ensure docker.service.d directory exists] *********************** skipping: [10.8.169.22] TASK [docker : Configure Docker service unit file] ***************************** skipping: [10.8.169.22] TASK [docker : Getting current systemd-udevd exec command] ********************* skipping: [10.8.169.22] TASK [docker : Assure systemd-udevd.service.d directory exists] **************** skipping: [10.8.169.22] TASK [docker : Create systemd-udevd override file] ***************************** skipping: [10.8.169.22] TASK [docker : stat] *********************************************************** ok: [10.8.169.22] TASK [docker : Set registry params] ******************************************** ok: [10.8.169.22] => (item={u'reg_conf_var': u'ADD_REGISTRY', u'reg_flag': u'--add-registry', u'reg_fact_val': [u'registry.ops.openshift.com', u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888', u'XXXXXXXXXXXXXXXXXXXXXXXXXXX']}) skipping: [10.8.169.22] => (item={u'reg_conf_var': u'BLOCK_REGISTRY', u'reg_flag': u'--block-registry', u'reg_fact_val': u''}) ok: [10.8.169.22] => (item={u'reg_conf_var': u'INSECURE_REGISTRY', u'reg_flag': u'--insecure-registry', u'reg_fact_val': [u'registry.ops.openshift.com', u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888']}) TASK [docker : Set Proxy Settings] ********************************************* ok: [10.8.169.22] => (item={u'reg_conf_var': u'HTTP_PROXY', u'reg_fact_val': u''}) ok: [10.8.169.22] => (item={u'reg_conf_var': u'HTTPS_PROXY', u'reg_fact_val': u''}) ok: [10.8.169.22] => (item={u'reg_conf_var': u'NO_PROXY', u'reg_fact_val': u''}) TASK [docker : Set various Docker options] ************************************* ok: [10.8.169.22] TASK [docker : Start the Docker service] *************************************** ok: [10.8.169.22] TASK [docker : set_fact] ******************************************************* ok: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** ok: [10.8.169.22] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** ok: [10.8.169.22] TASK [openshift_version : Use openshift.common.version fact as version to configure if already installed] *** skipping: [10.8.169.22] TASK [openshift_version : Set rpm version to configure if openshift_pkg_version specified] *** skipping: [10.8.169.22] TASK [openshift_version : Gather common package version] *********************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : Set containerized version to configure if openshift_image_tag specified] *** skipping: [10.8.169.22] TASK [openshift_version : Set containerized version to configure if openshift_release specified] *** skipping: [10.8.169.22] TASK [openshift_version : Lookup latest containerized version if no version specified] *** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : Set precise containerized version to configure if openshift_release specified] *** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** ok: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.22] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.22] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.22] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.22] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.22] TASK [openshift_version : debug] *********************************************** ok: [10.8.169.22] => { "openshift_release": "VARIABLE IS NOT DEFINED!" } TASK [openshift_version : debug] *********************************************** ok: [10.8.169.22] => { "openshift_image_tag": "v3.3" } TASK [openshift_version : debug] *********************************************** ok: [10.8.169.22] => { "openshift_pkg_version": "-3.3.1.9" } TASK [openshift_common : fail] ************************************************* skipping: [10.8.169.22] TASK [openshift_common : fail] ************************************************* skipping: [10.8.169.22] TASK [openshift_common : fail] ************************************************* skipping: [10.8.169.22] TASK [openshift_common : fail] ************************************************* skipping: [10.8.169.22] TASK [openshift_common : Set common Cluster facts] ***************************** ok: [10.8.169.22] TASK [openshift_common : Install the base package for versioning] ************** skipping: [10.8.169.22] TASK [openshift_common : Set version facts] ************************************ ok: [10.8.169.22] TASK [openshift_common : set_fact] ********************************************* ok: [10.8.169.22] TASK [openshift_common : Set hostname] ***************************************** changed: [10.8.169.22] TASK [openshift_examples : Create local temp dir for OpenShift examples copy] ** changed: [10.8.169.22 -> localhost] TASK [openshift_examples : Create tar of OpenShift examples] ******************* changed: [10.8.169.22 -> localhost] TASK [openshift_examples : Create the remote OpenShift examples directory] ***** changed: [10.8.169.22] TASK [openshift_examples : Unarchive the OpenShift examples on the remote] ***** changed: [10.8.169.22] TASK [openshift_examples : Cleanup the OpenShift Examples temp dir] ************ changed: [10.8.169.22 -> localhost] TASK [openshift_examples : Modify registry paths if registry_url is not XXXXXXXXXXXXXXXXXXXXXXXXXXX] *** changed: [10.8.169.22] TASK [openshift_examples : Import RHEL streams] ******************************** ok: [10.8.169.22] => (item=/etc/origin/examples/image-streams/image-streams-rhel7.json) ok: [10.8.169.22] => (item=/etc/origin/examples/image-streams/dotnet_imagestreams.json) TASK [openshift_examples : Import Centos Image streams] ************************ skipping: [10.8.169.22] TASK [openshift_examples : Import db templates] ******************************** ok: [10.8.169.22] TASK [openshift_examples : Remove defunct quickstart template files] *********** changed: [10.8.169.22] => (item=/etc/origin/examples/quickstart-templates/nodejs.json) changed: [10.8.169.22] => (item=/etc/origin/examples/quickstart-templates/cakephp.json) changed: [10.8.169.22] => (item=/etc/origin/examples/quickstart-templates/dancer.json) changed: [10.8.169.22] => (item=/etc/origin/examples/quickstart-templates/django.json) TASK [openshift_examples : Remove defunct quickstart templates from openshift namespace] *** ok: [10.8.169.22] => (item=nodejs-example) ok: [10.8.169.22] => (item=cakephp-example) ok: [10.8.169.22] => (item=dancer-example) ok: [10.8.169.22] => (item=django-example) TASK [openshift_examples : Import quickstart-templates] ************************ ok: [10.8.169.22] TASK [openshift_examples : Remove old xPaas template files] ******************** ok: [10.8.169.22] => (item=/etc/origin/examples/xpaas-templates/sso70-basic.json) TASK [openshift_examples : Remove old xPaas templates from openshift namespace] ok: [10.8.169.22] => (item=sso70-basic) TASK [openshift_examples : Import xPaas image streams] ************************* ok: [10.8.169.22] TASK [openshift_examples : Import xPaas templates] ***************************** ok: [10.8.169.22] TASK [openshift_repos : assert] ************************************************ skipping: [10.8.169.22] TASK [openshift_repos : Ensure libselinux-python is installed] ***************** skipping: [10.8.169.22] TASK [openshift_repos : Create any additional repos that are defined] ********** skipping: [10.8.169.22] TASK [openshift_repos : Remove the additional repos if no longer defined] ****** skipping: [10.8.169.22] TASK [openshift_repos : Configure origin gpg keys if needed] ******************* skipping: [10.8.169.22] TASK [openshift_repos : Configure origin yum repositories RHEL/CentOS] ********* skipping: [10.8.169.22] TASK [openshift_repos : assert] ************************************************ skipping: [10.8.169.22] TASK [openshift_repos : Ensure libselinux-python is installed] ***************** skipping: [10.8.169.22] TASK [openshift_repos : Create any additional repos that are defined] ********** skipping: [10.8.169.22] TASK [openshift_repos : Remove the additional repos if no longer defined] ****** skipping: [10.8.169.22] TASK [openshift_repos : Configure origin gpg keys if needed] ******************* skipping: [10.8.169.22] TASK [openshift_repos : Configure origin yum repositories RHEL/CentOS] ********* skipping: [10.8.169.22] TASK [os_firewall : Assert - Do not use firewalld on Atomic Host] ************** skipping: [10.8.169.22] TASK [os_firewall : Install firewalld packages] ******************************** skipping: [10.8.169.22] TASK [os_firewall : Ensure iptables services are not enabled] ****************** ok: [10.8.169.22] => (item=iptables) ok: [10.8.169.22] => (item=ip6tables) TASK [os_firewall : Start and enable firewalld service] ************************ ok: [10.8.169.22] TASK [os_firewall : need to pause here, otherwise the firewalld service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.22] TASK [os_firewall : Add firewalld allow rules] ********************************* TASK [os_firewall : Remove firewalld allow rules] ****************************** TASK [os_firewall : Ensure firewalld service is not enabled] ******************* skipping: [10.8.169.22] TASK [os_firewall : Install iptables packages] ********************************* skipping: [10.8.169.22] => (item=[]) TASK [os_firewall : Start and enable iptables service] ************************* skipping: [10.8.169.22] TASK [os_firewall : need to pause here, otherwise the iptables service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.22] TASK [os_firewall : Add iptables allow rules] ********************************** TASK [os_firewall : Remove iptables rules] ************************************* TASK [openshift_hosted_templates : Create local temp dir for OpenShift hosted templates copy] *** changed: [10.8.169.22 -> localhost] TASK [openshift_hosted_templates : Create tar of OpenShift examples] *********** changed: [10.8.169.22 -> localhost] TASK [openshift_hosted_templates : Create remote OpenShift hosted templates directory] *** changed: [10.8.169.22] TASK [openshift_hosted_templates : Unarchive the OpenShift hosted templates on the remote] *** changed: [10.8.169.22] TASK [openshift_hosted_templates : Cleanup the OpenShift hosted templates temp dir] *** changed: [10.8.169.22 -> localhost] TASK [openshift_hosted_templates : Modify registry paths if registry_url is not XXXXXXXXXXXXXXXXXXXXXXXXXXX] *** changed: [10.8.169.22] TASK [openshift_hosted_templates : Create temp directory for kubeconfig] ******* ok: [10.8.169.22] TASK [openshift_hosted_templates : Record kubeconfig tmp dir] ****************** ok: [10.8.169.22] TASK [openshift_hosted_templates : Copy the admin client config(s)] ************ ok: [10.8.169.22] TASK [openshift_hosted_templates : Create or update hosted templates] ********** changed: [10.8.169.22] TASK [openshift_hosted_templates : Delete temp directory] ********************** ok: [10.8.169.22] TASK [openshift_manageiq : fail] *********************************************** skipping: [10.8.169.22] TASK [openshift_manageiq : Copy Configuration to temporary conf] *************** ok: [10.8.169.22] TASK [openshift_manageiq : Add Management Infrastructure project] ************** changed: [10.8.169.22] TASK [openshift_manageiq : Create Admin Service Account] *********************** changed: [10.8.169.22] TASK [openshift_manageiq : Create Image Inspector Service Account] ************* changed: [10.8.169.22] TASK [openshift_manageiq : Create Cluster Role] ******************************** changed: [10.8.169.22] TASK [openshift_manageiq : Create Hawkular Metrics Admin Cluster Role] ********* changed: [10.8.169.22] TASK [openshift_manageiq : Configure role/user permissions] ******************** changed: [10.8.169.22] => (item=policy add-role-to-user -n management-infra admin -z management-admin) changed: [10.8.169.22] => (item=policy add-role-to-user -n management-infra management-infra-admin -z management-admin) changed: [10.8.169.22] => (item=policy add-cluster-role-to-user cluster-reader system:serviceaccount:management-infra:management-admin) changed: [10.8.169.22] => (item=policy add-scc-to-user privileged system:serviceaccount:management-infra:management-admin) changed: [10.8.169.22] => (item=policy add-cluster-role-to-user system:image-puller system:serviceaccount:management-infra:inspector-admin) changed: [10.8.169.22] => (item=policy add-scc-to-user privileged system:serviceaccount:management-infra:inspector-admin) changed: [10.8.169.22] => (item=policy add-cluster-role-to-user self-provisioner system:serviceaccount:management-infra:management-admin) changed: [10.8.169.22] => (item=policy add-cluster-role-to-user hawkular-metrics-admin system:serviceaccount:management-infra:management-admin) TASK [openshift_manageiq : Configure 3_2 role/user permissions] **************** changed: [10.8.169.22] => (item=policy add-cluster-role-to-user system:image-auditor system:serviceaccount:management-infra:management-admin) TASK [openshift_manageiq : Clean temporary configuration file] ***************** changed: [10.8.169.22] TASK [os_firewall : Assert - Do not use firewalld on Atomic Host] ************** skipping: [10.8.169.22] TASK [os_firewall : Install firewalld packages] ******************************** skipping: [10.8.169.22] TASK [os_firewall : Ensure iptables services are not enabled] ****************** ok: [10.8.169.22] => (item=iptables) ok: [10.8.169.22] => (item=ip6tables) TASK [os_firewall : Start and enable firewalld service] ************************ ok: [10.8.169.22] TASK [os_firewall : need to pause here, otherwise the firewalld service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.22] TASK [os_firewall : Add firewalld allow rules] ********************************* changed: [10.8.169.22] => (item={u'port': u'9090/tcp', u'service': u'cockpit-ws'}) TASK [os_firewall : Remove firewalld allow rules] ****************************** TASK [os_firewall : Ensure firewalld service is not enabled] ******************* skipping: [10.8.169.22] TASK [os_firewall : Install iptables packages] ********************************* skipping: [10.8.169.22] => (item=[]) TASK [os_firewall : Start and enable iptables service] ************************* skipping: [10.8.169.22] TASK [os_firewall : need to pause here, otherwise the iptables service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.22] TASK [os_firewall : Add iptables allow rules] ********************************** skipping: [10.8.169.22] => (item={u'port': u'9090/tcp', u'service': u'cockpit-ws'}) TASK [os_firewall : Remove iptables rules] ************************************* TASK [cockpit : Install cockpit-ws] ******************************************** skipping: [10.8.169.22] => (item=[]) TASK [cockpit : Enable cockpit-ws] ********************************************* skipping: [10.8.169.22] TASK [flannel_register : Assures /etc/flannel dir exists] ********************** skipping: [10.8.169.22] TASK [flannel_register : Generate etcd configuration for etcd] ***************** skipping: [10.8.169.22] TASK [flannel_register : Insert flannel configuration into etcd] *************** skipping: [10.8.169.22] PLAY [Gather and set facts for node hosts] ************************************* TASK [setup] ******************************************************************* ok: [10.8.169.22] ok: [10.8.169.215] ok: [10.8.169.210] TASK [set_fact] **************************************************************** skipping: [10.8.169.22] skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_facts : Detecting Operating System] **************************** ok: [10.8.169.215] ok: [10.8.169.210] ok: [10.8.169.22] TASK [openshift_facts : set_fact] ********************************************** ok: [10.8.169.215] ok: [10.8.169.210] ok: [10.8.169.22] TASK [openshift_facts : set_fact] ********************************************** ok: [10.8.169.210] ok: [10.8.169.215] ok: [10.8.169.22] TASK [openshift_facts : Ensure various deps are installed] ********************* ok: [10.8.169.22] => (item=[u'iproute', u'python-dbus', u'python-six', u'PyYAML', u'yum-utils']) ok: [10.8.169.215] => (item=[u'iproute', u'python-dbus', u'python-six', u'PyYAML', u'yum-utils']) ok: [10.8.169.210] => (item=[u'iproute', u'python-dbus', u'python-six', u'PyYAML', u'yum-utils']) TASK [openshift_facts : Gather Cluster facts and set is_containerized if needed] *** ok: [10.8.169.215] ok: [10.8.169.22] ok: [10.8.169.210] TASK [openshift_facts : Set repoquery command] ********************************* ok: [10.8.169.210] ok: [10.8.169.215] ok: [10.8.169.22] TASK [openshift_facts] ********************************************************* changed: [10.8.169.210] changed: [10.8.169.215] changed: [10.8.169.22] PLAY [Create temp directory for syncing certs] ********************************* TASK [Create local temp directory for syncing certs] *************************** ok: [localhost -> localhost] PLAY [Evaluate node groups] **************************************************** TASK [setup] ******************************************************************* ok: [localhost] TASK [Evaluate oo_containerized_master_nodes] ********************************** skipping: [localhost] => (item=10.8.169.210) changed: [localhost] => (item=10.8.169.22) skipping: [localhost] => (item=10.8.169.215) PLAY [Configure containerized nodes] ******************************************* TASK [setup] ******************************************************************* ok: [10.8.169.22] TASK [openshift_facts : Detecting Operating System] **************************** ok: [10.8.169.22] TASK [openshift_facts : set_fact] ********************************************** ok: [10.8.169.22] TASK [openshift_facts : set_fact] ********************************************** ok: [10.8.169.22] TASK [openshift_facts : Ensure various deps are installed] ********************* ok: [10.8.169.22] => (item=[u'iproute', u'python-dbus', u'python-six', u'PyYAML', u'yum-utils']) TASK [openshift_facts : Gather Cluster facts and set is_containerized if needed] *** changed: [10.8.169.22] TASK [openshift_facts : Set repoquery command] ********************************* ok: [10.8.169.22] TASK [openshift_repos : assert] ************************************************ skipping: [10.8.169.22] TASK [openshift_repos : Ensure libselinux-python is installed] ***************** skipping: [10.8.169.22] TASK [openshift_repos : Create any additional repos that are defined] ********** skipping: [10.8.169.22] TASK [openshift_repos : Remove the additional repos if no longer defined] ****** skipping: [10.8.169.22] TASK [openshift_repos : Configure origin gpg keys if needed] ******************* skipping: [10.8.169.22] TASK [openshift_repos : Configure origin yum repositories RHEL/CentOS] ********* skipping: [10.8.169.22] TASK [openshift_repos : assert] ************************************************ skipping: [10.8.169.22] TASK [openshift_repos : Ensure libselinux-python is installed] ***************** skipping: [10.8.169.22] TASK [openshift_repos : Create any additional repos that are defined] ********** skipping: [10.8.169.22] TASK [openshift_repos : Remove the additional repos if no longer defined] ****** skipping: [10.8.169.22] TASK [openshift_repos : Configure origin gpg keys if needed] ******************* skipping: [10.8.169.22] TASK [openshift_repos : Configure origin yum repositories RHEL/CentOS] ********* skipping: [10.8.169.22] TASK [openshift_docker_facts : Set docker facts] ******************************* ok: [10.8.169.22] => (item={u'local_facts': {u'blocked_registries': u'', u'additional_registries': u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888,registry.ops.openshift.com', u'log_driver': u'', u'disable_push_dockerhub': u'', u'hosted_registry_insecure': False, u'hosted_registry_network': u'172.30.0.0/16', u'log_options': u'', u'insecure_registries': u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888,registry.ops.openshift.com', u'options': u''}, u'role': u'docker'}) TASK [openshift_docker_facts : set_fact] *************************************** ok: [10.8.169.22] TASK [openshift_docker_facts : set_fact] *************************************** skipping: [10.8.169.22] TASK [openshift_docker_facts : set_fact] *************************************** ok: [10.8.169.22] TASK [os_firewall : Assert - Do not use firewalld on Atomic Host] ************** skipping: [10.8.169.22] TASK [os_firewall : Install firewalld packages] ******************************** skipping: [10.8.169.22] TASK [os_firewall : Ensure iptables services are not enabled] ****************** ok: [10.8.169.22] => (item=iptables) ok: [10.8.169.22] => (item=ip6tables) TASK [os_firewall : Start and enable firewalld service] ************************ ok: [10.8.169.22] TASK [os_firewall : need to pause here, otherwise the firewalld service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.22] TASK [os_firewall : Add firewalld allow rules] ********************************* TASK [os_firewall : Remove firewalld allow rules] ****************************** TASK [os_firewall : Ensure firewalld service is not enabled] ******************* skipping: [10.8.169.22] TASK [os_firewall : Install iptables packages] ********************************* skipping: [10.8.169.22] => (item=[]) TASK [os_firewall : Start and enable iptables service] ************************* skipping: [10.8.169.22] TASK [os_firewall : need to pause here, otherwise the iptables service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.22] TASK [os_firewall : Add iptables allow rules] ********************************** TASK [os_firewall : Remove iptables rules] ************************************* TASK [docker : Get current installed Docker version] *************************** ok: [10.8.169.22] TASK [docker : Error out if Docker pre-installed but too old] ****************** skipping: [10.8.169.22] TASK [docker : Error out if requested Docker is too old] *********************** skipping: [10.8.169.22] TASK [docker : Get latest available version of Docker] ************************* skipping: [10.8.169.22] TASK [docker : Fail if Docker version requested but downgrade is required] ***** skipping: [10.8.169.22] TASK [docker : Error out if attempting to upgrade Docker across the 1.10 boundary] *** skipping: [10.8.169.22] TASK [docker : Install Docker] ************************************************* ok: [10.8.169.22] TASK [docker : Ensure docker.service.d directory exists] *********************** skipping: [10.8.169.22] TASK [docker : Configure Docker service unit file] ***************************** skipping: [10.8.169.22] TASK [docker : Getting current systemd-udevd exec command] ********************* skipping: [10.8.169.22] TASK [docker : Assure systemd-udevd.service.d directory exists] **************** skipping: [10.8.169.22] TASK [docker : Create systemd-udevd override file] ***************************** skipping: [10.8.169.22] TASK [docker : stat] *********************************************************** ok: [10.8.169.22] TASK [docker : Set registry params] ******************************************** skipping: [10.8.169.22] => (item={u'reg_conf_var': u'BLOCK_REGISTRY', u'reg_flag': u'--block-registry', u'reg_fact_val': u''}) ok: [10.8.169.22] => (item={u'reg_conf_var': u'ADD_REGISTRY', u'reg_flag': u'--add-registry', u'reg_fact_val': [u'registry.ops.openshift.com', u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888', u'XXXXXXXXXXXXXXXXXXXXXXXXXXX']}) ok: [10.8.169.22] => (item={u'reg_conf_var': u'INSECURE_REGISTRY', u'reg_flag': u'--insecure-registry', u'reg_fact_val': [u'registry.ops.openshift.com', u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888']}) TASK [docker : Set Proxy Settings] ********************************************* ok: [10.8.169.22] => (item={u'reg_conf_var': u'HTTP_PROXY', u'reg_fact_val': u''}) ok: [10.8.169.22] => (item={u'reg_conf_var': u'HTTPS_PROXY', u'reg_fact_val': u''}) ok: [10.8.169.22] => (item={u'reg_conf_var': u'NO_PROXY', u'reg_fact_val': u''}) TASK [docker : Set various Docker options] ************************************* ok: [10.8.169.22] TASK [docker : Start the Docker service] *************************************** ok: [10.8.169.22] TASK [docker : set_fact] ******************************************************* ok: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** ok: [10.8.169.22] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** ok: [10.8.169.22] TASK [openshift_version : Use openshift.common.version fact as version to configure if already installed] *** skipping: [10.8.169.22] TASK [openshift_version : Set rpm version to configure if openshift_pkg_version specified] *** skipping: [10.8.169.22] TASK [openshift_version : Gather common package version] *********************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : Set containerized version to configure if openshift_image_tag specified] *** skipping: [10.8.169.22] TASK [openshift_version : Set containerized version to configure if openshift_release specified] *** skipping: [10.8.169.22] TASK [openshift_version : Lookup latest containerized version if no version specified] *** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : Set precise containerized version to configure if openshift_release specified] *** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** ok: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.22] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.22] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.22] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.22] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.22] TASK [openshift_version : debug] *********************************************** ok: [10.8.169.22] => { "openshift_release": "VARIABLE IS NOT DEFINED!" } TASK [openshift_version : debug] *********************************************** ok: [10.8.169.22] => { "openshift_image_tag": "v3.3" } TASK [openshift_version : debug] *********************************************** ok: [10.8.169.22] => { "openshift_pkg_version": "-3.3.1.9" } TASK [openshift_common : fail] ************************************************* skipping: [10.8.169.22] TASK [openshift_common : fail] ************************************************* skipping: [10.8.169.22] TASK [openshift_common : fail] ************************************************* skipping: [10.8.169.22] TASK [openshift_common : fail] ************************************************* skipping: [10.8.169.22] TASK [openshift_common : Set common Cluster facts] ***************************** ok: [10.8.169.22] TASK [openshift_common : Install the base package for versioning] ************** skipping: [10.8.169.22] TASK [openshift_common : Set version facts] ************************************ ok: [10.8.169.22] TASK [openshift_common : set_fact] ********************************************* ok: [10.8.169.22] TASK [openshift_common : Set hostname] ***************************************** changed: [10.8.169.22] TASK [openshift_clock : Set clock facts] *************************************** ok: [10.8.169.22] TASK [openshift_clock : Install ntp package] *********************************** skipping: [10.8.169.22] TASK [openshift_clock : Start and enable ntpd/chronyd] ************************* changed: [10.8.169.22] TASK [openshift_repos : assert] ************************************************ skipping: [10.8.169.22] TASK [openshift_repos : Ensure libselinux-python is installed] ***************** skipping: [10.8.169.22] TASK [openshift_repos : Create any additional repos that are defined] ********** skipping: [10.8.169.22] TASK [openshift_repos : Remove the additional repos if no longer defined] ****** skipping: [10.8.169.22] TASK [openshift_repos : Configure origin gpg keys if needed] ******************* skipping: [10.8.169.22] TASK [openshift_repos : Configure origin yum repositories RHEL/CentOS] ********* skipping: [10.8.169.22] TASK [os_firewall : Assert - Do not use firewalld on Atomic Host] ************** skipping: [10.8.169.22] TASK [os_firewall : Install firewalld packages] ******************************** skipping: [10.8.169.22] TASK [os_firewall : Ensure iptables services are not enabled] ****************** ok: [10.8.169.22] => (item=iptables) ok: [10.8.169.22] => (item=ip6tables) TASK [os_firewall : Start and enable firewalld service] ************************ ok: [10.8.169.22] TASK [os_firewall : need to pause here, otherwise the firewalld service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.22] TASK [os_firewall : Add firewalld allow rules] ********************************* TASK [os_firewall : Remove firewalld allow rules] ****************************** TASK [os_firewall : Ensure firewalld service is not enabled] ******************* skipping: [10.8.169.22] TASK [os_firewall : Install iptables packages] ********************************* skipping: [10.8.169.22] => (item=[]) TASK [os_firewall : Start and enable iptables service] ************************* skipping: [10.8.169.22] TASK [os_firewall : need to pause here, otherwise the iptables service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.22] TASK [os_firewall : Add iptables allow rules] ********************************** TASK [os_firewall : Remove iptables rules] ************************************* TASK [os_firewall : Assert - Do not use firewalld on Atomic Host] ************** skipping: [10.8.169.22] TASK [os_firewall : Install firewalld packages] ******************************** skipping: [10.8.169.22] TASK [os_firewall : Ensure iptables services are not enabled] ****************** ok: [10.8.169.22] => (item=iptables) ok: [10.8.169.22] => (item=ip6tables) TASK [os_firewall : Start and enable firewalld service] ************************ ok: [10.8.169.22] TASK [os_firewall : need to pause here, otherwise the firewalld service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.22] TASK [os_firewall : Add firewalld allow rules] ********************************* TASK [os_firewall : Remove firewalld allow rules] ****************************** TASK [os_firewall : Ensure firewalld service is not enabled] ******************* skipping: [10.8.169.22] TASK [os_firewall : Install iptables packages] ********************************* skipping: [10.8.169.22] => (item=[]) TASK [os_firewall : Start and enable iptables service] ************************* skipping: [10.8.169.22] TASK [os_firewall : need to pause here, otherwise the iptables service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.22] TASK [os_firewall : Add iptables allow rules] ********************************** TASK [os_firewall : Remove iptables rules] ************************************* TASK [docker : Get current installed Docker version] *************************** ok: [10.8.169.22] TASK [docker : Error out if Docker pre-installed but too old] ****************** skipping: [10.8.169.22] TASK [docker : Error out if requested Docker is too old] *********************** skipping: [10.8.169.22] TASK [docker : Get latest available version of Docker] ************************* skipping: [10.8.169.22] TASK [docker : Fail if Docker version requested but downgrade is required] ***** skipping: [10.8.169.22] TASK [docker : Error out if attempting to upgrade Docker across the 1.10 boundary] *** skipping: [10.8.169.22] TASK [docker : Install Docker] ************************************************* ok: [10.8.169.22] TASK [docker : Ensure docker.service.d directory exists] *********************** skipping: [10.8.169.22] TASK [docker : Configure Docker service unit file] ***************************** skipping: [10.8.169.22] TASK [docker : Getting current systemd-udevd exec command] ********************* skipping: [10.8.169.22] TASK [docker : Assure systemd-udevd.service.d directory exists] **************** skipping: [10.8.169.22] TASK [docker : Create systemd-udevd override file] ***************************** skipping: [10.8.169.22] TASK [docker : stat] *********************************************************** ok: [10.8.169.22] TASK [docker : Set registry params] ******************************************** skipping: [10.8.169.22] => (item={u'reg_conf_var': u'BLOCK_REGISTRY', u'reg_flag': u'--block-registry', u'reg_fact_val': u''}) ok: [10.8.169.22] => (item={u'reg_conf_var': u'ADD_REGISTRY', u'reg_flag': u'--add-registry', u'reg_fact_val': [u'registry.ops.openshift.com', u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888', u'XXXXXXXXXXXXXXXXXXXXXXXXXXX']}) ok: [10.8.169.22] => (item={u'reg_conf_var': u'INSECURE_REGISTRY', u'reg_flag': u'--insecure-registry', u'reg_fact_val': [u'registry.ops.openshift.com', u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888']}) TASK [docker : Set Proxy Settings] ********************************************* ok: [10.8.169.22] => (item={u'reg_conf_var': u'HTTP_PROXY', u'reg_fact_val': u''}) ok: [10.8.169.22] => (item={u'reg_conf_var': u'HTTPS_PROXY', u'reg_fact_val': u''}) ok: [10.8.169.22] => (item={u'reg_conf_var': u'NO_PROXY', u'reg_fact_val': u''}) TASK [docker : Set various Docker options] ************************************* ok: [10.8.169.22] TASK [docker : Start the Docker service] *************************************** ok: [10.8.169.22] TASK [docker : set_fact] ******************************************************* ok: [10.8.169.22] TASK [openshift_node_certificates : Ensure CA certificate exists on openshift_ca_host] *** ok: [10.8.169.22 -> 10.8.169.22] TASK [openshift_node_certificates : fail] ************************************** skipping: [10.8.169.22] TASK [openshift_node_certificates : Check status of node certificates] ********* ok: [10.8.169.22] => (item=system:node:172.16.132.155.crt) ok: [10.8.169.22] => (item=system:node:172.16.132.155.key) ok: [10.8.169.22] => (item=system:node:172.16.132.155.kubeconfig) ok: [10.8.169.22] => (item=ca.crt) ok: [10.8.169.22] => (item=server.key) ok: [10.8.169.22] => (item=server.crt) TASK [openshift_node_certificates : set_fact] ********************************** ok: [10.8.169.22] TASK [openshift_node_certificates : Create openshift_generated_configs_dir if it does not exist] *** changed: [10.8.169.22 -> 10.8.169.22] TASK [openshift_node_certificates : Generate the node client config] *********** changed: [10.8.169.22 -> 10.8.169.22] TASK [openshift_node_certificates : Generate the node server certificate] ****** changed: [10.8.169.22 -> 10.8.169.22] TASK [openshift_node_certificates : Create local temp directory for syncing certs] *** ok: [10.8.169.22 -> localhost] TASK [openshift_node_certificates : Create a tarball of the node config directories] *** changed: [10.8.169.22 -> 10.8.169.22] TASK [openshift_node_certificates : Retrieve the node config tarballs from the master] *** changed: [10.8.169.22 -> 10.8.169.22] TASK [openshift_node_certificates : Ensure certificate directory exists] ******* changed: [10.8.169.22] TASK [openshift_node_certificates : Unarchive the tarball on the node] ********* changed: [10.8.169.22] TASK [openshift_node_certificates : file] ************************************** ok: [10.8.169.22 -> localhost] TASK [openshift_node_certificates : Copy OpenShift CA to system CA trust] ****** changed: [10.8.169.22] => (item={u'cert': u'/etc/origin/node/ca.crt', u'id': u'openshift'}) TASK [openshift_cloud_provider : Set cloud provider facts] ********************* ok: [10.8.169.22] TASK [openshift_cloud_provider : Create cloudprovider config dir] ************** skipping: [10.8.169.22] TASK [openshift_cloud_provider : fail] ***************************************** skipping: [10.8.169.22] TASK [openshift_cloud_provider : Create cloud config] ************************** skipping: [10.8.169.22] TASK [openshift_cloud_provider : Create cloud config file] ********************* skipping: [10.8.169.22] TASK [openshift_cloud_provider : Configure AWS cloud provider] ***************** skipping: [10.8.169.22] TASK [openshift_cloud_provider : Create cloud config file] ********************* skipping: [10.8.169.22] TASK [openshift_cloud_provider : Configure GCE cloud provider] ***************** skipping: [10.8.169.22] TASK [openshift_repos : assert] ************************************************ skipping: [10.8.169.22] TASK [openshift_repos : Ensure libselinux-python is installed] ***************** skipping: [10.8.169.22] TASK [openshift_repos : Create any additional repos that are defined] ********** skipping: [10.8.169.22] TASK [openshift_repos : Remove the additional repos if no longer defined] ****** skipping: [10.8.169.22] TASK [openshift_repos : Configure origin gpg keys if needed] ******************* skipping: [10.8.169.22] TASK [openshift_repos : Configure origin yum repositories RHEL/CentOS] ********* skipping: [10.8.169.22] TASK [openshift_repos : assert] ************************************************ skipping: [10.8.169.22] TASK [openshift_repos : Ensure libselinux-python is installed] ***************** skipping: [10.8.169.22] TASK [openshift_repos : Create any additional repos that are defined] ********** skipping: [10.8.169.22] TASK [openshift_repos : Remove the additional repos if no longer defined] ****** skipping: [10.8.169.22] TASK [openshift_repos : Configure origin gpg keys if needed] ******************* skipping: [10.8.169.22] TASK [openshift_repos : Configure origin yum repositories RHEL/CentOS] ********* skipping: [10.8.169.22] TASK [os_firewall : Assert - Do not use firewalld on Atomic Host] ************** skipping: [10.8.169.22] TASK [os_firewall : Install firewalld packages] ******************************** skipping: [10.8.169.22] TASK [os_firewall : Ensure iptables services are not enabled] ****************** ok: [10.8.169.22] => (item=iptables) ok: [10.8.169.22] => (item=ip6tables) TASK [os_firewall : Start and enable firewalld service] ************************ ok: [10.8.169.22] TASK [os_firewall : need to pause here, otherwise the firewalld service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.22] TASK [os_firewall : Add firewalld allow rules] ********************************* TASK [os_firewall : Remove firewalld allow rules] ****************************** TASK [os_firewall : Ensure firewalld service is not enabled] ******************* skipping: [10.8.169.22] TASK [os_firewall : Install iptables packages] ********************************* skipping: [10.8.169.22] => (item=[]) TASK [os_firewall : Start and enable iptables service] ************************* skipping: [10.8.169.22] TASK [os_firewall : need to pause here, otherwise the iptables service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.22] TASK [os_firewall : Add iptables allow rules] ********************************** TASK [os_firewall : Remove iptables rules] ************************************* TASK [openshift_node_dnsmasq : Check for NetworkManager service] *************** ok: [10.8.169.22] TASK [openshift_node_dnsmasq : Set fact using_network_manager] ***************** ok: [10.8.169.22] TASK [openshift_node_dnsmasq : Install dnsmasq] ******************************** ok: [10.8.169.22] TASK [openshift_node_dnsmasq : Install dnsmasq configuration] ****************** changed: [10.8.169.22] TASK [openshift_node_dnsmasq : Deploy additional dnsmasq.conf] ***************** skipping: [10.8.169.22] TASK [openshift_node_dnsmasq : Enable dnsmasq] ********************************* changed: [10.8.169.22] TASK [openshift_node_dnsmasq : Install network manager dispatch script] ******** changed: [10.8.169.22] RUNNING HANDLER [openshift_node_certificates : update ca trust] **************** changed: [10.8.169.22] RUNNING HANDLER [openshift_node_certificates : restart docker after updating ca trust] *** changed: [10.8.169.22] RUNNING HANDLER [openshift_node_dnsmasq : restart NetworkManager] ************** changed: [10.8.169.22] RUNNING HANDLER [openshift_node_dnsmasq : restart dnsmasq] ********************* changed: [10.8.169.22] TASK [openshift_node_dnsmasq : fail] ******************************************* skipping: [10.8.169.22] TASK [os_firewall : Assert - Do not use firewalld on Atomic Host] ************** skipping: [10.8.169.22] TASK [os_firewall : Install firewalld packages] ******************************** skipping: [10.8.169.22] TASK [os_firewall : Ensure iptables services are not enabled] ****************** ok: [10.8.169.22] => (item=iptables) ok: [10.8.169.22] => (item=ip6tables) TASK [os_firewall : Start and enable firewalld service] ************************ ok: [10.8.169.22] TASK [os_firewall : need to pause here, otherwise the firewalld service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.22] TASK [os_firewall : Add firewalld allow rules] ********************************* changed: [10.8.169.22] => (item={u'port': u'10250/tcp', u'service': u'Kubernetes kubelet'}) changed: [10.8.169.22] => (item={u'port': u'80/tcp', u'service': u'http'}) changed: [10.8.169.22] => (item={u'port': u'443/tcp', u'service': u'https'}) changed: [10.8.169.22] => (item={u'port': u'10255/tcp', u'service': u'Openshift kubelet ReadOnlyPort'}) changed: [10.8.169.22] => (item={u'port': u'10255/udp', u'service': u'Openshift kubelet ReadOnlyPort udp'}) changed: [10.8.169.22] => (item={u'when': u'openshift.node.use_openshift_sdn | bool', u'port': u'4789/udp', u'service': u'OpenShift OVS sdn'}) TASK [os_firewall : Remove firewalld allow rules] ****************************** TASK [os_firewall : Ensure firewalld service is not enabled] ******************* skipping: [10.8.169.22] TASK [os_firewall : Install iptables packages] ********************************* skipping: [10.8.169.22] => (item=[]) TASK [os_firewall : Start and enable iptables service] ************************* skipping: [10.8.169.22] TASK [os_firewall : need to pause here, otherwise the iptables service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.22] TASK [os_firewall : Add iptables allow rules] ********************************** skipping: [10.8.169.22] => (item={u'port': u'80/tcp', u'service': u'http'}) skipping: [10.8.169.22] => (item={u'port': u'10250/tcp', u'service': u'Kubernetes kubelet'}) skipping: [10.8.169.22] => (item={u'when': u'openshift.node.use_openshift_sdn | bool', u'port': u'4789/udp', u'service': u'OpenShift OVS sdn'}) skipping: [10.8.169.22] => (item={u'port': u'10255/udp', u'service': u'Openshift kubelet ReadOnlyPort udp'}) skipping: [10.8.169.22] => (item={u'port': u'10255/tcp', u'service': u'Openshift kubelet ReadOnlyPort'}) skipping: [10.8.169.22] => (item={u'port': u'443/tcp', u'service': u'https'}) TASK [os_firewall : Remove iptables rules] ************************************* TASK [openshift_node : fail] *************************************************** skipping: [10.8.169.22] TASK [openshift_node : Set node facts] ***************************************** ok: [10.8.169.22] => (item={u'local_facts': {u'labels': {}}, u'role': u'node'}) changed: [10.8.169.22] => (item={u'local_facts': {u'node_image': u'', u'schedulable': u'', u'sdn_mtu': u'', u'env_vars': u'', u'labels': {'region': u'infra', 'zone': u'default'}, u'dns_ip': u'172.16.132.155', u'set_node_ip': u'', u'debug_level': u'2', u'storage_plugin_deps': u'', u'proxy_mode': u'iptables', u'iptables_sync_period': u'', u'local_quota_per_fsgroup': u'', u'ovs_image': u'', u'kubelet_args': u'', u'annotations': u'', u'registry_url': u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888/openshift3/ose-${component}:${version}'}, u'role': u'node'}) TASK [openshift_node : Install Node package] *********************************** skipping: [10.8.169.22] TASK [openshift_node : Check for tuned package] ******************************** ok: [10.8.169.22] TASK [openshift_node : Set atomic-guest tuned profile] ************************* skipping: [10.8.169.22] TASK [openshift_node : Install sdn-ovs package] ******************************** skipping: [10.8.169.22] TASK [openshift_node : Pre-pull node image] ************************************ changed: [10.8.169.22] TASK [openshift_node : Pre-pull openvswitch image] ***************************** changed: [10.8.169.22] TASK [openshift_node : Install Node dependencies docker service file] ********** changed: [10.8.169.22] TASK [openshift_node : Install Node docker service file] *********************** changed: [10.8.169.22] TASK [openshift_node : Create the openvswitch service env file] **************** changed: [10.8.169.22] TASK [openshift_node : Create OpenvSwitch service.d directory] ***************** changed: [10.8.169.22] TASK [openshift_node : Install OpenvSwitch service OOM fix] ******************** changed: [10.8.169.22] TASK [openshift_node : Install OpenvSwitch docker service file] **************** changed: [10.8.169.22] TASK [openshift_node : Configure Node settings] ******************************** changed: [10.8.169.22] => (item={u'regex': u'^OPTIONS=', u'line': u'OPTIONS=--loglevel=2'}) changed: [10.8.169.22] => (item={u'regex': u'^CONFIG_FILE=', u'line': u'CONFIG_FILE=/etc/origin/node/node-config.yaml'}) changed: [10.8.169.22] => (item={u'regex': u'^IMAGE_VERSION=', u'line': u'IMAGE_VERSION=v3.3'}) TASK [openshift_node : Configure Proxy Settings] ******************************* skipping: [10.8.169.22] => (item={u'regex': u'^HTTPS_PROXY=', u'line': u'HTTPS_PROXY='}) skipping: [10.8.169.22] => (item={u'regex': u'^HTTP_PROXY=', u'line': u'HTTP_PROXY='}) skipping: [10.8.169.22] => (item={u'regex': u'^NO_PROXY=', u'line': u'NO_PROXY=,172.30.0.0/16,10.128.0.0/14'}) TASK [openshift_node : Reload systemd units] *********************************** changed: [10.8.169.22] TASK [openshift_node : Persist net.ipv4.ip_forward sysctl entry] *************** changed: [10.8.169.22] TASK [openshift_node : Start and enable openvswitch docker service] ************ changed: [10.8.169.22] TASK [openshift_node : set_fact] *********************************************** ok: [10.8.169.22] TASK [openshift_node : file] *************************************************** skipping: [10.8.169.22] TASK [openshift_node : Create the Node config] ********************************* changed: [10.8.169.22] TASK [openshift_node : Configure AWS Cloud Provider Settings] ****************** skipping: [10.8.169.22] => (item=(censored due to no_log)) skipping: [10.8.169.22] => (item=(censored due to no_log)) TASK [openshift_node : Configure Node Environment Variables] ******************* TASK [openshift_node : Install NFS storage plugin dependencies] **************** changed: [10.8.169.22] TASK [openshift_node : Check for existence of seboolean] *********************** ok: [10.8.169.22] => (item=virt_use_nfs) ok: [10.8.169.22] => (item=virt_sandbox_use_nfs) TASK [openshift_node : Set seboolean to allow nfs storage plugin access from containers] *** skipping: [10.8.169.22] => (item={'_ansible_parsed': True, u'cmd': [u'getsebool', u'virt_sandbox_use_nfs'], u'end': u'2017-01-10 18:23:14.427186', '_ansible_no_log': False, u'stdout': u'virt_use_nfs --> off', '_ansible_item_result': True, u'changed': False, 'item': u'virt_sandbox_use_nfs', u'delta': u'0:00:00.002541', u'stderr': u'', u'rc': 0, 'invocation': {'module_name': u'command', u'module_args': {u'creates': None, u'executable': None, u'_uses_shell': False, u'_raw_params': u'getsebool virt_sandbox_use_nfs', u'removes': None, u'warn': True, u'chdir': None}}, 'stdout_lines': [u'virt_use_nfs --> off'], 'failed_when_result': False, u'start': u'2017-01-10 18:23:14.424645', u'warnings': [], 'failed': False}) changed: [10.8.169.22] => (item={'_ansible_parsed': True, u'cmd': [u'getsebool', u'virt_use_nfs'], u'end': u'2017-01-10 18:23:13.952075', '_ansible_no_log': False, u'stdout': u'virt_use_nfs --> off', '_ansible_item_result': True, u'changed': False, 'item': u'virt_use_nfs', u'delta': u'0:00:00.002807', u'stderr': u'', u'rc': 0, 'invocation': {'module_name': u'command', u'module_args': {u'creates': None, u'executable': None, u'_uses_shell': False, u'_raw_params': u'getsebool virt_use_nfs', u'removes': None, u'warn': True, u'chdir': None}}, 'stdout_lines': [u'virt_use_nfs --> off'], 'failed_when_result': False, u'start': u'2017-01-10 18:23:13.949268', u'warnings': [], 'failed': False}) TASK [openshift_node : Install GlusterFS storage plugin dependencies] ********** changed: [10.8.169.22] TASK [openshift_node : Check for existence of virt_use_fusefs seboolean] ******* ok: [10.8.169.22] TASK [openshift_node : Set seboolean to allow gluster storage plugin access from containers] *** changed: [10.8.169.22] TASK [openshift_node : Check for existence of virt_sandbox_use_fusefs seboolean] *** ok: [10.8.169.22] TASK [openshift_node : Set seboolean to allow gluster storage plugin access from containers(sandbox)] *** changed: [10.8.169.22] TASK [openshift_node : Install Ceph storage plugin dependencies] *************** changed: [10.8.169.22] TASK [openshift_node : Install iSCSI storage plugin dependencies] ************** changed: [10.8.169.22] TASK [openshift_node : Wait for master API to become available before proceeding] *** ok: [10.8.169.22] TASK [openshift_node : Start and enable node dep] ****************************** changed: [10.8.169.22] TASK [openshift_node : Start and enable node] ********************************** changed: [10.8.169.22] TASK [openshift_node : set_fact] *********************************************** ok: [10.8.169.22] RUNNING HANDLER [openshift_node : restart openvswitch] ************************* skipping: [10.8.169.22] RUNNING HANDLER [openshift_node : restart node] ******************************** skipping: [10.8.169.22] PLAY [Configure nodes] ********************************************************* TASK [setup] ******************************************************************* ok: [10.8.169.210] ok: [10.8.169.215] TASK [openshift_facts : Detecting Operating System] **************************** ok: [10.8.169.215] ok: [10.8.169.210] TASK [openshift_facts : set_fact] ********************************************** ok: [10.8.169.215] ok: [10.8.169.210] TASK [openshift_facts : set_fact] ********************************************** ok: [10.8.169.210] ok: [10.8.169.215] TASK [openshift_facts : Ensure various deps are installed] ********************* ok: [10.8.169.215] => (item=[u'iproute', u'python-dbus', u'python-six', u'PyYAML', u'yum-utils']) ok: [10.8.169.210] => (item=[u'iproute', u'python-dbus', u'python-six', u'PyYAML', u'yum-utils']) TASK [openshift_facts : Gather Cluster facts and set is_containerized if needed] *** changed: [10.8.169.210] changed: [10.8.169.215] TASK [openshift_facts : Set repoquery command] ********************************* ok: [10.8.169.215] ok: [10.8.169.210] TASK [openshift_repos : assert] ************************************************ skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_repos : Ensure libselinux-python is installed] ***************** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_repos : Create any additional repos that are defined] ********** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_repos : Remove the additional repos if no longer defined] ****** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_repos : Configure origin gpg keys if needed] ******************* skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_repos : Configure origin yum repositories RHEL/CentOS] ********* skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_repos : assert] ************************************************ skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_repos : Ensure libselinux-python is installed] ***************** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_repos : Create any additional repos that are defined] ********** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_repos : Remove the additional repos if no longer defined] ****** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_repos : Configure origin gpg keys if needed] ******************* skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_repos : Configure origin yum repositories RHEL/CentOS] ********* skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_docker_facts : Set docker facts] ******************************* ok: [10.8.169.210] => (item={u'local_facts': {u'blocked_registries': u'', u'additional_registries': u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888,registry.ops.openshift.com', u'log_driver': u'', u'disable_push_dockerhub': u'', u'hosted_registry_insecure': False, u'hosted_registry_network': u'172.30.0.0/16', u'log_options': u'', u'insecure_registries': u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888,registry.ops.openshift.com', u'options': u''}, u'role': u'docker'}) ok: [10.8.169.215] => (item={u'local_facts': {u'blocked_registries': u'', u'additional_registries': u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888,registry.ops.openshift.com', u'log_driver': u'', u'disable_push_dockerhub': u'', u'hosted_registry_insecure': False, u'hosted_registry_network': u'172.30.0.0/16', u'log_options': u'', u'insecure_registries': u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888,registry.ops.openshift.com', u'options': u''}, u'role': u'docker'}) TASK [openshift_docker_facts : set_fact] *************************************** ok: [10.8.169.210] ok: [10.8.169.215] TASK [openshift_docker_facts : set_fact] *************************************** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_docker_facts : set_fact] *************************************** ok: [10.8.169.210] ok: [10.8.169.215] TASK [os_firewall : Assert - Do not use firewalld on Atomic Host] ************** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [os_firewall : Install firewalld packages] ******************************** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [os_firewall : Ensure iptables services are not enabled] ****************** ok: [10.8.169.210] => (item=iptables) ok: [10.8.169.215] => (item=iptables) ok: [10.8.169.210] => (item=ip6tables) ok: [10.8.169.215] => (item=ip6tables) TASK [os_firewall : Start and enable firewalld service] ************************ ok: [10.8.169.210] ok: [10.8.169.215] TASK [os_firewall : need to pause here, otherwise the firewalld service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.210] TASK [os_firewall : Add firewalld allow rules] ********************************* TASK [os_firewall : Remove firewalld allow rules] ****************************** TASK [os_firewall : Ensure firewalld service is not enabled] ******************* skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [os_firewall : Install iptables packages] ********************************* skipping: [10.8.169.215] => (item=[]) skipping: [10.8.169.210] => (item=[]) TASK [os_firewall : Start and enable iptables service] ************************* skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [os_firewall : need to pause here, otherwise the iptables service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.210] TASK [os_firewall : Add iptables allow rules] ********************************** TASK [os_firewall : Remove iptables rules] ************************************* TASK [docker : Get current installed Docker version] *************************** ok: [10.8.169.215] ok: [10.8.169.210] TASK [docker : Error out if Docker pre-installed but too old] ****************** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [docker : Error out if requested Docker is too old] *********************** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [docker : Get latest available version of Docker] ************************* skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [docker : Fail if Docker version requested but downgrade is required] ***** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [docker : Error out if attempting to upgrade Docker across the 1.10 boundary] *** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [docker : Install Docker] ************************************************* ok: [10.8.169.215] ok: [10.8.169.210] TASK [docker : Ensure docker.service.d directory exists] *********************** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [docker : Configure Docker service unit file] ***************************** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [docker : Getting current systemd-udevd exec command] ********************* skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [docker : Assure systemd-udevd.service.d directory exists] **************** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [docker : Create systemd-udevd override file] ***************************** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [docker : stat] *********************************************************** ok: [10.8.169.210] ok: [10.8.169.215] TASK [docker : Set registry params] ******************************************** skipping: [10.8.169.210] => (item={u'reg_conf_var': u'BLOCK_REGISTRY', u'reg_flag': u'--block-registry', u'reg_fact_val': u''}) ok: [10.8.169.210] => (item={u'reg_conf_var': u'ADD_REGISTRY', u'reg_flag': u'--add-registry', u'reg_fact_val': [u'registry.ops.openshift.com', u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888', u'XXXXXXXXXXXXXXXXXXXXXXXXXXX']}) skipping: [10.8.169.215] => (item={u'reg_conf_var': u'BLOCK_REGISTRY', u'reg_flag': u'--block-registry', u'reg_fact_val': u''}) ok: [10.8.169.215] => (item={u'reg_conf_var': u'ADD_REGISTRY', u'reg_flag': u'--add-registry', u'reg_fact_val': [u'registry.ops.openshift.com', u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888', u'XXXXXXXXXXXXXXXXXXXXXXXXXXX']}) ok: [10.8.169.215] => (item={u'reg_conf_var': u'INSECURE_REGISTRY', u'reg_flag': u'--insecure-registry', u'reg_fact_val': [u'registry.ops.openshift.com', u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888']}) ok: [10.8.169.210] => (item={u'reg_conf_var': u'INSECURE_REGISTRY', u'reg_flag': u'--insecure-registry', u'reg_fact_val': [u'registry.ops.openshift.com', u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888']}) TASK [docker : Set Proxy Settings] ********************************************* ok: [10.8.169.210] => (item={u'reg_conf_var': u'HTTP_PROXY', u'reg_fact_val': u''}) ok: [10.8.169.215] => (item={u'reg_conf_var': u'HTTP_PROXY', u'reg_fact_val': u''}) ok: [10.8.169.210] => (item={u'reg_conf_var': u'HTTPS_PROXY', u'reg_fact_val': u''}) ok: [10.8.169.215] => (item={u'reg_conf_var': u'HTTPS_PROXY', u'reg_fact_val': u''}) ok: [10.8.169.210] => (item={u'reg_conf_var': u'NO_PROXY', u'reg_fact_val': u''}) ok: [10.8.169.215] => (item={u'reg_conf_var': u'NO_PROXY', u'reg_fact_val': u''}) TASK [docker : Set various Docker options] ************************************* ok: [10.8.169.215] ok: [10.8.169.210] TASK [docker : Start the Docker service] *************************************** ok: [10.8.169.215] ok: [10.8.169.210] TASK [docker : set_fact] ******************************************************* ok: [10.8.169.215] ok: [10.8.169.210] TASK [openshift_version : set_fact] ******************************************** ok: [10.8.169.210] ok: [10.8.169.215] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_version : set_fact] ******************************************** ok: [10.8.169.210] ok: [10.8.169.215] TASK [openshift_version : Use openshift.common.version fact as version to configure if already installed] *** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_version : Set rpm version to configure if openshift_pkg_version specified] *** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_version : Gather common package version] *********************** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_version : Set containerized version to configure if openshift_image_tag specified] *** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_version : Set containerized version to configure if openshift_release specified] *** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_version : Lookup latest containerized version if no version specified] *** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_version : Set precise containerized version to configure if openshift_release specified] *** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_version : set_fact] ******************************************** ok: [10.8.169.210] ok: [10.8.169.215] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_version : debug] *********************************************** ok: [10.8.169.215] => { "openshift_release": "VARIABLE IS NOT DEFINED!" } ok: [10.8.169.210] => { "openshift_release": "VARIABLE IS NOT DEFINED!" } TASK [openshift_version : debug] *********************************************** ok: [10.8.169.215] => { "openshift_image_tag": "v3.3" } ok: [10.8.169.210] => { "openshift_image_tag": "v3.3" } TASK [openshift_version : debug] *********************************************** ok: [10.8.169.210] => { "openshift_pkg_version": "-3.3.1.9" } ok: [10.8.169.215] => { "openshift_pkg_version": "-3.3.1.9" } TASK [openshift_common : fail] ************************************************* skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_common : fail] ************************************************* skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_common : fail] ************************************************* skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_common : fail] ************************************************* skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_common : Set common Cluster facts] ***************************** changed: [10.8.169.210] changed: [10.8.169.215] TASK [openshift_common : Install the base package for versioning] ************** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_common : Set version facts] ************************************ ok: [10.8.169.215] ok: [10.8.169.210] TASK [openshift_common : set_fact] ********************************************* ok: [10.8.169.210] ok: [10.8.169.215] TASK [openshift_common : Set hostname] ***************************************** changed: [10.8.169.215] changed: [10.8.169.210] TASK [openshift_clock : Set clock facts] *************************************** changed: [10.8.169.215] changed: [10.8.169.210] TASK [openshift_clock : Install ntp package] *********************************** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_clock : Start and enable ntpd/chronyd] ************************* changed: [10.8.169.215] changed: [10.8.169.210] TASK [openshift_repos : assert] ************************************************ skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_repos : Ensure libselinux-python is installed] ***************** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_repos : Create any additional repos that are defined] ********** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_repos : Remove the additional repos if no longer defined] ****** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_repos : Configure origin gpg keys if needed] ******************* skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_repos : Configure origin yum repositories RHEL/CentOS] ********* skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [os_firewall : Assert - Do not use firewalld on Atomic Host] ************** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [os_firewall : Install firewalld packages] ******************************** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [os_firewall : Ensure iptables services are not enabled] ****************** ok: [10.8.169.210] => (item=iptables) ok: [10.8.169.215] => (item=iptables) ok: [10.8.169.215] => (item=ip6tables) ok: [10.8.169.210] => (item=ip6tables) TASK [os_firewall : Start and enable firewalld service] ************************ ok: [10.8.169.215] ok: [10.8.169.210] TASK [os_firewall : need to pause here, otherwise the firewalld service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.210] TASK [os_firewall : Add firewalld allow rules] ********************************* TASK [os_firewall : Remove firewalld allow rules] ****************************** TASK [os_firewall : Ensure firewalld service is not enabled] ******************* skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [os_firewall : Install iptables packages] ********************************* skipping: [10.8.169.210] => (item=[]) skipping: [10.8.169.215] => (item=[]) TASK [os_firewall : Start and enable iptables service] ************************* skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [os_firewall : need to pause here, otherwise the iptables service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.210] TASK [os_firewall : Add iptables allow rules] ********************************** TASK [os_firewall : Remove iptables rules] ************************************* TASK [os_firewall : Assert - Do not use firewalld on Atomic Host] ************** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [os_firewall : Install firewalld packages] ******************************** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [os_firewall : Ensure iptables services are not enabled] ****************** ok: [10.8.169.215] => (item=iptables) ok: [10.8.169.210] => (item=iptables) ok: [10.8.169.215] => (item=ip6tables) ok: [10.8.169.210] => (item=ip6tables) TASK [os_firewall : Start and enable firewalld service] ************************ ok: [10.8.169.215] ok: [10.8.169.210] TASK [os_firewall : need to pause here, otherwise the firewalld service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.210] TASK [os_firewall : Add firewalld allow rules] ********************************* TASK [os_firewall : Remove firewalld allow rules] ****************************** TASK [os_firewall : Ensure firewalld service is not enabled] ******************* skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [os_firewall : Install iptables packages] ********************************* skipping: [10.8.169.210] => (item=[]) skipping: [10.8.169.215] => (item=[]) TASK [os_firewall : Start and enable iptables service] ************************* skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [os_firewall : need to pause here, otherwise the iptables service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.210] TASK [os_firewall : Add iptables allow rules] ********************************** TASK [os_firewall : Remove iptables rules] ************************************* TASK [docker : Get current installed Docker version] *************************** ok: [10.8.169.210] ok: [10.8.169.215] TASK [docker : Error out if Docker pre-installed but too old] ****************** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [docker : Error out if requested Docker is too old] *********************** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [docker : Get latest available version of Docker] ************************* skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [docker : Fail if Docker version requested but downgrade is required] ***** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [docker : Error out if attempting to upgrade Docker across the 1.10 boundary] *** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [docker : Install Docker] ************************************************* ok: [10.8.169.210] ok: [10.8.169.215] TASK [docker : Ensure docker.service.d directory exists] *********************** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [docker : Configure Docker service unit file] ***************************** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [docker : Getting current systemd-udevd exec command] ********************* skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [docker : Assure systemd-udevd.service.d directory exists] **************** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [docker : Create systemd-udevd override file] ***************************** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [docker : stat] *********************************************************** ok: [10.8.169.215] ok: [10.8.169.210] TASK [docker : Set registry params] ******************************************** skipping: [10.8.169.210] => (item={u'reg_conf_var': u'BLOCK_REGISTRY', u'reg_flag': u'--block-registry', u'reg_fact_val': u''}) skipping: [10.8.169.215] => (item={u'reg_conf_var': u'BLOCK_REGISTRY', u'reg_flag': u'--block-registry', u'reg_fact_val': u''}) ok: [10.8.169.215] => (item={u'reg_conf_var': u'ADD_REGISTRY', u'reg_flag': u'--add-registry', u'reg_fact_val': [u'registry.ops.openshift.com', u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888', u'XXXXXXXXXXXXXXXXXXXXXXXXXXX']}) ok: [10.8.169.210] => (item={u'reg_conf_var': u'ADD_REGISTRY', u'reg_flag': u'--add-registry', u'reg_fact_val': [u'registry.ops.openshift.com', u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888', u'XXXXXXXXXXXXXXXXXXXXXXXXXXX']}) ok: [10.8.169.210] => (item={u'reg_conf_var': u'INSECURE_REGISTRY', u'reg_flag': u'--insecure-registry', u'reg_fact_val': [u'registry.ops.openshift.com', u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888']}) ok: [10.8.169.215] => (item={u'reg_conf_var': u'INSECURE_REGISTRY', u'reg_flag': u'--insecure-registry', u'reg_fact_val': [u'registry.ops.openshift.com', u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888']}) TASK [docker : Set Proxy Settings] ********************************************* ok: [10.8.169.210] => (item={u'reg_conf_var': u'HTTP_PROXY', u'reg_fact_val': u''}) ok: [10.8.169.215] => (item={u'reg_conf_var': u'HTTP_PROXY', u'reg_fact_val': u''}) ok: [10.8.169.210] => (item={u'reg_conf_var': u'HTTPS_PROXY', u'reg_fact_val': u''}) ok: [10.8.169.215] => (item={u'reg_conf_var': u'HTTPS_PROXY', u'reg_fact_val': u''}) ok: [10.8.169.210] => (item={u'reg_conf_var': u'NO_PROXY', u'reg_fact_val': u''}) ok: [10.8.169.215] => (item={u'reg_conf_var': u'NO_PROXY', u'reg_fact_val': u''}) TASK [docker : Set various Docker options] ************************************* ok: [10.8.169.210] ok: [10.8.169.215] TASK [docker : Start the Docker service] *************************************** ok: [10.8.169.215] ok: [10.8.169.210] TASK [docker : set_fact] ******************************************************* ok: [10.8.169.210] ok: [10.8.169.215] TASK [openshift_node_certificates : Ensure CA certificate exists on openshift_ca_host] *** ok: [10.8.169.210 -> 10.8.169.22] TASK [openshift_node_certificates : fail] ************************************** skipping: [10.8.169.210] TASK [openshift_node_certificates : Check status of node certificates] ********* ok: [10.8.169.215] => (item=system:node:172.16.132.150.crt) ok: [10.8.169.210] => (item=system:node:172.16.132.148.crt) ok: [10.8.169.210] => (item=system:node:172.16.132.148.key) ok: [10.8.169.215] => (item=system:node:172.16.132.150.key) ok: [10.8.169.215] => (item=system:node:172.16.132.150.kubeconfig) ok: [10.8.169.210] => (item=system:node:172.16.132.148.kubeconfig) ok: [10.8.169.210] => (item=ca.crt) ok: [10.8.169.215] => (item=ca.crt) ok: [10.8.169.210] => (item=server.key) ok: [10.8.169.215] => (item=server.key) ok: [10.8.169.215] => (item=server.crt) ok: [10.8.169.210] => (item=server.crt) TASK [openshift_node_certificates : set_fact] ********************************** ok: [10.8.169.215] ok: [10.8.169.210] TASK [openshift_node_certificates : Create openshift_generated_configs_dir if it does not exist] *** ok: [10.8.169.215 -> 10.8.169.22] ok: [10.8.169.210 -> 10.8.169.22] TASK [openshift_node_certificates : Generate the node client config] *********** changed: [10.8.169.215 -> 10.8.169.22] changed: [10.8.169.210 -> 10.8.169.22] TASK [openshift_node_certificates : Generate the node server certificate] ****** changed: [10.8.169.210 -> 10.8.169.22] changed: [10.8.169.215 -> 10.8.169.22] TASK [openshift_node_certificates : Create local temp directory for syncing certs] *** ok: [10.8.169.215 -> localhost] ok: [10.8.169.210 -> localhost] TASK [openshift_node_certificates : Create a tarball of the node config directories] *** changed: [10.8.169.215 -> 10.8.169.22] changed: [10.8.169.210 -> 10.8.169.22] TASK [openshift_node_certificates : Retrieve the node config tarballs from the master] *** changed: [10.8.169.210 -> 10.8.169.22] changed: [10.8.169.215 -> 10.8.169.22] TASK [openshift_node_certificates : Ensure certificate directory exists] ******* changed: [10.8.169.210] changed: [10.8.169.215] TASK [openshift_node_certificates : Unarchive the tarball on the node] ********* changed: [10.8.169.210] changed: [10.8.169.215] TASK [openshift_node_certificates : file] ************************************** ok: [10.8.169.210 -> localhost] ok: [10.8.169.215 -> localhost] TASK [openshift_node_certificates : Copy OpenShift CA to system CA trust] ****** changed: [10.8.169.210] => (item={u'cert': u'/etc/origin/node/ca.crt', u'id': u'openshift'}) changed: [10.8.169.215] => (item={u'cert': u'/etc/origin/node/ca.crt', u'id': u'openshift'}) TASK [openshift_cloud_provider : Set cloud provider facts] ********************* changed: [10.8.169.215] changed: [10.8.169.210] TASK [openshift_cloud_provider : Create cloudprovider config dir] ************** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_cloud_provider : fail] ***************************************** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_cloud_provider : Create cloud config] ************************** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_cloud_provider : Create cloud config file] ********************* skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_cloud_provider : Configure AWS cloud provider] ***************** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_cloud_provider : Create cloud config file] ********************* skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_cloud_provider : Configure GCE cloud provider] ***************** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_repos : assert] ************************************************ skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_repos : Ensure libselinux-python is installed] ***************** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_repos : Create any additional repos that are defined] ********** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_repos : Remove the additional repos if no longer defined] ****** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_repos : Configure origin gpg keys if needed] ******************* skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_repos : Configure origin yum repositories RHEL/CentOS] ********* skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_repos : assert] ************************************************ skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_repos : Ensure libselinux-python is installed] ***************** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_repos : Create any additional repos that are defined] ********** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_repos : Remove the additional repos if no longer defined] ****** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_repos : Configure origin gpg keys if needed] ******************* skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_repos : Configure origin yum repositories RHEL/CentOS] ********* skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [os_firewall : Assert - Do not use firewalld on Atomic Host] ************** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [os_firewall : Install firewalld packages] ******************************** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [os_firewall : Ensure iptables services are not enabled] ****************** ok: [10.8.169.210] => (item=iptables) ok: [10.8.169.215] => (item=iptables) ok: [10.8.169.210] => (item=ip6tables) ok: [10.8.169.215] => (item=ip6tables) TASK [os_firewall : Start and enable firewalld service] ************************ ok: [10.8.169.210] ok: [10.8.169.215] TASK [os_firewall : need to pause here, otherwise the firewalld service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.210] TASK [os_firewall : Add firewalld allow rules] ********************************* TASK [os_firewall : Remove firewalld allow rules] ****************************** TASK [os_firewall : Ensure firewalld service is not enabled] ******************* skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [os_firewall : Install iptables packages] ********************************* skipping: [10.8.169.210] => (item=[]) skipping: [10.8.169.215] => (item=[]) TASK [os_firewall : Start and enable iptables service] ************************* skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [os_firewall : need to pause here, otherwise the iptables service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.210] TASK [os_firewall : Add iptables allow rules] ********************************** TASK [os_firewall : Remove iptables rules] ************************************* TASK [openshift_node_dnsmasq : Check for NetworkManager service] *************** ok: [10.8.169.215] ok: [10.8.169.210] TASK [openshift_node_dnsmasq : Set fact using_network_manager] ***************** ok: [10.8.169.210] ok: [10.8.169.215] TASK [openshift_node_dnsmasq : Install dnsmasq] ******************************** ok: [10.8.169.210] ok: [10.8.169.215] TASK [openshift_node_dnsmasq : Install dnsmasq configuration] ****************** changed: [10.8.169.215] changed: [10.8.169.210] TASK [openshift_node_dnsmasq : Deploy additional dnsmasq.conf] ***************** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_node_dnsmasq : Enable dnsmasq] ********************************* changed: [10.8.169.210] changed: [10.8.169.215] TASK [openshift_node_dnsmasq : Install network manager dispatch script] ******** changed: [10.8.169.210] changed: [10.8.169.215] RUNNING HANDLER [openshift_node_certificates : update ca trust] **************** changed: [10.8.169.210] changed: [10.8.169.215] RUNNING HANDLER [openshift_node_certificates : restart docker after updating ca trust] *** changed: [10.8.169.215] changed: [10.8.169.210] RUNNING HANDLER [openshift_node_dnsmasq : restart NetworkManager] ************** changed: [10.8.169.215] changed: [10.8.169.210] RUNNING HANDLER [openshift_node_dnsmasq : restart dnsmasq] ********************* changed: [10.8.169.215] changed: [10.8.169.210] TASK [openshift_node_dnsmasq : fail] ******************************************* skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [os_firewall : Assert - Do not use firewalld on Atomic Host] ************** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [os_firewall : Install firewalld packages] ******************************** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [os_firewall : Ensure iptables services are not enabled] ****************** ok: [10.8.169.215] => (item=iptables) ok: [10.8.169.210] => (item=iptables) ok: [10.8.169.210] => (item=ip6tables) ok: [10.8.169.215] => (item=ip6tables) TASK [os_firewall : Start and enable firewalld service] ************************ ok: [10.8.169.210] ok: [10.8.169.215] TASK [os_firewall : need to pause here, otherwise the firewalld service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.210] TASK [os_firewall : Add firewalld allow rules] ********************************* changed: [10.8.169.215] => (item={u'port': u'10250/tcp', u'service': u'Kubernetes kubelet'}) changed: [10.8.169.210] => (item={u'port': u'10250/tcp', u'service': u'Kubernetes kubelet'}) changed: [10.8.169.210] => (item={u'port': u'80/tcp', u'service': u'http'}) changed: [10.8.169.215] => (item={u'port': u'80/tcp', u'service': u'http'}) changed: [10.8.169.210] => (item={u'port': u'443/tcp', u'service': u'https'}) changed: [10.8.169.215] => (item={u'port': u'443/tcp', u'service': u'https'}) changed: [10.8.169.210] => (item={u'port': u'10255/tcp', u'service': u'Openshift kubelet ReadOnlyPort'}) changed: [10.8.169.215] => (item={u'port': u'10255/tcp', u'service': u'Openshift kubelet ReadOnlyPort'}) changed: [10.8.169.210] => (item={u'port': u'10255/udp', u'service': u'Openshift kubelet ReadOnlyPort udp'}) changed: [10.8.169.215] => (item={u'port': u'10255/udp', u'service': u'Openshift kubelet ReadOnlyPort udp'}) changed: [10.8.169.210] => (item={u'when': u'openshift.node.use_openshift_sdn | bool', u'port': u'4789/udp', u'service': u'OpenShift OVS sdn'}) changed: [10.8.169.215] => (item={u'when': u'openshift.node.use_openshift_sdn | bool', u'port': u'4789/udp', u'service': u'OpenShift OVS sdn'}) TASK [os_firewall : Remove firewalld allow rules] ****************************** TASK [os_firewall : Ensure firewalld service is not enabled] ******************* skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [os_firewall : Install iptables packages] ********************************* skipping: [10.8.169.215] => (item=[]) skipping: [10.8.169.210] => (item=[]) TASK [os_firewall : Start and enable iptables service] ************************* skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [os_firewall : need to pause here, otherwise the iptables service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.210] TASK [os_firewall : Add iptables allow rules] ********************************** skipping: [10.8.169.210] => (item={u'when': u'openshift.node.use_openshift_sdn | bool', u'port': u'4789/udp', u'service': u'OpenShift OVS sdn'}) skipping: [10.8.169.210] => (item={u'port': u'10255/udp', u'service': u'Openshift kubelet ReadOnlyPort udp'}) skipping: [10.8.169.215] => (item={u'when': u'openshift.node.use_openshift_sdn | bool', u'port': u'4789/udp', u'service': u'OpenShift OVS sdn'}) skipping: [10.8.169.215] => (item={u'port': u'10255/udp', u'service': u'Openshift kubelet ReadOnlyPort udp'}) skipping: [10.8.169.215] => (item={u'port': u'10255/tcp', u'service': u'Openshift kubelet ReadOnlyPort'}) skipping: [10.8.169.215] => (item={u'port': u'443/tcp', u'service': u'https'}) skipping: [10.8.169.210] => (item={u'port': u'10255/tcp', u'service': u'Openshift kubelet ReadOnlyPort'}) skipping: [10.8.169.210] => (item={u'port': u'443/tcp', u'service': u'https'}) skipping: [10.8.169.215] => (item={u'port': u'80/tcp', u'service': u'http'}) skipping: [10.8.169.215] => (item={u'port': u'10250/tcp', u'service': u'Kubernetes kubelet'}) skipping: [10.8.169.210] => (item={u'port': u'80/tcp', u'service': u'http'}) skipping: [10.8.169.210] => (item={u'port': u'10250/tcp', u'service': u'Kubernetes kubelet'}) TASK [os_firewall : Remove iptables rules] ************************************* TASK [openshift_node : fail] *************************************************** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_node : Set node facts] ***************************************** ok: [10.8.169.215] => (item={u'local_facts': {u'labels': {}}, u'role': u'node'}) ok: [10.8.169.210] => (item={u'local_facts': {u'labels': {}}, u'role': u'node'}) changed: [10.8.169.215] => (item={u'local_facts': {u'node_image': u'', u'schedulable': u'', u'sdn_mtu': u'', u'env_vars': u'', u'labels': {'region': u'primary', 'zone': u'west'}, u'dns_ip': u'172.16.132.150', u'set_node_ip': u'', u'debug_level': u'2', u'storage_plugin_deps': u'', u'proxy_mode': u'iptables', u'iptables_sync_period': u'', u'local_quota_per_fsgroup': u'', u'ovs_image': u'', u'kubelet_args': u'', u'annotations': u'', u'registry_url': u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888/openshift3/ose-${component}:${version}'}, u'role': u'node'}) changed: [10.8.169.210] => (item={u'local_facts': {u'node_image': u'', u'schedulable': u'', u'sdn_mtu': u'', u'env_vars': u'', u'labels': {'region': u'primary', 'zone': u'east'}, u'dns_ip': u'172.16.132.148', u'set_node_ip': u'', u'debug_level': u'2', u'storage_plugin_deps': u'', u'proxy_mode': u'iptables', u'iptables_sync_period': u'', u'local_quota_per_fsgroup': u'', u'ovs_image': u'', u'kubelet_args': u'', u'annotations': u'', u'registry_url': u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888/openshift3/ose-${component}:${version}'}, u'role': u'node'}) TASK [openshift_node : Install Node package] *********************************** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_node : Check for tuned package] ******************************** ok: [10.8.169.210] ok: [10.8.169.215] TASK [openshift_node : Set atomic-guest tuned profile] ************************* skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_node : Install sdn-ovs package] ******************************** skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_node : Pre-pull node image] ************************************ changed: [10.8.169.215] changed: [10.8.169.210] TASK [openshift_node : Pre-pull openvswitch image] ***************************** changed: [10.8.169.210] changed: [10.8.169.215] TASK [openshift_node : Install Node dependencies docker service file] ********** changed: [10.8.169.210] changed: [10.8.169.215] TASK [openshift_node : Install Node docker service file] *********************** changed: [10.8.169.210] changed: [10.8.169.215] TASK [openshift_node : Create the openvswitch service env file] **************** changed: [10.8.169.210] changed: [10.8.169.215] TASK [openshift_node : Create OpenvSwitch service.d directory] ***************** changed: [10.8.169.210] changed: [10.8.169.215] TASK [openshift_node : Install OpenvSwitch service OOM fix] ******************** changed: [10.8.169.215] changed: [10.8.169.210] TASK [openshift_node : Install OpenvSwitch docker service file] **************** changed: [10.8.169.210] changed: [10.8.169.215] TASK [openshift_node : Configure Node settings] ******************************** changed: [10.8.169.210] => (item={u'regex': u'^OPTIONS=', u'line': u'OPTIONS=--loglevel=2'}) changed: [10.8.169.215] => (item={u'regex': u'^OPTIONS=', u'line': u'OPTIONS=--loglevel=2'}) changed: [10.8.169.210] => (item={u'regex': u'^CONFIG_FILE=', u'line': u'CONFIG_FILE=/etc/origin/node/node-config.yaml'}) changed: [10.8.169.215] => (item={u'regex': u'^CONFIG_FILE=', u'line': u'CONFIG_FILE=/etc/origin/node/node-config.yaml'}) changed: [10.8.169.210] => (item={u'regex': u'^IMAGE_VERSION=', u'line': u'IMAGE_VERSION=v3.3'}) changed: [10.8.169.215] => (item={u'regex': u'^IMAGE_VERSION=', u'line': u'IMAGE_VERSION=v3.3'}) TASK [openshift_node : Configure Proxy Settings] ******************************* skipping: [10.8.169.215] => (item={u'regex': u'^NO_PROXY=', u'line': u'NO_PROXY=,172.30.0.0/16,10.128.0.0/14'}) skipping: [10.8.169.215] => (item={u'regex': u'^HTTPS_PROXY=', u'line': u'HTTPS_PROXY='}) skipping: [10.8.169.215] => (item={u'regex': u'^HTTP_PROXY=', u'line': u'HTTP_PROXY='}) skipping: [10.8.169.210] => (item={u'regex': u'^NO_PROXY=', u'line': u'NO_PROXY=,172.30.0.0/16,10.128.0.0/14'}) skipping: [10.8.169.210] => (item={u'regex': u'^HTTPS_PROXY=', u'line': u'HTTPS_PROXY='}) skipping: [10.8.169.210] => (item={u'regex': u'^HTTP_PROXY=', u'line': u'HTTP_PROXY='}) TASK [openshift_node : Reload systemd units] *********************************** changed: [10.8.169.215] changed: [10.8.169.210] TASK [openshift_node : Persist net.ipv4.ip_forward sysctl entry] *************** changed: [10.8.169.215] changed: [10.8.169.210] TASK [openshift_node : Start and enable openvswitch docker service] ************ changed: [10.8.169.210] changed: [10.8.169.215] TASK [openshift_node : set_fact] *********************************************** ok: [10.8.169.215] ok: [10.8.169.210] TASK [openshift_node : file] *************************************************** skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_node : Create the Node config] ********************************* changed: [10.8.169.210] changed: [10.8.169.215] TASK [openshift_node : Configure AWS Cloud Provider Settings] ****************** skipping: [10.8.169.210] => (item=(censored due to no_log)) skipping: [10.8.169.210] => (item=(censored due to no_log)) skipping: [10.8.169.215] => (item=(censored due to no_log)) skipping: [10.8.169.215] => (item=(censored due to no_log)) TASK [openshift_node : Configure Node Environment Variables] ******************* TASK [openshift_node : Install NFS storage plugin dependencies] **************** changed: [10.8.169.210] changed: [10.8.169.215] TASK [openshift_node : Check for existence of seboolean] *********************** ok: [10.8.169.215] => (item=virt_use_nfs) ok: [10.8.169.210] => (item=virt_use_nfs) ok: [10.8.169.210] => (item=virt_sandbox_use_nfs) ok: [10.8.169.215] => (item=virt_sandbox_use_nfs) TASK [openshift_node : Set seboolean to allow nfs storage plugin access from containers] *** skipping: [10.8.169.210] => (item={'_ansible_parsed': True, u'cmd': [u'getsebool', u'virt_sandbox_use_nfs'], u'end': u'2017-01-10 18:27:00.102827', '_ansible_no_log': False, u'stdout': u'virt_use_nfs --> off', '_ansible_item_result': True, u'changed': False, 'item': u'virt_sandbox_use_nfs', u'delta': u'0:00:00.002425', u'stderr': u'', u'rc': 0, 'invocation': {'module_name': u'command', u'module_args': {u'creates': None, u'executable': None, u'_uses_shell': False, u'_raw_params': u'getsebool virt_sandbox_use_nfs', u'removes': None, u'warn': True, u'chdir': None}}, 'stdout_lines': [u'virt_use_nfs --> off'], 'failed_when_result': False, u'start': u'2017-01-10 18:27:00.100402', u'warnings': [], 'failed': False}) changed: [10.8.169.210] => (item={'_ansible_parsed': True, u'cmd': [u'getsebool', u'virt_use_nfs'], u'end': u'2017-01-10 18:26:59.613210', '_ansible_no_log': False, u'stdout': u'virt_use_nfs --> off', '_ansible_item_result': True, u'changed': False, 'item': u'virt_use_nfs', u'delta': u'0:00:00.002305', u'stderr': u'', u'rc': 0, 'invocation': {'module_name': u'command', u'module_args': {u'creates': None, u'executable': None, u'_uses_shell': False, u'_raw_params': u'getsebool virt_use_nfs', u'removes': None, u'warn': True, u'chdir': None}}, 'stdout_lines': [u'virt_use_nfs --> off'], 'failed_when_result': False, u'start': u'2017-01-10 18:26:59.610905', u'warnings': [], 'failed': False}) skipping: [10.8.169.215] => (item={'_ansible_parsed': True, u'cmd': [u'getsebool', u'virt_sandbox_use_nfs'], u'end': u'2017-01-10 18:27:00.067967', '_ansible_no_log': False, u'stdout': u'virt_use_nfs --> off', '_ansible_item_result': True, u'changed': False, 'item': u'virt_sandbox_use_nfs', u'delta': u'0:00:00.002327', u'stderr': u'', u'rc': 0, 'invocation': {'module_name': u'command', u'module_args': {u'creates': None, u'executable': None, u'_uses_shell': False, u'_raw_params': u'getsebool virt_sandbox_use_nfs', u'removes': None, u'warn': True, u'chdir': None}}, 'stdout_lines': [u'virt_use_nfs --> off'], 'failed_when_result': False, u'start': u'2017-01-10 18:27:00.065640', u'warnings': [], 'failed': False}) changed: [10.8.169.215] => (item={'_ansible_parsed': True, u'cmd': [u'getsebool', u'virt_use_nfs'], u'end': u'2017-01-10 18:26:59.573903', '_ansible_no_log': False, u'stdout': u'virt_use_nfs --> off', '_ansible_item_result': True, u'changed': False, 'item': u'virt_use_nfs', u'delta': u'0:00:00.002505', u'stderr': u'', u'rc': 0, 'invocation': {'module_name': u'command', u'module_args': {u'creates': None, u'executable': None, u'_uses_shell': False, u'_raw_params': u'getsebool virt_use_nfs', u'removes': None, u'warn': True, u'chdir': None}}, 'stdout_lines': [u'virt_use_nfs --> off'], 'failed_when_result': False, u'start': u'2017-01-10 18:26:59.571398', u'warnings': [], 'failed': False}) TASK [openshift_node : Install GlusterFS storage plugin dependencies] ********** changed: [10.8.169.210] changed: [10.8.169.215] TASK [openshift_node : Check for existence of virt_use_fusefs seboolean] ******* ok: [10.8.169.210] ok: [10.8.169.215] TASK [openshift_node : Set seboolean to allow gluster storage plugin access from containers] *** changed: [10.8.169.215] changed: [10.8.169.210] TASK [openshift_node : Check for existence of virt_sandbox_use_fusefs seboolean] *** ok: [10.8.169.215] ok: [10.8.169.210] TASK [openshift_node : Set seboolean to allow gluster storage plugin access from containers(sandbox)] *** changed: [10.8.169.210] changed: [10.8.169.215] TASK [openshift_node : Install Ceph storage plugin dependencies] *************** changed: [10.8.169.210] changed: [10.8.169.215] TASK [openshift_node : Install iSCSI storage plugin dependencies] ************** changed: [10.8.169.210] changed: [10.8.169.215] TASK [openshift_node : Wait for master API to become available before proceeding] *** ok: [10.8.169.210] ok: [10.8.169.215] TASK [openshift_node : Start and enable node dep] ****************************** changed: [10.8.169.215] changed: [10.8.169.210] TASK [openshift_node : Start and enable node] ********************************** changed: [10.8.169.215] changed: [10.8.169.210] TASK [openshift_node : set_fact] *********************************************** ok: [10.8.169.210] ok: [10.8.169.215] RUNNING HANDLER [openshift_node : restart openvswitch] ************************* skipping: [10.8.169.210] skipping: [10.8.169.215] RUNNING HANDLER [openshift_node : restart node] ******************************** skipping: [10.8.169.215] skipping: [10.8.169.210] PLAY [Additional node config] ************************************************** TASK [setup] ******************************************************************* ok: [10.8.169.215] ok: [10.8.169.210] ok: [10.8.169.22] TASK [openshift_facts : Detecting Operating System] **************************** skipping: [10.8.169.210] skipping: [10.8.169.22] skipping: [10.8.169.215] TASK [openshift_facts : set_fact] ********************************************** skipping: [10.8.169.210] skipping: [10.8.169.22] skipping: [10.8.169.215] TASK [openshift_facts : set_fact] ********************************************** skipping: [10.8.169.210] skipping: [10.8.169.215] skipping: [10.8.169.22] TASK [openshift_facts : Ensure various deps are installed] ********************* skipping: [10.8.169.210] => (item=[]) skipping: [10.8.169.215] => (item=[]) skipping: [10.8.169.22] => (item=[]) TASK [openshift_facts : Gather Cluster facts and set is_containerized if needed] *** skipping: [10.8.169.22] skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_facts : Set repoquery command] ********************************* skipping: [10.8.169.210] skipping: [10.8.169.22] skipping: [10.8.169.215] TASK [openshift_facts : Detecting Operating System] **************************** skipping: [10.8.169.210] skipping: [10.8.169.215] skipping: [10.8.169.22] TASK [openshift_facts : set_fact] ********************************************** skipping: [10.8.169.210] skipping: [10.8.169.22] skipping: [10.8.169.215] TASK [openshift_facts : set_fact] ********************************************** skipping: [10.8.169.22] skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_facts : Ensure various deps are installed] ********************* skipping: [10.8.169.215] => (item=[]) skipping: [10.8.169.210] => (item=[]) skipping: [10.8.169.22] => (item=[]) TASK [openshift_facts : Gather Cluster facts and set is_containerized if needed] *** skipping: [10.8.169.22] skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_facts : Set repoquery command] ********************************* skipping: [10.8.169.210] skipping: [10.8.169.215] skipping: [10.8.169.22] TASK [openshift_etcd_facts : openshift_facts] ********************************** skipping: [10.8.169.22] skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_facts : Detecting Operating System] **************************** skipping: [10.8.169.22] skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_facts : set_fact] ********************************************** skipping: [10.8.169.22] skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_facts : set_fact] ********************************************** skipping: [10.8.169.215] skipping: [10.8.169.210] skipping: [10.8.169.22] TASK [openshift_facts : Ensure various deps are installed] ********************* skipping: [10.8.169.22] => (item=[]) skipping: [10.8.169.210] => (item=[]) skipping: [10.8.169.215] => (item=[]) TASK [openshift_facts : Gather Cluster facts and set is_containerized if needed] *** skipping: [10.8.169.22] skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_facts : Set repoquery command] ********************************* skipping: [10.8.169.22] skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_repos : assert] ************************************************ skipping: [10.8.169.210] skipping: [10.8.169.22] skipping: [10.8.169.215] TASK [openshift_repos : Ensure libselinux-python is installed] ***************** skipping: [10.8.169.22] skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_repos : Create any additional repos that are defined] ********** skipping: [10.8.169.22] skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_repos : Remove the additional repos if no longer defined] ****** skipping: [10.8.169.22] skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_repos : Configure origin gpg keys if needed] ******************* skipping: [10.8.169.215] skipping: [10.8.169.22] skipping: [10.8.169.210] TASK [openshift_repos : Configure origin yum repositories RHEL/CentOS] ********* skipping: [10.8.169.215] skipping: [10.8.169.22] skipping: [10.8.169.210] TASK [etcd_client_certificates : Ensure CA certificate exists on etcd_ca_host] * skipping: [10.8.169.210] TASK [etcd_client_certificates : fail] ***************************************** skipping: [10.8.169.210] TASK [etcd_client_certificates : Check status of external etcd certificatees] ** skipping: [10.8.169.22] => (item=flannel.etcd-ca.crt) skipping: [10.8.169.22] => (item=flannel.etcd-client.key) skipping: [10.8.169.22] => (item=flannel.etcd-client.crt) skipping: [10.8.169.215] => (item=flannel.etcd-ca.crt) skipping: [10.8.169.210] => (item=flannel.etcd-ca.crt) skipping: [10.8.169.215] => (item=flannel.etcd-client.key) skipping: [10.8.169.215] => (item=flannel.etcd-client.crt) skipping: [10.8.169.210] => (item=flannel.etcd-client.key) skipping: [10.8.169.210] => (item=flannel.etcd-client.crt) TASK [etcd_client_certificates : set_fact] ************************************* skipping: [10.8.169.210] skipping: [10.8.169.22] skipping: [10.8.169.215] TASK [etcd_client_certificates : Ensure generated_certs directory present] ***** skipping: [10.8.169.210] skipping: [10.8.169.22] skipping: [10.8.169.215] TASK [etcd_client_certificates : Create the client csr] ************************ skipping: [10.8.169.22] skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [etcd_client_certificates : Sign and create the client crt] *************** skipping: [10.8.169.215] skipping: [10.8.169.22] skipping: [10.8.169.210] TASK [etcd_client_certificates : file] ***************************************** skipping: [10.8.169.22] skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [etcd_client_certificates : Create local temp directory for syncing certs] skipping: [10.8.169.210] skipping: [10.8.169.215] skipping: [10.8.169.22] TASK [etcd_client_certificates : Create a tarball of the etcd certs] *********** skipping: [10.8.169.22] skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [etcd_client_certificates : Retrieve the etcd cert tarballs] ************** skipping: [10.8.169.210] skipping: [10.8.169.215] skipping: [10.8.169.22] TASK [etcd_client_certificates : Ensure certificate directory exists] ********** skipping: [10.8.169.215] skipping: [10.8.169.210] skipping: [10.8.169.22] TASK [etcd_client_certificates : Unarchive etcd cert tarballs] ***************** skipping: [10.8.169.210] skipping: [10.8.169.22] skipping: [10.8.169.215] TASK [etcd_client_certificates : file] ***************************************** skipping: [10.8.169.210] => (item=flannel.etcd-ca.crt) skipping: [10.8.169.210] => (item=flannel.etcd-client.key) skipping: [10.8.169.210] => (item=flannel.etcd-client.crt) skipping: [10.8.169.215] => (item=flannel.etcd-ca.crt) skipping: [10.8.169.215] => (item=flannel.etcd-client.key) skipping: [10.8.169.215] => (item=flannel.etcd-client.crt) skipping: [10.8.169.22] => (item=flannel.etcd-ca.crt) skipping: [10.8.169.22] => (item=flannel.etcd-client.key) skipping: [10.8.169.22] => (item=flannel.etcd-client.crt) TASK [etcd_client_certificates : Delete temporary directory] ******************* skipping: [10.8.169.215] skipping: [10.8.169.22] skipping: [10.8.169.210] TASK [flannel : Install flannel] *********************************************** skipping: [10.8.169.210] skipping: [10.8.169.215] skipping: [10.8.169.22] TASK [flannel : Set flannel etcd options] ************************************** skipping: [10.8.169.215] => (item={u'regexp': u'^(FLANNEL_ETCD_PREFIX=)', u'line': u'\\1/openshift.com/network'}) skipping: [10.8.169.215] => (item={u'regexp': u'^(FLANNEL_ETCD_KEY=)', u'line': u'\\1/openshift.com/network'}) skipping: [10.8.169.215] => (item={u'regexp': u'^(FLANNEL_ETCD_ENDPOINTS=)', u'line': u'\\1https://172.16.132.155:4001'}) skipping: [10.8.169.22] => (item={u'regexp': u'^(FLANNEL_ETCD_PREFIX=)', u'line': u'\\1/openshift.com/network'}) skipping: [10.8.169.22] => (item={u'regexp': u'^(FLANNEL_ETCD_KEY=)', u'line': u'\\1/openshift.com/network'}) skipping: [10.8.169.22] => (item={u'regexp': u'^(FLANNEL_ETCD_ENDPOINTS=)', u'line': u'\\1https://172.16.132.155:4001'}) skipping: [10.8.169.22] => (item={u'regexp': u'^(FLANNEL_ETCD=)', u'line': u'\\1https://172.16.132.155:4001'}) skipping: [10.8.169.215] => (item={u'regexp': u'^(FLANNEL_ETCD=)', u'line': u'\\1https://172.16.132.155:4001'}) skipping: [10.8.169.210] => (item={u'regexp': u'^(FLANNEL_ETCD_PREFIX=)', u'line': u'\\1/openshift.com/network'}) skipping: [10.8.169.210] => (item={u'regexp': u'^(FLANNEL_ETCD_KEY=)', u'line': u'\\1/openshift.com/network'}) skipping: [10.8.169.210] => (item={u'regexp': u'^(FLANNEL_ETCD_ENDPOINTS=)', u'line': u'\\1https://172.16.132.155:4001'}) skipping: [10.8.169.210] => (item={u'regexp': u'^(FLANNEL_ETCD=)', u'line': u'\\1https://172.16.132.155:4001'}) TASK [flannel : Set flannel options] ******************************************* skipping: [10.8.169.22] skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [flannel : Enable flanneld] *********************************************** skipping: [10.8.169.210] skipping: [10.8.169.215] skipping: [10.8.169.22] TASK [flannel : Remove docker bridge ip] *************************************** skipping: [10.8.169.215] skipping: [10.8.169.210] skipping: [10.8.169.22] TASK [nuage_ca : Install openssl] ********************************************** skipping: [10.8.169.215] skipping: [10.8.169.210] skipping: [10.8.169.22] TASK [nuage_ca : Create CA directory] ****************************************** skipping: [10.8.169.210] TASK [nuage_ca : Create certificate directory] ********************************* skipping: [10.8.169.210] TASK [nuage_ca : Check if the CA key already exists] *************************** skipping: [10.8.169.210] skipping: [10.8.169.22] skipping: [10.8.169.215] TASK [nuage_ca : Create CA key] ************************************************ skipping: [10.8.169.210] TASK [nuage_ca : Check if the CA crt already exists] *************************** skipping: [10.8.169.22] skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [nuage_ca : Create CA crt] ************************************************ skipping: [10.8.169.210] TASK [nuage_ca : Create the serial file] *************************************** skipping: [10.8.169.210] TASK [nuage_ca : Copy SSL config file] ***************************************** skipping: [10.8.169.210] TASK [openshift_facts : Detecting Operating System] **************************** skipping: [10.8.169.22] skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [openshift_facts : set_fact] ********************************************** skipping: [10.8.169.22] skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_facts : set_fact] ********************************************** skipping: [10.8.169.210] skipping: [10.8.169.22] skipping: [10.8.169.215] TASK [openshift_facts : Ensure various deps are installed] ********************* skipping: [10.8.169.22] => (item=[]) skipping: [10.8.169.210] => (item=[]) skipping: [10.8.169.215] => (item=[]) TASK [openshift_facts : Gather Cluster facts and set is_containerized if needed] *** skipping: [10.8.169.22] skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [openshift_facts : Set repoquery command] ********************************* skipping: [10.8.169.22] skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [os_firewall : Assert - Do not use firewalld on Atomic Host] ************** skipping: [10.8.169.210] skipping: [10.8.169.215] skipping: [10.8.169.22] TASK [os_firewall : Install firewalld packages] ******************************** skipping: [10.8.169.210] skipping: [10.8.169.22] skipping: [10.8.169.215] TASK [os_firewall : Ensure iptables services are not enabled] ****************** skipping: [10.8.169.210] => (item=ip6tables) skipping: [10.8.169.210] => (item=iptables) skipping: [10.8.169.22] => (item=ip6tables) skipping: [10.8.169.22] => (item=iptables) skipping: [10.8.169.215] => (item=ip6tables) skipping: [10.8.169.215] => (item=iptables) TASK [os_firewall : Start and enable firewalld service] ************************ skipping: [10.8.169.210] skipping: [10.8.169.22] skipping: [10.8.169.215] TASK [os_firewall : need to pause here, otherwise the firewalld service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.210] TASK [os_firewall : Add firewalld allow rules] ********************************* skipping: [10.8.169.210] => (item={u'port': u'9443/tcp', u'service': u'nuage-monitor'}) skipping: [10.8.169.210] => (item={u'port': u'4789/udp', u'service': u'vxlan'}) skipping: [10.8.169.22] => (item={u'port': u'9443/tcp', u'service': u'nuage-monitor'}) skipping: [10.8.169.22] => (item={u'port': u'4789/udp', u'service': u'vxlan'}) skipping: [10.8.169.215] => (item={u'port': u'9443/tcp', u'service': u'nuage-monitor'}) skipping: [10.8.169.215] => (item={u'port': u'4789/udp', u'service': u'vxlan'}) TASK [os_firewall : Remove firewalld allow rules] ****************************** TASK [os_firewall : Ensure firewalld service is not enabled] ******************* skipping: [10.8.169.22] skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [os_firewall : Install iptables packages] ********************************* skipping: [10.8.169.215] => (item=[]) skipping: [10.8.169.22] => (item=[]) skipping: [10.8.169.210] => (item=[]) TASK [os_firewall : Start and enable iptables service] ************************* skipping: [10.8.169.210] skipping: [10.8.169.22] skipping: [10.8.169.215] TASK [os_firewall : need to pause here, otherwise the iptables service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.210] TASK [os_firewall : Add iptables allow rules] ********************************** skipping: [10.8.169.215] => (item={u'port': u'9443/tcp', u'service': u'nuage-monitor'}) skipping: [10.8.169.215] => (item={u'port': u'4789/udp', u'service': u'vxlan'}) skipping: [10.8.169.210] => (item={u'port': u'9443/tcp', u'service': u'nuage-monitor'}) skipping: [10.8.169.210] => (item={u'port': u'4789/udp', u'service': u'vxlan'}) skipping: [10.8.169.22] => (item={u'port': u'9443/tcp', u'service': u'nuage-monitor'}) skipping: [10.8.169.22] => (item={u'port': u'4789/udp', u'service': u'vxlan'}) TASK [os_firewall : Remove iptables rules] ************************************* TASK [nuage_node : Install Nuage VRS] ****************************************** skipping: [10.8.169.215] skipping: [10.8.169.210] skipping: [10.8.169.22] TASK [nuage_node : Set the uplink interface] *********************************** skipping: [10.8.169.210] skipping: [10.8.169.22] skipping: [10.8.169.215] TASK [nuage_node : Set the Active Controller] ********************************** skipping: [10.8.169.22] skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [nuage_node : Set the Standby Controller] ********************************* skipping: [10.8.169.22] skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [nuage_node : Install plugin rpm] ***************************************** skipping: [10.8.169.22] skipping: [10.8.169.210] skipping: [10.8.169.215] TASK [nuage_node : Copy the certificates and keys] ***************************** skipping: [10.8.169.210] => (item=nuage.kubeconfig) skipping: [10.8.169.210] => (item=nuage.key) skipping: [10.8.169.215] => (item=nuage.kubeconfig) skipping: [10.8.169.215] => (item=nuage.key) skipping: [10.8.169.215] => (item=nuage.crt) skipping: [10.8.169.215] => (item=ca.crt) skipping: [10.8.169.22] => (item=nuage.kubeconfig) skipping: [10.8.169.22] => (item=nuage.key) skipping: [10.8.169.22] => (item=nuage.crt) skipping: [10.8.169.210] => (item=nuage.crt) skipping: [10.8.169.210] => (item=ca.crt) skipping: [10.8.169.22] => (item=ca.crt) TASK [nuage_node : Create a directory to hold the certificates] **************** skipping: [10.8.169.210] skipping: [10.8.169.22] skipping: [10.8.169.215] TASK [nuage_node : Create the key] ********************************************* skipping: [10.8.169.210] skipping: [10.8.169.215] skipping: [10.8.169.22] TASK [nuage_node : Create the req file] **************************************** skipping: [10.8.169.210] skipping: [10.8.169.22] skipping: [10.8.169.215] TASK [nuage_node : Generate the crt file] ************************************** skipping: [10.8.169.22] skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [nuage_node : Remove the req file] **************************************** skipping: [10.8.169.215] skipping: [10.8.169.210] skipping: [10.8.169.22] TASK [nuage_node : Copy nuage CA crt] ****************************************** skipping: [10.8.169.22] skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [nuage_node : Archive the certificate dir] ******************************** skipping: [10.8.169.215] skipping: [10.8.169.210] skipping: [10.8.169.22] TASK [nuage_node : Create a temp directory for the certificates] *************** skipping: [10.8.169.22] skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [nuage_node : Download the certificates] ********************************** skipping: [10.8.169.215] skipping: [10.8.169.22] skipping: [10.8.169.210] TASK [nuage_node : Extract the certificates] *********************************** skipping: [10.8.169.210] skipping: [10.8.169.215] skipping: [10.8.169.22] TASK [nuage_node : Delete the certificates after copy] ************************* skipping: [10.8.169.22] skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [nuage_node : Delete the temp directory] ********************************** skipping: [10.8.169.215] skipping: [10.8.169.210] skipping: [10.8.169.22] TASK [nuage_node : Set the vsp-openshift.yaml] ********************************* skipping: [10.8.169.22] skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [nuage_node : IPtables | Get iptables rules] ****************************** [DEPRECATION WARNING]: always_run is deprecated. Use check_mode = no instead.. This feature will be removed in version 2.4. Deprecation warnings can be disabled by setting deprecation_warnings=False in ansible.cfg. [DEPRECATION WARNING]: always_run is deprecated. Use check_mode = no instead.. This feature will be removed in version 2.4. Deprecation warnings can be disabled by setting deprecation_warnings=False in ansible.cfg. [DEPRECATION WARNING]: always_run is deprecated. Use check_mode = no instead.. This feature will be removed in version 2.4. Deprecation warnings can be disabled by setting deprecation_warnings=False in ansible.cfg. skipping: [10.8.169.22] skipping: [10.8.169.215] skipping: [10.8.169.210] TASK [nuage_node : Allow traffic from overlay to underlay] ********************* skipping: [10.8.169.215] skipping: [10.8.169.210] skipping: [10.8.169.22] TASK [nuage_node : Allow traffic from underlay to overlay] ********************* skipping: [10.8.169.210] skipping: [10.8.169.22] skipping: [10.8.169.215] TASK [nickhammond.logrotate : nickhammond.logrotate | Install logrotate] ******* ok: [10.8.169.22] ok: [10.8.169.210] ok: [10.8.169.215] TASK [nickhammond.logrotate : nickhammond.logrotate | Setup logrotate.d scripts] *** TASK [openshift_manage_node : Create temp directory for kubeconfig] ************ ok: [10.8.169.210 -> 10.8.169.22] TASK [openshift_manage_node : set_fact] **************************************** ok: [10.8.169.210 -> 10.8.169.22] TASK [openshift_manage_node : Copy the admin client config(s)] ***************** ok: [10.8.169.210 -> 10.8.169.22] TASK [openshift_manage_node : Wait for master API to become available before proceeding] *** ok: [10.8.169.210 -> 10.8.169.22] TASK [openshift_manage_node : Wait for Node Registration] ********************** ok: [10.8.169.22 -> 10.8.169.22] ok: [10.8.169.210 -> 10.8.169.22] ok: [10.8.169.215 -> 10.8.169.22] TASK [openshift_manage_node : Set node schedulability] ************************* changed: [10.8.169.210 -> 10.8.169.22] changed: [10.8.169.215 -> 10.8.169.22] changed: [10.8.169.22 -> 10.8.169.22] TASK [openshift_manage_node : Label nodes] ************************************* changed: [10.8.169.215 -> 10.8.169.22] changed: [10.8.169.210 -> 10.8.169.22] changed: [10.8.169.22 -> 10.8.169.22] TASK [openshift_manage_node : Delete temp directory] *************************** ok: [10.8.169.210 -> 10.8.169.22] TASK [Create group for deployment type] **************************************** ok: [10.8.169.22] ok: [10.8.169.215] ok: [10.8.169.210] PLAY [Delete temporary directory on localhost] ********************************* TASK [file] ******************************************************************** ok: [localhost] PLAY [Create persistent volumes] *********************************************** TASK [setup] ******************************************************************* ok: [10.8.169.22] TASK [openshift_facts : Detecting Operating System] **************************** skipping: [10.8.169.22] TASK [openshift_facts : set_fact] ********************************************** skipping: [10.8.169.22] TASK [openshift_facts : set_fact] ********************************************** skipping: [10.8.169.22] TASK [openshift_facts : Ensure various deps are installed] ********************* skipping: [10.8.169.22] => (item=[]) TASK [openshift_facts : Gather Cluster facts and set is_containerized if needed] *** skipping: [10.8.169.22] TASK [openshift_facts : Set repoquery command] ********************************* skipping: [10.8.169.22] TASK [openshift_facts : Detecting Operating System] **************************** skipping: [10.8.169.22] TASK [openshift_facts : set_fact] ********************************************** skipping: [10.8.169.22] TASK [openshift_facts : set_fact] ********************************************** skipping: [10.8.169.22] TASK [openshift_facts : Ensure various deps are installed] ********************* skipping: [10.8.169.22] => (item=[]) TASK [openshift_facts : Gather Cluster facts and set is_containerized if needed] *** skipping: [10.8.169.22] TASK [openshift_facts : Set repoquery command] ********************************* skipping: [10.8.169.22] TASK [openshift_repos : assert] ************************************************ skipping: [10.8.169.22] TASK [openshift_repos : Ensure libselinux-python is installed] ***************** skipping: [10.8.169.22] TASK [openshift_repos : Create any additional repos that are defined] ********** skipping: [10.8.169.22] TASK [openshift_repos : Remove the additional repos if no longer defined] ****** skipping: [10.8.169.22] TASK [openshift_repos : Configure origin gpg keys if needed] ******************* skipping: [10.8.169.22] TASK [openshift_repos : Configure origin yum repositories RHEL/CentOS] ********* skipping: [10.8.169.22] TASK [openshift_facts : Detecting Operating System] **************************** skipping: [10.8.169.22] TASK [openshift_facts : set_fact] ********************************************** skipping: [10.8.169.22] TASK [openshift_facts : set_fact] ********************************************** skipping: [10.8.169.22] TASK [openshift_facts : Ensure various deps are installed] ********************* skipping: [10.8.169.22] => (item=[]) TASK [openshift_facts : Gather Cluster facts and set is_containerized if needed] *** skipping: [10.8.169.22] TASK [openshift_facts : Set repoquery command] ********************************* skipping: [10.8.169.22] TASK [openshift_repos : assert] ************************************************ skipping: [10.8.169.22] TASK [openshift_repos : Ensure libselinux-python is installed] ***************** skipping: [10.8.169.22] TASK [openshift_repos : Create any additional repos that are defined] ********** skipping: [10.8.169.22] TASK [openshift_repos : Remove the additional repos if no longer defined] ****** skipping: [10.8.169.22] TASK [openshift_repos : Configure origin gpg keys if needed] ******************* skipping: [10.8.169.22] TASK [openshift_repos : Configure origin yum repositories RHEL/CentOS] ********* skipping: [10.8.169.22] TASK [openshift_facts : Detecting Operating System] **************************** skipping: [10.8.169.22] TASK [openshift_facts : set_fact] ********************************************** skipping: [10.8.169.22] TASK [openshift_facts : set_fact] ********************************************** skipping: [10.8.169.22] TASK [openshift_facts : Ensure various deps are installed] ********************* skipping: [10.8.169.22] => (item=[]) TASK [openshift_facts : Gather Cluster facts and set is_containerized if needed] *** skipping: [10.8.169.22] TASK [openshift_facts : Set repoquery command] ********************************* skipping: [10.8.169.22] TASK [openshift_docker_facts : Set docker facts] ******************************* skipping: [10.8.169.22] => (item={u'local_facts': {u'blocked_registries': u'', u'additional_registries': u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888,registry.ops.openshift.com', u'log_driver': u'', u'disable_push_dockerhub': u'', u'hosted_registry_insecure': False, u'hosted_registry_network': u'172.30.0.0/16', u'log_options': u'', u'insecure_registries': u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888,registry.ops.openshift.com', u'options': u''}, u'role': u'docker'}) TASK [openshift_docker_facts : set_fact] *************************************** skipping: [10.8.169.22] TASK [openshift_docker_facts : set_fact] *************************************** skipping: [10.8.169.22] TASK [openshift_docker_facts : set_fact] *************************************** skipping: [10.8.169.22] TASK [openshift_facts : Detecting Operating System] **************************** skipping: [10.8.169.22] TASK [openshift_facts : set_fact] ********************************************** skipping: [10.8.169.22] TASK [openshift_facts : set_fact] ********************************************** skipping: [10.8.169.22] TASK [openshift_facts : Ensure various deps are installed] ********************* skipping: [10.8.169.22] => (item=[]) TASK [openshift_facts : Gather Cluster facts and set is_containerized if needed] *** skipping: [10.8.169.22] TASK [openshift_facts : Set repoquery command] ********************************* skipping: [10.8.169.22] TASK [os_firewall : Assert - Do not use firewalld on Atomic Host] ************** skipping: [10.8.169.22] TASK [os_firewall : Install firewalld packages] ******************************** skipping: [10.8.169.22] TASK [os_firewall : Ensure iptables services are not enabled] ****************** skipping: [10.8.169.22] => (item=ip6tables) skipping: [10.8.169.22] => (item=iptables) TASK [os_firewall : Start and enable firewalld service] ************************ skipping: [10.8.169.22] TASK [os_firewall : need to pause here, otherwise the firewalld service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.22] TASK [os_firewall : Add firewalld allow rules] ********************************* TASK [os_firewall : Remove firewalld allow rules] ****************************** TASK [os_firewall : Ensure firewalld service is not enabled] ******************* skipping: [10.8.169.22] TASK [os_firewall : Install iptables packages] ********************************* skipping: [10.8.169.22] => (item=[]) TASK [os_firewall : Start and enable iptables service] ************************* skipping: [10.8.169.22] TASK [os_firewall : need to pause here, otherwise the iptables service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.22] TASK [os_firewall : Add iptables allow rules] ********************************** TASK [os_firewall : Remove iptables rules] ************************************* TASK [docker : Get current installed Docker version] *************************** skipping: [10.8.169.22] TASK [docker : Error out if Docker pre-installed but too old] ****************** skipping: [10.8.169.22] TASK [docker : Error out if requested Docker is too old] *********************** skipping: [10.8.169.22] TASK [docker : Get latest available version of Docker] ************************* skipping: [10.8.169.22] TASK [docker : Fail if Docker version requested but downgrade is required] ***** skipping: [10.8.169.22] TASK [docker : Error out if attempting to upgrade Docker across the 1.10 boundary] *** skipping: [10.8.169.22] TASK [docker : Install Docker] ************************************************* skipping: [10.8.169.22] TASK [docker : Ensure docker.service.d directory exists] *********************** skipping: [10.8.169.22] TASK [docker : Configure Docker service unit file] ***************************** skipping: [10.8.169.22] TASK [docker : Getting current systemd-udevd exec command] ********************* skipping: [10.8.169.22] TASK [docker : Assure systemd-udevd.service.d directory exists] **************** skipping: [10.8.169.22] TASK [docker : Create systemd-udevd override file] ***************************** skipping: [10.8.169.22] TASK [docker : stat] *********************************************************** skipping: [10.8.169.22] TASK [docker : Set registry params] ******************************************** skipping: [10.8.169.22] => (item={u'reg_conf_var': u'INSECURE_REGISTRY', u'reg_flag': u'--insecure-registry', u'reg_fact_val': [u'registry.ops.openshift.com', u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888']}) skipping: [10.8.169.22] => (item={u'reg_conf_var': u'BLOCK_REGISTRY', u'reg_flag': u'--block-registry', u'reg_fact_val': u''}) skipping: [10.8.169.22] => (item={u'reg_conf_var': u'ADD_REGISTRY', u'reg_flag': u'--add-registry', u'reg_fact_val': [u'registry.ops.openshift.com', u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888', u'XXXXXXXXXXXXXXXXXXXXXXXXXXX']}) TASK [docker : Set Proxy Settings] ********************************************* skipping: [10.8.169.22] => (item={u'reg_conf_var': u'HTTP_PROXY', u'reg_fact_val': u''}) skipping: [10.8.169.22] => (item={u'reg_conf_var': u'NO_PROXY', u'reg_fact_val': u''}) skipping: [10.8.169.22] => (item={u'reg_conf_var': u'HTTPS_PROXY', u'reg_fact_val': u''}) TASK [docker : Set various Docker options] ************************************* skipping: [10.8.169.22] TASK [docker : Start the Docker service] *************************************** skipping: [10.8.169.22] TASK [docker : set_fact] ******************************************************* skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : Use openshift.common.version fact as version to configure if already installed] *** skipping: [10.8.169.22] TASK [openshift_version : Set rpm version to configure if openshift_pkg_version specified] *** skipping: [10.8.169.22] TASK [openshift_version : Gather common package version] *********************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : Set containerized version to configure if openshift_image_tag specified] *** skipping: [10.8.169.22] TASK [openshift_version : Set containerized version to configure if openshift_release specified] *** skipping: [10.8.169.22] TASK [openshift_version : Lookup latest containerized version if no version specified] *** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : Set precise containerized version to configure if openshift_release specified] *** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.22] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.22] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.22] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.22] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.22] TASK [openshift_version : debug] *********************************************** skipping: [10.8.169.22] TASK [openshift_version : debug] *********************************************** skipping: [10.8.169.22] TASK [openshift_version : debug] *********************************************** skipping: [10.8.169.22] TASK [openshift_common : fail] ************************************************* skipping: [10.8.169.22] TASK [openshift_common : fail] ************************************************* skipping: [10.8.169.22] TASK [openshift_common : fail] ************************************************* skipping: [10.8.169.22] TASK [openshift_common : fail] ************************************************* skipping: [10.8.169.22] TASK [openshift_common : Set common Cluster facts] ***************************** skipping: [10.8.169.22] TASK [openshift_common : Install the base package for versioning] ************** skipping: [10.8.169.22] TASK [openshift_common : Set version facts] ************************************ skipping: [10.8.169.22] TASK [openshift_common : set_fact] ********************************************* skipping: [10.8.169.22] TASK [openshift_common : Set hostname] ***************************************** skipping: [10.8.169.22] TASK [openshift_facts : Detecting Operating System] **************************** skipping: [10.8.169.22] TASK [openshift_facts : set_fact] ********************************************** skipping: [10.8.169.22] TASK [openshift_facts : set_fact] ********************************************** skipping: [10.8.169.22] TASK [openshift_facts : Ensure various deps are installed] ********************* skipping: [10.8.169.22] => (item=[]) TASK [openshift_facts : Gather Cluster facts and set is_containerized if needed] *** skipping: [10.8.169.22] TASK [openshift_facts : Set repoquery command] ********************************* skipping: [10.8.169.22] TASK [openshift_hosted_facts : set_fact] *************************************** skipping: [10.8.169.22] TASK [openshift_hosted_facts : set_fact] *************************************** skipping: [10.8.169.22] TASK [openshift_hosted_facts : Set hosted facts] ******************************* skipping: [10.8.169.22] TASK [openshift_persistent_volumes : Create temp directory for volume definitions] *** skipping: [10.8.169.22] TASK [openshift_persistent_volumes : Copy the admin client config(s)] ********** skipping: [10.8.169.22] TASK [openshift_persistent_volumes : Deploy PersistentVolume definitions] ****** skipping: [10.8.169.22] TASK [openshift_persistent_volumes : Create PersistentVolumes] ***************** skipping: [10.8.169.22] TASK [openshift_persistent_volumes : Deploy PersistentVolumeClaim definitions] * skipping: [10.8.169.22] TASK [openshift_persistent_volumes : Create PersistentVolumeClaims] ************ skipping: [10.8.169.22] TASK [openshift_persistent_volumes : Delete temp directory] ******************** skipping: [10.8.169.22] PLAY [Create Hosted Resources] ************************************************* TASK [setup] ******************************************************************* ok: [10.8.169.22] TASK [set_fact] **************************************************************** ok: [10.8.169.22] TASK [set_fact] **************************************************************** ok: [10.8.169.22] TASK [openshift_facts : Detecting Operating System] **************************** ok: [10.8.169.22] TASK [openshift_facts : set_fact] ********************************************** ok: [10.8.169.22] TASK [openshift_facts : set_fact] ********************************************** ok: [10.8.169.22] TASK [openshift_facts : Ensure various deps are installed] ********************* ok: [10.8.169.22] => (item=[u'iproute', u'python-dbus', u'python-six', u'PyYAML', u'yum-utils']) TASK [openshift_facts : Gather Cluster facts and set is_containerized if needed] *** changed: [10.8.169.22] TASK [openshift_facts : Set repoquery command] ********************************* ok: [10.8.169.22] TASK [openshift_repos : assert] ************************************************ skipping: [10.8.169.22] TASK [openshift_repos : Ensure libselinux-python is installed] ***************** skipping: [10.8.169.22] TASK [openshift_repos : Create any additional repos that are defined] ********** skipping: [10.8.169.22] TASK [openshift_repos : Remove the additional repos if no longer defined] ****** skipping: [10.8.169.22] TASK [openshift_repos : Configure origin gpg keys if needed] ******************* skipping: [10.8.169.22] TASK [openshift_repos : Configure origin yum repositories RHEL/CentOS] ********* skipping: [10.8.169.22] TASK [openshift_docker_facts : Set docker facts] ******************************* ok: [10.8.169.22] => (item={u'local_facts': {u'blocked_registries': u'', u'additional_registries': u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888,registry.ops.openshift.com', u'log_driver': u'', u'disable_push_dockerhub': u'', u'hosted_registry_insecure': False, u'hosted_registry_network': u'172.30.0.0/16', u'log_options': u'', u'insecure_registries': u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888,registry.ops.openshift.com', u'options': u''}, u'role': u'docker'}) TASK [openshift_docker_facts : set_fact] *************************************** ok: [10.8.169.22] TASK [openshift_docker_facts : set_fact] *************************************** skipping: [10.8.169.22] TASK [openshift_docker_facts : set_fact] *************************************** ok: [10.8.169.22] TASK [os_firewall : Assert - Do not use firewalld on Atomic Host] ************** skipping: [10.8.169.22] TASK [os_firewall : Install firewalld packages] ******************************** skipping: [10.8.169.22] TASK [os_firewall : Ensure iptables services are not enabled] ****************** ok: [10.8.169.22] => (item=iptables) ok: [10.8.169.22] => (item=ip6tables) TASK [os_firewall : Start and enable firewalld service] ************************ ok: [10.8.169.22] TASK [os_firewall : need to pause here, otherwise the firewalld service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.22] TASK [os_firewall : Add firewalld allow rules] ********************************* TASK [os_firewall : Remove firewalld allow rules] ****************************** TASK [os_firewall : Ensure firewalld service is not enabled] ******************* skipping: [10.8.169.22] TASK [os_firewall : Install iptables packages] ********************************* skipping: [10.8.169.22] => (item=[]) TASK [os_firewall : Start and enable iptables service] ************************* skipping: [10.8.169.22] TASK [os_firewall : need to pause here, otherwise the iptables service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.22] TASK [os_firewall : Add iptables allow rules] ********************************** TASK [os_firewall : Remove iptables rules] ************************************* TASK [docker : Get current installed Docker version] *************************** ok: [10.8.169.22] TASK [docker : Error out if Docker pre-installed but too old] ****************** skipping: [10.8.169.22] TASK [docker : Error out if requested Docker is too old] *********************** skipping: [10.8.169.22] TASK [docker : Get latest available version of Docker] ************************* skipping: [10.8.169.22] TASK [docker : Fail if Docker version requested but downgrade is required] ***** skipping: [10.8.169.22] TASK [docker : Error out if attempting to upgrade Docker across the 1.10 boundary] *** skipping: [10.8.169.22] TASK [docker : Install Docker] ************************************************* ok: [10.8.169.22] TASK [docker : Ensure docker.service.d directory exists] *********************** skipping: [10.8.169.22] TASK [docker : Configure Docker service unit file] ***************************** skipping: [10.8.169.22] TASK [docker : Getting current systemd-udevd exec command] ********************* skipping: [10.8.169.22] TASK [docker : Assure systemd-udevd.service.d directory exists] **************** skipping: [10.8.169.22] TASK [docker : Create systemd-udevd override file] ***************************** skipping: [10.8.169.22] TASK [docker : stat] *********************************************************** ok: [10.8.169.22] TASK [docker : Set registry params] ******************************************** skipping: [10.8.169.22] => (item={u'reg_conf_var': u'BLOCK_REGISTRY', u'reg_flag': u'--block-registry', u'reg_fact_val': u''}) ok: [10.8.169.22] => (item={u'reg_conf_var': u'ADD_REGISTRY', u'reg_flag': u'--add-registry', u'reg_fact_val': [u'registry.ops.openshift.com', u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888', u'XXXXXXXXXXXXXXXXXXXXXXXXXXX']}) ok: [10.8.169.22] => (item={u'reg_conf_var': u'INSECURE_REGISTRY', u'reg_flag': u'--insecure-registry', u'reg_fact_val': [u'registry.ops.openshift.com', u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888']}) TASK [docker : Set Proxy Settings] ********************************************* ok: [10.8.169.22] => (item={u'reg_conf_var': u'HTTP_PROXY', u'reg_fact_val': u''}) ok: [10.8.169.22] => (item={u'reg_conf_var': u'HTTPS_PROXY', u'reg_fact_val': u''}) ok: [10.8.169.22] => (item={u'reg_conf_var': u'NO_PROXY', u'reg_fact_val': u''}) TASK [docker : Set various Docker options] ************************************* ok: [10.8.169.22] TASK [docker : Start the Docker service] *************************************** ok: [10.8.169.22] TASK [docker : set_fact] ******************************************************* ok: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** ok: [10.8.169.22] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** ok: [10.8.169.22] TASK [openshift_version : Use openshift.common.version fact as version to configure if already installed] *** skipping: [10.8.169.22] TASK [openshift_version : Set rpm version to configure if openshift_pkg_version specified] *** skipping: [10.8.169.22] TASK [openshift_version : Gather common package version] *********************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : Set containerized version to configure if openshift_image_tag specified] *** skipping: [10.8.169.22] TASK [openshift_version : Set containerized version to configure if openshift_release specified] *** skipping: [10.8.169.22] TASK [openshift_version : Lookup latest containerized version if no version specified] *** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : Set precise containerized version to configure if openshift_release specified] *** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** ok: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.22] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.22] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.22] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.22] TASK [openshift_version : fail] ************************************************ skipping: [10.8.169.22] TASK [openshift_version : debug] *********************************************** ok: [10.8.169.22] => { "openshift_release": "VARIABLE IS NOT DEFINED!" } TASK [openshift_version : debug] *********************************************** ok: [10.8.169.22] => { "openshift_image_tag": "v3.3" } TASK [openshift_version : debug] *********************************************** ok: [10.8.169.22] => { "openshift_pkg_version": "-3.3.1.9" } TASK [os_firewall : Assert - Do not use firewalld on Atomic Host] ************** skipping: [10.8.169.22] TASK [os_firewall : Install firewalld packages] ******************************** skipping: [10.8.169.22] TASK [os_firewall : Ensure iptables services are not enabled] ****************** ok: [10.8.169.22] => (item=iptables) ok: [10.8.169.22] => (item=ip6tables) TASK [os_firewall : Start and enable firewalld service] ************************ ok: [10.8.169.22] TASK [os_firewall : need to pause here, otherwise the firewalld service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.22] TASK [os_firewall : Add firewalld allow rules] ********************************* TASK [os_firewall : Remove firewalld allow rules] ****************************** TASK [os_firewall : Ensure firewalld service is not enabled] ******************* skipping: [10.8.169.22] TASK [os_firewall : Install iptables packages] ********************************* skipping: [10.8.169.22] => (item=[]) TASK [os_firewall : Start and enable iptables service] ************************* skipping: [10.8.169.22] TASK [os_firewall : need to pause here, otherwise the iptables service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.22] TASK [os_firewall : Add iptables allow rules] ********************************** TASK [os_firewall : Remove iptables rules] ************************************* TASK [docker : Get current installed Docker version] *************************** ok: [10.8.169.22] TASK [docker : Error out if Docker pre-installed but too old] ****************** skipping: [10.8.169.22] TASK [docker : Error out if requested Docker is too old] *********************** skipping: [10.8.169.22] TASK [docker : Get latest available version of Docker] ************************* skipping: [10.8.169.22] TASK [docker : Fail if Docker version requested but downgrade is required] ***** skipping: [10.8.169.22] TASK [docker : Error out if attempting to upgrade Docker across the 1.10 boundary] *** skipping: [10.8.169.22] TASK [docker : Install Docker] ************************************************* ok: [10.8.169.22] TASK [docker : Ensure docker.service.d directory exists] *********************** skipping: [10.8.169.22] TASK [docker : Configure Docker service unit file] ***************************** skipping: [10.8.169.22] TASK [docker : Getting current systemd-udevd exec command] ********************* skipping: [10.8.169.22] TASK [docker : Assure systemd-udevd.service.d directory exists] **************** skipping: [10.8.169.22] TASK [docker : Create systemd-udevd override file] ***************************** skipping: [10.8.169.22] TASK [docker : stat] *********************************************************** ok: [10.8.169.22] TASK [docker : Set registry params] ******************************************** ok: [10.8.169.22] => (item={u'reg_conf_var': u'ADD_REGISTRY', u'reg_flag': u'--add-registry', u'reg_fact_val': [u'registry.ops.openshift.com', u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888', u'XXXXXXXXXXXXXXXXXXXXXXXXXXX']}) skipping: [10.8.169.22] => (item={u'reg_conf_var': u'BLOCK_REGISTRY', u'reg_flag': u'--block-registry', u'reg_fact_val': u''}) ok: [10.8.169.22] => (item={u'reg_conf_var': u'INSECURE_REGISTRY', u'reg_flag': u'--insecure-registry', u'reg_fact_val': [u'registry.ops.openshift.com', u'brew-pulp-docker01.web.prod.ext.phx2.redhat.com:8888']}) TASK [docker : Set Proxy Settings] ********************************************* ok: [10.8.169.22] => (item={u'reg_conf_var': u'HTTP_PROXY', u'reg_fact_val': u''}) ok: [10.8.169.22] => (item={u'reg_conf_var': u'HTTPS_PROXY', u'reg_fact_val': u''}) ok: [10.8.169.22] => (item={u'reg_conf_var': u'NO_PROXY', u'reg_fact_val': u''}) TASK [docker : Set various Docker options] ************************************* ok: [10.8.169.22] TASK [docker : Start the Docker service] *************************************** ok: [10.8.169.22] TASK [docker : set_fact] ******************************************************* ok: [10.8.169.22] TASK [openshift_repos : assert] ************************************************ skipping: [10.8.169.22] TASK [openshift_repos : Ensure libselinux-python is installed] ***************** skipping: [10.8.169.22] TASK [openshift_repos : Create any additional repos that are defined] ********** skipping: [10.8.169.22] TASK [openshift_repos : Remove the additional repos if no longer defined] ****** skipping: [10.8.169.22] TASK [openshift_repos : Configure origin gpg keys if needed] ******************* skipping: [10.8.169.22] TASK [openshift_repos : Configure origin yum repositories RHEL/CentOS] ********* skipping: [10.8.169.22] TASK [openshift_repos : assert] ************************************************ skipping: [10.8.169.22] TASK [openshift_repos : Ensure libselinux-python is installed] ***************** skipping: [10.8.169.22] TASK [openshift_repos : Create any additional repos that are defined] ********** skipping: [10.8.169.22] TASK [openshift_repos : Remove the additional repos if no longer defined] ****** skipping: [10.8.169.22] TASK [openshift_repos : Configure origin gpg keys if needed] ******************* skipping: [10.8.169.22] TASK [openshift_repos : Configure origin yum repositories RHEL/CentOS] ********* skipping: [10.8.169.22] TASK [os_firewall : Assert - Do not use firewalld on Atomic Host] ************** skipping: [10.8.169.22] TASK [os_firewall : Install firewalld packages] ******************************** skipping: [10.8.169.22] TASK [os_firewall : Ensure iptables services are not enabled] ****************** ok: [10.8.169.22] => (item=iptables) ok: [10.8.169.22] => (item=ip6tables) TASK [os_firewall : Start and enable firewalld service] ************************ ok: [10.8.169.22] TASK [os_firewall : need to pause here, otherwise the firewalld service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.22] TASK [os_firewall : Add firewalld allow rules] ********************************* TASK [os_firewall : Remove firewalld allow rules] ****************************** TASK [os_firewall : Ensure firewalld service is not enabled] ******************* skipping: [10.8.169.22] TASK [os_firewall : Install iptables packages] ********************************* skipping: [10.8.169.22] => (item=[]) TASK [os_firewall : Start and enable iptables service] ************************* skipping: [10.8.169.22] TASK [os_firewall : need to pause here, otherwise the iptables service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.22] TASK [os_firewall : Add iptables allow rules] ********************************** TASK [os_firewall : Remove iptables rules] ************************************* TASK [openshift_common : fail] ************************************************* skipping: [10.8.169.22] TASK [openshift_common : fail] ************************************************* skipping: [10.8.169.22] TASK [openshift_common : fail] ************************************************* skipping: [10.8.169.22] TASK [openshift_common : fail] ************************************************* skipping: [10.8.169.22] TASK [openshift_common : Set common Cluster facts] ***************************** ok: [10.8.169.22] TASK [openshift_common : Install the base package for versioning] ************** skipping: [10.8.169.22] TASK [openshift_common : Set version facts] ************************************ ok: [10.8.169.22] TASK [openshift_common : set_fact] ********************************************* ok: [10.8.169.22] TASK [openshift_common : Set hostname] ***************************************** changed: [10.8.169.22] TASK [openshift_cli_facts : openshift_facts] *********************************** ok: [10.8.169.22] TASK [openshift_cli : Install clients] ***************************************** skipping: [10.8.169.22] TASK [openshift_cli : Pull CLI Image] ****************************************** ok: [10.8.169.22] TASK [openshift_cli : Copy client binaries/symlinks out of CLI image for use on the host] *** changed: [10.8.169.22] TASK [openshift_cli : Reload facts to pick up installed OpenShift version] ***** ok: [10.8.169.22] TASK [openshift_cli : Install bash completion for oc tools] ******************** skipping: [10.8.169.22] TASK [openshift_hosted_facts : set_fact] *************************************** skipping: [10.8.169.22] TASK [openshift_hosted_facts : set_fact] *************************************** skipping: [10.8.169.22] TASK [openshift_hosted_facts : Set hosted facts] ******************************* changed: [10.8.169.22] TASK [openshift_projects : Create temp directory for kubeconfig] *************** ok: [10.8.169.22] TASK [openshift_projects : Copy the admin client config(s)] ******************** ok: [10.8.169.22] TASK [openshift_projects : Determine if projects exist] ************************ ok: [10.8.169.22] => (item={'key': 'default', 'value': {'default_node_selector': ''}}) ok: [10.8.169.22] => (item={'key': 'logging', 'value': {'default_node_selector': ''}}) ok: [10.8.169.22] => (item={'key': u'openshift-infra', 'value': {'default_node_selector': u''}}) TASK [openshift_projects : Create projects] ************************************ skipping: [10.8.169.22] => (item={'_ansible_parsed': True, u'cmd': [u'/usr/local/bin/oc', u'--config=/tmp/openshift-ansible-jgYBWG/admin.kubeconfig', u'get', u'projects', u'default', u'-o', u'json'], u'end': u'2017-01-10 18:30:08.472093', '_ansible_no_log': False, u'stdout': u'{\n "kind": "Project",\n "apiVersion": "v1",\n "metadata": {\n "name": "default",\n "selfLink": "/oapi/v1/projects/default",\n "uid": "19945ebe-d78b-11e6-84d4-fa163e61d47e",\n "resourceVersion": "587",\n "creationTimestamp": "2017-01-10T23:18:31Z",\n "annotations": {\n "openshift.io/sa.initialized-roles": "true",\n "openshift.io/sa.scc.mcs": "s0:c1,c0",\n "openshift.io/sa.scc.supplemental-groups": "1000000000/10000",\n "openshift.io/sa.scc.uid-range": "1000000000/10000"\n }\n },\n "spec": {\n "finalizers": [\n "kubernetes",\n "openshift.io/origin"\n ]\n },\n "status": {\n "phase": "Active"\n }\n}', '_ansible_item_result': True, u'changed': False, 'item': {'value': {'default_node_selector': u''}, 'key': u'default'}, u'delta': u'0:00:00.108825', u'stderr': u'', u'rc': 0, 'invocation': {'module_name': u'command', u'module_args': {u'creates': None, u'executable': None, u'_uses_shell': False, u'_raw_params': u'/usr/local/bin/oc --config=/tmp/openshift-ansible-jgYBWG/admin.kubeconfig get projects default -o json', u'removes': None, u'warn': True, u'chdir': None}}, 'stdout_lines': [u'{', u' "kind": "Project",', u' "apiVersion": "v1",', u' "metadata": {', u' "name": "default",', u' "selfLink": "/oapi/v1/projects/default",', u' "uid": "19945ebe-d78b-11e6-84d4-fa163e61d47e",', u' "resourceVersion": "587",', u' "creationTimestamp": "2017-01-10T23:18:31Z",', u' "annotations": {', u' "openshift.io/sa.initialized-roles": "true",', u' "openshift.io/sa.scc.mcs": "s0:c1,c0",', u' "openshift.io/sa.scc.supplemental-groups": "1000000000/10000",', u' "openshift.io/sa.scc.uid-range": "1000000000/10000"', u' }', u' },', u' "spec": {', u' "finalizers": [', u' "kubernetes",', u' "openshift.io/origin"', u' ]', u' },', u' "status": {', u' "phase": "Active"', u' }', u'}'], 'failed_when_result': False, u'start': u'2017-01-10 18:30:08.363268', u'warnings': [], 'failed': False}) skipping: [10.8.169.22] => (item={'_ansible_parsed': True, u'cmd': [u'/usr/local/bin/oc', u'--config=/tmp/openshift-ansible-jgYBWG/admin.kubeconfig', u'get', u'projects', u'openshift-infra', u'-o', u'json'], u'end': u'2017-01-10 18:30:09.368170', '_ansible_no_log': False, u'stdout': u'{\n "kind": "Project",\n "apiVersion": "v1",\n "metadata": {\n "name": "openshift-infra",\n "selfLink": "/oapi/v1/projects/openshift-infra",\n "uid": "1b0b60da-d78b-11e6-84d4-fa163e61d47e",\n "resourceVersion": "590",\n "creationTimestamp": "2017-01-10T23:18:33Z",\n "annotations": {\n "openshift.io/sa.initialized-roles": "true",\n "openshift.io/sa.scc.mcs": "s0:c5,c0",\n "openshift.io/sa.scc.supplemental-groups": "1000020000/10000",\n "openshift.io/sa.scc.uid-range": "1000020000/10000"\n }\n },\n "spec": {\n "finalizers": [\n "kubernetes",\n "openshift.io/origin"\n ]\n },\n "status": {\n "phase": "Active"\n }\n}', '_ansible_item_result': True, u'changed': False, 'item': {'value': {'default_node_selector': u''}, 'key': u'openshift-infra'}, u'delta': u'0:00:00.103067', u'stderr': u'', u'rc': 0, 'invocation': {'module_name': u'command', u'module_args': {u'creates': None, u'executable': None, u'_uses_shell': False, u'_raw_params': u'/usr/local/bin/oc --config=/tmp/openshift-ansible-jgYBWG/admin.kubeconfig get projects openshift-infra -o json', u'removes': None, u'warn': True, u'chdir': None}}, 'stdout_lines': [u'{', u' "kind": "Project",', u' "apiVersion": "v1",', u' "metadata": {', u' "name": "openshift-infra",', u' "selfLink": "/oapi/v1/projects/openshift-infra",', u' "uid": "1b0b60da-d78b-11e6-84d4-fa163e61d47e",', u' "resourceVersion": "590",', u' "creationTimestamp": "2017-01-10T23:18:33Z",', u' "annotations": {', u' "openshift.io/sa.initialized-roles": "true",', u' "openshift.io/sa.scc.mcs": "s0:c5,c0",', u' "openshift.io/sa.scc.supplemental-groups": "1000020000/10000",', u' "openshift.io/sa.scc.uid-range": "1000020000/10000"', u' }', u' },', u' "spec": {', u' "finalizers": [', u' "kubernetes",', u' "openshift.io/origin"', u' ]', u' },', u' "status": {', u' "phase": "Active"', u' }', u'}'], 'failed_when_result': False, u'start': u'2017-01-10 18:30:09.265103', u'warnings': [], 'failed': False}) changed: [10.8.169.22] => (item={'_ansible_parsed': True, u'cmd': [u'/usr/local/bin/oc', u'--config=/tmp/openshift-ansible-jgYBWG/admin.kubeconfig', u'get', u'projects', u'logging', u'-o', u'json'], u'end': u'2017-01-10 18:30:08.872307', '_ansible_no_log': False, u'stdout': u'', '_ansible_item_result': True, u'changed': False, 'item': {'value': {'default_node_selector': u''}, 'key': u'logging'}, u'delta': u'0:00:00.104497', u'stderr': u'Error from server: namespaces "logging" not found', u'rc': 1, 'invocation': {'module_name': u'command', u'module_args': {u'creates': None, u'executable': None, u'_uses_shell': False, u'_raw_params': u'/usr/local/bin/oc --config=/tmp/openshift-ansible-jgYBWG/admin.kubeconfig get projects logging -o json', u'removes': None, u'warn': True, u'chdir': None}}, 'stdout_lines': [], 'failed_when_result': False, u'start': u'2017-01-10 18:30:08.767810', u'warnings': [], 'failed': False}) TASK [openshift_projects : Update project default node selector if necessary] ** skipping: [10.8.169.22] => (item={'_ansible_parsed': True, u'cmd': [u'/usr/local/bin/oc', u'--config=/tmp/openshift-ansible-jgYBWG/admin.kubeconfig', u'get', u'projects', u'logging', u'-o', u'json'], u'end': u'2017-01-10 18:30:08.872307', '_ansible_no_log': False, u'stdout': u'', '_ansible_item_result': True, u'changed': False, 'item': {'value': {'default_node_selector': u''}, 'key': u'logging'}, u'delta': u'0:00:00.104497', u'stderr': u'Error from server: namespaces "logging" not found', u'rc': 1, 'invocation': {'module_name': u'command', u'module_args': {u'creates': None, u'executable': None, u'_uses_shell': False, u'_raw_params': u'/usr/local/bin/oc --config=/tmp/openshift-ansible-jgYBWG/admin.kubeconfig get projects logging -o json', u'removes': None, u'warn': True, u'chdir': None}}, 'stdout_lines': [], 'failed_when_result': False, u'start': u'2017-01-10 18:30:08.767810', u'warnings': [], 'failed': False}) changed: [10.8.169.22] => (item={'_ansible_parsed': True, u'cmd': [u'/usr/local/bin/oc', u'--config=/tmp/openshift-ansible-jgYBWG/admin.kubeconfig', u'get', u'projects', u'default', u'-o', u'json'], u'end': u'2017-01-10 18:30:08.472093', '_ansible_no_log': False, u'stdout': u'{\n "kind": "Project",\n "apiVersion": "v1",\n "metadata": {\n "name": "default",\n "selfLink": "/oapi/v1/projects/default",\n "uid": "19945ebe-d78b-11e6-84d4-fa163e61d47e",\n "resourceVersion": "587",\n "creationTimestamp": "2017-01-10T23:18:31Z",\n "annotations": {\n "openshift.io/sa.initialized-roles": "true",\n "openshift.io/sa.scc.mcs": "s0:c1,c0",\n "openshift.io/sa.scc.supplemental-groups": "1000000000/10000",\n "openshift.io/sa.scc.uid-range": "1000000000/10000"\n }\n },\n "spec": {\n "finalizers": [\n "kubernetes",\n "openshift.io/origin"\n ]\n },\n "status": {\n "phase": "Active"\n }\n}', '_ansible_item_result': True, u'changed': False, 'item': {'value': {'default_node_selector': u''}, 'key': u'default'}, u'delta': u'0:00:00.108825', u'stderr': u'', u'rc': 0, 'invocation': {'module_name': u'command', u'module_args': {u'creates': None, u'executable': None, u'_uses_shell': False, u'_raw_params': u'/usr/local/bin/oc --config=/tmp/openshift-ansible-jgYBWG/admin.kubeconfig get projects default -o json', u'removes': None, u'warn': True, u'chdir': None}}, 'stdout_lines': [u'{', u' "kind": "Project",', u' "apiVersion": "v1",', u' "metadata": {', u' "name": "default",', u' "selfLink": "/oapi/v1/projects/default",', u' "uid": "19945ebe-d78b-11e6-84d4-fa163e61d47e",', u' "resourceVersion": "587",', u' "creationTimestamp": "2017-01-10T23:18:31Z",', u' "annotations": {', u' "openshift.io/sa.initialized-roles": "true",', u' "openshift.io/sa.scc.mcs": "s0:c1,c0",', u' "openshift.io/sa.scc.supplemental-groups": "1000000000/10000",', u' "openshift.io/sa.scc.uid-range": "1000000000/10000"', u' }', u' },', u' "spec": {', u' "finalizers": [', u' "kubernetes",', u' "openshift.io/origin"', u' ]', u' },', u' "status": {', u' "phase": "Active"', u' }', u'}'], 'failed_when_result': False, u'start': u'2017-01-10 18:30:08.363268', u'warnings': [], 'failed': False}) changed: [10.8.169.22] => (item={'_ansible_parsed': True, u'cmd': [u'/usr/local/bin/oc', u'--config=/tmp/openshift-ansible-jgYBWG/admin.kubeconfig', u'get', u'projects', u'openshift-infra', u'-o', u'json'], u'end': u'2017-01-10 18:30:09.368170', '_ansible_no_log': False, u'stdout': u'{\n "kind": "Project",\n "apiVersion": "v1",\n "metadata": {\n "name": "openshift-infra",\n "selfLink": "/oapi/v1/projects/openshift-infra",\n "uid": "1b0b60da-d78b-11e6-84d4-fa163e61d47e",\n "resourceVersion": "590",\n "creationTimestamp": "2017-01-10T23:18:33Z",\n "annotations": {\n "openshift.io/sa.initialized-roles": "true",\n "openshift.io/sa.scc.mcs": "s0:c5,c0",\n "openshift.io/sa.scc.supplemental-groups": "1000020000/10000",\n "openshift.io/sa.scc.uid-range": "1000020000/10000"\n }\n },\n "spec": {\n "finalizers": [\n "kubernetes",\n "openshift.io/origin"\n ]\n },\n "status": {\n "phase": "Active"\n }\n}', '_ansible_item_result': True, u'changed': False, 'item': {'value': {'default_node_selector': u''}, 'key': u'openshift-infra'}, u'delta': u'0:00:00.103067', u'stderr': u'', u'rc': 0, 'invocation': {'module_name': u'command', u'module_args': {u'creates': None, u'executable': None, u'_uses_shell': False, u'_raw_params': u'/usr/local/bin/oc --config=/tmp/openshift-ansible-jgYBWG/admin.kubeconfig get projects openshift-infra -o json', u'removes': None, u'warn': True, u'chdir': None}}, 'stdout_lines': [u'{', u' "kind": "Project",', u' "apiVersion": "v1",', u' "metadata": {', u' "name": "openshift-infra",', u' "selfLink": "/oapi/v1/projects/openshift-infra",', u' "uid": "1b0b60da-d78b-11e6-84d4-fa163e61d47e",', u' "resourceVersion": "590",', u' "creationTimestamp": "2017-01-10T23:18:33Z",', u' "annotations": {', u' "openshift.io/sa.initialized-roles": "true",', u' "openshift.io/sa.scc.mcs": "s0:c5,c0",', u' "openshift.io/sa.scc.supplemental-groups": "1000020000/10000",', u' "openshift.io/sa.scc.uid-range": "1000020000/10000"', u' }', u' },', u' "spec": {', u' "finalizers": [', u' "kubernetes",', u' "openshift.io/origin"', u' ]', u' },', u' "status": {', u' "phase": "Active"', u' }', u'}'], 'failed_when_result': False, u'start': u'2017-01-10 18:30:09.265103', u'warnings': [], 'failed': False}) TASK [openshift_projects : Delete temp directory] ****************************** ok: [10.8.169.22] TASK [openshift_serviceaccounts : test if service accounts exists] ************* ok: [10.8.169.22] => (item=router) TASK [openshift_serviceaccounts : create the service account] ****************** changed: [10.8.169.22] => (item=[u'router', {'_ansible_parsed': True, u'cmd': [u'/usr/local/bin/oc', u'get', u'sa', u'router', u'-n', u'default'], u'end': u'2017-01-10 18:30:12.582697', '_ansible_no_log': False, u'stdout': u'', '_ansible_item_result': True, u'rc': 1, 'item': u'router', u'delta': u'0:00:00.097965', u'stderr': u'Error from server: serviceaccounts "router" not found', u'changed': False, 'invocation': {'module_name': u'command', u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': False, u'_raw_params': u'/usr/local/bin/oc get sa router -n default', u'removes': None, u'creates': None, u'chdir': None}}, 'stdout_lines': [], 'failed_when_result': False, u'start': u'2017-01-10 18:30:12.484732', u'warnings': [], 'failed': False}]) TASK [openshift_serviceaccounts : test if scc needs to be updated] ************* ok: [10.8.169.22] => (item=hostnetwork) TASK [openshift_serviceaccounts : Grant the user access to the appropriate scc] changed: [10.8.169.22] => (item=[u'router', {'_ansible_parsed': True, u'cmd': [u'/usr/local/bin/oc', u'get', u'scc', u'hostnetwork', u'-o', u'yaml'], u'end': u'2017-01-10 18:30:13.737930', '_ansible_no_log': False, u'stdout': u'allowHostDirVolumePlugin: false\nallowHostIPC: false\nallowHostNetwork: true\nallowHostPID: false\nallowHostPorts: true\nallowPrivilegedContainer: false\nallowedCapabilities: null\napiVersion: v1\ndefaultAddCapabilities: null\nfsGroup:\n type: MustRunAs\nkind: SecurityContextConstraints\nmetadata:\n annotations:\n kubernetes.io/description: hostnetwork allows using host networking and host ports\n but still requires pods to be run with a UID and SELinux context that are allocated\n to the namespace.\n creationTimestamp: 2017-01-10T23:18:32Z\n name: hostnetwork\n resourceVersion: "104"\n selfLink: /api/v1/securitycontextconstraints/hostnetwork\n uid: 1a9c5cab-d78b-11e6-84d4-fa163e61d47e\npriority: null\nreadOnlyRootFilesystem: false\nrequiredDropCapabilities:\n- KILL\n- MKNOD\n- SYS_CHROOT\n- SETUID\n- SETGID\nrunAsUser:\n type: MustRunAsRange\nseLinuxContext:\n type: MustRunAs\nsupplementalGroups:\n type: MustRunAs\nvolumes:\n- configMap\n- downwardAPI\n- emptyDir\n- persistentVolumeClaim\n- secret', '_ansible_item_result': True, u'rc': 0, 'item': u'hostnetwork', u'delta': u'0:00:00.098686', u'stderr': u'', u'changed': False, 'invocation': {'module_name': u'command', u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': False, u'_raw_params': u'/usr/local/bin/oc get scc hostnetwork -o yaml', u'removes': None, u'creates': None, u'chdir': None}}, 'stdout_lines': [u'allowHostDirVolumePlugin: false', u'allowHostIPC: false', u'allowHostNetwork: true', u'allowHostPID: false', u'allowHostPorts: true', u'allowPrivilegedContainer: false', u'allowedCapabilities: null', u'apiVersion: v1', u'defaultAddCapabilities: null', u'fsGroup:', u' type: MustRunAs', u'kind: SecurityContextConstraints', u'metadata:', u' annotations:', u' kubernetes.io/description: hostnetwork allows using host networking and host ports', u' but still requires pods to be run with a UID and SELinux context that are allocated', u' to the namespace.', u' creationTimestamp: 2017-01-10T23:18:32Z', u' name: hostnetwork', u' resourceVersion: "104"', u' selfLink: /api/v1/securitycontextconstraints/hostnetwork', u' uid: 1a9c5cab-d78b-11e6-84d4-fa163e61d47e', u'priority: null', u'readOnlyRootFilesystem: false', u'requiredDropCapabilities:', u'- KILL', u'- MKNOD', u'- SYS_CHROOT', u'- SETUID', u'- SETGID', u'runAsUser:', u' type: MustRunAsRange', u'seLinuxContext:', u' type: MustRunAs', u'supplementalGroups:', u' type: MustRunAs', u'volumes:', u'- configMap', u'- downwardAPI', u'- emptyDir', u'- persistentVolumeClaim', u'- secret'], 'failed_when_result': False, u'start': u'2017-01-10 18:30:13.639244', u'warnings': [], 'failed': False}]) TASK [openshift_serviceaccounts : tmp dir for openshift] *********************** skipping: [10.8.169.22] TASK [openshift_serviceaccounts : Create service account configs] ************** skipping: [10.8.169.22] => (item=router) TASK [openshift_serviceaccounts : Get current security context constraints] **** skipping: [10.8.169.22] TASK [openshift_serviceaccounts : Add security context constraint for {{ item }}] *** skipping: [10.8.169.22] => (item=[u'router', {'_ansible_parsed': True, u'cmd': [u'/usr/local/bin/oc', u'get', u'scc', u'hostnetwork', u'-o', u'yaml'], u'end': u'2017-01-10 18:30:13.737930', '_ansible_no_log': False, u'stdout': u'allowHostDirVolumePlugin: false\nallowHostIPC: false\nallowHostNetwork: true\nallowHostPID: false\nallowHostPorts: true\nallowPrivilegedContainer: false\nallowedCapabilities: null\napiVersion: v1\ndefaultAddCapabilities: null\nfsGroup:\n type: MustRunAs\nkind: SecurityContextConstraints\nmetadata:\n annotations:\n kubernetes.io/description: hostnetwork allows using host networking and host ports\n but still requires pods to be run with a UID and SELinux context that are allocated\n to the namespace.\n creationTimestamp: 2017-01-10T23:18:32Z\n name: hostnetwork\n resourceVersion: "104"\n selfLink: /api/v1/securitycontextconstraints/hostnetwork\n uid: 1a9c5cab-d78b-11e6-84d4-fa163e61d47e\npriority: null\nreadOnlyRootFilesystem: false\nrequiredDropCapabilities:\n- KILL\n- MKNOD\n- SYS_CHROOT\n- SETUID\n- SETGID\nrunAsUser:\n type: MustRunAsRange\nseLinuxContext:\n type: MustRunAs\nsupplementalGroups:\n type: MustRunAs\nvolumes:\n- configMap\n- downwardAPI\n- emptyDir\n- persistentVolumeClaim\n- secret', '_ansible_item_result': True, u'rc': 0, 'item': u'hostnetwork', u'delta': u'0:00:00.098686', u'stderr': u'', u'changed': False, 'invocation': {'module_name': u'command', u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': False, u'_raw_params': u'/usr/local/bin/oc get scc hostnetwork -o yaml', u'removes': None, u'creates': None, u'chdir': None}}, 'stdout_lines': [u'allowHostDirVolumePlugin: false', u'allowHostIPC: false', u'allowHostNetwork: true', u'allowHostPID: false', u'allowHostPorts: true', u'allowPrivilegedContainer: false', u'allowedCapabilities: null', u'apiVersion: v1', u'defaultAddCapabilities: null', u'fsGroup:', u' type: MustRunAs', u'kind: SecurityContextConstraints', u'metadata:', u' annotations:', u' kubernetes.io/description: hostnetwork allows using host networking and host ports', u' but still requires pods to be run with a UID and SELinux context that are allocated', u' to the namespace.', u' creationTimestamp: 2017-01-10T23:18:32Z', u' name: hostnetwork', u' resourceVersion: "104"', u' selfLink: /api/v1/securitycontextconstraints/hostnetwork', u' uid: 1a9c5cab-d78b-11e6-84d4-fa163e61d47e', u'priority: null', u'readOnlyRootFilesystem: false', u'requiredDropCapabilities:', u'- KILL', u'- MKNOD', u'- SYS_CHROOT', u'- SETUID', u'- SETGID', u'runAsUser:', u' type: MustRunAsRange', u'seLinuxContext:', u' type: MustRunAs', u'supplementalGroups:', u' type: MustRunAs', u'volumes:', u'- configMap', u'- downwardAPI', u'- emptyDir', u'- persistentVolumeClaim', u'- secret'], 'failed_when_result': False, u'start': u'2017-01-10 18:30:13.639244', u'warnings': [], 'failed': False}]) TASK [openshift_serviceaccounts : Apply new scc rules for service accounts] **** skipping: [10.8.169.22] TASK [openshift_serviceaccounts : test if service accounts exists] ************* skipping: [10.8.169.22] => (item=registry) skipping: [10.8.169.22] => (item=router) TASK [openshift_serviceaccounts : create the service account] ****************** skipping: [10.8.169.22] => (item=[u'registry', {'skipped': True, '_ansible_no_log': False, 'skip_reason': u'Conditional check failed', '_ansible_item_result': True, 'item': u'registry', 'changed': False}]) skipping: [10.8.169.22] => (item=[u'router', {'skipped': True, '_ansible_no_log': False, 'skip_reason': u'Conditional check failed', '_ansible_item_result': True, 'item': u'router', 'changed': False}]) TASK [openshift_serviceaccounts : test if scc needs to be updated] ************* skipping: [10.8.169.22] => (item=privileged) TASK [openshift_serviceaccounts : Grant the user access to the appropriate scc] skipping: [10.8.169.22] => (item=[u'router', {'skipped': True, '_ansible_no_log': False, 'skip_reason': u'Conditional check failed', '_ansible_item_result': True, 'item': u'privileged', 'changed': False}]) skipping: [10.8.169.22] => (item=[u'registry', {'skipped': True, '_ansible_no_log': False, 'skip_reason': u'Conditional check failed', '_ansible_item_result': True, 'item': u'privileged', 'changed': False}]) TASK [openshift_serviceaccounts : tmp dir for openshift] *********************** skipping: [10.8.169.22] TASK [openshift_serviceaccounts : Create service account configs] ************** skipping: [10.8.169.22] => (item=registry) skipping: [10.8.169.22] => (item=router) TASK [openshift_serviceaccounts : Get current security context constraints] **** skipping: [10.8.169.22] TASK [openshift_serviceaccounts : Add security context constraint for {{ item }}] *** skipping: [10.8.169.22] => (item=[u'router', {'skipped': True, '_ansible_no_log': False, 'skip_reason': u'Conditional check failed', '_ansible_item_result': True, 'item': u'privileged', 'changed': False}]) skipping: [10.8.169.22] => (item=[u'registry', {'skipped': True, '_ansible_no_log': False, 'skip_reason': u'Conditional check failed', '_ansible_item_result': True, 'item': u'privileged', 'changed': False}]) TASK [openshift_serviceaccounts : Apply new scc rules for service accounts] **** skipping: [10.8.169.22] TASK [openshift_hosted : Create temp directory for kubeconfig] ***************** ok: [10.8.169.22] TASK [openshift_hosted : set_fact] ********************************************* ok: [10.8.169.22] TASK [openshift_hosted : Copy the admin client config(s)] ********************** ok: [10.8.169.22] TASK [openshift_hosted : fail] ************************************************* skipping: [10.8.169.22] TASK [openshift_hosted : Read router certificate and key] ********************** skipping: [10.8.169.22] => (item=) skipping: [10.8.169.22] => (item=) skipping: [10.8.169.22] => (item=) TASK [openshift_hosted : Persist certificate contents] ************************* skipping: [10.8.169.22] TASK [openshift_hosted : Create PEM certificate] ******************************* skipping: [10.8.169.22] TASK [openshift_hosted : Retrieve list of openshift nodes matching router selector] *** ok: [10.8.169.22] TASK [openshift_hosted : set_fact] ********************************************* ok: [10.8.169.22] TASK [openshift_hosted : Create OpenShift router] ****************************** changed: [10.8.169.22] TASK [openshift_hosted : Retrieve list of openshift nodes matching registry selector] *** ok: [10.8.169.22] TASK [openshift_hosted : set_fact] ********************************************* ok: [10.8.169.22] TASK [openshift_hosted : set_fact] ********************************************* skipping: [10.8.169.22] TASK [openshift_hosted : set_fact] ********************************************* ok: [10.8.169.22] TASK [openshift_hosted : set_fact] ********************************************* ok: [10.8.169.22] TASK [openshift_hosted : Create OpenShift registry] **************************** changed: [10.8.169.22] TASK [openshift_hosted : include] ********************************************** included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/openshift-ansible/roles/openshift_hosted/tasks/registry/secure.yml for 10.8.169.22 TASK [openshift_hosted : Create passthrough route for docker-registry] ********* changed: [10.8.169.22] TASK [openshift_hosted : Determine if registry certificate must be created] **** ok: [10.8.169.22] => (item=registry.crt) ok: [10.8.169.22] => (item=registry.key) TASK [openshift_hosted : Retrieve registry service IP] ************************* ok: [10.8.169.22] TASK [openshift_hosted : set_fact] ********************************************* ok: [10.8.169.22] TASK [openshift_hosted : Create registry certificates if they do not exist] **** changed: [10.8.169.22] TASK [openshift_hosted : Create the secret for the registry certificates] ****** changed: [10.8.169.22] TASK [openshift_hosted : Add the secret to the registry's pod service accounts] changed: [10.8.169.22] => (item=registry) changed: [10.8.169.22] => (item=default) TASK [openshift_hosted : Determine if registry-certificates secret volume attached] *** ok: [10.8.169.22] TASK [openshift_hosted : Attach registry-certificates secret volume] *********** changed: [10.8.169.22] TASK [openshift_hosted : Determine if registry environment variables must be set] *** ok: [10.8.169.22] TASK [openshift_hosted : Configure certificates in registry deplomentConfig] *** changed: [10.8.169.22] TASK [openshift_hosted : Determine if registry liveness probe scheme is HTTPS] * ok: [10.8.169.22] TASK [openshift_hosted : Update registry liveness probe from HTTP to HTTPS] **** changed: [10.8.169.22] TASK [openshift_hosted : Determine if registry readiness probe scheme is HTTPS] ok: [10.8.169.22] TASK [openshift_hosted : Update registry readiness probe from HTTP to HTTPS] *** changed: [10.8.169.22] TASK [openshift_hosted : include] ********************************************** skipping: [10.8.169.22] TASK [openshift_hosted : include] ********************************************** skipping: [10.8.169.22] TASK [openshift_hosted : Delete temp directory] ******************************** ok: [10.8.169.22] TASK [openshift_repos : assert] ************************************************ skipping: [10.8.169.22] TASK [openshift_repos : Ensure libselinux-python is installed] ***************** skipping: [10.8.169.22] TASK [openshift_repos : Create any additional repos that are defined] ********** skipping: [10.8.169.22] TASK [openshift_repos : Remove the additional repos if no longer defined] ****** skipping: [10.8.169.22] TASK [openshift_repos : Configure origin gpg keys if needed] ******************* skipping: [10.8.169.22] TASK [openshift_repos : Configure origin yum repositories RHEL/CentOS] ********* skipping: [10.8.169.22] TASK [openshift_repos : assert] ************************************************ skipping: [10.8.169.22] TASK [openshift_repos : Ensure libselinux-python is installed] ***************** skipping: [10.8.169.22] TASK [openshift_repos : Create any additional repos that are defined] ********** skipping: [10.8.169.22] TASK [openshift_repos : Remove the additional repos if no longer defined] ****** skipping: [10.8.169.22] TASK [openshift_repos : Configure origin gpg keys if needed] ******************* skipping: [10.8.169.22] TASK [openshift_repos : Configure origin yum repositories RHEL/CentOS] ********* skipping: [10.8.169.22] TASK [os_firewall : Assert - Do not use firewalld on Atomic Host] ************** skipping: [10.8.169.22] TASK [os_firewall : Install firewalld packages] ******************************** skipping: [10.8.169.22] TASK [os_firewall : Ensure iptables services are not enabled] ****************** skipping: [10.8.169.22] => (item=ip6tables) skipping: [10.8.169.22] => (item=iptables) TASK [os_firewall : Start and enable firewalld service] ************************ skipping: [10.8.169.22] TASK [os_firewall : need to pause here, otherwise the firewalld service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.22] TASK [os_firewall : Add firewalld allow rules] ********************************* TASK [os_firewall : Remove firewalld allow rules] ****************************** TASK [os_firewall : Ensure firewalld service is not enabled] ******************* skipping: [10.8.169.22] TASK [os_firewall : Install iptables packages] ********************************* skipping: [10.8.169.22] => (item=[]) TASK [os_firewall : Start and enable iptables service] ************************* skipping: [10.8.169.22] TASK [os_firewall : need to pause here, otherwise the iptables service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.22] TASK [os_firewall : Add iptables allow rules] ********************************** TASK [os_firewall : Remove iptables rules] ************************************* TASK [openshift_examples : Create local temp dir for OpenShift examples copy] ** skipping: [10.8.169.22] TASK [openshift_examples : Create tar of OpenShift examples] ******************* skipping: [10.8.169.22] TASK [openshift_examples : Create the remote OpenShift examples directory] ***** skipping: [10.8.169.22] TASK [openshift_examples : Unarchive the OpenShift examples on the remote] ***** skipping: [10.8.169.22] TASK [openshift_examples : Cleanup the OpenShift Examples temp dir] ************ skipping: [10.8.169.22] TASK [openshift_examples : Modify registry paths if registry_url is not XXXXXXXXXXXXXXXXXXXXXXXXXXX] *** skipping: [10.8.169.22] TASK [openshift_examples : Import RHEL streams] ******************************** skipping: [10.8.169.22] => (item=/etc/origin/examples/image-streams/dotnet_imagestreams.json) skipping: [10.8.169.22] => (item=/etc/origin/examples/image-streams/image-streams-rhel7.json) TASK [openshift_examples : Import Centos Image streams] ************************ skipping: [10.8.169.22] TASK [openshift_examples : Import db templates] ******************************** skipping: [10.8.169.22] TASK [openshift_examples : Remove defunct quickstart template files] *********** skipping: [10.8.169.22] => (item=/etc/origin/examples/quickstart-templates/nodejs.json) skipping: [10.8.169.22] => (item=/etc/origin/examples/quickstart-templates/django.json) skipping: [10.8.169.22] => (item=/etc/origin/examples/quickstart-templates/dancer.json) skipping: [10.8.169.22] => (item=/etc/origin/examples/quickstart-templates/cakephp.json) TASK [openshift_examples : Remove defunct quickstart templates from openshift namespace] *** skipping: [10.8.169.22] => (item=nodejs-example) skipping: [10.8.169.22] => (item=cakephp-example) skipping: [10.8.169.22] => (item=dancer-example) skipping: [10.8.169.22] => (item=django-example) TASK [openshift_examples : Import quickstart-templates] ************************ skipping: [10.8.169.22] TASK [openshift_examples : Remove old xPaas template files] ******************** skipping: [10.8.169.22] => (item=/etc/origin/examples/xpaas-templates/sso70-basic.json) TASK [openshift_examples : Remove old xPaas templates from openshift namespace] skipping: [10.8.169.22] => (item=sso70-basic) TASK [openshift_examples : Import xPaas image streams] ************************* skipping: [10.8.169.22] TASK [openshift_examples : Import xPaas templates] ***************************** skipping: [10.8.169.22] TASK [openshift_master_facts : Migrate legacy osm_default_subdomain fact] ****** skipping: [10.8.169.22] TASK [openshift_master_facts : fail] ******************************************* skipping: [10.8.169.22] TASK [openshift_master_facts : set_fact] *************************************** skipping: [10.8.169.22] TASK [openshift_master_facts : set_fact] *************************************** skipping: [10.8.169.22] TASK [openshift_master_facts : Set master facts] ******************************* skipping: [10.8.169.22] TASK [openshift_master_facts : Determine if scheduler config present] ********** skipping: [10.8.169.22] TASK [openshift_master_facts : set_fact] *************************************** skipping: [10.8.169.22] TASK [openshift_master_facts : Retrieve current scheduler config] ************** skipping: [10.8.169.22] TASK [openshift_master_facts : set_fact] *************************************** skipping: [10.8.169.22] TASK [openshift_master_facts : fail] ******************************************* skipping: [10.8.169.22] TASK [openshift_master_facts : set_fact] *************************************** skipping: [10.8.169.22] TASK [openshift_metrics : Create temp directory for kubeconfig] **************** skipping: [10.8.169.22] TASK [openshift_metrics : Record kubeconfig tmp dir] *************************** skipping: [10.8.169.22] TASK [openshift_metrics : Copy the admin client config(s)] ********************* skipping: [10.8.169.22] TASK [openshift_metrics : Set hosted metrics facts] **************************** skipping: [10.8.169.22] TASK [openshift_metrics : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_metrics : Check for existing metrics pods] ********************* skipping: [10.8.169.22] => (item=metrics-infra=hawkular-cassandra) skipping: [10.8.169.22] => (item=metrics-infra=heapster) skipping: [10.8.169.22] => (item=metrics-infra=hawkular-metrics) TASK [openshift_metrics : Check for previous deployer] ************************* skipping: [10.8.169.22] TASK [openshift_metrics : Record current deployment status] ******************** skipping: [10.8.169.22] TASK [openshift_metrics : Set deployment mode] ********************************* skipping: [10.8.169.22] TASK [openshift_metrics : Test if metrics-deployer service account exists] ***** skipping: [10.8.169.22] TASK [openshift_metrics : Create metrics-deployer Service Account] ************* skipping: [10.8.169.22] TASK [openshift_metrics : Test edit permissions] ******************************* skipping: [10.8.169.22] TASK [openshift_metrics : Add edit permission to the openshift-infra project to metrics-deployer SA] *** skipping: [10.8.169.22] TASK [openshift_metrics : Test hawkular view permissions] ********************** skipping: [10.8.169.22] TASK [openshift_metrics : Add view permissions to hawkular SA] ***************** skipping: [10.8.169.22] TASK [openshift_metrics : Test cluster-reader permissions] ********************* skipping: [10.8.169.22] TASK [openshift_metrics : Add cluster-reader permission to the openshift-infra project to heapster SA] *** skipping: [10.8.169.22] TASK [openshift_metrics : Create metrics-deployer secret] ********************** skipping: [10.8.169.22] TASK [openshift_metrics : Build metrics deployer command] ********************** skipping: [10.8.169.22] TASK [openshift_metrics : Deploy Metrics] ************************************** skipping: [10.8.169.22] TASK [openshift_metrics : set_fact] ******************************************** skipping: [10.8.169.22] TASK [openshift_metrics : Wait for image pull and deployer pod] **************** skipping: [10.8.169.22] TASK [openshift_metrics : Configure master for metrics] ************************ skipping: [10.8.169.22] TASK [openshift_metrics : Delete temp directory] ******************************* skipping: [10.8.169.22] TASK [openshift_repos : assert] ************************************************ skipping: [10.8.169.22] TASK [openshift_repos : Ensure libselinux-python is installed] ***************** skipping: [10.8.169.22] TASK [openshift_repos : Create any additional repos that are defined] ********** skipping: [10.8.169.22] TASK [openshift_repos : Remove the additional repos if no longer defined] ****** skipping: [10.8.169.22] TASK [openshift_repos : Configure origin gpg keys if needed] ******************* skipping: [10.8.169.22] TASK [openshift_repos : Configure origin yum repositories RHEL/CentOS] ********* skipping: [10.8.169.22] TASK [openshift_repos : assert] ************************************************ skipping: [10.8.169.22] TASK [openshift_repos : Ensure libselinux-python is installed] ***************** skipping: [10.8.169.22] TASK [openshift_repos : Create any additional repos that are defined] ********** skipping: [10.8.169.22] TASK [openshift_repos : Remove the additional repos if no longer defined] ****** skipping: [10.8.169.22] TASK [openshift_repos : Configure origin gpg keys if needed] ******************* skipping: [10.8.169.22] TASK [openshift_repos : Configure origin yum repositories RHEL/CentOS] ********* skipping: [10.8.169.22] TASK [os_firewall : Assert - Do not use firewalld on Atomic Host] ************** skipping: [10.8.169.22] TASK [os_firewall : Install firewalld packages] ******************************** skipping: [10.8.169.22] TASK [os_firewall : Ensure iptables services are not enabled] ****************** skipping: [10.8.169.22] => (item=ip6tables) skipping: [10.8.169.22] => (item=iptables) TASK [os_firewall : Start and enable firewalld service] ************************ skipping: [10.8.169.22] TASK [os_firewall : need to pause here, otherwise the firewalld service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.22] TASK [os_firewall : Add firewalld allow rules] ********************************* TASK [os_firewall : Remove firewalld allow rules] ****************************** TASK [os_firewall : Ensure firewalld service is not enabled] ******************* skipping: [10.8.169.22] TASK [os_firewall : Install iptables packages] ********************************* skipping: [10.8.169.22] => (item=[]) TASK [os_firewall : Start and enable iptables service] ************************* skipping: [10.8.169.22] TASK [os_firewall : need to pause here, otherwise the iptables service starting can sometimes cause ssh to fail] *** skipping: [10.8.169.22] TASK [os_firewall : Add iptables allow rules] ********************************** TASK [os_firewall : Remove iptables rules] ************************************* TASK [openshift_hosted_logging : Cleanup logging deployment] ******************* skipping: [10.8.169.22] TASK [openshift_hosted_logging : Deploy logging] ******************************* skipping: [10.8.169.22] TASK [cockpit-ui : Create temp directory for kubeconfig] *********************** skipping: [10.8.169.22] TASK [cockpit-ui : set_fact] *************************************************** skipping: [10.8.169.22] TASK [cockpit-ui : Copy the admin client config(s)] **************************** skipping: [10.8.169.22] TASK [cockpit-ui : Determine if docker-registry service exists] **************** skipping: [10.8.169.22] TASK [cockpit-ui : Create passthrough route for docker-registry] *************** skipping: [10.8.169.22] TASK [cockpit-ui : Create passthrough route for registry-console] ************** skipping: [10.8.169.22] TASK [cockpit-ui : Retrieve docker-registry route] ***************************** skipping: [10.8.169.22] TASK [cockpit-ui : Retrieve cockpit kube url] ********************************** skipping: [10.8.169.22] TASK [cockpit-ui : Deploy registry-console] ************************************ skipping: [10.8.169.22] TASK [cockpit-ui : Delete temp directory] ************************************** skipping: [10.8.169.22] PLAY RECAP ********************************************************************* 10.8.169.210 : ok=175 changed=54 unreachable=0 failed=0 10.8.169.215 : ok=169 changed=54 unreachable=0 failed=0 10.8.169.22 : ok=467 changed=126 unreachable=0 failed=0 localhost : ok=13 changed=1 unreachable=0 failed=0 [EnvInject] - Injecting environment variables from a build step. [EnvInject] - Injecting as environment variables the properties content GH_STATE=success GH_DESC="openshift-ansible install passed" GH_CONTEXT=aos-ci-jenkins/OS_3.3_containerized [EnvInject] - Variables injected successfully. [openshift-ansible-slave] $ python /tmp/hudson2126168883643737859.py Updating status for ae88c63dda7c433cd2b42ad7d9527b483579dc4e [EnvInject] - Injecting environment variables from a build step. [EnvInject] - Injecting as environment variables the properties content GH_STATE=pending GH_DESC="e2e tests in progress" GH_CONTEXT=aos-ci-jenkins/OS_3.3_containerized_e2e_tests ALL_GH_CONTEXT="aos-ci-jenkins/OS_3.3_containerized, aos-ci-jenkins/OS_3.3_containerized_e2e_tests" [EnvInject] - Variables injected successfully. [openshift-ansible-slave] $ python /tmp/hudson9015640478724369929.py Updating status for ae88c63dda7c433cd2b42ad7d9527b483579dc4e [EnvInject] - Injecting environment variables from a build step. [EnvInject] - Injecting as environment variables the properties content GH_DESC="e2e tests failed" GH_CONTEXT=aos-ci-jenkins/OS_3.3_containerized_e2e_tests [EnvInject] - Variables injected successfully. [openshift-ansible-slave] $ /bin/bash /tmp/hudson7204145595311449379.sh + chmod 600 /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/config/keys/ci-factory + ansible-playbook -vvv -i /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/ci-factory/utils/central_ci_dynamic_hosts.py --private-key=/home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/config/keys/ci-factory /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/playbooks/openshift/run-e2e-tests.yml Using /etc/ansible/ansible.cfg as config file statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/playbooks/openshift/roles/e2e/tasks/enable_pulp_repo.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/playbooks/openshift/roles/e2e/tasks/get_e2e_binary.yml statically included: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/playbooks/openshift/roles/e2e/tasks/execute_e2e_tests.yml PLAYBOOK: run-e2e-tests.yml **************************************************** 1 plays in /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/playbooks/openshift/run-e2e-tests.yml PLAY [masters] ***************************************************************** TASK [setup] ******************************************************************* Using module file /usr/lib/python2.7/site-packages/ansible/modules/core/system/setup.py <10.8.169.22> ESTABLISH SSH CONNECTION FOR USER: root <10.8.169.22> SSH: EXEC ssh -C -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o 'IdentityFile="/home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/config/keys/ci-factory"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/%h-%r 10.8.169.22 '/bin/sh -c '"'"'( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1484091056.35-279511475131085 `" && echo ansible-tmp-1484091056.35-279511475131085="` echo $HOME/.ansible/tmp/ansible-tmp-1484091056.35-279511475131085 `" ) && sleep 0'"'"'' <10.8.169.22> PUT /tmp/tmpSPQuz0 TO /root/.ansible/tmp/ansible-tmp-1484091056.35-279511475131085/setup.py <10.8.169.22> SSH: EXEC sftp -b - -C -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o 'IdentityFile="/home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/config/keys/ci-factory"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/%h-%r '[10.8.169.22]' <10.8.169.22> ESTABLISH SSH CONNECTION FOR USER: root <10.8.169.22> SSH: EXEC ssh -C -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o 'IdentityFile="/home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/config/keys/ci-factory"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/%h-%r 10.8.169.22 '/bin/sh -c '"'"'chmod u+x /root/.ansible/tmp/ansible-tmp-1484091056.35-279511475131085/ /root/.ansible/tmp/ansible-tmp-1484091056.35-279511475131085/setup.py && sleep 0'"'"'' <10.8.169.22> ESTABLISH SSH CONNECTION FOR USER: root <10.8.169.22> SSH: EXEC ssh -C -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o 'IdentityFile="/home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/config/keys/ci-factory"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/%h-%r -tt 10.8.169.22 '/bin/sh -c '"'"'/usr/bin/python /root/.ansible/tmp/ansible-tmp-1484091056.35-279511475131085/setup.py; rm -rf "/root/.ansible/tmp/ansible-tmp-1484091056.35-279511475131085/" > /dev/null 2>&1 && sleep 0'"'"'' ok: [10.8.169.22] TASK [e2e : Enable puddle ose repo on all nodes] ******************************* task path: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/playbooks/openshift/roles/e2e/tasks/enable_pulp_repo.yml:1 skipping: [10.8.169.22] => { "changed": false, "skip_reason": "Conditional check failed", "skipped": true } TASK [e2e : Clean all repos] *************************************************** task path: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/playbooks/openshift/roles/e2e/tasks/enable_pulp_repo.yml:11 skipping: [10.8.169.22] => { "changed": false, "skip_reason": "Conditional check failed", "skipped": true } TASK [e2e : Install atomic-openshift-tests] ************************************ task path: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/playbooks/openshift/roles/e2e/tasks/get_e2e_binary.yml:1 Using module file /usr/lib/python2.7/site-packages/ansible/modules/core/packaging/os/yum.py <10.8.169.22> ESTABLISH SSH CONNECTION FOR USER: root <10.8.169.22> SSH: EXEC ssh -C -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o 'IdentityFile="/home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/config/keys/ci-factory"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/%h-%r 10.8.169.22 '/bin/sh -c '"'"'( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1484091058.02-277052638319465 `" && echo ansible-tmp-1484091058.02-277052638319465="` echo $HOME/.ansible/tmp/ansible-tmp-1484091058.02-277052638319465 `" ) && sleep 0'"'"'' <10.8.169.22> PUT /tmp/tmpc3LaOQ TO /root/.ansible/tmp/ansible-tmp-1484091058.02-277052638319465/yum.py <10.8.169.22> SSH: EXEC sftp -b - -C -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o 'IdentityFile="/home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/config/keys/ci-factory"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/%h-%r '[10.8.169.22]' <10.8.169.22> ESTABLISH SSH CONNECTION FOR USER: root <10.8.169.22> SSH: EXEC ssh -C -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o 'IdentityFile="/home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/config/keys/ci-factory"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/%h-%r 10.8.169.22 '/bin/sh -c '"'"'chmod u+x /root/.ansible/tmp/ansible-tmp-1484091058.02-277052638319465/ /root/.ansible/tmp/ansible-tmp-1484091058.02-277052638319465/yum.py && sleep 0'"'"'' <10.8.169.22> ESTABLISH SSH CONNECTION FOR USER: root <10.8.169.22> SSH: EXEC ssh -C -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o 'IdentityFile="/home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/config/keys/ci-factory"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/%h-%r -tt 10.8.169.22 '/bin/sh -c '"'"'/usr/bin/python /root/.ansible/tmp/ansible-tmp-1484091058.02-277052638319465/yum.py; rm -rf "/root/.ansible/tmp/ansible-tmp-1484091058.02-277052638319465/" > /dev/null 2>&1 && sleep 0'"'"'' changed: [10.8.169.22] => { "changed": true, "invocation": { "module_args": { "conf_file": null, "disable_gpg_check": false, "disablerepo": null, "enablerepo": null, "exclude": null, "install_repoquery": true, "list": null, "name": [ "atomic-openshift-tests" ], "state": "present", "update_cache": false, "validate_certs": true }, "module_name": "yum" }, "msg": "", "rc": 0, "results": [ "Loaded plugins: product-id, search-disabled-repos, subscription-manager\nThis system is not registered to Red Hat Subscription Management. You can use subscription-manager to register.\nResolving Dependencies\n--> Running transaction check\n---> Package atomic-openshift-tests.x86_64 0:3.3.1.9-1.git.0.a7f5265.el7 will be installed\n--> Processing Dependency: atomic-openshift = 3.3.1.9-1.git.0.a7f5265.el7 for package: atomic-openshift-tests-3.3.1.9-1.git.0.a7f5265.el7.x86_64\n--> Running transaction check\n---> Package atomic-openshift.x86_64 0:3.3.1.9-1.git.0.a7f5265.el7 will be installed\n--> Processing Dependency: atomic-openshift-clients = 3.3.1.9-1.git.0.a7f5265.el7 for package: atomic-openshift-3.3.1.9-1.git.0.a7f5265.el7.x86_64\n--> Running transaction check\n---> Package atomic-openshift-clients.x86_64 0:3.3.1.9-1.git.0.a7f5265.el7 will be installed\n--> Finished Dependency Resolution\n\nDependencies Resolved\n\n================================================================================\n Package Arch Version Repository\n Size\n================================================================================\nInstalling:\n atomic-openshift-tests x86_64 3.3.1.9-1.git.0.a7f5265.el7 origin 32 M\nInstalling for dependencies:\n atomic-openshift x86_64 3.3.1.9-1.git.0.a7f5265.el7 origin 32 M\n atomic-openshift-clients x86_64 3.3.1.9-1.git.0.a7f5265.el7 origin 15 M\n\nTransaction Summary\n================================================================================\nInstall 1 Package (+2 Dependent packages)\n\nTotal download size: 79 M\nInstalled size: 450 M\nDownloading packages:\n--------------------------------------------------------------------------------\nTotal 44 MB/s | 79 MB 00:01 \nRunning transaction check\nRunning transaction test\nTransaction test succeeded\nRunning transaction\n Installing : atomic-openshift-clients-3.3.1.9-1.git.0.a7f5265.el7.x86_6 1/3 \n Installing : atomic-openshift-3.3.1.9-1.git.0.a7f5265.el7.x86_64 2/3 \n Installing : atomic-openshift-tests-3.3.1.9-1.git.0.a7f5265.el7.x86_64 3/3 \n Verifying : atomic-openshift-clients-3.3.1.9-1.git.0.a7f5265.el7.x86_6 1/3 \n Verifying : atomic-openshift-3.3.1.9-1.git.0.a7f5265.el7.x86_64 2/3 \n Verifying : atomic-openshift-tests-3.3.1.9-1.git.0.a7f5265.el7.x86_64 3/3 \n\nInstalled:\n atomic-openshift-tests.x86_64 0:3.3.1.9-1.git.0.a7f5265.el7 \n\nDependency Installed:\n atomic-openshift.x86_64 0:3.3.1.9-1.git.0.a7f5265.el7 \n atomic-openshift-clients.x86_64 0:3.3.1.9-1.git.0.a7f5265.el7 \n\nComplete!\n" ] } TASK [e2e : Execute extended.tests binary with set of tests] ******************* task path: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/playbooks/openshift/roles/e2e/tasks/execute_e2e_tests.yml:2 Using module file /usr/lib/python2.7/site-packages/ansible/modules/core/commands/command.py <10.8.169.22> ESTABLISH SSH CONNECTION FOR USER: root <10.8.169.22> SSH: EXEC ssh -C -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o 'IdentityFile="/home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/config/keys/ci-factory"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/%h-%r 10.8.169.22 '/bin/sh -c '"'"'( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1484091075.57-24914133899522 `" && echo ansible-tmp-1484091075.57-24914133899522="` echo $HOME/.ansible/tmp/ansible-tmp-1484091075.57-24914133899522 `" ) && sleep 0'"'"'' <10.8.169.22> PUT /tmp/tmpqvdSqw TO /root/.ansible/tmp/ansible-tmp-1484091075.57-24914133899522/command.py <10.8.169.22> SSH: EXEC sftp -b - -C -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o 'IdentityFile="/home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/config/keys/ci-factory"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/%h-%r '[10.8.169.22]' <10.8.169.22> ESTABLISH SSH CONNECTION FOR USER: root <10.8.169.22> SSH: EXEC ssh -C -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o 'IdentityFile="/home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/config/keys/ci-factory"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/%h-%r 10.8.169.22 '/bin/sh -c '"'"'chmod u+x /root/.ansible/tmp/ansible-tmp-1484091075.57-24914133899522/ /root/.ansible/tmp/ansible-tmp-1484091075.57-24914133899522/command.py && sleep 0'"'"'' <10.8.169.22> ESTABLISH SSH CONNECTION FOR USER: root <10.8.169.22> SSH: EXEC ssh -C -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o 'IdentityFile="/home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/config/keys/ci-factory"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/%h-%r -tt 10.8.169.22 '/bin/sh -c '"'"'/usr/bin/python /root/.ansible/tmp/ansible-tmp-1484091075.57-24914133899522/command.py; rm -rf "/root/.ansible/tmp/ansible-tmp-1484091075.57-24914133899522/" > /dev/null 2>&1 && sleep 0'"'"'' changed: [10.8.169.22] => { "changed": true, "cmd": "KUBECONFIG=/etc/origin/master/admin.kubeconfig /usr/libexec/atomic-openshift/extended.test --ginkgo.v=True --ginkgo.skip=\"should becomes\" --ginkgo.focus=\"EmptyDir\" 1>e2e.log 2>&1", "delta": "0:02:35.456889", "end": "2017-01-10 18:33:51.510663", "invocation": { "module_args": { "_raw_params": "KUBECONFIG=/etc/origin/master/admin.kubeconfig /usr/libexec/atomic-openshift/extended.test --ginkgo.v=True --ginkgo.skip=\"should becomes\" --ginkgo.focus=\"EmptyDir\" 1>e2e.log 2>&1", "_uses_shell": true, "chdir": null, "creates": null, "executable": null, "removes": null, "warn": true }, "module_name": "command" }, "rc": 0, "start": "2017-01-10 18:31:16.053774", "stderr": "", "stdout": "", "stdout_lines": [], "warnings": [] } TASK [e2e : Print out results from running e2e tests] ************************** task path: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/playbooks/openshift/roles/e2e/tasks/execute_e2e_tests.yml:8 Using module file /usr/lib/python2.7/site-packages/ansible/modules/core/commands/command.py <10.8.169.22> ESTABLISH SSH CONNECTION FOR USER: root <10.8.169.22> SSH: EXEC ssh -C -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o 'IdentityFile="/home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/config/keys/ci-factory"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/%h-%r 10.8.169.22 '/bin/sh -c '"'"'( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1484091231.68-73815349709102 `" && echo ansible-tmp-1484091231.68-73815349709102="` echo $HOME/.ansible/tmp/ansible-tmp-1484091231.68-73815349709102 `" ) && sleep 0'"'"'' <10.8.169.22> PUT /tmp/tmpkDdnU4 TO /root/.ansible/tmp/ansible-tmp-1484091231.68-73815349709102/command.py <10.8.169.22> SSH: EXEC sftp -b - -C -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o 'IdentityFile="/home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/config/keys/ci-factory"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/%h-%r '[10.8.169.22]' <10.8.169.22> ESTABLISH SSH CONNECTION FOR USER: root <10.8.169.22> SSH: EXEC ssh -C -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o 'IdentityFile="/home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/config/keys/ci-factory"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/%h-%r 10.8.169.22 '/bin/sh -c '"'"'chmod u+x /root/.ansible/tmp/ansible-tmp-1484091231.68-73815349709102/ /root/.ansible/tmp/ansible-tmp-1484091231.68-73815349709102/command.py && sleep 0'"'"'' <10.8.169.22> ESTABLISH SSH CONNECTION FOR USER: root <10.8.169.22> SSH: EXEC ssh -C -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o 'IdentityFile="/home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/config/keys/ci-factory"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/%h-%r -tt 10.8.169.22 '/bin/sh -c '"'"'/usr/bin/python /root/.ansible/tmp/ansible-tmp-1484091231.68-73815349709102/command.py; rm -rf "/root/.ansible/tmp/ansible-tmp-1484091231.68-73815349709102/" > /dev/null 2>&1 && sleep 0'"'"'' changed: [10.8.169.22] => { "changed": true, "cmd": "cat e2e.log", "delta": "0:00:00.002899", "end": "2017-01-10 18:33:52.189312", "invocation": { "module_args": { "_raw_params": "cat e2e.log", "_uses_shell": true, "chdir": null, "creates": null, "executable": null, "removes": null, "warn": true }, "module_name": "command" }, "rc": 0, "start": "2017-01-10 18:33:52.186413", "stderr": "", "stdout": "Running Suite: Extended\n=======================\nRandom Seed: \u001b[1m1484091076\u001b[0m\nWill run \u001b[1m19\u001b[0m of \u001b[1m509\u001b[0m specs\n\nI0110 18:31:16.345825 23683 e2e.go:58] The --provider flag is not set. Treating as a conformance test. Some tests may not be run.\nJan 10 18:31:16.345: INFO: >>> kubeConfig: /etc/origin/master/admin.kubeconfig\n\nJan 10 18:31:16.348: INFO: Waiting up to 0 for all pods (need at least 0) in namespace 'kube-system' to be running and ready\nJan 10 18:31:16.383: INFO: Waiting for pods to enter Success, but no pods in \"kube-system\" match label map[name:e2e-image-puller]\nJan 10 18:31:16.383: INFO: 0 / 0 pods in namespace 'kube-system' are running and ready (0 seconds elapsed)\nJan 10 18:31:16.383: INFO: expected 0 pod replicas in namespace 'kube-system', 0 are Running and Ready.\nJan 10 18:31:16.384: INFO: Waiting for pods to enter Success, but no pods in \"kube-system\" match label map[name:e2e-image-puller]\nI0110 18:31:16.384692 23683 e2e.go:58] The --provider flag is not set. Treating as a conformance test. Some tests may not be run.\n\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\n\u001b[90m------------------------------\u001b[0m\n\u001b[0m[k8s.io] EmptyDir volumes\u001b[0m \u001b[90mwhen FSGroup is specified [Feature:FSGroup]\u001b[0m \n \u001b[1mnew files should be created with FSGroup ownership when container is root\u001b[0m\n \u001b[37m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:44\u001b[0m\n\u001b[1mSTEP\u001b[0m: Creating a kubernetes client\nJan 10 18:31:16.385: INFO: >>> kubeConfig: /etc/origin/master/admin.kubeconfig\n\n\u001b[1mSTEP\u001b[0m: Building a namespace api object\nJan 10 18:31:16.408: INFO: About to run a Kube e2e test, ensuring namespace is privileged\n\u001b[1mSTEP\u001b[0m: Waiting for a default service account to be provisioned in namespace\n\u001b[1mSTEP\u001b[0m: Creating a pod to test emptydir 0644 on tmpfs\nJan 10 18:31:16.447: INFO: Waiting up to 5m0s for pod pod-e1c30011-d78c-11e6-8a95-fa163e61d47e status to be success or failure\nJan 10 18:31:16.457: INFO: No Status.Info for container 'test-container' in pod 'pod-e1c30011-d78c-11e6-8a95-fa163e61d47e' yet\nJan 10 18:31:16.457: INFO: Waiting for pod pod-e1c30011-d78c-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-njmyz' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (9.678773ms elapsed)\nJan 10 18:31:18.459: INFO: Nil State.Terminated for container 'test-container' in pod 'pod-e1c30011-d78c-11e6-8a95-fa163e61d47e' in namespace 'e2e-tests-emptydir-njmyz' so far\nJan 10 18:31:18.459: INFO: Waiting for pod pod-e1c30011-d78c-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-njmyz' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (2.011909436s elapsed)\nJan 10 18:31:20.461: INFO: Nil State.Terminated for container 'test-container' in pod 'pod-e1c30011-d78c-11e6-8a95-fa163e61d47e' in namespace 'e2e-tests-emptydir-njmyz' so far\nJan 10 18:31:20.461: INFO: Waiting for pod pod-e1c30011-d78c-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-njmyz' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (4.014110983s elapsed)\n\u001b[1mSTEP\u001b[0m: Saw pod success\n\u001b[1mSTEP\u001b[0m: Trying to get logs from node 172.16.132.150 pod pod-e1c30011-d78c-11e6-8a95-fa163e61d47e container test-container: \n\u001b[1mSTEP\u001b[0m: Successfully fetched pod logs:mount type of \"/test-volume\": tmpfs\ncontent of file \"/test-volume/test-file\": mount-tester new file\n\nperms of file \"/test-volume/test-file\": -rw-rw----\nowner UID of \"/test-volume/test-file\": 0\nowner GID of \"/test-volume/test-file\": 123\n\n\n\u001b[1mSTEP\u001b[0m: Collecting resource usage data\nJan 10 18:31:22.486: INFO: Closed stop channel. Waiting for 3 workers\nJan 10 18:31:22.486: INFO: Closing worker for 172.16.132.155\nJan 10 18:31:22.486: INFO: Closing worker for 172.16.132.150\nJan 10 18:31:22.486: INFO: Closing worker for 172.16.132.148\nJan 10 18:31:22.486: INFO: Waitgroup finished.\nJan 10 18:31:22.486: INFO: Unknown output type: . Skipping.\nJan 10 18:31:22.486: INFO: Waiting up to 1m0s for all nodes to be ready\n\u001b[1mSTEP\u001b[0m: Destroying namespace \"e2e-tests-emptydir-njmyz\" for this suite.\n\n\u001b[32m• [SLOW TEST:11.113 seconds]\u001b[0m\n[k8s.io] EmptyDir volumes\n\u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:690\u001b[0m\n when FSGroup is specified [Feature:FSGroup]\n \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:61\u001b[0m\n new files should be created with FSGroup ownership when container is root\n \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:44\u001b[0m\n\u001b[90m------------------------------\u001b[0m\n\u001b[0m[k8s.io] EmptyDir volumes\u001b[0m \u001b[90mwhen FSGroup is specified [Feature:FSGroup]\u001b[0m \n \u001b[1mnew files should be created with FSGroup ownership when container is non-root\u001b[0m\n \u001b[37m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:48\u001b[0m\n\u001b[1mSTEP\u001b[0m: Creating a kubernetes client\nJan 10 18:31:27.498: INFO: >>> kubeConfig: /etc/origin/master/admin.kubeconfig\n\n\u001b[1mSTEP\u001b[0m: Building a namespace api object\nJan 10 18:31:27.530: INFO: About to run a Kube e2e test, ensuring namespace is privileged\n\u001b[1mSTEP\u001b[0m: Waiting for a default service account to be provisioned in namespace\n\u001b[1mSTEP\u001b[0m: Creating a pod to test emptydir 0644 on tmpfs\nJan 10 18:31:27.561: INFO: Waiting up to 5m0s for pod pod-e862e652-d78c-11e6-8a95-fa163e61d47e status to be success or failure\nJan 10 18:31:27.563: INFO: No Status.Info for container 'test-container' in pod 'pod-e862e652-d78c-11e6-8a95-fa163e61d47e' yet\nJan 10 18:31:27.563: INFO: Waiting for pod pod-e862e652-d78c-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-yd8ef' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (1.233326ms elapsed)\nJan 10 18:31:29.565: INFO: Nil State.Terminated for container 'test-container' in pod 'pod-e862e652-d78c-11e6-8a95-fa163e61d47e' in namespace 'e2e-tests-emptydir-yd8ef' so far\nJan 10 18:31:29.565: INFO: Waiting for pod pod-e862e652-d78c-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-yd8ef' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (2.003996573s elapsed)\nJan 10 18:31:31.568: INFO: Nil State.Terminated for container 'test-container' in pod 'pod-e862e652-d78c-11e6-8a95-fa163e61d47e' in namespace 'e2e-tests-emptydir-yd8ef' so far\nJan 10 18:31:31.568: INFO: Waiting for pod pod-e862e652-d78c-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-yd8ef' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (4.006466774s elapsed)\nJan 10 18:31:33.571: INFO: Nil State.Terminated for container 'test-container' in pod 'pod-e862e652-d78c-11e6-8a95-fa163e61d47e' in namespace 'e2e-tests-emptydir-yd8ef' so far\nJan 10 18:31:33.571: INFO: Waiting for pod pod-e862e652-d78c-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-yd8ef' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (6.009226209s elapsed)\n\u001b[1mSTEP\u001b[0m: Saw pod success\n\u001b[1mSTEP\u001b[0m: Trying to get logs from node 172.16.132.148 pod pod-e862e652-d78c-11e6-8a95-fa163e61d47e container test-container: \n\u001b[1mSTEP\u001b[0m: Successfully fetched pod logs:mount type of \"/test-volume\": tmpfs\ncontent of file \"/test-volume/test-file\": mount-tester new file\n\nperms of file \"/test-volume/test-file\": -rw-rw----\nowner UID of \"/test-volume/test-file\": 0\nowner GID of \"/test-volume/test-file\": 123\n\n\n\u001b[1mSTEP\u001b[0m: Collecting resource usage data\nJan 10 18:31:35.587: INFO: Closed stop channel. Waiting for 3 workers\nJan 10 18:31:35.587: INFO: Closing worker for 172.16.132.155\nJan 10 18:31:35.587: INFO: Closing worker for 172.16.132.150\nJan 10 18:31:35.587: INFO: Closing worker for 172.16.132.148\nJan 10 18:31:35.587: INFO: Waitgroup finished.\nJan 10 18:31:35.587: INFO: Unknown output type: . Skipping.\nJan 10 18:31:35.587: INFO: Waiting up to 1m0s for all nodes to be ready\n\u001b[1mSTEP\u001b[0m: Destroying namespace \"e2e-tests-emptydir-yd8ef\" for this suite.\n\n\u001b[32m• [SLOW TEST:13.097 seconds]\u001b[0m\n[k8s.io] EmptyDir volumes\n\u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:690\u001b[0m\n when FSGroup is specified [Feature:FSGroup]\n \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:61\u001b[0m\n new files should be created with FSGroup ownership when container is non-root\n \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:48\u001b[0m\n\u001b[90m------------------------------\u001b[0m\n\u001b[0m[k8s.io] EmptyDir volumes\u001b[0m \u001b[90mwhen FSGroup is specified [Feature:FSGroup]\u001b[0m \n \u001b[1mfiles with FSGroup ownership should support (root,0644,tmpfs)\u001b[0m\n \u001b[37m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:52\u001b[0m\n\u001b[1mSTEP\u001b[0m: Creating a kubernetes client\nJan 10 18:31:40.595: INFO: >>> kubeConfig: /etc/origin/master/admin.kubeconfig\n\n\u001b[1mSTEP\u001b[0m: Building a namespace api object\nJan 10 18:31:40.620: INFO: About to run a Kube e2e test, ensuring namespace is privileged\n\u001b[1mSTEP\u001b[0m: Waiting for a default service account to be provisioned in namespace\n\u001b[1mSTEP\u001b[0m: Creating a pod to test emptydir 0644 on tmpfs\nJan 10 18:31:40.667: INFO: Waiting up to 5m0s for pod pod-f031fdd8-d78c-11e6-8a95-fa163e61d47e status to be success or failure\nJan 10 18:31:40.669: INFO: No Status.Info for container 'test-container' in pod 'pod-f031fdd8-d78c-11e6-8a95-fa163e61d47e' yet\nJan 10 18:31:40.669: INFO: Waiting for pod pod-f031fdd8-d78c-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-emin5' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (2.723887ms elapsed)\n\u001b[1mSTEP\u001b[0m: Saw pod success\n\u001b[1mSTEP\u001b[0m: Trying to get logs from node 172.16.132.150 pod pod-f031fdd8-d78c-11e6-8a95-fa163e61d47e container test-container: \n\u001b[1mSTEP\u001b[0m: Successfully fetched pod logs:mount type of \"/test-volume\": tmpfs\ncontent of file \"/test-volume/test-file\": mount-tester new file\n\nperms of file \"/test-volume/test-file\": -rw-r--r--\n\n\n\u001b[1mSTEP\u001b[0m: Collecting resource usage data\nJan 10 18:31:42.687: INFO: Closed stop channel. Waiting for 3 workers\nJan 10 18:31:42.687: INFO: Closing worker for 172.16.132.155\nJan 10 18:31:42.687: INFO: Closing worker for 172.16.132.150\nJan 10 18:31:42.687: INFO: Closing worker for 172.16.132.148\nJan 10 18:31:42.687: INFO: Waitgroup finished.\nJan 10 18:31:42.687: INFO: Unknown output type: . Skipping.\nJan 10 18:31:42.687: INFO: Waiting up to 1m0s for all nodes to be ready\n\u001b[1mSTEP\u001b[0m: Destroying namespace \"e2e-tests-emptydir-emin5\" for this suite.\n\n\u001b[32m• [SLOW TEST:7.101 seconds]\u001b[0m\n[k8s.io] EmptyDir volumes\n\u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:690\u001b[0m\n when FSGroup is specified [Feature:FSGroup]\n \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:61\u001b[0m\n files with FSGroup ownership should support (root,0644,tmpfs)\n \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:52\u001b[0m\n\u001b[90m------------------------------\u001b[0m\n\u001b[0m[k8s.io] EmptyDir volumes\u001b[0m \u001b[90mwhen FSGroup is specified [Feature:FSGroup]\u001b[0m \n \u001b[1mvolume on default medium should have the correct mode using FSGroup\u001b[0m\n \u001b[37m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:56\u001b[0m\n\u001b[1mSTEP\u001b[0m: Creating a kubernetes client\nJan 10 18:31:47.696: INFO: >>> kubeConfig: /etc/origin/master/admin.kubeconfig\n\n\u001b[1mSTEP\u001b[0m: Building a namespace api object\nJan 10 18:31:47.719: INFO: About to run a Kube e2e test, ensuring namespace is privileged\n\u001b[1mSTEP\u001b[0m: Waiting for a default service account to be provisioned in namespace\n\u001b[1mSTEP\u001b[0m: Creating a pod to test emptydir volume type on node default medium\nJan 10 18:31:47.747: INFO: Waiting up to 5m0s for pod pod-f46afde4-d78c-11e6-8a95-fa163e61d47e status to be success or failure\nJan 10 18:31:47.761: INFO: No Status.Info for container 'test-container' in pod 'pod-f46afde4-d78c-11e6-8a95-fa163e61d47e' yet\nJan 10 18:31:47.761: INFO: Waiting for pod pod-f46afde4-d78c-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-oiv0k' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (14.862636ms elapsed)\n\u001b[1mSTEP\u001b[0m: Saw pod success\n\u001b[1mSTEP\u001b[0m: Trying to get logs from node 172.16.132.148 pod pod-f46afde4-d78c-11e6-8a95-fa163e61d47e container test-container: \n\u001b[1mSTEP\u001b[0m: Successfully fetched pod logs:mount type of \"/test-volume\": 1481003842\nperms of file \"/test-volume\": -rwxrwxrwx\n\n\n\u001b[1mSTEP\u001b[0m: Collecting resource usage data\nJan 10 18:31:49.776: INFO: Closed stop channel. Waiting for 3 workers\nJan 10 18:31:49.776: INFO: Closing worker for 172.16.132.155\nJan 10 18:31:49.776: INFO: Closing worker for 172.16.132.150\nJan 10 18:31:49.776: INFO: Closing worker for 172.16.132.148\nJan 10 18:31:49.776: INFO: Waitgroup finished.\nJan 10 18:31:49.776: INFO: Unknown output type: . Skipping.\nJan 10 18:31:49.776: INFO: Waiting up to 1m0s for all nodes to be ready\n\u001b[1mSTEP\u001b[0m: Destroying namespace \"e2e-tests-emptydir-oiv0k\" for this suite.\n\n\u001b[32m• [SLOW TEST:7.088 seconds]\u001b[0m\n[k8s.io] EmptyDir volumes\n\u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:690\u001b[0m\n when FSGroup is specified [Feature:FSGroup]\n \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:61\u001b[0m\n volume on default medium should have the correct mode using FSGroup\n \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:56\u001b[0m\n\u001b[90m------------------------------\u001b[0m\n\u001b[0m[k8s.io] EmptyDir volumes\u001b[0m \u001b[90mwhen FSGroup is specified [Feature:FSGroup]\u001b[0m \n \u001b[1mvolume on tmpfs should have the correct mode using FSGroup\u001b[0m\n \u001b[37m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:60\u001b[0m\n\u001b[1mSTEP\u001b[0m: Creating a kubernetes client\nJan 10 18:31:54.784: INFO: >>> kubeConfig: /etc/origin/master/admin.kubeconfig\n\n\u001b[1mSTEP\u001b[0m: Building a namespace api object\nJan 10 18:31:54.815: INFO: About to run a Kube e2e test, ensuring namespace is privileged\n\u001b[1mSTEP\u001b[0m: Waiting for a default service account to be provisioned in namespace\n\u001b[1mSTEP\u001b[0m: Creating a pod to test emptydir volume type on tmpfs\nJan 10 18:31:54.845: INFO: Waiting up to 5m0s for pod pod-f8a62e99-d78c-11e6-8a95-fa163e61d47e status to be success or failure\nJan 10 18:31:54.847: INFO: No Status.Info for container 'test-container' in pod 'pod-f8a62e99-d78c-11e6-8a95-fa163e61d47e' yet\nJan 10 18:31:54.847: INFO: Waiting for pod pod-f8a62e99-d78c-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-d5s9w' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (2.337756ms elapsed)\n\u001b[1mSTEP\u001b[0m: Saw pod success\n\u001b[1mSTEP\u001b[0m: Trying to get logs from node 172.16.132.150 pod pod-f8a62e99-d78c-11e6-8a95-fa163e61d47e container test-container: \n\u001b[1mSTEP\u001b[0m: Successfully fetched pod logs:mount type of \"/test-volume\": tmpfs\nperms of file \"/test-volume\": -rwxrwxrwx\n\n\n\u001b[1mSTEP\u001b[0m: Collecting resource usage data\nJan 10 18:31:56.869: INFO: Closed stop channel. Waiting for 3 workers\nJan 10 18:31:56.869: INFO: Closing worker for 172.16.132.155\nJan 10 18:31:56.869: INFO: Closing worker for 172.16.132.148\nJan 10 18:31:56.869: INFO: Closing worker for 172.16.132.150\nJan 10 18:31:56.869: INFO: Waitgroup finished.\nJan 10 18:31:56.869: INFO: Unknown output type: . Skipping.\nJan 10 18:31:56.869: INFO: Waiting up to 1m0s for all nodes to be ready\n\u001b[1mSTEP\u001b[0m: Destroying namespace \"e2e-tests-emptydir-d5s9w\" for this suite.\n\n\u001b[32m• [SLOW TEST:7.094 seconds]\u001b[0m\n[k8s.io] EmptyDir volumes\n\u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:690\u001b[0m\n when FSGroup is specified [Feature:FSGroup]\n \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:61\u001b[0m\n volume on tmpfs should have the correct mode using FSGroup\n \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:60\u001b[0m\n\u001b[90m------------------------------\u001b[0m\n\u001b[0m[k8s.io] EmptyDir volumes\u001b[0m \n \u001b[1mvolume on tmpfs should have the correct mode [Conformance]\u001b[0m\n \u001b[37m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:65\u001b[0m\n\u001b[1mSTEP\u001b[0m: Creating a kubernetes client\nJan 10 18:32:01.878: INFO: >>> kubeConfig: /etc/origin/master/admin.kubeconfig\n\n\u001b[1mSTEP\u001b[0m: Building a namespace api object\nJan 10 18:32:01.910: INFO: About to run a Kube e2e test, ensuring namespace is privileged\n\u001b[1mSTEP\u001b[0m: Waiting for a default service account to be provisioned in namespace\n\u001b[1mSTEP\u001b[0m: Creating a pod to test emptydir volume type on tmpfs\nJan 10 18:32:01.990: INFO: Waiting up to 5m0s for pod pod-fce6d939-d78c-11e6-8a95-fa163e61d47e status to be success or failure\nJan 10 18:32:01.995: INFO: No Status.Info for container 'test-container' in pod 'pod-fce6d939-d78c-11e6-8a95-fa163e61d47e' yet\nJan 10 18:32:01.995: INFO: Waiting for pod pod-fce6d939-d78c-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-4w1ts' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (5.113696ms elapsed)\nJan 10 18:32:03.998: INFO: Nil State.Terminated for container 'test-container' in pod 'pod-fce6d939-d78c-11e6-8a95-fa163e61d47e' in namespace 'e2e-tests-emptydir-4w1ts' so far\nJan 10 18:32:03.998: INFO: Waiting for pod pod-fce6d939-d78c-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-4w1ts' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (2.007629958s elapsed)\n\u001b[1mSTEP\u001b[0m: Saw pod success\n\u001b[1mSTEP\u001b[0m: Trying to get logs from node 172.16.132.148 pod pod-fce6d939-d78c-11e6-8a95-fa163e61d47e container test-container: \n\u001b[1mSTEP\u001b[0m: Successfully fetched pod logs:mount type of \"/test-volume\": tmpfs\nperms of file \"/test-volume\": -rwxrwxrwx\n\n\n\u001b[1mSTEP\u001b[0m: Collecting resource usage data\nJan 10 18:32:06.013: INFO: Closed stop channel. Waiting for 3 workers\nJan 10 18:32:06.013: INFO: Closing worker for 172.16.132.150\nJan 10 18:32:06.013: INFO: Closing worker for 172.16.132.148\nJan 10 18:32:06.013: INFO: Closing worker for 172.16.132.155\nJan 10 18:32:06.013: INFO: Waitgroup finished.\nJan 10 18:32:06.013: INFO: Unknown output type: . Skipping.\nJan 10 18:32:06.013: INFO: Waiting up to 1m0s for all nodes to be ready\n\u001b[1mSTEP\u001b[0m: Destroying namespace \"e2e-tests-emptydir-4w1ts\" for this suite.\n\n\u001b[32m• [SLOW TEST:9.145 seconds]\u001b[0m\n[k8s.io] EmptyDir volumes\n\u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:690\u001b[0m\n volume on tmpfs should have the correct mode [Conformance]\n \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:65\u001b[0m\n\u001b[90m------------------------------\u001b[0m\n\u001b[0m[k8s.io] EmptyDir volumes\u001b[0m \n \u001b[1mshould support (root,0644,tmpfs) [Conformance]\u001b[0m\n \u001b[37m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:69\u001b[0m\n\u001b[1mSTEP\u001b[0m: Creating a kubernetes client\nJan 10 18:32:11.023: INFO: >>> kubeConfig: /etc/origin/master/admin.kubeconfig\n\n\u001b[1mSTEP\u001b[0m: Building a namespace api object\nJan 10 18:32:11.047: INFO: About to run a Kube e2e test, ensuring namespace is privileged\n\u001b[1mSTEP\u001b[0m: Waiting for a default service account to be provisioned in namespace\n\u001b[1mSTEP\u001b[0m: Creating a pod to test emptydir 0644 on tmpfs\nJan 10 18:32:11.078: INFO: Waiting up to 5m0s for pod pod-0252e642-d78d-11e6-8a95-fa163e61d47e status to be success or failure\nJan 10 18:32:11.079: INFO: No Status.Info for container 'test-container' in pod 'pod-0252e642-d78d-11e6-8a95-fa163e61d47e' yet\nJan 10 18:32:11.079: INFO: Waiting for pod pod-0252e642-d78d-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-z8gry' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (1.232535ms elapsed)\n\u001b[1mSTEP\u001b[0m: Saw pod success\n\u001b[1mSTEP\u001b[0m: Trying to get logs from node 172.16.132.150 pod pod-0252e642-d78d-11e6-8a95-fa163e61d47e container test-container: \n\u001b[1mSTEP\u001b[0m: Successfully fetched pod logs:mount type of \"/test-volume\": tmpfs\ncontent of file \"/test-volume/test-file\": mount-tester new file\n\nperms of file \"/test-volume/test-file\": -rw-r--r--\n\n\n\u001b[1mSTEP\u001b[0m: Collecting resource usage data\nJan 10 18:32:13.097: INFO: Closed stop channel. Waiting for 3 workers\nJan 10 18:32:13.097: INFO: Closing worker for 172.16.132.155\nJan 10 18:32:13.097: INFO: Closing worker for 172.16.132.150\nJan 10 18:32:13.097: INFO: Closing worker for 172.16.132.148\nJan 10 18:32:13.097: INFO: Waitgroup finished.\nJan 10 18:32:13.097: INFO: Unknown output type: . Skipping.\nJan 10 18:32:13.097: INFO: Waiting up to 1m0s for all nodes to be ready\n\u001b[1mSTEP\u001b[0m: Destroying namespace \"e2e-tests-emptydir-z8gry\" for this suite.\n\n\u001b[32m• [SLOW TEST:7.084 seconds]\u001b[0m\n[k8s.io] EmptyDir volumes\n\u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:690\u001b[0m\n should support (root,0644,tmpfs) [Conformance]\n \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:69\u001b[0m\n\u001b[90m------------------------------\u001b[0m\n\u001b[0m[k8s.io] EmptyDir volumes\u001b[0m \n \u001b[1mshould support (root,0666,tmpfs) [Conformance]\u001b[0m\n \u001b[37m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:73\u001b[0m\n\u001b[1mSTEP\u001b[0m: Creating a kubernetes client\nJan 10 18:32:18.107: INFO: >>> kubeConfig: /etc/origin/master/admin.kubeconfig\n\n\u001b[1mSTEP\u001b[0m: Building a namespace api object\nJan 10 18:32:18.128: INFO: About to run a Kube e2e test, ensuring namespace is privileged\n\u001b[1mSTEP\u001b[0m: Waiting for a default service account to be provisioned in namespace\n\u001b[1mSTEP\u001b[0m: Creating a pod to test emptydir 0666 on tmpfs\nJan 10 18:32:18.228: INFO: Waiting up to 5m0s for pod pod-068d993f-d78d-11e6-8a95-fa163e61d47e status to be success or failure\nJan 10 18:32:18.231: INFO: No Status.Info for container 'test-container' in pod 'pod-068d993f-d78d-11e6-8a95-fa163e61d47e' yet\nJan 10 18:32:18.231: INFO: Waiting for pod pod-068d993f-d78d-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-03t1m' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (2.177997ms elapsed)\nJan 10 18:32:20.233: INFO: Nil State.Terminated for container 'test-container' in pod 'pod-068d993f-d78d-11e6-8a95-fa163e61d47e' in namespace 'e2e-tests-emptydir-03t1m' so far\nJan 10 18:32:20.233: INFO: Waiting for pod pod-068d993f-d78d-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-03t1m' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (2.004342222s elapsed)\n\u001b[1mSTEP\u001b[0m: Saw pod success\n\u001b[1mSTEP\u001b[0m: Trying to get logs from node 172.16.132.148 pod pod-068d993f-d78d-11e6-8a95-fa163e61d47e container test-container: \n\u001b[1mSTEP\u001b[0m: Successfully fetched pod logs:mount type of \"/test-volume\": tmpfs\ncontent of file \"/test-volume/test-file\": mount-tester new file\n\nperms of file \"/test-volume/test-file\": -rw-rw-rw-\n\n\n\u001b[1mSTEP\u001b[0m: Collecting resource usage data\nJan 10 18:32:22.247: INFO: Closed stop channel. Waiting for 3 workers\nJan 10 18:32:22.247: INFO: Closing worker for 172.16.132.155\nJan 10 18:32:22.247: INFO: Closing worker for 172.16.132.150\nJan 10 18:32:22.247: INFO: Closing worker for 172.16.132.148\nJan 10 18:32:22.247: INFO: Waitgroup finished.\nJan 10 18:32:22.247: INFO: Unknown output type: . Skipping.\nJan 10 18:32:22.247: INFO: Waiting up to 1m0s for all nodes to be ready\n\u001b[1mSTEP\u001b[0m: Destroying namespace \"e2e-tests-emptydir-03t1m\" for this suite.\n\n\u001b[32m• [SLOW TEST:9.149 seconds]\u001b[0m\n[k8s.io] EmptyDir volumes\n\u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:690\u001b[0m\n should support (root,0666,tmpfs) [Conformance]\n \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:73\u001b[0m\n\u001b[90m------------------------------\u001b[0m\n\u001b[0m[k8s.io] EmptyDir volumes\u001b[0m \n \u001b[1mshould support (root,0777,tmpfs) [Conformance]\u001b[0m\n \u001b[37m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:77\u001b[0m\n\u001b[1mSTEP\u001b[0m: Creating a kubernetes client\nJan 10 18:32:27.256: INFO: >>> kubeConfig: /etc/origin/master/admin.kubeconfig\n\n\u001b[1mSTEP\u001b[0m: Building a namespace api object\nJan 10 18:32:27.279: INFO: About to run a Kube e2e test, ensuring namespace is privileged\n\u001b[1mSTEP\u001b[0m: Waiting for a default service account to be provisioned in namespace\n\u001b[1mSTEP\u001b[0m: Creating a pod to test emptydir 0777 on tmpfs\nJan 10 18:32:27.328: INFO: Waiting up to 5m0s for pod pod-0c02aec2-d78d-11e6-8a95-fa163e61d47e status to be success or failure\nJan 10 18:32:27.337: INFO: No Status.Info for container 'test-container' in pod 'pod-0c02aec2-d78d-11e6-8a95-fa163e61d47e' yet\nJan 10 18:32:27.337: INFO: Waiting for pod pod-0c02aec2-d78d-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-zbsyv' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (8.617146ms elapsed)\n\u001b[1mSTEP\u001b[0m: Saw pod success\n\u001b[1mSTEP\u001b[0m: Trying to get logs from node 172.16.132.150 pod pod-0c02aec2-d78d-11e6-8a95-fa163e61d47e container test-container: \n\u001b[1mSTEP\u001b[0m: Successfully fetched pod logs:mount type of \"/test-volume\": tmpfs\ncontent of file \"/test-volume/test-file\": mount-tester new file\n\nperms of file \"/test-volume/test-file\": -rwxrwxrwx\n\n\n\u001b[1mSTEP\u001b[0m: Collecting resource usage data\nJan 10 18:32:29.351: INFO: Closed stop channel. Waiting for 3 workers\nJan 10 18:32:29.351: INFO: Closing worker for 172.16.132.155\nJan 10 18:32:29.351: INFO: Closing worker for 172.16.132.150\nJan 10 18:32:29.351: INFO: Closing worker for 172.16.132.148\nJan 10 18:32:29.351: INFO: Waitgroup finished.\nJan 10 18:32:29.351: INFO: Unknown output type: . Skipping.\nJan 10 18:32:29.351: INFO: Waiting up to 1m0s for all nodes to be ready\n\u001b[1mSTEP\u001b[0m: Destroying namespace \"e2e-tests-emptydir-zbsyv\" for this suite.\n\n\u001b[32m• [SLOW TEST:7.105 seconds]\u001b[0m\n[k8s.io] EmptyDir volumes\n\u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:690\u001b[0m\n should support (root,0777,tmpfs) [Conformance]\n \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:77\u001b[0m\n\u001b[90m------------------------------\u001b[0m\n\u001b[0m[k8s.io] EmptyDir volumes\u001b[0m \n \u001b[1mshould support (non-root,0644,tmpfs) [Conformance]\u001b[0m\n \u001b[37m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:81\u001b[0m\n\u001b[1mSTEP\u001b[0m: Creating a kubernetes client\nJan 10 18:32:34.361: INFO: >>> kubeConfig: /etc/origin/master/admin.kubeconfig\n\n\u001b[1mSTEP\u001b[0m: Building a namespace api object\nJan 10 18:32:34.392: INFO: About to run a Kube e2e test, ensuring namespace is privileged\n\u001b[1mSTEP\u001b[0m: Waiting for a default service account to be provisioned in namespace\n\u001b[1mSTEP\u001b[0m: Creating a pod to test emptydir 0644 on tmpfs\nJan 10 18:32:34.422: INFO: Waiting up to 5m0s for pod pod-103d0c8b-d78d-11e6-8a95-fa163e61d47e status to be success or failure\nJan 10 18:32:34.423: INFO: No Status.Info for container 'test-container' in pod 'pod-103d0c8b-d78d-11e6-8a95-fa163e61d47e' yet\nJan 10 18:32:34.423: INFO: Waiting for pod pod-103d0c8b-d78d-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-lsclc' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (1.246851ms elapsed)\nJan 10 18:32:36.426: INFO: Nil State.Terminated for container 'test-container' in pod 'pod-103d0c8b-d78d-11e6-8a95-fa163e61d47e' in namespace 'e2e-tests-emptydir-lsclc' so far\nJan 10 18:32:36.426: INFO: Waiting for pod pod-103d0c8b-d78d-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-lsclc' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (2.003866886s elapsed)\n\u001b[1mSTEP\u001b[0m: Saw pod success\n\u001b[1mSTEP\u001b[0m: Trying to get logs from node 172.16.132.148 pod pod-103d0c8b-d78d-11e6-8a95-fa163e61d47e container test-container: \n\u001b[1mSTEP\u001b[0m: Successfully fetched pod logs:mount type of \"/test-volume\": tmpfs\ncontent of file \"/test-volume/test-file\": mount-tester new file\n\nperms of file \"/test-volume/test-file\": -rw-r--r--\n\n\n\u001b[1mSTEP\u001b[0m: Collecting resource usage data\nJan 10 18:32:38.443: INFO: Closed stop channel. Waiting for 3 workers\nJan 10 18:32:38.443: INFO: Closing worker for 172.16.132.155\nJan 10 18:32:38.443: INFO: Closing worker for 172.16.132.148\nJan 10 18:32:38.443: INFO: Closing worker for 172.16.132.150\nJan 10 18:32:38.443: INFO: Waitgroup finished.\nJan 10 18:32:38.443: INFO: Unknown output type: . Skipping.\nJan 10 18:32:38.443: INFO: Waiting up to 1m0s for all nodes to be ready\n\u001b[1mSTEP\u001b[0m: Destroying namespace \"e2e-tests-emptydir-lsclc\" for this suite.\n\n\u001b[32m• [SLOW TEST:9.091 seconds]\u001b[0m\n[k8s.io] EmptyDir volumes\n\u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:690\u001b[0m\n should support (non-root,0644,tmpfs) [Conformance]\n \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:81\u001b[0m\n\u001b[90m------------------------------\u001b[0m\n\u001b[0m[k8s.io] EmptyDir volumes\u001b[0m \n \u001b[1mshould support (non-root,0666,tmpfs) [Conformance]\u001b[0m\n \u001b[37m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:85\u001b[0m\n\u001b[1mSTEP\u001b[0m: Creating a kubernetes client\nJan 10 18:32:43.452: INFO: >>> kubeConfig: /etc/origin/master/admin.kubeconfig\n\n\u001b[1mSTEP\u001b[0m: Building a namespace api object\nJan 10 18:32:43.490: INFO: About to run a Kube e2e test, ensuring namespace is privileged\n\u001b[1mSTEP\u001b[0m: Waiting for a default service account to be provisioned in namespace\n\u001b[1mSTEP\u001b[0m: Creating a pod to test emptydir 0666 on tmpfs\nJan 10 18:32:43.537: INFO: Waiting up to 5m0s for pod pod-15aa8a5f-d78d-11e6-8a95-fa163e61d47e status to be success or failure\nJan 10 18:32:43.538: INFO: No Status.Info for container 'test-container' in pod 'pod-15aa8a5f-d78d-11e6-8a95-fa163e61d47e' yet\nJan 10 18:32:43.538: INFO: Waiting for pod pod-15aa8a5f-d78d-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-ecfrt' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (1.597589ms elapsed)\nJan 10 18:32:45.541: INFO: Nil State.Terminated for container 'test-container' in pod 'pod-15aa8a5f-d78d-11e6-8a95-fa163e61d47e' in namespace 'e2e-tests-emptydir-ecfrt' so far\nJan 10 18:32:45.541: INFO: Waiting for pod pod-15aa8a5f-d78d-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-ecfrt' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (2.004118652s elapsed)\n\u001b[1mSTEP\u001b[0m: Saw pod success\n\u001b[1mSTEP\u001b[0m: Trying to get logs from node 172.16.132.150 pod pod-15aa8a5f-d78d-11e6-8a95-fa163e61d47e container test-container: \n\u001b[1mSTEP\u001b[0m: Successfully fetched pod logs:mount type of \"/test-volume\": tmpfs\ncontent of file \"/test-volume/test-file\": mount-tester new file\n\nperms of file \"/test-volume/test-file\": -rw-rw-rw-\n\n\n\u001b[1mSTEP\u001b[0m: Collecting resource usage data\nJan 10 18:32:47.591: INFO: Closed stop channel. Waiting for 3 workers\nJan 10 18:32:47.591: INFO: Closing worker for 172.16.132.155\nJan 10 18:32:47.591: INFO: Closing worker for 172.16.132.150\nJan 10 18:32:47.591: INFO: Closing worker for 172.16.132.148\nJan 10 18:32:47.591: INFO: Waitgroup finished.\nJan 10 18:32:47.591: INFO: Unknown output type: . Skipping.\nJan 10 18:32:47.591: INFO: Waiting up to 1m0s for all nodes to be ready\n\u001b[1mSTEP\u001b[0m: Destroying namespace \"e2e-tests-emptydir-ecfrt\" for this suite.\n\n\u001b[32m• [SLOW TEST:9.150 seconds]\u001b[0m\n[k8s.io] EmptyDir volumes\n\u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:690\u001b[0m\n should support (non-root,0666,tmpfs) [Conformance]\n \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:85\u001b[0m\n\u001b[90m------------------------------\u001b[0m\n\u001b[0m[k8s.io] EmptyDir volumes\u001b[0m \n \u001b[1mshould support (non-root,0777,tmpfs) [Conformance]\u001b[0m\n \u001b[37m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:89\u001b[0m\n\u001b[1mSTEP\u001b[0m: Creating a kubernetes client\nJan 10 18:32:52.602: INFO: >>> kubeConfig: /etc/origin/master/admin.kubeconfig\n\n\u001b[1mSTEP\u001b[0m: Building a namespace api object\nJan 10 18:32:52.624: INFO: About to run a Kube e2e test, ensuring namespace is privileged\n\u001b[1mSTEP\u001b[0m: Waiting for a default service account to be provisioned in namespace\n\u001b[1mSTEP\u001b[0m: Creating a pod to test emptydir 0777 on tmpfs\nJan 10 18:32:52.671: INFO: Waiting up to 5m0s for pod pod-1b1ce8f8-d78d-11e6-8a95-fa163e61d47e status to be success or failure\nJan 10 18:32:52.674: INFO: No Status.Info for container 'test-container' in pod 'pod-1b1ce8f8-d78d-11e6-8a95-fa163e61d47e' yet\nJan 10 18:32:52.674: INFO: Waiting for pod pod-1b1ce8f8-d78d-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-xx63m' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (3.125197ms elapsed)\n\u001b[1mSTEP\u001b[0m: Saw pod success\n\u001b[1mSTEP\u001b[0m: Trying to get logs from node 172.16.132.148 pod pod-1b1ce8f8-d78d-11e6-8a95-fa163e61d47e container test-container: \n\u001b[1mSTEP\u001b[0m: Successfully fetched pod logs:mount type of \"/test-volume\": tmpfs\ncontent of file \"/test-volume/test-file\": mount-tester new file\n\nperms of file \"/test-volume/test-file\": -rwxrwxrwx\n\n\n\u001b[1mSTEP\u001b[0m: Collecting resource usage data\nJan 10 18:32:54.691: INFO: Closed stop channel. Waiting for 3 workers\nJan 10 18:32:54.691: INFO: Closing worker for 172.16.132.150\nJan 10 18:32:54.691: INFO: Closing worker for 172.16.132.148\nJan 10 18:32:54.691: INFO: Closing worker for 172.16.132.155\nJan 10 18:32:54.691: INFO: Waitgroup finished.\nJan 10 18:32:54.691: INFO: Unknown output type: . Skipping.\nJan 10 18:32:54.691: INFO: Waiting up to 1m0s for all nodes to be ready\n\u001b[1mSTEP\u001b[0m: Destroying namespace \"e2e-tests-emptydir-xx63m\" for this suite.\n\n\u001b[32m• [SLOW TEST:7.099 seconds]\u001b[0m\n[k8s.io] EmptyDir volumes\n\u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:690\u001b[0m\n should support (non-root,0777,tmpfs) [Conformance]\n \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:89\u001b[0m\n\u001b[90m------------------------------\u001b[0m\n\u001b[0m[k8s.io] EmptyDir volumes\u001b[0m \n \u001b[1mvolume on default medium should have the correct mode [Conformance]\u001b[0m\n \u001b[37m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:93\u001b[0m\n\u001b[1mSTEP\u001b[0m: Creating a kubernetes client\nJan 10 18:32:59.700: INFO: >>> kubeConfig: /etc/origin/master/admin.kubeconfig\n\n\u001b[1mSTEP\u001b[0m: Building a namespace api object\nJan 10 18:32:59.772: INFO: About to run a Kube e2e test, ensuring namespace is privileged\n\u001b[1mSTEP\u001b[0m: Waiting for a default service account to be provisioned in namespace\n\u001b[1mSTEP\u001b[0m: Creating a pod to test emptydir volume type on node default medium\nJan 10 18:32:59.837: INFO: Waiting up to 5m0s for pod pod-1f62f945-d78d-11e6-8a95-fa163e61d47e status to be success or failure\nJan 10 18:32:59.853: INFO: No Status.Info for container 'test-container' in pod 'pod-1f62f945-d78d-11e6-8a95-fa163e61d47e' yet\nJan 10 18:32:59.853: INFO: Waiting for pod pod-1f62f945-d78d-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-n5tg3' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (15.838227ms elapsed)\n\u001b[1mSTEP\u001b[0m: Saw pod success\n\u001b[1mSTEP\u001b[0m: Trying to get logs from node 172.16.132.150 pod pod-1f62f945-d78d-11e6-8a95-fa163e61d47e container test-container: \n\u001b[1mSTEP\u001b[0m: Successfully fetched pod logs:mount type of \"/test-volume\": 1481003842\nperms of file \"/test-volume\": -rwxrwxrwx\n\n\n\u001b[1mSTEP\u001b[0m: Collecting resource usage data\nJan 10 18:33:01.870: INFO: Closed stop channel. Waiting for 3 workers\nJan 10 18:33:01.870: INFO: Closing worker for 172.16.132.155\nJan 10 18:33:01.870: INFO: Closing worker for 172.16.132.150\nJan 10 18:33:01.870: INFO: Closing worker for 172.16.132.148\nJan 10 18:33:01.870: INFO: Waitgroup finished.\nJan 10 18:33:01.870: INFO: Unknown output type: . Skipping.\nJan 10 18:33:01.870: INFO: Waiting up to 1m0s for all nodes to be ready\n\u001b[1mSTEP\u001b[0m: Destroying namespace \"e2e-tests-emptydir-n5tg3\" for this suite.\n\n\u001b[32m• [SLOW TEST:7.177 seconds]\u001b[0m\n[k8s.io] EmptyDir volumes\n\u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:690\u001b[0m\n volume on default medium should have the correct mode [Conformance]\n \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:93\u001b[0m\n\u001b[90m------------------------------\u001b[0m\n\u001b[0m[k8s.io] EmptyDir volumes\u001b[0m \n \u001b[1mshould support (root,0644,default) [Conformance]\u001b[0m\n \u001b[37m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:97\u001b[0m\n\u001b[1mSTEP\u001b[0m: Creating a kubernetes client\nJan 10 18:33:06.878: INFO: >>> kubeConfig: /etc/origin/master/admin.kubeconfig\n\n\u001b[1mSTEP\u001b[0m: Building a namespace api object\nJan 10 18:33:06.902: INFO: About to run a Kube e2e test, ensuring namespace is privileged\n\u001b[1mSTEP\u001b[0m: Waiting for a default service account to be provisioned in namespace\n\u001b[1mSTEP\u001b[0m: Creating a pod to test emptydir 0644 on node default medium\nJan 10 18:33:06.938: INFO: Waiting up to 5m0s for pod pod-239e56ae-d78d-11e6-8a95-fa163e61d47e status to be success or failure\nJan 10 18:33:06.940: INFO: No Status.Info for container 'test-container' in pod 'pod-239e56ae-d78d-11e6-8a95-fa163e61d47e' yet\nJan 10 18:33:06.940: INFO: Waiting for pod pod-239e56ae-d78d-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-xy29m' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (2.451514ms elapsed)\nJan 10 18:33:08.944: INFO: Nil State.Terminated for container 'test-container' in pod 'pod-239e56ae-d78d-11e6-8a95-fa163e61d47e' in namespace 'e2e-tests-emptydir-xy29m' so far\nJan 10 18:33:08.944: INFO: Waiting for pod pod-239e56ae-d78d-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-xy29m' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (2.005639937s elapsed)\n\u001b[1mSTEP\u001b[0m: Saw pod success\n\u001b[1mSTEP\u001b[0m: Trying to get logs from node 172.16.132.148 pod pod-239e56ae-d78d-11e6-8a95-fa163e61d47e container test-container: \n\u001b[1mSTEP\u001b[0m: Successfully fetched pod logs:mount type of \"/test-volume\": 1481003842\ncontent of file \"/test-volume/test-file\": mount-tester new file\n\nperms of file \"/test-volume/test-file\": -rw-r--r--\n\n\n\u001b[1mSTEP\u001b[0m: Collecting resource usage data\nJan 10 18:33:10.959: INFO: Closed stop channel. Waiting for 3 workers\nJan 10 18:33:10.959: INFO: Closing worker for 172.16.132.148\nJan 10 18:33:10.959: INFO: Closing worker for 172.16.132.155\nJan 10 18:33:10.959: INFO: Closing worker for 172.16.132.150\nJan 10 18:33:10.959: INFO: Waitgroup finished.\nJan 10 18:33:10.959: INFO: Unknown output type: . Skipping.\nJan 10 18:33:10.959: INFO: Waiting up to 1m0s for all nodes to be ready\n\u001b[1mSTEP\u001b[0m: Destroying namespace \"e2e-tests-emptydir-xy29m\" for this suite.\n\n\u001b[32m• [SLOW TEST:9.090 seconds]\u001b[0m\n[k8s.io] EmptyDir volumes\n\u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:690\u001b[0m\n should support (root,0644,default) [Conformance]\n \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:97\u001b[0m\n\u001b[90m------------------------------\u001b[0m\n\u001b[0m[k8s.io] EmptyDir volumes\u001b[0m \n \u001b[1mshould support (root,0666,default) [Conformance]\u001b[0m\n \u001b[37m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:101\u001b[0m\n\u001b[1mSTEP\u001b[0m: Creating a kubernetes client\nJan 10 18:33:15.968: INFO: >>> kubeConfig: /etc/origin/master/admin.kubeconfig\n\n\u001b[1mSTEP\u001b[0m: Building a namespace api object\nJan 10 18:33:16.035: INFO: About to run a Kube e2e test, ensuring namespace is privileged\n\u001b[1mSTEP\u001b[0m: Waiting for a default service account to be provisioned in namespace\n\u001b[1mSTEP\u001b[0m: Creating a pod to test emptydir 0666 on node default medium\nJan 10 18:33:16.077: INFO: Waiting up to 5m0s for pod pod-29111d02-d78d-11e6-8a95-fa163e61d47e status to be success or failure\nJan 10 18:33:16.081: INFO: No Status.Info for container 'test-container' in pod 'pod-29111d02-d78d-11e6-8a95-fa163e61d47e' yet\nJan 10 18:33:16.081: INFO: Waiting for pod pod-29111d02-d78d-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-u5046' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (3.792893ms elapsed)\n\u001b[1mSTEP\u001b[0m: Saw pod success\n\u001b[1mSTEP\u001b[0m: Trying to get logs from node 172.16.132.150 pod pod-29111d02-d78d-11e6-8a95-fa163e61d47e container test-container: \n\u001b[1mSTEP\u001b[0m: Successfully fetched pod logs:mount type of \"/test-volume\": 1481003842\ncontent of file \"/test-volume/test-file\": mount-tester new file\n\nperms of file \"/test-volume/test-file\": -rw-rw-rw-\n\n\n\u001b[1mSTEP\u001b[0m: Collecting resource usage data\nJan 10 18:33:18.101: INFO: Closed stop channel. Waiting for 3 workers\nJan 10 18:33:18.101: INFO: Closing worker for 172.16.132.150\nJan 10 18:33:18.101: INFO: Closing worker for 172.16.132.148\nJan 10 18:33:18.101: INFO: Closing worker for 172.16.132.155\nJan 10 18:33:18.101: INFO: Waitgroup finished.\nJan 10 18:33:18.101: INFO: Unknown output type: . Skipping.\nJan 10 18:33:18.101: INFO: Waiting up to 1m0s for all nodes to be ready\n\u001b[1mSTEP\u001b[0m: Destroying namespace \"e2e-tests-emptydir-u5046\" for this suite.\n\n\u001b[32m• [SLOW TEST:7.147 seconds]\u001b[0m\n[k8s.io] EmptyDir volumes\n\u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:690\u001b[0m\n should support (root,0666,default) [Conformance]\n \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:101\u001b[0m\n\u001b[90m------------------------------\u001b[0m\n\u001b[0m[k8s.io] EmptyDir volumes\u001b[0m \n \u001b[1mshould support (root,0777,default) [Conformance]\u001b[0m\n \u001b[37m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:105\u001b[0m\n\u001b[1mSTEP\u001b[0m: Creating a kubernetes client\nJan 10 18:33:23.116: INFO: >>> kubeConfig: /etc/origin/master/admin.kubeconfig\n\n\u001b[1mSTEP\u001b[0m: Building a namespace api object\nJan 10 18:33:23.138: INFO: About to run a Kube e2e test, ensuring namespace is privileged\n\u001b[1mSTEP\u001b[0m: Waiting for a default service account to be provisioned in namespace\n\u001b[1mSTEP\u001b[0m: Creating a pod to test emptydir 0777 on node default medium\nJan 10 18:33:23.165: INFO: Waiting up to 5m0s for pod pod-2d4ae08d-d78d-11e6-8a95-fa163e61d47e status to be success or failure\nJan 10 18:33:23.166: INFO: No Status.Info for container 'test-container' in pod 'pod-2d4ae08d-d78d-11e6-8a95-fa163e61d47e' yet\nJan 10 18:33:23.166: INFO: Waiting for pod pod-2d4ae08d-d78d-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-m69vd' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (1.189333ms elapsed)\n\u001b[1mSTEP\u001b[0m: Saw pod success\n\u001b[1mSTEP\u001b[0m: Trying to get logs from node 172.16.132.148 pod pod-2d4ae08d-d78d-11e6-8a95-fa163e61d47e container test-container: \n\u001b[1mSTEP\u001b[0m: Successfully fetched pod logs:mount type of \"/test-volume\": 1481003842\ncontent of file \"/test-volume/test-file\": mount-tester new file\n\nperms of file \"/test-volume/test-file\": -rwxrwxrwx\n\n\n\u001b[1mSTEP\u001b[0m: Collecting resource usage data\nJan 10 18:33:25.183: INFO: Closed stop channel. Waiting for 3 workers\nJan 10 18:33:25.183: INFO: Closing worker for 172.16.132.155\nJan 10 18:33:25.183: INFO: Closing worker for 172.16.132.150\nJan 10 18:33:25.183: INFO: Closing worker for 172.16.132.148\nJan 10 18:33:25.183: INFO: Waitgroup finished.\nJan 10 18:33:25.183: INFO: Unknown output type: . Skipping.\nJan 10 18:33:25.183: INFO: Waiting up to 1m0s for all nodes to be ready\n\u001b[1mSTEP\u001b[0m: Destroying namespace \"e2e-tests-emptydir-m69vd\" for this suite.\n\n\u001b[32m• [SLOW TEST:7.081 seconds]\u001b[0m\n[k8s.io] EmptyDir volumes\n\u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:690\u001b[0m\n should support (root,0777,default) [Conformance]\n \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:105\u001b[0m\n\u001b[90m------------------------------\u001b[0m\n\u001b[0m[k8s.io] EmptyDir volumes\u001b[0m \n \u001b[1mshould support (non-root,0644,default) [Conformance]\u001b[0m\n \u001b[37m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:109\u001b[0m\n\u001b[1mSTEP\u001b[0m: Creating a kubernetes client\nJan 10 18:33:30.197: INFO: >>> kubeConfig: /etc/origin/master/admin.kubeconfig\n\n\u001b[1mSTEP\u001b[0m: Building a namespace api object\nJan 10 18:33:30.227: INFO: About to run a Kube e2e test, ensuring namespace is privileged\n\u001b[1mSTEP\u001b[0m: Waiting for a default service account to be provisioned in namespace\n\u001b[1mSTEP\u001b[0m: Creating a pod to test emptydir 0644 on node default medium\nJan 10 18:33:30.272: INFO: Waiting up to 5m0s for pod pod-318732e2-d78d-11e6-8a95-fa163e61d47e status to be success or failure\nJan 10 18:33:30.277: INFO: No Status.Info for container 'test-container' in pod 'pod-318732e2-d78d-11e6-8a95-fa163e61d47e' yet\nJan 10 18:33:30.277: INFO: Waiting for pod pod-318732e2-d78d-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-ojjux' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (5.163181ms elapsed)\n\u001b[1mSTEP\u001b[0m: Saw pod success\n\u001b[1mSTEP\u001b[0m: Trying to get logs from node 172.16.132.150 pod pod-318732e2-d78d-11e6-8a95-fa163e61d47e container test-container: \n\u001b[1mSTEP\u001b[0m: Successfully fetched pod logs:mount type of \"/test-volume\": 1481003842\ncontent of file \"/test-volume/test-file\": mount-tester new file\n\nperms of file \"/test-volume/test-file\": -rw-r--r--\n\n\n\u001b[1mSTEP\u001b[0m: Collecting resource usage data\nJan 10 18:33:32.292: INFO: Closed stop channel. Waiting for 3 workers\nJan 10 18:33:32.292: INFO: Closing worker for 172.16.132.155\nJan 10 18:33:32.292: INFO: Closing worker for 172.16.132.150\nJan 10 18:33:32.292: INFO: Closing worker for 172.16.132.148\nJan 10 18:33:32.292: INFO: Waitgroup finished.\nJan 10 18:33:32.292: INFO: Unknown output type: . Skipping.\nJan 10 18:33:32.292: INFO: Waiting up to 1m0s for all nodes to be ready\n\u001b[1mSTEP\u001b[0m: Destroying namespace \"e2e-tests-emptydir-ojjux\" for this suite.\n\n\u001b[32m• [SLOW TEST:7.104 seconds]\u001b[0m\n[k8s.io] EmptyDir volumes\n\u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:690\u001b[0m\n should support (non-root,0644,default) [Conformance]\n \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:109\u001b[0m\n\u001b[90m------------------------------\u001b[0m\n\u001b[0m[k8s.io] EmptyDir volumes\u001b[0m \n \u001b[1mshould support (non-root,0666,default) [Conformance]\u001b[0m\n \u001b[37m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:113\u001b[0m\n\u001b[1mSTEP\u001b[0m: Creating a kubernetes client\nJan 10 18:33:37.301: INFO: >>> kubeConfig: /etc/origin/master/admin.kubeconfig\n\n\u001b[1mSTEP\u001b[0m: Building a namespace api object\nJan 10 18:33:37.327: INFO: About to run a Kube e2e test, ensuring namespace is privileged\n\u001b[1mSTEP\u001b[0m: Waiting for a default service account to be provisioned in namespace\n\u001b[1mSTEP\u001b[0m: Creating a pod to test emptydir 0666 on node default medium\nJan 10 18:33:37.358: INFO: Waiting up to 5m0s for pod pod-35bfd2f2-d78d-11e6-8a95-fa163e61d47e status to be success or failure\nJan 10 18:33:37.361: INFO: No Status.Info for container 'test-container' in pod 'pod-35bfd2f2-d78d-11e6-8a95-fa163e61d47e' yet\nJan 10 18:33:37.361: INFO: Waiting for pod pod-35bfd2f2-d78d-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-n3m8r' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (3.220896ms elapsed)\n\u001b[1mSTEP\u001b[0m: Saw pod success\n\u001b[1mSTEP\u001b[0m: Trying to get logs from node 172.16.132.148 pod pod-35bfd2f2-d78d-11e6-8a95-fa163e61d47e container test-container: \n\u001b[1mSTEP\u001b[0m: Successfully fetched pod logs:mount type of \"/test-volume\": 1481003842\ncontent of file \"/test-volume/test-file\": mount-tester new file\n\nperms of file \"/test-volume/test-file\": -rw-rw-rw-\n\n\n\u001b[1mSTEP\u001b[0m: Collecting resource usage data\nJan 10 18:33:39.376: INFO: Closed stop channel. Waiting for 3 workers\nJan 10 18:33:39.376: INFO: Closing worker for 172.16.132.155\nJan 10 18:33:39.376: INFO: Closing worker for 172.16.132.150\nJan 10 18:33:39.376: INFO: Closing worker for 172.16.132.148\nJan 10 18:33:39.376: INFO: Waitgroup finished.\nJan 10 18:33:39.376: INFO: Unknown output type: . Skipping.\nJan 10 18:33:39.376: INFO: Waiting up to 1m0s for all nodes to be ready\n\u001b[1mSTEP\u001b[0m: Destroying namespace \"e2e-tests-emptydir-n3m8r\" for this suite.\n\n\u001b[32m• [SLOW TEST:7.083 seconds]\u001b[0m\n[k8s.io] EmptyDir volumes\n\u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:690\u001b[0m\n should support (non-root,0666,default) [Conformance]\n \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:113\u001b[0m\n\u001b[90m------------------------------\u001b[0m\n\u001b[0m[k8s.io] EmptyDir volumes\u001b[0m \n \u001b[1mshould support (non-root,0777,default) [Conformance]\u001b[0m\n \u001b[37m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:117\u001b[0m\n\u001b[1mSTEP\u001b[0m: Creating a kubernetes client\nJan 10 18:33:44.384: INFO: >>> kubeConfig: /etc/origin/master/admin.kubeconfig\n\n\u001b[1mSTEP\u001b[0m: Building a namespace api object\nJan 10 18:33:44.407: INFO: About to run a Kube e2e test, ensuring namespace is privileged\n\u001b[1mSTEP\u001b[0m: Waiting for a default service account to be provisioned in namespace\n\u001b[1mSTEP\u001b[0m: Creating a pod to test emptydir 0777 on node default medium\nJan 10 18:33:44.449: INFO: Waiting up to 5m0s for pod pod-39f9eed1-d78d-11e6-8a95-fa163e61d47e status to be success or failure\nJan 10 18:33:44.464: INFO: Nil State.Terminated for container 'test-container' in pod 'pod-39f9eed1-d78d-11e6-8a95-fa163e61d47e' in namespace 'e2e-tests-emptydir-docy2' so far\nJan 10 18:33:44.464: INFO: Waiting for pod pod-39f9eed1-d78d-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-docy2' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (14.373228ms elapsed)\n\u001b[1mSTEP\u001b[0m: Saw pod success\n\u001b[1mSTEP\u001b[0m: Trying to get logs from node 172.16.132.150 pod pod-39f9eed1-d78d-11e6-8a95-fa163e61d47e container test-container: \n\u001b[1mSTEP\u001b[0m: Successfully fetched pod logs:mount type of \"/test-volume\": 1481003842\ncontent of file \"/test-volume/test-file\": mount-tester new file\n\nperms of file \"/test-volume/test-file\": -rwxrwxrwx\n\n\n\u001b[1mSTEP\u001b[0m: Collecting resource usage data\nJan 10 18:33:46.497: INFO: Closed stop channel. Waiting for 3 workers\nJan 10 18:33:46.497: INFO: Closing worker for 172.16.132.148\nJan 10 18:33:46.497: INFO: Closing worker for 172.16.132.150\nJan 10 18:33:46.497: INFO: Closing worker for 172.16.132.155\nJan 10 18:33:46.497: INFO: Waitgroup finished.\nJan 10 18:33:46.497: INFO: Unknown output type: . Skipping.\nJan 10 18:33:46.497: INFO: Waiting up to 1m0s for all nodes to be ready\n\u001b[1mSTEP\u001b[0m: Destroying namespace \"e2e-tests-emptydir-docy2\" for this suite.\n\n\u001b[32m• [SLOW TEST:7.121 seconds]\u001b[0m\n[k8s.io] EmptyDir volumes\n\u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:690\u001b[0m\n should support (non-root,0777,default) [Conformance]\n \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:117\u001b[0m\n\u001b[90m------------------------------\u001b[0m\n\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\n\u001b[1m\u001b[32mRan 19 of 509 Specs in 155.161 seconds\u001b[0m\n\u001b[1m\u001b[32mSUCCESS!\u001b[0m -- \u001b[32m\u001b[1m19 Passed\u001b[0m | \u001b[91m\u001b[1m0 Failed\u001b[0m | \u001b[33m\u001b[1m0 Pending\u001b[0m | \u001b[36m\u001b[1m490 Skipped\u001b[0m PASS", "stdout_lines": [ "Running Suite: Extended", "=======================", "Random Seed: \u001b[1m1484091076\u001b[0m", "Will run \u001b[1m19\u001b[0m of \u001b[1m509\u001b[0m specs", "", "I0110 18:31:16.345825 23683 e2e.go:58] The --provider flag is not set. Treating as a conformance test. Some tests may not be run.", "Jan 10 18:31:16.345: INFO: >>> kubeConfig: /etc/origin/master/admin.kubeconfig", "", "Jan 10 18:31:16.348: INFO: Waiting up to 0 for all pods (need at least 0) in namespace 'kube-system' to be running and ready", "Jan 10 18:31:16.383: INFO: Waiting for pods to enter Success, but no pods in \"kube-system\" match label map[name:e2e-image-puller]", "Jan 10 18:31:16.383: INFO: 0 / 0 pods in namespace 'kube-system' are running and ready (0 seconds elapsed)", "Jan 10 18:31:16.383: INFO: expected 0 pod replicas in namespace 'kube-system', 0 are Running and Ready.", "Jan 10 18:31:16.384: INFO: Waiting for pods to enter Success, but no pods in \"kube-system\" match label map[name:e2e-image-puller]", "I0110 18:31:16.384692 23683 e2e.go:58] The --provider flag is not set. Treating as a conformance test. Some tests may not be run.", "\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m", "\u001b[90m------------------------------\u001b[0m", "\u001b[0m[k8s.io] EmptyDir volumes\u001b[0m \u001b[90mwhen FSGroup is specified [Feature:FSGroup]\u001b[0m ", " \u001b[1mnew files should be created with FSGroup ownership when container is root\u001b[0m", " \u001b[37m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:44\u001b[0m", "\u001b[1mSTEP\u001b[0m: Creating a kubernetes client", "Jan 10 18:31:16.385: INFO: >>> kubeConfig: /etc/origin/master/admin.kubeconfig", "", "\u001b[1mSTEP\u001b[0m: Building a namespace api object", "Jan 10 18:31:16.408: INFO: About to run a Kube e2e test, ensuring namespace is privileged", "\u001b[1mSTEP\u001b[0m: Waiting for a default service account to be provisioned in namespace", "\u001b[1mSTEP\u001b[0m: Creating a pod to test emptydir 0644 on tmpfs", "Jan 10 18:31:16.447: INFO: Waiting up to 5m0s for pod pod-e1c30011-d78c-11e6-8a95-fa163e61d47e status to be success or failure", "Jan 10 18:31:16.457: INFO: No Status.Info for container 'test-container' in pod 'pod-e1c30011-d78c-11e6-8a95-fa163e61d47e' yet", "Jan 10 18:31:16.457: INFO: Waiting for pod pod-e1c30011-d78c-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-njmyz' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (9.678773ms elapsed)", "Jan 10 18:31:18.459: INFO: Nil State.Terminated for container 'test-container' in pod 'pod-e1c30011-d78c-11e6-8a95-fa163e61d47e' in namespace 'e2e-tests-emptydir-njmyz' so far", "Jan 10 18:31:18.459: INFO: Waiting for pod pod-e1c30011-d78c-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-njmyz' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (2.011909436s elapsed)", "Jan 10 18:31:20.461: INFO: Nil State.Terminated for container 'test-container' in pod 'pod-e1c30011-d78c-11e6-8a95-fa163e61d47e' in namespace 'e2e-tests-emptydir-njmyz' so far", "Jan 10 18:31:20.461: INFO: Waiting for pod pod-e1c30011-d78c-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-njmyz' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (4.014110983s elapsed)", "\u001b[1mSTEP\u001b[0m: Saw pod success", "\u001b[1mSTEP\u001b[0m: Trying to get logs from node 172.16.132.150 pod pod-e1c30011-d78c-11e6-8a95-fa163e61d47e container test-container: ", "\u001b[1mSTEP\u001b[0m: Successfully fetched pod logs:mount type of \"/test-volume\": tmpfs", "content of file \"/test-volume/test-file\": mount-tester new file", "", "perms of file \"/test-volume/test-file\": -rw-rw----", "owner UID of \"/test-volume/test-file\": 0", "owner GID of \"/test-volume/test-file\": 123", "", "", "\u001b[1mSTEP\u001b[0m: Collecting resource usage data", "Jan 10 18:31:22.486: INFO: Closed stop channel. Waiting for 3 workers", "Jan 10 18:31:22.486: INFO: Closing worker for 172.16.132.155", "Jan 10 18:31:22.486: INFO: Closing worker for 172.16.132.150", "Jan 10 18:31:22.486: INFO: Closing worker for 172.16.132.148", "Jan 10 18:31:22.486: INFO: Waitgroup finished.", "Jan 10 18:31:22.486: INFO: Unknown output type: . Skipping.", "Jan 10 18:31:22.486: INFO: Waiting up to 1m0s for all nodes to be ready", "\u001b[1mSTEP\u001b[0m: Destroying namespace \"e2e-tests-emptydir-njmyz\" for this suite.", "", "\u001b[32m• [SLOW TEST:11.113 seconds]\u001b[0m", "[k8s.io] EmptyDir volumes", "\u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:690\u001b[0m", " when FSGroup is specified [Feature:FSGroup]", " \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:61\u001b[0m", " new files should be created with FSGroup ownership when container is root", " \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:44\u001b[0m", "\u001b[90m------------------------------\u001b[0m", "\u001b[0m[k8s.io] EmptyDir volumes\u001b[0m \u001b[90mwhen FSGroup is specified [Feature:FSGroup]\u001b[0m ", " \u001b[1mnew files should be created with FSGroup ownership when container is non-root\u001b[0m", " \u001b[37m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:48\u001b[0m", "\u001b[1mSTEP\u001b[0m: Creating a kubernetes client", "Jan 10 18:31:27.498: INFO: >>> kubeConfig: /etc/origin/master/admin.kubeconfig", "", "\u001b[1mSTEP\u001b[0m: Building a namespace api object", "Jan 10 18:31:27.530: INFO: About to run a Kube e2e test, ensuring namespace is privileged", "\u001b[1mSTEP\u001b[0m: Waiting for a default service account to be provisioned in namespace", "\u001b[1mSTEP\u001b[0m: Creating a pod to test emptydir 0644 on tmpfs", "Jan 10 18:31:27.561: INFO: Waiting up to 5m0s for pod pod-e862e652-d78c-11e6-8a95-fa163e61d47e status to be success or failure", "Jan 10 18:31:27.563: INFO: No Status.Info for container 'test-container' in pod 'pod-e862e652-d78c-11e6-8a95-fa163e61d47e' yet", "Jan 10 18:31:27.563: INFO: Waiting for pod pod-e862e652-d78c-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-yd8ef' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (1.233326ms elapsed)", "Jan 10 18:31:29.565: INFO: Nil State.Terminated for container 'test-container' in pod 'pod-e862e652-d78c-11e6-8a95-fa163e61d47e' in namespace 'e2e-tests-emptydir-yd8ef' so far", "Jan 10 18:31:29.565: INFO: Waiting for pod pod-e862e652-d78c-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-yd8ef' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (2.003996573s elapsed)", "Jan 10 18:31:31.568: INFO: Nil State.Terminated for container 'test-container' in pod 'pod-e862e652-d78c-11e6-8a95-fa163e61d47e' in namespace 'e2e-tests-emptydir-yd8ef' so far", "Jan 10 18:31:31.568: INFO: Waiting for pod pod-e862e652-d78c-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-yd8ef' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (4.006466774s elapsed)", "Jan 10 18:31:33.571: INFO: Nil State.Terminated for container 'test-container' in pod 'pod-e862e652-d78c-11e6-8a95-fa163e61d47e' in namespace 'e2e-tests-emptydir-yd8ef' so far", "Jan 10 18:31:33.571: INFO: Waiting for pod pod-e862e652-d78c-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-yd8ef' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (6.009226209s elapsed)", "\u001b[1mSTEP\u001b[0m: Saw pod success", "\u001b[1mSTEP\u001b[0m: Trying to get logs from node 172.16.132.148 pod pod-e862e652-d78c-11e6-8a95-fa163e61d47e container test-container: ", "\u001b[1mSTEP\u001b[0m: Successfully fetched pod logs:mount type of \"/test-volume\": tmpfs", "content of file \"/test-volume/test-file\": mount-tester new file", "", "perms of file \"/test-volume/test-file\": -rw-rw----", "owner UID of \"/test-volume/test-file\": 0", "owner GID of \"/test-volume/test-file\": 123", "", "", "\u001b[1mSTEP\u001b[0m: Collecting resource usage data", "Jan 10 18:31:35.587: INFO: Closed stop channel. Waiting for 3 workers", "Jan 10 18:31:35.587: INFO: Closing worker for 172.16.132.155", "Jan 10 18:31:35.587: INFO: Closing worker for 172.16.132.150", "Jan 10 18:31:35.587: INFO: Closing worker for 172.16.132.148", "Jan 10 18:31:35.587: INFO: Waitgroup finished.", "Jan 10 18:31:35.587: INFO: Unknown output type: . Skipping.", "Jan 10 18:31:35.587: INFO: Waiting up to 1m0s for all nodes to be ready", "\u001b[1mSTEP\u001b[0m: Destroying namespace \"e2e-tests-emptydir-yd8ef\" for this suite.", "", "\u001b[32m• [SLOW TEST:13.097 seconds]\u001b[0m", "[k8s.io] EmptyDir volumes", "\u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:690\u001b[0m", " when FSGroup is specified [Feature:FSGroup]", " \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:61\u001b[0m", " new files should be created with FSGroup ownership when container is non-root", " \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:48\u001b[0m", "\u001b[90m------------------------------\u001b[0m", "\u001b[0m[k8s.io] EmptyDir volumes\u001b[0m \u001b[90mwhen FSGroup is specified [Feature:FSGroup]\u001b[0m ", " \u001b[1mfiles with FSGroup ownership should support (root,0644,tmpfs)\u001b[0m", " \u001b[37m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:52\u001b[0m", "\u001b[1mSTEP\u001b[0m: Creating a kubernetes client", "Jan 10 18:31:40.595: INFO: >>> kubeConfig: /etc/origin/master/admin.kubeconfig", "", "\u001b[1mSTEP\u001b[0m: Building a namespace api object", "Jan 10 18:31:40.620: INFO: About to run a Kube e2e test, ensuring namespace is privileged", "\u001b[1mSTEP\u001b[0m: Waiting for a default service account to be provisioned in namespace", "\u001b[1mSTEP\u001b[0m: Creating a pod to test emptydir 0644 on tmpfs", "Jan 10 18:31:40.667: INFO: Waiting up to 5m0s for pod pod-f031fdd8-d78c-11e6-8a95-fa163e61d47e status to be success or failure", "Jan 10 18:31:40.669: INFO: No Status.Info for container 'test-container' in pod 'pod-f031fdd8-d78c-11e6-8a95-fa163e61d47e' yet", "Jan 10 18:31:40.669: INFO: Waiting for pod pod-f031fdd8-d78c-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-emin5' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (2.723887ms elapsed)", "\u001b[1mSTEP\u001b[0m: Saw pod success", "\u001b[1mSTEP\u001b[0m: Trying to get logs from node 172.16.132.150 pod pod-f031fdd8-d78c-11e6-8a95-fa163e61d47e container test-container: ", "\u001b[1mSTEP\u001b[0m: Successfully fetched pod logs:mount type of \"/test-volume\": tmpfs", "content of file \"/test-volume/test-file\": mount-tester new file", "", "perms of file \"/test-volume/test-file\": -rw-r--r--", "", "", "\u001b[1mSTEP\u001b[0m: Collecting resource usage data", "Jan 10 18:31:42.687: INFO: Closed stop channel. Waiting for 3 workers", "Jan 10 18:31:42.687: INFO: Closing worker for 172.16.132.155", "Jan 10 18:31:42.687: INFO: Closing worker for 172.16.132.150", "Jan 10 18:31:42.687: INFO: Closing worker for 172.16.132.148", "Jan 10 18:31:42.687: INFO: Waitgroup finished.", "Jan 10 18:31:42.687: INFO: Unknown output type: . Skipping.", "Jan 10 18:31:42.687: INFO: Waiting up to 1m0s for all nodes to be ready", "\u001b[1mSTEP\u001b[0m: Destroying namespace \"e2e-tests-emptydir-emin5\" for this suite.", "", "\u001b[32m• [SLOW TEST:7.101 seconds]\u001b[0m", "[k8s.io] EmptyDir volumes", "\u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:690\u001b[0m", " when FSGroup is specified [Feature:FSGroup]", " \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:61\u001b[0m", " files with FSGroup ownership should support (root,0644,tmpfs)", " \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:52\u001b[0m", "\u001b[90m------------------------------\u001b[0m", "\u001b[0m[k8s.io] EmptyDir volumes\u001b[0m \u001b[90mwhen FSGroup is specified [Feature:FSGroup]\u001b[0m ", " \u001b[1mvolume on default medium should have the correct mode using FSGroup\u001b[0m", " \u001b[37m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:56\u001b[0m", "\u001b[1mSTEP\u001b[0m: Creating a kubernetes client", "Jan 10 18:31:47.696: INFO: >>> kubeConfig: /etc/origin/master/admin.kubeconfig", "", "\u001b[1mSTEP\u001b[0m: Building a namespace api object", "Jan 10 18:31:47.719: INFO: About to run a Kube e2e test, ensuring namespace is privileged", "\u001b[1mSTEP\u001b[0m: Waiting for a default service account to be provisioned in namespace", "\u001b[1mSTEP\u001b[0m: Creating a pod to test emptydir volume type on node default medium", "Jan 10 18:31:47.747: INFO: Waiting up to 5m0s for pod pod-f46afde4-d78c-11e6-8a95-fa163e61d47e status to be success or failure", "Jan 10 18:31:47.761: INFO: No Status.Info for container 'test-container' in pod 'pod-f46afde4-d78c-11e6-8a95-fa163e61d47e' yet", "Jan 10 18:31:47.761: INFO: Waiting for pod pod-f46afde4-d78c-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-oiv0k' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (14.862636ms elapsed)", "\u001b[1mSTEP\u001b[0m: Saw pod success", "\u001b[1mSTEP\u001b[0m: Trying to get logs from node 172.16.132.148 pod pod-f46afde4-d78c-11e6-8a95-fa163e61d47e container test-container: ", "\u001b[1mSTEP\u001b[0m: Successfully fetched pod logs:mount type of \"/test-volume\": 1481003842", "perms of file \"/test-volume\": -rwxrwxrwx", "", "", "\u001b[1mSTEP\u001b[0m: Collecting resource usage data", "Jan 10 18:31:49.776: INFO: Closed stop channel. Waiting for 3 workers", "Jan 10 18:31:49.776: INFO: Closing worker for 172.16.132.155", "Jan 10 18:31:49.776: INFO: Closing worker for 172.16.132.150", "Jan 10 18:31:49.776: INFO: Closing worker for 172.16.132.148", "Jan 10 18:31:49.776: INFO: Waitgroup finished.", "Jan 10 18:31:49.776: INFO: Unknown output type: . Skipping.", "Jan 10 18:31:49.776: INFO: Waiting up to 1m0s for all nodes to be ready", "\u001b[1mSTEP\u001b[0m: Destroying namespace \"e2e-tests-emptydir-oiv0k\" for this suite.", "", "\u001b[32m• [SLOW TEST:7.088 seconds]\u001b[0m", "[k8s.io] EmptyDir volumes", "\u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:690\u001b[0m", " when FSGroup is specified [Feature:FSGroup]", " \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:61\u001b[0m", " volume on default medium should have the correct mode using FSGroup", " \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:56\u001b[0m", "\u001b[90m------------------------------\u001b[0m", "\u001b[0m[k8s.io] EmptyDir volumes\u001b[0m \u001b[90mwhen FSGroup is specified [Feature:FSGroup]\u001b[0m ", " \u001b[1mvolume on tmpfs should have the correct mode using FSGroup\u001b[0m", " \u001b[37m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:60\u001b[0m", "\u001b[1mSTEP\u001b[0m: Creating a kubernetes client", "Jan 10 18:31:54.784: INFO: >>> kubeConfig: /etc/origin/master/admin.kubeconfig", "", "\u001b[1mSTEP\u001b[0m: Building a namespace api object", "Jan 10 18:31:54.815: INFO: About to run a Kube e2e test, ensuring namespace is privileged", "\u001b[1mSTEP\u001b[0m: Waiting for a default service account to be provisioned in namespace", "\u001b[1mSTEP\u001b[0m: Creating a pod to test emptydir volume type on tmpfs", "Jan 10 18:31:54.845: INFO: Waiting up to 5m0s for pod pod-f8a62e99-d78c-11e6-8a95-fa163e61d47e status to be success or failure", "Jan 10 18:31:54.847: INFO: No Status.Info for container 'test-container' in pod 'pod-f8a62e99-d78c-11e6-8a95-fa163e61d47e' yet", "Jan 10 18:31:54.847: INFO: Waiting for pod pod-f8a62e99-d78c-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-d5s9w' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (2.337756ms elapsed)", "\u001b[1mSTEP\u001b[0m: Saw pod success", "\u001b[1mSTEP\u001b[0m: Trying to get logs from node 172.16.132.150 pod pod-f8a62e99-d78c-11e6-8a95-fa163e61d47e container test-container: ", "\u001b[1mSTEP\u001b[0m: Successfully fetched pod logs:mount type of \"/test-volume\": tmpfs", "perms of file \"/test-volume\": -rwxrwxrwx", "", "", "\u001b[1mSTEP\u001b[0m: Collecting resource usage data", "Jan 10 18:31:56.869: INFO: Closed stop channel. Waiting for 3 workers", "Jan 10 18:31:56.869: INFO: Closing worker for 172.16.132.155", "Jan 10 18:31:56.869: INFO: Closing worker for 172.16.132.148", "Jan 10 18:31:56.869: INFO: Closing worker for 172.16.132.150", "Jan 10 18:31:56.869: INFO: Waitgroup finished.", "Jan 10 18:31:56.869: INFO: Unknown output type: . Skipping.", "Jan 10 18:31:56.869: INFO: Waiting up to 1m0s for all nodes to be ready", "\u001b[1mSTEP\u001b[0m: Destroying namespace \"e2e-tests-emptydir-d5s9w\" for this suite.", "", "\u001b[32m• [SLOW TEST:7.094 seconds]\u001b[0m", "[k8s.io] EmptyDir volumes", "\u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:690\u001b[0m", " when FSGroup is specified [Feature:FSGroup]", " \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:61\u001b[0m", " volume on tmpfs should have the correct mode using FSGroup", " \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:60\u001b[0m", "\u001b[90m------------------------------\u001b[0m", "\u001b[0m[k8s.io] EmptyDir volumes\u001b[0m ", " \u001b[1mvolume on tmpfs should have the correct mode [Conformance]\u001b[0m", " \u001b[37m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:65\u001b[0m", "\u001b[1mSTEP\u001b[0m: Creating a kubernetes client", "Jan 10 18:32:01.878: INFO: >>> kubeConfig: /etc/origin/master/admin.kubeconfig", "", "\u001b[1mSTEP\u001b[0m: Building a namespace api object", "Jan 10 18:32:01.910: INFO: About to run a Kube e2e test, ensuring namespace is privileged", "\u001b[1mSTEP\u001b[0m: Waiting for a default service account to be provisioned in namespace", "\u001b[1mSTEP\u001b[0m: Creating a pod to test emptydir volume type on tmpfs", "Jan 10 18:32:01.990: INFO: Waiting up to 5m0s for pod pod-fce6d939-d78c-11e6-8a95-fa163e61d47e status to be success or failure", "Jan 10 18:32:01.995: INFO: No Status.Info for container 'test-container' in pod 'pod-fce6d939-d78c-11e6-8a95-fa163e61d47e' yet", "Jan 10 18:32:01.995: INFO: Waiting for pod pod-fce6d939-d78c-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-4w1ts' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (5.113696ms elapsed)", "Jan 10 18:32:03.998: INFO: Nil State.Terminated for container 'test-container' in pod 'pod-fce6d939-d78c-11e6-8a95-fa163e61d47e' in namespace 'e2e-tests-emptydir-4w1ts' so far", "Jan 10 18:32:03.998: INFO: Waiting for pod pod-fce6d939-d78c-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-4w1ts' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (2.007629958s elapsed)", "\u001b[1mSTEP\u001b[0m: Saw pod success", "\u001b[1mSTEP\u001b[0m: Trying to get logs from node 172.16.132.148 pod pod-fce6d939-d78c-11e6-8a95-fa163e61d47e container test-container: ", "\u001b[1mSTEP\u001b[0m: Successfully fetched pod logs:mount type of \"/test-volume\": tmpfs", "perms of file \"/test-volume\": -rwxrwxrwx", "", "", "\u001b[1mSTEP\u001b[0m: Collecting resource usage data", "Jan 10 18:32:06.013: INFO: Closed stop channel. Waiting for 3 workers", "Jan 10 18:32:06.013: INFO: Closing worker for 172.16.132.150", "Jan 10 18:32:06.013: INFO: Closing worker for 172.16.132.148", "Jan 10 18:32:06.013: INFO: Closing worker for 172.16.132.155", "Jan 10 18:32:06.013: INFO: Waitgroup finished.", "Jan 10 18:32:06.013: INFO: Unknown output type: . Skipping.", "Jan 10 18:32:06.013: INFO: Waiting up to 1m0s for all nodes to be ready", "\u001b[1mSTEP\u001b[0m: Destroying namespace \"e2e-tests-emptydir-4w1ts\" for this suite.", "", "\u001b[32m• [SLOW TEST:9.145 seconds]\u001b[0m", "[k8s.io] EmptyDir volumes", "\u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:690\u001b[0m", " volume on tmpfs should have the correct mode [Conformance]", " \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:65\u001b[0m", "\u001b[90m------------------------------\u001b[0m", "\u001b[0m[k8s.io] EmptyDir volumes\u001b[0m ", " \u001b[1mshould support (root,0644,tmpfs) [Conformance]\u001b[0m", " \u001b[37m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:69\u001b[0m", "\u001b[1mSTEP\u001b[0m: Creating a kubernetes client", "Jan 10 18:32:11.023: INFO: >>> kubeConfig: /etc/origin/master/admin.kubeconfig", "", "\u001b[1mSTEP\u001b[0m: Building a namespace api object", "Jan 10 18:32:11.047: INFO: About to run a Kube e2e test, ensuring namespace is privileged", "\u001b[1mSTEP\u001b[0m: Waiting for a default service account to be provisioned in namespace", "\u001b[1mSTEP\u001b[0m: Creating a pod to test emptydir 0644 on tmpfs", "Jan 10 18:32:11.078: INFO: Waiting up to 5m0s for pod pod-0252e642-d78d-11e6-8a95-fa163e61d47e status to be success or failure", "Jan 10 18:32:11.079: INFO: No Status.Info for container 'test-container' in pod 'pod-0252e642-d78d-11e6-8a95-fa163e61d47e' yet", "Jan 10 18:32:11.079: INFO: Waiting for pod pod-0252e642-d78d-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-z8gry' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (1.232535ms elapsed)", "\u001b[1mSTEP\u001b[0m: Saw pod success", "\u001b[1mSTEP\u001b[0m: Trying to get logs from node 172.16.132.150 pod pod-0252e642-d78d-11e6-8a95-fa163e61d47e container test-container: ", "\u001b[1mSTEP\u001b[0m: Successfully fetched pod logs:mount type of \"/test-volume\": tmpfs", "content of file \"/test-volume/test-file\": mount-tester new file", "", "perms of file \"/test-volume/test-file\": -rw-r--r--", "", "", "\u001b[1mSTEP\u001b[0m: Collecting resource usage data", "Jan 10 18:32:13.097: INFO: Closed stop channel. Waiting for 3 workers", "Jan 10 18:32:13.097: INFO: Closing worker for 172.16.132.155", "Jan 10 18:32:13.097: INFO: Closing worker for 172.16.132.150", "Jan 10 18:32:13.097: INFO: Closing worker for 172.16.132.148", "Jan 10 18:32:13.097: INFO: Waitgroup finished.", "Jan 10 18:32:13.097: INFO: Unknown output type: . Skipping.", "Jan 10 18:32:13.097: INFO: Waiting up to 1m0s for all nodes to be ready", "\u001b[1mSTEP\u001b[0m: Destroying namespace \"e2e-tests-emptydir-z8gry\" for this suite.", "", "\u001b[32m• [SLOW TEST:7.084 seconds]\u001b[0m", "[k8s.io] EmptyDir volumes", "\u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:690\u001b[0m", " should support (root,0644,tmpfs) [Conformance]", " \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:69\u001b[0m", "\u001b[90m------------------------------\u001b[0m", "\u001b[0m[k8s.io] EmptyDir volumes\u001b[0m ", " \u001b[1mshould support (root,0666,tmpfs) [Conformance]\u001b[0m", " \u001b[37m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:73\u001b[0m", "\u001b[1mSTEP\u001b[0m: Creating a kubernetes client", "Jan 10 18:32:18.107: INFO: >>> kubeConfig: /etc/origin/master/admin.kubeconfig", "", "\u001b[1mSTEP\u001b[0m: Building a namespace api object", "Jan 10 18:32:18.128: INFO: About to run a Kube e2e test, ensuring namespace is privileged", "\u001b[1mSTEP\u001b[0m: Waiting for a default service account to be provisioned in namespace", "\u001b[1mSTEP\u001b[0m: Creating a pod to test emptydir 0666 on tmpfs", "Jan 10 18:32:18.228: INFO: Waiting up to 5m0s for pod pod-068d993f-d78d-11e6-8a95-fa163e61d47e status to be success or failure", "Jan 10 18:32:18.231: INFO: No Status.Info for container 'test-container' in pod 'pod-068d993f-d78d-11e6-8a95-fa163e61d47e' yet", "Jan 10 18:32:18.231: INFO: Waiting for pod pod-068d993f-d78d-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-03t1m' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (2.177997ms elapsed)", "Jan 10 18:32:20.233: INFO: Nil State.Terminated for container 'test-container' in pod 'pod-068d993f-d78d-11e6-8a95-fa163e61d47e' in namespace 'e2e-tests-emptydir-03t1m' so far", "Jan 10 18:32:20.233: INFO: Waiting for pod pod-068d993f-d78d-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-03t1m' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (2.004342222s elapsed)", "\u001b[1mSTEP\u001b[0m: Saw pod success", "\u001b[1mSTEP\u001b[0m: Trying to get logs from node 172.16.132.148 pod pod-068d993f-d78d-11e6-8a95-fa163e61d47e container test-container: ", "\u001b[1mSTEP\u001b[0m: Successfully fetched pod logs:mount type of \"/test-volume\": tmpfs", "content of file \"/test-volume/test-file\": mount-tester new file", "", "perms of file \"/test-volume/test-file\": -rw-rw-rw-", "", "", "\u001b[1mSTEP\u001b[0m: Collecting resource usage data", "Jan 10 18:32:22.247: INFO: Closed stop channel. Waiting for 3 workers", "Jan 10 18:32:22.247: INFO: Closing worker for 172.16.132.155", "Jan 10 18:32:22.247: INFO: Closing worker for 172.16.132.150", "Jan 10 18:32:22.247: INFO: Closing worker for 172.16.132.148", "Jan 10 18:32:22.247: INFO: Waitgroup finished.", "Jan 10 18:32:22.247: INFO: Unknown output type: . Skipping.", "Jan 10 18:32:22.247: INFO: Waiting up to 1m0s for all nodes to be ready", "\u001b[1mSTEP\u001b[0m: Destroying namespace \"e2e-tests-emptydir-03t1m\" for this suite.", "", "\u001b[32m• [SLOW TEST:9.149 seconds]\u001b[0m", "[k8s.io] EmptyDir volumes", "\u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:690\u001b[0m", " should support (root,0666,tmpfs) [Conformance]", " \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:73\u001b[0m", "\u001b[90m------------------------------\u001b[0m", "\u001b[0m[k8s.io] EmptyDir volumes\u001b[0m ", " \u001b[1mshould support (root,0777,tmpfs) [Conformance]\u001b[0m", " \u001b[37m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:77\u001b[0m", "\u001b[1mSTEP\u001b[0m: Creating a kubernetes client", "Jan 10 18:32:27.256: INFO: >>> kubeConfig: /etc/origin/master/admin.kubeconfig", "", "\u001b[1mSTEP\u001b[0m: Building a namespace api object", "Jan 10 18:32:27.279: INFO: About to run a Kube e2e test, ensuring namespace is privileged", "\u001b[1mSTEP\u001b[0m: Waiting for a default service account to be provisioned in namespace", "\u001b[1mSTEP\u001b[0m: Creating a pod to test emptydir 0777 on tmpfs", "Jan 10 18:32:27.328: INFO: Waiting up to 5m0s for pod pod-0c02aec2-d78d-11e6-8a95-fa163e61d47e status to be success or failure", "Jan 10 18:32:27.337: INFO: No Status.Info for container 'test-container' in pod 'pod-0c02aec2-d78d-11e6-8a95-fa163e61d47e' yet", "Jan 10 18:32:27.337: INFO: Waiting for pod pod-0c02aec2-d78d-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-zbsyv' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (8.617146ms elapsed)", "\u001b[1mSTEP\u001b[0m: Saw pod success", "\u001b[1mSTEP\u001b[0m: Trying to get logs from node 172.16.132.150 pod pod-0c02aec2-d78d-11e6-8a95-fa163e61d47e container test-container: ", "\u001b[1mSTEP\u001b[0m: Successfully fetched pod logs:mount type of \"/test-volume\": tmpfs", "content of file \"/test-volume/test-file\": mount-tester new file", "", "perms of file \"/test-volume/test-file\": -rwxrwxrwx", "", "", "\u001b[1mSTEP\u001b[0m: Collecting resource usage data", "Jan 10 18:32:29.351: INFO: Closed stop channel. Waiting for 3 workers", "Jan 10 18:32:29.351: INFO: Closing worker for 172.16.132.155", "Jan 10 18:32:29.351: INFO: Closing worker for 172.16.132.150", "Jan 10 18:32:29.351: INFO: Closing worker for 172.16.132.148", "Jan 10 18:32:29.351: INFO: Waitgroup finished.", "Jan 10 18:32:29.351: INFO: Unknown output type: . Skipping.", "Jan 10 18:32:29.351: INFO: Waiting up to 1m0s for all nodes to be ready", "\u001b[1mSTEP\u001b[0m: Destroying namespace \"e2e-tests-emptydir-zbsyv\" for this suite.", "", "\u001b[32m• [SLOW TEST:7.105 seconds]\u001b[0m", "[k8s.io] EmptyDir volumes", "\u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:690\u001b[0m", " should support (root,0777,tmpfs) [Conformance]", " \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:77\u001b[0m", "\u001b[90m------------------------------\u001b[0m", "\u001b[0m[k8s.io] EmptyDir volumes\u001b[0m ", " \u001b[1mshould support (non-root,0644,tmpfs) [Conformance]\u001b[0m", " \u001b[37m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:81\u001b[0m", "\u001b[1mSTEP\u001b[0m: Creating a kubernetes client", "Jan 10 18:32:34.361: INFO: >>> kubeConfig: /etc/origin/master/admin.kubeconfig", "", "\u001b[1mSTEP\u001b[0m: Building a namespace api object", "Jan 10 18:32:34.392: INFO: About to run a Kube e2e test, ensuring namespace is privileged", "\u001b[1mSTEP\u001b[0m: Waiting for a default service account to be provisioned in namespace", "\u001b[1mSTEP\u001b[0m: Creating a pod to test emptydir 0644 on tmpfs", "Jan 10 18:32:34.422: INFO: Waiting up to 5m0s for pod pod-103d0c8b-d78d-11e6-8a95-fa163e61d47e status to be success or failure", "Jan 10 18:32:34.423: INFO: No Status.Info for container 'test-container' in pod 'pod-103d0c8b-d78d-11e6-8a95-fa163e61d47e' yet", "Jan 10 18:32:34.423: INFO: Waiting for pod pod-103d0c8b-d78d-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-lsclc' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (1.246851ms elapsed)", "Jan 10 18:32:36.426: INFO: Nil State.Terminated for container 'test-container' in pod 'pod-103d0c8b-d78d-11e6-8a95-fa163e61d47e' in namespace 'e2e-tests-emptydir-lsclc' so far", "Jan 10 18:32:36.426: INFO: Waiting for pod pod-103d0c8b-d78d-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-lsclc' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (2.003866886s elapsed)", "\u001b[1mSTEP\u001b[0m: Saw pod success", "\u001b[1mSTEP\u001b[0m: Trying to get logs from node 172.16.132.148 pod pod-103d0c8b-d78d-11e6-8a95-fa163e61d47e container test-container: ", "\u001b[1mSTEP\u001b[0m: Successfully fetched pod logs:mount type of \"/test-volume\": tmpfs", "content of file \"/test-volume/test-file\": mount-tester new file", "", "perms of file \"/test-volume/test-file\": -rw-r--r--", "", "", "\u001b[1mSTEP\u001b[0m: Collecting resource usage data", "Jan 10 18:32:38.443: INFO: Closed stop channel. Waiting for 3 workers", "Jan 10 18:32:38.443: INFO: Closing worker for 172.16.132.155", "Jan 10 18:32:38.443: INFO: Closing worker for 172.16.132.148", "Jan 10 18:32:38.443: INFO: Closing worker for 172.16.132.150", "Jan 10 18:32:38.443: INFO: Waitgroup finished.", "Jan 10 18:32:38.443: INFO: Unknown output type: . Skipping.", "Jan 10 18:32:38.443: INFO: Waiting up to 1m0s for all nodes to be ready", "\u001b[1mSTEP\u001b[0m: Destroying namespace \"e2e-tests-emptydir-lsclc\" for this suite.", "", "\u001b[32m• [SLOW TEST:9.091 seconds]\u001b[0m", "[k8s.io] EmptyDir volumes", "\u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:690\u001b[0m", " should support (non-root,0644,tmpfs) [Conformance]", " \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:81\u001b[0m", "\u001b[90m------------------------------\u001b[0m", "\u001b[0m[k8s.io] EmptyDir volumes\u001b[0m ", " \u001b[1mshould support (non-root,0666,tmpfs) [Conformance]\u001b[0m", " \u001b[37m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:85\u001b[0m", "\u001b[1mSTEP\u001b[0m: Creating a kubernetes client", "Jan 10 18:32:43.452: INFO: >>> kubeConfig: /etc/origin/master/admin.kubeconfig", "", "\u001b[1mSTEP\u001b[0m: Building a namespace api object", "Jan 10 18:32:43.490: INFO: About to run a Kube e2e test, ensuring namespace is privileged", "\u001b[1mSTEP\u001b[0m: Waiting for a default service account to be provisioned in namespace", "\u001b[1mSTEP\u001b[0m: Creating a pod to test emptydir 0666 on tmpfs", "Jan 10 18:32:43.537: INFO: Waiting up to 5m0s for pod pod-15aa8a5f-d78d-11e6-8a95-fa163e61d47e status to be success or failure", "Jan 10 18:32:43.538: INFO: No Status.Info for container 'test-container' in pod 'pod-15aa8a5f-d78d-11e6-8a95-fa163e61d47e' yet", "Jan 10 18:32:43.538: INFO: Waiting for pod pod-15aa8a5f-d78d-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-ecfrt' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (1.597589ms elapsed)", "Jan 10 18:32:45.541: INFO: Nil State.Terminated for container 'test-container' in pod 'pod-15aa8a5f-d78d-11e6-8a95-fa163e61d47e' in namespace 'e2e-tests-emptydir-ecfrt' so far", "Jan 10 18:32:45.541: INFO: Waiting for pod pod-15aa8a5f-d78d-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-ecfrt' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (2.004118652s elapsed)", "\u001b[1mSTEP\u001b[0m: Saw pod success", "\u001b[1mSTEP\u001b[0m: Trying to get logs from node 172.16.132.150 pod pod-15aa8a5f-d78d-11e6-8a95-fa163e61d47e container test-container: ", "\u001b[1mSTEP\u001b[0m: Successfully fetched pod logs:mount type of \"/test-volume\": tmpfs", "content of file \"/test-volume/test-file\": mount-tester new file", "", "perms of file \"/test-volume/test-file\": -rw-rw-rw-", "", "", "\u001b[1mSTEP\u001b[0m: Collecting resource usage data", "Jan 10 18:32:47.591: INFO: Closed stop channel. Waiting for 3 workers", "Jan 10 18:32:47.591: INFO: Closing worker for 172.16.132.155", "Jan 10 18:32:47.591: INFO: Closing worker for 172.16.132.150", "Jan 10 18:32:47.591: INFO: Closing worker for 172.16.132.148", "Jan 10 18:32:47.591: INFO: Waitgroup finished.", "Jan 10 18:32:47.591: INFO: Unknown output type: . Skipping.", "Jan 10 18:32:47.591: INFO: Waiting up to 1m0s for all nodes to be ready", "\u001b[1mSTEP\u001b[0m: Destroying namespace \"e2e-tests-emptydir-ecfrt\" for this suite.", "", "\u001b[32m• [SLOW TEST:9.150 seconds]\u001b[0m", "[k8s.io] EmptyDir volumes", "\u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:690\u001b[0m", " should support (non-root,0666,tmpfs) [Conformance]", " \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:85\u001b[0m", "\u001b[90m------------------------------\u001b[0m", "\u001b[0m[k8s.io] EmptyDir volumes\u001b[0m ", " \u001b[1mshould support (non-root,0777,tmpfs) [Conformance]\u001b[0m", " \u001b[37m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:89\u001b[0m", "\u001b[1mSTEP\u001b[0m: Creating a kubernetes client", "Jan 10 18:32:52.602: INFO: >>> kubeConfig: /etc/origin/master/admin.kubeconfig", "", "\u001b[1mSTEP\u001b[0m: Building a namespace api object", "Jan 10 18:32:52.624: INFO: About to run a Kube e2e test, ensuring namespace is privileged", "\u001b[1mSTEP\u001b[0m: Waiting for a default service account to be provisioned in namespace", "\u001b[1mSTEP\u001b[0m: Creating a pod to test emptydir 0777 on tmpfs", "Jan 10 18:32:52.671: INFO: Waiting up to 5m0s for pod pod-1b1ce8f8-d78d-11e6-8a95-fa163e61d47e status to be success or failure", "Jan 10 18:32:52.674: INFO: No Status.Info for container 'test-container' in pod 'pod-1b1ce8f8-d78d-11e6-8a95-fa163e61d47e' yet", "Jan 10 18:32:52.674: INFO: Waiting for pod pod-1b1ce8f8-d78d-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-xx63m' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (3.125197ms elapsed)", "\u001b[1mSTEP\u001b[0m: Saw pod success", "\u001b[1mSTEP\u001b[0m: Trying to get logs from node 172.16.132.148 pod pod-1b1ce8f8-d78d-11e6-8a95-fa163e61d47e container test-container: ", "\u001b[1mSTEP\u001b[0m: Successfully fetched pod logs:mount type of \"/test-volume\": tmpfs", "content of file \"/test-volume/test-file\": mount-tester new file", "", "perms of file \"/test-volume/test-file\": -rwxrwxrwx", "", "", "\u001b[1mSTEP\u001b[0m: Collecting resource usage data", "Jan 10 18:32:54.691: INFO: Closed stop channel. Waiting for 3 workers", "Jan 10 18:32:54.691: INFO: Closing worker for 172.16.132.150", "Jan 10 18:32:54.691: INFO: Closing worker for 172.16.132.148", "Jan 10 18:32:54.691: INFO: Closing worker for 172.16.132.155", "Jan 10 18:32:54.691: INFO: Waitgroup finished.", "Jan 10 18:32:54.691: INFO: Unknown output type: . Skipping.", "Jan 10 18:32:54.691: INFO: Waiting up to 1m0s for all nodes to be ready", "\u001b[1mSTEP\u001b[0m: Destroying namespace \"e2e-tests-emptydir-xx63m\" for this suite.", "", "\u001b[32m• [SLOW TEST:7.099 seconds]\u001b[0m", "[k8s.io] EmptyDir volumes", "\u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:690\u001b[0m", " should support (non-root,0777,tmpfs) [Conformance]", " \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:89\u001b[0m", "\u001b[90m------------------------------\u001b[0m", "\u001b[0m[k8s.io] EmptyDir volumes\u001b[0m ", " \u001b[1mvolume on default medium should have the correct mode [Conformance]\u001b[0m", " \u001b[37m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:93\u001b[0m", "\u001b[1mSTEP\u001b[0m: Creating a kubernetes client", "Jan 10 18:32:59.700: INFO: >>> kubeConfig: /etc/origin/master/admin.kubeconfig", "", "\u001b[1mSTEP\u001b[0m: Building a namespace api object", "Jan 10 18:32:59.772: INFO: About to run a Kube e2e test, ensuring namespace is privileged", "\u001b[1mSTEP\u001b[0m: Waiting for a default service account to be provisioned in namespace", "\u001b[1mSTEP\u001b[0m: Creating a pod to test emptydir volume type on node default medium", "Jan 10 18:32:59.837: INFO: Waiting up to 5m0s for pod pod-1f62f945-d78d-11e6-8a95-fa163e61d47e status to be success or failure", "Jan 10 18:32:59.853: INFO: No Status.Info for container 'test-container' in pod 'pod-1f62f945-d78d-11e6-8a95-fa163e61d47e' yet", "Jan 10 18:32:59.853: INFO: Waiting for pod pod-1f62f945-d78d-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-n5tg3' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (15.838227ms elapsed)", "\u001b[1mSTEP\u001b[0m: Saw pod success", "\u001b[1mSTEP\u001b[0m: Trying to get logs from node 172.16.132.150 pod pod-1f62f945-d78d-11e6-8a95-fa163e61d47e container test-container: ", "\u001b[1mSTEP\u001b[0m: Successfully fetched pod logs:mount type of \"/test-volume\": 1481003842", "perms of file \"/test-volume\": -rwxrwxrwx", "", "", "\u001b[1mSTEP\u001b[0m: Collecting resource usage data", "Jan 10 18:33:01.870: INFO: Closed stop channel. Waiting for 3 workers", "Jan 10 18:33:01.870: INFO: Closing worker for 172.16.132.155", "Jan 10 18:33:01.870: INFO: Closing worker for 172.16.132.150", "Jan 10 18:33:01.870: INFO: Closing worker for 172.16.132.148", "Jan 10 18:33:01.870: INFO: Waitgroup finished.", "Jan 10 18:33:01.870: INFO: Unknown output type: . Skipping.", "Jan 10 18:33:01.870: INFO: Waiting up to 1m0s for all nodes to be ready", "\u001b[1mSTEP\u001b[0m: Destroying namespace \"e2e-tests-emptydir-n5tg3\" for this suite.", "", "\u001b[32m• [SLOW TEST:7.177 seconds]\u001b[0m", "[k8s.io] EmptyDir volumes", "\u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:690\u001b[0m", " volume on default medium should have the correct mode [Conformance]", " \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:93\u001b[0m", "\u001b[90m------------------------------\u001b[0m", "\u001b[0m[k8s.io] EmptyDir volumes\u001b[0m ", " \u001b[1mshould support (root,0644,default) [Conformance]\u001b[0m", " \u001b[37m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:97\u001b[0m", "\u001b[1mSTEP\u001b[0m: Creating a kubernetes client", "Jan 10 18:33:06.878: INFO: >>> kubeConfig: /etc/origin/master/admin.kubeconfig", "", "\u001b[1mSTEP\u001b[0m: Building a namespace api object", "Jan 10 18:33:06.902: INFO: About to run a Kube e2e test, ensuring namespace is privileged", "\u001b[1mSTEP\u001b[0m: Waiting for a default service account to be provisioned in namespace", "\u001b[1mSTEP\u001b[0m: Creating a pod to test emptydir 0644 on node default medium", "Jan 10 18:33:06.938: INFO: Waiting up to 5m0s for pod pod-239e56ae-d78d-11e6-8a95-fa163e61d47e status to be success or failure", "Jan 10 18:33:06.940: INFO: No Status.Info for container 'test-container' in pod 'pod-239e56ae-d78d-11e6-8a95-fa163e61d47e' yet", "Jan 10 18:33:06.940: INFO: Waiting for pod pod-239e56ae-d78d-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-xy29m' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (2.451514ms elapsed)", "Jan 10 18:33:08.944: INFO: Nil State.Terminated for container 'test-container' in pod 'pod-239e56ae-d78d-11e6-8a95-fa163e61d47e' in namespace 'e2e-tests-emptydir-xy29m' so far", "Jan 10 18:33:08.944: INFO: Waiting for pod pod-239e56ae-d78d-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-xy29m' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (2.005639937s elapsed)", "\u001b[1mSTEP\u001b[0m: Saw pod success", "\u001b[1mSTEP\u001b[0m: Trying to get logs from node 172.16.132.148 pod pod-239e56ae-d78d-11e6-8a95-fa163e61d47e container test-container: ", "\u001b[1mSTEP\u001b[0m: Successfully fetched pod logs:mount type of \"/test-volume\": 1481003842", "content of file \"/test-volume/test-file\": mount-tester new file", "", "perms of file \"/test-volume/test-file\": -rw-r--r--", "", "", "\u001b[1mSTEP\u001b[0m: Collecting resource usage data", "Jan 10 18:33:10.959: INFO: Closed stop channel. Waiting for 3 workers", "Jan 10 18:33:10.959: INFO: Closing worker for 172.16.132.148", "Jan 10 18:33:10.959: INFO: Closing worker for 172.16.132.155", "Jan 10 18:33:10.959: INFO: Closing worker for 172.16.132.150", "Jan 10 18:33:10.959: INFO: Waitgroup finished.", "Jan 10 18:33:10.959: INFO: Unknown output type: . Skipping.", "Jan 10 18:33:10.959: INFO: Waiting up to 1m0s for all nodes to be ready", "\u001b[1mSTEP\u001b[0m: Destroying namespace \"e2e-tests-emptydir-xy29m\" for this suite.", "", "\u001b[32m• [SLOW TEST:9.090 seconds]\u001b[0m", "[k8s.io] EmptyDir volumes", "\u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:690\u001b[0m", " should support (root,0644,default) [Conformance]", " \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:97\u001b[0m", "\u001b[90m------------------------------\u001b[0m", "\u001b[0m[k8s.io] EmptyDir volumes\u001b[0m ", " \u001b[1mshould support (root,0666,default) [Conformance]\u001b[0m", " \u001b[37m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:101\u001b[0m", "\u001b[1mSTEP\u001b[0m: Creating a kubernetes client", "Jan 10 18:33:15.968: INFO: >>> kubeConfig: /etc/origin/master/admin.kubeconfig", "", "\u001b[1mSTEP\u001b[0m: Building a namespace api object", "Jan 10 18:33:16.035: INFO: About to run a Kube e2e test, ensuring namespace is privileged", "\u001b[1mSTEP\u001b[0m: Waiting for a default service account to be provisioned in namespace", "\u001b[1mSTEP\u001b[0m: Creating a pod to test emptydir 0666 on node default medium", "Jan 10 18:33:16.077: INFO: Waiting up to 5m0s for pod pod-29111d02-d78d-11e6-8a95-fa163e61d47e status to be success or failure", "Jan 10 18:33:16.081: INFO: No Status.Info for container 'test-container' in pod 'pod-29111d02-d78d-11e6-8a95-fa163e61d47e' yet", "Jan 10 18:33:16.081: INFO: Waiting for pod pod-29111d02-d78d-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-u5046' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (3.792893ms elapsed)", "\u001b[1mSTEP\u001b[0m: Saw pod success", "\u001b[1mSTEP\u001b[0m: Trying to get logs from node 172.16.132.150 pod pod-29111d02-d78d-11e6-8a95-fa163e61d47e container test-container: ", "\u001b[1mSTEP\u001b[0m: Successfully fetched pod logs:mount type of \"/test-volume\": 1481003842", "content of file \"/test-volume/test-file\": mount-tester new file", "", "perms of file \"/test-volume/test-file\": -rw-rw-rw-", "", "", "\u001b[1mSTEP\u001b[0m: Collecting resource usage data", "Jan 10 18:33:18.101: INFO: Closed stop channel. Waiting for 3 workers", "Jan 10 18:33:18.101: INFO: Closing worker for 172.16.132.150", "Jan 10 18:33:18.101: INFO: Closing worker for 172.16.132.148", "Jan 10 18:33:18.101: INFO: Closing worker for 172.16.132.155", "Jan 10 18:33:18.101: INFO: Waitgroup finished.", "Jan 10 18:33:18.101: INFO: Unknown output type: . Skipping.", "Jan 10 18:33:18.101: INFO: Waiting up to 1m0s for all nodes to be ready", "\u001b[1mSTEP\u001b[0m: Destroying namespace \"e2e-tests-emptydir-u5046\" for this suite.", "", "\u001b[32m• [SLOW TEST:7.147 seconds]\u001b[0m", "[k8s.io] EmptyDir volumes", "\u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:690\u001b[0m", " should support (root,0666,default) [Conformance]", " \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:101\u001b[0m", "\u001b[90m------------------------------\u001b[0m", "\u001b[0m[k8s.io] EmptyDir volumes\u001b[0m ", " \u001b[1mshould support (root,0777,default) [Conformance]\u001b[0m", " \u001b[37m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:105\u001b[0m", "\u001b[1mSTEP\u001b[0m: Creating a kubernetes client", "Jan 10 18:33:23.116: INFO: >>> kubeConfig: /etc/origin/master/admin.kubeconfig", "", "\u001b[1mSTEP\u001b[0m: Building a namespace api object", "Jan 10 18:33:23.138: INFO: About to run a Kube e2e test, ensuring namespace is privileged", "\u001b[1mSTEP\u001b[0m: Waiting for a default service account to be provisioned in namespace", "\u001b[1mSTEP\u001b[0m: Creating a pod to test emptydir 0777 on node default medium", "Jan 10 18:33:23.165: INFO: Waiting up to 5m0s for pod pod-2d4ae08d-d78d-11e6-8a95-fa163e61d47e status to be success or failure", "Jan 10 18:33:23.166: INFO: No Status.Info for container 'test-container' in pod 'pod-2d4ae08d-d78d-11e6-8a95-fa163e61d47e' yet", "Jan 10 18:33:23.166: INFO: Waiting for pod pod-2d4ae08d-d78d-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-m69vd' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (1.189333ms elapsed)", "\u001b[1mSTEP\u001b[0m: Saw pod success", "\u001b[1mSTEP\u001b[0m: Trying to get logs from node 172.16.132.148 pod pod-2d4ae08d-d78d-11e6-8a95-fa163e61d47e container test-container: ", "\u001b[1mSTEP\u001b[0m: Successfully fetched pod logs:mount type of \"/test-volume\": 1481003842", "content of file \"/test-volume/test-file\": mount-tester new file", "", "perms of file \"/test-volume/test-file\": -rwxrwxrwx", "", "", "\u001b[1mSTEP\u001b[0m: Collecting resource usage data", "Jan 10 18:33:25.183: INFO: Closed stop channel. Waiting for 3 workers", "Jan 10 18:33:25.183: INFO: Closing worker for 172.16.132.155", "Jan 10 18:33:25.183: INFO: Closing worker for 172.16.132.150", "Jan 10 18:33:25.183: INFO: Closing worker for 172.16.132.148", "Jan 10 18:33:25.183: INFO: Waitgroup finished.", "Jan 10 18:33:25.183: INFO: Unknown output type: . Skipping.", "Jan 10 18:33:25.183: INFO: Waiting up to 1m0s for all nodes to be ready", "\u001b[1mSTEP\u001b[0m: Destroying namespace \"e2e-tests-emptydir-m69vd\" for this suite.", "", "\u001b[32m• [SLOW TEST:7.081 seconds]\u001b[0m", "[k8s.io] EmptyDir volumes", "\u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:690\u001b[0m", " should support (root,0777,default) [Conformance]", " \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:105\u001b[0m", "\u001b[90m------------------------------\u001b[0m", "\u001b[0m[k8s.io] EmptyDir volumes\u001b[0m ", " \u001b[1mshould support (non-root,0644,default) [Conformance]\u001b[0m", " \u001b[37m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:109\u001b[0m", "\u001b[1mSTEP\u001b[0m: Creating a kubernetes client", "Jan 10 18:33:30.197: INFO: >>> kubeConfig: /etc/origin/master/admin.kubeconfig", "", "\u001b[1mSTEP\u001b[0m: Building a namespace api object", "Jan 10 18:33:30.227: INFO: About to run a Kube e2e test, ensuring namespace is privileged", "\u001b[1mSTEP\u001b[0m: Waiting for a default service account to be provisioned in namespace", "\u001b[1mSTEP\u001b[0m: Creating a pod to test emptydir 0644 on node default medium", "Jan 10 18:33:30.272: INFO: Waiting up to 5m0s for pod pod-318732e2-d78d-11e6-8a95-fa163e61d47e status to be success or failure", "Jan 10 18:33:30.277: INFO: No Status.Info for container 'test-container' in pod 'pod-318732e2-d78d-11e6-8a95-fa163e61d47e' yet", "Jan 10 18:33:30.277: INFO: Waiting for pod pod-318732e2-d78d-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-ojjux' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (5.163181ms elapsed)", "\u001b[1mSTEP\u001b[0m: Saw pod success", "\u001b[1mSTEP\u001b[0m: Trying to get logs from node 172.16.132.150 pod pod-318732e2-d78d-11e6-8a95-fa163e61d47e container test-container: ", "\u001b[1mSTEP\u001b[0m: Successfully fetched pod logs:mount type of \"/test-volume\": 1481003842", "content of file \"/test-volume/test-file\": mount-tester new file", "", "perms of file \"/test-volume/test-file\": -rw-r--r--", "", "", "\u001b[1mSTEP\u001b[0m: Collecting resource usage data", "Jan 10 18:33:32.292: INFO: Closed stop channel. Waiting for 3 workers", "Jan 10 18:33:32.292: INFO: Closing worker for 172.16.132.155", "Jan 10 18:33:32.292: INFO: Closing worker for 172.16.132.150", "Jan 10 18:33:32.292: INFO: Closing worker for 172.16.132.148", "Jan 10 18:33:32.292: INFO: Waitgroup finished.", "Jan 10 18:33:32.292: INFO: Unknown output type: . Skipping.", "Jan 10 18:33:32.292: INFO: Waiting up to 1m0s for all nodes to be ready", "\u001b[1mSTEP\u001b[0m: Destroying namespace \"e2e-tests-emptydir-ojjux\" for this suite.", "", "\u001b[32m• [SLOW TEST:7.104 seconds]\u001b[0m", "[k8s.io] EmptyDir volumes", "\u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:690\u001b[0m", " should support (non-root,0644,default) [Conformance]", " \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:109\u001b[0m", "\u001b[90m------------------------------\u001b[0m", "\u001b[0m[k8s.io] EmptyDir volumes\u001b[0m ", " \u001b[1mshould support (non-root,0666,default) [Conformance]\u001b[0m", " \u001b[37m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:113\u001b[0m", "\u001b[1mSTEP\u001b[0m: Creating a kubernetes client", "Jan 10 18:33:37.301: INFO: >>> kubeConfig: /etc/origin/master/admin.kubeconfig", "", "\u001b[1mSTEP\u001b[0m: Building a namespace api object", "Jan 10 18:33:37.327: INFO: About to run a Kube e2e test, ensuring namespace is privileged", "\u001b[1mSTEP\u001b[0m: Waiting for a default service account to be provisioned in namespace", "\u001b[1mSTEP\u001b[0m: Creating a pod to test emptydir 0666 on node default medium", "Jan 10 18:33:37.358: INFO: Waiting up to 5m0s for pod pod-35bfd2f2-d78d-11e6-8a95-fa163e61d47e status to be success or failure", "Jan 10 18:33:37.361: INFO: No Status.Info for container 'test-container' in pod 'pod-35bfd2f2-d78d-11e6-8a95-fa163e61d47e' yet", "Jan 10 18:33:37.361: INFO: Waiting for pod pod-35bfd2f2-d78d-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-n3m8r' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (3.220896ms elapsed)", "\u001b[1mSTEP\u001b[0m: Saw pod success", "\u001b[1mSTEP\u001b[0m: Trying to get logs from node 172.16.132.148 pod pod-35bfd2f2-d78d-11e6-8a95-fa163e61d47e container test-container: ", "\u001b[1mSTEP\u001b[0m: Successfully fetched pod logs:mount type of \"/test-volume\": 1481003842", "content of file \"/test-volume/test-file\": mount-tester new file", "", "perms of file \"/test-volume/test-file\": -rw-rw-rw-", "", "", "\u001b[1mSTEP\u001b[0m: Collecting resource usage data", "Jan 10 18:33:39.376: INFO: Closed stop channel. Waiting for 3 workers", "Jan 10 18:33:39.376: INFO: Closing worker for 172.16.132.155", "Jan 10 18:33:39.376: INFO: Closing worker for 172.16.132.150", "Jan 10 18:33:39.376: INFO: Closing worker for 172.16.132.148", "Jan 10 18:33:39.376: INFO: Waitgroup finished.", "Jan 10 18:33:39.376: INFO: Unknown output type: . Skipping.", "Jan 10 18:33:39.376: INFO: Waiting up to 1m0s for all nodes to be ready", "\u001b[1mSTEP\u001b[0m: Destroying namespace \"e2e-tests-emptydir-n3m8r\" for this suite.", "", "\u001b[32m• [SLOW TEST:7.083 seconds]\u001b[0m", "[k8s.io] EmptyDir volumes", "\u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:690\u001b[0m", " should support (non-root,0666,default) [Conformance]", " \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:113\u001b[0m", "\u001b[90m------------------------------\u001b[0m", "\u001b[0m[k8s.io] EmptyDir volumes\u001b[0m ", " \u001b[1mshould support (non-root,0777,default) [Conformance]\u001b[0m", " \u001b[37m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:117\u001b[0m", "\u001b[1mSTEP\u001b[0m: Creating a kubernetes client", "Jan 10 18:33:44.384: INFO: >>> kubeConfig: /etc/origin/master/admin.kubeconfig", "", "\u001b[1mSTEP\u001b[0m: Building a namespace api object", "Jan 10 18:33:44.407: INFO: About to run a Kube e2e test, ensuring namespace is privileged", "\u001b[1mSTEP\u001b[0m: Waiting for a default service account to be provisioned in namespace", "\u001b[1mSTEP\u001b[0m: Creating a pod to test emptydir 0777 on node default medium", "Jan 10 18:33:44.449: INFO: Waiting up to 5m0s for pod pod-39f9eed1-d78d-11e6-8a95-fa163e61d47e status to be success or failure", "Jan 10 18:33:44.464: INFO: Nil State.Terminated for container 'test-container' in pod 'pod-39f9eed1-d78d-11e6-8a95-fa163e61d47e' in namespace 'e2e-tests-emptydir-docy2' so far", "Jan 10 18:33:44.464: INFO: Waiting for pod pod-39f9eed1-d78d-11e6-8a95-fa163e61d47e in namespace 'e2e-tests-emptydir-docy2' status to be 'success or failure'(found phase: \"Pending\", readiness: false) (14.373228ms elapsed)", "\u001b[1mSTEP\u001b[0m: Saw pod success", "\u001b[1mSTEP\u001b[0m: Trying to get logs from node 172.16.132.150 pod pod-39f9eed1-d78d-11e6-8a95-fa163e61d47e container test-container: ", "\u001b[1mSTEP\u001b[0m: Successfully fetched pod logs:mount type of \"/test-volume\": 1481003842", "content of file \"/test-volume/test-file\": mount-tester new file", "", "perms of file \"/test-volume/test-file\": -rwxrwxrwx", "", "", "\u001b[1mSTEP\u001b[0m: Collecting resource usage data", "Jan 10 18:33:46.497: INFO: Closed stop channel. Waiting for 3 workers", "Jan 10 18:33:46.497: INFO: Closing worker for 172.16.132.148", "Jan 10 18:33:46.497: INFO: Closing worker for 172.16.132.150", "Jan 10 18:33:46.497: INFO: Closing worker for 172.16.132.155", "Jan 10 18:33:46.497: INFO: Waitgroup finished.", "Jan 10 18:33:46.497: INFO: Unknown output type: . Skipping.", "Jan 10 18:33:46.497: INFO: Waiting up to 1m0s for all nodes to be ready", "\u001b[1mSTEP\u001b[0m: Destroying namespace \"e2e-tests-emptydir-docy2\" for this suite.", "", "\u001b[32m• [SLOW TEST:7.121 seconds]\u001b[0m", "[k8s.io] EmptyDir volumes", "\u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:690\u001b[0m", " should support (non-root,0777,default) [Conformance]", " \u001b[90m/builddir/build/BUILD/atomic-openshift-git-0.a7f5265/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/empty_dir.go:117\u001b[0m", "\u001b[90m------------------------------\u001b[0m", "\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m\u001b[36mS\u001b[0m", "\u001b[1m\u001b[32mRan 19 of 509 Specs in 155.161 seconds\u001b[0m", "\u001b[1m\u001b[32mSUCCESS!\u001b[0m -- \u001b[32m\u001b[1m19 Passed\u001b[0m | \u001b[91m\u001b[1m0 Failed\u001b[0m | \u001b[33m\u001b[1m0 Pending\u001b[0m | \u001b[36m\u001b[1m490 Skipped\u001b[0m PASS" ], "warnings": [] } TASK [e2e : Grab e2e.log file from test run] *********************************** task path: /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/playbooks/openshift/roles/e2e/tasks/execute_e2e_tests.yml:13 Using module file /usr/lib/python2.7/site-packages/ansible/modules/core/files/stat.py <10.8.169.22> ESTABLISH SSH CONNECTION FOR USER: root <10.8.169.22> SSH: EXEC ssh -C -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o 'IdentityFile="/home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/config/keys/ci-factory"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/%h-%r 10.8.169.22 '/bin/sh -c '"'"'( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1484091232.68-158023703250022 `" && echo ansible-tmp-1484091232.68-158023703250022="` echo $HOME/.ansible/tmp/ansible-tmp-1484091232.68-158023703250022 `" ) && sleep 0'"'"'' <10.8.169.22> PUT /tmp/tmpsLaRar TO /root/.ansible/tmp/ansible-tmp-1484091232.68-158023703250022/stat.py <10.8.169.22> SSH: EXEC sftp -b - -C -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o 'IdentityFile="/home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/config/keys/ci-factory"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/%h-%r '[10.8.169.22]' <10.8.169.22> ESTABLISH SSH CONNECTION FOR USER: root <10.8.169.22> SSH: EXEC ssh -C -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o 'IdentityFile="/home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/config/keys/ci-factory"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/%h-%r 10.8.169.22 '/bin/sh -c '"'"'chmod u+x /root/.ansible/tmp/ansible-tmp-1484091232.68-158023703250022/ /root/.ansible/tmp/ansible-tmp-1484091232.68-158023703250022/stat.py && sleep 0'"'"'' <10.8.169.22> ESTABLISH SSH CONNECTION FOR USER: root <10.8.169.22> SSH: EXEC ssh -C -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o 'IdentityFile="/home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/config/keys/ci-factory"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/%h-%r -tt 10.8.169.22 '/bin/sh -c '"'"'/usr/bin/python /root/.ansible/tmp/ansible-tmp-1484091232.68-158023703250022/stat.py; rm -rf "/root/.ansible/tmp/ansible-tmp-1484091232.68-158023703250022/" > /dev/null 2>&1 && sleep 0'"'"'' <10.8.169.22> FETCH /root/e2e.log TO /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/e2e.log <10.8.169.22> SSH: EXEC scp -C -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o 'IdentityFile="/home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/atomic-ci-jobs/project/config/keys/ci-factory"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/%h-%r '[10.8.169.22]:/root/e2e.log' /home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/e2e.log changed: [10.8.169.22] => { "changed": true, "checksum": "38427ab40cae4cb3a3c7fb8abf91b541decd8a71", "dest": "/home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/e2e.log", "failed": false, "failed_when_result": false, "invocation": { "module_args": { "dest": "/home/jenkins/workspace/openshift-ansible-3-test-matrix/CONTAINERIZED/_containerized/OSE_VER/3.3/PYTHON/System-CPython-2.7/TOPOLOGY/openshift-cluster-containerized/TargetBranch/master/nodes/openshift-ansible-slave/e2e.log", "fail_on_missing": true, "flat": true, "src": "/root/e2e.log" }, "module_name": "fetch" }, "md5sum": "722d101bc50cad681bdd87241f3b9b10", "remote_checksum": "38427ab40cae4cb3a3c7fb8abf91b541decd8a71", "remote_md5sum": null } PLAY RECAP ********************************************************************* 10.8.169.22 : ok=5 changed=4 unreachable=0 failed=0 [EnvInject] - Injecting environment variables from a build step. [EnvInject] - Injecting as environment variables the properties content GH_STATE=success GH_DESC="e2e tests passed" GH_CONTEXT=aos-ci-jenkins/OS_3.3_containerized_e2e_tests [EnvInject] - Variables injected successfully. [openshift-ansible-slave] $ python /tmp/hudson4780746396192583352.py Updating status for ae88c63dda7c433cd2b42ad7d9527b483579dc4e [openshift-ansible-slave] $ python /tmp/hudson4484054981170037208.py [PostBuildScript] - Execution post build scripts. [openshift-ansible-slave] $ /bin/bash /tmp/hudson8498363982444158633.sh <<<<<<<<<<<<<<<<<<<< END jenkins-openshift-ansible-3-test-matrix-CONTAINERIZED=_containerized,OSE_VER=3.3,PYTHON=System-CPython-2.7,TOPOLOGY=openshift-cluster-containerized,TargetBranch=master,nodes=openshift-ansible-slave-688 >>>>>>>>>>>>>>>>>>