installcentos's People
Forkers
levkov trifonnt cloudseabr glamperi sriharsha-at-git frgomes julienbell thato-mokaila rajeshyukta robbie1977 aldwindelgado syafiqfiqq itabara yestechgroup msghaleb mostmark carlosthe19916 ediri sagh0900 koreset bsk01 cleygraf bidfa01 frank75888 schiznik isnuryusuf itechdhaval dmaple-gci kpritam oschangkai jimmyadepeju jcperezamin sharp-pixel ganeshcodehub flobaader dmurat waydenie venerari navidaro etannous bashkirtsevich fguadalix thol-voleak ryannix123 lenocruz bschraven warmchang mohan-raj boy12371 oliversommer caruccio feng-du kindlyops boogermann baileyvw danielraq marcelomrwin lebrisg toanalien bowenislandsong gabrielstein kpsimoulis paddy235 gunnarmorling permalost curvc gjaminon qingfeng786 robert-greathouse troii matteomori agentelinux ggrames xiaoruiguo minikiller ashokkalakoti isreehari macauleycheng demoon abelkour mysticaltech greenwich157 bittarello slaterx beefee alishaika concrain voravitl dpalomar pietrangelo congto ethereumgeeks albertoclarit sureshgaikwad pradeep-kaloori razvanpat pirovskih thakurmhn scm20008 ddefrancescoinstallcentos's Issues
No Attribute 'oo_glusterfs_to_config'
I am getting the following error : " fatal: [console.isreehari.io]: FAILED! => {"failed": true, "msg": "'dict object' has no attribute 'oo_glusterfs_to_config'"} "
Can't deploy apps that have persist volume claims
I finally got my cluster to build correctly with the service catalog installed. I believe what was hanging my install up was that port 80 and 443 were not available to the outside world. Once that was cleared up, my cluster installed with the service broker installed.
My new issue is that I can't deploy any apps from the service catalog that have persistence i.e. cake-php-mysql non-ephemeral. Ephermal apps deploy just fine.
Recreate:
- sh createPV.sh, which is located in /tmp of installcentos
- Run oc get pv to confirm the existence of pv claims. Confirmed to work
- Deploy app that needs persistence. e.g. Jenkins, php-mysql, etc
- Run oc get pv again to confirm app has made a claim. Confirmed to work
Sample output:
=> sourcing 20-validate-variables.sh ... => sourcing 25-validate-replication-variables.sh ... => sourcing 30-base-config.sh ... ---> 18:23:21 Processing basic MySQL configuration files ... => sourcing 60-replication-config.sh ... => sourcing 70-s2i-config.sh ... ---> 18:23:21 Processing additional arbitrary MySQL configuration provided by s2i ... => sourcing 40-paas.cnf ... => sourcing 50-my-tuning.cnf ... ---> 18:23:21 Initializing database ... ---> 18:23:21 Running mysqld --initialize-insecure ... mysqld: Can't create directory '/var/lib/mysql/data/' (Errcode: 17 - File exists) 2018-04-14T18:23:21.872387Z 0 [Warning] TIMESTAMP with implicit DEFAULT value is deprecated. Please use --explicit_defaults_for_timestamp server option (see documentation for more details). 2018-04-14T18:23:21.890075Z 0 [Warning] Duplicate ignore-db-dir directory name 'lost+found' found in the config file(s). Ignoring the duplicate. 2018-04-14T18:23:21.890196Z 0 [ERROR] Aborting
Apps fail to deploy each time, although there is a claim. What could be happening?
Registry-console, Metrics and logging not working
Thanks for this repo.
I tried to follow this repo (and video mentioned in README) and My installation went successful and I am able to login to openshift web console but when I looked in registry-console, metrics and logging, all them are failed to start.
Installation logs:
INSTALLER STATUS *******************************************************************************************************
Initialization : Complete (0:00:25)
Health Check : Complete (0:00:02)
etcd Install : Complete (0:01:17)
Master Install : Complete (0:03:25)
Master Additional Install : Complete (0:00:53)
Node Install : Complete (0:02:46)
Hosted Install : Complete (0:01:38)
Web Console Install : Complete (0:00:29)
Metrics Install : Complete (0:02:53)
Logging Install : Complete (0:04:34)
Service Catalog Install : Complete (0:01:57)
Adding password for user root
cluster role "cluster-admin" added: "root"
******
* Your console is https://console.192.168.31.135.nip.io:8443
* Your username is root
* Your password is password
*
* Login using:
*
$ oc login -u root -p password https://console.192.168.31.135.nip.io:8443/
******
Login successful.
You have access to the following projects and can switch between them with 'oc project <projectname>':
* default
kube-public
kube-service-catalog
kube-system
logging
management-infra
openshift
openshift-infra
openshift-node
openshift-template-service-broker
openshift-web-console
Using project "default".
oc status -v
In project default on server https://console.192.168.31.135.nip.io:8443
https://docker-registry-default.apps.192.168.31.135.nip.io (passthrough) (svc/docker-registry)
dc/docker-registry deploys docker.io/openshift/origin-docker-registry:v3.9.0
deployment #1 deployed 28 minutes ago - 1 pod
svc/kubernetes - 172.30.0.1 ports 443->8443, 53->8053, 53->8053
https://registry-console-default.apps.192.168.31.135.nip.io (passthrough) (svc/registry-console)
dc/registry-console deploys docker.io/cockpit/kubernetes:latest
deployment #1 failed 28 minutes ago: config change
svc/router - 172.30.33.136 ports 80, 443, 1936
dc/router deploys docker.io/openshift/origin-haproxy-router:v3.9.0
deployment #1 deployed 29 minutes ago - 1 pod
Info:
* pod/registry-console-1-deploy has no liveness probe to verify pods are still running.
try: oc set probe pod/registry-console-1-deploy --liveness ...
View details with 'oc describe <resource>/<name>' or list everything with 'oc get all'.
oc adm diagnostics
oc get all --all-namespaces
NAMESPACE NAME DESIRED CURRENT READY UP-TO-DATE AVAILABLE NODE SELECTOR AGE
kube-service-catalog ds/apiserver 1 1 1 1 1 node-role.kubernetes.io/master=true 33m
kube-service-catalog ds/controller-manager 1 1 1 1 1 node-role.kubernetes.io/master=true 32m
logging ds/logging-fluentd 1 1 1 1 1 logging-infra-fluentd=true 34m
openshift-template-service-broker ds/apiserver 1 1 1 1 1 region=infra 32m
NAMESPACE NAME DESIRED CURRENT UP-TO-DATE AVAILABLE AGE
openshift-web-console deploy/webconsole 1 1 1 1 41m
NAMESPACE NAME DESIRED CURRENT READY AGE
openshift-web-console rs/webconsole-84466b9d97 1 1 1 41m
NAMESPACE NAME REVISION DESIRED CURRENT TRIGGERED BY
default deploymentconfigs/docker-registry 1 1 1 config
default deploymentconfigs/registry-console 1 1 0 config
default deploymentconfigs/router 1 1 1 config
logging deploymentconfigs/logging-curator 1 1 1 config
logging deploymentconfigs/logging-es-data-master-nck7hzw3 1 1 0
logging deploymentconfigs/logging-kibana 1 1 0 config
openshift-diagnostic-appcreate-6cgjp deploymentconfigs/diagnostic-appcreate 1 0 1 config
NAMESPACE NAME DOCKER REPO TAGS UPDATED
default imagestreams/registry-console docker-registry.default.svc:5000/default/registry-console latest 42 minutes ago
openshift imagestreams/dotnet docker-registry.default.svc:5000/openshift/dotnet 2.0,latest About an hour ago
openshift imagestreams/dotnet-runtime docker-registry.default.svc:5000/openshift/dotnet-runtime 2.0,latest About an hour ago
openshift imagestreams/httpd docker-registry.default.svc:5000/openshift/httpd latest,2.4 About an hour ago
openshift imagestreams/jenkins docker-registry.default.svc:5000/openshift/jenkins 1,2,latest About an hour ago
openshift imagestreams/mariadb docker-registry.default.svc:5000/openshift/mariadb 10.1,10.2,latest About an hour ago
openshift imagestreams/mongodb docker-registry.default.svc:5000/openshift/mongodb 2.4,2.6,3.2 + 2 more... About an hour ago
openshift imagestreams/mysql docker-registry.default.svc:5000/openshift/mysql 5.5,5.6,5.7 + 1 more... About an hour ago
openshift imagestreams/nginx docker-registry.default.svc:5000/openshift/nginx 1.12,1.8,latest + 1 more... About an hour ago
openshift imagestreams/nodejs docker-registry.default.svc:5000/openshift/nodejs 0.10,4,6 + 2 more... About an hour ago
openshift imagestreams/perl docker-registry.default.svc:5000/openshift/perl 5.16,5.20,5.24 + 1 more... About an hour ago
openshift imagestreams/php docker-registry.default.svc:5000/openshift/php 7.1,latest,5.5 + 2 more... About an hour ago
openshift imagestreams/postgresql docker-registry.default.svc:5000/openshift/postgresql 9.2,9.4,9.5 + 2 more... About an hour ago
openshift imagestreams/python docker-registry.default.svc:5000/openshift/python 3.3,3.4,3.5 + 3 more... About an hour ago
openshift imagestreams/redis docker-registry.default.svc:5000/openshift/redis latest,3.2 About an hour ago
openshift imagestreams/ruby docker-registry.default.svc:5000/openshift/ruby 2.0,2.2,2.3 + 2 more... About an hour ago
openshift imagestreams/wildfly docker-registry.default.svc:5000/openshift/wildfly 10.1,8.1,9.0 + 2 more... About an hour ago
NAMESPACE NAME HOST/PORT PATH SERVICES PORT TERMINATION WILDCARD
default routes/docker-registry docker-registry-default.apps.192.168.31.135.nip.io docker-registry <all> passthrough None
default routes/registry-console registry-console-default.apps.192.168.31.135.nip.io registry-console <all> passthrough None
kube-service-catalog routes/apiserver apiserver-kube-service-catalog.apps.192.168.31.135.nip.io apiserver secure passthrough None
logging routes/logging-kibana kibana.apps.192.168.31.135.nip.io logging-kibana <all> reencrypt/Redirect None
openshift-infra routes/hawkular-metrics hawkular-metrics.apps.192.168.31.135.nip.io hawkular-metrics <all> reencrypt None
NAMESPACE NAME READY STATUS RESTARTS AGE
default po/docker-registry-1-6ng6b 1/1 Running 0 42m
default po/registry-console-1-deploy 0/1 Error 0 42m
default po/router-1-bc68m 1/1 Running 0 43m
kube-service-catalog po/apiserver-6kzm6 1/1 Running 0 33m
kube-service-catalog po/controller-manager-hdkbj 1/1 Running 0 32m
logging po/logging-curator-1-g62b5 1/1 Running 7 35m
logging po/logging-es-data-master-nck7hzw3-1-deploy 0/1 Error 0 34m
logging po/logging-fluentd-b9b6k 1/1 Running 0 34m
logging po/logging-kibana-1-deploy 0/1 Error 0 35m
openshift-diagnostic-appcreate-6cgjp po/diagnostic-appcreate-1-vj6xn 1/1 Running 0 4m
openshift-infra po/hawkular-cassandra-1-f98nf 0/1 CrashLoopBackOff 11 39m
openshift-infra po/hawkular-metrics-s8jln 0/1 CrashLoopBackOff 14 39m
openshift-infra po/heapster-c4kdn 0/1 Running 4 38m
openshift-template-service-broker po/apiserver-2zqfs 1/1 Running 0 32m
openshift-web-console po/webconsole-84466b9d97-c5rzb 1/1 Running 2 41m
NAMESPACE NAME DESIRED CURRENT READY AGE
default rc/docker-registry-1 1 1 1 42m
default rc/registry-console-1 0 0 0 42m
default rc/router-1 1 1 1 43m
logging rc/logging-curator-1 1 1 1 35m
logging rc/logging-es-data-master-nck7hzw3-1 0 0 0 34m
logging rc/logging-kibana-1 0 0 0 35m
openshift-diagnostic-appcreate-6cgjp rc/diagnostic-appcreate-1 1 1 1 4m
openshift-infra rc/hawkular-cassandra-1 1 1 0 39m
openshift-infra rc/hawkular-metrics 1 1 0 39m
openshift-infra rc/heapster 1 1 0 39m
NAMESPACE NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE
default svc/docker-registry ClusterIP 172.30.33.55 <none> 5000/TCP 43m
default svc/kubernetes ClusterIP 172.30.0.1 <none> 443/TCP,53/UDP,53/TCP 48m
default svc/registry-console ClusterIP 172.30.164.99 <none> 9000/TCP 42m
default svc/router ClusterIP 172.30.33.136 <none> 80/TCP,443/TCP,1936/TCP 43m
kube-service-catalog svc/apiserver ClusterIP 172.30.111.155 <none> 443/TCP 33m
logging svc/logging-es ClusterIP 172.30.29.28 <none> 9200/TCP 36m
logging svc/logging-es-cluster ClusterIP 172.30.207.66 <none> 9300/TCP 36m
logging svc/logging-es-prometheus ClusterIP 172.30.8.32 <none> 443/TCP 36m
logging svc/logging-kibana ClusterIP 172.30.54.197 <none> 443/TCP 36m
openshift-infra svc/hawkular-cassandra ClusterIP 172.30.180.60 <none> 9042/TCP,9160/TCP,7000/TCP,7001/TCP 40m
openshift-infra svc/hawkular-cassandra-nodes ClusterIP None <none> 9042/TCP,9160/TCP,7000/TCP,7001/TCP 40m
openshift-infra svc/hawkular-metrics ClusterIP 172.30.187.119 <none> 443/TCP 40m
openshift-infra svc/heapster ClusterIP 172.30.163.27 <none> 80/TCP 39m
openshift-template-service-broker svc/apiserver ClusterIP 172.30.255.151 <none> 443/TCP 32m
openshift-web-console svc/webconsole ClusterIP 172.30.207.143 <none> 443/TCP 41m
is openshift-ansible broken ?
on a fresh bare metal install of centos 7.4 following your instruction i got that in gathering_facts
PLAY [Ensure that all non-node hosts are accessible]
TASK [Gathering Facts]
failed! => {"msg": "module (setup) is missing interpreter line"}
looks like a bug of ansible when a setup.py is present in the playbook
ansible/ansible#34109
renaming installcentos/openshift-ansible/utils/setup.{py,cfg} and installcentos/openshift-ansible/setup.{py,cfg}
solve the problem
and i am able to finish the installation properly
the question is how could it work for you or any user of openshift-ansible ?
should i create an issue for openshift-ansible ?
or is it something with my config ?
Fail on finding control plane 60 retries after install
This is a welcome find! And if I can get past my error I can use this to setup a single node well.
I am using this on a MAC and have VMWare Fusion for a 2 CPU 8+ GB VM with 150 GB drive with / and swap as the YT video depicted. I am just using 192.168.1.54.nip.io kind of a URL to keep locally. I get through everything fine installing into a minimum CentOS 7.4 (added KDE desktop) as root until I get to the following piece below.
Waiting for all control plane pods to be come ready (60 retries).
This is when it runs '/usr/bin/oc' to get control plane pods ready. I am not sure what the issue is and am curious if someone can help me figure it out.
Unable to pull etcd image from "registry.fedoraproject.org/latest/etcd"
fatal: [10.8.87.185]: FAILED! => {"changed": false, "cmd": ["docker", "pull", "registry.fedoraproject.org/latest/etcd"], "delta": "0:00:01.606210", "end": "2018-08-29 06:33:59.951167", "msg": "non-zero return code", "rc": 1, "start": "2018-08-29 06:33:58.344957", "stderr": "error pulling image configuration: Get https://cdn.registry.fedoraproject.org/v2/f27/etcd/blobs/sha256:6c1ecdb769b7e0d21f87aa7d5bd0bbd7d456ff802ac03edb1903e65002d89ea8: EOF", "stderr_lines": ["error pulling image configuration: Get https://cdn.registry.fedoraproject.org/v2/f27/etcd/blobs/sha256:6c1ecdb769b7e0d21f87aa7d5bd0bbd7d456ff802ac03edb1903e65002d89ea8: EOF"], "stdout": "Using default tag: latest\nTrying to pull repository registry.fedoraproject.org/latest/etcd ... \nlatest: Pulling from registry.fedoraproject.org/latest/etcd\nff3dab903f92: Pulling fs layer\nf34e1517e3dd: Pulling fs layer", "stdout_lines": ["Using default tag: latest", "Trying to pull repository registry.fedoraproject.org/latest/etcd ... ", "latest: Pulling from registry.fedoraproject.org/latest/etcd", "ff3dab903f92: Pulling fs layer", "f34e1517e3dd: Pulling fs layer"]}
Service Catalog Install Failed.
Dear @gshipley
My Name is Shuichi Ikegami.
your contents is very nice!
but, I executed a task and the following message occurred.
TASK [openshift_service_catalog : wait for api server to be ready]
FAILED - RETRYING: wait for api server to be ready (1 retries left).
0curl: (7) Failed connect to apiserver.kube-service-catalog.svc:443; Connection refused"], "stdout": "", "stdout_lines": []}
to retry, use: --limit @/root/installcentos/openshift-ansible/playbooks/byo/config.retry
Connection to the Openshift server seems to have problems with name resolution, so it resolved, but how to retry Ansible's Playbook?
Persistent volumes?
It doesn't look like the scripts provision persistent volumes for PVCs. Is this the case?
./tmp/createPV.sh volume names are bugged
Here is the output I am getting:
persistentvolume "vol1" created
created volume 1
persistentvolume "vol21" created
created volume 2
persistentvolume "vol321" created
created volume 3
persistentvolume "vol4321" created
created volume 4
persistentvolume "vol54321" created
created volume 5
persistentvolume "vol654321" created
created volume 6
persistentvolume "vol7654321" created
created volume 7
persistentvolume "vol87654321" created
created volume 8
persistentvolume "vol987654321" created
created volume 9
persistentvolume "vol10987654321" created
created volume 10
'oc get pvc' outputs:
NAME CAPACITY ACCESS MODES RECLAIM POLICY STATUS CLAIM STORAGECLASS REASON AGE
vol1 500Gi RWO,RWX Recycle Bound gitlab/gitlab-ce-redis-data 2m
vol10987654321 500Gi RWO,RWX Recycle Available 1m
vol21 500Gi RWO,RWX Recycle Bound gitlab/gitlab-ce-etc 2m
vol321 500Gi RWO,RWX Recycle Bound gitlab/gitlab-ce-data 2m
vol4321 500Gi RWO,RWX Recycle Bound gitlab/gitlab-ce-postgresql 2m
vol54321 500Gi RWO,RWX Recycle Available 1m
vol654321 500Gi RWO,RWX Recycle Available 1m
vol7654321 500Gi RWO,RWX Recycle Available 1m
vol87654321 500Gi RWO,RWX Recycle Available 1m
vol987654321 500Gi RWO,RWX Recycle Available 1m
[Question] Upgrade OpenShift version
Is it possible to use this script to upgrade a previously installed OpenShift instance with the same script?
Failed Install and odd STDOUT Messages
Hi, I'm on a fresh VM with CentOS 7 (not sure if that's ideal) and I'm seeing some odd messages upon execution of the script:
$ curl https://raw.githubusercontent.com/gshipley/installcentos/master/install-openshift.sh -a | /bin/bash
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
100 5126 100 5126 0 0 17026 0 --:--:-- --:--:-- --:--:-- 17086
* OpenShift version: echo "* Your IP is $IP "
******
Loaded plugins: fastestmirror
Loading mirror speeds from cached hostfile
* base: mirror.rackspace.com
Notice how the * OpenShift version: echo "* Your IP is $IP "
bit of things is garbled? Has anyone else encountered this? Also, not using the -a
switch, my terminal never actually prompted me for mandatory variables :/ Maybe I'm missing some key package to run this wrapper?
Also, here's some error snippets from the install process:
TASK [Evaluate groups - Fail if no etcd hosts group is defined] ****************
fatal: [localhost]: FAILED! => {"changed": false, "msg": "Running etcd as an embedded service is no longer supported. If this is a new install please define an 'etcd' group with either one, three or five hosts. These hosts may be the same hosts as your masters. If this is an upgrade please see https://docs.openshift.com/container-platform/latest/install_config/upgrading/migrating_embedded_etcd.html for documentation on how to migrate from embedded to external etcd.\n"}
to retry, use: --limit @/home/kentos/openshift-ansible/playbooks/prerequisites.retry
.
.
.
Usage:
htpasswd [-cimBdpsDv] [-C cost] passwordfile username
htpasswd -b[cmBdpsDv] [-C cost] passwordfile username password
htpasswd -n[imBdps] [-C cost] username
htpasswd -nb[mBdps] [-C cost] username password
-c Create a new file.
.
.
.
/bin/bash: line 155: oc: command not found
Failed to restart origin-master-api.service: Unit not found.
******
* Your console is https://console.97.116.33.228.nip.io:echo "* Your password is $PASSWORD "
* Your username is echo "******"
* Your password is echo "* Your domain is $DOMAIN "
*
* Login using:
*
$ oc login -u echo "******" -p echo "* Your domain is $DOMAIN " https://console.97.116.44.1.nip.io:echo "* Your password is $PASSWORD "/
******
/bin/bash: line 170: oc: command not found
RPM install fails due to wrong format of VERSION variable
Current versioning variable (VERSION=v3.9.0) fails when doing a RPM install (containerized=False).
Specific line where it fails is:
https://github.com/openshift/openshift-ansible/blob/master/roles/openshift_cli/tasks/main.yml#L3
as that would result in following (non-existent) package name to be installed "origin-clientsv3.9.0" while it should be "origin-clients-3.9.0"
To solve this, the correct way to set version in inventory file would be:
openshift_release=v3.9.0
openshift_pkg_version=-3.9.0
Ansible in service_catalog is always fail
I can't success run the script. It always failed at the service_catalog.
It will check the "https://apiserver.kube-service-catalog.svc/healthz" but I find there is no kubenetes POD or service for this address.
I use pure Centos 7.2 minimal version.
Below is my output of ansible message and kubectl get all.
TASK [openshift_service_catalog : wait for api server to be ready] ***********************************************************************************************************
FAILED - RETRYING: wait for api server to be ready (60 retries left).
FAILED - RETRYING: wait for api server to be ready (59 retries left).
FAILED - RETRYING: wait for api server to be ready (58 retries left).
fatal: [192.168.200.40]: FAILED! => {"attempts": 60, "changed": false, "content": "", "msg": "Status code was -1 and not [200]: Request failed: <urlopen error [Errno 111] Con
nection refused>", "redirected": false, "status": -1, "url": "https://apiserver.kube-service-catalog.svc/healthz"}
to retry, use: --limit @/root/openshift-ansible/playbooks/byo/config.retry
[root@openshift ~]# kubectl get all
NAME DOCKER REPO TAGS UPDATED
imagestreams/registry-console docker-registry.default.svc:5000/default/registry-console latest 3 hours ago
NAME REVISION DESIRED CURRENT TRIGGERED BY
deploymentconfigs/docker-registry 1 1 1 config
deploymentconfigs/registry-console 1 1 0 config
deploymentconfigs/router 5 1 1 config
NAME HOST/PORT PATH SERVICES PORT TERMINATION WILDCARD
routes/docker-registry docker-registry-default.apps.macauley.com docker-registry passthrough None
routes/registry-console registry-console-default.apps.macauley.com registry-console passthrough None
NAME READY STATUS RESTARTS AGE
po/docker-registry-1-b996v 0/1 Error 0 2h
po/router-5-mqzs2 1/1 Running 0 15m
NAME DESIRED CURRENT READY AGE
rc/docker-registry-1 1 1 0 2h
rc/registry-console-1 0 0 0 2h
rc/router-1 0 0 0 2h
rc/router-2 0 0 0 2h
rc/router-3 0 0 0 1h
rc/router-4 0 0 0 33m
rc/router-5 1 1 1 15m
NAME CLUSTER-IP EXTERNAL-IP PORT(S) AGE
svc/docker-registry 172.30.82.155 5000/TCP 2h
svc/kubernetes 172.30.0.1 443/TCP,53/UDP,53/TCP 3h
svc/registry-console 172.30.12.197 9000/TCP 2h
svc/router 172.30.224.129 80/TCP,443/TCP,1936/TCP 2h
TSB not available
Let's collect all the info we can ...
python-passlib rpm must be installed on control host
when i install openshift origin 3.7.1 yesterday, i found ansible is 2.5.0,then following error show.
TASK [openshift_metrics : Check that python-passlib is available on the control host] ***
fatal: [125.222.244.17]: FAILED! => {
"assertion": "'not installed' not in passlib_result.stdout",
"changed": false,
"evaluated_to": false,
"msg": "python-passlib rpm must be installed on control host"
}
to retry, use: --limit @/root/project/installcentos/openshift-ansible/playbooks/byo/config.retry
thank for your help!
import_playbook' is not a valid attribute for a Play
hi ,
i am trying to deploy using the install script and i am getting this error:
import_playbook' is not a valid attribute for a Play
config.yml does not exist anymore
There is no file playbooks/byo/config.yaml
anymore.
Is playbooks/deploy_cluster.yml
the right one?
Persistent registry
Check if we could you something like hostPath
PV to make the registry persistent.
oc: command not found (htpasswd error?)
Looks to me like the oc client is not installed.
Im running the script on a fresh CentOS 7 installation after yum update. After running the script i get the following output:
INSTALLER STATUS ******************************************************************************************************************************
Initialization : Complete (0:00:18)
Health Check : Complete (0:00:02)
etcd Install : Complete (0:01:40)
Master Install : In Progress (0:00:07)
This phase can be restarted by running: playbooks/openshift-master/config.yml
Failure summary:
1. Hosts: x.x.x.x
Play: Create OpenShift certificates for master hosts
Task: Pull CLI Image
Message: non-zero return code
htpasswd: cannot modify file /etc/origin/master/htpasswd; use '-c' to create it
./install-openshift.sh: line 103: oc: command not found
Failed to restart origin-master-api.service: Unit not found.
******
* Your console is https://console.x.x.x:8443
* Your username is xxxxxx
* Your password is xxxxxxxxx
*
* Login using:
*
$ oc login -u koen -p Ql1ckl1k https://console.openshift.qlic.nl:8443/
******
./install-openshift.sh: line 118: oc: command not found
ElasticSearch logging crash
[root@openshift01 installcentos]# oc project logging
Now using project "logging" on server "https://openshift01:8443".
[root@openshift01 installcentos]# oc get pod
NAME READY STATUS RESTARTS AGE
logging-curator-1-2cqkr 0/1 CrashLoopBackOff 5 25m
logging-es-data-master-db07udx7-1-deploy 0/1 Error 0 24m
logging-fluentd-nd6nv 1/1 Running 0 25m
logging-kibana-1-gjfvj 2/2 Running 0 26m
[root@openshift01 installcentos]# oc logs logging-es-data-master-db07udx7-1-deploy
--> Scaling logging-es-data-master-db07udx7-1 to 1
error: update acceptor rejected logging-es-data-master-db07udx7-1: watch closed before Until timeout
--> Error listing events for replication controller logging-es-data-master-db07udx7-1: Get https://172.30.0.1:443/api/v1/namespaces/logging/events?fieldSelector=involvedObject.kind%3DReplicationController%2CinvolvedObject.name%3Dlogging-es-data-master-db07udx7-1%2CinvolvedObject.namespace%3Dlogging%2CinvolvedObject.uid%3D43dc018d-2dbe-11e8-9588-0800271bdf18: dial tcp 172.30.0.1:443: getsockopt: connection refused
oc: command not found & unable to connect to a server to handle "persistentvolumes"
I have 2 errors:
- oc: command not found
I had to copy the oc binary from https://github.com/openshift/origin/releases manually otherwise your script always ended up withoc: command not found
.
Does your script install the oc tools? Your video did not seem to require the oc binary before running the script. - Unable to connect to a server to handle "persistentvolumes"
Once I copiedoc
manually in my PATH, I re-ran the script and I got this error at the end
INSTALLER STATUS ****************************************************************************************************************************************************************************************************************************
Initialization : Complete (0:00:18)
Health Check : Complete (0:00:01)
etcd Install : Complete (0:00:31)
Master Install : Complete (0:02:29)
Master Additional Install : Complete (0:00:26)
Node Install : Complete (0:00:47)
Hosted Install : Complete (0:00:47)
Web Console Install : Complete (0:00:38)
Service Catalog Install : Complete (0:00:49)
Adding password for user pierrick
error: Missing or incomplete configuration info. Please login or point to an existing, complete config file:
1. Via the command-line flag --config
2. Via the KUBECONFIG environment variable
3. In your home directory as ~/.kube/config
To view or setup config directly use the 'config' command.
See 'oc adm policy add-cluster-role-to-user -h' for help and examples.
error: unable to connect to a server to handle "persistentvolumes": Missing or incomplete configuration info. Please login or point to an existing, complete config file:
1. Via the command-line flag --config
2. Via the KUBECONFIG environment variable
3. In your home directory as ~/.kube/config
To view or setup config directly use the 'config' command.
created volume 1
error: unable to connect to a server to handle "persistentvolumes": Missing or incomplete configuration info. Please login or point to an existing, complete config file:
1. Via the command-line flag --config
2. Via the KUBECONFIG environment variable
3. In your home directory as ~/.kube/config
To view or setup config directly use the 'config' command.
Openshift Excluder Failed to Install
TASK [openshift_excluder : Install openshift excluder] *************************
fatal: [console.techdope.io]: FAILED! => {"changed": false, "failed": true, "msg": "No package matching 'origin-excluder-1.4.0*' found available, installed or updated", "rc": 126, "results": ["No package matching 'origin-excluder-1.4.0*' found available, installed or updated"]}
What is the process to install master and node on dedicate VMs
First of all thank you for your contribution, I am able to install when I used a single VM for all the services.
But this script is not working when I used dedicated VMs for master and node. Is there any addition changes required or this project doesn't support a multi node cluster ?
Thanks,
Mahesh.
metrics/view archive link doesnt appear on openshift 3.9 webconsole
I have installed openshift-3.9 version and it doesnt show any view arhive/metrics link on the UI web console
Upgrade to 3.9
And test the installation process.
Persistent volumes
Provide several persistent volumes by default as part of the installation process.
Unable to restart service origin-master-api
Hi,
installation script failed with below error, any clue on this?
FAILED! => {"changed": false, "msg": "Unable to restart service origin-master-api: Job for origin-master-api.service failed because the control process exited with error code. See "systemctl status origin-master-api.service" and "journalctl -xe" for details.\n"}
openshift network is not working
Hello,
I have installed Openshift Origin 3.9 using the repo. And deployed a customized application in it. But when I try to access the application using route, it does not work.
I have installed openshift in a VM on Proxmox VE. (80G disk, 48G RAM, 1 socket & 6 cores, bridged networking)
When I check for the interfaces on my VM, (using ip a) I see that docker0, br0, ovs-system are in DOWN state, while tun0 is in UNKNOWN state.
This is the output of "ip a"
[root@openshift ~]# ip a
1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000
link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
inet 127.0.0.1/8 scope host lo
valid_lft forever preferred_lft forever
inet6 ::1/128 scope host
valid_lft forever preferred_lft forever
2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast state UP group default qlen 1000
link/ether 62:15:5b:9f:58:7b brd ff:ff:ff:ff:ff:ff
inet 192.168.10.216/24 brd 192.168.10.255 scope global noprefixroute eth0
valid_lft forever preferred_lft forever
inet6 fe80::a9de:8d6b:44be:2f4e/64 scope link noprefixroute
valid_lft forever preferred_lft forever
3: docker0: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500 qdisc noqueue state DOWN group default
link/ether 02:42:81:45:d4:52 brd ff:ff:ff:ff:ff:ff
inet 172.17.0.1/16 scope global docker0
valid_lft forever preferred_lft forever
4: ovs-system: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN group default qlen 1000
link/ether d2:ea:0c:27:5b:47 brd ff:ff:ff:ff:ff:ff
8: br0: <BROADCAST,MULTICAST> mtu 1450 qdisc noop state DOWN group default qlen 1000
link/ether 92:c8:a1:94:7a:45 brd ff:ff:ff:ff:ff:ff
9: vxlan_sys_4789: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 65520 qdisc noqueue master ovs-system state UNKNOWN group default qlen 1000
link/ether 2a:3b:49:ca:e3:99 brd ff:ff:ff:ff:ff:ff
inet6 fe80::283b:49ff:feca:e399/64 scope link
valid_lft forever preferred_lft forever
10: tun0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1450 qdisc noqueue state UNKNOWN group default qlen 1000
link/ether 16:b2:34:58:81:a4 brd ff:ff:ff:ff:ff:ff
inet 10.128.0.1/23 brd 10.128.1.255 scope global tun0
valid_lft forever preferred_lft forever
inet6 fe80::14b2:34ff:fe58:81a4/64 scope link
valid_lft forever preferred_lft forever
11: veth28b72cfe@if3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1450 qdisc noqueue master ovs-system state UP group default
link/ether 32:63:f5:60:ec:33 brd ff:ff:ff:ff:ff:ff link-netnsid 0
inet6 fe80::3063:f5ff:fe60:ec33/64 scope link
valid_lft forever preferred_lft forever
12: vethec8de5f5@if3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1450 qdisc noqueue master ovs-system state UP group default
link/ether f2:78:5a:7b:03:9b brd ff:ff:ff:ff:ff:ff link-netnsid 1
inet6 fe80::f078:5aff:fe7b:39b/64 scope link
valid_lft forever preferred_lft forever
13: vethcf4bab48@if3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1450 qdisc noqueue master ovs-system state UP group default
link/ether 12:2f:61:71:8b:8e brd ff:ff:ff:ff:ff:ff link-netnsid 2
inet6 fe80::102f:61ff:fe71:8b8e/64 scope link
valid_lft forever preferred_lft forever
14: vethba340af9@if3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1450 qdisc noqueue master ovs-system state UP group default
link/ether 46:68:af:7f:53:34 brd ff:ff:ff:ff:ff:ff link-netnsid 3
inet6 fe80::4468:afff:fe7f:5334/64 scope link
valid_lft forever preferred_lft forever
15: vethaa9933bb@if3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1450 qdisc noqueue master ovs-system state UP group default
link/ether aa:83:d9:c7:6c:a8 brd ff:ff:ff:ff:ff:ff link-netnsid 4
inet6 fe80::a883:d9ff:fec7:6ca8/64 scope link
valid_lft forever preferred_lft forever
16: vethd81db48b@if3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1450 qdisc noqueue master ovs-system state UP group default
link/ether 56:f8:df:0c:65:46 brd ff:ff:ff:ff:ff:ff link-netnsid 5
inet6 fe80::54f8:dfff:fe0c:6546/64 scope link
valid_lft forever preferred_lft forever
17: veth2d069743@if3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1450 qdisc noqueue master ovs-system state UP group default
link/ether d6:d8:bc:5d:18:40 brd ff:ff:ff:ff:ff:ff link-netnsid 6
inet6 fe80::d4d8:bcff:fe5d:1840/64 scope link
valid_lft forever preferred_lft forever
18: veth1eb83fb0@if3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1450 qdisc noqueue master ovs-system state UP group default
link/ether 8a:9d:b5:15:7a:72 brd ff:ff:ff:ff:ff:ff link-netnsid 7
inet6 fe80::889d:b5ff:fe15:7a72/64 scope link
valid_lft forever preferred_lft forever
19: veth07290fcc@if3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1450 qdisc noqueue master ovs-system state UP group default
link/ether 3a:da:06:eb:26:e9 brd ff:ff:ff:ff:ff:ff link-netnsid 8
inet6 fe80::38da:6ff:feeb:26e9/64 scope link
valid_lft forever preferred_lft forever
20: vetha2a639f2@if3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1450 qdisc noqueue master ovs-system state UP group default
link/ether 72:83:98:58:08:33 brd ff:ff:ff:ff:ff:ff link-netnsid 9
inet6 fe80::7083:98ff:fe58:833/64 scope link
valid_lft forever preferred_lft forever
21: vethd181d9e2@if3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1450 qdisc noqueue master ovs-system state UP group default
link/ether 66:a2:b8:70:4c:02 brd ff:ff:ff:ff:ff:ff link-netnsid 10
inet6 fe80::64a2:b8ff:fe70:4c02/64 scope link
valid_lft forever preferred_lft forever
22: veth0fbae888@if3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1450 qdisc noqueue master ovs-system state UP group default
link/ether a2:62:45:ed:73:13 brd ff:ff:ff:ff:ff:ff link-netnsid 11
inet6 fe80::a062:45ff:feed:7313/64 scope link
valid_lft forever preferred_lft forever
23: veth696055ef@if3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1450 qdisc noqueue master ovs-system state UP group default
link/ether 3a:2b:56:1d:6a:96 brd ff:ff:ff:ff:ff:ff link-netnsid 12
inet6 fe80::382b:56ff:fe1d:6a96/64 scope link
valid_lft forever preferred_lft forever
24: vethcb4caedb@if3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1450 qdisc noqueue master ovs-system state UP group default
link/ether c6:bf:79:14:bb:fd brd ff:ff:ff:ff:ff:ff link-netnsid 13
inet6 fe80::c4bf:79ff:fe14:bbfd/64 scope link
valid_lft forever preferred_lft forever
I haven't changed any default configuration while installing Openshift.
Can anyone please guide me through this.
Thank you.
unable to read config file: open /etc/origin/node/resolv.conf: no such file or directory
I am unable to install openshift all in one getting error while restarting node during openshift-node task /etc/origin/node/resolv.conf: no such file or directory
Checked /etc/origin/node/ directory
it contains following file
total 44
drwx------. 2 root root 4096 Jul 11 19:13 .
drwx------. 10 root root 4096 Jul 11 19:13 ..
-rw-r--r--. 1 root root 1070 Jul 11 19:12 ca.crt
-rw-------. 1 root root 1313 Jul 11 19:13 node-config.yaml
-rw-r--r--. 1 root root 63 Jul 11 19:13 node-dnsmasq.conf
-rw-r--r--. 1 root root 2258 Jul 11 19:12 server.crt
-rw-------. 1 root root 1675 Jul 11 19:12 server.key
-rw-r--r--. 1 root root 1139 Jul 11 19:12 system:node:xxxxxxxx.crt
-rw-------. 1 root root 1679 Jul 11 19:12 system:node:xxxxxx.key
-rw-------. 1 root root 5866 Jul 11 19:12 system:node:xxxxxxx.kubeconfig
cat node/node-dnsmasq.conf
server=/in-addr.arpa/127.0.0.1
server=/cluster.local/127.0.0.1
any idea why this is happening
error during the installation process, Service Catalog Install
After
# ansible-playbook -i ./installcentos/inventory.erb ./openshift-ansible/playbooks/byo/config.yml
..................
TASK [ansible_service_broker : create ansible-service-broker service] *******************************************************************************
fatal: [console.phoenixit.ru]: FAILED! => {"changed": false, "failed": true, "msg": {"cmd": "/usr/local/bin/oc replace -f /tmp/asb-ygaDEN -n openshi}
to retry, use: --limit @/root/openshift-ansible/playbooks/byo/config.retry
PLAY RECAP ******************************************************************************************************************************************
console.phoenixit.ru : ok=645 changed=89 unreachable=0 failed=1
localhost : ok=13 changed=0 unreachable=0 failed=0
INSTALLER STATUS ************************************************************************************************************************************
Initialization : Complete (0:00:40)
Health Check : Complete (0:00:03)
etcd Install : Complete (0:00:39)
Master Install : Complete (0:03:01)
Master Additional Install : Complete (0:00:22)
Node Install : Complete (0:01:39)
Hosted Install : Complete (0:01:52)
Service Catalog Install : In Progress (0:00:46)
This phase can be restarted by running: playbooks/openshift-service-catalog/config.yml
missing origin-clients & vol.yaml
Little feedback from my experience using you script.
- During the creation of persistent volume, the script need the file vol.yaml as template to create 200 PV.
The script do no copy vol.yaml from git to you local host , so you have to copy it manually. - Creating PV and login to openshift require the command "oc" from missing package origin-clients.
origin-clients is not a dependency or installed by the script. Can be fix by installing this one after ansible playbooks. Ansible playbooks take care to create the repository "CentOS OpenShift Origin" where the package origin-clients is available. - Run twice the script is disruptive for docker if you use $DISK.
If you want to use and additional disk for docker, the script take care to wipefs and delete /var/lib/docker folder. If you miss something and run the script a second time, the script will break the storage and docker will be not able to start. You end up with a busy storage unable to be deleted and docker not able to start.
Otherwise, this is a very good script and I am an happy user of OpenShift Origine, thanks to you !
Changing web-console public URL
Hi,
I used this project to setup the Openshift in one RHEL 7.5 vm. I had a typo in the env variable for domain and ended up configuring the console URL as " https://console.116.87.123.194.nip.io:8443/"
I would like to know if there is any way to change the public URL. I tried changing the /etc/origin/master/master-config.yaml and restarting the services, didn't help much.
Thanks for the help,
H
Installer fails on 'restart master controllers'
Hello,
Ran the installer on a fresh CentOS minimal install (ESXi VM, 80GB SSD, 16GB RAM) following video instructions.
Step where it fails:
RUNNING HANDLER [openshift_master : restart master controllers] ******************************************
FAILED - RETRYING: restart master controllers (5 retries left).
FAILED - RETRYING: restart master controllers (4 retries left).
FAILED - RETRYING: restart master controllers (3 retries left).
FAILED - RETRYING: restart master controllers (2 retries left).
FAILED - RETRYING: restart master controllers (1 retries left).
fatal: [10.0.0.12]: FAILED! => {"attempts": 5, "changed": true, "cmd": ["systemctl", "restart", "origin-master-controllers"], "delta": "0:00:10.058201", "end": "2018-06-22 16:47:39.196244", "msg": "non-zero return code", "rc": 1, "start": "2018-06-22 16:47:29.138043", "stderr": "Job for origin-master-controllers.service failed because the control process exited with error code. See \"systemctl status origin-master-controllers.service\" and \"journalctl -xe\" for details.", "stderr_lines": ["Job for origin-master-controllers.service failed because the control process exited with error code. See \"systemctl status origin-master-controllers.service\" and \"journalctl -xe\" for details."], "stdout": "", "stdout_lines": []}
to retry, use: --limit @/root/installcentos/openshift-ansible/playbooks/deploy_cluster.retry
Output from # systemctl status origin-master-controllers.service
:
● origin-master-controllers.service - Atomic OpenShift Master Controllers
Loaded: loaded (/etc/systemd/system/origin-master-controllers.service; enabled; vendor preset: disabled)
Active: activating (auto-restart) (Result: exit-code) since Fri 2018-06-22 20:25:03 EDT; 918ms ago
Docs: https://github.com/openshift/origin
Process: 4909 ExecStop=/usr/bin/docker stop origin-master-controllers (code=exited, status=1/FAILURE)
Process: 4827 ExecStartPost=/usr/bin/sleep 10 (code=exited, status=0/SUCCESS)
Process: 4826 ExecStart=/usr/bin/docker run --rm --privileged --net=host --name origin-master-controllers --env-file=/etc/sysconfig/origin-master-controllers -v /var/lib/origin:/var/lib/origin -v /var/run/docker.sock:/var/run/docker.sock -v /etc/origin:/etc/origin -v /etc/pki:/etc/pki:ro openshift/origin:${IMAGE_VERSION} start master controllers --config=${CONFIG_FILE} $OPTIONS (code=exited, status=255)
Process: 4818 ExecStartPre=/usr/bin/docker rm -f origin-master-controllers (code=exited, status=1/FAILURE)
Main PID: 4826 (code=exited, status=255)
Jun 22 20:25:03 quicksilver systemd[1]: Failed to start Atomic OpenShift Master Controllers.
Jun 22 20:25:03 quicksilver systemd[1]: Unit origin-master-controllers.service entered failed state.
Jun 22 20:25:03 quicksilver systemd[1]: origin-master-controllers.service failed.
Output from # journalctl -xe
:
journalctl-xe-output.txt
Thank you in advance!
Patching the single cluster?
I'm not sure how to apply this patch to run OpenWhisk on OpenShift. https://github.com/projectodd/openwhisk-openshift
minishift openshift config set --patch
'{"admissionConfig":
{"pluginConfig":
{"openshift.io/ImagePolicy":
{"configuration":
{"apiVersion": "v1",
"kind": "ImagePolicyConfig",
"resolveImages": "AttemptRewrite"}}}}}'
admin role
Ran this installer on a Centos build in AWS (Centos official 7), all went fine until the last steps setting up the user
solved adding user by
**sudo** htpasswd -b /etc/origin/master/htpasswd ${USERNAME} ${PASSWORD}
but do not have rights to add admin cluster role (have tried adding --as system:admin
)
So I can login in, but do not have an admin access role
3.10 support?
Has anyone tried install 3.10?
Error at Task - Evaluate Groups - Fail if no etcd hosts group is defined
Hello,
I am receiving the following error. How do i define an etcd group?
TASK [Evaluate groups - Fail if no etcd hosts group is defined] ***********************************************************************************************************************
fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "msg": "Running etcd as an embedded service is no longer supported. If this is a new install please define an 'etcd' group with either one or three hosts. These hosts may be the same hosts as your masters. If this is an upgrade you may set openshift_master_unsupported_embedded_etcd=true until a migration playbook becomes available.\n"}
to retry, use: --limit @/root/openshift-ansible/playbooks/byo/config.retry
PLAY RECAP ****************************************************************************************************************************************************************************
localhost : ok=1 changed=0 unreachable=0 failed=1
Failure summary:
- Hosts: localhost
Play: Populate config host groups
Task: Evaluate groups - Fail if no etcd hosts group is defined
Message: Running etcd as an embedded service is no longer supported. If this is a new install please define an 'etcd' group with either one or three hosts. These hosts may be the same hosts as your masters. If this is an upgrade you may set openshift_master_unsupported_embedded_etcd=true until a migration playbook becomes available.
DOMAIN name issue while installing
Following your video i install openshift with domain like
export DOMAIN=${DOMAIN:="$(curl -s ipinfo.io/ip).nip.io"}
above command produce result like 85.80.161.23.nip.io
After successful installation i am not able to access console, i verified domain and found that now system is showing values like below
77.79.13.222.nip.io
How to manage this situation?
scale up
Hi Grant,
This is not really an issue. I am trying to add a node to my installation using another vm on my laptop.
Is it possible to use the scaleup.yml playbook to add a node to this installation?
It looks like the scaleup.yaml tries to install 3.7.0 (instead of 3.7.1) which fails.
Thanks,
Mohammad
how to deploy cluster
how to deploy cluster
Cannot create new app
Cloning "https://github.com/openshift/nodejs-ex.git " ...
| error: fatal: unable to access 'https://github.com/openshift/nodejs-ex.git/': Could not resolve host: github.com; Unknown error
err during running the play book
An exception occurred during task execution. To see the full traceback, use -vvv. The error was: RepresenterError: cannot represent an object: htpasswd_auth
fatal: [console.techdope.io]: FAILED! => {"failed": true, "msg": "Unexpected failure during module execution.", "stdout": ""}
Allow web console on port 443
Allow web console to be accessible using port 443
Error on install and finaly Catalog only with 8 items
Hi @gshipley thank you for your great video tutorial, but I had an error while installing. After installation console is up and running but I can see only 8 items in the catalogue while you have 39 when you first open the openshift console
Here the log
TASK [template_service_broker : Apply template file] **********************************************************************************************************************************************************
fatal: [192.168.1.111]: FAILED! => {"changed": true, "cmd": "/usr/local/bin/oc process -f \"/tmp/tsb-ansible-kbTpfe/apiserver-template.yaml\" --param API_SERVER_CONFIG=\"kind: TemplateServiceBrokerConfig\napiVersion: config.templateservicebroker.openshift.io/v1\ntemplateNamespaces:\n- openshift\n\" --param IMAGE=\"docker.io/openshift/origin:latest\" --param NODE_SELECTOR='{\"region\": \"infra\"}' | /usr/local/bin/oc apply -f -", "delta": "0:00:00.207589", "end": "2018-03-17 06:51:21.036675", "msg": "non-zero return code", "rc": 1, "start": "2018-03-17 06:51:20.829086", "stderr": "Unable to connect to the server: x509: certificate signed by unknown authority\nerror: unable to process template: Post https://192.168.1.111:8443/apis/template.openshift.io/v1/namespaces/myproject/processedtemplates: x509: certificate signed by unknown authority", "stderr_lines": ["Unable to connect to the server: x509: certificate signed by unknown authority", "error: unable to process template: Post https://192.168.1.111:8443/apis/template.openshift.io/v1/namespaces/myproject/processedtemplates: x509: certificate signed by unknown authority"], "stdout": "", "stdout_lines": []}
to retry, use: --limit @/root/installcentos/openshift-ansible/playbooks/byo/config.retry
PLAY RECAP ****************************************************************************************************************************************************************************************************
192.168.1.111 : ok=635 changed=230 unreachable=0 failed=1
localhost : ok=13 changed=0 unreachable=0 failed=0
INSTALLER STATUS **********************************************************************************************************************************************************************************************
Initialization : Complete
Health Check : Complete
etcd Install : Complete
Master Install : Complete
Master Additional Install : Complete
Node Install : Complete
Hosted Install : Complete
Service Catalog Install : In Progress
This phase can be restarted by running: playbooks/byo/openshift-cluster/service-catalog.yml
openshift services are not getting started
Hi,
I am getting following error message after completion of ./install-openshift.sh. my environment is single host centos7 running as VM in KVM. please help
To view or setup config directly use the 'config' command.
created volume 199
chcon: can't apply partial context to unlabeled file ‘/mnt/data/vol200’
error: Missing or incomplete configuration info. Please login or point to an existing, complete config file:
- Via the command-line flag --config
- Via the KUBECONFIG environment variable
- In your home directory as ~/.kube/config
To view or setup config directly use the 'config' command.
created volume 200
- Your console is https://console.bruttech.com:8443
- Your username is admin
- Your password is admin123
- Login using:
$ oc login -u admin -p admin123 https://console.bruttech.com:8443/
error: dial tcp 192.168.1.168:8443: getsockopt: connection refused - verify you have provided the correct host and port and that the server is currently running.
OpenShift Failed installation at "Install Ceph storage plugin dependencies"
Enviornment: VM with CentOs 7.5
ASK [openshift_node : Install Ceph storage plugin dependencies] *********************************************************************************************************
FAILED - RETRYING: Install Ceph storage plugin dependencies (3 retries left).
FAILED - RETRYING: Install Ceph storage plugin dependencies (2 retries left).
FAILED - RETRYING: Install Ceph storage plugin dependencies (1 retries left).
fatal: [10.0.2.15]: FAILED! => {"attempts": 3, "changed": false, "module_stderr": "Shared connection to 10.0.2.15 closed.\r\n", "module_stdout": "error: rpmdb: BDB0113 Thread/proce
ss 28774/140193218025280 failed: BDB1507 Thread died in Berkeley DB library\r\nerror: db5 error(-30973) from dbenv->failchk: BDB0087 DB_RUNRECOVERY: Fatal error, run database recov
ery\r\nerror: cannot open Packages index using db5 - (-30973)\r\nerror: cannot open Packages database in /var/lib/rpm\r\nTraceback (most recent call last):\r\n File "/tmp/ansibl
e_xLSJTe/ansible_module_yum.py", line 1509, in \r\n main()\r\n File "/tmp/ansible_xLSJTe/ansible_module_yum.py", line 1470, in main\r\n my.conf\r\n File "/usr/l
ib/python2.7/site-packages/yum/init.py", line 1078, in \r\n conf = property(fget=lambda self: self.getConfig(),\r\n File "/usr/lib/python2.7/site-packages/yum/_
init__.py", line 349, in _getConfig\r\n startupconf = config.readStartupConfig(fn, root, releasever)\r\n File "/usr/lib/python2.7/site-packages/yum/config.py", line 1112, in
readStartupConfig\r\n startupconf.distroverpkg)\r\n File "/usr/lib/python2.7/site-packages/yum/config.py", line 1254, in _getsysver\r\n raise Errors.YumBaseError("Error:
" + str(e))\r\nyum.Errors.YumBaseError: Error: rpmdb open failed\r\n", "msg": "MODULE FAILURE", "rc": 1}
install fails during "RETRYING: Verify that TSB is running"
OS: Linux centos 3.10.0-693.17.1.el7.x86_64 #1 SMP Thu Jan 25 20:13:58 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux
Exports which are required by the installation step:
PASSWORD=rob
USERNAME=rob
DISK=/dev/sda
DOMAIN=192.168.0.135.nip.io
After starting following script:
curl https://raw.githubusercontent.com/gshipley/installcentos/master/install-openshift.sh | /bin/bash
which calls
ansible-playbook -i inventory.ini openshift-ansible/playbooks/byo/config.yml
The ansible playbook failed with following error:
...
TASK [template_service_broker : Verify that TSB is running] ********************
FAILED - RETRYING: Verify that TSB is running (120 retries left).
FAILED - RETRYING: Verify that TSB is running (119 retries left).
FAILED - RETRYING: Verify that TSB is running (118 retries left).
...
FAILED - RETRYING: Verify that TSB is running (2 retries left).
FAILED - RETRYING: Verify that TSB is running (1 retries left).
fatal: [192.168.0.135]: FAILED! => {"attempts": 120, "changed": false, "cmd": ["curl", "-k", "https://apiserver.openshift-template-service-broker.svc/healthz"], "delta": "0:00:01.066711", "end": "2018-03-07 13:44:18.024722", "msg": "non-zero return code", "rc": 7, "start": "2018-03-07 13:44:16.958011", "stderr": " % Total % Received % Xferd Average Speed Time Time Time Current\n Dload Upload Total Spent Left Speed\n\r 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0curl: (7) Failed connect to apiserver.openshift-template-service-broker.svc:443; Connection refused", "stderr_lines": [" % Total % Received % Xferd Average Speed Time Time Time Current", " Dload Upload Total Spent Left Speed", "", " 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0curl: (7) Failed connect to apiserver.openshift-template-service-broker.svc:443; Connection refused"], "stdout": "", "stdout_lines": []}
to retry, use: --limit @/root/openshift-ansible/playbooks/byo/config.retry
PLAY RECAP *********************************************************************
192.168.0.135 : ok=981 changed=351 unreachable=0 failed=1
localhost : ok=13 changed=0 unreachable=0 failed=0
INSTALLER STATUS ***************************************************************
Initialization : Complete
Health Check : Complete
etcd Install : Complete
Master Install : Complete
Master Additional Install : Complete
Node Install : Complete
Hosted Install : Complete
Metrics Install : Complete
Logging Install : Complete
Service Catalog Install : In Progress
This phase can be restarted by running: playbooks/byo/openshift-cluster/service-catalog.yml
Failure summary:
1. Hosts: 192.168.0.135
Play: Service Catalog
Task: Verify that TSB is running
Message: non-zero return code
Adding password for user rob
cluster role "cluster-admin" added: "rob"
******
* Your conosle is https://console.192.168.0.135.nip.io:8443
* Your username is rob
* Your password is rob
*
* Login using:
*
$ oc login -u rob -p rob https://console.192.168.0.135.nip.io:8443/
******
error: dial tcp 192.168.0.135:8443: getsockopt: connection refused - verify you have provided the correct host and port and that the server is currently running.
[root@centos ~]#
simplephp is build OK, but deploy is failed.
Log messages
--> Scaling simplephp-1 to 1
error: update acceptor rejected simplephp-1: pods for rc 'grantproject/simplephp-1' took longer than 600 seconds to become available
Openshift Task to disable legacy master service fails
Hello,
I have a CentOS minimal install VM running on Proxmox VE in which I am running the script provided by G. Shipley.
I am following G. Shipley's youtube video and also instructions given here.
I run into following error:
Task [Openshift_master: Disable the legacy maszter service if it exists]
fatal: [192.168.10.216]: FAILED! => {"changed": false, "msg": "Could not find the requested service origin-master: host"}
Although this error doesn't seem to affect any implementation or later working of openshift but I am curious to know why this appears. Even in my other implementions (1 master VM and 1 node VM) I have same error and several dependency failures.
Also, I have a doubt. According to the Openshift documentation, if we are using "containerized=true", then "docker run" command should be used to implement each service as a separate docker container. But this and many other documents simply follow ansible-playbook command which is suitable for RPM installation.
Any help is appreciated.
Thank you.
missing webconsole service
Hey guys I'm new to DevOps, please help me to point out what is potentially wrong here.
Thanks so much!
Environment
OS: Centos 7.4
RAM: 16 GB
CPU: 4
HD: 200 GB
Repo Steps:
- Setup variables DOMAIN USERNAME PASSWORD using openshift 3.9
- Change /etc/sysconfig/selinux SELINUX=enforcing
- Ran install-openshift.sh
- Visit webconsole i got following error:
missing service (service "webconsole" not found)
missing route (service "webconsole" not found)
Expect
Webconsole to show up when vising https://console.${DOMAIN}.com:8443
Extra Information
when i run docker ps i can see following containers running:
- origin-master-controllers
- origin-master-api
- etcd_container
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.