Coder Social home page Coder Social logo

installcentos's Introduction

Note: Update: November 2019 - I no longer work at Red Hat on the OpenShift team. Given that, I will not be updating this repo to work with OpenShift 4.x on baremetal. My future work will be centered on a more pure upstream kubernetes experience. If a community member wants to add support for 4.x, I will gladly accept a PR.

Install RedHat OKD 3.11 on your own server. For a local only install, it is suggested that you use CDK or MiniShift instead of this repo. This install method is targeted for a single node cluster that has a long life.

This repository is a set of scripts that will allow you easily install the latest version (3.11) of OKD in a single node fashion. What that means is that all of the services required for OKD to function (master, node, etcd, etc.) will all be installed on a single host. The script supports a custom hostname which you can provide using the interactive mode.

If you are wanting to install OCP on RDO (OpenStack)

Michel Peterson has created a wrapper script in his repo that will do all the heavy lifting for you. Check it out!

https://github.com/mpeterson/rdo-openshift-tools

Please do use a clean CentOS system, the script installs all necesary tools and packages including Ansible, container runtime, etc.

Warning about Let's Encrypt setup available on this project: Let's Encrypt only works if the IP is using publicly accessible IP and custom certificates." This feature doesn't work with OpenShift CLI for now.

Installation

  1. Create a VM as explained in https://www.youtube.com/watch?v=ZkFIozGY0IA (this video) by Grant Shipley

  2. Clone this repo

git clone https://github.com/okd-community-install/installcentos.git
  1. Execute the installation script
cd installcentos
./install-openshift.sh

Automation

  1. Define mandatory variables for the installation process
# Domain name to access the cluster
$ export DOMAIN=<public ip address>.nip.io

# User created after installation
$ export USERNAME=<current user name>

# Password for the user
$ export PASSWORD=password
  1. Define optional variables for the installation process
# Instead of using loopback, setup DeviceMapper on this disk.
# !! All data on the disk will be wiped out !!
$ export DISK="/dev/sda"
  1. Run the automagic installation script as root with the environment variable in place:
curl https://raw.githubusercontent.com/okd-community-install/installcentos/master/install-openshift.sh | INTERACTIVE=false /bin/bash

Development

For development it's possible to switch the script repo

# Change location of source repository
$ export SCRIPT_REPO="https://raw.githubusercontent.com/okd-community-install/installcentos/master"
$ curl $SCRIPT_REPO/install-openshift.sh | /bin/bash

Testing

The script is tested using the tooling in the validate directory.

To use the tooling, it's required to create file validate/env.sh with the DigitalOcean API key

export DIGITALOCEAN_TOKEN=""

and then run start.sh to start the provisioning. Once the ssh is connected to the server, the script will atatch to the tmux session running Ansible installer.

To destroy the infrastructure, run the stop.sh script.

installcentos's People

Contributors

accorvin avatar b-dantas avatar brigriffin avatar darkmukke avatar frgomes avatar gastaldi avatar gshipley avatar gunnarmorling avatar hirenvadalia avatar kmarilleau avatar koenhendriks avatar ltupin avatar marekjelen avatar navidaro avatar oveits avatar razvanpat avatar rkreienbuehl avatar ryannix123 avatar slaterx avatar slmingol avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

installcentos's Issues

Failed Install and odd STDOUT Messages

Hi, I'm on a fresh VM with CentOS 7 (not sure if that's ideal) and I'm seeing some odd messages upon execution of the script:

$  curl https://raw.githubusercontent.com/gshipley/installcentos/master/install-openshift.sh -a | /bin/bash
  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current
                                 Dload  Upload   Total   Spent    Left  Speed
100  5126  100  5126    0     0  17026      0 --:--:-- --:--:-- --:--:-- 17086

* OpenShift version: echo "* Your IP is $IP "
******
Loaded plugins: fastestmirror
Loading mirror speeds from cached hostfile
 * base: mirror.rackspace.com

Notice how the * OpenShift version: echo "* Your IP is $IP " bit of things is garbled? Has anyone else encountered this? Also, not using the -a switch, my terminal never actually prompted me for mandatory variables :/ Maybe I'm missing some key package to run this wrapper?

Also, here's some error snippets from the install process:

TASK [Evaluate groups - Fail if no etcd hosts group is defined] ****************
fatal: [localhost]: FAILED! => {"changed": false, "msg": "Running etcd as an embedded service is no longer supported. If this is a new install please define an 'etcd' group with either one, three or five hosts. These hosts may be the same hosts as your masters. If this is an upgrade please see https://docs.openshift.com/container-platform/latest/install_config/upgrading/migrating_embedded_etcd.html for documentation on how to migrate from embedded to external etcd.\n"}
	to retry, use: --limit @/home/kentos/openshift-ansible/playbooks/prerequisites.retry
.
.
.
Usage:
	htpasswd [-cimBdpsDv] [-C cost] passwordfile username
	htpasswd -b[cmBdpsDv] [-C cost] passwordfile username password

	htpasswd -n[imBdps] [-C cost] username
	htpasswd -nb[mBdps] [-C cost] username password
 -c  Create a new file.
.
.
.
/bin/bash: line 155: oc: command not found
Failed to restart origin-master-api.service: Unit not found.
******
* Your console is https://console.97.116.33.228.nip.io:echo "* Your password is $PASSWORD "
* Your username is echo "******"
* Your password is echo "* Your domain is $DOMAIN "
*
* Login using:
*
$ oc login -u echo "******" -p echo "* Your domain is $DOMAIN " https://console.97.116.44.1.nip.io:echo "* Your password is $PASSWORD "/
******
/bin/bash: line 170: oc: command not found

openshift services are not getting started

Hi,
I am getting following error message after completion of ./install-openshift.sh. my environment is single host centos7 running as VM in KVM. please help

To view or setup config directly use the 'config' command.
created volume 199
chcon: can't apply partial context to unlabeled file ‘/mnt/data/vol200’
error: Missing or incomplete configuration info. Please login or point to an existing, complete config file:

  1. Via the command-line flag --config
  2. Via the KUBECONFIG environment variable
  3. In your home directory as ~/.kube/config

To view or setup config directly use the 'config' command.
created volume 200


$ oc login -u admin -p admin123 https://console.bruttech.com:8443/


error: dial tcp 192.168.1.168:8443: getsockopt: connection refused - verify you have provided the correct host and port and that the server is currently running.

python-passlib rpm must be installed on control host

when i install openshift origin 3.7.1 yesterday, i found ansible is 2.5.0,then following error show.

TASK [openshift_metrics : Check that python-passlib is available on the control host] ***
fatal: [125.222.244.17]: FAILED! => {
    "assertion": "'not installed' not in passlib_result.stdout",
    "changed": false,
    "evaluated_to": false,
    "msg": "python-passlib rpm must be installed on control host"
}
        to retry, use: --limit @/root/project/installcentos/openshift-ansible/playbooks/byo/config.retry

thank for your help!

openshift network is not working

Hello,
I have installed Openshift Origin 3.9 using the repo. And deployed a customized application in it. But when I try to access the application using route, it does not work.
I have installed openshift in a VM on Proxmox VE. (80G disk, 48G RAM, 1 socket & 6 cores, bridged networking)
When I check for the interfaces on my VM, (using ip a) I see that docker0, br0, ovs-system are in DOWN state, while tun0 is in UNKNOWN state.
This is the output of "ip a"

[root@openshift ~]# ip a
1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000
link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
inet 127.0.0.1/8 scope host lo
valid_lft forever preferred_lft forever
inet6 ::1/128 scope host
valid_lft forever preferred_lft forever
2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast state UP group default qlen 1000
link/ether 62:15:5b:9f:58:7b brd ff:ff:ff:ff:ff:ff
inet 192.168.10.216/24 brd 192.168.10.255 scope global noprefixroute eth0
valid_lft forever preferred_lft forever
inet6 fe80::a9de:8d6b:44be:2f4e/64 scope link noprefixroute
valid_lft forever preferred_lft forever
3: docker0: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500 qdisc noqueue state DOWN group default
link/ether 02:42:81:45:d4:52 brd ff:ff:ff:ff:ff:ff
inet 172.17.0.1/16 scope global docker0
valid_lft forever preferred_lft forever
4: ovs-system: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN group default qlen 1000
link/ether d2:ea:0c:27:5b:47 brd ff:ff:ff:ff:ff:ff
8: br0: <BROADCAST,MULTICAST> mtu 1450 qdisc noop state DOWN group default qlen 1000
link/ether 92:c8:a1:94:7a:45 brd ff:ff:ff:ff:ff:ff
9: vxlan_sys_4789: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 65520 qdisc noqueue master ovs-system state UNKNOWN group default qlen 1000
link/ether 2a:3b:49:ca:e3:99 brd ff:ff:ff:ff:ff:ff
inet6 fe80::283b:49ff:feca:e399/64 scope link
valid_lft forever preferred_lft forever
10: tun0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1450 qdisc noqueue state UNKNOWN group default qlen 1000
link/ether 16:b2:34:58:81:a4 brd ff:ff:ff:ff:ff:ff
inet 10.128.0.1/23 brd 10.128.1.255 scope global tun0
valid_lft forever preferred_lft forever
inet6 fe80::14b2:34ff:fe58:81a4/64 scope link
valid_lft forever preferred_lft forever
11: veth28b72cfe@if3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1450 qdisc noqueue master ovs-system state UP group default
link/ether 32:63:f5:60:ec:33 brd ff:ff:ff:ff:ff:ff link-netnsid 0
inet6 fe80::3063:f5ff:fe60:ec33/64 scope link
valid_lft forever preferred_lft forever
12: vethec8de5f5@if3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1450 qdisc noqueue master ovs-system state UP group default
link/ether f2:78:5a:7b:03:9b brd ff:ff:ff:ff:ff:ff link-netnsid 1
inet6 fe80::f078:5aff:fe7b:39b/64 scope link
valid_lft forever preferred_lft forever
13: vethcf4bab48@if3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1450 qdisc noqueue master ovs-system state UP group default
link/ether 12:2f:61:71:8b:8e brd ff:ff:ff:ff:ff:ff link-netnsid 2
inet6 fe80::102f:61ff:fe71:8b8e/64 scope link
valid_lft forever preferred_lft forever
14: vethba340af9@if3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1450 qdisc noqueue master ovs-system state UP group default
link/ether 46:68:af:7f:53:34 brd ff:ff:ff:ff:ff:ff link-netnsid 3
inet6 fe80::4468:afff:fe7f:5334/64 scope link
valid_lft forever preferred_lft forever
15: vethaa9933bb@if3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1450 qdisc noqueue master ovs-system state UP group default
link/ether aa:83:d9:c7:6c:a8 brd ff:ff:ff:ff:ff:ff link-netnsid 4
inet6 fe80::a883:d9ff:fec7:6ca8/64 scope link
valid_lft forever preferred_lft forever
16: vethd81db48b@if3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1450 qdisc noqueue master ovs-system state UP group default
link/ether 56:f8:df:0c:65:46 brd ff:ff:ff:ff:ff:ff link-netnsid 5
inet6 fe80::54f8:dfff:fe0c:6546/64 scope link
valid_lft forever preferred_lft forever
17: veth2d069743@if3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1450 qdisc noqueue master ovs-system state UP group default
link/ether d6:d8:bc:5d:18:40 brd ff:ff:ff:ff:ff:ff link-netnsid 6
inet6 fe80::d4d8:bcff:fe5d:1840/64 scope link
valid_lft forever preferred_lft forever
18: veth1eb83fb0@if3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1450 qdisc noqueue master ovs-system state UP group default
link/ether 8a:9d:b5:15:7a:72 brd ff:ff:ff:ff:ff:ff link-netnsid 7
inet6 fe80::889d:b5ff:fe15:7a72/64 scope link
valid_lft forever preferred_lft forever
19: veth07290fcc@if3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1450 qdisc noqueue master ovs-system state UP group default
link/ether 3a:da:06:eb:26:e9 brd ff:ff:ff:ff:ff:ff link-netnsid 8
inet6 fe80::38da:6ff:feeb:26e9/64 scope link
valid_lft forever preferred_lft forever
20: vetha2a639f2@if3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1450 qdisc noqueue master ovs-system state UP group default
link/ether 72:83:98:58:08:33 brd ff:ff:ff:ff:ff:ff link-netnsid 9
inet6 fe80::7083:98ff:fe58:833/64 scope link
valid_lft forever preferred_lft forever
21: vethd181d9e2@if3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1450 qdisc noqueue master ovs-system state UP group default
link/ether 66:a2:b8:70:4c:02 brd ff:ff:ff:ff:ff:ff link-netnsid 10
inet6 fe80::64a2:b8ff:fe70:4c02/64 scope link
valid_lft forever preferred_lft forever
22: veth0fbae888@if3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1450 qdisc noqueue master ovs-system state UP group default
link/ether a2:62:45:ed:73:13 brd ff:ff:ff:ff:ff:ff link-netnsid 11
inet6 fe80::a062:45ff:feed:7313/64 scope link
valid_lft forever preferred_lft forever
23: veth696055ef@if3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1450 qdisc noqueue master ovs-system state UP group default
link/ether 3a:2b:56:1d:6a:96 brd ff:ff:ff:ff:ff:ff link-netnsid 12
inet6 fe80::382b:56ff:fe1d:6a96/64 scope link
valid_lft forever preferred_lft forever
24: vethcb4caedb@if3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1450 qdisc noqueue master ovs-system state UP group default
link/ether c6:bf:79:14:bb:fd brd ff:ff:ff:ff:ff:ff link-netnsid 13
inet6 fe80::c4bf:79ff:fe14:bbfd/64 scope link
valid_lft forever preferred_lft forever

I haven't changed any default configuration while installing Openshift.
Can anyone please guide me through this.
Thank you.

Can't deploy apps that have persist volume claims

I finally got my cluster to build correctly with the service catalog installed. I believe what was hanging my install up was that port 80 and 443 were not available to the outside world. Once that was cleared up, my cluster installed with the service broker installed.

My new issue is that I can't deploy any apps from the service catalog that have persistence i.e. cake-php-mysql non-ephemeral. Ephermal apps deploy just fine.

Recreate:

  1. sh createPV.sh, which is located in /tmp of installcentos
  2. Run oc get pv to confirm the existence of pv claims. Confirmed to work
  3. Deploy app that needs persistence. e.g. Jenkins, php-mysql, etc
  4. Run oc get pv again to confirm app has made a claim. Confirmed to work

Sample output:

=> sourcing 20-validate-variables.sh ... => sourcing 25-validate-replication-variables.sh ... => sourcing 30-base-config.sh ... ---> 18:23:21 Processing basic MySQL configuration files ... => sourcing 60-replication-config.sh ... => sourcing 70-s2i-config.sh ... ---> 18:23:21 Processing additional arbitrary MySQL configuration provided by s2i ... => sourcing 40-paas.cnf ... => sourcing 50-my-tuning.cnf ... ---> 18:23:21 Initializing database ... ---> 18:23:21 Running mysqld --initialize-insecure ... mysqld: Can't create directory '/var/lib/mysql/data/' (Errcode: 17 - File exists) 2018-04-14T18:23:21.872387Z 0 [Warning] TIMESTAMP with implicit DEFAULT value is deprecated. Please use --explicit_defaults_for_timestamp server option (see documentation for more details). 2018-04-14T18:23:21.890075Z 0 [Warning] Duplicate ignore-db-dir directory name 'lost+found' found in the config file(s). Ignoring the duplicate. 2018-04-14T18:23:21.890196Z 0 [ERROR] Aborting

Apps fail to deploy each time, although there is a claim. What could be happening?

Openshift Excluder Failed to Install

TASK [openshift_excluder : Install openshift excluder] *************************
fatal: [console.techdope.io]: FAILED! => {"changed": false, "failed": true, "msg": "No package matching 'origin-excluder-1.4.0*' found available, installed or updated", "rc": 126, "results": ["No package matching 'origin-excluder-1.4.0*' found available, installed or updated"]}

Registry-console, Metrics and logging not working

Thanks for this repo.

I tried to follow this repo (and video mentioned in README) and My installation went successful and I am able to login to openshift web console but when I looked in registry-console, metrics and logging, all them are failed to start.

Installation logs:

INSTALLER STATUS *******************************************************************************************************
Initialization             : Complete (0:00:25)
Health Check               : Complete (0:00:02)
etcd Install               : Complete (0:01:17)
Master Install             : Complete (0:03:25)
Master Additional Install  : Complete (0:00:53)
Node Install               : Complete (0:02:46)
Hosted Install             : Complete (0:01:38)
Web Console Install        : Complete (0:00:29)
Metrics Install            : Complete (0:02:53)
Logging Install            : Complete (0:04:34)
Service Catalog Install    : Complete (0:01:57)

Adding password for user root
cluster role "cluster-admin" added: "root"
******
* Your console is https://console.192.168.31.135.nip.io:8443
* Your username is root
* Your password is password
*
* Login using:
*
$ oc login -u root -p password https://console.192.168.31.135.nip.io:8443/
******
Login successful.

You have access to the following projects and can switch between them with 'oc project <projectname>':

  * default
    kube-public
    kube-service-catalog
    kube-system
    logging
    management-infra
    openshift
    openshift-infra
    openshift-node
    openshift-template-service-broker
    openshift-web-console

Using project "default".

oc status -v

In project default on server https://console.192.168.31.135.nip.io:8443

https://docker-registry-default.apps.192.168.31.135.nip.io (passthrough) (svc/docker-registry)
  dc/docker-registry deploys docker.io/openshift/origin-docker-registry:v3.9.0
    deployment #1 deployed 28 minutes ago - 1 pod

svc/kubernetes - 172.30.0.1 ports 443->8443, 53->8053, 53->8053

https://registry-console-default.apps.192.168.31.135.nip.io (passthrough) (svc/registry-console)
  dc/registry-console deploys docker.io/cockpit/kubernetes:latest
    deployment #1 failed 28 minutes ago: config change

svc/router - 172.30.33.136 ports 80, 443, 1936
  dc/router deploys docker.io/openshift/origin-haproxy-router:v3.9.0
    deployment #1 deployed 29 minutes ago - 1 pod

Info:
  * pod/registry-console-1-deploy has no liveness probe to verify pods are still running.
    try: oc set probe pod/registry-console-1-deploy --liveness ...

View details with 'oc describe <resource>/<name>' or list everything with 'oc get all'.

oc adm diagnostics

oc_adm_diagnostics.txt

oc get all --all-namespaces

NAMESPACE                           NAME                    DESIRED   CURRENT   READY     UP-TO-DATE   AVAILABLE   NODE SELECTOR                         AGE
kube-service-catalog                ds/apiserver            1         1         1         1            1           node-role.kubernetes.io/master=true   33m
kube-service-catalog                ds/controller-manager   1         1         1         1            1           node-role.kubernetes.io/master=true   32m
logging                             ds/logging-fluentd      1         1         1         1            1           logging-infra-fluentd=true            34m
openshift-template-service-broker   ds/apiserver            1         1         1         1            1           region=infra                          32m

NAMESPACE               NAME                DESIRED   CURRENT   UP-TO-DATE   AVAILABLE   AGE
openshift-web-console   deploy/webconsole   1         1         1            1           41m

NAMESPACE               NAME                       DESIRED   CURRENT   READY     AGE
openshift-web-console   rs/webconsole-84466b9d97   1         1         1         41m

NAMESPACE                              NAME                                                REVISION   DESIRED   CURRENT   TRIGGERED BY
default                                deploymentconfigs/docker-registry                   1          1         1         config
default                                deploymentconfigs/registry-console                  1          1         0         config
default                                deploymentconfigs/router                            1          1         1         config
logging                                deploymentconfigs/logging-curator                   1          1         1         config
logging                                deploymentconfigs/logging-es-data-master-nck7hzw3   1          1         0       
logging                                deploymentconfigs/logging-kibana                    1          1         0         config
openshift-diagnostic-appcreate-6cgjp   deploymentconfigs/diagnostic-appcreate              1          0         1         config

NAMESPACE   NAME                            DOCKER REPO                                                 TAGS                          UPDATED
default     imagestreams/registry-console   docker-registry.default.svc:5000/default/registry-console   latest                        42 minutes ago
openshift   imagestreams/dotnet             docker-registry.default.svc:5000/openshift/dotnet           2.0,latest                    About an hour ago
openshift   imagestreams/dotnet-runtime     docker-registry.default.svc:5000/openshift/dotnet-runtime   2.0,latest                    About an hour ago
openshift   imagestreams/httpd              docker-registry.default.svc:5000/openshift/httpd            latest,2.4                    About an hour ago
openshift   imagestreams/jenkins            docker-registry.default.svc:5000/openshift/jenkins          1,2,latest                    About an hour ago
openshift   imagestreams/mariadb            docker-registry.default.svc:5000/openshift/mariadb          10.1,10.2,latest              About an hour ago
openshift   imagestreams/mongodb            docker-registry.default.svc:5000/openshift/mongodb          2.4,2.6,3.2 + 2 more...       About an hour ago
openshift   imagestreams/mysql              docker-registry.default.svc:5000/openshift/mysql            5.5,5.6,5.7 + 1 more...       About an hour ago
openshift   imagestreams/nginx              docker-registry.default.svc:5000/openshift/nginx            1.12,1.8,latest + 1 more...   About an hour ago
openshift   imagestreams/nodejs             docker-registry.default.svc:5000/openshift/nodejs           0.10,4,6 + 2 more...          About an hour ago
openshift   imagestreams/perl               docker-registry.default.svc:5000/openshift/perl             5.16,5.20,5.24 + 1 more...    About an hour ago
openshift   imagestreams/php                docker-registry.default.svc:5000/openshift/php              7.1,latest,5.5 + 2 more...    About an hour ago
openshift   imagestreams/postgresql         docker-registry.default.svc:5000/openshift/postgresql       9.2,9.4,9.5 + 2 more...       About an hour ago
openshift   imagestreams/python             docker-registry.default.svc:5000/openshift/python           3.3,3.4,3.5 + 3 more...       About an hour ago
openshift   imagestreams/redis              docker-registry.default.svc:5000/openshift/redis            latest,3.2                    About an hour ago
openshift   imagestreams/ruby               docker-registry.default.svc:5000/openshift/ruby             2.0,2.2,2.3 + 2 more...       About an hour ago
openshift   imagestreams/wildfly            docker-registry.default.svc:5000/openshift/wildfly          10.1,8.1,9.0 + 2 more...      About an hour ago

NAMESPACE              NAME                      HOST/PORT                                                   PATH      SERVICES           PORT      TERMINATION          WILDCARD
default                routes/docker-registry    docker-registry-default.apps.192.168.31.135.nip.io                    docker-registry    <all>     passthrough          None
default                routes/registry-console   registry-console-default.apps.192.168.31.135.nip.io                   registry-console   <all>     passthrough          None
kube-service-catalog   routes/apiserver          apiserver-kube-service-catalog.apps.192.168.31.135.nip.io             apiserver          secure    passthrough          None
logging                routes/logging-kibana     kibana.apps.192.168.31.135.nip.io                                     logging-kibana     <all>     reencrypt/Redirect   None
openshift-infra        routes/hawkular-metrics   hawkular-metrics.apps.192.168.31.135.nip.io                           hawkular-metrics   <all>     reencrypt            None

NAMESPACE                              NAME                                          READY     STATUS             RESTARTS   AGE
default                                po/docker-registry-1-6ng6b                    1/1       Running            0          42m
default                                po/registry-console-1-deploy                  0/1       Error              0          42m
default                                po/router-1-bc68m                             1/1       Running            0          43m
kube-service-catalog                   po/apiserver-6kzm6                            1/1       Running            0          33m
kube-service-catalog                   po/controller-manager-hdkbj                   1/1       Running            0          32m
logging                                po/logging-curator-1-g62b5                    1/1       Running            7          35m
logging                                po/logging-es-data-master-nck7hzw3-1-deploy   0/1       Error              0          34m
logging                                po/logging-fluentd-b9b6k                      1/1       Running            0          34m
logging                                po/logging-kibana-1-deploy                    0/1       Error              0          35m
openshift-diagnostic-appcreate-6cgjp   po/diagnostic-appcreate-1-vj6xn               1/1       Running            0          4m
openshift-infra                        po/hawkular-cassandra-1-f98nf                 0/1       CrashLoopBackOff   11         39m
openshift-infra                        po/hawkular-metrics-s8jln                     0/1       CrashLoopBackOff   14         39m
openshift-infra                        po/heapster-c4kdn                             0/1       Running            4          38m
openshift-template-service-broker      po/apiserver-2zqfs                            1/1       Running            0          32m
openshift-web-console                  po/webconsole-84466b9d97-c5rzb                1/1       Running            2          41m

NAMESPACE                              NAME                                   DESIRED   CURRENT   READY     AGE
default                                rc/docker-registry-1                   1         1         1         42m
default                                rc/registry-console-1                  0         0         0         42m
default                                rc/router-1                            1         1         1         43m
logging                                rc/logging-curator-1                   1         1         1         35m
logging                                rc/logging-es-data-master-nck7hzw3-1   0         0         0         34m
logging                                rc/logging-kibana-1                    0         0         0         35m
openshift-diagnostic-appcreate-6cgjp   rc/diagnostic-appcreate-1              1         1         1         4m
openshift-infra                        rc/hawkular-cassandra-1                1         1         0         39m
openshift-infra                        rc/hawkular-metrics                    1         1         0         39m
openshift-infra                        rc/heapster                            1         1         0         39m

NAMESPACE                           NAME                           TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)                               AGE
default                             svc/docker-registry            ClusterIP   172.30.33.55     <none>        5000/TCP                              43m
default                             svc/kubernetes                 ClusterIP   172.30.0.1       <none>        443/TCP,53/UDP,53/TCP                 48m
default                             svc/registry-console           ClusterIP   172.30.164.99    <none>        9000/TCP                              42m
default                             svc/router                     ClusterIP   172.30.33.136    <none>        80/TCP,443/TCP,1936/TCP               43m
kube-service-catalog                svc/apiserver                  ClusterIP   172.30.111.155   <none>        443/TCP                               33m
logging                             svc/logging-es                 ClusterIP   172.30.29.28     <none>        9200/TCP                              36m
logging                             svc/logging-es-cluster         ClusterIP   172.30.207.66    <none>        9300/TCP                              36m
logging                             svc/logging-es-prometheus      ClusterIP   172.30.8.32      <none>        443/TCP                               36m
logging                             svc/logging-kibana             ClusterIP   172.30.54.197    <none>        443/TCP                               36m
openshift-infra                     svc/hawkular-cassandra         ClusterIP   172.30.180.60    <none>        9042/TCP,9160/TCP,7000/TCP,7001/TCP   40m
openshift-infra                     svc/hawkular-cassandra-nodes   ClusterIP   None             <none>        9042/TCP,9160/TCP,7000/TCP,7001/TCP   40m
openshift-infra                     svc/hawkular-metrics           ClusterIP   172.30.187.119   <none>        443/TCP                               40m
openshift-infra                     svc/heapster                   ClusterIP   172.30.163.27    <none>        80/TCP                                39m
openshift-template-service-broker   svc/apiserver                  ClusterIP   172.30.255.151   <none>        443/TCP                               32m
openshift-web-console               svc/webconsole                 ClusterIP   172.30.207.143   <none>        443/TCP                               41m

Unable to pull etcd image from "registry.fedoraproject.org/latest/etcd"

fatal: [10.8.87.185]: FAILED! => {"changed": false, "cmd": ["docker", "pull", "registry.fedoraproject.org/latest/etcd"], "delta": "0:00:01.606210", "end": "2018-08-29 06:33:59.951167", "msg": "non-zero return code", "rc": 1, "start": "2018-08-29 06:33:58.344957", "stderr": "error pulling image configuration: Get https://cdn.registry.fedoraproject.org/v2/f27/etcd/blobs/sha256:6c1ecdb769b7e0d21f87aa7d5bd0bbd7d456ff802ac03edb1903e65002d89ea8: EOF", "stderr_lines": ["error pulling image configuration: Get https://cdn.registry.fedoraproject.org/v2/f27/etcd/blobs/sha256:6c1ecdb769b7e0d21f87aa7d5bd0bbd7d456ff802ac03edb1903e65002d89ea8: EOF"], "stdout": "Using default tag: latest\nTrying to pull repository registry.fedoraproject.org/latest/etcd ... \nlatest: Pulling from registry.fedoraproject.org/latest/etcd\nff3dab903f92: Pulling fs layer\nf34e1517e3dd: Pulling fs layer", "stdout_lines": ["Using default tag: latest", "Trying to pull repository registry.fedoraproject.org/latest/etcd ... ", "latest: Pulling from registry.fedoraproject.org/latest/etcd", "ff3dab903f92: Pulling fs layer", "f34e1517e3dd: Pulling fs layer"]}

OpenShift Failed installation at "Install Ceph storage plugin dependencies"

Enviornment: VM with CentOs 7.5

ASK [openshift_node : Install Ceph storage plugin dependencies] *********************************************************************************************************
FAILED - RETRYING: Install Ceph storage plugin dependencies (3 retries left).
FAILED - RETRYING: Install Ceph storage plugin dependencies (2 retries left).
FAILED - RETRYING: Install Ceph storage plugin dependencies (1 retries left).
fatal: [10.0.2.15]: FAILED! => {"attempts": 3, "changed": false, "module_stderr": "Shared connection to 10.0.2.15 closed.\r\n", "module_stdout": "error: rpmdb: BDB0113 Thread/proce
ss 28774/140193218025280 failed: BDB1507 Thread died in Berkeley DB library\r\nerror: db5 error(-30973) from dbenv->failchk: BDB0087 DB_RUNRECOVERY: Fatal error, run database recov
ery\r\nerror: cannot open Packages index using db5 - (-30973)\r\nerror: cannot open Packages database in /var/lib/rpm\r\nTraceback (most recent call last):\r\n File "/tmp/ansibl
e_xLSJTe/ansible_module_yum.py", line 1509, in \r\n main()\r\n File "/tmp/ansible_xLSJTe/ansible_module_yum.py", line 1470, in main\r\n my.conf\r\n File "/usr/l
ib/python2.7/site-packages/yum/init.py", line 1078, in \r\n conf = property(fget=lambda self: self.getConfig(),\r\n File "/usr/lib/python2.7/site-packages/yum/_
init__.py", line 349, in _getConfig\r\n startupconf = config.readStartupConfig(fn, root, releasever)\r\n File "/usr/lib/python2.7/site-packages/yum/config.py", line 1112, in
readStartupConfig\r\n startupconf.distroverpkg)\r\n File "/usr/lib/python2.7/site-packages/yum/config.py", line 1254, in _getsysver\r\n raise Errors.YumBaseError("Error:
" + str(e))\r\nyum.Errors.YumBaseError: Error: rpmdb open failed\r\n", "msg": "MODULE FAILURE", "rc": 1}

Persistent registry

Check if we could you something like hostPath PV to make the registry persistent.

Installer fails on 'restart master controllers'

Hello,

Ran the installer on a fresh CentOS minimal install (ESXi VM, 80GB SSD, 16GB RAM) following video instructions.

Step where it fails:

RUNNING HANDLER [openshift_master : restart master controllers] ******************************************
FAILED - RETRYING: restart master controllers (5 retries left).
FAILED - RETRYING: restart master controllers (4 retries left).
FAILED - RETRYING: restart master controllers (3 retries left).
FAILED - RETRYING: restart master controllers (2 retries left).
FAILED - RETRYING: restart master controllers (1 retries left).
fatal: [10.0.0.12]: FAILED! => {"attempts": 5, "changed": true, "cmd": ["systemctl", "restart", "origin-master-controllers"], "delta": "0:00:10.058201", "end": "2018-06-22 16:47:39.196244", "msg": "non-zero return code", "rc": 1, "start": "2018-06-22 16:47:29.138043", "stderr": "Job for origin-master-controllers.service failed because the control process exited with error code. See \"systemctl status origin-master-controllers.service\" and \"journalctl -xe\" for details.", "stderr_lines": ["Job for origin-master-controllers.service failed because the control process exited with error code. See \"systemctl status origin-master-controllers.service\" and \"journalctl -xe\" for details."], "stdout": "", "stdout_lines": []}
        to retry, use: --limit @/root/installcentos/openshift-ansible/playbooks/deploy_cluster.retry

Output from # systemctl status origin-master-controllers.service:

● origin-master-controllers.service - Atomic OpenShift Master Controllers
   Loaded: loaded (/etc/systemd/system/origin-master-controllers.service; enabled; vendor preset: disabled)
   Active: activating (auto-restart) (Result: exit-code) since Fri 2018-06-22 20:25:03 EDT; 918ms ago
     Docs: https://github.com/openshift/origin
  Process: 4909 ExecStop=/usr/bin/docker stop origin-master-controllers (code=exited, status=1/FAILURE)
  Process: 4827 ExecStartPost=/usr/bin/sleep 10 (code=exited, status=0/SUCCESS)
  Process: 4826 ExecStart=/usr/bin/docker run --rm --privileged --net=host --name origin-master-controllers --env-file=/etc/sysconfig/origin-master-controllers -v /var/lib/origin:/var/lib/origin -v /var/run/docker.sock:/var/run/docker.sock -v /etc/origin:/etc/origin -v /etc/pki:/etc/pki:ro openshift/origin:${IMAGE_VERSION} start master controllers --config=${CONFIG_FILE} $OPTIONS (code=exited, status=255)
  Process: 4818 ExecStartPre=/usr/bin/docker rm -f origin-master-controllers (code=exited, status=1/FAILURE)
 Main PID: 4826 (code=exited, status=255)

Jun 22 20:25:03 quicksilver systemd[1]: Failed to start Atomic OpenShift Master Controllers.
Jun 22 20:25:03 quicksilver systemd[1]: Unit origin-master-controllers.service entered failed state.
Jun 22 20:25:03 quicksilver systemd[1]: origin-master-controllers.service failed.

Output from # journalctl -xe:
journalctl-xe-output.txt

Thank you in advance!

Fail on finding control plane 60 retries after install

This is a welcome find! And if I can get past my error I can use this to setup a single node well.

I am using this on a MAC and have VMWare Fusion for a 2 CPU 8+ GB VM with 150 GB drive with / and swap as the YT video depicted. I am just using 192.168.1.54.nip.io kind of a URL to keep locally. I get through everything fine installing into a minimum CentOS 7.4 (added KDE desktop) as root until I get to the following piece below.

Waiting for all control plane pods to be come ready (60 retries).

This is when it runs '/usr/bin/oc' to get control plane pods ready. I am not sure what the issue is and am curious if someone can help me figure it out.

Changing web-console public URL

Hi,
I used this project to setup the Openshift in one RHEL 7.5 vm. I had a typo in the env variable for domain and ended up configuring the console URL as " https://console.116.87.123.194.nip.io:8443/"

I would like to know if there is any way to change the public URL. I tried changing the /etc/origin/master/master-config.yaml and restarting the services, didn't help much.

Thanks for the help,
H

admin role

Ran this installer on a Centos build in AWS (Centos official 7), all went fine until the last steps setting up the user
solved adding user by
**sudo** htpasswd -b /etc/origin/master/htpasswd ${USERNAME} ${PASSWORD}
but do not have rights to add admin cluster role (have tried adding --as system:admin)

So I can login in, but do not have an admin access role

No Attribute 'oo_glusterfs_to_config'

I am getting the following error : " fatal: [console.isreehari.io]: FAILED! => {"failed": true, "msg": "'dict object' has no attribute 'oo_glusterfs_to_config'"} "

Persistent volumes

Provide several persistent volumes by default as part of the installation process.

is openshift-ansible broken ?

on a fresh bare metal install of centos 7.4 following your instruction i got that in gathering_facts

PLAY [Ensure that all non-node hosts are accessible] 
TASK [Gathering Facts]
failed! => {"msg": "module (setup) is missing interpreter line"}

looks like a bug of ansible when a setup.py is present in the playbook
ansible/ansible#34109
renaming installcentos/openshift-ansible/utils/setup.{py,cfg} and installcentos/openshift-ansible/setup.{py,cfg}
solve the problem
and i am able to finish the installation properly

the question is how could it work for you or any user of openshift-ansible ?
should i create an issue for openshift-ansible ?
or is it something with my config ?

oc: command not found (htpasswd error?)

Looks to me like the oc client is not installed.

Im running the script on a fresh CentOS 7 installation after yum update. After running the script i get the following output:

INSTALLER STATUS ******************************************************************************************************************************
Initialization             : Complete (0:00:18)
Health Check               : Complete (0:00:02)
etcd Install               : Complete (0:01:40)
Master Install             : In Progress (0:00:07)
        This phase can be restarted by running: playbooks/openshift-master/config.yml



Failure summary:


  1. Hosts:   x.x.x.x
     Play:     Create OpenShift certificates for master hosts
     Task:     Pull CLI Image
     Message:  non-zero return code
htpasswd: cannot modify file /etc/origin/master/htpasswd; use '-c' to create it
./install-openshift.sh: line 103: oc: command not found
Failed to restart origin-master-api.service: Unit not found.
******
* Your console is https://console.x.x.x:8443
* Your username is xxxxxx 
* Your password is xxxxxxxxx
*
* Login using:
*
$ oc login -u koen -p Ql1ckl1k https://console.openshift.qlic.nl:8443/
******
./install-openshift.sh: line 118: oc: command not found

DOMAIN name issue while installing

Following your video i install openshift with domain like

export DOMAIN=${DOMAIN:="$(curl -s ipinfo.io/ip).nip.io"}

above command produce result like 85.80.161.23.nip.io

After successful installation i am not able to access console, i verified domain and found that now system is showing values like below
77.79.13.222.nip.io

How to manage this situation?

Error on install and finaly Catalog only with 8 items

Hi @gshipley thank you for your great video tutorial, but I had an error while installing. After installation console is up and running but I can see only 8 items in the catalogue while you have 39 when you first open the openshift console

Here the log

TASK [template_service_broker : Apply template file] **********************************************************************************************************************************************************
fatal: [192.168.1.111]: FAILED! => {"changed": true, "cmd": "/usr/local/bin/oc process -f \"/tmp/tsb-ansible-kbTpfe/apiserver-template.yaml\" --param API_SERVER_CONFIG=\"kind: TemplateServiceBrokerConfig\napiVersion: config.templateservicebroker.openshift.io/v1\ntemplateNamespaces:\n- openshift\n\" --param IMAGE=\"docker.io/openshift/origin:latest\" --param NODE_SELECTOR='{\"region\": \"infra\"}' | /usr/local/bin/oc apply -f -", "delta": "0:00:00.207589", "end": "2018-03-17 06:51:21.036675", "msg": "non-zero return code", "rc": 1, "start": "2018-03-17 06:51:20.829086", "stderr": "Unable to connect to the server: x509: certificate signed by unknown authority\nerror: unable to process template: Post https://192.168.1.111:8443/apis/template.openshift.io/v1/namespaces/myproject/processedtemplates: x509: certificate signed by unknown authority", "stderr_lines": ["Unable to connect to the server: x509: certificate signed by unknown authority", "error: unable to process template: Post https://192.168.1.111:8443/apis/template.openshift.io/v1/namespaces/myproject/processedtemplates: x509: certificate signed by unknown authority"], "stdout": "", "stdout_lines": []}
        to retry, use: --limit @/root/installcentos/openshift-ansible/playbooks/byo/config.retry

PLAY RECAP ****************************************************************************************************************************************************************************************************
192.168.1.111              : ok=635  changed=230  unreachable=0    failed=1
localhost                  : ok=13   changed=0    unreachable=0    failed=0


INSTALLER STATUS **********************************************************************************************************************************************************************************************
Initialization             : Complete
Health Check               : Complete
etcd Install               : Complete
Master Install             : Complete
Master Additional Install  : Complete
Node Install               : Complete
Hosted Install             : Complete
Service Catalog Install    : In Progress
        This phase can be restarted by running: playbooks/byo/openshift-cluster/service-catalog.yml

And here my how my console looks like
image

Service Catalog Install Failed.

Dear @gshipley
My Name is Shuichi Ikegami.
your contents is very nice!
but, I executed a task and the following message occurred.

TASK [openshift_service_catalog : wait for api server to be ready]
FAILED - RETRYING: wait for api server to be ready (1 retries left).
0curl: (7) Failed connect to apiserver.kube-service-catalog.svc:443; Connection refused"], "stdout": "", "stdout_lines": []}
to retry, use: --limit @/root/installcentos/openshift-ansible/playbooks/byo/config.retry

Connection to the Openshift server seems to have problems with name resolution, so it resolved, but how to retry Ansible's Playbook?

Unable to restart service origin-master-api

Hi,

installation script failed with below error, any clue on this?

FAILED! => {"changed": false, "msg": "Unable to restart service origin-master-api: Job for origin-master-api.service failed because the control process exited with error code. See "systemctl status origin-master-api.service" and "journalctl -xe" for details.\n"}

err during running the play book

An exception occurred during task execution. To see the full traceback, use -vvv. The error was: RepresenterError: cannot represent an object: htpasswd_auth
fatal: [console.techdope.io]: FAILED! => {"failed": true, "msg": "Unexpected failure during module execution.", "stdout": ""}

What is the process to install master and node on dedicate VMs

First of all thank you for your contribution, I am able to install when I used a single VM for all the services.

But this script is not working when I used dedicated VMs for master and node. Is there any addition changes required or this project doesn't support a multi node cluster ?

Thanks,
Mahesh.

error during the installation process, Service Catalog Install

After
# ansible-playbook -i ./installcentos/inventory.erb ./openshift-ansible/playbooks/byo/config.yml
..................

TASK [ansible_service_broker : create ansible-service-broker service] *******************************************************************************
fatal: [console.phoenixit.ru]: FAILED! => {"changed": false, "failed": true, "msg": {"cmd": "/usr/local/bin/oc replace -f /tmp/asb-ygaDEN -n openshi}
        to retry, use: --limit @/root/openshift-ansible/playbooks/byo/config.retry

PLAY RECAP ******************************************************************************************************************************************
console.phoenixit.ru       : ok=645  changed=89   unreachable=0    failed=1
localhost                  : ok=13   changed=0    unreachable=0    failed=0


INSTALLER STATUS ************************************************************************************************************************************
Initialization             : Complete (0:00:40)
Health Check               : Complete (0:00:03)
etcd Install               : Complete (0:00:39)
Master Install             : Complete (0:03:01)
Master Additional Install  : Complete (0:00:22)
Node Install               : Complete (0:01:39)
Hosted Install             : Complete (0:01:52)
Service Catalog Install    : In Progress (0:00:46)
        This phase can be restarted by running: playbooks/openshift-service-catalog/config.yml

scale up

Hi Grant,

This is not really an issue. I am trying to add a node to my installation using another vm on my laptop.
Is it possible to use the scaleup.yml playbook to add a node to this installation?
It looks like the scaleup.yaml tries to install 3.7.0 (instead of 3.7.1) which fails.

Thanks,

Mohammad

Openshift Task to disable legacy master service fails

Hello,
I have a CentOS minimal install VM running on Proxmox VE in which I am running the script provided by G. Shipley.
I am following G. Shipley's youtube video and also instructions given here.
I run into following error:

Task [Openshift_master: Disable the legacy maszter service if it exists]
fatal: [192.168.10.216]: FAILED! => {"changed": false, "msg": "Could not find the requested service origin-master: host"}
Although this error doesn't seem to affect any implementation or later working of openshift but I am curious to know why this appears. Even in my other implementions (1 master VM and 1 node VM) I have same error and several dependency failures.

Also, I have a doubt. According to the Openshift documentation, if we are using "containerized=true", then "docker run" command should be used to implement each service as a separate docker container. But this and many other documents simply follow ansible-playbook command which is suitable for RPM installation.
Any help is appreciated.
Thank you.

missing origin-clients & vol.yaml

Little feedback from my experience using you script.

  1. During the creation of persistent volume, the script need the file vol.yaml as template to create 200 PV.
    The script do no copy vol.yaml from git to you local host , so you have to copy it manually.
  2. Creating PV and login to openshift require the command "oc" from missing package origin-clients.
    origin-clients is not a dependency or installed by the script. Can be fix by installing this one after ansible playbooks. Ansible playbooks take care to create the repository "CentOS OpenShift Origin" where the package origin-clients is available.
  3. Run twice the script is disruptive for docker if you use $DISK.
    If you want to use and additional disk for docker, the script take care to wipefs and delete /var/lib/docker folder. If you miss something and run the script a second time, the script will break the storage and docker will be not able to start. You end up with a busy storage unable to be deleted and docker not able to start.

Otherwise, this is a very good script and I am an happy user of OpenShift Origine, thanks to you !

oc: command not found & unable to connect to a server to handle "persistentvolumes"

I have 2 errors:

  1. oc: command not found
    I had to copy the oc binary from https://github.com/openshift/origin/releases manually otherwise your script always ended up with oc: command not found.
    Does your script install the oc tools? Your video did not seem to require the oc binary before running the script.
  2. Unable to connect to a server to handle "persistentvolumes"
    Once I copied oc manually in my PATH, I re-ran the script and I got this error at the end
INSTALLER STATUS ****************************************************************************************************************************************************************************************************************************
Initialization             : Complete (0:00:18)
Health Check               : Complete (0:00:01)
etcd Install               : Complete (0:00:31)
Master Install             : Complete (0:02:29)
Master Additional Install  : Complete (0:00:26)
Node Install               : Complete (0:00:47)
Hosted Install             : Complete (0:00:47)
Web Console Install        : Complete (0:00:38)
Service Catalog Install    : Complete (0:00:49)

Adding password for user pierrick
error: Missing or incomplete configuration info.  Please login or point to an existing, complete config file:

  1. Via the command-line flag --config
  2. Via the KUBECONFIG environment variable
  3. In your home directory as ~/.kube/config

To view or setup config directly use the 'config' command.
See 'oc adm policy add-cluster-role-to-user -h' for help and examples.
error: unable to connect to a server to handle "persistentvolumes": Missing or incomplete configuration info.  Please login or point to an existing, complete config file:

  1. Via the command-line flag --config
  2. Via the KUBECONFIG environment variable
  3. In your home directory as ~/.kube/config

To view or setup config directly use the 'config' command.
created volume 1
error: unable to connect to a server to handle "persistentvolumes": Missing or incomplete configuration info.  Please login or point to an existing, complete config file:

  1. Via the command-line flag --config
  2. Via the KUBECONFIG environment variable
  3. In your home directory as ~/.kube/config

To view or setup config directly use the 'config' command.

unable to read config file: open /etc/origin/node/resolv.conf: no such file or directory

I am unable to install openshift all in one getting error while restarting node during openshift-node task /etc/origin/node/resolv.conf: no such file or directory

Checked /etc/origin/node/ directory
it contains following file

total 44
drwx------. 2 root root 4096 Jul 11 19:13 .
drwx------. 10 root root 4096 Jul 11 19:13 ..
-rw-r--r--. 1 root root 1070 Jul 11 19:12 ca.crt
-rw-------. 1 root root 1313 Jul 11 19:13 node-config.yaml
-rw-r--r--. 1 root root 63 Jul 11 19:13 node-dnsmasq.conf
-rw-r--r--. 1 root root 2258 Jul 11 19:12 server.crt
-rw-------. 1 root root 1675 Jul 11 19:12 server.key
-rw-r--r--. 1 root root 1139 Jul 11 19:12 system:node:xxxxxxxx.crt
-rw-------. 1 root root 1679 Jul 11 19:12 system:node:xxxxxx.key
-rw-------. 1 root root 5866 Jul 11 19:12 system:node:xxxxxxx.kubeconfig

cat node/node-dnsmasq.conf
server=/in-addr.arpa/127.0.0.1
server=/cluster.local/127.0.0.1

any idea why this is happening

Persistent volumes?

It doesn't look like the scripts provision persistent volumes for PVCs. Is this the case?

Error at Task - Evaluate Groups - Fail if no etcd hosts group is defined

Hello,

I am receiving the following error. How do i define an etcd group?

TASK [Evaluate groups - Fail if no etcd hosts group is defined] ***********************************************************************************************************************
fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "msg": "Running etcd as an embedded service is no longer supported. If this is a new install please define an 'etcd' group with either one or three hosts. These hosts may be the same hosts as your masters. If this is an upgrade you may set openshift_master_unsupported_embedded_etcd=true until a migration playbook becomes available.\n"}
to retry, use: --limit @/root/openshift-ansible/playbooks/byo/config.retry

PLAY RECAP ****************************************************************************************************************************************************************************
localhost : ok=1 changed=0 unreachable=0 failed=1

Failure summary:

  1. Hosts: localhost
    Play: Populate config host groups
    Task: Evaluate groups - Fail if no etcd hosts group is defined
    Message: Running etcd as an embedded service is no longer supported. If this is a new install please define an 'etcd' group with either one or three hosts. These hosts may be the same hosts as your masters. If this is an upgrade you may set openshift_master_unsupported_embedded_etcd=true until a migration playbook becomes available.

ElasticSearch logging crash

[root@openshift01 installcentos]# oc project logging
Now using project "logging" on server "https://openshift01:8443".
[root@openshift01 installcentos]# oc get pod
NAME                                       READY     STATUS             RESTARTS   AGE
logging-curator-1-2cqkr                    0/1       CrashLoopBackOff   5          25m
logging-es-data-master-db07udx7-1-deploy   0/1       Error              0          24m
logging-fluentd-nd6nv                      1/1       Running            0          25m
logging-kibana-1-gjfvj                     2/2       Running            0          26m
[root@openshift01 installcentos]# oc logs logging-es-data-master-db07udx7-1-deploy

--> Scaling logging-es-data-master-db07udx7-1 to 1
error: update acceptor rejected logging-es-data-master-db07udx7-1: watch closed before Until timeout
--> Error listing events for replication controller logging-es-data-master-db07udx7-1: Get https://172.30.0.1:443/api/v1/namespaces/logging/events?fieldSelector=involvedObject.kind%3DReplicationController%2CinvolvedObject.name%3Dlogging-es-data-master-db07udx7-1%2CinvolvedObject.namespace%3Dlogging%2CinvolvedObject.uid%3D43dc018d-2dbe-11e8-9588-0800271bdf18: dial tcp 172.30.0.1:443: getsockopt: connection refused

missing webconsole service

Hey guys I'm new to DevOps, please help me to point out what is potentially wrong here.

Thanks so much!

Environment
OS: Centos 7.4
RAM: 16 GB
CPU: 4
HD: 200 GB

Repo Steps:

  1. Setup variables DOMAIN USERNAME PASSWORD using openshift 3.9
  2. Change /etc/sysconfig/selinux SELINUX=enforcing
  3. Ran install-openshift.sh
  4. Visit webconsole i got following error:

missing service (service "webconsole" not found)
missing route (service "webconsole" not found)

Expect
Webconsole to show up when vising https://console.${DOMAIN}.com:8443

Extra Information
when i run docker ps i can see following containers running:

  • origin-master-controllers
  • origin-master-api
  • etcd_container

simplephp is build OK, but deploy is failed.

Log messages
--> Scaling simplephp-1 to 1
error: update acceptor rejected simplephp-1: pods for rc 'grantproject/simplephp-1' took longer than 600 seconds to become available

./tmp/createPV.sh volume names are bugged

Here is the output I am getting:

persistentvolume "vol1" created
created volume 1
persistentvolume "vol21" created
created volume 2
persistentvolume "vol321" created
created volume 3
persistentvolume "vol4321" created
created volume 4
persistentvolume "vol54321" created
created volume 5
persistentvolume "vol654321" created
created volume 6
persistentvolume "vol7654321" created
created volume 7
persistentvolume "vol87654321" created
created volume 8
persistentvolume "vol987654321" created
created volume 9
persistentvolume "vol10987654321" created
created volume 10

'oc get pvc' outputs:

NAME             CAPACITY   ACCESS MODES   RECLAIM POLICY   STATUS      CLAIM                         STORAGECLASS   REASON    AGE
vol1             500Gi      RWO,RWX        Recycle          Bound       gitlab/gitlab-ce-redis-data                            2m
vol10987654321   500Gi      RWO,RWX        Recycle          Available                                                          1m
vol21            500Gi      RWO,RWX        Recycle          Bound       gitlab/gitlab-ce-etc                                   2m
vol321           500Gi      RWO,RWX        Recycle          Bound       gitlab/gitlab-ce-data                                  2m
vol4321          500Gi      RWO,RWX        Recycle          Bound       gitlab/gitlab-ce-postgresql                            2m
vol54321         500Gi      RWO,RWX        Recycle          Available                                                          1m
vol654321        500Gi      RWO,RWX        Recycle          Available                                                          1m
vol7654321       500Gi      RWO,RWX        Recycle          Available                                                          1m
vol87654321      500Gi      RWO,RWX        Recycle          Available                                                          1m
vol987654321     500Gi      RWO,RWX        Recycle          Available                                                          1m

RPM install fails due to wrong format of VERSION variable

Current versioning variable (VERSION=v3.9.0) fails when doing a RPM install (containerized=False).

Specific line where it fails is:
https://github.com/openshift/openshift-ansible/blob/master/roles/openshift_cli/tasks/main.yml#L3

as that would result in following (non-existent) package name to be installed "origin-clientsv3.9.0" while it should be "origin-clients-3.9.0"

To solve this, the correct way to set version in inventory file would be:
openshift_release=v3.9.0
openshift_pkg_version=-3.9.0

install fails during "RETRYING: Verify that TSB is running"

OS: Linux centos 3.10.0-693.17.1.el7.x86_64 #1 SMP Thu Jan 25 20:13:58 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux

Exports which are required by the installation step:

PASSWORD=rob
USERNAME=rob
DISK=/dev/sda
DOMAIN=192.168.0.135.nip.io

After starting following script:
curl https://raw.githubusercontent.com/gshipley/installcentos/master/install-openshift.sh | /bin/bash
which calls
ansible-playbook -i inventory.ini openshift-ansible/playbooks/byo/config.yml

The ansible playbook failed with following error:

...
TASK [template_service_broker : Verify that TSB is running] ********************
FAILED - RETRYING: Verify that TSB is running (120 retries left).
FAILED - RETRYING: Verify that TSB is running (119 retries left).
FAILED - RETRYING: Verify that TSB is running (118 retries left).
...
FAILED - RETRYING: Verify that TSB is running (2 retries left).
FAILED - RETRYING: Verify that TSB is running (1 retries left).
fatal: [192.168.0.135]: FAILED! => {"attempts": 120, "changed": false, "cmd": ["curl", "-k", "https://apiserver.openshift-template-service-broker.svc/healthz"], "delta": "0:00:01.066711", "end": "2018-03-07 13:44:18.024722", "msg": "non-zero return code", "rc": 7, "start": "2018-03-07 13:44:16.958011", "stderr": "  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current\n                                 Dload  Upload   Total   Spent    Left  Speed\n\r  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0curl: (7) Failed connect to apiserver.openshift-template-service-broker.svc:443; Connection refused", "stderr_lines": ["  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current", "                                 Dload  Upload   Total   Spent    Left  Speed", "", "  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0curl: (7) Failed connect to apiserver.openshift-template-service-broker.svc:443; Connection refused"], "stdout": "", "stdout_lines": []}
        to retry, use: --limit @/root/openshift-ansible/playbooks/byo/config.retry

PLAY RECAP *********************************************************************
192.168.0.135              : ok=981  changed=351  unreachable=0    failed=1
localhost                  : ok=13   changed=0    unreachable=0    failed=0


INSTALLER STATUS ***************************************************************
Initialization             : Complete
Health Check               : Complete
etcd Install               : Complete
Master Install             : Complete
Master Additional Install  : Complete
Node Install               : Complete
Hosted Install             : Complete
Metrics Install            : Complete
Logging Install            : Complete
Service Catalog Install    : In Progress
        This phase can be restarted by running: playbooks/byo/openshift-cluster/service-catalog.yml



Failure summary:


  1. Hosts:    192.168.0.135
     Play:     Service Catalog
     Task:     Verify that TSB is running
     Message:  non-zero return code
Adding password for user rob
cluster role "cluster-admin" added: "rob"
******
* Your conosle is https://console.192.168.0.135.nip.io:8443
* Your username is rob
* Your password is rob
*
* Login using:
*
$ oc login -u rob -p rob https://console.192.168.0.135.nip.io:8443/
******
error: dial tcp 192.168.0.135:8443: getsockopt: connection refused - verify you have provided the correct host and port and that the server is currently running.
[root@centos ~]#

Ansible in service_catalog is always fail

I can't success run the script. It always failed at the service_catalog.
It will check the "https://apiserver.kube-service-catalog.svc/healthz" but I find there is no kubenetes POD or service for this address.
I use pure Centos 7.2 minimal version.

Below is my output of ansible message and kubectl get all.

TASK [openshift_service_catalog : wait for api server to be ready] ***********************************************************************************************************
FAILED - RETRYING: wait for api server to be ready (60 retries left).
FAILED - RETRYING: wait for api server to be ready (59 retries left).
FAILED - RETRYING: wait for api server to be ready (58 retries left).
fatal: [192.168.200.40]: FAILED! => {"attempts": 60, "changed": false, "content": "", "msg": "Status code was -1 and not [200]: Request failed: <urlopen error [Errno 111] Con
nection refused>", "redirected": false, "status": -1, "url": "https://apiserver.kube-service-catalog.svc/healthz"}
to retry, use: --limit @/root/openshift-ansible/playbooks/byo/config.retry

[root@openshift ~]# kubectl get all
NAME DOCKER REPO TAGS UPDATED
imagestreams/registry-console docker-registry.default.svc:5000/default/registry-console latest 3 hours ago

NAME REVISION DESIRED CURRENT TRIGGERED BY
deploymentconfigs/docker-registry 1 1 1 config
deploymentconfigs/registry-console 1 1 0 config
deploymentconfigs/router 5 1 1 config

NAME HOST/PORT PATH SERVICES PORT TERMINATION WILDCARD
routes/docker-registry docker-registry-default.apps.macauley.com docker-registry passthrough None
routes/registry-console registry-console-default.apps.macauley.com registry-console passthrough None

NAME READY STATUS RESTARTS AGE
po/docker-registry-1-b996v 0/1 Error 0 2h
po/router-5-mqzs2 1/1 Running 0 15m

NAME DESIRED CURRENT READY AGE
rc/docker-registry-1 1 1 0 2h
rc/registry-console-1 0 0 0 2h
rc/router-1 0 0 0 2h
rc/router-2 0 0 0 2h
rc/router-3 0 0 0 1h
rc/router-4 0 0 0 33m
rc/router-5 1 1 1 15m

NAME CLUSTER-IP EXTERNAL-IP PORT(S) AGE
svc/docker-registry 172.30.82.155 5000/TCP 2h
svc/kubernetes 172.30.0.1 443/TCP,53/UDP,53/TCP 3h
svc/registry-console 172.30.12.197 9000/TCP 2h
svc/router 172.30.224.129 80/TCP,443/TCP,1936/TCP 2h

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.