- Requirements
- Caveats
- Known Issues
- Configuring the host inventory
- Creating the default variables for the hosts and host groups
- Running the ansible playbooks
- Post-ansible steps
- Overriding detected ip addresses and hostnames
-
ansible
- Tested using ansible-1.8.4-1.fc20.noarch, but should work with version 1.8+
- There is currently a known issue with ansible-1.9.0, you can downgrade to 1.8.4 on Fedora by installing one of the builds from Koji: http://koji.fedoraproject.org/koji/packageinfo?packageID=13842
- Available in Fedora channels
- Available for EL with EPEL and Optional channel
-
One or more RHEL 7.1 VMs
-
Either ssh key based auth for the root user or ssh key based auth for a user with sudo access (no password)
-
A checkout of openshift-ansible from https://github.com/openshift/openshift-ansible/
git clone https://github.com/openshift/openshift-ansible.git cd openshift-ansible
This ansible repo is currently under heavy revision for providing OSE support; the following items are highly likely to change before the OSE support is merged into the upstream repo:
- the current git branch for testing
- how the inventory file should be configured
- variables that need to be set
- bootstrapping steps
- other configuration steps
- Host subscriptions are not configurable yet, the hosts need to be pre-registered with subscription-manager or have the RHEL base repo pre-configured. If using subscription-manager the following commands will disable all but the rhel-7-server rhel-7-server-extras and rhel-server7-ose-beta repos:
subscription-manager repos --disable="*"
subscription-manager repos \
--enable="rhel-7-server-rpms" \
--enable="rhel-7-server-extras-rpms" \
--enable="rhel-7-server-ose-3.0-rpms"
- Configuration of router is not automated yet
- Configuration of docker-registry is not automated yet
Example inventory file for configuring one master and two nodes for the test environment. This can be configured in the default inventory file (/etc/ansible/hosts), or using a custom file and passing the --inventory option to ansible-playbook.
/etc/ansible/hosts:
# This is an example of a bring your own (byo) host inventory
# Create an OSEv3 group that contains the masters and nodes groups
[OSEv3:children]
masters
nodes
# Set variables common for all OSEv3 hosts
[OSEv3:vars]
# SSH user, this user should allow ssh based auth without requiring a password
ansible_ssh_user=root
# If ansible_ssh_user is not root, ansible_sudo must be set to true
#ansible_sudo=true
# To deploy origin, change deployment_type to origin
deployment_type=enterprise
# Pre-release additional repo
openshift_additional_repos=[{'id': 'ose-devel', 'name': 'ose-devel',
'baseurl':
'http://buildvm/puddle/build/OpenShiftEnterprise/3.0/latest/RH7-RHOSE-3.0/$basearch/os',
'enabled': 1, 'gpgcheck': 0}]
# Origin copr repo
#openshift_additional_repos=[{'id': 'openshift-origin-copr', 'name':
'OpenShift Origin COPR', 'baseurl':
'https://copr-be.cloud.fedoraproject.org/results/maxamillion/origin-next/epel-7-$basearch/',
'enabled': 1, 'gpgcheck': 1, gpgkey:
'https://copr-be.cloud.fedoraproject.org/results/maxamillion/origin-next/pubkey.gpg'}]
# host group for masters
[masters]
ose3-master.example.com
# host group for nodes
[nodes]
ose3-master.example.com
ose3-node[1:2].example.com
The hostnames above should resolve both from the hosts themselves and the host where ansible is running (if different).
A more complete example inventory file (hosts.ose.example) is available under the /inventory/byo
directory.
From the openshift-ansible checkout run:
ansible-playbook playbooks/byo/config.yml
Note: this assumes that the host inventory is /etc/ansible/hosts, if using a different inventory file use the -i option for ansible-playbook.
You should now be ready to follow the What's Next? section of the advanced installation guide to deploy your router, registry, and other components.
Some deployments will require that the user override the detected hostnames and ip addresses for the hosts. To see what the default values will be you can run the openshift_facts playbook:
ansible-playbook playbooks/byo/openshift_facts.yml
The output will be similar to:
ok: [10.3.9.45] => {
"result": {
"ansible_facts": {
"openshift": {
"common": {
"hostname": "jdetiber-osev3-ansible-005dcfa6-27c6-463d-9b95-ef059579befd.os1.phx2.redhat.com",
"ip": "172.16.4.79",
"public_hostname": "jdetiber-osev3-ansible-005dcfa6-27c6-463d-9b95-ef059579befd.os1.phx2.redhat.com",
"public_ip": "10.3.9.45",
"use_openshift_sdn": true
},
"provider": {
... <snip> ...
}
}
},
"changed": false,
"invocation": {
"module_args": "",
"module_name": "openshift_facts"
}
}
}
ok: [10.3.9.42] => {
"result": {
"ansible_facts": {
"openshift": {
"common": {
"hostname": "jdetiber-osev3-ansible-c6ae8cdc-ba0b-4a81-bb37-14549893f9d3.os1.phx2.redhat.com",
"ip": "172.16.4.75",
"public_hostname": "jdetiber-osev3-ansible-c6ae8cdc-ba0b-4a81-bb37-14549893f9d3.os1.phx2.redhat.com",
"public_ip": "10.3.9.42",
"use_openshift_sdn": true
},
"provider": {
...<snip>...
}
}
},
"changed": false,
"invocation": {
"module_args": "",
"module_name": "openshift_facts"
}
}
}
ok: [10.3.9.36] => {
"result": {
"ansible_facts": {
"openshift": {
"common": {
"hostname": "jdetiber-osev3-ansible-bc39a3d3-cdd7-42fe-9c12-9fac9b0ec320.os1.phx2.redhat.com",
"ip": "172.16.4.73",
"public_hostname": "jdetiber-osev3-ansible-bc39a3d3-cdd7-42fe-9c12-9fac9b0ec320.os1.phx2.redhat.com",
"public_ip": "10.3.9.36",
"use_openshift_sdn": true
},
"provider": {
...<snip>...
}
}
},
"changed": false,
"invocation": {
"module_args": "",
"module_name": "openshift_facts"
}
}
}
Now, we want to verify the detected common settings to verify that they are what we expect them to be (if not, we can override them).
- hostname
- Should resolve to the internal ip from the instances themselves.
- openshift_hostname will override.
- ip
- Should be the internal ip of the instance.
- openshift_ip will override.
- public hostname
- Should resolve to the external ip from hosts outside of the cloud
- provider openshift_public_hostname will override.
- public_ip
- Should be the externally accessible ip associated with the instance
- openshift_public_ip will override
- use_openshift_sdn
- Should be true unless the cloud is GCE.
- openshift_use_openshift_sdn overrides
To override the the defaults, you can set the variables in your inventory:
...snip...
[masters]
ose3-master.example.com openshift_ip=1.1.1.1 openshift_hostname=ose3-master.example.com openshift_public_ip=2.2.2.2 openshift_public_hostname=ose3-master.public.example.com
...snip...