| Commit message (Collapse) | Author | Age | Files | Lines |
|
|
|
| |
This issue was already addressed by #1355 fc2f776 but it has been accidentally reverted by 6d3e176.
|
| |
|
|
|
|
| |
which the checksum currently expects (#1384).
|
| |
|
| |
|
|\
| |
| | |
Fix issue when there are no infra nodes
|
| | |
|
|\ \
| | |
| | | |
Fix libvirt cluster creation
|
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | | |
Doing a `bin/cluster create libvirt …` fails with the following error:
```
TASK: [Warn user about bad openshift_hostname values] *************************
[lenaic-node-compute-cf3fb, lenaic-node-compute-f39d3, lenaic-node-infra-0407f, lenaic-master-fc5f9]
The hostname "lenaic-node-compute-cf3fb.example.com" for "lenaic-node-compute-cf3fb.example.com" doesn't resolve to an ip address owned by this host. Please set openshift_hostname variable to a hostname that when resolved on the host in question resolves to an IP address matching an interface on this host. This host will fail liveness checks for pods utilizing hostPorts, press CTRL-C to continue.:
```
because the `/etc/hosts` of the VM is containing:
```
[openshift@lenaic-node-compute-cf3fb ~]$ cat /etc/hosts
# Your system has configured 'manage_etc_hosts' as True.
# As a result, if you wish for changes to this file to persist
# then you will need to either
# a.) make changes to the master file in /etc/cloud/templates/hosts.redhat.tmpl
# b.) change or remove the value of 'manage_etc_hosts' in
# /etc/cloud/cloud.cfg or cloud-config from user-data
#
# The following lines are desirable for IPv4 capable hosts
127.0.0.1 lenaic-node-compute-cf3fb.example.com lenaic-node-compute-cf3fb
127.0.0.1 localhost.localdomain localhost
127.0.0.1 localhost4.localdomain4 localhost4
# The following lines are desirable for IPv6 capable hosts
::1 lenaic-node-compute-cf3fb.example.com lenaic-node-compute-cf3fb
::1 localhost.localdomain localhost
::1 localhost6.localdomain6 localhost6
```
Disabling `manage_etc_hosts` removes the lines for `lenaic-node-compute-cf3fb` from `/etc/hosts`
and makes than name resolved by the DNS to its real IP.
|
|/ / |
|
|\ \
| |/
|/| |
Add missing atomic- and openshift-enterprise
|
| |
| |
| |
| |
| |
| | |
Some checks related to *enterprise deployments were still only
looking for "enterprise" deployment_type. Update them to
cover also atomic-enterprise and openshift-enterprise deployment types.
|
|\ \
| | |
| | | |
Allow compression option to be set to empty for non compressed images
|
| |/
| |
| |
| | |
Support tgz and gzip compressed images
|
|\ \
| | |
| | | |
Additional overrides for cloud provider playbooks
|
| |/
| |
| |
| |
| | |
- sdn overrides
- allow overrides for use_flannel and use_fluentd
|
|\ \
| |/
|/| |
Refactor storage options
|
| | |
|
|/
|
|
|
| |
- set openshift_master_cluster_method=native for all cloud providers so
bin/cluster will build the ha masters correctly
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
- openshift_master role update
- infra_nodes was previously being set to num_infra, which is an integer
value when using the cloud providers, added a new variable osm_infra_nodes
that is expected to be a list of hosts
- if openshift_infra_nodes is not already set, create it from the nodes that
have the region=infra label.
- Cloud provider config playbook updates
- override openshift_router_selector for cloud providers to avoid using the
default of 'region=infra' when deployment_type is not 'online'
- Set openshift_infra_nodes to g_infra_host for cloud providers
|
|
|
|
|
|
| |
- Add g_infra_hosts (nodes with sub-type infra)
- Add g_compute_hosts (nodes with sub-type compute)
- Reduce duplication by re-using previously defined variables
|
|\
| |
| | |
Make bin/cluster able to spawn an OSE 3.1 cluster
|
| | |
|
|/ |
|
| |
|
|
|
|
|
|
|
|
|
| |
- Move debug_level into vars.yml and byo inventory
- change variables in cluster_hosts.yml to be g_* and update playbooks to use
those values directly instead of setting them indirectly
- added a new g_all_hosts entry in cluster_hosts to use in the update playbook
instead of unioning all host types within the playbook
- added a cluster_hosts.yml for the byo playbook
|
| |
|
| |
|
|\
| |
| | |
Removing env-host-type in preparation of env and environment changes.
|
| | |
|
|\ \
| | |
| | | |
Update for latest CentOS-7-x86_64-GenericCloud.
|
| | | |
|
| |/
| |
| |
| |
| |
| | |
- Use xz compressed image
- Update sha256 for new image
- Update docs to reflect new settings
|
|/ |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
The list playbook listed the IPs of the VMs without logging their role like:
TASK: [debug ] ************************************************************
ok: [10.64.109.37] => {
"msg": "public:10.64.109.37 private:192.168.165.5"
}
ok: [10.64.109.47] => {
"msg": "public:10.64.109.47 private:192.168.165.6"
}
ok: [10.64.109.36] => {
"msg": "public:10.64.109.36 private:192.168.165.4"
}
ok: [10.64.109.215] => {
"msg": "public:10.64.109.215 private:192.168.165.2"
}
The list playbook now prints the information in a more structured way with
a list of masters, a list of nodes and the subtype of the nodes like:
TASK: [debug ] ************************************************************
ok: [localhost] => {
"msg": {
"lenaicnewlist": {
"master": [
{
"name": "10.64.109.215",
"private IP": "192.168.165.2",
"public IP": "10.64.109.215",
"subtype": "default"
}
],
"node": [
{
"name": "10.64.109.47",
"private IP": "192.168.165.6",
"public IP": "10.64.109.47",
"subtype": "compute"
},
{
"name": "10.64.109.37",
"private IP": "192.168.165.5",
"public IP": "10.64.109.37",
"subtype": "compute"
},
{
"name": "10.64.109.36",
"private IP": "192.168.165.4",
"public IP": "10.64.109.36",
"subtype": "infra"
}
]
}
}
}
|
| |
|
| |
|
| |
|
|
|
| |
Use write_files to disable requiretty for the openshift user as suggested by @detiberm, fixes #773
|
|
|
| |
Set Defaults !requiretty so that ansible can run sudo without a terminal. Fixes #773
|
|
|
|
| |
It was timeouting on slower hardware.
|
|
|
|
|
|
|
|
|
|
|
|
| |
Using bootcmd in cloud-config lead to restarts prior to starting the
systemd-hostnamed, which was probable cause of the failure when DHCP
client was failing to send the hostname, and subsequently, the
ansible-opnshift was not able to identify the VM among the others when
checking DHCP leases. The failure looked like: following
10:17:31 failed: [localhost] => {"attempts": 60, "changed": true, "cmd": "virsh -c qemu:///system net-dhcp-leases openshift-ansible | egrep -c 'experiment-node-compute-453d0|experiment-node-compute-61e16'", "delta": "0:00:00.033061", "end": "2015-10-19 10:17:31.409434", "failed": true, "rc": 0, "start": "2015-10-19 10:17:31.376373", "warnings": []}
10:17:31 stdout: 1
10:17:31 msg: Task failed as maximum retries was encountered
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
The dnsmasq should not be resolving the example.com recursively, because
in case that we have /etc/NetworkManager/dnsmasq.d/libvirt_dnsmasq.conf:
server=/example.com/192.168.55.1
the dnsmasq will be asking itself, therefore a dns resolution loop is
created, which causes
Maximum number of concurrent DNS queries reached (max: 150)
and performance degradation of dns resolution on the whole hypervizor and
guests.
This patch will fix that in the domain.xml, which will cause adding
local=/example.com/
to the /var/lib/libvirt/dnsmasq/openshift-ansible.conf, effectively
fixing the problem.
|
| |
|
|\
| |
| | |
Set loglevel=2 as our default across the board
|
| | |
|
|/ |
|
| |
|
| |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
- Add support to bin/cluster for specifying etcd hosts
- defaults to 0, if no etcd hosts are selected, then configures embedded
etcd
- Updates for the byo inventory file for etcd and master as node by default
- Consolidation of cluster logic more centrally into common playbook
- Added etcd config support to playbooks
- Restructured byo playbooks to leverage the common openshift-cluster playbook
- Added support to common master playbook to generate and apply external etcd
client certs from the etcd ca
- start of refactor for better handling of master certs in a multi-master
environment.
- added the openshift_master_ca and openshift_master_certificates roles to
manage master certs instead of generating them in the openshift_master
role
- added etcd host groups to the cluster update playbooks
- aded better handling of host groups when they are either not present or are
empty.
- Update AWS readme
|