| Commit message (Collapse) | Author | Age | Files | Lines |
|
|
|
|
| |
* Added checks to make ci for yaml linting
* Modified y(a)ml files to pass lint checks
|
| |
|
| |
|
| |
|
| |
|
|\
| |
| | |
Fix metricsPublicURL only being set correctly on first master.
|
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| | |
Problem was caused by facts not being set for that master. To fix this
patch cleans up the calculation of metricsPublicURL in general. Because
this value is used in openshift_master to template into the master
config file, we now define these facts more clearly in
openshift_master_facts, and add a dependency on this to
openshift_metrics.
The calculation of default sub-domain is also changed to remove it from
system facts (as neither of these are facts about the system) and
instead use plain variables.
|
|\ \
| | |
| | | |
Drop 3.2 upgrade playbooks.
|
| | | |
|
|\ \ \
| | | |
| | | | |
Silence warnings when using some commands directly
|
| |/ / |
|
|/ / |
|
|\ \
| | |
| | | |
etcd_upgrade: Simplify package installation
|
| | | |
|
|\ \ \
| |/ /
|/| | |
Scheduler upgrades
|
| | | |
|
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | | |
- do not upgrade predicates if openshift_master_scheduler_predicates is
defined
- do not upgrade priorities if openshift_master_scheduler_priorities is
defined
- do not upgrade predicates/priorities unless they match known previous
default configs
- output WARNING to user if predictes/priorities are not updated during
install
|
| | | |
|
| | | |
|
|/ / |
|
| | |
|
|\ \
| | |
| | | |
upgrade_control_plane.yml: systemd_units.yaml needs the master facts
|
| | | |
|
|/ /
| |
| |
| |
| |
| |
| | |
inventory_hostname
When using a dynamic inventory inventory_hostname isn't guaranteed to be usable. We should use openshift.common.hostname which
already copes with this
|
|/
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
In 3.3 one of our services lays down a systemd drop-in for configuring
Docker networking to use lbr0. In 3.4, this has been changed but the
file must be cleaned up manually by us.
However, after removing the file docker requires a restart. This had big
implications particularly in containerized environments where upgrade is
a very fragile series of upgrading and service restarts.
To avoid double docker restarts, and thus double service restarts in
containerized environments, this change does the following:
- Skip restart during docker upgrade, if it is required. We will restart
on our own later.
- Skip containerized service restarts when we upgrade the services
themselves.
- Clean shutdown of all containerized services.
- Restart Docker. (always, previously this only happened if it needed an
upgrade)
- Ensure all containerized services are restarted.
- Restart rpm node services. (always)
- Mark node schedulable again.
At the end of this process, docker0 should be back on the system.
|
|\
| |
| | |
Update scheduler defaults
|
| | |
|
| | |
|
| | |
|
|/
|
|
|
| |
We require ansible >= 2.2.0 now. Updating version checking playbook to
reflect this change.
|
|
|
|
| |
Fixes Bug 1395945
|
|\
| |
| | |
Fix invalid embedded etcd fact in etcd upgrade playbook.
|
| |
| |
| |
| |
| |
| |
| |
| |
| | |
Fixes: https://bugzilla.redhat.com/show_bug.cgi?id=1398549
Was getting a different failure here complaining that openshift was not
in the facts, as we had not loaded facts for the first master during
playbook run. However this check was used recently in
upgrade_control_plane and should be more reliable.
|
|\ \
| |/
|/| |
Merge admission plugin configs
|
| |
| |
| |
| |
| |
| |
| |
| | |
Move the values in kube_admission_plugin_config up one level per
the new format from 1.3:
"The kubernetesMasterConfig.admissionConfig.pluginConfig should be moved
and merged into admissionConfig.pluginConfig."
|
| |
| |
| |
| | |
containerized.
|
|\ \
| |/
|/| |
Refactor to use Ansible package module
|
| |
| |
| |
| |
| | |
The Ansible package module will call the correct package manager for the
underlying OS.
|
|\ \
| | |
| | | |
Allow ansible to continue when a node is unaccessible or fails.
|
| | | |
|
| |/ |
|
|/ |
|
|\
| |
| | |
Check for bad versions of yum and subscription-manager.
|
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| | |
Use of yum and repoquery will output the given additional warning when
using newer versions of subscription-manager, with older versions of
yum. (RHEL 7.1) Installing/upgrading newer docker can pull this
subscription-manager in resulting in problems with older versions of
ansible and it's yum module, as well as any use of repoquery/yum
commands in our playbooks.
This change explicitly checks for the problem by using repoquery and
fails early if found. This is run early in both config and upgrade.
|
|\ \
| | |
| | | |
etcd upgrade playbooks
|
| | | |
|
| | | |
|
| | | |
|
| | |
| | |
| | |
| | |
| | | |
On Fedora we just blindly upgrade to the latest.
On RHEL we do stepwise upgrades 2.0,2.1,2.2,2.3,3.0
|
| |/ |
|