| Commit message (Collapse) | Author | Age | Files | Lines |
| |
|
|\
| |
| | |
Fix version handling in 3.8/3.9 control plane upgrades
|
| | |
|
| | |
|
|\ \
| |/
|/|
| |
| |
| |
| |
| |
| |
| |
| | |
Automatic merge from submit-queue.
Upgrade playbook improvements
Two changes in this PR:
* Avoid duplication in upgrade.yml - it now would simply run upgrade_control_plane and upgrade_nodes playbooks
* Restart controllers instead of start/stop if restart mode is 'services'
* For 'system' restart mode controllers are disabled before master upgrade and re-enabled after upgrades are finished
|
| |
| |
| |
| |
| |
| | |
mode is system
This removes a start/stop after control plane upgrade
|
| |
| |
| |
| | |
upgrade
|
| | |
|
|\ \
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | | |
Automatic merge from submit-queue.
upgrades: use openshift_version as a regexp when checking openshift.common.version
This would fix errors like 'Master running 3.9.0 must be upgraded to
3.9* before node upgrade can be run' during the upgrade
Signed-off-by: Vadim Rutkovsky <vrutkovs@redhat.com>
|
| |/
| |
| |
| |
| |
| |
| |
| |
| | |
openshift.common.version
This would fix errors like 'Master running 3.9.0 must be upgraded to
3.9* before node upgrade can be run' during the upgrade
Signed-off-by: Vadim Rutkovsky <vrutkovs@redhat.com>
|
|\ \
| | |
| | | |
Remove master_ha bool checks
|
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | | |
Most of these checks are no longer applicable to
the tasks on which they are applied.
This commit removes incorrect ha checks to ensure
services are restarted at appropriate times.
Fixes: https://bugzilla.redhat.com/show_bug.cgi?id=1500897
|
| |/
|/|
| |
| |
| |
| |
| |
| |
| |
| | |
Version 3.8 is an intermediate step when upgrading to 3.9 and
openshift_pkg_version may be set in the inventory for 3.9. This will
ensure there is not a conflict between the intermediate step and the
final version requested.
Bug 1540108
https://bugzilla.redhat.com/show_bug.cgi?id=1540108
|
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| | |
Due to complexities upgrading two versions
at the same time, openshift_image_tag was being
set incorrectly during control_plane upgrades.
This commit ensures that openshift_image_tag
is set correctly during this process.
Fixes: https://bugzilla.redhat.com/show_bug.cgi?id=1536839
|
|\ \
| | |
| | | |
Remove old assetConfig from master-config.yaml
|
| |/
| |
| |
| |
| |
| |
| |
| | |
The `assetConfig` is no longer used now that the console is split into
its own pod.
- Remove `assetConfig` on upgrades to 3.9
- Stop writing logging and metrics URLs to assetConfig for 3.9
|
|\ \
| |/
|/|
| |
| | |
Automatic merge from submit-queue.
Bug 1536217: Need to validate etcd serving certs before 3.9 upgrade
|
| |
| |
| |
| | |
etcd serving cert SAN.
|
|\ \
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | | |
Automatic merge from submit-queue.
Label masters with node-role.kubernetes.io/master
This is controlled by `openshift_master_node_label` var in openshift_manage_node
Related to https://bugzilla.redhat.com/show_bug.cgi?id=1535673
|
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | | |
labels
and scheduling status during upgrades
Signed-off-by: Vadim Rutkovsky <vrutkovs@redhat.com>
|
|\ \ \
| | | |
| | | |
| | | |
| | | | |
mgugino-upstream-stage/protect-upgrade-master-version
Protect master installed version during node upgrades
|
| |/ /
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | | |
Master version is unprotected during all upgrades.
During node-only upgrades, master version should
remain protected to ensure correct image tag is
set for node upgrades.
Fixes: https://bugzilla.redhat.com/show_bug.cgi?id=1536839
|
|\ \ \
| |/ /
|/| | |
Allow openshift-ansible image to deploy to GCP
|
| | | |
|
|\ \ \
| |/ /
|/| |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | | |
Automatic merge from submit-queue.
Ensure that node names are lowerecased before matching
The name field is always lowercase whereas the nodename may not have
been in the past. Ensure that we lowercase it before looking for
matching nodes.
Fixes https://bugzilla.redhat.com/show_bug.cgi?id=1396350
|
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | | |
The name field is always lowercase whereas the nodename may not have
been in the past. Ensure that we lowercase it before looking for
matching nodes.
Fixes https://bugzilla.redhat.com/show_bug.cgi?id=1396350
|
| |/
|/|
| |
| |
| |
| |
| |
| |
| |
| |
| |
| | |
Currently, openshift_facts requires pyyaml to be installed.
This package is installed via init/base_packages.yml, which
is currently called after init/facts.yml. This results
in a situation where installs will fail due to missing
python dependency.
This commit splits init/facts.yml into two, and
allows base_packages.yml to be run before the
openshift_facts.py plugin is executed.
|
|/
|
|
|
|
|
|
| |
Currently, users have no way to run preqrequisites.yml
on just newly added nodes during scaleup.
This commit ensures only the new nodes are changed during
scaleup as well as ensure prerequisites are run.
|
| |
|
| |
|
|
|
|
|
|
|
|
| |
There are some obsoletes in 3.9 packages which lead yum to upgrade
directly from 3.7 to 3.9 when you ask for 3.8 packages while 3.9 repos
are enabled. Since we'd like to allow people to run one playbook to
upgrade from 3.7 to 3.8 to 3.9 we need to exclude those packages when
upgrading to 3.8
|
| |
|
| |
|
| |
|
|\
| |
| |
| |
| |
| |
| |
| |
| |
| | |
Automatic merge from submit-queue.
Add the ability to specify a timeout for node drain operations
A timeout to wait for nodes to drain pods can be specified to ensure that the upgrade continues even if nodes fail to drain pods in the allowed time. The default value of 0 will wait indefinitely allowing the admin to investigate the root cause and ensuring that disruption budgets are respected. In practice the `oc adm drain` command will eventually error out, at least that's what we've seen in our large online clusters, when that happens a second attempt will be made to drain the nodes, if it fails again it will abort the upgrade for that node or for the entire cluster based on your defined `openshift_upgrade_nodes_max_fail_percentage`.
`openshift_upgrade_nodes_drain_timeout=0` is the default and will wait until all pods have been drained successfully
`openshift_upgrade_nodes_drain_timeout=600` would wait for 600s before moving on to the tasks which would forcefully stop pods such as stopping docker, node, and openvswitch.
|
| | |
|
|\ \
| | |
| | | |
Ensure that openshift_facts role is imported whenever we rely on
|
| | |
| | |
| | |
| | | |
openshift_client_binary
|
|\ \ \
| | | |
| | | |
| | | |
| | | | |
vrutkovs/3.9-upgrades-remove-openshift.common.service_type
3.9 upgrade: remove openshift.common.service_type
|
| | |/
| |/|
| | |
| | | |
See eb6b20fc9183cc2aae424c72efd1191b99110a93
|
|\ \ \
| | | |
| | | | |
Remove become statements
|
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | | |
After remove become:no statements on local_action tasks,
we need to ensure that the proper file permssions are
applied to local temp directories.
This reason for this is that the 'fetch' module
does not use 'become' for the localhost, just the remote
host.
Additionally, users may not wish for the localhost to
become during a fetch. local_action will execute with
whatever permissions are specified in inventory or via
cli.
|
| | |/
| |/|
| | |
| | |
| | | |
This commit removes become:no statements that break
the installer in various ways.
|
|/ /
| |
| |
| |
| | |
This commit limits common init code to exclude
oo_nodes_to_config during upgrade_control_plane runs.
|
|\ \
| |/
|/|
| |
| |
| |
| |
| | |
Automatic merge from submit-queue.
Remove last of openshift_node role meta-depends
Remove last non-taskless meta-depends from
openshift_node role.
|
| |
| |
| |
| |
| |
| |
| |
| | |
Remove last non-taskless meta-depends from
openshift_node role.
Remove variable 'openshift_node_upgrade_in_progress' as
it is no longer used.
|
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| | |
In Ansible 2.2, the include_role directive came into existence as
a Tech Preview. It is still a Tech Preview through Ansible 2.4
(and in current devel branch), but with a noteable change. The
default behavior switched from static: true to static: false
because that functionality moved to the newly introduced
import_role directive (in order to stay consistent with include*
being dynamic in nature and `import* being static in nature).
The dynamic include is considerably more memory intensive as it will
dynamically create a role import for every host in the inventory
list to be used. (Also worth noting, there is at the time of this
writing an object allocation inefficiency in the dynamic include
that can in certain situations amplify this effect considerably)
This change is meant to mitigate the pressure on memory for the
Ansible control host.
We need to evaluate where it makes sense to dynamically include roles
and revert back to dynamic inclusion if and where it makes sense to do
so.
|
| |
| |
| |
| |
| |
| |
| | |
Move openshift_deployment_type check into sanity_check
action plugin. Remove compatibility for deployment_type.
deployment_type has been deprecated for some time now.
|
|/
|
|
|
|
| |
Move more checks outside of init/main.yml for
speeding up upgrades and other operational plays that
need to run.
|
|
|
|
|
| |
We set these variables using facts in init, no need
to duplicate the logic all around the codebase.
|