| Commit message (Collapse) | Author | Age | Files | Lines |
|\
| |
| | |
Merged by openshift-bot
|
| |
| |
| |
| | |
Signed-off-by: Jose A. Rivera <jarrpa@redhat.com>
|
| |
| |
| |
| | |
Signed-off-by: Jose A. Rivera <jarrpa@redhat.com>
|
| |
| |
| |
| | |
Signed-off-by: Jose A. Rivera <jarrpa@redhat.com>
|
| |
| |
| |
| | |
Signed-off-by: Jose A. Rivera <jarrpa@redhat.com>
|
| |
| |
| |
| | |
Signed-off-by: Jose A. Rivera <jarrpa@redhat.com>
|
| |
| |
| |
| | |
Signed-off-by: Jose A. Rivera <jarrpa@redhat.com>
|
| |
| |
| |
| | |
Signed-off-by: Jose A. Rivera <jarrpa@redhat.com>
|
|\ \
| | |
| | | |
Temporarilly only migrate jobs as we were before
|
| | | |
|
| | |
| | |
| | |
| | |
| | |
| | | |
Created by command:
/usr/bin/tito tag --accept-auto-changelog --use-version=3.6.117
|
|\ \ \
| |/ /
|/| | |
Merged by openshift-bot
|
| | | |
|
|\ \ \
| | | |
| | | | |
Run storage upgrade pre and post master upgrade
|
| | | | |
|
|\ \ \ \
| | | | |
| | | | | |
Merged by openshift-bot
|
| | | | | |
|
| | | | |
| | | | |
| | | | |
| | | | |
| | | | |
| | | | | |
Created by command:
/usr/bin/tito tag --accept-auto-changelog --use-version=3.6.116
|
|\ \ \ \ \
| | | | | |
| | | | | | |
Merged by openshift-bot
|
| | | | | |
| | | | | |
| | | | | |
| | | | | |
| | | | | | |
If we have no master config assume that we're a clean install.
If we're a clean install and we're 3.6 or greater use etcd v3 storage.
|
|\ \ \ \ \ \
| | | | | | |
| | | | | | | |
Merged by openshift-bot
|
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | | |
bug 1460564. Fixes [BZ #1460564](https://bugzilla.redhat.com/show_bug.cgi?id=1460564).
Unfortunately, the defaults for Elasticsearch prior to v5 allow more
than one "node" to access the same configured storage volume(s).
This change forces this value to 1 to ensure we don't have an ES pod
starting up accessing a volume while another ES pod is shutting down
when reploying. This can lead to "1" directories being created in
`/elasticsearch/persistent/${CLUSTER_NAME}/data/${CLUSTER_NAME}/nodes/`.
By default ES uses a "0" directory there when only one node is accessing
it.
|
|\ \ \ \ \ \ \
| |_|_|/ / / /
|/| | | | | | |
Rename cockpit-shell -> cockpit-system
|
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | | |
This is a temporary workaround to make sure the action plugin in the
openshift_health_checker role can read the correct value specified in
the OPENSHIFT_IMAGE_TAG environment variable, instead of a jinja2
template expression.
|
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | | |
The package name has changed.
See
https://bugzilla.redhat.com/show_bug.cgi?id=1461689
https://bugzilla.redhat.com/show_bug.cgi?id=1419718
|
|\ \ \ \ \ \ \
| |_|_|_|/ / /
|/| | | | | | |
Update CloudForms templates for CF 4.5/CF 4.2
|
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | | |
'cloudforms42' for CF 4.2.
|
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | | |
'cloudforms45' for CF 4.5.
|
| | | | | | | |
|
| |/ / / / /
|/| | | | |
| | | | | |
| | | | | |
| | | | | |
| | | | | | |
Created by command:
/usr/bin/tito tag --accept-auto-changelog --use-version=3.6.114
|
| | | | | |
| | | | | |
| | | | | |
| | | | | |
| | | | | |
| | | | | | |
Created by command:
/usr/bin/tito tag --accept-auto-changelog --use-version=3.6.113
|
|\ \ \ \ \ \
| | | | | | |
| | | | | | | |
Merged by openshift-bot
|
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | | |
port the code that creates the external Elasticsearch routes to the
new logging roles
Have to suppress this error message:
SSL Problem illegal change cipher spec msg, conn state = 6, handshake state = 1
which is coming from the router health check, until
https://github.com/openshift/origin/issues/14515
is fixed - otherwise, the es log is spammed relentlessly
|
|\ \ \ \ \ \ \
| | | | | | | |
| | | | | | | | |
Merged by openshift-bot
|
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | | |
We cannot rely on the `watch.Until` call in the `rollout status`
subcommand for the time being, so we need to ignore the result of this
call. This will make the rollout status check best-effort, so we need to
follow it with a poll for the actual status of the rollout, which we can
extract from the `openshift.io/deployment.phase` annotation on the
ReplicationControllers. This annotation can have only three values --
`Running`, `Complete` and `Failed`. If we poll on this attribute until
we stop seeing `Running`, we can then inspect the last result for
`Failed`; if it's present, we have failed the deployment.
Signed-off-by: Steve Kuznetsov <skuznets@redhat.com>
|
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | | |
When deploying the hosted router and registry components, we need to
ensure that they correctly roll out. The previous checks were weak in
that they either simply waited for a set amount of time and/or did one
replica check. They would fail if the router or registry took longer to
deploy or if there were un-ready or failing replicas. The `oc rollout`
command group contains the `status` endpoint for internalizing all of
the logic for determining when a rollout has succeeded or failed, so
simply using this client call will ensure that the router and registry
correctly deploy.
Signed-off-by: Steve Kuznetsov <skuznets@redhat.com>
|
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | | |
Created by command:
/usr/bin/tito tag --accept-auto-changelog --use-version=3.6.112
|
|\ \ \ \ \ \ \ \
| | | | | | | | |
| | | | | | | | | |
Add the other featured audit-config paramters as example
|
| | | | | | | | | |
|
| | | | | | | | |
| | | | | | | | |
| | | | | | | | |
| | | | | | | | |
| | | | | | | | |
| | | | | | | | | |
Created by command:
/usr/bin/tito tag --accept-auto-changelog --use-version=3.6.111
|
|\ \ \ \ \ \ \ \ \
| |_|_|_|_|/ / / /
|/| | | | | | | | |
CFME as an OpenShift Pod
|
| | | | | | | | | |
|
| | | | | | | | | |
|
| | | | | | | | | |
|
| | |_|_|_|_|_|/
| |/| | | | | | |
|
|\ \ \ \ \ \ \ \
| | | | | | | | |
| | | | | | | | | |
System container honor provided ansible cfg
|
| | | | | | | | | |
|
| | |/ / / / / /
| |/| | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | | |
The provided configuration file was not being honored. This change adds
a new variable called ANSIBLE_CONFIG which points to the ansible.cfg
file within the container. By default it is set to the configuration
provided by openshift-ansible within the container at:
/usr/share/ansible/openshift-ansible/ansible.cfg
|
|\ \ \ \ \ \ \ \
| |/ / / / / / /
|/| | | | | | | |
Merged by openshift-bot
|
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | | |
adding
openshift_logging_fluentd_buffer_queue_limit: 1024
openshift_logging_fluentd_buffer_size_limit: 1m
openshift_logging_mux_buffer_queue_limit: 1024
openshift_logging_mux_buffer_size_limit: 1m
and setting the matched environment variables.
|