From be97433dd559a3bdae4baedda20a7f17bd47450b Mon Sep 17 00:00:00 2001
From: Russell Teague <rteague@redhat.com>
Date: Fri, 9 Dec 2016 14:40:43 -0500
Subject: YAML Linting

* Added checks to make ci for yaml linting
* Modified y(a)ml files to pass lint checks
---
 .../tasks/cleanup_logging.yaml                     |  86 ++---
 .../tasks/deploy_logging.yaml                      | 348 ++++++++++-----------
 roles/openshift_hosted_logging/vars/main.yaml      |   1 +
 3 files changed, 218 insertions(+), 217 deletions(-)

(limited to 'roles/openshift_hosted_logging')

diff --git a/roles/openshift_hosted_logging/tasks/cleanup_logging.yaml b/roles/openshift_hosted_logging/tasks/cleanup_logging.yaml
index 8754616d9..70b0d67a4 100644
--- a/roles/openshift_hosted_logging/tasks/cleanup_logging.yaml
+++ b/roles/openshift_hosted_logging/tasks/cleanup_logging.yaml
@@ -1,59 +1,59 @@
 ---
-  - name: Create temp directory for kubeconfig
-    command: mktemp -d /tmp/openshift-ansible-XXXXXX
-    register: mktemp
-    changed_when: False
+- name: Create temp directory for kubeconfig
+  command: mktemp -d /tmp/openshift-ansible-XXXXXX
+  register: mktemp
+  changed_when: False
 
-  - name: Copy the admin client config(s)
-    command: >
-      cp {{ openshift_master_config_dir }}/admin.kubeconfig {{ mktemp.stdout }}/admin.kubeconfig
-    changed_when: False
+- name: Copy the admin client config(s)
+  command: >
+    cp {{ openshift_master_config_dir }}/admin.kubeconfig {{ mktemp.stdout }}/admin.kubeconfig
+  changed_when: False
 
-  - name: "Checking for logging project"
-    command: "{{ openshift.common.client_binary }} --config={{ mktemp.stdout }}/admin.kubeconfig get project logging"
-    register: logging_project
-    failed_when: "'FAILED' in logging_project.stderr"
+- name: "Checking for logging project"
+  command: "{{ openshift.common.client_binary }} --config={{ mktemp.stdout }}/admin.kubeconfig get project logging"
+  register: logging_project
+  failed_when: "'FAILED' in logging_project.stderr"
 
-  - name: "Changing projects"
-    command: "{{ openshift.common.client_binary }} --config={{ mktemp.stdout }}/admin.kubeconfig project logging"
+- name: "Changing projects"
+  command: "{{ openshift.common.client_binary }} --config={{ mktemp.stdout }}/admin.kubeconfig project logging"
 
 
-  - name: "Cleanup any previous logging infrastructure"
-    command: "{{ openshift.common.client_binary }} --config={{ mktemp.stdout }}/admin.kubeconfig delete --ignore-not-found all --selector logging-infra={{ item }}"
-    with_items:
-      - kibana
-      - fluentd
-      - elasticsearch
-    ignore_errors: yes
+- name: "Cleanup any previous logging infrastructure"
+  command: "{{ openshift.common.client_binary }} --config={{ mktemp.stdout }}/admin.kubeconfig delete --ignore-not-found all --selector logging-infra={{ item }}"
+  with_items:
+    - kibana
+    - fluentd
+    - elasticsearch
+  ignore_errors: yes
 
-  - name: "Cleanup existing support infrastructure"
-    command: "{{ openshift.common.client_binary }} --config={{ mktemp.stdout }}/admin.kubeconfig delete --ignore-not-found all,sa,oauthclient --selector logging-infra=support"
-    ignore_errors: yes
+- name: "Cleanup existing support infrastructure"
+  command: "{{ openshift.common.client_binary }} --config={{ mktemp.stdout }}/admin.kubeconfig delete --ignore-not-found all,sa,oauthclient --selector logging-infra=support"
+  ignore_errors: yes
 
-  - name: "Cleanup existing secrets"
-    command: "{{ openshift.common.client_binary }} --config={{ mktemp.stdout }}/admin.kubeconfig delete secret logging-fluentd logging-elasticsearch logging-es-proxy logging-kibana logging-kibana-proxy logging-kibana-ops-proxy"
-    ignore_errors: yes
-    register: clean_result
-    failed_when: clean_result.rc == 1 and 'not found' not in clean_result.stderr
+- name: "Cleanup existing secrets"
+  command: "{{ openshift.common.client_binary }} --config={{ mktemp.stdout }}/admin.kubeconfig delete secret logging-fluentd logging-elasticsearch logging-es-proxy logging-kibana logging-kibana-proxy logging-kibana-ops-proxy"
+  ignore_errors: yes
+  register: clean_result
+  failed_when: clean_result.rc == 1 and 'not found' not in clean_result.stderr
 
-  - name: "Cleanup existing logging deployers"
-    command: "{{ openshift.common.client_binary }} --config={{ mktemp.stdout }}/admin.kubeconfig delete pods --all"
+- name: "Cleanup existing logging deployers"
+  command: "{{ openshift.common.client_binary }} --config={{ mktemp.stdout }}/admin.kubeconfig delete pods --all"
 
 
-  - name: "Cleanup logging project"
-    command: "{{ openshift.common.client_binary }} --config={{ mktemp.stdout }}/admin.kubeconfig delete project logging"
+- name: "Cleanup logging project"
+  command: "{{ openshift.common.client_binary }} --config={{ mktemp.stdout }}/admin.kubeconfig delete project logging"
 
 
-  - name: "Remove deployer template"
-    command: "{{ openshift.common.client_binary }} --config={{ mktemp.stdout }}/admin.kubeconfig delete template logging-deployer-template -n openshift"
-    register: delete_output
-    failed_when: delete_output.rc == 1 and 'exists' not in delete_output.stderr
+- name: "Remove deployer template"
+  command: "{{ openshift.common.client_binary }} --config={{ mktemp.stdout }}/admin.kubeconfig delete template logging-deployer-template -n openshift"
+  register: delete_output
+  failed_when: delete_output.rc == 1 and 'exists' not in delete_output.stderr
 
 
-  - name: Delete temp directory
-    file:
-      name: "{{ mktemp.stdout }}"
-      state: absent
-    changed_when: False
+- name: Delete temp directory
+  file:
+    name: "{{ mktemp.stdout }}"
+    state: absent
+  changed_when: False
 
-  - debug: msg="Success!"
+- debug: msg="Success!"
diff --git a/roles/openshift_hosted_logging/tasks/deploy_logging.yaml b/roles/openshift_hosted_logging/tasks/deploy_logging.yaml
index 625af9acd..513a74c69 100644
--- a/roles/openshift_hosted_logging/tasks/deploy_logging.yaml
+++ b/roles/openshift_hosted_logging/tasks/deploy_logging.yaml
@@ -1,175 +1,175 @@
 ---
-  - debug: msg="WARNING target_registry is deprecated, use openshift_hosted_logging_image_prefix instead"
-    when: target_registry is defined and target_registry
-
-  - fail: msg="This role requires the following vars to be defined. openshift_hosted_logging_master_public_url, openshift_hosted_logging_hostname, openshift_hosted_logging_elasticsearch_cluster_size"
-    when: "openshift_hosted_logging_hostname is not defined or
-          openshift_hosted_logging_elasticsearch_cluster_size is not defined or
-          openshift_hosted_logging_master_public_url is not defined"
-
-  - name: Create temp directory for kubeconfig
-    command: mktemp -d /tmp/openshift-ansible-XXXXXX
-    register: mktemp
-    changed_when: False
-
-  - name: Copy the admin client config(s)
-    command: >
-      cp {{ openshift_master_config_dir }}/admin.kubeconfig {{ mktemp.stdout }}/admin.kubeconfig
-    changed_when: False
-
-  - name: "Check for logging project already exists"
-    command: >
-      {{ openshift.common.client_binary }} --config={{ mktemp.stdout }}/admin.kubeconfig get project logging -o jsonpath='{.metadata.name}'
-    register: logging_project_result
-    ignore_errors: True
-
-  - name: "Create logging project"
-    command: >
-      {{ openshift.common.client_binary }} adm --config={{ mktemp.stdout }}/admin.kubeconfig new-project logging
-    when: logging_project_result.stdout == ""
-
-  - name: "Changing projects"
-    command: >
-      {{ openshift.common.client_binary }} --config={{ mktemp.stdout }}/admin.kubeconfig project logging
-
-  - name: "Creating logging deployer secret"
-    command: >
-      {{ openshift.common.client_binary }} --config={{ mktemp.stdout }}/admin.kubeconfig secrets new logging-deployer {{ openshift_hosted_logging_secret_vars | default('nothing=/dev/null') }}
-    register: secret_output
-    failed_when: "secret_output.rc == 1 and 'exists' not in secret_output.stderr"
-
-  - name: "Create templates for logging accounts and the deployer"
-    command: >
-      {{ openshift.common.client_binary }} create --config={{ mktemp.stdout }}/admin.kubeconfig
-      -f {{ hosted_base }}/logging-deployer.yaml
-      --config={{ mktemp.stdout }}/admin.kubeconfig
-      -n logging
-    register: logging_import_template
-    failed_when: "'already exists' not in logging_import_template.stderr and logging_import_template.rc != 0"
-    changed_when: "'created' in logging_import_template.stdout"
-
-  - name: "Process the logging accounts template"
-    shell: >
-      {{ openshift.common.client_binary }} --config={{ mktemp.stdout }}/admin.kubeconfig
-      process logging-deployer-account-template |  {{ openshift.common.client_binary }} --config={{ mktemp.stdout }}/admin.kubeconfig create -f -
-    register: process_deployer_accounts
-    failed_when: process_deployer_accounts.rc == 1 and 'already exists' not in process_deployer_accounts.stderr
-
-  - name: "Set permissions for logging-deployer service account"
-    command: >
-      {{ openshift.common.client_binary }} adm --config={{ mktemp.stdout }}/admin.kubeconfig
-      policy add-cluster-role-to-user oauth-editor system:serviceaccount:logging:logging-deployer
-    register: permiss_output
-    failed_when: "permiss_output.rc == 1 and 'exists' not in permiss_output.stderr"
-
-  - name: "Set permissions for fluentd"
-    command: >
-      {{ openshift.common.client_binary }} adm --config={{ mktemp.stdout }}/admin.kubeconfig
-      policy add-scc-to-user privileged system:serviceaccount:logging:aggregated-logging-fluentd
-    register: fluentd_output
-    failed_when: "fluentd_output.rc == 1 and 'exists' not in fluentd_output.stderr"
-
-  - name: "Set additional permissions for fluentd"
-    command: >
-      {{ openshift.common.client_binary }} adm policy --config={{ mktemp.stdout }}/admin.kubeconfig
-      add-cluster-role-to-user cluster-reader system:serviceaccount:logging:aggregated-logging-fluentd
-    register: fluentd2_output
-    failed_when: "fluentd2_output.rc == 1 and 'exists' not in fluentd2_output.stderr"
-
-  - name: "Add rolebinding-reader to aggregated-logging-elasticsearch"
-    command: >
-      {{ openshift.common.client_binary }} adm --config={{ mktemp.stdout }}/admin.kubeconfig
-      policy add-cluster-role-to-user rolebinding-reader \
-      system:serviceaccount:logging:aggregated-logging-elasticsearch
-    register: rolebinding_reader_output
-    failed_when: "rolebinding_reader_output == 1 and 'exists' not in rolebinding_reader_output.stderr"
-
-  - name: "Create ConfigMap for deployer parameters"
-    command: >
-      {{ openshift.common.client_binary}} --config={{ mktemp.stdout }}/admin.kubeconfig create configmap logging-deployer {{ deployer_cmap_params }}
-    register: deployer_configmap_output
-    failed_when: "deployer_configmap_output.rc == 1 and 'exists' not in deployer_configmap_output.stderr"
-
-  - name: "Process the deployer template"
-    shell:  "{{ openshift.common.client_binary }} --config={{ mktemp.stdout }}/admin.kubeconfig new-app logging-deployer-template {{ oc_new_app_values }}"
-    register: process_deployer
-    failed_when: process_deployer.rc == 1 and 'already exists' not in process_deployer.stderr
-
-  - name: "Wait for image pull and deployer pod"
-    shell:  "{{ openshift.common.client_binary }} --config={{ mktemp.stdout }}/admin.kubeconfig get pods | grep logging-deployer.*Completed"
-    register: result
-    until: result.rc == 0
-    retries: 20
-    delay: 15
-
-  - name: "Process imagestream template"
-    shell:  "{{ openshift.common.client_binary }} --config={{ mktemp.stdout }}/admin.kubeconfig new-app logging-imagestream-template {{ oc_new_app_values }}"
-    when: tr_or_ohlip is defined and insecure_registry is defined and insecure_registry
-    register: process_is
-    failed_when: process_is.rc == 1 and 'already exists' not in process_is.stderr
-
-  - name: "Set insecured registry"
-    command:  "{{ openshift.common.client_binary }} --config={{ mktemp.stdout }}/admin.kubeconfig annotate is --all  openshift.io/image.insecureRepository=true --overwrite"
-    when: tr_or_ohlip is defined and insecure_registry is defined and insecure_registry
-
-  - name: "Wait for imagestreams to become available"
-    shell:  "{{ openshift.common.client_binary }} --config={{ mktemp.stdout }}/admin.kubeconfig get is | grep logging-fluentd"
-    when: tr_or_ohlip is defined and insecure_registry is defined and insecure_registry
-    register: result
-    until: result.rc == 0
-    failed_when: result.rc == 1 and 'not found' not in result.stderr
-    retries: 20
-    delay: 5
-
-  - name: "Wait for component pods to be running"
-    shell:  "{{ openshift.common.client_binary }} --config={{ mktemp.stdout }}/admin.kubeconfig get pods -l component={{ item }} | grep Running"
-    with_items:
-      - es
-      - kibana
-      - curator
-    register: result
-    until: result.rc == 0
-    failed_when: result.rc == 1 or 'Error' in result.stderr
-    retries: 20
-    delay: 15
-
-  - name: "Wait for ops component pods to be running"
-    shell:  "{{ openshift.common.client_binary }} --config={{ mktemp.stdout }}/admin.kubeconfig get pods -l component={{ item }} | grep Running"
-    with_items:
-      - es-ops
-      - kibana-ops
-      - curator-ops
-    when: openshift_hosted_logging_enable_ops_cluster is defined and openshift_hosted_logging_enable_ops_cluster
-    register: result
-    until: result.rc == 0
-    failed_when: result.rc == 1 or 'Error' in result.stderr
-    retries: 20
-    delay: 15
-
-  - name: "Wait for fluentd DaemonSet to exist"
-    shell:  "{{ openshift.common.client_binary }} --config={{ mktemp.stdout }}/admin.kubeconfig get daemonset logging-fluentd"
-    register: result
-    until: result.rc == 0
-    failed_when: result.rc == 1 or 'Error' in result.stderr
-    retries: 20
-    delay: 5
-
-  - name: "Deploy fluentd by labeling the node"
-    shell:  "{{ openshift.common.client_binary }} --config={{ mktemp.stdout }}/admin.kubeconfig label node --overwrite=true {{ '-l' ~ openshift_hosted_logging_fluentd_nodeselector if openshift_hosted_logging_fluentd_nodeselector is defined else '--all' }} {{ openshift_hosted_logging_fluentd_nodeselector_label if openshift_hosted_logging_fluentd_nodeselector_label is defined else 'logging-infra-fluentd=true' }}"
-
-  - name: "Wait for fluentd to be running"
-    shell:  "{{ openshift.common.client_binary }} --config={{ mktemp.stdout }}/admin.kubeconfig get pods -l component=fluentd | grep Running"
-    register: result
-    until: result.rc == 0
-    failed_when: result.rc == 1 or 'Error' in result.stderr
-    retries: 20
-    delay: 15
-
-  - debug:
-      msg: "Logging components deployed. Note persistent volume for elasticsearch must be setup manually"
-
-  - name: Delete temp directory
-    file:
-      name: "{{ mktemp.stdout }}"
-      state: absent
-    changed_when: False
+- debug: msg="WARNING target_registry is deprecated, use openshift_hosted_logging_image_prefix instead"
+  when: target_registry is defined and target_registry
+
+- fail: msg="This role requires the following vars to be defined. openshift_hosted_logging_master_public_url, openshift_hosted_logging_hostname, openshift_hosted_logging_elasticsearch_cluster_size"
+  when: "openshift_hosted_logging_hostname is not defined or
+        openshift_hosted_logging_elasticsearch_cluster_size is not defined or
+        openshift_hosted_logging_master_public_url is not defined"
+
+- name: Create temp directory for kubeconfig
+  command: mktemp -d /tmp/openshift-ansible-XXXXXX
+  register: mktemp
+  changed_when: False
+
+- name: Copy the admin client config(s)
+  command: >
+    cp {{ openshift_master_config_dir }}/admin.kubeconfig {{ mktemp.stdout }}/admin.kubeconfig
+  changed_when: False
+
+- name: "Check for logging project already exists"
+  command: >
+    {{ openshift.common.client_binary }} --config={{ mktemp.stdout }}/admin.kubeconfig get project logging -o jsonpath='{.metadata.name}'
+  register: logging_project_result
+  ignore_errors: True
+
+- name: "Create logging project"
+  command: >
+    {{ openshift.common.client_binary }} adm --config={{ mktemp.stdout }}/admin.kubeconfig new-project logging
+  when: logging_project_result.stdout == ""
+
+- name: "Changing projects"
+  command: >
+    {{ openshift.common.client_binary }} --config={{ mktemp.stdout }}/admin.kubeconfig project logging
+
+- name: "Creating logging deployer secret"
+  command: >
+    {{ openshift.common.client_binary }} --config={{ mktemp.stdout }}/admin.kubeconfig secrets new logging-deployer {{ openshift_hosted_logging_secret_vars | default('nothing=/dev/null') }}
+  register: secret_output
+  failed_when: "secret_output.rc == 1 and 'exists' not in secret_output.stderr"
+
+- name: "Create templates for logging accounts and the deployer"
+  command: >
+    {{ openshift.common.client_binary }} create --config={{ mktemp.stdout }}/admin.kubeconfig
+    -f {{ hosted_base }}/logging-deployer.yaml
+    --config={{ mktemp.stdout }}/admin.kubeconfig
+    -n logging
+  register: logging_import_template
+  failed_when: "'already exists' not in logging_import_template.stderr and logging_import_template.rc != 0"
+  changed_when: "'created' in logging_import_template.stdout"
+
+- name: "Process the logging accounts template"
+  shell: >
+    {{ openshift.common.client_binary }} --config={{ mktemp.stdout }}/admin.kubeconfig
+    process logging-deployer-account-template |  {{ openshift.common.client_binary }} --config={{ mktemp.stdout }}/admin.kubeconfig create -f -
+  register: process_deployer_accounts
+  failed_when: process_deployer_accounts.rc == 1 and 'already exists' not in process_deployer_accounts.stderr
+
+- name: "Set permissions for logging-deployer service account"
+  command: >
+    {{ openshift.common.client_binary }} adm --config={{ mktemp.stdout }}/admin.kubeconfig
+    policy add-cluster-role-to-user oauth-editor system:serviceaccount:logging:logging-deployer
+  register: permiss_output
+  failed_when: "permiss_output.rc == 1 and 'exists' not in permiss_output.stderr"
+
+- name: "Set permissions for fluentd"
+  command: >
+    {{ openshift.common.client_binary }} adm --config={{ mktemp.stdout }}/admin.kubeconfig
+    policy add-scc-to-user privileged system:serviceaccount:logging:aggregated-logging-fluentd
+  register: fluentd_output
+  failed_when: "fluentd_output.rc == 1 and 'exists' not in fluentd_output.stderr"
+
+- name: "Set additional permissions for fluentd"
+  command: >
+    {{ openshift.common.client_binary }} adm policy --config={{ mktemp.stdout }}/admin.kubeconfig
+    add-cluster-role-to-user cluster-reader system:serviceaccount:logging:aggregated-logging-fluentd
+  register: fluentd2_output
+  failed_when: "fluentd2_output.rc == 1 and 'exists' not in fluentd2_output.stderr"
+
+- name: "Add rolebinding-reader to aggregated-logging-elasticsearch"
+  command: >
+    {{ openshift.common.client_binary }} adm --config={{ mktemp.stdout }}/admin.kubeconfig
+    policy add-cluster-role-to-user rolebinding-reader \
+    system:serviceaccount:logging:aggregated-logging-elasticsearch
+  register: rolebinding_reader_output
+  failed_when: "rolebinding_reader_output == 1 and 'exists' not in rolebinding_reader_output.stderr"
+
+- name: "Create ConfigMap for deployer parameters"
+  command: >
+    {{ openshift.common.client_binary}} --config={{ mktemp.stdout }}/admin.kubeconfig create configmap logging-deployer {{ deployer_cmap_params }}
+  register: deployer_configmap_output
+  failed_when: "deployer_configmap_output.rc == 1 and 'exists' not in deployer_configmap_output.stderr"
+
+- name: "Process the deployer template"
+  shell: "{{ openshift.common.client_binary }} --config={{ mktemp.stdout }}/admin.kubeconfig new-app logging-deployer-template {{ oc_new_app_values }}"
+  register: process_deployer
+  failed_when: process_deployer.rc == 1 and 'already exists' not in process_deployer.stderr
+
+- name: "Wait for image pull and deployer pod"
+  shell: "{{ openshift.common.client_binary }} --config={{ mktemp.stdout }}/admin.kubeconfig get pods | grep logging-deployer.*Completed"
+  register: result
+  until: result.rc == 0
+  retries: 20
+  delay: 15
+
+- name: "Process imagestream template"
+  shell: "{{ openshift.common.client_binary }} --config={{ mktemp.stdout }}/admin.kubeconfig new-app logging-imagestream-template {{ oc_new_app_values }}"
+  when: tr_or_ohlip is defined and insecure_registry is defined and insecure_registry
+  register: process_is
+  failed_when: process_is.rc == 1 and 'already exists' not in process_is.stderr
+
+- name: "Set insecured registry"
+  command: "{{ openshift.common.client_binary }} --config={{ mktemp.stdout }}/admin.kubeconfig annotate is --all  openshift.io/image.insecureRepository=true --overwrite"
+  when: tr_or_ohlip is defined and insecure_registry is defined and insecure_registry
+
+- name: "Wait for imagestreams to become available"
+  shell: "{{ openshift.common.client_binary }} --config={{ mktemp.stdout }}/admin.kubeconfig get is | grep logging-fluentd"
+  when: tr_or_ohlip is defined and insecure_registry is defined and insecure_registry
+  register: result
+  until: result.rc == 0
+  failed_when: result.rc == 1 and 'not found' not in result.stderr
+  retries: 20
+  delay: 5
+
+- name: "Wait for component pods to be running"
+  shell: "{{ openshift.common.client_binary }} --config={{ mktemp.stdout }}/admin.kubeconfig get pods -l component={{ item }} | grep Running"
+  with_items:
+    - es
+    - kibana
+    - curator
+  register: result
+  until: result.rc == 0
+  failed_when: result.rc == 1 or 'Error' in result.stderr
+  retries: 20
+  delay: 15
+
+- name: "Wait for ops component pods to be running"
+  shell: "{{ openshift.common.client_binary }} --config={{ mktemp.stdout }}/admin.kubeconfig get pods -l component={{ item }} | grep Running"
+  with_items:
+    - es-ops
+    - kibana-ops
+    - curator-ops
+  when: openshift_hosted_logging_enable_ops_cluster is defined and openshift_hosted_logging_enable_ops_cluster
+  register: result
+  until: result.rc == 0
+  failed_when: result.rc == 1 or 'Error' in result.stderr
+  retries: 20
+  delay: 15
+
+- name: "Wait for fluentd DaemonSet to exist"
+  shell: "{{ openshift.common.client_binary }} --config={{ mktemp.stdout }}/admin.kubeconfig get daemonset logging-fluentd"
+  register: result
+  until: result.rc == 0
+  failed_when: result.rc == 1 or 'Error' in result.stderr
+  retries: 20
+  delay: 5
+
+- name: "Deploy fluentd by labeling the node"
+  shell: "{{ openshift.common.client_binary }} --config={{ mktemp.stdout }}/admin.kubeconfig label node --overwrite=true {{ '-l' ~ openshift_hosted_logging_fluentd_nodeselector if openshift_hosted_logging_fluentd_nodeselector is defined else '--all' }} {{ openshift_hosted_logging_fluentd_nodeselector_label if openshift_hosted_logging_fluentd_nodeselector_label is defined else 'logging-infra-fluentd=true' }}"
+
+- name: "Wait for fluentd to be running"
+  shell: "{{ openshift.common.client_binary }} --config={{ mktemp.stdout }}/admin.kubeconfig get pods -l component=fluentd | grep Running"
+  register: result
+  until: result.rc == 0
+  failed_when: result.rc == 1 or 'Error' in result.stderr
+  retries: 20
+  delay: 15
+
+- debug:
+    msg: "Logging components deployed. Note persistent volume for elasticsearch must be setup manually"
+
+- name: Delete temp directory
+  file:
+    name: "{{ mktemp.stdout }}"
+    state: absent
+  changed_when: False
diff --git a/roles/openshift_hosted_logging/vars/main.yaml b/roles/openshift_hosted_logging/vars/main.yaml
index 11412733b..33320e9c8 100644
--- a/roles/openshift_hosted_logging/vars/main.yaml
+++ b/roles/openshift_hosted_logging/vars/main.yaml
@@ -1,3 +1,4 @@
+---
 tr_or_ohlip: "{{ openshift_hosted_logging_deployer_prefix | default(target_registry) | default(None) }}"
 ip_kv: "{{ '-p IMAGE_PREFIX=' ~ tr_or_ohlip | quote if tr_or_ohlip != '' else '' }}"
 iv_kv: "{{ '-p IMAGE_VERSION=' ~ openshift_hosted_logging_deployer_version | quote if openshift_hosted_logging_deployer_version | default(none) is not none else '' }}"
-- 
cgit v1.2.3