From a6b49265c5b287617d7971ceee443303bdb53174 Mon Sep 17 00:00:00 2001 From: Guillaume Abrioux Date: Sat, 9 Mar 2024 10:24:23 +0100 Subject: [PATCH] osd: drop openstack related tasks All of this should be addressed in custom separate playbooks if needed. Signed-off-by: Guillaume Abrioux --- group_vars/all.yml.sample | 58 ---------------- roles/ceph-defaults/defaults/main.yml | 58 ---------------- roles/ceph-osd/tasks/main.yml | 9 --- roles/ceph-osd/tasks/openstack_config.yml | 68 ------------------- roles/ceph-validate/tasks/check_pools.yml | 1 - .../all-in-one/container/group_vars/all | 1 - tests/functional/all-in-one/group_vars/all | 1 - .../all_daemons/container/group_vars/all | 12 ---- tests/functional/all_daemons/group_vars/all | 14 ---- .../all_daemons_ipv6/container/group_vars/all | 12 ---- .../all_daemons_ipv6/group_vars/all | 14 ---- tests/functional/docker2podman/group_vars/all | 14 +--- .../container/inventory/group_vars/all | 1 - .../external_clients/inventory/group_vars/all | 1 - .../lvm-osds/container/group_vars/all | 16 +---- tests/functional/lvm-osds/group_vars/all | 15 +--- tests/functional/ooo-collocation/hosts | 1 - tests/functional/podman/group_vars/all | 14 +--- .../shrink_mds/container/group_vars/all | 1 - .../shrink_mgr/container/group_vars/all | 1 - .../shrink_mon/container/group_vars/all | 1 - .../shrink_osd/container/group_vars/all | 1 - tests/functional/shrink_osd/group_vars/all | 1 - .../shrink_rbdmirror/container/group_vars/all | 1 - .../shrink_rgw/container/group_vars/all | 1 - .../subset_update/container/group_vars/all | 1 - tests/functional/subset_update/group_vars/all | 1 - 27 files changed, 4 insertions(+), 315 deletions(-) delete mode 100644 roles/ceph-osd/tasks/openstack_config.yml diff --git a/group_vars/all.yml.sample b/group_vars/all.yml.sample index 77d3536d1d..c146cdbb07 100644 --- a/group_vars/all.yml.sample +++ b/group_vars/all.yml.sample @@ -554,64 +554,6 @@ dummy: #docker_pull_timeout: "300s" -############# -# OPENSTACK # -############# -#openstack_config: false -# When pg_autoscale_mode is set to True, you must add the target_size_ratio key with a correct value -# `pg_num` and `pgp_num` keys will be ignored, even if specified. -# eg: -# openstack_glance_pool: -# name: "images" -# rule_name: "my_replicated_rule" -# application: "rbd" -# pg_autoscale_mode: false -# pg_num: 16 -# pgp_num: 16 -# target_size_ratio: 0.2 -#openstack_glance_pool: -# name: "images" -# application: "rbd" -#openstack_cinder_pool: -# name: "volumes" -# application: "rbd" -#openstack_nova_pool: -# name: "vms" -# application: "rbd" -#openstack_cinder_backup_pool: -# name: "backups" -# application: "rbd" -#openstack_gnocchi_pool: -# name: "metrics" -# application: "rbd" -#openstack_cephfs_data_pool: -# name: "manila_data" -# application: "cephfs" -#openstack_cephfs_metadata_pool: -# name: "manila_metadata" -# application: "cephfs" -#openstack_pools: -# - "{{ openstack_glance_pool }}" -# - "{{ openstack_cinder_pool }}" -# - "{{ openstack_nova_pool }}" -# - "{{ openstack_cinder_backup_pool }}" -# - "{{ openstack_gnocchi_pool }}" -# - "{{ openstack_cephfs_data_pool }}" -# - "{{ openstack_cephfs_metadata_pool }}" - - -# The value for 'key' can be a pre-generated key, -# e.g key: "AQDC2UxZH4yeLhAAgTaZb+4wDUlYOsr1OfZSpQ==" -# By default, keys will be auto-generated. -# -#openstack_keys: -# - { name: client.glance, caps: { mon: "profile rbd", osd: "profile rbd pool={{ openstack_cinder_pool.name }}, profile rbd pool={{ openstack_glance_pool.name }}"}, mode: "0600" } -# - { name: client.cinder, caps: { mon: "profile rbd", osd: "profile rbd pool={{ openstack_cinder_pool.name }}, profile rbd pool={{ openstack_nova_pool.name }}, profile rbd pool={{ openstack_glance_pool.name }}"}, mode: "0600" } -# - { name: client.cinder-backup, caps: { mon: "profile rbd", osd: "profile rbd pool={{ openstack_cinder_backup_pool.name }}"}, mode: "0600" } -# - { name: client.gnocchi, caps: { mon: "profile rbd", osd: "profile rbd pool={{ openstack_gnocchi_pool.name }}"}, mode: "0600", } -# - { name: client.openstack, caps: { mon: "profile rbd", osd: "profile rbd pool={{ openstack_glance_pool.name }}, profile rbd pool={{ openstack_nova_pool.name }}, profile rbd pool={{ openstack_cinder_pool.name }}, profile rbd pool={{ openstack_cinder_backup_pool.name }}"}, mode: "0600" } - - ############# # DASHBOARD # ############# diff --git a/roles/ceph-defaults/defaults/main.yml b/roles/ceph-defaults/defaults/main.yml index 79ef487dee..a6ff3229d8 100644 --- a/roles/ceph-defaults/defaults/main.yml +++ b/roles/ceph-defaults/defaults/main.yml @@ -546,64 +546,6 @@ docker_pull_retry: 3 docker_pull_timeout: "300s" -############# -# OPENSTACK # -############# -openstack_config: false -# When pg_autoscale_mode is set to True, you must add the target_size_ratio key with a correct value -# `pg_num` and `pgp_num` keys will be ignored, even if specified. -# eg: -# openstack_glance_pool: -# name: "images" -# rule_name: "my_replicated_rule" -# application: "rbd" -# pg_autoscale_mode: false -# pg_num: 16 -# pgp_num: 16 -# target_size_ratio: 0.2 -openstack_glance_pool: - name: "images" - application: "rbd" -openstack_cinder_pool: - name: "volumes" - application: "rbd" -openstack_nova_pool: - name: "vms" - application: "rbd" -openstack_cinder_backup_pool: - name: "backups" - application: "rbd" -openstack_gnocchi_pool: - name: "metrics" - application: "rbd" -openstack_cephfs_data_pool: - name: "manila_data" - application: "cephfs" -openstack_cephfs_metadata_pool: - name: "manila_metadata" - application: "cephfs" -openstack_pools: - - "{{ openstack_glance_pool }}" - - "{{ openstack_cinder_pool }}" - - "{{ openstack_nova_pool }}" - - "{{ openstack_cinder_backup_pool }}" - - "{{ openstack_gnocchi_pool }}" - - "{{ openstack_cephfs_data_pool }}" - - "{{ openstack_cephfs_metadata_pool }}" - - -# The value for 'key' can be a pre-generated key, -# e.g key: "AQDC2UxZH4yeLhAAgTaZb+4wDUlYOsr1OfZSpQ==" -# By default, keys will be auto-generated. -# -openstack_keys: - - { name: client.glance, caps: { mon: "profile rbd", osd: "profile rbd pool={{ openstack_cinder_pool.name }}, profile rbd pool={{ openstack_glance_pool.name }}"}, mode: "0600" } - - { name: client.cinder, caps: { mon: "profile rbd", osd: "profile rbd pool={{ openstack_cinder_pool.name }}, profile rbd pool={{ openstack_nova_pool.name }}, profile rbd pool={{ openstack_glance_pool.name }}"}, mode: "0600" } - - { name: client.cinder-backup, caps: { mon: "profile rbd", osd: "profile rbd pool={{ openstack_cinder_backup_pool.name }}"}, mode: "0600" } - - { name: client.gnocchi, caps: { mon: "profile rbd", osd: "profile rbd pool={{ openstack_gnocchi_pool.name }}"}, mode: "0600", } - - { name: client.openstack, caps: { mon: "profile rbd", osd: "profile rbd pool={{ openstack_glance_pool.name }}, profile rbd pool={{ openstack_nova_pool.name }}, profile rbd pool={{ openstack_cinder_pool.name }}, profile rbd pool={{ openstack_cinder_backup_pool.name }}"}, mode: "0600" } - - ############# # DASHBOARD # ############# diff --git a/roles/ceph-osd/tasks/main.yml b/roles/ceph-osd/tasks/main.yml index f51fc98459..ce709bc2f4 100644 --- a/roles/ceph-osd/tasks/main.yml +++ b/roles/ceph-osd/tasks/main.yml @@ -100,12 +100,3 @@ when: hostvars[groups[mon_group_name][0]]['crush_rule_config'] | default(crush_rule_config) | bool tags: wait_all_osds_up -# Create the pools listed in openstack_pools -- name: Include openstack_config.yml - ansible.builtin.include_tasks: openstack_config.yml - when: - - not add_osd | bool - - not rolling_update | default(False) | bool - - openstack_config | bool - - inventory_hostname == groups[osd_group_name] | last - tags: wait_all_osds_up diff --git a/roles/ceph-osd/tasks/openstack_config.yml b/roles/ceph-osd/tasks/openstack_config.yml deleted file mode 100644 index 6b276c2c0f..0000000000 --- a/roles/ceph-osd/tasks/openstack_config.yml +++ /dev/null @@ -1,68 +0,0 @@ ---- -- name: Pool related tasks - block: - - name: Create openstack pool(s) - ceph_pool: - name: "{{ item.name }}" - cluster: "{{ cluster }}" - pg_num: "{{ item.pg_num | default(omit) }}" - pgp_num: "{{ item.pgp_num | default(omit) }}" - size: "{{ item.size | default(omit) }}" - min_size: "{{ item.min_size | default(omit) }}" - pool_type: "{{ item.type | default('replicated') }}" - rule_name: "{{ item.rule_name | default(omit) }}" - erasure_profile: "{{ item.erasure_profile | default(omit) }}" - pg_autoscale_mode: "{{ item.pg_autoscale_mode | default(omit) }}" - target_size_ratio: "{{ item.target_size_ratio | default(omit) }}" - application: "{{ item.application | default(omit) }}" - with_items: "{{ openstack_pools }}" - delegate_to: "{{ groups[mon_group_name][0] }}" - environment: - CEPH_CONTAINER_IMAGE: "{{ ceph_docker_registry + '/' + ceph_docker_image + ':' + ceph_docker_image_tag if containerized_deployment | bool else None }}" - CEPH_CONTAINER_BINARY: "{{ container_binary }}" - -- name: Create openstack cephx key(s) - when: - - cephx | bool - - openstack_config | bool - block: - - name: Generate keys - ceph_key: - name: "{{ item.name }}" - caps: "{{ item.caps }}" - secret: "{{ item.key | default('') }}" - cluster: "{{ cluster }}" - mode: "{{ item.mode | default(ceph_keyring_permissions) }}" - environment: - CEPH_CONTAINER_IMAGE: "{{ ceph_docker_registry + '/' + ceph_docker_image + ':' + ceph_docker_image_tag if containerized_deployment | bool else None }}" - CEPH_CONTAINER_BINARY: "{{ container_binary }}" - with_items: "{{ openstack_keys }}" - delegate_to: "{{ groups[mon_group_name][0] }}" - no_log: "{{ no_log_on_ceph_key_tasks }}" - - - name: Get keys from monitors - ceph_key: - name: "{{ item.name }}" - cluster: "{{ cluster }}" - output_format: plain - state: info - environment: - CEPH_CONTAINER_IMAGE: "{{ ceph_docker_registry + '/' + ceph_docker_image + ':' + ceph_docker_image_tag if containerized_deployment | bool else None }}" - CEPH_CONTAINER_BINARY: "{{ container_binary }}" - register: _osp_keys - with_items: "{{ openstack_keys }}" - delegate_to: "{{ groups.get(mon_group_name)[0] }}" - no_log: "{{ no_log_on_ceph_key_tasks }}" - - - name: Copy ceph key(s) if needed - ansible.builtin.copy: - dest: "/etc/ceph/{{ cluster }}.{{ item.0.item.name }}.keyring" - content: "{{ item.0.stdout + '\n' }}" - owner: "{{ ceph_uid if containerized_deployment | bool else 'ceph' }}" - group: "{{ ceph_uid if containerized_deployment | bool else 'ceph' }}" - mode: "{{ item.0.item.mode | default(ceph_keyring_permissions) }}" - with_nested: - - "{{ _osp_keys.results }}" - - "{{ groups[mon_group_name] }}" - delegate_to: "{{ item.1 }}" - no_log: "{{ no_log_on_ceph_key_tasks }}" diff --git a/roles/ceph-validate/tasks/check_pools.yml b/roles/ceph-validate/tasks/check_pools.yml index bca9d26ff4..0acf7c9572 100644 --- a/roles/ceph-validate/tasks/check_pools.yml +++ b/roles/ceph-validate/tasks/check_pools.yml @@ -3,7 +3,6 @@ ansible.builtin.fail: msg: "You must set a target_size_ratio value on following pool: {{ item.name }}." with_items: - - "{{ openstack_pools | default([]) }}" - "{{ cephfs_pools | default([]) }}" - "{{ pools | default([]) }}" when: diff --git a/tests/functional/all-in-one/container/group_vars/all b/tests/functional/all-in-one/container/group_vars/all index ce364b0fea..21e9bc5d50 100644 --- a/tests/functional/all-in-one/container/group_vars/all +++ b/tests/functional/all-in-one/container/group_vars/all @@ -11,7 +11,6 @@ public_network: "192.168.19.0/24" cluster_network: "192.168.20.0/24" rgw_override_bucket_index_max_shards: 16 rgw_bucket_default_quota_max_objects: 1638400 -openstack_config: True dashboard_enabled: false ceph_conf_overrides: global: diff --git a/tests/functional/all-in-one/group_vars/all b/tests/functional/all-in-one/group_vars/all index 1b5ddd6348..e4967a8b14 100644 --- a/tests/functional/all-in-one/group_vars/all +++ b/tests/functional/all-in-one/group_vars/all @@ -4,7 +4,6 @@ ceph_origin: repository ceph_repository: community radosgw_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" ceph_mon_docker_subnet: "{{ public_network }}" -openstack_config: True dashboard_enabled: False public_network: "192.168.17.0/24" cluster_network: "192.168.18.0/24" diff --git a/tests/functional/all_daemons/container/group_vars/all b/tests/functional/all_daemons/container/group_vars/all index 4c45ec9c7d..7645c1ce8e 100644 --- a/tests/functional/all_daemons/container/group_vars/all +++ b/tests/functional/all_daemons/container/group_vars/all @@ -17,18 +17,6 @@ ceph_conf_overrides: mon_warn_on_pool_no_redundancy: false osd_pool_default_size: 1 mon_max_pg_per_osd: 300 -openstack_config: True -openstack_glance_pool: - name: "images" - size: 1 - target_size_ratio: 0.2 -openstack_cinder_pool: - name: "volumes" - rule_name: "HDD" - size: 1 -openstack_pools: - - "{{ openstack_glance_pool }}" - - "{{ openstack_cinder_pool }}" docker_pull_timeout: 600s handler_health_mon_check_delay: 10 handler_health_osd_check_delay: 10 diff --git a/tests/functional/all_daemons/group_vars/all b/tests/functional/all_daemons/group_vars/all index 63104689e7..f11618ce9a 100644 --- a/tests/functional/all_daemons/group_vars/all +++ b/tests/functional/all_daemons/group_vars/all @@ -11,20 +11,6 @@ ceph_conf_overrides: mon_warn_on_pool_no_redundancy: false osd_pool_default_size: 1 mon_max_pg_per_osd: 300 -openstack_config: True -openstack_glance_pool: - name: "images" - size: 1 - application: rbd - target_size_ratio: 0.2 -openstack_cinder_pool: - name: "volumes" - rule_name: "HDD" - size: 1 - application: rbd -openstack_pools: - - "{{ openstack_glance_pool }}" - - "{{ openstack_cinder_pool }}" handler_health_mon_check_delay: 10 handler_health_osd_check_delay: 10 mds_max_mds: 2 diff --git a/tests/functional/all_daemons_ipv6/container/group_vars/all b/tests/functional/all_daemons_ipv6/container/group_vars/all index 013fa975e2..f50a51cb2f 100644 --- a/tests/functional/all_daemons_ipv6/container/group_vars/all +++ b/tests/functional/all_daemons_ipv6/container/group_vars/all @@ -18,18 +18,6 @@ ceph_conf_overrides: mon_warn_on_pool_no_redundancy: false osd_pool_default_size: 1 mon_max_pg_per_osd: 300 -openstack_config: True -openstack_glance_pool: - name: "images" - size: 1 - target_size_ratio: 0.2 -openstack_cinder_pool: - name: "volumes" - rule_name: "HDD" - size: 1 -openstack_pools: - - "{{ openstack_glance_pool }}" - - "{{ openstack_cinder_pool }}" docker_pull_timeout: 600s handler_health_mon_check_delay: 10 handler_health_osd_check_delay: 10 diff --git a/tests/functional/all_daemons_ipv6/group_vars/all b/tests/functional/all_daemons_ipv6/group_vars/all index 204219f19e..ab16bb789c 100644 --- a/tests/functional/all_daemons_ipv6/group_vars/all +++ b/tests/functional/all_daemons_ipv6/group_vars/all @@ -12,20 +12,6 @@ ceph_conf_overrides: mon_warn_on_pool_no_redundancy: false osd_pool_default_size: 1 mon_max_pg_per_osd: 300 -openstack_config: True -openstack_glance_pool: - name: "images" - size: 1 - application: rbd - target_size_ratio: 0.2 -openstack_cinder_pool: - name: "volumes" - rule_name: "HDD" - size: 1 - application: rbd -openstack_pools: - - "{{ openstack_glance_pool }}" - - "{{ openstack_cinder_pool }}" handler_health_mon_check_delay: 10 handler_health_osd_check_delay: 10 mds_max_mds: 2 diff --git a/tests/functional/docker2podman/group_vars/all b/tests/functional/docker2podman/group_vars/all index 84d3dac2b3..b58d30a5d5 100644 --- a/tests/functional/docker2podman/group_vars/all +++ b/tests/functional/docker2podman/group_vars/all @@ -16,18 +16,6 @@ ceph_conf_overrides: mon_allow_pool_size_one: true mon_warn_on_pool_no_redundancy: false osd_pool_default_size: 1 -openstack_config: False -openstack_glance_pool: - name: "images" - rule_name: "HDD" - size: 1 -openstack_cinder_pool: - name: "volumes" - rule_name: "HDD" - size: 1 -openstack_pools: - - "{{ openstack_glance_pool }}" - - "{{ openstack_cinder_pool }}" handler_health_mon_check_delay: 10 handler_health_osd_check_delay: 10 dashboard_admin_password: $sX!cD$rYU6qR^B! @@ -38,4 +26,4 @@ ceph_docker_image_tag: latest-main node_exporter_container_image: "quay.io/prometheus/node-exporter:v0.17.0" prometheus_container_image: "quay.io/prometheus/prometheus:v2.7.2" alertmanager_container_image: "quay.io/prometheus/alertmanager:v0.16.2" -grafana_container_image: "quay.io/ceph/ceph-grafana:6.7.4" \ No newline at end of file +grafana_container_image: "quay.io/ceph/ceph-grafana:6.7.4" diff --git a/tests/functional/external_clients/container/inventory/group_vars/all b/tests/functional/external_clients/container/inventory/group_vars/all index 0bf378374b..20cd47ae85 100644 --- a/tests/functional/external_clients/container/inventory/group_vars/all +++ b/tests/functional/external_clients/container/inventory/group_vars/all @@ -11,7 +11,6 @@ public_network: "192.168.31.0/24" cluster_network: "192.168.32.0/24" rgw_override_bucket_index_max_shards: 16 rgw_bucket_default_quota_max_objects: 1638400 -openstack_config: True dashboard_enabled: false ceph_conf_overrides: global: diff --git a/tests/functional/external_clients/inventory/group_vars/all b/tests/functional/external_clients/inventory/group_vars/all index 1152b04cef..29c0ed43ba 100644 --- a/tests/functional/external_clients/inventory/group_vars/all +++ b/tests/functional/external_clients/inventory/group_vars/all @@ -4,7 +4,6 @@ ceph_origin: repository ceph_repository: community radosgw_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" ceph_mon_docker_subnet: "{{ public_network }}" -openstack_config: True dashboard_enabled: False public_network: "192.168.31.0/24" cluster_network: "192.168.32.0/24" diff --git a/tests/functional/lvm-osds/container/group_vars/all b/tests/functional/lvm-osds/container/group_vars/all index 225a5cffe0..c9786a4dd8 100644 --- a/tests/functional/lvm-osds/container/group_vars/all +++ b/tests/functional/lvm-osds/container/group_vars/all @@ -21,20 +21,6 @@ ceph_conf_overrides: dashboard_enabled: False handler_health_mon_check_delay: 10 handler_health_osd_check_delay: 10 -openstack_config: True -openstack_glance_pool: - name: "images" - type: 3 - size: 1 - application: rbd - target_size_ratio: 0.2 -openstack_cinder_pool: - name: "volumes" - size: 1 - application: rbd -openstack_pools: - - "{{ openstack_glance_pool }}" - - "{{ openstack_cinder_pool }}" ceph_docker_registry: quay.io ceph_docker_image: ceph/daemon-base -ceph_docker_image_tag: latest-main \ No newline at end of file +ceph_docker_image_tag: latest-main diff --git a/tests/functional/lvm-osds/group_vars/all b/tests/functional/lvm-osds/group_vars/all index c8066f0132..c39cfd433d 100644 --- a/tests/functional/lvm-osds/group_vars/all +++ b/tests/functional/lvm-osds/group_vars/all @@ -16,17 +16,4 @@ ceph_conf_overrides: dashboard_enabled: False handler_health_mon_check_delay: 10 handler_health_osd_check_delay: 10 -openstack_config: True -openstack_glance_pool: - name: "images" - type: 3 - size: 1 - application: rbd - target_size_ratio: 0.2 -openstack_cinder_pool: - name: "volumes" - size: 1 - application: rbd -openstack_pools: - - "{{ openstack_glance_pool }}" - - "{{ openstack_cinder_pool }}" \ No newline at end of file + diff --git a/tests/functional/ooo-collocation/hosts b/tests/functional/ooo-collocation/hosts index 812bd8c26a..327f7b0232 100644 --- a/tests/functional/ooo-collocation/hosts +++ b/tests/functional/ooo-collocation/hosts @@ -37,7 +37,6 @@ all: monitor_address_block: 192.168.95.0/24 monitor_secret: AQBSV4xaAAAAABAALqm4vRHcITs4/041TwluMg== ntp_service_enabled: false - openstack_config: true openstack_keys: - {key: AQAN0RdbAAAAABAA3CpSKRVDrENjkOSunEFZ0A==, mgr_cap: 'allow *', mode: '0600', mon_cap: 'allow r', name: client.openstack, osd_cap: "allow class-read object_prefix rbd_children, allow rwx pool=volumes, allow rwx pool=backups, allow rwx pool=vms, allow rwx pool=images, allow rwx pool=metrics"} - {key: AQAN0RdbAAAAABAAtV5Dq28z4H6XxwhaNEaFZg==, mds_cap: 'allow *', mgr_cap: 'allow *', mode: '0600', mon_cap: 'allow r, allow command "auth del", allow command "auth caps", allow command "auth get", allow command "auth get-or-create"', name: client.manila, osd_cap: 'allow rw'} diff --git a/tests/functional/podman/group_vars/all b/tests/functional/podman/group_vars/all index 29709d562c..5ffd6f0f05 100644 --- a/tests/functional/podman/group_vars/all +++ b/tests/functional/podman/group_vars/all @@ -15,18 +15,6 @@ ceph_conf_overrides: mon_allow_pool_size_one: true mon_warn_on_pool_no_redundancy: false osd_pool_default_size: 1 -openstack_config: True -openstack_glance_pool: - name: "images" - rule_name: "HDD" - size: 1 -openstack_cinder_pool: - name: "volumes" - rule_name: "HDD" - size: 1 -openstack_pools: - - "{{ openstack_glance_pool }}" - - "{{ openstack_cinder_pool }}" handler_health_mon_check_delay: 10 handler_health_osd_check_delay: 10 dashboard_admin_password: $sX!cD$rYU6qR^B! @@ -37,4 +25,4 @@ ceph_docker_image_tag: latest-main node_exporter_container_image: "quay.io/prometheus/node-exporter:v0.17.0" prometheus_container_image: "quay.io/prometheus/prometheus:v2.7.2" alertmanager_container_image: "quay.io/prometheus/alertmanager:v0.16.2" -grafana_container_image: "quay.io/ceph/ceph-grafana:6.7.4" \ No newline at end of file +grafana_container_image: "quay.io/ceph/ceph-grafana:6.7.4" diff --git a/tests/functional/shrink_mds/container/group_vars/all b/tests/functional/shrink_mds/container/group_vars/all index e0cc1865e7..4eb3151d08 100644 --- a/tests/functional/shrink_mds/container/group_vars/all +++ b/tests/functional/shrink_mds/container/group_vars/all @@ -12,7 +12,6 @@ ceph_conf_overrides: mon_allow_pool_size_one: true mon_warn_on_pool_no_redundancy: false osd_pool_default_size: 1 -openstack_config: False dashboard_enabled: False copy_admin_key: True ceph_docker_registry: quay.io diff --git a/tests/functional/shrink_mgr/container/group_vars/all b/tests/functional/shrink_mgr/container/group_vars/all index 671bc426b4..8ddb84b9a8 100644 --- a/tests/functional/shrink_mgr/container/group_vars/all +++ b/tests/functional/shrink_mgr/container/group_vars/all @@ -12,7 +12,6 @@ ceph_conf_overrides: mon_allow_pool_size_one: true mon_warn_on_pool_no_redundancy: false osd_pool_default_size: 1 -openstack_config: False dashboard_enabled: False ceph_docker_registry: quay.io ceph_docker_image: ceph/daemon-base diff --git a/tests/functional/shrink_mon/container/group_vars/all b/tests/functional/shrink_mon/container/group_vars/all index 7415150db5..d1e5d2182f 100644 --- a/tests/functional/shrink_mon/container/group_vars/all +++ b/tests/functional/shrink_mon/container/group_vars/all @@ -12,7 +12,6 @@ ceph_conf_overrides: mon_allow_pool_size_one: true mon_warn_on_pool_no_redundancy: false osd_pool_default_size: 1 -openstack_config: False dashboard_enabled: False ceph_docker_registry: quay.io ceph_docker_image: ceph/daemon-base diff --git a/tests/functional/shrink_osd/container/group_vars/all b/tests/functional/shrink_osd/container/group_vars/all index 1ae9358193..5c56a1c18d 100644 --- a/tests/functional/shrink_osd/container/group_vars/all +++ b/tests/functional/shrink_osd/container/group_vars/all @@ -12,7 +12,6 @@ ceph_conf_overrides: mon_allow_pool_size_one: true mon_warn_on_pool_no_redundancy: false osd_pool_default_size: 1 -openstack_config: False dashboard_enabled: False copy_admin_key: True ceph_docker_registry: quay.io diff --git a/tests/functional/shrink_osd/group_vars/all b/tests/functional/shrink_osd/group_vars/all index d6999b17ee..956f325f85 100644 --- a/tests/functional/shrink_osd/group_vars/all +++ b/tests/functional/shrink_osd/group_vars/all @@ -6,6 +6,5 @@ cluster_network: "192.168.72.0/24" ceph_conf_overrides: global: osd_pool_default_size: 3 -openstack_config: False dashboard_enabled: False copy_admin_key: True \ No newline at end of file diff --git a/tests/functional/shrink_rbdmirror/container/group_vars/all b/tests/functional/shrink_rbdmirror/container/group_vars/all index 3fe10aa207..0ec43c7533 100644 --- a/tests/functional/shrink_rbdmirror/container/group_vars/all +++ b/tests/functional/shrink_rbdmirror/container/group_vars/all @@ -11,7 +11,6 @@ ceph_conf_overrides: mon_allow_pool_size_one: true mon_warn_on_pool_no_redundancy: false osd_pool_default_size: 1 -openstack_config: False dashboard_enabled: False copy_admin_key: True ceph_docker_registry: quay.io diff --git a/tests/functional/shrink_rgw/container/group_vars/all b/tests/functional/shrink_rgw/container/group_vars/all index 9aafb5541b..8b918527f2 100644 --- a/tests/functional/shrink_rgw/container/group_vars/all +++ b/tests/functional/shrink_rgw/container/group_vars/all @@ -13,7 +13,6 @@ ceph_conf_overrides: mon_allow_pool_size_one: true mon_warn_on_pool_no_redundancy: false osd_pool_default_size: 1 -openstack_config: False dashboard_enabled: False copy_admin_key: True ceph_docker_registry: quay.io diff --git a/tests/functional/subset_update/container/group_vars/all b/tests/functional/subset_update/container/group_vars/all index ad86c04010..e0085ed78c 100644 --- a/tests/functional/subset_update/container/group_vars/all +++ b/tests/functional/subset_update/container/group_vars/all @@ -17,7 +17,6 @@ ceph_conf_overrides: mon_warn_on_pool_no_redundancy: false osd_pool_default_size: 1 mon_max_pg_per_osd: 300 -openstack_config: false docker_pull_timeout: 600s handler_health_mon_check_delay: 10 handler_health_osd_check_delay: 10 diff --git a/tests/functional/subset_update/group_vars/all b/tests/functional/subset_update/group_vars/all index 03d44a831f..e1bf727f32 100644 --- a/tests/functional/subset_update/group_vars/all +++ b/tests/functional/subset_update/group_vars/all @@ -11,7 +11,6 @@ ceph_conf_overrides: mon_warn_on_pool_no_redundancy: false osd_pool_default_size: 1 mon_max_pg_per_osd: 300 -openstack_config: false handler_health_mon_check_delay: 10 handler_health_osd_check_delay: 10 mds_max_mds: 2