From b20acab937a682c2e19f65a12e000cd52151388f Mon Sep 17 00:00:00 2001 From: Sanjay Chari Date: Wed, 9 Nov 2022 11:44:44 +0530 Subject: [PATCH] Remove option all in dynamic workloads Dynamic workloads has an option all in browbeat-config.yaml, which doesn't really makes sense because it runs only VM and trunk dynamic workloads. This patch replaces all with an option called all_vm_and_trunk, for better readability. Change-Id: I9e3a73a9fb98fe46c830ad56b6a3cdd4a8c2b863 --- browbeat-config.yaml | 12 ++++----- .../dynamic-workloads/dynamic_workload.py | 26 ++++++++++++------- 2 files changed, 22 insertions(+), 16 deletions(-) diff --git a/browbeat-config.yaml b/browbeat-config.yaml index 0ec7745fb..81a186658 100644 --- a/browbeat-config.yaml +++ b/browbeat-config.yaml @@ -709,23 +709,23 @@ workloads: # created as part of rally context for dynamic workloads. These external # networks will be used in a round robin fashion by the iterations. num_external_networks: 16 - # workloads can be 'all', a single workload(Eg. : create_delete_servers), + # workloads can contain a single workload(Eg. : create_delete_servers), # or a comma separated string(Eg. : create_delete_servers,migrate_servers). # Currently supported workloads : create_delete_servers, migrate_servers # swap_floating_ips_between_servers, stop_start_servers, create_loadbalancers, # delete_loadbalancers, delete_members_random_lb, pod_fip_simulation, # add_subports_to_random_trunks, delete_subports_from_random_trunks, - # swap_floating_ips_between_random_subports, provider_netcreate_nova_boot_ping, - # provider_net_nova_boot_ping, provider_net_nova_delete, e2e_kube_burner, ocp_on_osp - # Note: Octavia, Provider, e2e_kube_burner and ocp_on_osp scenarios are not included in 'all' - # by default, and have to be included separately. + # swap_floating_ips_between_random_subports, + # all_vm_and_trunk(will run all vm and trunk dynamic workloads), + # provider_netcreate_nova_boot_ping, provider_net_nova_boot_ping, provider_net_nova_delete, + # e2e_kube_burner, ocp_on_osp # Steps for running ocp_on_osp workload, given that underlying openstack has been # deployed with necessary dependencies: # 1) Pass your pull secret in ocp_on_osp/vars/shift_stack_vars.yaml # 2) If you want to change any default params(like ocp_version, worker count, master and worker flavors, etc), refer to ocp_on_osp/vars/shift_stack_vars.yml # 3) Run "ansible-playbook -vvv ocp_on_osp/ocp_bootstrap.yml" # 4) And then run ocp_on_osp workload. NOTE: use "ansible-playbook -vvv ocp_on_osp/ocp_cleanup.yml" for the clean up. - workloads: all + workloads: all_vm_and_trunk file: rally/rally-plugins/dynamic-workloads/dynamic_workload.yml - name: dynamic-workloads-min diff --git a/rally/rally-plugins/dynamic-workloads/dynamic_workload.py b/rally/rally-plugins/dynamic-workloads/dynamic_workload.py index 6ea2b1807..84863f694 100644 --- a/rally/rally-plugins/dynamic-workloads/dynamic_workload.py +++ b/rally/rally-plugins/dynamic-workloads/dynamic_workload.py @@ -79,7 +79,9 @@ class DynamicWorkload(vm.VMDynamicScenario, trunk.TrunkDynamicScenario, self.security_group = self.create_sec_group_with_icmp_ssh() self.log_info("security group {} created for this iteration".format(self.security_group)) - if(workloads == "all" or "migrate_servers" in workloads_list or + run_all_vm_and_trunk_workloads = "all_vm_and_trunk" in workloads_list + + if(run_all_vm_and_trunk_workloads or "migrate_servers" in workloads_list or "swap_floating_ips_between_servers" in workloads_list or "stop_start_servers" in workloads_list or "pod_fip_simulation" in workloads_list or @@ -103,38 +105,42 @@ class DynamicWorkload(vm.VMDynamicScenario, trunk.TrunkDynamicScenario, except AttributeError: DynamicWorkload.browbeat_dir = os.getcwd() - if workloads == "all" or "create_delete_servers" in workloads_list: + if run_all_vm_and_trunk_workloads or "create_delete_servers" in workloads_list: self.boot_servers(smallest_image, smallest_flavor, num_create_vms, subnet_create_args=subnet_create_args) self.delete_random_servers(num_delete_vms) - if(workloads == "all" or "migrate_servers" in workloads_list or + if(run_all_vm_and_trunk_workloads or "migrate_servers" in workloads_list or "swap_floating_ips_between_servers" in workloads_list or "stop_start_servers" in workloads_list): self.boot_servers_with_fip(smallest_image, smallest_flavor, context_ext_net_id, num_vms_to_create_with_fip, network_create_args, subnet_create_args, **kwargs) - if workloads == "all" or "migrate_servers" in workloads_list: + if run_all_vm_and_trunk_workloads or "migrate_servers" in workloads_list: self.migrate_servers_with_fip(num_vms_to_migrate) - if workloads == "all" or "swap_floating_ips_between_servers" in workloads_list: + if(run_all_vm_and_trunk_workloads or + "swap_floating_ips_between_servers" in workloads_list): self.swap_floating_ips_between_servers() - if workloads == "all" or "stop_start_servers" in workloads_list: + if run_all_vm_and_trunk_workloads or "stop_start_servers" in workloads_list: self.stop_start_servers_with_fip(num_stop_start_vms) - if workloads == "all" or "pod_fip_simulation" in workloads_list: + if run_all_vm_and_trunk_workloads or "pod_fip_simulation" in workloads_list: self.pod_fip_simulation(context_ext_net_id, trunk_image, trunk_flavor, smallest_image, smallest_flavor, num_initial_subports, num_trunk_vms) - if workloads == "all" or "add_subports_to_random_trunks" in workloads_list: + if(run_all_vm_and_trunk_workloads or + "add_subports_to_random_trunks" in workloads_list): self.add_subports_to_random_trunks(num_add_subports_trunks, num_add_subports) - if workloads == "all" or "delete_subports_from_random_trunks" in workloads_list: + if(run_all_vm_and_trunk_workloads or + "delete_subports_from_random_trunks" in workloads_list): self.delete_subports_from_random_trunks(num_delete_subports_trunks, num_delete_subports) - if workloads == "all" or "swap_floating_ips_between_random_subports" in workloads_list: + if(run_all_vm_and_trunk_workloads or + "swap_floating_ips_between_random_subports" in workloads_list): self.swap_floating_ips_between_random_subports() if "create_loadbalancers" in workloads_list: