diff --git a/ansible/roles/aodh/templates/aodh.conf.j2 b/ansible/roles/aodh/templates/aodh.conf.j2 index 3664309f17..cbd03db031 100644 --- a/ansible/roles/aodh/templates/aodh.conf.j2 +++ b/ansible/roles/aodh/templates/aodh.conf.j2 @@ -32,7 +32,7 @@ cafile = {{ openstack_cacert }} region_name = {{ openstack_region_name }} [oslo_middleware] -enable_proxy_headers_parsing = True +enable_proxy_headers_parsing = true {% if aodh_policy_file is defined %} [oslo_policy] diff --git a/ansible/roles/barbican/templates/barbican.conf.j2 b/ansible/roles/barbican/templates/barbican.conf.j2 index afc5b98a8b..90ce5e9031 100644 --- a/ansible/roles/barbican/templates/barbican.conf.j2 +++ b/ansible/roles/barbican/templates/barbican.conf.j2 @@ -11,7 +11,7 @@ host_href = {{ barbican_public_endpoint }} backlog = 4096 -db_auto_create = False +db_auto_create = false transport_url = {{ rpc_transport_url }} @@ -52,7 +52,7 @@ kek = '{{ barbican_crypto_key }}' [keystone_notifications] -enable = True +enable = true {% if enable_keystone | bool %} topic = barbican_notifications {% endif %} @@ -97,7 +97,7 @@ rabbit_quorum_queue = true {% endif %} [oslo_middleware] -enable_proxy_headers_parsing = True +enable_proxy_headers_parsing = true {% if barbican_policy_file is defined %} [oslo_policy] diff --git a/ansible/roles/blazar/templates/blazar.conf.j2 b/ansible/roles/blazar/templates/blazar.conf.j2 index b2ce275775..df438a62fd 100644 --- a/ansible/roles/blazar/templates/blazar.conf.j2 +++ b/ansible/roles/blazar/templates/blazar.conf.j2 @@ -29,7 +29,7 @@ user_domain_id = default project_name = service username = {{ blazar_keystone_user }} password = {{ blazar_keystone_password }} -service_token_roles_required = True +service_token_roles_required = true cafile = {{ openstack_cacert }} region_name = {{ openstack_region_name }} diff --git a/ansible/roles/ceilometer/templates/ceilometer.conf.j2 b/ansible/roles/ceilometer/templates/ceilometer.conf.j2 index 404afc7ae3..f03100e3ec 100644 --- a/ansible/roles/ceilometer/templates/ceilometer.conf.j2 +++ b/ansible/roles/ceilometer/templates/ceilometer.conf.j2 @@ -55,7 +55,7 @@ policy_file = {{ ceilometer_policy_file }} [cache] backend = oslo_cache.memcache_pool -enabled = True +enabled = true memcache_servers = {% for host in groups['memcached'] %}{{ 'api' | kolla_address(host) | put_address_in_context('memcache') }}:{{ memcached_port }}{% if not loop.last %},{% endif %}{% endfor %} [oslo_concurrency] diff --git a/ansible/roles/cinder/templates/cinder.conf.j2 b/ansible/roles/cinder/templates/cinder.conf.j2 index 649e05f357..732db76a54 100644 --- a/ansible/roles/cinder/templates/cinder.conf.j2 +++ b/ansible/roles/cinder/templates/cinder.conf.j2 @@ -7,9 +7,9 @@ log_file = cinder-api.log {% endif %} use_forwarded_for = true -# Set use_stderr to False or the logs will also be sent to stderr +# Set use_stderr to false or the logs will also be sent to stderr # and collected by Docker -use_stderr = False +use_stderr = false my_ip = {{ api_interface_address }} @@ -92,7 +92,7 @@ rabbit_quorum_queue = true {% endif %} [oslo_middleware] -enable_proxy_headers_parsing = True +enable_proxy_headers_parsing = true {% if cinder_policy_file is defined %} [oslo_policy] @@ -164,7 +164,7 @@ rbd_user = {{ backend.user }} rbd_cluster_name = {{ backend.cluster }} rbd_keyring_conf = /etc/ceph/{{ backend.cluster }}.client.{{ backend.user }}.keyring rbd_secret_uuid = {{ cinder_rbd_secret_uuid }} -report_discard_supported = True +report_discard_supported = true {% if backend.availability_zone is defined %} backend_availability_zone = {{ backend.availability_zone }} {% endif %} @@ -177,9 +177,9 @@ backend_availability_zone = {{ backend.availability_zone }} volume_driver = cinder.volume.drivers.nfs.NfsDriver volume_backend_name = {{ cinder_backend_nfs_name }} nfs_shares_config = /etc/cinder/nfs_shares -nfs_snapshot_support = True -nas_secure_file_permissions = False -nas_secure_file_operations = False +nfs_snapshot_support = true +nas_secure_file_permissions = false +nas_secure_file_operations = false {% endif %} {% if cinder_backend_vmwarevc_vmdk | bool %} @@ -189,7 +189,7 @@ vmware_host_ip = {{ vmware_vcenter_host_ip }} vmware_host_username = {{ vmware_vcenter_host_username }} vmware_host_password = {{ vmware_vcenter_host_password }} vmware_cluster_name = {{ vmware_vcenter_cluster_name }} -vmware_insecure = True +vmware_insecure = true {% endif %} {% if cinder_backend_vmware_vstorage_object | bool %} @@ -199,7 +199,7 @@ vmware_host_ip = {{ vmware_vcenter_host_ip }} vmware_host_username = {{ vmware_vcenter_host_username }} vmware_host_password = {{ vmware_vcenter_host_password }} vmware_cluster_name = {{ vmware_vcenter_cluster_name }} -vmware_insecure = True +vmware_insecure = true {% endif %} {% if enable_cinder_backend_quobyte | bool %} diff --git a/ansible/roles/cloudkitty/templates/cloudkitty.conf.j2 b/ansible/roles/cloudkitty/templates/cloudkitty.conf.j2 index 71fd481ea8..413eb5c9c6 100644 --- a/ansible/roles/cloudkitty/templates/cloudkitty.conf.j2 +++ b/ansible/roles/cloudkitty/templates/cloudkitty.conf.j2 @@ -34,7 +34,7 @@ memcache_secret_key = {{ memcache_secret_key }} memcached_servers = {% for host in groups['memcached'] %}{{ 'api' | kolla_address(host) | put_address_in_context('memcache') }}:{{ memcached_port }}{% if not loop.last %},{% endif %}{% endfor %} [oslo_middleware] -enable_proxy_headers_parsing = True +enable_proxy_headers_parsing = true [oslo_concurrency] lock_path = /var/lib/cloudkitty/tmp diff --git a/ansible/roles/designate/templates/designate.conf.j2 b/ansible/roles/designate/templates/designate.conf.j2 index 4fce643e7a..27cc901e43 100644 --- a/ansible/roles/designate/templates/designate.conf.j2 +++ b/ansible/roles/designate/templates/designate.conf.j2 @@ -14,8 +14,8 @@ workers = {{ designate_central_workers }} listen = {{ api_interface_address | put_address_in_context('url') }}:{{ designate_api_listen_port }} api_base_uri = {{ designate_internal_endpoint }} workers = {{ designate_api_workers }} -enable_api_admin = True -enable_host_header = True +enable_api_admin = true +enable_host_header = true enabled_extensions_admin = quotas, reports [keystone_authtoken] @@ -29,7 +29,7 @@ project_name = service username = {{ designate_keystone_user }} password = {{ designate_keystone_password }} http_connect_timeout = 60 -service_token_roles_required = True +service_token_roles_required = true cafile = {{ openstack_cacert }} region_name = {{ openstack_region_name }} @@ -109,7 +109,7 @@ rabbit_quorum_queue = true lock_path = /var/lib/designate/tmp [oslo_middleware] -enable_proxy_headers_parsing = True +enable_proxy_headers_parsing = true {% if designate_policy_file is defined %} [oslo_policy] diff --git a/ansible/roles/glance/templates/glance-api.conf.j2 b/ansible/roles/glance/templates/glance-api.conf.j2 index 03ecb21203..b4766cb9b9 100644 --- a/ansible/roles/glance/templates/glance-api.conf.j2 +++ b/ansible/roles/glance/templates/glance-api.conf.j2 @@ -78,12 +78,12 @@ rbd_store_ceph_conf = /etc/ceph/{{ backend.cluster }}.conf [swift] swift_store_container = glance swift_store_multiple_containers_seed = 0 -swift_store_multi_tenant = False -swift_store_create_container_on_put = True +swift_store_multi_tenant = false +swift_store_create_container_on_put = true swift_store_region = {{ openstack_region_name }} default_swift_reference = swift swift_store_config_file = /etc/glance/glance-swift.conf -swift_store_auth_insecure = True +swift_store_auth_insecure = true {% endif %} {% if glance_backend_s3 | bool %} @@ -100,7 +100,7 @@ vmware_server_host = {{ vmware_vcenter_host_ip }} vmware_server_username = {{ vmware_vcenter_host_username }} vmware_server_password = {{ vmware_vcenter_host_password }} vmware_datastores = {{ vmware_vcenter_name }}:{{ vmware_datastore_name }} -vmware_insecure = True +vmware_insecure = true {% endif %} [os_glance_tasks_store] @@ -110,7 +110,7 @@ filesystem_store_datadir = /var/lib/glance/tasks_work_dir filesystem_store_datadir = /var/lib/glance/staging [oslo_middleware] -enable_proxy_headers_parsing = True +enable_proxy_headers_parsing = true [oslo_concurrency] lock_path = /var/lib/glance/tmp diff --git a/ansible/roles/gnocchi/templates/gnocchi.conf.j2 b/ansible/roles/gnocchi/templates/gnocchi.conf.j2 index fab81554e2..aaf33c579e 100644 --- a/ansible/roles/gnocchi/templates/gnocchi.conf.j2 +++ b/ansible/roles/gnocchi/templates/gnocchi.conf.j2 @@ -19,7 +19,7 @@ middlewares = keystonemiddleware.auth_token.AuthProtocol auth_mode = keystone [oslo_middleware] -enable_proxy_headers_parsing = True +enable_proxy_headers_parsing = true [database] connection = mysql+pymysql://{{ gnocchi_database_user }}:{{ gnocchi_database_password }}@{{ gnocchi_database_address }}/{{ gnocchi_database_name }} diff --git a/ansible/roles/heat/templates/heat.conf.j2 b/ansible/roles/heat/templates/heat.conf.j2 index 0bf5e331c4..793c42d10e 100644 --- a/ansible/roles/heat/templates/heat.conf.j2 +++ b/ansible/roles/heat/templates/heat.conf.j2 @@ -48,7 +48,7 @@ memcached_servers = {% for host in groups['memcached'] %}{{ 'api' | kolla_addres [cache] backend = oslo_cache.memcache_pool -enabled = True +enabled = true memcache_servers = {% for host in groups['memcached'] %}{{ 'api' | kolla_address(host) | put_address_in_context('memcache') }}:{{ memcached_port }}{% if not loop.last %},{% endif %}{% endfor %} @@ -94,7 +94,7 @@ endpoint_type = internalURL ca_file = {{ openstack_cacert }} [oslo_middleware] -enable_proxy_headers_parsing = True +enable_proxy_headers_parsing = true {% if enable_osprofiler | bool %} [profiler] diff --git a/ansible/roles/ironic/templates/ironic.conf.j2 b/ansible/roles/ironic/templates/ironic.conf.j2 index f4a1fc82a0..94096f6ff4 100644 --- a/ansible/roles/ironic/templates/ironic.conf.j2 +++ b/ansible/roles/ironic/templates/ironic.conf.j2 @@ -18,7 +18,7 @@ my_ip = {{ api_interface_address }} notification_level = info {% endif %} -rbac_service_role_elevated_access = True +rbac_service_role_elevated_access = true [oslo_messaging_notifications] transport_url = {{ notify_transport_url }} @@ -217,7 +217,7 @@ http_root = /var/lib/ironic/httpboot http_url = {{ ironic_http_url }} [oslo_middleware] -enable_proxy_headers_parsing = True +enable_proxy_headers_parsing = true {% if not enable_neutron | bool %} [dhcp] diff --git a/ansible/roles/keystone/templates/keystone.conf.j2 b/ansible/roles/keystone/templates/keystone.conf.j2 index b418e4e568..af43cfc939 100644 --- a/ansible/roles/keystone/templates/keystone.conf.j2 +++ b/ansible/roles/keystone/templates/keystone.conf.j2 @@ -5,10 +5,10 @@ transport_url = {{ rpc_transport_url }} # NOTE(elemoine) log_dir alone does not work for Keystone log_file = /var/log/kolla/keystone/keystone.log -use_stderr = True +use_stderr = true [oslo_middleware] -enable_proxy_headers_parsing = True +enable_proxy_headers_parsing = true {% if keystone_policy_file is defined %} [oslo_policy] @@ -28,7 +28,7 @@ domain_config_dir = /etc/keystone/domains {% endif %} [token] -revoke_by_id = False +revoke_by_id = false provider = fernet expiration = {{ fernet_token_expiry }} allow_expired_window = {{ fernet_token_allow_expired_window }} @@ -47,7 +47,7 @@ max_active_keys = {{ ((fernet_token_expiry | int + [cache] backend = oslo_cache.memcache_pool -enabled = True +enabled = true memcache_servers = {% for host in groups['memcached'] %}{{ 'api' | kolla_address(host) | put_address_in_context('memcache') }}:{{ memcached_port }}{% if not loop.last %},{% endif %}{% endfor %} [oslo_messaging_notifications] diff --git a/ansible/roles/magnum/templates/magnum.conf.j2 b/ansible/roles/magnum/templates/magnum.conf.j2 index 1e696664a5..c9c720af8a 100644 --- a/ansible/roles/magnum/templates/magnum.conf.j2 +++ b/ansible/roles/magnum/templates/magnum.conf.j2 @@ -108,7 +108,7 @@ cluster_user_trust = {{ enable_cluster_user_trust }} lock_path = /var/lib/magnum/tmp [oslo_middleware] -enable_proxy_headers_parsing = True +enable_proxy_headers_parsing = true [certificates] {% if enable_barbican | bool %} diff --git a/ansible/roles/manila/templates/manila-share.conf.j2 b/ansible/roles/manila/templates/manila-share.conf.j2 index 2a4a48f364..5bae7eedd6 100644 --- a/ansible/roles/manila/templates/manila-share.conf.j2 +++ b/ansible/roles/manila/templates/manila-share.conf.j2 @@ -78,7 +78,7 @@ interface_driver = manila.network.linux.interface.OVSInterfaceDriver interface_driver = manila.network.linux.interface.BridgeInterfaceDriver {% endif %} -driver_handles_share_servers = True +driver_handles_share_servers = true service_instance_password = {{ manila_service_instance_password }} service_instance_user = {{ manila_service_instance_user }} @@ -96,7 +96,7 @@ policy_file = {{ manila_policy_file }} [hnas1] share_backend_name = HNAS1 share_driver = manila.share.drivers.hitachi.hnas.driver.HitachiHNASDriver -driver_handles_share_servers = False +driver_handles_share_servers = false hitachi_hnas_ip = {{ hnas_ip }} hitachi_hnas_user = {{ hnas_user }} hitachi_hnas_password = {{ hnas_password }} @@ -109,7 +109,7 @@ hitachi_hnas_file_system_name = {{ hnas_file_system_name }} {% for backend in manila_ceph_backends %} {% if backend.driver == 'cephfsnative' %} [{{ backend.name }}] -driver_handles_share_servers = False +driver_handles_share_servers = false share_backend_name = {{ backend.share_name }} share_driver = manila.share.drivers.cephfs.driver.CephFSDriver cephfs_conf_path = /etc/ceph/{{ backend.cluster }}.conf @@ -126,7 +126,7 @@ cephfs_filesystem_name = {{ manila_cephfs_filesystem_name }} {% for backend in manila_ceph_backends %} {% if backend.driver == 'cephfsnfs' %} [{{ backend.name }}] -driver_handles_share_servers = False +driver_handles_share_servers = false share_backend_name = {{ backend.share_name }} share_driver = manila.share.drivers.cephfs.driver.CephFSDriver cephfs_protocol_helper_type = NFS @@ -136,7 +136,7 @@ cephfs_cluster_name = {{ backend.cluster }} {% if manila_cephfs_filesystem_name | length %} cephfs_filesystem_name = {{ manila_cephfs_filesystem_name }} {% endif %} -cephfs_ganesha_server_is_remote= False +cephfs_ganesha_server_is_remote= false cephfs_ganesha_server_ip = {{ api_interface_address }} {% endif %} {% endfor %} @@ -144,7 +144,7 @@ cephfs_ganesha_server_ip = {{ api_interface_address }} {% if enable_manila_backend_glusterfs_nfs | bool %} [glusterfsnfs1] -driver_handles_share_servers = False +driver_handles_share_servers = false share_backend_name = GLUSTERFSNFS1 share_driver = manila.share.drivers.glusterfs.GlusterfsShareDriver glusterfs_share_layout = {{ manila_glusterfs_share_layout }} diff --git a/ansible/roles/manila/templates/manila.conf.j2 b/ansible/roles/manila/templates/manila.conf.j2 index 2edf5b5ca2..da78f6a129 100644 --- a/ansible/roles/manila/templates/manila.conf.j2 +++ b/ansible/roles/manila/templates/manila.conf.j2 @@ -70,7 +70,7 @@ rabbit_quorum_queue = true {% endif %} [oslo_middleware] -enable_proxy_headers_parsing = True +enable_proxy_headers_parsing = true {% if manila_policy_file is defined %} [oslo_policy] diff --git a/ansible/roles/masakari/templates/masakari-monitors.conf.j2 b/ansible/roles/masakari/templates/masakari-monitors.conf.j2 index 7a4f573928..f52afd5037 100644 --- a/ansible/roles/masakari/templates/masakari-monitors.conf.j2 +++ b/ansible/roles/masakari/templates/masakari-monitors.conf.j2 @@ -26,13 +26,13 @@ connection_uri = "qemu+tcp://{{ migration_interface_address | put_address_in_con a better default choice. This limitation may be lifted in the near future (Xena+). #} -restrict_to_remotes = True +restrict_to_remotes = true {# NOTE(yoctozepto): ``disable_ipmi_check`` is due to ``restrict_to_remotes`` above. See https://bugs.launchpad.net/masakari-monitors/+bug/1933203 #} -disable_ipmi_check = True +disable_ipmi_check = true {% if inventory_hostname in groups['hacluster'] %} pacemaker_node_type = cluster corosync_multicast_interfaces = {{ api_interface }} diff --git a/ansible/roles/masakari/templates/masakari.conf.j2 b/ansible/roles/masakari/templates/masakari.conf.j2 index 03c6ac695b..b96e6c6d1a 100644 --- a/ansible/roles/masakari/templates/masakari.conf.j2 +++ b/ansible/roles/masakari/templates/masakari.conf.j2 @@ -31,7 +31,7 @@ user_domain_name = {{ default_user_domain_name }} project_name = service username = {{ masakari_keystone_user }} password = {{ masakari_keystone_password }} -service_token_roles_required = True +service_token_roles_required = true region_name = {{ openstack_region_name }} cafile = {{ openstack_cacert }} @@ -64,7 +64,7 @@ rabbit_quorum_queue = true {% endif %} [oslo_middleware] -enable_proxy_headers_parsing = True +enable_proxy_headers_parsing = true {% if masakari_policy_file is defined %} [oslo_policy] diff --git a/ansible/roles/mistral/templates/mistral.conf.j2 b/ansible/roles/mistral/templates/mistral.conf.j2 index 600e921a57..5ce7a4a0e5 100644 --- a/ansible/roles/mistral/templates/mistral.conf.j2 +++ b/ansible/roles/mistral/templates/mistral.conf.j2 @@ -10,9 +10,9 @@ log_file = /var/log/kolla/mistral/mistral-event-engine.log log_file = /var/log/kolla/mistral/mistral-executor.log {% endif %} -# NOTE(elemoine): set use_stderr to False or the logs will also be sent to +# NOTE(elemoine): set use_stderr to false or the logs will also be sent to # stderr and collected by Docker -use_stderr = False +use_stderr = false transport_url = {{ rpc_transport_url }} diff --git a/ansible/roles/neutron/templates/dhcp_agent.ini.j2 b/ansible/roles/neutron/templates/dhcp_agent.ini.j2 index 7b96582402..42d37e88f8 100644 --- a/ansible/roles/neutron/templates/dhcp_agent.ini.j2 +++ b/ansible/roles/neutron/templates/dhcp_agent.ini.j2 @@ -8,10 +8,10 @@ dnsmasq_dns_servers = {{ neutron_dnsmasq_dns_servers }} {% if neutron_plugin_agent == 'vmware_dvs' %} ovs_integration_bridge = {{ neutron_bridge_name }} interface_driver = openvswitch -enable_metadata_network = True +enable_metadata_network = true dhcp_driver = vmware_nsx.plugins.dvs.dhcp.Dnsmasq -use_namespaces = True -ovs_use_veth = False +use_namespaces = true +ovs_use_veth = false {% if vmware_dvs_dhcp_override_mac != '' %} dhcp_override_mac = {{ vmware_dvs_dhcp_override_mac }} {% endif %} diff --git a/ansible/roles/neutron/templates/fwaas_driver.ini.j2 b/ansible/roles/neutron/templates/fwaas_driver.ini.j2 index b0df0858fb..28bb824ba1 100644 --- a/ansible/roles/neutron/templates/fwaas_driver.ini.j2 +++ b/ansible/roles/neutron/templates/fwaas_driver.ini.j2 @@ -1,5 +1,5 @@ [fwaas] -enabled = True +enabled = true {% if neutron_plugin_agent == 'vmware_nsxv' %} driver = vmware_nsxv_edge {% else %} diff --git a/ansible/roles/neutron/templates/ml2_conf.ini.j2 b/ansible/roles/neutron/templates/ml2_conf.ini.j2 index 0e34477691..8b857625d8 100644 --- a/ansible/roles/neutron/templates/ml2_conf.ini.j2 +++ b/ansible/roles/neutron/templates/ml2_conf.ini.j2 @@ -38,7 +38,7 @@ max_header_size = 38 [ovn] ovn_nb_connection = {{ ovn_nb_connection }} ovn_sb_connection = {{ ovn_sb_connection }} -ovn_metadata_enabled = True +ovn_metadata_enabled = true enable_distributed_floating_ip = {{ neutron_ovn_distributed_fip | bool }} -ovn_emit_need_to_frag = True +ovn_emit_need_to_frag = true {% endif %} diff --git a/ansible/roles/neutron/templates/neutron.conf.j2 b/ansible/roles/neutron/templates/neutron.conf.j2 index 515b42ce0a..7b5aea818e 100644 --- a/ansible/roles/neutron/templates/neutron.conf.j2 +++ b/ansible/roles/neutron/templates/neutron.conf.j2 @@ -4,9 +4,9 @@ debug = {{ neutron_logging_debug }} log_dir = /var/log/kolla/neutron -# NOTE(elemoine): set use_stderr to False or the logs will also be sent to +# NOTE(elemoine): set use_stderr to false or the logs will also be sent to # stderr and collected by Docker -use_stderr = False +use_stderr = false {% if neutron_enable_tls_backend | bool %} bind_host = 127.0.0.1 {% else %} @@ -38,10 +38,10 @@ host = {{ ansible_facts.hostname }}_{{ item }} core_plugin = vmware_nsx.plugin.NsxVPlugin {% elif neutron_plugin_agent == 'vmware_nsxv3' %} core_plugin = vmware_nsx.plugin.NsxV3Plugin -dhcp_agent_notification = False +dhcp_agent_notification = false {% elif neutron_plugin_agent == 'vmware_nsxp' %} core_plugin = vmware_nsx.plugin.NsxPolicyPlugin -dhcp_agent_notification = False +dhcp_agent_notification = false {% elif neutron_plugin_agent == 'vmware_dvs' %} core_plugin = vmware_nsx.plugin.NsxDvsPlugin {% else %} @@ -58,7 +58,7 @@ max_l3_agents_per_router = {{ max_l3_agents_per_router }} transport_url = {{ rpc_transport_url }} {% if enable_neutron_dvr | bool %} -router_distributed = True +router_distributed = true {% endif %} dns_domain = {{ neutron_dns_domain }} @@ -90,7 +90,7 @@ endpoint_type = internal cafile = {{ openstack_cacert }} [oslo_middleware] -enable_proxy_headers_parsing = True +enable_proxy_headers_parsing = true [oslo_concurrency] lock_path = /var/lib/neutron/tmp @@ -168,7 +168,7 @@ user_domain_id = {{ default_user_domain_id }} project_name = service username = {{ designate_keystone_user }} password = {{ designate_keystone_password }} -allow_reverse_dns_lookup = True +allow_reverse_dns_lookup = true ipv4_ptr_zone_prefix_size = 24 ipv6_ptr_zone_prefix_size = 116 cafile = {{ openstack_cacert }} diff --git a/ansible/roles/neutron/templates/neutron_taas.conf.j2 b/ansible/roles/neutron/templates/neutron_taas.conf.j2 index 5031871834..95ddb24d20 100644 --- a/ansible/roles/neutron/templates/neutron_taas.conf.j2 +++ b/ansible/roles/neutron/templates/neutron_taas.conf.j2 @@ -3,4 +3,4 @@ service_provider = TAAS:TAAS:neutron_taas.services.taas.service_drivers.taas_rpc [taas] driver = neutron_taas.services.taas.drivers.linux.ovs_taas.OvsTaasDriver -enabled = True +enabled = true diff --git a/ansible/roles/neutron/templates/openvswitch_agent.ini.j2 b/ansible/roles/neutron/templates/openvswitch_agent.ini.j2 index 8ac25af7e1..f908c7f146 100644 --- a/ansible/roles/neutron/templates/openvswitch_agent.ini.j2 +++ b/ansible/roles/neutron/templates/openvswitch_agent.ini.j2 @@ -4,7 +4,7 @@ tunnel_types = vxlan l2_population = true arp_responder = true {% if enable_neutron_dvr | bool %} -enable_distributed_routing = True +enable_distributed_routing = true {% endif %} {% if neutron_agent_extensions %} extensions = {{ neutron_agent_extensions|map(attribute='name')|join(',') }} diff --git a/ansible/roles/nova-cell/templates/nova.conf.d/libvirt.conf.j2 b/ansible/roles/nova-cell/templates/nova.conf.d/libvirt.conf.j2 index 7c23ff8306..35b538a31d 100644 --- a/ansible/roles/nova-cell/templates/nova.conf.d/libvirt.conf.j2 +++ b/ansible/roles/nova-cell/templates/nova.conf.d/libvirt.conf.j2 @@ -25,8 +25,8 @@ virt_type = {{ nova_compute_virt_type }} cpu_mode = {{ nova_libvirt_cpu_mode }} {% endif %} {% if enable_multipathd | bool %} -volume_use_multipath = True +volume_use_multipath = true {% endif %} num_pcie_ports = 16 [workarounds] -skip_cpu_compare_on_dest = True +skip_cpu_compare_on_dest = true diff --git a/ansible/roles/nova-cell/templates/nova.conf.j2 b/ansible/roles/nova-cell/templates/nova.conf.j2 index 8aedbd7ea7..cdacedf4ae 100644 --- a/ansible/roles/nova-cell/templates/nova.conf.j2 +++ b/ansible/roles/nova-cell/templates/nova.conf.j2 @@ -27,7 +27,7 @@ compute_driver = libvirt.LibvirtDriver my_ip = {{ api_interface_address }} {% if enable_ceilometer | bool %} -instance_usage_audit = True +instance_usage_audit = true instance_usage_audit_period = hour {% if enable_watcher | bool %} compute_monitors=nova.compute.monitors.cpu.virt_driver diff --git a/ansible/roles/nova/templates/nova.conf.j2 b/ansible/roles/nova/templates/nova.conf.j2 index 8ebfc8664b..3a35519568 100644 --- a/ansible/roles/nova/templates/nova.conf.j2 +++ b/ansible/roles/nova/templates/nova.conf.j2 @@ -19,7 +19,7 @@ allow_resize_to_same_host = true my_ip = {{ api_interface_address }} {% if enable_ceilometer | bool %} -instance_usage_audit = True +instance_usage_audit = true instance_usage_audit_period = hour {% endif %} @@ -37,7 +37,7 @@ enabled_filters = ComputeFilter,ComputeCapabilitiesFilter,ImagePropertiesFilter {% if enable_cells | bool %} # When in superconductor mode, nova-compute can't send instance # info updates to the scheduler, so just disable it. -track_instance_changes = False +track_instance_changes = false {% endif %} {% if vendordata_file_path is defined %} @@ -50,7 +50,7 @@ vendordata_jsonfile_path = /etc/nova/vendordata.json workers = {{ nova_superconductor_workers }} [oslo_middleware] -enable_proxy_headers_parsing = True +enable_proxy_headers_parsing = true [oslo_concurrency] lock_path = /var/lib/nova/tmp @@ -105,7 +105,7 @@ max_retries = -1 [cache] backend = oslo_cache.memcache_pool -enabled = True +enabled = true memcache_servers = {% for host in groups['memcached'] %}{{ 'api' | kolla_address(host) | put_address_in_context('memcache') }}:{{ memcached_port }}{% if not loop.last %},{% endif %}{% endfor %} diff --git a/ansible/roles/placement/templates/placement.conf.j2 b/ansible/roles/placement/templates/placement.conf.j2 index 10914fde13..4c325e2fcf 100644 --- a/ansible/roles/placement/templates/placement.conf.j2 +++ b/ansible/roles/placement/templates/placement.conf.j2 @@ -6,7 +6,7 @@ log_file = /var/log/kolla/placement/{{ service_name }}.log state_path = /var/lib/placement [oslo_middleware] -enable_proxy_headers_parsing = True +enable_proxy_headers_parsing = true [oslo_concurrency] lock_path = /var/lib/placement/tmp @@ -20,7 +20,7 @@ max_retries = -1 [cache] backend = oslo_cache.memcache_pool -enabled = True +enabled = true memcache_servers = {% for host in groups['memcached'] %}{{ 'api' | kolla_address(host) | put_address_in_context('memcache') }}:{{ memcached_port }}{% if not loop.last %},{% endif %}{% endfor %} diff --git a/ansible/roles/swift/templates/container.conf.j2 b/ansible/roles/swift/templates/container.conf.j2 index 1c6a95b7e0..63a9515f57 100644 --- a/ansible/roles/swift/templates/container.conf.j2 +++ b/ansible/roles/swift/templates/container.conf.j2 @@ -22,7 +22,7 @@ recon_cache_path = /var/cache/swift [app:container-server] use = egg:swift#container -allow_versions = True +allow_versions = true {% if swift_has_replication_network %} replication_server = {{ service_name == 'swift-container-replication-server' }} {% endif %} diff --git a/ansible/roles/swift/templates/proxy-server.conf.j2 b/ansible/roles/swift/templates/proxy-server.conf.j2 index 4b6982a334..0a808ea90d 100644 --- a/ansible/roles/swift/templates/proxy-server.conf.j2 +++ b/ansible/roles/swift/templates/proxy-server.conf.j2 @@ -91,7 +91,7 @@ use = egg:swift#dlo [filter:versioned_writes] use = egg:swift#versioned_writes -allow_versioned_writes = True +allow_versioned_writes = true {% if enable_swift_s3api | bool %} [filter:s3api] diff --git a/ansible/roles/tacker/templates/tacker.conf.j2 b/ansible/roles/tacker/templates/tacker.conf.j2 index 5dfad235f0..ed091f0947 100644 --- a/ansible/roles/tacker/templates/tacker.conf.j2 +++ b/ansible/roles/tacker/templates/tacker.conf.j2 @@ -18,7 +18,7 @@ stack_retry_wait = 10 {% if enable_barbican | bool %} [vim_keys] -use_barbican = True +use_barbican = true {% endif %} [tacker] diff --git a/ansible/roles/trove/templates/trove-guestagent.conf.j2 b/ansible/roles/trove/templates/trove-guestagent.conf.j2 index 64d4a636da..ca29734aa6 100644 --- a/ansible/roles/trove/templates/trove-guestagent.conf.j2 +++ b/ansible/roles/trove/templates/trove-guestagent.conf.j2 @@ -4,7 +4,7 @@ log_dir = /var/log/trove/ transport_url = {{ rpc_transport_url }} control_exchange = trove root_grant = ALL -root_grant_option = True +root_grant_option = true debug = {{ trove_logging_debug }} [service_credentials] diff --git a/ansible/roles/watcher/templates/watcher.conf.j2 b/ansible/roles/watcher/templates/watcher.conf.j2 index 5e9ad92658..2afa97cb30 100644 --- a/ansible/roles/watcher/templates/watcher.conf.j2 +++ b/ansible/roles/watcher/templates/watcher.conf.j2 @@ -28,7 +28,7 @@ user_domain_id = {{ default_user_domain_id }} project_name = service username = {{ watcher_keystone_user }} password = {{ watcher_keystone_password }} -service_token_roles_required = True +service_token_roles_required = true cafile = {{ openstack_cacert }} region_name = {{ openstack_region_name }} diff --git a/ansible/roles/zun/templates/zun.conf.j2 b/ansible/roles/zun/templates/zun.conf.j2 index 8caa33f49d..30e456285c 100644 --- a/ansible/roles/zun/templates/zun.conf.j2 +++ b/ansible/roles/zun/templates/zun.conf.j2 @@ -36,7 +36,7 @@ user_domain_id = {{ default_user_domain_id }} project_name = service username = {{ zun_keystone_user }} password = {{ zun_keystone_password }} -service_token_roles_required = True +service_token_roles_required = true region_name = {{ openstack_region_name }} cafile = {{ openstack_cacert }} @@ -59,7 +59,7 @@ user_domain_id = {{ default_user_domain_id }} project_name = service username = {{ zun_keystone_user }} password = {{ zun_keystone_password }} -service_token_roles_required = True +service_token_roles_required = true region_name = {{ openstack_region_name }} cafile = {{ openstack_cacert }}