octavia_tempest_plugin.tests.scenario.v2.test_traffic_ops

Source code for octavia_tempest_plugin.tests.scenario.v2.test_traffic_ops

# Copyright 2018 GoDaddy
#
#    Licensed under the Apache License, Version 2.0 (the "License"); you may
#    not use this file except in compliance with the License. You may obtain
#    a copy of the License at
#
#         http://www.apache.org/licenses/LICENSE-2.0
#
#    Unless required by applicable law or agreed to in writing, software
#    distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
#    WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
#    License for the specific language governing permissions and limitations
#    under the License.

import testtools

from oslo_log import log as logging
from tempest import config
from tempest.lib.common.utils import data_utils
from tempest.lib import decorators

from octavia_tempest_plugin.common import constants as const
from octavia_tempest_plugin.tests import test_base
from octavia_tempest_plugin.tests import waiters

CONF = config.CONF
LOG = logging.getLogger(__name__)


[docs]@testtools.skipUnless( CONF.validation.run_validation, 'Traffic tests will not work without run_validation enabled.') class TrafficOperationsScenarioTest(test_base.LoadBalancerBaseTestWithCompute):
[docs] @classmethod def resource_setup(cls): """Setup resources needed by the tests.""" super(TrafficOperationsScenarioTest, cls).resource_setup() lb_name = data_utils.rand_name("lb_member_lb1_operations") lb_kwargs = {const.PROVIDER: CONF.load_balancer.provider, const.NAME: lb_name} # TODO(rm_work): Make this work with ipv6 and split this test for both ip_version = 4 cls._setup_lb_network_kwargs(lb_kwargs, ip_version) lb = cls.mem_lb_client.create_loadbalancer(**lb_kwargs) cls.lb_id = lb[const.ID] cls.addClassResourceCleanup( cls.mem_lb_client.cleanup_loadbalancer, cls.lb_id) if CONF.validation.connect_method == 'floating': port_id = lb[const.VIP_PORT_ID] result = cls.lb_mem_float_ip_client.create_floatingip( floating_network_id=CONF.network.public_network_id, port_id=port_id) floating_ip = result['floatingip'] LOG.info('lb1_floating_ip: {}'.format(floating_ip)) cls.addClassResourceCleanup( waiters.wait_for_not_found, cls.lb_mem_float_ip_client.delete_floatingip, cls.lb_mem_float_ip_client.show_floatingip, floatingip_id=floating_ip['id']) cls.lb_vip_address = floating_ip['floating_ip_address'] else: cls.lb_vip_address = lb[const.VIP_ADDRESS] waiters.wait_for_status(cls.mem_lb_client.show_loadbalancer, cls.lb_id, const.PROVISIONING_STATUS, const.ACTIVE, CONF.load_balancer.lb_build_interval, CONF.load_balancer.lb_build_timeout) listener_name = data_utils.rand_name("lb_member_listener1_operations") listener_kwargs = { const.NAME: listener_name, const.PROTOCOL: const.HTTP, const.PROTOCOL_PORT: '80', const.LOADBALANCER_ID: cls.lb_id, } listener = cls.mem_listener_client.create_listener(**listener_kwargs) cls.listener_id = listener[const.ID] cls.addClassResourceCleanup( cls.mem_listener_client.cleanup_listener, cls.listener_id, lb_client=cls.mem_lb_client, lb_id=cls.lb_id) waiters.wait_for_status(cls.mem_lb_client.show_loadbalancer, cls.lb_id, const.PROVISIONING_STATUS, const.ACTIVE, CONF.load_balancer.build_interval, CONF.load_balancer.build_timeout) pool_name = data_utils.rand_name("lb_member_pool1_operations") pool_kwargs = { const.NAME: pool_name, const.PROTOCOL: const.HTTP, const.LB_ALGORITHM: const.LB_ALGORITHM_ROUND_ROBIN, const.LISTENER_ID: cls.listener_id, } pool = cls.mem_pool_client.create_pool(**pool_kwargs) cls.pool_id = pool[const.ID] cls.addClassResourceCleanup( cls.mem_pool_client.cleanup_pool, cls.pool_id, lb_client=cls.mem_lb_client, lb_id=cls.lb_id) waiters.wait_for_status(cls.mem_lb_client.show_loadbalancer, cls.lb_id, const.PROVISIONING_STATUS, const.ACTIVE, CONF.load_balancer.build_interval, CONF.load_balancer.build_timeout)
[docs] @testtools.skipIf(CONF.load_balancer.test_with_noop, 'Traffic tests will not work in noop mode.') @decorators.idempotent_id('6751135d-e15a-4e22-89f4-bfcc3408d424') def test_basic_traffic(self): """Tests sending traffic through a loadbalancer * Set up members on a loadbalancer. * Test traffic to ensure it is balanced properly. """ # Set up Member 1 for Webserver 1 member1_name = data_utils.rand_name("lb_member_member1-traffic") member1_kwargs = { const.POOL_ID: self.pool_id, const.NAME: member1_name, const.ADMIN_STATE_UP: True, const.ADDRESS: self.webserver1_ip, const.PROTOCOL_PORT: 80, } if self.lb_member_1_subnet: member1_kwargs[const.SUBNET_ID] = self.lb_member_1_subnet[const.ID] member1 = self.mem_member_client.create_member( **member1_kwargs) self.addCleanup( self.mem_member_client.cleanup_member, member1[const.ID], pool_id=self.pool_id, lb_client=self.mem_lb_client, lb_id=self.lb_id) waiters.wait_for_status( self.mem_lb_client.show_loadbalancer, self.lb_id, const.PROVISIONING_STATUS, const.ACTIVE, CONF.load_balancer.check_interval, CONF.load_balancer.check_timeout) # Set up Member 2 for Webserver 2 member2_name = data_utils.rand_name("lb_member_member2-traffic") member2_kwargs = { const.POOL_ID: self.pool_id, const.NAME: member2_name, const.ADMIN_STATE_UP: True, const.ADDRESS: self.webserver2_ip, const.PROTOCOL_PORT: 80, } if self.lb_member_2_subnet: member2_kwargs[const.SUBNET_ID] = self.lb_member_2_subnet[const.ID] member2 = self.mem_member_client.create_member( **member2_kwargs) self.addCleanup( self.mem_member_client.cleanup_member, member2[const.ID], pool_id=self.pool_id, lb_client=self.mem_lb_client, lb_id=self.lb_id) waiters.wait_for_status( self.mem_lb_client.show_loadbalancer, self.lb_id, const.PROVISIONING_STATUS, const.ACTIVE, CONF.load_balancer.check_interval, CONF.load_balancer.check_timeout) # Send some traffic self.check_members_balanced(self.lb_vip_address)
[docs] @decorators.idempotent_id('a16f8eb4-a77c-4b0e-8b1b-91c237039713') def test_healthmonitor_traffic(self): """Tests traffic is correctly routed based on healthmonitor status * Create three members: * One should be working, and ONLINE with a healthmonitor (passing) * One should be working, and ERROR with a healthmonitor (failing) * One should be disabled, and OFFLINE with a healthmonitor * Verify members are in their correct respective operating statuses. * Verify that traffic is balanced evenly between the working members. * Create a fully populated healthmonitor. * Verify members are in their correct respective operating statuses. * Verify that traffic is balanced *unevenly*. * Delete the healthmonitor. * Verify members are in their correct respective operating statuses. * Verify that traffic is balanced evenly between the working members. """ member1_name = data_utils.rand_name("lb_member_member1-hm-traffic") member1_kwargs = { const.POOL_ID: self.pool_id, const.NAME: member1_name, const.ADMIN_STATE_UP: True, const.ADDRESS: self.webserver1_ip, const.PROTOCOL_PORT: 80, } if self.lb_member_1_subnet: member1_kwargs[const.SUBNET_ID] = self.lb_member_1_subnet[const.ID] member1 = self.mem_member_client.create_member( **member1_kwargs) member1_id = member1[const.ID] self.addCleanup( self.mem_member_client.cleanup_member, member1_id, pool_id=self.pool_id, lb_client=self.mem_lb_client, lb_id=self.lb_id) waiters.wait_for_status( self.mem_lb_client.show_loadbalancer, self.lb_id, const.PROVISIONING_STATUS, const.ACTIVE, CONF.load_balancer.check_interval, CONF.load_balancer.check_timeout) # Set up Member 2 for Webserver 2 member2_name = data_utils.rand_name("lb_member_member2-hm-traffic") member2_kwargs = { const.POOL_ID: self.pool_id, const.NAME: member2_name, const.ADMIN_STATE_UP: True, const.ADDRESS: self.webserver2_ip, const.PROTOCOL_PORT: 80, const.MONITOR_PORT: 9999, # We want this to go offline with a HM } if self.lb_member_2_subnet: member2_kwargs[const.SUBNET_ID] = self.lb_member_2_subnet[const.ID] member2 = self.mem_member_client.create_member( **member2_kwargs) member2_id = member2[const.ID] self.addCleanup( self.mem_member_client.cleanup_member, member2_id, pool_id=self.pool_id, lb_client=self.mem_lb_client, lb_id=self.lb_id) waiters.wait_for_status( self.mem_lb_client.show_loadbalancer, self.lb_id, const.PROVISIONING_STATUS, const.ACTIVE, CONF.load_balancer.check_interval, CONF.load_balancer.check_timeout) # Set up Member 3 as a non-existent disabled node member3_name = data_utils.rand_name("lb_member_member3-hm-traffic") member3_kwargs = { const.POOL_ID: self.pool_id, const.NAME: member3_name, const.ADMIN_STATE_UP: False, const.ADDRESS: '192.0.2.1', const.PROTOCOL_PORT: 80, } member3 = self.mem_member_client.create_member( **member3_kwargs) member3_id = member3[const.ID] self.addCleanup( self.mem_member_client.cleanup_member, member3_id, pool_id=self.pool_id, lb_client=self.mem_lb_client, lb_id=self.lb_id) waiters.wait_for_status( self.mem_lb_client.show_loadbalancer, self.lb_id, const.PROVISIONING_STATUS, const.ACTIVE, CONF.load_balancer.check_interval, CONF.load_balancer.check_timeout) # Wait for members to adjust to the correct OPERATING_STATUS waiters.wait_for_status( self.mem_member_client.show_member, member1_id, const.OPERATING_STATUS, const.NO_MONITOR, CONF.load_balancer.build_interval, CONF.load_balancer.build_timeout, pool_id=self.pool_id) waiters.wait_for_status( self.mem_member_client.show_member, member2_id, const.OPERATING_STATUS, const.NO_MONITOR, CONF.load_balancer.build_interval, CONF.load_balancer.build_timeout, pool_id=self.pool_id) waiters.wait_for_status( self.mem_member_client.show_member, member3_id, const.OPERATING_STATUS, const.OFFLINE, CONF.load_balancer.build_interval, CONF.load_balancer.build_timeout, pool_id=self.pool_id) # Send some traffic and verify it is balanced self.check_members_balanced(self.lb_vip_address, traffic_member_count=2) # Create the healthmonitor hm_name = data_utils.rand_name("lb_member_hm1-hm-traffic") hm_kwargs = { const.POOL_ID: self.pool_id, const.NAME: hm_name, const.TYPE: const.HEALTH_MONITOR_HTTP, const.DELAY: 2, const.TIMEOUT: 2, const.MAX_RETRIES: 2, const.MAX_RETRIES_DOWN: 2, const.HTTP_METHOD: const.GET, const.URL_PATH: '/', const.EXPECTED_CODES: '200', const.ADMIN_STATE_UP: True, } hm = self.mem_healthmonitor_client.create_healthmonitor(**hm_kwargs) self.addCleanup( self.mem_healthmonitor_client.cleanup_healthmonitor, hm[const.ID], lb_client=self.mem_lb_client, lb_id=self.lb_id) waiters.wait_for_status( self.mem_lb_client.show_loadbalancer, self.lb_id, const.PROVISIONING_STATUS, const.ACTIVE, CONF.load_balancer.build_interval, CONF.load_balancer.build_timeout) hm = waiters.wait_for_status( self.mem_healthmonitor_client.show_healthmonitor, hm[const.ID], const.PROVISIONING_STATUS, const.ACTIVE, CONF.load_balancer.build_interval, CONF.load_balancer.build_timeout) # Wait for members to adjust to the correct OPERATING_STATUS waiters.wait_for_status( self.mem_member_client.show_member, member1_id, const.OPERATING_STATUS, const.ONLINE, CONF.load_balancer.build_interval, CONF.load_balancer.build_timeout, pool_id=self.pool_id) waiters.wait_for_status( self.mem_member_client.show_member, member2_id, const.OPERATING_STATUS, const.ERROR, CONF.load_balancer.build_interval, CONF.load_balancer.build_timeout, pool_id=self.pool_id) waiters.wait_for_status( self.mem_member_client.show_member, member3_id, const.OPERATING_STATUS, const.OFFLINE, CONF.load_balancer.build_interval, CONF.load_balancer.build_timeout, pool_id=self.pool_id) # Send some traffic and verify it is *unbalanced*, as expected self.check_members_balanced(self.lb_vip_address, traffic_member_count=1) # Delete the healthmonitor self.mem_healthmonitor_client.delete_healthmonitor(hm[const.ID]) waiters.wait_for_deleted_status_or_not_found( self.mem_healthmonitor_client.show_healthmonitor, hm[const.ID], const.PROVISIONING_STATUS, CONF.load_balancer.check_interval, CONF.load_balancer.check_timeout) # Wait for members to adjust to the correct OPERATING_STATUS waiters.wait_for_status( self.mem_member_client.show_member, member1_id, const.OPERATING_STATUS, const.NO_MONITOR, CONF.load_balancer.build_interval, CONF.load_balancer.build_timeout, pool_id=self.pool_id) waiters.wait_for_status( self.mem_member_client.show_member, member2_id, const.OPERATING_STATUS, const.NO_MONITOR, CONF.load_balancer.build_interval, CONF.load_balancer.build_timeout, pool_id=self.pool_id) waiters.wait_for_status( self.mem_member_client.show_member, member3_id, const.OPERATING_STATUS, const.OFFLINE, CONF.load_balancer.build_interval, CONF.load_balancer.build_timeout, pool_id=self.pool_id) # Send some traffic and verify it is balanced again self.check_members_balanced(self.lb_vip_address)
[docs] @decorators.idempotent_id('3558186d-6dcd-4d9d-b7f7-adc190b66149') def test_l7policies_and_l7rules(self): """Tests sending traffic through a loadbalancer with l7rules * Create an extra pool. * Put one member on the default pool, and one on the second pool. * Create a policy/rule to redirect to the second pool. * Create a policy/rule to redirect to the identity URI. * Create a policy/rule to reject connections. * Test traffic to ensure it goes to the correct place. """ # Create a second pool pool_name = data_utils.rand_name("lb_member_pool2_l7redirect") pool_kwargs = { const.NAME: pool_name, const.PROTOCOL: const.HTTP, const.LB_ALGORITHM: const.LB_ALGORITHM_ROUND_ROBIN, const.LOADBALANCER_ID: self.lb_id, } pool = self.mem_pool_client.create_pool(**pool_kwargs) pool_id = pool[const.ID] self.addCleanup( self.mem_pool_client.cleanup_pool, pool_id, lb_client=self.mem_lb_client, lb_id=self.lb_id) waiters.wait_for_status(self.mem_lb_client.show_loadbalancer, self.lb_id, const.PROVISIONING_STATUS, const.ACTIVE, CONF.load_balancer.build_interval, CONF.load_balancer.build_timeout) # Set up Member 1 for Webserver 1 on the default pool member1_name = data_utils.rand_name("lb_member_member1-l7redirect") member1_kwargs = { const.POOL_ID: self.pool_id, const.NAME: member1_name, const.ADMIN_STATE_UP: True, const.ADDRESS: self.webserver1_ip, const.PROTOCOL_PORT: 80, } if self.lb_member_1_subnet: member1_kwargs[const.SUBNET_ID] = self.lb_member_1_subnet[const.ID] member1 = self.mem_member_client.create_member( **member1_kwargs) self.addCleanup( self.mem_member_client.cleanup_member, member1[const.ID], pool_id=self.pool_id, lb_client=self.mem_lb_client, lb_id=self.lb_id) waiters.wait_for_status( self.mem_lb_client.show_loadbalancer, self.lb_id, const.PROVISIONING_STATUS, const.ACTIVE, CONF.load_balancer.check_interval, CONF.load_balancer.check_timeout) # Set up Member 2 for Webserver 2 on the alternate pool member2_name = data_utils.rand_name("lb_member_member2-l7redirect") member2_kwargs = { const.POOL_ID: pool_id, const.NAME: member2_name, const.ADMIN_STATE_UP: True, const.ADDRESS: self.webserver2_ip, const.PROTOCOL_PORT: 80, } if self.lb_member_2_subnet: member2_kwargs[const.SUBNET_ID] = self.lb_member_2_subnet[const.ID] member2 = self.mem_member_client.create_member( **member2_kwargs) self.addCleanup( self.mem_member_client.cleanup_member, member2[const.ID], pool_id=self.pool_id, lb_client=self.mem_lb_client, lb_id=self.lb_id) waiters.wait_for_status( self.mem_lb_client.show_loadbalancer, self.lb_id, const.PROVISIONING_STATUS, const.ACTIVE, CONF.load_balancer.check_interval, CONF.load_balancer.check_timeout) # Create the l7policy to redirect to the alternate pool l7policy1_name = data_utils.rand_name("lb_member_l7policy1-l7redirect") l7policy1_description = data_utils.arbitrary_string(size=255) l7policy1_kwargs = { const.LISTENER_ID: self.listener_id, const.NAME: l7policy1_name, const.DESCRIPTION: l7policy1_description, const.ADMIN_STATE_UP: True, const.POSITION: 1, const.ACTION: const.REDIRECT_TO_POOL, const.REDIRECT_POOL_ID: pool_id, } l7policy1 = self.mem_l7policy_client.create_l7policy( **l7policy1_kwargs) self.addCleanup( self.mem_l7policy_client.cleanup_l7policy, l7policy1[const.ID], lb_client=self.mem_lb_client, lb_id=self.lb_id) waiters.wait_for_status( self.mem_lb_client.show_loadbalancer, self.lb_id, const.PROVISIONING_STATUS, const.ACTIVE, CONF.load_balancer.build_interval, CONF.load_balancer.build_timeout) # Redirect slow queries to the alternate pool l7rule1_kwargs = { const.L7POLICY_ID: l7policy1[const.ID], const.ADMIN_STATE_UP: True, const.TYPE: const.PATH, const.VALUE: '/slow', const.COMPARE_TYPE: const.STARTS_WITH, const.INVERT: False, } l7rule1 = self.mem_l7rule_client.create_l7rule(**l7rule1_kwargs) self.addCleanup( self.mem_l7rule_client.cleanup_l7rule, l7rule1[const.ID], l7policy_id=l7rule1_kwargs[const.L7POLICY_ID], lb_client=self.mem_lb_client, lb_id=self.lb_id) waiters.wait_for_status( self.mem_lb_client.show_loadbalancer, self.lb_id, const.PROVISIONING_STATUS, const.ACTIVE, CONF.load_balancer.build_interval, CONF.load_balancer.build_timeout) # Create the l7policy to redirect to the identity URI l7policy2_name = data_utils.rand_name("lb_member_l7policy2-l7redirect") l7policy2_description = data_utils.arbitrary_string(size=255) l7policy2_kwargs = { const.LISTENER_ID: self.listener_id, const.NAME: l7policy2_name, const.DESCRIPTION: l7policy2_description, const.ADMIN_STATE_UP: True, const.POSITION: 1, const.ACTION: const.REDIRECT_TO_URL, const.REDIRECT_URL: CONF.identity.uri_v3, } l7policy2 = self.mem_l7policy_client.create_l7policy( **l7policy2_kwargs) self.addCleanup( self.mem_l7policy_client.cleanup_l7policy, l7policy2[const.ID], lb_client=self.mem_lb_client, lb_id=self.lb_id) waiters.wait_for_status( self.mem_lb_client.show_loadbalancer, self.lb_id, const.PROVISIONING_STATUS, const.ACTIVE, CONF.load_balancer.build_interval, CONF.load_balancer.build_timeout) # Redirect queries for 'turtles' to identity l7rule2_kwargs = { const.L7POLICY_ID: l7policy2[const.ID], const.ADMIN_STATE_UP: True, const.TYPE: const.PATH, const.VALUE: '/turtles', const.COMPARE_TYPE: const.EQUAL_TO, const.INVERT: False, } l7rule2 = self.mem_l7rule_client.create_l7rule(**l7rule2_kwargs) self.addCleanup( self.mem_l7rule_client.cleanup_l7rule, l7rule2[const.ID], l7policy_id=l7rule2_kwargs[const.L7POLICY_ID], lb_client=self.mem_lb_client, lb_id=self.lb_id) waiters.wait_for_status( self.mem_lb_client.show_loadbalancer, self.lb_id, const.PROVISIONING_STATUS, const.ACTIVE, CONF.load_balancer.build_interval, CONF.load_balancer.build_timeout) # Create the l7policy to reject requests l7policy3_name = data_utils.rand_name("lb_member_l7policy3-l7redirect") l7policy3_description = data_utils.arbitrary_string(size=255) l7policy3_kwargs = { const.LISTENER_ID: self.listener_id, const.NAME: l7policy3_name, const.DESCRIPTION: l7policy3_description, const.ADMIN_STATE_UP: True, const.POSITION: 1, const.ACTION: const.REJECT, } l7policy3 = self.mem_l7policy_client.create_l7policy( **l7policy3_kwargs) self.addCleanup( self.mem_l7policy_client.cleanup_l7policy, l7policy3[const.ID], lb_client=self.mem_lb_client, lb_id=self.lb_id) waiters.wait_for_status( self.mem_lb_client.show_loadbalancer, self.lb_id, const.PROVISIONING_STATUS, const.ACTIVE, CONF.load_balancer.build_interval, CONF.load_balancer.build_timeout) # Reject requests that include the header data 'reject=true' l7rule3_kwargs = { const.L7POLICY_ID: l7policy3[const.ID], const.ADMIN_STATE_UP: True, const.TYPE: const.HEADER, const.KEY: 'reject', const.VALUE: 'true', const.COMPARE_TYPE: const.EQUAL_TO, const.INVERT: False, } l7rule3 = self.mem_l7rule_client.create_l7rule(**l7rule3_kwargs) self.addCleanup( self.mem_l7rule_client.cleanup_l7rule, l7rule3[const.ID], l7policy_id=l7rule3_kwargs[const.L7POLICY_ID], lb_client=self.mem_lb_client, lb_id=self.lb_id) waiters.wait_for_status( self.mem_lb_client.show_loadbalancer, self.lb_id, const.PROVISIONING_STATUS, const.ACTIVE, CONF.load_balancer.build_interval, CONF.load_balancer.build_timeout) # Assert that normal traffic goes to pool1->member1 url_for_member1 = 'http://{}/'.format(self.lb_vip_address) self.assertConsistentResponse((200, self.webserver1_response), url_for_member1) # Assert that slow traffic goes to pool2->member2 url_for_member2 = 'http://{}/slow?delay=1s'.format(self.lb_vip_address) self.assertConsistentResponse((200, self.webserver2_response), url_for_member2) # Assert that /turtles is redirected to identity url_for_identity = 'http://{}/turtles'.format(self.lb_vip_address) self.assertConsistentResponse((302, CONF.identity.uri_v3), url_for_identity, redirect=True) # Assert that traffic with header 'reject=true' is rejected self.assertConsistentResponse((403, None), url_for_member1, headers={'reject': 'true'})
[docs] @testtools.skipIf(CONF.load_balancer.test_with_noop, 'Traffic tests will not work in noop mode.') @testtools.skipUnless(CONF.load_balancer.test_with_ipv6, 'Mixed IPv4/IPv6 member test requires IPv6.') @decorators.idempotent_id('20b6b671-0101-4bed-a249-9af6ee3aa6d9') def test_mixed_ipv4_ipv6_members_traffic(self): """Tests traffic through a loadbalancer with IPv4 and IPv6 members. * Set up members on a loadbalancer. * Test traffic to ensure it is balanced properly. """ # Set up Member 1 for Webserver 1 member1_name = data_utils.rand_name("lb_member_member1-traffic") member1_kwargs = { const.POOL_ID: self.pool_id, const.NAME: member1_name, const.ADMIN_STATE_UP: True, const.ADDRESS: self.webserver1_ip, const.PROTOCOL_PORT: 80, } if self.lb_member_1_subnet: member1_kwargs[const.SUBNET_ID] = self.lb_member_1_subnet[const.ID] member1 = self.mem_member_client.create_member( **member1_kwargs) self.addCleanup( self.mem_member_client.cleanup_member, member1[const.ID], pool_id=self.pool_id, lb_client=self.mem_lb_client, lb_id=self.lb_id) waiters.wait_for_status( self.mem_lb_client.show_loadbalancer, self.lb_id, const.PROVISIONING_STATUS, const.ACTIVE, CONF.load_balancer.check_interval, CONF.load_balancer.check_timeout) # Set up Member 2 for Webserver 2 member2_name = data_utils.rand_name("lb_member_member2-traffic") member2_kwargs = { const.POOL_ID: self.pool_id, const.NAME: member2_name, const.ADMIN_STATE_UP: True, const.ADDRESS: self.webserver2_ipv6, const.PROTOCOL_PORT: 80, } if self.lb_member_2_ipv6_subnet: member2_kwargs[const.SUBNET_ID] = ( self.lb_member_2_ipv6_subnet[const.ID]) member2 = self.mem_member_client.create_member( **member2_kwargs) self.addCleanup( self.mem_member_client.cleanup_member, member2[const.ID], pool_id=self.pool_id, lb_client=self.mem_lb_client, lb_id=self.lb_id) waiters.wait_for_status( self.mem_lb_client.show_loadbalancer, self.lb_id, const.PROVISIONING_STATUS, const.ACTIVE, CONF.load_balancer.check_interval, CONF.load_balancer.check_timeout) # Send some traffic self.check_members_balanced(self.lb_vip_address)
Creative Commons Attribution 3.0 License

Except where otherwise noted, this document is licensed under Creative Commons Attribution 3.0 License. See all OpenStack Legal Documents.