mirror of https://github.com/apache/cloudstack.git
CLOUDSTACK-8439:Automation for feature CPU/Memory overcommit
CLOUDSTACK-8439:Automation for feature CPU/Memory overcommit This closes #226
This commit is contained in:
parent
5d9f851deb
commit
896f6b1e5e
|
|
@ -0,0 +1,458 @@
|
|||
# Licensed to the Apache Software Foundation (ASF) under one
|
||||
# or more contributor license agreements. See the NOTICE file
|
||||
# distributed with this work for additional information
|
||||
# regarding copyright ownership. The ASF licenses this file
|
||||
# to you under the Apache License, Version 2.0 (the
|
||||
# "License"); you may not use this file except in compliance
|
||||
# with the License. You may obtain a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing,
|
||||
# software distributed under the License is distributed on an
|
||||
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
||||
# KIND, either express or implied. See the License for the
|
||||
# specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
""" test for feature "CPU and MEM OVERCOMMIT"
|
||||
"""
|
||||
|
||||
from marvin.cloudstackTestCase import cloudstackTestCase
|
||||
from marvin.lib.base import (Account,
|
||||
Capacities,
|
||||
Cluster,
|
||||
Configurations,
|
||||
FAILED,
|
||||
Host,
|
||||
PASS,
|
||||
ServiceOffering,
|
||||
time,
|
||||
VirtualMachine,)
|
||||
from marvin.lib.utils import cleanup_resources, validateList
|
||||
from marvin.lib.common import (get_zone,
|
||||
get_domain,
|
||||
get_template)
|
||||
from marvin.sshClient import SshClient
|
||||
from nose.plugins.attrib import attr
|
||||
|
||||
|
||||
def ssh_xen_host(password, ipaddr, instance_name):
|
||||
"""Ssh into xen host and get vm mem details"""
|
||||
mem = []
|
||||
sshClient = SshClient(
|
||||
ipaddr,
|
||||
22,
|
||||
"root",
|
||||
password
|
||||
)
|
||||
command = "xe vm-list params=all name-label=%s" % instance_name
|
||||
vm_detail = sshClient.execute(command)
|
||||
max_str = vm_detail[17].split(":")
|
||||
min_str = vm_detail[20].split(":")
|
||||
max = int(max_str[1])
|
||||
min = int(min_str[1])
|
||||
mem.append(max)
|
||||
mem.append(min)
|
||||
return mem
|
||||
|
||||
|
||||
def ssh_kvm_host(password, ipaddr, instance_name):
|
||||
"""Ssh into kvm host and get vm mem details"""
|
||||
mem = []
|
||||
sshClient = SshClient(
|
||||
ipaddr,
|
||||
22,
|
||||
"root",
|
||||
password
|
||||
)
|
||||
|
||||
command = "virsh dominfo %s" % instance_name
|
||||
vm_detail = sshClient.execute(command)
|
||||
max = vm_detail[7].split()
|
||||
min = vm_detail[8].split()
|
||||
mem.append(int(max[2]))
|
||||
mem.append(int(min[2]))
|
||||
return mem
|
||||
|
||||
|
||||
def capacity_parser(capacity):
|
||||
cpu = []
|
||||
mem = []
|
||||
cpu.append(capacity[0].capacitytotal)
|
||||
cpu.append(capacity[0].capacityused)
|
||||
cpu.append(capacity[0].percentused)
|
||||
mem.append(capacity[1].capacitytotal)
|
||||
mem.append(capacity[1].capacityused)
|
||||
mem.append(capacity[1].percentused)
|
||||
return cpu, mem
|
||||
|
||||
|
||||
class Overcommit (cloudstackTestCase):
|
||||
|
||||
@classmethod
|
||||
def setUpClass(cls):
|
||||
testClient = super(Overcommit, cls).getClsTestClient()
|
||||
cls.apiclient = testClient.getApiClient()
|
||||
cls.testdata = testClient.getParsedTestDataConfig()
|
||||
# Get Zone,Domain and templates
|
||||
cls.domain = get_domain(cls.apiclient)
|
||||
cls.zone = get_zone(cls.apiclient)
|
||||
cls.testdata["mode"] = cls.zone.networktype
|
||||
cls.testdata["configurableData"]["password"] = "xenroot"
|
||||
cls.hypervisor = testClient.getHypervisorInfo()
|
||||
|
||||
cls.template = get_template(
|
||||
cls.apiclient,
|
||||
cls.zone.id,
|
||||
cls.testdata["ostype"])
|
||||
cls.testdata["template"]["ostypeid"] = cls.template.ostypeid
|
||||
list_conf = Configurations.list(cls.apiclient,
|
||||
name="capacity.check.period"
|
||||
)
|
||||
cls.wait_time = 5 + int(list_conf[0].value) / 1000
|
||||
if cls.template == FAILED:
|
||||
cls.fail(
|
||||
"get_template() failed to return template with description \
|
||||
%s" %
|
||||
cls.testdata["ostype"])
|
||||
cls._cleanup = []
|
||||
try:
|
||||
cls.account = Account.create(cls.apiclient,
|
||||
cls.testdata["account"],
|
||||
domainid=cls.domain.id
|
||||
)
|
||||
cls._cleanup.append(cls.account)
|
||||
|
||||
cls.service_offering = ServiceOffering.create(
|
||||
cls.apiclient,
|
||||
cls.testdata["service_offerings"]["small"])
|
||||
|
||||
cls._cleanup.append(cls.service_offering)
|
||||
|
||||
cls.deployVmResponse = VirtualMachine.create(
|
||||
cls.apiclient,
|
||||
services=cls.testdata["virtual_machine"],
|
||||
accountid=cls.account.name,
|
||||
domainid=cls.account.domainid,
|
||||
serviceofferingid=cls.service_offering.id,
|
||||
templateid=cls.template.id,
|
||||
zoneid=cls.zone.id,
|
||||
)
|
||||
|
||||
except Exception as e:
|
||||
cls.tearDownClass()
|
||||
raise e
|
||||
|
||||
return
|
||||
|
||||
@classmethod
|
||||
def tearDownClass(cls):
|
||||
try:
|
||||
cleanup_resources(cls.apiclient, cls._cleanup)
|
||||
except Exception as e:
|
||||
raise Exception("Warning:Exception during cleanup: %s" % e)
|
||||
|
||||
@attr(
|
||||
tags=[
|
||||
"simulator",
|
||||
"devcloud",
|
||||
"basic",
|
||||
"advanced"],
|
||||
required_hardware="false")
|
||||
def test_01_cluster_settings(self):
|
||||
"""change cpu/mem.overprovisioning.factor at cluster level and
|
||||
verify the change """
|
||||
listHost = Host.list(self.apiclient,
|
||||
id=self.deployVmResponse.hostid
|
||||
)
|
||||
self.assertEqual(
|
||||
validateList(listHost)[0],
|
||||
PASS,
|
||||
"check list host response for host id %s" %
|
||||
self.deployVmResponse.hostid)
|
||||
Configurations.update(self.apiclient,
|
||||
clusterid=listHost[0].clusterid,
|
||||
name="mem.overprovisioning.factor",
|
||||
value=2)
|
||||
|
||||
Configurations.update(self.apiclient,
|
||||
clusterid=listHost[0].clusterid,
|
||||
name="cpu.overprovisioning.factor",
|
||||
value=3)
|
||||
|
||||
list_cluster = Cluster.list(self.apiclient,
|
||||
id=listHost[0].clusterid)
|
||||
self.assertEqual(
|
||||
validateList(list_cluster)[0],
|
||||
PASS,
|
||||
"check list cluster response for cluster id %s" %
|
||||
listHost[0].clusterid)
|
||||
self.assertEqual(int(list_cluster[0].cpuovercommitratio),
|
||||
3,
|
||||
"check the cpu overcommit value at cluster level ")
|
||||
|
||||
self.assertEqual(int(list_cluster[0].memoryovercommitratio),
|
||||
2,
|
||||
"check memory overcommit value at cluster level")
|
||||
|
||||
Configurations.update(self.apiclient,
|
||||
clusterid=listHost[0].clusterid,
|
||||
name="mem.overprovisioning.factor",
|
||||
value=1)
|
||||
|
||||
Configurations.update(self.apiclient,
|
||||
clusterid=listHost[0].clusterid,
|
||||
name="cpu.overprovisioning.factor",
|
||||
value=1)
|
||||
list_cluster1 = Cluster.list(self.apiclient,
|
||||
id=listHost[0].clusterid)
|
||||
self.assertEqual(
|
||||
validateList(list_cluster1)[0],
|
||||
PASS,
|
||||
"check the list cluster response for id %s" %
|
||||
listHost[0].clusterid)
|
||||
self.assertEqual(int(list_cluster1[0].cpuovercommitratio),
|
||||
1,
|
||||
"check the cpu overcommit value at cluster level ")
|
||||
|
||||
self.assertEqual(int(list_cluster1[0].memoryovercommitratio),
|
||||
1,
|
||||
"check memory overcommit value at cluster level")
|
||||
|
||||
@attr(
|
||||
tags=["simulator",
|
||||
"devcloud",
|
||||
"basic",
|
||||
"advanced"],
|
||||
required_hardware="true")
|
||||
def test_02_Overcommit_factor(self):
|
||||
"""change mem.overprovisioning.factor and verify vm memory """
|
||||
|
||||
listHost = Host.list(self.apiclient,
|
||||
id=self.deployVmResponse.hostid
|
||||
)
|
||||
self.assertEqual(
|
||||
validateList(listHost)[0],
|
||||
PASS,
|
||||
"check list host for host id %s" %
|
||||
self.deployVmResponse.hostid)
|
||||
if listHost[0].hypervisor.lower() not in ['kvm', 'xenserver']:
|
||||
self.skipTest(
|
||||
"Skiping test because of not supported hypervisor type %s" %
|
||||
listHost[0].hypervisor)
|
||||
|
||||
Configurations.update(self.apiclient,
|
||||
clusterid=listHost[0].clusterid,
|
||||
name="mem.overprovisioning.factor",
|
||||
value=1)
|
||||
|
||||
self.deployVmResponse.stop(self.apiclient)
|
||||
self.deployVmResponse.start(self.apiclient)
|
||||
|
||||
if listHost[0].hypervisor.lower() == 'xenserver':
|
||||
|
||||
k = ssh_xen_host(
|
||||
self.testdata["configurableData"]["password"],
|
||||
listHost[0].ipaddress,
|
||||
self.deployVmResponse.instancename)
|
||||
|
||||
elif listHost[0].hypervisor.lower() == 'kvm':
|
||||
|
||||
k = ssh_kvm_host(
|
||||
self.testdata["configurableData"]["password"],
|
||||
listHost[0].ipaddress,
|
||||
self.deployVmResponse.instancename)
|
||||
|
||||
self.assertEqual(k[0],
|
||||
k[1],
|
||||
"Check static max ,min on host for overcommit 1 ")
|
||||
|
||||
Configurations.update(self.apiclient,
|
||||
clusterid=listHost[0].clusterid,
|
||||
name="mem.overprovisioning.factor",
|
||||
value=2)
|
||||
|
||||
self.deployVmResponse.stop(self.apiclient)
|
||||
self.deployVmResponse.start(self.apiclient)
|
||||
|
||||
if listHost[0].hypervisor.lower() == 'xenserver':
|
||||
k1 = ssh_xen_host(
|
||||
self.testdata["configurableData"]["password"],
|
||||
listHost[0].ipaddress,
|
||||
self.deployVmResponse.instancename)
|
||||
|
||||
elif listHost[0].hypervisor.lower() == 'kvm':
|
||||
time.sleep(200)
|
||||
k1 = ssh_kvm_host(
|
||||
self.testdata["configurableData"]["password"],
|
||||
listHost[0].ipaddress,
|
||||
self.deployVmResponse.instancename)
|
||||
self.assertEqual(k1[0],
|
||||
2 * k1[1],
|
||||
"Check static max ,min on host for overcommit 2")
|
||||
|
||||
@attr(
|
||||
tags=[
|
||||
"simulator",
|
||||
"devcloud",
|
||||
"basic",
|
||||
"advanced"],
|
||||
required_hardware="false")
|
||||
def test_03_cluste_capacity_check(self):
|
||||
"""change cpu/mem.overprovisioning.factor at cluster level and
|
||||
verify cluster capacity """
|
||||
|
||||
listHost = Host.list(self.apiclient,
|
||||
id=self.deployVmResponse.hostid
|
||||
)
|
||||
self.assertEqual(
|
||||
validateList(listHost)[0],
|
||||
PASS,
|
||||
"check list host for host id %s" %
|
||||
self.deployVmResponse.hostid)
|
||||
|
||||
Configurations.update(self.apiclient,
|
||||
clusterid=listHost[0].clusterid,
|
||||
name="mem.overprovisioning.factor",
|
||||
value=1)
|
||||
Configurations.update(self.apiclient,
|
||||
clusterid=listHost[0].clusterid,
|
||||
name="cpu.overprovisioning.factor",
|
||||
value=1)
|
||||
|
||||
time.sleep(self.wait_time)
|
||||
|
||||
capacity = Capacities.list(self.apiclient,
|
||||
clusterid=listHost[0].clusterid)
|
||||
self.assertEqual(
|
||||
validateList(capacity)[0],
|
||||
PASS,
|
||||
"check list capacity response for cluster id %s" %
|
||||
listHost[0].clusterid)
|
||||
cpu, mem = capacity_parser(capacity)
|
||||
|
||||
Configurations.update(self.apiclient,
|
||||
clusterid=listHost[0].clusterid,
|
||||
name="mem.overprovisioning.factor",
|
||||
value=2)
|
||||
Configurations.update(self.apiclient,
|
||||
clusterid=listHost[0].clusterid,
|
||||
name="cpu.overprovisioning.factor",
|
||||
value=2)
|
||||
|
||||
time.sleep(self.wait_time)
|
||||
|
||||
capacity1 = Capacities.list(self.apiclient,
|
||||
clusterid=listHost[0].clusterid)
|
||||
self.assertEqual(
|
||||
validateList(capacity1)[0],
|
||||
PASS,
|
||||
"check list capacity response for cluster id %s" %
|
||||
listHost[0].clusterid)
|
||||
cpu1, mem1 = capacity_parser(capacity1)
|
||||
self.assertEqual(2 * cpu[0],
|
||||
cpu1[0],
|
||||
"check total capacity ")
|
||||
self.assertEqual(2 * cpu[1],
|
||||
cpu1[1],
|
||||
"check capacity used")
|
||||
self.assertEqual(cpu[2],
|
||||
cpu1[2],
|
||||
"check capacity % used")
|
||||
|
||||
self.assertEqual(2 * mem[0],
|
||||
mem1[0],
|
||||
"check mem total capacity ")
|
||||
self.assertEqual(2 * mem[1],
|
||||
mem1[1],
|
||||
"check mem capacity used")
|
||||
self.assertEqual(mem[2],
|
||||
mem1[2],
|
||||
"check mem capacity % used")
|
||||
|
||||
@attr(
|
||||
tags=[
|
||||
"simulator",
|
||||
"devcloud",
|
||||
"basic",
|
||||
"advanced"],
|
||||
required_hardware="false")
|
||||
def test_04_zone_capacity_check(self):
|
||||
"""change cpu/mem.overprovisioning.factor at cluster level for
|
||||
all cluster in a zone and verify capacity at zone level """
|
||||
list_cluster = Cluster.list(self.apiclient,
|
||||
zoneid=self.zone.id)
|
||||
self.assertEqual(
|
||||
validateList(list_cluster)[0],
|
||||
PASS,
|
||||
"check list cluster response for zone id %s" %
|
||||
self.zone.id)
|
||||
k = len(list_cluster)
|
||||
for id in xrange(k):
|
||||
Configurations.update(self.apiclient,
|
||||
clusterid=list_cluster[id].id,
|
||||
name="mem.overprovisioning.factor",
|
||||
value=1)
|
||||
Configurations.update(self.apiclient,
|
||||
clusterid=list_cluster[id].id,
|
||||
name="cpu.overprovisioning.factor",
|
||||
value=1)
|
||||
|
||||
time.sleep(self.wait_time)
|
||||
|
||||
capacity = Capacities.list(self.apiclient,
|
||||
zoneid=self.zone.id)
|
||||
self.assertEqual(
|
||||
validateList(capacity)[0],
|
||||
PASS,
|
||||
"check list capacity response for zone id %s" %
|
||||
self.zone.id)
|
||||
cpu, mem = capacity_parser(capacity)
|
||||
for id in xrange(k):
|
||||
Configurations.update(self.apiclient,
|
||||
clusterid=list_cluster[id].id,
|
||||
name="mem.overprovisioning.factor",
|
||||
value=2)
|
||||
Configurations.update(self.apiclient,
|
||||
clusterid=list_cluster[id].id,
|
||||
name="cpu.overprovisioning.factor",
|
||||
value=2)
|
||||
|
||||
time.sleep(self.wait_time)
|
||||
|
||||
capacity1 = Capacities.list(self.apiclient,
|
||||
zoneid=self.zone.id)
|
||||
self.assertEqual(validateList(capacity1)[0],
|
||||
PASS,
|
||||
"check list capacity for zone id %s" % self.zone.id)
|
||||
|
||||
cpu1, mem1 = capacity_parser(capacity1)
|
||||
self.assertEqual(2 * cpu[0],
|
||||
cpu1[0],
|
||||
"check total capacity ")
|
||||
self.assertEqual(2 * cpu[1],
|
||||
cpu1[1],
|
||||
"check capacity used")
|
||||
self.assertEqual(cpu[2],
|
||||
cpu1[2],
|
||||
"check capacity % used")
|
||||
|
||||
self.assertEqual(2 * mem[0],
|
||||
mem1[0],
|
||||
"check mem total capacity ")
|
||||
self.assertEqual(2 * mem[1],
|
||||
mem1[1],
|
||||
"check mem capacity used")
|
||||
self.assertEqual(mem[2],
|
||||
mem1[2],
|
||||
"check mem capacity % used")
|
||||
for id in xrange(k):
|
||||
Configurations.update(self.apiclient,
|
||||
clusterid=list_cluster[id].id,
|
||||
name="mem.overprovisioning.factor",
|
||||
value=1)
|
||||
Configurations.update(self.apiclient,
|
||||
clusterid=list_cluster[id].id,
|
||||
name="cpu.overprovisioning.factor",
|
||||
value=1)
|
||||
Loading…
Reference in New Issue