sclp-s390: Add memory hotplug SCLPs
Add memory information to read SCP info and add handlers for Read Storage Element Information, Attach Storage Element, Assign Storage and Unassign Storage. Signed-off-by: Matthew Rosato <mjrosato@linux.vnet.ibm.com> Signed-off-by: Christian Borntraeger <borntraeger@de.ibm.com>
This commit is contained in:
parent
e7f1314f97
commit
1def6656b6
259
hw/s390x/sclp.c
259
hw/s390x/sclp.c
@ -16,7 +16,8 @@
|
||||
#include "sysemu/kvm.h"
|
||||
#include "exec/memory.h"
|
||||
#include "sysemu/sysemu.h"
|
||||
|
||||
#include "exec/address-spaces.h"
|
||||
#include "qemu/config-file.h"
|
||||
#include "hw/s390x/sclp.h"
|
||||
#include "hw/s390x/event-facility.h"
|
||||
|
||||
@ -33,10 +34,19 @@ static inline SCLPEventFacility *get_event_facility(void)
|
||||
static void read_SCP_info(SCCB *sccb)
|
||||
{
|
||||
ReadInfo *read_info = (ReadInfo *) sccb;
|
||||
sclpMemoryHotplugDev *mhd = get_sclp_memory_hotplug_dev();
|
||||
CPUState *cpu;
|
||||
int shift = 0;
|
||||
int cpu_count = 0;
|
||||
int i = 0;
|
||||
int increment_size = 20;
|
||||
int rnsize, rnmax;
|
||||
QemuOpts *opts = qemu_opts_find(qemu_find_opts("memory"), NULL);
|
||||
int slots = qemu_opt_get_number(opts, "slots", 0);
|
||||
int max_avail_slots = s390_get_memslot_count(kvm_state);
|
||||
|
||||
if (slots > max_avail_slots) {
|
||||
slots = max_avail_slots;
|
||||
}
|
||||
|
||||
CPU_FOREACH(cpu) {
|
||||
cpu_count++;
|
||||
@ -54,14 +64,235 @@ static void read_SCP_info(SCCB *sccb)
|
||||
|
||||
read_info->facilities = cpu_to_be64(SCLP_HAS_CPU_INFO);
|
||||
|
||||
while ((ram_size >> (20 + shift)) > 65535) {
|
||||
shift++;
|
||||
/*
|
||||
* The storage increment size is a multiple of 1M and is a power of 2.
|
||||
* The number of storage increments must be MAX_STORAGE_INCREMENTS or fewer.
|
||||
*/
|
||||
while ((ram_size >> increment_size) > MAX_STORAGE_INCREMENTS) {
|
||||
increment_size++;
|
||||
}
|
||||
read_info->rnmax = cpu_to_be16(ram_size >> (20 + shift));
|
||||
read_info->rnsize = 1 << shift;
|
||||
rnmax = ram_size >> increment_size;
|
||||
|
||||
/* Memory Hotplug is only supported for the ccw machine type */
|
||||
if (mhd) {
|
||||
while ((mhd->standby_mem_size >> increment_size) >
|
||||
MAX_STORAGE_INCREMENTS) {
|
||||
increment_size++;
|
||||
}
|
||||
assert(increment_size == mhd->increment_size);
|
||||
|
||||
mhd->standby_subregion_size = MEM_SECTION_SIZE;
|
||||
/* Deduct the memory slot already used for core */
|
||||
if (slots > 0) {
|
||||
while ((mhd->standby_subregion_size * (slots - 1)
|
||||
< mhd->standby_mem_size)) {
|
||||
mhd->standby_subregion_size = mhd->standby_subregion_size << 1;
|
||||
}
|
||||
}
|
||||
/*
|
||||
* Initialize mapping of guest standby memory sections indicating which
|
||||
* are and are not online. Assume all standby memory begins offline.
|
||||
*/
|
||||
if (mhd->standby_state_map == 0) {
|
||||
if (mhd->standby_mem_size % mhd->standby_subregion_size) {
|
||||
mhd->standby_state_map = g_malloc0((mhd->standby_mem_size /
|
||||
mhd->standby_subregion_size + 1) *
|
||||
(mhd->standby_subregion_size /
|
||||
MEM_SECTION_SIZE));
|
||||
} else {
|
||||
mhd->standby_state_map = g_malloc0(mhd->standby_mem_size /
|
||||
MEM_SECTION_SIZE);
|
||||
}
|
||||
}
|
||||
mhd->padded_ram_size = ram_size + mhd->pad_size;
|
||||
mhd->rzm = 1 << mhd->increment_size;
|
||||
rnmax = ((ram_size + mhd->standby_mem_size + mhd->pad_size)
|
||||
>> mhd->increment_size);
|
||||
|
||||
read_info->facilities |= cpu_to_be64(SCLP_FC_ASSIGN_ATTACH_READ_STOR);
|
||||
}
|
||||
|
||||
rnsize = 1 << (increment_size - 20);
|
||||
if (rnsize <= 128) {
|
||||
read_info->rnsize = rnsize;
|
||||
} else {
|
||||
read_info->rnsize = 0;
|
||||
read_info->rnsize2 = cpu_to_be32(rnsize);
|
||||
}
|
||||
|
||||
if (rnmax < 0x10000) {
|
||||
read_info->rnmax = cpu_to_be16(rnmax);
|
||||
} else {
|
||||
read_info->rnmax = cpu_to_be16(0);
|
||||
read_info->rnmax2 = cpu_to_be64(rnmax);
|
||||
}
|
||||
|
||||
sccb->h.response_code = cpu_to_be16(SCLP_RC_NORMAL_READ_COMPLETION);
|
||||
}
|
||||
|
||||
static void read_storage_element0_info(SCCB *sccb)
|
||||
{
|
||||
int i, assigned;
|
||||
int subincrement_id = SCLP_STARTING_SUBINCREMENT_ID;
|
||||
ReadStorageElementInfo *storage_info = (ReadStorageElementInfo *) sccb;
|
||||
sclpMemoryHotplugDev *mhd = get_sclp_memory_hotplug_dev();
|
||||
|
||||
assert(mhd);
|
||||
|
||||
if ((ram_size >> mhd->increment_size) >= 0x10000) {
|
||||
sccb->h.response_code = cpu_to_be16(SCLP_RC_SCCB_BOUNDARY_VIOLATION);
|
||||
return;
|
||||
}
|
||||
|
||||
/* Return information regarding core memory */
|
||||
storage_info->max_id = cpu_to_be16(mhd->standby_mem_size ? 1 : 0);
|
||||
assigned = ram_size >> mhd->increment_size;
|
||||
storage_info->assigned = cpu_to_be16(assigned);
|
||||
|
||||
for (i = 0; i < assigned; i++) {
|
||||
storage_info->entries[i] = cpu_to_be32(subincrement_id);
|
||||
subincrement_id += SCLP_INCREMENT_UNIT;
|
||||
}
|
||||
sccb->h.response_code = cpu_to_be16(SCLP_RC_NORMAL_READ_COMPLETION);
|
||||
}
|
||||
|
||||
static void read_storage_element1_info(SCCB *sccb)
|
||||
{
|
||||
ReadStorageElementInfo *storage_info = (ReadStorageElementInfo *) sccb;
|
||||
sclpMemoryHotplugDev *mhd = get_sclp_memory_hotplug_dev();
|
||||
|
||||
assert(mhd);
|
||||
|
||||
if ((mhd->standby_mem_size >> mhd->increment_size) >= 0x10000) {
|
||||
sccb->h.response_code = cpu_to_be16(SCLP_RC_SCCB_BOUNDARY_VIOLATION);
|
||||
return;
|
||||
}
|
||||
|
||||
/* Return information regarding standby memory */
|
||||
storage_info->max_id = cpu_to_be16(mhd->standby_mem_size ? 1 : 0);
|
||||
storage_info->assigned = cpu_to_be16(mhd->standby_mem_size >>
|
||||
mhd->increment_size);
|
||||
storage_info->standby = cpu_to_be16(mhd->standby_mem_size >>
|
||||
mhd->increment_size);
|
||||
sccb->h.response_code = cpu_to_be16(SCLP_RC_STANDBY_READ_COMPLETION);
|
||||
}
|
||||
|
||||
static void attach_storage_element(SCCB *sccb, uint16_t element)
|
||||
{
|
||||
int i, assigned, subincrement_id;
|
||||
AttachStorageElement *attach_info = (AttachStorageElement *) sccb;
|
||||
sclpMemoryHotplugDev *mhd = get_sclp_memory_hotplug_dev();
|
||||
|
||||
assert(mhd);
|
||||
|
||||
if (element != 1) {
|
||||
sccb->h.response_code = cpu_to_be16(SCLP_RC_INVALID_SCLP_COMMAND);
|
||||
return;
|
||||
}
|
||||
|
||||
assigned = mhd->standby_mem_size >> mhd->increment_size;
|
||||
attach_info->assigned = cpu_to_be16(assigned);
|
||||
subincrement_id = ((ram_size >> mhd->increment_size) << 16)
|
||||
+ SCLP_STARTING_SUBINCREMENT_ID;
|
||||
for (i = 0; i < assigned; i++) {
|
||||
attach_info->entries[i] = cpu_to_be32(subincrement_id);
|
||||
subincrement_id += SCLP_INCREMENT_UNIT;
|
||||
}
|
||||
sccb->h.response_code = cpu_to_be16(SCLP_RC_NORMAL_COMPLETION);
|
||||
}
|
||||
|
||||
static void assign_storage(SCCB *sccb)
|
||||
{
|
||||
MemoryRegion *mr = NULL;
|
||||
uint64_t this_subregion_size;
|
||||
AssignStorage *assign_info = (AssignStorage *) sccb;
|
||||
sclpMemoryHotplugDev *mhd = get_sclp_memory_hotplug_dev();
|
||||
assert(mhd);
|
||||
ram_addr_t assign_addr = (assign_info->rn - 1) * mhd->rzm;
|
||||
MemoryRegion *sysmem = get_system_memory();
|
||||
|
||||
if ((assign_addr % MEM_SECTION_SIZE == 0) &&
|
||||
(assign_addr >= mhd->padded_ram_size)) {
|
||||
/* Re-use existing memory region if found */
|
||||
mr = memory_region_find(sysmem, assign_addr, 1).mr;
|
||||
if (!mr) {
|
||||
|
||||
MemoryRegion *standby_ram = g_new(MemoryRegion, 1);
|
||||
|
||||
/* offset to align to standby_subregion_size for allocation */
|
||||
ram_addr_t offset = assign_addr -
|
||||
(assign_addr - mhd->padded_ram_size)
|
||||
% mhd->standby_subregion_size;
|
||||
|
||||
/* strlen("standby.ram") + 4 (Max of KVM_MEMORY_SLOTS) + NULL */
|
||||
char id[16];
|
||||
snprintf(id, 16, "standby.ram%d",
|
||||
(int)((offset - mhd->padded_ram_size) /
|
||||
mhd->standby_subregion_size) + 1);
|
||||
|
||||
/* Allocate a subregion of the calculated standby_subregion_size */
|
||||
if (offset + mhd->standby_subregion_size >
|
||||
mhd->padded_ram_size + mhd->standby_mem_size) {
|
||||
this_subregion_size = mhd->padded_ram_size +
|
||||
mhd->standby_mem_size - offset;
|
||||
} else {
|
||||
this_subregion_size = mhd->standby_subregion_size;
|
||||
}
|
||||
|
||||
memory_region_init_ram(standby_ram, NULL, id, this_subregion_size);
|
||||
vmstate_register_ram_global(standby_ram);
|
||||
memory_region_add_subregion(sysmem, offset, standby_ram);
|
||||
}
|
||||
/* The specified subregion is no longer in standby */
|
||||
mhd->standby_state_map[(assign_addr - mhd->padded_ram_size)
|
||||
/ MEM_SECTION_SIZE] = 1;
|
||||
}
|
||||
sccb->h.response_code = cpu_to_be16(SCLP_RC_NORMAL_COMPLETION);
|
||||
}
|
||||
|
||||
static void unassign_storage(SCCB *sccb)
|
||||
{
|
||||
MemoryRegion *mr = NULL;
|
||||
AssignStorage *assign_info = (AssignStorage *) sccb;
|
||||
sclpMemoryHotplugDev *mhd = get_sclp_memory_hotplug_dev();
|
||||
assert(mhd);
|
||||
ram_addr_t unassign_addr = (assign_info->rn - 1) * mhd->rzm;
|
||||
MemoryRegion *sysmem = get_system_memory();
|
||||
|
||||
/* if the addr is a multiple of 256 MB */
|
||||
if ((unassign_addr % MEM_SECTION_SIZE == 0) &&
|
||||
(unassign_addr >= mhd->padded_ram_size)) {
|
||||
mhd->standby_state_map[(unassign_addr -
|
||||
mhd->padded_ram_size) / MEM_SECTION_SIZE] = 0;
|
||||
|
||||
/* find the specified memory region and destroy it */
|
||||
mr = memory_region_find(sysmem, unassign_addr, 1).mr;
|
||||
if (mr) {
|
||||
int i;
|
||||
int is_removable = 1;
|
||||
ram_addr_t map_offset = (unassign_addr - mhd->padded_ram_size -
|
||||
(unassign_addr - mhd->padded_ram_size)
|
||||
% mhd->standby_subregion_size);
|
||||
/* Mark all affected subregions as 'standby' once again */
|
||||
for (i = 0;
|
||||
i < (mhd->standby_subregion_size / MEM_SECTION_SIZE);
|
||||
i++) {
|
||||
|
||||
if (mhd->standby_state_map[i + map_offset / MEM_SECTION_SIZE]) {
|
||||
is_removable = 0;
|
||||
break;
|
||||
}
|
||||
}
|
||||
if (is_removable) {
|
||||
memory_region_del_subregion(sysmem, mr);
|
||||
object_unparent(OBJECT(mr));
|
||||
g_free(mr);
|
||||
}
|
||||
}
|
||||
}
|
||||
sccb->h.response_code = cpu_to_be16(SCLP_RC_NORMAL_COMPLETION);
|
||||
}
|
||||
|
||||
/* Provide information about the CPU */
|
||||
static void sclp_read_cpu_info(SCCB *sccb)
|
||||
{
|
||||
@ -103,6 +334,22 @@ static void sclp_execute(SCCB *sccb, uint32_t code)
|
||||
case SCLP_CMDW_READ_CPU_INFO:
|
||||
sclp_read_cpu_info(sccb);
|
||||
break;
|
||||
case SCLP_READ_STORAGE_ELEMENT_INFO:
|
||||
if (code & 0xff00) {
|
||||
read_storage_element1_info(sccb);
|
||||
} else {
|
||||
read_storage_element0_info(sccb);
|
||||
}
|
||||
break;
|
||||
case SCLP_ATTACH_STORAGE_ELEMENT:
|
||||
attach_storage_element(sccb, (code & 0xff00) >> 8);
|
||||
break;
|
||||
case SCLP_ASSIGN_STORAGE:
|
||||
assign_storage(sccb);
|
||||
break;
|
||||
case SCLP_UNASSIGN_STORAGE:
|
||||
unassign_storage(sccb);
|
||||
break;
|
||||
default:
|
||||
efc->command_handler(ef, sccb, code);
|
||||
break;
|
||||
|
@ -1062,6 +1062,7 @@ static inline void cpu_inject_crw_mchk(S390CPU *cpu)
|
||||
|
||||
/* from s390-virtio-ccw */
|
||||
#define MEM_SECTION_SIZE 0x10000000UL
|
||||
#define MAX_AVAIL_SLOTS 32
|
||||
|
||||
/* fpu_helper.c */
|
||||
uint32_t set_cc_nz_f32(float32 v);
|
||||
@ -1085,6 +1086,7 @@ void kvm_s390_enable_css_support(S390CPU *cpu);
|
||||
int kvm_s390_assign_subch_ioeventfd(EventNotifier *notifier, uint32_t sch,
|
||||
int vq, bool assign);
|
||||
int kvm_s390_cpu_restart(S390CPU *cpu);
|
||||
int kvm_s390_get_memslot_count(KVMState *s);
|
||||
void kvm_s390_clear_cmma_callback(void *opaque);
|
||||
#else
|
||||
static inline void kvm_s390_io_interrupt(uint16_t subchannel_id,
|
||||
@ -1112,6 +1114,10 @@ static inline int kvm_s390_cpu_restart(S390CPU *cpu)
|
||||
static inline void kvm_s390_clear_cmma_callback(void *opaque)
|
||||
{
|
||||
}
|
||||
static inline int kvm_s390_get_memslot_count(KVMState *s)
|
||||
{
|
||||
return MAX_AVAIL_SLOTS;
|
||||
}
|
||||
#endif
|
||||
|
||||
static inline void cmma_reset(S390CPU *cpu)
|
||||
@ -1130,6 +1136,15 @@ static inline int s390_cpu_restart(S390CPU *cpu)
|
||||
return -ENOSYS;
|
||||
}
|
||||
|
||||
static inline int s390_get_memslot_count(KVMState *s)
|
||||
{
|
||||
if (kvm_enabled()) {
|
||||
return kvm_s390_get_memslot_count(s);
|
||||
} else {
|
||||
return MAX_AVAIL_SLOTS;
|
||||
}
|
||||
}
|
||||
|
||||
void s390_io_interrupt(uint16_t subchannel_id, uint16_t subchannel_nr,
|
||||
uint32_t io_int_parm, uint32_t io_int_word);
|
||||
void s390_crw_mchk(void);
|
||||
|
@ -1315,3 +1315,8 @@ int kvm_s390_assign_subch_ioeventfd(EventNotifier *notifier, uint32_t sch,
|
||||
}
|
||||
return kvm_vm_ioctl(kvm_state, KVM_IOEVENTFD, &kick);
|
||||
}
|
||||
|
||||
int kvm_s390_get_memslot_count(KVMState *s)
|
||||
{
|
||||
return kvm_check_extension(s, KVM_CAP_NR_MEMSLOTS);
|
||||
}
|
||||
|
Loading…
Reference in New Issue
Block a user