11f070489SIgor Mammedov /* 21f070489SIgor Mammedov * QEMU Host Memory Backend 31f070489SIgor Mammedov * 41f070489SIgor Mammedov * Copyright (C) 2013-2014 Red Hat Inc 51f070489SIgor Mammedov * 61f070489SIgor Mammedov * Authors: 71f070489SIgor Mammedov * Igor Mammedov <imammedo@redhat.com> 81f070489SIgor Mammedov * 91f070489SIgor Mammedov * This work is licensed under the terms of the GNU GPL, version 2 or later. 101f070489SIgor Mammedov * See the COPYING file in the top-level directory. 111f070489SIgor Mammedov */ 129c058332SPeter Maydell #include "qemu/osdep.h" 131f070489SIgor Mammedov #include "sysemu/hostmem.h" 146b269967SEduardo Habkost #include "hw/boards.h" 15da34e65cSMarkus Armbruster #include "qapi/error.h" 161f070489SIgor Mammedov #include "qapi/visitor.h" 174cf1b76bSHu Tao #include "qapi-types.h" 184cf1b76bSHu Tao #include "qapi-visit.h" 191f070489SIgor Mammedov #include "qemu/config-file.h" 201f070489SIgor Mammedov #include "qom/object_interfaces.h" 211f070489SIgor Mammedov 224cf1b76bSHu Tao #ifdef CONFIG_NUMA 234cf1b76bSHu Tao #include <numaif.h> 244cf1b76bSHu Tao QEMU_BUILD_BUG_ON(HOST_MEM_POLICY_DEFAULT != MPOL_DEFAULT); 254cf1b76bSHu Tao QEMU_BUILD_BUG_ON(HOST_MEM_POLICY_PREFERRED != MPOL_PREFERRED); 264cf1b76bSHu Tao QEMU_BUILD_BUG_ON(HOST_MEM_POLICY_BIND != MPOL_BIND); 274cf1b76bSHu Tao QEMU_BUILD_BUG_ON(HOST_MEM_POLICY_INTERLEAVE != MPOL_INTERLEAVE); 284cf1b76bSHu Tao #endif 294cf1b76bSHu Tao 301f070489SIgor Mammedov static void 31d7bce999SEric Blake host_memory_backend_get_size(Object *obj, Visitor *v, const char *name, 32d7bce999SEric Blake void *opaque, Error **errp) 331f070489SIgor Mammedov { 341f070489SIgor Mammedov HostMemoryBackend *backend = MEMORY_BACKEND(obj); 351f070489SIgor Mammedov uint64_t value = backend->size; 361f070489SIgor Mammedov 3751e72bc1SEric Blake visit_type_size(v, name, &value, errp); 381f070489SIgor Mammedov } 391f070489SIgor Mammedov 401f070489SIgor Mammedov static void 41d7bce999SEric Blake host_memory_backend_set_size(Object *obj, Visitor *v, const char *name, 42d7bce999SEric Blake void *opaque, Error **errp) 431f070489SIgor Mammedov { 441f070489SIgor Mammedov HostMemoryBackend *backend = MEMORY_BACKEND(obj); 451f070489SIgor Mammedov Error *local_err = NULL; 461f070489SIgor Mammedov uint64_t value; 471f070489SIgor Mammedov 481f070489SIgor Mammedov if (memory_region_size(&backend->mr)) { 491f070489SIgor Mammedov error_setg(&local_err, "cannot change property value"); 501f070489SIgor Mammedov goto out; 511f070489SIgor Mammedov } 521f070489SIgor Mammedov 5351e72bc1SEric Blake visit_type_size(v, name, &value, &local_err); 541f070489SIgor Mammedov if (local_err) { 551f070489SIgor Mammedov goto out; 561f070489SIgor Mammedov } 571f070489SIgor Mammedov if (!value) { 581f070489SIgor Mammedov error_setg(&local_err, "Property '%s.%s' doesn't take value '%" 591f070489SIgor Mammedov PRIu64 "'", object_get_typename(obj), name, value); 601f070489SIgor Mammedov goto out; 611f070489SIgor Mammedov } 621f070489SIgor Mammedov backend->size = value; 631f070489SIgor Mammedov out: 641f070489SIgor Mammedov error_propagate(errp, local_err); 651f070489SIgor Mammedov } 661f070489SIgor Mammedov 674cf1b76bSHu Tao static void 68d7bce999SEric Blake host_memory_backend_get_host_nodes(Object *obj, Visitor *v, const char *name, 69d7bce999SEric Blake void *opaque, Error **errp) 704cf1b76bSHu Tao { 714cf1b76bSHu Tao HostMemoryBackend *backend = MEMORY_BACKEND(obj); 724cf1b76bSHu Tao uint16List *host_nodes = NULL; 734cf1b76bSHu Tao uint16List **node = &host_nodes; 744cf1b76bSHu Tao unsigned long value; 754cf1b76bSHu Tao 764cf1b76bSHu Tao value = find_first_bit(backend->host_nodes, MAX_NODES); 771454d33fSXiao Guangrong if (value == MAX_NODES) { 78*658ae5a7SMarkus Armbruster return; 791454d33fSXiao Guangrong } 804cf1b76bSHu Tao 81*658ae5a7SMarkus Armbruster *node = g_malloc0(sizeof(**node)); 82*658ae5a7SMarkus Armbruster (*node)->value = value; 83*658ae5a7SMarkus Armbruster node = &(*node)->next; 84*658ae5a7SMarkus Armbruster 854cf1b76bSHu Tao do { 864cf1b76bSHu Tao value = find_next_bit(backend->host_nodes, MAX_NODES, value + 1); 874cf1b76bSHu Tao if (value == MAX_NODES) { 884cf1b76bSHu Tao break; 894cf1b76bSHu Tao } 904cf1b76bSHu Tao 91*658ae5a7SMarkus Armbruster *node = g_malloc0(sizeof(**node)); 92*658ae5a7SMarkus Armbruster (*node)->value = value; 93*658ae5a7SMarkus Armbruster node = &(*node)->next; 944cf1b76bSHu Tao } while (true); 954cf1b76bSHu Tao 9651e72bc1SEric Blake visit_type_uint16List(v, name, &host_nodes, errp); 974cf1b76bSHu Tao } 984cf1b76bSHu Tao 994cf1b76bSHu Tao static void 100d7bce999SEric Blake host_memory_backend_set_host_nodes(Object *obj, Visitor *v, const char *name, 101d7bce999SEric Blake void *opaque, Error **errp) 1024cf1b76bSHu Tao { 1034cf1b76bSHu Tao #ifdef CONFIG_NUMA 1044cf1b76bSHu Tao HostMemoryBackend *backend = MEMORY_BACKEND(obj); 1054cf1b76bSHu Tao uint16List *l = NULL; 1064cf1b76bSHu Tao 10751e72bc1SEric Blake visit_type_uint16List(v, name, &l, errp); 1084cf1b76bSHu Tao 1094cf1b76bSHu Tao while (l) { 1104cf1b76bSHu Tao bitmap_set(backend->host_nodes, l->value, 1); 1114cf1b76bSHu Tao l = l->next; 1124cf1b76bSHu Tao } 1134cf1b76bSHu Tao #else 1144cf1b76bSHu Tao error_setg(errp, "NUMA node binding are not supported by this QEMU"); 1154cf1b76bSHu Tao #endif 1164cf1b76bSHu Tao } 1174cf1b76bSHu Tao 118a3590dacSDaniel P. Berrange static int 119a3590dacSDaniel P. Berrange host_memory_backend_get_policy(Object *obj, Error **errp G_GNUC_UNUSED) 1204cf1b76bSHu Tao { 1214cf1b76bSHu Tao HostMemoryBackend *backend = MEMORY_BACKEND(obj); 122a3590dacSDaniel P. Berrange return backend->policy; 1234cf1b76bSHu Tao } 1244cf1b76bSHu Tao 1254cf1b76bSHu Tao static void 126a3590dacSDaniel P. Berrange host_memory_backend_set_policy(Object *obj, int policy, Error **errp) 1274cf1b76bSHu Tao { 1284cf1b76bSHu Tao HostMemoryBackend *backend = MEMORY_BACKEND(obj); 1294cf1b76bSHu Tao backend->policy = policy; 1304cf1b76bSHu Tao 1314cf1b76bSHu Tao #ifndef CONFIG_NUMA 1324cf1b76bSHu Tao if (policy != HOST_MEM_POLICY_DEFAULT) { 1334cf1b76bSHu Tao error_setg(errp, "NUMA policies are not supported by this QEMU"); 1344cf1b76bSHu Tao } 1354cf1b76bSHu Tao #endif 1364cf1b76bSHu Tao } 1374cf1b76bSHu Tao 138605d0a94SPaolo Bonzini static bool host_memory_backend_get_merge(Object *obj, Error **errp) 139605d0a94SPaolo Bonzini { 140605d0a94SPaolo Bonzini HostMemoryBackend *backend = MEMORY_BACKEND(obj); 141605d0a94SPaolo Bonzini 142605d0a94SPaolo Bonzini return backend->merge; 143605d0a94SPaolo Bonzini } 144605d0a94SPaolo Bonzini 145605d0a94SPaolo Bonzini static void host_memory_backend_set_merge(Object *obj, bool value, Error **errp) 146605d0a94SPaolo Bonzini { 147605d0a94SPaolo Bonzini HostMemoryBackend *backend = MEMORY_BACKEND(obj); 148605d0a94SPaolo Bonzini 149605d0a94SPaolo Bonzini if (!memory_region_size(&backend->mr)) { 150605d0a94SPaolo Bonzini backend->merge = value; 151605d0a94SPaolo Bonzini return; 152605d0a94SPaolo Bonzini } 153605d0a94SPaolo Bonzini 154605d0a94SPaolo Bonzini if (value != backend->merge) { 155605d0a94SPaolo Bonzini void *ptr = memory_region_get_ram_ptr(&backend->mr); 156605d0a94SPaolo Bonzini uint64_t sz = memory_region_size(&backend->mr); 157605d0a94SPaolo Bonzini 158605d0a94SPaolo Bonzini qemu_madvise(ptr, sz, 159605d0a94SPaolo Bonzini value ? QEMU_MADV_MERGEABLE : QEMU_MADV_UNMERGEABLE); 160605d0a94SPaolo Bonzini backend->merge = value; 161605d0a94SPaolo Bonzini } 162605d0a94SPaolo Bonzini } 163605d0a94SPaolo Bonzini 164605d0a94SPaolo Bonzini static bool host_memory_backend_get_dump(Object *obj, Error **errp) 165605d0a94SPaolo Bonzini { 166605d0a94SPaolo Bonzini HostMemoryBackend *backend = MEMORY_BACKEND(obj); 167605d0a94SPaolo Bonzini 168605d0a94SPaolo Bonzini return backend->dump; 169605d0a94SPaolo Bonzini } 170605d0a94SPaolo Bonzini 171605d0a94SPaolo Bonzini static void host_memory_backend_set_dump(Object *obj, bool value, Error **errp) 172605d0a94SPaolo Bonzini { 173605d0a94SPaolo Bonzini HostMemoryBackend *backend = MEMORY_BACKEND(obj); 174605d0a94SPaolo Bonzini 175605d0a94SPaolo Bonzini if (!memory_region_size(&backend->mr)) { 176605d0a94SPaolo Bonzini backend->dump = value; 177605d0a94SPaolo Bonzini return; 178605d0a94SPaolo Bonzini } 179605d0a94SPaolo Bonzini 180605d0a94SPaolo Bonzini if (value != backend->dump) { 181605d0a94SPaolo Bonzini void *ptr = memory_region_get_ram_ptr(&backend->mr); 182605d0a94SPaolo Bonzini uint64_t sz = memory_region_size(&backend->mr); 183605d0a94SPaolo Bonzini 184605d0a94SPaolo Bonzini qemu_madvise(ptr, sz, 185605d0a94SPaolo Bonzini value ? QEMU_MADV_DODUMP : QEMU_MADV_DONTDUMP); 186605d0a94SPaolo Bonzini backend->dump = value; 187605d0a94SPaolo Bonzini } 188605d0a94SPaolo Bonzini } 189605d0a94SPaolo Bonzini 190a35ba7beSPaolo Bonzini static bool host_memory_backend_get_prealloc(Object *obj, Error **errp) 191a35ba7beSPaolo Bonzini { 192a35ba7beSPaolo Bonzini HostMemoryBackend *backend = MEMORY_BACKEND(obj); 193a35ba7beSPaolo Bonzini 194a35ba7beSPaolo Bonzini return backend->prealloc || backend->force_prealloc; 195a35ba7beSPaolo Bonzini } 196a35ba7beSPaolo Bonzini 197a35ba7beSPaolo Bonzini static void host_memory_backend_set_prealloc(Object *obj, bool value, 198a35ba7beSPaolo Bonzini Error **errp) 199a35ba7beSPaolo Bonzini { 200056b68afSIgor Mammedov Error *local_err = NULL; 201a35ba7beSPaolo Bonzini HostMemoryBackend *backend = MEMORY_BACKEND(obj); 202a35ba7beSPaolo Bonzini 203a35ba7beSPaolo Bonzini if (backend->force_prealloc) { 204a35ba7beSPaolo Bonzini if (value) { 205a35ba7beSPaolo Bonzini error_setg(errp, 206a35ba7beSPaolo Bonzini "remove -mem-prealloc to use the prealloc property"); 207a35ba7beSPaolo Bonzini return; 208a35ba7beSPaolo Bonzini } 209a35ba7beSPaolo Bonzini } 210a35ba7beSPaolo Bonzini 211a35ba7beSPaolo Bonzini if (!memory_region_size(&backend->mr)) { 212a35ba7beSPaolo Bonzini backend->prealloc = value; 213a35ba7beSPaolo Bonzini return; 214a35ba7beSPaolo Bonzini } 215a35ba7beSPaolo Bonzini 216a35ba7beSPaolo Bonzini if (value && !backend->prealloc) { 217a35ba7beSPaolo Bonzini int fd = memory_region_get_fd(&backend->mr); 218a35ba7beSPaolo Bonzini void *ptr = memory_region_get_ram_ptr(&backend->mr); 219a35ba7beSPaolo Bonzini uint64_t sz = memory_region_size(&backend->mr); 220a35ba7beSPaolo Bonzini 2211e356fc1SJitendra Kolhe os_mem_prealloc(fd, ptr, sz, smp_cpus, &local_err); 222056b68afSIgor Mammedov if (local_err) { 223056b68afSIgor Mammedov error_propagate(errp, local_err); 224056b68afSIgor Mammedov return; 225056b68afSIgor Mammedov } 226a35ba7beSPaolo Bonzini backend->prealloc = true; 227a35ba7beSPaolo Bonzini } 228a35ba7beSPaolo Bonzini } 229a35ba7beSPaolo Bonzini 23058f4662cSHu Tao static void host_memory_backend_init(Object *obj) 2311f070489SIgor Mammedov { 232605d0a94SPaolo Bonzini HostMemoryBackend *backend = MEMORY_BACKEND(obj); 2336b269967SEduardo Habkost MachineState *machine = MACHINE(qdev_get_machine()); 234605d0a94SPaolo Bonzini 2356b269967SEduardo Habkost backend->merge = machine_mem_merge(machine); 2366b269967SEduardo Habkost backend->dump = machine_dump_guest_core(machine); 237a35ba7beSPaolo Bonzini backend->prealloc = mem_prealloc; 2381f070489SIgor Mammedov } 2391f070489SIgor Mammedov 2401f070489SIgor Mammedov MemoryRegion * 2411f070489SIgor Mammedov host_memory_backend_get_memory(HostMemoryBackend *backend, Error **errp) 2421f070489SIgor Mammedov { 2431f070489SIgor Mammedov return memory_region_size(&backend->mr) ? &backend->mr : NULL; 2441f070489SIgor Mammedov } 2451f070489SIgor Mammedov 2462aece63cSXiao Guangrong void host_memory_backend_set_mapped(HostMemoryBackend *backend, bool mapped) 2472aece63cSXiao Guangrong { 2482aece63cSXiao Guangrong backend->is_mapped = mapped; 2492aece63cSXiao Guangrong } 2502aece63cSXiao Guangrong 2512aece63cSXiao Guangrong bool host_memory_backend_is_mapped(HostMemoryBackend *backend) 2522aece63cSXiao Guangrong { 2532aece63cSXiao Guangrong return backend->is_mapped; 2542aece63cSXiao Guangrong } 2552aece63cSXiao Guangrong 256bd9262d9SHu Tao static void 257bd9262d9SHu Tao host_memory_backend_memory_complete(UserCreatable *uc, Error **errp) 258bd9262d9SHu Tao { 259bd9262d9SHu Tao HostMemoryBackend *backend = MEMORY_BACKEND(uc); 260bd9262d9SHu Tao HostMemoryBackendClass *bc = MEMORY_BACKEND_GET_CLASS(uc); 261605d0a94SPaolo Bonzini Error *local_err = NULL; 262605d0a94SPaolo Bonzini void *ptr; 263605d0a94SPaolo Bonzini uint64_t sz; 264bd9262d9SHu Tao 265bd9262d9SHu Tao if (bc->alloc) { 266605d0a94SPaolo Bonzini bc->alloc(backend, &local_err); 267605d0a94SPaolo Bonzini if (local_err) { 268056b68afSIgor Mammedov goto out; 269605d0a94SPaolo Bonzini } 270605d0a94SPaolo Bonzini 271605d0a94SPaolo Bonzini ptr = memory_region_get_ram_ptr(&backend->mr); 272605d0a94SPaolo Bonzini sz = memory_region_size(&backend->mr); 273605d0a94SPaolo Bonzini 274605d0a94SPaolo Bonzini if (backend->merge) { 275605d0a94SPaolo Bonzini qemu_madvise(ptr, sz, QEMU_MADV_MERGEABLE); 276605d0a94SPaolo Bonzini } 277605d0a94SPaolo Bonzini if (!backend->dump) { 278605d0a94SPaolo Bonzini qemu_madvise(ptr, sz, QEMU_MADV_DONTDUMP); 279605d0a94SPaolo Bonzini } 2804cf1b76bSHu Tao #ifdef CONFIG_NUMA 2814cf1b76bSHu Tao unsigned long lastbit = find_last_bit(backend->host_nodes, MAX_NODES); 2824cf1b76bSHu Tao /* lastbit == MAX_NODES means maxnode = 0 */ 2834cf1b76bSHu Tao unsigned long maxnode = (lastbit + 1) % (MAX_NODES + 1); 2844cf1b76bSHu Tao /* ensure policy won't be ignored in case memory is preallocated 2854cf1b76bSHu Tao * before mbind(). note: MPOL_MF_STRICT is ignored on hugepages so 2864cf1b76bSHu Tao * this doesn't catch hugepage case. */ 287288d3322SMichael S. Tsirkin unsigned flags = MPOL_MF_STRICT | MPOL_MF_MOVE; 2884cf1b76bSHu Tao 2894cf1b76bSHu Tao /* check for invalid host-nodes and policies and give more verbose 2904cf1b76bSHu Tao * error messages than mbind(). */ 2914cf1b76bSHu Tao if (maxnode && backend->policy == MPOL_DEFAULT) { 2924cf1b76bSHu Tao error_setg(errp, "host-nodes must be empty for policy default," 2934cf1b76bSHu Tao " or you should explicitly specify a policy other" 2944cf1b76bSHu Tao " than default"); 2954cf1b76bSHu Tao return; 2964cf1b76bSHu Tao } else if (maxnode == 0 && backend->policy != MPOL_DEFAULT) { 2974cf1b76bSHu Tao error_setg(errp, "host-nodes must be set for policy %s", 2984cf1b76bSHu Tao HostMemPolicy_lookup[backend->policy]); 2994cf1b76bSHu Tao return; 3004cf1b76bSHu Tao } 3014cf1b76bSHu Tao 3024cf1b76bSHu Tao /* We can have up to MAX_NODES nodes, but we need to pass maxnode+1 3034cf1b76bSHu Tao * as argument to mbind() due to an old Linux bug (feature?) which 3044cf1b76bSHu Tao * cuts off the last specified node. This means backend->host_nodes 3054cf1b76bSHu Tao * must have MAX_NODES+1 bits available. 3064cf1b76bSHu Tao */ 3074cf1b76bSHu Tao assert(sizeof(backend->host_nodes) >= 3084cf1b76bSHu Tao BITS_TO_LONGS(MAX_NODES + 1) * sizeof(unsigned long)); 3094cf1b76bSHu Tao assert(maxnode <= MAX_NODES); 3104cf1b76bSHu Tao if (mbind(ptr, sz, backend->policy, 3114cf1b76bSHu Tao maxnode ? backend->host_nodes : NULL, maxnode + 1, flags)) { 312a3567ba1SPavel Fedin if (backend->policy != MPOL_DEFAULT || errno != ENOSYS) { 3134cf1b76bSHu Tao error_setg_errno(errp, errno, 3144cf1b76bSHu Tao "cannot bind memory to host NUMA nodes"); 3154cf1b76bSHu Tao return; 3164cf1b76bSHu Tao } 317a3567ba1SPavel Fedin } 3184cf1b76bSHu Tao #endif 3194cf1b76bSHu Tao /* Preallocate memory after the NUMA policy has been instantiated. 3204cf1b76bSHu Tao * This is necessary to guarantee memory is allocated with 3214cf1b76bSHu Tao * specified NUMA policy in place. 3224cf1b76bSHu Tao */ 323a35ba7beSPaolo Bonzini if (backend->prealloc) { 324056b68afSIgor Mammedov os_mem_prealloc(memory_region_get_fd(&backend->mr), ptr, sz, 3251e356fc1SJitendra Kolhe smp_cpus, &local_err); 326056b68afSIgor Mammedov if (local_err) { 327056b68afSIgor Mammedov goto out; 328a35ba7beSPaolo Bonzini } 329bd9262d9SHu Tao } 330bd9262d9SHu Tao } 331056b68afSIgor Mammedov out: 332056b68afSIgor Mammedov error_propagate(errp, local_err); 333056b68afSIgor Mammedov } 334bd9262d9SHu Tao 33536bce5caSLin Ma static bool 33636bce5caSLin Ma host_memory_backend_can_be_deleted(UserCreatable *uc, Error **errp) 33736bce5caSLin Ma { 3382aece63cSXiao Guangrong if (host_memory_backend_is_mapped(MEMORY_BACKEND(uc))) { 33936bce5caSLin Ma return false; 34036bce5caSLin Ma } else { 34136bce5caSLin Ma return true; 34236bce5caSLin Ma } 34336bce5caSLin Ma } 34436bce5caSLin Ma 345e1ff3c67SIgor Mammedov static char *get_id(Object *o, Error **errp) 346e1ff3c67SIgor Mammedov { 347e1ff3c67SIgor Mammedov HostMemoryBackend *backend = MEMORY_BACKEND(o); 348e1ff3c67SIgor Mammedov 349e1ff3c67SIgor Mammedov return g_strdup(backend->id); 350e1ff3c67SIgor Mammedov } 351e1ff3c67SIgor Mammedov 352e1ff3c67SIgor Mammedov static void set_id(Object *o, const char *str, Error **errp) 353e1ff3c67SIgor Mammedov { 354e1ff3c67SIgor Mammedov HostMemoryBackend *backend = MEMORY_BACKEND(o); 355e1ff3c67SIgor Mammedov 356e1ff3c67SIgor Mammedov if (backend->id) { 357e1ff3c67SIgor Mammedov error_setg(errp, "cannot change property value"); 358e1ff3c67SIgor Mammedov return; 359e1ff3c67SIgor Mammedov } 360e1ff3c67SIgor Mammedov backend->id = g_strdup(str); 361e1ff3c67SIgor Mammedov } 362e1ff3c67SIgor Mammedov 363bd9262d9SHu Tao static void 364bd9262d9SHu Tao host_memory_backend_class_init(ObjectClass *oc, void *data) 365bd9262d9SHu Tao { 366bd9262d9SHu Tao UserCreatableClass *ucc = USER_CREATABLE_CLASS(oc); 367bd9262d9SHu Tao 368bd9262d9SHu Tao ucc->complete = host_memory_backend_memory_complete; 36936bce5caSLin Ma ucc->can_be_deleted = host_memory_backend_can_be_deleted; 370e62834caSEduardo Habkost 371e62834caSEduardo Habkost object_class_property_add_bool(oc, "merge", 372e62834caSEduardo Habkost host_memory_backend_get_merge, 373e62834caSEduardo Habkost host_memory_backend_set_merge, &error_abort); 374e62834caSEduardo Habkost object_class_property_add_bool(oc, "dump", 375e62834caSEduardo Habkost host_memory_backend_get_dump, 376e62834caSEduardo Habkost host_memory_backend_set_dump, &error_abort); 377e62834caSEduardo Habkost object_class_property_add_bool(oc, "prealloc", 378e62834caSEduardo Habkost host_memory_backend_get_prealloc, 379e62834caSEduardo Habkost host_memory_backend_set_prealloc, &error_abort); 380e62834caSEduardo Habkost object_class_property_add(oc, "size", "int", 381e62834caSEduardo Habkost host_memory_backend_get_size, 382e62834caSEduardo Habkost host_memory_backend_set_size, 383e62834caSEduardo Habkost NULL, NULL, &error_abort); 384e62834caSEduardo Habkost object_class_property_add(oc, "host-nodes", "int", 385e62834caSEduardo Habkost host_memory_backend_get_host_nodes, 386e62834caSEduardo Habkost host_memory_backend_set_host_nodes, 387e62834caSEduardo Habkost NULL, NULL, &error_abort); 388e62834caSEduardo Habkost object_class_property_add_enum(oc, "policy", "HostMemPolicy", 389e62834caSEduardo Habkost HostMemPolicy_lookup, 390e62834caSEduardo Habkost host_memory_backend_get_policy, 391e62834caSEduardo Habkost host_memory_backend_set_policy, &error_abort); 392e1ff3c67SIgor Mammedov object_class_property_add_str(oc, "id", get_id, set_id, &error_abort); 393e1ff3c67SIgor Mammedov } 394e1ff3c67SIgor Mammedov 395e1ff3c67SIgor Mammedov static void host_memory_backend_finalize(Object *o) 396e1ff3c67SIgor Mammedov { 397e1ff3c67SIgor Mammedov HostMemoryBackend *backend = MEMORY_BACKEND(o); 398e1ff3c67SIgor Mammedov g_free(backend->id); 399bd9262d9SHu Tao } 400bd9262d9SHu Tao 40158f4662cSHu Tao static const TypeInfo host_memory_backend_info = { 4021f070489SIgor Mammedov .name = TYPE_MEMORY_BACKEND, 4031f070489SIgor Mammedov .parent = TYPE_OBJECT, 4041f070489SIgor Mammedov .abstract = true, 4051f070489SIgor Mammedov .class_size = sizeof(HostMemoryBackendClass), 406bd9262d9SHu Tao .class_init = host_memory_backend_class_init, 4071f070489SIgor Mammedov .instance_size = sizeof(HostMemoryBackend), 40858f4662cSHu Tao .instance_init = host_memory_backend_init, 409e1ff3c67SIgor Mammedov .instance_finalize = host_memory_backend_finalize, 4101f070489SIgor Mammedov .interfaces = (InterfaceInfo[]) { 4111f070489SIgor Mammedov { TYPE_USER_CREATABLE }, 4121f070489SIgor Mammedov { } 4131f070489SIgor Mammedov } 4141f070489SIgor Mammedov }; 4151f070489SIgor Mammedov 4161f070489SIgor Mammedov static void register_types(void) 4171f070489SIgor Mammedov { 41858f4662cSHu Tao type_register_static(&host_memory_backend_info); 4191f070489SIgor Mammedov } 4201f070489SIgor Mammedov 4211f070489SIgor Mammedov type_init(register_types); 422