11f070489SIgor Mammedov /* 21f070489SIgor Mammedov * QEMU Host Memory Backend 31f070489SIgor Mammedov * 41f070489SIgor Mammedov * Copyright (C) 2013-2014 Red Hat Inc 51f070489SIgor Mammedov * 61f070489SIgor Mammedov * Authors: 71f070489SIgor Mammedov * Igor Mammedov <imammedo@redhat.com> 81f070489SIgor Mammedov * 91f070489SIgor Mammedov * This work is licensed under the terms of the GNU GPL, version 2 or later. 101f070489SIgor Mammedov * See the COPYING file in the top-level directory. 111f070489SIgor Mammedov */ 129c058332SPeter Maydell #include "qemu/osdep.h" 131f070489SIgor Mammedov #include "sysemu/hostmem.h" 146b269967SEduardo Habkost #include "hw/boards.h" 15da34e65cSMarkus Armbruster #include "qapi/error.h" 161f070489SIgor Mammedov #include "qapi/visitor.h" 174cf1b76bSHu Tao #include "qapi-types.h" 184cf1b76bSHu Tao #include "qapi-visit.h" 191f070489SIgor Mammedov #include "qemu/config-file.h" 201f070489SIgor Mammedov #include "qom/object_interfaces.h" 211f070489SIgor Mammedov 224cf1b76bSHu Tao #ifdef CONFIG_NUMA 234cf1b76bSHu Tao #include <numaif.h> 244cf1b76bSHu Tao QEMU_BUILD_BUG_ON(HOST_MEM_POLICY_DEFAULT != MPOL_DEFAULT); 254cf1b76bSHu Tao QEMU_BUILD_BUG_ON(HOST_MEM_POLICY_PREFERRED != MPOL_PREFERRED); 264cf1b76bSHu Tao QEMU_BUILD_BUG_ON(HOST_MEM_POLICY_BIND != MPOL_BIND); 274cf1b76bSHu Tao QEMU_BUILD_BUG_ON(HOST_MEM_POLICY_INTERLEAVE != MPOL_INTERLEAVE); 284cf1b76bSHu Tao #endif 294cf1b76bSHu Tao 301f070489SIgor Mammedov static void 31d7bce999SEric Blake host_memory_backend_get_size(Object *obj, Visitor *v, const char *name, 32d7bce999SEric Blake void *opaque, Error **errp) 331f070489SIgor Mammedov { 341f070489SIgor Mammedov HostMemoryBackend *backend = MEMORY_BACKEND(obj); 351f070489SIgor Mammedov uint64_t value = backend->size; 361f070489SIgor Mammedov 3751e72bc1SEric Blake visit_type_size(v, name, &value, errp); 381f070489SIgor Mammedov } 391f070489SIgor Mammedov 401f070489SIgor Mammedov static void 41d7bce999SEric Blake host_memory_backend_set_size(Object *obj, Visitor *v, const char *name, 42d7bce999SEric Blake void *opaque, Error **errp) 431f070489SIgor Mammedov { 441f070489SIgor Mammedov HostMemoryBackend *backend = MEMORY_BACKEND(obj); 451f070489SIgor Mammedov Error *local_err = NULL; 461f070489SIgor Mammedov uint64_t value; 471f070489SIgor Mammedov 48*6f4c60e4SPeter Xu if (host_memory_backend_mr_inited(backend)) { 491f070489SIgor Mammedov error_setg(&local_err, "cannot change property value"); 501f070489SIgor Mammedov goto out; 511f070489SIgor Mammedov } 521f070489SIgor Mammedov 5351e72bc1SEric Blake visit_type_size(v, name, &value, &local_err); 541f070489SIgor Mammedov if (local_err) { 551f070489SIgor Mammedov goto out; 561f070489SIgor Mammedov } 571f070489SIgor Mammedov if (!value) { 581f070489SIgor Mammedov error_setg(&local_err, "Property '%s.%s' doesn't take value '%" 591f070489SIgor Mammedov PRIu64 "'", object_get_typename(obj), name, value); 601f070489SIgor Mammedov goto out; 611f070489SIgor Mammedov } 621f070489SIgor Mammedov backend->size = value; 631f070489SIgor Mammedov out: 641f070489SIgor Mammedov error_propagate(errp, local_err); 651f070489SIgor Mammedov } 661f070489SIgor Mammedov 674cf1b76bSHu Tao static void 68d7bce999SEric Blake host_memory_backend_get_host_nodes(Object *obj, Visitor *v, const char *name, 69d7bce999SEric Blake void *opaque, Error **errp) 704cf1b76bSHu Tao { 714cf1b76bSHu Tao HostMemoryBackend *backend = MEMORY_BACKEND(obj); 724cf1b76bSHu Tao uint16List *host_nodes = NULL; 734cf1b76bSHu Tao uint16List **node = &host_nodes; 744cf1b76bSHu Tao unsigned long value; 754cf1b76bSHu Tao 764cf1b76bSHu Tao value = find_first_bit(backend->host_nodes, MAX_NODES); 771454d33fSXiao Guangrong if (value == MAX_NODES) { 78658ae5a7SMarkus Armbruster return; 791454d33fSXiao Guangrong } 804cf1b76bSHu Tao 81658ae5a7SMarkus Armbruster *node = g_malloc0(sizeof(**node)); 82658ae5a7SMarkus Armbruster (*node)->value = value; 83658ae5a7SMarkus Armbruster node = &(*node)->next; 84658ae5a7SMarkus Armbruster 854cf1b76bSHu Tao do { 864cf1b76bSHu Tao value = find_next_bit(backend->host_nodes, MAX_NODES, value + 1); 874cf1b76bSHu Tao if (value == MAX_NODES) { 884cf1b76bSHu Tao break; 894cf1b76bSHu Tao } 904cf1b76bSHu Tao 91658ae5a7SMarkus Armbruster *node = g_malloc0(sizeof(**node)); 92658ae5a7SMarkus Armbruster (*node)->value = value; 93658ae5a7SMarkus Armbruster node = &(*node)->next; 944cf1b76bSHu Tao } while (true); 954cf1b76bSHu Tao 9651e72bc1SEric Blake visit_type_uint16List(v, name, &host_nodes, errp); 974cf1b76bSHu Tao } 984cf1b76bSHu Tao 994cf1b76bSHu Tao static void 100d7bce999SEric Blake host_memory_backend_set_host_nodes(Object *obj, Visitor *v, const char *name, 101d7bce999SEric Blake void *opaque, Error **errp) 1024cf1b76bSHu Tao { 1034cf1b76bSHu Tao #ifdef CONFIG_NUMA 1044cf1b76bSHu Tao HostMemoryBackend *backend = MEMORY_BACKEND(obj); 1054cf1b76bSHu Tao uint16List *l = NULL; 1064cf1b76bSHu Tao 10751e72bc1SEric Blake visit_type_uint16List(v, name, &l, errp); 1084cf1b76bSHu Tao 1094cf1b76bSHu Tao while (l) { 1104cf1b76bSHu Tao bitmap_set(backend->host_nodes, l->value, 1); 1114cf1b76bSHu Tao l = l->next; 1124cf1b76bSHu Tao } 1134cf1b76bSHu Tao #else 1144cf1b76bSHu Tao error_setg(errp, "NUMA node binding are not supported by this QEMU"); 1154cf1b76bSHu Tao #endif 1164cf1b76bSHu Tao } 1174cf1b76bSHu Tao 118a3590dacSDaniel P. Berrange static int 119a3590dacSDaniel P. Berrange host_memory_backend_get_policy(Object *obj, Error **errp G_GNUC_UNUSED) 1204cf1b76bSHu Tao { 1214cf1b76bSHu Tao HostMemoryBackend *backend = MEMORY_BACKEND(obj); 122a3590dacSDaniel P. Berrange return backend->policy; 1234cf1b76bSHu Tao } 1244cf1b76bSHu Tao 1254cf1b76bSHu Tao static void 126a3590dacSDaniel P. Berrange host_memory_backend_set_policy(Object *obj, int policy, Error **errp) 1274cf1b76bSHu Tao { 1284cf1b76bSHu Tao HostMemoryBackend *backend = MEMORY_BACKEND(obj); 1294cf1b76bSHu Tao backend->policy = policy; 1304cf1b76bSHu Tao 1314cf1b76bSHu Tao #ifndef CONFIG_NUMA 1324cf1b76bSHu Tao if (policy != HOST_MEM_POLICY_DEFAULT) { 1334cf1b76bSHu Tao error_setg(errp, "NUMA policies are not supported by this QEMU"); 1344cf1b76bSHu Tao } 1354cf1b76bSHu Tao #endif 1364cf1b76bSHu Tao } 1374cf1b76bSHu Tao 138605d0a94SPaolo Bonzini static bool host_memory_backend_get_merge(Object *obj, Error **errp) 139605d0a94SPaolo Bonzini { 140605d0a94SPaolo Bonzini HostMemoryBackend *backend = MEMORY_BACKEND(obj); 141605d0a94SPaolo Bonzini 142605d0a94SPaolo Bonzini return backend->merge; 143605d0a94SPaolo Bonzini } 144605d0a94SPaolo Bonzini 145605d0a94SPaolo Bonzini static void host_memory_backend_set_merge(Object *obj, bool value, Error **errp) 146605d0a94SPaolo Bonzini { 147605d0a94SPaolo Bonzini HostMemoryBackend *backend = MEMORY_BACKEND(obj); 148605d0a94SPaolo Bonzini 149*6f4c60e4SPeter Xu if (!host_memory_backend_mr_inited(backend)) { 150605d0a94SPaolo Bonzini backend->merge = value; 151605d0a94SPaolo Bonzini return; 152605d0a94SPaolo Bonzini } 153605d0a94SPaolo Bonzini 154605d0a94SPaolo Bonzini if (value != backend->merge) { 155605d0a94SPaolo Bonzini void *ptr = memory_region_get_ram_ptr(&backend->mr); 156605d0a94SPaolo Bonzini uint64_t sz = memory_region_size(&backend->mr); 157605d0a94SPaolo Bonzini 158605d0a94SPaolo Bonzini qemu_madvise(ptr, sz, 159605d0a94SPaolo Bonzini value ? QEMU_MADV_MERGEABLE : QEMU_MADV_UNMERGEABLE); 160605d0a94SPaolo Bonzini backend->merge = value; 161605d0a94SPaolo Bonzini } 162605d0a94SPaolo Bonzini } 163605d0a94SPaolo Bonzini 164605d0a94SPaolo Bonzini static bool host_memory_backend_get_dump(Object *obj, Error **errp) 165605d0a94SPaolo Bonzini { 166605d0a94SPaolo Bonzini HostMemoryBackend *backend = MEMORY_BACKEND(obj); 167605d0a94SPaolo Bonzini 168605d0a94SPaolo Bonzini return backend->dump; 169605d0a94SPaolo Bonzini } 170605d0a94SPaolo Bonzini 171605d0a94SPaolo Bonzini static void host_memory_backend_set_dump(Object *obj, bool value, Error **errp) 172605d0a94SPaolo Bonzini { 173605d0a94SPaolo Bonzini HostMemoryBackend *backend = MEMORY_BACKEND(obj); 174605d0a94SPaolo Bonzini 175*6f4c60e4SPeter Xu if (!host_memory_backend_mr_inited(backend)) { 176605d0a94SPaolo Bonzini backend->dump = value; 177605d0a94SPaolo Bonzini return; 178605d0a94SPaolo Bonzini } 179605d0a94SPaolo Bonzini 180605d0a94SPaolo Bonzini if (value != backend->dump) { 181605d0a94SPaolo Bonzini void *ptr = memory_region_get_ram_ptr(&backend->mr); 182605d0a94SPaolo Bonzini uint64_t sz = memory_region_size(&backend->mr); 183605d0a94SPaolo Bonzini 184605d0a94SPaolo Bonzini qemu_madvise(ptr, sz, 185605d0a94SPaolo Bonzini value ? QEMU_MADV_DODUMP : QEMU_MADV_DONTDUMP); 186605d0a94SPaolo Bonzini backend->dump = value; 187605d0a94SPaolo Bonzini } 188605d0a94SPaolo Bonzini } 189605d0a94SPaolo Bonzini 190a35ba7beSPaolo Bonzini static bool host_memory_backend_get_prealloc(Object *obj, Error **errp) 191a35ba7beSPaolo Bonzini { 192a35ba7beSPaolo Bonzini HostMemoryBackend *backend = MEMORY_BACKEND(obj); 193a35ba7beSPaolo Bonzini 194a35ba7beSPaolo Bonzini return backend->prealloc || backend->force_prealloc; 195a35ba7beSPaolo Bonzini } 196a35ba7beSPaolo Bonzini 197a35ba7beSPaolo Bonzini static void host_memory_backend_set_prealloc(Object *obj, bool value, 198a35ba7beSPaolo Bonzini Error **errp) 199a35ba7beSPaolo Bonzini { 200056b68afSIgor Mammedov Error *local_err = NULL; 201a35ba7beSPaolo Bonzini HostMemoryBackend *backend = MEMORY_BACKEND(obj); 202a35ba7beSPaolo Bonzini 203a35ba7beSPaolo Bonzini if (backend->force_prealloc) { 204a35ba7beSPaolo Bonzini if (value) { 205a35ba7beSPaolo Bonzini error_setg(errp, 206a35ba7beSPaolo Bonzini "remove -mem-prealloc to use the prealloc property"); 207a35ba7beSPaolo Bonzini return; 208a35ba7beSPaolo Bonzini } 209a35ba7beSPaolo Bonzini } 210a35ba7beSPaolo Bonzini 211*6f4c60e4SPeter Xu if (!host_memory_backend_mr_inited(backend)) { 212a35ba7beSPaolo Bonzini backend->prealloc = value; 213a35ba7beSPaolo Bonzini return; 214a35ba7beSPaolo Bonzini } 215a35ba7beSPaolo Bonzini 216a35ba7beSPaolo Bonzini if (value && !backend->prealloc) { 217a35ba7beSPaolo Bonzini int fd = memory_region_get_fd(&backend->mr); 218a35ba7beSPaolo Bonzini void *ptr = memory_region_get_ram_ptr(&backend->mr); 219a35ba7beSPaolo Bonzini uint64_t sz = memory_region_size(&backend->mr); 220a35ba7beSPaolo Bonzini 2211e356fc1SJitendra Kolhe os_mem_prealloc(fd, ptr, sz, smp_cpus, &local_err); 222056b68afSIgor Mammedov if (local_err) { 223056b68afSIgor Mammedov error_propagate(errp, local_err); 224056b68afSIgor Mammedov return; 225056b68afSIgor Mammedov } 226a35ba7beSPaolo Bonzini backend->prealloc = true; 227a35ba7beSPaolo Bonzini } 228a35ba7beSPaolo Bonzini } 229a35ba7beSPaolo Bonzini 23058f4662cSHu Tao static void host_memory_backend_init(Object *obj) 2311f070489SIgor Mammedov { 232605d0a94SPaolo Bonzini HostMemoryBackend *backend = MEMORY_BACKEND(obj); 2336b269967SEduardo Habkost MachineState *machine = MACHINE(qdev_get_machine()); 234605d0a94SPaolo Bonzini 2356b269967SEduardo Habkost backend->merge = machine_mem_merge(machine); 2366b269967SEduardo Habkost backend->dump = machine_dump_guest_core(machine); 237a35ba7beSPaolo Bonzini backend->prealloc = mem_prealloc; 2381f070489SIgor Mammedov } 2391f070489SIgor Mammedov 2404728b574SPeter Xu bool host_memory_backend_mr_inited(HostMemoryBackend *backend) 2414728b574SPeter Xu { 2424728b574SPeter Xu /* 2434728b574SPeter Xu * NOTE: We forbid zero-length memory backend, so here zero means 2444728b574SPeter Xu * "we haven't inited the backend memory region yet". 2454728b574SPeter Xu */ 2464728b574SPeter Xu return memory_region_size(&backend->mr) != 0; 2474728b574SPeter Xu } 2484728b574SPeter Xu 2491f070489SIgor Mammedov MemoryRegion * 2501f070489SIgor Mammedov host_memory_backend_get_memory(HostMemoryBackend *backend, Error **errp) 2511f070489SIgor Mammedov { 252*6f4c60e4SPeter Xu return host_memory_backend_mr_inited(backend) ? &backend->mr : NULL; 2531f070489SIgor Mammedov } 2541f070489SIgor Mammedov 2552aece63cSXiao Guangrong void host_memory_backend_set_mapped(HostMemoryBackend *backend, bool mapped) 2562aece63cSXiao Guangrong { 2572aece63cSXiao Guangrong backend->is_mapped = mapped; 2582aece63cSXiao Guangrong } 2592aece63cSXiao Guangrong 2602aece63cSXiao Guangrong bool host_memory_backend_is_mapped(HostMemoryBackend *backend) 2612aece63cSXiao Guangrong { 2622aece63cSXiao Guangrong return backend->is_mapped; 2632aece63cSXiao Guangrong } 2642aece63cSXiao Guangrong 265bd9262d9SHu Tao static void 266bd9262d9SHu Tao host_memory_backend_memory_complete(UserCreatable *uc, Error **errp) 267bd9262d9SHu Tao { 268bd9262d9SHu Tao HostMemoryBackend *backend = MEMORY_BACKEND(uc); 269bd9262d9SHu Tao HostMemoryBackendClass *bc = MEMORY_BACKEND_GET_CLASS(uc); 270605d0a94SPaolo Bonzini Error *local_err = NULL; 271605d0a94SPaolo Bonzini void *ptr; 272605d0a94SPaolo Bonzini uint64_t sz; 273bd9262d9SHu Tao 274bd9262d9SHu Tao if (bc->alloc) { 275605d0a94SPaolo Bonzini bc->alloc(backend, &local_err); 276605d0a94SPaolo Bonzini if (local_err) { 277056b68afSIgor Mammedov goto out; 278605d0a94SPaolo Bonzini } 279605d0a94SPaolo Bonzini 280605d0a94SPaolo Bonzini ptr = memory_region_get_ram_ptr(&backend->mr); 281605d0a94SPaolo Bonzini sz = memory_region_size(&backend->mr); 282605d0a94SPaolo Bonzini 283605d0a94SPaolo Bonzini if (backend->merge) { 284605d0a94SPaolo Bonzini qemu_madvise(ptr, sz, QEMU_MADV_MERGEABLE); 285605d0a94SPaolo Bonzini } 286605d0a94SPaolo Bonzini if (!backend->dump) { 287605d0a94SPaolo Bonzini qemu_madvise(ptr, sz, QEMU_MADV_DONTDUMP); 288605d0a94SPaolo Bonzini } 2894cf1b76bSHu Tao #ifdef CONFIG_NUMA 2904cf1b76bSHu Tao unsigned long lastbit = find_last_bit(backend->host_nodes, MAX_NODES); 2914cf1b76bSHu Tao /* lastbit == MAX_NODES means maxnode = 0 */ 2924cf1b76bSHu Tao unsigned long maxnode = (lastbit + 1) % (MAX_NODES + 1); 2934cf1b76bSHu Tao /* ensure policy won't be ignored in case memory is preallocated 2944cf1b76bSHu Tao * before mbind(). note: MPOL_MF_STRICT is ignored on hugepages so 2954cf1b76bSHu Tao * this doesn't catch hugepage case. */ 296288d3322SMichael S. Tsirkin unsigned flags = MPOL_MF_STRICT | MPOL_MF_MOVE; 2974cf1b76bSHu Tao 2984cf1b76bSHu Tao /* check for invalid host-nodes and policies and give more verbose 2994cf1b76bSHu Tao * error messages than mbind(). */ 3004cf1b76bSHu Tao if (maxnode && backend->policy == MPOL_DEFAULT) { 3014cf1b76bSHu Tao error_setg(errp, "host-nodes must be empty for policy default," 3024cf1b76bSHu Tao " or you should explicitly specify a policy other" 3034cf1b76bSHu Tao " than default"); 3044cf1b76bSHu Tao return; 3054cf1b76bSHu Tao } else if (maxnode == 0 && backend->policy != MPOL_DEFAULT) { 3064cf1b76bSHu Tao error_setg(errp, "host-nodes must be set for policy %s", 3074cf1b76bSHu Tao HostMemPolicy_lookup[backend->policy]); 3084cf1b76bSHu Tao return; 3094cf1b76bSHu Tao } 3104cf1b76bSHu Tao 3114cf1b76bSHu Tao /* We can have up to MAX_NODES nodes, but we need to pass maxnode+1 3124cf1b76bSHu Tao * as argument to mbind() due to an old Linux bug (feature?) which 3134cf1b76bSHu Tao * cuts off the last specified node. This means backend->host_nodes 3144cf1b76bSHu Tao * must have MAX_NODES+1 bits available. 3154cf1b76bSHu Tao */ 3164cf1b76bSHu Tao assert(sizeof(backend->host_nodes) >= 3174cf1b76bSHu Tao BITS_TO_LONGS(MAX_NODES + 1) * sizeof(unsigned long)); 3184cf1b76bSHu Tao assert(maxnode <= MAX_NODES); 3194cf1b76bSHu Tao if (mbind(ptr, sz, backend->policy, 3204cf1b76bSHu Tao maxnode ? backend->host_nodes : NULL, maxnode + 1, flags)) { 321a3567ba1SPavel Fedin if (backend->policy != MPOL_DEFAULT || errno != ENOSYS) { 3224cf1b76bSHu Tao error_setg_errno(errp, errno, 3234cf1b76bSHu Tao "cannot bind memory to host NUMA nodes"); 3244cf1b76bSHu Tao return; 3254cf1b76bSHu Tao } 326a3567ba1SPavel Fedin } 3274cf1b76bSHu Tao #endif 3284cf1b76bSHu Tao /* Preallocate memory after the NUMA policy has been instantiated. 3294cf1b76bSHu Tao * This is necessary to guarantee memory is allocated with 3304cf1b76bSHu Tao * specified NUMA policy in place. 3314cf1b76bSHu Tao */ 332a35ba7beSPaolo Bonzini if (backend->prealloc) { 333056b68afSIgor Mammedov os_mem_prealloc(memory_region_get_fd(&backend->mr), ptr, sz, 3341e356fc1SJitendra Kolhe smp_cpus, &local_err); 335056b68afSIgor Mammedov if (local_err) { 336056b68afSIgor Mammedov goto out; 337a35ba7beSPaolo Bonzini } 338bd9262d9SHu Tao } 339bd9262d9SHu Tao } 340056b68afSIgor Mammedov out: 341056b68afSIgor Mammedov error_propagate(errp, local_err); 342056b68afSIgor Mammedov } 343bd9262d9SHu Tao 34436bce5caSLin Ma static bool 34536bce5caSLin Ma host_memory_backend_can_be_deleted(UserCreatable *uc, Error **errp) 34636bce5caSLin Ma { 3472aece63cSXiao Guangrong if (host_memory_backend_is_mapped(MEMORY_BACKEND(uc))) { 34836bce5caSLin Ma return false; 34936bce5caSLin Ma } else { 35036bce5caSLin Ma return true; 35136bce5caSLin Ma } 35236bce5caSLin Ma } 35336bce5caSLin Ma 354e1ff3c67SIgor Mammedov static char *get_id(Object *o, Error **errp) 355e1ff3c67SIgor Mammedov { 356e1ff3c67SIgor Mammedov HostMemoryBackend *backend = MEMORY_BACKEND(o); 357e1ff3c67SIgor Mammedov 358e1ff3c67SIgor Mammedov return g_strdup(backend->id); 359e1ff3c67SIgor Mammedov } 360e1ff3c67SIgor Mammedov 361e1ff3c67SIgor Mammedov static void set_id(Object *o, const char *str, Error **errp) 362e1ff3c67SIgor Mammedov { 363e1ff3c67SIgor Mammedov HostMemoryBackend *backend = MEMORY_BACKEND(o); 364e1ff3c67SIgor Mammedov 365e1ff3c67SIgor Mammedov if (backend->id) { 366e1ff3c67SIgor Mammedov error_setg(errp, "cannot change property value"); 367e1ff3c67SIgor Mammedov return; 368e1ff3c67SIgor Mammedov } 369e1ff3c67SIgor Mammedov backend->id = g_strdup(str); 370e1ff3c67SIgor Mammedov } 371e1ff3c67SIgor Mammedov 372bd9262d9SHu Tao static void 373bd9262d9SHu Tao host_memory_backend_class_init(ObjectClass *oc, void *data) 374bd9262d9SHu Tao { 375bd9262d9SHu Tao UserCreatableClass *ucc = USER_CREATABLE_CLASS(oc); 376bd9262d9SHu Tao 377bd9262d9SHu Tao ucc->complete = host_memory_backend_memory_complete; 37836bce5caSLin Ma ucc->can_be_deleted = host_memory_backend_can_be_deleted; 379e62834caSEduardo Habkost 380e62834caSEduardo Habkost object_class_property_add_bool(oc, "merge", 381e62834caSEduardo Habkost host_memory_backend_get_merge, 382e62834caSEduardo Habkost host_memory_backend_set_merge, &error_abort); 383e62834caSEduardo Habkost object_class_property_add_bool(oc, "dump", 384e62834caSEduardo Habkost host_memory_backend_get_dump, 385e62834caSEduardo Habkost host_memory_backend_set_dump, &error_abort); 386e62834caSEduardo Habkost object_class_property_add_bool(oc, "prealloc", 387e62834caSEduardo Habkost host_memory_backend_get_prealloc, 388e62834caSEduardo Habkost host_memory_backend_set_prealloc, &error_abort); 389e62834caSEduardo Habkost object_class_property_add(oc, "size", "int", 390e62834caSEduardo Habkost host_memory_backend_get_size, 391e62834caSEduardo Habkost host_memory_backend_set_size, 392e62834caSEduardo Habkost NULL, NULL, &error_abort); 393e62834caSEduardo Habkost object_class_property_add(oc, "host-nodes", "int", 394e62834caSEduardo Habkost host_memory_backend_get_host_nodes, 395e62834caSEduardo Habkost host_memory_backend_set_host_nodes, 396e62834caSEduardo Habkost NULL, NULL, &error_abort); 397e62834caSEduardo Habkost object_class_property_add_enum(oc, "policy", "HostMemPolicy", 398e62834caSEduardo Habkost HostMemPolicy_lookup, 399e62834caSEduardo Habkost host_memory_backend_get_policy, 400e62834caSEduardo Habkost host_memory_backend_set_policy, &error_abort); 401e1ff3c67SIgor Mammedov object_class_property_add_str(oc, "id", get_id, set_id, &error_abort); 402e1ff3c67SIgor Mammedov } 403e1ff3c67SIgor Mammedov 404e1ff3c67SIgor Mammedov static void host_memory_backend_finalize(Object *o) 405e1ff3c67SIgor Mammedov { 406e1ff3c67SIgor Mammedov HostMemoryBackend *backend = MEMORY_BACKEND(o); 407e1ff3c67SIgor Mammedov g_free(backend->id); 408bd9262d9SHu Tao } 409bd9262d9SHu Tao 41058f4662cSHu Tao static const TypeInfo host_memory_backend_info = { 4111f070489SIgor Mammedov .name = TYPE_MEMORY_BACKEND, 4121f070489SIgor Mammedov .parent = TYPE_OBJECT, 4131f070489SIgor Mammedov .abstract = true, 4141f070489SIgor Mammedov .class_size = sizeof(HostMemoryBackendClass), 415bd9262d9SHu Tao .class_init = host_memory_backend_class_init, 4161f070489SIgor Mammedov .instance_size = sizeof(HostMemoryBackend), 41758f4662cSHu Tao .instance_init = host_memory_backend_init, 418e1ff3c67SIgor Mammedov .instance_finalize = host_memory_backend_finalize, 4191f070489SIgor Mammedov .interfaces = (InterfaceInfo[]) { 4201f070489SIgor Mammedov { TYPE_USER_CREATABLE }, 4211f070489SIgor Mammedov { } 4221f070489SIgor Mammedov } 4231f070489SIgor Mammedov }; 4241f070489SIgor Mammedov 4251f070489SIgor Mammedov static void register_types(void) 4261f070489SIgor Mammedov { 42758f4662cSHu Tao type_register_static(&host_memory_backend_info); 4281f070489SIgor Mammedov } 4291f070489SIgor Mammedov 4301f070489SIgor Mammedov type_init(register_types); 431