/xen-4.10.0-shim-comet/xen/common/ |
A D | tmem.c | 166 if ( pool == NULL || !is_persistent(pool) ) in tmem_free() 182 if ( pool != NULL && is_persistent(pool) ) in tmem_alloc_page() 196 if ( pool == NULL || !is_persistent(pool) ) in tmem_free_page() 266 pool = obj->pool; in pgp_alloc() 325 pool = pgp->us.obj->pool; in pgp_free() 462 pool = objnode->obj->pool; in rtn_free() 552 pool = obj->pool; in obj_free() 621 obj->pool = pool; in obj_alloc() 906 struct tmem_pool *pool = obj->pool; in tmem_try_to_evict_pgp() local 969 pool = obj->pool; in tmem_evict() [all …]
|
A D | xmalloc_tlsf.c | 303 if ( pool == NULL ) in xmem_pool_create() 320 strlcpy(pool->name, name, sizeof(pool->name)); in xmem_pool_create() 331 return pool; in xmem_pool_create() 344 + (pool->num_regions - 1) * pool->grow_size; in xmem_pool_get_total_size() 366 pool->name, pool, xmem_pool_get_used_size(pool)); in xmem_pool_destroy() 385 if ( (region = pool->get_mem(pool->init_size)) == NULL ) in xmem_pool_alloc() 387 ADD_REGION(region, pool->init_size, pool); in xmem_pool_alloc() 404 if ( pool->max_size && (pool->init_size + in xmem_pool_alloc() 405 pool->num_regions * pool->grow_size in xmem_pool_alloc() 409 if ( (region = pool->get_mem(pool->grow_size)) == NULL ) in xmem_pool_alloc() [all …]
|
A D | tmem_control.c | 377 if ( pool == NULL ) in tmemc_get_pool() 385 out.uuid[0] = pool->uuid[0]; in tmemc_get_pool() 386 out.uuid[1] = pool->uuid[1]; in tmemc_get_pool() 431 xen_tmem_pool_info_t pool; in tmemc_set_pools() local 436 if ( pool.n_pages ) in tmemc_set_pools() 439 rc = do_tmem_new_pool(cli_id, pool.id, pool.flags.raw, in tmemc_set_pools() 440 pool.uuid[0], pool.uuid[1]); in tmemc_set_pools() 444 pool.id = rc; in tmemc_set_pools() 480 xen_tmem_pool_info_t pool; in tmemc_auth_pools() local 485 if ( pool.n_pages ) in tmemc_auth_pools() [all …]
|
A D | cpupool.c | 115 void cpupool_put(struct cpupool *pool) in cpupool_put() argument 117 if ( !atomic_dec_and_test(&pool->refcnt) ) in cpupool_put() 119 scheduler_free(pool->sched); in cpupool_put() 120 free_cpupool_struct(pool); in cpupool_put()
|
A D | schedule.c | 1332 struct cpupool *pool; in sched_adjust_global() local 1343 pool = cpupool_get_by_id(op->cpupool_id); in sched_adjust_global() 1344 if ( pool == NULL ) in sched_adjust_global() 1347 rc = ((op->sched_id == pool->sched->sched_id) in sched_adjust_global() 1348 ? SCHED_OP(pool->sched, adjust_global, op) : -EINVAL); in sched_adjust_global() 1350 cpupool_put(pool); in sched_adjust_global()
|
A D | Kconfig | 72 collective pool of memory shared across guests. The end result is
|
/xen-4.10.0-shim-comet/tools/libxc/ |
A D | xc_tmem.c | 158 xen_tmem_pool_info_t pool = { in xc_tmem_auth() local 259 pool->n_pages = 0; in xc_tmem_save() 261 if ( write_exact(io_fd, pool, sizeof(*pool)) ) in xc_tmem_save() 308 pool->n_pages - j, domid, pool->id, checksum); in xc_tmem_save() 368 xen_tmem_pool_info_t pool = { in xc_tmem_restore_new_pool() local 415 if ( read_exact(io_fd, &pool, sizeof(pool)) ) in xc_tmem_restore() 418 if ( xc_tmem_restore_new_pool(xch, domid, pool.id, pool.flags.raw, in xc_tmem_restore() 419 pool.uuid[0], pool.uuid[1]) < 0 ) in xc_tmem_restore() 422 if ( pool.n_pages <= 0 ) in xc_tmem_restore() 460 if ( pool.n_pages ) in xc_tmem_restore() [all …]
|
/xen-4.10.0-shim-comet/tools/xl/ |
A D | xl_cpupool.c | 246 const char *pool = NULL; in main_cpupoollist() local 306 const char *pool; in main_cpupooldestroy() local 313 pool = argv[optind]; in main_cpupooldestroy() 332 const char *pool; in main_cpupoolrename() local 340 pool = argv[optind++]; in main_cpupoolrename() 361 const char *pool; in main_cpupoolcpuadd() local 376 pool = argv[optind++]; in main_cpupoolcpuadd() 399 const char *pool; in main_cpupoolcpuremove() local 414 pool = argv[optind++]; in main_cpupoolcpuremove() 439 const char *pool; in main_cpupoolmigrate() local [all …]
|
/xen-4.10.0-shim-comet/xen/include/xen/ |
A D | xmalloc.h | 87 void xmem_pool_destroy(struct xmem_pool *pool); 94 void *xmem_pool_alloc(unsigned long size, struct xmem_pool *pool); 100 int xmem_pool_maxalloc(struct xmem_pool *pool); 107 void xmem_pool_free(void *ptr, struct xmem_pool *pool); 114 unsigned long xmem_pool_get_used_size(struct xmem_pool *pool); 124 unsigned long xmem_pool_get_total_size(struct xmem_pool *pool);
|
A D | sched.h | 912 void cpupool_put(struct cpupool *pool);
|
/xen-4.10.0-shim-comet/docs/features/ |
A D | sched_rtds.pandoc | 34 xl cpupool-create name=\"pool-rt\" sched=\"rtds\" cpus=[4,5,6,8] 78 * create a few virtual machines a move them in and out of the pool, 79 * create a few virtual machines, directly inside the pool, and verify
|
/xen-4.10.0-shim-comet/docs/man/ |
A D | xl.pod.1.in | 1063 =item B<-p [pool]> : list all domains and sched params for [pool] 1069 =item B<-p [pool] -s> : List sched params for [pool] 1071 =item B<-p [pool] -s [params]> : Set sched params for [pool] 1073 =item B<-p [pool] -d>... : Illegal 1228 cpu-pool. Scheduling does not cross cpu-pool boundaries, so each cpu-pool has 1268 =item B<cpupool-destroy> I<cpu-pool> 1270 Deactivates a cpu pool. 1273 =item B<cpupool-rename> I<cpu-pool> <newname> 1275 Renames a cpu-pool to I<newname>. 1305 Domain-0 can't be moved to another cpu-pool. [all …]
|
A D | xl.conf.pod.5 | 180 Normal memory and freeable pool of ephemeral pages (tmem) is used when
|
A D | xl-numa-placement.pod.7 | 59 is to pool together the pCPUs spanning the node and put the domain in 60 such a I<cpupool> with the "pool=" config option (as documented in our
|
A D | xen-tscmode.pod.7 | 173 synchronize TSC across a data center or even a pool of machines. As 268 across all the machines in a pool or data center, a migration may "break"
|
/xen-4.10.0-shim-comet/xen/tools/kconfig/ |
A D | zconf.gperf | 4 %define string-pool-name kconf_id_strings
|
/xen-4.10.0-shim-comet/docs/misc/ |
A D | xenpaging.txt | 10 pages back to the pool of available memory. Once the guests wants to
|
A D | grant-tables.txt | 364 To release a batch back to the shared pool: 374 To release a claimed reference back to the reserve pool:
|
A D | xen-command-line.markdown | 468 pool appear the same to guests to allow successful live migration. 476 help with cpuid levelling across a pool of hosts. Setting a bit in 479 is important to ensure that all hosts in a pool appear the same to 488 help with cpuid levelling across a pool of hosts. See the description 497 masks to help with cpuid levelling across a pool of hosts. See the
|
/xen-4.10.0-shim-comet/tools/python/xen/lowlevel/xc/ |
A D | xc.c | 1742 uint32_t pool; in pyxc_cpupool_getinfo() local 1746 for (pool = 0;;) in pyxc_cpupool_getinfo() 1748 info = xc_cpupool_getinfo(self->xc_handle, pool); in pyxc_cpupool_getinfo() 1757 pool = info->cpupool_id + 1; in pyxc_cpupool_getinfo()
|
/xen-4.10.0-shim-comet/tools/golang/xenlight/ |
A D | xenlight.go | 433 func (Ctx *Context) CpupoolInfo(Poolid uint32) (pool CpupoolInfo) { 448 pool = c_cpupool.toGo()
|
/xen-4.10.0-shim-comet/xen/include/public/ |
A D | sysctl.h | 812 XEN_GUEST_HANDLE_64(xen_tmem_pool_info_t) pool; /* OUT for */
|
/xen-4.10.0-shim-comet/tools/xenstore/ |
A D | talloc_guide.txt | 13 The new talloc is a hierarchical, reference counted memory pool system
|
/xen-4.10.0-shim-comet/ |
A D | SUPPORT.md | 251 with each pool having the capability
|
/xen-4.10.0-shim-comet/stubdom/ |
A D | lwip.patch-cvs | 700 /* for every pool: */ 716 LWIP_DEBUGF(MEMP_DEBUG | 2, ("memp_malloc: out of memory in pool %s\n", memp_desc[type]));
|