1 /******************************************************************************
2 * vm_event.c
3 *
4 * VM event support.
5 *
6 * Copyright (c) 2009 Citrix Systems, Inc. (Patrick Colp)
7 *
8 * This program is free software; you can redistribute it and/or modify
9 * it under the terms of the GNU General Public License as published by
10 * the Free Software Foundation; either version 2 of the License, or
11 * (at your option) any later version.
12 *
13 * This program is distributed in the hope that it will be useful,
14 * but WITHOUT ANY WARRANTY; without even the implied warranty of
15 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
16 * GNU General Public License for more details.
17 *
18 * You should have received a copy of the GNU General Public License
19 * along with this program; If not, see <http://www.gnu.org/licenses/>.
20 */
21
22
23 #include <xen/sched.h>
24 #include <xen/event.h>
25 #include <xen/wait.h>
26 #include <xen/vm_event.h>
27 #include <xen/mem_access.h>
28 #include <asm/p2m.h>
29 #include <asm/monitor.h>
30 #include <asm/vm_event.h>
31 #include <xsm/xsm.h>
32
33 /* for public/io/ring.h macros */
34 #define xen_mb() smp_mb()
35 #define xen_rmb() smp_rmb()
36 #define xen_wmb() smp_wmb()
37
38 #define vm_event_ring_lock_init(_ved) spin_lock_init(&(_ved)->ring_lock)
39 #define vm_event_ring_lock(_ved) spin_lock(&(_ved)->ring_lock)
40 #define vm_event_ring_unlock(_ved) spin_unlock(&(_ved)->ring_lock)
41
vm_event_enable(struct domain * d,struct xen_domctl_vm_event_op * vec,struct vm_event_domain ** ved,int pause_flag,int param,xen_event_channel_notification_t notification_fn)42 static int vm_event_enable(
43 struct domain *d,
44 struct xen_domctl_vm_event_op *vec,
45 struct vm_event_domain **ved,
46 int pause_flag,
47 int param,
48 xen_event_channel_notification_t notification_fn)
49 {
50 int rc;
51 unsigned long ring_gfn = d->arch.hvm_domain.params[param];
52
53 if ( !*ved )
54 *ved = xzalloc(struct vm_event_domain);
55 if ( !*ved )
56 return -ENOMEM;
57
58 /* Only one helper at a time. If the helper crashed,
59 * the ring is in an undefined state and so is the guest.
60 */
61 if ( (*ved)->ring_page )
62 return -EBUSY;;
63
64 /* The parameter defaults to zero, and it should be
65 * set to something */
66 if ( ring_gfn == 0 )
67 return -ENOSYS;
68
69 vm_event_ring_lock_init(*ved);
70 vm_event_ring_lock(*ved);
71
72 rc = vm_event_init_domain(d);
73
74 if ( rc < 0 )
75 goto err;
76
77 rc = prepare_ring_for_helper(d, ring_gfn, &(*ved)->ring_pg_struct,
78 &(*ved)->ring_page);
79 if ( rc < 0 )
80 goto err;
81
82 /* Set the number of currently blocked vCPUs to 0. */
83 (*ved)->blocked = 0;
84
85 /* Allocate event channel */
86 rc = alloc_unbound_xen_event_channel(d, 0, current->domain->domain_id,
87 notification_fn);
88 if ( rc < 0 )
89 goto err;
90
91 (*ved)->xen_port = vec->port = rc;
92
93 /* Prepare ring buffer */
94 FRONT_RING_INIT(&(*ved)->front_ring,
95 (vm_event_sring_t *)(*ved)->ring_page,
96 PAGE_SIZE);
97
98 /* Save the pause flag for this particular ring. */
99 (*ved)->pause_flag = pause_flag;
100
101 /* Initialize the last-chance wait queue. */
102 init_waitqueue_head(&(*ved)->wq);
103
104 vm_event_ring_unlock(*ved);
105 return 0;
106
107 err:
108 destroy_ring_for_helper(&(*ved)->ring_page,
109 (*ved)->ring_pg_struct);
110 vm_event_ring_unlock(*ved);
111 xfree(*ved);
112 *ved = NULL;
113
114 return rc;
115 }
116
vm_event_ring_available(struct vm_event_domain * ved)117 static unsigned int vm_event_ring_available(struct vm_event_domain *ved)
118 {
119 int avail_req = RING_FREE_REQUESTS(&ved->front_ring);
120 avail_req -= ved->target_producers;
121 avail_req -= ved->foreign_producers;
122
123 BUG_ON(avail_req < 0);
124
125 return avail_req;
126 }
127
128 /*
129 * vm_event_wake_blocked() will wakeup vcpus waiting for room in the
130 * ring. These vCPUs were paused on their way out after placing an event,
131 * but need to be resumed where the ring is capable of processing at least
132 * one event from them.
133 */
vm_event_wake_blocked(struct domain * d,struct vm_event_domain * ved)134 static void vm_event_wake_blocked(struct domain *d, struct vm_event_domain *ved)
135 {
136 struct vcpu *v;
137 unsigned int avail_req = vm_event_ring_available(ved);
138
139 if ( avail_req == 0 || ved->blocked == 0 )
140 return;
141
142 /* We remember which vcpu last woke up to avoid scanning always linearly
143 * from zero and starving higher-numbered vcpus under high load */
144 if ( d->vcpu )
145 {
146 int i, j, k;
147
148 for (i = ved->last_vcpu_wake_up + 1, j = 0; j < d->max_vcpus; i++, j++)
149 {
150 k = i % d->max_vcpus;
151 v = d->vcpu[k];
152 if ( !v )
153 continue;
154
155 if ( !(ved->blocked) || avail_req == 0 )
156 break;
157
158 if ( test_and_clear_bit(ved->pause_flag, &v->pause_flags) )
159 {
160 vcpu_unpause(v);
161 avail_req--;
162 ved->blocked--;
163 ved->last_vcpu_wake_up = k;
164 }
165 }
166 }
167 }
168
169 /*
170 * In the event that a vCPU attempted to place an event in the ring and
171 * was unable to do so, it is queued on a wait queue. These are woken as
172 * needed, and take precedence over the blocked vCPUs.
173 */
vm_event_wake_queued(struct domain * d,struct vm_event_domain * ved)174 static void vm_event_wake_queued(struct domain *d, struct vm_event_domain *ved)
175 {
176 unsigned int avail_req = vm_event_ring_available(ved);
177
178 if ( avail_req > 0 )
179 wake_up_nr(&ved->wq, avail_req);
180 }
181
182 /*
183 * vm_event_wake() will wakeup all vcpus waiting for the ring to
184 * become available. If we have queued vCPUs, they get top priority. We
185 * are guaranteed that they will go through code paths that will eventually
186 * call vm_event_wake() again, ensuring that any blocked vCPUs will get
187 * unpaused once all the queued vCPUs have made it through.
188 */
vm_event_wake(struct domain * d,struct vm_event_domain * ved)189 void vm_event_wake(struct domain *d, struct vm_event_domain *ved)
190 {
191 if (!list_empty(&ved->wq.list))
192 vm_event_wake_queued(d, ved);
193 else
194 vm_event_wake_blocked(d, ved);
195 }
196
vm_event_disable(struct domain * d,struct vm_event_domain ** ved)197 static int vm_event_disable(struct domain *d, struct vm_event_domain **ved)
198 {
199 if ( vm_event_check_ring(*ved) )
200 {
201 struct vcpu *v;
202
203 vm_event_ring_lock(*ved);
204
205 if ( !list_empty(&(*ved)->wq.list) )
206 {
207 vm_event_ring_unlock(*ved);
208 return -EBUSY;
209 }
210
211 /* Free domU's event channel and leave the other one unbound */
212 free_xen_event_channel(d, (*ved)->xen_port);
213
214 /* Unblock all vCPUs */
215 for_each_vcpu ( d, v )
216 {
217 if ( test_and_clear_bit((*ved)->pause_flag, &v->pause_flags) )
218 {
219 vcpu_unpause(v);
220 (*ved)->blocked--;
221 }
222 }
223
224 destroy_ring_for_helper(&(*ved)->ring_page,
225 (*ved)->ring_pg_struct);
226
227 vm_event_cleanup_domain(d);
228
229 vm_event_ring_unlock(*ved);
230 }
231
232 xfree(*ved);
233 *ved = NULL;
234
235 return 0;
236 }
237
vm_event_release_slot(struct domain * d,struct vm_event_domain * ved)238 static inline void vm_event_release_slot(struct domain *d,
239 struct vm_event_domain *ved)
240 {
241 /* Update the accounting */
242 if ( current->domain == d )
243 ved->target_producers--;
244 else
245 ved->foreign_producers--;
246
247 /* Kick any waiters */
248 vm_event_wake(d, ved);
249 }
250
251 /*
252 * vm_event_mark_and_pause() tags vcpu and put it to sleep.
253 * The vcpu will resume execution in vm_event_wake_blocked().
254 */
vm_event_mark_and_pause(struct vcpu * v,struct vm_event_domain * ved)255 void vm_event_mark_and_pause(struct vcpu *v, struct vm_event_domain *ved)
256 {
257 if ( !test_and_set_bit(ved->pause_flag, &v->pause_flags) )
258 {
259 vcpu_pause_nosync(v);
260 ved->blocked++;
261 }
262 }
263
264 /*
265 * This must be preceded by a call to claim_slot(), and is guaranteed to
266 * succeed. As a side-effect however, the vCPU may be paused if the ring is
267 * overly full and its continued execution would cause stalling and excessive
268 * waiting. The vCPU will be automatically unpaused when the ring clears.
269 */
vm_event_put_request(struct domain * d,struct vm_event_domain * ved,vm_event_request_t * req)270 void vm_event_put_request(struct domain *d,
271 struct vm_event_domain *ved,
272 vm_event_request_t *req)
273 {
274 vm_event_front_ring_t *front_ring;
275 int free_req;
276 unsigned int avail_req;
277 RING_IDX req_prod;
278 struct vcpu *curr = current;
279
280 if( !vm_event_check_ring(ved))
281 return;
282
283 if ( curr->domain != d )
284 {
285 req->flags |= VM_EVENT_FLAG_FOREIGN;
286 #ifndef NDEBUG
287 if ( !(req->flags & VM_EVENT_FLAG_VCPU_PAUSED) )
288 gdprintk(XENLOG_G_WARNING, "d%dv%d was not paused.\n",
289 d->domain_id, req->vcpu_id);
290 #endif
291 }
292
293 req->version = VM_EVENT_INTERFACE_VERSION;
294
295 vm_event_ring_lock(ved);
296
297 /* Due to the reservations, this step must succeed. */
298 front_ring = &ved->front_ring;
299 free_req = RING_FREE_REQUESTS(front_ring);
300 ASSERT(free_req > 0);
301
302 /* Copy request */
303 req_prod = front_ring->req_prod_pvt;
304 memcpy(RING_GET_REQUEST(front_ring, req_prod), req, sizeof(*req));
305 req_prod++;
306
307 /* Update ring */
308 front_ring->req_prod_pvt = req_prod;
309 RING_PUSH_REQUESTS(front_ring);
310
311 /* We've actually *used* our reservation, so release the slot. */
312 vm_event_release_slot(d, ved);
313
314 /* Give this vCPU a black eye if necessary, on the way out.
315 * See the comments above wake_blocked() for more information
316 * on how this mechanism works to avoid waiting. */
317 avail_req = vm_event_ring_available(ved);
318 if( curr->domain == d && avail_req < d->max_vcpus &&
319 !atomic_read(&curr->vm_event_pause_count) )
320 vm_event_mark_and_pause(curr, ved);
321
322 vm_event_ring_unlock(ved);
323
324 notify_via_xen_event_channel(d, ved->xen_port);
325 }
326
vm_event_get_response(struct domain * d,struct vm_event_domain * ved,vm_event_response_t * rsp)327 int vm_event_get_response(struct domain *d, struct vm_event_domain *ved,
328 vm_event_response_t *rsp)
329 {
330 vm_event_front_ring_t *front_ring;
331 RING_IDX rsp_cons;
332
333 vm_event_ring_lock(ved);
334
335 front_ring = &ved->front_ring;
336 rsp_cons = front_ring->rsp_cons;
337
338 if ( !RING_HAS_UNCONSUMED_RESPONSES(front_ring) )
339 {
340 vm_event_ring_unlock(ved);
341 return 0;
342 }
343
344 /* Copy response */
345 memcpy(rsp, RING_GET_RESPONSE(front_ring, rsp_cons), sizeof(*rsp));
346 rsp_cons++;
347
348 /* Update ring */
349 front_ring->rsp_cons = rsp_cons;
350 front_ring->sring->rsp_event = rsp_cons + 1;
351
352 /* Kick any waiters -- since we've just consumed an event,
353 * there may be additional space available in the ring. */
354 vm_event_wake(d, ved);
355
356 vm_event_ring_unlock(ved);
357
358 return 1;
359 }
360
361 /*
362 * Pull all responses from the given ring and unpause the corresponding vCPU
363 * if required. Based on the response type, here we can also call custom
364 * handlers.
365 *
366 * Note: responses are handled the same way regardless of which ring they
367 * arrive on.
368 */
vm_event_resume(struct domain * d,struct vm_event_domain * ved)369 void vm_event_resume(struct domain *d, struct vm_event_domain *ved)
370 {
371 vm_event_response_t rsp;
372
373 /*
374 * vm_event_resume() runs in either XEN_DOMCTL_VM_EVENT_OP_*, or
375 * EVTCHN_send context from the introspection consumer. Both contexts
376 * are guaranteed not to be the subject of vm_event responses.
377 * While we could ASSERT(v != current) for each VCPU in d in the loop
378 * below, this covers the case where we would need to iterate over all
379 * of them more succintly.
380 */
381 ASSERT(d != current->domain);
382
383 /* Pull all responses off the ring. */
384 while ( vm_event_get_response(d, ved, &rsp) )
385 {
386 struct vcpu *v;
387
388 if ( rsp.version != VM_EVENT_INTERFACE_VERSION )
389 {
390 printk(XENLOG_G_WARNING "vm_event interface version mismatch\n");
391 continue;
392 }
393
394 /* Validate the vcpu_id in the response. */
395 if ( (rsp.vcpu_id >= d->max_vcpus) || !d->vcpu[rsp.vcpu_id] )
396 continue;
397
398 v = d->vcpu[rsp.vcpu_id];
399
400 /*
401 * In some cases the response type needs extra handling, so here
402 * we call the appropriate handlers.
403 */
404
405 /* Check flags which apply only when the vCPU is paused */
406 if ( atomic_read(&v->vm_event_pause_count) )
407 {
408 #ifdef CONFIG_HAS_MEM_PAGING
409 if ( rsp.reason == VM_EVENT_REASON_MEM_PAGING )
410 p2m_mem_paging_resume(d, &rsp);
411 #endif
412
413 /*
414 * Check emulation flags in the arch-specific handler only, as it
415 * has to set arch-specific flags when supported, and to avoid
416 * bitmask overhead when it isn't supported.
417 */
418 vm_event_emulate_check(v, &rsp);
419
420 /*
421 * Check in arch-specific handler to avoid bitmask overhead when
422 * not supported.
423 */
424 vm_event_register_write_resume(v, &rsp);
425
426 /*
427 * Check in arch-specific handler to avoid bitmask overhead when
428 * not supported.
429 */
430 vm_event_toggle_singlestep(d, v, &rsp);
431
432 /* Check for altp2m switch */
433 if ( rsp.flags & VM_EVENT_FLAG_ALTERNATE_P2M )
434 p2m_altp2m_check(v, rsp.altp2m_idx);
435
436 if ( rsp.flags & VM_EVENT_FLAG_SET_REGISTERS )
437 vm_event_set_registers(v, &rsp);
438
439 if ( rsp.flags & VM_EVENT_FLAG_GET_NEXT_INTERRUPT )
440 vm_event_monitor_next_interrupt(v);
441
442 if ( rsp.flags & VM_EVENT_FLAG_VCPU_PAUSED )
443 vm_event_vcpu_unpause(v);
444 }
445 }
446 }
447
vm_event_cancel_slot(struct domain * d,struct vm_event_domain * ved)448 void vm_event_cancel_slot(struct domain *d, struct vm_event_domain *ved)
449 {
450 if( !vm_event_check_ring(ved) )
451 return;
452
453 vm_event_ring_lock(ved);
454 vm_event_release_slot(d, ved);
455 vm_event_ring_unlock(ved);
456 }
457
vm_event_grab_slot(struct vm_event_domain * ved,int foreign)458 static int vm_event_grab_slot(struct vm_event_domain *ved, int foreign)
459 {
460 unsigned int avail_req;
461
462 if ( !ved->ring_page )
463 return -ENOSYS;
464
465 vm_event_ring_lock(ved);
466
467 avail_req = vm_event_ring_available(ved);
468 if ( avail_req == 0 )
469 {
470 vm_event_ring_unlock(ved);
471 return -EBUSY;
472 }
473
474 if ( !foreign )
475 ved->target_producers++;
476 else
477 ved->foreign_producers++;
478
479 vm_event_ring_unlock(ved);
480
481 return 0;
482 }
483
484 /* Simple try_grab wrapper for use in the wait_event() macro. */
vm_event_wait_try_grab(struct vm_event_domain * ved,int * rc)485 static int vm_event_wait_try_grab(struct vm_event_domain *ved, int *rc)
486 {
487 *rc = vm_event_grab_slot(ved, 0);
488 return *rc;
489 }
490
491 /* Call vm_event_grab_slot() until the ring doesn't exist, or is available. */
vm_event_wait_slot(struct vm_event_domain * ved)492 static int vm_event_wait_slot(struct vm_event_domain *ved)
493 {
494 int rc = -EBUSY;
495 wait_event(ved->wq, vm_event_wait_try_grab(ved, &rc) != -EBUSY);
496 return rc;
497 }
498
vm_event_check_ring(struct vm_event_domain * ved)499 bool_t vm_event_check_ring(struct vm_event_domain *ved)
500 {
501 return (ved && ved->ring_page);
502 }
503
504 /*
505 * Determines whether or not the current vCPU belongs to the target domain,
506 * and calls the appropriate wait function. If it is a guest vCPU, then we
507 * use vm_event_wait_slot() to reserve a slot. As long as there is a ring,
508 * this function will always return 0 for a guest. For a non-guest, we check
509 * for space and return -EBUSY if the ring is not available.
510 *
511 * Return codes: -ENOSYS: the ring is not yet configured
512 * -EBUSY: the ring is busy
513 * 0: a spot has been reserved
514 *
515 */
__vm_event_claim_slot(struct domain * d,struct vm_event_domain * ved,bool_t allow_sleep)516 int __vm_event_claim_slot(struct domain *d, struct vm_event_domain *ved,
517 bool_t allow_sleep)
518 {
519 if ( !vm_event_check_ring(ved) )
520 return -EOPNOTSUPP;
521
522 if ( (current->domain == d) && allow_sleep )
523 return vm_event_wait_slot(ved);
524 else
525 return vm_event_grab_slot(ved, (current->domain != d));
526 }
527
528 #ifdef CONFIG_HAS_MEM_PAGING
529 /* Registered with Xen-bound event channel for incoming notifications. */
mem_paging_notification(struct vcpu * v,unsigned int port)530 static void mem_paging_notification(struct vcpu *v, unsigned int port)
531 {
532 struct domain *domain = v->domain;
533
534 if ( likely(vm_event_check_ring(domain->vm_event_paging)) )
535 vm_event_resume(domain, domain->vm_event_paging);
536 }
537 #endif
538
539 /* Registered with Xen-bound event channel for incoming notifications. */
monitor_notification(struct vcpu * v,unsigned int port)540 static void monitor_notification(struct vcpu *v, unsigned int port)
541 {
542 struct domain *domain = v->domain;
543
544 if ( likely(vm_event_check_ring(domain->vm_event_monitor)) )
545 vm_event_resume(domain, domain->vm_event_monitor);
546 }
547
548 #ifdef CONFIG_HAS_MEM_SHARING
549 /* Registered with Xen-bound event channel for incoming notifications. */
mem_sharing_notification(struct vcpu * v,unsigned int port)550 static void mem_sharing_notification(struct vcpu *v, unsigned int port)
551 {
552 struct domain *domain = v->domain;
553
554 if ( likely(vm_event_check_ring(domain->vm_event_share)) )
555 vm_event_resume(domain, domain->vm_event_share);
556 }
557 #endif
558
559 /* Clean up on domain destruction */
vm_event_cleanup(struct domain * d)560 void vm_event_cleanup(struct domain *d)
561 {
562 #ifdef CONFIG_HAS_MEM_PAGING
563 if ( vm_event_check_ring(d->vm_event_paging) )
564 {
565 /* Destroying the wait queue head means waking up all
566 * queued vcpus. This will drain the list, allowing
567 * the disable routine to complete. It will also drop
568 * all domain refs the wait-queued vcpus are holding.
569 * Finally, because this code path involves previously
570 * pausing the domain (domain_kill), unpausing the
571 * vcpus causes no harm. */
572 destroy_waitqueue_head(&d->vm_event_paging->wq);
573 (void)vm_event_disable(d, &d->vm_event_paging);
574 }
575 #endif
576 if ( vm_event_check_ring(d->vm_event_monitor) )
577 {
578 destroy_waitqueue_head(&d->vm_event_monitor->wq);
579 (void)vm_event_disable(d, &d->vm_event_monitor);
580 }
581 #ifdef CONFIG_HAS_MEM_SHARING
582 if ( vm_event_check_ring(d->vm_event_share) )
583 {
584 destroy_waitqueue_head(&d->vm_event_share->wq);
585 (void)vm_event_disable(d, &d->vm_event_share);
586 }
587 #endif
588 }
589
vm_event_domctl(struct domain * d,struct xen_domctl_vm_event_op * vec,XEN_GUEST_HANDLE_PARAM (void)u_domctl)590 int vm_event_domctl(struct domain *d, struct xen_domctl_vm_event_op *vec,
591 XEN_GUEST_HANDLE_PARAM(void) u_domctl)
592 {
593 int rc;
594
595 rc = xsm_vm_event_control(XSM_PRIV, d, vec->mode, vec->op);
596 if ( rc )
597 return rc;
598
599 if ( unlikely(d == current->domain) ) /* no domain_pause() */
600 {
601 gdprintk(XENLOG_INFO, "Tried to do a memory event op on itself.\n");
602 return -EINVAL;
603 }
604
605 if ( unlikely(d->is_dying) )
606 {
607 gdprintk(XENLOG_INFO, "Ignoring memory event op on dying domain %u\n",
608 d->domain_id);
609 return 0;
610 }
611
612 if ( unlikely(d->vcpu == NULL) || unlikely(d->vcpu[0] == NULL) )
613 {
614 gdprintk(XENLOG_INFO,
615 "Memory event op on a domain (%u) with no vcpus\n",
616 d->domain_id);
617 return -EINVAL;
618 }
619
620 rc = -ENOSYS;
621
622 switch ( vec->mode )
623 {
624 #ifdef CONFIG_HAS_MEM_PAGING
625 case XEN_DOMCTL_VM_EVENT_OP_PAGING:
626 {
627 rc = -EINVAL;
628
629 switch( vec->op )
630 {
631 case XEN_VM_EVENT_ENABLE:
632 {
633 struct p2m_domain *p2m = p2m_get_hostp2m(d);
634
635 rc = -EOPNOTSUPP;
636 /* hvm fixme: p2m_is_foreign types need addressing */
637 if ( is_hvm_domain(hardware_domain) )
638 break;
639
640 rc = -ENODEV;
641 /* Only HAP is supported */
642 if ( !hap_enabled(d) )
643 break;
644
645 /* No paging if iommu is used */
646 rc = -EMLINK;
647 if ( unlikely(need_iommu(d)) )
648 break;
649
650 rc = -EXDEV;
651 /* Disallow paging in a PoD guest */
652 if ( p2m->pod.entry_count )
653 break;
654
655 /* domain_pause() not required here, see XSA-99 */
656 rc = vm_event_enable(d, vec, &d->vm_event_paging, _VPF_mem_paging,
657 HVM_PARAM_PAGING_RING_PFN,
658 mem_paging_notification);
659 }
660 break;
661
662 case XEN_VM_EVENT_DISABLE:
663 if ( vm_event_check_ring(d->vm_event_paging) )
664 {
665 domain_pause(d);
666 rc = vm_event_disable(d, &d->vm_event_paging);
667 domain_unpause(d);
668 }
669 break;
670
671 case XEN_VM_EVENT_RESUME:
672 if ( vm_event_check_ring(d->vm_event_paging) )
673 vm_event_resume(d, d->vm_event_paging);
674 else
675 rc = -ENODEV;
676 break;
677
678 default:
679 rc = -ENOSYS;
680 break;
681 }
682 }
683 break;
684 #endif
685
686 case XEN_DOMCTL_VM_EVENT_OP_MONITOR:
687 {
688 rc = -EINVAL;
689
690 switch( vec->op )
691 {
692 case XEN_VM_EVENT_ENABLE:
693 /* domain_pause() not required here, see XSA-99 */
694 rc = arch_monitor_init_domain(d);
695 if ( rc )
696 break;
697 rc = vm_event_enable(d, vec, &d->vm_event_monitor, _VPF_mem_access,
698 HVM_PARAM_MONITOR_RING_PFN,
699 monitor_notification);
700 break;
701
702 case XEN_VM_EVENT_DISABLE:
703 if ( vm_event_check_ring(d->vm_event_monitor) )
704 {
705 domain_pause(d);
706 rc = vm_event_disable(d, &d->vm_event_monitor);
707 arch_monitor_cleanup_domain(d);
708 domain_unpause(d);
709 }
710 break;
711
712 case XEN_VM_EVENT_RESUME:
713 if ( vm_event_check_ring(d->vm_event_monitor) )
714 vm_event_resume(d, d->vm_event_monitor);
715 else
716 rc = -ENODEV;
717 break;
718
719 default:
720 rc = -ENOSYS;
721 break;
722 }
723 }
724 break;
725
726 #ifdef CONFIG_HAS_MEM_SHARING
727 case XEN_DOMCTL_VM_EVENT_OP_SHARING:
728 {
729 rc = -EINVAL;
730
731 switch( vec->op )
732 {
733 case XEN_VM_EVENT_ENABLE:
734 rc = -EOPNOTSUPP;
735 /* hvm fixme: p2m_is_foreign types need addressing */
736 if ( is_hvm_domain(hardware_domain) )
737 break;
738
739 rc = -ENODEV;
740 /* Only HAP is supported */
741 if ( !hap_enabled(d) )
742 break;
743
744 /* domain_pause() not required here, see XSA-99 */
745 rc = vm_event_enable(d, vec, &d->vm_event_share, _VPF_mem_sharing,
746 HVM_PARAM_SHARING_RING_PFN,
747 mem_sharing_notification);
748 break;
749
750 case XEN_VM_EVENT_DISABLE:
751 if ( vm_event_check_ring(d->vm_event_share) )
752 {
753 domain_pause(d);
754 rc = vm_event_disable(d, &d->vm_event_share);
755 domain_unpause(d);
756 }
757 break;
758
759 case XEN_VM_EVENT_RESUME:
760 if ( vm_event_check_ring(d->vm_event_share) )
761 vm_event_resume(d, d->vm_event_share);
762 else
763 rc = -ENODEV;
764 break;
765
766 default:
767 rc = -ENOSYS;
768 break;
769 }
770 }
771 break;
772 #endif
773
774 default:
775 rc = -ENOSYS;
776 }
777
778 return rc;
779 }
780
vm_event_vcpu_pause(struct vcpu * v)781 void vm_event_vcpu_pause(struct vcpu *v)
782 {
783 ASSERT(v == current);
784
785 atomic_inc(&v->vm_event_pause_count);
786 vcpu_pause_nosync(v);
787 }
788
vm_event_vcpu_unpause(struct vcpu * v)789 void vm_event_vcpu_unpause(struct vcpu *v)
790 {
791 int old, new, prev = v->vm_event_pause_count.counter;
792
793 /*
794 * All unpause requests as a result of toolstack responses.
795 * Prevent underflow of the vcpu pause count.
796 */
797 do
798 {
799 old = prev;
800 new = old - 1;
801
802 if ( new < 0 )
803 {
804 printk(XENLOG_G_WARNING
805 "%pv vm_event: Too many unpause attempts\n", v);
806 return;
807 }
808
809 prev = cmpxchg(&v->vm_event_pause_count.counter, old, new);
810 } while ( prev != old );
811
812 vcpu_unpause(v);
813 }
814
815 /*
816 * Local variables:
817 * mode: C
818 * c-file-style: "BSD"
819 * c-basic-offset: 4
820 * indent-tabs-mode: nil
821 * End:
822 */
823