1 /******************************************************************************
2  * vm_event.c
3  *
4  * VM event support.
5  *
6  * Copyright (c) 2009 Citrix Systems, Inc. (Patrick Colp)
7  *
8  * This program is free software; you can redistribute it and/or modify
9  * it under the terms of the GNU General Public License as published by
10  * the Free Software Foundation; either version 2 of the License, or
11  * (at your option) any later version.
12  *
13  * This program is distributed in the hope that it will be useful,
14  * but WITHOUT ANY WARRANTY; without even the implied warranty of
15  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
16  * GNU General Public License for more details.
17  *
18  * You should have received a copy of the GNU General Public License
19  * along with this program; If not, see <http://www.gnu.org/licenses/>.
20  */
21 
22 
23 #include <xen/sched.h>
24 #include <xen/event.h>
25 #include <xen/wait.h>
26 #include <xen/vm_event.h>
27 #include <xen/mem_access.h>
28 #include <asm/p2m.h>
29 #include <asm/monitor.h>
30 #include <asm/vm_event.h>
31 #include <xsm/xsm.h>
32 
33 /* for public/io/ring.h macros */
34 #define xen_mb()   smp_mb()
35 #define xen_rmb()  smp_rmb()
36 #define xen_wmb()  smp_wmb()
37 
38 #define vm_event_ring_lock_init(_ved)  spin_lock_init(&(_ved)->ring_lock)
39 #define vm_event_ring_lock(_ved)       spin_lock(&(_ved)->ring_lock)
40 #define vm_event_ring_unlock(_ved)     spin_unlock(&(_ved)->ring_lock)
41 
vm_event_enable(struct domain * d,struct xen_domctl_vm_event_op * vec,struct vm_event_domain ** ved,int pause_flag,int param,xen_event_channel_notification_t notification_fn)42 static int vm_event_enable(
43     struct domain *d,
44     struct xen_domctl_vm_event_op *vec,
45     struct vm_event_domain **ved,
46     int pause_flag,
47     int param,
48     xen_event_channel_notification_t notification_fn)
49 {
50     int rc;
51     unsigned long ring_gfn = d->arch.hvm_domain.params[param];
52 
53     if ( !*ved )
54         *ved = xzalloc(struct vm_event_domain);
55     if ( !*ved )
56         return -ENOMEM;
57 
58     /* Only one helper at a time. If the helper crashed,
59      * the ring is in an undefined state and so is the guest.
60      */
61     if ( (*ved)->ring_page )
62         return -EBUSY;;
63 
64     /* The parameter defaults to zero, and it should be
65      * set to something */
66     if ( ring_gfn == 0 )
67         return -ENOSYS;
68 
69     vm_event_ring_lock_init(*ved);
70     vm_event_ring_lock(*ved);
71 
72     rc = vm_event_init_domain(d);
73 
74     if ( rc < 0 )
75         goto err;
76 
77     rc = prepare_ring_for_helper(d, ring_gfn, &(*ved)->ring_pg_struct,
78                                  &(*ved)->ring_page);
79     if ( rc < 0 )
80         goto err;
81 
82     /* Set the number of currently blocked vCPUs to 0. */
83     (*ved)->blocked = 0;
84 
85     /* Allocate event channel */
86     rc = alloc_unbound_xen_event_channel(d, 0, current->domain->domain_id,
87                                          notification_fn);
88     if ( rc < 0 )
89         goto err;
90 
91     (*ved)->xen_port = vec->port = rc;
92 
93     /* Prepare ring buffer */
94     FRONT_RING_INIT(&(*ved)->front_ring,
95                     (vm_event_sring_t *)(*ved)->ring_page,
96                     PAGE_SIZE);
97 
98     /* Save the pause flag for this particular ring. */
99     (*ved)->pause_flag = pause_flag;
100 
101     /* Initialize the last-chance wait queue. */
102     init_waitqueue_head(&(*ved)->wq);
103 
104     vm_event_ring_unlock(*ved);
105     return 0;
106 
107  err:
108     destroy_ring_for_helper(&(*ved)->ring_page,
109                             (*ved)->ring_pg_struct);
110     vm_event_ring_unlock(*ved);
111     xfree(*ved);
112     *ved = NULL;
113 
114     return rc;
115 }
116 
vm_event_ring_available(struct vm_event_domain * ved)117 static unsigned int vm_event_ring_available(struct vm_event_domain *ved)
118 {
119     int avail_req = RING_FREE_REQUESTS(&ved->front_ring);
120     avail_req -= ved->target_producers;
121     avail_req -= ved->foreign_producers;
122 
123     BUG_ON(avail_req < 0);
124 
125     return avail_req;
126 }
127 
128 /*
129  * vm_event_wake_blocked() will wakeup vcpus waiting for room in the
130  * ring. These vCPUs were paused on their way out after placing an event,
131  * but need to be resumed where the ring is capable of processing at least
132  * one event from them.
133  */
vm_event_wake_blocked(struct domain * d,struct vm_event_domain * ved)134 static void vm_event_wake_blocked(struct domain *d, struct vm_event_domain *ved)
135 {
136     struct vcpu *v;
137     unsigned int avail_req = vm_event_ring_available(ved);
138 
139     if ( avail_req == 0 || ved->blocked == 0 )
140         return;
141 
142     /* We remember which vcpu last woke up to avoid scanning always linearly
143      * from zero and starving higher-numbered vcpus under high load */
144     if ( d->vcpu )
145     {
146         int i, j, k;
147 
148         for (i = ved->last_vcpu_wake_up + 1, j = 0; j < d->max_vcpus; i++, j++)
149         {
150             k = i % d->max_vcpus;
151             v = d->vcpu[k];
152             if ( !v )
153                 continue;
154 
155             if ( !(ved->blocked) || avail_req == 0 )
156                break;
157 
158             if ( test_and_clear_bit(ved->pause_flag, &v->pause_flags) )
159             {
160                 vcpu_unpause(v);
161                 avail_req--;
162                 ved->blocked--;
163                 ved->last_vcpu_wake_up = k;
164             }
165         }
166     }
167 }
168 
169 /*
170  * In the event that a vCPU attempted to place an event in the ring and
171  * was unable to do so, it is queued on a wait queue.  These are woken as
172  * needed, and take precedence over the blocked vCPUs.
173  */
vm_event_wake_queued(struct domain * d,struct vm_event_domain * ved)174 static void vm_event_wake_queued(struct domain *d, struct vm_event_domain *ved)
175 {
176     unsigned int avail_req = vm_event_ring_available(ved);
177 
178     if ( avail_req > 0 )
179         wake_up_nr(&ved->wq, avail_req);
180 }
181 
182 /*
183  * vm_event_wake() will wakeup all vcpus waiting for the ring to
184  * become available.  If we have queued vCPUs, they get top priority. We
185  * are guaranteed that they will go through code paths that will eventually
186  * call vm_event_wake() again, ensuring that any blocked vCPUs will get
187  * unpaused once all the queued vCPUs have made it through.
188  */
vm_event_wake(struct domain * d,struct vm_event_domain * ved)189 void vm_event_wake(struct domain *d, struct vm_event_domain *ved)
190 {
191     if (!list_empty(&ved->wq.list))
192         vm_event_wake_queued(d, ved);
193     else
194         vm_event_wake_blocked(d, ved);
195 }
196 
vm_event_disable(struct domain * d,struct vm_event_domain ** ved)197 static int vm_event_disable(struct domain *d, struct vm_event_domain **ved)
198 {
199     if ( vm_event_check_ring(*ved) )
200     {
201         struct vcpu *v;
202 
203         vm_event_ring_lock(*ved);
204 
205         if ( !list_empty(&(*ved)->wq.list) )
206         {
207             vm_event_ring_unlock(*ved);
208             return -EBUSY;
209         }
210 
211         /* Free domU's event channel and leave the other one unbound */
212         free_xen_event_channel(d, (*ved)->xen_port);
213 
214         /* Unblock all vCPUs */
215         for_each_vcpu ( d, v )
216         {
217             if ( test_and_clear_bit((*ved)->pause_flag, &v->pause_flags) )
218             {
219                 vcpu_unpause(v);
220                 (*ved)->blocked--;
221             }
222         }
223 
224         destroy_ring_for_helper(&(*ved)->ring_page,
225                                 (*ved)->ring_pg_struct);
226 
227         vm_event_cleanup_domain(d);
228 
229         vm_event_ring_unlock(*ved);
230     }
231 
232     xfree(*ved);
233     *ved = NULL;
234 
235     return 0;
236 }
237 
vm_event_release_slot(struct domain * d,struct vm_event_domain * ved)238 static inline void vm_event_release_slot(struct domain *d,
239                                          struct vm_event_domain *ved)
240 {
241     /* Update the accounting */
242     if ( current->domain == d )
243         ved->target_producers--;
244     else
245         ved->foreign_producers--;
246 
247     /* Kick any waiters */
248     vm_event_wake(d, ved);
249 }
250 
251 /*
252  * vm_event_mark_and_pause() tags vcpu and put it to sleep.
253  * The vcpu will resume execution in vm_event_wake_blocked().
254  */
vm_event_mark_and_pause(struct vcpu * v,struct vm_event_domain * ved)255 void vm_event_mark_and_pause(struct vcpu *v, struct vm_event_domain *ved)
256 {
257     if ( !test_and_set_bit(ved->pause_flag, &v->pause_flags) )
258     {
259         vcpu_pause_nosync(v);
260         ved->blocked++;
261     }
262 }
263 
264 /*
265  * This must be preceded by a call to claim_slot(), and is guaranteed to
266  * succeed.  As a side-effect however, the vCPU may be paused if the ring is
267  * overly full and its continued execution would cause stalling and excessive
268  * waiting.  The vCPU will be automatically unpaused when the ring clears.
269  */
vm_event_put_request(struct domain * d,struct vm_event_domain * ved,vm_event_request_t * req)270 void vm_event_put_request(struct domain *d,
271                           struct vm_event_domain *ved,
272                           vm_event_request_t *req)
273 {
274     vm_event_front_ring_t *front_ring;
275     int free_req;
276     unsigned int avail_req;
277     RING_IDX req_prod;
278     struct vcpu *curr = current;
279 
280     if( !vm_event_check_ring(ved))
281         return;
282 
283     if ( curr->domain != d )
284     {
285         req->flags |= VM_EVENT_FLAG_FOREIGN;
286 #ifndef NDEBUG
287         if ( !(req->flags & VM_EVENT_FLAG_VCPU_PAUSED) )
288             gdprintk(XENLOG_G_WARNING, "d%dv%d was not paused.\n",
289                      d->domain_id, req->vcpu_id);
290 #endif
291     }
292 
293     req->version = VM_EVENT_INTERFACE_VERSION;
294 
295     vm_event_ring_lock(ved);
296 
297     /* Due to the reservations, this step must succeed. */
298     front_ring = &ved->front_ring;
299     free_req = RING_FREE_REQUESTS(front_ring);
300     ASSERT(free_req > 0);
301 
302     /* Copy request */
303     req_prod = front_ring->req_prod_pvt;
304     memcpy(RING_GET_REQUEST(front_ring, req_prod), req, sizeof(*req));
305     req_prod++;
306 
307     /* Update ring */
308     front_ring->req_prod_pvt = req_prod;
309     RING_PUSH_REQUESTS(front_ring);
310 
311     /* We've actually *used* our reservation, so release the slot. */
312     vm_event_release_slot(d, ved);
313 
314     /* Give this vCPU a black eye if necessary, on the way out.
315      * See the comments above wake_blocked() for more information
316      * on how this mechanism works to avoid waiting. */
317     avail_req = vm_event_ring_available(ved);
318     if( curr->domain == d && avail_req < d->max_vcpus &&
319         !atomic_read(&curr->vm_event_pause_count) )
320         vm_event_mark_and_pause(curr, ved);
321 
322     vm_event_ring_unlock(ved);
323 
324     notify_via_xen_event_channel(d, ved->xen_port);
325 }
326 
vm_event_get_response(struct domain * d,struct vm_event_domain * ved,vm_event_response_t * rsp)327 int vm_event_get_response(struct domain *d, struct vm_event_domain *ved,
328                           vm_event_response_t *rsp)
329 {
330     vm_event_front_ring_t *front_ring;
331     RING_IDX rsp_cons;
332 
333     vm_event_ring_lock(ved);
334 
335     front_ring = &ved->front_ring;
336     rsp_cons = front_ring->rsp_cons;
337 
338     if ( !RING_HAS_UNCONSUMED_RESPONSES(front_ring) )
339     {
340         vm_event_ring_unlock(ved);
341         return 0;
342     }
343 
344     /* Copy response */
345     memcpy(rsp, RING_GET_RESPONSE(front_ring, rsp_cons), sizeof(*rsp));
346     rsp_cons++;
347 
348     /* Update ring */
349     front_ring->rsp_cons = rsp_cons;
350     front_ring->sring->rsp_event = rsp_cons + 1;
351 
352     /* Kick any waiters -- since we've just consumed an event,
353      * there may be additional space available in the ring. */
354     vm_event_wake(d, ved);
355 
356     vm_event_ring_unlock(ved);
357 
358     return 1;
359 }
360 
361 /*
362  * Pull all responses from the given ring and unpause the corresponding vCPU
363  * if required. Based on the response type, here we can also call custom
364  * handlers.
365  *
366  * Note: responses are handled the same way regardless of which ring they
367  * arrive on.
368  */
vm_event_resume(struct domain * d,struct vm_event_domain * ved)369 void vm_event_resume(struct domain *d, struct vm_event_domain *ved)
370 {
371     vm_event_response_t rsp;
372 
373     /*
374      * vm_event_resume() runs in either XEN_DOMCTL_VM_EVENT_OP_*, or
375      * EVTCHN_send context from the introspection consumer. Both contexts
376      * are guaranteed not to be the subject of vm_event responses.
377      * While we could ASSERT(v != current) for each VCPU in d in the loop
378      * below, this covers the case where we would need to iterate over all
379      * of them more succintly.
380      */
381     ASSERT(d != current->domain);
382 
383     /* Pull all responses off the ring. */
384     while ( vm_event_get_response(d, ved, &rsp) )
385     {
386         struct vcpu *v;
387 
388         if ( rsp.version != VM_EVENT_INTERFACE_VERSION )
389         {
390             printk(XENLOG_G_WARNING "vm_event interface version mismatch\n");
391             continue;
392         }
393 
394         /* Validate the vcpu_id in the response. */
395         if ( (rsp.vcpu_id >= d->max_vcpus) || !d->vcpu[rsp.vcpu_id] )
396             continue;
397 
398         v = d->vcpu[rsp.vcpu_id];
399 
400         /*
401          * In some cases the response type needs extra handling, so here
402          * we call the appropriate handlers.
403          */
404 
405         /* Check flags which apply only when the vCPU is paused */
406         if ( atomic_read(&v->vm_event_pause_count) )
407         {
408 #ifdef CONFIG_HAS_MEM_PAGING
409             if ( rsp.reason == VM_EVENT_REASON_MEM_PAGING )
410                 p2m_mem_paging_resume(d, &rsp);
411 #endif
412 
413             /*
414              * Check emulation flags in the arch-specific handler only, as it
415              * has to set arch-specific flags when supported, and to avoid
416              * bitmask overhead when it isn't supported.
417              */
418             vm_event_emulate_check(v, &rsp);
419 
420             /*
421              * Check in arch-specific handler to avoid bitmask overhead when
422              * not supported.
423              */
424             vm_event_register_write_resume(v, &rsp);
425 
426             /*
427              * Check in arch-specific handler to avoid bitmask overhead when
428              * not supported.
429              */
430             vm_event_toggle_singlestep(d, v, &rsp);
431 
432             /* Check for altp2m switch */
433             if ( rsp.flags & VM_EVENT_FLAG_ALTERNATE_P2M )
434                 p2m_altp2m_check(v, rsp.altp2m_idx);
435 
436             if ( rsp.flags & VM_EVENT_FLAG_SET_REGISTERS )
437                 vm_event_set_registers(v, &rsp);
438 
439             if ( rsp.flags & VM_EVENT_FLAG_GET_NEXT_INTERRUPT )
440                 vm_event_monitor_next_interrupt(v);
441 
442             if ( rsp.flags & VM_EVENT_FLAG_VCPU_PAUSED )
443                 vm_event_vcpu_unpause(v);
444         }
445     }
446 }
447 
vm_event_cancel_slot(struct domain * d,struct vm_event_domain * ved)448 void vm_event_cancel_slot(struct domain *d, struct vm_event_domain *ved)
449 {
450     if( !vm_event_check_ring(ved) )
451         return;
452 
453     vm_event_ring_lock(ved);
454     vm_event_release_slot(d, ved);
455     vm_event_ring_unlock(ved);
456 }
457 
vm_event_grab_slot(struct vm_event_domain * ved,int foreign)458 static int vm_event_grab_slot(struct vm_event_domain *ved, int foreign)
459 {
460     unsigned int avail_req;
461 
462     if ( !ved->ring_page )
463         return -ENOSYS;
464 
465     vm_event_ring_lock(ved);
466 
467     avail_req = vm_event_ring_available(ved);
468     if ( avail_req == 0 )
469     {
470         vm_event_ring_unlock(ved);
471         return -EBUSY;
472     }
473 
474     if ( !foreign )
475         ved->target_producers++;
476     else
477         ved->foreign_producers++;
478 
479     vm_event_ring_unlock(ved);
480 
481     return 0;
482 }
483 
484 /* Simple try_grab wrapper for use in the wait_event() macro. */
vm_event_wait_try_grab(struct vm_event_domain * ved,int * rc)485 static int vm_event_wait_try_grab(struct vm_event_domain *ved, int *rc)
486 {
487     *rc = vm_event_grab_slot(ved, 0);
488     return *rc;
489 }
490 
491 /* Call vm_event_grab_slot() until the ring doesn't exist, or is available. */
vm_event_wait_slot(struct vm_event_domain * ved)492 static int vm_event_wait_slot(struct vm_event_domain *ved)
493 {
494     int rc = -EBUSY;
495     wait_event(ved->wq, vm_event_wait_try_grab(ved, &rc) != -EBUSY);
496     return rc;
497 }
498 
vm_event_check_ring(struct vm_event_domain * ved)499 bool_t vm_event_check_ring(struct vm_event_domain *ved)
500 {
501     return (ved && ved->ring_page);
502 }
503 
504 /*
505  * Determines whether or not the current vCPU belongs to the target domain,
506  * and calls the appropriate wait function.  If it is a guest vCPU, then we
507  * use vm_event_wait_slot() to reserve a slot.  As long as there is a ring,
508  * this function will always return 0 for a guest.  For a non-guest, we check
509  * for space and return -EBUSY if the ring is not available.
510  *
511  * Return codes: -ENOSYS: the ring is not yet configured
512  *               -EBUSY: the ring is busy
513  *               0: a spot has been reserved
514  *
515  */
__vm_event_claim_slot(struct domain * d,struct vm_event_domain * ved,bool_t allow_sleep)516 int __vm_event_claim_slot(struct domain *d, struct vm_event_domain *ved,
517                           bool_t allow_sleep)
518 {
519     if ( !vm_event_check_ring(ved) )
520         return -EOPNOTSUPP;
521 
522     if ( (current->domain == d) && allow_sleep )
523         return vm_event_wait_slot(ved);
524     else
525         return vm_event_grab_slot(ved, (current->domain != d));
526 }
527 
528 #ifdef CONFIG_HAS_MEM_PAGING
529 /* Registered with Xen-bound event channel for incoming notifications. */
mem_paging_notification(struct vcpu * v,unsigned int port)530 static void mem_paging_notification(struct vcpu *v, unsigned int port)
531 {
532     struct domain *domain = v->domain;
533 
534     if ( likely(vm_event_check_ring(domain->vm_event_paging)) )
535         vm_event_resume(domain, domain->vm_event_paging);
536 }
537 #endif
538 
539 /* Registered with Xen-bound event channel for incoming notifications. */
monitor_notification(struct vcpu * v,unsigned int port)540 static void monitor_notification(struct vcpu *v, unsigned int port)
541 {
542     struct domain *domain = v->domain;
543 
544     if ( likely(vm_event_check_ring(domain->vm_event_monitor)) )
545         vm_event_resume(domain, domain->vm_event_monitor);
546 }
547 
548 #ifdef CONFIG_HAS_MEM_SHARING
549 /* Registered with Xen-bound event channel for incoming notifications. */
mem_sharing_notification(struct vcpu * v,unsigned int port)550 static void mem_sharing_notification(struct vcpu *v, unsigned int port)
551 {
552     struct domain *domain = v->domain;
553 
554     if ( likely(vm_event_check_ring(domain->vm_event_share)) )
555         vm_event_resume(domain, domain->vm_event_share);
556 }
557 #endif
558 
559 /* Clean up on domain destruction */
vm_event_cleanup(struct domain * d)560 void vm_event_cleanup(struct domain *d)
561 {
562 #ifdef CONFIG_HAS_MEM_PAGING
563     if ( vm_event_check_ring(d->vm_event_paging) )
564     {
565         /* Destroying the wait queue head means waking up all
566          * queued vcpus. This will drain the list, allowing
567          * the disable routine to complete. It will also drop
568          * all domain refs the wait-queued vcpus are holding.
569          * Finally, because this code path involves previously
570          * pausing the domain (domain_kill), unpausing the
571          * vcpus causes no harm. */
572         destroy_waitqueue_head(&d->vm_event_paging->wq);
573         (void)vm_event_disable(d, &d->vm_event_paging);
574     }
575 #endif
576     if ( vm_event_check_ring(d->vm_event_monitor) )
577     {
578         destroy_waitqueue_head(&d->vm_event_monitor->wq);
579         (void)vm_event_disable(d, &d->vm_event_monitor);
580     }
581 #ifdef CONFIG_HAS_MEM_SHARING
582     if ( vm_event_check_ring(d->vm_event_share) )
583     {
584         destroy_waitqueue_head(&d->vm_event_share->wq);
585         (void)vm_event_disable(d, &d->vm_event_share);
586     }
587 #endif
588 }
589 
vm_event_domctl(struct domain * d,struct xen_domctl_vm_event_op * vec,XEN_GUEST_HANDLE_PARAM (void)u_domctl)590 int vm_event_domctl(struct domain *d, struct xen_domctl_vm_event_op *vec,
591                     XEN_GUEST_HANDLE_PARAM(void) u_domctl)
592 {
593     int rc;
594 
595     rc = xsm_vm_event_control(XSM_PRIV, d, vec->mode, vec->op);
596     if ( rc )
597         return rc;
598 
599     if ( unlikely(d == current->domain) ) /* no domain_pause() */
600     {
601         gdprintk(XENLOG_INFO, "Tried to do a memory event op on itself.\n");
602         return -EINVAL;
603     }
604 
605     if ( unlikely(d->is_dying) )
606     {
607         gdprintk(XENLOG_INFO, "Ignoring memory event op on dying domain %u\n",
608                  d->domain_id);
609         return 0;
610     }
611 
612     if ( unlikely(d->vcpu == NULL) || unlikely(d->vcpu[0] == NULL) )
613     {
614         gdprintk(XENLOG_INFO,
615                  "Memory event op on a domain (%u) with no vcpus\n",
616                  d->domain_id);
617         return -EINVAL;
618     }
619 
620     rc = -ENOSYS;
621 
622     switch ( vec->mode )
623     {
624 #ifdef CONFIG_HAS_MEM_PAGING
625     case XEN_DOMCTL_VM_EVENT_OP_PAGING:
626     {
627         rc = -EINVAL;
628 
629         switch( vec->op )
630         {
631         case XEN_VM_EVENT_ENABLE:
632         {
633             struct p2m_domain *p2m = p2m_get_hostp2m(d);
634 
635             rc = -EOPNOTSUPP;
636             /* hvm fixme: p2m_is_foreign types need addressing */
637             if ( is_hvm_domain(hardware_domain) )
638                 break;
639 
640             rc = -ENODEV;
641             /* Only HAP is supported */
642             if ( !hap_enabled(d) )
643                 break;
644 
645             /* No paging if iommu is used */
646             rc = -EMLINK;
647             if ( unlikely(need_iommu(d)) )
648                 break;
649 
650             rc = -EXDEV;
651             /* Disallow paging in a PoD guest */
652             if ( p2m->pod.entry_count )
653                 break;
654 
655             /* domain_pause() not required here, see XSA-99 */
656             rc = vm_event_enable(d, vec, &d->vm_event_paging, _VPF_mem_paging,
657                                  HVM_PARAM_PAGING_RING_PFN,
658                                  mem_paging_notification);
659         }
660         break;
661 
662         case XEN_VM_EVENT_DISABLE:
663             if ( vm_event_check_ring(d->vm_event_paging) )
664             {
665                 domain_pause(d);
666                 rc = vm_event_disable(d, &d->vm_event_paging);
667                 domain_unpause(d);
668             }
669             break;
670 
671         case XEN_VM_EVENT_RESUME:
672             if ( vm_event_check_ring(d->vm_event_paging) )
673                 vm_event_resume(d, d->vm_event_paging);
674             else
675                 rc = -ENODEV;
676             break;
677 
678         default:
679             rc = -ENOSYS;
680             break;
681         }
682     }
683     break;
684 #endif
685 
686     case XEN_DOMCTL_VM_EVENT_OP_MONITOR:
687     {
688         rc = -EINVAL;
689 
690         switch( vec->op )
691         {
692         case XEN_VM_EVENT_ENABLE:
693             /* domain_pause() not required here, see XSA-99 */
694             rc = arch_monitor_init_domain(d);
695             if ( rc )
696                 break;
697             rc = vm_event_enable(d, vec, &d->vm_event_monitor, _VPF_mem_access,
698                                  HVM_PARAM_MONITOR_RING_PFN,
699                                  monitor_notification);
700             break;
701 
702         case XEN_VM_EVENT_DISABLE:
703             if ( vm_event_check_ring(d->vm_event_monitor) )
704             {
705                 domain_pause(d);
706                 rc = vm_event_disable(d, &d->vm_event_monitor);
707                 arch_monitor_cleanup_domain(d);
708                 domain_unpause(d);
709             }
710             break;
711 
712         case XEN_VM_EVENT_RESUME:
713             if ( vm_event_check_ring(d->vm_event_monitor) )
714                 vm_event_resume(d, d->vm_event_monitor);
715             else
716                 rc = -ENODEV;
717             break;
718 
719         default:
720             rc = -ENOSYS;
721             break;
722         }
723     }
724     break;
725 
726 #ifdef CONFIG_HAS_MEM_SHARING
727     case XEN_DOMCTL_VM_EVENT_OP_SHARING:
728     {
729         rc = -EINVAL;
730 
731         switch( vec->op )
732         {
733         case XEN_VM_EVENT_ENABLE:
734             rc = -EOPNOTSUPP;
735             /* hvm fixme: p2m_is_foreign types need addressing */
736             if ( is_hvm_domain(hardware_domain) )
737                 break;
738 
739             rc = -ENODEV;
740             /* Only HAP is supported */
741             if ( !hap_enabled(d) )
742                 break;
743 
744             /* domain_pause() not required here, see XSA-99 */
745             rc = vm_event_enable(d, vec, &d->vm_event_share, _VPF_mem_sharing,
746                                  HVM_PARAM_SHARING_RING_PFN,
747                                  mem_sharing_notification);
748             break;
749 
750         case XEN_VM_EVENT_DISABLE:
751             if ( vm_event_check_ring(d->vm_event_share) )
752             {
753                 domain_pause(d);
754                 rc = vm_event_disable(d, &d->vm_event_share);
755                 domain_unpause(d);
756             }
757             break;
758 
759         case XEN_VM_EVENT_RESUME:
760             if ( vm_event_check_ring(d->vm_event_share) )
761                 vm_event_resume(d, d->vm_event_share);
762             else
763                 rc = -ENODEV;
764             break;
765 
766         default:
767             rc = -ENOSYS;
768             break;
769         }
770     }
771     break;
772 #endif
773 
774     default:
775         rc = -ENOSYS;
776     }
777 
778     return rc;
779 }
780 
vm_event_vcpu_pause(struct vcpu * v)781 void vm_event_vcpu_pause(struct vcpu *v)
782 {
783     ASSERT(v == current);
784 
785     atomic_inc(&v->vm_event_pause_count);
786     vcpu_pause_nosync(v);
787 }
788 
vm_event_vcpu_unpause(struct vcpu * v)789 void vm_event_vcpu_unpause(struct vcpu *v)
790 {
791     int old, new, prev = v->vm_event_pause_count.counter;
792 
793     /*
794      * All unpause requests as a result of toolstack responses.
795      * Prevent underflow of the vcpu pause count.
796      */
797     do
798     {
799         old = prev;
800         new = old - 1;
801 
802         if ( new < 0 )
803         {
804             printk(XENLOG_G_WARNING
805                    "%pv vm_event: Too many unpause attempts\n", v);
806             return;
807         }
808 
809         prev = cmpxchg(&v->vm_event_pause_count.counter, old, new);
810     } while ( prev != old );
811 
812     vcpu_unpause(v);
813 }
814 
815 /*
816  * Local variables:
817  * mode: C
818  * c-file-style: "BSD"
819  * c-basic-offset: 4
820  * indent-tabs-mode: nil
821  * End:
822  */
823