Lines Matching refs:hvm_domain

52                           &d->arch.hvm_domain.ioreq_server.list,  in hvm_io_pending()
132 &d->arch.hvm_domain.ioreq_server.list, in handle_hvm_io_completion()
190 for ( i = 0; i < sizeof(d->arch.hvm_domain.ioreq_gfn.mask) * 8; i++ ) in hvm_alloc_ioreq_gfn()
192 if ( test_and_clear_bit(i, &d->arch.hvm_domain.ioreq_gfn.mask) ) in hvm_alloc_ioreq_gfn()
194 *gfn = d->arch.hvm_domain.ioreq_gfn.base + i; in hvm_alloc_ioreq_gfn()
205 unsigned int i = gfn - d->arch.hvm_domain.ioreq_gfn.base; in hvm_free_ioreq_gfn()
208 set_bit(i, &d->arch.hvm_domain.ioreq_gfn.mask); in hvm_free_ioreq_gfn()
248 spin_lock_recursive(&d->arch.hvm_domain.ioreq_server.lock); in is_ioreq_server_page()
251 &d->arch.hvm_domain.ioreq_server.list, in is_ioreq_server_page()
262 spin_unlock_recursive(&d->arch.hvm_domain.ioreq_server.lock); in is_ioreq_server_page()
335 d->arch.hvm_domain.params[HVM_PARAM_BUFIOREQ_EVTCHN] = in hvm_ioreq_server_add_vcpu()
450 d->arch.hvm_domain.params[HVM_PARAM_IOREQ_PFN], in hvm_ioreq_server_setup_pages()
451 d->arch.hvm_domain.params[HVM_PARAM_BUFIOREQ_PFN]); in hvm_ioreq_server_setup_pages()
662 ASSERT(spin_is_locked(&d->arch.hvm_domain.ioreq_server.lock)); in next_ioservid()
664 id = d->arch.hvm_domain.ioreq_server.id; in next_ioservid()
671 &d->arch.hvm_domain.ioreq_server.list, in next_ioservid()
678 d->arch.hvm_domain.ioreq_server.id = id; in next_ioservid()
699 spin_lock_recursive(&d->arch.hvm_domain.ioreq_server.lock); in hvm_create_ioreq_server()
702 if ( is_default && d->arch.hvm_domain.default_ioreq_server != NULL ) in hvm_create_ioreq_server()
711 &d->arch.hvm_domain.ioreq_server.list); in hvm_create_ioreq_server()
715 d->arch.hvm_domain.default_ioreq_server = s; in hvm_create_ioreq_server()
722 spin_unlock_recursive(&d->arch.hvm_domain.ioreq_server.lock); in hvm_create_ioreq_server()
729 spin_unlock_recursive(&d->arch.hvm_domain.ioreq_server.lock); in hvm_create_ioreq_server()
742 spin_lock_recursive(&d->arch.hvm_domain.ioreq_server.lock); in hvm_destroy_ioreq_server()
746 &d->arch.hvm_domain.ioreq_server.list, in hvm_destroy_ioreq_server()
749 if ( s == d->arch.hvm_domain.default_ioreq_server ) in hvm_destroy_ioreq_server()
773 spin_unlock_recursive(&d->arch.hvm_domain.ioreq_server.lock); in hvm_destroy_ioreq_server()
786 spin_lock_recursive(&d->arch.hvm_domain.ioreq_server.lock); in hvm_get_ioreq_server_info()
790 &d->arch.hvm_domain.ioreq_server.list, in hvm_get_ioreq_server_info()
793 if ( s == d->arch.hvm_domain.default_ioreq_server ) in hvm_get_ioreq_server_info()
811 spin_unlock_recursive(&d->arch.hvm_domain.ioreq_server.lock); in hvm_get_ioreq_server_info()
826 spin_lock_recursive(&d->arch.hvm_domain.ioreq_server.lock); in hvm_map_io_range_to_ioreq_server()
830 &d->arch.hvm_domain.ioreq_server.list, in hvm_map_io_range_to_ioreq_server()
833 if ( s == d->arch.hvm_domain.default_ioreq_server ) in hvm_map_io_range_to_ioreq_server()
866 spin_unlock_recursive(&d->arch.hvm_domain.ioreq_server.lock); in hvm_map_io_range_to_ioreq_server()
881 spin_lock_recursive(&d->arch.hvm_domain.ioreq_server.lock); in hvm_unmap_io_range_from_ioreq_server()
885 &d->arch.hvm_domain.ioreq_server.list, in hvm_unmap_io_range_from_ioreq_server()
888 if ( s == d->arch.hvm_domain.default_ioreq_server ) in hvm_unmap_io_range_from_ioreq_server()
921 spin_unlock_recursive(&d->arch.hvm_domain.ioreq_server.lock); in hvm_unmap_io_range_from_ioreq_server()
946 spin_lock_recursive(&d->arch.hvm_domain.ioreq_server.lock); in hvm_map_mem_type_to_ioreq_server()
950 &d->arch.hvm_domain.ioreq_server.list, in hvm_map_mem_type_to_ioreq_server()
953 if ( s == d->arch.hvm_domain.default_ioreq_server ) in hvm_map_mem_type_to_ioreq_server()
963 spin_unlock_recursive(&d->arch.hvm_domain.ioreq_server.lock); in hvm_map_mem_type_to_ioreq_server()
982 spin_lock_recursive(&d->arch.hvm_domain.ioreq_server.lock); in hvm_set_ioreq_server_state()
986 &d->arch.hvm_domain.ioreq_server.list ) in hvm_set_ioreq_server_state()
992 if ( s == d->arch.hvm_domain.default_ioreq_server ) in hvm_set_ioreq_server_state()
1011 spin_unlock_recursive(&d->arch.hvm_domain.ioreq_server.lock); in hvm_set_ioreq_server_state()
1020 spin_lock_recursive(&d->arch.hvm_domain.ioreq_server.lock); in hvm_all_ioreq_servers_add_vcpu()
1023 &d->arch.hvm_domain.ioreq_server.list, in hvm_all_ioreq_servers_add_vcpu()
1026 bool is_default = (s == d->arch.hvm_domain.default_ioreq_server); in hvm_all_ioreq_servers_add_vcpu()
1033 spin_unlock_recursive(&d->arch.hvm_domain.ioreq_server.lock); in hvm_all_ioreq_servers_add_vcpu()
1039 &d->arch.hvm_domain.ioreq_server.list, in hvm_all_ioreq_servers_add_vcpu()
1043 spin_unlock_recursive(&d->arch.hvm_domain.ioreq_server.lock); in hvm_all_ioreq_servers_add_vcpu()
1052 spin_lock_recursive(&d->arch.hvm_domain.ioreq_server.lock); in hvm_all_ioreq_servers_remove_vcpu()
1055 &d->arch.hvm_domain.ioreq_server.list, in hvm_all_ioreq_servers_remove_vcpu()
1059 spin_unlock_recursive(&d->arch.hvm_domain.ioreq_server.lock); in hvm_all_ioreq_servers_remove_vcpu()
1066 spin_lock_recursive(&d->arch.hvm_domain.ioreq_server.lock); in hvm_destroy_all_ioreq_servers()
1072 &d->arch.hvm_domain.ioreq_server.list, in hvm_destroy_all_ioreq_servers()
1075 bool is_default = (s == d->arch.hvm_domain.default_ioreq_server); in hvm_destroy_all_ioreq_servers()
1080 d->arch.hvm_domain.default_ioreq_server = NULL; in hvm_destroy_all_ioreq_servers()
1089 spin_unlock_recursive(&d->arch.hvm_domain.ioreq_server.lock); in hvm_destroy_all_ioreq_servers()
1113 spin_lock_recursive(&d->arch.hvm_domain.ioreq_server.lock); in hvm_set_dm_domain()
1120 s = d->arch.hvm_domain.default_ioreq_server; in hvm_set_dm_domain()
1144 d->arch.hvm_domain.params[HVM_PARAM_BUFIOREQ_EVTCHN] = in hvm_set_dm_domain()
1162 spin_unlock_recursive(&d->arch.hvm_domain.ioreq_server.lock); in hvm_set_dm_domain()
1174 if ( list_empty(&d->arch.hvm_domain.ioreq_server.list) ) in hvm_select_ioreq_server()
1178 return d->arch.hvm_domain.default_ioreq_server; in hvm_select_ioreq_server()
1180 cf8 = d->arch.hvm_domain.pci_cf8; in hvm_select_ioreq_server()
1219 &d->arch.hvm_domain.ioreq_server.list, in hvm_select_ioreq_server()
1224 if ( s == d->arch.hvm_domain.default_ioreq_server ) in hvm_select_ioreq_server()
1260 return d->arch.hvm_domain.default_ioreq_server; in hvm_select_ioreq_server()
1422 &d->arch.hvm_domain.ioreq_server.list, in hvm_broadcast_ioreq()
1436 d->arch.hvm_domain.pci_cf8 = *val; in hvm_access_cf8()
1444 spin_lock_init(&d->arch.hvm_domain.ioreq_server.lock); in hvm_ioreq_init()
1445 INIT_LIST_HEAD(&d->arch.hvm_domain.ioreq_server.list); in hvm_ioreq_init()