/linux-6.3-rc2/drivers/gpu/drm/vmwgfx/ |
A D | vmwgfx_page_dirty.c | 83 struct vmw_bo_dirty *dirty = vbo->dirty; in vmw_bo_dirty_scan_pagetable() 168 vmw_bo_dirty_scan_pagetable(vbo); in vmw_bo_dirty_scan() 170 vmw_bo_dirty_scan_mkwrite(vbo); in vmw_bo_dirty_scan() 269 vbo->dirty = dirty; in vmw_bo_dirty_add() 293 vbo->dirty = NULL; in vmw_bo_dirty_release() 383 struct vmw_buffer_object *vbo = in vmw_bo_vm_mkwrite() local 384 container_of(bo, typeof(*vbo), base); in vmw_bo_vm_mkwrite() 403 if (vbo->dirty && vbo->dirty->method == VMW_BO_DIRTY_MKWRITE && in vmw_bo_vm_mkwrite() 422 struct vmw_buffer_object *vbo = in vmw_bo_vm_fault() local 435 if (vbo->dirty) { in vmw_bo_vm_fault() [all …]
|
A D | vmwgfx_validation.c | 176 struct vmw_buffer_object *vbo) in vmw_validation_find_bo_dup() argument 267 struct vmw_buffer_object *vbo, in vmw_validation_add_bo() argument 424 val->new_backup = vbo; in vmw_validation_res_switch_backup() 529 struct vmw_buffer_object *vbo = in vmw_validation_bo_validate_single() local 540 if (vbo->base.pin_count > 0) in vmw_validation_bo_validate_single() 581 struct vmw_buffer_object *vbo = in vmw_validation_bo_validate() local 608 ret = vmw_bo_dirty_add(vbo); in vmw_validation_bo_validate() 617 if (vbo->dirty) in vmw_validation_bo_validate() 618 vmw_bo_dirty_scan(vbo); in vmw_validation_bo_validate() 892 struct vmw_buffer_object *vbo = in vmw_validation_bo_backoff() local [all …]
|
A D | vmwgfx_bo.c | 373 if (vbo->map.bo == NULL) in vmw_bo_unmap() 376 ttm_bo_kunmap(&vbo->map); in vmw_bo_unmap() 623 struct vmw_buffer_object *vbo; in vmw_user_bo_synccpu_ioctl() local 641 vmw_bo_unreference(&vbo); in vmw_user_bo_synccpu_ioctl() 642 drm_gem_object_put(&vbo->base.base); in vmw_user_bo_synccpu_ioctl() 774 struct vmw_buffer_object *vbo; in vmw_dumb_create() local 796 &vbo); in vmw_dumb_create() 798 drm_gem_object_put(&vbo->base.base); in vmw_dumb_create() 831 struct vmw_buffer_object *vbo; in vmw_bo_move_notify() local 845 vmw_bo_unmap(vbo); in vmw_bo_move_notify() [all …]
|
A D | vmwgfx_gem.c | 68 struct vmw_buffer_object *vbo = vmw_buffer_object(bo); in vmw_gem_pin_private() local 75 vmw_bo_pin_reserved(vbo, do_pin); in vmw_gem_pin_private() 127 struct vmw_buffer_object *vbo = vmw_buffer_object(bo); in vmw_gem_destroy() local 129 WARN_ON(vbo->dirty); in vmw_gem_destroy() 130 WARN_ON(!RB_EMPTY_ROOT(&vbo->res_tree)); in vmw_gem_destroy() 131 vmw_bo_unmap(vbo); in vmw_gem_destroy() 132 drm_gem_object_release(&vbo->base.base); in vmw_gem_destroy() 133 kfree(vbo); in vmw_gem_destroy() 168 struct vmw_buffer_object *vbo; in vmw_gem_object_create_ioctl() local 173 req->size, &handle, &vbo); in vmw_gem_object_create_ioctl() [all …]
|
A D | vmwgfx_resource.c | 746 .bo = &vbo->base, in vmw_resource_unbind_list() 962 struct vmw_buffer_object *vbo = NULL; in vmw_resource_pin() local 965 vbo = res->backup; in vmw_resource_pin() 970 if (!vbo->base.pin_count) { in vmw_resource_pin() 972 (&vbo->base, in vmw_resource_pin() 976 ttm_bo_unreserve(&vbo->base); in vmw_resource_pin() 982 vmw_bo_pin_reserved(vbo, true); in vmw_resource_pin() 985 if (vbo) in vmw_resource_pin() 986 ttm_bo_unreserve(&vbo->base); in vmw_resource_pin() 1023 vmw_bo_pin_reserved(vbo, false); in vmw_resource_unpin() [all …]
|
A D | vmwgfx_drv.h | 933 int i = ARRAY_SIZE(vbo->res_prios); in vmw_bo_prio_adjust() 936 if (vbo->res_prios[i]) { in vmw_bo_prio_adjust() 937 vbo->base.priority = i; in vmw_bo_prio_adjust() 942 vbo->base.priority = 3; in vmw_bo_prio_adjust() 956 if (vbo->res_prios[prio]++ == 0) in vmw_bo_prio_add() 957 vmw_bo_prio_adjust(vbo); in vmw_bo_prio_add() 971 if (--vbo->res_prios[prio] == 0) in vmw_bo_prio_del() 972 vmw_bo_prio_adjust(vbo); in vmw_bo_prio_del() 1526 void vmw_bo_dirty_scan(struct vmw_buffer_object *vbo); 1527 int vmw_bo_dirty_add(struct vmw_buffer_object *vbo); [all …]
|
A D | vmwgfx_validation.h | 162 struct vmw_buffer_object *vbo, 182 struct vmw_buffer_object *vbo,
|
A D | vmwgfx_drv.c | 389 struct vmw_buffer_object *vbo; in vmw_dummy_query_bo_create() local 401 &vmw_bo_bo_free, &vbo); in vmw_dummy_query_bo_create() 405 ret = ttm_bo_reserve(&vbo->base, false, true, NULL); in vmw_dummy_query_bo_create() 407 vmw_bo_pin_reserved(vbo, true); in vmw_dummy_query_bo_create() 409 ret = ttm_bo_kmap(&vbo->base, 0, 1, &map); in vmw_dummy_query_bo_create() 417 vmw_bo_pin_reserved(vbo, false); in vmw_dummy_query_bo_create() 418 ttm_bo_unreserve(&vbo->base); in vmw_dummy_query_bo_create() 422 vmw_bo_unreference(&vbo); in vmw_dummy_query_bo_create() 424 dev_priv->dummy_query_bo = vbo; in vmw_dummy_query_bo_create()
|
A D | vmwgfx_execbuf.c | 68 struct vmw_buffer_object *vbo; member 1172 reloc->vbo = vmw_bo; in vmw_translate_mob_ptr() 1227 reloc->vbo = vmw_bo; in vmw_translate_guest_ptr() 1704 struct vmw_buffer_object *vbo; in vmw_cmd_res_switch_backup() local 1712 ret = vmw_translate_mob_ptr(dev_priv, sw_context, buf_id, &vbo); in vmw_cmd_res_switch_backup() 1716 vmw_validation_res_switch_backup(sw_context->ctx, info, vbo, in vmw_cmd_res_switch_backup() 3757 bo = &reloc->vbo->base; in vmw_apply_relocations()
|
/linux-6.3-rc2/fs/ntfs3/ |
A D | dir.c | 343 if (vbo + off < pos) in ntfs_read_hdr() 349 ctx->pos = vbo + off; in ntfs_read_hdr() 368 u64 vbo; in ntfs_readdir() local 436 vbo = (u64)bit << index_bits; in ntfs_readdir() 437 if (vbo >= i_size) { in ntfs_readdir() 451 vbo = (u64)bit << index_bits; in ntfs_readdir() 452 if (vbo >= i_size) { in ntfs_readdir() 494 u64 vbo = 0; in ntfs_dir_count() local 542 if (vbo >= i_size) in ntfs_dir_count() 552 vbo = (u64)bit << index_bits; in ntfs_dir_count() [all …]
|
A D | attrib.c | 1222 u64 vbo; in attr_data_read_resident() local 1258 u64 vbo; in attr_data_write_resident() local 1383 vbo[1] = frame << 2; in attr_wof_frame_info() 1433 if (vbo[1]) { in attr_wof_frame_info() 1435 vbo[0] = vbo[1] - bytes_per_off; in attr_wof_frame_info() 1439 vbo[0] = 0; in attr_wof_frame_info() 1883 if (vbo > data_size) in attr_collapse_range() 2067 valid_size = vbo; in attr_collapse_range() 2126 from = vbo; in attr_punch_hole() 2145 bytes += vbo; in attr_punch_hole() [all …]
|
A D | fsntfs.c | 773 u64 vbo; in ntfs_clear_mft_tail() local 785 vbo = (u64)from * rs; in ntfs_clear_mft_tail() 1192 lbo = vbo + sbi->mft.lbo; in ntfs_read_run_nb() 1319 off = vbo & sbi->cluster_mask; in ntfs_get_bh() 1498 vbo = vbo & ~511ull; in ntfs_bio_pages() 1499 bytes = lbo - vbo; in ntfs_bio_pages() 1501 vcn = vbo >> cluster_bits; in ntfs_bio_pages() 1506 off = vbo & sbi->cluster_mask; in ntfs_bio_pages() 1523 off = vbo & (PAGE_SIZE - 1); in ntfs_bio_pages() 1532 vbo += add; in ntfs_bio_pages() [all …]
|
A D | file.c | 182 pgoff_t idx = vbo >> PAGE_SHIFT; in ntfs_zero_range() 183 u32 from = vbo & (PAGE_SIZE - 1); in ntfs_zero_range() 441 loff_t end = vbo + len; in ntfs_fallocate() 504 err = attr_punch_hole(ni, vbo, len, &frame_size); in ntfs_fallocate() 511 vbo_a = (vbo + mask) & ~mask; in ntfs_fallocate() 515 if (tmp > vbo) { in ntfs_fallocate() 516 err = ntfs_zero_range(inode, vbo, tmp); in ntfs_fallocate() 521 if (vbo < end_a && end_a < end) { in ntfs_fallocate() 553 err = attr_collapse_range(ni, vbo, len); in ntfs_fallocate() 569 err = attr_insert_range(ni, vbo, len); in ntfs_fallocate() [all …]
|
A D | frecord.c | 942 u64 vbo; in ni_ins_attr_ext() local 1952 end = vbo + len; in ni_fiemap() 1959 while (vbo < end) { in ni_fiemap() 2026 if (vbo + bytes >= end) in ni_fiemap() 2027 bytes = end - vbo; in ni_fiemap() 2037 if (vbo + dlen >= end) in ni_fiemap() 2049 vbo = valid; in ni_fiemap() 2069 vbo += bytes; in ni_fiemap() 2180 u64 vbo; in ni_decompress_file() local 2211 for (vbo = 0; vbo < i_size; vbo += bytes) { in ni_decompress_file() [all …]
|
A D | fslog.c | 928 if (off == vbo) { in alloc_rsttbl_from_idx() 950 if (off == vbo) { in alloc_rsttbl_from_idx() 1049 return vbo; in lsn_to_vbo() 1111 u32 vbo; member 1182 u32 skip, vbo; in log_read_rst() local 1187 vbo = 0; in log_read_rst() 1190 vbo = 512; in log_read_rst() 1195 for (; vbo < l_size; vbo = 2 * vbo + skip, skip = 0) { in log_read_rst() 1230 info->vbo = vbo; in log_read_rst() 1458 if (end <= vbo) in next_log_lsn() [all …]
|
A D | index.c | 263 vbo = off & ~(size_t)sbi->block_mask; in bmp_buf_get() 265 bbuf->new_valid = vbo + blocksize; in bmp_buf_get() 271 if (vbo >= valid_size) { in bmp_buf_get() 273 } else if (vbo + blocksize > valid_size) { in bmp_buf_get() 375 size_t vbo = from >> 3; in scan_nres_bitmap() local 390 vcn = vbo >> sbi->cluster_bits; in scan_nres_bitmap() 423 if (vbo >= valid_size) { in scan_nres_bitmap() 431 if (vbo + blocksize > data_size) in scan_nres_bitmap() 432 nbits = 8 * (data_size - vbo); in scan_nres_bitmap() 439 *ret += 8 * vbo; in scan_nres_bitmap() [all …]
|
A D | bitmap.c | 507 size_t wpos, wbit, iw, vbo; in wnd_rescan() local 516 vbo = 0; in wnd_rescan() 527 vbo * 8 - prev_tail, in wnd_rescan() 542 u32 off = vbo & sbi->cluster_mask; in wnd_rescan() 544 if (!run_lookup_entry(&wnd->run, vbo >> cluster_bits, in wnd_rescan() 568 wbit = vbo * 8; in wnd_rescan() 612 vbo += blocksize; in wnd_rescan() 679 size_t vbo; in wnd_map() local 687 vbo = (u64)iw << sb->s_blocksize_bits; in wnd_map() 1371 u64 vbo, lbo, bytes; in wnd_extend() local [all …]
|
A D | record.c | 121 u64 vbo = (u64)mi->rno << sbi->record_bits; in mi_read() local 133 err = ntfs_read_bh(sbi, run, vbo, &rec->rhdr, bpr, &mi->nb); in mi_read() 152 vbo >> sbi->cluster_bits); in mi_read() 162 err = ntfs_read_bh(sbi, run, vbo, &rec->rhdr, bpr, &mi->nb); in mi_read() 358 u64 vbo = (u64)rno << sbi->record_bits; in mi_format_new() local 397 err = ntfs_get_bh(sbi, &ni->file.run, vbo, sbi->record_size, in mi_format_new()
|
A D | inode.c | 569 vcn = vbo >> cluster_bits; in ntfs_get_block_vbo() 570 off = vbo & sbi->cluster_mask; in ntfs_get_block_vbo() 605 if (vbo >= valid) in ntfs_get_block_vbo() 612 if (vbo >= valid) in ntfs_get_block_vbo() 615 if (vbo + bytes > valid) { in ntfs_get_block_vbo() 616 ni->i_valid = vbo + bytes; in ntfs_get_block_vbo() 619 } else if (vbo >= valid) { in ntfs_get_block_vbo() 634 u32 voff = valid - vbo; in ntfs_get_block_vbo() 762 loff_t vbo = iocb->ki_pos; in ntfs_direct_IO() local 780 end = vbo + ret; in ntfs_direct_IO() [all …]
|
A D | ntfs_fs.h | 448 int attr_collapse_range(struct ntfs_inode *ni, u64 vbo, u64 bytes); 449 int attr_insert_range(struct ntfs_inode *ni, u64 vbo, u64 bytes); 450 int attr_punch_hole(struct ntfs_inode *ni, u64 vbo, u64 bytes, u32 *frame_size); 560 __u64 vbo, __u64 len); 614 u64 vbo, const void *buf, size_t bytes, int sync); 616 const struct runs_tree *run, u64 vbo); 618 u64 vbo, void *buf, u32 bytes, struct ntfs_buffers *nb); 619 int ntfs_read_bh(struct ntfs_sb_info *sbi, const struct runs_tree *run, u64 vbo, 622 int ntfs_get_bh(struct ntfs_sb_info *sbi, const struct runs_tree *run, u64 vbo, 627 struct page **pages, u32 nr_pages, u64 vbo, u32 bytes, [all …]
|
/linux-6.3-rc2/drivers/gpu/drm/imx/ipuv3/ |
A D | ipuv3-plane.c | 377 unsigned long eba, ubo, vbo, old_ubo, old_vbo, alpha_eba; in ipu_plane_atomic_check() local 466 vbo = drm_plane_state_to_vbo(new_state); in ipu_plane_atomic_check() 468 if (vbo & 0x7 || vbo > 0xfffff8) in ipu_plane_atomic_check() 473 if (vbo != old_vbo) in ipu_plane_atomic_check() 586 unsigned long eba, ubo, vbo; in ipu_plane_atomic_update() local 710 vbo = drm_plane_state_to_vbo(new_state); in ipu_plane_atomic_update() 714 swap(ubo, vbo); in ipu_plane_atomic_update() 717 fb->pitches[1], ubo, vbo); in ipu_plane_atomic_update() 720 "phy = %lu %lu %lu, x = %d, y = %d", eba, ubo, vbo, in ipu_plane_atomic_update()
|
/linux-6.3-rc2/drivers/gpu/drm/vc4/ |
A D | vc4_validate.c | 899 struct drm_gem_dma_object *vbo = in validate_gl_shader_rec() local 908 to_vc4_bo(&vbo->base)->write_seqno); in validate_gl_shader_rec() 913 if (vbo->base.size < offset || in validate_gl_shader_rec() 914 vbo->base.size - offset < attr_size) { in validate_gl_shader_rec() 916 offset, attr_size, vbo->base.size); in validate_gl_shader_rec() 921 max_index = ((vbo->base.size - offset - attr_size) / in validate_gl_shader_rec() 931 *(uint32_t *)(pkt_v + o) = vbo->dma_addr + offset; in validate_gl_shader_rec()
|