| /drivers/s390/char/ |
| A D | sclp_rw.c | 52 buffer->sccb = sccb; in sclp_make_buffer() 59 buffer->htab = htab; in sclp_make_buffer() 65 return buffer; in sclp_make_buffer() 155 buffer->char_sum += buffer->current_length; in sclp_finalize_mto() 230 if (buffer->current_length >= buffer->columns) in sclp_write() 235 } while (buffer->current_length % buffer->htab); in sclp_write() 294 buffer->current_length >= buffer->columns) in sclp_write() 368 buffer->callback(buffer, -EIO); in sclp_writedata_callback() 416 buffer->callback(buffer, rc); in sclp_writedata_callback() 439 buffer->request.callback_data = buffer; in sclp_emit_buffer() [all …]
|
| /drivers/gpu/drm/etnaviv/ |
| A D | etnaviv_buffer.c | 29 BUG_ON(buffer->user_size >= buffer->size); in OUT() 40 buffer->user_size = ALIGN(buffer->user_size, 8); in CMD_LOAD_STATE() 51 buffer->user_size = ALIGN(buffer->user_size, 8); in CMD_END() 59 buffer->user_size = ALIGN(buffer->user_size, 8); in CMD_WAIT() 67 buffer->user_size = ALIGN(buffer->user_size, 8); in CMD_LINK() 77 buffer->user_size = ALIGN(buffer->user_size, 8); in CMD_STALL() 165 struct etnaviv_cmdbuf *buffer = &gpu->buffer; in etnaviv_buffer_init() local 182 struct etnaviv_cmdbuf *buffer = &gpu->buffer; in etnaviv_buffer_config_mmuv2() local 217 struct etnaviv_cmdbuf *buffer = &gpu->buffer; in etnaviv_buffer_config_pta() local 235 struct etnaviv_cmdbuf *buffer = &gpu->buffer; in etnaviv_buffer_end() local [all …]
|
| /drivers/media/tuners/ |
| A D | tea5767.c | 147 div = ((buffer[0] & 0x3f) << 8) | buffer[1]; in tea5767_status_dump() 165 buffer[1] = div & 0xff; in tea5767_status_dump() 198 buffer[2] = 0; in set_radio_freq() 211 buffer[3] = 0; in set_radio_freq() 228 buffer[4] = 0; in set_radio_freq() 267 buffer[1] = div & 0xff; in set_radio_freq() 289 memset(buffer, 0, 5); in tea5767_read_status() 360 buffer[4] = 0; in tea5767_standby() 381 if (buffer[0] == buffer[1] && buffer[0] == buffer[2] && in tea5767_autodetection() 382 buffer[0] == buffer[3] && buffer[0] == buffer[4]) { in tea5767_autodetection() [all …]
|
| A D | tea5761.c | 127 div = ((buffer[2] & 0x3f) << 8) | buffer[3]; in tea5761_status_dump() 164 buffer[1] = (div >> 8) & 0x3f; in __set_radio_freq() 165 buffer[2] = div & 0xff; in __set_radio_freq() 168 tea5761_status_dump(buffer); in __set_radio_freq() 203 memset(buffer, 0, 16); in tea5761_read_status() 236 unsigned char buffer[16]; in tea5761_get_status() local 252 unsigned char buffer[16]; in tea5761_get_rf_strength() local 264 unsigned char buffer[16]; in tea5761_autodetection() local 273 if ((buffer[13] != 0x2b) || (buffer[14] != 0x57) || (buffer[15] != 0x061)) { in tea5761_autodetection() 275 buffer[13], buffer[14], buffer[15]); in tea5761_autodetection() [all …]
|
| /drivers/pnp/ |
| A D | interface.c | 42 if (buffer->stop || buffer->error) in pnp_printf() 45 res = vsnprintf(buffer->curr, buffer->len - buffer->size, fmt, args); in pnp_printf() 47 if (buffer->size + res >= buffer->len) { in pnp_printf() 217 buffer = kzalloc(sizeof(*buffer), GFP_KERNEL); in options_show() 218 if (!buffer) in options_show() 222 buffer->buffer = buf; in options_show() 223 buffer->curr = buffer->buffer; in options_show() 260 buffer = kzalloc(sizeof(*buffer), GFP_KERNEL); in resources_show() 261 if (!buffer) in resources_show() 265 buffer->buffer = buf; in resources_show() [all …]
|
| /drivers/gpu/drm/ |
| A D | drm_client.c | 178 drm_gem_vunmap(buffer->gem, &buffer->map); in drm_client_buffer_delete() 182 kfree(buffer); in drm_client_buffer_delete() 196 buffer = kzalloc(sizeof(*buffer), GFP_KERNEL); in drm_client_buffer_create() 197 if (!buffer) in drm_client_buffer_create() 312 ret = drm_gem_vmap(buffer->gem, &buffer->map); in drm_client_buffer_vmap() 331 drm_gem_vunmap(buffer->gem, &buffer->map); in drm_client_buffer_vunmap() 342 ret = drm_mode_rmfb(buffer->client->dev, buffer->fb->base.id, buffer->client->file); in drm_client_buffer_rmfb() 431 if (!buffer) in drm_client_framebuffer_delete() 452 if (!buffer || !buffer->fb || !buffer->fb->funcs->dirty) in drm_client_framebuffer_flush() 463 return buffer->fb->funcs->dirty(buffer->fb, buffer->client->file, in drm_client_framebuffer_flush() [all …]
|
| /drivers/dma-buf/heaps/ |
| A D | cma_heap.c | 134 invalidate_kernel_vmap_range(buffer->vaddr, buffer->len); in cma_heap_dma_buf_begin_cpu_access() 155 flush_kernel_vmap_range(buffer->vaddr, buffer->len); in cma_heap_dma_buf_end_cpu_access() 201 vaddr = vmap(buffer->pages, buffer->pagecount, VM_MAP, PAGE_KERNEL); in cma_heap_do_vmap() 216 buffer->vmap_cnt++; in cma_heap_vmap() 227 buffer->vmap_cnt++; in cma_heap_vmap() 262 cma_release(cma_heap->cma, buffer->cma_pages, buffer->pagecount); in cma_heap_dma_buf_release() 263 kfree(buffer); in cma_heap_dma_buf_release() 295 buffer = kzalloc(sizeof(*buffer), GFP_KERNEL); in cma_heap_allocate() 296 if (!buffer) in cma_heap_allocate() 301 buffer->len = size; in cma_heap_allocate() [all …]
|
| A D | system_heap.c | 150 if (buffer->vmap_cnt) in system_heap_dma_buf_begin_cpu_access() 151 invalidate_kernel_vmap_range(buffer->vaddr, buffer->len); in system_heap_dma_buf_begin_cpu_access() 171 if (buffer->vmap_cnt) in system_heap_dma_buf_end_cpu_access() 172 flush_kernel_vmap_range(buffer->vaddr, buffer->len); in system_heap_dma_buf_end_cpu_access() 240 buffer->vmap_cnt++; in system_heap_vmap() 252 buffer->vmap_cnt++; in system_heap_vmap() 287 kfree(buffer); in system_heap_dma_buf_release() 339 buffer = kzalloc(sizeof(*buffer), GFP_KERNEL); in system_heap_allocate() 340 if (!buffer) in system_heap_allocate() 346 buffer->len = len; in system_heap_allocate() [all …]
|
| /drivers/iio/ |
| A D | industrialio-buffer.c | 358 if (!buffer || !buffer->access || !buffer->access->remove_from) in iio_pop_from_buffer() 720 buffer->access->set_length(buffer, val); in length_store() 722 if (buffer->length && buffer->length < buffer->watermark) in length_store() 723 buffer->watermark = buffer->length; in length_store() 1485 iio_attr->buffer = buffer; in iio_buffer_wrap_attr() 1703 priv->buffer = buffer; in iio_buffer_attach_dmabuf() 1893 buffer->access->lock_queue(buffer); in iio_buffer_enqueue_dmabuf() 2036 ib->buffer = buffer; in iio_device_buffer_getfd() 2452 buffer->access->release(buffer); in iio_buffer_release() 2508 buffer = iio_buffer_get(buffer); in iio_device_attach_buffer() [all …]
|
| /drivers/s390/net/ |
| A D | qeth_core_mpc.h | 18 #define QETH_IPA_PDU_LEN_TOTAL(buffer) (buffer + 0x0e) argument 19 #define QETH_IPA_PDU_LEN_PDU1(buffer) (buffer + 0x26) argument 20 #define QETH_IPA_PDU_LEN_PDU2(buffer) (buffer + 0x29) argument 21 #define QETH_IPA_PDU_LEN_PDU3(buffer) (buffer + 0x3a) argument 23 #define QETH_IPA_CMD_DEST_ADDR(buffer) (buffer + 0x2c) argument 919 #define QETH_ULP_SETUP_CUA(buffer) (buffer + 0x68) argument 940 #define QETH_IDX_ACT_PNO(buffer) (buffer+0x0b) argument 949 #define QETH_IDX_REPLY_LEVEL(buffer) (buffer + 0x12) argument 961 (buffer + *(buffer + (*(buffer + 0x0b)) + \ 962 *(buffer + *(buffer + 0x0b) + 0x11) + 0x07)) [all …]
|
| /drivers/android/ |
| A D | binder_alloc.c | 66 return binder_buffer_next(buffer)->user_data - buffer->user_data; in binder_alloc_buffer_size() 152 return buffer; in binder_alloc_prepare_to_free_locked() 607 return buffer; in binder_alloc_new_buf_locked() 698 return buffer; in binder_alloc_new_buf() 736 kfree(buffer); in binder_delete_free_buffer() 786 buffer = prev; in binder_free_buf_locked() 928 buffer = kzalloc(sizeof(*buffer), GFP_KERNEL); in binder_alloc_mmap_handler() 929 if (!buffer) { in binder_alloc_mmap_handler() 995 kfree(buffer); in binder_alloc_deferred_release() 1053 buffer->data_size, buffer->offsets_size, in binder_alloc_print_allocated() [all …]
|
| /drivers/net/ethernet/hisilicon/hibmcge/ |
| A D | hbg_txrx.c | 13 #define buffer_to_dma_dir(buffer) (((buffer)->dir == HBG_DIR_RX) ? \ argument 39 buffer->skb->data, buffer->skb_len, in hbg_dma_map() 60 dma_unmap_single(&priv->pdev->dev, buffer->skb_dma, buffer->skb_len, in hbg_dma_unmap() 62 buffer->skb_dma = 0; in hbg_dma_unmap() 68 u32 ip_offset = buffer->skb->network_header - buffer->skb->mac_header; in hbg_init_tx_desc() 110 buffer->skb = skb; in hbg_net_start_xmit() 135 buffer->skb = NULL; in hbg_buffer_free_skb() 140 u32 len = hbg_spec_max_frame_len(buffer->priv, buffer->dir); in hbg_buffer_alloc_skb() 441 buffer->skb->protocol = eth_type_trans(buffer->skb, in hbg_napi_rx_poll() 505 buffer->dir = dir; in hbg_ring_init() [all …]
|
| /drivers/scsi/ |
| A D | sr_vendor.c | 151 cgc.buffer = buffer; in sr_set_blocklength() 198 cgc.buffer = buffer; in sr_cd_check() 206 if ((buffer[0] << 8) + buffer[1] < 0x0a) { in sr_cd_check() 212 sector = buffer[11] + (buffer[10] << 8) + in sr_cd_check() 213 (buffer[9] << 16) + (buffer[8] << 24); in sr_cd_check() 225 cgc.buffer = buffer; in sr_cd_check() 254 cgc.buffer = buffer; in sr_cd_check() 282 cgc.buffer = buffer; in sr_cd_check() 300 cgc.buffer = buffer; in sr_cd_check() 309 sector = buffer[11] + (buffer[10] << 8) + in sr_cd_check() [all …]
|
| /drivers/gpu/drm/nouveau/nvkm/subdev/fault/ |
| A D | tu102.c | 36 struct nvkm_fault_buffer *buffer = container_of(inth, typeof(*buffer), inth); in tu102_fault_buffer_notify() local 38 nvkm_event_ntfy(&buffer->fault->event, buffer->id, NVKM_FAULT_BUFFER_EVENT_PENDING); in tu102_fault_buffer_notify() 46 nvkm_inth_allow(&buffer->inth); in tu102_fault_buffer_intr() 48 nvkm_inth_block(&buffer->inth); in tu102_fault_buffer_intr() 81 buffer->get = 0xb83008 + foff; in tu102_fault_buffer_info() 82 buffer->put = 0xb8300c + foff; in tu102_fault_buffer_info() 121 if (fault->buffer[0]) in tu102_fault_fini() 122 fault->func->buffer.fini(fault->buffer[0]); in tu102_fault_fini() 132 fault->func->buffer.init(fault->buffer[0]); in tu102_fault_init() 165 .buffer.nr = 2, [all …]
|
| A D | gv100.c | 34 struct nvkm_fault_buffer *buffer = fault->buffer[0]; in gv100_fault_buffer_process() local 44 const u32 base = get * buffer->fault->func->buffer.entry_size; in gv100_fault_buffer_process() 55 if (++get == buffer->entries) in gv100_fault_buffer_process() 115 buffer->get = 0x100e2c + foff; in gv100_fault_buffer_info() 168 if (fault->buffer[0]) { in gv100_fault_intr() 175 if (fault->buffer[1]) { in gv100_fault_intr() 192 if (fault->buffer[0]) in gv100_fault_fini() 193 fault->func->buffer.fini(fault->buffer[0]); in gv100_fault_fini() 202 fault->func->buffer.init(fault->buffer[0]); in gv100_fault_init() 220 .buffer.nr = 2, [all …]
|
| A D | gp100.c | 54 return nvkm_memory_bar2(buffer->mem); in gp100_fault_buffer_pin() 60 buffer->entries = nvkm_rd32(buffer->fault->subdev.device, 0x002a78); in gp100_fault_buffer_info() 61 buffer->get = 0x002a7c; in gp100_fault_buffer_info() 62 buffer->put = 0x002a80; in gp100_fault_buffer_info() 74 .buffer.nr = 1, 75 .buffer.entry_size = 32, 76 .buffer.info = gp100_fault_buffer_info, 77 .buffer.pin = gp100_fault_buffer_pin, 78 .buffer.init = gp100_fault_buffer_init, 79 .buffer.fini = gp100_fault_buffer_fini, [all …]
|
| A D | user.c | 42 return nvkm_uevent_add(uevent, &buffer->fault->event, buffer->id, in nvkm_ufault_uevent() 51 struct nvkm_device *device = buffer->fault->subdev.device; in nvkm_ufault_map() 54 *size = nvkm_memory_size(buffer->mem); in nvkm_ufault_map() 62 buffer->fault->func->buffer.fini(buffer); in nvkm_ufault_fini() 70 buffer->fault->func->buffer.init(buffer); in nvkm_ufault_init() 97 struct nvkm_fault_buffer *buffer = fault->buffer[fault->func->user.rp]; in nvkm_ufault_new() local 101 args->v0.entries = buffer->entries; in nvkm_ufault_new() 102 args->v0.get = buffer->get; in nvkm_ufault_new() 103 args->v0.put = buffer->put; in nvkm_ufault_new() 107 nvkm_object_ctor(&nvkm_ufault, oclass, &buffer->object); in nvkm_ufault_new() [all …]
|
| /drivers/iio/buffer/ |
| A D | industrialio-buffer-dmaengine.c | 45 return container_of(buffer, struct dmaengine_buffer, queue.buffer); in iio_buffer_to_dmaengine_buffer() 192 struct iio_buffer *buffer = to_iio_dev_attr(attr)->buffer; in iio_dmaengine_buffer_get_length_align() local 269 iio_buffer_put(buffer); in iio_dmaengine_buffer_free() 299 if (IS_ERR(buffer)) in __iio_dmaengine_buffer_setup_ext() 312 return buffer; in __iio_dmaengine_buffer_setup_ext() 343 if (IS_ERR(buffer)) in iio_dmaengine_buffer_setup_ext() 346 return buffer; in iio_dmaengine_buffer_setup_ext() 375 if (IS_ERR(buffer)) in devm_iio_dmaengine_buffer_setup_ext() 379 buffer); in devm_iio_dmaengine_buffer_setup_ext() 412 if (IS_ERR(buffer)) in devm_iio_dmaengine_buffer_setup_with_handle() [all …]
|
| /drivers/usb/misc/ |
| A D | cytherm.c | 85 if (!buffer) in brightness_store() 106 kfree(buffer); in brightness_store() 128 if (!buffer) in temp_show() 143 kfree(buffer); in temp_show() 163 if (!buffer) in button_show() 173 kfree(buffer); in button_show() 192 if (!buffer) in port0_show() 201 kfree(buffer); in port0_show() 217 if (!buffer) in port0_store() 247 if (!buffer) in port1_show() [all …]
|
| /drivers/net/ethernet/sfc/siena/ |
| A D | tx_common.c | 39 tx_queue->buffer = kcalloc(entries, sizeof(*tx_queue->buffer), in efx_siena_probe_tx_queue() 41 if (!tx_queue->buffer) in efx_siena_probe_tx_queue() 102 if (!tx_queue->buffer) in efx_siena_remove_tx_queue() 129 dma_addr_t unmap_addr = buffer->dma_addr - buffer->dma_offset; in efx_dequeue_buffer() 166 buffer->len = 0; in efx_dequeue_buffer() 167 buffer->flags = 0; in efx_dequeue_buffer() 184 buffer = &tx_queue->buffer[tx_queue->read_count & tx_queue->ptr_mask]; in efx_siena_fini_tx_queue() 210 struct efx_tx_buffer *buffer = &tx_queue->buffer[read_ptr]; in efx_dequeue_buffers() local 312 return buffer; in efx_siena_tx_map_chunk() 378 buffer->dma_offset = buffer->dma_addr - unmap_addr; in efx_siena_tx_map_data() [all …]
|
| /drivers/net/ethernet/sfc/ |
| A D | tx_common.c | 39 tx_queue->buffer = kcalloc(entries, sizeof(*tx_queue->buffer), in efx_probe_tx_queue() 41 if (!tx_queue->buffer) in efx_probe_tx_queue() 112 if (!tx_queue->buffer) in efx_fini_tx_queue() 121 buffer = &tx_queue->buffer[tx_queue->read_count & tx_queue->ptr_mask]; in efx_fini_tx_queue() 166 dma_addr_t unmap_addr = buffer->dma_addr - buffer->dma_offset; in efx_dequeue_buffer() 213 buffer->len = 0; in efx_dequeue_buffer() 214 buffer->flags = 0; in efx_dequeue_buffer() 237 struct efx_tx_buffer *buffer = &tx_queue->buffer[read_ptr]; in efx_dequeue_buffers() local 352 return buffer; in efx_tx_map_chunk() 418 buffer->dma_offset = buffer->dma_addr - unmap_addr; in efx_tx_map_data() [all …]
|
| /drivers/md/dm-vdo/ |
| A D | encodings.c | 258 decode_u32_le(buffer, offset, &unused); in decode_volume_geometry() 261 decode_u64_le(buffer, offset, &nonce); in decode_volume_geometry() 284 decode_u32_le(buffer, offset, &mem); in decode_volume_geometry() 286 sparse = buffer[(*offset)++]; in decode_volume_geometry() 339 struct block_map_page *page = buffer; in vdo_format_block_map_page() 341 memset(buffer, 0, VDO_BLOCK_SIZE); in vdo_format_block_map_page() 594 decode_u64_le(buffer, offset, &count); in decode_slab_depot_state_2_0() 617 zone_count = buffer[(*offset)++]; in decode_slab_depot_state_2_0() 984 buffer[(*offset)++] = partition->id; in encode_layout() 987 encode_u64_le(buffer, offset, 0); in encode_layout() [all …]
|
| /drivers/acpi/acpica/ |
| A D | exstorob.c | 34 u8 *buffer; in acpi_ex_store_buffer_to_buffer() local 46 buffer = ACPI_CAST_PTR(u8, source_desc->buffer.pointer); in acpi_ex_store_buffer_to_buffer() 47 length = source_desc->buffer.length; in acpi_ex_store_buffer_to_buffer() 56 if (!target_desc->buffer.pointer) { in acpi_ex_store_buffer_to_buffer() 60 target_desc->buffer.length = length; in acpi_ex_store_buffer_to_buffer() 70 target_desc->buffer.length); in acpi_ex_store_buffer_to_buffer() 71 memcpy(target_desc->buffer.pointer, buffer, length); in acpi_ex_store_buffer_to_buffer() 98 memcpy(target_desc->buffer.pointer, buffer, in acpi_ex_store_buffer_to_buffer() 99 target_desc->buffer.length); in acpi_ex_store_buffer_to_buffer() 108 target_desc->buffer.flags = source_desc->buffer.flags; in acpi_ex_store_buffer_to_buffer() [all …]
|
| /drivers/gpu/drm/nouveau/ |
| A D | nouveau_svm.c | 70 } buffer[]; member 477 if (!buffer->fault[buffer->fault_nr]) { in nouveau_svm_fault_cache() 483 buffer->fault[buffer->fault_nr] = fault; in nouveau_svm_fault_cache() 486 fault = buffer->fault[buffer->fault_nr++]; in nouveau_svm_fault_cache() 720 struct nouveau_svm *svm = container_of(buffer, typeof(*svm), buffer[buffer->id]); in nouveau_svm_fault() 733 if (buffer->get == buffer->put) { in nouveau_svm_fault() 736 if (buffer->get == buffer->put) in nouveau_svm_fault() 742 while (buffer->get != buffer->put) { in nouveau_svm_fault() 744 if (++buffer->get == buffer->entries) in nouveau_svm_fault() 754 sort(buffer->fault, buffer->fault_nr, sizeof(*buffer->fault), in nouveau_svm_fault() [all …]
|
| /drivers/net/ethernet/sfc/falcon/ |
| A D | tx.c | 39 buffer->unmap_len = 0; in ef4_tx_get_copy_buffer() 50 dma_addr_t unmap_addr = buffer->dma_addr - buffer->dma_offset; in ef4_dequeue_buffer() 57 buffer->unmap_len = 0; in ef4_dequeue_buffer() 69 buffer->len = 0; in ef4_dequeue_buffer() 70 buffer->flags = 0; in ef4_dequeue_buffer() 166 buffer->skb = skb; in ef4_enqueue_skb_copy() 194 return buffer; in ef4_tx_map_chunk() 233 buffer->dma_offset = buffer->dma_addr - unmap_addr; in ef4_tx_map_data() 239 buffer->skb = skb; in ef4_tx_map_data() 359 struct ef4_tx_buffer *buffer = &tx_queue->buffer[read_ptr]; in ef4_dequeue_buffers() local [all …]
|