Home
last modified time | relevance | path

Searched refs:buffer (Results 1 – 25 of 1584) sorted by relevance

12345678910>>...64

/drivers/s390/char/
A Dsclp_rw.c52 buffer->sccb = sccb; in sclp_make_buffer()
59 buffer->htab = htab; in sclp_make_buffer()
65 return buffer; in sclp_make_buffer()
155 buffer->char_sum += buffer->current_length; in sclp_finalize_mto()
230 if (buffer->current_length >= buffer->columns) in sclp_write()
235 } while (buffer->current_length % buffer->htab); in sclp_write()
294 buffer->current_length >= buffer->columns) in sclp_write()
368 buffer->callback(buffer, -EIO); in sclp_writedata_callback()
416 buffer->callback(buffer, rc); in sclp_writedata_callback()
439 buffer->request.callback_data = buffer; in sclp_emit_buffer()
[all …]
/drivers/gpu/drm/etnaviv/
A Detnaviv_buffer.c29 BUG_ON(buffer->user_size >= buffer->size); in OUT()
40 buffer->user_size = ALIGN(buffer->user_size, 8); in CMD_LOAD_STATE()
51 buffer->user_size = ALIGN(buffer->user_size, 8); in CMD_END()
59 buffer->user_size = ALIGN(buffer->user_size, 8); in CMD_WAIT()
67 buffer->user_size = ALIGN(buffer->user_size, 8); in CMD_LINK()
77 buffer->user_size = ALIGN(buffer->user_size, 8); in CMD_STALL()
165 struct etnaviv_cmdbuf *buffer = &gpu->buffer; in etnaviv_buffer_init() local
182 struct etnaviv_cmdbuf *buffer = &gpu->buffer; in etnaviv_buffer_config_mmuv2() local
217 struct etnaviv_cmdbuf *buffer = &gpu->buffer; in etnaviv_buffer_config_pta() local
235 struct etnaviv_cmdbuf *buffer = &gpu->buffer; in etnaviv_buffer_end() local
[all …]
/drivers/media/tuners/
A Dtea5767.c147 div = ((buffer[0] & 0x3f) << 8) | buffer[1]; in tea5767_status_dump()
165 buffer[1] = div & 0xff; in tea5767_status_dump()
198 buffer[2] = 0; in set_radio_freq()
211 buffer[3] = 0; in set_radio_freq()
228 buffer[4] = 0; in set_radio_freq()
267 buffer[1] = div & 0xff; in set_radio_freq()
289 memset(buffer, 0, 5); in tea5767_read_status()
360 buffer[4] = 0; in tea5767_standby()
381 if (buffer[0] == buffer[1] && buffer[0] == buffer[2] && in tea5767_autodetection()
382 buffer[0] == buffer[3] && buffer[0] == buffer[4]) { in tea5767_autodetection()
[all …]
A Dtea5761.c127 div = ((buffer[2] & 0x3f) << 8) | buffer[3]; in tea5761_status_dump()
164 buffer[1] = (div >> 8) & 0x3f; in __set_radio_freq()
165 buffer[2] = div & 0xff; in __set_radio_freq()
168 tea5761_status_dump(buffer); in __set_radio_freq()
203 memset(buffer, 0, 16); in tea5761_read_status()
236 unsigned char buffer[16]; in tea5761_get_status() local
252 unsigned char buffer[16]; in tea5761_get_rf_strength() local
264 unsigned char buffer[16]; in tea5761_autodetection() local
273 if ((buffer[13] != 0x2b) || (buffer[14] != 0x57) || (buffer[15] != 0x061)) { in tea5761_autodetection()
275 buffer[13], buffer[14], buffer[15]); in tea5761_autodetection()
[all …]
/drivers/pnp/
A Dinterface.c42 if (buffer->stop || buffer->error) in pnp_printf()
45 res = vsnprintf(buffer->curr, buffer->len - buffer->size, fmt, args); in pnp_printf()
47 if (buffer->size + res >= buffer->len) { in pnp_printf()
217 buffer = kzalloc(sizeof(*buffer), GFP_KERNEL); in options_show()
218 if (!buffer) in options_show()
222 buffer->buffer = buf; in options_show()
223 buffer->curr = buffer->buffer; in options_show()
260 buffer = kzalloc(sizeof(*buffer), GFP_KERNEL); in resources_show()
261 if (!buffer) in resources_show()
265 buffer->buffer = buf; in resources_show()
[all …]
/drivers/gpu/drm/
A Ddrm_client.c178 drm_gem_vunmap(buffer->gem, &buffer->map); in drm_client_buffer_delete()
182 kfree(buffer); in drm_client_buffer_delete()
196 buffer = kzalloc(sizeof(*buffer), GFP_KERNEL); in drm_client_buffer_create()
197 if (!buffer) in drm_client_buffer_create()
312 ret = drm_gem_vmap(buffer->gem, &buffer->map); in drm_client_buffer_vmap()
331 drm_gem_vunmap(buffer->gem, &buffer->map); in drm_client_buffer_vunmap()
342 ret = drm_mode_rmfb(buffer->client->dev, buffer->fb->base.id, buffer->client->file); in drm_client_buffer_rmfb()
431 if (!buffer) in drm_client_framebuffer_delete()
452 if (!buffer || !buffer->fb || !buffer->fb->funcs->dirty) in drm_client_framebuffer_flush()
463 return buffer->fb->funcs->dirty(buffer->fb, buffer->client->file, in drm_client_framebuffer_flush()
[all …]
/drivers/dma-buf/heaps/
A Dcma_heap.c134 invalidate_kernel_vmap_range(buffer->vaddr, buffer->len); in cma_heap_dma_buf_begin_cpu_access()
155 flush_kernel_vmap_range(buffer->vaddr, buffer->len); in cma_heap_dma_buf_end_cpu_access()
201 vaddr = vmap(buffer->pages, buffer->pagecount, VM_MAP, PAGE_KERNEL); in cma_heap_do_vmap()
216 buffer->vmap_cnt++; in cma_heap_vmap()
227 buffer->vmap_cnt++; in cma_heap_vmap()
262 cma_release(cma_heap->cma, buffer->cma_pages, buffer->pagecount); in cma_heap_dma_buf_release()
263 kfree(buffer); in cma_heap_dma_buf_release()
295 buffer = kzalloc(sizeof(*buffer), GFP_KERNEL); in cma_heap_allocate()
296 if (!buffer) in cma_heap_allocate()
301 buffer->len = size; in cma_heap_allocate()
[all …]
A Dsystem_heap.c150 if (buffer->vmap_cnt) in system_heap_dma_buf_begin_cpu_access()
151 invalidate_kernel_vmap_range(buffer->vaddr, buffer->len); in system_heap_dma_buf_begin_cpu_access()
171 if (buffer->vmap_cnt) in system_heap_dma_buf_end_cpu_access()
172 flush_kernel_vmap_range(buffer->vaddr, buffer->len); in system_heap_dma_buf_end_cpu_access()
240 buffer->vmap_cnt++; in system_heap_vmap()
252 buffer->vmap_cnt++; in system_heap_vmap()
287 kfree(buffer); in system_heap_dma_buf_release()
339 buffer = kzalloc(sizeof(*buffer), GFP_KERNEL); in system_heap_allocate()
340 if (!buffer) in system_heap_allocate()
346 buffer->len = len; in system_heap_allocate()
[all …]
/drivers/iio/
A Dindustrialio-buffer.c358 if (!buffer || !buffer->access || !buffer->access->remove_from) in iio_pop_from_buffer()
720 buffer->access->set_length(buffer, val); in length_store()
722 if (buffer->length && buffer->length < buffer->watermark) in length_store()
723 buffer->watermark = buffer->length; in length_store()
1485 iio_attr->buffer = buffer; in iio_buffer_wrap_attr()
1703 priv->buffer = buffer; in iio_buffer_attach_dmabuf()
1893 buffer->access->lock_queue(buffer); in iio_buffer_enqueue_dmabuf()
2036 ib->buffer = buffer; in iio_device_buffer_getfd()
2452 buffer->access->release(buffer); in iio_buffer_release()
2508 buffer = iio_buffer_get(buffer); in iio_device_attach_buffer()
[all …]
/drivers/s390/net/
A Dqeth_core_mpc.h18 #define QETH_IPA_PDU_LEN_TOTAL(buffer) (buffer + 0x0e) argument
19 #define QETH_IPA_PDU_LEN_PDU1(buffer) (buffer + 0x26) argument
20 #define QETH_IPA_PDU_LEN_PDU2(buffer) (buffer + 0x29) argument
21 #define QETH_IPA_PDU_LEN_PDU3(buffer) (buffer + 0x3a) argument
23 #define QETH_IPA_CMD_DEST_ADDR(buffer) (buffer + 0x2c) argument
919 #define QETH_ULP_SETUP_CUA(buffer) (buffer + 0x68) argument
940 #define QETH_IDX_ACT_PNO(buffer) (buffer+0x0b) argument
949 #define QETH_IDX_REPLY_LEVEL(buffer) (buffer + 0x12) argument
961 (buffer + *(buffer + (*(buffer + 0x0b)) + \
962 *(buffer + *(buffer + 0x0b) + 0x11) + 0x07))
[all …]
/drivers/android/
A Dbinder_alloc.c66 return binder_buffer_next(buffer)->user_data - buffer->user_data; in binder_alloc_buffer_size()
152 return buffer; in binder_alloc_prepare_to_free_locked()
607 return buffer; in binder_alloc_new_buf_locked()
698 return buffer; in binder_alloc_new_buf()
736 kfree(buffer); in binder_delete_free_buffer()
786 buffer = prev; in binder_free_buf_locked()
928 buffer = kzalloc(sizeof(*buffer), GFP_KERNEL); in binder_alloc_mmap_handler()
929 if (!buffer) { in binder_alloc_mmap_handler()
995 kfree(buffer); in binder_alloc_deferred_release()
1053 buffer->data_size, buffer->offsets_size, in binder_alloc_print_allocated()
[all …]
/drivers/net/ethernet/hisilicon/hibmcge/
A Dhbg_txrx.c13 #define buffer_to_dma_dir(buffer) (((buffer)->dir == HBG_DIR_RX) ? \ argument
39 buffer->skb->data, buffer->skb_len, in hbg_dma_map()
60 dma_unmap_single(&priv->pdev->dev, buffer->skb_dma, buffer->skb_len, in hbg_dma_unmap()
62 buffer->skb_dma = 0; in hbg_dma_unmap()
68 u32 ip_offset = buffer->skb->network_header - buffer->skb->mac_header; in hbg_init_tx_desc()
110 buffer->skb = skb; in hbg_net_start_xmit()
135 buffer->skb = NULL; in hbg_buffer_free_skb()
140 u32 len = hbg_spec_max_frame_len(buffer->priv, buffer->dir); in hbg_buffer_alloc_skb()
441 buffer->skb->protocol = eth_type_trans(buffer->skb, in hbg_napi_rx_poll()
505 buffer->dir = dir; in hbg_ring_init()
[all …]
/drivers/scsi/
A Dsr_vendor.c151 cgc.buffer = buffer; in sr_set_blocklength()
198 cgc.buffer = buffer; in sr_cd_check()
206 if ((buffer[0] << 8) + buffer[1] < 0x0a) { in sr_cd_check()
212 sector = buffer[11] + (buffer[10] << 8) + in sr_cd_check()
213 (buffer[9] << 16) + (buffer[8] << 24); in sr_cd_check()
225 cgc.buffer = buffer; in sr_cd_check()
254 cgc.buffer = buffer; in sr_cd_check()
282 cgc.buffer = buffer; in sr_cd_check()
300 cgc.buffer = buffer; in sr_cd_check()
309 sector = buffer[11] + (buffer[10] << 8) + in sr_cd_check()
[all …]
/drivers/gpu/drm/nouveau/nvkm/subdev/fault/
A Dtu102.c36 struct nvkm_fault_buffer *buffer = container_of(inth, typeof(*buffer), inth); in tu102_fault_buffer_notify() local
38 nvkm_event_ntfy(&buffer->fault->event, buffer->id, NVKM_FAULT_BUFFER_EVENT_PENDING); in tu102_fault_buffer_notify()
46 nvkm_inth_allow(&buffer->inth); in tu102_fault_buffer_intr()
48 nvkm_inth_block(&buffer->inth); in tu102_fault_buffer_intr()
81 buffer->get = 0xb83008 + foff; in tu102_fault_buffer_info()
82 buffer->put = 0xb8300c + foff; in tu102_fault_buffer_info()
121 if (fault->buffer[0]) in tu102_fault_fini()
122 fault->func->buffer.fini(fault->buffer[0]); in tu102_fault_fini()
132 fault->func->buffer.init(fault->buffer[0]); in tu102_fault_init()
165 .buffer.nr = 2,
[all …]
A Dgv100.c34 struct nvkm_fault_buffer *buffer = fault->buffer[0]; in gv100_fault_buffer_process() local
44 const u32 base = get * buffer->fault->func->buffer.entry_size; in gv100_fault_buffer_process()
55 if (++get == buffer->entries) in gv100_fault_buffer_process()
115 buffer->get = 0x100e2c + foff; in gv100_fault_buffer_info()
168 if (fault->buffer[0]) { in gv100_fault_intr()
175 if (fault->buffer[1]) { in gv100_fault_intr()
192 if (fault->buffer[0]) in gv100_fault_fini()
193 fault->func->buffer.fini(fault->buffer[0]); in gv100_fault_fini()
202 fault->func->buffer.init(fault->buffer[0]); in gv100_fault_init()
220 .buffer.nr = 2,
[all …]
A Dgp100.c54 return nvkm_memory_bar2(buffer->mem); in gp100_fault_buffer_pin()
60 buffer->entries = nvkm_rd32(buffer->fault->subdev.device, 0x002a78); in gp100_fault_buffer_info()
61 buffer->get = 0x002a7c; in gp100_fault_buffer_info()
62 buffer->put = 0x002a80; in gp100_fault_buffer_info()
74 .buffer.nr = 1,
75 .buffer.entry_size = 32,
76 .buffer.info = gp100_fault_buffer_info,
77 .buffer.pin = gp100_fault_buffer_pin,
78 .buffer.init = gp100_fault_buffer_init,
79 .buffer.fini = gp100_fault_buffer_fini,
[all …]
A Duser.c42 return nvkm_uevent_add(uevent, &buffer->fault->event, buffer->id, in nvkm_ufault_uevent()
51 struct nvkm_device *device = buffer->fault->subdev.device; in nvkm_ufault_map()
54 *size = nvkm_memory_size(buffer->mem); in nvkm_ufault_map()
62 buffer->fault->func->buffer.fini(buffer); in nvkm_ufault_fini()
70 buffer->fault->func->buffer.init(buffer); in nvkm_ufault_init()
97 struct nvkm_fault_buffer *buffer = fault->buffer[fault->func->user.rp]; in nvkm_ufault_new() local
101 args->v0.entries = buffer->entries; in nvkm_ufault_new()
102 args->v0.get = buffer->get; in nvkm_ufault_new()
103 args->v0.put = buffer->put; in nvkm_ufault_new()
107 nvkm_object_ctor(&nvkm_ufault, oclass, &buffer->object); in nvkm_ufault_new()
[all …]
/drivers/iio/buffer/
A Dindustrialio-buffer-dmaengine.c45 return container_of(buffer, struct dmaengine_buffer, queue.buffer); in iio_buffer_to_dmaengine_buffer()
192 struct iio_buffer *buffer = to_iio_dev_attr(attr)->buffer; in iio_dmaengine_buffer_get_length_align() local
269 iio_buffer_put(buffer); in iio_dmaengine_buffer_free()
299 if (IS_ERR(buffer)) in __iio_dmaengine_buffer_setup_ext()
312 return buffer; in __iio_dmaengine_buffer_setup_ext()
343 if (IS_ERR(buffer)) in iio_dmaengine_buffer_setup_ext()
346 return buffer; in iio_dmaengine_buffer_setup_ext()
375 if (IS_ERR(buffer)) in devm_iio_dmaengine_buffer_setup_ext()
379 buffer); in devm_iio_dmaengine_buffer_setup_ext()
412 if (IS_ERR(buffer)) in devm_iio_dmaengine_buffer_setup_with_handle()
[all …]
/drivers/usb/misc/
A Dcytherm.c85 if (!buffer) in brightness_store()
106 kfree(buffer); in brightness_store()
128 if (!buffer) in temp_show()
143 kfree(buffer); in temp_show()
163 if (!buffer) in button_show()
173 kfree(buffer); in button_show()
192 if (!buffer) in port0_show()
201 kfree(buffer); in port0_show()
217 if (!buffer) in port0_store()
247 if (!buffer) in port1_show()
[all …]
/drivers/net/ethernet/sfc/siena/
A Dtx_common.c39 tx_queue->buffer = kcalloc(entries, sizeof(*tx_queue->buffer), in efx_siena_probe_tx_queue()
41 if (!tx_queue->buffer) in efx_siena_probe_tx_queue()
102 if (!tx_queue->buffer) in efx_siena_remove_tx_queue()
129 dma_addr_t unmap_addr = buffer->dma_addr - buffer->dma_offset; in efx_dequeue_buffer()
166 buffer->len = 0; in efx_dequeue_buffer()
167 buffer->flags = 0; in efx_dequeue_buffer()
184 buffer = &tx_queue->buffer[tx_queue->read_count & tx_queue->ptr_mask]; in efx_siena_fini_tx_queue()
210 struct efx_tx_buffer *buffer = &tx_queue->buffer[read_ptr]; in efx_dequeue_buffers() local
312 return buffer; in efx_siena_tx_map_chunk()
378 buffer->dma_offset = buffer->dma_addr - unmap_addr; in efx_siena_tx_map_data()
[all …]
/drivers/net/ethernet/sfc/
A Dtx_common.c39 tx_queue->buffer = kcalloc(entries, sizeof(*tx_queue->buffer), in efx_probe_tx_queue()
41 if (!tx_queue->buffer) in efx_probe_tx_queue()
112 if (!tx_queue->buffer) in efx_fini_tx_queue()
121 buffer = &tx_queue->buffer[tx_queue->read_count & tx_queue->ptr_mask]; in efx_fini_tx_queue()
166 dma_addr_t unmap_addr = buffer->dma_addr - buffer->dma_offset; in efx_dequeue_buffer()
213 buffer->len = 0; in efx_dequeue_buffer()
214 buffer->flags = 0; in efx_dequeue_buffer()
237 struct efx_tx_buffer *buffer = &tx_queue->buffer[read_ptr]; in efx_dequeue_buffers() local
352 return buffer; in efx_tx_map_chunk()
418 buffer->dma_offset = buffer->dma_addr - unmap_addr; in efx_tx_map_data()
[all …]
/drivers/md/dm-vdo/
A Dencodings.c258 decode_u32_le(buffer, offset, &unused); in decode_volume_geometry()
261 decode_u64_le(buffer, offset, &nonce); in decode_volume_geometry()
284 decode_u32_le(buffer, offset, &mem); in decode_volume_geometry()
286 sparse = buffer[(*offset)++]; in decode_volume_geometry()
339 struct block_map_page *page = buffer; in vdo_format_block_map_page()
341 memset(buffer, 0, VDO_BLOCK_SIZE); in vdo_format_block_map_page()
594 decode_u64_le(buffer, offset, &count); in decode_slab_depot_state_2_0()
617 zone_count = buffer[(*offset)++]; in decode_slab_depot_state_2_0()
984 buffer[(*offset)++] = partition->id; in encode_layout()
987 encode_u64_le(buffer, offset, 0); in encode_layout()
[all …]
/drivers/acpi/acpica/
A Dexstorob.c34 u8 *buffer; in acpi_ex_store_buffer_to_buffer() local
46 buffer = ACPI_CAST_PTR(u8, source_desc->buffer.pointer); in acpi_ex_store_buffer_to_buffer()
47 length = source_desc->buffer.length; in acpi_ex_store_buffer_to_buffer()
56 if (!target_desc->buffer.pointer) { in acpi_ex_store_buffer_to_buffer()
60 target_desc->buffer.length = length; in acpi_ex_store_buffer_to_buffer()
70 target_desc->buffer.length); in acpi_ex_store_buffer_to_buffer()
71 memcpy(target_desc->buffer.pointer, buffer, length); in acpi_ex_store_buffer_to_buffer()
98 memcpy(target_desc->buffer.pointer, buffer, in acpi_ex_store_buffer_to_buffer()
99 target_desc->buffer.length); in acpi_ex_store_buffer_to_buffer()
108 target_desc->buffer.flags = source_desc->buffer.flags; in acpi_ex_store_buffer_to_buffer()
[all …]
/drivers/gpu/drm/nouveau/
A Dnouveau_svm.c70 } buffer[]; member
477 if (!buffer->fault[buffer->fault_nr]) { in nouveau_svm_fault_cache()
483 buffer->fault[buffer->fault_nr] = fault; in nouveau_svm_fault_cache()
486 fault = buffer->fault[buffer->fault_nr++]; in nouveau_svm_fault_cache()
720 struct nouveau_svm *svm = container_of(buffer, typeof(*svm), buffer[buffer->id]); in nouveau_svm_fault()
733 if (buffer->get == buffer->put) { in nouveau_svm_fault()
736 if (buffer->get == buffer->put) in nouveau_svm_fault()
742 while (buffer->get != buffer->put) { in nouveau_svm_fault()
744 if (++buffer->get == buffer->entries) in nouveau_svm_fault()
754 sort(buffer->fault, buffer->fault_nr, sizeof(*buffer->fault), in nouveau_svm_fault()
[all …]
/drivers/net/ethernet/sfc/falcon/
A Dtx.c39 buffer->unmap_len = 0; in ef4_tx_get_copy_buffer()
50 dma_addr_t unmap_addr = buffer->dma_addr - buffer->dma_offset; in ef4_dequeue_buffer()
57 buffer->unmap_len = 0; in ef4_dequeue_buffer()
69 buffer->len = 0; in ef4_dequeue_buffer()
70 buffer->flags = 0; in ef4_dequeue_buffer()
166 buffer->skb = skb; in ef4_enqueue_skb_copy()
194 return buffer; in ef4_tx_map_chunk()
233 buffer->dma_offset = buffer->dma_addr - unmap_addr; in ef4_tx_map_data()
239 buffer->skb = skb; in ef4_tx_map_data()
359 struct ef4_tx_buffer *buffer = &tx_queue->buffer[read_ptr]; in ef4_dequeue_buffers() local
[all …]

Completed in 67 milliseconds

12345678910>>...64