Searched refs:dma_align (Results 1 – 11 of 11) sorted by relevance
559 dma_addr_t dma_align; in enetc_cbd_alloc_data_mem() local570 dma_align = ALIGN(*dma, ENETC_CBD_DATA_MEM_ALIGN); in enetc_cbd_alloc_data_mem()573 cbd->addr[0] = cpu_to_le32(lower_32_bits(dma_align)); in enetc_cbd_alloc_data_mem()574 cbd->addr[1] = cpu_to_le32(upper_32_bits(dma_align)); in enetc_cbd_alloc_data_mem()
196 dma_addr_t dma_align; in ntmp_fill_request_hdr() local199 dma_align = ALIGN(dma, NTMP_DATA_ADDR_ALIGN); in ntmp_fill_request_hdr()200 cbd->req_hdr.addr = cpu_to_le64(dma_align); in ntmp_fill_request_hdr()
84 unsigned int dma_align; member
369 args->pitch = ALIGN(pitch, dpsub->dma_align); in zynqmp_dpsub_dumb_create()385 cmd.pitches[i] = ALIGN(cmd.pitches[i], dpsub->dma_align); in zynqmp_dpsub_fb_create()
1390 dpsub->dma_align = 1 << layer->dmas[0].chan->device->copy_align; in zynqmp_disp_probe()
573 unsigned long dma_align = dma_get_cache_alignment(); in vb2_dc_get_userptr() local576 if (!IS_ALIGNED(vaddr | size, dma_align)) { in vb2_dc_get_userptr()577 pr_debug("user data must be aligned to %lu bytes\n", dma_align); in vb2_dc_get_userptr()
919 size_t dma_align = dma_get_cache_alignment(); in spi_qup_can_dma() local923 if (!IS_ALIGNED((size_t)xfer->rx_buf, dma_align) || in spi_qup_can_dma()931 if (!IS_ALIGNED((size_t)xfer->tx_buf, dma_align) || in spi_qup_can_dma()
853 u64 dma_align = (u64)(uintptr_t)xqspi->rxbuf; in zynqmp_qspi_setuprxdma() local856 ((dma_align & GQSPI_DMA_UNALIGN) != 0x0)) { in zynqmp_qspi_setuprxdma()
1413 u64 dma_align = (u64)(uintptr_t)buf; in cqspi_read() local1424 virt_addr_valid(buf) && ((dma_align & CQSPI_DMA_UNALIGN) == 0)) in cqspi_read()
2407 md->properties.dma_align = attr.dma_align; in mport_cdev_add()
2600 attr->dma_align = 0; in tsi721_query_mport()
Completed in 40 milliseconds