| /linux/include/linux/ |
| A D | folio_queue.h | 51 folio_batch_init(&folioq->vec); in folioq_init() 52 folioq->next = NULL; in folioq_init() 53 folioq->prev = NULL; in folioq_init() 54 folioq->marks = 0; in folioq_init() 55 folioq->marks2 = 0; in folioq_init() 56 folioq->marks3 = 0; in folioq_init() 93 return folioq_count(folioq) >= folioq_nr_slots(folioq); in folioq_full() 119 set_bit(slot, &folioq->marks); in folioq_mark() 261 folioq_mark(folioq, slot); in folioq_append_mark() 291 return folioq->orders[slot]; in folioq_folio_order() [all …]
|
| A D | iov_iter.h | 151 const struct folio_queue *folioq = iter->folioq; in iterate_folioq() local 155 if (slot == folioq_nr_slots(folioq)) { in iterate_folioq() 157 folioq = folioq->next; in iterate_folioq() 162 struct folio *folio = folioq_folio(folioq, slot); in iterate_folioq() 170 fsize = folioq_folio_size(folioq, slot); in iterate_folioq() 182 if (slot == folioq_nr_slots(folioq) && folioq->next) { in iterate_folioq() 183 folioq = folioq->next; in iterate_folioq() 192 iter->folioq = folioq; in iterate_folioq()
|
| A D | uio.h | 71 const struct folio_queue *folioq; member 292 const struct folio_queue *folioq,
|
| /linux/fs/netfs/ |
| A D | read_pgpriv2.c | 30 folioq_mark3(folioq, slot); in netfs_pgpriv2_mark_copy_to_cache() 42 while (folioq) { in netfs_pgpriv2_cancel() 43 if (!folioq->marks3) { in netfs_pgpriv2_cancel() 44 folioq = folioq->next; in netfs_pgpriv2_cancel() 154 for (folioq = rreq->buffer; folioq; folioq = folioq->next) { in netfs_pgpriv2_write_to_the_cache() 155 if (folioq->marks3) { in netfs_pgpriv2_write_to_the_cache() 160 if (!folioq) in netfs_pgpriv2_write_to_the_cache() 180 if (!folioq->marks3) { in netfs_pgpriv2_write_to_the_cache() 181 folioq = folioq->next; in netfs_pgpriv2_write_to_the_cache() 182 if (!folioq) in netfs_pgpriv2_write_to_the_cache() [all …]
|
| A D | read_collect.c | 36 struct folio_queue *folioq, in netfs_unlock_read_folio() argument 81 folioq_clear(folioq, slot); in netfs_unlock_read_folio() 109 if (WARN_ON_ONCE(!folioq) || in netfs_consume_read_data() 117 if (folioq) { in netfs_consume_read_data() 121 folioq->orders[0], folioq->orders[1], in netfs_consume_read_data() 122 folioq->orders[2], folioq->orders[3]); in netfs_consume_read_data() 178 folioq_mark2(folioq, slot); in netfs_consume_read_data() 219 if (slot >= folioq_nr_slots(folioq)) { in netfs_consume_read_data() 221 folioq = folioq->next; in netfs_consume_read_data() 222 subreq->curr_folioq = folioq; in netfs_consume_read_data() [all …]
|
| A D | buffered_read.c | 78 folioq->vec.nr = nr; in netfs_load_buffer_from_ra() 92 folioq_clear(folioq, i); in netfs_load_buffer_from_ra() 164 subreq->io_iter.folioq = subreq->io_iter.folioq->next; in netfs_prepare_read_iterator() 362 folioq = kmalloc(sizeof(*folioq), GFP_KERNEL); in netfs_prime_buffer() 363 if (!folioq) in netfs_prime_buffer() 366 folioq_init(folioq); in netfs_prime_buffer() 367 rreq->buffer = folioq; in netfs_prime_buffer() 439 folioq = kmalloc(sizeof(*folioq), GFP_KERNEL); in netfs_create_singular_buffer() 440 if (!folioq) in netfs_create_singular_buffer() 444 folioq_init(folioq); in netfs_create_singular_buffer() [all …]
|
| A D | iterator.c | 199 const struct folio_queue *folioq = iter->folioq; in netfs_limit_folioq() local 210 if (slot >= folioq_nr_slots(folioq)) { in netfs_limit_folioq() 211 folioq = folioq->next; in netfs_limit_folioq() 217 size_t flen = folioq_folio_size(folioq, slot); in netfs_limit_folioq() 230 if (slot >= folioq_nr_slots(folioq)) { in netfs_limit_folioq() 231 folioq = folioq->next; in netfs_limit_folioq() 234 } while (folioq); in netfs_limit_folioq()
|
| A D | write_collect.c | 86 struct folio_queue *folioq = wreq->buffer; in netfs_writeback_unlock_folios() local 96 if (slot >= folioq_nr_slots(folioq)) { in netfs_writeback_unlock_folios() 97 folioq = netfs_delete_buffer_head(wreq); in netfs_writeback_unlock_folios() 107 folio = folioq_folio(folioq, slot); in netfs_writeback_unlock_folios() 134 folioq_clear(folioq, slot); in netfs_writeback_unlock_folios() 136 if (slot >= folioq_nr_slots(folioq)) { in netfs_writeback_unlock_folios() 137 if (READ_ONCE(wreq->buffer_tail) == folioq) in netfs_writeback_unlock_folios() 139 folioq = netfs_delete_buffer_head(wreq); in netfs_writeback_unlock_folios() 147 wreq->buffer = folioq; in netfs_writeback_unlock_folios()
|
| A D | misc.c | 50 if (rreq->io_iter.folioq == prev && in netfs_buffer_make_space() 52 rreq->io_iter.folioq = tail; in netfs_buffer_make_space()
|
| A D | read_retry.c | 18 subreq->curr_folioq = (struct folio_queue *)io_iter->folioq; in netfs_reissue_read()
|
| A D | internal.h | 98 struct folio_queue *folioq,
|
| A D | write_issue.c | 163 wreq_iter->folioq_slot >= folioq_nr_slots(wreq_iter->folioq)) { in netfs_prepare_write()
|
| /linux/Documentation/core-api/ |
| A D | folio_queue.rst | 64 void folioq_init(struct folio_queue *folioq); 77 unsigned int folioq_append(struct folio_queue *folioq, 80 unsigned int folioq_append_mark(struct folio_queue *folioq, 91 void folioq_clear(struct folio_queue *folioq, unsigned int slot); 104 struct folio *folioq_folio(const struct folio_queue *folioq, 113 size_t folioq_folio_size(const struct folio_queue *folioq, 126 unsigned int folioq_nr_slots(const struct folio_queue *folioq); 128 unsigned int folioq_count(struct folio_queue *folioq); 130 bool folioq_full(struct folio_queue *folioq); 152 void folioq_mark(struct folio_queue *folioq, unsigned int slot); [all …]
|
| /linux/lib/ |
| A D | iov_iter.c | 542 folioq = folioq->next; in iov_iter_folioq_advance() 555 folioq = folioq->next; in iov_iter_folioq_advance() 562 i->folioq = folioq; in iov_iter_folioq_advance() 594 folioq = folioq->prev; in iov_iter_folioq_revert() 608 i->folioq = folioq; in iov_iter_folioq_revert() 733 .folioq = folioq, in iov_iter_folio_queue() 1009 folioq = folioq->next; in iter_folioq_get_pages() 1045 folioq = folioq->next; in iter_folioq_get_pages() 1053 iter->folioq = folioq; in iter_folioq_get_pages() 1592 folioq = folioq->next; in iov_iter_extract_folioq_pages() [all …]
|
| A D | kunit_iov_iter.c | 373 for (folioq = data; folioq; folioq = next) { in iov_kunit_destroy_folioq() 374 next = folioq->next; in iov_kunit_destroy_folioq() 376 if (folioq_folio(folioq, i)) in iov_kunit_destroy_folioq() 378 kfree(folioq); in iov_kunit_destroy_folioq() 384 struct folio_queue *folioq, in iov_kunit_load_folioq() argument 407 struct folio_queue *folioq; in iov_kunit_create_folioq() local 412 folioq_init(folioq); in iov_kunit_create_folioq() 413 return folioq; in iov_kunit_create_folioq() 423 struct folio_queue *folioq; in iov_kunit_copy_to_folioq() local 485 struct folio_queue *folioq; in iov_kunit_copy_from_folioq() local [all …]
|
| A D | scatterlist.c | 1275 const struct folio_queue *folioq = iter->folioq; in extract_folioq_to_sg() local 1281 BUG_ON(!folioq); in extract_folioq_to_sg() 1283 if (slot >= folioq_nr_slots(folioq)) { in extract_folioq_to_sg() 1284 folioq = folioq->next; in extract_folioq_to_sg() 1285 if (WARN_ON_ONCE(!folioq)) in extract_folioq_to_sg() 1291 struct folio *folio = folioq_folio(folioq, slot); in extract_folioq_to_sg() 1292 size_t fsize = folioq_folio_size(folioq, slot); in extract_folioq_to_sg() 1308 if (slot >= folioq_nr_slots(folioq)) { in extract_folioq_to_sg() 1309 if (!folioq->next) { in extract_folioq_to_sg() 1313 folioq = folioq->next; in extract_folioq_to_sg() [all …]
|
| /linux/fs/smb/client/ |
| A D | smbdirect.c | 2536 const struct folio_queue *folioq = iter->folioq; in smb_extract_folioq_to_rdma() local 2541 BUG_ON(!folioq); in smb_extract_folioq_to_rdma() 2543 if (slot >= folioq_nr_slots(folioq)) { in smb_extract_folioq_to_rdma() 2544 folioq = folioq->next; in smb_extract_folioq_to_rdma() 2545 if (WARN_ON_ONCE(!folioq)) in smb_extract_folioq_to_rdma() 2551 struct folio *folio = folioq_folio(folioq, slot); in smb_extract_folioq_to_rdma() 2552 size_t fsize = folioq_folio_size(folioq, slot); in smb_extract_folioq_to_rdma() 2567 if (slot >= folioq_nr_slots(folioq)) { in smb_extract_folioq_to_rdma() 2568 if (!folioq->next) { in smb_extract_folioq_to_rdma() 2572 folioq = folioq->next; in smb_extract_folioq_to_rdma() [all …]
|
| A D | smb2ops.c | 4393 struct folio_queue *folioq; in cifs_clear_folioq_buffer() local 4395 while ((folioq = buffer)) { in cifs_clear_folioq_buffer() 4396 for (int s = 0; s < folioq_count(folioq); s++) in cifs_clear_folioq_buffer() 4397 if (folioq_is_marked(folioq, s)) in cifs_clear_folioq_buffer() 4398 folio_put(folioq_folio(folioq, s)); in cifs_clear_folioq_buffer() 4399 buffer = folioq->next; in cifs_clear_folioq_buffer() 4400 kfree(folioq); in cifs_clear_folioq_buffer() 4601 cifs_copy_folioq_to_iter(struct folio_queue *folioq, size_t data_size, in cifs_copy_folioq_to_iter() argument 4604 for (; folioq; folioq = folioq->next) { in cifs_copy_folioq_to_iter() 4605 for (int s = 0; s < folioq_count(folioq); s++) { in cifs_copy_folioq_to_iter() [all …]
|