| /fs/xfs/ |
| A D | xfs_extent_busy.c | 163 xfs_extlen_t flen, in xfs_extent_busy_update_extent() argument 169 xfs_agblock_t fend = fbno + flen; in xfs_extent_busy_update_extent() 311 xfs_extlen_t flen, in xfs_extent_busy_reuse() argument 317 ASSERT(flen > 0); in xfs_extent_busy_reuse() 327 if (fbno + flen <= bbno) { in xfs_extent_busy_reuse() 365 xfs_extlen_t flen; in xfs_extent_busy_trim() local 373 flen = *len; in xfs_extent_busy_trim() 375 while (rbp && flen >= minlen) { in xfs_extent_busy_trim() 510 flen = fend - fbno; in xfs_extent_busy_trim() 517 *len = flen; in xfs_extent_busy_trim() [all …]
|
| A D | xfs_discard.c | 235 xfs_extlen_t flen; in xfs_trim_gather_extents() local 237 error = xfs_alloc_get_rec(cur, &fbno, &flen, &i); in xfs_trim_gather_extents() 252 tcur->count = flen; in xfs_trim_gather_extents() 261 if (fbno + flen < tcur->start) { in xfs_trim_gather_extents() 262 trace_xfs_discard_exclude(pag_group(pag), fbno, flen); in xfs_trim_gather_extents() 266 trace_xfs_discard_exclude(pag_group(pag), fbno, flen); in xfs_trim_gather_extents() 276 flen -= tcur->start - fbno; in xfs_trim_gather_extents() 279 if (fbno + flen > tcur->end + 1) in xfs_trim_gather_extents() 280 flen = tcur->end - fbno + 1; in xfs_trim_gather_extents() 283 if (flen < tcur->minlen) { in xfs_trim_gather_extents() [all …]
|
| A D | xfs_extent_busy.h | 54 xfs_extlen_t flen, bool userdata);
|
| A D | xfs_fsmap.c | 216 xfs_extlen_t flen = 0; in xfs_getfsmap_is_shared() local 231 XFS_BB_TO_FSBT(mp, frec->len_daddr), &fbno, &flen, in xfs_getfsmap_is_shared() 238 *stat = flen > 0; in xfs_getfsmap_is_shared()
|
| /fs/netfs/ |
| A D | read_pgpriv2.c | 22 size_t fsize = folio_size(folio), flen = fsize; in netfs_pgpriv2_copy_folio() local 44 if (flen > i_size - fpos) { in netfs_pgpriv2_copy_folio() 45 flen = i_size - fpos; in netfs_pgpriv2_copy_folio() 47 } else if (flen == i_size - fpos) { in netfs_pgpriv2_copy_folio() 51 _debug("folio %zx %zx", flen, fsize); in netfs_pgpriv2_copy_folio() 63 cache->submit_len = flen; in netfs_pgpriv2_copy_folio() 90 if (flen < fsize) in netfs_pgpriv2_copy_folio() 186 size_t fsize, flen; in netfs_pgpriv2_unlock_copied_folios() local 196 flen = fsize; in netfs_pgpriv2_unlock_copied_folios() 198 fend = min_t(unsigned long long, fpos + flen, creq->i_size); in netfs_pgpriv2_unlock_copied_folios()
|
| A D | write_issue.c | 364 flen = foff + finfo->dirty_len; in netfs_write_folio() 370 if (flen > i_size - fpos) in netfs_write_folio() 371 flen = i_size - fpos; in netfs_write_folio() 372 } else if (flen > i_size - fpos) { in netfs_write_folio() 373 flen = i_size - fpos; in netfs_write_folio() 380 flen -= foff; in netfs_write_folio() 456 stream->submit_len = flen; in netfs_write_folio() 516 if (foff + flen < fsize) in netfs_write_folio() 753 flen = folio_size(folio); in netfs_write_folio_single() 755 flen = wreq->i_size - fpos; in netfs_write_folio_single() [all …]
|
| A D | iterator.c | 170 size_t offset, flen, len; in netfs_limit_xarray() local 178 flen = folio_size(folio); in netfs_limit_xarray() 180 len = min(max_size, flen - offset); in netfs_limit_xarray() 217 size_t flen = folioq_folio_size(folioq, slot); in netfs_limit_folioq() local 219 if (start_offset < flen) { in netfs_limit_folioq() 220 span += flen - start_offset; in netfs_limit_folioq() 224 start_offset -= flen; in netfs_limit_folioq()
|
| A D | buffered_write.c | 156 size_t flen; in netfs_perform_write() local 185 flen = folio_size(folio); in netfs_perform_write() 188 part = min_t(size_t, flen - offset, part); in netfs_perform_write() 239 folio_zero_segment(folio, offset + copied, flen); in netfs_perform_write() 247 if (!maybe_trouble && offset == 0 && part >= flen) { in netfs_perform_write() 298 if (offset == 0 && copied == flen) { in netfs_perform_write() 329 if (finfo->dirty_offset == 0 && finfo->dirty_len == flen) { in netfs_perform_write() 348 ret = filemap_write_and_wait_range(mapping, fpos, fpos + flen - 1); in netfs_perform_write() 366 offset + copied == flen, in netfs_perform_write()
|
| A D | buffered_read.c | 408 size_t flen = folio_size(folio); in netfs_read_gaps() local 409 size_t nr_bvec = flen / PAGE_SIZE + 2; in netfs_read_gaps() 415 rreq = netfs_alloc_request(mapping, file, folio_pos(folio), flen, NETFS_READ_GAPS); in netfs_read_gaps() 455 if (to < flen) in netfs_read_gaps() 456 bvec_set_folio(&bvec[i++], folio, flen - to, to); in netfs_read_gaps() 458 rreq->submitted = rreq->start + flen; in netfs_read_gaps() 723 size_t flen = folio_size(folio); in netfs_prefetch_for_write() local 726 _enter("%zx @%llx", flen, start); in netfs_prefetch_for_write() 730 rreq = netfs_alloc_request(mapping, file, start, flen, in netfs_prefetch_for_write() 744 trace_netfs_read(rreq, start, flen, netfs_read_trace_prefetch_for_write); in netfs_prefetch_for_write()
|
| A D | misc.c | 215 size_t flen = folio_size(folio); in netfs_invalidate_folio() local 219 if (offset == 0 && length == flen) { in netfs_invalidate_folio() 223 end = umin(fpos + flen, i_size); in netfs_invalidate_folio() 235 if (offset == 0 && length >= flen) in netfs_invalidate_folio()
|
| A D | write_collect.c | 141 size_t fsize, flen; in netfs_writeback_unlock_folios() local 152 flen = finfo ? finfo->dirty_offset + finfo->dirty_len : fsize; in netfs_writeback_unlock_folios() 154 fend = min_t(unsigned long long, fpos + flen, wreq->i_size); in netfs_writeback_unlock_folios()
|
| /fs/udf/ |
| A D | dir.c | 39 int flen; in udf_readdir() local 106 flen = udf_get_filename(sb, iter.name, in udf_readdir() 108 if (flen < 0) in udf_readdir() 113 if (!dir_emit(ctx, fname, flen, iblock, DT_UNKNOWN)) in udf_readdir()
|
| A D | namei.c | 55 int flen; in udf_fiiter_find_entry() local 86 flen = udf_get_filename(sb, iter->name, in udf_fiiter_find_entry() 88 if (flen < 0) { in udf_fiiter_find_entry() 89 ret = flen; in udf_fiiter_find_entry() 93 if (udf_match(flen, fname, child->len, child->name)) in udf_fiiter_find_entry()
|
| /fs/xfs/libxfs/ |
| A D | xfs_alloc.c | 660 nflen1 = flen - rlen; in xfs_alloc_fixup_trees() 664 nflen1 = flen - rlen; in xfs_alloc_fixup_trees() 1295 flen = 0; in xfs_alloc_ag_vextent_small() 1298 *flenp = flen; in xfs_alloc_ag_vextent_small() 1365 tlen = flen; in xfs_alloc_ag_vextent_exact() 1898 (rlen > flen || in xfs_alloc_ag_vextent_size() 1912 bestflen = flen; in xfs_alloc_ag_vextent_size() 1934 (rlen > flen || in xfs_alloc_ag_vextent_size() 1943 bestflen = flen; in xfs_alloc_ag_vextent_size() 1959 flen = bestflen; in xfs_alloc_ag_vextent_size() [all …]
|
| A D | xfs_refcount.c | 1645 xfs_extlen_t *flen, in xfs_refcount_find_shared() argument 1657 *flen = 0; in xfs_refcount_find_shared() 1713 *flen = min(tmp.rc_blockcount, agbno + aglen - *fbno); in xfs_refcount_find_shared() 1718 while (*fbno + *flen < agbno + aglen) { in xfs_refcount_find_shared() 1734 tmp.rc_startblock != *fbno + *flen) in xfs_refcount_find_shared() 1736 *flen = min(*flen + tmp.rc_blockcount, agbno + aglen - *fbno); in xfs_refcount_find_shared() 1740 trace_xfs_refcount_find_shared_result(cur, *fbno, *flen); in xfs_refcount_find_shared()
|
| A D | xfs_refcount.h | 91 xfs_extlen_t *flen, bool find_end_of_shared);
|
| /fs/xfs/scrub/ |
| A D | alloc.c | 64 xfs_extlen_t flen; in xchk_allocbt_xref_other() local 83 error = xfs_alloc_get_rec(*pcur, &fbno, &flen, &has_otherrec); in xchk_allocbt_xref_other() 91 if (fbno != agbno || flen != len) in xchk_allocbt_xref_other()
|
| A D | rtrmap.c | 175 xfs_extlen_t flen; in xchk_rtrmapbt_xref_rtrefc() local 192 irec->rm_blockcount, &fbno, &flen, false); in xchk_rtrmapbt_xref_rtrefc() 195 if (flen != 0 && (!is_inode || is_attr || is_bmbt || is_unwritten)) in xchk_rtrmapbt_xref_rtrefc()
|
| A D | rmap.c | 85 xfs_extlen_t flen; in xchk_rmapbt_xref_refc() local 102 irec->rm_blockcount, &fbno, &flen, false); in xchk_rmapbt_xref_refc() 105 if (flen != 0 && (non_inode || is_attr || is_bmbt || is_unwritten)) in xchk_rmapbt_xref_refc()
|
| A D | bmap_repair.c | 107 xfs_extlen_t flen; in xrep_bmap_discover_shared() local 117 error = xfs_refcount_find_shared(cur, agbno, blockcount, &fbno, &flen, in xrep_bmap_discover_shared()
|
| /fs/nfsd/ |
| A D | vfs.c | 1586 char *fname, int flen, struct nfsd_attrs *attrs, in nfsd_create() argument 1593 trace_nfsd_vfs_create(rqstp, fhp, type, fname, flen); in nfsd_create() 1595 if (isdotent(fname, flen)) in nfsd_create() 1687 char *fname, int flen, in nfsd_symlink() argument 1695 trace_nfsd_vfs_symlink(rqstp, fhp, fname, flen, path); in nfsd_symlink() 1698 if (!flen || path[0] == '\0') in nfsd_symlink() 1701 if (isdotent(fname, flen)) in nfsd_symlink() 1897 if (!flen || isdotent(fname, flen) || !tlen || isdotent(tname, tlen)) in nfsd_rename() 2033 char *fname, int flen) in nfsd_unlink() argument 2041 trace_nfsd_vfs_unlink(rqstp, fhp, fname, flen); in nfsd_unlink() [all …]
|
| A D | xdr.h | 49 unsigned int flen; member 65 unsigned int flen; member
|
| A D | xdr3.h | 67 unsigned int flen; member 83 unsigned int flen; member
|
| A D | nfsxdr.c | 361 &args->fname, &args->flen) && in nfssvc_decode_renameargs() 382 if (!svcxdr_decode_diropargs(xdr, &args->ffh, &args->fname, &args->flen)) in nfssvc_decode_symlinkargs()
|
| /fs/ntfs3/ |
| A D | attrib.c | 155 CLST flen, vcn0 = vcn, pre = pre_alloc ? *pre_alloc : 0; in attr_allocate_clusters() local 159 err = ntfs_look_for_free_space(sbi, lcn, len + pre, &lcn, &flen, in attr_allocate_clusters() 177 *new_len = flen; in attr_allocate_clusters() 181 if (!run_add_entry(run, vcn, lcn, flen, opt & ALLOCATE_MFT)) { in attr_allocate_clusters() 193 (sector_t)flen << shift, in attr_allocate_clusters() 199 vcn += flen; in attr_allocate_clusters() 201 if (flen >= len || (opt & ALLOCATE_MFT) || in attr_allocate_clusters() 207 len -= flen; in attr_allocate_clusters()
|