Home
last modified time | relevance | path

Searched refs:gl (Results 1 – 25 of 41) sorted by relevance

12

/linux/fs/gfs2/
A Dglock.c258 GLOCK_BUG_ON(gl, gl->gl_lockref.count < 2); in __gfs2_glock_queue_work()
687 gl->gl_target = gl->gl_state; in finish_xmote()
696 do_xmote(gl, gh, gl->gl_target); in finish_xmote()
767 GLOCK_BUG_ON(gl, gl->gl_state == target); in do_xmote()
768 GLOCK_BUG_ON(gl, gl->gl_state == gl->gl_target); in do_xmote()
903 gl->gl_target = gl->gl_demote_state; in run_queue()
917 do_xmote(gl, gh, gl->gl_target); in run_queue()
1066 finish_xmote(gl, gl->gl_reply); in glock_work_func()
1443 gl->gl_ops->go_callback(gl, remote); in handle_callback()
2550 struct gfs2_glock *gl = gi->gl; in gfs2_glock_iter_next() local
[all …]
A Dglock.h148 spin_lock(&gl->gl_lockref.lock); in gfs2_glock_is_locked_by_me()
160 spin_unlock(&gl->gl_lockref.lock); in gfs2_glock_is_locked_by_me()
305 spin_lock(&gl->gl_lockref.lock); in glock_set_object()
306 if (gfs2_assert_warn(gl->gl_name.ln_sbd, gl->gl_object == NULL)) in glock_set_object()
307 gfs2_dump_glock(NULL, gl, true); in glock_set_object()
308 gl->gl_object = object; in glock_set_object()
332 spin_lock(&gl->gl_lockref.lock); in glock_clear_object()
333 if (gl->gl_object == object) in glock_clear_object()
334 gl->gl_object = NULL; in glock_clear_object()
342 spin_lock(&gl->gl_lockref.lock); in gfs2_holder_allow_demote()
[all …]
A Dglops.c44 gl->gl_name.ln_type, gl->gl_name.ln_number, in gfs2_ail_error()
77 gfs2_ail_error(gl, bh); in __gfs2_ail_flush()
82 GLOCK_BUG_ON(gl, !fsync && atomic_read(&gl->gl_ail_count)); in __gfs2_ail_flush()
198 GLOCK_BUG_ON(gl, gl->gl_state != LM_ST_EXCLUSIVE); in rgrp_go_sync()
248 ip = gl->gl_object; in gfs2_glock2inode()
260 rgd = gl->gl_object; in gfs2_glock2rgrp()
313 GLOCK_BUG_ON(gl, gl->gl_state != LM_ST_EXCLUSIVE); in inode_go_sync()
315 gfs2_log_flush(gl->gl_name.ln_sbd, gl, GFS2_LOG_HEAD_FLUSH_NORMAL | in inode_go_sync()
327 gfs2_ail_empty_gl(gl); in inode_go_sync()
389 if (sdp->sd_jindex == gl->gl_object || sdp->sd_rindex == gl->gl_object) in inode_go_demote_ok()
[all …]
A Dlock_dlm.c110 dstamp = gl->gl_dstamp; in gfs2_update_request_times()
127 if ((gl->gl_lksb.sb_flags & DLM_SBF_VALNOTVALID) && gl->gl_lksb.sb_lvbptr) in gdlm_ast()
132 if (gl->gl_ops->go_free) in gdlm_ast()
133 gl->gl_ops->go_free(gl); in gdlm_ast()
134 gfs2_glock_free(gl); in gdlm_ast()
151 ret = gl->gl_req; in gdlm_ast()
293 gfs2_glock_free(gl); in gdlm_put_lock()
304 gfs2_glock_free(gl); in gdlm_put_lock()
311 gfs2_glock_free(gl); in gdlm_put_lock()
316 NULL, gl); in gdlm_put_lock()
[all …]
A Dtrace_gfs2.h94 TP_ARGS(gl, new_state),
109 __entry->glnum = gl->gl_name.ln_number;
110 __entry->gltype = gl->gl_name.ln_type;
115 __entry->flags = gl->gl_flags | (gl->gl_object ? (1UL<<GLF_OBJECT) : 0);
131 TP_PROTO(const struct gfs2_glock *gl),
133 TP_ARGS(gl),
145 __entry->gltype = gl->gl_name.ln_type;
148 __entry->flags = gl->gl_flags | (gl->gl_object ? (1UL<<GLF_OBJECT) : 0);
165 TP_ARGS(gl, remote),
183 __entry->flags = gl->gl_flags | (gl->gl_object ? (1UL<<GLF_OBJECT) : 0);
[all …]
A Dmain.c53 struct gfs2_glock *gl = foo; in gfs2_init_glock_once() local
55 spin_lock_init(&gl->gl_lockref.lock); in gfs2_init_glock_once()
56 INIT_LIST_HEAD(&gl->gl_holders); in gfs2_init_glock_once()
57 INIT_LIST_HEAD(&gl->gl_lru); in gfs2_init_glock_once()
58 INIT_LIST_HEAD(&gl->gl_ail_list); in gfs2_init_glock_once()
59 atomic_set(&gl->gl_ail_count, 0); in gfs2_init_glock_once()
60 atomic_set(&gl->gl_revokes, 0); in gfs2_init_glock_once()
65 struct gfs2_glock *gl = foo; in gfs2_init_gl_aspace_once() local
66 struct address_space *mapping = (struct address_space *)(gl + 1); in gfs2_init_gl_aspace_once()
68 gfs2_init_glock_once(gl); in gfs2_init_gl_aspace_once()
A Dmeta_io.c114 struct address_space *mapping = gfs2_glock2aspace(gl); in gfs2_getbuf()
115 struct gfs2_sbd *sdp = gl->gl_name.ln_sbd; in gfs2_getbuf()
187 bh = gfs2_getbuf(gl, blkno, CREATE); in gfs2_meta_new()
257 struct gfs2_sbd *sdp = gl->gl_name.ln_sbd; in gfs2_meta_read()
266 *bhp = bh = gfs2_getbuf(gl, blkno, CREATE); in gfs2_meta_read()
279 bh = gfs2_getbuf(gl, blkno + 1, CREATE); in gfs2_meta_read()
484 struct gfs2_glock *gl = ip->i_gl; in gfs2_meta_buffer() local
492 ret = gfs2_meta_read(gl, num, DIO_WAIT, rahead, &bh); in gfs2_meta_buffer()
513 struct gfs2_sbd *sdp = gl->gl_name.ln_sbd; in gfs2_meta_ra()
525 first_bh = gfs2_getbuf(gl, dblock, CREATE); in gfs2_meta_ra()
[all …]
A Dincore.h38 typedef void (*gfs2_glop_bh_t) (struct gfs2_glock *gl, unsigned int ret);
218 int (*go_sync) (struct gfs2_glock *gl);
219 int (*go_xmote_bh)(struct gfs2_glock *gl);
220 void (*go_inval) (struct gfs2_glock *gl, int flags);
221 int (*go_demote_ok) (const struct gfs2_glock *gl);
223 void (*go_dump)(struct seq_file *seq, struct gfs2_glock *gl,
225 void (*go_callback)(struct gfs2_glock *gl, bool remote);
226 void (*go_free)(struct gfs2_glock *gl);
856 gl->gl_stats.stats[which]++; in gfs2_glstats_inc()
861 const struct gfs2_sbd *sdp = gl->gl_name.ln_sbd; in gfs2_sbstats_inc()
[all …]
A Dtrans.c163 static struct gfs2_bufdata *gfs2_alloc_bufdata(struct gfs2_glock *gl, in gfs2_alloc_bufdata() argument
170 bd->bd_gl = gl; in gfs2_alloc_bufdata()
192 void gfs2_trans_add_data(struct gfs2_glock *gl, struct buffer_head *bh) in gfs2_trans_add_data() argument
195 struct gfs2_sbd *sdp = gl->gl_name.ln_sbd; in gfs2_trans_add_data()
209 bd = gfs2_alloc_bufdata(gl, bh); in gfs2_trans_add_data()
215 gfs2_assert(sdp, bd->bd_gl == gl); in gfs2_trans_add_data()
229 void gfs2_trans_add_meta(struct gfs2_glock *gl, struct buffer_head *bh) in gfs2_trans_add_meta() argument
232 struct gfs2_sbd *sdp = gl->gl_name.ln_sbd; in gfs2_trans_add_meta()
250 bd = gfs2_alloc_bufdata(gl, bh); in gfs2_trans_add_meta()
257 gfs2_assert(sdp, bd->bd_gl == gl); in gfs2_trans_add_meta()
A Dmeta_io.h51 extern struct buffer_head *gfs2_meta_new(struct gfs2_glock *gl, u64 blkno);
52 extern int gfs2_meta_read(struct gfs2_glock *gl, u64 blkno, int flags,
55 extern struct buffer_head *gfs2_getbuf(struct gfs2_glock *gl, u64 blkno,
73 struct buffer_head *gfs2_meta_ra(struct gfs2_glock *gl, u64 dblock, u32 extlen);
A Dlops.c73 struct gfs2_glock *gl = bd->bd_gl; in maybe_release_space() local
74 struct gfs2_sbd *sdp = gl->gl_name.ln_sbd; in maybe_release_space()
122 struct gfs2_glock *gl = bd->bd_gl; in gfs2_unpin() local
124 atomic_inc(&gl->gl_ail_count); in gfs2_unpin()
796 struct gfs2_glock *gl = ip->i_gl; in buf_lo_scan_elements() local
819 bh_ip = gfs2_meta_new(gl, blkno); in buf_lo_scan_elements()
906 struct gfs2_glock *gl; in gfs2_drain_revokes() local
911 gl = bd->bd_gl; in gfs2_drain_revokes()
912 gfs2_glock_remove_revoke(gl); in gfs2_drain_revokes()
1020 struct gfs2_glock *gl = ip->i_gl; in databuf_lo_scan_elements() local
[all …]
A Drgrp.c719 struct gfs2_glock *gl; in gfs2_clear_rgrpd() local
723 gl = rgd->rd_gl; in gfs2_clear_rgrpd()
727 if (gl) { in gfs2_clear_rgrpd()
733 glock_clear_object(gl, rgd); in gfs2_clear_rgrpd()
734 gfs2_glock_put(gl); in gfs2_clear_rgrpd()
1035 struct gfs2_glock *gl = ip->i_gl; in gfs2_rindex_update() local
1199 struct gfs2_glock *gl = gh->gh_gl; in gfs2_rgrp_go_instantiate() local
1845 struct gfs2_glock *gl; in try_rgrp_unlink() local
1879 ip = gl->gl_object; in try_rgrp_unlink()
1882 gfs2_glock_put(gl); in try_rgrp_unlink()
[all …]
A Dsuper.c947 gfs2_glock_hold(gl); in gfs2_drop_inode()
948 if (!gfs2_queue_delete_work(gl, 0)) in gfs2_drop_inode()
949 gfs2_glock_queue_put(gl); in gfs2_drop_inode()
1076 struct gfs2_glock *gl = ip->i_gl; in gfs2_final_release_pages() local
1082 clear_bit(GLF_LFLUSH, &gl->gl_flags); in gfs2_final_release_pages()
1083 clear_bit(GLF_DIRTY, &gl->gl_flags); in gfs2_final_release_pages()
1148 gfs2_glock_queue_put(gl); in gfs2_glock_put_eventually()
1150 gfs2_glock_put(gl); in gfs2_glock_put_eventually()
1408 glock_clear_object(gl, ip); in gfs2_evict_inode()
1413 gfs2_glock_hold(gl); in gfs2_evict_inode()
[all …]
A Dlog.c97 struct gfs2_glock *gl = NULL; in gfs2_ail1_start_one() local
126 if (gl == bd->bd_gl) in gfs2_ail1_start_one()
128 gl = bd->bd_gl; in gfs2_ail1_start_one()
768 struct gfs2_glock *gl = bd->bd_gl; in gfs2_add_revoke() local
771 if (atomic_inc_return(&gl->gl_revokes) == 1) in gfs2_add_revoke()
772 gfs2_glock_hold(gl); in gfs2_add_revoke()
777 set_bit(GLF_LFLUSH, &gl->gl_flags); in gfs2_add_revoke()
783 if (atomic_dec_return(&gl->gl_revokes) == 0) { in gfs2_glock_remove_revoke()
784 clear_bit(GLF_LFLUSH, &gl->gl_flags); in gfs2_glock_remove_revoke()
785 gfs2_glock_queue_put(gl); in gfs2_glock_remove_revoke()
[all …]
A Dglops.h25 extern int gfs2_inode_metasync(struct gfs2_glock *gl);
26 extern void gfs2_ail_flush(struct gfs2_glock *gl, bool fsync);
A Dtrans.h44 extern void gfs2_trans_add_data(struct gfs2_glock *gl, struct buffer_head *bh);
45 extern void gfs2_trans_add_meta(struct gfs2_glock *gl, struct buffer_head *bh);
A Dlog.h86 extern void gfs2_log_flush(struct gfs2_sbd *sdp, struct gfs2_glock *gl,
94 extern void gfs2_glock_remove_revoke(struct gfs2_glock *gl);
/linux/drivers/target/iscsi/cxgbit/
A Dcxgbit_main.c231 gl->frags[i].page, in cxgbit_copy_frags()
232 gl->frags[i].offset, in cxgbit_copy_frags()
233 gl->frags[i].size); in cxgbit_copy_frags()
238 get_page(gl->frags[gl->nfrags - 1].page); in cxgbit_copy_frags()
260 if (unlikely(gl->nfrags > 1)) in cxgbit_lro_add_packet_gl()
308 pdu_cb->frags += gl->nfrags; in cxgbit_lro_add_packet_gl()
413 if ((gl && (((skb_shinfo(skb)->nr_frags + gl->nfrags) > in cxgbit_lro_receive()
420 if (gl) in cxgbit_lro_receive()
462 rpl = gl ? (struct cpl_tx_data *)gl->va : in cxgbit_uld_lro_rx_handler()
474 if (!gl) { in cxgbit_uld_lro_rx_handler()
[all …]
/linux/drivers/net/ethernet/chelsio/cxgb4vf/
A Dsge.c1489 get_page(gl->frags[gl->nfrags - 1].page); in copy_frags()
1524 skb_copy_to_linear_data(skb, gl->va, gl->tot_len); in t4vf_pktgl_to_skb()
1553 frag = gl->nfrags - 1; in t4vf_pktgl_free()
1578 t4vf_pktgl_free(gl); in do_gro()
1633 do_gro(rxq, gl, pkt); in t4vf_ethrx_handler()
1642 t4vf_pktgl_free(gl); in t4vf_ethrx_handler()
1780 struct pkt_gl gl; in process_responses() local
1802 gl.tot_len = len; in process_responses()
1820 gl.nfrags = frag+1; in process_responses()
1830 gl.va = (page_address(gl.frags[0].page) + in process_responses()
[all …]
/linux/drivers/net/ethernet/chelsio/inline_crypto/chtls/
A Dchtls_main.c339 static struct sk_buff *copy_gl_to_skb_pkt(const struct pkt_gl *gl, in copy_gl_to_skb_pkt() argument
349 skb = alloc_skb(gl->tot_len + sizeof(struct cpl_pass_accept_req) in copy_gl_to_skb_pkt()
353 __skb_put(skb, gl->tot_len + sizeof(struct cpl_pass_accept_req) in copy_gl_to_skb_pkt()
358 , gl->va + pktshift, in copy_gl_to_skb_pkt()
359 gl->tot_len - pktshift); in copy_gl_to_skb_pkt()
365 const struct pkt_gl *gl, const __be64 *rsp) in chtls_recv_packet() argument
371 skb = copy_gl_to_skb_pkt(gl, rsp, cdev->lldi->sge_pktshift); in chtls_recv_packet()
439 const struct pkt_gl *gl) in chtls_uld_rx_handler() argument
448 if (chtls_recv_packet(cdev, gl, rsp) < 0) in chtls_uld_rx_handler()
453 if (!gl) in chtls_uld_rx_handler()
[all …]
/linux/tools/testing/cxl/test/
A Dmem.c59 struct cxl_mbox_get_log *gl = cmd->payload_in; in mock_get_log() local
60 u32 offset = le32_to_cpu(gl->offset); in mock_get_log()
61 u32 length = le32_to_cpu(gl->length); in mock_get_log()
65 if (cmd->size_in < sizeof(*gl)) in mock_get_log()
71 if (!uuid_equal(&gl->uuid, &uuid)) in mock_get_log()
/linux/drivers/infiniband/hw/cxgb4/
A Ddevice.c1137 gl->va + pktshift, in copy_gl_to_skb_pkt()
1138 gl->tot_len - pktshift); in copy_gl_to_skb_pkt()
1167 const struct pkt_gl *gl) in c4iw_uld_rx_handler() argument
1174 if (gl == NULL) { in c4iw_uld_rx_handler()
1183 } else if (gl == CXGB4_MSG_AN) { in c4iw_uld_rx_handler()
1189 } else if (unlikely(*(u8 *)rsp != *(u8 *)gl->va)) { in c4iw_uld_rx_handler()
1190 if (recv_rx_pkt(dev, gl, rsp)) in c4iw_uld_rx_handler()
1194 pci_name(ctx->lldi.pdev), gl->va, in c4iw_uld_rx_handler()
1196 be64_to_cpu(*(__force __be64 *)gl->va), in c4iw_uld_rx_handler()
1197 gl->tot_len); in c4iw_uld_rx_handler()
[all …]
/linux/drivers/net/ethernet/chelsio/cxgb4/
A Dsge.c3328 gl->frags[i].offset, in copy_frags()
3329 gl->frags[i].size); in copy_frags()
3332 get_page(gl->frags[gl->nfrags - 1].page); in copy_frags()
3358 __skb_put(skb, gl->tot_len); in cxgb4_pktgl_to_skb()
3359 skb_copy_to_linear_data(skb, gl->va, gl->tot_len); in cxgb4_pktgl_to_skb()
3367 copy_frags(skb, gl, pull_len); in cxgb4_pktgl_to_skb()
3368 skb->len = gl->tot_len; in cxgb4_pktgl_to_skb()
3388 for (p = gl->frags, n = gl->nfrags - 1; n--; p++) in t4_pktgl_free()
3403 t4_pktgl_free(gl); in handle_trace_pkt()
3452 t4_pktgl_free(gl); in do_gro()
[all …]
A Dcxgb4_uld.h483 const struct pkt_gl *gl);
487 const struct pkt_gl *gl,
529 struct sk_buff *cxgb4_pktgl_to_skb(const struct pkt_gl *gl,
A Dcxgb4_uld.c75 const struct pkt_gl *gl) in uldrx_handler() argument
88 rsp, gl, &q->lro_mgr, in uldrx_handler()
92 rsp, gl); in uldrx_handler()
99 if (!gl) in uldrx_handler()
101 else if (gl == CXGB4_MSG_AN) in uldrx_handler()

Completed in 770 milliseconds

12