Lines Matching refs:fq

640 static inline void dpaa_assign_wq(struct dpaa_fq *fq, int idx)  in dpaa_assign_wq()  argument
642 switch (fq->fq_type) { in dpaa_assign_wq()
645 fq->wq = 1; in dpaa_assign_wq()
649 fq->wq = 5; in dpaa_assign_wq()
653 fq->wq = 6; in dpaa_assign_wq()
659 fq->wq = 6; in dpaa_assign_wq()
663 fq->wq = 2; in dpaa_assign_wq()
667 fq->wq = 1; in dpaa_assign_wq()
671 fq->wq = 0; in dpaa_assign_wq()
680 fq->fq_type, fq->fqid); in dpaa_assign_wq()
887 struct dpaa_fq *fq, in dpaa_setup_ingress() argument
890 fq->fq_base = *template; in dpaa_setup_ingress()
891 fq->net_dev = priv->net_dev; in dpaa_setup_ingress()
893 fq->flags = QMAN_FQ_FLAG_NO_ENQUEUE; in dpaa_setup_ingress()
894 fq->channel = priv->channel; in dpaa_setup_ingress()
898 struct dpaa_fq *fq, in dpaa_setup_egress() argument
902 fq->fq_base = *template; in dpaa_setup_egress()
903 fq->net_dev = priv->net_dev; in dpaa_setup_egress()
906 fq->flags = QMAN_FQ_FLAG_TO_DCPORTAL; in dpaa_setup_egress()
907 fq->channel = (u16)fman_port_get_qman_channel_id(port); in dpaa_setup_egress()
909 fq->flags = QMAN_FQ_FLAG_NO_MODIFY; in dpaa_setup_egress()
920 struct dpaa_fq *fq; in dpaa_fq_setup() local
930 list_for_each_entry(fq, &priv->dpaa_fq_list, list) { in dpaa_fq_setup()
931 switch (fq->fq_type) { in dpaa_fq_setup()
933 dpaa_setup_ingress(priv, fq, &fq_cbs->rx_defq); in dpaa_fq_setup()
936 dpaa_setup_ingress(priv, fq, &fq_cbs->rx_errq); in dpaa_fq_setup()
941 dpaa_setup_ingress(priv, fq, &fq_cbs->rx_defq); in dpaa_fq_setup()
942 fq->channel = channels[portal_cnt++ % num_portals]; in dpaa_fq_setup()
945 dpaa_setup_egress(priv, fq, tx_port, in dpaa_fq_setup()
951 priv->egress_fqs[egress_cnt++] = &fq->fq_base; in dpaa_fq_setup()
954 priv->conf_fqs[conf_cnt++] = &fq->fq_base; in dpaa_fq_setup()
957 dpaa_setup_ingress(priv, fq, &fq_cbs->tx_defq); in dpaa_fq_setup()
960 dpaa_setup_ingress(priv, fq, &fq_cbs->tx_errq); in dpaa_fq_setup()
971 list_for_each_entry(fq, &priv->dpaa_fq_list, list) { in dpaa_fq_setup()
972 if (fq->fq_type != FQ_TYPE_TX) in dpaa_fq_setup()
974 priv->egress_fqs[egress_cnt++] = &fq->fq_base; in dpaa_fq_setup()
999 struct qman_fq *fq; in dpaa_fq_init() local
1016 fq = &dpaa_fq->fq_base; in dpaa_fq_init()
1122 err = qman_init_fq(fq, QMAN_INITFQ_FLAG_SCHED, &initfq); in dpaa_fq_init()
1125 qman_fq_fqid(fq), err); in dpaa_fq_init()
1126 qman_destroy_fq(fq); in dpaa_fq_init()
1131 dpaa_fq->fqid = qman_fq_fqid(fq); in dpaa_fq_init()
1155 static int dpaa_fq_free_entry(struct device *dev, struct qman_fq *fq) in dpaa_fq_free_entry() argument
1163 dpaa_fq = container_of(fq, struct dpaa_fq, fq_base); in dpaa_fq_free_entry()
1167 err = qman_retire_fq(fq, NULL); in dpaa_fq_free_entry()
1170 qman_fq_fqid(fq), err); in dpaa_fq_free_entry()
1172 error = qman_oos_fq(fq); in dpaa_fq_free_entry()
1175 qman_fq_fqid(fq), error); in dpaa_fq_free_entry()
1186 qman_destroy_fq(fq); in dpaa_fq_free_entry()
2451 struct qman_fq *fq, in rx_error_dqrr() argument
2455 struct dpaa_fq *dpaa_fq = container_of(fq, struct dpaa_fq, fq_base); in rx_error_dqrr()
2473 dpaa_rx_error(net_dev, priv, percpu_priv, &dq->fd, fq->fqid); in rx_error_dqrr()
2641 struct qman_fq *fq, in rx_default_dqrr() argument
2667 dpaa_fq = container_of(fq, struct dpaa_fq, fq_base); in rx_default_dqrr()
2677 trace_dpaa_rx_fd(net_dev, fq, &dq->fd); in rx_default_dqrr()
2800 struct qman_fq *fq, in conf_error_dqrr() argument
2808 net_dev = ((struct dpaa_fq *)fq)->net_dev; in conf_error_dqrr()
2816 dpaa_tx_error(net_dev, priv, percpu_priv, &dq->fd, fq->fqid); in conf_error_dqrr()
2822 struct qman_fq *fq, in conf_dflt_dqrr() argument
2830 net_dev = ((struct dpaa_fq *)fq)->net_dev; in conf_dflt_dqrr()
2834 trace_dpaa_tx_conf_fd(net_dev, fq, &dq->fd); in conf_dflt_dqrr()
2841 dpaa_tx_conf(net_dev, priv, percpu_priv, &dq->fd, fq->fqid); in conf_dflt_dqrr()
2847 struct qman_fq *fq, in egress_ern() argument
2856 net_dev = ((struct dpaa_fq *)fq)->net_dev; in egress_ern()