Home
last modified time | relevance | path

Searched refs:fq (Results 1 – 14 of 14) sorted by relevance

/linux-4.19.296/include/net/
Dfq_impl.h15 static void fq_adjust_removal(struct fq *fq, in fq_adjust_removal() argument
24 fq->backlog--; in fq_adjust_removal()
25 fq->memory_usage -= skb->truesize; in fq_adjust_removal()
28 static void fq_rejigger_backlog(struct fq *fq, struct fq_flow *flow) in fq_rejigger_backlog() argument
37 list_for_each_entry_continue(i, &fq->backlogs, backlogchain) in fq_rejigger_backlog()
46 static struct sk_buff *fq_flow_dequeue(struct fq *fq, in fq_flow_dequeue() argument
51 lockdep_assert_held(&fq->lock); in fq_flow_dequeue()
57 fq_adjust_removal(fq, flow, skb); in fq_flow_dequeue()
58 fq_rejigger_backlog(fq, flow); in fq_flow_dequeue()
63 static struct sk_buff *fq_tin_dequeue(struct fq *fq, in fq_tin_dequeue() argument
[all …]
Dipv6_frag.h33 struct frag_queue *fq = container_of(q, struct frag_queue, q); in ip6frag_init() local
37 fq->ecn = 0; in ip6frag_init()
48 const struct inet_frag_queue *fq = data; in ip6frag_obj_hashfn() local
50 return jhash2((const u32 *)&fq->key.v6, in ip6frag_obj_hashfn()
58 const struct inet_frag_queue *fq = ptr; in ip6frag_obj_cmpfn() local
60 return !!memcmp(&fq->key, key, sizeof(*key)); in ip6frag_obj_cmpfn()
64 ip6frag_expire_frag_queue(struct net *net, struct frag_queue *fq) in ip6frag_expire_frag_queue() argument
70 spin_lock(&fq->q.lock); in ip6frag_expire_frag_queue()
72 if (fq->q.flags & INET_FRAG_COMPLETE) in ip6frag_expire_frag_queue()
75 inet_frag_kill(&fq->q); in ip6frag_expire_frag_queue()
[all …]
Dfq.h68 struct fq { struct
84 typedef struct sk_buff *fq_tin_dequeue_t(struct fq *, argument
88 typedef void fq_skb_free_t(struct fq *,
94 typedef bool fq_skb_filter_t(struct fq *,
100 typedef struct fq_flow *fq_flow_get_default_t(struct fq *,
Dxdp_sock.h30 struct xsk_queue *fq; member
/linux-4.19.296/block/
Dblk-flush.c97 struct blk_flush_queue *fq, unsigned int flags);
166 struct blk_flush_queue *fq, in blk_flush_complete_seq() argument
170 struct list_head *pending = &fq->flush_queue[fq->flush_pending_idx]; in blk_flush_complete_seq()
188 fq->flush_pending_since = jiffies; in blk_flush_complete_seq()
193 list_move_tail(&rq->flush.list, &fq->flush_data_in_flight); in blk_flush_complete_seq()
217 kicked = blk_kick_flush(q, fq, cmd_flags); in blk_flush_complete_seq()
228 struct blk_flush_queue *fq = blk_get_flush_queue(q, flush_rq->mq_ctx); in flush_end_io() local
234 spin_lock_irqsave(&fq->mq_flush_lock, flags); in flush_end_io()
237 fq->rq_status = error; in flush_end_io()
238 spin_unlock_irqrestore(&fq->mq_flush_lock, flags); in flush_end_io()
[all …]
Dblk.h118 return blk_mq_map_queue(q, ctx->cpu)->fq; in blk_get_flush_queue()
119 return q->fq; in blk_get_flush_queue()
130 return hctx->fq->flush_rq == req; in is_flush_rq()
Dblk-core.c593 struct blk_flush_queue *fq = blk_get_flush_queue(q, NULL); in __blk_drain_queue() local
598 if (fq) in __blk_drain_queue()
599 drain |= !list_empty(&fq->flush_queue[i]); in __blk_drain_queue()
1174 q->fq = blk_alloc_flush_queue(q, NUMA_NO_NODE, q->cmd_size, GFP_KERNEL); in blk_init_allocated_queue()
1175 if (!q->fq) in blk_init_allocated_queue()
1178 if (q->init_rq_fn && q->init_rq_fn(q, q->fq->flush_rq, GFP_KERNEL)) in blk_init_allocated_queue()
1200 q->exit_rq_fn(q, q->fq->flush_rq); in blk_init_allocated_queue()
1202 blk_free_flush_queue(q->fq); in blk_init_allocated_queue()
1203 q->fq = NULL; in blk_init_allocated_queue()
2823 struct blk_flush_queue *fq = blk_get_flush_queue(q, NULL); in elv_next_request() local
[all …]
Dblk-mq-sysfs.c28 blk_free_flush_queue(hctx->fq); in blk_mq_hw_sysfs_release()
Dblk-sysfs.c849 q->exit_rq_fn(q, q->fq->flush_rq); in __blk_release_queue()
850 blk_free_flush_queue(q->fq); in __blk_release_queue()
Dblk-mq.c2174 set->ops->exit_request(set, hctx->fq->flush_rq, hctx_idx); in blk_mq_exit_hctx()
2238 hctx->fq = blk_alloc_flush_queue(q, hctx->numa_node, set->cmd_size, in blk_mq_init_hctx()
2240 if (!hctx->fq) in blk_mq_init_hctx()
2243 if (blk_mq_init_request(set, hctx->fq->flush_rq, hctx_idx, node)) in blk_mq_init_hctx()
2254 blk_free_flush_queue(hctx->fq); in blk_mq_init_hctx()
/linux-4.19.296/include/soc/fsl/
Dqman.h299 } __packed fq; /* FQRN/FQRNI/FQRL/FQPN */ member
690 struct qman_fq *fq,
697 typedef void (*qman_cb_mr)(struct qman_portal *qm, struct qman_fq *fq,
957 int qman_create_fq(u32 fqid, u32 flags, struct qman_fq *fq);
967 void qman_destroy_fq(struct qman_fq *fq);
973 u32 qman_fq_fqid(struct qman_fq *fq);
1009 int qman_init_fq(struct qman_fq *fq, u32 flags, struct qm_mcc_initfq *opts);
1018 int qman_schedule_fq(struct qman_fq *fq);
1038 int qman_retire_fq(struct qman_fq *fq, u32 *flags);
1047 int qman_oos_fq(struct qman_fq *fq);
[all …]
/linux-4.19.296/include/linux/
Diova.h87 struct iova_fq __percpu *fq; /* Flush Queue */ member
Dblk-mq.h31 struct blk_flush_queue *fq; member
Dblkdev.h641 struct blk_flush_queue *fq; member