/linux-4.19.296/include/net/ |
D | fq_impl.h | 15 static void fq_adjust_removal(struct fq *fq, in fq_adjust_removal() argument 24 fq->backlog--; in fq_adjust_removal() 25 fq->memory_usage -= skb->truesize; in fq_adjust_removal() 28 static void fq_rejigger_backlog(struct fq *fq, struct fq_flow *flow) in fq_rejigger_backlog() argument 37 list_for_each_entry_continue(i, &fq->backlogs, backlogchain) in fq_rejigger_backlog() 46 static struct sk_buff *fq_flow_dequeue(struct fq *fq, in fq_flow_dequeue() argument 51 lockdep_assert_held(&fq->lock); in fq_flow_dequeue() 57 fq_adjust_removal(fq, flow, skb); in fq_flow_dequeue() 58 fq_rejigger_backlog(fq, flow); in fq_flow_dequeue() 63 static struct sk_buff *fq_tin_dequeue(struct fq *fq, in fq_tin_dequeue() argument [all …]
|
D | ipv6_frag.h | 33 struct frag_queue *fq = container_of(q, struct frag_queue, q); in ip6frag_init() local 37 fq->ecn = 0; in ip6frag_init() 48 const struct inet_frag_queue *fq = data; in ip6frag_obj_hashfn() local 50 return jhash2((const u32 *)&fq->key.v6, in ip6frag_obj_hashfn() 58 const struct inet_frag_queue *fq = ptr; in ip6frag_obj_cmpfn() local 60 return !!memcmp(&fq->key, key, sizeof(*key)); in ip6frag_obj_cmpfn() 64 ip6frag_expire_frag_queue(struct net *net, struct frag_queue *fq) in ip6frag_expire_frag_queue() argument 70 spin_lock(&fq->q.lock); in ip6frag_expire_frag_queue() 72 if (fq->q.flags & INET_FRAG_COMPLETE) in ip6frag_expire_frag_queue() 75 inet_frag_kill(&fq->q); in ip6frag_expire_frag_queue() [all …]
|
D | fq.h | 68 struct fq { struct 84 typedef struct sk_buff *fq_tin_dequeue_t(struct fq *, argument 88 typedef void fq_skb_free_t(struct fq *, 94 typedef bool fq_skb_filter_t(struct fq *, 100 typedef struct fq_flow *fq_flow_get_default_t(struct fq *,
|
D | xdp_sock.h | 30 struct xsk_queue *fq; member
|
/linux-4.19.296/block/ |
D | blk-flush.c | 97 struct blk_flush_queue *fq, unsigned int flags); 166 struct blk_flush_queue *fq, in blk_flush_complete_seq() argument 170 struct list_head *pending = &fq->flush_queue[fq->flush_pending_idx]; in blk_flush_complete_seq() 188 fq->flush_pending_since = jiffies; in blk_flush_complete_seq() 193 list_move_tail(&rq->flush.list, &fq->flush_data_in_flight); in blk_flush_complete_seq() 217 kicked = blk_kick_flush(q, fq, cmd_flags); in blk_flush_complete_seq() 228 struct blk_flush_queue *fq = blk_get_flush_queue(q, flush_rq->mq_ctx); in flush_end_io() local 234 spin_lock_irqsave(&fq->mq_flush_lock, flags); in flush_end_io() 237 fq->rq_status = error; in flush_end_io() 238 spin_unlock_irqrestore(&fq->mq_flush_lock, flags); in flush_end_io() [all …]
|
D | blk.h | 118 return blk_mq_map_queue(q, ctx->cpu)->fq; in blk_get_flush_queue() 119 return q->fq; in blk_get_flush_queue() 130 return hctx->fq->flush_rq == req; in is_flush_rq()
|
D | blk-core.c | 593 struct blk_flush_queue *fq = blk_get_flush_queue(q, NULL); in __blk_drain_queue() local 598 if (fq) in __blk_drain_queue() 599 drain |= !list_empty(&fq->flush_queue[i]); in __blk_drain_queue() 1174 q->fq = blk_alloc_flush_queue(q, NUMA_NO_NODE, q->cmd_size, GFP_KERNEL); in blk_init_allocated_queue() 1175 if (!q->fq) in blk_init_allocated_queue() 1178 if (q->init_rq_fn && q->init_rq_fn(q, q->fq->flush_rq, GFP_KERNEL)) in blk_init_allocated_queue() 1200 q->exit_rq_fn(q, q->fq->flush_rq); in blk_init_allocated_queue() 1202 blk_free_flush_queue(q->fq); in blk_init_allocated_queue() 1203 q->fq = NULL; in blk_init_allocated_queue() 2823 struct blk_flush_queue *fq = blk_get_flush_queue(q, NULL); in elv_next_request() local [all …]
|
D | blk-mq-sysfs.c | 28 blk_free_flush_queue(hctx->fq); in blk_mq_hw_sysfs_release()
|
D | blk-sysfs.c | 849 q->exit_rq_fn(q, q->fq->flush_rq); in __blk_release_queue() 850 blk_free_flush_queue(q->fq); in __blk_release_queue()
|
D | blk-mq.c | 2174 set->ops->exit_request(set, hctx->fq->flush_rq, hctx_idx); in blk_mq_exit_hctx() 2238 hctx->fq = blk_alloc_flush_queue(q, hctx->numa_node, set->cmd_size, in blk_mq_init_hctx() 2240 if (!hctx->fq) in blk_mq_init_hctx() 2243 if (blk_mq_init_request(set, hctx->fq->flush_rq, hctx_idx, node)) in blk_mq_init_hctx() 2254 blk_free_flush_queue(hctx->fq); in blk_mq_init_hctx()
|
/linux-4.19.296/include/soc/fsl/ |
D | qman.h | 299 } __packed fq; /* FQRN/FQRNI/FQRL/FQPN */ member 690 struct qman_fq *fq, 697 typedef void (*qman_cb_mr)(struct qman_portal *qm, struct qman_fq *fq, 957 int qman_create_fq(u32 fqid, u32 flags, struct qman_fq *fq); 967 void qman_destroy_fq(struct qman_fq *fq); 973 u32 qman_fq_fqid(struct qman_fq *fq); 1009 int qman_init_fq(struct qman_fq *fq, u32 flags, struct qm_mcc_initfq *opts); 1018 int qman_schedule_fq(struct qman_fq *fq); 1038 int qman_retire_fq(struct qman_fq *fq, u32 *flags); 1047 int qman_oos_fq(struct qman_fq *fq); [all …]
|
/linux-4.19.296/include/linux/ |
D | iova.h | 87 struct iova_fq __percpu *fq; /* Flush Queue */ member
|
D | blk-mq.h | 31 struct blk_flush_queue *fq; member
|
D | blkdev.h | 641 struct blk_flush_queue *fq; member
|