Home
last modified time | relevance | path

Searched refs:request (Results 1 – 25 of 298) sorted by relevance

12345678910>>...12

/linux-4.19.296/drivers/s390/char/
Dtape_std.c38 struct tape_request * request = from_timer(request, t, timer); in tape_std_assign_timeout() local
39 struct tape_device * device = request->device; in tape_std_assign_timeout()
46 rc = tape_cancel_io(device, request); in tape_std_assign_timeout()
56 struct tape_request *request; in tape_std_assign() local
58 request = tape_alloc_request(2, 11); in tape_std_assign()
59 if (IS_ERR(request)) in tape_std_assign()
60 return PTR_ERR(request); in tape_std_assign()
62 request->op = TO_ASSIGN; in tape_std_assign()
63 tape_ccw_cc(request->cpaddr, ASSIGN, 11, request->cpdata); in tape_std_assign()
64 tape_ccw_end(request->cpaddr + 1, NOP, 0, NULL); in tape_std_assign()
[all …]
Dtape_34xx.c56 static void __tape_34xx_medium_sense(struct tape_request *request) in __tape_34xx_medium_sense() argument
58 struct tape_device *device = request->device; in __tape_34xx_medium_sense()
61 if (request->rc == 0) { in __tape_34xx_medium_sense()
62 sense = request->cpdata; in __tape_34xx_medium_sense()
81 request->rc); in __tape_34xx_medium_sense()
82 tape_free_request(request); in __tape_34xx_medium_sense()
87 struct tape_request *request; in tape_34xx_medium_sense() local
90 request = tape_alloc_request(1, 32); in tape_34xx_medium_sense()
91 if (IS_ERR(request)) { in tape_34xx_medium_sense()
93 return PTR_ERR(request); in tape_34xx_medium_sense()
[all …]
Dtape_3590.c204 struct tape_request *request; in tape_3592_kekl_query() local
213 request = tape_alloc_request(2, sizeof(*order)); in tape_3592_kekl_query()
214 if (IS_ERR(request)) { in tape_3592_kekl_query()
215 rc = PTR_ERR(request); in tape_3592_kekl_query()
218 order = request->cpdata; in tape_3592_kekl_query()
222 request->op = TO_KEKL_QUERY; in tape_3592_kekl_query()
223 tape_ccw_cc(request->cpaddr, PERF_SUBSYS_FUNC, sizeof(*order), order); in tape_3592_kekl_query()
224 tape_ccw_end(request->cpaddr + 1, READ_SS_DATA, sizeof(*int_kekls), in tape_3592_kekl_query()
226 rc = tape_do_io(device, request); in tape_3592_kekl_query()
233 tape_free_request(request); in tape_3592_kekl_query()
[all …]
Dtape_core.c292 __tape_cancel_io(struct tape_device *device, struct tape_request *request) in __tape_cancel_io() argument
298 if (request->callback == NULL) in __tape_cancel_io()
303 rc = ccw_device_clear(device->cdev, (long) request); in __tape_cancel_io()
307 request->status = TAPE_REQUEST_DONE; in __tape_cancel_io()
310 request->status = TAPE_REQUEST_CANCEL; in __tape_cancel_io()
641 struct tape_request * request; in __tape_discard_requests() local
645 request = list_entry(l, struct tape_request, list); in __tape_discard_requests()
646 if (request->status == TAPE_REQUEST_IN_IO) in __tape_discard_requests()
647 request->status = TAPE_REQUEST_DONE; in __tape_discard_requests()
648 list_del(&request->list); in __tape_discard_requests()
[all …]
Dsclp_vt220.c100 static int __sclp_vt220_emit(struct sclp_vt220_request *request);
121 sclp_vt220_process_queue(struct sclp_vt220_request *request) in sclp_vt220_process_queue() argument
128 page = request->sclp_req.sccb; in sclp_vt220_process_queue()
131 list_del(&request->list); in sclp_vt220_process_queue()
134 request = NULL; in sclp_vt220_process_queue()
136 request = list_entry(sclp_vt220_outqueue.next, in sclp_vt220_process_queue()
138 if (!request || sclp_vt220_suspended) { in sclp_vt220_process_queue()
144 } while (__sclp_vt220_emit(request)); in sclp_vt220_process_queue()
145 if (request == NULL && sclp_vt220_flush_later) in sclp_vt220_process_queue()
157 sclp_vt220_callback(struct sclp_req *request, void *data) in sclp_vt220_callback() argument
[all …]
Dsclp_sdias.c53 static void sdias_callback(struct sclp_req *request, void *data) in sdias_callback() argument
107 struct sclp_req request; in sclp_sdias_blk_count() local
113 memset(&request, 0, sizeof(request)); in sclp_sdias_blk_count()
123 request.sccb = &sccb; in sclp_sdias_blk_count()
124 request.command = SCLP_CMDW_WRITE_EVENT_DATA; in sclp_sdias_blk_count()
125 request.status = SCLP_REQ_FILLED; in sclp_sdias_blk_count()
126 request.callback = sdias_callback; in sclp_sdias_blk_count()
128 rc = sdias_sclp_send(&request); in sclp_sdias_blk_count()
166 struct sclp_req request; in sclp_sdias_copy() local
172 memset(&request, 0, sizeof(request)); in sclp_sdias_copy()
[all …]
Dsclp_async.c26 static struct sclp_req *request; variable
109 request->command = SCLP_CMDW_WRITE_EVENT_DATA; in sclp_async_send_wait()
110 request->sccb = sccb; in sclp_async_send_wait()
111 request->status = SCLP_REQ_FILLED; in sclp_async_send_wait()
122 rc = sclp_add_request(request); in sclp_async_send_wait()
126 while (request->status != SCLP_REQ_DONE && in sclp_async_send_wait()
127 request->status != SCLP_REQ_FAILED) { in sclp_async_send_wait()
131 if (request->status != SCLP_REQ_DONE) in sclp_async_send_wait()
134 request->sccb)->header.response_code; in sclp_async_send_wait()
156 request = kzalloc(sizeof(struct sclp_req), GFP_KERNEL); in sclp_async_init()
[all …]
/linux-4.19.296/include/linux/
Delevator.h26 typedef enum elv_merge (elevator_merge_fn) (struct request_queue *, struct request **,
29 typedef void (elevator_merge_req_fn) (struct request_queue *, struct request *, struct request *);
31 typedef void (elevator_merged_fn) (struct request_queue *, struct request *, enum elv_merge);
34 struct request *, struct bio *);
37 struct request *, struct request *);
40 struct request *, struct bio *);
44 typedef void (elevator_add_req_fn) (struct request_queue *, struct request *);
45 typedef struct request *(elevator_request_list_fn) (struct request_queue *, struct request *);
46 typedef void (elevator_completed_req_fn) (struct request_queue *, struct request *);
51 typedef int (elevator_set_req_fn) (struct request_queue *, struct request *,
[all …]
Dblk-mq.h96 struct request *rq;
104 typedef enum blk_eh_timer_return (timeout_fn)(struct request *, bool);
107 typedef int (init_request_fn)(struct blk_mq_tag_set *set, struct request *,
109 typedef void (exit_request_fn)(struct blk_mq_tag_set *set, struct request *,
112 typedef void (busy_iter_fn)(struct blk_mq_hw_ctx *, struct request *, void *,
114 typedef void (busy_tag_iter_fn)(struct request *, void *, bool);
117 typedef void (cleanup_rq_fn)(struct request *);
167 void (*initialize_rq_fn)(struct request *rq);
182 void (*show_rq)(struct seq_file *m, struct request *rq);
221 void blk_mq_free_request(struct request *rq);
[all …]
Dblkdev.h37 struct request;
67 typedef void (rq_end_io_fn)(struct request *, blk_status_t);
158 struct request { struct
274 struct request *next_rq; argument
291 static inline bool blk_rq_is_scsi(struct request *rq) in blk_rq_is_scsi()
296 static inline bool blk_rq_is_private(struct request *rq) in blk_rq_is_private()
301 static inline bool blk_rq_is_passthrough(struct request *rq) in blk_rq_is_passthrough()
313 static inline unsigned short req_get_ioprio(struct request *req) in req_get_ioprio()
325 typedef int (prep_rq_fn) (struct request_queue *, struct request *);
326 typedef void (unprep_rq_fn) (struct request_queue *, struct request *);
[all …]
/linux-4.19.296/block/
Dblk.h30 struct request *flush_rq;
36 struct request *orig_rq;
128 is_flush_rq(struct request *req, struct blk_mq_hw_ctx *hctx) in is_flush_rq()
141 void blk_rq_bio_prep(struct request_queue *q, struct request *rq,
180 void blk_add_timer(struct request *req);
181 void blk_delete_timer(struct request *);
184 bool bio_attempt_front_merge(struct request_queue *q, struct request *req,
186 bool bio_attempt_back_merge(struct request_queue *q, struct request *req,
188 bool bio_attempt_discard_merge(struct request_queue *q, struct request *req,
192 struct request **same_queue_rq);
[all …]
Ddeadline-iosched.c40 struct request *next_rq[2];
54 deadline_rb_root(struct deadline_data *dd, struct request *rq) in deadline_rb_root()
62 static inline struct request *
63 deadline_latter_request(struct request *rq) in deadline_latter_request()
74 deadline_add_rq_rb(struct deadline_data *dd, struct request *rq) in deadline_add_rq_rb()
82 deadline_del_rq_rb(struct deadline_data *dd, struct request *rq) in deadline_del_rq_rb()
96 deadline_add_request(struct request_queue *q, struct request *rq) in deadline_add_request()
119 static void deadline_remove_request(struct request_queue *q, struct request *rq) in deadline_remove_request()
128 deadline_merge(struct request_queue *q, struct request **req, struct bio *bio) in deadline_merge()
131 struct request *__rq; in deadline_merge()
[all …]
Dnoop-iosched.c15 static void noop_merged_requests(struct request_queue *q, struct request *rq, in noop_merged_requests()
16 struct request *next) in noop_merged_requests()
24 struct request *rq; in noop_dispatch()
26 rq = list_first_entry_or_null(&nd->queue, struct request, queuelist); in noop_dispatch()
35 static void noop_add_request(struct request_queue *q, struct request *rq) in noop_add_request()
42 static struct request *
43 noop_former_request(struct request_queue *q, struct request *rq) in noop_former_request()
52 static struct request *
53 noop_latter_request(struct request_queue *q, struct request *rq) in noop_latter_request()
Dblk-mq-sched.h11 void blk_mq_sched_assign_ioc(struct request *rq, struct bio *bio);
13 void blk_mq_sched_request_inserted(struct request *rq);
15 struct request **merged_request);
17 bool blk_mq_sched_try_insert_merge(struct request_queue *q, struct request *rq);
21 void blk_mq_sched_insert_request(struct request *rq, bool at_head,
42 blk_mq_sched_allow_merge(struct request_queue *q, struct request *rq, in blk_mq_sched_allow_merge()
53 static inline void blk_mq_sched_completed_request(struct request *rq) in blk_mq_sched_completed_request()
61 static inline void blk_mq_sched_started_request(struct request *rq) in blk_mq_sched_started_request()
70 static inline void blk_mq_sched_requeue_request(struct request *rq) in blk_mq_sched_requeue_request()
Dbsg-lib.c44 static int bsg_transport_fill_hdr(struct request *rq, struct sg_io_v4 *hdr, in bsg_transport_fill_hdr()
50 job->request = memdup_user(uptr64(hdr->request), hdr->request_len); in bsg_transport_fill_hdr()
52 return PTR_ERR_OR_ZERO(job->request); in bsg_transport_fill_hdr()
55 static int bsg_transport_complete_rq(struct request *rq, struct sg_io_v4 *hdr) in bsg_transport_complete_rq()
104 static void bsg_transport_free_rq(struct request *rq) in bsg_transport_free_rq()
108 kfree(job->request); in bsg_transport_free_rq()
125 struct request *rq = blk_mq_rq_from_pdu(job); in bsg_teardown_job()
168 static void bsg_softirq_done(struct request *rq) in bsg_softirq_done()
175 static int bsg_map_buffer(struct bsg_buffer *buf, struct request *req) in bsg_map_buffer()
195 static bool bsg_prepare_job(struct device *dev, struct request *req) in bsg_prepare_job()
[all …]
Dblk-rq-qos.h29 void (*track)(struct rq_qos *, struct request *, struct bio *);
30 void (*issue)(struct rq_qos *, struct request *);
31 void (*requeue)(struct rq_qos *, struct request *);
32 void (*done)(struct rq_qos *, struct request *);
99 void rq_qos_done(struct request_queue *, struct request *);
100 void rq_qos_issue(struct request_queue *, struct request *);
101 void rq_qos_requeue(struct request_queue *, struct request *);
104 void rq_qos_track(struct request_queue *q, struct request *, struct bio *);
Dmq-deadline.c49 struct request *next_rq[2];
67 deadline_rb_root(struct deadline_data *dd, struct request *rq) in deadline_rb_root()
75 static inline struct request *
76 deadline_latter_request(struct request *rq) in deadline_latter_request()
87 deadline_add_rq_rb(struct deadline_data *dd, struct request *rq) in deadline_add_rq_rb()
95 deadline_del_rq_rb(struct deadline_data *dd, struct request *rq) in deadline_del_rq_rb()
108 static void deadline_remove_request(struct request_queue *q, struct request *rq) in deadline_remove_request()
125 static void dd_request_merged(struct request_queue *q, struct request *req, in dd_request_merged()
139 static void dd_merged_requests(struct request_queue *q, struct request *req, in dd_merged_requests()
140 struct request *next) in dd_merged_requests()
[all …]
/linux-4.19.296/drivers/s390/net/
Dism_drv.c63 cmd.request.hdr.cmd = cmd_code; in ism_cmd_simple()
64 cmd.request.hdr.len = sizeof(cmd.request); in ism_cmd_simple()
74 cmd.request.hdr.cmd = ISM_QUERY_INFO; in query_info()
75 cmd.request.hdr.len = sizeof(cmd.request); in query_info()
98 cmd.request.hdr.cmd = ISM_REG_SBA; in register_sba()
99 cmd.request.hdr.len = sizeof(cmd.request); in register_sba()
100 cmd.request.sba = dma_handle; in register_sba()
125 cmd.request.hdr.cmd = ISM_REG_IEQ; in register_ieq()
126 cmd.request.hdr.len = sizeof(cmd.request); in register_ieq()
127 cmd.request.ieq = dma_handle; in register_ieq()
[all …]
Dism.h52 } request; member
63 } request; member
72 } request; member
82 } request; member
107 } request; member
122 } request; member
136 } request; member
146 } request; member
155 } request; member
165 } request; member
/linux-4.19.296/drivers/thunderbolt/
Dicm.c177 const struct icm_pkg_header *req_hdr = req->request; in icm_match()
200 static int icm_request(struct tb *tb, const void *request, size_t request_size, in icm_request() argument
217 req->request = request; in icm_request()
259 struct icm_fr_pkg_get_topology request = { in icm_fr_get_route() local
270 ret = icm_request(tb, &request, sizeof(request), switches, in icm_fr_get_route()
317 struct icm_pkg_driver_ready request = { in icm_fr_driver_ready() local
323 ret = icm_request(tb, &request, sizeof(request), &reply, sizeof(reply), in icm_fr_driver_ready()
336 struct icm_fr_pkg_approve_device request; in icm_fr_approve_switch() local
340 memset(&request, 0, sizeof(request)); in icm_fr_approve_switch()
341 memcpy(&request.ep_uuid, sw->uuid, sizeof(request.ep_uuid)); in icm_fr_approve_switch()
[all …]
/linux-4.19.296/drivers/s390/cio/
Dchsc_sch.c56 struct chsc_request *request = private->request; in chsc_subchannel_irq() local
64 if (!request) { in chsc_subchannel_irq()
69 private->request = NULL; in chsc_subchannel_irq()
70 memcpy(&request->irb, irb, sizeof(*irb)); in chsc_subchannel_irq()
72 complete(&request->completion); in chsc_subchannel_irq()
110 if (private->request) { in chsc_subchannel_remove()
111 complete(&private->request->completion); in chsc_subchannel_remove()
238 struct chsc_request *request) in chsc_async() argument
250 if (private->request) { in chsc_async()
268 private->request = request; in chsc_async()
[all …]
/linux-4.19.296/drivers/isdn/hardware/eicon/
Ddadapter.c112 if (d->request) { in diva_didd_add_descriptor()
113 MAdapter.request = d->request; in diva_didd_add_descriptor()
114 dprintf = (DIVA_DI_PRINTF)d->request; in diva_didd_add_descriptor()
116 DBG_TRC(("DIMAINT registered, dprintf=%08x", d->request)) in diva_didd_add_descriptor()
120 MAdapter.request = (IDI_CALL)no_printf; in diva_didd_add_descriptor()
132 DBG_TRC(("Add adapter[%d], request=%08x", (i + 1), d->request)) in diva_didd_add_descriptor()
145 static int diva_didd_remove_descriptor(IDI_CALL request) { in diva_didd_remove_descriptor() argument
148 if (request == MAdapter.request) { in diva_didd_remove_descriptor()
152 MAdapter.request = (IDI_CALL)no_printf; in diva_didd_remove_descriptor()
156 if (HandleTable[i].request == request) { in diva_didd_remove_descriptor()
[all …]
/linux-4.19.296/drivers/reset/tegra/
Dreset-bpmp.c24 struct mrq_reset_request request; in tegra_bpmp_reset_common() local
28 memset(&request, 0, sizeof(request)); in tegra_bpmp_reset_common()
29 request.cmd = command; in tegra_bpmp_reset_common()
30 request.reset_id = id; in tegra_bpmp_reset_common()
34 msg.tx.data = &request; in tegra_bpmp_reset_common()
35 msg.tx.size = sizeof(request); in tegra_bpmp_reset_common()
/linux-4.19.296/drivers/clk/tegra/
Dclk-bpmp.c65 struct mrq_clk_request request; in tegra_bpmp_clk_transfer() local
67 void *req = &request; in tegra_bpmp_clk_transfer()
70 memset(&request, 0, sizeof(request)); in tegra_bpmp_clk_transfer()
71 request.cmd_and_id = (clk->cmd << 24) | clk->id; in tegra_bpmp_clk_transfer()
84 msg.tx.data = &request; in tegra_bpmp_clk_transfer()
85 msg.tx.size = sizeof(request); in tegra_bpmp_clk_transfer()
151 struct cmd_clk_get_rate_request request; in tegra_bpmp_clk_recalc_rate() local
158 msg.tx.data = &request; in tegra_bpmp_clk_recalc_rate()
159 msg.tx.size = sizeof(request); in tegra_bpmp_clk_recalc_rate()
175 struct cmd_clk_round_rate_request request; in tegra_bpmp_clk_round_rate() local
[all …]
/linux-4.19.296/drivers/s390/block/
Dscm_blk.c47 kfree(scmrq->request); in __scm_free_rq()
81 scmrq->request = kcalloc(nr_requests_per_io, sizeof(scmrq->request[0]), in __scm_alloc_rq()
83 if (!scmrq->request) in __scm_alloc_rq()
132 for (i = 0; i < nr_requests_per_io && scmrq->request[i]; i++) { in scm_request_done()
146 static bool scm_permit_request(struct scm_blk_dev *bdev, struct request *req) in scm_permit_request()
183 int pos = scmrq->aob->request.msb_count; in scm_request_prepare()
185 struct request *req = scmrq->request[pos]; in scm_request_prepare()
195 scmrq->aob->request.msb_count++; in scm_request_prepare()
213 struct request *req) in scm_request_set()
215 scmrq->request[scmrq->aob->request.msb_count] = req; in scm_request_set()
[all …]

12345678910>>...12