Merge branch 'work.iov_iter' of git://git.kernel.org/pub/scm/linux/kernel/git/viro/vfs
[linux-2.6-microblaze.git] / drivers / s390 / cio / vfio_ccw_fsm.c
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * Finite state machine for vfio-ccw device handling
4  *
5  * Copyright IBM Corp. 2017
6  * Copyright Red Hat, Inc. 2019
7  *
8  * Author(s): Dong Jia Shi <bjsdjshi@linux.vnet.ibm.com>
9  *            Cornelia Huck <cohuck@redhat.com>
10  */
11
12 #include <linux/vfio.h>
13 #include <linux/mdev.h>
14
15 #include "ioasm.h"
16 #include "vfio_ccw_private.h"
17
18 static int fsm_io_helper(struct vfio_ccw_private *private)
19 {
20         struct subchannel *sch;
21         union orb *orb;
22         int ccode;
23         __u8 lpm;
24         unsigned long flags;
25         int ret;
26
27         sch = private->sch;
28
29         spin_lock_irqsave(sch->lock, flags);
30
31         orb = cp_get_orb(&private->cp, (u32)(addr_t)sch, sch->lpm);
32         if (!orb) {
33                 ret = -EIO;
34                 goto out;
35         }
36
37         VFIO_CCW_TRACE_EVENT(5, "stIO");
38         VFIO_CCW_TRACE_EVENT(5, dev_name(&sch->dev));
39
40         /* Issue "Start Subchannel" */
41         ccode = ssch(sch->schid, orb);
42
43         VFIO_CCW_HEX_EVENT(5, &ccode, sizeof(ccode));
44
45         switch (ccode) {
46         case 0:
47                 /*
48                  * Initialize device status information
49                  */
50                 sch->schib.scsw.cmd.actl |= SCSW_ACTL_START_PEND;
51                 ret = 0;
52                 private->state = VFIO_CCW_STATE_CP_PENDING;
53                 break;
54         case 1:         /* Status pending */
55         case 2:         /* Busy */
56                 ret = -EBUSY;
57                 break;
58         case 3:         /* Device/path not operational */
59         {
60                 lpm = orb->cmd.lpm;
61                 if (lpm != 0)
62                         sch->lpm &= ~lpm;
63                 else
64                         sch->lpm = 0;
65
66                 if (cio_update_schib(sch))
67                         ret = -ENODEV;
68                 else
69                         ret = sch->lpm ? -EACCES : -ENODEV;
70                 break;
71         }
72         default:
73                 ret = ccode;
74         }
75 out:
76         spin_unlock_irqrestore(sch->lock, flags);
77         return ret;
78 }
79
80 static int fsm_do_halt(struct vfio_ccw_private *private)
81 {
82         struct subchannel *sch;
83         unsigned long flags;
84         int ccode;
85         int ret;
86
87         sch = private->sch;
88
89         spin_lock_irqsave(sch->lock, flags);
90
91         VFIO_CCW_TRACE_EVENT(2, "haltIO");
92         VFIO_CCW_TRACE_EVENT(2, dev_name(&sch->dev));
93
94         /* Issue "Halt Subchannel" */
95         ccode = hsch(sch->schid);
96
97         VFIO_CCW_HEX_EVENT(2, &ccode, sizeof(ccode));
98
99         switch (ccode) {
100         case 0:
101                 /*
102                  * Initialize device status information
103                  */
104                 sch->schib.scsw.cmd.actl |= SCSW_ACTL_HALT_PEND;
105                 ret = 0;
106                 break;
107         case 1:         /* Status pending */
108         case 2:         /* Busy */
109                 ret = -EBUSY;
110                 break;
111         case 3:         /* Device not operational */
112                 ret = -ENODEV;
113                 break;
114         default:
115                 ret = ccode;
116         }
117         spin_unlock_irqrestore(sch->lock, flags);
118         return ret;
119 }
120
121 static int fsm_do_clear(struct vfio_ccw_private *private)
122 {
123         struct subchannel *sch;
124         unsigned long flags;
125         int ccode;
126         int ret;
127
128         sch = private->sch;
129
130         spin_lock_irqsave(sch->lock, flags);
131
132         VFIO_CCW_TRACE_EVENT(2, "clearIO");
133         VFIO_CCW_TRACE_EVENT(2, dev_name(&sch->dev));
134
135         /* Issue "Clear Subchannel" */
136         ccode = csch(sch->schid);
137
138         VFIO_CCW_HEX_EVENT(2, &ccode, sizeof(ccode));
139
140         switch (ccode) {
141         case 0:
142                 /*
143                  * Initialize device status information
144                  */
145                 sch->schib.scsw.cmd.actl = SCSW_ACTL_CLEAR_PEND;
146                 /* TODO: check what else we might need to clear */
147                 ret = 0;
148                 break;
149         case 3:         /* Device not operational */
150                 ret = -ENODEV;
151                 break;
152         default:
153                 ret = ccode;
154         }
155         spin_unlock_irqrestore(sch->lock, flags);
156         return ret;
157 }
158
159 static void fsm_notoper(struct vfio_ccw_private *private,
160                         enum vfio_ccw_event event)
161 {
162         struct subchannel *sch = private->sch;
163
164         VFIO_CCW_TRACE_EVENT(2, "notoper");
165         VFIO_CCW_TRACE_EVENT(2, dev_name(&sch->dev));
166
167         /*
168          * TODO:
169          * Probably we should send the machine check to the guest.
170          */
171         css_sched_sch_todo(sch, SCH_TODO_UNREG);
172         private->state = VFIO_CCW_STATE_NOT_OPER;
173 }
174
175 /*
176  * No operation action.
177  */
178 static void fsm_nop(struct vfio_ccw_private *private,
179                     enum vfio_ccw_event event)
180 {
181 }
182
183 static void fsm_io_error(struct vfio_ccw_private *private,
184                          enum vfio_ccw_event event)
185 {
186         pr_err("vfio-ccw: FSM: I/O request from state:%d\n", private->state);
187         private->io_region->ret_code = -EIO;
188 }
189
190 static void fsm_io_busy(struct vfio_ccw_private *private,
191                         enum vfio_ccw_event event)
192 {
193         private->io_region->ret_code = -EBUSY;
194 }
195
196 static void fsm_io_retry(struct vfio_ccw_private *private,
197                          enum vfio_ccw_event event)
198 {
199         private->io_region->ret_code = -EAGAIN;
200 }
201
202 static void fsm_async_error(struct vfio_ccw_private *private,
203                             enum vfio_ccw_event event)
204 {
205         struct ccw_cmd_region *cmd_region = private->cmd_region;
206
207         pr_err("vfio-ccw: FSM: %s request from state:%d\n",
208                cmd_region->command == VFIO_CCW_ASYNC_CMD_HSCH ? "halt" :
209                cmd_region->command == VFIO_CCW_ASYNC_CMD_CSCH ? "clear" :
210                "<unknown>", private->state);
211         cmd_region->ret_code = -EIO;
212 }
213
214 static void fsm_async_retry(struct vfio_ccw_private *private,
215                             enum vfio_ccw_event event)
216 {
217         private->cmd_region->ret_code = -EAGAIN;
218 }
219
220 static void fsm_disabled_irq(struct vfio_ccw_private *private,
221                              enum vfio_ccw_event event)
222 {
223         struct subchannel *sch = private->sch;
224
225         /*
226          * An interrupt in a disabled state means a previous disable was not
227          * successful - should not happen, but we try to disable again.
228          */
229         cio_disable_subchannel(sch);
230 }
231 inline struct subchannel_id get_schid(struct vfio_ccw_private *p)
232 {
233         return p->sch->schid;
234 }
235
236 /*
237  * Deal with the ccw command request from the userspace.
238  */
239 static void fsm_io_request(struct vfio_ccw_private *private,
240                            enum vfio_ccw_event event)
241 {
242         union orb *orb;
243         union scsw *scsw = &private->scsw;
244         struct ccw_io_region *io_region = private->io_region;
245         struct mdev_device *mdev = private->mdev;
246         char *errstr = "request";
247         struct subchannel_id schid = get_schid(private);
248
249         private->state = VFIO_CCW_STATE_CP_PROCESSING;
250         memcpy(scsw, io_region->scsw_area, sizeof(*scsw));
251
252         if (scsw->cmd.fctl & SCSW_FCTL_START_FUNC) {
253                 orb = (union orb *)io_region->orb_area;
254
255                 /* Don't try to build a cp if transport mode is specified. */
256                 if (orb->tm.b) {
257                         io_region->ret_code = -EOPNOTSUPP;
258                         VFIO_CCW_MSG_EVENT(2,
259                                            "%pUl (%x.%x.%04x): transport mode\n",
260                                            mdev_uuid(mdev), schid.cssid,
261                                            schid.ssid, schid.sch_no);
262                         errstr = "transport mode";
263                         goto err_out;
264                 }
265                 io_region->ret_code = cp_init(&private->cp, mdev_dev(mdev),
266                                               orb);
267                 if (io_region->ret_code) {
268                         VFIO_CCW_MSG_EVENT(2,
269                                            "%pUl (%x.%x.%04x): cp_init=%d\n",
270                                            mdev_uuid(mdev), schid.cssid,
271                                            schid.ssid, schid.sch_no,
272                                            io_region->ret_code);
273                         errstr = "cp init";
274                         goto err_out;
275                 }
276
277                 io_region->ret_code = cp_prefetch(&private->cp);
278                 if (io_region->ret_code) {
279                         VFIO_CCW_MSG_EVENT(2,
280                                            "%pUl (%x.%x.%04x): cp_prefetch=%d\n",
281                                            mdev_uuid(mdev), schid.cssid,
282                                            schid.ssid, schid.sch_no,
283                                            io_region->ret_code);
284                         errstr = "cp prefetch";
285                         cp_free(&private->cp);
286                         goto err_out;
287                 }
288
289                 /* Start channel program and wait for I/O interrupt. */
290                 io_region->ret_code = fsm_io_helper(private);
291                 if (io_region->ret_code) {
292                         VFIO_CCW_MSG_EVENT(2,
293                                            "%pUl (%x.%x.%04x): fsm_io_helper=%d\n",
294                                            mdev_uuid(mdev), schid.cssid,
295                                            schid.ssid, schid.sch_no,
296                                            io_region->ret_code);
297                         errstr = "cp fsm_io_helper";
298                         cp_free(&private->cp);
299                         goto err_out;
300                 }
301                 return;
302         } else if (scsw->cmd.fctl & SCSW_FCTL_HALT_FUNC) {
303                 VFIO_CCW_MSG_EVENT(2,
304                                    "%pUl (%x.%x.%04x): halt on io_region\n",
305                                    mdev_uuid(mdev), schid.cssid,
306                                    schid.ssid, schid.sch_no);
307                 /* halt is handled via the async cmd region */
308                 io_region->ret_code = -EOPNOTSUPP;
309                 goto err_out;
310         } else if (scsw->cmd.fctl & SCSW_FCTL_CLEAR_FUNC) {
311                 VFIO_CCW_MSG_EVENT(2,
312                                    "%pUl (%x.%x.%04x): clear on io_region\n",
313                                    mdev_uuid(mdev), schid.cssid,
314                                    schid.ssid, schid.sch_no);
315                 /* clear is handled via the async cmd region */
316                 io_region->ret_code = -EOPNOTSUPP;
317                 goto err_out;
318         }
319
320 err_out:
321         private->state = VFIO_CCW_STATE_IDLE;
322         trace_vfio_ccw_fsm_io_request(scsw->cmd.fctl, schid,
323                                       io_region->ret_code, errstr);
324 }
325
326 /*
327  * Deal with an async request from userspace.
328  */
329 static void fsm_async_request(struct vfio_ccw_private *private,
330                               enum vfio_ccw_event event)
331 {
332         struct ccw_cmd_region *cmd_region = private->cmd_region;
333
334         switch (cmd_region->command) {
335         case VFIO_CCW_ASYNC_CMD_HSCH:
336                 cmd_region->ret_code = fsm_do_halt(private);
337                 break;
338         case VFIO_CCW_ASYNC_CMD_CSCH:
339                 cmd_region->ret_code = fsm_do_clear(private);
340                 break;
341         default:
342                 /* should not happen? */
343                 cmd_region->ret_code = -EINVAL;
344         }
345
346         trace_vfio_ccw_fsm_async_request(get_schid(private),
347                                          cmd_region->command,
348                                          cmd_region->ret_code);
349 }
350
351 /*
352  * Got an interrupt for a normal io (state busy).
353  */
354 static void fsm_irq(struct vfio_ccw_private *private,
355                     enum vfio_ccw_event event)
356 {
357         struct irb *irb = this_cpu_ptr(&cio_irb);
358
359         VFIO_CCW_TRACE_EVENT(6, "IRQ");
360         VFIO_CCW_TRACE_EVENT(6, dev_name(&private->sch->dev));
361
362         memcpy(&private->irb, irb, sizeof(*irb));
363
364         queue_work(vfio_ccw_work_q, &private->io_work);
365
366         if (private->completion)
367                 complete(private->completion);
368 }
369
370 /*
371  * Device statemachine
372  */
373 fsm_func_t *vfio_ccw_jumptable[NR_VFIO_CCW_STATES][NR_VFIO_CCW_EVENTS] = {
374         [VFIO_CCW_STATE_NOT_OPER] = {
375                 [VFIO_CCW_EVENT_NOT_OPER]       = fsm_nop,
376                 [VFIO_CCW_EVENT_IO_REQ]         = fsm_io_error,
377                 [VFIO_CCW_EVENT_ASYNC_REQ]      = fsm_async_error,
378                 [VFIO_CCW_EVENT_INTERRUPT]      = fsm_disabled_irq,
379         },
380         [VFIO_CCW_STATE_STANDBY] = {
381                 [VFIO_CCW_EVENT_NOT_OPER]       = fsm_notoper,
382                 [VFIO_CCW_EVENT_IO_REQ]         = fsm_io_error,
383                 [VFIO_CCW_EVENT_ASYNC_REQ]      = fsm_async_error,
384                 [VFIO_CCW_EVENT_INTERRUPT]      = fsm_irq,
385         },
386         [VFIO_CCW_STATE_IDLE] = {
387                 [VFIO_CCW_EVENT_NOT_OPER]       = fsm_notoper,
388                 [VFIO_CCW_EVENT_IO_REQ]         = fsm_io_request,
389                 [VFIO_CCW_EVENT_ASYNC_REQ]      = fsm_async_request,
390                 [VFIO_CCW_EVENT_INTERRUPT]      = fsm_irq,
391         },
392         [VFIO_CCW_STATE_CP_PROCESSING] = {
393                 [VFIO_CCW_EVENT_NOT_OPER]       = fsm_notoper,
394                 [VFIO_CCW_EVENT_IO_REQ]         = fsm_io_retry,
395                 [VFIO_CCW_EVENT_ASYNC_REQ]      = fsm_async_retry,
396                 [VFIO_CCW_EVENT_INTERRUPT]      = fsm_irq,
397         },
398         [VFIO_CCW_STATE_CP_PENDING] = {
399                 [VFIO_CCW_EVENT_NOT_OPER]       = fsm_notoper,
400                 [VFIO_CCW_EVENT_IO_REQ]         = fsm_io_busy,
401                 [VFIO_CCW_EVENT_ASYNC_REQ]      = fsm_async_request,
402                 [VFIO_CCW_EVENT_INTERRUPT]      = fsm_irq,
403         },
404 };