1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * Copyright IBM Corp. 2016, 2023
4  * Author(s): Martin Schwidefsky <schwidefsky@de.ibm.com>
5  *
6  * Adjunct processor bus, queue related code.
7  */
8 
9 #define KMSG_COMPONENT "ap"
10 #define pr_fmt(fmt) KMSG_COMPONENT ": " fmt
11 
12 #include <linux/init.h>
13 #include <linux/slab.h>
14 #include <asm/facility.h>
15 
16 #include "ap_bus.h"
17 #include "ap_debug.h"
18 
19 static void __ap_flush_queue(struct ap_queue *aq);
20 
21 /*
22  * some AP queue helper functions
23  */
24 
ap_q_supports_bind(struct ap_queue * aq)25 static inline bool ap_q_supports_bind(struct ap_queue *aq)
26 {
27 	return ap_test_bit(&aq->card->functions, AP_FUNC_EP11) ||
28 		ap_test_bit(&aq->card->functions, AP_FUNC_ACCEL);
29 }
30 
ap_q_supports_assoc(struct ap_queue * aq)31 static inline bool ap_q_supports_assoc(struct ap_queue *aq)
32 {
33 	return ap_test_bit(&aq->card->functions, AP_FUNC_EP11);
34 }
35 
36 /**
37  * ap_queue_enable_irq(): Enable interrupt support on this AP queue.
38  * @aq: The AP queue
39  * @ind: the notification indicator byte
40  *
41  * Enables interruption on AP queue via ap_aqic(). Based on the return
42  * value it waits a while and tests the AP queue if interrupts
43  * have been switched on using ap_test_queue().
44  */
ap_queue_enable_irq(struct ap_queue * aq,void * ind)45 static int ap_queue_enable_irq(struct ap_queue *aq, void *ind)
46 {
47 	union ap_qirq_ctrl qirqctrl = { .value = 0 };
48 	struct ap_queue_status status;
49 
50 	qirqctrl.ir = 1;
51 	qirqctrl.isc = AP_ISC;
52 	status = ap_aqic(aq->qid, qirqctrl, virt_to_phys(ind));
53 	if (status.async)
54 		return -EPERM;
55 	switch (status.response_code) {
56 	case AP_RESPONSE_NORMAL:
57 	case AP_RESPONSE_OTHERWISE_CHANGED:
58 		return 0;
59 	case AP_RESPONSE_Q_NOT_AVAIL:
60 	case AP_RESPONSE_DECONFIGURED:
61 	case AP_RESPONSE_CHECKSTOPPED:
62 	case AP_RESPONSE_INVALID_ADDRESS:
63 		pr_err("Registering adapter interrupts for AP device %02x.%04x failed\n",
64 		       AP_QID_CARD(aq->qid),
65 		       AP_QID_QUEUE(aq->qid));
66 		return -EOPNOTSUPP;
67 	case AP_RESPONSE_RESET_IN_PROGRESS:
68 	case AP_RESPONSE_BUSY:
69 	default:
70 		return -EBUSY;
71 	}
72 }
73 
74 /**
75  * __ap_send(): Send message to adjunct processor queue.
76  * @qid: The AP queue number
77  * @psmid: The program supplied message identifier
78  * @msg: The message text
79  * @msglen: The message length
80  * @special: Special Bit
81  *
82  * Returns AP queue status structure.
83  * Condition code 1 on NQAP can't happen because the L bit is 1.
84  * Condition code 2 on NQAP also means the send is incomplete,
85  * because a segment boundary was reached. The NQAP is repeated.
86  */
87 static inline struct ap_queue_status
__ap_send(ap_qid_t qid,unsigned long psmid,void * msg,size_t msglen,int special)88 __ap_send(ap_qid_t qid, unsigned long psmid, void *msg, size_t msglen,
89 	  int special)
90 {
91 	if (special)
92 		qid |= 0x400000UL;
93 	return ap_nqap(qid, psmid, msg, msglen);
94 }
95 
96 /* State machine definitions and helpers */
97 
ap_sm_nop(struct ap_queue * aq)98 static enum ap_sm_wait ap_sm_nop(struct ap_queue *aq)
99 {
100 	return AP_SM_WAIT_NONE;
101 }
102 
103 /**
104  * ap_sm_recv(): Receive pending reply messages from an AP queue but do
105  *	not change the state of the device.
106  * @aq: pointer to the AP queue
107  *
108  * Returns AP_SM_WAIT_NONE, AP_SM_WAIT_AGAIN, or AP_SM_WAIT_INTERRUPT
109  */
ap_sm_recv(struct ap_queue * aq)110 static struct ap_queue_status ap_sm_recv(struct ap_queue *aq)
111 {
112 	struct ap_queue_status status;
113 	struct ap_message *ap_msg;
114 	bool found = false;
115 	size_t reslen;
116 	unsigned long resgr0 = 0;
117 	int parts = 0;
118 
119 	/*
120 	 * DQAP loop until response code and resgr0 indicate that
121 	 * the msg is totally received. As we use the very same buffer
122 	 * the msg is overwritten with each invocation. That's intended
123 	 * and the receiver of the msg is informed with a msg rc code
124 	 * of EMSGSIZE in such a case.
125 	 */
126 	do {
127 		status = ap_dqap(aq->qid, &aq->reply->psmid,
128 				 aq->reply->msg, aq->reply->bufsize,
129 				 &aq->reply->len, &reslen, &resgr0);
130 		parts++;
131 	} while (status.response_code == 0xFF && resgr0 != 0);
132 
133 	switch (status.response_code) {
134 	case AP_RESPONSE_NORMAL:
135 		aq->queue_count = max_t(int, 0, aq->queue_count - 1);
136 		if (!status.queue_empty && !aq->queue_count)
137 			aq->queue_count++;
138 		if (aq->queue_count > 0)
139 			mod_timer(&aq->timeout,
140 				  jiffies + aq->request_timeout);
141 		list_for_each_entry(ap_msg, &aq->pendingq, list) {
142 			if (ap_msg->psmid != aq->reply->psmid)
143 				continue;
144 			list_del_init(&ap_msg->list);
145 			aq->pendingq_count--;
146 			if (parts > 1) {
147 				ap_msg->rc = -EMSGSIZE;
148 				ap_msg->receive(aq, ap_msg, NULL);
149 			} else {
150 				ap_msg->receive(aq, ap_msg, aq->reply);
151 			}
152 			found = true;
153 			break;
154 		}
155 		if (!found) {
156 			AP_DBF_WARN("%s unassociated reply psmid=0x%016lx on 0x%02x.%04x\n",
157 				    __func__, aq->reply->psmid,
158 				    AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid));
159 		}
160 		fallthrough;
161 	case AP_RESPONSE_NO_PENDING_REPLY:
162 		if (!status.queue_empty || aq->queue_count <= 0)
163 			break;
164 		/* The card shouldn't forget requests but who knows. */
165 		aq->queue_count = 0;
166 		list_splice_init(&aq->pendingq, &aq->requestq);
167 		aq->requestq_count += aq->pendingq_count;
168 		aq->pendingq_count = 0;
169 		break;
170 	default:
171 		break;
172 	}
173 	return status;
174 }
175 
176 /**
177  * ap_sm_read(): Receive pending reply messages from an AP queue.
178  * @aq: pointer to the AP queue
179  *
180  * Returns AP_SM_WAIT_NONE, AP_SM_WAIT_AGAIN, or AP_SM_WAIT_INTERRUPT
181  */
ap_sm_read(struct ap_queue * aq)182 static enum ap_sm_wait ap_sm_read(struct ap_queue *aq)
183 {
184 	struct ap_queue_status status;
185 
186 	if (!aq->reply)
187 		return AP_SM_WAIT_NONE;
188 	status = ap_sm_recv(aq);
189 	if (status.async)
190 		return AP_SM_WAIT_NONE;
191 	switch (status.response_code) {
192 	case AP_RESPONSE_NORMAL:
193 		if (aq->queue_count > 0) {
194 			aq->sm_state = AP_SM_STATE_WORKING;
195 			return AP_SM_WAIT_AGAIN;
196 		}
197 		aq->sm_state = AP_SM_STATE_IDLE;
198 		return AP_SM_WAIT_NONE;
199 	case AP_RESPONSE_NO_PENDING_REPLY:
200 		if (aq->queue_count > 0)
201 			return aq->interrupt ?
202 				AP_SM_WAIT_INTERRUPT : AP_SM_WAIT_HIGH_TIMEOUT;
203 		aq->sm_state = AP_SM_STATE_IDLE;
204 		return AP_SM_WAIT_NONE;
205 	default:
206 		aq->dev_state = AP_DEV_STATE_ERROR;
207 		aq->last_err_rc = status.response_code;
208 		AP_DBF_WARN("%s RC 0x%02x on 0x%02x.%04x -> AP_DEV_STATE_ERROR\n",
209 			    __func__, status.response_code,
210 			    AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid));
211 		return AP_SM_WAIT_NONE;
212 	}
213 }
214 
215 /**
216  * ap_sm_write(): Send messages from the request queue to an AP queue.
217  * @aq: pointer to the AP queue
218  *
219  * Returns AP_SM_WAIT_NONE, AP_SM_WAIT_AGAIN, or AP_SM_WAIT_INTERRUPT
220  */
ap_sm_write(struct ap_queue * aq)221 static enum ap_sm_wait ap_sm_write(struct ap_queue *aq)
222 {
223 	struct ap_queue_status status;
224 	struct ap_message *ap_msg;
225 	ap_qid_t qid = aq->qid;
226 
227 	if (aq->requestq_count <= 0)
228 		return AP_SM_WAIT_NONE;
229 
230 	/* Start the next request on the queue. */
231 	ap_msg = list_entry(aq->requestq.next, struct ap_message, list);
232 	status = __ap_send(qid, ap_msg->psmid,
233 			   ap_msg->msg, ap_msg->len,
234 			   ap_msg->flags & AP_MSG_FLAG_SPECIAL);
235 	if (status.async)
236 		return AP_SM_WAIT_NONE;
237 	switch (status.response_code) {
238 	case AP_RESPONSE_NORMAL:
239 		aq->queue_count = max_t(int, 1, aq->queue_count + 1);
240 		if (aq->queue_count == 1)
241 			mod_timer(&aq->timeout, jiffies + aq->request_timeout);
242 		list_move_tail(&ap_msg->list, &aq->pendingq);
243 		aq->requestq_count--;
244 		aq->pendingq_count++;
245 		if (aq->queue_count < aq->card->queue_depth) {
246 			aq->sm_state = AP_SM_STATE_WORKING;
247 			return AP_SM_WAIT_AGAIN;
248 		}
249 		fallthrough;
250 	case AP_RESPONSE_Q_FULL:
251 		aq->sm_state = AP_SM_STATE_QUEUE_FULL;
252 		return aq->interrupt ?
253 			AP_SM_WAIT_INTERRUPT : AP_SM_WAIT_HIGH_TIMEOUT;
254 	case AP_RESPONSE_RESET_IN_PROGRESS:
255 		aq->sm_state = AP_SM_STATE_RESET_WAIT;
256 		return AP_SM_WAIT_LOW_TIMEOUT;
257 	case AP_RESPONSE_INVALID_DOMAIN:
258 		AP_DBF_WARN("%s RESPONSE_INVALID_DOMAIN on NQAP\n", __func__);
259 		fallthrough;
260 	case AP_RESPONSE_MESSAGE_TOO_BIG:
261 	case AP_RESPONSE_REQ_FAC_NOT_INST:
262 		list_del_init(&ap_msg->list);
263 		aq->requestq_count--;
264 		ap_msg->rc = -EINVAL;
265 		ap_msg->receive(aq, ap_msg, NULL);
266 		return AP_SM_WAIT_AGAIN;
267 	default:
268 		aq->dev_state = AP_DEV_STATE_ERROR;
269 		aq->last_err_rc = status.response_code;
270 		AP_DBF_WARN("%s RC 0x%02x on 0x%02x.%04x -> AP_DEV_STATE_ERROR\n",
271 			    __func__, status.response_code,
272 			    AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid));
273 		return AP_SM_WAIT_NONE;
274 	}
275 }
276 
277 /**
278  * ap_sm_read_write(): Send and receive messages to/from an AP queue.
279  * @aq: pointer to the AP queue
280  *
281  * Returns AP_SM_WAIT_NONE, AP_SM_WAIT_AGAIN, or AP_SM_WAIT_INTERRUPT
282  */
ap_sm_read_write(struct ap_queue * aq)283 static enum ap_sm_wait ap_sm_read_write(struct ap_queue *aq)
284 {
285 	return min(ap_sm_read(aq), ap_sm_write(aq));
286 }
287 
288 /**
289  * ap_sm_reset(): Reset an AP queue.
290  * @aq: The AP queue
291  *
292  * Submit the Reset command to an AP queue.
293  */
ap_sm_reset(struct ap_queue * aq)294 static enum ap_sm_wait ap_sm_reset(struct ap_queue *aq)
295 {
296 	struct ap_queue_status status;
297 
298 	status = ap_rapq(aq->qid, aq->rapq_fbit);
299 	if (status.async)
300 		return AP_SM_WAIT_NONE;
301 	switch (status.response_code) {
302 	case AP_RESPONSE_NORMAL:
303 	case AP_RESPONSE_RESET_IN_PROGRESS:
304 		aq->sm_state = AP_SM_STATE_RESET_WAIT;
305 		aq->interrupt = false;
306 		aq->rapq_fbit = 0;
307 		return AP_SM_WAIT_LOW_TIMEOUT;
308 	default:
309 		aq->dev_state = AP_DEV_STATE_ERROR;
310 		aq->last_err_rc = status.response_code;
311 		AP_DBF_WARN("%s RC 0x%02x on 0x%02x.%04x -> AP_DEV_STATE_ERROR\n",
312 			    __func__, status.response_code,
313 			    AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid));
314 		return AP_SM_WAIT_NONE;
315 	}
316 }
317 
318 /**
319  * ap_sm_reset_wait(): Test queue for completion of the reset operation
320  * @aq: pointer to the AP queue
321  *
322  * Returns AP_POLL_IMMEDIATELY, AP_POLL_AFTER_TIMEROUT or 0.
323  */
ap_sm_reset_wait(struct ap_queue * aq)324 static enum ap_sm_wait ap_sm_reset_wait(struct ap_queue *aq)
325 {
326 	struct ap_queue_status status;
327 	void *lsi_ptr;
328 
329 	if (aq->queue_count > 0 && aq->reply)
330 		/* Try to read a completed message and get the status */
331 		status = ap_sm_recv(aq);
332 	else
333 		/* Get the status with TAPQ */
334 		status = ap_tapq(aq->qid, NULL);
335 
336 	switch (status.response_code) {
337 	case AP_RESPONSE_NORMAL:
338 		lsi_ptr = ap_airq_ptr();
339 		if (lsi_ptr && ap_queue_enable_irq(aq, lsi_ptr) == 0)
340 			aq->sm_state = AP_SM_STATE_SETIRQ_WAIT;
341 		else
342 			aq->sm_state = (aq->queue_count > 0) ?
343 				AP_SM_STATE_WORKING : AP_SM_STATE_IDLE;
344 		return AP_SM_WAIT_AGAIN;
345 	case AP_RESPONSE_BUSY:
346 	case AP_RESPONSE_RESET_IN_PROGRESS:
347 		return AP_SM_WAIT_LOW_TIMEOUT;
348 	case AP_RESPONSE_Q_NOT_AVAIL:
349 	case AP_RESPONSE_DECONFIGURED:
350 	case AP_RESPONSE_CHECKSTOPPED:
351 	default:
352 		aq->dev_state = AP_DEV_STATE_ERROR;
353 		aq->last_err_rc = status.response_code;
354 		AP_DBF_WARN("%s RC 0x%02x on 0x%02x.%04x -> AP_DEV_STATE_ERROR\n",
355 			    __func__, status.response_code,
356 			    AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid));
357 		return AP_SM_WAIT_NONE;
358 	}
359 }
360 
361 /**
362  * ap_sm_setirq_wait(): Test queue for completion of the irq enablement
363  * @aq: pointer to the AP queue
364  *
365  * Returns AP_POLL_IMMEDIATELY, AP_POLL_AFTER_TIMEROUT or 0.
366  */
ap_sm_setirq_wait(struct ap_queue * aq)367 static enum ap_sm_wait ap_sm_setirq_wait(struct ap_queue *aq)
368 {
369 	struct ap_queue_status status;
370 
371 	if (aq->queue_count > 0 && aq->reply)
372 		/* Try to read a completed message and get the status */
373 		status = ap_sm_recv(aq);
374 	else
375 		/* Get the status with TAPQ */
376 		status = ap_tapq(aq->qid, NULL);
377 
378 	if (status.irq_enabled == 1) {
379 		/* Irqs are now enabled */
380 		aq->interrupt = true;
381 		aq->sm_state = (aq->queue_count > 0) ?
382 			AP_SM_STATE_WORKING : AP_SM_STATE_IDLE;
383 	}
384 
385 	switch (status.response_code) {
386 	case AP_RESPONSE_NORMAL:
387 		if (aq->queue_count > 0)
388 			return AP_SM_WAIT_AGAIN;
389 		fallthrough;
390 	case AP_RESPONSE_NO_PENDING_REPLY:
391 		return AP_SM_WAIT_LOW_TIMEOUT;
392 	default:
393 		aq->dev_state = AP_DEV_STATE_ERROR;
394 		aq->last_err_rc = status.response_code;
395 		AP_DBF_WARN("%s RC 0x%02x on 0x%02x.%04x -> AP_DEV_STATE_ERROR\n",
396 			    __func__, status.response_code,
397 			    AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid));
398 		return AP_SM_WAIT_NONE;
399 	}
400 }
401 
402 /**
403  * ap_sm_assoc_wait(): Test queue for completion of a pending
404  *		       association request.
405  * @aq: pointer to the AP queue
406  */
ap_sm_assoc_wait(struct ap_queue * aq)407 static enum ap_sm_wait ap_sm_assoc_wait(struct ap_queue *aq)
408 {
409 	struct ap_queue_status status;
410 	struct ap_tapq_gr2 info;
411 
412 	status = ap_test_queue(aq->qid, 1, &info);
413 	/* handle asynchronous error on this queue */
414 	if (status.async && status.response_code) {
415 		aq->dev_state = AP_DEV_STATE_ERROR;
416 		aq->last_err_rc = status.response_code;
417 		AP_DBF_WARN("%s asynch RC 0x%02x on 0x%02x.%04x -> AP_DEV_STATE_ERROR\n",
418 			    __func__, status.response_code,
419 			    AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid));
420 		return AP_SM_WAIT_NONE;
421 	}
422 	if (status.response_code > AP_RESPONSE_BUSY) {
423 		aq->dev_state = AP_DEV_STATE_ERROR;
424 		aq->last_err_rc = status.response_code;
425 		AP_DBF_WARN("%s RC 0x%02x on 0x%02x.%04x -> AP_DEV_STATE_ERROR\n",
426 			    __func__, status.response_code,
427 			    AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid));
428 		return AP_SM_WAIT_NONE;
429 	}
430 
431 	/* check bs bits */
432 	switch (info.bs) {
433 	case AP_BS_Q_USABLE:
434 		/* association is through */
435 		aq->sm_state = AP_SM_STATE_IDLE;
436 		AP_DBF_DBG("%s queue 0x%02x.%04x associated with %u\n",
437 			   __func__, AP_QID_CARD(aq->qid),
438 			   AP_QID_QUEUE(aq->qid), aq->assoc_idx);
439 		return AP_SM_WAIT_NONE;
440 	case AP_BS_Q_USABLE_NO_SECURE_KEY:
441 		/* association still pending */
442 		return AP_SM_WAIT_LOW_TIMEOUT;
443 	default:
444 		/* reset from 'outside' happened or no idea at all */
445 		aq->assoc_idx = ASSOC_IDX_INVALID;
446 		aq->dev_state = AP_DEV_STATE_ERROR;
447 		aq->last_err_rc = status.response_code;
448 		AP_DBF_WARN("%s bs 0x%02x on 0x%02x.%04x -> AP_DEV_STATE_ERROR\n",
449 			    __func__, info.bs,
450 			    AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid));
451 		return AP_SM_WAIT_NONE;
452 	}
453 }
454 
455 /*
456  * AP state machine jump table
457  */
458 static ap_func_t *ap_jumptable[NR_AP_SM_STATES][NR_AP_SM_EVENTS] = {
459 	[AP_SM_STATE_RESET_START] = {
460 		[AP_SM_EVENT_POLL] = ap_sm_reset,
461 		[AP_SM_EVENT_TIMEOUT] = ap_sm_nop,
462 	},
463 	[AP_SM_STATE_RESET_WAIT] = {
464 		[AP_SM_EVENT_POLL] = ap_sm_reset_wait,
465 		[AP_SM_EVENT_TIMEOUT] = ap_sm_nop,
466 	},
467 	[AP_SM_STATE_SETIRQ_WAIT] = {
468 		[AP_SM_EVENT_POLL] = ap_sm_setirq_wait,
469 		[AP_SM_EVENT_TIMEOUT] = ap_sm_nop,
470 	},
471 	[AP_SM_STATE_IDLE] = {
472 		[AP_SM_EVENT_POLL] = ap_sm_write,
473 		[AP_SM_EVENT_TIMEOUT] = ap_sm_nop,
474 	},
475 	[AP_SM_STATE_WORKING] = {
476 		[AP_SM_EVENT_POLL] = ap_sm_read_write,
477 		[AP_SM_EVENT_TIMEOUT] = ap_sm_reset,
478 	},
479 	[AP_SM_STATE_QUEUE_FULL] = {
480 		[AP_SM_EVENT_POLL] = ap_sm_read,
481 		[AP_SM_EVENT_TIMEOUT] = ap_sm_reset,
482 	},
483 	[AP_SM_STATE_ASSOC_WAIT] = {
484 		[AP_SM_EVENT_POLL] = ap_sm_assoc_wait,
485 		[AP_SM_EVENT_TIMEOUT] = ap_sm_reset,
486 	},
487 };
488 
ap_sm_event(struct ap_queue * aq,enum ap_sm_event event)489 enum ap_sm_wait ap_sm_event(struct ap_queue *aq, enum ap_sm_event event)
490 {
491 	if (aq->config && !aq->chkstop &&
492 	    aq->dev_state > AP_DEV_STATE_UNINITIATED)
493 		return ap_jumptable[aq->sm_state][event](aq);
494 	else
495 		return AP_SM_WAIT_NONE;
496 }
497 
ap_sm_event_loop(struct ap_queue * aq,enum ap_sm_event event)498 enum ap_sm_wait ap_sm_event_loop(struct ap_queue *aq, enum ap_sm_event event)
499 {
500 	enum ap_sm_wait wait;
501 
502 	while ((wait = ap_sm_event(aq, event)) == AP_SM_WAIT_AGAIN)
503 		;
504 	return wait;
505 }
506 
507 /*
508  * AP queue related attributes.
509  */
request_count_show(struct device * dev,struct device_attribute * attr,char * buf)510 static ssize_t request_count_show(struct device *dev,
511 				  struct device_attribute *attr,
512 				  char *buf)
513 {
514 	struct ap_queue *aq = to_ap_queue(dev);
515 	bool valid = false;
516 	u64 req_cnt;
517 
518 	spin_lock_bh(&aq->lock);
519 	if (aq->dev_state > AP_DEV_STATE_UNINITIATED) {
520 		req_cnt = aq->total_request_count;
521 		valid = true;
522 	}
523 	spin_unlock_bh(&aq->lock);
524 
525 	if (valid)
526 		return sysfs_emit(buf, "%llu\n", req_cnt);
527 	else
528 		return sysfs_emit(buf, "-\n");
529 }
530 
request_count_store(struct device * dev,struct device_attribute * attr,const char * buf,size_t count)531 static ssize_t request_count_store(struct device *dev,
532 				   struct device_attribute *attr,
533 				   const char *buf, size_t count)
534 {
535 	struct ap_queue *aq = to_ap_queue(dev);
536 
537 	spin_lock_bh(&aq->lock);
538 	aq->total_request_count = 0;
539 	spin_unlock_bh(&aq->lock);
540 
541 	return count;
542 }
543 
544 static DEVICE_ATTR_RW(request_count);
545 
requestq_count_show(struct device * dev,struct device_attribute * attr,char * buf)546 static ssize_t requestq_count_show(struct device *dev,
547 				   struct device_attribute *attr, char *buf)
548 {
549 	struct ap_queue *aq = to_ap_queue(dev);
550 	unsigned int reqq_cnt = 0;
551 
552 	spin_lock_bh(&aq->lock);
553 	if (aq->dev_state > AP_DEV_STATE_UNINITIATED)
554 		reqq_cnt = aq->requestq_count;
555 	spin_unlock_bh(&aq->lock);
556 	return sysfs_emit(buf, "%d\n", reqq_cnt);
557 }
558 
559 static DEVICE_ATTR_RO(requestq_count);
560 
pendingq_count_show(struct device * dev,struct device_attribute * attr,char * buf)561 static ssize_t pendingq_count_show(struct device *dev,
562 				   struct device_attribute *attr, char *buf)
563 {
564 	struct ap_queue *aq = to_ap_queue(dev);
565 	unsigned int penq_cnt = 0;
566 
567 	spin_lock_bh(&aq->lock);
568 	if (aq->dev_state > AP_DEV_STATE_UNINITIATED)
569 		penq_cnt = aq->pendingq_count;
570 	spin_unlock_bh(&aq->lock);
571 	return sysfs_emit(buf, "%d\n", penq_cnt);
572 }
573 
574 static DEVICE_ATTR_RO(pendingq_count);
575 
reset_show(struct device * dev,struct device_attribute * attr,char * buf)576 static ssize_t reset_show(struct device *dev,
577 			  struct device_attribute *attr, char *buf)
578 {
579 	struct ap_queue *aq = to_ap_queue(dev);
580 	int rc = 0;
581 
582 	spin_lock_bh(&aq->lock);
583 	switch (aq->sm_state) {
584 	case AP_SM_STATE_RESET_START:
585 	case AP_SM_STATE_RESET_WAIT:
586 		rc = sysfs_emit(buf, "Reset in progress.\n");
587 		break;
588 	case AP_SM_STATE_WORKING:
589 	case AP_SM_STATE_QUEUE_FULL:
590 		rc = sysfs_emit(buf, "Reset Timer armed.\n");
591 		break;
592 	default:
593 		rc = sysfs_emit(buf, "No Reset Timer set.\n");
594 	}
595 	spin_unlock_bh(&aq->lock);
596 	return rc;
597 }
598 
reset_store(struct device * dev,struct device_attribute * attr,const char * buf,size_t count)599 static ssize_t reset_store(struct device *dev,
600 			   struct device_attribute *attr,
601 			   const char *buf, size_t count)
602 {
603 	struct ap_queue *aq = to_ap_queue(dev);
604 
605 	spin_lock_bh(&aq->lock);
606 	__ap_flush_queue(aq);
607 	aq->sm_state = AP_SM_STATE_RESET_START;
608 	ap_wait(ap_sm_event(aq, AP_SM_EVENT_POLL));
609 	spin_unlock_bh(&aq->lock);
610 
611 	AP_DBF_INFO("%s reset queue=%02x.%04x triggered by user\n",
612 		    __func__, AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid));
613 
614 	return count;
615 }
616 
617 static DEVICE_ATTR_RW(reset);
618 
interrupt_show(struct device * dev,struct device_attribute * attr,char * buf)619 static ssize_t interrupt_show(struct device *dev,
620 			      struct device_attribute *attr, char *buf)
621 {
622 	struct ap_queue *aq = to_ap_queue(dev);
623 	int rc = 0;
624 
625 	spin_lock_bh(&aq->lock);
626 	if (aq->sm_state == AP_SM_STATE_SETIRQ_WAIT)
627 		rc = sysfs_emit(buf, "Enable Interrupt pending.\n");
628 	else if (aq->interrupt)
629 		rc = sysfs_emit(buf, "Interrupts enabled.\n");
630 	else
631 		rc = sysfs_emit(buf, "Interrupts disabled.\n");
632 	spin_unlock_bh(&aq->lock);
633 	return rc;
634 }
635 
636 static DEVICE_ATTR_RO(interrupt);
637 
config_show(struct device * dev,struct device_attribute * attr,char * buf)638 static ssize_t config_show(struct device *dev,
639 			   struct device_attribute *attr, char *buf)
640 {
641 	struct ap_queue *aq = to_ap_queue(dev);
642 	int rc;
643 
644 	spin_lock_bh(&aq->lock);
645 	rc = sysfs_emit(buf, "%d\n", aq->config ? 1 : 0);
646 	spin_unlock_bh(&aq->lock);
647 	return rc;
648 }
649 
650 static DEVICE_ATTR_RO(config);
651 
chkstop_show(struct device * dev,struct device_attribute * attr,char * buf)652 static ssize_t chkstop_show(struct device *dev,
653 			    struct device_attribute *attr, char *buf)
654 {
655 	struct ap_queue *aq = to_ap_queue(dev);
656 	int rc;
657 
658 	spin_lock_bh(&aq->lock);
659 	rc = sysfs_emit(buf, "%d\n", aq->chkstop ? 1 : 0);
660 	spin_unlock_bh(&aq->lock);
661 	return rc;
662 }
663 
664 static DEVICE_ATTR_RO(chkstop);
665 
ap_functions_show(struct device * dev,struct device_attribute * attr,char * buf)666 static ssize_t ap_functions_show(struct device *dev,
667 				 struct device_attribute *attr, char *buf)
668 {
669 	struct ap_queue *aq = to_ap_queue(dev);
670 	struct ap_queue_status status;
671 	struct ap_tapq_gr2 info;
672 
673 	status = ap_test_queue(aq->qid, 1, &info);
674 	if (status.response_code > AP_RESPONSE_BUSY) {
675 		AP_DBF_DBG("%s RC 0x%02x on tapq(0x%02x.%04x)\n",
676 			   __func__, status.response_code,
677 			   AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid));
678 		return -EIO;
679 	}
680 
681 	return sysfs_emit(buf, "0x%08X\n", info.fac);
682 }
683 
684 static DEVICE_ATTR_RO(ap_functions);
685 
686 #ifdef CONFIG_ZCRYPT_DEBUG
states_show(struct device * dev,struct device_attribute * attr,char * buf)687 static ssize_t states_show(struct device *dev,
688 			   struct device_attribute *attr, char *buf)
689 {
690 	struct ap_queue *aq = to_ap_queue(dev);
691 	int rc = 0;
692 
693 	spin_lock_bh(&aq->lock);
694 	/* queue device state */
695 	switch (aq->dev_state) {
696 	case AP_DEV_STATE_UNINITIATED:
697 		rc = sysfs_emit(buf, "UNINITIATED\n");
698 		break;
699 	case AP_DEV_STATE_OPERATING:
700 		rc = sysfs_emit(buf, "OPERATING");
701 		break;
702 	case AP_DEV_STATE_SHUTDOWN:
703 		rc = sysfs_emit(buf, "SHUTDOWN");
704 		break;
705 	case AP_DEV_STATE_ERROR:
706 		rc = sysfs_emit(buf, "ERROR");
707 		break;
708 	default:
709 		rc = sysfs_emit(buf, "UNKNOWN");
710 	}
711 	/* state machine state */
712 	if (aq->dev_state) {
713 		switch (aq->sm_state) {
714 		case AP_SM_STATE_RESET_START:
715 			rc += sysfs_emit_at(buf, rc, " [RESET_START]\n");
716 			break;
717 		case AP_SM_STATE_RESET_WAIT:
718 			rc += sysfs_emit_at(buf, rc, " [RESET_WAIT]\n");
719 			break;
720 		case AP_SM_STATE_SETIRQ_WAIT:
721 			rc += sysfs_emit_at(buf, rc, " [SETIRQ_WAIT]\n");
722 			break;
723 		case AP_SM_STATE_IDLE:
724 			rc += sysfs_emit_at(buf, rc, " [IDLE]\n");
725 			break;
726 		case AP_SM_STATE_WORKING:
727 			rc += sysfs_emit_at(buf, rc, " [WORKING]\n");
728 			break;
729 		case AP_SM_STATE_QUEUE_FULL:
730 			rc += sysfs_emit_at(buf, rc, " [FULL]\n");
731 			break;
732 		case AP_SM_STATE_ASSOC_WAIT:
733 			rc += sysfs_emit_at(buf, rc, " [ASSOC_WAIT]\n");
734 			break;
735 		default:
736 			rc += sysfs_emit_at(buf, rc, " [UNKNOWN]\n");
737 		}
738 	}
739 	spin_unlock_bh(&aq->lock);
740 
741 	return rc;
742 }
743 static DEVICE_ATTR_RO(states);
744 
last_err_rc_show(struct device * dev,struct device_attribute * attr,char * buf)745 static ssize_t last_err_rc_show(struct device *dev,
746 				struct device_attribute *attr, char *buf)
747 {
748 	struct ap_queue *aq = to_ap_queue(dev);
749 	int rc;
750 
751 	spin_lock_bh(&aq->lock);
752 	rc = aq->last_err_rc;
753 	spin_unlock_bh(&aq->lock);
754 
755 	switch (rc) {
756 	case AP_RESPONSE_NORMAL:
757 		return sysfs_emit(buf, "NORMAL\n");
758 	case AP_RESPONSE_Q_NOT_AVAIL:
759 		return sysfs_emit(buf, "Q_NOT_AVAIL\n");
760 	case AP_RESPONSE_RESET_IN_PROGRESS:
761 		return sysfs_emit(buf, "RESET_IN_PROGRESS\n");
762 	case AP_RESPONSE_DECONFIGURED:
763 		return sysfs_emit(buf, "DECONFIGURED\n");
764 	case AP_RESPONSE_CHECKSTOPPED:
765 		return sysfs_emit(buf, "CHECKSTOPPED\n");
766 	case AP_RESPONSE_BUSY:
767 		return sysfs_emit(buf, "BUSY\n");
768 	case AP_RESPONSE_INVALID_ADDRESS:
769 		return sysfs_emit(buf, "INVALID_ADDRESS\n");
770 	case AP_RESPONSE_OTHERWISE_CHANGED:
771 		return sysfs_emit(buf, "OTHERWISE_CHANGED\n");
772 	case AP_RESPONSE_Q_FULL:
773 		return sysfs_emit(buf, "Q_FULL/NO_PENDING_REPLY\n");
774 	case AP_RESPONSE_INDEX_TOO_BIG:
775 		return sysfs_emit(buf, "INDEX_TOO_BIG\n");
776 	case AP_RESPONSE_NO_FIRST_PART:
777 		return sysfs_emit(buf, "NO_FIRST_PART\n");
778 	case AP_RESPONSE_MESSAGE_TOO_BIG:
779 		return sysfs_emit(buf, "MESSAGE_TOO_BIG\n");
780 	case AP_RESPONSE_REQ_FAC_NOT_INST:
781 		return sysfs_emit(buf, "REQ_FAC_NOT_INST\n");
782 	default:
783 		return sysfs_emit(buf, "response code %d\n", rc);
784 	}
785 }
786 static DEVICE_ATTR_RO(last_err_rc);
787 #endif
788 
789 static struct attribute *ap_queue_dev_attrs[] = {
790 	&dev_attr_request_count.attr,
791 	&dev_attr_requestq_count.attr,
792 	&dev_attr_pendingq_count.attr,
793 	&dev_attr_reset.attr,
794 	&dev_attr_interrupt.attr,
795 	&dev_attr_config.attr,
796 	&dev_attr_chkstop.attr,
797 	&dev_attr_ap_functions.attr,
798 #ifdef CONFIG_ZCRYPT_DEBUG
799 	&dev_attr_states.attr,
800 	&dev_attr_last_err_rc.attr,
801 #endif
802 	NULL
803 };
804 
805 static struct attribute_group ap_queue_dev_attr_group = {
806 	.attrs = ap_queue_dev_attrs
807 };
808 
809 static const struct attribute_group *ap_queue_dev_attr_groups[] = {
810 	&ap_queue_dev_attr_group,
811 	NULL
812 };
813 
814 static struct device_type ap_queue_type = {
815 	.name = "ap_queue",
816 	.groups = ap_queue_dev_attr_groups,
817 };
818 
se_bind_show(struct device * dev,struct device_attribute * attr,char * buf)819 static ssize_t se_bind_show(struct device *dev,
820 			    struct device_attribute *attr, char *buf)
821 {
822 	struct ap_queue *aq = to_ap_queue(dev);
823 	struct ap_queue_status status;
824 	struct ap_tapq_gr2 info;
825 
826 	if (!ap_q_supports_bind(aq))
827 		return sysfs_emit(buf, "-\n");
828 
829 	status = ap_test_queue(aq->qid, 1, &info);
830 	if (status.response_code > AP_RESPONSE_BUSY) {
831 		AP_DBF_DBG("%s RC 0x%02x on tapq(0x%02x.%04x)\n",
832 			   __func__, status.response_code,
833 			   AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid));
834 		return -EIO;
835 	}
836 	switch (info.bs) {
837 	case AP_BS_Q_USABLE:
838 	case AP_BS_Q_USABLE_NO_SECURE_KEY:
839 		return sysfs_emit(buf, "bound\n");
840 	default:
841 		return sysfs_emit(buf, "unbound\n");
842 	}
843 }
844 
se_bind_store(struct device * dev,struct device_attribute * attr,const char * buf,size_t count)845 static ssize_t se_bind_store(struct device *dev,
846 			     struct device_attribute *attr,
847 			     const char *buf, size_t count)
848 {
849 	struct ap_queue *aq = to_ap_queue(dev);
850 	struct ap_queue_status status;
851 	bool value;
852 	int rc;
853 
854 	if (!ap_q_supports_bind(aq))
855 		return -EINVAL;
856 
857 	/* only 0 (unbind) and 1 (bind) allowed */
858 	rc = kstrtobool(buf, &value);
859 	if (rc)
860 		return rc;
861 
862 	if (value) {
863 		/* bind, do BAPQ */
864 		spin_lock_bh(&aq->lock);
865 		if (aq->sm_state < AP_SM_STATE_IDLE) {
866 			spin_unlock_bh(&aq->lock);
867 			return -EBUSY;
868 		}
869 		status = ap_bapq(aq->qid);
870 		spin_unlock_bh(&aq->lock);
871 		if (status.response_code) {
872 			AP_DBF_WARN("%s RC 0x%02x on bapq(0x%02x.%04x)\n",
873 				    __func__, status.response_code,
874 				    AP_QID_CARD(aq->qid),
875 				    AP_QID_QUEUE(aq->qid));
876 			return -EIO;
877 		}
878 	} else {
879 		/* unbind, set F bit arg and trigger RAPQ */
880 		spin_lock_bh(&aq->lock);
881 		__ap_flush_queue(aq);
882 		aq->rapq_fbit = 1;
883 		aq->assoc_idx = ASSOC_IDX_INVALID;
884 		aq->sm_state = AP_SM_STATE_RESET_START;
885 		ap_wait(ap_sm_event(aq, AP_SM_EVENT_POLL));
886 		spin_unlock_bh(&aq->lock);
887 	}
888 
889 	return count;
890 }
891 
892 static DEVICE_ATTR_RW(se_bind);
893 
se_associate_show(struct device * dev,struct device_attribute * attr,char * buf)894 static ssize_t se_associate_show(struct device *dev,
895 				 struct device_attribute *attr, char *buf)
896 {
897 	struct ap_queue *aq = to_ap_queue(dev);
898 	struct ap_queue_status status;
899 	struct ap_tapq_gr2 info;
900 
901 	if (!ap_q_supports_assoc(aq))
902 		return sysfs_emit(buf, "-\n");
903 
904 	status = ap_test_queue(aq->qid, 1, &info);
905 	if (status.response_code > AP_RESPONSE_BUSY) {
906 		AP_DBF_DBG("%s RC 0x%02x on tapq(0x%02x.%04x)\n",
907 			   __func__, status.response_code,
908 			   AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid));
909 		return -EIO;
910 	}
911 
912 	switch (info.bs) {
913 	case AP_BS_Q_USABLE:
914 		if (aq->assoc_idx == ASSOC_IDX_INVALID) {
915 			AP_DBF_WARN("%s AP_BS_Q_USABLE but invalid assoc_idx\n", __func__);
916 			return -EIO;
917 		}
918 		return sysfs_emit(buf, "associated %u\n", aq->assoc_idx);
919 	case AP_BS_Q_USABLE_NO_SECURE_KEY:
920 		if (aq->assoc_idx != ASSOC_IDX_INVALID)
921 			return sysfs_emit(buf, "association pending\n");
922 		fallthrough;
923 	default:
924 		return sysfs_emit(buf, "unassociated\n");
925 	}
926 }
927 
se_associate_store(struct device * dev,struct device_attribute * attr,const char * buf,size_t count)928 static ssize_t se_associate_store(struct device *dev,
929 				  struct device_attribute *attr,
930 				  const char *buf, size_t count)
931 {
932 	struct ap_queue *aq = to_ap_queue(dev);
933 	struct ap_queue_status status;
934 	unsigned int value;
935 	int rc;
936 
937 	if (!ap_q_supports_assoc(aq))
938 		return -EINVAL;
939 
940 	/* association index needs to be >= 0 */
941 	rc = kstrtouint(buf, 0, &value);
942 	if (rc)
943 		return rc;
944 	if (value >= ASSOC_IDX_INVALID)
945 		return -EINVAL;
946 
947 	spin_lock_bh(&aq->lock);
948 
949 	/* sm should be in idle state */
950 	if (aq->sm_state != AP_SM_STATE_IDLE) {
951 		spin_unlock_bh(&aq->lock);
952 		return -EBUSY;
953 	}
954 
955 	/* already associated or association pending ? */
956 	if (aq->assoc_idx != ASSOC_IDX_INVALID) {
957 		spin_unlock_bh(&aq->lock);
958 		return -EINVAL;
959 	}
960 
961 	/* trigger the asynchronous association request */
962 	status = ap_aapq(aq->qid, value);
963 	switch (status.response_code) {
964 	case AP_RESPONSE_NORMAL:
965 	case AP_RESPONSE_STATE_CHANGE_IN_PROGRESS:
966 		aq->sm_state = AP_SM_STATE_ASSOC_WAIT;
967 		aq->assoc_idx = value;
968 		ap_wait(ap_sm_event(aq, AP_SM_EVENT_POLL));
969 		spin_unlock_bh(&aq->lock);
970 		break;
971 	default:
972 		spin_unlock_bh(&aq->lock);
973 		AP_DBF_WARN("%s RC 0x%02x on aapq(0x%02x.%04x)\n",
974 			    __func__, status.response_code,
975 			    AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid));
976 		return -EIO;
977 	}
978 
979 	return count;
980 }
981 
982 static DEVICE_ATTR_RW(se_associate);
983 
984 static struct attribute *ap_queue_dev_sb_attrs[] = {
985 	&dev_attr_se_bind.attr,
986 	&dev_attr_se_associate.attr,
987 	NULL
988 };
989 
990 static struct attribute_group ap_queue_dev_sb_attr_group = {
991 	.attrs = ap_queue_dev_sb_attrs
992 };
993 
994 static const struct attribute_group *ap_queue_dev_sb_attr_groups[] = {
995 	&ap_queue_dev_sb_attr_group,
996 	NULL
997 };
998 
ap_queue_device_release(struct device * dev)999 static void ap_queue_device_release(struct device *dev)
1000 {
1001 	struct ap_queue *aq = to_ap_queue(dev);
1002 
1003 	spin_lock_bh(&ap_queues_lock);
1004 	hash_del(&aq->hnode);
1005 	spin_unlock_bh(&ap_queues_lock);
1006 
1007 	kfree(aq);
1008 }
1009 
ap_queue_create(ap_qid_t qid,int device_type)1010 struct ap_queue *ap_queue_create(ap_qid_t qid, int device_type)
1011 {
1012 	struct ap_queue *aq;
1013 
1014 	aq = kzalloc(sizeof(*aq), GFP_KERNEL);
1015 	if (!aq)
1016 		return NULL;
1017 	aq->ap_dev.device.release = ap_queue_device_release;
1018 	aq->ap_dev.device.type = &ap_queue_type;
1019 	aq->ap_dev.device_type = device_type;
1020 	// add optional SE secure binding attributes group
1021 	if (ap_sb_available() && is_prot_virt_guest())
1022 		aq->ap_dev.device.groups = ap_queue_dev_sb_attr_groups;
1023 	aq->qid = qid;
1024 	aq->interrupt = false;
1025 	spin_lock_init(&aq->lock);
1026 	INIT_LIST_HEAD(&aq->pendingq);
1027 	INIT_LIST_HEAD(&aq->requestq);
1028 	timer_setup(&aq->timeout, ap_request_timeout, 0);
1029 
1030 	return aq;
1031 }
1032 
ap_queue_init_reply(struct ap_queue * aq,struct ap_message * reply)1033 void ap_queue_init_reply(struct ap_queue *aq, struct ap_message *reply)
1034 {
1035 	aq->reply = reply;
1036 
1037 	spin_lock_bh(&aq->lock);
1038 	ap_wait(ap_sm_event(aq, AP_SM_EVENT_POLL));
1039 	spin_unlock_bh(&aq->lock);
1040 }
1041 EXPORT_SYMBOL(ap_queue_init_reply);
1042 
1043 /**
1044  * ap_queue_message(): Queue a request to an AP device.
1045  * @aq: The AP device to queue the message to
1046  * @ap_msg: The message that is to be added
1047  */
ap_queue_message(struct ap_queue * aq,struct ap_message * ap_msg)1048 int ap_queue_message(struct ap_queue *aq, struct ap_message *ap_msg)
1049 {
1050 	int rc = 0;
1051 
1052 	/* msg needs to have a valid receive-callback */
1053 	BUG_ON(!ap_msg->receive);
1054 
1055 	spin_lock_bh(&aq->lock);
1056 
1057 	/* only allow to queue new messages if device state is ok */
1058 	if (aq->dev_state == AP_DEV_STATE_OPERATING) {
1059 		list_add_tail(&ap_msg->list, &aq->requestq);
1060 		aq->requestq_count++;
1061 		aq->total_request_count++;
1062 		atomic64_inc(&aq->card->total_request_count);
1063 	} else {
1064 		rc = -ENODEV;
1065 	}
1066 
1067 	/* Send/receive as many request from the queue as possible. */
1068 	ap_wait(ap_sm_event_loop(aq, AP_SM_EVENT_POLL));
1069 
1070 	spin_unlock_bh(&aq->lock);
1071 
1072 	return rc;
1073 }
1074 EXPORT_SYMBOL(ap_queue_message);
1075 
1076 /**
1077  * ap_cancel_message(): Cancel a crypto request.
1078  * @aq: The AP device that has the message queued
1079  * @ap_msg: The message that is to be removed
1080  *
1081  * Cancel a crypto request. This is done by removing the request
1082  * from the device pending or request queue. Note that the
1083  * request stays on the AP queue. When it finishes the message
1084  * reply will be discarded because the psmid can't be found.
1085  */
ap_cancel_message(struct ap_queue * aq,struct ap_message * ap_msg)1086 void ap_cancel_message(struct ap_queue *aq, struct ap_message *ap_msg)
1087 {
1088 	struct ap_message *tmp;
1089 
1090 	spin_lock_bh(&aq->lock);
1091 	if (!list_empty(&ap_msg->list)) {
1092 		list_for_each_entry(tmp, &aq->pendingq, list)
1093 			if (tmp->psmid == ap_msg->psmid) {
1094 				aq->pendingq_count--;
1095 				goto found;
1096 			}
1097 		aq->requestq_count--;
1098 found:
1099 		list_del_init(&ap_msg->list);
1100 	}
1101 	spin_unlock_bh(&aq->lock);
1102 }
1103 EXPORT_SYMBOL(ap_cancel_message);
1104 
1105 /**
1106  * __ap_flush_queue(): Flush requests.
1107  * @aq: Pointer to the AP queue
1108  *
1109  * Flush all requests from the request/pending queue of an AP device.
1110  */
__ap_flush_queue(struct ap_queue * aq)1111 static void __ap_flush_queue(struct ap_queue *aq)
1112 {
1113 	struct ap_message *ap_msg, *next;
1114 
1115 	list_for_each_entry_safe(ap_msg, next, &aq->pendingq, list) {
1116 		list_del_init(&ap_msg->list);
1117 		aq->pendingq_count--;
1118 		ap_msg->rc = -EAGAIN;
1119 		ap_msg->receive(aq, ap_msg, NULL);
1120 	}
1121 	list_for_each_entry_safe(ap_msg, next, &aq->requestq, list) {
1122 		list_del_init(&ap_msg->list);
1123 		aq->requestq_count--;
1124 		ap_msg->rc = -EAGAIN;
1125 		ap_msg->receive(aq, ap_msg, NULL);
1126 	}
1127 	aq->queue_count = 0;
1128 }
1129 
ap_flush_queue(struct ap_queue * aq)1130 void ap_flush_queue(struct ap_queue *aq)
1131 {
1132 	spin_lock_bh(&aq->lock);
1133 	__ap_flush_queue(aq);
1134 	spin_unlock_bh(&aq->lock);
1135 }
1136 EXPORT_SYMBOL(ap_flush_queue);
1137 
ap_queue_prepare_remove(struct ap_queue * aq)1138 void ap_queue_prepare_remove(struct ap_queue *aq)
1139 {
1140 	spin_lock_bh(&aq->lock);
1141 	/* flush queue */
1142 	__ap_flush_queue(aq);
1143 	/* move queue device state to SHUTDOWN in progress */
1144 	aq->dev_state = AP_DEV_STATE_SHUTDOWN;
1145 	spin_unlock_bh(&aq->lock);
1146 	del_timer_sync(&aq->timeout);
1147 }
1148 
ap_queue_remove(struct ap_queue * aq)1149 void ap_queue_remove(struct ap_queue *aq)
1150 {
1151 	/*
1152 	 * all messages have been flushed and the device state
1153 	 * is SHUTDOWN. Now reset with zero which also clears
1154 	 * the irq registration and move the device state
1155 	 * to the initial value AP_DEV_STATE_UNINITIATED.
1156 	 */
1157 	spin_lock_bh(&aq->lock);
1158 	ap_zapq(aq->qid, 0);
1159 	aq->dev_state = AP_DEV_STATE_UNINITIATED;
1160 	spin_unlock_bh(&aq->lock);
1161 }
1162 
ap_queue_init_state(struct ap_queue * aq)1163 void ap_queue_init_state(struct ap_queue *aq)
1164 {
1165 	spin_lock_bh(&aq->lock);
1166 	aq->dev_state = AP_DEV_STATE_OPERATING;
1167 	aq->sm_state = AP_SM_STATE_RESET_START;
1168 	aq->last_err_rc = 0;
1169 	aq->assoc_idx = ASSOC_IDX_INVALID;
1170 	ap_wait(ap_sm_event(aq, AP_SM_EVENT_POLL));
1171 	spin_unlock_bh(&aq->lock);
1172 }
1173 EXPORT_SYMBOL(ap_queue_init_state);
1174