1 // SPDX-License-Identifier: GPL-2.0
3 * Shared Memory Communications over RDMA (SMC-R) and RoCE
5 * Socket Closing - normal and abnormal
7 * Copyright IBM Corp. 2016
9 * Author(s): Ursula Braun <ubraun@linux.vnet.ibm.com>
12 #include <linux/workqueue.h>
13 #include <linux/sched/signal.h>
21 #include "smc_close.h"
23 #define SMC_CLOSE_WAIT_LISTEN_CLCSOCK_TIME (5 * HZ)
25 /* release the clcsock that is assigned to the smc_sock */
26 void smc_clcsock_release(struct smc_sock *smc)
30 if (smc->listen_smc && current_work() != &smc->smc_listen_work)
31 cancel_work_sync(&smc->smc_listen_work);
32 mutex_lock(&smc->clcsock_release_lock);
38 mutex_unlock(&smc->clcsock_release_lock);
41 static void smc_close_cleanup_listen(struct sock *parent)
45 /* Close non-accepted connections */
46 while ((sk = smc_accept_dequeue(parent, NULL)))
47 smc_close_non_accepted(sk);
50 /* wait for sndbuf data being transmitted */
51 static void smc_close_stream_wait(struct smc_sock *smc, long timeout)
53 DEFINE_WAIT_FUNC(wait, woken_wake_function);
54 struct sock *sk = &smc->sk;
59 if (!smc_tx_prepared_sends(&smc->conn))
62 smc->wait_close_tx_prepared = 1;
63 add_wait_queue(sk_sleep(sk), &wait);
64 while (!signal_pending(current) && timeout) {
67 rc = sk_wait_event(sk, &timeout,
68 !smc_tx_prepared_sends(&smc->conn) ||
69 sk->sk_err == ECONNABORTED ||
70 sk->sk_err == ECONNRESET ||
76 remove_wait_queue(sk_sleep(sk), &wait);
77 smc->wait_close_tx_prepared = 0;
80 void smc_close_wake_tx_prepared(struct smc_sock *smc)
82 if (smc->wait_close_tx_prepared)
83 /* wake up socket closing */
84 smc->sk.sk_state_change(&smc->sk);
87 static int smc_close_wr(struct smc_connection *conn)
89 conn->local_tx_ctrl.conn_state_flags.peer_done_writing = 1;
91 return smc_cdc_get_slot_and_msg_send(conn);
94 static int smc_close_final(struct smc_connection *conn)
96 if (atomic_read(&conn->bytes_to_rcv))
97 conn->local_tx_ctrl.conn_state_flags.peer_conn_abort = 1;
99 conn->local_tx_ctrl.conn_state_flags.peer_conn_closed = 1;
103 return smc_cdc_get_slot_and_msg_send(conn);
106 int smc_close_abort(struct smc_connection *conn)
108 conn->local_tx_ctrl.conn_state_flags.peer_conn_abort = 1;
110 return smc_cdc_get_slot_and_msg_send(conn);
113 /* terminate smc socket abnormally - active abort
114 * link group is terminated, i.e. RDMA communication no longer possible
116 void smc_close_active_abort(struct smc_sock *smc)
118 struct sock *sk = &smc->sk;
119 bool release_clcsock = false;
121 if (sk->sk_state != SMC_INIT && smc->clcsock && smc->clcsock->sk) {
122 sk->sk_err = ECONNABORTED;
123 if (smc->clcsock && smc->clcsock->sk)
124 tcp_abort(smc->clcsock->sk, ECONNABORTED);
126 switch (sk->sk_state) {
128 sk->sk_state = SMC_PEERABORTWAIT;
130 cancel_delayed_work_sync(&smc->conn.tx_work);
132 sk->sk_state = SMC_CLOSED;
133 sock_put(sk); /* passive closing */
135 case SMC_APPCLOSEWAIT1:
136 case SMC_APPCLOSEWAIT2:
138 cancel_delayed_work_sync(&smc->conn.tx_work);
140 sk->sk_state = SMC_CLOSED;
141 sock_put(sk); /* postponed passive closing */
143 case SMC_PEERCLOSEWAIT1:
144 case SMC_PEERCLOSEWAIT2:
145 case SMC_PEERFINCLOSEWAIT:
146 sk->sk_state = SMC_CLOSED;
147 smc_conn_free(&smc->conn);
148 release_clcsock = true;
149 sock_put(sk); /* passive closing */
151 case SMC_PROCESSABORT:
152 case SMC_APPFINCLOSEWAIT:
153 sk->sk_state = SMC_CLOSED;
156 case SMC_PEERABORTWAIT:
161 sock_set_flag(sk, SOCK_DEAD);
162 sk->sk_state_change(sk);
164 if (release_clcsock) {
166 smc_clcsock_release(smc);
171 static inline bool smc_close_sent_any_close(struct smc_connection *conn)
173 return conn->local_tx_ctrl.conn_state_flags.peer_conn_abort ||
174 conn->local_tx_ctrl.conn_state_flags.peer_conn_closed;
177 int smc_close_active(struct smc_sock *smc)
179 struct smc_cdc_conn_state_flags *txflags =
180 &smc->conn.local_tx_ctrl.conn_state_flags;
181 struct smc_connection *conn = &smc->conn;
182 struct sock *sk = &smc->sk;
187 timeout = current->flags & PF_EXITING ?
188 0 : sock_flag(sk, SOCK_LINGER) ?
189 sk->sk_lingertime : SMC_MAX_STREAM_WAIT_TIMEOUT;
191 old_state = sk->sk_state;
193 switch (sk->sk_state) {
195 sk->sk_state = SMC_CLOSED;
198 sk->sk_state = SMC_CLOSED;
199 sk->sk_state_change(sk); /* wake up accept */
200 if (smc->clcsock && smc->clcsock->sk) {
201 rc = kernel_sock_shutdown(smc->clcsock, SHUT_RDWR);
202 /* wake up kernel_accept of smc_tcp_listen_worker */
203 smc->clcsock->sk->sk_data_ready(smc->clcsock->sk);
205 smc_close_cleanup_listen(sk);
207 flush_work(&smc->tcp_listen_work);
211 smc_close_stream_wait(smc, timeout);
213 cancel_delayed_work_sync(&conn->tx_work);
215 if (sk->sk_state == SMC_ACTIVE) {
216 /* send close request */
217 rc = smc_close_final(conn);
218 sk->sk_state = SMC_PEERCLOSEWAIT1;
220 /* peer event has changed the state */
224 case SMC_APPFINCLOSEWAIT:
225 /* socket already shutdown wr or both (active close) */
226 if (txflags->peer_done_writing &&
227 !smc_close_sent_any_close(conn)) {
228 /* just shutdown wr done, send close request */
229 rc = smc_close_final(conn);
231 sk->sk_state = SMC_CLOSED;
233 case SMC_APPCLOSEWAIT1:
234 case SMC_APPCLOSEWAIT2:
235 if (!smc_cdc_rxed_any_close(conn))
236 smc_close_stream_wait(smc, timeout);
238 cancel_delayed_work_sync(&conn->tx_work);
240 if (sk->sk_state != SMC_APPCLOSEWAIT1 &&
241 sk->sk_state != SMC_APPCLOSEWAIT2)
243 /* confirm close from peer */
244 rc = smc_close_final(conn);
245 if (smc_cdc_rxed_any_close(conn)) {
246 /* peer has closed the socket already */
247 sk->sk_state = SMC_CLOSED;
248 sock_put(sk); /* postponed passive closing */
250 /* peer has just issued a shutdown write */
251 sk->sk_state = SMC_PEERFINCLOSEWAIT;
254 case SMC_PEERCLOSEWAIT1:
255 case SMC_PEERCLOSEWAIT2:
256 if (txflags->peer_done_writing &&
257 !smc_close_sent_any_close(conn)) {
258 /* just shutdown wr done, send close request */
259 rc = smc_close_final(conn);
261 /* peer sending PeerConnectionClosed will cause transition */
263 case SMC_PEERFINCLOSEWAIT:
264 /* peer sending PeerConnectionClosed will cause transition */
266 case SMC_PROCESSABORT:
267 rc = smc_close_abort(conn);
268 sk->sk_state = SMC_CLOSED;
270 case SMC_PEERABORTWAIT:
271 sk->sk_state = SMC_CLOSED;
274 /* nothing to do, add tracing in future patch */
278 if (old_state != sk->sk_state)
279 sk->sk_state_change(sk);
283 static void smc_close_passive_abort_received(struct smc_sock *smc)
285 struct smc_cdc_conn_state_flags *txflags =
286 &smc->conn.local_tx_ctrl.conn_state_flags;
287 struct sock *sk = &smc->sk;
289 switch (sk->sk_state) {
292 case SMC_APPCLOSEWAIT1:
293 sk->sk_state = SMC_PROCESSABORT;
294 sock_put(sk); /* passive closing */
296 case SMC_APPFINCLOSEWAIT:
297 sk->sk_state = SMC_PROCESSABORT;
299 case SMC_PEERCLOSEWAIT1:
300 case SMC_PEERCLOSEWAIT2:
301 if (txflags->peer_done_writing &&
302 !smc_close_sent_any_close(&smc->conn))
303 /* just shutdown, but not yet closed locally */
304 sk->sk_state = SMC_PROCESSABORT;
306 sk->sk_state = SMC_CLOSED;
307 sock_put(sk); /* passive closing */
309 case SMC_APPCLOSEWAIT2:
310 case SMC_PEERFINCLOSEWAIT:
311 sk->sk_state = SMC_CLOSED;
312 sock_put(sk); /* passive closing */
314 case SMC_PEERABORTWAIT:
315 sk->sk_state = SMC_CLOSED;
317 case SMC_PROCESSABORT:
318 /* nothing to do, add tracing in future patch */
323 /* Either some kind of closing has been received: peer_conn_closed,
324 * peer_conn_abort, or peer_done_writing
325 * or the link group of the connection terminates abnormally.
327 static void smc_close_passive_work(struct work_struct *work)
329 struct smc_connection *conn = container_of(work,
330 struct smc_connection,
332 struct smc_sock *smc = container_of(conn, struct smc_sock, conn);
333 struct smc_cdc_conn_state_flags *rxflags;
334 bool release_clcsock = false;
335 struct sock *sk = &smc->sk;
339 old_state = sk->sk_state;
341 rxflags = &conn->local_rx_ctrl.conn_state_flags;
342 if (rxflags->peer_conn_abort) {
343 /* peer has not received all data */
344 smc_close_passive_abort_received(smc);
345 release_sock(&smc->sk);
346 cancel_delayed_work_sync(&conn->tx_work);
351 switch (sk->sk_state) {
353 sk->sk_state = SMC_APPCLOSEWAIT1;
356 sk->sk_state = SMC_APPCLOSEWAIT1;
357 /* postpone sock_put() for passive closing to cover
358 * received SEND_SHUTDOWN as well
361 case SMC_PEERCLOSEWAIT1:
362 if (rxflags->peer_done_writing)
363 sk->sk_state = SMC_PEERCLOSEWAIT2;
365 /* to check for closing */
366 case SMC_PEERCLOSEWAIT2:
367 if (!smc_cdc_rxed_any_close(conn))
369 if (sock_flag(sk, SOCK_DEAD) &&
370 smc_close_sent_any_close(conn)) {
371 /* smc_release has already been called locally */
372 sk->sk_state = SMC_CLOSED;
374 /* just shutdown, but not yet closed locally */
375 sk->sk_state = SMC_APPFINCLOSEWAIT;
377 sock_put(sk); /* passive closing */
379 case SMC_PEERFINCLOSEWAIT:
380 if (smc_cdc_rxed_any_close(conn)) {
381 sk->sk_state = SMC_CLOSED;
382 sock_put(sk); /* passive closing */
385 case SMC_APPCLOSEWAIT1:
386 case SMC_APPCLOSEWAIT2:
387 /* postpone sock_put() for passive closing to cover
388 * received SEND_SHUTDOWN as well
391 case SMC_APPFINCLOSEWAIT:
392 case SMC_PEERABORTWAIT:
393 case SMC_PROCESSABORT:
395 /* nothing to do, add tracing in future patch */
400 sk->sk_data_ready(sk); /* wakeup blocked rcvbuf consumers */
401 sk->sk_write_space(sk); /* wakeup blocked sndbuf producers */
403 if (old_state != sk->sk_state) {
404 sk->sk_state_change(sk);
405 if ((sk->sk_state == SMC_CLOSED) &&
406 (sock_flag(sk, SOCK_DEAD) || !sk->sk_socket)) {
409 release_clcsock = true;
414 smc_clcsock_release(smc);
415 sock_put(sk); /* sock_hold done by schedulers of close_work */
418 int smc_close_shutdown_write(struct smc_sock *smc)
420 struct smc_connection *conn = &smc->conn;
421 struct sock *sk = &smc->sk;
426 timeout = current->flags & PF_EXITING ?
427 0 : sock_flag(sk, SOCK_LINGER) ?
428 sk->sk_lingertime : SMC_MAX_STREAM_WAIT_TIMEOUT;
430 old_state = sk->sk_state;
432 switch (sk->sk_state) {
434 smc_close_stream_wait(smc, timeout);
436 cancel_delayed_work_sync(&conn->tx_work);
438 if (sk->sk_state != SMC_ACTIVE)
440 /* send close wr request */
441 rc = smc_close_wr(conn);
442 sk->sk_state = SMC_PEERCLOSEWAIT1;
444 case SMC_APPCLOSEWAIT1:
446 if (!smc_cdc_rxed_any_close(conn))
447 smc_close_stream_wait(smc, timeout);
449 cancel_delayed_work_sync(&conn->tx_work);
451 if (sk->sk_state != SMC_APPCLOSEWAIT1)
453 /* confirm close from peer */
454 rc = smc_close_wr(conn);
455 sk->sk_state = SMC_APPCLOSEWAIT2;
457 case SMC_APPCLOSEWAIT2:
458 case SMC_PEERFINCLOSEWAIT:
459 case SMC_PEERCLOSEWAIT1:
460 case SMC_PEERCLOSEWAIT2:
461 case SMC_APPFINCLOSEWAIT:
462 case SMC_PROCESSABORT:
463 case SMC_PEERABORTWAIT:
464 /* nothing to do, add tracing in future patch */
468 if (old_state != sk->sk_state)
469 sk->sk_state_change(sk);
473 /* Initialize close properties on connection establishment. */
474 void smc_close_init(struct smc_sock *smc)
476 INIT_WORK(&smc->conn.close_work, smc_close_passive_work);