@@ -429,11 +429,11 @@ static void mptcp_check_data_fin_ack(struct sock *sk)
429
429
430
430
switch (sk -> sk_state ) {
431
431
case TCP_FIN_WAIT1 :
432
- inet_sk_state_store (sk , TCP_FIN_WAIT2 );
432
+ mptcp_set_state (sk , TCP_FIN_WAIT2 );
433
433
break ;
434
434
case TCP_CLOSING :
435
435
case TCP_LAST_ACK :
436
- inet_sk_state_store (sk , TCP_CLOSE );
436
+ mptcp_set_state (sk , TCP_CLOSE );
437
437
break ;
438
438
}
439
439
@@ -594,13 +594,13 @@ static bool mptcp_check_data_fin(struct sock *sk)
594
594
595
595
switch (sk -> sk_state ) {
596
596
case TCP_ESTABLISHED :
597
- inet_sk_state_store (sk , TCP_CLOSE_WAIT );
597
+ mptcp_set_state (sk , TCP_CLOSE_WAIT );
598
598
break ;
599
599
case TCP_FIN_WAIT1 :
600
- inet_sk_state_store (sk , TCP_CLOSING );
600
+ mptcp_set_state (sk , TCP_CLOSING );
601
601
break ;
602
602
case TCP_FIN_WAIT2 :
603
- inet_sk_state_store (sk , TCP_CLOSE );
603
+ mptcp_set_state (sk , TCP_CLOSE );
604
604
break ;
605
605
default :
606
606
/* Other states not expected */
@@ -775,7 +775,7 @@ static bool __mptcp_subflow_error_report(struct sock *sk, struct sock *ssk)
775
775
*/
776
776
ssk_state = inet_sk_state_load (ssk );
777
777
if (ssk_state == TCP_CLOSE && !sock_flag (sk , SOCK_DEAD ))
778
- inet_sk_state_store (sk , ssk_state );
778
+ mptcp_set_state (sk , ssk_state );
779
779
WRITE_ONCE (sk -> sk_err , - err );
780
780
781
781
/* This barrier is coupled with smp_rmb() in mptcp_poll() */
@@ -2463,7 +2463,7 @@ static void __mptcp_close_ssk(struct sock *sk, struct sock *ssk,
2463
2463
inet_sk_state_load (msk -> first ) == TCP_CLOSE ) {
2464
2464
if (sk -> sk_state != TCP_ESTABLISHED ||
2465
2465
msk -> in_accept_queue || sock_flag (sk , SOCK_DEAD )) {
2466
- inet_sk_state_store (sk , TCP_CLOSE );
2466
+ mptcp_set_state (sk , TCP_CLOSE );
2467
2467
mptcp_close_wake_up (sk );
2468
2468
} else {
2469
2469
mptcp_start_tout_timer (sk );
@@ -2558,7 +2558,7 @@ static void mptcp_check_fastclose(struct mptcp_sock *msk)
2558
2558
WRITE_ONCE (sk -> sk_err , ECONNRESET );
2559
2559
}
2560
2560
2561
- inet_sk_state_store (sk , TCP_CLOSE );
2561
+ mptcp_set_state (sk , TCP_CLOSE );
2562
2562
WRITE_ONCE (sk -> sk_shutdown , SHUTDOWN_MASK );
2563
2563
smp_mb__before_atomic (); /* SHUTDOWN must be visible first */
2564
2564
set_bit (MPTCP_WORK_CLOSE_SUBFLOW , & msk -> flags );
@@ -2693,7 +2693,7 @@ static void mptcp_do_fastclose(struct sock *sk)
2693
2693
struct mptcp_subflow_context * subflow , * tmp ;
2694
2694
struct mptcp_sock * msk = mptcp_sk (sk );
2695
2695
2696
- inet_sk_state_store (sk , TCP_CLOSE );
2696
+ mptcp_set_state (sk , TCP_CLOSE );
2697
2697
mptcp_for_each_subflow_safe (msk , subflow , tmp )
2698
2698
__mptcp_close_ssk (sk , mptcp_subflow_tcp_sock (subflow ),
2699
2699
subflow , MPTCP_CF_FASTCLOSE );
@@ -2911,7 +2911,7 @@ static int mptcp_close_state(struct sock *sk)
2911
2911
int next = (int )new_state [sk -> sk_state ];
2912
2912
int ns = next & TCP_STATE_MASK ;
2913
2913
2914
- inet_sk_state_store (sk , ns );
2914
+ mptcp_set_state (sk , ns );
2915
2915
2916
2916
return next & TCP_ACTION_FIN ;
2917
2917
}
@@ -3022,7 +3022,7 @@ bool __mptcp_close(struct sock *sk, long timeout)
3022
3022
3023
3023
if ((1 << sk -> sk_state ) & (TCPF_LISTEN | TCPF_CLOSE )) {
3024
3024
mptcp_check_listen_stop (sk );
3025
- inet_sk_state_store (sk , TCP_CLOSE );
3025
+ mptcp_set_state (sk , TCP_CLOSE );
3026
3026
goto cleanup ;
3027
3027
}
3028
3028
@@ -3065,7 +3065,7 @@ bool __mptcp_close(struct sock *sk, long timeout)
3065
3065
* state, let's not keep resources busy for no reasons
3066
3066
*/
3067
3067
if (subflows_alive == 0 )
3068
- inet_sk_state_store (sk , TCP_CLOSE );
3068
+ mptcp_set_state (sk , TCP_CLOSE );
3069
3069
3070
3070
sock_hold (sk );
3071
3071
pr_debug ("msk=%p state=%d" , sk , sk -> sk_state );
@@ -3131,7 +3131,7 @@ static int mptcp_disconnect(struct sock *sk, int flags)
3131
3131
return - EBUSY ;
3132
3132
3133
3133
mptcp_check_listen_stop (sk );
3134
- inet_sk_state_store (sk , TCP_CLOSE );
3134
+ mptcp_set_state (sk , TCP_CLOSE );
3135
3135
3136
3136
mptcp_stop_rtx_timer (sk );
3137
3137
mptcp_stop_tout_timer (sk );
@@ -3219,7 +3219,7 @@ struct sock *mptcp_sk_clone_init(const struct sock *sk,
3219
3219
/* this can't race with mptcp_close(), as the msk is
3220
3220
* not yet exposted to user-space
3221
3221
*/
3222
- inet_sk_state_store (nsk , TCP_ESTABLISHED );
3222
+ mptcp_set_state (nsk , TCP_ESTABLISHED );
3223
3223
3224
3224
/* The msk maintain a ref to each subflow in the connections list */
3225
3225
WRITE_ONCE (msk -> first , ssk );
@@ -3640,7 +3640,7 @@ static int mptcp_connect(struct sock *sk, struct sockaddr *uaddr, int addr_len)
3640
3640
if (IS_ERR (ssk ))
3641
3641
return PTR_ERR (ssk );
3642
3642
3643
- inet_sk_state_store (sk , TCP_SYN_SENT );
3643
+ mptcp_set_state (sk , TCP_SYN_SENT );
3644
3644
subflow = mptcp_subflow_ctx (ssk );
3645
3645
#ifdef CONFIG_TCP_MD5SIG
3646
3646
/* no MPTCP if MD5SIG is enabled on this socket or we may run out of
@@ -3690,7 +3690,7 @@ static int mptcp_connect(struct sock *sk, struct sockaddr *uaddr, int addr_len)
3690
3690
if (unlikely (err )) {
3691
3691
/* avoid leaving a dangling token in an unconnected socket */
3692
3692
mptcp_token_destroy (msk );
3693
- inet_sk_state_store (sk , TCP_CLOSE );
3693
+ mptcp_set_state (sk , TCP_CLOSE );
3694
3694
return err ;
3695
3695
}
3696
3696
@@ -3779,13 +3779,13 @@ static int mptcp_listen(struct socket *sock, int backlog)
3779
3779
goto unlock ;
3780
3780
}
3781
3781
3782
- inet_sk_state_store (sk , TCP_LISTEN );
3782
+ mptcp_set_state (sk , TCP_LISTEN );
3783
3783
sock_set_flag (sk , SOCK_RCU_FREE );
3784
3784
3785
3785
lock_sock (ssk );
3786
3786
err = __inet_listen_sk (ssk , backlog );
3787
3787
release_sock (ssk );
3788
- inet_sk_state_store (sk , inet_sk_state_load (ssk ));
3788
+ mptcp_set_state (sk , inet_sk_state_load (ssk ));
3789
3789
3790
3790
if (!err ) {
3791
3791
sock_prot_inuse_add (sock_net (sk ), sk -> sk_prot , 1 );
@@ -3863,7 +3863,7 @@ static int mptcp_stream_accept(struct socket *sock, struct socket *newsock,
3863
3863
__mptcp_close_ssk (newsk , msk -> first ,
3864
3864
mptcp_subflow_ctx (msk -> first ), 0 );
3865
3865
if (unlikely (list_is_singular (& msk -> conn_list )))
3866
- inet_sk_state_store (newsk , TCP_CLOSE );
3866
+ mptcp_set_state (newsk , TCP_CLOSE );
3867
3867
}
3868
3868
} else {
3869
3869
MPTCP_INC_STATS (sock_net (ssk ),
0 commit comments