@@ -405,7 +405,7 @@ static bool __mptcp_move_skb(struct mptcp_sock *msk, struct sock *ssk,
405
405
return false;
406
406
}
407
407
408
- static void mptcp_stop_timer (struct sock * sk )
408
+ static void mptcp_stop_rtx_timer (struct sock * sk )
409
409
{
410
410
struct inet_connection_sock * icsk = inet_csk (sk );
411
411
@@ -911,12 +911,12 @@ static void __mptcp_flush_join_list(struct sock *sk, struct list_head *join_list
911
911
}
912
912
}
913
913
914
- static bool mptcp_timer_pending (struct sock * sk )
914
+ static bool mptcp_rtx_timer_pending (struct sock * sk )
915
915
{
916
916
return timer_pending (& inet_csk (sk )-> icsk_retransmit_timer );
917
917
}
918
918
919
- static void mptcp_reset_timer (struct sock * sk )
919
+ static void mptcp_reset_rtx_timer (struct sock * sk )
920
920
{
921
921
struct inet_connection_sock * icsk = inet_csk (sk );
922
922
unsigned long tout ;
@@ -1050,10 +1050,10 @@ static void __mptcp_clean_una(struct sock *sk)
1050
1050
out :
1051
1051
if (snd_una == READ_ONCE (msk -> snd_nxt ) &&
1052
1052
snd_una == READ_ONCE (msk -> write_seq )) {
1053
- if (mptcp_timer_pending (sk ) && !mptcp_data_fin_enabled (msk ))
1054
- mptcp_stop_timer (sk );
1053
+ if (mptcp_rtx_timer_pending (sk ) && !mptcp_data_fin_enabled (msk ))
1054
+ mptcp_stop_rtx_timer (sk );
1055
1055
} else {
1056
- mptcp_reset_timer (sk );
1056
+ mptcp_reset_rtx_timer (sk );
1057
1057
}
1058
1058
}
1059
1059
@@ -1626,8 +1626,8 @@ void __mptcp_push_pending(struct sock *sk, unsigned int flags)
1626
1626
mptcp_push_release (ssk , & info );
1627
1627
1628
1628
/* ensure the rtx timer is running */
1629
- if (!mptcp_timer_pending (sk ))
1630
- mptcp_reset_timer (sk );
1629
+ if (!mptcp_rtx_timer_pending (sk ))
1630
+ mptcp_reset_rtx_timer (sk );
1631
1631
if (do_check_data_fin )
1632
1632
mptcp_check_send_data_fin (sk );
1633
1633
}
@@ -1690,8 +1690,8 @@ static void __mptcp_subflow_push_pending(struct sock *sk, struct sock *ssk, bool
1690
1690
if (copied ) {
1691
1691
tcp_push (ssk , 0 , info .mss_now , tcp_sk (ssk )-> nonagle ,
1692
1692
info .size_goal );
1693
- if (!mptcp_timer_pending (sk ))
1694
- mptcp_reset_timer (sk );
1693
+ if (!mptcp_rtx_timer_pending (sk ))
1694
+ mptcp_reset_rtx_timer (sk );
1695
1695
1696
1696
if (msk -> snd_data_fin_enable &&
1697
1697
msk -> snd_nxt + 1 == msk -> write_seq )
@@ -2260,7 +2260,7 @@ static void mptcp_retransmit_timer(struct timer_list *t)
2260
2260
sock_put (sk );
2261
2261
}
2262
2262
2263
- static void mptcp_timeout_timer (struct timer_list * t )
2263
+ static void mptcp_tout_timer (struct timer_list * t )
2264
2264
{
2265
2265
struct sock * sk = from_timer (sk , t , sk_timer );
2266
2266
@@ -2629,14 +2629,14 @@ static void __mptcp_retrans(struct sock *sk)
2629
2629
reset_timer :
2630
2630
mptcp_check_and_set_pending (sk );
2631
2631
2632
- if (!mptcp_timer_pending (sk ))
2633
- mptcp_reset_timer (sk );
2632
+ if (!mptcp_rtx_timer_pending (sk ))
2633
+ mptcp_reset_rtx_timer (sk );
2634
2634
}
2635
2635
2636
2636
/* schedule the timeout timer for the relevant event: either close timeout
2637
2637
* or mp_fail timeout. The close timeout takes precedence on the mp_fail one
2638
2638
*/
2639
- void mptcp_reset_timeout (struct mptcp_sock * msk , unsigned long fail_tout )
2639
+ void mptcp_reset_tout_timer (struct mptcp_sock * msk , unsigned long fail_tout )
2640
2640
{
2641
2641
struct sock * sk = (struct sock * )msk ;
2642
2642
unsigned long timeout , close_timeout ;
@@ -2669,7 +2669,7 @@ static void mptcp_mp_fail_no_response(struct mptcp_sock *msk)
2669
2669
WRITE_ONCE (mptcp_subflow_ctx (ssk )-> fail_tout , 0 );
2670
2670
unlock_sock_fast (ssk , slow );
2671
2671
2672
- mptcp_reset_timeout (msk , 0 );
2672
+ mptcp_reset_tout_timer (msk , 0 );
2673
2673
}
2674
2674
2675
2675
static void mptcp_do_fastclose (struct sock * sk )
@@ -2758,7 +2758,7 @@ static void __mptcp_init_sock(struct sock *sk)
2758
2758
2759
2759
/* re-use the csk retrans timer for MPTCP-level retrans */
2760
2760
timer_setup (& msk -> sk .icsk_retransmit_timer , mptcp_retransmit_timer , 0 );
2761
- timer_setup (& sk -> sk_timer , mptcp_timeout_timer , 0 );
2761
+ timer_setup (& sk -> sk_timer , mptcp_tout_timer , 0 );
2762
2762
}
2763
2763
2764
2764
static void mptcp_ca_reset (struct sock * sk )
@@ -2849,8 +2849,8 @@ void mptcp_subflow_shutdown(struct sock *sk, struct sock *ssk, int how)
2849
2849
} else {
2850
2850
pr_debug ("Sending DATA_FIN on subflow %p" , ssk );
2851
2851
tcp_send_ack (ssk );
2852
- if (!mptcp_timer_pending (sk ))
2853
- mptcp_reset_timer (sk );
2852
+ if (!mptcp_rtx_timer_pending (sk ))
2853
+ mptcp_reset_rtx_timer (sk );
2854
2854
}
2855
2855
break ;
2856
2856
}
@@ -2933,7 +2933,7 @@ static void __mptcp_destroy_sock(struct sock *sk)
2933
2933
2934
2934
might_sleep ();
2935
2935
2936
- mptcp_stop_timer (sk );
2936
+ mptcp_stop_rtx_timer (sk );
2937
2937
sk_stop_timer (sk , & sk -> sk_timer );
2938
2938
msk -> pm .status = 0 ;
2939
2939
mptcp_release_sched (msk );
@@ -3053,7 +3053,7 @@ bool __mptcp_close(struct sock *sk, long timeout)
3053
3053
__mptcp_destroy_sock (sk );
3054
3054
do_cancel_work = true;
3055
3055
} else {
3056
- mptcp_reset_timeout (msk , 0 );
3056
+ mptcp_reset_tout_timer (msk , 0 );
3057
3057
}
3058
3058
3059
3059
return do_cancel_work ;
@@ -3116,7 +3116,7 @@ static int mptcp_disconnect(struct sock *sk, int flags)
3116
3116
mptcp_check_listen_stop (sk );
3117
3117
inet_sk_state_store (sk , TCP_CLOSE );
3118
3118
3119
- mptcp_stop_timer (sk );
3119
+ mptcp_stop_rtx_timer (sk );
3120
3120
sk_stop_timer (sk , & sk -> sk_timer );
3121
3121
3122
3122
if (msk -> token )
0 commit comments