Lines Matching refs:tcp_sock

247 				struct tcp_sock *tp = tcp_sk(sk);  in tcp_measure_rcv_mss()
340 static void tcp_ecn_queue_cwr(struct tcp_sock *tp) in tcp_ecn_queue_cwr()
360 static void tcp_ecn_withdraw_cwr(struct tcp_sock *tp) in tcp_ecn_withdraw_cwr()
367 struct tcp_sock *tp = tcp_sk(sk); in __tcp_ecn_check_ce()
403 static void tcp_ecn_rcv_synack(struct tcp_sock *tp, const struct tcphdr *th) in tcp_ecn_rcv_synack()
409 static void tcp_ecn_rcv_syn(struct tcp_sock *tp, const struct tcphdr *th) in tcp_ecn_rcv_syn()
415 static bool tcp_ecn_rcv_ecn_echo(const struct tcp_sock *tp, const struct tcphdr *th) in tcp_ecn_rcv_ecn_echo()
429 const struct tcp_sock *tp = tcp_sk(sk); in tcp_sndbuf_expand()
488 const struct tcp_sock *tp = tcp_sk(sk); in __tcp_grow_window()
525 struct tcp_sock *tp = tcp_sk(sk); in tcp_grow_window()
565 struct tcp_sock *tp = tcp_sk(sk); in tcp_init_buffer_space()
602 struct tcp_sock *tp = tcp_sk(sk); in tcp_clamp_window()
630 const struct tcp_sock *tp = tcp_sk(sk); in tcp_initialize_rcv_mss()
652 static void tcp_rcv_rtt_update(struct tcp_sock *tp, u32 sample, int win_dep) in tcp_rcv_rtt_update()
684 static inline void tcp_rcv_rtt_measure(struct tcp_sock *tp) in tcp_rcv_rtt_measure()
702 static s32 tcp_rtt_tsopt_us(const struct tcp_sock *tp) in tcp_rtt_tsopt_us()
722 struct tcp_sock *tp = tcp_sk(sk); in tcp_rcv_rtt_measure_ts()
743 struct tcp_sock *tp = tcp_sk(sk); in tcp_rcv_space_adjust()
822 struct tcp_sock *tp = tcp_sk(sk); in tcp_event_data_recv()
877 struct tcp_sock *tp = tcp_sk(sk); in tcp_rtt_estimator()
945 const struct tcp_sock *tp = tcp_sk(sk); in tcp_update_pacing_rate()
982 const struct tcp_sock *tp = tcp_sk(sk); in tcp_set_rto()
1007 __u32 tcp_init_cwnd(const struct tcp_sock *tp, const struct dst_entry *dst) in tcp_init_cwnd()
1036 static u32 tcp_dsack_seen(struct tcp_sock *tp, u32 start_seq, in tcp_dsack_seen()
1082 struct tcp_sock *tp = tcp_sk(sk); in tcp_check_sack_reordering()
1115 static void tcp_verify_retransmit_hint(struct tcp_sock *tp, struct sk_buff *skb) in tcp_verify_retransmit_hint()
1127 static void tcp_notify_skb_loss_event(struct tcp_sock *tp, const struct sk_buff *skb) in tcp_notify_skb_loss_event()
1135 struct tcp_sock *tp = tcp_sk(sk); in tcp_mark_skb_lost()
1158 static void tcp_count_delivered(struct tcp_sock *tp, u32 delivered, in tcp_count_delivered()
1259 static bool tcp_is_sackblock_valid(struct tcp_sock *tp, bool is_dsack, in tcp_is_sackblock_valid()
1300 struct tcp_sock *tp = tcp_sk(sk); in tcp_check_dsack()
1398 struct tcp_sock *tp = tcp_sk(sk); in tcp_sacktag_one()
1483 struct tcp_sock *tp = tcp_sk(sk); in tcp_shifted_skb()
1596 struct tcp_sock *tp = tcp_sk(sk); in tcp_shift_skb_data()
1735 struct tcp_sock *tp = tcp_sk(sk); in tcp_sacktag_walk()
1850 static int tcp_sack_cache_ok(const struct tcp_sock *tp, const struct tcp_sack_block *cache) in tcp_sack_cache_ok()
1859 struct tcp_sock *tp = tcp_sk(sk); in tcp_sacktag_write_queue()
2054 static bool tcp_limit_reno_sacked(struct tcp_sock *tp) in tcp_limit_reno_sacked()
2074 struct tcp_sock *tp = tcp_sk(sk); in tcp_check_reno_reordering()
2090 struct tcp_sock *tp = tcp_sk(sk); in tcp_add_reno_sack()
2107 struct tcp_sock *tp = tcp_sk(sk); in tcp_remove_reno_sacks()
2122 static inline void tcp_reset_reno_sack(struct tcp_sock *tp) in tcp_reset_reno_sack()
2127 void tcp_clear_retrans(struct tcp_sock *tp) in tcp_clear_retrans()
2140 static inline void tcp_init_undo(struct tcp_sock *tp) in tcp_init_undo()
2167 struct tcp_sock *tp = tcp_sk(sk); in tcp_timeout_mark_lost()
2199 struct tcp_sock *tp = tcp_sk(sk); in tcp_enter_loss()
2257 struct tcp_sock *tp = tcp_sk(sk); in tcp_check_sack_reneging()
2280 static inline int tcp_dupack_heuristics(const struct tcp_sock *tp) in tcp_dupack_heuristics()
2384 struct tcp_sock *tp = tcp_sk(sk); in tcp_time_to_recover()
2404 struct tcp_sock *tp = tcp_sk(sk); in tcp_mark_head_lost()
2450 struct tcp_sock *tp = tcp_sk(sk); in tcp_update_scoreboard()
2461 static bool tcp_tsopt_ecr_before(const struct tcp_sock *tp, u32 when) in tcp_tsopt_ecr_before()
2470 static bool tcp_skb_spurious_retrans(const struct tcp_sock *tp, in tcp_skb_spurious_retrans()
2480 static inline bool tcp_packet_delayed(const struct tcp_sock *tp) in tcp_packet_delayed()
2518 const struct tcp_sock *tp = tcp_sk(sk); in tcp_any_retrans_done()
2544 struct tcp_sock *tp = tcp_sk(sk); in DBGUNDO()
2570 struct tcp_sock *tp = tcp_sk(sk); in tcp_undo_cwnd_reduction()
2597 static inline bool tcp_may_undo(const struct tcp_sock *tp) in tcp_may_undo()
2604 struct tcp_sock *tp = tcp_sk(sk); in tcp_is_non_sack_preventing_reopen()
2620 struct tcp_sock *tp = tcp_sk(sk); in tcp_try_undo_recovery()
2649 struct tcp_sock *tp = tcp_sk(sk); in tcp_try_undo_dsack()
2665 struct tcp_sock *tp = tcp_sk(sk); in tcp_try_undo_loss()
2698 struct tcp_sock *tp = tcp_sk(sk); in tcp_init_cwnd_reduction()
2712 struct tcp_sock *tp = tcp_sk(sk); in tcp_cwnd_reduction()
2738 struct tcp_sock *tp = tcp_sk(sk); in tcp_end_cwnd_reduction()
2755 struct tcp_sock *tp = tcp_sk(sk); in tcp_enter_cwr()
2768 struct tcp_sock *tp = tcp_sk(sk); in tcp_try_keep_open()
2782 struct tcp_sock *tp = tcp_sk(sk); in tcp_try_to_open()
2808 struct tcp_sock *tp = tcp_sk(sk); in tcp_mtup_probe_success()
2842 struct tcp_sock *tp = tcp_sk(sk); in tcp_non_congestion_loss_retransmit()
2860 struct tcp_sock *tp = tcp_sk(sk); in tcp_simple_retransmit()
2905 struct tcp_sock *tp = tcp_sk(sk); in tcp_enter_recovery()
2929 static void tcp_update_rto_time(struct tcp_sock *tp) in tcp_update_rto_time()
2943 struct tcp_sock *tp = tcp_sk(sk); in tcp_process_loss()
2995 struct tcp_sock *tp = tcp_sk(sk); in tcp_force_fast_retransmit()
3005 struct tcp_sock *tp = tcp_sk(sk); in tcp_try_undo_partial()
3037 struct tcp_sock *tp = tcp_sk(sk); in tcp_identify_packet_loss()
3070 struct tcp_sock *tp = tcp_sk(sk); in tcp_fastretrans_alert()
3189 struct tcp_sock *tp = tcp_sk(sk); in tcp_update_rtt_min()
3206 const struct tcp_sock *tp = tcp_sk(sk); in tcp_ack_update_rtt()
3270 struct tcp_sock *tp = tcp_sk(sk); in tcp_rearm_rto()
3306 struct tcp_sock *tp = tcp_sk(sk); in tcp_tso_acked()
3352 struct tcp_sock *tp = tcp_sk(sk); in tcp_clean_rtx_queue()
3556 const struct tcp_sock *tp = tcp_sk(sk); in tcp_ack_probe()
3626 static inline bool tcp_may_update_window(const struct tcp_sock *tp, in tcp_may_update_window()
3635 static void tcp_snd_sne_update(struct tcp_sock *tp, u32 ack) in tcp_snd_sne_update()
3653 static void tcp_snd_una_update(struct tcp_sock *tp, u32 ack) in tcp_snd_una_update()
3663 static void tcp_rcv_sne_update(struct tcp_sock *tp, u32 seq) in tcp_rcv_sne_update()
3681 static void tcp_rcv_nxt_update(struct tcp_sock *tp, u32 seq) in tcp_rcv_nxt_update()
3699 struct tcp_sock *tp = tcp_sk(sk); in tcp_ack_update_window()
3779 struct tcp_sock *tp = tcp_sk(sk); in tcp_send_challenge_ack()
3811 static void tcp_store_ts_recent(struct tcp_sock *tp) in tcp_store_ts_recent()
3817 static void tcp_replace_ts_recent(struct tcp_sock *tp, u32 seq) in tcp_replace_ts_recent()
3837 struct tcp_sock *tp = tcp_sk(sk); in tcp_process_tlp_ack()
3879 struct tcp_sock *tp = tcp_sk(sk); in tcp_xmit_recovery()
3898 struct tcp_sock *tp = tcp_sk(sk); in tcp_newly_delivered()
3913 struct tcp_sock *tp = tcp_sk(sk); in tcp_ack()
4320 static bool tcp_parse_aligned_timestamp(struct tcp_sock *tp, const struct tcphdr *th) in tcp_parse_aligned_timestamp()
4344 const struct tcphdr *th, struct tcp_sock *tp) in tcp_fast_parse_options()
4462 const struct tcp_sock *tp = tcp_sk(sk); in tcp_disordered_ack_check()
4508 static enum skb_drop_reason tcp_sequence(const struct tcp_sock *tp, in tcp_sequence()
4580 struct tcp_sock *tp = tcp_sk(sk); in tcp_fin()
4661 struct tcp_sock *tp = tcp_sk(sk); in tcp_dsack_set()
4681 struct tcp_sock *tp = tcp_sk(sk); in tcp_dsack_extend()
4714 struct tcp_sock *tp = tcp_sk(sk); in tcp_send_dupack()
4737 static void tcp_sack_maybe_coalesce(struct tcp_sock *tp) in tcp_sack_maybe_coalesce()
4765 struct tcp_sock *tp = tcp_sk(sk); in tcp_sack_compress_send_ack()
4792 struct tcp_sock *tp = tcp_sk(sk); in tcp_sack_new_ofo_skb()
4839 static void tcp_sack_remove(struct tcp_sock *tp) in tcp_sack_remove()
4948 struct tcp_sock *tp = tcp_sk(sk); in tcp_ofo_queue()
5015 struct tcp_sock *tp = tcp_sk(sk); in tcp_data_queue_ofo()
5237 struct tcp_sock *tp = tcp_sk(sk); in tcp_data_queue()
5527 struct tcp_sock *tp = tcp_sk(sk); in tcp_collapse_ofo_queue()
5587 struct tcp_sock *tp = tcp_sk(sk); in tcp_prune_ofo_queue()
5641 struct tcp_sock *tp = tcp_sk(sk); in tcp_prune_queue()
5684 const struct tcp_sock *tp = tcp_sk(sk); in tcp_should_expand_sndbuf()
5719 struct tcp_sock *tp = tcp_sk(sk); in tcp_new_space()
5762 struct tcp_sock *tp = tcp_sk(sk); in __tcp_ack_snd_check()
5848 struct tcp_sock *tp = tcp_sk(sk); in tcp_check_urg()
5914 struct tcp_sock *tp = tcp_sk(sk); in tcp_urg()
5947 const struct tcp_sock *tp = tcp_sk(sk); in tcp_reset_check()
5960 struct tcp_sock *tp = tcp_sk(sk); in tcp_validate_incoming()
6120 struct tcp_sock *tp = tcp_sk(sk); in tcp_rcv_established()
6308 struct tcp_sock *tp = tcp_sk(sk); in tcp_init_transfer()
6335 struct tcp_sock *tp = tcp_sk(sk); in tcp_finish_connect()
6367 struct tcp_sock *tp = tcp_sk(sk); in tcp_rcv_fastopen_synack()
6427 static void smc_check_reset_syn(struct tcp_sock *tp) in smc_check_reset_syn()
6439 struct tcp_sock *tp = tcp_sk(sk); in tcp_try_undo_spurious_syn()
6456 struct tcp_sock *tp = tcp_sk(sk); in tcp_rcv_synsent_state_process()
6702 struct tcp_sock *tp = tcp_sk(sk); in tcp_rcv_synrecv_state_fastopen()
6751 struct tcp_sock *tp = tcp_sk(sk); in tcp_rcv_state_process()
7178 struct tcp_sock *tp = tcp_sk(sk); in tcp_get_syncookie_mss()
7207 struct tcp_sock *tp = tcp_sk(sk); in tcp_conn_request()