Lines Matching +full:network +full:- +full:oriented

1 /* SPDX-License-Identifier: GPL-2.0-or-later */
3 * NET Generic infrastructure for INET connection oriented protocols.
56 /** inet_connection_sock - INET connection oriented sock
76 * @icsk_ext_hdr_len: Network protocol overhead (IP/IPv6 options)
79 * @icsk_probes_tstamp: Probe timestamp (cleared by non-zero window ack)
154 return (void *)inet_csk(sk)->icsk_ca_priv; in inet_csk_ca()
179 inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_SCHED; in inet_csk_schedule_ack()
184 return inet_csk(sk)->icsk_ack.pending & ICSK_ACK_SCHED; in inet_csk_ack_scheduled()
189 memset(&inet_csk(sk)->icsk_ack, 0, sizeof(inet_csk(sk)->icsk_ack)); in inet_csk_delack_init()
200 smp_store_release(&icsk->icsk_pending, 0); in inet_csk_clear_xmit_timer()
202 sk_stop_timer(sk, &icsk->icsk_retransmit_timer); in inet_csk_clear_xmit_timer()
205 smp_store_release(&icsk->icsk_ack.pending, 0); in inet_csk_clear_xmit_timer()
206 icsk->icsk_ack.retry = 0; in inet_csk_clear_xmit_timer()
208 sk_stop_timer(sk, &icsk->icsk_delack_timer); in inet_csk_clear_xmit_timer()
232 smp_store_release(&icsk->icsk_pending, what); in inet_csk_reset_xmit_timer()
233 icsk->icsk_timeout = jiffies + when; in inet_csk_reset_xmit_timer()
234 sk_reset_timer(sk, &icsk->icsk_retransmit_timer, icsk->icsk_timeout); in inet_csk_reset_xmit_timer()
236 smp_store_release(&icsk->icsk_ack.pending, in inet_csk_reset_xmit_timer()
237 icsk->icsk_ack.pending | ICSK_ACK_TIMER); in inet_csk_reset_xmit_timer()
238 icsk->icsk_ack.timeout = jiffies + when; in inet_csk_reset_xmit_timer()
239 sk_reset_timer(sk, &icsk->icsk_delack_timer, icsk->icsk_ack.timeout); in inet_csk_reset_xmit_timer()
249 u64 when = (u64)icsk->icsk_rto << icsk->icsk_backoff; in inet_csk_rto_backoff()
275 reqsk_queue_added(&inet_csk(sk)->icsk_accept_queue); in inet_csk_reqsk_queue_added()
280 return reqsk_queue_len(&inet_csk(sk)->icsk_accept_queue); in inet_csk_reqsk_queue_len()
285 return inet_csk_reqsk_queue_len(sk) > READ_ONCE(sk->sk_max_ack_backlog); in inet_csk_reqsk_queue_is_full()
294 u64 timeout = (u64)req->timeout << req->num_timeout; in reqsk_timeout()
303 this_cpu_inc(*sk->sk_prot->orphan_count); in inet_csk_prepare_for_destroy_sock()
314 return !reqsk_queue_empty(&inet_csk(sk)->icsk_accept_queue) ? in inet_csk_listen_poll()
331 inet_csk(sk)->icsk_ack.pingpong = in inet_csk_enter_pingpong_mode()
332 READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_pingpong_thresh); in inet_csk_enter_pingpong_mode()
337 inet_csk(sk)->icsk_ack.pingpong = 0; in inet_csk_exit_pingpong_mode()
342 return inet_csk(sk)->icsk_ack.pingpong >= in inet_csk_in_pingpong_mode()
343 READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_pingpong_thresh); in inet_csk_in_pingpong_mode()
350 if (icsk->icsk_ack.pingpong < U8_MAX) in inet_csk_inc_pingpong_cnt()
351 icsk->icsk_ack.pingpong++; in inet_csk_inc_pingpong_cnt()
356 return inet_test_bit(IS_ICSK, sk) && !!inet_csk(sk)->icsk_ulp_ops; in inet_csk_has_ulp()
363 spin_lock_init(&icsk->icsk_accept_queue.rskq_lock); in inet_init_csk_locks()
364 spin_lock_init(&icsk->icsk_accept_queue.fastopenq.lock); in inet_init_csk_locks()