Lines Matching refs:sock_net
215 dev = dev_get_by_index_rcu(sock_net(sk), skb->skb_iif); in tcp_gro_dev_warn()
428 min(sndmem, sock_net(sk)->ipv4.sysctl_tcp_wmem[2])); in tcp_sndbuf_expand()
462 int window = tcp_win_from_space(sk, sock_net(sk)->ipv4.sysctl_tcp_rmem[2]) >> 1; in __tcp_grow_window()
506 int tcp_app_win = sock_net(sk)->ipv4.sysctl_tcp_app_win; in tcp_init_buffer_space()
545 struct net *net = sock_net(sk); in tcp_clamp_window()
696 if (sock_net(sk)->ipv4.sysctl_tcp_moderate_rcvbuf && in tcp_rcv_space_adjust()
717 sock_net(sk)->ipv4.sysctl_tcp_rmem[2]); in tcp_rcv_space_adjust()
882 rate *= sock_net(sk)->ipv4.sysctl_tcp_pacing_ss_ratio; in tcp_update_pacing_rate()
884 rate *= sock_net(sk)->ipv4.sysctl_tcp_pacing_ca_ratio; in tcp_update_pacing_rate()
1014 sock_net(sk)->ipv4.sysctl_tcp_max_reordering); in tcp_check_sack_reordering()
1019 NET_INC_STATS(sock_net(sk), in tcp_check_sack_reordering()
1059 NET_ADD_STATS(sock_net(sk), LINUX_MIB_TCPLOSTRETRANSMIT, in tcp_mark_skb_lost()
1219 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPDSACKRECV); in tcp_check_dsack()
1226 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPDSACKOFORECV); in tcp_check_dsack()
1233 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPDSACKIGNOREDDUBIOUS); in tcp_check_dsack()
1237 NET_ADD_STATS(sock_net(sk), LINUX_MIB_TCPDSACKRECVSEGS, dup_segs); in tcp_check_dsack()
1440 NET_INC_STATS(sock_net(sk), LINUX_MIB_SACKSHIFTED); in tcp_shifted_skb()
1467 NET_INC_STATS(sock_net(sk), LINUX_MIB_SACKMERGED); in tcp_shifted_skb()
1636 NET_INC_STATS(sock_net(sk), LINUX_MIB_SACKSHIFTFALLBACK); in tcp_shift_skb_data()
1828 NET_INC_STATS(sock_net(sk), mib_idx); in tcp_sacktag_write_queue()
1991 sock_net(sk)->ipv4.sysctl_tcp_max_reordering); in tcp_check_reno_reordering()
1993 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPRENOREORDER); in tcp_check_reno_reordering()
2056 return sock_net(sk)->ipv4.sysctl_tcp_recovery & TCP_RACK_LOSS_DETECTION; in tcp_is_rack()
2072 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPSACKRENEGING); in tcp_timeout_mark_lost()
2098 struct net *net = sock_net(sk); in tcp_enter_loss()
2488 NET_INC_STATS(sock_net(sk), mib_idx); in tcp_try_undo_recovery()
2515 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPDSACKUNDO); in tcp_try_undo_dsack()
2530 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPLOSSUNDO); in tcp_try_undo_loss()
2532 NET_INC_STATS(sock_net(sk), in tcp_try_undo_loss()
2661 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPMTUPFAIL); in tcp_mtup_probe_failed()
2681 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPMTUPSUCCESS); in tcp_mtup_probe_success()
2736 NET_INC_STATS(sock_net(sk), mib_idx); in tcp_enter_recovery()
2829 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPPARTIALUNDO); in tcp_try_undo_partial()
2990 u32 wlen = sock_net(sk)->ipv4.sysctl_tcp_min_rtt_wlen * HZ; in tcp_update_rtt_min()
3400 if (tcp_sk(sk)->reordering > sock_net(sk)->ipv4.sysctl_tcp_reordering) in tcp_may_raise_cwnd()
3548 struct net *net = sock_net(sk); in tcp_send_challenge_ack()
3619 NET_INC_STATS(sock_net(sk), in tcp_process_tlp_ack()
3660 const struct net *net = sock_net(sk); in tcp_newly_delivered()
3748 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPHPACKS); in tcp_ack()
3755 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPPUREACKS); in tcp_ack()
4349 if (tcp_is_sack(tp) && sock_net(sk)->ipv4.sysctl_tcp_dsack) { in tcp_dsack_set()
4357 NET_INC_STATS(sock_net(sk), mib_idx); in tcp_dsack_set()
4384 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPDUPLICATEDATAREHASH); in tcp_rcv_spurious_retrans()
4394 NET_INC_STATS(sock_net(sk), LINUX_MIB_DELAYEDACKLOST); in tcp_send_dupack()
4397 if (tcp_is_sack(tp) && sock_net(sk)->ipv4.sysctl_tcp_dsack) { in tcp_send_dupack()
4453 NET_ADD_STATS(sock_net(sk), LINUX_MIB_TCPACKCOMPRESSED, in tcp_sack_compress_send_ack()
4586 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPRCVCOALESCE); in tcp_try_coalesce()
4704 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPOFODROP); in tcp_data_queue_ofo()
4715 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPOFOQUEUE); in tcp_data_queue_ofo()
4767 NET_INC_STATS(sock_net(sk), in tcp_data_queue_ofo()
4786 NET_INC_STATS(sock_net(sk), in tcp_data_queue_ofo()
4815 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPOFOMERGE); in tcp_data_queue_ofo()
4881 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPRCVQDROP); in tcp_send_rcvq()
4943 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPZEROWINDOWDROP); in tcp_data_queue()
4952 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPRCVQDROP); in tcp_data_queue()
4988 NET_INC_STATS(sock_net(sk), LINUX_MIB_DELAYEDACKLOST); in tcp_data_queue()
5011 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPZEROWINDOWDROP); in tcp_data_queue()
5040 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPRCVCOLLAPSED); in tcp_collapse_one()
5245 NET_INC_STATS(sock_net(sk), LINUX_MIB_OFOPRUNED); in tcp_prune_ofo_queue()
5285 NET_INC_STATS(sock_net(sk), LINUX_MIB_PRUNECALLED); in tcp_prune_queue()
5318 NET_INC_STATS(sock_net(sk), LINUX_MIB_RCVPRUNED); in tcp_prune_queue()
5412 tp->compressed_ack >= sock_net(sk)->ipv4.sysctl_tcp_comp_sack_nr) in __tcp_ack_snd_check()
5433 delay = min_t(unsigned long, sock_net(sk)->ipv4.sysctl_tcp_comp_sack_delay_ns, in __tcp_ack_snd_check()
5437 sock_net(sk)->ipv4.sysctl_tcp_comp_sack_slack_ns, in __tcp_ack_snd_check()
5465 if (ptr && !sock_net(sk)->ipv4.sysctl_tcp_stdurg) in tcp_check_urg()
5578 if (tcp_fast_parse_options(sock_net(sk), skb, th, tp) && in tcp_validate_incoming()
5582 NET_INC_STATS(sock_net(sk), LINUX_MIB_PAWSESTABREJECTED); in tcp_validate_incoming()
5583 if (!tcp_oow_rate_limited(sock_net(sk), skb, in tcp_validate_incoming()
5603 if (!tcp_oow_rate_limited(sock_net(sk), skb, in tcp_validate_incoming()
5666 TCP_INC_STATS(sock_net(sk), TCP_MIB_INERRS); in tcp_validate_incoming()
5667 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPSYNCHALLENGE); in tcp_validate_incoming()
5794 TCP_INC_STATS(sock_net(sk), TCP_MIB_INERRS); in tcp_rcv_established()
5818 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPHPHITS); in tcp_rcv_established()
5876 TCP_INC_STATS(sock_net(sk), TCP_MIB_CSUMERRORS); in tcp_rcv_established()
5877 TCP_INC_STATS(sock_net(sk), TCP_MIB_INERRS); in tcp_rcv_established()
5956 tcp_parse_options(sock_net(sk), synack, &opt, 0, NULL); in tcp_rcv_fastopen_synack()
5990 NET_INC_STATS(sock_net(sk), in tcp_rcv_fastopen_synack()
5996 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPFASTOPENACTIVE); in tcp_rcv_fastopen_synack()
6041 tcp_parse_options(sock_net(sk), skb, &tp->rx_opt, 0, &foc); in tcp_rcv_synsent_state_process()
6067 NET_INC_STATS(sock_net(sk), in tcp_rcv_synsent_state_process()
6452 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPABORTONDATA); in tcp_rcv_state_process()
6461 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPABORTONDATA); in tcp_rcv_state_process()
6522 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPABORTONDATA); in tcp_rcv_state_process()
6585 const struct net *net = sock_net(listen_sk); in tcp_ecn_create_request()
6646 write_pnet(&ireq->ireq_net, sock_net(sk_listener)); in inet_reqsk_alloc()
6662 struct net *net = sock_net(sk); in tcp_syn_flood_action()
6668 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPREQQFULLDOCOOKIES); in tcp_syn_flood_action()
6671 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPREQQFULLDROP); in tcp_syn_flood_action()
6723 if (sock_net(sk)->ipv4.sysctl_tcp_syncookies != 2 && in tcp_get_syncookie_mss()
6731 NET_INC_STATS(sock_net(sk), LINUX_MIB_LISTENOVERFLOWS); in tcp_get_syncookie_mss()
6751 struct net *net = sock_net(sk); in tcp_conn_request()
6770 NET_INC_STATS(sock_net(sk), LINUX_MIB_LISTENOVERFLOWS); in tcp_conn_request()
6788 tcp_parse_options(sock_net(sk), skb, &tmp_opt, 0, in tcp_conn_request()