Lines Matching refs:sock_net
145 dev = dev_get_by_index_rcu(sock_net(sk), skb->skb_iif); in tcp_gro_dev_warn()
356 sk->sk_sndbuf = min(sndmem, sock_net(sk)->ipv4.sysctl_tcp_wmem[2]); in tcp_sndbuf_expand()
390 int window = tcp_win_from_space(sk, sock_net(sk)->ipv4.sysctl_tcp_rmem[2]) >> 1; in __tcp_grow_window()
441 if (sock_net(sk)->ipv4.sysctl_tcp_moderate_rcvbuf) in tcp_fixup_rcvbuf()
445 sk->sk_rcvbuf = min(rcvmem, sock_net(sk)->ipv4.sysctl_tcp_rmem[2]); in tcp_fixup_rcvbuf()
453 int tcp_app_win = sock_net(sk)->ipv4.sysctl_tcp_app_win; in tcp_init_buffer_space()
493 struct net *net = sock_net(sk); in tcp_clamp_window()
641 if (sock_net(sk)->ipv4.sysctl_tcp_moderate_rcvbuf && in tcp_rcv_space_adjust()
662 sock_net(sk)->ipv4.sysctl_tcp_rmem[2]); in tcp_rcv_space_adjust()
823 rate *= sock_net(sk)->ipv4.sysctl_tcp_pacing_ss_ratio; in tcp_update_pacing_rate()
825 rate *= sock_net(sk)->ipv4.sysctl_tcp_pacing_ca_ratio; in tcp_update_pacing_rate()
913 sock_net(sk)->ipv4.sysctl_tcp_max_reordering); in tcp_check_sack_reordering()
918 NET_INC_STATS(sock_net(sk), in tcp_check_sack_reordering()
1111 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPDSACKRECV); in tcp_check_dsack()
1120 NET_INC_STATS(sock_net(sk), in tcp_check_dsack()
1338 NET_INC_STATS(sock_net(sk), LINUX_MIB_SACKSHIFTED); in tcp_shifted_skb()
1365 NET_INC_STATS(sock_net(sk), LINUX_MIB_SACKMERGED); in tcp_shifted_skb()
1520 NET_INC_STATS(sock_net(sk), LINUX_MIB_SACKSHIFTFALLBACK); in tcp_shift_skb_data()
1719 NET_INC_STATS(sock_net(sk), mib_idx); in tcp_sacktag_write_queue()
1880 sock_net(sk)->ipv4.sysctl_tcp_max_reordering); in tcp_check_reno_reordering()
1882 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPRENOREORDER); in tcp_check_reno_reordering()
1940 return sock_net(sk)->ipv4.sysctl_tcp_recovery & TCP_RACK_LOSS_DETECTION; in tcp_is_rack()
1956 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPSACKRENEGING); in tcp_timeout_mark_lost()
1982 struct net *net = sock_net(sk); in tcp_enter_loss()
2388 NET_INC_STATS(sock_net(sk), mib_idx); in tcp_try_undo_recovery()
2415 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPDSACKUNDO); in tcp_try_undo_dsack()
2430 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPLOSSUNDO); in tcp_try_undo_loss()
2432 NET_INC_STATS(sock_net(sk), in tcp_try_undo_loss()
2561 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPMTUPFAIL); in tcp_mtup_probe_failed()
2581 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPMTUPSUCCESS); in tcp_mtup_probe_success()
2642 NET_INC_STATS(sock_net(sk), mib_idx); in tcp_enter_recovery()
2735 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPPARTIALUNDO); in tcp_try_undo_partial()
2897 u32 wlen = sock_net(sk)->ipv4.sysctl_tcp_min_rtt_wlen * HZ; in tcp_update_rtt_min()
3298 if (tcp_sk(sk)->reordering > sock_net(sk)->ipv4.sysctl_tcp_reordering) in tcp_may_raise_cwnd()
3446 struct net *net = sock_net(sk); in tcp_send_challenge_ack()
3516 NET_INC_STATS(sock_net(sk), in tcp_process_tlp_ack()
3557 const struct net *net = sock_net(sk); in tcp_newly_delivered()
3644 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPHPACKS); in tcp_ack()
3651 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPPUREACKS); in tcp_ack()
4176 if (tcp_is_sack(tp) && sock_net(sk)->ipv4.sysctl_tcp_dsack) { in tcp_dsack_set()
4184 NET_INC_STATS(sock_net(sk), mib_idx); in tcp_dsack_set()
4208 NET_INC_STATS(sock_net(sk), LINUX_MIB_DELAYEDACKLOST); in tcp_send_dupack()
4211 if (tcp_is_sack(tp) && sock_net(sk)->ipv4.sysctl_tcp_dsack) { in tcp_send_dupack()
4366 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPRCVCOALESCE); in tcp_try_coalesce()
4487 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPOFODROP); in tcp_data_queue_ofo()
4496 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPOFOQUEUE); in tcp_data_queue_ofo()
4546 NET_INC_STATS(sock_net(sk), in tcp_data_queue_ofo()
4565 NET_INC_STATS(sock_net(sk), in tcp_data_queue_ofo()
4594 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPOFOMERGE); in tcp_data_queue_ofo()
4657 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPRCVQDROP); in tcp_send_rcvq()
4716 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPZEROWINDOWDROP); in tcp_data_queue()
4725 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPRCVQDROP); in tcp_data_queue()
4759 NET_INC_STATS(sock_net(sk), LINUX_MIB_DELAYEDACKLOST); in tcp_data_queue()
4786 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPZEROWINDOWDROP); in tcp_data_queue()
4815 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPRCVCOLLAPSED); in tcp_collapse_one()
5018 NET_INC_STATS(sock_net(sk), LINUX_MIB_OFOPRUNED); in tcp_prune_ofo_queue()
5060 NET_INC_STATS(sock_net(sk), LINUX_MIB_PRUNECALLED); in tcp_prune_queue()
5093 NET_INC_STATS(sock_net(sk), LINUX_MIB_RCVPRUNED); in tcp_prune_queue()
5196 tp->compressed_ack >= sock_net(sk)->ipv4.sysctl_tcp_comp_sack_nr) in __tcp_ack_snd_check()
5209 delay = min_t(unsigned long, sock_net(sk)->ipv4.sysctl_tcp_comp_sack_delay_ns, in __tcp_ack_snd_check()
5240 if (ptr && !sock_net(sk)->ipv4.sysctl_tcp_stdurg) in tcp_check_urg()
5353 if (tcp_fast_parse_options(sock_net(sk), skb, th, tp) && in tcp_validate_incoming()
5357 NET_INC_STATS(sock_net(sk), LINUX_MIB_PAWSESTABREJECTED); in tcp_validate_incoming()
5358 if (!tcp_oow_rate_limited(sock_net(sk), skb, in tcp_validate_incoming()
5378 if (!tcp_oow_rate_limited(sock_net(sk), skb, in tcp_validate_incoming()
5441 TCP_INC_STATS(sock_net(sk), TCP_MIB_INERRS); in tcp_validate_incoming()
5442 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPSYNCHALLENGE); in tcp_validate_incoming()
5567 TCP_INC_STATS(sock_net(sk), TCP_MIB_INERRS); in tcp_rcv_established()
5591 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPHPHITS); in tcp_rcv_established()
5647 TCP_INC_STATS(sock_net(sk), TCP_MIB_CSUMERRORS); in tcp_rcv_established()
5648 TCP_INC_STATS(sock_net(sk), TCP_MIB_INERRS); in tcp_rcv_established()
5699 tcp_parse_options(sock_net(sk), synack, &opt, 0, NULL); in tcp_rcv_fastopen_synack()
5729 NET_INC_STATS(sock_net(sk), in tcp_rcv_fastopen_synack()
5735 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPFASTOPENACTIVE); in tcp_rcv_fastopen_synack()
5765 tcp_parse_options(sock_net(sk), skb, &tp->rx_opt, 0, &foc); in tcp_rcv_synsent_state_process()
5785 NET_INC_STATS(sock_net(sk), in tcp_rcv_synsent_state_process()
6153 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPABORTONDATA); in tcp_rcv_state_process()
6162 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPABORTONDATA); in tcp_rcv_state_process()
6220 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPABORTONDATA); in tcp_rcv_state_process()
6278 const struct net *net = sock_net(listen_sk); in tcp_ecn_create_request()
6340 write_pnet(&ireq->ireq_net, sock_net(sk_listener)); in inet_reqsk_alloc()
6358 struct net *net = sock_net(sk); in tcp_syn_flood_action()
6364 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPREQQFULLDOCOOKIES); in tcp_syn_flood_action()
6367 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPREQQFULLDROP); in tcp_syn_flood_action()
6403 struct net *net = sock_net(sk); in tcp_conn_request()
6422 NET_INC_STATS(sock_net(sk), LINUX_MIB_LISTENOVERFLOWS); in tcp_conn_request()
6436 tcp_parse_options(sock_net(sk), skb, &tmp_opt, 0, in tcp_conn_request()