Home
last modified time | relevance | path

Searched refs:sk_listener (Results 1 – 13 of 13) sorted by relevance

/Linux-v5.15/include/net/
Drequest_sock.h86 reqsk_alloc(const struct request_sock_ops *ops, struct sock *sk_listener, in reqsk_alloc() argument
96 if (unlikely(!refcount_inc_not_zero(&sk_listener->sk_refcnt))) { in reqsk_alloc()
100 req->rsk_listener = sk_listener; in reqsk_alloc()
103 req_to_sk(req)->sk_prot = sk_listener->sk_prot; in reqsk_alloc()
Dmptcp.h183 const struct sock *sk_listener,
271 const struct sock *sk_listener, in mptcp_subflow_init_cookie_req() argument
Dinet_sock.h340 struct sock *sk_listener,
Dsock.h2718 static inline bool sk_listener(const struct sock *sk) in sk_listener() function
Dtcp.h1454 const struct sock *sk_listener,
/Linux-v5.15/net/mptcp/
Dsubflow.c105 static void subflow_init_req(struct request_sock *req, const struct sock *sk_listener) in subflow_init_req() argument
111 subflow_req->csum_reqd = mptcp_is_checksum_enabled(sock_net(sk_listener)); in subflow_init_req()
112 subflow_req->allow_join_id0 = mptcp_allow_join_id0(sock_net(sk_listener)); in subflow_init_req()
138 const struct sock *sk_listener, in subflow_check_req() argument
141 struct mptcp_subflow_context *listener = mptcp_subflow_ctx(sk_listener); in subflow_check_req()
152 if (rcu_access_pointer(tcp_sk(sk_listener)->md5sig_info)) in subflow_check_req()
156 mptcp_get_options(sk_listener, skb, &mp_opt); in subflow_check_req()
215 if (subflow_use_different_sport(subflow_req->msk, sk_listener)) { in subflow_check_req()
217 ntohs(inet_sk(sk_listener)->inet_sport), in subflow_check_req()
219 if (!mptcp_pm_sport_in_anno_list(subflow_req->msk, sk_listener)) { in subflow_check_req()
[all …]
/Linux-v5.15/net/ipv4/
Dinet_connection_sock.c800 struct sock *sk_listener = req->rsk_listener; in reqsk_timer_handler() local
806 if (inet_sk_state_load(sk_listener) != TCP_LISTEN) { in reqsk_timer_handler()
809 nsk = reuseport_migrate_sock(sk_listener, req_to_sk(req), NULL); in reqsk_timer_handler()
827 sk_listener = nsk; in reqsk_timer_handler()
830 icsk = inet_csk(sk_listener); in reqsk_timer_handler()
831 net = sock_net(sk_listener); in reqsk_timer_handler()
852 if ((qlen << 1) > max(8U, READ_ONCE(sk_listener->sk_max_ack_backlog))) { in reqsk_timer_handler()
867 !inet_rtx_syn_ack(sk_listener, req) || in reqsk_timer_handler()
881 inet_csk_reqsk_queue_drop(sk_listener, nreq); in reqsk_timer_handler()
Dtcp_minisocks.c361 const struct sock *sk_listener, in tcp_openreq_init_rwin() argument
365 const struct tcp_sock *tp = tcp_sk(sk_listener); in tcp_openreq_init_rwin()
366 int full_space = tcp_full_space(sk_listener); in tcp_openreq_init_rwin()
378 if (sk_listener->sk_userlocks & SOCK_RCVBUF_LOCK && in tcp_openreq_init_rwin()
389 tcp_select_initial_window(sk_listener, full_space, in tcp_openreq_init_rwin()
Dtcp_input.c6692 struct sock *sk_listener, in inet_reqsk_alloc() argument
6695 struct request_sock *req = reqsk_alloc(ops, sk_listener, in inet_reqsk_alloc()
6707 write_pnet(&ireq->ireq_net, sock_net(sk_listener)); in inet_reqsk_alloc()
6708 ireq->ireq_family = sk_listener->sk_family; in inet_reqsk_alloc()
Dtcp_ipv4.c1470 const struct sock *sk_listener, in tcp_v4_init_req() argument
1474 struct net *net = sock_net(sk_listener); in tcp_v4_init_req()
/Linux-v5.15/net/ipv6/
Dtcp_ipv6.c824 const struct sock *sk_listener, in tcp_v6_init_req() argument
829 const struct ipv6_pinfo *np = tcp_inet6_sk(sk_listener); in tcp_v6_init_req()
835 if ((!sk_listener->sk_bound_dev_if || l3_slave) && in tcp_v6_init_req()
840 (ipv6_opt_accepted(sk_listener, skb, &TCP_SKB_CB(skb)->header.h6) || in tcp_v6_init_req()
/Linux-v5.15/net/sched/
Dsch_fq.c281 if (!sk || sk_listener(sk)) { in fq_classify()
/Linux-v5.15/security/selinux/
Dhooks.c5781 if (sk_listener(sk)) in selinux_ip_output()
5898 !(sk && sk_listener(sk))) in selinux_ip_postroute()
5915 } else if (sk_listener(sk)) { in selinux_ip_postroute()