Lines Matching refs:pctx

87 static void pdp_context_delete(struct pdp_ctx *pctx);
154 static bool gtp_check_ms_ipv4(struct sk_buff *skb, struct pdp_ctx *pctx, in gtp_check_ms_ipv4() argument
165 return iph->daddr == pctx->ms_addr_ip4.s_addr; in gtp_check_ms_ipv4()
167 return iph->saddr == pctx->ms_addr_ip4.s_addr; in gtp_check_ms_ipv4()
173 static bool gtp_check_ms(struct sk_buff *skb, struct pdp_ctx *pctx, in gtp_check_ms() argument
178 return gtp_check_ms_ipv4(skb, pctx, hdrlen, role); in gtp_check_ms()
183 static int gtp_rx(struct pdp_ctx *pctx, struct sk_buff *skb, in gtp_rx() argument
188 if (!gtp_check_ms(skb, pctx, hdrlen, role)) { in gtp_rx()
189 netdev_dbg(pctx->dev, "No PDP ctx for this MS\n"); in gtp_rx()
195 !net_eq(sock_net(pctx->sk), dev_net(pctx->dev)))) in gtp_rx()
198 netdev_dbg(pctx->dev, "forwarding packet from GGSN to uplink\n"); in gtp_rx()
206 skb->dev = pctx->dev; in gtp_rx()
208 stats = this_cpu_ptr(pctx->dev->tstats); in gtp_rx()
224 struct pdp_ctx *pctx; in gtp0_udp_encap_recv() local
237 pctx = gtp0_pdp_find(gtp, be64_to_cpu(gtp0->tid)); in gtp0_udp_encap_recv()
238 if (!pctx) { in gtp0_udp_encap_recv()
243 return gtp_rx(pctx, skb, hdrlen, gtp->role); in gtp0_udp_encap_recv()
251 struct pdp_ctx *pctx; in gtp1u_udp_encap_recv() local
279 pctx = gtp1_pdp_find(gtp, ntohl(gtp1->tid)); in gtp1u_udp_encap_recv()
280 if (!pctx) { in gtp1u_udp_encap_recv()
285 return gtp_rx(pctx, skb, hdrlen, gtp->role); in gtp1u_udp_encap_recv()
405 static inline void gtp0_push_header(struct sk_buff *skb, struct pdp_ctx *pctx) in gtp0_push_header() argument
415 gtp0->seq = htons((atomic_inc_return(&pctx->tx_seq) - 1) % 0xffff); in gtp0_push_header()
416 gtp0->flow = htons(pctx->u.v0.flow); in gtp0_push_header()
419 gtp0->tid = cpu_to_be64(pctx->u.v0.tid); in gtp0_push_header()
422 static inline void gtp1_push_header(struct sk_buff *skb, struct pdp_ctx *pctx) in gtp1_push_header() argument
438 gtp1->tid = htonl(pctx->u.v1.o_tei); in gtp1_push_header()
450 struct pdp_ctx *pctx; member
457 switch (pktinfo->pctx->gtp_version) { in gtp_push_header()
460 gtp0_push_header(skb, pktinfo->pctx); in gtp_push_header()
464 gtp1_push_header(skb, pktinfo->pctx); in gtp_push_header()
471 struct pdp_ctx *pctx, struct rtable *rt, in gtp_set_pktinfo_ipv4() argument
477 pktinfo->pctx = pctx; in gtp_set_pktinfo_ipv4()
487 struct pdp_ctx *pctx; in gtp_build_skb_ip4() local
499 pctx = ipv4_pdp_find(gtp, iph->saddr); in gtp_build_skb_ip4()
501 pctx = ipv4_pdp_find(gtp, iph->daddr); in gtp_build_skb_ip4()
503 if (!pctx) { in gtp_build_skb_ip4()
508 netdev_dbg(dev, "found PDP context %p\n", pctx); in gtp_build_skb_ip4()
510 rt = ip4_route_output_gtp(&fl4, pctx->sk, pctx->peer_addr_ip4.s_addr); in gtp_build_skb_ip4()
513 &pctx->peer_addr_ip4.s_addr); in gtp_build_skb_ip4()
520 &pctx->peer_addr_ip4.s_addr); in gtp_build_skb_ip4()
532 switch (pctx->gtp_version) { in gtp_build_skb_ip4()
555 gtp_set_pktinfo_ipv4(pktinfo, pctx->sk, iph, pctx, rt, &fl4, dev); in gtp_build_skb_ip4()
777 struct pdp_ctx *pctx; in gtp_hashtable_free() local
781 hlist_for_each_entry_rcu(pctx, &gtp->tid_hash[i], hlist_tid) in gtp_hashtable_free()
782 pdp_context_delete(pctx); in gtp_hashtable_free()
902 static void ipv4_pdp_fill(struct pdp_ctx *pctx, struct genl_info *info) in ipv4_pdp_fill() argument
904 pctx->gtp_version = nla_get_u32(info->attrs[GTPA_VERSION]); in ipv4_pdp_fill()
905 pctx->af = AF_INET; in ipv4_pdp_fill()
906 pctx->peer_addr_ip4.s_addr = in ipv4_pdp_fill()
908 pctx->ms_addr_ip4.s_addr = in ipv4_pdp_fill()
911 switch (pctx->gtp_version) { in ipv4_pdp_fill()
917 pctx->u.v0.tid = nla_get_u64(info->attrs[GTPA_TID]); in ipv4_pdp_fill()
918 pctx->u.v0.flow = nla_get_u16(info->attrs[GTPA_FLOW]); in ipv4_pdp_fill()
921 pctx->u.v1.i_tei = nla_get_u32(info->attrs[GTPA_I_TEI]); in ipv4_pdp_fill()
922 pctx->u.v1.o_tei = nla_get_u32(info->attrs[GTPA_O_TEI]); in ipv4_pdp_fill()
934 struct pdp_ctx *pctx; in ipv4_pdp_add() local
941 hlist_for_each_entry_rcu(pctx, &gtp->addr_hash[hash_ms], hlist_addr) { in ipv4_pdp_add()
942 if (pctx->ms_addr_ip4.s_addr == ms_addr) { in ipv4_pdp_add()
954 ipv4_pdp_fill(pctx, info); in ipv4_pdp_add()
956 if (pctx->gtp_version == GTP_V0) in ipv4_pdp_add()
958 pctx->u.v0.tid, pctx); in ipv4_pdp_add()
959 else if (pctx->gtp_version == GTP_V1) in ipv4_pdp_add()
961 pctx->u.v1.i_tei, pctx->u.v1.o_tei, pctx); in ipv4_pdp_add()
967 pctx = kmalloc(sizeof(*pctx), GFP_ATOMIC); in ipv4_pdp_add()
968 if (pctx == NULL) in ipv4_pdp_add()
972 pctx->sk = sk; in ipv4_pdp_add()
973 pctx->dev = gtp->dev; in ipv4_pdp_add()
974 ipv4_pdp_fill(pctx, info); in ipv4_pdp_add()
975 atomic_set(&pctx->tx_seq, 0); in ipv4_pdp_add()
977 switch (pctx->gtp_version) { in ipv4_pdp_add()
984 hash_tid = gtp0_hashfn(pctx->u.v0.tid) % gtp->hash_size; in ipv4_pdp_add()
987 hash_tid = gtp1u_hashfn(pctx->u.v1.i_tei) % gtp->hash_size; in ipv4_pdp_add()
991 hlist_add_head_rcu(&pctx->hlist_addr, &gtp->addr_hash[hash_ms]); in ipv4_pdp_add()
992 hlist_add_head_rcu(&pctx->hlist_tid, &gtp->tid_hash[hash_tid]); in ipv4_pdp_add()
994 switch (pctx->gtp_version) { in ipv4_pdp_add()
997 pctx->u.v0.tid, &pctx->peer_addr_ip4, in ipv4_pdp_add()
998 &pctx->ms_addr_ip4, pctx); in ipv4_pdp_add()
1002 pctx->u.v1.i_tei, pctx->u.v1.o_tei, in ipv4_pdp_add()
1003 &pctx->peer_addr_ip4, &pctx->ms_addr_ip4, pctx); in ipv4_pdp_add()
1012 struct pdp_ctx *pctx = container_of(head, struct pdp_ctx, rcu_head); in pdp_context_free() local
1014 sock_put(pctx->sk); in pdp_context_free()
1015 kfree(pctx); in pdp_context_free()
1018 static void pdp_context_delete(struct pdp_ctx *pctx) in pdp_context_delete() argument
1020 hlist_del_rcu(&pctx->hlist_tid); in pdp_context_delete()
1021 hlist_del_rcu(&pctx->hlist_addr); in pdp_context_delete()
1022 call_rcu(&pctx->rcu_head, pdp_context_free); in pdp_context_delete()
1112 struct pdp_ctx *pctx; in gtp_find_pdp() local
1115 pctx = gtp_find_pdp_by_link(net, nla); in gtp_find_pdp()
1117 pctx = ERR_PTR(-EINVAL); in gtp_find_pdp()
1119 if (!pctx) in gtp_find_pdp()
1120 pctx = ERR_PTR(-ENOENT); in gtp_find_pdp()
1122 return pctx; in gtp_find_pdp()
1127 struct pdp_ctx *pctx; in gtp_genl_del_pdp() local
1135 pctx = gtp_find_pdp(sock_net(skb->sk), info->attrs); in gtp_genl_del_pdp()
1136 if (IS_ERR(pctx)) { in gtp_genl_del_pdp()
1137 err = PTR_ERR(pctx); in gtp_genl_del_pdp()
1141 if (pctx->gtp_version == GTP_V0) in gtp_genl_del_pdp()
1142 netdev_dbg(pctx->dev, "GTPv0-U: deleting tunnel id = %llx (pdp %p)\n", in gtp_genl_del_pdp()
1143 pctx->u.v0.tid, pctx); in gtp_genl_del_pdp()
1144 else if (pctx->gtp_version == GTP_V1) in gtp_genl_del_pdp()
1145 netdev_dbg(pctx->dev, "GTPv1-U: deleting tunnel id = %x/%x (pdp %p)\n", in gtp_genl_del_pdp()
1146 pctx->u.v1.i_tei, pctx->u.v1.o_tei, pctx); in gtp_genl_del_pdp()
1148 pdp_context_delete(pctx); in gtp_genl_del_pdp()
1158 u32 type, struct pdp_ctx *pctx) in gtp_genl_fill_info() argument
1167 if (nla_put_u32(skb, GTPA_VERSION, pctx->gtp_version) || in gtp_genl_fill_info()
1168 nla_put_be32(skb, GTPA_PEER_ADDRESS, pctx->peer_addr_ip4.s_addr) || in gtp_genl_fill_info()
1169 nla_put_be32(skb, GTPA_MS_ADDRESS, pctx->ms_addr_ip4.s_addr)) in gtp_genl_fill_info()
1172 switch (pctx->gtp_version) { in gtp_genl_fill_info()
1174 if (nla_put_u64_64bit(skb, GTPA_TID, pctx->u.v0.tid, GTPA_PAD) || in gtp_genl_fill_info()
1175 nla_put_u16(skb, GTPA_FLOW, pctx->u.v0.flow)) in gtp_genl_fill_info()
1179 if (nla_put_u32(skb, GTPA_I_TEI, pctx->u.v1.i_tei) || in gtp_genl_fill_info()
1180 nla_put_u32(skb, GTPA_O_TEI, pctx->u.v1.o_tei)) in gtp_genl_fill_info()
1195 struct pdp_ctx *pctx = NULL; in gtp_genl_get_pdp() local
1204 pctx = gtp_find_pdp(sock_net(skb->sk), info->attrs); in gtp_genl_get_pdp()
1205 if (IS_ERR(pctx)) { in gtp_genl_get_pdp()
1206 err = PTR_ERR(pctx); in gtp_genl_get_pdp()
1217 info->snd_seq, info->nlhdr->nlmsg_type, pctx); in gtp_genl_get_pdp()
1239 struct pdp_ctx *pctx; in gtp_genl_dump_pdp() local
1251 hlist_for_each_entry_rcu(pctx, &gtp->tid_hash[i], hlist_tid) { in gtp_genl_dump_pdp()
1252 if (tid && tid != pctx->u.tid) in gtp_genl_dump_pdp()
1260 cb->nlh->nlmsg_type, pctx); in gtp_genl_dump_pdp()
1263 cb->args[1] = pctx->u.tid; in gtp_genl_dump_pdp()