Lines Matching refs:pctx
106 static void pdp_context_delete(struct pdp_ctx *pctx);
173 static bool gtp_check_ms_ipv4(struct sk_buff *skb, struct pdp_ctx *pctx, in gtp_check_ms_ipv4() argument
184 return iph->daddr == pctx->ms_addr_ip4.s_addr; in gtp_check_ms_ipv4()
186 return iph->saddr == pctx->ms_addr_ip4.s_addr; in gtp_check_ms_ipv4()
192 static bool gtp_check_ms(struct sk_buff *skb, struct pdp_ctx *pctx, in gtp_check_ms() argument
197 return gtp_check_ms_ipv4(skb, pctx, hdrlen, role); in gtp_check_ms()
202 static int gtp_rx(struct pdp_ctx *pctx, struct sk_buff *skb, in gtp_rx() argument
205 if (!gtp_check_ms(skb, pctx, hdrlen, role)) { in gtp_rx()
206 netdev_dbg(pctx->dev, "No PDP ctx for this MS\n"); in gtp_rx()
212 !net_eq(sock_net(pctx->sk), dev_net(pctx->dev)))) { in gtp_rx()
213 pctx->dev->stats.rx_length_errors++; in gtp_rx()
217 netdev_dbg(pctx->dev, "forwarding packet from GGSN to uplink\n"); in gtp_rx()
226 skb->dev = pctx->dev; in gtp_rx()
228 dev_sw_netstats_rx_add(pctx->dev, skb->len); in gtp_rx()
234 pctx->dev->stats.rx_dropped++; in gtp_rx()
413 struct pdp_ctx *pctx; in gtp0_udp_encap_recv() local
436 pctx = gtp0_pdp_find(gtp, be64_to_cpu(gtp0->tid)); in gtp0_udp_encap_recv()
437 if (!pctx) { in gtp0_udp_encap_recv()
442 return gtp_rx(pctx, skb, hdrlen, gtp->role); in gtp0_udp_encap_recv()
575 struct pdp_ctx *pctx; in gtp1u_udp_encap_recv() local
613 pctx = gtp1_pdp_find(gtp, ntohl(gtp1->tid)); in gtp1u_udp_encap_recv()
614 if (!pctx) { in gtp1u_udp_encap_recv()
619 return gtp_rx(pctx, skb, hdrlen, gtp->role); in gtp1u_udp_encap_recv()
735 static inline void gtp0_push_header(struct sk_buff *skb, struct pdp_ctx *pctx) in gtp0_push_header() argument
745 gtp0->seq = htons((atomic_inc_return(&pctx->tx_seq) - 1) % 0xffff); in gtp0_push_header()
746 gtp0->flow = htons(pctx->u.v0.flow); in gtp0_push_header()
749 gtp0->tid = cpu_to_be64(pctx->u.v0.tid); in gtp0_push_header()
752 static inline void gtp1_push_header(struct sk_buff *skb, struct pdp_ctx *pctx) in gtp1_push_header() argument
768 gtp1->tid = htonl(pctx->u.v1.o_tei); in gtp1_push_header()
780 struct pdp_ctx *pctx; member
787 switch (pktinfo->pctx->gtp_version) { in gtp_push_header()
790 gtp0_push_header(skb, pktinfo->pctx); in gtp_push_header()
794 gtp1_push_header(skb, pktinfo->pctx); in gtp_push_header()
801 struct pdp_ctx *pctx, struct rtable *rt, in gtp_set_pktinfo_ipv4() argument
807 pktinfo->pctx = pctx; in gtp_set_pktinfo_ipv4()
817 struct pdp_ctx *pctx; in gtp_build_skb_ip4() local
829 pctx = ipv4_pdp_find(gtp, iph->saddr); in gtp_build_skb_ip4()
831 pctx = ipv4_pdp_find(gtp, iph->daddr); in gtp_build_skb_ip4()
833 if (!pctx) { in gtp_build_skb_ip4()
838 netdev_dbg(dev, "found PDP context %p\n", pctx); in gtp_build_skb_ip4()
840 rt = ip4_route_output_gtp(&fl4, pctx->sk, pctx->peer_addr_ip4.s_addr, in gtp_build_skb_ip4()
841 inet_sk(pctx->sk)->inet_saddr); in gtp_build_skb_ip4()
844 &pctx->peer_addr_ip4.s_addr); in gtp_build_skb_ip4()
851 &pctx->peer_addr_ip4.s_addr); in gtp_build_skb_ip4()
861 switch (pctx->gtp_version) { in gtp_build_skb_ip4()
884 gtp_set_pktinfo_ipv4(pktinfo, pctx->sk, iph, pctx, rt, &fl4, dev); in gtp_build_skb_ip4()
931 !net_eq(sock_net(pktinfo.pctx->sk), in gtp_dev_xmit()
1114 struct pdp_ctx *pctx; in gtp_dellink() local
1118 hlist_for_each_entry_rcu(pctx, >p->tid_hash[i], hlist_tid) in gtp_dellink()
1119 pdp_context_delete(pctx); in gtp_dellink()
1311 static void ipv4_pdp_fill(struct pdp_ctx *pctx, struct genl_info *info) in ipv4_pdp_fill() argument
1313 pctx->gtp_version = nla_get_u32(info->attrs[GTPA_VERSION]); in ipv4_pdp_fill()
1314 pctx->af = AF_INET; in ipv4_pdp_fill()
1315 pctx->peer_addr_ip4.s_addr = in ipv4_pdp_fill()
1317 pctx->ms_addr_ip4.s_addr = in ipv4_pdp_fill()
1320 switch (pctx->gtp_version) { in ipv4_pdp_fill()
1326 pctx->u.v0.tid = nla_get_u64(info->attrs[GTPA_TID]); in ipv4_pdp_fill()
1327 pctx->u.v0.flow = nla_get_u16(info->attrs[GTPA_FLOW]); in ipv4_pdp_fill()
1330 pctx->u.v1.i_tei = nla_get_u32(info->attrs[GTPA_I_TEI]); in ipv4_pdp_fill()
1331 pctx->u.v1.o_tei = nla_get_u32(info->attrs[GTPA_O_TEI]); in ipv4_pdp_fill()
1341 struct pdp_ctx *pctx, *pctx_tid = NULL; in gtp_pdp_add() local
1352 pctx = ipv4_pdp_find(gtp, ms_addr); in gtp_pdp_add()
1353 if (pctx) in gtp_pdp_add()
1370 if (pctx && pctx_tid) in gtp_pdp_add()
1372 if (!pctx) in gtp_pdp_add()
1373 pctx = pctx_tid; in gtp_pdp_add()
1375 ipv4_pdp_fill(pctx, info); in gtp_pdp_add()
1377 if (pctx->gtp_version == GTP_V0) in gtp_pdp_add()
1379 pctx->u.v0.tid, pctx); in gtp_pdp_add()
1380 else if (pctx->gtp_version == GTP_V1) in gtp_pdp_add()
1382 pctx->u.v1.i_tei, pctx->u.v1.o_tei, pctx); in gtp_pdp_add()
1384 return pctx; in gtp_pdp_add()
1388 pctx = kmalloc(sizeof(*pctx), GFP_ATOMIC); in gtp_pdp_add()
1389 if (pctx == NULL) in gtp_pdp_add()
1393 pctx->sk = sk; in gtp_pdp_add()
1394 pctx->dev = gtp->dev; in gtp_pdp_add()
1395 ipv4_pdp_fill(pctx, info); in gtp_pdp_add()
1396 atomic_set(&pctx->tx_seq, 0); in gtp_pdp_add()
1398 switch (pctx->gtp_version) { in gtp_pdp_add()
1405 hash_tid = gtp0_hashfn(pctx->u.v0.tid) % gtp->hash_size; in gtp_pdp_add()
1408 hash_tid = gtp1u_hashfn(pctx->u.v1.i_tei) % gtp->hash_size; in gtp_pdp_add()
1412 hlist_add_head_rcu(&pctx->hlist_addr, >p->addr_hash[hash_ms]); in gtp_pdp_add()
1413 hlist_add_head_rcu(&pctx->hlist_tid, >p->tid_hash[hash_tid]); in gtp_pdp_add()
1415 switch (pctx->gtp_version) { in gtp_pdp_add()
1418 pctx->u.v0.tid, &pctx->peer_addr_ip4, in gtp_pdp_add()
1419 &pctx->ms_addr_ip4, pctx); in gtp_pdp_add()
1423 pctx->u.v1.i_tei, pctx->u.v1.o_tei, in gtp_pdp_add()
1424 &pctx->peer_addr_ip4, &pctx->ms_addr_ip4, pctx); in gtp_pdp_add()
1428 return pctx; in gtp_pdp_add()
1433 struct pdp_ctx *pctx = container_of(head, struct pdp_ctx, rcu_head); in pdp_context_free() local
1435 sock_put(pctx->sk); in pdp_context_free()
1436 kfree(pctx); in pdp_context_free()
1439 static void pdp_context_delete(struct pdp_ctx *pctx) in pdp_context_delete() argument
1441 hlist_del_rcu(&pctx->hlist_tid); in pdp_context_delete()
1442 hlist_del_rcu(&pctx->hlist_addr); in pdp_context_delete()
1443 call_rcu(&pctx->rcu_head, pdp_context_free); in pdp_context_delete()
1446 static int gtp_tunnel_notify(struct pdp_ctx *pctx, u8 cmd, gfp_t allocation);
1451 struct pdp_ctx *pctx; in gtp_genl_new_pdp() local
1500 pctx = gtp_pdp_add(gtp, sk, info); in gtp_genl_new_pdp()
1501 if (IS_ERR(pctx)) { in gtp_genl_new_pdp()
1502 err = PTR_ERR(pctx); in gtp_genl_new_pdp()
1504 gtp_tunnel_notify(pctx, GTP_CMD_NEWPDP, GFP_KERNEL); in gtp_genl_new_pdp()
1540 struct pdp_ctx *pctx; in gtp_find_pdp() local
1543 pctx = gtp_find_pdp_by_link(net, nla); in gtp_find_pdp()
1545 pctx = ERR_PTR(-EINVAL); in gtp_find_pdp()
1547 if (!pctx) in gtp_find_pdp()
1548 pctx = ERR_PTR(-ENOENT); in gtp_find_pdp()
1550 return pctx; in gtp_find_pdp()
1555 struct pdp_ctx *pctx; in gtp_genl_del_pdp() local
1563 pctx = gtp_find_pdp(sock_net(skb->sk), info->attrs); in gtp_genl_del_pdp()
1564 if (IS_ERR(pctx)) { in gtp_genl_del_pdp()
1565 err = PTR_ERR(pctx); in gtp_genl_del_pdp()
1569 if (pctx->gtp_version == GTP_V0) in gtp_genl_del_pdp()
1570 netdev_dbg(pctx->dev, "GTPv0-U: deleting tunnel id = %llx (pdp %p)\n", in gtp_genl_del_pdp()
1571 pctx->u.v0.tid, pctx); in gtp_genl_del_pdp()
1572 else if (pctx->gtp_version == GTP_V1) in gtp_genl_del_pdp()
1573 netdev_dbg(pctx->dev, "GTPv1-U: deleting tunnel id = %x/%x (pdp %p)\n", in gtp_genl_del_pdp()
1574 pctx->u.v1.i_tei, pctx->u.v1.o_tei, pctx); in gtp_genl_del_pdp()
1576 gtp_tunnel_notify(pctx, GTP_CMD_DELPDP, GFP_ATOMIC); in gtp_genl_del_pdp()
1577 pdp_context_delete(pctx); in gtp_genl_del_pdp()
1585 int flags, u32 type, struct pdp_ctx *pctx) in gtp_genl_fill_info() argument
1594 if (nla_put_u32(skb, GTPA_VERSION, pctx->gtp_version) || in gtp_genl_fill_info()
1595 nla_put_u32(skb, GTPA_LINK, pctx->dev->ifindex) || in gtp_genl_fill_info()
1596 nla_put_be32(skb, GTPA_PEER_ADDRESS, pctx->peer_addr_ip4.s_addr) || in gtp_genl_fill_info()
1597 nla_put_be32(skb, GTPA_MS_ADDRESS, pctx->ms_addr_ip4.s_addr)) in gtp_genl_fill_info()
1600 switch (pctx->gtp_version) { in gtp_genl_fill_info()
1602 if (nla_put_u64_64bit(skb, GTPA_TID, pctx->u.v0.tid, GTPA_PAD) || in gtp_genl_fill_info()
1603 nla_put_u16(skb, GTPA_FLOW, pctx->u.v0.flow)) in gtp_genl_fill_info()
1607 if (nla_put_u32(skb, GTPA_I_TEI, pctx->u.v1.i_tei) || in gtp_genl_fill_info()
1608 nla_put_u32(skb, GTPA_O_TEI, pctx->u.v1.o_tei)) in gtp_genl_fill_info()
1621 static int gtp_tunnel_notify(struct pdp_ctx *pctx, u8 cmd, gfp_t allocation) in gtp_tunnel_notify() argument
1630 ret = gtp_genl_fill_info(msg, 0, 0, 0, cmd, pctx); in gtp_tunnel_notify()
1636 ret = genlmsg_multicast_netns(>p_genl_family, dev_net(pctx->dev), msg, in gtp_tunnel_notify()
1643 struct pdp_ctx *pctx = NULL; in gtp_genl_get_pdp() local
1652 pctx = gtp_find_pdp(sock_net(skb->sk), info->attrs); in gtp_genl_get_pdp()
1653 if (IS_ERR(pctx)) { in gtp_genl_get_pdp()
1654 err = PTR_ERR(pctx); in gtp_genl_get_pdp()
1665 0, info->nlhdr->nlmsg_type, pctx); in gtp_genl_get_pdp()
1685 struct pdp_ctx *pctx; in gtp_genl_dump_pdp() local
1702 hlist_for_each_entry_rcu(pctx, >p->tid_hash[i], in gtp_genl_dump_pdp()
1709 cb->nlh->nlmsg_type, pctx)) { in gtp_genl_dump_pdp()