/Linux-v4.19/fs/reiserfs/ |
D | do_balan.c | 19 static inline void buffer_info_init_left(struct tree_balance *tb, in buffer_info_init_left() argument 22 bi->tb = tb; in buffer_info_init_left() 23 bi->bi_bh = tb->L[0]; in buffer_info_init_left() 24 bi->bi_parent = tb->FL[0]; in buffer_info_init_left() 25 bi->bi_position = get_left_neighbor_position(tb, 0); in buffer_info_init_left() 28 static inline void buffer_info_init_right(struct tree_balance *tb, in buffer_info_init_right() argument 31 bi->tb = tb; in buffer_info_init_right() 32 bi->bi_bh = tb->R[0]; in buffer_info_init_right() 33 bi->bi_parent = tb->FR[0]; in buffer_info_init_right() 34 bi->bi_position = get_right_neighbor_position(tb, 0); in buffer_info_init_right() [all …]
|
D | fix_node.c | 51 static void create_virtual_node(struct tree_balance *tb, int h) in create_virtual_node() argument 54 struct virtual_node *vn = tb->tb_vn; in create_virtual_node() 58 Sh = PATH_H_PBUFFER(tb->tb_path, h); in create_virtual_node() 62 MAX_CHILD_SIZE(Sh) - B_FREE_SPACE(Sh) + tb->insert_size[h]; in create_virtual_node() 76 vn->vn_vi = (struct virtual_item *)(tb->tb_vn + 1); in create_virtual_node() 115 op_create_vi(vn, vi, is_affected, tb->insert_size[0]); in create_virtual_node() 116 if (tb->vn_buf + tb->vn_buf_size < vn->vn_free_ptr) in create_virtual_node() 117 reiserfs_panic(tb->tb_sb, "vs-8030", in create_virtual_node() 125 vn->vn_vi[new_num].vi_item_len += tb->insert_size[0]; in create_virtual_node() 137 vi->vi_item_len = tb->insert_size[0]; in create_virtual_node() [all …]
|
D | ibalance.c | 28 struct tree_balance *tb, in internal_define_dest_src_infos() argument 41 src_bi->tb = tb; in internal_define_dest_src_infos() 42 src_bi->bi_bh = PATH_H_PBUFFER(tb->tb_path, h); in internal_define_dest_src_infos() 43 src_bi->bi_parent = PATH_H_PPARENT(tb->tb_path, h); in internal_define_dest_src_infos() 44 src_bi->bi_position = PATH_H_POSITION(tb->tb_path, h + 1); in internal_define_dest_src_infos() 45 dest_bi->tb = tb; in internal_define_dest_src_infos() 46 dest_bi->bi_bh = tb->L[h]; in internal_define_dest_src_infos() 47 dest_bi->bi_parent = tb->FL[h]; in internal_define_dest_src_infos() 48 dest_bi->bi_position = get_left_neighbor_position(tb, h); in internal_define_dest_src_infos() 49 *d_key = tb->lkey[h]; in internal_define_dest_src_infos() [all …]
|
D | prints.c | 640 void store_print_tb(struct tree_balance *tb) in store_print_tb() argument 646 if (!tb) in store_print_tb() 654 REISERFS_SB(tb->tb_sb)->s_do_balance, in store_print_tb() 655 tb->tb_mode, PATH_LAST_POSITION(tb->tb_path), in store_print_tb() 656 tb->tb_path->pos_in_item); in store_print_tb() 658 for (h = 0; h < ARRAY_SIZE(tb->insert_size); h++) { in store_print_tb() 659 if (PATH_H_PATH_OFFSET(tb->tb_path, h) <= in store_print_tb() 660 tb->tb_path->path_length in store_print_tb() 661 && PATH_H_PATH_OFFSET(tb->tb_path, in store_print_tb() 663 tbSh = PATH_H_PBUFFER(tb->tb_path, h); in store_print_tb() [all …]
|
D | lbalance.c | 392 do_balance_mark_leaf_dirty(dest_bi->tb, dest, 0); in leaf_copy_items_entirely() 405 do_balance_mark_internal_dirty(dest_bi->tb, dest_bi->bi_parent, in leaf_copy_items_entirely() 629 static void leaf_define_dest_src_infos(int shift_mode, struct tree_balance *tb, in leaf_define_dest_src_infos() argument 641 src_bi->tb = tb; in leaf_define_dest_src_infos() 642 src_bi->bi_bh = PATH_PLAST_BUFFER(tb->tb_path); in leaf_define_dest_src_infos() 643 src_bi->bi_parent = PATH_H_PPARENT(tb->tb_path, 0); in leaf_define_dest_src_infos() 646 src_bi->bi_position = PATH_H_B_ITEM_ORDER(tb->tb_path, 0); in leaf_define_dest_src_infos() 647 dest_bi->tb = tb; in leaf_define_dest_src_infos() 648 dest_bi->bi_bh = tb->L[0]; in leaf_define_dest_src_infos() 649 dest_bi->bi_parent = tb->FL[0]; in leaf_define_dest_src_infos() [all …]
|
/Linux-v4.19/drivers/thunderbolt/ |
D | domain.c | 127 struct tb *tb = container_of(dev, struct tb, dev); in boot_acl_show() local 132 uuids = kcalloc(tb->nboot_acl, sizeof(uuid_t), GFP_KERNEL); in boot_acl_show() 136 pm_runtime_get_sync(&tb->dev); in boot_acl_show() 138 if (mutex_lock_interruptible(&tb->lock)) { in boot_acl_show() 142 ret = tb->cm_ops->get_boot_acl(tb, uuids, tb->nboot_acl); in boot_acl_show() 144 mutex_unlock(&tb->lock); in boot_acl_show() 147 mutex_unlock(&tb->lock); in boot_acl_show() 149 for (ret = 0, i = 0; i < tb->nboot_acl; i++) { in boot_acl_show() 155 i < tb->nboot_acl - 1 ? "," : "\n"); in boot_acl_show() 159 pm_runtime_mark_last_busy(&tb->dev); in boot_acl_show() [all …]
|
D | tb.h | 90 struct tb *tb; member 179 struct tb *tb; member 217 int (*driver_ready)(struct tb *tb); 218 int (*start)(struct tb *tb); 219 void (*stop)(struct tb *tb); 220 int (*suspend_noirq)(struct tb *tb); 221 int (*resume_noirq)(struct tb *tb); 222 int (*suspend)(struct tb *tb); 223 void (*complete)(struct tb *tb); 224 int (*runtime_suspend)(struct tb *tb); [all …]
|
D | tb.c | 66 sw = tb_switch_alloc(port->sw->tb, &port->sw->dev, in tb_scan_port() 91 static void tb_free_invalid_tunnels(struct tb *tb) in tb_free_invalid_tunnels() argument 93 struct tb_cm *tcm = tb_priv(tb); in tb_free_invalid_tunnels() 173 static void tb_activate_pcie_devices(struct tb *tb) in tb_activate_pcie_devices() argument 182 struct tb_cm *tcm = tb_priv(tb); in tb_activate_pcie_devices() 185 for (i = 1; i <= tb->root_switch->config.max_port_number; i++) { in tb_activate_pcie_devices() 186 if (tb_is_upstream_port(&tb->root_switch->ports[i])) in tb_activate_pcie_devices() 188 if (tb->root_switch->ports[i].config.type != TB_TYPE_PORT) in tb_activate_pcie_devices() 190 if (!tb->root_switch->ports[i].remote) in tb_activate_pcie_devices() 192 sw = tb->root_switch->ports[i].remote->sw; in tb_activate_pcie_devices() [all …]
|
D | icm.c | 80 bool (*is_supported)(struct tb *tb); 81 int (*get_mode)(struct tb *tb); 82 int (*get_route)(struct tb *tb, u8 link, u8 depth, u64 *route); 83 void (*save_devices)(struct tb *tb); 84 int (*driver_ready)(struct tb *tb, 87 void (*device_connected)(struct tb *tb, 89 void (*device_disconnected)(struct tb *tb, 91 void (*xdomain_connected)(struct tb *tb, 93 void (*xdomain_disconnected)(struct tb *tb, 100 struct tb *tb; member [all …]
|
/Linux-v4.19/tools/testing/selftests/powerpc/pmu/ebb/ |
D | trace.c | 17 struct trace_buffer *tb; in trace_buffer_allocate() local 19 if (size < sizeof(*tb)) { in trace_buffer_allocate() 24 tb = mmap(NULL, size, PROT_READ | PROT_WRITE, in trace_buffer_allocate() 26 if (tb == MAP_FAILED) { in trace_buffer_allocate() 31 tb->size = size; in trace_buffer_allocate() 32 tb->tail = tb->data; in trace_buffer_allocate() 33 tb->overflow = false; in trace_buffer_allocate() 35 return tb; in trace_buffer_allocate() 38 static bool trace_check_bounds(struct trace_buffer *tb, void *p) in trace_check_bounds() argument 40 return p < ((void *)tb + tb->size); in trace_check_bounds() [all …]
|
/Linux-v4.19/net/netfilter/ |
D | nfnetlink_cthelper.c | 79 struct nlattr *tb[NFCTH_TUPLE_MAX+1]; in nfnl_cthelper_parse_tuple() local 81 err = nla_parse_nested(tb, NFCTH_TUPLE_MAX, attr, in nfnl_cthelper_parse_tuple() 86 if (!tb[NFCTH_TUPLE_L3PROTONUM] || !tb[NFCTH_TUPLE_L4PROTONUM]) in nfnl_cthelper_parse_tuple() 92 tuple->src.l3num = ntohs(nla_get_be16(tb[NFCTH_TUPLE_L3PROTONUM])); in nfnl_cthelper_parse_tuple() 93 tuple->dst.protonum = nla_get_u8(tb[NFCTH_TUPLE_L4PROTONUM]); in nfnl_cthelper_parse_tuple() 140 struct nlattr *tb[NFCTH_POLICY_MAX+1]; in nfnl_cthelper_expect_policy() local 142 err = nla_parse_nested(tb, NFCTH_POLICY_MAX, attr, in nfnl_cthelper_expect_policy() 147 if (!tb[NFCTH_POLICY_NAME] || in nfnl_cthelper_expect_policy() 148 !tb[NFCTH_POLICY_EXPECT_MAX] || in nfnl_cthelper_expect_policy() 149 !tb[NFCTH_POLICY_EXPECT_TIMEOUT]) in nfnl_cthelper_expect_policy() [all …]
|
D | nft_tunnel.c | 54 const struct nlattr * const tb[]) in nft_tunnel_get_init() argument 59 if (!tb[NFTA_TUNNEL_KEY] && in nft_tunnel_get_init() 60 !tb[NFTA_TUNNEL_DREG]) in nft_tunnel_get_init() 63 priv->key = ntohl(nla_get_be32(tb[NFTA_TUNNEL_KEY])); in nft_tunnel_get_init() 75 priv->dreg = nft_parse_register(tb[NFTA_TUNNEL_DREG]); in nft_tunnel_get_init() 136 struct nlattr *tb[NFTA_TUNNEL_KEY_IP_MAX + 1]; in nft_tunnel_obj_ip_init() local 139 err = nla_parse_nested(tb, NFTA_TUNNEL_KEY_IP_MAX, attr, in nft_tunnel_obj_ip_init() 144 if (!tb[NFTA_TUNNEL_KEY_IP_DST]) in nft_tunnel_obj_ip_init() 147 if (tb[NFTA_TUNNEL_KEY_IP_SRC]) in nft_tunnel_obj_ip_init() 148 info->key.u.ipv4.src = nla_get_be32(tb[NFTA_TUNNEL_KEY_IP_SRC]); in nft_tunnel_obj_ip_init() [all …]
|
D | nft_hash.c | 124 const struct nlattr * const tb[]) in nft_jhash_init() argument 130 if (!tb[NFTA_HASH_SREG] || in nft_jhash_init() 131 !tb[NFTA_HASH_DREG] || in nft_jhash_init() 132 !tb[NFTA_HASH_LEN] || in nft_jhash_init() 133 !tb[NFTA_HASH_MODULUS]) in nft_jhash_init() 136 if (tb[NFTA_HASH_OFFSET]) in nft_jhash_init() 137 priv->offset = ntohl(nla_get_be32(tb[NFTA_HASH_OFFSET])); in nft_jhash_init() 139 priv->sreg = nft_parse_register(tb[NFTA_HASH_SREG]); in nft_jhash_init() 140 priv->dreg = nft_parse_register(tb[NFTA_HASH_DREG]); in nft_jhash_init() 142 err = nft_parse_u32_check(tb[NFTA_HASH_LEN], U8_MAX, &len); in nft_jhash_init() [all …]
|
D | nft_exthdr.c | 222 const struct nlattr * const tb[]) in nft_exthdr_init() argument 228 if (!tb[NFTA_EXTHDR_DREG] || in nft_exthdr_init() 229 !tb[NFTA_EXTHDR_TYPE] || in nft_exthdr_init() 230 !tb[NFTA_EXTHDR_OFFSET] || in nft_exthdr_init() 231 !tb[NFTA_EXTHDR_LEN]) in nft_exthdr_init() 234 err = nft_parse_u32_check(tb[NFTA_EXTHDR_OFFSET], U8_MAX, &offset); in nft_exthdr_init() 238 err = nft_parse_u32_check(tb[NFTA_EXTHDR_LEN], U8_MAX, &len); in nft_exthdr_init() 242 if (tb[NFTA_EXTHDR_FLAGS]) { in nft_exthdr_init() 243 err = nft_parse_u32_check(tb[NFTA_EXTHDR_FLAGS], U8_MAX, &flags); in nft_exthdr_init() 251 if (tb[NFTA_EXTHDR_OP]) { in nft_exthdr_init() [all …]
|
/Linux-v4.19/drivers/iio/common/st_sensors/ |
D | st_sensors_spi.c | 29 static int st_sensors_spi_read(struct st_sensor_transfer_buffer *tb, in st_sensors_spi_read() argument 36 .tx_buf = tb->tx_buf, in st_sensors_spi_read() 41 .rx_buf = tb->rx_buf, in st_sensors_spi_read() 47 mutex_lock(&tb->buf_lock); in st_sensors_spi_read() 49 tb->tx_buf[0] = reg_addr | ST_SENSORS_SPI_MULTIREAD; in st_sensors_spi_read() 51 tb->tx_buf[0] = reg_addr | ST_SENSORS_SPI_READ; in st_sensors_spi_read() 57 memcpy(data, tb->rx_buf, len); in st_sensors_spi_read() 58 mutex_unlock(&tb->buf_lock); in st_sensors_spi_read() 62 mutex_unlock(&tb->buf_lock); in st_sensors_spi_read() 66 static int st_sensors_spi_read_byte(struct st_sensor_transfer_buffer *tb, in st_sensors_spi_read_byte() argument [all …]
|
/Linux-v4.19/arch/powerpc/kernel/ |
D | watchdog.c | 115 u64 tb = get_tb(); in wd_lockup_ipi() local 119 cpu, tb, per_cpu(wd_timer_tb, cpu), in wd_lockup_ipi() 120 tb_to_ns(tb - per_cpu(wd_timer_tb, cpu)) / 1000000); in wd_lockup_ipi() 131 static void set_cpumask_stuck(const struct cpumask *cpumask, u64 tb) in set_cpumask_stuck() argument 136 wd_smp_last_reset_tb = tb; in set_cpumask_stuck() 142 static void set_cpu_stuck(int cpu, u64 tb) in set_cpu_stuck() argument 144 set_cpumask_stuck(cpumask_of(cpu), tb); in set_cpu_stuck() 147 static void watchdog_smp_panic(int cpu, u64 tb) in watchdog_smp_panic() argument 154 if ((s64)(tb - wd_smp_last_reset_tb) < (s64)wd_smp_panic_timeout_tb) in watchdog_smp_panic() 164 cpu, tb, wd_smp_last_reset_tb, in watchdog_smp_panic() [all …]
|
/Linux-v4.19/arch/sparc/mm/ |
D | tlb.c | 26 struct tlb_batch *tb = &get_cpu_var(tlb_batch); in flush_tlb_pending() local 27 struct mm_struct *mm = tb->mm; in flush_tlb_pending() 29 if (!tb->tlb_nr) in flush_tlb_pending() 32 flush_tsb_user(tb); in flush_tlb_pending() 35 if (tb->tlb_nr == 1) { in flush_tlb_pending() 36 global_flush_tlb_page(mm, tb->vaddrs[0]); in flush_tlb_pending() 39 smp_flush_tlb_pending(tb->mm, tb->tlb_nr, in flush_tlb_pending() 40 &tb->vaddrs[0]); in flush_tlb_pending() 42 __flush_tlb_pending(CTX_HWBITS(tb->mm->context), in flush_tlb_pending() 43 tb->tlb_nr, &tb->vaddrs[0]); in flush_tlb_pending() [all …]
|
/Linux-v4.19/drivers/target/ |
D | target_core_hba.c | 53 struct target_backend *tb, *old; in transport_backend_register() local 55 tb = kzalloc(sizeof(*tb), GFP_KERNEL); in transport_backend_register() 56 if (!tb) in transport_backend_register() 58 tb->ops = ops; in transport_backend_register() 65 kfree(tb); in transport_backend_register() 69 target_setup_backend_cits(tb); in transport_backend_register() 70 list_add_tail(&tb->list, &backend_list); in transport_backend_register() 81 struct target_backend *tb; in target_backend_unregister() local 84 list_for_each_entry(tb, &backend_list, list) { in target_backend_unregister() 85 if (tb->ops == ops) { in target_backend_unregister() [all …]
|
/Linux-v4.19/net/netfilter/ipset/ |
D | ip_set_hash_ipportip.c | 111 hash_ipportip4_uadt(struct ip_set *set, struct nlattr *tb[], in hash_ipportip4_uadt() argument 122 if (tb[IPSET_ATTR_LINENO]) in hash_ipportip4_uadt() 123 *lineno = nla_get_u32(tb[IPSET_ATTR_LINENO]); in hash_ipportip4_uadt() 125 if (unlikely(!tb[IPSET_ATTR_IP] || !tb[IPSET_ATTR_IP2] || in hash_ipportip4_uadt() 126 !ip_set_attr_netorder(tb, IPSET_ATTR_PORT) || in hash_ipportip4_uadt() 127 !ip_set_optattr_netorder(tb, IPSET_ATTR_PORT_TO))) in hash_ipportip4_uadt() 130 ret = ip_set_get_ipaddr4(tb[IPSET_ATTR_IP], &e.ip); in hash_ipportip4_uadt() 134 ret = ip_set_get_extensions(set, tb, &ext); in hash_ipportip4_uadt() 138 ret = ip_set_get_ipaddr4(tb[IPSET_ATTR_IP2], &e.ip2); in hash_ipportip4_uadt() 142 e.port = nla_get_be16(tb[IPSET_ATTR_PORT]); in hash_ipportip4_uadt() [all …]
|
D | ip_set_hash_ipportnet.c | 163 hash_ipportnet4_uadt(struct ip_set *set, struct nlattr *tb[], in hash_ipportnet4_uadt() argument 176 if (tb[IPSET_ATTR_LINENO]) in hash_ipportnet4_uadt() 177 *lineno = nla_get_u32(tb[IPSET_ATTR_LINENO]); in hash_ipportnet4_uadt() 179 if (unlikely(!tb[IPSET_ATTR_IP] || !tb[IPSET_ATTR_IP2] || in hash_ipportnet4_uadt() 180 !ip_set_attr_netorder(tb, IPSET_ATTR_PORT) || in hash_ipportnet4_uadt() 181 !ip_set_optattr_netorder(tb, IPSET_ATTR_PORT_TO) || in hash_ipportnet4_uadt() 182 !ip_set_optattr_netorder(tb, IPSET_ATTR_CADT_FLAGS))) in hash_ipportnet4_uadt() 185 ret = ip_set_get_hostipaddr4(tb[IPSET_ATTR_IP], &ip); in hash_ipportnet4_uadt() 189 ret = ip_set_get_extensions(set, tb, &ext); in hash_ipportnet4_uadt() 193 ret = ip_set_get_hostipaddr4(tb[IPSET_ATTR_IP2], &ip2_from); in hash_ipportnet4_uadt() [all …]
|
D | ip_set_hash_ipport.c | 108 hash_ipport4_uadt(struct ip_set *set, struct nlattr *tb[], in hash_ipport4_uadt() argument 119 if (tb[IPSET_ATTR_LINENO]) in hash_ipport4_uadt() 120 *lineno = nla_get_u32(tb[IPSET_ATTR_LINENO]); in hash_ipport4_uadt() 122 if (unlikely(!tb[IPSET_ATTR_IP] || in hash_ipport4_uadt() 123 !ip_set_attr_netorder(tb, IPSET_ATTR_PORT) || in hash_ipport4_uadt() 124 !ip_set_optattr_netorder(tb, IPSET_ATTR_PORT_TO))) in hash_ipport4_uadt() 127 ret = ip_set_get_ipaddr4(tb[IPSET_ATTR_IP], &e.ip); in hash_ipport4_uadt() 131 ret = ip_set_get_extensions(set, tb, &ext); in hash_ipport4_uadt() 135 e.port = nla_get_be16(tb[IPSET_ATTR_PORT]); in hash_ipport4_uadt() 137 if (tb[IPSET_ATTR_PROTO]) { in hash_ipport4_uadt() [all …]
|
D | ip_set_hash_ipmac.c | 116 hash_ipmac4_uadt(struct ip_set *set, struct nlattr *tb[], in hash_ipmac4_uadt() argument 124 if (unlikely(!tb[IPSET_ATTR_IP] || in hash_ipmac4_uadt() 125 !tb[IPSET_ATTR_ETHER] || in hash_ipmac4_uadt() 126 nla_len(tb[IPSET_ATTR_ETHER]) != ETH_ALEN || in hash_ipmac4_uadt() 127 !ip_set_optattr_netorder(tb, IPSET_ATTR_TIMEOUT) || in hash_ipmac4_uadt() 128 !ip_set_optattr_netorder(tb, IPSET_ATTR_PACKETS) || in hash_ipmac4_uadt() 129 !ip_set_optattr_netorder(tb, IPSET_ATTR_BYTES) || in hash_ipmac4_uadt() 130 !ip_set_optattr_netorder(tb, IPSET_ATTR_SKBMARK) || in hash_ipmac4_uadt() 131 !ip_set_optattr_netorder(tb, IPSET_ATTR_SKBPRIO) || in hash_ipmac4_uadt() 132 !ip_set_optattr_netorder(tb, IPSET_ATTR_SKBQUEUE))) in hash_ipmac4_uadt() [all …]
|
D | ip_set_hash_netportnet.c | 180 hash_netportnet4_uadt(struct ip_set *set, struct nlattr *tb[], in hash_netportnet4_uadt() argument 192 if (tb[IPSET_ATTR_LINENO]) in hash_netportnet4_uadt() 193 *lineno = nla_get_u32(tb[IPSET_ATTR_LINENO]); in hash_netportnet4_uadt() 196 if (unlikely(!tb[IPSET_ATTR_IP] || !tb[IPSET_ATTR_IP2] || in hash_netportnet4_uadt() 197 !ip_set_attr_netorder(tb, IPSET_ATTR_PORT) || in hash_netportnet4_uadt() 198 !ip_set_optattr_netorder(tb, IPSET_ATTR_PORT_TO) || in hash_netportnet4_uadt() 199 !ip_set_optattr_netorder(tb, IPSET_ATTR_CADT_FLAGS))) in hash_netportnet4_uadt() 202 ret = ip_set_get_hostipaddr4(tb[IPSET_ATTR_IP], &ip); in hash_netportnet4_uadt() 206 ret = ip_set_get_hostipaddr4(tb[IPSET_ATTR_IP2], &ip2_from); in hash_netportnet4_uadt() 210 ret = ip_set_get_extensions(set, tb, &ext); in hash_netportnet4_uadt() [all …]
|
D | ip_set_hash_netport.c | 157 hash_netport4_uadt(struct ip_set *set, struct nlattr *tb[], in hash_netport4_uadt() argument 169 if (tb[IPSET_ATTR_LINENO]) in hash_netport4_uadt() 170 *lineno = nla_get_u32(tb[IPSET_ATTR_LINENO]); in hash_netport4_uadt() 172 if (unlikely(!tb[IPSET_ATTR_IP] || in hash_netport4_uadt() 173 !ip_set_attr_netorder(tb, IPSET_ATTR_PORT) || in hash_netport4_uadt() 174 !ip_set_optattr_netorder(tb, IPSET_ATTR_PORT_TO) || in hash_netport4_uadt() 175 !ip_set_optattr_netorder(tb, IPSET_ATTR_CADT_FLAGS))) in hash_netport4_uadt() 178 ret = ip_set_get_hostipaddr4(tb[IPSET_ATTR_IP], &ip); in hash_netport4_uadt() 182 ret = ip_set_get_extensions(set, tb, &ext); in hash_netport4_uadt() 186 if (tb[IPSET_ATTR_CIDR]) { in hash_netport4_uadt() [all …]
|
/Linux-v4.19/drivers/net/wireless/ti/wlcore/ |
D | testmode.c | 72 static int wl1271_tm_cmd_test(struct wl1271 *wl, struct nlattr *tb[]) in wl1271_tm_cmd_test() argument 81 if (!tb[WL1271_TM_ATTR_DATA]) in wl1271_tm_cmd_test() 84 buf = nla_data(tb[WL1271_TM_ATTR_DATA]); in wl1271_tm_cmd_test() 85 buf_len = nla_len(tb[WL1271_TM_ATTR_DATA]); in wl1271_tm_cmd_test() 87 if (tb[WL1271_TM_ATTR_ANSWER]) in wl1271_tm_cmd_test() 88 answer = nla_get_u8(tb[WL1271_TM_ATTR_ANSWER]); in wl1271_tm_cmd_test() 154 static int wl1271_tm_cmd_interrogate(struct wl1271 *wl, struct nlattr *tb[]) in wl1271_tm_cmd_interrogate() argument 163 if (!tb[WL1271_TM_ATTR_IE_ID]) in wl1271_tm_cmd_interrogate() 166 ie_id = nla_get_u8(tb[WL1271_TM_ATTR_IE_ID]); in wl1271_tm_cmd_interrogate() 221 static int wl1271_tm_cmd_configure(struct wl1271 *wl, struct nlattr *tb[]) in wl1271_tm_cmd_configure() argument [all …]
|