/Linux-v5.4/fs/reiserfs/ |
D | do_balan.c | 19 static inline void buffer_info_init_left(struct tree_balance *tb, in buffer_info_init_left() argument 22 bi->tb = tb; in buffer_info_init_left() 23 bi->bi_bh = tb->L[0]; in buffer_info_init_left() 24 bi->bi_parent = tb->FL[0]; in buffer_info_init_left() 25 bi->bi_position = get_left_neighbor_position(tb, 0); in buffer_info_init_left() 28 static inline void buffer_info_init_right(struct tree_balance *tb, in buffer_info_init_right() argument 31 bi->tb = tb; in buffer_info_init_right() 32 bi->bi_bh = tb->R[0]; in buffer_info_init_right() 33 bi->bi_parent = tb->FR[0]; in buffer_info_init_right() 34 bi->bi_position = get_right_neighbor_position(tb, 0); in buffer_info_init_right() [all …]
|
D | fix_node.c | 51 static void create_virtual_node(struct tree_balance *tb, int h) in create_virtual_node() argument 54 struct virtual_node *vn = tb->tb_vn; in create_virtual_node() 58 Sh = PATH_H_PBUFFER(tb->tb_path, h); in create_virtual_node() 62 MAX_CHILD_SIZE(Sh) - B_FREE_SPACE(Sh) + tb->insert_size[h]; in create_virtual_node() 76 vn->vn_vi = (struct virtual_item *)(tb->tb_vn + 1); in create_virtual_node() 115 op_create_vi(vn, vi, is_affected, tb->insert_size[0]); in create_virtual_node() 116 if (tb->vn_buf + tb->vn_buf_size < vn->vn_free_ptr) in create_virtual_node() 117 reiserfs_panic(tb->tb_sb, "vs-8030", in create_virtual_node() 125 vn->vn_vi[new_num].vi_item_len += tb->insert_size[0]; in create_virtual_node() 137 vi->vi_item_len = tb->insert_size[0]; in create_virtual_node() [all …]
|
D | ibalance.c | 28 struct tree_balance *tb, in internal_define_dest_src_infos() argument 41 src_bi->tb = tb; in internal_define_dest_src_infos() 42 src_bi->bi_bh = PATH_H_PBUFFER(tb->tb_path, h); in internal_define_dest_src_infos() 43 src_bi->bi_parent = PATH_H_PPARENT(tb->tb_path, h); in internal_define_dest_src_infos() 44 src_bi->bi_position = PATH_H_POSITION(tb->tb_path, h + 1); in internal_define_dest_src_infos() 45 dest_bi->tb = tb; in internal_define_dest_src_infos() 46 dest_bi->bi_bh = tb->L[h]; in internal_define_dest_src_infos() 47 dest_bi->bi_parent = tb->FL[h]; in internal_define_dest_src_infos() 48 dest_bi->bi_position = get_left_neighbor_position(tb, h); in internal_define_dest_src_infos() 49 *d_key = tb->lkey[h]; in internal_define_dest_src_infos() [all …]
|
D | prints.c | 640 void store_print_tb(struct tree_balance *tb) in store_print_tb() argument 646 if (!tb) in store_print_tb() 654 REISERFS_SB(tb->tb_sb)->s_do_balance, in store_print_tb() 655 tb->tb_mode, PATH_LAST_POSITION(tb->tb_path), in store_print_tb() 656 tb->tb_path->pos_in_item); in store_print_tb() 658 for (h = 0; h < ARRAY_SIZE(tb->insert_size); h++) { in store_print_tb() 659 if (PATH_H_PATH_OFFSET(tb->tb_path, h) <= in store_print_tb() 660 tb->tb_path->path_length in store_print_tb() 661 && PATH_H_PATH_OFFSET(tb->tb_path, in store_print_tb() 663 tbSh = PATH_H_PBUFFER(tb->tb_path, h); in store_print_tb() [all …]
|
D | lbalance.c | 392 do_balance_mark_leaf_dirty(dest_bi->tb, dest, 0); in leaf_copy_items_entirely() 405 do_balance_mark_internal_dirty(dest_bi->tb, dest_bi->bi_parent, in leaf_copy_items_entirely() 629 static void leaf_define_dest_src_infos(int shift_mode, struct tree_balance *tb, in leaf_define_dest_src_infos() argument 641 src_bi->tb = tb; in leaf_define_dest_src_infos() 642 src_bi->bi_bh = PATH_PLAST_BUFFER(tb->tb_path); in leaf_define_dest_src_infos() 643 src_bi->bi_parent = PATH_H_PPARENT(tb->tb_path, 0); in leaf_define_dest_src_infos() 646 src_bi->bi_position = PATH_H_B_ITEM_ORDER(tb->tb_path, 0); in leaf_define_dest_src_infos() 647 dest_bi->tb = tb; in leaf_define_dest_src_infos() 648 dest_bi->bi_bh = tb->L[0]; in leaf_define_dest_src_infos() 649 dest_bi->bi_parent = tb->FL[0]; in leaf_define_dest_src_infos() [all …]
|
/Linux-v5.4/drivers/thunderbolt/ |
D | domain.c | 126 struct tb *tb = container_of(dev, struct tb, dev); in boot_acl_show() local 131 uuids = kcalloc(tb->nboot_acl, sizeof(uuid_t), GFP_KERNEL); in boot_acl_show() 135 pm_runtime_get_sync(&tb->dev); in boot_acl_show() 137 if (mutex_lock_interruptible(&tb->lock)) { in boot_acl_show() 141 ret = tb->cm_ops->get_boot_acl(tb, uuids, tb->nboot_acl); in boot_acl_show() 143 mutex_unlock(&tb->lock); in boot_acl_show() 146 mutex_unlock(&tb->lock); in boot_acl_show() 148 for (ret = 0, i = 0; i < tb->nboot_acl; i++) { in boot_acl_show() 154 i < tb->nboot_acl - 1 ? "," : "\n"); in boot_acl_show() 158 pm_runtime_mark_last_busy(&tb->dev); in boot_acl_show() [all …]
|
D | tb.h | 93 struct tb *tb; member 223 struct tb *tb; member 269 int (*driver_ready)(struct tb *tb); 270 int (*start)(struct tb *tb); 271 void (*stop)(struct tb *tb); 272 int (*suspend_noirq)(struct tb *tb); 273 int (*resume_noirq)(struct tb *tb); 274 int (*suspend)(struct tb *tb); 275 void (*complete)(struct tb *tb); 276 int (*runtime_suspend)(struct tb *tb); [all …]
|
D | tb.c | 33 struct tb *tb; member 41 static void tb_queue_hotplug(struct tb *tb, u64 route, u8 port, bool unplug) in tb_queue_hotplug() argument 49 ev->tb = tb; in tb_queue_hotplug() 54 queue_work(tb->wq, &ev->work); in tb_queue_hotplug() 61 struct tb *tb = sw->tb; in tb_discover_tunnels() local 62 struct tb_cm *tcm = tb_priv(tb); in tb_discover_tunnels() 72 tunnel = tb_tunnel_discover_dp(tb, port); in tb_discover_tunnels() 76 tunnel = tb_tunnel_discover_pci(tb, port); in tb_discover_tunnels() 107 struct tb *tb = sw->tb; in tb_scan_xdomain() local 112 xd = tb_xdomain_find_by_route(tb, route); in tb_scan_xdomain() [all …]
|
D | icm.c | 83 bool (*is_supported)(struct tb *tb); 84 int (*cio_reset)(struct tb *tb); 85 int (*get_mode)(struct tb *tb); 86 int (*get_route)(struct tb *tb, u8 link, u8 depth, u64 *route); 87 void (*save_devices)(struct tb *tb); 88 int (*driver_ready)(struct tb *tb, 91 void (*set_uuid)(struct tb *tb); 92 void (*device_connected)(struct tb *tb, 94 void (*device_disconnected)(struct tb *tb, 96 void (*xdomain_connected)(struct tb *tb, [all …]
|
D | tunnel.h | 34 struct tb *tb; member 45 struct tb_tunnel *tb_tunnel_discover_pci(struct tb *tb, struct tb_port *down); 46 struct tb_tunnel *tb_tunnel_alloc_pci(struct tb *tb, struct tb_port *up, 48 struct tb_tunnel *tb_tunnel_discover_dp(struct tb *tb, struct tb_port *in); 49 struct tb_tunnel *tb_tunnel_alloc_dp(struct tb *tb, struct tb_port *in, 51 struct tb_tunnel *tb_tunnel_alloc_dma(struct tb *tb, struct tb_port *nhi,
|
/Linux-v5.4/tools/bpf/bpftool/ |
D | netlink_dumper.c | 14 static void xdp_dump_prog_id(struct nlattr **tb, int attr, in xdp_dump_prog_id() argument 18 if (!tb[attr]) in xdp_dump_prog_id() 24 NET_DUMP_UINT("id", " id %u", libbpf_nla_getattr_u32(tb[attr])) in xdp_dump_prog_id() 32 struct nlattr *tb[IFLA_XDP_MAX + 1]; in do_xdp_dump_one() local 35 if (libbpf_nla_parse_nested(tb, IFLA_XDP_MAX, attr, NULL) < 0) in do_xdp_dump_one() 38 if (!tb[IFLA_XDP_ATTACHED]) in do_xdp_dump_one() 41 mode = libbpf_nla_getattr_u8(tb[IFLA_XDP_ATTACHED]); in do_xdp_dump_one() 55 xdp_dump_prog_id(tb, IFLA_XDP_SKB_PROG_ID, "generic", true); in do_xdp_dump_one() 56 xdp_dump_prog_id(tb, IFLA_XDP_DRV_PROG_ID, "driver", true); in do_xdp_dump_one() 57 xdp_dump_prog_id(tb, IFLA_XDP_HW_PROG_ID, "offload", true); in do_xdp_dump_one() [all …]
|
/Linux-v5.4/tools/testing/selftests/powerpc/pmu/ebb/ |
D | trace.c | 17 struct trace_buffer *tb; in trace_buffer_allocate() local 19 if (size < sizeof(*tb)) { in trace_buffer_allocate() 24 tb = mmap(NULL, size, PROT_READ | PROT_WRITE, in trace_buffer_allocate() 26 if (tb == MAP_FAILED) { in trace_buffer_allocate() 31 tb->size = size; in trace_buffer_allocate() 32 tb->tail = tb->data; in trace_buffer_allocate() 33 tb->overflow = false; in trace_buffer_allocate() 35 return tb; in trace_buffer_allocate() 38 static bool trace_check_bounds(struct trace_buffer *tb, void *p) in trace_check_bounds() argument 40 return p < ((void *)tb + tb->size); in trace_check_bounds() [all …]
|
/Linux-v5.4/net/netfilter/ |
D | nfnetlink_cthelper.c | 76 struct nlattr *tb[NFCTH_TUPLE_MAX+1]; in nfnl_cthelper_parse_tuple() local 78 err = nla_parse_nested_deprecated(tb, NFCTH_TUPLE_MAX, attr, in nfnl_cthelper_parse_tuple() 83 if (!tb[NFCTH_TUPLE_L3PROTONUM] || !tb[NFCTH_TUPLE_L4PROTONUM]) in nfnl_cthelper_parse_tuple() 89 tuple->src.l3num = ntohs(nla_get_be16(tb[NFCTH_TUPLE_L3PROTONUM])); in nfnl_cthelper_parse_tuple() 90 tuple->dst.protonum = nla_get_u8(tb[NFCTH_TUPLE_L4PROTONUM]); in nfnl_cthelper_parse_tuple() 137 struct nlattr *tb[NFCTH_POLICY_MAX+1]; in nfnl_cthelper_expect_policy() local 139 err = nla_parse_nested_deprecated(tb, NFCTH_POLICY_MAX, attr, in nfnl_cthelper_expect_policy() 144 if (!tb[NFCTH_POLICY_NAME] || in nfnl_cthelper_expect_policy() 145 !tb[NFCTH_POLICY_EXPECT_MAX] || in nfnl_cthelper_expect_policy() 146 !tb[NFCTH_POLICY_EXPECT_TIMEOUT]) in nfnl_cthelper_expect_policy() [all …]
|
D | nft_tunnel.c | 74 const struct nlattr * const tb[]) in nft_tunnel_get_init() argument 79 if (!tb[NFTA_TUNNEL_KEY] && in nft_tunnel_get_init() 80 !tb[NFTA_TUNNEL_DREG]) in nft_tunnel_get_init() 83 priv->key = ntohl(nla_get_be32(tb[NFTA_TUNNEL_KEY])); in nft_tunnel_get_init() 95 priv->dreg = nft_parse_register(tb[NFTA_TUNNEL_DREG]); in nft_tunnel_get_init() 97 if (tb[NFTA_TUNNEL_MODE]) { in nft_tunnel_get_init() 98 priv->mode = ntohl(nla_get_be32(tb[NFTA_TUNNEL_MODE])); in nft_tunnel_get_init() 166 struct nlattr *tb[NFTA_TUNNEL_KEY_IP_MAX + 1]; in nft_tunnel_obj_ip_init() local 169 err = nla_parse_nested_deprecated(tb, NFTA_TUNNEL_KEY_IP_MAX, attr, in nft_tunnel_obj_ip_init() 174 if (!tb[NFTA_TUNNEL_KEY_IP_DST]) in nft_tunnel_obj_ip_init() [all …]
|
D | nft_hash.c | 71 const struct nlattr * const tb[]) in nft_jhash_init() argument 77 if (!tb[NFTA_HASH_SREG] || in nft_jhash_init() 78 !tb[NFTA_HASH_DREG] || in nft_jhash_init() 79 !tb[NFTA_HASH_LEN] || in nft_jhash_init() 80 !tb[NFTA_HASH_MODULUS]) in nft_jhash_init() 83 if (tb[NFTA_HASH_OFFSET]) in nft_jhash_init() 84 priv->offset = ntohl(nla_get_be32(tb[NFTA_HASH_OFFSET])); in nft_jhash_init() 86 priv->sreg = nft_parse_register(tb[NFTA_HASH_SREG]); in nft_jhash_init() 87 priv->dreg = nft_parse_register(tb[NFTA_HASH_DREG]); in nft_jhash_init() 89 err = nft_parse_u32_check(tb[NFTA_HASH_LEN], U8_MAX, &len); in nft_jhash_init() [all …]
|
/Linux-v5.4/crypto/asymmetric_keys/ |
D | asym_tpm.c | 34 static int tpm_loadkey2(struct tpm_buf *tb, in tpm_loadkey2() argument 50 ret = oiap(tb, &authhandle, enonce); in tpm_loadkey2() 71 INIT_BUF(tb); in tpm_loadkey2() 72 store16(tb, TPM_TAG_RQU_AUTH1_COMMAND); in tpm_loadkey2() 73 store32(tb, TPM_LOADKEY2_SIZE + keybloblen); in tpm_loadkey2() 74 store32(tb, TPM_ORD_LOADKEY2); in tpm_loadkey2() 75 store32(tb, keyhandle); in tpm_loadkey2() 76 storebytes(tb, keyblob, keybloblen); in tpm_loadkey2() 77 store32(tb, authhandle); in tpm_loadkey2() 78 storebytes(tb, nonceodd, TPM_NONCE_SIZE); in tpm_loadkey2() [all …]
|
/Linux-v5.4/arch/powerpc/kernel/ |
D | watchdog.c | 115 u64 tb = get_tb(); in wd_lockup_ipi() local 119 cpu, tb, per_cpu(wd_timer_tb, cpu), in wd_lockup_ipi() 120 tb_to_ns(tb - per_cpu(wd_timer_tb, cpu)) / 1000000); in wd_lockup_ipi() 131 static void set_cpumask_stuck(const struct cpumask *cpumask, u64 tb) in set_cpumask_stuck() argument 136 wd_smp_last_reset_tb = tb; in set_cpumask_stuck() 142 static void set_cpu_stuck(int cpu, u64 tb) in set_cpu_stuck() argument 144 set_cpumask_stuck(cpumask_of(cpu), tb); in set_cpu_stuck() 147 static void watchdog_smp_panic(int cpu, u64 tb) in watchdog_smp_panic() argument 154 if ((s64)(tb - wd_smp_last_reset_tb) < (s64)wd_smp_panic_timeout_tb) in watchdog_smp_panic() 164 cpu, tb, wd_smp_last_reset_tb, in watchdog_smp_panic() [all …]
|
/Linux-v5.4/arch/sparc/mm/ |
D | tlb.c | 26 struct tlb_batch *tb = &get_cpu_var(tlb_batch); in flush_tlb_pending() local 27 struct mm_struct *mm = tb->mm; in flush_tlb_pending() 29 if (!tb->tlb_nr) in flush_tlb_pending() 32 flush_tsb_user(tb); in flush_tlb_pending() 35 if (tb->tlb_nr == 1) { in flush_tlb_pending() 36 global_flush_tlb_page(mm, tb->vaddrs[0]); in flush_tlb_pending() 39 smp_flush_tlb_pending(tb->mm, tb->tlb_nr, in flush_tlb_pending() 40 &tb->vaddrs[0]); in flush_tlb_pending() 42 __flush_tlb_pending(CTX_HWBITS(tb->mm->context), in flush_tlb_pending() 43 tb->tlb_nr, &tb->vaddrs[0]); in flush_tlb_pending() [all …]
|
/Linux-v5.4/drivers/target/ |
D | target_core_hba.c | 40 struct target_backend *tb, *old; in transport_backend_register() local 42 tb = kzalloc(sizeof(*tb), GFP_KERNEL); in transport_backend_register() 43 if (!tb) in transport_backend_register() 45 tb->ops = ops; in transport_backend_register() 52 kfree(tb); in transport_backend_register() 56 target_setup_backend_cits(tb); in transport_backend_register() 57 list_add_tail(&tb->list, &backend_list); in transport_backend_register() 68 struct target_backend *tb; in target_backend_unregister() local 71 list_for_each_entry(tb, &backend_list, list) { in target_backend_unregister() 72 if (tb->ops == ops) { in target_backend_unregister() [all …]
|
/Linux-v5.4/net/netfilter/ipset/ |
D | ip_set_hash_ipportip.c | 107 hash_ipportip4_uadt(struct ip_set *set, struct nlattr *tb[], in hash_ipportip4_uadt() argument 118 if (tb[IPSET_ATTR_LINENO]) in hash_ipportip4_uadt() 119 *lineno = nla_get_u32(tb[IPSET_ATTR_LINENO]); in hash_ipportip4_uadt() 121 if (unlikely(!tb[IPSET_ATTR_IP] || !tb[IPSET_ATTR_IP2] || in hash_ipportip4_uadt() 122 !ip_set_attr_netorder(tb, IPSET_ATTR_PORT) || in hash_ipportip4_uadt() 123 !ip_set_optattr_netorder(tb, IPSET_ATTR_PORT_TO))) in hash_ipportip4_uadt() 126 ret = ip_set_get_ipaddr4(tb[IPSET_ATTR_IP], &e.ip); in hash_ipportip4_uadt() 130 ret = ip_set_get_extensions(set, tb, &ext); in hash_ipportip4_uadt() 134 ret = ip_set_get_ipaddr4(tb[IPSET_ATTR_IP2], &e.ip2); in hash_ipportip4_uadt() 138 e.port = nla_get_be16(tb[IPSET_ATTR_PORT]); in hash_ipportip4_uadt() [all …]
|
D | ip_set_hash_ipportnet.c | 159 hash_ipportnet4_uadt(struct ip_set *set, struct nlattr *tb[], in hash_ipportnet4_uadt() argument 172 if (tb[IPSET_ATTR_LINENO]) in hash_ipportnet4_uadt() 173 *lineno = nla_get_u32(tb[IPSET_ATTR_LINENO]); in hash_ipportnet4_uadt() 175 if (unlikely(!tb[IPSET_ATTR_IP] || !tb[IPSET_ATTR_IP2] || in hash_ipportnet4_uadt() 176 !ip_set_attr_netorder(tb, IPSET_ATTR_PORT) || in hash_ipportnet4_uadt() 177 !ip_set_optattr_netorder(tb, IPSET_ATTR_PORT_TO) || in hash_ipportnet4_uadt() 178 !ip_set_optattr_netorder(tb, IPSET_ATTR_CADT_FLAGS))) in hash_ipportnet4_uadt() 181 ret = ip_set_get_hostipaddr4(tb[IPSET_ATTR_IP], &ip); in hash_ipportnet4_uadt() 185 ret = ip_set_get_extensions(set, tb, &ext); in hash_ipportnet4_uadt() 189 ret = ip_set_get_hostipaddr4(tb[IPSET_ATTR_IP2], &ip2_from); in hash_ipportnet4_uadt() [all …]
|
D | ip_set_hash_ipport.c | 104 hash_ipport4_uadt(struct ip_set *set, struct nlattr *tb[], in hash_ipport4_uadt() argument 115 if (tb[IPSET_ATTR_LINENO]) in hash_ipport4_uadt() 116 *lineno = nla_get_u32(tb[IPSET_ATTR_LINENO]); in hash_ipport4_uadt() 118 if (unlikely(!tb[IPSET_ATTR_IP] || in hash_ipport4_uadt() 119 !ip_set_attr_netorder(tb, IPSET_ATTR_PORT) || in hash_ipport4_uadt() 120 !ip_set_optattr_netorder(tb, IPSET_ATTR_PORT_TO))) in hash_ipport4_uadt() 123 ret = ip_set_get_ipaddr4(tb[IPSET_ATTR_IP], &e.ip); in hash_ipport4_uadt() 127 ret = ip_set_get_extensions(set, tb, &ext); in hash_ipport4_uadt() 131 e.port = nla_get_be16(tb[IPSET_ATTR_PORT]); in hash_ipport4_uadt() 133 if (tb[IPSET_ATTR_PROTO]) { in hash_ipport4_uadt() [all …]
|
D | ip_set_hash_ipmac.c | 110 hash_ipmac4_uadt(struct ip_set *set, struct nlattr *tb[], in hash_ipmac4_uadt() argument 118 if (unlikely(!tb[IPSET_ATTR_IP] || in hash_ipmac4_uadt() 119 !tb[IPSET_ATTR_ETHER] || in hash_ipmac4_uadt() 120 nla_len(tb[IPSET_ATTR_ETHER]) != ETH_ALEN || in hash_ipmac4_uadt() 121 !ip_set_optattr_netorder(tb, IPSET_ATTR_TIMEOUT) || in hash_ipmac4_uadt() 122 !ip_set_optattr_netorder(tb, IPSET_ATTR_PACKETS) || in hash_ipmac4_uadt() 123 !ip_set_optattr_netorder(tb, IPSET_ATTR_BYTES) || in hash_ipmac4_uadt() 124 !ip_set_optattr_netorder(tb, IPSET_ATTR_SKBMARK) || in hash_ipmac4_uadt() 125 !ip_set_optattr_netorder(tb, IPSET_ATTR_SKBPRIO) || in hash_ipmac4_uadt() 126 !ip_set_optattr_netorder(tb, IPSET_ATTR_SKBQUEUE))) in hash_ipmac4_uadt() [all …]
|
D | ip_set_hash_netportnet.c | 176 hash_netportnet4_uadt(struct ip_set *set, struct nlattr *tb[], in hash_netportnet4_uadt() argument 188 if (tb[IPSET_ATTR_LINENO]) in hash_netportnet4_uadt() 189 *lineno = nla_get_u32(tb[IPSET_ATTR_LINENO]); in hash_netportnet4_uadt() 192 if (unlikely(!tb[IPSET_ATTR_IP] || !tb[IPSET_ATTR_IP2] || in hash_netportnet4_uadt() 193 !ip_set_attr_netorder(tb, IPSET_ATTR_PORT) || in hash_netportnet4_uadt() 194 !ip_set_optattr_netorder(tb, IPSET_ATTR_PORT_TO) || in hash_netportnet4_uadt() 195 !ip_set_optattr_netorder(tb, IPSET_ATTR_CADT_FLAGS))) in hash_netportnet4_uadt() 198 ret = ip_set_get_hostipaddr4(tb[IPSET_ATTR_IP], &ip); in hash_netportnet4_uadt() 202 ret = ip_set_get_hostipaddr4(tb[IPSET_ATTR_IP2], &ip2_from); in hash_netportnet4_uadt() 206 ret = ip_set_get_extensions(set, tb, &ext); in hash_netportnet4_uadt() [all …]
|
/Linux-v5.4/drivers/net/wireless/ti/wlcore/ |
D | testmode.c | 58 static int wl1271_tm_cmd_test(struct wl1271 *wl, struct nlattr *tb[]) in wl1271_tm_cmd_test() argument 67 if (!tb[WL1271_TM_ATTR_DATA]) in wl1271_tm_cmd_test() 70 buf = nla_data(tb[WL1271_TM_ATTR_DATA]); in wl1271_tm_cmd_test() 71 buf_len = nla_len(tb[WL1271_TM_ATTR_DATA]); in wl1271_tm_cmd_test() 73 if (tb[WL1271_TM_ATTR_ANSWER]) in wl1271_tm_cmd_test() 74 answer = nla_get_u8(tb[WL1271_TM_ATTR_ANSWER]); in wl1271_tm_cmd_test() 140 static int wl1271_tm_cmd_interrogate(struct wl1271 *wl, struct nlattr *tb[]) in wl1271_tm_cmd_interrogate() argument 149 if (!tb[WL1271_TM_ATTR_IE_ID]) in wl1271_tm_cmd_interrogate() 152 ie_id = nla_get_u8(tb[WL1271_TM_ATTR_IE_ID]); in wl1271_tm_cmd_interrogate() 207 static int wl1271_tm_cmd_configure(struct wl1271 *wl, struct nlattr *tb[]) in wl1271_tm_cmd_configure() argument [all …]
|