Home
last modified time | relevance | path

Searched refs:tb (Results 1 – 25 of 405) sorted by relevance

12345678910>>...17

/Linux-v5.4/fs/reiserfs/
Ddo_balan.c19 static inline void buffer_info_init_left(struct tree_balance *tb, in buffer_info_init_left() argument
22 bi->tb = tb; in buffer_info_init_left()
23 bi->bi_bh = tb->L[0]; in buffer_info_init_left()
24 bi->bi_parent = tb->FL[0]; in buffer_info_init_left()
25 bi->bi_position = get_left_neighbor_position(tb, 0); in buffer_info_init_left()
28 static inline void buffer_info_init_right(struct tree_balance *tb, in buffer_info_init_right() argument
31 bi->tb = tb; in buffer_info_init_right()
32 bi->bi_bh = tb->R[0]; in buffer_info_init_right()
33 bi->bi_parent = tb->FR[0]; in buffer_info_init_right()
34 bi->bi_position = get_right_neighbor_position(tb, 0); in buffer_info_init_right()
[all …]
Dfix_node.c51 static void create_virtual_node(struct tree_balance *tb, int h) in create_virtual_node() argument
54 struct virtual_node *vn = tb->tb_vn; in create_virtual_node()
58 Sh = PATH_H_PBUFFER(tb->tb_path, h); in create_virtual_node()
62 MAX_CHILD_SIZE(Sh) - B_FREE_SPACE(Sh) + tb->insert_size[h]; in create_virtual_node()
76 vn->vn_vi = (struct virtual_item *)(tb->tb_vn + 1); in create_virtual_node()
115 op_create_vi(vn, vi, is_affected, tb->insert_size[0]); in create_virtual_node()
116 if (tb->vn_buf + tb->vn_buf_size < vn->vn_free_ptr) in create_virtual_node()
117 reiserfs_panic(tb->tb_sb, "vs-8030", in create_virtual_node()
125 vn->vn_vi[new_num].vi_item_len += tb->insert_size[0]; in create_virtual_node()
137 vi->vi_item_len = tb->insert_size[0]; in create_virtual_node()
[all …]
Dibalance.c28 struct tree_balance *tb, in internal_define_dest_src_infos() argument
41 src_bi->tb = tb; in internal_define_dest_src_infos()
42 src_bi->bi_bh = PATH_H_PBUFFER(tb->tb_path, h); in internal_define_dest_src_infos()
43 src_bi->bi_parent = PATH_H_PPARENT(tb->tb_path, h); in internal_define_dest_src_infos()
44 src_bi->bi_position = PATH_H_POSITION(tb->tb_path, h + 1); in internal_define_dest_src_infos()
45 dest_bi->tb = tb; in internal_define_dest_src_infos()
46 dest_bi->bi_bh = tb->L[h]; in internal_define_dest_src_infos()
47 dest_bi->bi_parent = tb->FL[h]; in internal_define_dest_src_infos()
48 dest_bi->bi_position = get_left_neighbor_position(tb, h); in internal_define_dest_src_infos()
49 *d_key = tb->lkey[h]; in internal_define_dest_src_infos()
[all …]
Dprints.c640 void store_print_tb(struct tree_balance *tb) in store_print_tb() argument
646 if (!tb) in store_print_tb()
654 REISERFS_SB(tb->tb_sb)->s_do_balance, in store_print_tb()
655 tb->tb_mode, PATH_LAST_POSITION(tb->tb_path), in store_print_tb()
656 tb->tb_path->pos_in_item); in store_print_tb()
658 for (h = 0; h < ARRAY_SIZE(tb->insert_size); h++) { in store_print_tb()
659 if (PATH_H_PATH_OFFSET(tb->tb_path, h) <= in store_print_tb()
660 tb->tb_path->path_length in store_print_tb()
661 && PATH_H_PATH_OFFSET(tb->tb_path, in store_print_tb()
663 tbSh = PATH_H_PBUFFER(tb->tb_path, h); in store_print_tb()
[all …]
Dlbalance.c392 do_balance_mark_leaf_dirty(dest_bi->tb, dest, 0); in leaf_copy_items_entirely()
405 do_balance_mark_internal_dirty(dest_bi->tb, dest_bi->bi_parent, in leaf_copy_items_entirely()
629 static void leaf_define_dest_src_infos(int shift_mode, struct tree_balance *tb, in leaf_define_dest_src_infos() argument
641 src_bi->tb = tb; in leaf_define_dest_src_infos()
642 src_bi->bi_bh = PATH_PLAST_BUFFER(tb->tb_path); in leaf_define_dest_src_infos()
643 src_bi->bi_parent = PATH_H_PPARENT(tb->tb_path, 0); in leaf_define_dest_src_infos()
646 src_bi->bi_position = PATH_H_B_ITEM_ORDER(tb->tb_path, 0); in leaf_define_dest_src_infos()
647 dest_bi->tb = tb; in leaf_define_dest_src_infos()
648 dest_bi->bi_bh = tb->L[0]; in leaf_define_dest_src_infos()
649 dest_bi->bi_parent = tb->FL[0]; in leaf_define_dest_src_infos()
[all …]
/Linux-v5.4/drivers/thunderbolt/
Ddomain.c126 struct tb *tb = container_of(dev, struct tb, dev); in boot_acl_show() local
131 uuids = kcalloc(tb->nboot_acl, sizeof(uuid_t), GFP_KERNEL); in boot_acl_show()
135 pm_runtime_get_sync(&tb->dev); in boot_acl_show()
137 if (mutex_lock_interruptible(&tb->lock)) { in boot_acl_show()
141 ret = tb->cm_ops->get_boot_acl(tb, uuids, tb->nboot_acl); in boot_acl_show()
143 mutex_unlock(&tb->lock); in boot_acl_show()
146 mutex_unlock(&tb->lock); in boot_acl_show()
148 for (ret = 0, i = 0; i < tb->nboot_acl; i++) { in boot_acl_show()
154 i < tb->nboot_acl - 1 ? "," : "\n"); in boot_acl_show()
158 pm_runtime_mark_last_busy(&tb->dev); in boot_acl_show()
[all …]
Dtb.h93 struct tb *tb; member
223 struct tb *tb; member
269 int (*driver_ready)(struct tb *tb);
270 int (*start)(struct tb *tb);
271 void (*stop)(struct tb *tb);
272 int (*suspend_noirq)(struct tb *tb);
273 int (*resume_noirq)(struct tb *tb);
274 int (*suspend)(struct tb *tb);
275 void (*complete)(struct tb *tb);
276 int (*runtime_suspend)(struct tb *tb);
[all …]
Dtb.c33 struct tb *tb; member
41 static void tb_queue_hotplug(struct tb *tb, u64 route, u8 port, bool unplug) in tb_queue_hotplug() argument
49 ev->tb = tb; in tb_queue_hotplug()
54 queue_work(tb->wq, &ev->work); in tb_queue_hotplug()
61 struct tb *tb = sw->tb; in tb_discover_tunnels() local
62 struct tb_cm *tcm = tb_priv(tb); in tb_discover_tunnels()
72 tunnel = tb_tunnel_discover_dp(tb, port); in tb_discover_tunnels()
76 tunnel = tb_tunnel_discover_pci(tb, port); in tb_discover_tunnels()
107 struct tb *tb = sw->tb; in tb_scan_xdomain() local
112 xd = tb_xdomain_find_by_route(tb, route); in tb_scan_xdomain()
[all …]
Dicm.c83 bool (*is_supported)(struct tb *tb);
84 int (*cio_reset)(struct tb *tb);
85 int (*get_mode)(struct tb *tb);
86 int (*get_route)(struct tb *tb, u8 link, u8 depth, u64 *route);
87 void (*save_devices)(struct tb *tb);
88 int (*driver_ready)(struct tb *tb,
91 void (*set_uuid)(struct tb *tb);
92 void (*device_connected)(struct tb *tb,
94 void (*device_disconnected)(struct tb *tb,
96 void (*xdomain_connected)(struct tb *tb,
[all …]
Dtunnel.h34 struct tb *tb; member
45 struct tb_tunnel *tb_tunnel_discover_pci(struct tb *tb, struct tb_port *down);
46 struct tb_tunnel *tb_tunnel_alloc_pci(struct tb *tb, struct tb_port *up,
48 struct tb_tunnel *tb_tunnel_discover_dp(struct tb *tb, struct tb_port *in);
49 struct tb_tunnel *tb_tunnel_alloc_dp(struct tb *tb, struct tb_port *in,
51 struct tb_tunnel *tb_tunnel_alloc_dma(struct tb *tb, struct tb_port *nhi,
/Linux-v5.4/tools/bpf/bpftool/
Dnetlink_dumper.c14 static void xdp_dump_prog_id(struct nlattr **tb, int attr, in xdp_dump_prog_id() argument
18 if (!tb[attr]) in xdp_dump_prog_id()
24 NET_DUMP_UINT("id", " id %u", libbpf_nla_getattr_u32(tb[attr])) in xdp_dump_prog_id()
32 struct nlattr *tb[IFLA_XDP_MAX + 1]; in do_xdp_dump_one() local
35 if (libbpf_nla_parse_nested(tb, IFLA_XDP_MAX, attr, NULL) < 0) in do_xdp_dump_one()
38 if (!tb[IFLA_XDP_ATTACHED]) in do_xdp_dump_one()
41 mode = libbpf_nla_getattr_u8(tb[IFLA_XDP_ATTACHED]); in do_xdp_dump_one()
55 xdp_dump_prog_id(tb, IFLA_XDP_SKB_PROG_ID, "generic", true); in do_xdp_dump_one()
56 xdp_dump_prog_id(tb, IFLA_XDP_DRV_PROG_ID, "driver", true); in do_xdp_dump_one()
57 xdp_dump_prog_id(tb, IFLA_XDP_HW_PROG_ID, "offload", true); in do_xdp_dump_one()
[all …]
/Linux-v5.4/tools/testing/selftests/powerpc/pmu/ebb/
Dtrace.c17 struct trace_buffer *tb; in trace_buffer_allocate() local
19 if (size < sizeof(*tb)) { in trace_buffer_allocate()
24 tb = mmap(NULL, size, PROT_READ | PROT_WRITE, in trace_buffer_allocate()
26 if (tb == MAP_FAILED) { in trace_buffer_allocate()
31 tb->size = size; in trace_buffer_allocate()
32 tb->tail = tb->data; in trace_buffer_allocate()
33 tb->overflow = false; in trace_buffer_allocate()
35 return tb; in trace_buffer_allocate()
38 static bool trace_check_bounds(struct trace_buffer *tb, void *p) in trace_check_bounds() argument
40 return p < ((void *)tb + tb->size); in trace_check_bounds()
[all …]
/Linux-v5.4/net/netfilter/
Dnfnetlink_cthelper.c76 struct nlattr *tb[NFCTH_TUPLE_MAX+1]; in nfnl_cthelper_parse_tuple() local
78 err = nla_parse_nested_deprecated(tb, NFCTH_TUPLE_MAX, attr, in nfnl_cthelper_parse_tuple()
83 if (!tb[NFCTH_TUPLE_L3PROTONUM] || !tb[NFCTH_TUPLE_L4PROTONUM]) in nfnl_cthelper_parse_tuple()
89 tuple->src.l3num = ntohs(nla_get_be16(tb[NFCTH_TUPLE_L3PROTONUM])); in nfnl_cthelper_parse_tuple()
90 tuple->dst.protonum = nla_get_u8(tb[NFCTH_TUPLE_L4PROTONUM]); in nfnl_cthelper_parse_tuple()
137 struct nlattr *tb[NFCTH_POLICY_MAX+1]; in nfnl_cthelper_expect_policy() local
139 err = nla_parse_nested_deprecated(tb, NFCTH_POLICY_MAX, attr, in nfnl_cthelper_expect_policy()
144 if (!tb[NFCTH_POLICY_NAME] || in nfnl_cthelper_expect_policy()
145 !tb[NFCTH_POLICY_EXPECT_MAX] || in nfnl_cthelper_expect_policy()
146 !tb[NFCTH_POLICY_EXPECT_TIMEOUT]) in nfnl_cthelper_expect_policy()
[all …]
Dnft_tunnel.c74 const struct nlattr * const tb[]) in nft_tunnel_get_init() argument
79 if (!tb[NFTA_TUNNEL_KEY] && in nft_tunnel_get_init()
80 !tb[NFTA_TUNNEL_DREG]) in nft_tunnel_get_init()
83 priv->key = ntohl(nla_get_be32(tb[NFTA_TUNNEL_KEY])); in nft_tunnel_get_init()
95 priv->dreg = nft_parse_register(tb[NFTA_TUNNEL_DREG]); in nft_tunnel_get_init()
97 if (tb[NFTA_TUNNEL_MODE]) { in nft_tunnel_get_init()
98 priv->mode = ntohl(nla_get_be32(tb[NFTA_TUNNEL_MODE])); in nft_tunnel_get_init()
166 struct nlattr *tb[NFTA_TUNNEL_KEY_IP_MAX + 1]; in nft_tunnel_obj_ip_init() local
169 err = nla_parse_nested_deprecated(tb, NFTA_TUNNEL_KEY_IP_MAX, attr, in nft_tunnel_obj_ip_init()
174 if (!tb[NFTA_TUNNEL_KEY_IP_DST]) in nft_tunnel_obj_ip_init()
[all …]
Dnft_hash.c71 const struct nlattr * const tb[]) in nft_jhash_init() argument
77 if (!tb[NFTA_HASH_SREG] || in nft_jhash_init()
78 !tb[NFTA_HASH_DREG] || in nft_jhash_init()
79 !tb[NFTA_HASH_LEN] || in nft_jhash_init()
80 !tb[NFTA_HASH_MODULUS]) in nft_jhash_init()
83 if (tb[NFTA_HASH_OFFSET]) in nft_jhash_init()
84 priv->offset = ntohl(nla_get_be32(tb[NFTA_HASH_OFFSET])); in nft_jhash_init()
86 priv->sreg = nft_parse_register(tb[NFTA_HASH_SREG]); in nft_jhash_init()
87 priv->dreg = nft_parse_register(tb[NFTA_HASH_DREG]); in nft_jhash_init()
89 err = nft_parse_u32_check(tb[NFTA_HASH_LEN], U8_MAX, &len); in nft_jhash_init()
[all …]
/Linux-v5.4/crypto/asymmetric_keys/
Dasym_tpm.c34 static int tpm_loadkey2(struct tpm_buf *tb, in tpm_loadkey2() argument
50 ret = oiap(tb, &authhandle, enonce); in tpm_loadkey2()
71 INIT_BUF(tb); in tpm_loadkey2()
72 store16(tb, TPM_TAG_RQU_AUTH1_COMMAND); in tpm_loadkey2()
73 store32(tb, TPM_LOADKEY2_SIZE + keybloblen); in tpm_loadkey2()
74 store32(tb, TPM_ORD_LOADKEY2); in tpm_loadkey2()
75 store32(tb, keyhandle); in tpm_loadkey2()
76 storebytes(tb, keyblob, keybloblen); in tpm_loadkey2()
77 store32(tb, authhandle); in tpm_loadkey2()
78 storebytes(tb, nonceodd, TPM_NONCE_SIZE); in tpm_loadkey2()
[all …]
/Linux-v5.4/arch/powerpc/kernel/
Dwatchdog.c115 u64 tb = get_tb(); in wd_lockup_ipi() local
119 cpu, tb, per_cpu(wd_timer_tb, cpu), in wd_lockup_ipi()
120 tb_to_ns(tb - per_cpu(wd_timer_tb, cpu)) / 1000000); in wd_lockup_ipi()
131 static void set_cpumask_stuck(const struct cpumask *cpumask, u64 tb) in set_cpumask_stuck() argument
136 wd_smp_last_reset_tb = tb; in set_cpumask_stuck()
142 static void set_cpu_stuck(int cpu, u64 tb) in set_cpu_stuck() argument
144 set_cpumask_stuck(cpumask_of(cpu), tb); in set_cpu_stuck()
147 static void watchdog_smp_panic(int cpu, u64 tb) in watchdog_smp_panic() argument
154 if ((s64)(tb - wd_smp_last_reset_tb) < (s64)wd_smp_panic_timeout_tb) in watchdog_smp_panic()
164 cpu, tb, wd_smp_last_reset_tb, in watchdog_smp_panic()
[all …]
/Linux-v5.4/arch/sparc/mm/
Dtlb.c26 struct tlb_batch *tb = &get_cpu_var(tlb_batch); in flush_tlb_pending() local
27 struct mm_struct *mm = tb->mm; in flush_tlb_pending()
29 if (!tb->tlb_nr) in flush_tlb_pending()
32 flush_tsb_user(tb); in flush_tlb_pending()
35 if (tb->tlb_nr == 1) { in flush_tlb_pending()
36 global_flush_tlb_page(mm, tb->vaddrs[0]); in flush_tlb_pending()
39 smp_flush_tlb_pending(tb->mm, tb->tlb_nr, in flush_tlb_pending()
40 &tb->vaddrs[0]); in flush_tlb_pending()
42 __flush_tlb_pending(CTX_HWBITS(tb->mm->context), in flush_tlb_pending()
43 tb->tlb_nr, &tb->vaddrs[0]); in flush_tlb_pending()
[all …]
/Linux-v5.4/drivers/target/
Dtarget_core_hba.c40 struct target_backend *tb, *old; in transport_backend_register() local
42 tb = kzalloc(sizeof(*tb), GFP_KERNEL); in transport_backend_register()
43 if (!tb) in transport_backend_register()
45 tb->ops = ops; in transport_backend_register()
52 kfree(tb); in transport_backend_register()
56 target_setup_backend_cits(tb); in transport_backend_register()
57 list_add_tail(&tb->list, &backend_list); in transport_backend_register()
68 struct target_backend *tb; in target_backend_unregister() local
71 list_for_each_entry(tb, &backend_list, list) { in target_backend_unregister()
72 if (tb->ops == ops) { in target_backend_unregister()
[all …]
/Linux-v5.4/net/netfilter/ipset/
Dip_set_hash_ipportip.c107 hash_ipportip4_uadt(struct ip_set *set, struct nlattr *tb[], in hash_ipportip4_uadt() argument
118 if (tb[IPSET_ATTR_LINENO]) in hash_ipportip4_uadt()
119 *lineno = nla_get_u32(tb[IPSET_ATTR_LINENO]); in hash_ipportip4_uadt()
121 if (unlikely(!tb[IPSET_ATTR_IP] || !tb[IPSET_ATTR_IP2] || in hash_ipportip4_uadt()
122 !ip_set_attr_netorder(tb, IPSET_ATTR_PORT) || in hash_ipportip4_uadt()
123 !ip_set_optattr_netorder(tb, IPSET_ATTR_PORT_TO))) in hash_ipportip4_uadt()
126 ret = ip_set_get_ipaddr4(tb[IPSET_ATTR_IP], &e.ip); in hash_ipportip4_uadt()
130 ret = ip_set_get_extensions(set, tb, &ext); in hash_ipportip4_uadt()
134 ret = ip_set_get_ipaddr4(tb[IPSET_ATTR_IP2], &e.ip2); in hash_ipportip4_uadt()
138 e.port = nla_get_be16(tb[IPSET_ATTR_PORT]); in hash_ipportip4_uadt()
[all …]
Dip_set_hash_ipportnet.c159 hash_ipportnet4_uadt(struct ip_set *set, struct nlattr *tb[], in hash_ipportnet4_uadt() argument
172 if (tb[IPSET_ATTR_LINENO]) in hash_ipportnet4_uadt()
173 *lineno = nla_get_u32(tb[IPSET_ATTR_LINENO]); in hash_ipportnet4_uadt()
175 if (unlikely(!tb[IPSET_ATTR_IP] || !tb[IPSET_ATTR_IP2] || in hash_ipportnet4_uadt()
176 !ip_set_attr_netorder(tb, IPSET_ATTR_PORT) || in hash_ipportnet4_uadt()
177 !ip_set_optattr_netorder(tb, IPSET_ATTR_PORT_TO) || in hash_ipportnet4_uadt()
178 !ip_set_optattr_netorder(tb, IPSET_ATTR_CADT_FLAGS))) in hash_ipportnet4_uadt()
181 ret = ip_set_get_hostipaddr4(tb[IPSET_ATTR_IP], &ip); in hash_ipportnet4_uadt()
185 ret = ip_set_get_extensions(set, tb, &ext); in hash_ipportnet4_uadt()
189 ret = ip_set_get_hostipaddr4(tb[IPSET_ATTR_IP2], &ip2_from); in hash_ipportnet4_uadt()
[all …]
Dip_set_hash_ipport.c104 hash_ipport4_uadt(struct ip_set *set, struct nlattr *tb[], in hash_ipport4_uadt() argument
115 if (tb[IPSET_ATTR_LINENO]) in hash_ipport4_uadt()
116 *lineno = nla_get_u32(tb[IPSET_ATTR_LINENO]); in hash_ipport4_uadt()
118 if (unlikely(!tb[IPSET_ATTR_IP] || in hash_ipport4_uadt()
119 !ip_set_attr_netorder(tb, IPSET_ATTR_PORT) || in hash_ipport4_uadt()
120 !ip_set_optattr_netorder(tb, IPSET_ATTR_PORT_TO))) in hash_ipport4_uadt()
123 ret = ip_set_get_ipaddr4(tb[IPSET_ATTR_IP], &e.ip); in hash_ipport4_uadt()
127 ret = ip_set_get_extensions(set, tb, &ext); in hash_ipport4_uadt()
131 e.port = nla_get_be16(tb[IPSET_ATTR_PORT]); in hash_ipport4_uadt()
133 if (tb[IPSET_ATTR_PROTO]) { in hash_ipport4_uadt()
[all …]
Dip_set_hash_ipmac.c110 hash_ipmac4_uadt(struct ip_set *set, struct nlattr *tb[], in hash_ipmac4_uadt() argument
118 if (unlikely(!tb[IPSET_ATTR_IP] || in hash_ipmac4_uadt()
119 !tb[IPSET_ATTR_ETHER] || in hash_ipmac4_uadt()
120 nla_len(tb[IPSET_ATTR_ETHER]) != ETH_ALEN || in hash_ipmac4_uadt()
121 !ip_set_optattr_netorder(tb, IPSET_ATTR_TIMEOUT) || in hash_ipmac4_uadt()
122 !ip_set_optattr_netorder(tb, IPSET_ATTR_PACKETS) || in hash_ipmac4_uadt()
123 !ip_set_optattr_netorder(tb, IPSET_ATTR_BYTES) || in hash_ipmac4_uadt()
124 !ip_set_optattr_netorder(tb, IPSET_ATTR_SKBMARK) || in hash_ipmac4_uadt()
125 !ip_set_optattr_netorder(tb, IPSET_ATTR_SKBPRIO) || in hash_ipmac4_uadt()
126 !ip_set_optattr_netorder(tb, IPSET_ATTR_SKBQUEUE))) in hash_ipmac4_uadt()
[all …]
Dip_set_hash_netportnet.c176 hash_netportnet4_uadt(struct ip_set *set, struct nlattr *tb[], in hash_netportnet4_uadt() argument
188 if (tb[IPSET_ATTR_LINENO]) in hash_netportnet4_uadt()
189 *lineno = nla_get_u32(tb[IPSET_ATTR_LINENO]); in hash_netportnet4_uadt()
192 if (unlikely(!tb[IPSET_ATTR_IP] || !tb[IPSET_ATTR_IP2] || in hash_netportnet4_uadt()
193 !ip_set_attr_netorder(tb, IPSET_ATTR_PORT) || in hash_netportnet4_uadt()
194 !ip_set_optattr_netorder(tb, IPSET_ATTR_PORT_TO) || in hash_netportnet4_uadt()
195 !ip_set_optattr_netorder(tb, IPSET_ATTR_CADT_FLAGS))) in hash_netportnet4_uadt()
198 ret = ip_set_get_hostipaddr4(tb[IPSET_ATTR_IP], &ip); in hash_netportnet4_uadt()
202 ret = ip_set_get_hostipaddr4(tb[IPSET_ATTR_IP2], &ip2_from); in hash_netportnet4_uadt()
206 ret = ip_set_get_extensions(set, tb, &ext); in hash_netportnet4_uadt()
[all …]
/Linux-v5.4/drivers/net/wireless/ti/wlcore/
Dtestmode.c58 static int wl1271_tm_cmd_test(struct wl1271 *wl, struct nlattr *tb[]) in wl1271_tm_cmd_test() argument
67 if (!tb[WL1271_TM_ATTR_DATA]) in wl1271_tm_cmd_test()
70 buf = nla_data(tb[WL1271_TM_ATTR_DATA]); in wl1271_tm_cmd_test()
71 buf_len = nla_len(tb[WL1271_TM_ATTR_DATA]); in wl1271_tm_cmd_test()
73 if (tb[WL1271_TM_ATTR_ANSWER]) in wl1271_tm_cmd_test()
74 answer = nla_get_u8(tb[WL1271_TM_ATTR_ANSWER]); in wl1271_tm_cmd_test()
140 static int wl1271_tm_cmd_interrogate(struct wl1271 *wl, struct nlattr *tb[]) in wl1271_tm_cmd_interrogate() argument
149 if (!tb[WL1271_TM_ATTR_IE_ID]) in wl1271_tm_cmd_interrogate()
152 ie_id = nla_get_u8(tb[WL1271_TM_ATTR_IE_ID]); in wl1271_tm_cmd_interrogate()
207 static int wl1271_tm_cmd_configure(struct wl1271 *wl, struct nlattr *tb[]) in wl1271_tm_cmd_configure() argument
[all …]

12345678910>>...17