Searched refs:CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS (Results 1 – 25 of 39) sorted by relevance
12
85 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) in is_link_local_ether_addr()103 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) in is_zero_ether_addr()121 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) in is_multicast_ether_addr()135 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && BITS_PER_LONG == 64 in is_multicast_ether_addr_64bits()281 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) in ether_addr_copy()320 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) in ether_addr_equal()350 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && BITS_PER_LONG == 64 in ether_addr_equal_64bits()374 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) in ether_addr_equal_unaligned()491 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && BITS_PER_LONG == 64 in compare_ether_header()
25 #ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS80 #ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS in siphash()94 #ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS133 #ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS in hsiphash()
725 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) in compare_vlan_header()
12 #ifdef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS17 # ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS25 # ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS
18 #ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS in flat_get_addr_from_rp()27 #ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS in flat_put_addr_at_rp()
72 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) in __crypto_memneq_generic()96 #ifdef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS in __crypto_memneq_16()
976 if (IS_ENABLED(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) || in crypto_inc()993 if (!IS_ENABLED(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS)) { in __crypto_xor()
84 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) in lzo1x_decompress_safe()185 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) in lzo1x_decompress_safe()227 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) in lzo1x_decompress_safe()
93 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && defined(LZO_USE_CTZ64) in lzo1x_1_do_compress()113 #elif defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && defined(LZO_USE_CTZ32) in lzo1x_1_do_compress()
442 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && BITS_PER_LONG == 64 in ipv6_masked_addr_cmp()489 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && BITS_PER_LONG == 64 in __ipv6_addr_set_half()517 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && BITS_PER_LONG == 64 in ipv6_addr_equal()530 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && BITS_PER_LONG == 64579 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && BITS_PER_LONG == 64 in ipv6_addr_any()591 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && BITS_PER_LONG == 64 in ipv6_addr_hash()615 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && BITS_PER_LONG == 64 in ipv6_addr_loopback()632 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && BITS_PER_LONG == 64 in ipv6_addr_v4mapped()716 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && BITS_PER_LONG == 64736 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && BITS_PER_LONG == 64 in __ipv6_addr_diff()
443 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && BITS_PER_LONG == 64 in ipv6_addr_is_ll_all_nodes()455 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && BITS_PER_LONG == 64 in ipv6_addr_is_ll_all_routers()472 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && BITS_PER_LONG == 64 in ipv6_addr_is_solict_mult()
1346 #ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS in nla_need_padding_for_64bit()1386 #ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS in nla_total_size_64bit()
53 #ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS
13 #ifdef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS
84 #ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS284 #ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS442 #ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS
16 return IS_ENABLED(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) || in alignment_ok()
187 #ifdef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS in strscpy()
50 #ifdef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS in needs_unaligned_copy()
206 if (IS_ENABLED(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && in crypto_xor()224 if (IS_ENABLED(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && in crypto_xor_cpy()
80 unaligned accesses (CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS=y). One could
269 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && BITS_PER_LONG == 64 in rds_addr_cmp()
155 #ifdef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS259 made dependent on CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS like so::261 #ifdef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS
278 #ifdef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS in inflate_fast()
2260 #ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS in dw_mci_push_data16()2296 #ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS in dw_mci_pull_data16()2343 #ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS in dw_mci_push_data32()2379 #ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS in dw_mci_pull_data32()2427 #ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS in dw_mci_push_data64()2463 #ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS in dw_mci_pull_data64()
245 #if !defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) in brcmf_skb_is_iapp()254 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) in brcmf_skb_is_iapp()