Searched refs:CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS (Results 1 – 25 of 40) sorted by relevance
12
12 #ifdef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS17 # ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS25 # ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS
10 #ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS in flat_get_addr_from_rp()19 #ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS in flat_put_addr_at_rp()
82 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) in is_link_local_ether_addr()100 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) in is_zero_ether_addr()118 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) in is_multicast_ether_addr()132 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && BITS_PER_LONG == 64 in is_multicast_ether_addr_64bits()289 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) in ether_addr_copy()328 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) in ether_addr_equal()358 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && BITS_PER_LONG == 64 in ether_addr_equal_64bits()382 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) in ether_addr_equal_unaligned()511 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && BITS_PER_LONG == 64 in compare_ether_header()
30 #ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS85 #ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS in siphash()99 #ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS138 #ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS in hsiphash()
27 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && BITS_PER_LONG == 64 in nf_inet_addr_cmp()44 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && BITS_PER_LONG == 64 in nf_inet_addr_mask()
759 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) in compare_vlan_header()
72 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) in __crypto_memneq_generic()96 #ifdef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS in __crypto_memneq_16()
95 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) in lzo1x_decompress_safe()213 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) in lzo1x_decompress_safe()255 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) in lzo1x_decompress_safe()
56 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && \ in lzo1x_1_do_compress()166 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && defined(LZO_USE_CTZ64) in lzo1x_1_do_compress()186 #elif defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && defined(LZO_USE_CTZ32) in lzo1x_1_do_compress()
500 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && BITS_PER_LONG == 64 in ipv6_masked_addr_cmp()547 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && BITS_PER_LONG == 64 in __ipv6_addr_set_half()575 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && BITS_PER_LONG == 64 in ipv6_addr_equal()588 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && BITS_PER_LONG == 64637 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && BITS_PER_LONG == 64 in ipv6_addr_any()649 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && BITS_PER_LONG == 64 in ipv6_addr_hash()673 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && BITS_PER_LONG == 64 in ipv6_addr_loopback()690 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && BITS_PER_LONG == 64 in ipv6_addr_v4mapped()779 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && BITS_PER_LONG == 64799 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && BITS_PER_LONG == 64 in __ipv6_addr_diff()
444 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && BITS_PER_LONG == 64 in ipv6_addr_is_ll_all_nodes()456 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && BITS_PER_LONG == 64 in ipv6_addr_is_ll_all_routers()473 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && BITS_PER_LONG == 64 in ipv6_addr_is_solict_mult()488 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && BITS_PER_LONG == 64 in ipv6_addr_is_all_snoopers()
1867 #ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS in nla_need_padding_for_64bit()1907 #ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS in nla_total_size_64bit()
141 if (IS_ENABLED(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && in crypto_xor()159 if (IS_ENABLED(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && in crypto_xor_cpy()
47 #ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS
15 #ifdef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS
84 #ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS284 #ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS442 #ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS
38 #ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS in is_aligned()
47 #ifdef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS in needs_unaligned_copy()
28 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) in efx_rx_buf_hash()
80 unaligned accesses (CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS=y). One could
270 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && BITS_PER_LONG == 64 in rds_addr_cmp()
155 #ifdef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS259 made dependent on CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS like so::261 #ifdef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS
257 #ifdef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS in inflate_fast()
2234 #ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS in dw_mci_push_data16()2270 #ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS in dw_mci_pull_data16()2317 #ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS in dw_mci_push_data32()2353 #ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS in dw_mci_pull_data32()2401 #ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS in dw_mci_push_data64()2437 #ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS in dw_mci_pull_data64()
41 # define CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS 1 macro1019 #ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS in do_test_single()