/Linux-v5.4/fs/fscache/ |
D | stats.c | 142 atomic_read(&fscache_n_cookie_index), in fscache_stats_show() 143 atomic_read(&fscache_n_cookie_data), in fscache_stats_show() 144 atomic_read(&fscache_n_cookie_special)); in fscache_stats_show() 147 atomic_read(&fscache_n_object_alloc), in fscache_stats_show() 148 atomic_read(&fscache_n_object_no_alloc), in fscache_stats_show() 149 atomic_read(&fscache_n_object_avail), in fscache_stats_show() 150 atomic_read(&fscache_n_object_dead)); in fscache_stats_show() 152 atomic_read(&fscache_n_checkaux_none), in fscache_stats_show() 153 atomic_read(&fscache_n_checkaux_okay), in fscache_stats_show() 154 atomic_read(&fscache_n_checkaux_update), in fscache_stats_show() [all …]
|
D | histogram.c | 37 n[0] = atomic_read(&fscache_obj_instantiate_histogram[index]); in fscache_histogram_show() 38 n[1] = atomic_read(&fscache_ops_histogram[index]); in fscache_histogram_show() 39 n[2] = atomic_read(&fscache_objs_histogram[index]); in fscache_histogram_show() 40 n[3] = atomic_read(&fscache_retrieval_delay_histogram[index]); in fscache_histogram_show() 41 n[4] = atomic_read(&fscache_retrieval_histogram[index]); in fscache_histogram_show()
|
/Linux-v5.4/net/netfilter/ipvs/ |
D | ip_vs_nq.c | 45 return atomic_read(&dest->activeconns) + 1; in ip_vs_nq_dest_overhead() 77 !atomic_read(&dest->weight)) in ip_vs_nq_schedule() 83 if (atomic_read(&dest->activeconns) == 0) { in ip_vs_nq_schedule() 90 ((__s64)loh * atomic_read(&dest->weight) > in ip_vs_nq_schedule() 91 (__s64)doh * atomic_read(&least->weight))) { in ip_vs_nq_schedule() 107 atomic_read(&least->activeconns), in ip_vs_nq_schedule() 109 atomic_read(&least->weight), loh); in ip_vs_nq_schedule()
|
D | ip_vs_lblcr.c | 173 if ((atomic_read(&least->weight) > 0) in ip_vs_dest_set_min() 189 if (((__s64)loh * atomic_read(&dest->weight) > in ip_vs_dest_set_min() 190 (__s64)doh * atomic_read(&least->weight)) in ip_vs_dest_set_min() 202 atomic_read(&least->activeconns), in ip_vs_dest_set_min() 204 atomic_read(&least->weight), loh); in ip_vs_dest_set_min() 222 if (atomic_read(&most->weight) > 0) { in ip_vs_dest_set_max() 235 if (((__s64)moh * atomic_read(&dest->weight) < in ip_vs_dest_set_max() 236 (__s64)doh * atomic_read(&most->weight)) in ip_vs_dest_set_max() 237 && (atomic_read(&dest->weight) > 0)) { in ip_vs_dest_set_max() 247 atomic_read(&most->activeconns), in ip_vs_dest_set_max() [all …]
|
D | ip_vs_sed.c | 49 return atomic_read(&dest->activeconns) + 1; in ip_vs_sed_dest_overhead() 80 atomic_read(&dest->weight) > 0) { in ip_vs_sed_schedule() 97 if ((__s64)loh * atomic_read(&dest->weight) > in ip_vs_sed_schedule() 98 (__s64)doh * atomic_read(&least->weight)) { in ip_vs_sed_schedule() 108 atomic_read(&least->activeconns), in ip_vs_sed_schedule() 110 atomic_read(&least->weight), loh); in ip_vs_sed_schedule()
|
D | ip_vs_wlc.c | 52 atomic_read(&dest->weight) > 0) { in ip_vs_wlc_schedule() 69 if ((__s64)loh * atomic_read(&dest->weight) > in ip_vs_wlc_schedule() 70 (__s64)doh * atomic_read(&least->weight)) { in ip_vs_wlc_schedule() 80 atomic_read(&least->activeconns), in ip_vs_wlc_schedule() 82 atomic_read(&least->weight), loh); in ip_vs_wlc_schedule()
|
D | ip_vs_fo.c | 34 atomic_read(&dest->weight) > hw) { in ip_vs_fo_schedule() 36 hw = atomic_read(&dest->weight); in ip_vs_fo_schedule() 44 atomic_read(&hweight->activeconns), in ip_vs_fo_schedule() 45 atomic_read(&hweight->weight)); in ip_vs_fo_schedule()
|
D | ip_vs_ovf.c | 36 w = atomic_read(&dest->weight); in ip_vs_ovf_schedule() 38 atomic_read(&dest->activeconns) > w || in ip_vs_ovf_schedule() 51 atomic_read(&h->activeconns), in ip_vs_ovf_schedule() 52 atomic_read(&h->weight)); in ip_vs_ovf_schedule()
|
D | ip_vs_lblc.c | 311 if (atomic_read(&tbl->entries) <= tbl->max_size) { in ip_vs_lblc_check_expire() 316 goal = (atomic_read(&tbl->entries) - tbl->max_size)*4/3; in ip_vs_lblc_check_expire() 420 if (atomic_read(&dest->weight) > 0) { in __ip_vs_lblc_schedule() 437 if ((__s64)loh * atomic_read(&dest->weight) > in __ip_vs_lblc_schedule() 438 (__s64)doh * atomic_read(&least->weight)) { in __ip_vs_lblc_schedule() 448 atomic_read(&least->activeconns), in __ip_vs_lblc_schedule() 450 atomic_read(&least->weight), loh); in __ip_vs_lblc_schedule() 463 if (atomic_read(&dest->activeconns) > atomic_read(&dest->weight)) { in is_overloaded() 467 if (atomic_read(&d->activeconns)*2 in is_overloaded() 468 < atomic_read(&d->weight)) { in is_overloaded() [all …]
|
/Linux-v5.4/drivers/infiniband/hw/vmw_pvrdma/ |
D | pvrdma_ring.h | 71 const unsigned int idx = atomic_read(var); in pvrdma_idx() 80 __u32 idx = atomic_read(var) + 1; /* Increment. */ in pvrdma_idx_ring_inc() 89 const __u32 tail = atomic_read(&r->prod_tail); in pvrdma_idx_ring_has_space() 90 const __u32 head = atomic_read(&r->cons_head); in pvrdma_idx_ring_has_space() 103 const __u32 tail = atomic_read(&r->prod_tail); in pvrdma_idx_ring_has_data() 104 const __u32 head = atomic_read(&r->cons_head); in pvrdma_idx_ring_has_data()
|
/Linux-v5.4/drivers/lightnvm/ |
D | pblk-rl.c | 31 rb_space = atomic_read(&rl->rb_space); in pblk_rl_is_limit() 38 int rb_user_cnt = atomic_read(&rl->rb_user_cnt); in pblk_rl_user_may_insert() 39 int rb_space = atomic_read(&rl->rb_space); in pblk_rl_user_may_insert() 52 int rb_space = atomic_read(&rl->rb_space); in pblk_rl_inserted() 60 int rb_gc_cnt = atomic_read(&rl->rb_gc_cnt); in pblk_rl_gc_may_insert() 100 return atomic_read(&rl->free_blocks); in pblk_rl_nr_free_blks() 105 return atomic_read(&rl->free_user_blocks); in pblk_rl_nr_user_free_blks() 113 int werr_gc_needed = atomic_read(&rl->werr_lines); in __pblk_rl_update_rates() 161 int blk_in_line = atomic_read(&line->blk_in_line); in pblk_rl_free_lines_inc() 173 int blk_in_line = atomic_read(&line->blk_in_line); in pblk_rl_free_lines_dec() [all …]
|
/Linux-v5.4/drivers/crypto/bcm/ |
D | util.c | 374 atomic_read(&ipriv->session_count)); in spu_debugfs_read() 377 atomic_read(&ipriv->stream_count)); in spu_debugfs_read() 380 atomic_read(&ipriv->setkey_cnt[SPU_OP_CIPHER])); in spu_debugfs_read() 383 atomic_read(&ipriv->op_counts[SPU_OP_CIPHER])); in spu_debugfs_read() 386 op_cnt = atomic_read(&ipriv->cipher_cnt[alg][mode]); in spu_debugfs_read() 397 atomic_read(&ipriv->op_counts[SPU_OP_HASH])); in spu_debugfs_read() 399 op_cnt = atomic_read(&ipriv->hash_cnt[alg]); in spu_debugfs_read() 409 atomic_read(&ipriv->setkey_cnt[SPU_OP_HMAC])); in spu_debugfs_read() 412 atomic_read(&ipriv->op_counts[SPU_OP_HMAC])); in spu_debugfs_read() 414 op_cnt = atomic_read(&ipriv->hmac_cnt[alg]); in spu_debugfs_read() [all …]
|
/Linux-v5.4/sound/core/seq/ |
D | seq_lock.c | 16 if (atomic_read(lockp) < 0) { in snd_use_lock_sync_helper() 17 pr_warn("ALSA: seq_lock: lock trouble [counter = %d] in %s:%d\n", atomic_read(lockp), file, line); in snd_use_lock_sync_helper() 20 while (atomic_read(lockp) > 0) { in snd_use_lock_sync_helper() 22 pr_warn("ALSA: seq_lock: waiting [%d left] in %s:%d\n", atomic_read(lockp), file, line); in snd_use_lock_sync_helper()
|
/Linux-v5.4/kernel/sched/ |
D | membarrier.c | 40 atomic_read(&mm->membarrier_state)); in ipi_sync_rq_state() 141 if (!(atomic_read(&mm->membarrier_state) & in membarrier_private_expedited() 145 if (!(atomic_read(&mm->membarrier_state) & in membarrier_private_expedited() 150 if (atomic_read(&mm->mm_users) == 1 || num_online_cpus() == 1) in membarrier_private_expedited() 202 int membarrier_state = atomic_read(&mm->membarrier_state); in sync_runqueues_membarrier_state() 206 if (atomic_read(&mm->mm_users) == 1 || num_online_cpus() == 1) { in sync_runqueues_membarrier_state() 265 if (atomic_read(&mm->membarrier_state) & in membarrier_register_global_expedited() 298 if ((atomic_read(&mm->membarrier_state) & ready_state) == ready_state) in membarrier_register_private_expedited()
|
/Linux-v5.4/arch/openrisc/kernel/ |
D | sync-timer.c | 53 while (atomic_read(&count_count_start) != 1) in synchronise_count_master() 74 while (atomic_read(&count_count_stop) != 1) in synchronise_count_master() 104 while (atomic_read(&count_count_start) != 2) in synchronise_count_slave() 114 while (atomic_read(&count_count_stop) != 2) in synchronise_count_slave()
|
/Linux-v5.4/arch/mips/kernel/ |
D | sync-r4k.c | 50 while (atomic_read(&count_count_start) != 1) in synchronise_count_master() 71 while (atomic_read(&count_count_stop) != 1) in synchronise_count_master() 101 while (atomic_read(&count_count_start) != 2) in synchronise_count_slave() 111 while (atomic_read(&count_count_stop) != 2) in synchronise_count_slave()
|
/Linux-v5.4/drivers/s390/scsi/ |
D | zfcp_erp.c | 87 if (atomic_read(&zfcp_sdev->status) & ZFCP_STATUS_COMMON_ERP_INUSE) in zfcp_erp_action_dismiss_lun() 95 if (atomic_read(&port->status) & ZFCP_STATUS_COMMON_ERP_INUSE) in zfcp_erp_action_dismiss_port() 110 if (atomic_read(&adapter->status) & ZFCP_STATUS_COMMON_ERP_INUSE) in zfcp_erp_action_dismiss_adapter() 130 if (atomic_read(&zsdev->status) & ZFCP_STATUS_COMMON_ERP_FAILED) in zfcp_erp_handle_failed() 134 if (atomic_read(&port->status) & ZFCP_STATUS_COMMON_ERP_FAILED) in zfcp_erp_handle_failed() 138 if (atomic_read(&port->status) & in zfcp_erp_handle_failed() 147 if (atomic_read(&adapter->status) & in zfcp_erp_handle_failed() 172 l_status = atomic_read(&zfcp_sdev->status); in zfcp_erp_required_act() 175 p_status = atomic_read(&port->status); in zfcp_erp_required_act() 183 p_status = atomic_read(&port->status); in zfcp_erp_required_act() [all …]
|
/Linux-v5.4/kernel/ |
D | cred.c | 75 return atomic_read(&cred->subscribers); in read_cred_subscribers() 101 atomic_read(&cred->usage) != 0 || in put_cred_rcu() 106 atomic_read(&cred->usage), in put_cred_rcu() 109 if (atomic_read(&cred->usage) != 0) in put_cred_rcu() 111 cred, atomic_read(&cred->usage)); in put_cred_rcu() 135 atomic_read(&cred->usage), in __put_cred() 138 BUG_ON(atomic_read(&cred->usage) != 0); in __put_cred() 162 atomic_read(&tsk->cred->usage), in exit_creds() 349 p->cred, atomic_read(&p->cred->usage), in copy_creds() 440 atomic_read(&new->usage), in commit_creds() [all …]
|
/Linux-v5.4/include/asm-generic/ |
D | qspinlock.h | 26 return atomic_read(&lock->val); in queued_spin_is_locked() 41 return !atomic_read(&lock.val); in queued_spin_value_unlocked() 51 return atomic_read(&lock->val) & ~_Q_LOCKED_MASK; in queued_spin_is_contended() 60 u32 val = atomic_read(&lock->val); in queued_spin_trylock()
|
/Linux-v5.4/net/mac80211/ |
D | led.h | 18 if (!atomic_read(&local->rx_led_active)) in ieee80211_led_rx() 29 if (!atomic_read(&local->tx_led_active)) in ieee80211_led_tx() 78 if (ieee80211_is_data(fc) && atomic_read(&local->tpt_led_active)) in ieee80211_tpt_led_trig_tx() 87 if (ieee80211_is_data(fc) && atomic_read(&local->tpt_led_active)) in ieee80211_tpt_led_trig_rx()
|
/Linux-v5.4/net/batman-adv/ |
D | gateway_common.c | 140 gw_mode = atomic_read(&bat_priv->gw.mode); in batadv_gw_tvlv_container_update() 148 down = atomic_read(&bat_priv->gw.bandwidth_down); in batadv_gw_tvlv_container_update() 149 up = atomic_read(&bat_priv->gw.bandwidth_up); in batadv_gw_tvlv_container_update() 177 down_curr = (unsigned int)atomic_read(&bat_priv->gw.bandwidth_down); in batadv_gw_bandwidth_set() 178 up_curr = (unsigned int)atomic_read(&bat_priv->gw.bandwidth_up); in batadv_gw_bandwidth_set() 246 atomic_read(&bat_priv->gw.mode) == BATADV_GW_MODE_CLIENT) in batadv_gw_tvlv_ogm_handler_v1()
|
/Linux-v5.4/fs/f2fs/ |
D | debug.c | 45 si->ext_tree = atomic_read(&sbi->total_ext_tree); in update_general_status() 46 si->zombie_tree = atomic_read(&sbi->total_zombie_tree); in update_general_status() 47 si->ext_node = atomic_read(&sbi->total_ext_node); in update_general_status() 59 si->aw_cnt = atomic_read(&sbi->aw_cnt); in update_general_status() 60 si->vw_cnt = atomic_read(&sbi->vw_cnt); in update_general_status() 61 si->max_aw_cnt = atomic_read(&sbi->max_aw_cnt); in update_general_status() 62 si->max_vw_cnt = atomic_read(&sbi->max_vw_cnt); in update_general_status() 72 atomic_read(&SM_I(sbi)->fcc_info->issued_flush); in update_general_status() 74 atomic_read(&SM_I(sbi)->fcc_info->queued_flush); in update_general_status() 80 atomic_read(&SM_I(sbi)->dcc_info->issued_discard); in update_general_status() [all …]
|
/Linux-v5.4/arch/x86/kernel/ |
D | tsc_sync.c | 334 while (atomic_read(&start_count) != cpus - 1) { in check_tsc_sync_source() 335 if (atomic_read(&skip_test) > 0) { in check_tsc_sync_source() 349 while (atomic_read(&stop_count) != cpus-1) in check_tsc_sync_source() 393 if (atomic_read(&test_runs) > 0) in check_tsc_sync_source() 431 while (atomic_read(&start_count) != cpus) in check_tsc_sync_target() 449 while (atomic_read(&stop_count) != cpus) in check_tsc_sync_target() 462 if (!atomic_read(&test_runs)) in check_tsc_sync_target()
|
/Linux-v5.4/fs/btrfs/ |
D | locking.c | 41 WARN_ON(atomic_read(&eb->spinning_readers) == 0); in btrfs_assert_spinning_readers_put() 57 BUG_ON(!atomic_read(&eb->read_locks)); in btrfs_assert_tree_read_locked() 210 if (eb->blocking_writers || atomic_read(&eb->blocking_readers)) in btrfs_try_tree_write_lock() 214 if (eb->blocking_writers || atomic_read(&eb->blocking_readers)) { in btrfs_try_tree_write_lock() 264 WARN_ON(atomic_read(&eb->blocking_readers) == 0); in btrfs_tree_read_unlock_blocking() 284 wait_event(eb->read_lock_wq, atomic_read(&eb->blocking_readers) == 0); in btrfs_tree_lock() 287 if (atomic_read(&eb->blocking_readers) || eb->blocking_writers) { in btrfs_tree_lock()
|
/Linux-v5.4/net/rxrpc/ |
D | conn_object.c | 110 if (!conn || atomic_read(&conn->usage) == 0) in rxrpc_find_connection_rcu() 120 if (!conn || atomic_read(&conn->usage) == 0) { in rxrpc_find_connection_rcu() 285 int n = atomic_read(&conn->usage); in rxrpc_see_connection() 355 _enter("{%d,u=%d}", conn->debug_id, atomic_read(&conn->usage)); in rxrpc_destroy_connection() 357 ASSERTCMP(atomic_read(&conn->usage), ==, 0); in rxrpc_destroy_connection() 396 ASSERTCMP(atomic_read(&conn->usage), >, 0); in rxrpc_service_connection_reaper() 397 if (likely(atomic_read(&conn->usage) > 1)) in rxrpc_service_connection_reaper() 409 conn->debug_id, atomic_read(&conn->usage), in rxrpc_service_connection_reaper() 446 ASSERTCMP(atomic_read(&conn->usage), ==, 0); in rxrpc_service_connection_reaper() 474 conn, atomic_read(&conn->usage)); in rxrpc_destroy_all_connections() [all …]
|