Home
last modified time | relevance | path

Searched refs:qs (Results 1 – 25 of 38) sorted by relevance

12

/Linux-v5.4/fs/ocfs2/cluster/
Dquorum.c96 struct o2quo_state *qs = &o2quo_state; in o2quo_make_decision() local
98 spin_lock(&qs->qs_lock); in o2quo_make_decision()
100 lowest_hb = find_first_bit(qs->qs_hb_bm, O2NM_MAX_NODES); in o2quo_make_decision()
102 lowest_reachable = test_bit(lowest_hb, qs->qs_conn_bm); in o2quo_make_decision()
105 "lowest: %d (%sreachable)\n", qs->qs_heartbeating, in o2quo_make_decision()
106 qs->qs_connected, lowest_hb, lowest_reachable ? "" : "un"); in o2quo_make_decision()
108 if (!test_bit(o2nm_this_node(), qs->qs_hb_bm) || in o2quo_make_decision()
109 qs->qs_heartbeating == 1) in o2quo_make_decision()
112 if (qs->qs_heartbeating & 1) { in o2quo_make_decision()
115 quorum = (qs->qs_heartbeating + 1)/2; in o2quo_make_decision()
[all …]
/Linux-v5.4/kernel/bpf/
Dqueue_stack_maps.c30 static bool queue_stack_map_is_empty(struct bpf_queue_stack *qs) in queue_stack_map_is_empty() argument
32 return qs->head == qs->tail; in queue_stack_map_is_empty()
35 static bool queue_stack_map_is_full(struct bpf_queue_stack *qs) in queue_stack_map_is_full() argument
37 u32 head = qs->head + 1; in queue_stack_map_is_full()
39 if (unlikely(head >= qs->size)) in queue_stack_map_is_full()
42 return head == qs->tail; in queue_stack_map_is_full()
71 struct bpf_queue_stack *qs; in queue_stack_map_alloc() local
75 cost = queue_size = sizeof(*qs) + size * attr->value_size; in queue_stack_map_alloc()
81 qs = bpf_map_area_alloc(queue_size, numa_node); in queue_stack_map_alloc()
82 if (!qs) { in queue_stack_map_alloc()
[all …]
/Linux-v5.4/drivers/net/ethernet/chelsio/cxgb3/
Dsge.c730 static void init_qset_cntxt(struct sge_qset *qs, unsigned int id) in init_qset_cntxt() argument
732 qs->rspq.cntxt_id = id; in init_qset_cntxt()
733 qs->fl[0].cntxt_id = 2 * id; in init_qset_cntxt()
734 qs->fl[1].cntxt_id = 2 * id + 1; in init_qset_cntxt()
735 qs->txq[TXQ_ETH].cntxt_id = FW_TUNNEL_SGEEC_START + id; in init_qset_cntxt()
736 qs->txq[TXQ_ETH].token = FW_TUNNEL_TID_START + id; in init_qset_cntxt()
737 qs->txq[TXQ_OFLD].cntxt_id = FW_OFLD_SGEEC_START + id; in init_qset_cntxt()
738 qs->txq[TXQ_CTRL].cntxt_id = FW_CTRL_SGEEC_START + id; in init_qset_cntxt()
739 qs->txq[TXQ_CTRL].token = FW_CTRL_TID_START + id; in init_qset_cntxt()
1249 struct sge_qset *qs, struct sge_txq *q) in t3_stop_tx_queue() argument
[all …]
Dcxgb3_main.c411 adap->sge.qs[qidx]. in request_msix_data_irqs()
414 &adap->sge.qs[qidx]); in request_msix_data_irqs()
418 &adap->sge.qs[qidx]); in request_msix_data_irqs()
438 &adapter->sge.qs[i]); in free_irq_resources()
448 while (adap->sge.qs[0].rspq.offload_pkts < init_cnt + n) { in await_mgmt_replies()
461 unsigned long cnt = adap->sge.qs[0].rspq.offload_pkts; in init_tp_parity()
597 struct sge_qset *qs = &adap->sge.qs[i]; in ring_dbs() local
599 if (qs->adap) in ring_dbs()
601 t3_write_reg(adap, A_SG_KDOORBELL, F_SELEGRCNTX | V_EGRCNTX(qs->txq[j].cntxt_id)); in ring_dbs()
610 struct sge_qset *qs = &adap->sge.qs[i]; in init_napi() local
[all …]
Dadapter.h68 struct sge_qset *qs; member
216 struct sge_qset qs[SGE_QSETS]; member
324 void t3_update_qset_coalesce(struct sge_qset *qs, const struct qset_params *p);
/Linux-v5.4/drivers/net/ethernet/cavium/thunder/
Dnicvf_queues.c381 struct queue_set *qs = nic->qs; in nicvf_refill_rbdr() local
382 int rbdr_idx = qs->rbdr_cnt; in nicvf_refill_rbdr()
394 rbdr = &qs->rbdr[rbdr_idx]; in nicvf_refill_rbdr()
403 if (qcount >= (qs->rbdr_len - 1)) in nicvf_refill_rbdr()
406 refill_rb_cnt = qs->rbdr_len - qcount - 1; in nicvf_refill_rbdr()
627 struct queue_set *qs, int qidx) in nicvf_reclaim_snd_queue() argument
639 struct queue_set *qs, int qidx) in nicvf_reclaim_rcv_queue() argument
649 struct queue_set *qs, int qidx) in nicvf_reclaim_cmp_queue() argument
744 static void nicvf_rcv_queue_config(struct nicvf *nic, struct queue_set *qs, in nicvf_rcv_queue_config() argument
751 rq = &qs->rq[qidx]; in nicvf_rcv_queue_config()
[all …]
Dnicvf_ethtool.c218 for (qidx = 0; qidx < nic->qs->rq_cnt; qidx++) { in nicvf_get_qset_strings()
226 for (qidx = 0; qidx < nic->qs->sq_cnt; qidx++) { in nicvf_get_qset_strings()
283 (nic->qs->rq_cnt + nic->qs->sq_cnt); in nicvf_get_sset_count()
291 (snic->qs->rq_cnt + snic->qs->sq_cnt); in nicvf_get_sset_count()
307 for (qidx = 0; qidx < nic->qs->rq_cnt; qidx++) { in nicvf_get_qset_stats()
310 *((*data)++) = ((u64 *)&nic->qs->rq[qidx].stats) in nicvf_get_qset_stats()
314 for (qidx = 0; qidx < nic->qs->sq_cnt; qidx++) { in nicvf_get_qset_stats()
317 *((*data)++) = ((u64 *)&nic->qs->sq[qidx].stats) in nicvf_get_qset_stats()
472 struct queue_set *qs = nic->qs; in nicvf_get_ringparam() local
475 ring->rx_pending = qs->cq_len; in nicvf_get_ringparam()
[all …]
Dnicvf_main.c321 mbx.cpi_cfg.rq_cnt = nic->qs->rq_cnt; in nicvf_config_cpi()
443 nic->snicvf[sqs]->qs->rq_cnt = MAX_RCV_QUEUES_PER_QS; in nicvf_request_sqs()
446 nic->snicvf[sqs]->qs->rq_cnt = rx_queues; in nicvf_request_sqs()
451 nic->snicvf[sqs]->qs->sq_cnt = MAX_SND_QUEUES_PER_QS; in nicvf_request_sqs()
454 nic->snicvf[sqs]->qs->sq_cnt = tx_queues; in nicvf_request_sqs()
458 nic->snicvf[sqs]->qs->cq_cnt = in nicvf_request_sqs()
459 max(nic->snicvf[sqs]->qs->rq_cnt, nic->snicvf[sqs]->qs->sq_cnt); in nicvf_request_sqs()
670 sq = &nic->qs->sq[cqe_tx->sq_idx]; in nicvf_snd_pkt_handler()
855 struct queue_set *qs = nic->qs; in nicvf_cq_intr_handler() local
856 struct cmp_queue *cq = &qs->cq[cq_idx]; in nicvf_cq_intr_handler()
[all …]
/Linux-v5.4/drivers/soc/qcom/
Dsocinfo.c413 struct qcom_socinfo *qs; in qcom_socinfo_probe() local
424 qs = devm_kzalloc(&pdev->dev, sizeof(*qs), GFP_KERNEL); in qcom_socinfo_probe()
425 if (!qs) in qcom_socinfo_probe()
428 qs->attr.family = "Snapdragon"; in qcom_socinfo_probe()
429 qs->attr.machine = socinfo_machine(&pdev->dev, in qcom_socinfo_probe()
431 qs->attr.revision = devm_kasprintf(&pdev->dev, GFP_KERNEL, "%u.%u", in qcom_socinfo_probe()
435 qs->attr.serial_number = devm_kasprintf(&pdev->dev, GFP_KERNEL, in qcom_socinfo_probe()
439 qs->soc_dev = soc_device_register(&qs->attr); in qcom_socinfo_probe()
440 if (IS_ERR(qs->soc_dev)) in qcom_socinfo_probe()
441 return PTR_ERR(qs->soc_dev); in qcom_socinfo_probe()
[all …]
/Linux-v5.4/fs/qnx4/
Dinode.c46 struct qnx4_sb_info *qs; in qnx4_remount() local
49 qs = qnx4_sb(sb); in qnx4_remount()
50 qs->Version = QNX4_VERSION; in qnx4_remount()
192 struct qnx4_sb_info *qs; in qnx4_fill_super() local
194 qs = kzalloc(sizeof(struct qnx4_sb_info), GFP_KERNEL); in qnx4_fill_super()
195 if (!qs) in qnx4_fill_super()
197 s->s_fs_info = qs; in qnx4_fill_super()
241 struct qnx4_sb_info *qs = qnx4_sb(sb); in qnx4_kill_sb() local
243 if (qs) { in qnx4_kill_sb()
244 kfree(qs->BitMap); in qnx4_kill_sb()
[all …]
/Linux-v5.4/drivers/net/ethernet/chelsio/cxgb4vf/
Dcxgb4vf_main.c369 int qs, msi; in name_msix_vecs() local
371 for (qs = 0, msi = MSIX_IQFLINT; qs < pi->nqsets; qs++, msi++) { in name_msix_vecs()
373 "%s-%d", dev->name, qs); in name_msix_vecs()
636 int qs; in setup_sge_queues() local
638 for (qs = 0; qs < pi->nqsets; qs++, rxq++, txq++) { in setup_sge_queues()
646 netdev_get_tx_queue(dev, qs), in setup_sge_queues()
651 rxq->rspq.idx = qs; in setup_sge_queues()
667 int qs; in setup_sge_queues() local
669 for (qs = 0; qs < pi->nqsets; qs++, rxq++, txq++) { in setup_sge_queues()
709 int qs, err; in setup_rss() local
[all …]
/Linux-v5.4/drivers/net/ethernet/hisilicon/hns/
Dhns_ae_adapt.c113 ae_handle->qs = (struct hnae_queue **)(&ae_handle->qs + 1); in hns_ae_get_handle()
115 ae_handle->qs[i] = &ring_pair_cb->q; in hns_ae_get_handle()
116 ae_handle->qs[i]->rx_ring.q = ae_handle->qs[i]; in hns_ae_get_handle()
117 ae_handle->qs[i]->tx_ring.q = ae_handle->qs[i]; in hns_ae_get_handle()
147 hns_ae_get_ring_pair(handle->qs[i])->used_by_vf = 0; in hns_ae_put_handle()
161 ret = hns_rcb_wait_tx_ring_clean(handle->qs[i]); in hns_ae_wait_flow_down()
193 hns_rcb_ring_enable_hw(handle->qs[i], val); in hns_ae_ring_enable_all()
319 q = handle->qs[i]; in hns_ae_set_mtu()
347 hns_rcb_int_clr_hw(handle->qs[k], in hns_ae_start()
350 hns_rcbv2_int_clr_hw(handle->qs[k], in hns_ae_start()
[all …]
Dhnae.c297 hnae_fini_queue(handle->qs[i]); in hnae_reinit_handle()
303 ret = hnae_init_queue(handle, handle->qs[i], handle->dev); in hnae_reinit_handle()
310 hnae_fini_queue(handle->qs[j]); in hnae_reinit_handle()
349 ret = hnae_init_queue(handle, handle->qs[i], dev); in hnae_get_handle()
362 hnae_fini_queue(handle->qs[j]); in hnae_get_handle()
376 hnae_fini_queue(h->qs[i]); in hnae_put_handle()
Dhns_dsaf_rcb.h134 void hns_rcb_wait_fbd_clean(struct hnae_queue **qs, int q_num, u32 flag);
135 int hns_rcb_wait_tx_ring_clean(struct hnae_queue *qs);
Dhns_dsaf_rcb.c40 void hns_rcb_wait_fbd_clean(struct hnae_queue **qs, int q_num, u32 flag) in hns_rcb_wait_fbd_clean() argument
49 fbd_num += dsaf_read_dev(qs[i], in hns_rcb_wait_fbd_clean()
52 fbd_num += dsaf_read_dev(qs[i], in hns_rcb_wait_fbd_clean()
61 dev_err(qs[i]->handle->owner_dev, in hns_rcb_wait_fbd_clean()
65 int hns_rcb_wait_tx_ring_clean(struct hnae_queue *qs) in hns_rcb_wait_tx_ring_clean() argument
70 tail = dsaf_read_dev(&qs->tx_ring, RCB_REG_TAIL); in hns_rcb_wait_tx_ring_clean()
73 head = dsaf_read_dev(&qs->tx_ring, RCB_REG_HEAD); in hns_rcb_wait_tx_ring_clean()
81 dev_err(qs->dev->dev, "rcb wait timeout, head not equal to tail.\n"); in hns_rcb_wait_tx_ring_clean()
Dhns_enet.c371 hnae_queue_xmit(priv->ae_handle->qs[skb->queue_mapping], buf_num); in hns_nic_net_xmit_hw()
1657 ring = &h->qs[i]->rx_ring; in hns_nic_clear_all_rx_fetch()
1908 tx_bytes += h->qs[idx]->tx_ring.stats.tx_bytes; in hns_nic_get_stats64()
1909 tx_pkts += h->qs[idx]->tx_ring.stats.tx_pkts; in hns_nic_get_stats64()
1910 rx_bytes += h->qs[idx]->rx_ring.stats.rx_bytes; in hns_nic_get_stats64()
1911 rx_pkts += h->qs[idx]->rx_ring.stats.rx_pkts; in hns_nic_get_stats64()
2009 i, h->qs[i]->tx_ring.next_to_clean); in hns_nic_dump()
2011 i, h->qs[i]->tx_ring.next_to_use); in hns_nic_dump()
2013 i, h->qs[i]->rx_ring.next_to_clean); in hns_nic_dump()
2015 i, h->qs[i]->rx_ring.next_to_use); in hns_nic_dump()
[all …]
Dhnae.h567 struct hnae_queue **qs; /* array base of all queues */ member
671 ring = &h->qs[i]->rx_ring; in hnae_reinit_all_ring_desc()
686 ring = &h->qs[i]->rx_ring; in hnae_reinit_all_ring_page_off()
/Linux-v5.4/drivers/staging/fieldbus/anybuss/
Dhost.c382 struct kfifo qs[3]; member
875 static bool qs_have_work(struct kfifo *qs, size_t num) in qs_have_work() argument
881 for (i = 0; i < num; i++, qs++) { in qs_have_work()
882 ret = kfifo_out_peek(qs, &t, sizeof(t)); in qs_have_work()
892 struct kfifo *qs = cd->qs; in process_qs() local
893 size_t nqs = ARRAY_SIZE(cd->qs); in process_qs()
895 for (i = 0; i < nqs; i++, qs++) in process_qs()
896 process_q(cd, qs); in process_qs()
968 struct kfifo *qs = cd->qs; in qthread_fn() local
969 size_t nqs = ARRAY_SIZE(cd->qs); in qthread_fn()
[all …]
/Linux-v5.4/fs/qnx6/
Dinode.c305 struct qnx6_sb_info *qs; in qnx6_fill_super() local
310 qs = kzalloc(sizeof(struct qnx6_sb_info), GFP_KERNEL); in qnx6_fill_super()
311 if (!qs) in qnx6_fill_super()
313 s->s_fs_info = qs; in qnx6_fill_super()
480 kfree(qs); in qnx6_fill_super()
487 struct qnx6_sb_info *qs = QNX6_SB(sb); in qnx6_put_super() local
488 brelse(qs->sb_buf); in qnx6_put_super()
489 iput(qs->longfile); in qnx6_put_super()
490 iput(qs->inodes); in qnx6_put_super()
491 kfree(qs); in qnx6_put_super()
/Linux-v5.4/Documentation/devicetree/bindings/net/
Dmscc-ocelot.txt14 - "qs"
63 reg-names = "sys", "rew", "qs", "ptp", "port0", "port1",
/Linux-v5.4/net/sched/
Dsch_fq_codel.c620 struct gnet_stats_queue qs = { 0 }; in fq_codel_dump_class_stats() local
647 qs.qlen++; in fq_codel_dump_class_stats()
652 qs.backlog = q->backlogs[idx]; in fq_codel_dump_class_stats()
653 qs.drops = 0; in fq_codel_dump_class_stats()
655 if (gnet_stats_copy_queue(d, NULL, &qs, qs.qlen) < 0) in fq_codel_dump_class_stats()
Dsch_sfq.c858 struct gnet_stats_queue qs = { 0 }; in sfq_dump_class_stats() local
865 qs.qlen = slot->qlen; in sfq_dump_class_stats()
866 qs.backlog = slot->backlog; in sfq_dump_class_stats()
868 if (gnet_stats_copy_queue(d, NULL, &qs, qs.qlen) < 0) in sfq_dump_class_stats()
/Linux-v5.4/drivers/md/
Ddm-cache-policy-smq.c257 struct ilist qs[MAX_LEVELS]; member
278 l_init(q->qs + i); in q_init()
302 l_add_tail(q->es, q->qs + e->level, e); in q_push()
312 l_add_head(q->es, q->qs + e->level, e); in q_push_front()
322 l_add_before(q->es, q->qs + e->level, old, e); in q_push_before()
327 l_del(q->es, q->qs + e->level, e); in q_del()
343 for (e = l_head(q->es, q->qs + level); e; e = l_next(q->es, e)) { in q_peek()
377 for (e = l_head(q->es, q->qs + level); e; e = l_next(q->es, e)) in __redist_pop_from()
379 l_del(q->es, q->qs + e->level, e); in __redist_pop_from()
436 l = q->qs + level; in q_redistribute()
[all …]
/Linux-v5.4/arch/s390/include/uapi/asm/
Druntime_instr.h22 __u32 qs : 1; member
/Linux-v5.4/include/uapi/sound/
Dhdspm.h40 qs enumerator

12