Home
last modified time | relevance | path

Searched refs:q (Results 1 – 25 of 1317) sorted by relevance

12345678910>>...53

/Linux-v5.15/crypto/
Dgf128mul.c56 #define gf128mul_dat(q) { \ argument
57 q(0x00), q(0x01), q(0x02), q(0x03), q(0x04), q(0x05), q(0x06), q(0x07),\
58 q(0x08), q(0x09), q(0x0a), q(0x0b), q(0x0c), q(0x0d), q(0x0e), q(0x0f),\
59 q(0x10), q(0x11), q(0x12), q(0x13), q(0x14), q(0x15), q(0x16), q(0x17),\
60 q(0x18), q(0x19), q(0x1a), q(0x1b), q(0x1c), q(0x1d), q(0x1e), q(0x1f),\
61 q(0x20), q(0x21), q(0x22), q(0x23), q(0x24), q(0x25), q(0x26), q(0x27),\
62 q(0x28), q(0x29), q(0x2a), q(0x2b), q(0x2c), q(0x2d), q(0x2e), q(0x2f),\
63 q(0x30), q(0x31), q(0x32), q(0x33), q(0x34), q(0x35), q(0x36), q(0x37),\
64 q(0x38), q(0x39), q(0x3a), q(0x3b), q(0x3c), q(0x3d), q(0x3e), q(0x3f),\
65 q(0x40), q(0x41), q(0x42), q(0x43), q(0x44), q(0x45), q(0x46), q(0x47),\
[all …]
/Linux-v5.15/Documentation/networking/
Dtls-offload-layers.svg1q-0.609375 -0.359375 -1.203125 -0.359375q-0.546875 0 -0.96875 0.328125q-0.421875 0.328125 -0.60937…
/Linux-v5.15/drivers/media/v4l2-core/
Dvideobuf-core.c50 #define CALL(q, f, arg...) \ argument
51 ((q->int_ops->f) ? q->int_ops->f(arg) : 0)
52 #define CALLPTR(q, f, arg...) \ argument
53 ((q->int_ops->f) ? q->int_ops->f(arg) : NULL)
55 struct videobuf_buffer *videobuf_alloc_vb(struct videobuf_queue *q) in videobuf_alloc_vb() argument
59 BUG_ON(q->msize < sizeof(*vb)); in videobuf_alloc_vb()
61 if (!q->int_ops || !q->int_ops->alloc_vb) { in videobuf_alloc_vb()
66 vb = q->int_ops->alloc_vb(q->msize); in videobuf_alloc_vb()
76 static int state_neither_active_nor_queued(struct videobuf_queue *q, in state_neither_active_nor_queued() argument
82 spin_lock_irqsave(q->irqlock, flags); in state_neither_active_nor_queued()
[all …]
/Linux-v5.15/sound/core/seq/
Dseq_queue.c50 static int queue_list_add(struct snd_seq_queue *q) in queue_list_add() argument
58 queue_list[i] = q; in queue_list_add()
59 q->queue = i; in queue_list_add()
71 struct snd_seq_queue *q; in queue_list_remove() local
75 q = queue_list[id]; in queue_list_remove()
76 if (q) { in queue_list_remove()
77 spin_lock(&q->owner_lock); in queue_list_remove()
78 if (q->owner == client) { in queue_list_remove()
80 q->klocked = 1; in queue_list_remove()
81 spin_unlock(&q->owner_lock); in queue_list_remove()
[all …]
/Linux-v5.15/sound/core/seq/oss/
Dseq_oss_readq.c35 struct seq_oss_readq *q; in snd_seq_oss_readq_new() local
37 q = kzalloc(sizeof(*q), GFP_KERNEL); in snd_seq_oss_readq_new()
38 if (!q) in snd_seq_oss_readq_new()
41 q->q = kcalloc(maxlen, sizeof(union evrec), GFP_KERNEL); in snd_seq_oss_readq_new()
42 if (!q->q) { in snd_seq_oss_readq_new()
43 kfree(q); in snd_seq_oss_readq_new()
47 q->maxlen = maxlen; in snd_seq_oss_readq_new()
48 q->qlen = 0; in snd_seq_oss_readq_new()
49 q->head = q->tail = 0; in snd_seq_oss_readq_new()
50 init_waitqueue_head(&q->midi_sleep); in snd_seq_oss_readq_new()
[all …]
Dseq_oss_writeq.c27 struct seq_oss_writeq *q; in snd_seq_oss_writeq_new() local
30 q = kzalloc(sizeof(*q), GFP_KERNEL); in snd_seq_oss_writeq_new()
31 if (!q) in snd_seq_oss_writeq_new()
33 q->dp = dp; in snd_seq_oss_writeq_new()
34 q->maxlen = maxlen; in snd_seq_oss_writeq_new()
35 spin_lock_init(&q->sync_lock); in snd_seq_oss_writeq_new()
36 q->sync_event_put = 0; in snd_seq_oss_writeq_new()
37 q->sync_time = 0; in snd_seq_oss_writeq_new()
38 init_waitqueue_head(&q->sync_sleep); in snd_seq_oss_writeq_new()
47 return q; in snd_seq_oss_writeq_new()
[all …]
/Linux-v5.15/net/sched/
Dsch_choke.c76 static unsigned int choke_len(const struct choke_sched_data *q) in choke_len() argument
78 return (q->tail - q->head) & q->tab_mask; in choke_len()
82 static int use_ecn(const struct choke_sched_data *q) in use_ecn() argument
84 return q->flags & TC_RED_ECN; in use_ecn()
88 static int use_harddrop(const struct choke_sched_data *q) in use_harddrop() argument
90 return q->flags & TC_RED_HARDDROP; in use_harddrop()
94 static void choke_zap_head_holes(struct choke_sched_data *q) in choke_zap_head_holes() argument
97 q->head = (q->head + 1) & q->tab_mask; in choke_zap_head_holes()
98 if (q->head == q->tail) in choke_zap_head_holes()
100 } while (q->tab[q->head] == NULL); in choke_zap_head_holes()
[all …]
Dsch_netem.c200 static bool loss_4state(struct netem_sched_data *q) in loss_4state() argument
202 struct clgstate *clg = &q->clg; in loss_4state()
265 static bool loss_gilb_ell(struct netem_sched_data *q) in loss_gilb_ell() argument
267 struct clgstate *clg = &q->clg; in loss_gilb_ell()
286 static bool loss_event(struct netem_sched_data *q) in loss_event() argument
288 switch (q->loss_model) { in loss_event()
291 return q->loss && q->loss >= get_crandom(&q->loss_cor); in loss_event()
299 return loss_4state(q); in loss_event()
307 return loss_gilb_ell(q); in loss_event()
345 static u64 packet_time_ns(u64 len, const struct netem_sched_data *q) in packet_time_ns() argument
[all …]
Dsch_sfq.c150 static inline struct sfq_head *sfq_dep_head(struct sfq_sched_data *q, sfq_index val) in sfq_dep_head() argument
153 return &q->slots[val].dep; in sfq_dep_head()
154 return &q->dep[val - SFQ_MAX_FLOWS]; in sfq_dep_head()
157 static unsigned int sfq_hash(const struct sfq_sched_data *q, in sfq_hash() argument
160 return skb_get_hash_perturb(skb, &q->perturbation) & (q->divisor - 1); in sfq_hash()
166 struct sfq_sched_data *q = qdisc_priv(sch); in sfq_classify() local
173 TC_H_MIN(skb->priority) <= q->divisor) in sfq_classify()
176 fl = rcu_dereference_bh(q->filter_list); in sfq_classify()
178 return sfq_hash(q, skb) + 1; in sfq_classify()
194 if (TC_H_MIN(res.classid) <= q->divisor) in sfq_classify()
[all …]
Dsch_sfb.c123 static void increment_one_qlen(u32 sfbhash, u32 slot, struct sfb_sched_data *q) in increment_one_qlen() argument
126 struct sfb_bucket *b = &q->bins[slot].bins[0][0]; in increment_one_qlen()
138 static void increment_qlen(const struct sk_buff *skb, struct sfb_sched_data *q) in increment_qlen() argument
144 increment_one_qlen(sfbhash, 0, q); in increment_qlen()
148 increment_one_qlen(sfbhash, 1, q); in increment_qlen()
152 struct sfb_sched_data *q) in decrement_one_qlen() argument
155 struct sfb_bucket *b = &q->bins[slot].bins[0][0]; in decrement_one_qlen()
167 static void decrement_qlen(const struct sk_buff *skb, struct sfb_sched_data *q) in decrement_qlen() argument
173 decrement_one_qlen(sfbhash, 0, q); in decrement_qlen()
177 decrement_one_qlen(sfbhash, 1, q); in decrement_qlen()
[all …]
Dsch_fq_pie.c73 static unsigned int fq_pie_hash(const struct fq_pie_sched_data *q, in fq_pie_hash() argument
76 return reciprocal_scale(skb_get_hash(skb), q->flows_cnt); in fq_pie_hash()
82 struct fq_pie_sched_data *q = qdisc_priv(sch); in fq_pie_classify() local
89 TC_H_MIN(skb->priority) <= q->flows_cnt) in fq_pie_classify()
92 filter = rcu_dereference_bh(q->filter_list); in fq_pie_classify()
94 return fq_pie_hash(q, skb) + 1; in fq_pie_classify()
110 if (TC_H_MIN(res.classid) <= q->flows_cnt) in fq_pie_classify()
131 struct fq_pie_sched_data *q = qdisc_priv(sch); in fq_pie_qdisc_enqueue() local
149 sel_flow = &q->flows[idx]; in fq_pie_qdisc_enqueue()
152 memory_limited = q->memory_usage > q->memory_limit + skb->truesize; in fq_pie_qdisc_enqueue()
[all …]
Dsch_skbprio.c40 static u16 calc_new_high_prio(const struct skbprio_sched_data *q) in calc_new_high_prio() argument
44 for (prio = q->highest_prio - 1; prio >= q->lowest_prio; prio--) { in calc_new_high_prio()
45 if (!skb_queue_empty(&q->qdiscs[prio])) in calc_new_high_prio()
53 static u16 calc_new_low_prio(const struct skbprio_sched_data *q) in calc_new_low_prio() argument
57 for (prio = q->lowest_prio + 1; prio <= q->highest_prio; prio++) { in calc_new_low_prio()
58 if (!skb_queue_empty(&q->qdiscs[prio])) in calc_new_low_prio()
72 struct skbprio_sched_data *q = qdisc_priv(sch); in skbprio_enqueue() local
81 qdisc = &q->qdiscs[prio]; in skbprio_enqueue()
82 if (sch->q.qlen < sch->limit) { in skbprio_enqueue()
85 q->qstats[prio].backlog += qdisc_pkt_len(skb); in skbprio_enqueue()
[all …]
Dsch_fq.c172 static void fq_flow_unset_throttled(struct fq_sched_data *q, struct fq_flow *f) in fq_flow_unset_throttled() argument
174 rb_erase(&f->rate_node, &q->delayed); in fq_flow_unset_throttled()
175 q->throttled_flows--; in fq_flow_unset_throttled()
176 fq_flow_add_tail(&q->old_flows, f); in fq_flow_unset_throttled()
179 static void fq_flow_set_throttled(struct fq_sched_data *q, struct fq_flow *f) in fq_flow_set_throttled() argument
181 struct rb_node **p = &q->delayed.rb_node, *parent = NULL; in fq_flow_set_throttled()
194 rb_insert_color(&f->rate_node, &q->delayed); in fq_flow_set_throttled()
195 q->throttled_flows++; in fq_flow_set_throttled()
196 q->stat_throttled++; in fq_flow_set_throttled()
199 if (q->time_next_delayed_flow > f->time_next_packet) in fq_flow_set_throttled()
[all …]
Dsch_fq_codel.c70 static unsigned int fq_codel_hash(const struct fq_codel_sched_data *q, in fq_codel_hash() argument
73 return reciprocal_scale(skb_get_hash(skb), q->flows_cnt); in fq_codel_hash()
79 struct fq_codel_sched_data *q = qdisc_priv(sch); in fq_codel_classify() local
86 TC_H_MIN(skb->priority) <= q->flows_cnt) in fq_codel_classify()
89 filter = rcu_dereference_bh(q->filter_list); in fq_codel_classify()
91 return fq_codel_hash(q, skb) + 1; in fq_codel_classify()
107 if (TC_H_MIN(res.classid) <= q->flows_cnt) in fq_codel_classify()
140 struct fq_codel_sched_data *q = qdisc_priv(sch); in fq_codel_drop() local
154 for (i = 0; i < q->flows_cnt; i++) { in fq_codel_drop()
155 if (q->backlogs[i] > maxbacklog) { in fq_codel_drop()
[all …]
/Linux-v5.15/net/xdp/
Dxsk_queue.h114 static inline bool xskq_cons_read_addr_unchecked(struct xsk_queue *q, u64 *addr) in xskq_cons_read_addr_unchecked() argument
116 struct xdp_umem_ring *ring = (struct xdp_umem_ring *)q->ring; in xskq_cons_read_addr_unchecked()
118 if (q->cached_cons != q->cached_prod) { in xskq_cons_read_addr_unchecked()
119 u32 idx = q->cached_cons & q->ring_mask; in xskq_cons_read_addr_unchecked()
175 static inline bool xskq_cons_is_valid_desc(struct xsk_queue *q, in xskq_cons_is_valid_desc() argument
180 q->invalid_descs++; in xskq_cons_is_valid_desc()
186 static inline bool xskq_cons_read_desc(struct xsk_queue *q, in xskq_cons_read_desc() argument
190 while (q->cached_cons != q->cached_prod) { in xskq_cons_read_desc()
191 struct xdp_rxtx_ring *ring = (struct xdp_rxtx_ring *)q->ring; in xskq_cons_read_desc()
192 u32 idx = q->cached_cons & q->ring_mask; in xskq_cons_read_desc()
[all …]
/Linux-v5.15/block/
Dblk-sysfs.c61 static ssize_t queue_requests_show(struct request_queue *q, char *page) in queue_requests_show() argument
63 return queue_var_show(q->nr_requests, page); in queue_requests_show()
67 queue_requests_store(struct request_queue *q, const char *page, size_t count) in queue_requests_store() argument
72 if (!queue_is_mq(q)) in queue_requests_store()
82 err = blk_mq_update_nr_requests(q, nr); in queue_requests_store()
89 static ssize_t queue_ra_show(struct request_queue *q, char *page) in queue_ra_show() argument
93 if (!q->disk) in queue_ra_show()
95 ra_kb = q->disk->bdi->ra_pages << (PAGE_SHIFT - 10); in queue_ra_show()
100 queue_ra_store(struct request_queue *q, const char *page, size_t count) in queue_ra_store() argument
105 if (!q->disk) in queue_ra_store()
[all …]
Dblk-pm.c31 void blk_pm_runtime_init(struct request_queue *q, struct device *dev) in blk_pm_runtime_init() argument
33 q->dev = dev; in blk_pm_runtime_init()
34 q->rpm_status = RPM_ACTIVE; in blk_pm_runtime_init()
35 pm_runtime_set_autosuspend_delay(q->dev, -1); in blk_pm_runtime_init()
36 pm_runtime_use_autosuspend(q->dev); in blk_pm_runtime_init()
61 int blk_pre_runtime_suspend(struct request_queue *q) in blk_pre_runtime_suspend() argument
65 if (!q->dev) in blk_pre_runtime_suspend()
68 WARN_ON_ONCE(q->rpm_status != RPM_ACTIVE); in blk_pre_runtime_suspend()
70 spin_lock_irq(&q->queue_lock); in blk_pre_runtime_suspend()
71 q->rpm_status = RPM_SUSPENDING; in blk_pre_runtime_suspend()
[all …]
/Linux-v5.15/drivers/media/common/videobuf2/
Dvideobuf2-core.c37 #define dprintk(q, level, fmt, arg...) \ argument
40 pr_info("[%s] %s: " fmt, (q)->name, __func__, \
93 #define log_qop(q, op) \ argument
94 dprintk(q, 2, "call_qop(%s)%s\n", #op, \
95 (q)->ops->op ? "" : " (nop)")
97 #define call_qop(q, op, args...) \ argument
101 log_qop(q, op); \
102 err = (q)->ops->op ? (q)->ops->op(args) : 0; \
104 (q)->cnt_ ## op++; \
108 #define call_void_qop(q, op, args...) \ argument
[all …]
/Linux-v5.15/drivers/infiniband/sw/rxe/
Drxe_queue.h64 void rxe_queue_reset(struct rxe_queue *q);
69 int rxe_queue_resize(struct rxe_queue *q, unsigned int *num_elem_p,
79 static inline int next_index(struct rxe_queue *q, int index) in next_index() argument
81 return (index + 1) & q->buf->index_mask; in next_index()
84 static inline int queue_empty(struct rxe_queue *q, enum queue_type type) in queue_empty() argument
92 prod = smp_load_acquire(&q->buf->producer_index); in queue_empty()
93 cons = q->index; in queue_empty()
96 prod = q->index; in queue_empty()
98 cons = smp_load_acquire(&q->buf->consumer_index); in queue_empty()
101 prod = q->buf->producer_index; in queue_empty()
[all …]
/Linux-v5.15/drivers/spi/
Dspi-fsl-qspi.c278 static inline int needs_swap_endian(struct fsl_qspi *q) in needs_swap_endian() argument
280 return q->devtype_data->quirks & QUADSPI_QUIRK_SWAP_ENDIAN; in needs_swap_endian()
283 static inline int needs_4x_clock(struct fsl_qspi *q) in needs_4x_clock() argument
285 return q->devtype_data->quirks & QUADSPI_QUIRK_4X_INT_CLK; in needs_4x_clock()
288 static inline int needs_fill_txfifo(struct fsl_qspi *q) in needs_fill_txfifo() argument
290 return q->devtype_data->quirks & QUADSPI_QUIRK_TKT253890; in needs_fill_txfifo()
293 static inline int needs_wakeup_wait_mode(struct fsl_qspi *q) in needs_wakeup_wait_mode() argument
295 return q->devtype_data->quirks & QUADSPI_QUIRK_TKT245618; in needs_wakeup_wait_mode()
298 static inline int needs_amba_base_offset(struct fsl_qspi *q) in needs_amba_base_offset() argument
300 return !(q->devtype_data->quirks & QUADSPI_QUIRK_BASE_INTERNAL); in needs_amba_base_offset()
[all …]
/Linux-v5.15/drivers/s390/cio/
Dqdio_main.c115 static int qdio_do_eqbs(struct qdio_q *q, unsigned char *state, in qdio_do_eqbs() argument
118 int tmp_count = count, tmp_start = start, nr = q->nr; in qdio_do_eqbs()
121 qperf_inc(q, eqbs); in qdio_do_eqbs()
123 if (!q->is_input_q) in qdio_do_eqbs()
124 nr += q->irq_ptr->nr_input_qs; in qdio_do_eqbs()
126 ccq = do_eqbs(q->irq_ptr->sch_token, state, nr, &tmp_start, &tmp_count, in qdio_do_eqbs()
136 qperf_inc(q, eqbs_partial); in qdio_do_eqbs()
137 DBF_DEV_EVENT(DBF_INFO, q->irq_ptr, "EQBS part:%02x", in qdio_do_eqbs()
142 DBF_DEV_EVENT(DBF_WARN, q->irq_ptr, "EQBS again:%2d", ccq); in qdio_do_eqbs()
145 DBF_ERROR("%4x ccq:%3d", SCH_NO(q), ccq); in qdio_do_eqbs()
[all …]
/Linux-v5.15/drivers/net/wireless/mediatek/mt76/
Dsdio.c23 struct mt76_queue *q = &dev->q_rx[qid]; in mt76s_alloc_rx_queue() local
25 spin_lock_init(&q->lock); in mt76s_alloc_rx_queue()
26 q->entry = devm_kcalloc(dev->dev, in mt76s_alloc_rx_queue()
27 MT_NUM_RX_ENTRIES, sizeof(*q->entry), in mt76s_alloc_rx_queue()
29 if (!q->entry) in mt76s_alloc_rx_queue()
32 q->ndesc = MT_NUM_RX_ENTRIES; in mt76s_alloc_rx_queue()
33 q->head = q->tail = 0; in mt76s_alloc_rx_queue()
34 q->queued = 0; in mt76s_alloc_rx_queue()
41 struct mt76_queue *q; in mt76s_alloc_tx_queue() local
43 q = devm_kzalloc(dev->dev, sizeof(*q), GFP_KERNEL); in mt76s_alloc_tx_queue()
[all …]
Ddma.c83 mt76_dma_sync_idx(struct mt76_dev *dev, struct mt76_queue *q) in mt76_dma_sync_idx() argument
85 writel(q->desc_dma, &q->regs->desc_base); in mt76_dma_sync_idx()
86 writel(q->ndesc, &q->regs->ring_size); in mt76_dma_sync_idx()
87 q->head = readl(&q->regs->dma_idx); in mt76_dma_sync_idx()
88 q->tail = q->head; in mt76_dma_sync_idx()
92 mt76_dma_queue_reset(struct mt76_dev *dev, struct mt76_queue *q) in mt76_dma_queue_reset() argument
96 if (!q) in mt76_dma_queue_reset()
100 for (i = 0; i < q->ndesc; i++) in mt76_dma_queue_reset()
101 q->desc[i].ctrl = cpu_to_le32(MT_DMA_CTL_DMA_DONE); in mt76_dma_queue_reset()
103 writel(0, &q->regs->cpu_idx); in mt76_dma_queue_reset()
[all …]
/Linux-v5.15/drivers/net/wireless/broadcom/b43/
Dpio.c24 static u16 generate_cookie(struct b43_pio_txqueue *q, in generate_cookie() argument
37 cookie = (((u16)q->index + 1) << 12); in generate_cookie()
49 struct b43_pio_txqueue *q = NULL; in parse_cookie() local
54 q = pio->tx_queue_AC_BK; in parse_cookie()
57 q = pio->tx_queue_AC_BE; in parse_cookie()
60 q = pio->tx_queue_AC_VI; in parse_cookie()
63 q = pio->tx_queue_AC_VO; in parse_cookie()
66 q = pio->tx_queue_mcast; in parse_cookie()
69 if (B43_WARN_ON(!q)) in parse_cookie()
72 if (B43_WARN_ON(pack_index >= ARRAY_SIZE(q->packets))) in parse_cookie()
[all …]
/Linux-v5.15/drivers/gpu/drm/amd/amdkfd/
Dkfd_queue.c27 void print_queue_properties(struct queue_properties *q) in print_queue_properties() argument
29 if (!q) in print_queue_properties()
33 pr_debug("Queue Type: %u\n", q->type); in print_queue_properties()
34 pr_debug("Queue Size: %llu\n", q->queue_size); in print_queue_properties()
35 pr_debug("Queue percent: %u\n", q->queue_percent); in print_queue_properties()
36 pr_debug("Queue Address: 0x%llX\n", q->queue_address); in print_queue_properties()
37 pr_debug("Queue Id: %u\n", q->queue_id); in print_queue_properties()
38 pr_debug("Queue Process Vmid: %u\n", q->vmid); in print_queue_properties()
39 pr_debug("Queue Read Pointer: 0x%px\n", q->read_ptr); in print_queue_properties()
40 pr_debug("Queue Write Pointer: 0x%px\n", q->write_ptr); in print_queue_properties()
[all …]

12345678910>>...53