Lines Matching full:offload

9 #include <linux/can/rx-offload.h>
24 can_rx_offload_le(struct can_rx_offload *offload, in can_rx_offload_le() argument
27 if (offload->inc) in can_rx_offload_le()
34 can_rx_offload_inc(struct can_rx_offload *offload, unsigned int *val) in can_rx_offload_inc() argument
36 if (offload->inc) in can_rx_offload_inc()
44 struct can_rx_offload *offload = container_of(napi, in can_rx_offload_napi_poll() local
47 struct net_device *dev = offload->dev; in can_rx_offload_napi_poll()
53 (skb = skb_dequeue(&offload->skb_queue))) { in can_rx_offload_napi_poll()
66 if (!skb_queue_empty(&offload->skb_queue)) in can_rx_offload_napi_poll()
67 napi_reschedule(&offload->napi); in can_rx_offload_napi_poll()
70 can_led_event(offload->dev, CAN_LED_EVENT_RX); in can_rx_offload_napi_poll()
120 * @offload: pointer to rx_offload context
140 can_rx_offload_offload_one(struct can_rx_offload *offload, unsigned int n) in can_rx_offload_offload_one() argument
148 if (unlikely(skb_queue_len(&offload->skb_queue) > in can_rx_offload_offload_one()
149 offload->skb_queue_len_max)) in can_rx_offload_offload_one()
152 skb = offload->mailbox_read(offload, n, &timestamp, drop); in can_rx_offload_offload_one()
161 offload->dev->stats.rx_dropped++; in can_rx_offload_offload_one()
162 offload->dev->stats.rx_fifo_errors++; in can_rx_offload_offload_one()
174 int can_rx_offload_irq_offload_timestamp(struct can_rx_offload *offload, in can_rx_offload_irq_offload_timestamp() argument
180 for (i = offload->mb_first; in can_rx_offload_irq_offload_timestamp()
181 can_rx_offload_le(offload, i, offload->mb_last); in can_rx_offload_irq_offload_timestamp()
182 can_rx_offload_inc(offload, &i)) { in can_rx_offload_irq_offload_timestamp()
188 skb = can_rx_offload_offload_one(offload, i); in can_rx_offload_irq_offload_timestamp()
192 __skb_queue_add_sort(&offload->skb_irq_queue, skb, in can_rx_offload_irq_offload_timestamp()
201 int can_rx_offload_irq_offload_fifo(struct can_rx_offload *offload) in can_rx_offload_irq_offload_fifo() argument
207 skb = can_rx_offload_offload_one(offload, 0); in can_rx_offload_irq_offload_fifo()
213 __skb_queue_tail(&offload->skb_irq_queue, skb); in can_rx_offload_irq_offload_fifo()
221 int can_rx_offload_queue_sorted(struct can_rx_offload *offload, in can_rx_offload_queue_sorted() argument
226 if (skb_queue_len(&offload->skb_queue) > in can_rx_offload_queue_sorted()
227 offload->skb_queue_len_max) { in can_rx_offload_queue_sorted()
235 __skb_queue_add_sort(&offload->skb_irq_queue, skb, in can_rx_offload_queue_sorted()
242 unsigned int can_rx_offload_get_echo_skb(struct can_rx_offload *offload, in can_rx_offload_get_echo_skb() argument
246 struct net_device *dev = offload->dev; in can_rx_offload_get_echo_skb()
256 err = can_rx_offload_queue_sorted(offload, skb, timestamp); in can_rx_offload_get_echo_skb()
266 int can_rx_offload_queue_tail(struct can_rx_offload *offload, in can_rx_offload_queue_tail() argument
269 if (skb_queue_len(&offload->skb_queue) > in can_rx_offload_queue_tail()
270 offload->skb_queue_len_max) { in can_rx_offload_queue_tail()
275 __skb_queue_tail(&offload->skb_irq_queue, skb); in can_rx_offload_queue_tail()
281 void can_rx_offload_irq_finish(struct can_rx_offload *offload) in can_rx_offload_irq_finish() argument
286 if (skb_queue_empty_lockless(&offload->skb_irq_queue)) in can_rx_offload_irq_finish()
289 spin_lock_irqsave(&offload->skb_queue.lock, flags); in can_rx_offload_irq_finish()
290 skb_queue_splice_tail_init(&offload->skb_irq_queue, &offload->skb_queue); in can_rx_offload_irq_finish()
291 spin_unlock_irqrestore(&offload->skb_queue.lock, flags); in can_rx_offload_irq_finish()
293 queue_len = skb_queue_len(&offload->skb_queue); in can_rx_offload_irq_finish()
294 if (queue_len > offload->skb_queue_len_max / 8) in can_rx_offload_irq_finish()
295 netdev_dbg(offload->dev, "%s: queue_len=%d\n", in can_rx_offload_irq_finish()
298 napi_schedule(&offload->napi); in can_rx_offload_irq_finish()
302 void can_rx_offload_threaded_irq_finish(struct can_rx_offload *offload) in can_rx_offload_threaded_irq_finish() argument
307 if (skb_queue_empty_lockless(&offload->skb_irq_queue)) in can_rx_offload_threaded_irq_finish()
310 spin_lock_irqsave(&offload->skb_queue.lock, flags); in can_rx_offload_threaded_irq_finish()
311 skb_queue_splice_tail_init(&offload->skb_irq_queue, &offload->skb_queue); in can_rx_offload_threaded_irq_finish()
312 spin_unlock_irqrestore(&offload->skb_queue.lock, flags); in can_rx_offload_threaded_irq_finish()
314 queue_len = skb_queue_len(&offload->skb_queue); in can_rx_offload_threaded_irq_finish()
315 if (queue_len > offload->skb_queue_len_max / 8) in can_rx_offload_threaded_irq_finish()
316 netdev_dbg(offload->dev, "%s: queue_len=%d\n", in can_rx_offload_threaded_irq_finish()
320 napi_schedule(&offload->napi); in can_rx_offload_threaded_irq_finish()
326 struct can_rx_offload *offload, in can_rx_offload_init_queue() argument
329 offload->dev = dev; in can_rx_offload_init_queue()
332 offload->skb_queue_len_max = 2 << fls(weight); in can_rx_offload_init_queue()
333 offload->skb_queue_len_max *= 4; in can_rx_offload_init_queue()
334 skb_queue_head_init(&offload->skb_queue); in can_rx_offload_init_queue()
335 __skb_queue_head_init(&offload->skb_irq_queue); in can_rx_offload_init_queue()
337 netif_napi_add(dev, &offload->napi, can_rx_offload_napi_poll, weight); in can_rx_offload_init_queue()
340 __func__, offload->skb_queue_len_max); in can_rx_offload_init_queue()
346 struct can_rx_offload *offload) in can_rx_offload_add_timestamp() argument
350 if (offload->mb_first > BITS_PER_LONG_LONG || in can_rx_offload_add_timestamp()
351 offload->mb_last > BITS_PER_LONG_LONG || !offload->mailbox_read) in can_rx_offload_add_timestamp()
354 if (offload->mb_first < offload->mb_last) { in can_rx_offload_add_timestamp()
355 offload->inc = true; in can_rx_offload_add_timestamp()
356 weight = offload->mb_last - offload->mb_first; in can_rx_offload_add_timestamp()
358 offload->inc = false; in can_rx_offload_add_timestamp()
359 weight = offload->mb_first - offload->mb_last; in can_rx_offload_add_timestamp()
362 return can_rx_offload_init_queue(dev, offload, weight); in can_rx_offload_add_timestamp()
367 struct can_rx_offload *offload, unsigned int weight) in can_rx_offload_add_fifo() argument
369 if (!offload->mailbox_read) in can_rx_offload_add_fifo()
372 return can_rx_offload_init_queue(dev, offload, weight); in can_rx_offload_add_fifo()
377 struct can_rx_offload *offload, in can_rx_offload_add_manual() argument
380 if (offload->mailbox_read) in can_rx_offload_add_manual()
383 return can_rx_offload_init_queue(dev, offload, weight); in can_rx_offload_add_manual()
387 void can_rx_offload_enable(struct can_rx_offload *offload) in can_rx_offload_enable() argument
389 napi_enable(&offload->napi); in can_rx_offload_enable()
393 void can_rx_offload_del(struct can_rx_offload *offload) in can_rx_offload_del() argument
395 netif_napi_del(&offload->napi); in can_rx_offload_del()
396 skb_queue_purge(&offload->skb_queue); in can_rx_offload_del()
397 __skb_queue_purge(&offload->skb_irq_queue); in can_rx_offload_del()