Lines Matching refs:skb_out
1205 struct sk_buff *skb_out; in cdc_ncm_fill_tx_frame() local
1231 skb_out = ctx->tx_curr_skb; in cdc_ncm_fill_tx_frame()
1234 if (!skb_out) { in cdc_ncm_fill_tx_frame()
1237 skb_out = alloc_skb(ctx->tx_curr_size, GFP_ATOMIC); in cdc_ncm_fill_tx_frame()
1243 if (skb_out == NULL) { in cdc_ncm_fill_tx_frame()
1249 if (skb_out == NULL) { in cdc_ncm_fill_tx_frame()
1259 skb_out = alloc_skb(ctx->tx_curr_size, GFP_ATOMIC); in cdc_ncm_fill_tx_frame()
1262 if (skb_out == NULL) { in cdc_ncm_fill_tx_frame()
1273 nth.nth16 = skb_put_zero(skb_out, sizeof(struct usb_cdc_ncm_nth16)); in cdc_ncm_fill_tx_frame()
1279 nth.nth32 = skb_put_zero(skb_out, sizeof(struct usb_cdc_ncm_nth32)); in cdc_ncm_fill_tx_frame()
1306 ndp.ndp16 = cdc_ncm_ndp16(ctx, skb_out, sign, skb->len + ctx->tx_modulus + ctx->tx_remainder); in cdc_ncm_fill_tx_frame()
1308 ndp.ndp32 = cdc_ncm_ndp32(ctx, skb_out, sign, skb->len + ctx->tx_modulus + ctx->tx_remainder); in cdc_ncm_fill_tx_frame()
1311 cdc_ncm_align_tail(skb_out, ctx->tx_modulus, ctx->tx_remainder, ctx->tx_curr_size); in cdc_ncm_fill_tx_frame()
1315 skb_out->len + skb->len + delayed_ndp_size > ctx->tx_curr_size) { in cdc_ncm_fill_tx_frame()
1343 ndp.ndp16->dpe16[index].wDatagramIndex = cpu_to_le16(skb_out->len); in cdc_ncm_fill_tx_frame()
1350 ndp.ndp32->dpe32[index].dwDatagramIndex = cpu_to_le32(skb_out->len); in cdc_ncm_fill_tx_frame()
1353 skb_put_data(skb_out, skb->data, skb->len); in cdc_ncm_fill_tx_frame()
1378 ctx->tx_curr_skb = skb_out; in cdc_ncm_fill_tx_frame()
1384 ctx->tx_curr_skb = skb_out; in cdc_ncm_fill_tx_frame()
1400 nth.nth16 = (struct usb_cdc_ncm_nth16 *)skb_out->data; in cdc_ncm_fill_tx_frame()
1401 cdc_ncm_align_tail(skb_out, ctx->tx_ndp_modulus, 0, ctx->tx_curr_size - ctx->max_ndp_size); in cdc_ncm_fill_tx_frame()
1402 nth.nth16->wNdpIndex = cpu_to_le16(skb_out->len); in cdc_ncm_fill_tx_frame()
1403 skb_put_data(skb_out, ctx->delayed_ndp16, ctx->max_ndp_size); in cdc_ncm_fill_tx_frame()
1408 nth.nth32 = (struct usb_cdc_ncm_nth32 *)skb_out->data; in cdc_ncm_fill_tx_frame()
1409 cdc_ncm_align_tail(skb_out, ctx->tx_ndp_modulus, 0, ctx->tx_curr_size - ctx->max_ndp_size); in cdc_ncm_fill_tx_frame()
1410 nth.nth32->dwNdpIndex = cpu_to_le32(skb_out->len); in cdc_ncm_fill_tx_frame()
1411 skb_put_data(skb_out, ctx->delayed_ndp32, ctx->max_ndp_size); in cdc_ncm_fill_tx_frame()
1427 skb_out->len > ctx->min_tx_pkt) { in cdc_ncm_fill_tx_frame()
1428 padding_count = ctx->tx_curr_size - skb_out->len; in cdc_ncm_fill_tx_frame()
1430 skb_put_zero(skb_out, padding_count); in cdc_ncm_fill_tx_frame()
1431 } else if (skb_out->len < ctx->tx_curr_size && in cdc_ncm_fill_tx_frame()
1432 (skb_out->len % dev->maxpacket) == 0) { in cdc_ncm_fill_tx_frame()
1433 skb_put_u8(skb_out, 0); /* force short packet */ in cdc_ncm_fill_tx_frame()
1438 nth.nth16 = (struct usb_cdc_ncm_nth16 *)skb_out->data; in cdc_ncm_fill_tx_frame()
1439 nth.nth16->wBlockLength = cpu_to_le16(skb_out->len); in cdc_ncm_fill_tx_frame()
1441 nth.nth32 = (struct usb_cdc_ncm_nth32 *)skb_out->data; in cdc_ncm_fill_tx_frame()
1442 nth.nth32->dwBlockLength = cpu_to_le32(skb_out->len); in cdc_ncm_fill_tx_frame()
1449 ctx->tx_overhead += skb_out->len - ctx->tx_curr_frame_payload; in cdc_ncm_fill_tx_frame()
1456 usbnet_set_skb_tx_stats(skb_out, n, in cdc_ncm_fill_tx_frame()
1457 (long)ctx->tx_curr_frame_payload - skb_out->len); in cdc_ncm_fill_tx_frame()
1459 return skb_out; in cdc_ncm_fill_tx_frame()
1512 struct sk_buff *skb_out; in cdc_ncm_tx_fixup() local
1528 skb_out = cdc_ncm_fill_tx_frame(dev, skb, cpu_to_le32(USB_CDC_NCM_NDP16_NOCRC_SIGN)); in cdc_ncm_tx_fixup()
1530 skb_out = cdc_ncm_fill_tx_frame(dev, skb, cpu_to_le32(USB_CDC_NCM_NDP32_NOCRC_SIGN)); in cdc_ncm_tx_fixup()
1533 return skb_out; in cdc_ncm_tx_fixup()