Lines Matching refs:skb_out

1192 	struct sk_buff *skb_out;  in cdc_ncm_fill_tx_frame()  local
1215 skb_out = ctx->tx_curr_skb; in cdc_ncm_fill_tx_frame()
1218 if (!skb_out) { in cdc_ncm_fill_tx_frame()
1221 skb_out = alloc_skb(ctx->tx_curr_size, GFP_ATOMIC); in cdc_ncm_fill_tx_frame()
1227 if (skb_out == NULL) { in cdc_ncm_fill_tx_frame()
1233 if (skb_out == NULL) { in cdc_ncm_fill_tx_frame()
1243 skb_out = alloc_skb(ctx->tx_curr_size, GFP_ATOMIC); in cdc_ncm_fill_tx_frame()
1246 if (skb_out == NULL) { in cdc_ncm_fill_tx_frame()
1257 nth.nth16 = skb_put_zero(skb_out, sizeof(struct usb_cdc_ncm_nth16)); in cdc_ncm_fill_tx_frame()
1263 nth.nth32 = skb_put_zero(skb_out, sizeof(struct usb_cdc_ncm_nth32)); in cdc_ncm_fill_tx_frame()
1290 ndp.ndp16 = cdc_ncm_ndp16(ctx, skb_out, sign, skb->len + ctx->tx_modulus + ctx->tx_remainder); in cdc_ncm_fill_tx_frame()
1292 ndp.ndp32 = cdc_ncm_ndp32(ctx, skb_out, sign, skb->len + ctx->tx_modulus + ctx->tx_remainder); in cdc_ncm_fill_tx_frame()
1295 cdc_ncm_align_tail(skb_out, ctx->tx_modulus, ctx->tx_remainder, ctx->tx_curr_size); in cdc_ncm_fill_tx_frame()
1299 skb_out->len + skb->len + delayed_ndp_size > ctx->tx_curr_size) { in cdc_ncm_fill_tx_frame()
1327 ndp.ndp16->dpe16[index].wDatagramIndex = cpu_to_le16(skb_out->len); in cdc_ncm_fill_tx_frame()
1334 ndp.ndp32->dpe32[index].dwDatagramIndex = cpu_to_le32(skb_out->len); in cdc_ncm_fill_tx_frame()
1337 skb_put_data(skb_out, skb->data, skb->len); in cdc_ncm_fill_tx_frame()
1362 ctx->tx_curr_skb = skb_out; in cdc_ncm_fill_tx_frame()
1368 ctx->tx_curr_skb = skb_out; in cdc_ncm_fill_tx_frame()
1384 nth.nth16 = (struct usb_cdc_ncm_nth16 *)skb_out->data; in cdc_ncm_fill_tx_frame()
1385 cdc_ncm_align_tail(skb_out, ctx->tx_ndp_modulus, 0, ctx->tx_curr_size - ctx->max_ndp_size); in cdc_ncm_fill_tx_frame()
1386 nth.nth16->wNdpIndex = cpu_to_le16(skb_out->len); in cdc_ncm_fill_tx_frame()
1387 skb_put_data(skb_out, ctx->delayed_ndp16, ctx->max_ndp_size); in cdc_ncm_fill_tx_frame()
1392 nth.nth32 = (struct usb_cdc_ncm_nth32 *)skb_out->data; in cdc_ncm_fill_tx_frame()
1393 cdc_ncm_align_tail(skb_out, ctx->tx_ndp_modulus, 0, ctx->tx_curr_size - ctx->max_ndp_size); in cdc_ncm_fill_tx_frame()
1394 nth.nth32->dwNdpIndex = cpu_to_le32(skb_out->len); in cdc_ncm_fill_tx_frame()
1395 skb_put_data(skb_out, ctx->delayed_ndp32, ctx->max_ndp_size); in cdc_ncm_fill_tx_frame()
1411 skb_out->len > ctx->min_tx_pkt) { in cdc_ncm_fill_tx_frame()
1412 padding_count = ctx->tx_curr_size - skb_out->len; in cdc_ncm_fill_tx_frame()
1413 skb_put_zero(skb_out, padding_count); in cdc_ncm_fill_tx_frame()
1414 } else if (skb_out->len < ctx->tx_curr_size && in cdc_ncm_fill_tx_frame()
1415 (skb_out->len % dev->maxpacket) == 0) { in cdc_ncm_fill_tx_frame()
1416 skb_put_u8(skb_out, 0); /* force short packet */ in cdc_ncm_fill_tx_frame()
1421 nth.nth16 = (struct usb_cdc_ncm_nth16 *)skb_out->data; in cdc_ncm_fill_tx_frame()
1422 nth.nth16->wBlockLength = cpu_to_le16(skb_out->len); in cdc_ncm_fill_tx_frame()
1424 nth.nth32 = (struct usb_cdc_ncm_nth32 *)skb_out->data; in cdc_ncm_fill_tx_frame()
1425 nth.nth32->dwBlockLength = cpu_to_le32(skb_out->len); in cdc_ncm_fill_tx_frame()
1432 ctx->tx_overhead += skb_out->len - ctx->tx_curr_frame_payload; in cdc_ncm_fill_tx_frame()
1439 usbnet_set_skb_tx_stats(skb_out, n, in cdc_ncm_fill_tx_frame()
1440 (long)ctx->tx_curr_frame_payload - skb_out->len); in cdc_ncm_fill_tx_frame()
1442 return skb_out; in cdc_ncm_fill_tx_frame()
1495 struct sk_buff *skb_out; in cdc_ncm_tx_fixup() local
1511 skb_out = cdc_ncm_fill_tx_frame(dev, skb, cpu_to_le32(USB_CDC_NCM_NDP16_NOCRC_SIGN)); in cdc_ncm_tx_fixup()
1513 skb_out = cdc_ncm_fill_tx_frame(dev, skb, cpu_to_le32(USB_CDC_NCM_NDP32_NOCRC_SIGN)); in cdc_ncm_tx_fixup()
1516 return skb_out; in cdc_ncm_tx_fixup()