Lines Matching refs:skb_out
1208 struct sk_buff *skb_out; in cdc_ncm_fill_tx_frame() local
1234 skb_out = ctx->tx_curr_skb; in cdc_ncm_fill_tx_frame()
1237 if (!skb_out) { in cdc_ncm_fill_tx_frame()
1240 skb_out = alloc_skb(ctx->tx_curr_size, GFP_ATOMIC); in cdc_ncm_fill_tx_frame()
1246 if (skb_out == NULL) { in cdc_ncm_fill_tx_frame()
1252 if (skb_out == NULL) { in cdc_ncm_fill_tx_frame()
1262 skb_out = alloc_skb(ctx->tx_curr_size, GFP_ATOMIC); in cdc_ncm_fill_tx_frame()
1265 if (skb_out == NULL) { in cdc_ncm_fill_tx_frame()
1276 nth.nth16 = skb_put_zero(skb_out, sizeof(struct usb_cdc_ncm_nth16)); in cdc_ncm_fill_tx_frame()
1282 nth.nth32 = skb_put_zero(skb_out, sizeof(struct usb_cdc_ncm_nth32)); in cdc_ncm_fill_tx_frame()
1309 ndp.ndp16 = cdc_ncm_ndp16(ctx, skb_out, sign, skb->len + ctx->tx_modulus + ctx->tx_remainder); in cdc_ncm_fill_tx_frame()
1311 ndp.ndp32 = cdc_ncm_ndp32(ctx, skb_out, sign, skb->len + ctx->tx_modulus + ctx->tx_remainder); in cdc_ncm_fill_tx_frame()
1314 cdc_ncm_align_tail(skb_out, ctx->tx_modulus, ctx->tx_remainder, ctx->tx_curr_size); in cdc_ncm_fill_tx_frame()
1318 skb_out->len + skb->len + delayed_ndp_size > ctx->tx_curr_size) { in cdc_ncm_fill_tx_frame()
1346 ndp.ndp16->dpe16[index].wDatagramIndex = cpu_to_le16(skb_out->len); in cdc_ncm_fill_tx_frame()
1353 ndp.ndp32->dpe32[index].dwDatagramIndex = cpu_to_le32(skb_out->len); in cdc_ncm_fill_tx_frame()
1356 skb_put_data(skb_out, skb->data, skb->len); in cdc_ncm_fill_tx_frame()
1381 ctx->tx_curr_skb = skb_out; in cdc_ncm_fill_tx_frame()
1387 ctx->tx_curr_skb = skb_out; in cdc_ncm_fill_tx_frame()
1403 nth.nth16 = (struct usb_cdc_ncm_nth16 *)skb_out->data; in cdc_ncm_fill_tx_frame()
1404 cdc_ncm_align_tail(skb_out, ctx->tx_ndp_modulus, 0, ctx->tx_curr_size - ctx->max_ndp_size); in cdc_ncm_fill_tx_frame()
1405 nth.nth16->wNdpIndex = cpu_to_le16(skb_out->len); in cdc_ncm_fill_tx_frame()
1406 skb_put_data(skb_out, ctx->delayed_ndp16, ctx->max_ndp_size); in cdc_ncm_fill_tx_frame()
1411 nth.nth32 = (struct usb_cdc_ncm_nth32 *)skb_out->data; in cdc_ncm_fill_tx_frame()
1412 cdc_ncm_align_tail(skb_out, ctx->tx_ndp_modulus, 0, ctx->tx_curr_size - ctx->max_ndp_size); in cdc_ncm_fill_tx_frame()
1413 nth.nth32->dwNdpIndex = cpu_to_le32(skb_out->len); in cdc_ncm_fill_tx_frame()
1414 skb_put_data(skb_out, ctx->delayed_ndp32, ctx->max_ndp_size); in cdc_ncm_fill_tx_frame()
1430 skb_out->len > ctx->min_tx_pkt) { in cdc_ncm_fill_tx_frame()
1431 padding_count = ctx->tx_curr_size - skb_out->len; in cdc_ncm_fill_tx_frame()
1433 skb_put_zero(skb_out, padding_count); in cdc_ncm_fill_tx_frame()
1434 } else if (skb_out->len < ctx->tx_curr_size && in cdc_ncm_fill_tx_frame()
1435 (skb_out->len % dev->maxpacket) == 0) { in cdc_ncm_fill_tx_frame()
1436 skb_put_u8(skb_out, 0); /* force short packet */ in cdc_ncm_fill_tx_frame()
1441 nth.nth16 = (struct usb_cdc_ncm_nth16 *)skb_out->data; in cdc_ncm_fill_tx_frame()
1442 nth.nth16->wBlockLength = cpu_to_le16(skb_out->len); in cdc_ncm_fill_tx_frame()
1444 nth.nth32 = (struct usb_cdc_ncm_nth32 *)skb_out->data; in cdc_ncm_fill_tx_frame()
1445 nth.nth32->dwBlockLength = cpu_to_le32(skb_out->len); in cdc_ncm_fill_tx_frame()
1452 ctx->tx_overhead += skb_out->len - ctx->tx_curr_frame_payload; in cdc_ncm_fill_tx_frame()
1459 usbnet_set_skb_tx_stats(skb_out, n, in cdc_ncm_fill_tx_frame()
1460 (long)ctx->tx_curr_frame_payload - skb_out->len); in cdc_ncm_fill_tx_frame()
1462 return skb_out; in cdc_ncm_fill_tx_frame()
1515 struct sk_buff *skb_out; in cdc_ncm_tx_fixup() local
1531 skb_out = cdc_ncm_fill_tx_frame(dev, skb, cpu_to_le32(USB_CDC_NCM_NDP16_NOCRC_SIGN)); in cdc_ncm_tx_fixup()
1533 skb_out = cdc_ncm_fill_tx_frame(dev, skb, cpu_to_le32(USB_CDC_NCM_NDP32_NOCRC_SIGN)); in cdc_ncm_tx_fixup()
1536 return skb_out; in cdc_ncm_tx_fixup()