skb_data
unsigned char* skb_data;
skb_data = skb->data + 4; /* skip 4-byte cell header */
skb_data = skb->data;
if (((unsigned long)skb_data) & 0x3) {
memcpy(data, skb_data, skb_len);
data = skb_data;
u8 *skb_data = skb->data;
memmove(skb->data, skb_data, skb_len);
struct skb_data *entry;
entry = (struct skb_data *)skb->cb;
struct skb_data *entry;
entry = (struct skb_data *)tx_skb->cb;
struct skb_data *entry;
entry = (struct skb_data *)skb->cb;
unsigned char *skb_data, u16 headlen,
dma_addr = dma_map_single(sq->pdev, skb_data, headlen,
__be32 *skb_data;
skb_data = (__be32 *)skb->data;
memcpy(skb_data, &hdr, 4);
skb_data[i] = cpu_to_be32(readl(data + i * 4));
struct skb_data *entry;
entry = (struct skb_data *)skb->cb;
struct skb_data *entry = (struct skb_data *)skb->cb;
struct skb_data *entry = (struct skb_data *)skb->cb;
struct skb_data *entry = (struct skb_data *)newsk->cb;
memset(skb->cb, 0, sizeof(struct skb_data));
struct skb_data *entry = (struct skb_data *)skb->cb;
struct skb_data *entry = (struct skb_data *)skb->cb;
static struct skb_data *lan78xx_tx_buf_fill(struct lan78xx_net *dev,
struct skb_data *entry = (struct skb_data *)tx_buf->cb;
struct skb_data *entry;
struct skb_data *entry;
entry = (struct skb_data *)(rx_buf->cb);
struct skb_data *entry;
entry = (struct skb_data *)buf->cb;
struct skb_data *entry;
entry = (struct skb_data *)buf->cb;
struct skb_data *entry = (struct skb_data *) skb->cb;
struct skb_data *entry;
entry = (struct skb_data *) skb->cb;
struct skb_data *entry = (struct skb_data *)skb->cb;
struct skb_data *entry;
entry = (struct skb_data *) skb->cb;
sizeof_field(struct sk_buff, cb) < sizeof(struct skb_data));
memset(skb->cb, 0, sizeof(struct skb_data));
struct skb_data *entry = (struct skb_data *) newsk->cb;
struct skb_data *entry = (struct skb_data *) skb->cb;
struct skb_data *entry;
entry = (struct skb_data *) skb->cb;
struct skb_data *entry = (struct skb_data *) skb->cb;
struct skb_data *entry;
entry = (struct skb_data *) skb->cb;
struct sk_buff *skb_data;
skb_data = card->rx_buf_list[rd_index];
if (!skb_data)
mwifiex_unmap_pci_memory(adapter, skb_data, DMA_FROM_DEVICE);
rx_len = get_unaligned_le16(skb_data->data);
dev_kfree_skb_any(skb_data);
skb_put(skb_data, rx_len);
skb_pull(skb_data, adapter->intf_hdr_len);
skb_queue_tail(&adapter->rx_data_q, skb_data);
mwifiex_handle_rx_packet(adapter, skb_data);
struct rtw89_tx_skb_data *skb_data = RTW89_TX_SKB_CB(skb);
rcu_assign_pointer(skb_data->wait, wait);
struct rtw89_tx_skb_data *skb_data)
return rcu_access_pointer(skb_data->wait);
struct rtw89_tx_skb_data *skb_data,
wait = rcu_dereference(skb_data->wait);
struct rtw89_tx_skb_data *skb_data;
skb_data = RTW89_TX_SKB_CB(skb);
txcnt != skb_data->tx_pkt_cnt_lmt)
struct rtw89_tx_skb_data *skb_data = RTW89_TX_SKB_CB(skb);
return rtw89_core_is_tx_wait(rtwdev, skb_data) ||
struct rtw89_tx_skb_data *skb_data = RTW89_TX_SKB_CB(skb);
if (rtw89_core_tx_wait_complete(rtwdev, skb_data, tx_status))
struct rtw89_tx_skb_data *skb_data;
skb_data = RTW89_TX_SKB_CB(skb);
idx = skb_data->tx_rpt_sn;
struct rtw89_tx_skb_data *skb_data = RTW89_TX_SKB_CB(skb);
if (rtw89_core_tx_wait_complete(rtwdev, skb_data, tx_status))
struct rtw89_tx_skb_data *skb_data;
skb_data = RTW89_TX_SKB_CB(skb);
skb_data->tx_rpt_sn = tx_req->desc_info.sn;
skb_data->tx_pkt_cnt_lmt = tx_req->desc_info.tx_cnt_lmt;
struct skb_data *entry = (struct skb_data *) skb->cb;
__bpf_md_ptr(void *, skb_data);
case offsetof(struct bpf_sock_ops, skb_data):
case bpf_ctx_range_ptr(struct bpf_sock_ops, skb_data):
__bpf_md_ptr(void *, skb_data);
th = skops->skb_data;
th = skops->skb_data;
th = skops->skb_data;
th = skops->skb_data;
th = skops->skb_data;
narrow_load("sockops", bpf_sock_ops, skb_data);
: __imm_const(skb_data, offsetof(struct __sk_buff, data)), \
__imm_const(skb_data, offsetof(struct __sk_buff, data)),