task_data
struct ql4_task_data *task_data = task->dd_data;
task_data->iocb_req_cnt = 1;
if (!qla4xxx_space_in_req_ring(ha, task_data->iocb_req_cnt))
passthru_iocb->hdr.entryCount = task_data->iocb_req_cnt;
if (task_data->req_len) {
memcpy((uint8_t *)task_data->req_buffer +
cpu_to_le32(LSDW(task_data->req_dma));
cpu_to_le32(MSDW(task_data->req_dma));
if (task_data->resp_len) {
cpu_to_le32(LSDW(task_data->resp_dma));
cpu_to_le32(MSDW(task_data->resp_dma));
cpu_to_le32(task_data->resp_len);
ha->iocb_cnt += task_data->iocb_req_cnt;
ha->req_q_count -= task_data->iocb_req_cnt;
struct ql4_task_data *task_data;
task_data = task->dd_data;
memcpy(&task_data->sts, sts_entry, sizeof(struct passthru_status));
ha->iocb_cnt -= task_data->iocb_req_cnt;
queue_work(ha->task_wq, &task_data->task_work);
struct ql4_task_data *task_data;
task_data = container_of(wdata, struct ql4_task_data, task_work);
ha = task_data->ha;
task = task_data->task;
sts = &task_data->sts;
DEBUG3(qla4xxx_dump_buffer(task_data->resp_buffer, 64));
hdr = (struct iscsi_hdr *)task_data->resp_buffer;
data = task_data->resp_buffer + hdr_len;
data_len = task_data->resp_len - hdr_len;
struct ql4_task_data *task_data;
task_data = task->dd_data;
memset(task_data, 0, sizeof(struct ql4_task_data));
task_data->ha = ha;
task_data->task = task;
task_data->data_dma = dma_map_single(&ha->pdev->dev, task->data,
if (dma_mapping_error(&ha->pdev->dev, task_data->data_dma))
task_data->resp_len = task->conn->max_recv_dlength + hdr_len;
task_data->resp_buffer = dma_alloc_coherent(&ha->pdev->dev,
task_data->resp_len,
&task_data->resp_dma,
if (!task_data->resp_buffer)
task_data->req_len = task->data_count + hdr_len;
task_data->req_buffer = dma_alloc_coherent(&ha->pdev->dev,
task_data->req_len,
&task_data->req_dma,
if (!task_data->req_buffer)
task->hdr = task_data->req_buffer;
INIT_WORK(&task_data->task_work, qla4xxx_task_work);
if (task_data->resp_buffer)
dma_free_coherent(&ha->pdev->dev, task_data->resp_len,
task_data->resp_buffer, task_data->resp_dma);
if (task_data->req_buffer)
dma_free_coherent(&ha->pdev->dev, task_data->req_len,
task_data->req_buffer, task_data->req_dma);
struct ql4_task_data *task_data;
task_data = task->dd_data;
dma_unmap_single(&ha->pdev->dev, task_data->data_dma,
dma_free_coherent(&ha->pdev->dev, task_data->resp_len,
task_data->resp_buffer, task_data->resp_dma);
dma_free_coherent(&ha->pdev->dev, task_data->req_len,
task_data->req_buffer, task_data->req_dma);
desc->data = task_data;
_dsp_create_task_tree(chip,task_data,dest,size);
static void _dsp_create_task_tree (struct snd_cs46xx *chip, u32 * task_data,
spdst, task_data[i]);
writel(task_data[i],spdst);
cs46xx_dsp_create_task_tree (struct snd_cs46xx *chip, char * name, u32 * task_data,
struct task_data data;
bpf_map__set_max_entries(skel->maps.task_data, con->map_nr_entries);
bpf_map__set_max_entries(skel->maps.task_data, 1);
int task_fd = bpf_map__fd(skel->maps.task_data);
struct task_data data = {
__uint(value_size, sizeof(struct task_data));
p = bpf_map_lookup_elem(&task_data, &pid);
if (bpf_map_update_elem(&task_data, &pid, &data, BPF_NOEXIST) == -E2BIG)
} task_data SEC(".maps");