list_size
end = p + list_size;
const unsigned long list_size,
(cid_list->list_size -
info->compatible_id_list.list_size = cid_list->list_size;
cid_list->list_size = cid_list_size;
int list_size,
for (; ind_offset < list_size; ind_offset++) {
int list_size;
list_size = adev->gfx.rlc.reg_list_size_bytes >> 2;
list_size = list_size >> 1;
WREG32(mmRLC_GPM_SCRATCH_DATA, list_size);
int list_size,
for (; indirect_offset < list_size; indirect_offset++) {
int list_size = 0;
list_size = adev->gfx.rlc.reg_list_size_bytes >> 2;
list_size = list_size >> 1;
WREG32(SOC15_REG_OFFSET(GC, 0, mmRLC_GPM_SCRATCH_DATA), list_size);
int list_size;
list_size = (list->next - list->buf.addr) >> 4;
list_size);
u32 list_size;
list_size(&dev->free_buff_list), \
list_size(&dev->rec_buff_list)); }
int list_size;
list_size = filter_list_len *
del_list = kzalloc(list_size, GFP_ATOMIC);
memset(del_list, 0, list_size);
list_size = filter_list_len *
add_list = kzalloc(list_size, GFP_ATOMIC);
memset(add_list, 0, list_size);
int list_size;
list_size = 0;
list_size++;
if (list_size > max_list_size) {
list_size, max_list_size);
list_size = max_list_size;
vlans = kvcalloc(list_size, sizeof(*vlans), GFP_KERNEL);
if (i >= list_size)
err = mlx5_modify_nic_vport_vlans(fs->mdev, vlans, list_size);
int list_size = 0;
list_size++;
list_size);
int list_size = 0;
list_size++;
if (list_size > max_list_size) {
list_size);
int list_size = 0;
list_size++;
list_size);
int list_size = 0;
list_size++;
if (list_size > max_list_size) {
list_size);
int *list_size)
req_list_size = *list_size;
*list_size = req_list_size;
int list_size)
if (list_size > max_list_size)
list_size * MLX5_ST_SZ_BYTES(mac_address_layout);
allowed_list_size, list_size);
for (i = 0; i < list_size; i++) {
int list_size)
if (list_size > max_list_size)
list_size * MLX5_ST_SZ_BYTES(vlan_layout);
allowed_list_size, list_size);
for (i = 0; i < list_size; i++) {
u32 list_size, new_size;
ret = of_property_read_u32(cur, cells_name, &list_size);
for (i = 0; i < list_size; i++)
while (map_len > (list_size + 1) && !match) {
for (i = 0; i < list_size; i++, map_len--)
if (i < list_size) {
out_args->args_count = list_size = new_size;
size_t list_size;
list_size = struct_size(list, cd00, count);
list_size = struct_size(list, cd01, count);
list = devm_kzalloc(&priv->wdev->dev, list_size, GFP_KERNEL);
bus_size_t list_size;
list_size = sizeof(struct ahd_dma_seg) * AHD_NSEG;
list_size = sizeof(struct ahd_dma64_seg) * AHD_NSEG;
return (list_size);
tempcnt = (u8)list_size(&dcb->srb_going_list);
if (list_size(&dcb->srb_going_list) > 1) {
seq_printf(m, "Nr of DCBs: %i\n", list_size(&acb->dcb_list));
list_size(&dcb->srb_waiting_list));
list_size(&dcb->srb_going_list));
pos->max_command <= list_size(&pos->srb_going_list)) {
if (dcb->max_command <= list_size(&dcb->srb_going_list) ||
u32 list_size;
a->last_read = a->list_size - 1;
| (a->list_size << MU_ILC_NUMBER_SHIFT);
dw |= MU_OLC_ENTRY_4_DW | (a->list_size << MU_OLC_NUMBER_SHIFT);
a->list_size = num_requests + ESAS2R_LIST_EXTRA;
a->inbound_list_md.size = a->list_size *
a->outbound_list_md.size = a->list_size *
for (i = 0; i < a->list_size; i++) {
if (unlikely(rspput_ptr >= a->list_size)) {
if (rspget_ptr >= a->list_size)
a->last_read = a->list_size - 1;
if (a->last_write >= a->list_size) {
int list_size;
dma_sync_single_for_cpu(kdev->dev, list_dma, info->list_size,
for (list = list_cpu; list < list_cpu + (info->list_size / sizeof(u32));
memset(list_cpu, 0, info->list_size);
dma_sync_single_for_device(kdev->dev, list_dma, info->list_size,
int list_size, mem_size;
list_size = info->list_entries;
list_size *= ACC_LIST_ENTRY_WORDS * sizeof(u32);
info->list_size = list_size;
mem_size = PAGE_ALIGN(list_size * 2);
acc->list_cpu[1] = list_mem + list_size;
acc->list_dma[1] = list_dma + list_size;
u32 list_size = isc->formats_list_size;
for (i = 0, j = 0; i < list_size; i++) {
unsigned long list_size = PAGE_SIZE / sizeof(struct page *);
sz2pin = min_t(unsigned long, npages, list_size);
for (name = buf; list_size; name += slen) {
slen = strnlen(name, list_size) + 1;
if (WARN_ON(slen > list_size)) {
list_size -= slen;
ssize_t list_size, size, value_size = 0;
list_size = vfs_listxattr(old, NULL, 0);
if (list_size <= 0) {
if (list_size == -EOPNOTSUPP)
return list_size;
buf = kvzalloc(list_size, GFP_KERNEL);
list_size = vfs_listxattr(old, buf, list_size);
if (list_size <= 0) {
error = list_size;
u32 list_size; /* Size of list, including ID strings */
int x509_load_certificate_list(const u8 cert_list[], const unsigned long list_size,
int *list_size);
int list_size);
int list_size);
int idx, list_size = 0;
if (list_size++)
static bool match_num_list(int num, int *list, int list_size)
for (int i = 0; i < list_size; ++i)
static bool match_str_list(const char *str, char **list, int list_size)
for (int i = 0; i < list_size; ++i)
if (list_size == max_size) {
list[list_size].pid = get_pid(buf);
list[list_size].tgid = get_tgid(buf);
list[list_size].comm = get_comm(buf);
list[list_size].txt = malloc(len+1);
if (!list[list_size].txt) {
memcpy(list[list_size].txt, buf, len);
len = remove_pattern(&ts_nsec_pattern, list[list_size].txt, len);
list[list_size].txt[len] = 0;
list[list_size].len = len;
list[list_size].num = 1;
list[list_size].page_num = get_page_num(buf);
list[list_size].stacktrace = strchr(list[list_size].txt, '\n') ?: "";
if (*list[list_size].stacktrace == '\n')
list[list_size].stacktrace++;
list[list_size].ts_nsec = get_ts_nsec(buf);
list[list_size].allocator = get_allocator(buf, ext_buf);
list_size++;
if (list_size % 1000 == 0) {
printf("loaded %d\r", list_size);
static int *parse_nums_list(char *arg_str, int *list_size)
*list_size = size;
printf("loaded %d\n", list_size);
qsort(list, list_size, sizeof(list[0]), compare_cull_condition);
for (i = count = 0; i < list_size; i++) {
static int list_size;