Symbol: part
arch/arm/include/asm/hardware/cache-l2x0.h
162
void l2x0_pmu_register(void __iomem *base, u32 part);
arch/arm/include/asm/hardware/cache-l2x0.h
166
static inline void l2x0_pmu_register(void __iomem *base, u32 part) {}
arch/arm/mach-omap2/cm.h
56
int (*wait_module_ready)(u8 part, s16 prcm_mod, u16 idlest_reg,
arch/arm/mach-omap2/cm.h
58
int (*wait_module_idle)(u8 part, s16 prcm_mod, u16 idlest_reg,
arch/arm/mach-omap2/cm.h
60
void (*module_enable)(u8 mode, u8 part, u16 inst, u16 clkctrl_offs);
arch/arm/mach-omap2/cm.h
61
void (*module_disable)(u8 part, u16 inst, u16 clkctrl_offs);
arch/arm/mach-omap2/cm.h
62
u32 (*xlate_clkctrl)(u8 part, u16 inst, u16 clkctrl_offs);
arch/arm/mach-omap2/cm.h
67
int omap_cm_wait_module_ready(u8 part, s16 prcm_mod, u16 idlest_reg,
arch/arm/mach-omap2/cm.h
69
int omap_cm_wait_module_idle(u8 part, s16 prcm_mod, u16 idlest_reg,
arch/arm/mach-omap2/cm.h
71
int omap_cm_module_enable(u8 mode, u8 part, u16 inst, u16 clkctrl_offs);
arch/arm/mach-omap2/cm.h
72
int omap_cm_module_disable(u8 part, u16 inst, u16 clkctrl_offs);
arch/arm/mach-omap2/cm.h
73
u32 omap_cm_xlate_clkctrl(u8 part, u16 inst, u16 clkctrl_offs);
arch/arm/mach-omap2/cm2xxx.c
148
static int omap2xxx_cm_wait_module_ready(u8 part, s16 prcm_mod, u16 idlest_id,
arch/arm/mach-omap2/cm33xx.c
224
static int am33xx_cm_wait_module_ready(u8 part, s16 inst, u16 clkctrl_offs,
arch/arm/mach-omap2/cm33xx.c
247
static int am33xx_cm_wait_module_idle(u8 part, s16 inst, u16 clkctrl_offs,
arch/arm/mach-omap2/cm33xx.c
268
static void am33xx_cm_module_enable(u8 mode, u8 part, u16 inst,
arch/arm/mach-omap2/cm33xx.c
287
static void am33xx_cm_module_disable(u8 part, u16 inst, u16 clkctrl_offs)
arch/arm/mach-omap2/cm33xx.c
351
static u32 am33xx_cm_xlate_clkctrl(u8 part, u16 inst, u16 offset)
arch/arm/mach-omap2/cm3xxx.c
88
static int omap3xxx_cm_wait_module_ready(u8 part, s16 prcm_mod, u16 idlest_id,
arch/arm/mach-omap2/cm_common.c
107
int omap_cm_wait_module_idle(u8 part, s16 prcm_mod, u16 idlest_reg,
arch/arm/mach-omap2/cm_common.c
116
return cm_ll_data->wait_module_idle(part, prcm_mod, idlest_reg,
arch/arm/mach-omap2/cm_common.c
131
int omap_cm_module_enable(u8 mode, u8 part, u16 inst, u16 clkctrl_offs)
arch/arm/mach-omap2/cm_common.c
139
cm_ll_data->module_enable(mode, part, inst, clkctrl_offs);
arch/arm/mach-omap2/cm_common.c
153
int omap_cm_module_disable(u8 part, u16 inst, u16 clkctrl_offs)
arch/arm/mach-omap2/cm_common.c
161
cm_ll_data->module_disable(part, inst, clkctrl_offs);
arch/arm/mach-omap2/cm_common.c
165
u32 omap_cm_xlate_clkctrl(u8 part, u16 inst, u16 clkctrl_offs)
arch/arm/mach-omap2/cm_common.c
172
return cm_ll_data->xlate_clkctrl(part, inst, clkctrl_offs);
arch/arm/mach-omap2/cm_common.c
81
int omap_cm_wait_module_ready(u8 part, s16 prcm_mod, u16 idlest_reg,
arch/arm/mach-omap2/cm_common.c
90
return cm_ll_data->wait_module_ready(part, prcm_mod, idlest_reg,
arch/arm/mach-omap2/cminst44xx.c
102
static bool _is_module_ready(u8 part, u16 inst, u16 clkctrl_offs)
arch/arm/mach-omap2/cminst44xx.c
106
v = _clkctrl_idlest(part, inst, clkctrl_offs);
arch/arm/mach-omap2/cminst44xx.c
113
static u32 omap4_cminst_read_inst_reg(u8 part, u16 inst, u16 idx)
arch/arm/mach-omap2/cminst44xx.c
115
BUG_ON(part >= OMAP4_MAX_PRCM_PARTITIONS ||
arch/arm/mach-omap2/cminst44xx.c
116
part == OMAP4430_INVALID_PRCM_PARTITION ||
arch/arm/mach-omap2/cminst44xx.c
117
!_cm_bases[part].va);
arch/arm/mach-omap2/cminst44xx.c
118
return readl_relaxed(_cm_bases[part].va + inst + idx);
arch/arm/mach-omap2/cminst44xx.c
122
static void omap4_cminst_write_inst_reg(u32 val, u8 part, u16 inst, u16 idx)
arch/arm/mach-omap2/cminst44xx.c
124
BUG_ON(part >= OMAP4_MAX_PRCM_PARTITIONS ||
arch/arm/mach-omap2/cminst44xx.c
125
part == OMAP4430_INVALID_PRCM_PARTITION ||
arch/arm/mach-omap2/cminst44xx.c
126
!_cm_bases[part].va);
arch/arm/mach-omap2/cminst44xx.c
127
writel_relaxed(val, _cm_bases[part].va + inst + idx);
arch/arm/mach-omap2/cminst44xx.c
131
static u32 omap4_cminst_rmw_inst_reg_bits(u32 mask, u32 bits, u8 part, u16 inst,
arch/arm/mach-omap2/cminst44xx.c
136
v = omap4_cminst_read_inst_reg(part, inst, idx);
arch/arm/mach-omap2/cminst44xx.c
139
omap4_cminst_write_inst_reg(v, part, inst, idx);
arch/arm/mach-omap2/cminst44xx.c
144
static u32 omap4_cminst_set_inst_reg_bits(u32 bits, u8 part, u16 inst, s16 idx)
arch/arm/mach-omap2/cminst44xx.c
146
return omap4_cminst_rmw_inst_reg_bits(bits, bits, part, inst, idx);
arch/arm/mach-omap2/cminst44xx.c
149
static u32 omap4_cminst_clear_inst_reg_bits(u32 bits, u8 part, u16 inst,
arch/arm/mach-omap2/cminst44xx.c
152
return omap4_cminst_rmw_inst_reg_bits(bits, 0x0, part, inst, idx);
arch/arm/mach-omap2/cminst44xx.c
155
static u32 omap4_cminst_read_inst_reg_bits(u8 part, u16 inst, s16 idx, u32 mask)
arch/arm/mach-omap2/cminst44xx.c
159
v = omap4_cminst_read_inst_reg(part, inst, idx);
arch/arm/mach-omap2/cminst44xx.c
180
static void _clktrctrl_write(u8 c, u8 part, u16 inst, u16 cdoffs)
arch/arm/mach-omap2/cminst44xx.c
184
v = omap4_cminst_read_inst_reg(part, inst, cdoffs + OMAP4_CM_CLKSTCTRL);
arch/arm/mach-omap2/cminst44xx.c
187
omap4_cminst_write_inst_reg(v, part, inst, cdoffs + OMAP4_CM_CLKSTCTRL);
arch/arm/mach-omap2/cminst44xx.c
199
static bool omap4_cminst_is_clkdm_in_hwsup(u8 part, u16 inst, u16 cdoffs)
arch/arm/mach-omap2/cminst44xx.c
203
v = omap4_cminst_read_inst_reg(part, inst, cdoffs + OMAP4_CM_CLKSTCTRL);
arch/arm/mach-omap2/cminst44xx.c
219
static void omap4_cminst_clkdm_enable_hwsup(u8 part, u16 inst, u16 cdoffs)
arch/arm/mach-omap2/cminst44xx.c
221
_clktrctrl_write(OMAP34XX_CLKSTCTRL_ENABLE_AUTO, part, inst, cdoffs);
arch/arm/mach-omap2/cminst44xx.c
234
static void omap4_cminst_clkdm_disable_hwsup(u8 part, u16 inst, u16 cdoffs)
arch/arm/mach-omap2/cminst44xx.c
236
_clktrctrl_write(OMAP34XX_CLKSTCTRL_DISABLE_AUTO, part, inst, cdoffs);
arch/arm/mach-omap2/cminst44xx.c
248
static void omap4_cminst_clkdm_force_wakeup(u8 part, u16 inst, u16 cdoffs)
arch/arm/mach-omap2/cminst44xx.c
250
_clktrctrl_write(OMAP34XX_CLKSTCTRL_FORCE_WAKEUP, part, inst, cdoffs);
arch/arm/mach-omap2/cminst44xx.c
257
static void omap4_cminst_clkdm_force_sleep(u8 part, u16 inst, u16 cdoffs)
arch/arm/mach-omap2/cminst44xx.c
259
_clktrctrl_write(OMAP34XX_CLKSTCTRL_FORCE_SLEEP, part, inst, cdoffs);
arch/arm/mach-omap2/cminst44xx.c
274
static int omap4_cminst_wait_module_ready(u8 part, s16 inst, u16 clkctrl_offs,
arch/arm/mach-omap2/cminst44xx.c
279
omap_test_timeout(_is_module_ready(part, inst, clkctrl_offs),
arch/arm/mach-omap2/cminst44xx.c
297
static int omap4_cminst_wait_module_idle(u8 part, s16 inst, u16 clkctrl_offs,
arch/arm/mach-omap2/cminst44xx.c
302
omap_test_timeout((_clkctrl_idlest(part, inst, clkctrl_offs) ==
arch/arm/mach-omap2/cminst44xx.c
318
static void omap4_cminst_module_enable(u8 mode, u8 part, u16 inst,
arch/arm/mach-omap2/cminst44xx.c
323
v = omap4_cminst_read_inst_reg(part, inst, clkctrl_offs);
arch/arm/mach-omap2/cminst44xx.c
326
omap4_cminst_write_inst_reg(v, part, inst, clkctrl_offs);
arch/arm/mach-omap2/cminst44xx.c
337
static void omap4_cminst_module_disable(u8 part, u16 inst, u16 clkctrl_offs)
arch/arm/mach-omap2/cminst44xx.c
341
v = omap4_cminst_read_inst_reg(part, inst, clkctrl_offs);
arch/arm/mach-omap2/cminst44xx.c
343
omap4_cminst_write_inst_reg(v, part, inst, clkctrl_offs);
arch/arm/mach-omap2/cminst44xx.c
476
static u32 omap4_cminst_xlate_clkctrl(u8 part, u16 inst, u16 offset)
arch/arm/mach-omap2/cminst44xx.c
478
return _cm_bases[part].pa + inst + offset;
arch/arm/mach-omap2/cminst44xx.c
74
static u32 omap4_cminst_read_inst_reg(u8 part, u16 inst, u16 idx);
arch/arm/mach-omap2/cminst44xx.c
85
static u32 _clkctrl_idlest(u8 part, u16 inst, u16 clkctrl_offs)
arch/arm/mach-omap2/cminst44xx.c
87
u32 v = omap4_cminst_read_inst_reg(part, inst, clkctrl_offs);
arch/arm/mach-omap2/prm.h
137
bool (*was_any_context_lost_old)(u8 part, s16 inst, u16 idx);
arch/arm/mach-omap2/prm.h
138
void (*clear_context_loss_flags_old)(u8 part, s16 inst, u16 idx);
arch/arm/mach-omap2/prm.h
140
int (*assert_hardreset)(u8 shift, u8 part, s16 prm_mod, u16 offset);
arch/arm/mach-omap2/prm.h
141
int (*deassert_hardreset)(u8 shift, u8 st_shift, u8 part, s16 prm_mod,
arch/arm/mach-omap2/prm.h
143
int (*is_hardreset_asserted)(u8 shift, u8 part, s16 prm_mod,
arch/arm/mach-omap2/prm.h
154
int omap_prm_assert_hardreset(u8 shift, u8 part, s16 prm_mod, u16 offset);
arch/arm/mach-omap2/prm.h
155
int omap_prm_deassert_hardreset(u8 shift, u8 st_shift, u8 part, s16 prm_mod,
arch/arm/mach-omap2/prm.h
157
int omap_prm_is_hardreset_asserted(u8 shift, u8 part, s16 prm_mod, u16 offset);
arch/arm/mach-omap2/prm.h
158
extern bool prm_was_any_context_lost_old(u8 part, s16 inst, u16 idx);
arch/arm/mach-omap2/prm.h
159
extern void prm_clear_context_loss_flags_old(u8 part, s16 inst, u16 idx);
arch/arm/mach-omap2/prm2xxx_3xxx.c
33
int omap2_prm_is_hardreset_asserted(u8 shift, u8 part, s16 prm_mod, u16 offset)
arch/arm/mach-omap2/prm2xxx_3xxx.c
53
int omap2_prm_assert_hardreset(u8 shift, u8 part, s16 prm_mod, u16 offset)
arch/arm/mach-omap2/prm2xxx_3xxx.c
82
int omap2_prm_deassert_hardreset(u8 rst_shift, u8 st_shift, u8 part,
arch/arm/mach-omap2/prm2xxx_3xxx.h
100
int omap2_prm_is_hardreset_asserted(u8 shift, u8 part, s16 prm_mod, u16 offset);
arch/arm/mach-omap2/prm2xxx_3xxx.h
101
int omap2_prm_assert_hardreset(u8 shift, u8 part, s16 prm_mod,
arch/arm/mach-omap2/prm2xxx_3xxx.h
103
int omap2_prm_deassert_hardreset(u8 rst_shift, u8 st_shift, u8 part,
arch/arm/mach-omap2/prm33xx.c
111
static int am33xx_prm_deassert_hardreset(u8 shift, u8 st_shift, u8 part,
arch/arm/mach-omap2/prm33xx.c
56
static int am33xx_prm_is_hardreset_asserted(u8 shift, u8 part, s16 inst,
arch/arm/mach-omap2/prm33xx.c
82
static int am33xx_prm_assert_hardreset(u8 shift, u8 part, s16 inst,
arch/arm/mach-omap2/prm44xx.c
404
static bool omap44xx_prm_was_any_context_lost_old(u8 part, s16 inst, u16 idx)
arch/arm/mach-omap2/prm44xx.c
406
return (omap4_prminst_read_inst_reg(part, inst, idx)) ? 1 : 0;
arch/arm/mach-omap2/prm44xx.c
419
static void omap44xx_prm_clear_context_loss_flags_old(u8 part, s16 inst,
arch/arm/mach-omap2/prm44xx.c
422
omap4_prminst_write_inst_reg(0xffffffff, part, inst, idx);
arch/arm/mach-omap2/prm_common.c
365
bool prm_was_any_context_lost_old(u8 part, s16 inst, u16 idx)
arch/arm/mach-omap2/prm_common.c
370
ret = prm_ll_data->was_any_context_lost_old(part, inst, idx);
arch/arm/mach-omap2/prm_common.c
389
void prm_clear_context_loss_flags_old(u8 part, s16 inst, u16 idx)
arch/arm/mach-omap2/prm_common.c
392
prm_ll_data->clear_context_loss_flags_old(part, inst, idx);
arch/arm/mach-omap2/prm_common.c
407
int omap_prm_assert_hardreset(u8 shift, u8 part, s16 prm_mod, u16 offset)
arch/arm/mach-omap2/prm_common.c
415
return prm_ll_data->assert_hardreset(shift, part, prm_mod, offset);
arch/arm/mach-omap2/prm_common.c
429
int omap_prm_deassert_hardreset(u8 shift, u8 st_shift, u8 part, s16 prm_mod,
arch/arm/mach-omap2/prm_common.c
438
return prm_ll_data->deassert_hardreset(shift, st_shift, part, prm_mod,
arch/arm/mach-omap2/prm_common.c
451
int omap_prm_is_hardreset_asserted(u8 shift, u8 part, s16 prm_mod, u16 offset)
arch/arm/mach-omap2/prm_common.c
459
return prm_ll_data->is_hardreset_asserted(shift, part, prm_mod, offset);
arch/arm/mach-omap2/prminst44xx.c
104
v = omap4_prminst_read_inst_reg(part, inst, rstctrl_offs);
arch/arm/mach-omap2/prminst44xx.c
123
int omap4_prminst_assert_hardreset(u8 shift, u8 part, s16 inst,
arch/arm/mach-omap2/prminst44xx.c
128
omap4_prminst_rmw_inst_reg_bits(mask, mask, part, inst, rstctrl_offs);
arch/arm/mach-omap2/prminst44xx.c
152
int omap4_prminst_deassert_hardreset(u8 shift, u8 st_shift, u8 part, s16 inst,
arch/arm/mach-omap2/prminst44xx.c
160
if (omap4_prminst_is_hardreset_asserted(shift, part, inst,
arch/arm/mach-omap2/prminst44xx.c
165
omap4_prminst_rmw_inst_reg_bits(0xffffffff, st_mask, part, inst,
arch/arm/mach-omap2/prminst44xx.c
168
omap4_prminst_rmw_inst_reg_bits(mask, 0, part, inst, rstctrl_offs);
arch/arm/mach-omap2/prminst44xx.c
170
omap_test_timeout(omap4_prminst_is_hardreset_asserted(st_shift, part,
arch/arm/mach-omap2/prminst44xx.c
58
u32 omap4_prminst_read_inst_reg(u8 part, s16 inst, u16 idx)
arch/arm/mach-omap2/prminst44xx.c
60
BUG_ON(part >= OMAP4_MAX_PRCM_PARTITIONS ||
arch/arm/mach-omap2/prminst44xx.c
61
part == OMAP4430_INVALID_PRCM_PARTITION ||
arch/arm/mach-omap2/prminst44xx.c
62
!_prm_bases[part].va);
arch/arm/mach-omap2/prminst44xx.c
63
return readl_relaxed(_prm_bases[part].va + inst + idx);
arch/arm/mach-omap2/prminst44xx.c
67
void omap4_prminst_write_inst_reg(u32 val, u8 part, s16 inst, u16 idx)
arch/arm/mach-omap2/prminst44xx.c
69
BUG_ON(part >= OMAP4_MAX_PRCM_PARTITIONS ||
arch/arm/mach-omap2/prminst44xx.c
70
part == OMAP4430_INVALID_PRCM_PARTITION ||
arch/arm/mach-omap2/prminst44xx.c
71
!_prm_bases[part].va);
arch/arm/mach-omap2/prminst44xx.c
72
writel_relaxed(val, _prm_bases[part].va + inst + idx);
arch/arm/mach-omap2/prminst44xx.c
76
u32 omap4_prminst_rmw_inst_reg_bits(u32 mask, u32 bits, u8 part, s16 inst,
arch/arm/mach-omap2/prminst44xx.c
81
v = omap4_prminst_read_inst_reg(part, inst, idx);
arch/arm/mach-omap2/prminst44xx.c
84
omap4_prminst_write_inst_reg(v, part, inst, idx);
arch/arm/mach-omap2/prminst44xx.c
99
int omap4_prminst_is_hardreset_asserted(u8 shift, u8 part, s16 inst,
arch/arm/mach-omap2/prminst44xx.h
20
extern u32 omap4_prminst_read_inst_reg(u8 part, s16 inst, u16 idx);
arch/arm/mach-omap2/prminst44xx.h
21
extern void omap4_prminst_write_inst_reg(u32 val, u8 part, s16 inst, u16 idx);
arch/arm/mach-omap2/prminst44xx.h
22
extern u32 omap4_prminst_rmw_inst_reg_bits(u32 mask, u32 bits, u8 part,
arch/arm/mach-omap2/prminst44xx.h
27
extern int omap4_prminst_is_hardreset_asserted(u8 shift, u8 part, s16 inst,
arch/arm/mach-omap2/prminst44xx.h
29
extern int omap4_prminst_assert_hardreset(u8 shift, u8 part, s16 inst,
arch/arm/mach-omap2/prminst44xx.h
31
int omap4_prminst_deassert_hardreset(u8 shift, u8 st_shift, u8 part,
arch/arm/mm/cache-l2x0-pmu.c
474
void __init l2x0_pmu_register(void __iomem *base, u32 part)
arch/arm/mm/cache-l2x0-pmu.c
489
switch (part & L2X0_CACHE_ID_PART_MASK) {
arch/mips/include/asm/xtalk/xwidget.h
118
u32 part;
arch/powerpc/include/asm/nvram.h
78
extern int __init nvram_init_os_partition(struct nvram_os_partition *part);
arch/powerpc/include/asm/nvram.h
84
extern int nvram_read_partition(struct nvram_os_partition *part, char *buff,
arch/powerpc/include/asm/nvram.h
89
extern int nvram_write_os_partition(struct nvram_os_partition *part,
arch/powerpc/kernel/nvram_64.c
180
int nvram_write_os_partition(struct nvram_os_partition *part,
arch/powerpc/kernel/nvram_64.c
189
if (part->index == -1)
arch/powerpc/kernel/nvram_64.c
192
if (length > part->size)
arch/powerpc/kernel/nvram_64.c
193
length = part->size;
arch/powerpc/kernel/nvram_64.c
198
tmp_index = part->index;
arch/powerpc/kernel/nvram_64.c
219
int nvram_read_partition(struct nvram_os_partition *part, char *buff,
arch/powerpc/kernel/nvram_64.c
227
if (part->index == -1)
arch/powerpc/kernel/nvram_64.c
230
if (length > part->size)
arch/powerpc/kernel/nvram_64.c
231
length = part->size;
arch/powerpc/kernel/nvram_64.c
233
tmp_index = part->index;
arch/powerpc/kernel/nvram_64.c
235
if (part->os_partition) {
arch/powerpc/kernel/nvram_64.c
249
if (part->os_partition) {
arch/powerpc/kernel/nvram_64.c
274
int __init nvram_init_os_partition(struct nvram_os_partition *part)
arch/powerpc/kernel/nvram_64.c
280
p = nvram_find_partition(part->name, NVRAM_SIG_OS, &size);
arch/powerpc/kernel/nvram_64.c
283
if (p && size < part->min_size) {
arch/powerpc/kernel/nvram_64.c
285
" removing it...\n", part->name);
arch/powerpc/kernel/nvram_64.c
286
nvram_remove_partition(part->name, NVRAM_SIG_OS, NULL);
arch/powerpc/kernel/nvram_64.c
292
p = nvram_create_partition(part->name, NVRAM_SIG_OS,
arch/powerpc/kernel/nvram_64.c
293
part->req_size, part->min_size);
arch/powerpc/kernel/nvram_64.c
297
part->name);
arch/powerpc/kernel/nvram_64.c
300
p = nvram_create_partition(part->name, NVRAM_SIG_OS,
arch/powerpc/kernel/nvram_64.c
301
part->req_size, part->min_size);
arch/powerpc/kernel/nvram_64.c
307
" partition, err %d\n", part->name, (int)p);
arch/powerpc/kernel/nvram_64.c
311
part->index = p;
arch/powerpc/kernel/nvram_64.c
312
part->size = nvram_get_partition_size(p) - sizeof(struct err_log_info);
arch/powerpc/kernel/nvram_64.c
392
if (record->part > 1 || (record->type != PSTORE_TYPE_DMESG))
arch/powerpc/kernel/nvram_64.c
412
record->id = record->part;
arch/powerpc/kernel/nvram_64.c
425
struct nvram_os_partition *part = NULL;
arch/powerpc/kernel/nvram_64.c
434
part = &oops_log_partition;
arch/powerpc/kernel/nvram_64.c
439
part = &common_partition;
arch/powerpc/kernel/nvram_64.c
447
part = &rtas_log_partition;
arch/powerpc/kernel/nvram_64.c
454
part = &of_config_partition;
arch/powerpc/kernel/nvram_64.c
464
part = &skiboot_partition;
arch/powerpc/kernel/nvram_64.c
475
if (!part->os_partition) {
arch/powerpc/kernel/nvram_64.c
476
p = nvram_find_partition(part->name, sig, &size);
arch/powerpc/kernel/nvram_64.c
479
"err %d\n", part->name, (int)p);
arch/powerpc/kernel/nvram_64.c
482
part->index = p;
arch/powerpc/kernel/nvram_64.c
483
part->size = size;
arch/powerpc/kernel/nvram_64.c
486
buff = kmalloc(part->size, GFP_KERNEL);
arch/powerpc/kernel/nvram_64.c
491
if (nvram_read_partition(part, buff, part->size, &err_type, &id_no)) {
arch/powerpc/kernel/nvram_64.c
498
if (part->os_partition)
arch/powerpc/kernel/nvram_64.c
531
return part->size;
arch/powerpc/kernel/nvram_64.c
724
static int __init nvram_write_header(struct nvram_partition * part)
arch/powerpc/kernel/nvram_64.c
730
memcpy(&phead, &part->header, NVRAM_HEADER_LEN);
arch/powerpc/kernel/nvram_64.c
733
tmp_index = part->index;
arch/powerpc/kernel/nvram_64.c
758
static int __init nvram_can_remove_partition(struct nvram_partition *part,
arch/powerpc/kernel/nvram_64.c
761
if (part->header.signature != sig)
arch/powerpc/kernel/nvram_64.c
764
if (strncmp(name, part->header.name, 12))
arch/powerpc/kernel/nvram_64.c
769
if (!strncmp(*except, part->header.name, 12))
arch/powerpc/kernel/nvram_64.c
788
struct nvram_partition *part, *prev, *tmp;
arch/powerpc/kernel/nvram_64.c
791
list_for_each_entry(part, &nvram_partitions, partition) {
arch/powerpc/kernel/nvram_64.c
792
if (!nvram_can_remove_partition(part, name, sig, exceptions))
arch/powerpc/kernel/nvram_64.c
796
part->header.signature = NVRAM_SIG_FREE;
arch/powerpc/kernel/nvram_64.c
797
memset(part->header.name, 'w', 12);
arch/powerpc/kernel/nvram_64.c
798
part->header.checksum = nvram_checksum(&part->header);
arch/powerpc/kernel/nvram_64.c
799
rc = nvram_write_header(part);
arch/powerpc/kernel/nvram_64.c
808
list_for_each_entry_safe(part, tmp, &nvram_partitions, partition) {
arch/powerpc/kernel/nvram_64.c
809
if (part->header.signature != NVRAM_SIG_FREE) {
arch/powerpc/kernel/nvram_64.c
814
prev->header.length += part->header.length;
arch/powerpc/kernel/nvram_64.c
821
list_del(&part->partition);
arch/powerpc/kernel/nvram_64.c
822
kfree(part);
arch/powerpc/kernel/nvram_64.c
824
prev = part;
arch/powerpc/kernel/nvram_64.c
846
struct nvram_partition *part;
arch/powerpc/kernel/nvram_64.c
874
list_for_each_entry(part, &nvram_partitions, partition) {
arch/powerpc/kernel/nvram_64.c
875
if (part->header.signature != NVRAM_SIG_FREE)
arch/powerpc/kernel/nvram_64.c
878
if (part->header.length >= req_size) {
arch/powerpc/kernel/nvram_64.c
880
free_part = part;
arch/powerpc/kernel/nvram_64.c
883
if (part->header.length > size &&
arch/powerpc/kernel/nvram_64.c
884
part->header.length >= min_size) {
arch/powerpc/kernel/nvram_64.c
885
size = part->header.length;
arch/powerpc/kernel/nvram_64.c
886
free_part = part;
arch/powerpc/kernel/nvram_64.c
952
struct nvram_partition *part;
arch/powerpc/kernel/nvram_64.c
954
list_for_each_entry(part, &nvram_partitions, partition) {
arch/powerpc/kernel/nvram_64.c
955
if (part->index + NVRAM_HEADER_LEN == data_index)
arch/powerpc/kernel/nvram_64.c
956
return (part->header.length - 1) * NVRAM_BLOCK_LEN;
arch/powerpc/kernel/ptrace/ptrace32.c
117
u32 part;
arch/powerpc/kernel/ptrace/ptrace32.c
125
part = 1; /* want the 2nd half of the register (right-most). */
arch/powerpc/kernel/ptrace/ptrace32.c
127
part = 0; /* want the 1st half of the register (left-most). */
arch/powerpc/kernel/ptrace/ptrace32.c
146
reg32bits = ((u32*)&tmp)[part];
arch/powerpc/platforms/powermac/setup.c
323
void note_bootable_part(dev_t dev, int part, int goodness);
arch/powerpc/platforms/powermac/setup.c
330
void __ref note_bootable_part(dev_t dev, int part, int goodness)
arch/powerpc/platforms/powermac/setup.c
343
ROOT_DEV = dev + part;
arch/x86/platform/uv/bios_uv.c
101
*region = part.region_size;
arch/x86/platform/uv/bios_uv.c
86
union partition_info_u part;
arch/x86/platform/uv/bios_uv.c
93
part.val = v0;
arch/x86/platform/uv/bios_uv.c
95
*uvtype = part.hub_version;
arch/x86/platform/uv/bios_uv.c
97
*partid = part.partition_id;
arch/x86/platform/uv/bios_uv.c
99
*coher = part.coherence_id;
block/bdev.c
773
static int blkdev_get_part(struct block_device *part, blk_mode_t mode)
block/bdev.c
775
struct gendisk *disk = part->bd_disk;
block/bdev.c
778
ret = blkdev_get_whole(bdev_whole(part), mode);
block/bdev.c
783
if (!bdev_nr_sectors(part))
block/bdev.c
786
if (!atomic_read(&part->bd_openers)) {
block/bdev.c
788
set_init_blocksize(part);
block/bdev.c
790
atomic_inc(&part->bd_openers);
block/bdev.c
794
blkdev_put_whole(bdev_whole(part));
block/bdev.c
823
static void blkdev_put_part(struct block_device *part)
block/bdev.c
825
struct block_device *whole = bdev_whole(part);
block/bdev.c
827
if (atomic_dec_and_test(&part->bd_openers)) {
block/bdev.c
828
blkdev_flush_mapping(part);
block/blk-core.c
1024
void update_io_ticks(struct block_device *part, unsigned long now, bool end)
block/blk-core.c
1028
stamp = READ_ONCE(part->bd_stamp);
block/blk-core.c
1030
likely(try_cmpxchg(&part->bd_stamp, &stamp, now)) &&
block/blk-core.c
1031
(end || bdev_count_inflight(part)))
block/blk-core.c
1032
__part_stat_add(part, io_ticks, now - stamp);
block/blk-core.c
1034
if (bdev_is_partition(part)) {
block/blk-core.c
1035
part = bdev_whole(part);
block/blk-core.c
506
bool should_fail_request(struct block_device *part, unsigned int bytes)
block/blk-core.c
508
return bdev_test_flag(part, BD_MAKE_IT_FAIL) &&
block/blk-flush.c
126
struct block_device *part = rq->q->disk->part0;
block/blk-flush.c
129
part_stat_inc(part, ios[STAT_FLUSH]);
block/blk-flush.c
130
part_stat_add(part, nsecs[STAT_FLUSH],
block/blk-merge.c
725
part_stat_inc(req->part, merges[op_stat_group(req_op(req))]);
block/blk-merge.c
726
part_stat_local_dec(req->part,
block/blk-merge.c
939
part_stat_inc(req->part, merges[op_stat_group(req_op(req))]);
block/blk-mq.c
105
void blk_mq_in_driver_rw(struct block_device *part, unsigned int inflight[2])
block/blk-mq.c
107
struct mq_inflight mi = { .part = part };
block/blk-mq.c
1082
update_io_ticks(req->part, jiffies, true);
block/blk-mq.c
1083
part_stat_inc(req->part, ios[sgrp]);
block/blk-mq.c
1084
part_stat_add(req->part, nsecs[sgrp], now - req->start_time_ns);
block/blk-mq.c
1085
part_stat_local_dec(req->part,
block/blk-mq.c
109
blk_mq_queue_tag_busy_iter(bdev_get_queue(part), blk_mq_check_in_driver,
block/blk-mq.c
1140
req->part = req->bio->bi_bdev;
block/blk-mq.c
1142
req->part = req->q->disk->part0;
block/blk-mq.c
1145
update_io_ticks(req->part, jiffies, false);
block/blk-mq.c
1146
part_stat_local_inc(req->part, in_flight[op_is_write(req_op(req))]);
block/blk-mq.c
436
rq->part = NULL;
block/blk-mq.c
862
part_stat_add(req->part, sectors[sgrp], bytes >> 9);
block/blk-mq.c
89
struct block_device *part;
block/blk-mq.c
98
(!bdev_is_partition(mi->part) || rq->part == mi->part) &&
block/blk-mq.h
262
void blk_mq_in_driver_rw(struct block_device *part, unsigned int inflight[2]);
block/blk.h
463
void update_io_ticks(struct block_device *part, unsigned long now, bool end);
block/blk.h
596
void drop_partition(struct block_device *part);
block/blk.h
633
bool should_fail_request(struct block_device *part, unsigned int bytes);
block/blk.h
635
static inline bool should_fail_request(struct block_device *part,
block/early-lookup.c
155
int part;
block/early-lookup.c
181
part = simple_strtoul(p, NULL, 10);
block/early-lookup.c
183
*devt = blk_lookup_devt(s, part);
block/early-lookup.c
191
*devt = blk_lookup_devt(s, part);
block/early-lookup.c
280
struct block_device *part;
block/early-lookup.c
296
xa_for_each(&disk->part_tbl, idx, part) {
block/early-lookup.c
297
if (!bdev_nr_sectors(part))
block/early-lookup.c
300
bdev_is_partition(part) ? " " : "",
block/early-lookup.c
301
bdevt_str(part->bd_dev, devt_buf),
block/early-lookup.c
302
bdev_nr_sectors(part) >> 1, part,
block/early-lookup.c
303
part->bd_meta_info ?
block/early-lookup.c
304
part->bd_meta_info->uuid : "");
block/early-lookup.c
305
if (bdev_is_partition(part))
block/genhd.c
107
static void part_stat_read_all(struct block_device *part,
block/genhd.c
114
struct disk_stats *ptr = per_cpu_ptr(part->bd_stats, cpu);
block/genhd.c
128
static void bdev_count_inflight_rw(struct block_device *part,
block/genhd.c
136
blk_mq_in_driver_rw(part, inflight);
block/genhd.c
141
read += part_stat_local_read_cpu(part, in_flight[READ], cpu);
block/genhd.c
142
write += part_stat_local_read_cpu(part, in_flight[WRITE], cpu);
block/genhd.c
1434
struct block_device *part;
block/genhd.c
1438
part = xa_load(&disk->part_tbl, partno);
block/genhd.c
1439
if (part)
block/genhd.c
1440
devt = part->bd_dev;
block/genhd.c
163
unsigned int bdev_count_inflight(struct block_device *part)
block/genhd.c
167
bdev_count_inflight_rw(part, inflight, false);
block/genhd.c
344
struct block_device *part;
block/genhd.c
348
xa_for_each(&disk->part_tbl, idx, part) {
block/genhd.c
349
if (bdev_is_partition(part) && !bdev_nr_sectors(part))
block/genhd.c
351
if (!kobject_get_unless_zero(&part->bd_device.kobj))
block/genhd.c
355
kobject_uevent(bdev_kobj(part), action);
block/genhd.c
356
put_device(&part->bd_device);
block/genhd.c
696
struct block_device *part;
block/genhd.c
711
xa_for_each(&disk->part_tbl, idx, part)
block/genhd.c
712
bdev_unhash(part);
block/genhd.c
729
xa_for_each_start(&disk->part_tbl, idx, part, 1)
block/genhd.c
730
drop_partition(part);
block/genhd.c
969
struct block_device *part;
block/genhd.c
976
xa_for_each(&sgp->part_tbl, idx, part) {
block/genhd.c
977
if (!bdev_nr_sectors(part))
block/genhd.c
980
MAJOR(part->bd_dev), MINOR(part->bd_dev),
block/genhd.c
981
bdev_nr_sectors(part) >> 1, part);
block/partitions/acorn.c
65
struct riscix_part part[8];
block/partitions/acorn.c
86
int part;
block/partitions/acorn.c
91
for (part = 0; part < 8; part++) {
block/partitions/acorn.c
92
if (rr->part[part].one &&
block/partitions/acorn.c
93
memcmp(rr->part[part].name, "All\0", 4)) {
block/partitions/acorn.c
95
le32_to_cpu(rr->part[part].start),
block/partitions/acorn.c
96
le32_to_cpu(rr->part[part].length));
block/partitions/acorn.c
98
strlcat(state->pp_buf, rr->part[part].name, PAGE_SIZE);
block/partitions/amiga.c
144
state->disk->disk_name, part);
block/partitions/amiga.c
165
state->disk->disk_name, part,
block/partitions/amiga.c
171
state->disk->disk_name, part,
block/partitions/amiga.c
46
int part, res = 0;
block/partitions/amiga.c
93
for (part = 1; (s32) blk>0 && part<=16; part++, put_dev_sector(sect)) {
block/partitions/amiga.c
97
state->disk->disk_name, blk, part);
block/partitions/atari.c
103
if (!(xrs->part[0].flg & 1)) {
block/partitions/atari.c
110
partsect + be32_to_cpu(xrs->part[0].st),
block/partitions/atari.c
111
be32_to_cpu(xrs->part[0].siz));
block/partitions/atari.c
113
if (!(xrs->part[1].flg & 1)) {
block/partitions/atari.c
118
if (memcmp( xrs->part[1].id, "XGM", 3 ) != 0) {
block/partitions/atari.c
124
partsect = be32_to_cpu(xrs->part[1].st) + extensect;
block/partitions/atari.c
59
if (!VALID_PARTITION(&rs->part[0], hd_size) &&
block/partitions/atari.c
60
!VALID_PARTITION(&rs->part[1], hd_size) &&
block/partitions/atari.c
61
!VALID_PARTITION(&rs->part[2], hd_size) &&
block/partitions/atari.c
62
!VALID_PARTITION(&rs->part[3], hd_size)) {
block/partitions/atari.c
72
pi = &rs->part[0];
block/partitions/atari.c
74
for (slot = 1; pi < &rs->part[4] && slot < state->limit; slot++, pi++) {
block/partitions/atari.h
31
struct partition_info part[4];
block/partitions/core.c
255
const struct block_device *part = dev_to_bdev(dev);
block/partitions/core.c
257
add_uevent_var(env, "PARTN=%u", bdev_partno(part));
block/partitions/core.c
258
if (part->bd_meta_info && part->bd_meta_info->volname[0])
block/partitions/core.c
259
add_uevent_var(env, "PARTNAME=%s", part->bd_meta_info->volname);
block/partitions/core.c
260
if (part->bd_meta_info && part->bd_meta_info->uuid[0])
block/partitions/core.c
261
add_uevent_var(env, "PARTUUID=%s", part->bd_meta_info->uuid);
block/partitions/core.c
272
void drop_partition(struct block_device *part)
block/partitions/core.c
274
lockdep_assert_held(&part->bd_disk->open_mutex);
block/partitions/core.c
276
xa_erase(&part->bd_disk->part_tbl, bdev_partno(part));
block/partitions/core.c
277
kobject_put(part->bd_holder_dir);
block/partitions/core.c
279
device_del(&part->bd_device);
block/partitions/core.c
280
put_device(&part->bd_device);
block/partitions/core.c
410
struct block_device *part;
block/partitions/core.c
415
xa_for_each_start(&disk->part_tbl, idx, part, 1) {
block/partitions/core.c
416
if (bdev_partno(part) != skip_partno &&
block/partitions/core.c
417
start < part->bd_start_sect + bdev_nr_sectors(part) &&
block/partitions/core.c
418
start + length > part->bd_start_sect) {
block/partitions/core.c
431
struct block_device *part;
block/partitions/core.c
450
part = add_partition(disk, partno, start, length,
block/partitions/core.c
452
ret = PTR_ERR_OR_ZERO(part);
block/partitions/core.c
460
struct block_device *part = NULL;
block/partitions/core.c
464
part = xa_load(&disk->part_tbl, partno);
block/partitions/core.c
465
if (!part)
block/partitions/core.c
469
if (atomic_read(&part->bd_openers))
block/partitions/core.c
481
bdev_unhash(part);
block/partitions/core.c
482
invalidate_bdev(part);
block/partitions/core.c
483
drop_partition(part);
block/partitions/core.c
493
struct block_device *part = NULL;
block/partitions/core.c
497
part = xa_load(&disk->part_tbl, partno);
block/partitions/core.c
498
if (!part)
block/partitions/core.c
502
if (start != part->bd_start_sect)
block/partitions/core.c
509
bdev_set_nr_sectors(part, length);
block/partitions/core.c
535
struct block_device *part;
block/partitions/core.c
565
part = add_partition(disk, p, from, size, state->parts[p].flags,
block/partitions/core.c
567
if (IS_ERR(part)) {
block/partitions/core.c
568
if (PTR_ERR(part) != -ENXIO) {
block/partitions/core.c
570
disk->disk_name, p, part);
block/partitions/core.c
577
md_autodetect_dev(part->bd_dev);
block/partitions/core.c
645
struct block_device *part;
block/partitions/core.c
660
xa_for_each_start(&disk->part_tbl, idx, part, 1) {
block/partitions/core.c
666
bdev_unhash(part);
block/partitions/core.c
673
WARN_ON_ONCE(atomic_read(&part->bd_openers));
block/partitions/core.c
674
invalidate_bdev(part);
block/partitions/core.c
675
drop_partition(part);
block/partitions/efi.c
140
static inline int pmbr_part_valid(gpt_mbr_record *part)
block/partitions/efi.c
142
if (part->os_type != EFI_PMBR_OSTYPE_EFI_GPT)
block/partitions/efi.c
146
if (le32_to_cpu(part->starting_lba) != GPT_PRIMARY_PARTITION_TABLE_LBA)
block/partitions/efi.c
175
int i, part = 0, ret = 0; /* invalid by default */
block/partitions/efi.c
183
part = i;
block/partitions/efi.c
215
sz = le32_to_cpu(mbr->partition_record[part].size_in_lba);
block/partitions/ldm.c
1192
if ((v->vblk.part.disk_id == vb->vblk.part.disk_id) &&
block/partitions/ldm.c
1193
(v->vblk.part.start > vb->vblk.part.start)) {
block/partitions/ldm.c
574
struct vblk_part *part;
block/partitions/ldm.c
590
part = &vb->vblk.part;
block/partitions/ldm.c
592
if (part->disk_id != disk->obj_id)
block/partitions/ldm.c
596
part->start, part->size);
block/partitions/ldm.c
922
struct vblk_part *part;
block/partitions/ldm.c
969
part = &vb->vblk.part;
block/partitions/ldm.c
970
part->start = get_unaligned_be64(buffer + 0x24 + r_name);
block/partitions/ldm.c
971
part->volume_offset = get_unaligned_be64(buffer + 0x2C + r_name);
block/partitions/ldm.c
972
part->size = ldm_get_vnum(buffer + 0x34 + r_name);
block/partitions/ldm.c
973
part->parent_id = ldm_get_vnum(buffer + 0x34 + r_size);
block/partitions/ldm.c
974
part->disk_id = ldm_get_vnum(buffer + 0x34 + r_parent);
block/partitions/ldm.c
976
part->partnum = buffer[0x35 + r_diskid];
block/partitions/ldm.c
978
part->partnum = 0;
block/partitions/ldm.h
175
struct vblk_part part;
block/partitions/mac.c
100
be32_to_cpu(part->start_block) * (secsize/512),
block/partitions/mac.c
101
be32_to_cpu(part->block_count) * (secsize/512));
block/partitions/mac.c
103
if (!strncasecmp(part->type, "Linux_RAID", 10))
block/partitions/mac.c
113
mac_fix_string(part->processor, 16);
block/partitions/mac.c
114
mac_fix_string(part->name, 32);
block/partitions/mac.c
115
mac_fix_string(part->type, 32);
block/partitions/mac.c
117
if ((be32_to_cpu(part->status) & MAC_STATUS_BOOTABLE)
block/partitions/mac.c
118
&& strcasecmp(part->processor, "powerpc") == 0)
block/partitions/mac.c
121
if (strcasecmp(part->type, "Apple_UNIX_SVR2") == 0
block/partitions/mac.c
122
|| (strncasecmp(part->type, "Linux", 5) == 0
block/partitions/mac.c
123
&& strcasecmp(part->type, "Linux_swap") != 0)) {
block/partitions/mac.c
127
l = strnlen(part->name, sizeof(part->name));
block/partitions/mac.c
128
if (strncmp(part->name, "/", sizeof(part->name)) == 0)
block/partitions/mac.c
131
if (strncasecmp(part->name + i, "root",
block/partitions/mac.c
137
if (strncasecmp(part->name, "swap", 4) == 0)
block/partitions/mac.c
16
extern void note_bootable_part(dev_t dev, int part, int goodness);
block/partitions/mac.c
43
struct mac_partition *part;
block/partitions/mac.c
71
if (partoffset + sizeof(*part) > datasize) {
block/partitions/mac.c
75
part = (struct mac_partition *) (data + partoffset);
block/partitions/mac.c
76
if (be16_to_cpu(part->signature) != MAC_PARTITION_MAGIC) {
block/partitions/mac.c
80
blocks_in_map = be32_to_cpu(part->map_count);
block/partitions/mac.c
96
part = (struct mac_partition *) (data + pos%512);
block/partitions/mac.c
97
if (be16_to_cpu(part->signature) != MAC_PARTITION_MAGIC)
crypto/krb5/selftest.c
51
unsigned int part = umin(len, 32);
crypto/krb5/selftest.c
53
pr_notice("[%x] %04x: %*phN\n", n, index, part, p + off);
crypto/krb5/selftest.c
54
index += part;
crypto/krb5/selftest.c
55
off += part;
crypto/krb5/selftest.c
56
len -= part;
drivers/accel/amdxdna/aie2_pci.c
276
hwctx->start_col = action->part.start_col;
drivers/accel/amdxdna/aie2_pci.c
277
hwctx->num_col = action->part.ncols;
drivers/accel/amdxdna/aie2_solver.c
295
action->part.start_col = snode->pt_node->start_col;
drivers/accel/amdxdna/aie2_solver.c
296
action->part.ncols = snode->pt_node->ncols;
drivers/accel/amdxdna/aie2_solver.h
64
struct aie_part part;
drivers/acpi/acpica/utmath.c
100
operand_ovl.part.hi = operand_ovl.part.lo;
drivers/acpi/acpica/utmath.c
101
operand_ovl.part.lo = 0;
drivers/acpi/acpica/utmath.c
104
ACPI_SHIFT_LEFT_64_BY_32(operand_ovl.part.hi,
drivers/acpi/acpica/utmath.c
105
operand_ovl.part.lo, count);
drivers/acpi/acpica/utmath.c
137
operand_ovl.part.lo = operand_ovl.part.hi;
drivers/acpi/acpica/utmath.c
138
operand_ovl.part.hi = 0;
drivers/acpi/acpica/utmath.c
141
ACPI_SHIFT_RIGHT_64_BY_32(operand_ovl.part.hi,
drivers/acpi/acpica/utmath.c
142
operand_ovl.part.lo, count);
drivers/acpi/acpica/utmath.c
23
struct uint64_struct part;
drivers/acpi/acpica/utmath.c
278
ACPI_DIV_64_BY_32(0, dividend_ovl.part.hi, divisor,
drivers/acpi/acpica/utmath.c
279
quotient.part.hi, remainder32);
drivers/acpi/acpica/utmath.c
281
ACPI_DIV_64_BY_32(remainder32, dividend_ovl.part.lo, divisor,
drivers/acpi/acpica/utmath.c
282
quotient.part.lo, remainder32);
drivers/acpi/acpica/utmath.c
336
if (divisor.part.hi == 0) {
drivers/acpi/acpica/utmath.c
341
remainder.part.hi = 0;
drivers/acpi/acpica/utmath.c
347
ACPI_DIV_64_BY_32(0, dividend.part.hi, divisor.part.lo,
drivers/acpi/acpica/utmath.c
348
quotient.part.hi, partial1);
drivers/acpi/acpica/utmath.c
350
ACPI_DIV_64_BY_32(partial1, dividend.part.lo, divisor.part.lo,
drivers/acpi/acpica/utmath.c
351
quotient.part.lo, remainder.part.lo);
drivers/acpi/acpica/utmath.c
359
quotient.part.hi = 0;
drivers/acpi/acpica/utmath.c
366
ACPI_SHIFT_RIGHT_64(normalized_divisor.part.hi,
drivers/acpi/acpica/utmath.c
367
normalized_divisor.part.lo);
drivers/acpi/acpica/utmath.c
368
ACPI_SHIFT_RIGHT_64(normalized_dividend.part.hi,
drivers/acpi/acpica/utmath.c
369
normalized_dividend.part.lo);
drivers/acpi/acpica/utmath.c
371
} while (normalized_divisor.part.hi != 0);
drivers/acpi/acpica/utmath.c
375
ACPI_DIV_64_BY_32(normalized_dividend.part.hi,
drivers/acpi/acpica/utmath.c
376
normalized_dividend.part.lo,
drivers/acpi/acpica/utmath.c
377
normalized_divisor.part.lo, quotient.part.lo,
drivers/acpi/acpica/utmath.c
384
partial1 = quotient.part.lo * divisor.part.hi;
drivers/acpi/acpica/utmath.c
385
partial2.full = (u64) quotient.part.lo * divisor.part.lo;
drivers/acpi/acpica/utmath.c
386
partial3.full = (u64) partial2.part.hi + partial1;
drivers/acpi/acpica/utmath.c
388
remainder.part.hi = partial3.part.lo;
drivers/acpi/acpica/utmath.c
389
remainder.part.lo = partial2.part.lo;
drivers/acpi/acpica/utmath.c
391
if (partial3.part.hi == 0) {
drivers/acpi/acpica/utmath.c
392
if (partial3.part.lo >= dividend.part.hi) {
drivers/acpi/acpica/utmath.c
393
if (partial3.part.lo == dividend.part.hi) {
drivers/acpi/acpica/utmath.c
394
if (partial2.part.lo > dividend.part.lo) {
drivers/acpi/acpica/utmath.c
395
quotient.part.lo--;
drivers/acpi/acpica/utmath.c
399
quotient.part.lo--;
drivers/acpi/acpica/utmath.c
405
remainder.part.hi = (u32)-((s32)remainder.part.hi);
drivers/acpi/acpica/utmath.c
406
remainder.part.lo = (u32)-((s32)remainder.part.lo);
drivers/acpi/acpica/utmath.c
408
if (remainder.part.lo) {
drivers/acpi/acpica/utmath.c
409
remainder.part.hi--;
drivers/acpi/acpica/utmath.c
62
ACPI_MUL_64_BY_32(0, multiplicand_ovl.part.hi, multiplier,
drivers/acpi/acpica/utmath.c
63
product.part.hi, carry32);
drivers/acpi/acpica/utmath.c
65
ACPI_MUL_64_BY_32(0, multiplicand_ovl.part.lo, multiplier,
drivers/acpi/acpica/utmath.c
66
product.part.lo, carry32);
drivers/acpi/acpica/utmath.c
68
product.part.hi += carry32;
drivers/char/ipmi/ssif_bmc.c
338
static void calculate_response_part_pec(struct ssif_part_buffer *part)
drivers/char/ipmi/ssif_bmc.c
340
u8 addr = part->address;
drivers/char/ipmi/ssif_bmc.c
343
part->pec = i2c_smbus_pec(0, &addr, 1);
drivers/char/ipmi/ssif_bmc.c
345
part->pec = i2c_smbus_pec(part->pec, &part->smbus_cmd, 1);
drivers/char/ipmi/ssif_bmc.c
348
part->pec = i2c_smbus_pec(part->pec, &addr, 1);
drivers/char/ipmi/ssif_bmc.c
349
part->pec = i2c_smbus_pec(part->pec, &part->length, 1);
drivers/char/ipmi/ssif_bmc.c
350
if (part->length)
drivers/char/ipmi/ssif_bmc.c
351
part->pec = i2c_smbus_pec(part->pec, part->payload, part->length);
drivers/char/ipmi/ssif_bmc.c
356
struct ssif_part_buffer *part = &ssif_bmc->part_buf;
drivers/char/ipmi/ssif_bmc.c
358
part->address = GET_8BIT_ADDR(ssif_bmc->client->addr);
drivers/char/ipmi/ssif_bmc.c
359
part->length = (u8)ssif_bmc->response.len;
drivers/char/ipmi/ssif_bmc.c
362
memset(part->payload + part->length, 0, MAX_PAYLOAD_PER_TRANSACTION - part->length);
drivers/char/ipmi/ssif_bmc.c
363
memcpy(&part->payload[0], &ssif_bmc->response.payload[0], part->length);
drivers/char/ipmi/ssif_bmc.c
368
struct ssif_part_buffer *part = &ssif_bmc->part_buf;
drivers/char/ipmi/ssif_bmc.c
371
part->address = GET_8BIT_ADDR(ssif_bmc->client->addr);
drivers/char/ipmi/ssif_bmc.c
372
switch (part->smbus_cmd) {
drivers/char/ipmi/ssif_bmc.c
381
part->length = MAX_PAYLOAD_PER_TRANSACTION;
drivers/char/ipmi/ssif_bmc.c
385
part->payload[0] = 0x00; /* Start Flag */
drivers/char/ipmi/ssif_bmc.c
386
part->payload[1] = 0x01; /* Start Flag */
drivers/char/ipmi/ssif_bmc.c
388
memcpy(&part->payload[2], &ssif_bmc->response.payload[0], part_len);
drivers/char/ipmi/ssif_bmc.c
405
memset(&part->payload[0], 0, MAX_PAYLOAD_PER_TRANSACTION);
drivers/char/ipmi/ssif_bmc.c
406
part->length = ssif_bmc->remain_len + 1;
drivers/char/ipmi/ssif_bmc.c
409
part->payload[0] = ssif_bmc->block_num;
drivers/char/ipmi/ssif_bmc.c
417
part->length = MAX_PAYLOAD_PER_TRANSACTION;
drivers/char/ipmi/ssif_bmc.c
419
part->payload[0] = ssif_bmc->block_num;
drivers/char/ipmi/ssif_bmc.c
424
memcpy(&part->payload[1], ssif_bmc->response.payload + ssif_bmc->nbytes_processed,
drivers/char/ipmi/ssif_bmc.c
431
__func__, part->smbus_cmd);
drivers/char/ipmi/ssif_bmc.c
462
struct ssif_part_buffer *part = &ssif_bmc->part_buf;
drivers/char/ipmi/ssif_bmc.c
465
if (part->index < part->length)
drivers/char/ipmi/ssif_bmc.c
466
*val = part->payload[part->index];
drivers/char/ipmi/ssif_bmc.c
467
else if (part->index == part->length && ssif_bmc->pec_support)
drivers/char/ipmi/ssif_bmc.c
468
*val = part->pec;
drivers/char/ipmi/ssif_bmc.c
472
part->index++;
drivers/char/ipmi/ssif_bmc.c
497
struct ssif_part_buffer *part = &ssif_bmc->part_buf;
drivers/char/ipmi/ssif_bmc.c
502
if (part->index == part->length) {
drivers/char/ipmi/ssif_bmc.c
509
if (part->index != part->length + 1) {
drivers/char/ipmi/ssif_bmc.c
516
part->pec = part->payload[part->length];
drivers/char/ipmi/ssif_bmc.c
519
cpec = i2c_smbus_pec(cpec, &part->smbus_cmd, 1);
drivers/char/ipmi/ssif_bmc.c
520
cpec = i2c_smbus_pec(cpec, &part->length, 1);
drivers/char/ipmi/ssif_bmc.c
530
if (part->length)
drivers/char/ipmi/ssif_bmc.c
531
cpec = i2c_smbus_pec(cpec, part->payload, part->length);
drivers/char/ipmi/ssif_bmc.c
533
if (cpec != part->pec)
drivers/char/ipmi/ssif_bmc.c
542
struct ssif_part_buffer *part = &ssif_bmc->part_buf;
drivers/char/ipmi/ssif_bmc.c
545
switch (part->smbus_cmd) {
drivers/char/ipmi/ssif_bmc.c
548
ssif_bmc->request.len = part->length;
drivers/char/ipmi/ssif_bmc.c
549
memcpy(ssif_bmc->request.payload, part->payload, part->length);
drivers/char/ipmi/ssif_bmc.c
558
len = ssif_bmc->request.len + part->length;
drivers/char/ipmi/ssif_bmc.c
567
part->payload, part->length);
drivers/char/ipmi/ssif_bmc.c
568
ssif_bmc->request.len += part->length;
drivers/char/ipmi/ssif_bmc.c
574
__func__, part->smbus_cmd);
drivers/char/nvram.c
294
int part, offset;
drivers/char/nvram.c
296
if (copy_from_user(&part, (void __user *)arg,
drivers/char/nvram.c
297
sizeof(part)) != 0)
drivers/char/nvram.c
299
if (part < pmac_nvram_OF || part > pmac_nvram_NR)
drivers/char/nvram.c
301
offset = pmac_get_partition(part);
drivers/cxl/core/cdat.c
269
struct resource *res = &cxlds->part[i].res;
drivers/cxl/core/cdat.c
277
&cxlds->part[i].perf);
drivers/cxl/core/cdat.c
375
struct cxl_dpa_perf *perf = &cxlds->part[i].perf;
drivers/cxl/core/cdat.c
383
struct cxl_dpa_perf *perf = &cxlds->part[i].perf;
drivers/cxl/core/cdat.c
588
if (cxled->part < 0)
drivers/cxl/core/cdat.c
590
perf = &cxlds->part[cxled->part].perf;
drivers/cxl/core/hdm.c
221
const struct resource *part_res = &cxlds->part[i].res;
drivers/cxl/core/hdm.c
398
if (cxled->part < 0)
drivers/cxl/core/hdm.c
400
if (resource_contains(&cxlds->part[i].res, res)) {
drivers/cxl/core/hdm.c
401
cxled->part = i;
drivers/cxl/core/hdm.c
405
if (cxled->part < 0)
drivers/cxl/core/hdm.c
473
const struct cxl_dpa_part_info *part = &info->part[i];
drivers/cxl/core/hdm.c
476
cxlds->part[i].perf.qos_class = CXL_QOS_CLASS_INVALID;
drivers/cxl/core/hdm.c
477
cxlds->part[i].mode = part->mode;
drivers/cxl/core/hdm.c
481
const struct cxl_dpa_part_info *prev = &info->part[i - 1];
drivers/cxl/core/hdm.c
483
if (prev->range.end + 1 != part->range.start)
drivers/cxl/core/hdm.c
486
rc = add_dpa_res(dev, &cxlds->dpa_res, &cxlds->part[i].res,
drivers/cxl/core/hdm.c
487
part->range.start, range_len(&part->range),
drivers/cxl/core/hdm.c
488
cxl_mode_name(part->mode));
drivers/cxl/core/hdm.c
575
int part;
drivers/cxl/core/hdm.c
581
for (part = 0; part < cxlds->nr_partitions; part++)
drivers/cxl/core/hdm.c
582
if (cxlds->part[part].mode == mode)
drivers/cxl/core/hdm.c
585
if (part >= cxlds->nr_partitions) {
drivers/cxl/core/hdm.c
590
if (!resource_size(&cxlds->part[part].res)) {
drivers/cxl/core/hdm.c
595
cxled->part = part;
drivers/cxl/core/hdm.c
607
int part;
drivers/cxl/core/hdm.c
621
part = cxled->part;
drivers/cxl/core/hdm.c
622
if (part < 0) {
drivers/cxl/core/hdm.c
627
res = &cxlds->part[part].res;
drivers/cxl/core/hdm.c
644
for (int i = part; i; i--) {
drivers/cxl/core/hdm.c
645
prev = &cxlds->part[i - 1].res;
drivers/cxl/core/mbox.c
1317
info->part[i].range = (struct range) {
drivers/cxl/core/mbox.c
1321
info->part[i].mode = mode;
drivers/cxl/core/memdev.c
218
const struct resource *res = &cxlds->part[i].res;
drivers/cxl/core/memdev.c
227
if (rc == -EFAULT && cxlds->part[i].mode == CXL_PARTMODE_RAM)
drivers/cxl/core/memdev.c
427
if (cxlds->part[i].mode == CXL_PARTMODE_PMEM)
drivers/cxl/core/memdev.c
428
return &cxlds->part[i].perf;
drivers/cxl/core/memdev.c
452
if (cxlds->part[0].mode != CXL_PARTMODE_RAM)
drivers/cxl/core/memdev.c
454
return &cxlds->part[0].perf;
drivers/cxl/core/memdev.c
81
if (cxlds->part[0].mode != CXL_PARTMODE_RAM)
drivers/cxl/core/memdev.c
83
return resource_size(&cxlds->part[0].res);
drivers/cxl/core/port.c
203
int part = READ_ONCE(cxled->part);
drivers/cxl/core/port.c
206
if (part < 0)
drivers/cxl/core/port.c
209
desc = cxlds->part[part].res.name;
drivers/cxl/core/port.c
2097
cxled->part = -1;
drivers/cxl/core/region.c
1999
if (cxled->part < 0) {
drivers/cxl/core/region.c
2004
if (cxlds->part[cxled->part].mode != cxlr->mode) {
drivers/cxl/core/region.c
2185
cxled->part = -1;
drivers/cxl/core/region.c
2802
int part;
drivers/cxl/core/region.c
2815
if (ctx->part < 0)
drivers/cxl/core/region.c
2822
for (int i = ctx->part; i < cxlds->nr_partitions; i++) {
drivers/cxl/core/region.c
2823
res = &cxlds->part[i].res;
drivers/cxl/core/region.c
2834
if (rc == -EFAULT && cxlds->part[i].mode == CXL_PARTMODE_RAM)
drivers/cxl/core/region.c
2862
mode = cxlds->part[cxled->part].mode;
drivers/cxl/core/region.c
2885
ctx->part = cxled->part;
drivers/cxl/core/region.c
2899
.part = -1,
drivers/cxl/core/region.c
3902
int rc, part = READ_ONCE(cxled->part);
drivers/cxl/core/region.c
3906
cxlr = __create_region(cxlrd, cxlds->part[part].mode,
drivers/cxl/cxl.h
403
int part;
drivers/cxl/cxlmem.h
123
} part[CXL_NR_PARTITIONS_MAX];
drivers/cxl/cxlmem.h
447
struct cxl_dpa_partition part[CXL_NR_PARTITIONS_MAX];
drivers/cxl/cxlmem.h
464
if (cxlds->part[i].mode == CXL_PARTMODE_PMEM)
drivers/cxl/cxlmem.h
465
return resource_size(&cxlds->part[i].res);
drivers/firmware/efi/efi-pstore.c
106
&record->type, &part, &cnt, &time) == 4) {
drivers/firmware/efi/efi-pstore.c
107
record->id = generic_id(time, part, cnt);
drivers/firmware/efi/efi-pstore.c
108
record->part = part;
drivers/firmware/efi/efi-pstore.c
115
&record->type, &part, &time) == 3) {
drivers/firmware/efi/efi-pstore.c
121
record->id = generic_id(time, part, 0);
drivers/firmware/efi/efi-pstore.c
122
record->part = part;
drivers/firmware/efi/efi-pstore.c
209
record->id = generic_id(record->time.tv_sec, record->part,
drivers/firmware/efi/efi-pstore.c
216
record->type, record->part, record->count,
drivers/firmware/efi/efi-pstore.c
76
static inline u64 generic_id(u64 timestamp, unsigned int part, int count)
drivers/firmware/efi/efi-pstore.c
78
return (timestamp * 100 + part) * 1000 + count;
drivers/firmware/efi/efi-pstore.c
88
unsigned int part;
drivers/firmware/efi/efi-pstore.c
94
&record->type, &part, &cnt, &time, &data_type) == 5) {
drivers/firmware/efi/efi-pstore.c
95
record->id = generic_id(time, part, cnt);
drivers/firmware/efi/efi-pstore.c
96
record->part = part;
drivers/gpu/drm/i915/display/intel_dp_hdcp.c
228
int i, u32 *part)
drivers/gpu/drm/i915/display/intel_dp_hdcp.c
237
DP_AUX_HDCP_V_PRIME(i), part,
drivers/gpu/drm/i915/display/intel_hdcp_shim.h
82
int i, u32 *part);
drivers/gpu/drm/i915/display/intel_hdmi.c
1495
int i, u32 *part)
drivers/gpu/drm/i915/display/intel_hdmi.c
1504
part, DRM_HDCP_V_PRIME_PART_LEN);
drivers/gpu/drm/i915/i915_perf.c
892
u32 part = oa_buf_end - (u8 *)report32;
drivers/gpu/drm/i915/i915_perf.c
895
if (report_size <= part) {
drivers/gpu/drm/i915/i915_perf.c
898
memset(report32, 0, part);
drivers/gpu/drm/i915/i915_perf.c
899
memset(oa_buf_base, 0, report_size - part);
drivers/gpu/drm/nouveau/dispnv50/wndw.c
93
args.nv50.part = NV50_DMA_V0_PART_256;
drivers/gpu/drm/nouveau/dispnv50/wndw.c
97
args.nv50.part = NV50_DMA_V0_PART_256;
drivers/gpu/drm/nouveau/include/nvif/cl0002.h
33
__u8 part;
drivers/gpu/drm/nouveau/include/nvkm/subdev/mmu.h
18
bool part:1; /* Region was split from an allocated region by map(). */
drivers/gpu/drm/nouveau/nvkm/engine/dma/usernv50.c
100
part = args->v0.part;
drivers/gpu/drm/nouveau/nvkm/engine/dma/usernv50.c
107
part = NV50_DMA_V0_PART_256;
drivers/gpu/drm/nouveau/nvkm/engine/dma/usernv50.c
112
part = NV50_DMA_V0_PART_VM;
drivers/gpu/drm/nouveau/nvkm/engine/dma/usernv50.c
119
if (user > 2 || part > 2 || comp > 3 || kind > 0x7f)
drivers/gpu/drm/nouveau/nvkm/engine/dma/usernv50.c
123
dmaobj->flags5 = (part << 16);
drivers/gpu/drm/nouveau/nvkm/engine/dma/usernv50.c
78
u32 user, part, comp, kind;
drivers/gpu/drm/nouveau/nvkm/engine/dma/usernv50.c
97
args->v0.priv, args->v0.part, args->v0.comp,
drivers/gpu/drm/nouveau/nvkm/subdev/devinit/gt215.c
122
u32 part = (addr / mmio[2]) & 7;
drivers/gpu/drm/nouveau/nvkm/subdev/devinit/gt215.c
125
if (part >= hweight8((init->r001540 >> 16) & 0xff))
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgf100.c
112
u32 part = nvkm_rd32(device, 0x022438), i;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgf100.c
119
for (i = 0; (magic & 0x80000000) && i < part; addr += 0x1000, i++) {
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
793
MUX_METHOD_DATA_ELEMENT *part)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
802
part->status = 0xffff;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
832
part->acpiId = id;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
833
part->mode = value;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
834
part->status = 0;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c
249
if (ret = -ENOENT, !vma || vma->addr != addr || vma->part) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c
251
vma ? vma->addr : ~0ULL, vma ? vma->part : 0);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1009
vma->part = true;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1019
tmp->part = true;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1036
vma->part ? 'P' : '-',
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1241
if (vma->addr == addr && vma->part && (prev = node(vma, prev))) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1247
if (!next->part ||
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1426
if (vma->part && (prev = node(vma, prev)) && prev->mapped)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1428
if ((next = node(vma, next)) && (!next->part || next->mapped))
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1636
BUG_ON(vma->part);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1647
while ((next = node(next, next)) && next->part &&
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1666
} while (next && next->part);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1677
} while ((next = node(vma, next)) && next->part);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1915
if (inst && vmm && vmm->func->part) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1917
vmm->func->part(vmm, inst);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
675
u64 part = ALIGN(addr, next) - addr;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
676
if (size - part >= next)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
677
block = (part >> page[i].shift) << page[i].shift;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
852
new->part = vma->part;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.h
149
void (*part)(struct nvkm_vmm *, struct nvkm_memory *inst);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
374
.part = gf100_vmm_part,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
389
.part = gf100_vmm_part,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
282
.part = gf100_vmm_part,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgk104.c
70
.part = gf100_vmm_part,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgk104.c
85
.part = gf100_vmm_part,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgk20a.c
39
.part = gf100_vmm_part,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgk20a.c
54
.part = gf100_vmm_part,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgm200.c
112
.part = gf100_vmm_part,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgm200.c
128
.part = gf100_vmm_part,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgm20b.c
27
.part = gf100_vmm_part,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgm20b.c
43
.part = gf100_vmm_part,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
596
.part = gf100_vmm_part,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp10b.c
27
.part = gf100_vmm_part,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgv100.c
65
.part = gf100_vmm_part,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmmcp77.c
27
.part = nv50_vmm_part,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c
369
.part = nv50_vmm_part,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmtu102.c
56
.part = gf100_vmm_part,
drivers/gpu/drm/xe/xe_oa.c
379
u32 part = oa_buf_end - report;
drivers/gpu/drm/xe/xe_oa.c
382
if (report_size <= part) {
drivers/gpu/drm/xe/xe_oa.c
385
memset(report, 0, part);
drivers/gpu/drm/xe/xe_oa.c
386
memset(oa_buf_base, 0, report_size - part);
drivers/i3c/device.c
268
u16 manuf, part, ext_info;
drivers/i3c/device.c
274
part = I3C_PID_PART_ID(devinfo.pid);
drivers/i3c/device.c
288
(rndpid || id->part_id != part))
drivers/i3c/master.c
285
u16 manuf, part, ext;
drivers/i3c/master.c
289
part = I3C_PID_PART_ID(devinfo.pid);
drivers/i3c/master.c
297
devinfo.dcr, manuf, part, ext);
drivers/i3c/master.c
316
u16 manuf, part, ext;
drivers/i3c/master.c
321
part = I3C_PID_PART_ID(devinfo.pid);
drivers/i3c/master.c
330
devinfo.dcr, manuf, part, ext);
drivers/iio/light/si1145.c
1300
if (part_id != data->part_info->part) {
drivers/iio/light/si1145.c
1302
part_id, data->part_info->part);
drivers/iio/light/si1145.c
152
u8 part;
drivers/infiniband/sw/siw/siw_qp_tx.c
88
unsigned long part = bytes - (PAGE_SIZE - off);
drivers/infiniband/sw/siw/siw_qp_tx.c
90
memcpy(paddr, buffer + off, part);
drivers/infiniband/sw/siw/siw_qp_tx.c
93
p = siw_get_page(mem, sge, part, &pbl_idx);
drivers/infiniband/sw/siw/siw_qp_tx.c
98
memcpy(paddr + part, buffer, bytes - part);
drivers/irqchip/irq-gic-v3.c
2098
struct partition_affinity *part;
drivers/irqchip/irq-gic-v3.c
2101
part = &parts[part_idx];
drivers/irqchip/irq-gic-v3.c
2103
part->partition_id = of_fwnode_handle(child_part);
drivers/irqchip/irq-gic-v3.c
2134
cpumask_set_cpu(cpu, &part->mask);
drivers/macintosh/smu.c
1018
const struct smu_sdbp_header *part;
drivers/macintosh/smu.c
1035
part = of_get_property(smu->of_node, pname, size);
drivers/macintosh/smu.c
1036
if (part == NULL) {
drivers/macintosh/smu.c
1038
part = smu_create_sdb_partition(id);
drivers/macintosh/smu.c
1039
if (part != NULL && size)
drivers/macintosh/smu.c
1040
*size = part->len << 2;
drivers/macintosh/smu.c
1043
return part;
drivers/macintosh/smu.c
1126
const struct smu_sdbp_header *part;
drivers/macintosh/smu.c
1127
part = __smu_get_sdb_partition(hdr.cmd, NULL, 1);
drivers/macintosh/smu.c
1128
if (part == NULL)
drivers/macintosh/smu.c
1130
else if (IS_ERR(part))
drivers/macintosh/smu.c
1131
return PTR_ERR(part);
drivers/md/dm-vdo/encodings.c
927
struct partition *part = layout->head;
drivers/md/dm-vdo/encodings.c
929
layout->head = part->next;
drivers/md/dm-vdo/encodings.c
930
vdo_free(part);
drivers/md/md.c
10761
void md_autostart_arrays(int part)
drivers/md/md.c
10798
autorun_devices(part);
drivers/md/md.c
7086
static void autorun_devices(int part)
drivers/md/md.c
7112
if (part) {
drivers/md/md.h
1028
void md_autostart_arrays(int part);
drivers/media/platform/renesas/vsp1/vsp1_video.c
261
struct vsp1_partition *part = &pipe->part_table[partition];
drivers/media/platform/renesas/vsp1/vsp1_video.c
266
vsp1_entity_configure_partition(entity, pipe, part, dl, dlb);
drivers/misc/bh1770glc.c
509
u8 manu, part;
drivers/misc/bh1770glc.c
519
part = (u8)ret;
drivers/misc/bh1770glc.c
521
chip->revision = (part & BH1770_REV_MASK) >> BH1770_REV_SHIFT;
drivers/misc/bh1770glc.c
527
((part & BH1770_PART_MASK) == BH1770_PART)) {
drivers/misc/bh1770glc.c
533
((part & BH1770_PART_MASK) == BH1770_PART)) {
drivers/misc/sgi-xp/xpc.h
637
extern int xpc_partition_disengaged_from_timer(struct xpc_partition *part);
drivers/misc/sgi-xp/xpc.h
665
xpc_wakeup_channel_mgr(struct xpc_partition *part)
drivers/misc/sgi-xp/xpc.h
667
if (atomic_inc_return(&part->channel_mgr_requests) == 1)
drivers/misc/sgi-xp/xpc.h
668
wake_up(&part->channel_mgr_wq);
drivers/misc/sgi-xp/xpc.h
699
xpc_part_deref(struct xpc_partition *part)
drivers/misc/sgi-xp/xpc.h
701
s32 refs = atomic_dec_return(&part->references);
drivers/misc/sgi-xp/xpc.h
704
if (refs == 0 && part->setup_state == XPC_P_SS_WTEARDOWN)
drivers/misc/sgi-xp/xpc.h
705
wake_up(&part->teardown_wq);
drivers/misc/sgi-xp/xpc.h
709
xpc_part_ref(struct xpc_partition *part)
drivers/misc/sgi-xp/xpc.h
713
atomic_inc(&part->references);
drivers/misc/sgi-xp/xpc.h
714
setup = (part->setup_state == XPC_P_SS_SETUP);
drivers/misc/sgi-xp/xpc.h
716
xpc_part_deref(part);
drivers/misc/sgi-xp/xpc_channel.c
1000
struct xpc_partition *part = &xpc_partitions[partid];
drivers/misc/sgi-xp/xpc_channel.c
1004
DBUG_ON(ch_number < 0 || ch_number >= part->nchannels);
drivers/misc/sgi-xp/xpc_channel.c
1006
ch = &part->channels[ch_number];
drivers/misc/sgi-xp/xpc_channel.c
101
if (part->act_state == XPC_P_AS_DEACTIVATING) {
drivers/misc/sgi-xp/xpc_channel.c
157
atomic_dec(&part->nchannels_active);
drivers/misc/sgi-xp/xpc_channel.c
168
if (part->act_state != XPC_P_AS_DEACTIVATING) {
drivers/misc/sgi-xp/xpc_channel.c
170
spin_lock(&part->chctl_lock);
drivers/misc/sgi-xp/xpc_channel.c
171
part->chctl.flags[ch->number] |=
drivers/misc/sgi-xp/xpc_channel.c
173
spin_unlock(&part->chctl_lock);
drivers/misc/sgi-xp/xpc_channel.c
183
xpc_process_openclose_chctl_flags(struct xpc_partition *part, int ch_number,
drivers/misc/sgi-xp/xpc_channel.c
188
&part->remote_openclose_args[ch_number];
drivers/misc/sgi-xp/xpc_channel.c
189
struct xpc_channel *ch = &part->channels[ch_number];
drivers/misc/sgi-xp/xpc_channel.c
238
if (part->chctl.flags[ch_number] &
drivers/misc/sgi-xp/xpc_channel.c
242
spin_lock(&part->chctl_lock);
drivers/misc/sgi-xp/xpc_channel.c
243
part->chctl.flags[ch_number] |=
drivers/misc/sgi-xp/xpc_channel.c
245
spin_unlock(&part->chctl_lock);
drivers/misc/sgi-xp/xpc_channel.c
253
atomic_inc(&part->nchannels_active);
drivers/misc/sgi-xp/xpc_channel.c
289
DBUG_ON(part->act_state != XPC_P_AS_DEACTIVATING);
drivers/misc/sgi-xp/xpc_channel.c
296
if (part->chctl.flags[ch_number] &
drivers/misc/sgi-xp/xpc_channel.c
300
spin_lock(&part->chctl_lock);
drivers/misc/sgi-xp/xpc_channel.c
301
part->chctl.flags[ch_number] |=
drivers/misc/sgi-xp/xpc_channel.c
303
spin_unlock(&part->chctl_lock);
drivers/misc/sgi-xp/xpc_channel.c
323
if (part->act_state == XPC_P_AS_DEACTIVATING ||
drivers/misc/sgi-xp/xpc_channel.c
362
atomic_inc(&part->nchannels_active);
drivers/misc/sgi-xp/xpc_channel.c
546
xpc_process_sent_chctl_flags(struct xpc_partition *part)
drivers/misc/sgi-xp/xpc_channel.c
554
chctl.all_flags = xpc_arch_ops.get_chctl_all_flags(part);
drivers/misc/sgi-xp/xpc_channel.c
563
for (ch_number = 0; ch_number < part->nchannels; ch_number++) {
drivers/misc/sgi-xp/xpc_channel.c
564
ch = &part->channels[ch_number];
drivers/misc/sgi-xp/xpc_channel.c
572
xpc_process_openclose_chctl_flags(part, ch_number,
drivers/misc/sgi-xp/xpc_channel.c
585
if (part->act_state == XPC_P_AS_DEACTIVATING)
drivers/misc/sgi-xp/xpc_channel.c
603
xpc_arch_ops.process_msg_chctl_flags(part, ch_number);
drivers/misc/sgi-xp/xpc_channel.c
617
xpc_partition_going_down(struct xpc_partition *part, enum xp_retval reason)
drivers/misc/sgi-xp/xpc_channel.c
624
XPC_PARTID(part), reason);
drivers/misc/sgi-xp/xpc_channel.c
626
if (!xpc_part_ref(part)) {
drivers/misc/sgi-xp/xpc_channel.c
633
for (ch_number = 0; ch_number < part->nchannels; ch_number++) {
drivers/misc/sgi-xp/xpc_channel.c
634
ch = &part->channels[ch_number];
drivers/misc/sgi-xp/xpc_channel.c
645
xpc_wakeup_channel_mgr(part);
drivers/misc/sgi-xp/xpc_channel.c
647
xpc_part_deref(part);
drivers/misc/sgi-xp/xpc_channel.c
658
struct xpc_partition *part;
drivers/misc/sgi-xp/xpc_channel.c
663
part = &xpc_partitions[partid];
drivers/misc/sgi-xp/xpc_channel.c
665
if (xpc_part_ref(part)) {
drivers/misc/sgi-xp/xpc_channel.c
670
xpc_wakeup_channel_mgr(part);
drivers/misc/sgi-xp/xpc_channel.c
671
xpc_part_deref(part);
drivers/misc/sgi-xp/xpc_channel.c
711
struct xpc_partition *part;
drivers/misc/sgi-xp/xpc_channel.c
718
part = &xpc_partitions[partid];
drivers/misc/sgi-xp/xpc_channel.c
720
if (xpc_part_ref(part)) {
drivers/misc/sgi-xp/xpc_channel.c
721
ch = &part->channels[ch_number];
drivers/misc/sgi-xp/xpc_channel.c
736
xpc_part_deref(part);
drivers/misc/sgi-xp/xpc_channel.c
82
struct xpc_partition *part = &xpc_partitions[ch->partid];
drivers/misc/sgi-xp/xpc_channel.c
875
struct xpc_partition *part = &xpc_partitions[partid];
drivers/misc/sgi-xp/xpc_channel.c
882
DBUG_ON(ch_number < 0 || ch_number >= part->nchannels);
drivers/misc/sgi-xp/xpc_channel.c
885
if (xpc_part_ref(part)) {
drivers/misc/sgi-xp/xpc_channel.c
886
ret = xpc_arch_ops.send_payload(&part->channels[ch_number],
drivers/misc/sgi-xp/xpc_channel.c
888
xpc_part_deref(part);
drivers/misc/sgi-xp/xpc_channel.c
926
struct xpc_partition *part = &xpc_partitions[partid];
drivers/misc/sgi-xp/xpc_channel.c
933
DBUG_ON(ch_number < 0 || ch_number >= part->nchannels);
drivers/misc/sgi-xp/xpc_channel.c
937
if (xpc_part_ref(part)) {
drivers/misc/sgi-xp/xpc_channel.c
938
ret = xpc_arch_ops.send_payload(&part->channels[ch_number],
drivers/misc/sgi-xp/xpc_channel.c
940
xpc_part_deref(part);
drivers/misc/sgi-xp/xpc_main.c
1075
struct xpc_partition *part;
drivers/misc/sgi-xp/xpc_main.c
1090
part = &xpc_partitions[partid];
drivers/misc/sgi-xp/xpc_main.c
1093
part->act_state != XPC_P_AS_INACTIVE) {
drivers/misc/sgi-xp/xpc_main.c
1094
xpc_arch_ops.request_partition_deactivation(part);
drivers/misc/sgi-xp/xpc_main.c
1095
xpc_arch_ops.indicate_partition_disengaged(part);
drivers/misc/sgi-xp/xpc_main.c
167
struct xpc_partition *part = timer_container_of(part, t,
drivers/misc/sgi-xp/xpc_main.c
170
DBUG_ON(time_is_after_jiffies(part->disengage_timeout));
drivers/misc/sgi-xp/xpc_main.c
172
xpc_partition_disengaged_from_timer(part);
drivers/misc/sgi-xp/xpc_main.c
174
DBUG_ON(part->disengage_timeout != 0);
drivers/misc/sgi-xp/xpc_main.c
175
DBUG_ON(xpc_arch_ops.partition_engaged(XPC_PARTID(part)));
drivers/misc/sgi-xp/xpc_main.c
217
struct xpc_partition *part;
drivers/misc/sgi-xp/xpc_main.c
229
part = &xpc_partitions[partid];
drivers/misc/sgi-xp/xpc_main.c
231
if (part->act_state == XPC_P_AS_INACTIVE ||
drivers/misc/sgi-xp/xpc_main.c
232
part->act_state == XPC_P_AS_DEACTIVATING) {
drivers/misc/sgi-xp/xpc_main.c
236
ret = xpc_arch_ops.get_remote_heartbeat(part);
drivers/misc/sgi-xp/xpc_main.c
238
XPC_DEACTIVATE_PARTITION(part, ret);
drivers/misc/sgi-xp/xpc_main.c
330
xpc_channel_mgr(struct xpc_partition *part)
drivers/misc/sgi-xp/xpc_main.c
332
while (part->act_state != XPC_P_AS_DEACTIVATING ||
drivers/misc/sgi-xp/xpc_main.c
333
atomic_read(&part->nchannels_active) > 0 ||
drivers/misc/sgi-xp/xpc_main.c
334
!xpc_partition_disengaged(part)) {
drivers/misc/sgi-xp/xpc_main.c
336
xpc_process_sent_chctl_flags(part);
drivers/misc/sgi-xp/xpc_main.c
351
atomic_dec(&part->channel_mgr_requests);
drivers/misc/sgi-xp/xpc_main.c
352
(void)wait_event_interruptible(part->channel_mgr_wq,
drivers/misc/sgi-xp/xpc_main.c
353
(atomic_read(&part->channel_mgr_requests) > 0 ||
drivers/misc/sgi-xp/xpc_main.c
354
part->chctl.all_flags != 0 ||
drivers/misc/sgi-xp/xpc_main.c
355
(part->act_state == XPC_P_AS_DEACTIVATING &&
drivers/misc/sgi-xp/xpc_main.c
356
atomic_read(&part->nchannels_active) == 0 &&
drivers/misc/sgi-xp/xpc_main.c
357
xpc_partition_disengaged(part))));
drivers/misc/sgi-xp/xpc_main.c
358
atomic_set(&part->channel_mgr_requests, 1);
drivers/misc/sgi-xp/xpc_main.c
391
xpc_setup_ch_structures(struct xpc_partition *part)
drivers/misc/sgi-xp/xpc_main.c
396
short partid = XPC_PARTID(part);
drivers/misc/sgi-xp/xpc_main.c
402
DBUG_ON(part->channels != NULL);
drivers/misc/sgi-xp/xpc_main.c
403
part->channels = kzalloc_objs(struct xpc_channel, XPC_MAX_NCHANNELS);
drivers/misc/sgi-xp/xpc_main.c
404
if (part->channels == NULL) {
drivers/misc/sgi-xp/xpc_main.c
411
part->remote_openclose_args =
drivers/misc/sgi-xp/xpc_main.c
413
GFP_KERNEL, &part->
drivers/misc/sgi-xp/xpc_main.c
415
if (part->remote_openclose_args == NULL) {
drivers/misc/sgi-xp/xpc_main.c
421
part->chctl.all_flags = 0;
drivers/misc/sgi-xp/xpc_main.c
422
spin_lock_init(&part->chctl_lock);
drivers/misc/sgi-xp/xpc_main.c
424
atomic_set(&part->channel_mgr_requests, 1);
drivers/misc/sgi-xp/xpc_main.c
425
init_waitqueue_head(&part->channel_mgr_wq);
drivers/misc/sgi-xp/xpc_main.c
427
part->nchannels = XPC_MAX_NCHANNELS;
drivers/misc/sgi-xp/xpc_main.c
429
atomic_set(&part->nchannels_active, 0);
drivers/misc/sgi-xp/xpc_main.c
430
atomic_set(&part->nchannels_engaged, 0);
drivers/misc/sgi-xp/xpc_main.c
432
for (ch_number = 0; ch_number < part->nchannels; ch_number++) {
drivers/misc/sgi-xp/xpc_main.c
433
ch = &part->channels[ch_number];
drivers/misc/sgi-xp/xpc_main.c
454
ret = xpc_arch_ops.setup_ch_structures(part);
drivers/misc/sgi-xp/xpc_main.c
462
part->setup_state = XPC_P_SS_SETUP;
drivers/misc/sgi-xp/xpc_main.c
468
kfree(part->remote_openclose_args_base);
drivers/misc/sgi-xp/xpc_main.c
469
part->remote_openclose_args = NULL;
drivers/misc/sgi-xp/xpc_main.c
471
kfree(part->channels);
drivers/misc/sgi-xp/xpc_main.c
472
part->channels = NULL;
drivers/misc/sgi-xp/xpc_main.c
481
xpc_teardown_ch_structures(struct xpc_partition *part)
drivers/misc/sgi-xp/xpc_main.c
483
DBUG_ON(atomic_read(&part->nchannels_engaged) != 0);
drivers/misc/sgi-xp/xpc_main.c
484
DBUG_ON(atomic_read(&part->nchannels_active) != 0);
drivers/misc/sgi-xp/xpc_main.c
491
DBUG_ON(part->setup_state != XPC_P_SS_SETUP);
drivers/misc/sgi-xp/xpc_main.c
492
part->setup_state = XPC_P_SS_WTEARDOWN;
drivers/misc/sgi-xp/xpc_main.c
494
wait_event(part->teardown_wq, (atomic_read(&part->references) == 0));
drivers/misc/sgi-xp/xpc_main.c
498
xpc_arch_ops.teardown_ch_structures(part);
drivers/misc/sgi-xp/xpc_main.c
500
kfree(part->remote_openclose_args_base);
drivers/misc/sgi-xp/xpc_main.c
501
part->remote_openclose_args = NULL;
drivers/misc/sgi-xp/xpc_main.c
502
kfree(part->channels);
drivers/misc/sgi-xp/xpc_main.c
503
part->channels = NULL;
drivers/misc/sgi-xp/xpc_main.c
505
part->setup_state = XPC_P_SS_TORNDOWN;
drivers/misc/sgi-xp/xpc_main.c
523
struct xpc_partition *part = &xpc_partitions[partid];
drivers/misc/sgi-xp/xpc_main.c
528
spin_lock_irqsave(&part->act_lock, irq_flags);
drivers/misc/sgi-xp/xpc_main.c
530
if (part->act_state == XPC_P_AS_DEACTIVATING) {
drivers/misc/sgi-xp/xpc_main.c
531
part->act_state = XPC_P_AS_INACTIVE;
drivers/misc/sgi-xp/xpc_main.c
532
spin_unlock_irqrestore(&part->act_lock, irq_flags);
drivers/misc/sgi-xp/xpc_main.c
533
part->remote_rp_pa = 0;
drivers/misc/sgi-xp/xpc_main.c
538
DBUG_ON(part->act_state != XPC_P_AS_ACTIVATION_REQ);
drivers/misc/sgi-xp/xpc_main.c
539
part->act_state = XPC_P_AS_ACTIVATING;
drivers/misc/sgi-xp/xpc_main.c
541
XPC_SET_REASON(part, 0, 0);
drivers/misc/sgi-xp/xpc_main.c
542
spin_unlock_irqrestore(&part->act_lock, irq_flags);
drivers/misc/sgi-xp/xpc_main.c
548
if (xpc_setup_ch_structures(part) == xpSuccess) {
drivers/misc/sgi-xp/xpc_main.c
549
(void)xpc_part_ref(part); /* this will always succeed */
drivers/misc/sgi-xp/xpc_main.c
551
if (xpc_arch_ops.make_first_contact(part) == xpSuccess) {
drivers/misc/sgi-xp/xpc_main.c
552
xpc_mark_partition_active(part);
drivers/misc/sgi-xp/xpc_main.c
553
xpc_channel_mgr(part);
drivers/misc/sgi-xp/xpc_main.c
557
xpc_part_deref(part);
drivers/misc/sgi-xp/xpc_main.c
558
xpc_teardown_ch_structures(part);
drivers/misc/sgi-xp/xpc_main.c
562
xpc_mark_partition_inactive(part);
drivers/misc/sgi-xp/xpc_main.c
564
if (part->reason == xpReactivating) {
drivers/misc/sgi-xp/xpc_main.c
566
xpc_arch_ops.request_partition_reactivation(part);
drivers/misc/sgi-xp/xpc_main.c
573
xpc_activate_partition(struct xpc_partition *part)
drivers/misc/sgi-xp/xpc_main.c
575
short partid = XPC_PARTID(part);
drivers/misc/sgi-xp/xpc_main.c
579
spin_lock_irqsave(&part->act_lock, irq_flags);
drivers/misc/sgi-xp/xpc_main.c
581
DBUG_ON(part->act_state != XPC_P_AS_INACTIVE);
drivers/misc/sgi-xp/xpc_main.c
583
part->act_state = XPC_P_AS_ACTIVATION_REQ;
drivers/misc/sgi-xp/xpc_main.c
584
XPC_SET_REASON(part, xpCloneKThread, __LINE__);
drivers/misc/sgi-xp/xpc_main.c
586
spin_unlock_irqrestore(&part->act_lock, irq_flags);
drivers/misc/sgi-xp/xpc_main.c
591
spin_lock_irqsave(&part->act_lock, irq_flags);
drivers/misc/sgi-xp/xpc_main.c
592
part->act_state = XPC_P_AS_INACTIVE;
drivers/misc/sgi-xp/xpc_main.c
593
XPC_SET_REASON(part, xpCloneKThreadFailed, __LINE__);
drivers/misc/sgi-xp/xpc_main.c
594
spin_unlock_irqrestore(&part->act_lock, irq_flags);
drivers/misc/sgi-xp/xpc_main.c
637
xpc_kthread_waitmsgs(struct xpc_partition *part, struct xpc_channel *ch)
drivers/misc/sgi-xp/xpc_main.c
674
struct xpc_partition *part = &xpc_partitions[partid];
drivers/misc/sgi-xp/xpc_main.c
684
ch = &part->channels[ch_number];
drivers/misc/sgi-xp/xpc_main.c
716
xpc_kthread_waitmsgs(part, ch);
drivers/misc/sgi-xp/xpc_main.c
735
atomic_dec_return(&part->nchannels_engaged) == 0) {
drivers/misc/sgi-xp/xpc_main.c
736
xpc_arch_ops.indicate_partition_disengaged(part);
drivers/misc/sgi-xp/xpc_main.c
744
xpc_part_deref(part);
drivers/misc/sgi-xp/xpc_main.c
766
struct xpc_partition *part = &xpc_partitions[ch->partid];
drivers/misc/sgi-xp/xpc_main.c
790
atomic_inc_return(&part->nchannels_engaged) == 1) {
drivers/misc/sgi-xp/xpc_main.c
791
xpc_arch_ops.indicate_partition_engaged(part);
drivers/misc/sgi-xp/xpc_main.c
793
(void)xpc_part_ref(part);
drivers/misc/sgi-xp/xpc_main.c
812
atomic_dec_return(&part->nchannels_engaged) == 0) {
drivers/misc/sgi-xp/xpc_main.c
813
indicate_partition_disengaged(part);
drivers/misc/sgi-xp/xpc_main.c
816
xpc_part_deref(part);
drivers/misc/sgi-xp/xpc_main.c
840
struct xpc_partition *part;
drivers/misc/sgi-xp/xpc_main.c
846
part = &xpc_partitions[partid];
drivers/misc/sgi-xp/xpc_main.c
848
if (!xpc_part_ref(part))
drivers/misc/sgi-xp/xpc_main.c
851
ch = &part->channels[ch_number];
drivers/misc/sgi-xp/xpc_main.c
854
xpc_part_deref(part);
drivers/misc/sgi-xp/xpc_main.c
865
if (part->act_state != XPC_P_AS_DEACTIVATING) {
drivers/misc/sgi-xp/xpc_main.c
866
spin_lock(&part->chctl_lock);
drivers/misc/sgi-xp/xpc_main.c
867
part->chctl.flags[ch->number] |=
drivers/misc/sgi-xp/xpc_main.c
869
spin_unlock(&part->chctl_lock);
drivers/misc/sgi-xp/xpc_main.c
879
xpc_wakeup_channel_mgr(part);
drivers/misc/sgi-xp/xpc_main.c
881
xpc_part_deref(part);
drivers/misc/sgi-xp/xpc_main.c
889
struct xpc_partition *part;
drivers/misc/sgi-xp/xpc_main.c
906
part = &xpc_partitions[partid];
drivers/misc/sgi-xp/xpc_main.c
908
DBUG_ON((u64)part != L1_CACHE_ALIGN((u64)part));
drivers/misc/sgi-xp/xpc_main.c
910
part->activate_IRQ_rcvd = 0;
drivers/misc/sgi-xp/xpc_main.c
911
spin_lock_init(&part->act_lock);
drivers/misc/sgi-xp/xpc_main.c
912
part->act_state = XPC_P_AS_INACTIVE;
drivers/misc/sgi-xp/xpc_main.c
913
XPC_SET_REASON(part, 0, 0);
drivers/misc/sgi-xp/xpc_main.c
915
timer_setup(&part->disengage_timer,
drivers/misc/sgi-xp/xpc_main.c
918
part->setup_state = XPC_P_SS_UNSET;
drivers/misc/sgi-xp/xpc_main.c
919
init_waitqueue_head(&part->teardown_wq);
drivers/misc/sgi-xp/xpc_main.c
920
atomic_set(&part->references, 0);
drivers/misc/sgi-xp/xpc_main.c
938
struct xpc_partition *part;
drivers/misc/sgi-xp/xpc_main.c
970
part = &xpc_partitions[partid];
drivers/misc/sgi-xp/xpc_main.c
972
if (xpc_partition_disengaged(part) &&
drivers/misc/sgi-xp/xpc_main.c
973
part->act_state == XPC_P_AS_INACTIVE) {
drivers/misc/sgi-xp/xpc_main.c
979
XPC_DEACTIVATE_PARTITION(part, reason);
drivers/misc/sgi-xp/xpc_main.c
981
if (part->disengage_timeout > disengage_timeout)
drivers/misc/sgi-xp/xpc_main.c
982
disengage_timeout = part->disengage_timeout;
drivers/misc/sgi-xp/xpc_partition.c
265
static int __xpc_partition_disengaged(struct xpc_partition *part,
drivers/misc/sgi-xp/xpc_partition.c
268
short partid = XPC_PARTID(part);
drivers/misc/sgi-xp/xpc_partition.c
272
if (part->disengage_timeout) {
drivers/misc/sgi-xp/xpc_partition.c
274
if (time_is_after_jiffies(part->disengage_timeout)) {
drivers/misc/sgi-xp/xpc_partition.c
290
part->disengage_timeout = 0;
drivers/misc/sgi-xp/xpc_partition.c
294
timer_delete_sync(&part->disengage_timer);
drivers/misc/sgi-xp/xpc_partition.c
296
DBUG_ON(part->act_state != XPC_P_AS_DEACTIVATING &&
drivers/misc/sgi-xp/xpc_partition.c
297
part->act_state != XPC_P_AS_INACTIVE);
drivers/misc/sgi-xp/xpc_partition.c
298
if (part->act_state != XPC_P_AS_INACTIVE)
drivers/misc/sgi-xp/xpc_partition.c
299
xpc_wakeup_channel_mgr(part);
drivers/misc/sgi-xp/xpc_partition.c
301
xpc_arch_ops.cancel_partition_deactivation_request(part);
drivers/misc/sgi-xp/xpc_partition.c
306
int xpc_partition_disengaged(struct xpc_partition *part)
drivers/misc/sgi-xp/xpc_partition.c
308
return __xpc_partition_disengaged(part, false);
drivers/misc/sgi-xp/xpc_partition.c
311
int xpc_partition_disengaged_from_timer(struct xpc_partition *part)
drivers/misc/sgi-xp/xpc_partition.c
313
return __xpc_partition_disengaged(part, true);
drivers/misc/sgi-xp/xpc_partition.c
320
xpc_mark_partition_active(struct xpc_partition *part)
drivers/misc/sgi-xp/xpc_partition.c
325
dev_dbg(xpc_part, "setting partition %d to ACTIVE\n", XPC_PARTID(part));
drivers/misc/sgi-xp/xpc_partition.c
327
spin_lock_irqsave(&part->act_lock, irq_flags);
drivers/misc/sgi-xp/xpc_partition.c
328
if (part->act_state == XPC_P_AS_ACTIVATING) {
drivers/misc/sgi-xp/xpc_partition.c
329
part->act_state = XPC_P_AS_ACTIVE;
drivers/misc/sgi-xp/xpc_partition.c
332
DBUG_ON(part->reason == xpSuccess);
drivers/misc/sgi-xp/xpc_partition.c
333
ret = part->reason;
drivers/misc/sgi-xp/xpc_partition.c
335
spin_unlock_irqrestore(&part->act_lock, irq_flags);
drivers/misc/sgi-xp/xpc_partition.c
344
xpc_deactivate_partition(const int line, struct xpc_partition *part,
drivers/misc/sgi-xp/xpc_partition.c
349
spin_lock_irqsave(&part->act_lock, irq_flags);
drivers/misc/sgi-xp/xpc_partition.c
351
if (part->act_state == XPC_P_AS_INACTIVE) {
drivers/misc/sgi-xp/xpc_partition.c
352
XPC_SET_REASON(part, reason, line);
drivers/misc/sgi-xp/xpc_partition.c
353
spin_unlock_irqrestore(&part->act_lock, irq_flags);
drivers/misc/sgi-xp/xpc_partition.c
356
xpc_arch_ops.request_partition_reactivation(part);
drivers/misc/sgi-xp/xpc_partition.c
360
if (part->act_state == XPC_P_AS_DEACTIVATING) {
drivers/misc/sgi-xp/xpc_partition.c
361
if ((part->reason == xpUnloading && reason != xpUnloading) ||
drivers/misc/sgi-xp/xpc_partition.c
363
XPC_SET_REASON(part, reason, line);
drivers/misc/sgi-xp/xpc_partition.c
365
spin_unlock_irqrestore(&part->act_lock, irq_flags);
drivers/misc/sgi-xp/xpc_partition.c
369
part->act_state = XPC_P_AS_DEACTIVATING;
drivers/misc/sgi-xp/xpc_partition.c
370
XPC_SET_REASON(part, reason, line);
drivers/misc/sgi-xp/xpc_partition.c
372
spin_unlock_irqrestore(&part->act_lock, irq_flags);
drivers/misc/sgi-xp/xpc_partition.c
375
xpc_arch_ops.request_partition_deactivation(part);
drivers/misc/sgi-xp/xpc_partition.c
378
part->disengage_timeout = jiffies + (xpc_disengage_timelimit * HZ);
drivers/misc/sgi-xp/xpc_partition.c
379
part->disengage_timer.expires = part->disengage_timeout;
drivers/misc/sgi-xp/xpc_partition.c
380
add_timer(&part->disengage_timer);
drivers/misc/sgi-xp/xpc_partition.c
383
XPC_PARTID(part), reason);
drivers/misc/sgi-xp/xpc_partition.c
385
xpc_partition_going_down(part, reason);
drivers/misc/sgi-xp/xpc_partition.c
392
xpc_mark_partition_inactive(struct xpc_partition *part)
drivers/misc/sgi-xp/xpc_partition.c
397
XPC_PARTID(part));
drivers/misc/sgi-xp/xpc_partition.c
399
spin_lock_irqsave(&part->act_lock, irq_flags);
drivers/misc/sgi-xp/xpc_partition.c
400
part->act_state = XPC_P_AS_INACTIVE;
drivers/misc/sgi-xp/xpc_partition.c
401
spin_unlock_irqrestore(&part->act_lock, irq_flags);
drivers/misc/sgi-xp/xpc_partition.c
402
part->remote_rp_pa = 0;
drivers/misc/sgi-xp/xpc_partition.c
532
struct xpc_partition *part;
drivers/misc/sgi-xp/xpc_partition.c
535
part = &xpc_partitions[partid];
drivers/misc/sgi-xp/xpc_partition.c
536
if (part->remote_rp_pa == 0)
drivers/misc/sgi-xp/xpc_partition.c
541
part_nasid_pa = (unsigned long)XPC_RP_PART_NASIDS(part->remote_rp_pa);
drivers/misc/sgi-xp/xpc_uv.c
1171
xpc_send_chctl_local_msgrequest_uv(struct xpc_partition *part, int ch_number)
drivers/misc/sgi-xp/xpc_uv.c
1175
spin_lock_irqsave(&part->chctl_lock, irq_flags);
drivers/misc/sgi-xp/xpc_uv.c
1176
part->chctl.flags[ch_number] |= XPC_CHCTL_MSGREQUEST;
drivers/misc/sgi-xp/xpc_uv.c
1177
spin_unlock_irqrestore(&part->chctl_lock, irq_flags);
drivers/misc/sgi-xp/xpc_uv.c
1179
xpc_wakeup_channel_mgr(part);
drivers/misc/sgi-xp/xpc_uv.c
1194
xpc_indicate_partition_engaged_uv(struct xpc_partition *part)
drivers/misc/sgi-xp/xpc_uv.c
1198
xpc_send_activate_IRQ_part_uv(part, &msg, sizeof(msg),
drivers/misc/sgi-xp/xpc_uv.c
1203
xpc_indicate_partition_disengaged_uv(struct xpc_partition *part)
drivers/misc/sgi-xp/xpc_uv.c
1207
xpc_send_activate_IRQ_part_uv(part, &msg, sizeof(msg),
drivers/misc/sgi-xp/xpc_uv.c
1321
xpc_handle_notify_mq_msg_uv(struct xpc_partition *part,
drivers/misc/sgi-xp/xpc_uv.c
1324
struct xpc_partition_uv *part_uv = &part->sn.uv;
drivers/misc/sgi-xp/xpc_uv.c
1331
if (unlikely(ch_number >= part->nchannels)) {
drivers/misc/sgi-xp/xpc_uv.c
1334
ch_number, XPC_PARTID(part));
drivers/misc/sgi-xp/xpc_uv.c
1348
ch = &part->channels[ch_number];
drivers/misc/sgi-xp/xpc_uv.c
1384
xpc_send_chctl_local_msgrequest_uv(part, ch->number);
drivers/misc/sgi-xp/xpc_uv.c
1394
struct xpc_partition *part;
drivers/misc/sgi-xp/xpc_uv.c
1404
part = &xpc_partitions[partid];
drivers/misc/sgi-xp/xpc_uv.c
1406
if (xpc_part_ref(part)) {
drivers/misc/sgi-xp/xpc_uv.c
1407
xpc_handle_notify_mq_msg_uv(part, msg);
drivers/misc/sgi-xp/xpc_uv.c
1408
xpc_part_deref(part);
drivers/misc/sgi-xp/xpc_uv.c
1425
xpc_process_msg_chctl_flags_uv(struct xpc_partition *part, int ch_number)
drivers/misc/sgi-xp/xpc_uv.c
1427
struct xpc_channel *ch = &part->channels[ch_number];
drivers/misc/sgi-xp/xpc_uv.c
304
struct xpc_partition *part;
drivers/misc/sgi-xp/xpc_uv.c
311
part = &xpc_partitions[partid];
drivers/misc/sgi-xp/xpc_uv.c
313
if (part->sn.uv.act_state_req == 0)
drivers/misc/sgi-xp/xpc_uv.c
319
act_state_req = part->sn.uv.act_state_req;
drivers/misc/sgi-xp/xpc_uv.c
320
part->sn.uv.act_state_req = 0;
drivers/misc/sgi-xp/xpc_uv.c
324
if (part->act_state == XPC_P_AS_INACTIVE)
drivers/misc/sgi-xp/xpc_uv.c
325
xpc_activate_partition(part);
drivers/misc/sgi-xp/xpc_uv.c
326
else if (part->act_state == XPC_P_AS_DEACTIVATING)
drivers/misc/sgi-xp/xpc_uv.c
327
XPC_DEACTIVATE_PARTITION(part, xpReactivating);
drivers/misc/sgi-xp/xpc_uv.c
330
if (part->act_state == XPC_P_AS_INACTIVE)
drivers/misc/sgi-xp/xpc_uv.c
331
xpc_activate_partition(part);
drivers/misc/sgi-xp/xpc_uv.c
333
XPC_DEACTIVATE_PARTITION(part, xpReactivating);
drivers/misc/sgi-xp/xpc_uv.c
336
XPC_DEACTIVATE_PARTITION(part, part->sn.uv.reason);
drivers/misc/sgi-xp/xpc_uv.c
351
xpc_handle_activate_mq_msg_uv(struct xpc_partition *part,
drivers/misc/sgi-xp/xpc_uv.c
357
struct xpc_partition_uv *part_uv = &part->sn.uv;
drivers/misc/sgi-xp/xpc_uv.c
382
part->remote_rp_pa = msg->rp_gpa; /* !!! _pa is _gpa */
drivers/misc/sgi-xp/xpc_uv.c
383
part->remote_rp_ts_jiffies = msg_hdr->rp_ts_jiffies;
drivers/misc/sgi-xp/xpc_uv.c
424
args = &part->remote_openclose_args[msg->ch_number];
drivers/misc/sgi-xp/xpc_uv.c
427
spin_lock_irqsave(&part->chctl_lock, irq_flags);
drivers/misc/sgi-xp/xpc_uv.c
428
part->chctl.flags[msg->ch_number] |= XPC_CHCTL_CLOSEREQUEST;
drivers/misc/sgi-xp/xpc_uv.c
429
spin_unlock_irqrestore(&part->chctl_lock, irq_flags);
drivers/misc/sgi-xp/xpc_uv.c
431
xpc_wakeup_channel_mgr(part);
drivers/misc/sgi-xp/xpc_uv.c
444
spin_lock_irqsave(&part->chctl_lock, irq_flags);
drivers/misc/sgi-xp/xpc_uv.c
445
part->chctl.flags[msg->ch_number] |= XPC_CHCTL_CLOSEREPLY;
drivers/misc/sgi-xp/xpc_uv.c
446
spin_unlock_irqrestore(&part->chctl_lock, irq_flags);
drivers/misc/sgi-xp/xpc_uv.c
448
xpc_wakeup_channel_mgr(part);
drivers/misc/sgi-xp/xpc_uv.c
460
args = &part->remote_openclose_args[msg->ch_number];
drivers/misc/sgi-xp/xpc_uv.c
464
spin_lock_irqsave(&part->chctl_lock, irq_flags);
drivers/misc/sgi-xp/xpc_uv.c
465
part->chctl.flags[msg->ch_number] |= XPC_CHCTL_OPENREQUEST;
drivers/misc/sgi-xp/xpc_uv.c
466
spin_unlock_irqrestore(&part->chctl_lock, irq_flags);
drivers/misc/sgi-xp/xpc_uv.c
468
xpc_wakeup_channel_mgr(part);
drivers/misc/sgi-xp/xpc_uv.c
479
args = &part->remote_openclose_args[msg->ch_number];
drivers/misc/sgi-xp/xpc_uv.c
484
spin_lock_irqsave(&part->chctl_lock, irq_flags);
drivers/misc/sgi-xp/xpc_uv.c
485
part->chctl.flags[msg->ch_number] |= XPC_CHCTL_OPENREPLY;
drivers/misc/sgi-xp/xpc_uv.c
486
spin_unlock_irqrestore(&part->chctl_lock, irq_flags);
drivers/misc/sgi-xp/xpc_uv.c
488
xpc_wakeup_channel_mgr(part);
drivers/misc/sgi-xp/xpc_uv.c
499
spin_lock_irqsave(&part->chctl_lock, irq_flags);
drivers/misc/sgi-xp/xpc_uv.c
500
part->chctl.flags[msg->ch_number] |= XPC_CHCTL_OPENCOMPLETE;
drivers/misc/sgi-xp/xpc_uv.c
501
spin_unlock_irqrestore(&part->chctl_lock, irq_flags);
drivers/misc/sgi-xp/xpc_uv.c
503
xpc_wakeup_channel_mgr(part);
drivers/misc/sgi-xp/xpc_uv.c
520
"from partition=%d\n", msg_hdr->type, XPC_PARTID(part));
drivers/misc/sgi-xp/xpc_uv.c
534
if (msg_hdr->rp_ts_jiffies != part->remote_rp_ts_jiffies &&
drivers/misc/sgi-xp/xpc_uv.c
535
part->remote_rp_ts_jiffies != 0) {
drivers/misc/sgi-xp/xpc_uv.c
555
struct xpc_partition *part;
drivers/misc/sgi-xp/xpc_uv.c
570
part = &xpc_partitions[partid];
drivers/misc/sgi-xp/xpc_uv.c
572
part_referenced = xpc_part_ref(part);
drivers/misc/sgi-xp/xpc_uv.c
573
xpc_handle_activate_mq_msg_uv(part, msg_hdr,
drivers/misc/sgi-xp/xpc_uv.c
577
xpc_part_deref(part);
drivers/misc/sgi-xp/xpc_uv.c
604
xpc_send_activate_IRQ_uv(struct xpc_partition *part, void *msg, size_t msg_size,
drivers/misc/sgi-xp/xpc_uv.c
608
struct xpc_partition_uv *part_uv = &part->sn.uv;
drivers/misc/sgi-xp/xpc_uv.c
617
msg_hdr->act_state = part->act_state;
drivers/misc/sgi-xp/xpc_uv.c
659
xpc_send_activate_IRQ_part_uv(struct xpc_partition *part, void *msg,
drivers/misc/sgi-xp/xpc_uv.c
664
ret = xpc_send_activate_IRQ_uv(part, msg, msg_size, msg_type);
drivers/misc/sgi-xp/xpc_uv.c
666
XPC_DEACTIVATE_PARTITION(part, ret);
drivers/misc/sgi-xp/xpc_uv.c
673
struct xpc_partition *part = &xpc_partitions[ch->partid];
drivers/misc/sgi-xp/xpc_uv.c
676
ret = xpc_send_activate_IRQ_uv(part, msg, msg_size, msg_type);
drivers/misc/sgi-xp/xpc_uv.c
681
XPC_DEACTIVATE_PARTITION(part, ret);
drivers/misc/sgi-xp/xpc_uv.c
689
xpc_send_local_activate_IRQ_uv(struct xpc_partition *part, int act_state_req)
drivers/misc/sgi-xp/xpc_uv.c
692
struct xpc_partition_uv *part_uv = &part->sn.uv;
drivers/misc/sgi-xp/xpc_uv.c
788
xpc_get_remote_heartbeat_uv(struct xpc_partition *part)
drivers/misc/sgi-xp/xpc_uv.c
790
struct xpc_partition_uv *part_uv = &part->sn.uv;
drivers/misc/sgi-xp/xpc_uv.c
799
if (part_uv->cached_heartbeat.value == part->last_heartbeat &&
drivers/misc/sgi-xp/xpc_uv.c
804
part->last_heartbeat = part_uv->cached_heartbeat.value;
drivers/misc/sgi-xp/xpc_uv.c
814
struct xpc_partition *part = &xpc_partitions[partid];
drivers/misc/sgi-xp/xpc_uv.c
817
part->remote_rp_pa = remote_rp_gpa; /* !!! _pa here is really _gpa */
drivers/misc/sgi-xp/xpc_uv.c
818
part->remote_rp_ts_jiffies = remote_rp->ts_jiffies;
drivers/misc/sgi-xp/xpc_uv.c
819
part->sn.uv.heartbeat_gpa = remote_rp->sn.uv.heartbeat_gpa;
drivers/misc/sgi-xp/xpc_uv.c
820
part->sn.uv.activate_gru_mq_desc_gpa =
drivers/misc/sgi-xp/xpc_uv.c
827
if (part->sn.uv.remote_act_state == XPC_P_AS_INACTIVE) {
drivers/misc/sgi-xp/xpc_uv.c
832
xpc_send_activate_IRQ_part_uv(part, &msg, sizeof(msg),
drivers/misc/sgi-xp/xpc_uv.c
836
if (part->act_state == XPC_P_AS_INACTIVE)
drivers/misc/sgi-xp/xpc_uv.c
837
xpc_send_local_activate_IRQ_uv(part, XPC_P_ASR_ACTIVATE_UV);
drivers/misc/sgi-xp/xpc_uv.c
841
xpc_request_partition_reactivation_uv(struct xpc_partition *part)
drivers/misc/sgi-xp/xpc_uv.c
843
xpc_send_local_activate_IRQ_uv(part, XPC_P_ASR_ACTIVATE_UV);
drivers/misc/sgi-xp/xpc_uv.c
847
xpc_request_partition_deactivation_uv(struct xpc_partition *part)
drivers/misc/sgi-xp/xpc_uv.c
855
if (part->sn.uv.remote_act_state != XPC_P_AS_DEACTIVATING &&
drivers/misc/sgi-xp/xpc_uv.c
856
part->sn.uv.remote_act_state != XPC_P_AS_INACTIVE) {
drivers/misc/sgi-xp/xpc_uv.c
858
msg.reason = part->reason;
drivers/misc/sgi-xp/xpc_uv.c
859
xpc_send_activate_IRQ_part_uv(part, &msg, sizeof(msg),
drivers/misc/sgi-xp/xpc_uv.c
865
xpc_cancel_partition_deactivation_request_uv(struct xpc_partition *part)
drivers/misc/sgi-xp/xpc_uv.c
929
xpc_setup_ch_structures_uv(struct xpc_partition *part)
drivers/misc/sgi-xp/xpc_uv.c
934
for (ch_number = 0; ch_number < part->nchannels; ch_number++) {
drivers/misc/sgi-xp/xpc_uv.c
935
ch_uv = &part->channels[ch_number].sn.uv;
drivers/misc/sgi-xp/xpc_uv.c
948
xpc_teardown_ch_structures_uv(struct xpc_partition *part)
drivers/misc/sgi-xp/xpc_uv.c
955
xpc_make_first_contact_uv(struct xpc_partition *part)
drivers/misc/sgi-xp/xpc_uv.c
964
xpc_send_activate_IRQ_part_uv(part, &msg, sizeof(msg),
drivers/misc/sgi-xp/xpc_uv.c
967
while (!((part->sn.uv.remote_act_state == XPC_P_AS_ACTIVATING) ||
drivers/misc/sgi-xp/xpc_uv.c
968
(part->sn.uv.remote_act_state == XPC_P_AS_ACTIVE))) {
drivers/misc/sgi-xp/xpc_uv.c
971
"partition %d\n", XPC_PARTID(part));
drivers/misc/sgi-xp/xpc_uv.c
976
if (part->act_state == XPC_P_AS_DEACTIVATING)
drivers/misc/sgi-xp/xpc_uv.c
977
return part->reason;
drivers/misc/sgi-xp/xpc_uv.c
984
xpc_get_chctl_all_flags_uv(struct xpc_partition *part)
drivers/misc/sgi-xp/xpc_uv.c
989
spin_lock_irqsave(&part->chctl_lock, irq_flags);
drivers/misc/sgi-xp/xpc_uv.c
990
chctl = part->chctl;
drivers/misc/sgi-xp/xpc_uv.c
992
part->chctl.all_flags = 0;
drivers/misc/sgi-xp/xpc_uv.c
994
spin_unlock_irqrestore(&part->chctl_lock, irq_flags);
drivers/misc/sram-exec.c
113
mutex_unlock(&part->lock);
drivers/misc/sram-exec.c
118
mutex_unlock(&part->lock);
drivers/misc/sram-exec.c
23
struct sram_partition *part)
drivers/misc/sram-exec.c
25
unsigned long base = (unsigned long)part->base;
drivers/misc/sram-exec.c
37
int sram_add_protect_exec(struct sram_partition *part)
drivers/misc/sram-exec.c
40
list_add_tail(&part->list, &exec_pool_list);
drivers/misc/sram-exec.c
76
struct sram_partition *part = NULL, *p;
drivers/misc/sram-exec.c
85
part = p;
drivers/misc/sram-exec.c
89
if (!part)
drivers/misc/sram-exec.c
95
base = (unsigned long)part->base;
drivers/misc/sram-exec.c
98
mutex_lock(&part->lock);
drivers/misc/sram.c
101
mutex_init(&part->lock);
drivers/misc/sram.c
116
part->base = virt_base;
drivers/misc/sram.c
118
part->base = sram->virt_base + block->start;
drivers/misc/sram.c
122
ret = sram_add_pool(sram, block, start, part);
drivers/misc/sram.c
127
ret = sram_add_export(sram, block, start, part);
drivers/misc/sram.c
132
ret = sram_check_protect_exec(sram, block, part);
drivers/misc/sram.c
136
ret = sram_add_pool(sram, block, start, part);
drivers/misc/sram.c
140
sram_add_protect_exec(part);
drivers/misc/sram.c
150
struct sram_partition *part;
drivers/misc/sram.c
155
part = &sram->partition[sram->partitions - 1];
drivers/misc/sram.c
156
for (; sram->partitions; sram->partitions--, part--) {
drivers/misc/sram.c
157
if (part->battr.size)
drivers/misc/sram.c
158
device_remove_bin_file(sram->dev, &part->battr);
drivers/misc/sram.c
160
if (part->pool &&
drivers/misc/sram.c
161
gen_pool_avail(part->pool) < gen_pool_size(part->pool))
drivers/misc/sram.c
29
struct sram_partition *part;
drivers/misc/sram.c
32
part = (struct sram_partition *)container_of(attr, struct sram_partition, battr);
drivers/misc/sram.c
34
mutex_lock(&part->lock);
drivers/misc/sram.c
35
memcpy_fromio(buf, part->base + pos, count);
drivers/misc/sram.c
36
mutex_unlock(&part->lock);
drivers/misc/sram.c
45
struct sram_partition *part;
drivers/misc/sram.c
48
part = (struct sram_partition *)container_of(attr, struct sram_partition, battr);
drivers/misc/sram.c
50
mutex_lock(&part->lock);
drivers/misc/sram.c
51
memcpy_toio(part->base + pos, buf, count);
drivers/misc/sram.c
52
mutex_unlock(&part->lock);
drivers/misc/sram.c
58
phys_addr_t start, struct sram_partition *part)
drivers/misc/sram.c
62
part->pool = devm_gen_pool_create(sram->dev, ilog2(SRAM_GRANULARITY),
drivers/misc/sram.c
64
if (IS_ERR(part->pool))
drivers/misc/sram.c
65
return PTR_ERR(part->pool);
drivers/misc/sram.c
67
ret = gen_pool_add_virt(part->pool, (unsigned long)part->base, start,
drivers/misc/sram.c
78
phys_addr_t start, struct sram_partition *part)
drivers/misc/sram.c
80
sysfs_bin_attr_init(&part->battr);
drivers/misc/sram.c
81
part->battr.attr.name = devm_kasprintf(sram->dev, GFP_KERNEL,
drivers/misc/sram.c
84
if (!part->battr.attr.name)
drivers/misc/sram.c
87
part->battr.attr.mode = S_IRUSR | S_IWUSR;
drivers/misc/sram.c
88
part->battr.read = sram_read;
drivers/misc/sram.c
89
part->battr.write = sram_write;
drivers/misc/sram.c
90
part->battr.size = block->size;
drivers/misc/sram.c
92
return device_create_bin_file(sram->dev, &part->battr);
drivers/misc/sram.c
99
struct sram_partition *part = &sram->partition[sram->partitions];
drivers/misc/sram.h
48
struct sram_partition *part);
drivers/misc/sram.h
49
int sram_add_protect_exec(struct sram_partition *part);
drivers/misc/sram.h
53
struct sram_partition *part)
drivers/misc/sram.h
58
static inline int sram_add_protect_exec(struct sram_partition *part)
drivers/mmc/core/block.c
118
struct list_head part;
drivers/mmc/core/block.c
2597
INIT_LIST_HEAD(&md->part);
drivers/mmc/core/block.c
2694
list_add(&part_md->part, &md->part);
drivers/mmc/core/block.c
288
list_for_each_entry(part_md, &md->part, part)
drivers/mmc/core/block.c
3011
if (card->part[idx].area_type & MMC_BLK_DATA_AREA_RPMB) {
drivers/mmc/core/block.c
3019
card->part[idx].part_cfg,
drivers/mmc/core/block.c
3020
card->part[idx].size >> 9,
drivers/mmc/core/block.c
3021
card->part[idx].name);
drivers/mmc/core/block.c
3024
} else if (card->part[idx].size) {
drivers/mmc/core/block.c
3026
card->part[idx].part_cfg,
drivers/mmc/core/block.c
3027
card->part[idx].size >> 9,
drivers/mmc/core/block.c
3028
card->part[idx].force_ro,
drivers/mmc/core/block.c
3029
card->part[idx].name,
drivers/mmc/core/block.c
3030
card->part[idx].area_type);
drivers/mmc/core/block.c
3064
list_for_each_safe(pos, q, &md->part) {
drivers/mmc/core/block.c
3065
part_md = list_entry(pos, struct mmc_blk_data, part);
drivers/mmc/core/block.c
3345
list_for_each_entry(part_md, &md->part, part) {
drivers/mmc/core/block.c
3377
list_for_each_entry(part_md, &md->part, part) {
drivers/mmc/core/mmc.c
322
card->part[card->nr_parts].size = size;
drivers/mmc/core/mmc.c
323
card->part[card->nr_parts].part_cfg = part_cfg;
drivers/mmc/core/mmc.c
324
sprintf(card->part[card->nr_parts].name, name, idx);
drivers/mmc/core/mmc.c
325
card->part[card->nr_parts].force_ro = ro;
drivers/mmc/core/mmc.c
326
card->part[card->nr_parts].area_type = area_type;
drivers/mtd/ftl.c
1000
part->bam_cache = NULL;
drivers/mtd/ftl.c
151
static int scan_header(partition_t *part)
drivers/mtd/ftl.c
157
part->header.FormattedSize = 0;
drivers/mtd/ftl.c
158
max_offset = (0x100000<part->mbd.mtd->size)?0x100000:part->mbd.mtd->size;
drivers/mtd/ftl.c
162
offset += part->mbd.mtd->erasesize ? : 0x2000) {
drivers/mtd/ftl.c
164
err = mtd_read(part->mbd.mtd, offset, sizeof(header), &ret,
drivers/mtd/ftl.c
183
if ((1 << header.EraseUnitSize) != part->mbd.mtd->erasesize) {
drivers/mtd/ftl.c
185
1 << header.EraseUnitSize,part->mbd.mtd->erasesize);
drivers/mtd/ftl.c
188
part->header = header;
drivers/mtd/ftl.c
192
static int build_maps(partition_t *part)
drivers/mtd/ftl.c
202
part->DataUnits = le16_to_cpu(part->header.NumEraseUnits) -
drivers/mtd/ftl.c
203
part->header.NumTransferUnits;
drivers/mtd/ftl.c
204
part->EUNInfo = kmalloc_objs(struct eun_info_t, part->DataUnits);
drivers/mtd/ftl.c
205
if (!part->EUNInfo)
drivers/mtd/ftl.c
207
for (i = 0; i < part->DataUnits; i++)
drivers/mtd/ftl.c
208
part->EUNInfo[i].Offset = 0xffffffff;
drivers/mtd/ftl.c
209
part->XferInfo =
drivers/mtd/ftl.c
210
kmalloc_objs(struct xfer_info_t, part->header.NumTransferUnits);
drivers/mtd/ftl.c
211
if (!part->XferInfo)
drivers/mtd/ftl.c
215
for (i = 0; i < le16_to_cpu(part->header.NumEraseUnits); i++) {
drivers/mtd/ftl.c
216
offset = ((i + le16_to_cpu(part->header.FirstPhysicalEUN))
drivers/mtd/ftl.c
217
<< part->header.EraseUnitSize);
drivers/mtd/ftl.c
218
ret = mtd_read(part->mbd.mtd, offset, sizeof(header), &retval,
drivers/mtd/ftl.c
227
if (hdr_ok && (le16_to_cpu(header.LogicalEUN) < part->DataUnits) &&
drivers/mtd/ftl.c
228
(part->EUNInfo[le16_to_cpu(header.LogicalEUN)].Offset == 0xffffffff)) {
drivers/mtd/ftl.c
229
part->EUNInfo[le16_to_cpu(header.LogicalEUN)].Offset = offset;
drivers/mtd/ftl.c
230
part->EUNInfo[le16_to_cpu(header.LogicalEUN)].EraseCount =
drivers/mtd/ftl.c
234
if (xtrans == part->header.NumTransferUnits) {
drivers/mtd/ftl.c
240
part->XferInfo[xtrans].state = XFER_PREPARED;
drivers/mtd/ftl.c
241
part->XferInfo[xtrans].EraseCount = le32_to_cpu(header.EraseCount);
drivers/mtd/ftl.c
243
part->XferInfo[xtrans].state = XFER_UNKNOWN;
drivers/mtd/ftl.c
245
part->XferInfo[xtrans].EraseCount =
drivers/mtd/ftl.c
246
le32_to_cpu(part->header.EraseCount);
drivers/mtd/ftl.c
248
part->XferInfo[xtrans].Offset = offset;
drivers/mtd/ftl.c
253
header = part->header;
drivers/mtd/ftl.c
263
part->VirtualBlockMap = vmalloc_array(blocks, sizeof(uint32_t));
drivers/mtd/ftl.c
264
if (!part->VirtualBlockMap)
drivers/mtd/ftl.c
267
memset(part->VirtualBlockMap, 0xff, blocks * sizeof(uint32_t));
drivers/mtd/ftl.c
268
part->BlocksPerUnit = (1 << header.EraseUnitSize) >> header.BlockSize;
drivers/mtd/ftl.c
270
part->bam_cache = kmalloc_array(part->BlocksPerUnit, sizeof(uint32_t),
drivers/mtd/ftl.c
272
if (!part->bam_cache)
drivers/mtd/ftl.c
275
part->bam_index = 0xffff;
drivers/mtd/ftl.c
276
part->FreeTotal = 0;
drivers/mtd/ftl.c
278
for (i = 0; i < part->DataUnits; i++) {
drivers/mtd/ftl.c
279
part->EUNInfo[i].Free = 0;
drivers/mtd/ftl.c
280
part->EUNInfo[i].Deleted = 0;
drivers/mtd/ftl.c
281
offset = part->EUNInfo[i].Offset + le32_to_cpu(header.BAMOffset);
drivers/mtd/ftl.c
283
ret = mtd_read(part->mbd.mtd, offset,
drivers/mtd/ftl.c
284
part->BlocksPerUnit * sizeof(uint32_t), &retval,
drivers/mtd/ftl.c
285
(unsigned char *)part->bam_cache);
drivers/mtd/ftl.c
290
for (j = 0; j < part->BlocksPerUnit; j++) {
drivers/mtd/ftl.c
291
if (BLOCK_FREE(le32_to_cpu(part->bam_cache[j]))) {
drivers/mtd/ftl.c
292
part->EUNInfo[i].Free++;
drivers/mtd/ftl.c
293
part->FreeTotal++;
drivers/mtd/ftl.c
294
} else if ((BLOCK_TYPE(le32_to_cpu(part->bam_cache[j])) == BLOCK_DATA) &&
drivers/mtd/ftl.c
295
(BLOCK_NUMBER(le32_to_cpu(part->bam_cache[j])) < blocks))
drivers/mtd/ftl.c
296
part->VirtualBlockMap[BLOCK_NUMBER(le32_to_cpu(part->bam_cache[j]))] =
drivers/mtd/ftl.c
298
else if (BLOCK_DELETED(le32_to_cpu(part->bam_cache[j])))
drivers/mtd/ftl.c
299
part->EUNInfo[i].Deleted++;
drivers/mtd/ftl.c
307
kfree(part->bam_cache);
drivers/mtd/ftl.c
309
vfree(part->VirtualBlockMap);
drivers/mtd/ftl.c
311
kfree(part->XferInfo);
drivers/mtd/ftl.c
313
kfree(part->EUNInfo);
drivers/mtd/ftl.c
325
static int erase_xfer(partition_t *part,
drivers/mtd/ftl.c
332
xfer = &part->XferInfo[xfernum];
drivers/mtd/ftl.c
344
erase->len = 1ULL << part->header.EraseUnitSize;
drivers/mtd/ftl.c
346
ret = mtd_erase(part->mbd.mtd, erase);
drivers/mtd/ftl.c
367
static int prepare_xfer(partition_t *part, int i)
drivers/mtd/ftl.c
376
xfer = &part->XferInfo[i];
drivers/mtd/ftl.c
382
header = part->header;
drivers/mtd/ftl.c
386
ret = mtd_write(part->mbd.mtd, xfer->Offset, sizeof(header), &retlen,
drivers/mtd/ftl.c
394
nbam = DIV_ROUND_UP(part->BlocksPerUnit * sizeof(uint32_t) +
drivers/mtd/ftl.c
395
le32_to_cpu(part->header.BAMOffset), SECTOR_SIZE);
drivers/mtd/ftl.c
397
offset = xfer->Offset + le32_to_cpu(part->header.BAMOffset);
drivers/mtd/ftl.c
402
ret = mtd_write(part->mbd.mtd, offset, sizeof(uint32_t), &retlen,
drivers/mtd/ftl.c
425
static int copy_erase_unit(partition_t *part, uint16_t srcunit,
drivers/mtd/ftl.c
438
eun = &part->EUNInfo[srcunit];
drivers/mtd/ftl.c
439
xfer = &part->XferInfo[xferunit];
drivers/mtd/ftl.c
445
if (part->bam_index != srcunit) {
drivers/mtd/ftl.c
447
offset = eun->Offset + le32_to_cpu(part->header.BAMOffset);
drivers/mtd/ftl.c
449
ret = mtd_read(part->mbd.mtd, offset,
drivers/mtd/ftl.c
450
part->BlocksPerUnit * sizeof(uint32_t), &retlen,
drivers/mtd/ftl.c
451
(u_char *)(part->bam_cache));
drivers/mtd/ftl.c
454
part->bam_index = 0xffff;
drivers/mtd/ftl.c
467
ret = mtd_write(part->mbd.mtd, offset, sizeof(uint16_t), &retlen,
drivers/mtd/ftl.c
480
for (i = 0; i < part->BlocksPerUnit; i++) {
drivers/mtd/ftl.c
481
switch (BLOCK_TYPE(le32_to_cpu(part->bam_cache[i]))) {
drivers/mtd/ftl.c
487
ret = mtd_read(part->mbd.mtd, src, SECTOR_SIZE, &retlen,
drivers/mtd/ftl.c
495
ret = mtd_write(part->mbd.mtd, dest, SECTOR_SIZE, &retlen,
drivers/mtd/ftl.c
505
part->bam_cache[i] = cpu_to_le32(0xffffffff);
drivers/mtd/ftl.c
514
ret = mtd_write(part->mbd.mtd,
drivers/mtd/ftl.c
515
xfer->Offset + le32_to_cpu(part->header.BAMOffset),
drivers/mtd/ftl.c
516
part->BlocksPerUnit * sizeof(int32_t),
drivers/mtd/ftl.c
518
(u_char *)part->bam_cache);
drivers/mtd/ftl.c
526
ret = mtd_write(part->mbd.mtd, xfer->Offset + 20, sizeof(uint16_t),
drivers/mtd/ftl.c
538
part->FreeTotal -= eun->Free;
drivers/mtd/ftl.c
539
part->FreeTotal += free;
drivers/mtd/ftl.c
544
part->bam_index = srcunit;
drivers/mtd/ftl.c
565
static int reclaim_block(partition_t *part)
drivers/mtd/ftl.c
572
pr_debug("NumTransferUnits == %x\n", part->header.NumTransferUnits);
drivers/mtd/ftl.c
577
for (i = 0; i < part->header.NumTransferUnits; i++) {
drivers/mtd/ftl.c
579
if (part->XferInfo[i].state == XFER_UNKNOWN) {
drivers/mtd/ftl.c
582
erase_xfer(part, i);
drivers/mtd/ftl.c
584
if (part->XferInfo[i].state == XFER_ERASING) {
drivers/mtd/ftl.c
589
else if (part->XferInfo[i].state == XFER_ERASED) {
drivers/mtd/ftl.c
592
prepare_xfer(part, i);
drivers/mtd/ftl.c
594
if (part->XferInfo[i].state == XFER_PREPARED) {
drivers/mtd/ftl.c
597
if (part->XferInfo[i].EraseCount <= best) {
drivers/mtd/ftl.c
598
best = part->XferInfo[i].EraseCount;
drivers/mtd/ftl.c
603
pr_debug("XferInfo[%d].state == %x\n",i, part->XferInfo[i].state);
drivers/mtd/ftl.c
610
mtd_sync(part->mbd.mtd);
drivers/mtd/ftl.c
629
for (i = 0; i < part->DataUnits; i++)
drivers/mtd/ftl.c
630
if (part->EUNInfo[i].EraseCount <= best) {
drivers/mtd/ftl.c
631
best = part->EUNInfo[i].EraseCount;
drivers/mtd/ftl.c
636
for (i = 0; i < part->DataUnits; i++)
drivers/mtd/ftl.c
637
if (part->EUNInfo[i].Deleted >= best) {
drivers/mtd/ftl.c
638
best = part->EUNInfo[i].Deleted;
drivers/mtd/ftl.c
653
ret = copy_erase_unit(part, eun, xfer);
drivers/mtd/ftl.c
655
erase_xfer(part, xfer);
drivers/mtd/ftl.c
672
static void dump_lists(partition_t *part)
drivers/mtd/ftl.c
675
printk(KERN_DEBUG "ftl_cs: Free total = %d\n", part->FreeTotal);
drivers/mtd/ftl.c
676
for (i = 0; i < part->DataUnits; i++)
drivers/mtd/ftl.c
679
part->EUNInfo[i].Offset >> part->header.EraseUnitSize,
drivers/mtd/ftl.c
680
part->EUNInfo[i].Free, part->EUNInfo[i].Deleted);
drivers/mtd/ftl.c
684
static uint32_t find_free(partition_t *part)
drivers/mtd/ftl.c
692
stop = (part->bam_index == 0xffff) ? 0 : part->bam_index;
drivers/mtd/ftl.c
695
if (part->EUNInfo[eun].Free != 0) break;
drivers/mtd/ftl.c
697
if (++eun == part->DataUnits) eun = 0;
drivers/mtd/ftl.c
700
if (part->EUNInfo[eun].Free == 0)
drivers/mtd/ftl.c
704
if (eun != part->bam_index) {
drivers/mtd/ftl.c
706
part->bam_index = 0xffff;
drivers/mtd/ftl.c
708
ret = mtd_read(part->mbd.mtd,
drivers/mtd/ftl.c
709
part->EUNInfo[eun].Offset + le32_to_cpu(part->header.BAMOffset),
drivers/mtd/ftl.c
710
part->BlocksPerUnit * sizeof(uint32_t),
drivers/mtd/ftl.c
712
(u_char *)(part->bam_cache));
drivers/mtd/ftl.c
718
part->bam_index = eun;
drivers/mtd/ftl.c
722
for (blk = 0; blk < part->BlocksPerUnit; blk++)
drivers/mtd/ftl.c
723
if (BLOCK_FREE(le32_to_cpu(part->bam_cache[blk]))) break;
drivers/mtd/ftl.c
724
if (blk == part->BlocksPerUnit) {
drivers/mtd/ftl.c
728
dump_lists(part);
drivers/mtd/ftl.c
745
static int ftl_read(partition_t *part, caddr_t buffer,
drivers/mtd/ftl.c
754
part, sector, nblocks);
drivers/mtd/ftl.c
755
if (!(part->state & FTL_FORMATTED)) {
drivers/mtd/ftl.c
759
bsize = 1 << part->header.EraseUnitSize;
drivers/mtd/ftl.c
762
if (((sector+i) * SECTOR_SIZE) >= le32_to_cpu(part->header.FormattedSize)) {
drivers/mtd/ftl.c
766
log_addr = part->VirtualBlockMap[sector+i];
drivers/mtd/ftl.c
770
offset = (part->EUNInfo[log_addr / bsize].Offset
drivers/mtd/ftl.c
772
ret = mtd_read(part->mbd.mtd, offset, SECTOR_SIZE, &retlen,
drivers/mtd/ftl.c
791
static int set_bam_entry(partition_t *part, uint32_t log_addr,
drivers/mtd/ftl.c
803
part, log_addr, virt_addr);
drivers/mtd/ftl.c
804
bsize = 1 << part->header.EraseUnitSize;
drivers/mtd/ftl.c
807
offset = (part->EUNInfo[eun].Offset + blk * sizeof(uint32_t) +
drivers/mtd/ftl.c
808
le32_to_cpu(part->header.BAMOffset));
drivers/mtd/ftl.c
811
ret = mtd_read(part->mbd.mtd, offset, sizeof(uint32_t), &retlen,
drivers/mtd/ftl.c
832
if (part->bam_index == eun) {
drivers/mtd/ftl.c
834
if (le32_to_cpu(part->bam_cache[blk]) != old_addr) {
drivers/mtd/ftl.c
841
le32_to_cpu(part->bam_cache[blk]), old_addr);
drivers/mtd/ftl.c
846
part->bam_cache[blk] = le_virt_addr;
drivers/mtd/ftl.c
848
ret = mtd_write(part->mbd.mtd, offset, sizeof(uint32_t), &retlen,
drivers/mtd/ftl.c
859
static int ftl_write(partition_t *part, caddr_t buffer,
drivers/mtd/ftl.c
868
part, sector, nblocks);
drivers/mtd/ftl.c
869
if (!(part->state & FTL_FORMATTED)) {
drivers/mtd/ftl.c
874
while (part->FreeTotal < nblocks) {
drivers/mtd/ftl.c
875
ret = reclaim_block(part);
drivers/mtd/ftl.c
880
bsize = 1 << part->header.EraseUnitSize;
drivers/mtd/ftl.c
884
if (virt_addr >= le32_to_cpu(part->header.FormattedSize)) {
drivers/mtd/ftl.c
890
blk = find_free(part);
drivers/mtd/ftl.c
900
log_addr = part->bam_index * bsize + blk * SECTOR_SIZE;
drivers/mtd/ftl.c
901
part->EUNInfo[part->bam_index].Free--;
drivers/mtd/ftl.c
902
part->FreeTotal--;
drivers/mtd/ftl.c
903
if (set_bam_entry(part, log_addr, 0xfffffffe))
drivers/mtd/ftl.c
905
part->EUNInfo[part->bam_index].Deleted++;
drivers/mtd/ftl.c
906
offset = (part->EUNInfo[part->bam_index].Offset +
drivers/mtd/ftl.c
908
ret = mtd_write(part->mbd.mtd, offset, SECTOR_SIZE, &retlen, buffer);
drivers/mtd/ftl.c
919
old_addr = part->VirtualBlockMap[sector+i];
drivers/mtd/ftl.c
921
part->VirtualBlockMap[sector+i] = 0xffffffff;
drivers/mtd/ftl.c
922
part->EUNInfo[old_addr/bsize].Deleted++;
drivers/mtd/ftl.c
923
if (set_bam_entry(part, old_addr, 0))
drivers/mtd/ftl.c
928
if (set_bam_entry(part, log_addr, virt_addr))
drivers/mtd/ftl.c
930
part->VirtualBlockMap[sector+i] = log_addr;
drivers/mtd/ftl.c
931
part->EUNInfo[part->bam_index].Deleted--;
drivers/mtd/ftl.c
941
partition_t *part = container_of(dev, struct partition_t, mbd);
drivers/mtd/ftl.c
945
sect = le32_to_cpu(part->header.FormattedSize)/SECTOR_SIZE;
drivers/mtd/ftl.c
969
partition_t *part = container_of(dev, struct partition_t, mbd);
drivers/mtd/ftl.c
970
uint32_t bsize = 1 << part->header.EraseUnitSize;
drivers/mtd/ftl.c
976
uint32_t old_addr = part->VirtualBlockMap[sector];
drivers/mtd/ftl.c
978
part->VirtualBlockMap[sector] = 0xffffffff;
drivers/mtd/ftl.c
979
part->EUNInfo[old_addr/bsize].Deleted++;
drivers/mtd/ftl.c
980
if (set_bam_entry(part, old_addr, 0))
drivers/mtd/ftl.c
991
static void ftl_freepart(partition_t *part)
drivers/mtd/ftl.c
993
vfree(part->VirtualBlockMap);
drivers/mtd/ftl.c
994
part->VirtualBlockMap = NULL;
drivers/mtd/ftl.c
995
kfree(part->EUNInfo);
drivers/mtd/ftl.c
996
part->EUNInfo = NULL;
drivers/mtd/ftl.c
997
kfree(part->XferInfo);
drivers/mtd/ftl.c
998
part->XferInfo = NULL;
drivers/mtd/ftl.c
999
kfree(part->bam_cache);
drivers/mtd/mtdpart.c
100
child->part.offset = cur_offset;
drivers/mtd/mtdpart.c
103
child->part.offset += wr_alignment - remainder;
drivers/mtd/mtdpart.c
107
child->part.offset);
drivers/mtd/mtdpart.c
110
if (child->part.offset == MTDPART_OFS_RETAIN) {
drivers/mtd/mtdpart.c
111
child->part.offset = cur_offset;
drivers/mtd/mtdpart.c
112
if (parent_size - child->part.offset >= child->part.size) {
drivers/mtd/mtdpart.c
113
child->part.size = parent_size - child->part.offset -
drivers/mtd/mtdpart.c
114
child->part.size;
drivers/mtd/mtdpart.c
117
part->name, parent_size - child->part.offset,
drivers/mtd/mtdpart.c
118
child->part.size);
drivers/mtd/mtdpart.c
123
if (child->part.size == MTDPART_SIZ_FULL)
drivers/mtd/mtdpart.c
124
child->part.size = parent_size - child->part.offset;
drivers/mtd/mtdpart.c
127
child->part.offset, child->part.offset + child->part.size,
drivers/mtd/mtdpart.c
131
if (child->part.offset >= parent_size) {
drivers/mtd/mtdpart.c
133
child->part.offset = 0;
drivers/mtd/mtdpart.c
134
child->part.size = 0;
drivers/mtd/mtdpart.c
139
part->name);
drivers/mtd/mtdpart.c
142
if (child->part.offset + child->part.size > parent->size) {
drivers/mtd/mtdpart.c
143
child->part.size = parent_size - child->part.offset;
drivers/mtd/mtdpart.c
145
part->name, parent->name, child->part.size);
drivers/mtd/mtdpart.c
151
u64 end = child->part.offset + child->part.size;
drivers/mtd/mtdpart.c
156
for (i = 0; i < max && regions[i].offset <= child->part.offset;
drivers/mtd/mtdpart.c
190
part->name);
drivers/mtd/mtdpart.c
193
tmp = mtd_get_master_ofs(child, 0) + child->part.size;
drivers/mtd/mtdpart.c
198
part->name);
drivers/mtd/mtdpart.c
201
child->size = child->part.size;
drivers/mtd/mtdpart.c
209
while (offs < child->part.size) {
drivers/mtd/mtdpart.c
227
return sysfs_emit(buf, "%lld\n", mtd->part.offset);
drivers/mtd/mtdpart.c
250
parent->part.size : parent->size;
drivers/mtd/mtdpart.c
251
struct mtd_partition part;
drivers/mtd/mtdpart.c
266
memset(&part, 0, sizeof(part));
drivers/mtd/mtdpart.c
267
part.name = name;
drivers/mtd/mtdpart.c
268
part.size = length;
drivers/mtd/mtdpart.c
269
part.offset = offset;
drivers/mtd/mtdpart.c
271
child = allocate_partition(parent, &part, -1, offset);
drivers/mtd/mtdpart.c
276
list_add_tail(&child->part.node, &parent->partitions);
drivers/mtd/mtdpart.c
289
list_del(&child->part.node);
drivers/mtd/mtdpart.c
310
list_for_each_entry_safe(child, next, &mtd->partitions, part.node) {
drivers/mtd/mtdpart.c
318
list_del_init(&mtd->part.node);
drivers/mtd/mtdpart.c
335
list_for_each_entry_safe(child, next, &mtd->partitions, part.node) {
drivers/mtd/mtdpart.c
340
list_del_init(&child->part.node);
drivers/mtd/mtdpart.c
37
WARN_ON(!list_empty(&mtd->part.node));
drivers/mtd/mtdpart.c
373
list_for_each_entry(child, &mtd->partitions, part.node) {
drivers/mtd/mtdpart.c
413
list_add_tail(&child->part.node, &parent->partitions);
drivers/mtd/mtdpart.c
419
list_del(&child->part.node);
drivers/mtd/mtdpart.c
42
const struct mtd_partition *part,
drivers/mtd/mtdpart.c
438
cur_offset = child->part.offset + child->part.size;
drivers/mtd/mtdpart.c
49
parent->part.size : parent->size;
drivers/mtd/mtdpart.c
57
name = kstrdup(part->name, GFP_KERNEL);
drivers/mtd/mtdpart.c
68
child->part.flags = parent->flags & ~part->mask_flags;
drivers/mtd/mtdpart.c
69
child->part.flags |= part->add_flags;
drivers/mtd/mtdpart.c
70
child->flags = child->part.flags;
drivers/mtd/mtdpart.c
71
child->part.size = part->size;
drivers/mtd/mtdpart.c
91
child->dev.of_node = part->of_node;
drivers/mtd/mtdpart.c
93
child->part.offset = part->offset;
drivers/mtd/mtdpart.c
96
if (child->part.offset == MTDPART_OFS_APPEND)
drivers/mtd/mtdpart.c
97
child->part.offset = cur_offset;
drivers/mtd/mtdpart.c
98
if (child->part.offset == MTDPART_OFS_NXTBLK) {
drivers/mtd/mtdswap.c
1338
unsigned long part;
drivers/mtd/mtdswap.c
1348
if (kstrtoul(this_opt, 0, &part) < 0)
drivers/mtd/mtdswap.c
1351
if (mtd->index == part)
drivers/mtd/mtdswap.c
1355
if (mtd->index != part)
drivers/mtd/mtdswap.c
1414
MTDSWAP_PREFIX, part, swap_size / 1024, spare_cnt, bad_blocks);
drivers/mtd/parsers/afs.c
120
u_int off, struct mtd_partition *part)
drivers/mtd/parsers/afs.c
201
part->name = kstrdup(iis.name, GFP_KERNEL);
drivers/mtd/parsers/afs.c
202
if (!part->name)
drivers/mtd/parsers/afs.c
205
part->size = (iis.length + mtd->erasesize - 1) & ~(mtd->erasesize - 1);
drivers/mtd/parsers/afs.c
206
part->offset = img_ptr;
drivers/mtd/parsers/afs.c
207
part->mask_flags = 0;
drivers/mtd/parsers/afs.c
210
img_ptr, part->size / 1024,
drivers/mtd/parsers/afs.c
211
iis.imageNumber, part->name);
drivers/mtd/parsers/afs.c
217
u_int off, struct mtd_partition *part)
drivers/mtd/parsers/afs.c
312
part->name = kstrdup(name, GFP_KERNEL);
drivers/mtd/parsers/afs.c
313
if (!part->name)
drivers/mtd/parsers/afs.c
315
part->offset = region_start;
drivers/mtd/parsers/afs.c
316
part->size = region_end - region_start;
drivers/mtd/parsers/afs.c
317
part->mask_flags = 0;
drivers/mtd/parsers/bcm47xxpart.c
54
static void bcm47xxpart_add_part(struct mtd_partition *part, const char *name,
drivers/mtd/parsers/bcm47xxpart.c
57
part->name = name;
drivers/mtd/parsers/bcm47xxpart.c
58
part->offset = offset;
drivers/mtd/parsers/bcm47xxpart.c
59
part->mask_flags = mask_flags;
drivers/mtd/parsers/cmdlinepart.c
331
struct cmdline_mtd_partition *part;
drivers/mtd/parsers/cmdlinepart.c
345
for (part = partitions; part; part = part->next) {
drivers/mtd/parsers/cmdlinepart.c
346
if ((!mtd_id) || (!strcmp(part->mtd_id, mtd_id)))
drivers/mtd/parsers/cmdlinepart.c
350
if (!part)
drivers/mtd/parsers/cmdlinepart.c
353
for (i = 0, offset = 0; i < part->num_parts; i++) {
drivers/mtd/parsers/cmdlinepart.c
354
if (part->parts[i].offset == OFFSET_CONTINUOUS)
drivers/mtd/parsers/cmdlinepart.c
355
part->parts[i].offset = offset;
drivers/mtd/parsers/cmdlinepart.c
357
offset = part->parts[i].offset;
drivers/mtd/parsers/cmdlinepart.c
359
if (part->parts[i].size == SIZE_REMAINING)
drivers/mtd/parsers/cmdlinepart.c
360
part->parts[i].size = master->size - offset;
drivers/mtd/parsers/cmdlinepart.c
362
if (offset + part->parts[i].size > master->size) {
drivers/mtd/parsers/cmdlinepart.c
364
part->mtd_id);
drivers/mtd/parsers/cmdlinepart.c
365
part->parts[i].size = master->size - offset;
drivers/mtd/parsers/cmdlinepart.c
367
offset += part->parts[i].size;
drivers/mtd/parsers/cmdlinepart.c
369
if (part->parts[i].size == 0) {
drivers/mtd/parsers/cmdlinepart.c
371
part->mtd_id);
drivers/mtd/parsers/cmdlinepart.c
372
part->num_parts--;
drivers/mtd/parsers/cmdlinepart.c
373
memmove(&part->parts[i], &part->parts[i + 1],
drivers/mtd/parsers/cmdlinepart.c
374
sizeof(*part->parts) * (part->num_parts - i));
drivers/mtd/parsers/cmdlinepart.c
379
*pparts = kmemdup(part->parts, sizeof(*part->parts) * part->num_parts,
drivers/mtd/parsers/cmdlinepart.c
384
return part->num_parts;
drivers/mtd/parsers/ofpart_core.c
236
} *part;
drivers/mtd/parsers/ofpart_core.c
244
part = of_get_property(dp, "partitions", &plen);
drivers/mtd/parsers/ofpart_core.c
245
if (!part)
drivers/mtd/parsers/ofpart_core.c
250
nr_parts = plen / sizeof(part[0]);
drivers/mtd/parsers/ofpart_core.c
259
parts[i].offset = be32_to_cpu(part->offset);
drivers/mtd/parsers/ofpart_core.c
260
parts[i].size = be32_to_cpu(part->len) & ~1;
drivers/mtd/parsers/ofpart_core.c
262
if (be32_to_cpu(part->len) & 1)
drivers/mtd/parsers/ofpart_core.c
275
part++;
drivers/mtd/parsers/parser_trx.c
100
part = &parts[curr_part++];
drivers/mtd/parsers/parser_trx.c
101
part->name = parser_trx_data_part_name(mtd, trx.offset[i]);
drivers/mtd/parsers/parser_trx.c
102
part->offset = trx.offset[i];
drivers/mtd/parsers/parser_trx.c
56
struct mtd_partition *part;
drivers/mtd/parsers/parser_trx.c
86
part = &parts[curr_part++];
drivers/mtd/parsers/parser_trx.c
87
part->name = "loader";
drivers/mtd/parsers/parser_trx.c
88
part->offset = trx.offset[i];
drivers/mtd/parsers/parser_trx.c
93
part = &parts[curr_part++];
drivers/mtd/parsers/parser_trx.c
94
part->name = "linux";
drivers/mtd/parsers/parser_trx.c
95
part->offset = trx.offset[i];
drivers/mtd/rfd_ftl.c
100
if (le16_to_cpu(part->header_cache[0]) != RFD_MAGIC) {
drivers/mtd/rfd_ftl.c
107
for (i=0; i<part->data_sectors_per_block; i++) {
drivers/mtd/rfd_ftl.c
110
entry = le16_to_cpu(part->header_cache[HEADER_MAP_OFFSET + i]);
drivers/mtd/rfd_ftl.c
123
if (entry >= part->sector_count) {
drivers/mtd/rfd_ftl.c
127
part->mbd.mtd->name, block_no, i, entry);
drivers/mtd/rfd_ftl.c
131
if (part->sector_map[entry] != -1) {
drivers/mtd/rfd_ftl.c
134
part->mbd.mtd->name, entry);
drivers/mtd/rfd_ftl.c
135
part->errors = 1;
drivers/mtd/rfd_ftl.c
139
part->sector_map[entry] = block->offset +
drivers/mtd/rfd_ftl.c
140
(i + part->header_sectors_per_block) * SECTOR_SIZE;
drivers/mtd/rfd_ftl.c
145
if (block->free_sectors == part->data_sectors_per_block)
drivers/mtd/rfd_ftl.c
146
part->reserved_block = block_no;
drivers/mtd/rfd_ftl.c
151
static int scan_header(struct partition *part)
drivers/mtd/rfd_ftl.c
158
sectors_per_block = part->block_size / SECTOR_SIZE;
drivers/mtd/rfd_ftl.c
159
part->total_blocks = (u32)part->mbd.mtd->size / part->block_size;
drivers/mtd/rfd_ftl.c
161
if (part->total_blocks < 2)
drivers/mtd/rfd_ftl.c
165
part->header_sectors_per_block =
drivers/mtd/rfd_ftl.c
169
part->data_sectors_per_block = sectors_per_block -
drivers/mtd/rfd_ftl.c
170
part->header_sectors_per_block;
drivers/mtd/rfd_ftl.c
172
part->header_size = (HEADER_MAP_OFFSET +
drivers/mtd/rfd_ftl.c
173
part->data_sectors_per_block) * sizeof(u16);
drivers/mtd/rfd_ftl.c
175
part->cylinders = (part->data_sectors_per_block *
drivers/mtd/rfd_ftl.c
176
(part->total_blocks - 1) - 1) / SECTORS_PER_TRACK;
drivers/mtd/rfd_ftl.c
178
part->sector_count = part->cylinders * SECTORS_PER_TRACK;
drivers/mtd/rfd_ftl.c
180
part->current_block = -1;
drivers/mtd/rfd_ftl.c
181
part->reserved_block = -1;
drivers/mtd/rfd_ftl.c
182
part->is_reclaiming = 0;
drivers/mtd/rfd_ftl.c
184
part->header_cache = kmalloc(part->header_size, GFP_KERNEL);
drivers/mtd/rfd_ftl.c
185
if (!part->header_cache)
drivers/mtd/rfd_ftl.c
188
part->blocks = kzalloc_objs(struct block, part->total_blocks);
drivers/mtd/rfd_ftl.c
189
if (!part->blocks)
drivers/mtd/rfd_ftl.c
192
part->sector_map = vmalloc_array(part->sector_count,
drivers/mtd/rfd_ftl.c
194
if (!part->sector_map)
drivers/mtd/rfd_ftl.c
197
for (i=0; i<part->sector_count; i++)
drivers/mtd/rfd_ftl.c
198
part->sector_map[i] = -1;
drivers/mtd/rfd_ftl.c
200
for (i=0, blocks_found=0; i<part->total_blocks; i++) {
drivers/mtd/rfd_ftl.c
201
rc = mtd_read(part->mbd.mtd, i * part->block_size,
drivers/mtd/rfd_ftl.c
202
part->header_size, &retlen,
drivers/mtd/rfd_ftl.c
203
(u_char *)part->header_cache);
drivers/mtd/rfd_ftl.c
205
if (!rc && retlen != part->header_size)
drivers/mtd/rfd_ftl.c
211
if (!build_block_map(part, i))
drivers/mtd/rfd_ftl.c
217
part->mbd.mtd->name);
drivers/mtd/rfd_ftl.c
222
if (part->reserved_block == -1) {
drivers/mtd/rfd_ftl.c
224
part->mbd.mtd->name);
drivers/mtd/rfd_ftl.c
226
part->errors = 1;
drivers/mtd/rfd_ftl.c
232
vfree(part->sector_map);
drivers/mtd/rfd_ftl.c
233
kfree(part->header_cache);
drivers/mtd/rfd_ftl.c
234
kfree(part->blocks);
drivers/mtd/rfd_ftl.c
241
struct partition *part = container_of(dev, struct partition, mbd);
drivers/mtd/rfd_ftl.c
246
if (sector >= part->sector_count)
drivers/mtd/rfd_ftl.c
249
addr = part->sector_map[sector];
drivers/mtd/rfd_ftl.c
251
rc = mtd_read(part->mbd.mtd, addr, SECTOR_SIZE, &retlen,
drivers/mtd/rfd_ftl.c
258
"0x%lx\n", part->mbd.mtd->name, addr);
drivers/mtd/rfd_ftl.c
267
static int erase_block(struct partition *part, int block)
drivers/mtd/rfd_ftl.c
276
erase->addr = part->blocks[block].offset;
drivers/mtd/rfd_ftl.c
277
erase->len = part->block_size;
drivers/mtd/rfd_ftl.c
279
part->blocks[block].state = BLOCK_ERASING;
drivers/mtd/rfd_ftl.c
280
part->blocks[block].free_sectors = 0;
drivers/mtd/rfd_ftl.c
282
rc = mtd_erase(part->mbd.mtd, erase);
drivers/mtd/rfd_ftl.c
286
(unsigned long long)erase->len, part->mbd.mtd->name);
drivers/mtd/rfd_ftl.c
287
part->blocks[block].state = BLOCK_FAILED;
drivers/mtd/rfd_ftl.c
288
part->blocks[block].free_sectors = 0;
drivers/mtd/rfd_ftl.c
289
part->blocks[block].used_sectors = 0;
drivers/mtd/rfd_ftl.c
294
part->blocks[block].state = BLOCK_ERASED;
drivers/mtd/rfd_ftl.c
295
part->blocks[block].free_sectors = part->data_sectors_per_block;
drivers/mtd/rfd_ftl.c
296
part->blocks[block].used_sectors = 0;
drivers/mtd/rfd_ftl.c
297
part->blocks[block].erases++;
drivers/mtd/rfd_ftl.c
299
rc = mtd_write(part->mbd.mtd, part->blocks[block].offset,
drivers/mtd/rfd_ftl.c
306
part->mbd.mtd->name, part->blocks[block].offset);
drivers/mtd/rfd_ftl.c
307
part->blocks[block].state = BLOCK_FAILED;
drivers/mtd/rfd_ftl.c
309
part->blocks[block].state = BLOCK_OK;
drivers/mtd/rfd_ftl.c
318
static int move_block_contents(struct partition *part, int block_no, u_long *old_sector)
drivers/mtd/rfd_ftl.c
325
part->is_reclaiming = 1;
drivers/mtd/rfd_ftl.c
331
map = kmalloc(part->header_size, GFP_KERNEL);
drivers/mtd/rfd_ftl.c
335
rc = mtd_read(part->mbd.mtd, part->blocks[block_no].offset,
drivers/mtd/rfd_ftl.c
336
part->header_size, &retlen, (u_char *)map);
drivers/mtd/rfd_ftl.c
338
if (!rc && retlen != part->header_size)
drivers/mtd/rfd_ftl.c
343
"0x%lx\n", part->mbd.mtd->name,
drivers/mtd/rfd_ftl.c
344
part->blocks[block_no].offset);
drivers/mtd/rfd_ftl.c
349
for (i=0; i<part->data_sectors_per_block; i++) {
drivers/mtd/rfd_ftl.c
361
if (entry >= part->sector_count)
drivers/mtd/rfd_ftl.c
364
addr = part->blocks[block_no].offset +
drivers/mtd/rfd_ftl.c
365
(i + part->header_sectors_per_block) * SECTOR_SIZE;
drivers/mtd/rfd_ftl.c
369
if (!part->blocks[block_no].used_sectors--) {
drivers/mtd/rfd_ftl.c
370
rc = erase_block(part, block_no);
drivers/mtd/rfd_ftl.c
375
rc = mtd_read(part->mbd.mtd, addr, SECTOR_SIZE, &retlen,
drivers/mtd/rfd_ftl.c
384
part->mbd.mtd->name);
drivers/mtd/rfd_ftl.c
389
rc = rfd_ftl_writesect((struct mtd_blktrans_dev*)part,
drivers/mtd/rfd_ftl.c
401
part->is_reclaiming = 0;
drivers/mtd/rfd_ftl.c
406
static int reclaim_block(struct partition *part, u_long *old_sector)
drivers/mtd/rfd_ftl.c
412
mtd_sync(part->mbd.mtd);
drivers/mtd/rfd_ftl.c
417
old_sector_block = *old_sector / part->block_size;
drivers/mtd/rfd_ftl.c
421
for (block=0; block<part->total_blocks; block++) {
drivers/mtd/rfd_ftl.c
424
if (block == part->reserved_block)
drivers/mtd/rfd_ftl.c
432
if (part->blocks[block].free_sectors)
drivers/mtd/rfd_ftl.c
435
this_score = part->blocks[block].used_sectors;
drivers/mtd/rfd_ftl.c
441
if (part->blocks[block].used_sectors ==
drivers/mtd/rfd_ftl.c
442
part->data_sectors_per_block)
drivers/mtd/rfd_ftl.c
446
this_score += part->blocks[block].erases;
drivers/mtd/rfd_ftl.c
457
part->current_block = -1;
drivers/mtd/rfd_ftl.c
458
part->reserved_block = best_block;
drivers/mtd/rfd_ftl.c
462
part->blocks[best_block].used_sectors,
drivers/mtd/rfd_ftl.c
463
part->blocks[best_block].free_sectors);
drivers/mtd/rfd_ftl.c
465
if (part->blocks[best_block].used_sectors)
drivers/mtd/rfd_ftl.c
466
rc = move_block_contents(part, best_block, old_sector);
drivers/mtd/rfd_ftl.c
468
rc = erase_block(part, best_block);
drivers/mtd/rfd_ftl.c
478
static int find_free_block(struct partition *part)
drivers/mtd/rfd_ftl.c
482
block = part->current_block == -1 ?
drivers/mtd/rfd_ftl.c
483
jiffies % part->total_blocks : part->current_block;
drivers/mtd/rfd_ftl.c
487
if (part->blocks[block].free_sectors &&
drivers/mtd/rfd_ftl.c
488
block != part->reserved_block)
drivers/mtd/rfd_ftl.c
491
if (part->blocks[block].state == BLOCK_UNUSED)
drivers/mtd/rfd_ftl.c
492
erase_block(part, block);
drivers/mtd/rfd_ftl.c
494
if (++block >= part->total_blocks)
drivers/mtd/rfd_ftl.c
502
static int find_writable_block(struct partition *part, u_long *old_sector)
drivers/mtd/rfd_ftl.c
507
block = find_free_block(part);
drivers/mtd/rfd_ftl.c
510
if (!part->is_reclaiming) {
drivers/mtd/rfd_ftl.c
511
rc = reclaim_block(part, old_sector);
drivers/mtd/rfd_ftl.c
515
block = find_free_block(part);
drivers/mtd/rfd_ftl.c
524
rc = mtd_read(part->mbd.mtd, part->blocks[block].offset,
drivers/mtd/rfd_ftl.c
525
part->header_size, &retlen,
drivers/mtd/rfd_ftl.c
526
(u_char *)part->header_cache);
drivers/mtd/rfd_ftl.c
528
if (!rc && retlen != part->header_size)
drivers/mtd/rfd_ftl.c
533
"0x%lx\n", part->mbd.mtd->name,
drivers/mtd/rfd_ftl.c
534
part->blocks[block].offset);
drivers/mtd/rfd_ftl.c
538
part->current_block = block;
drivers/mtd/rfd_ftl.c
544
static int mark_sector_deleted(struct partition *part, u_long old_addr)
drivers/mtd/rfd_ftl.c
551
block = old_addr / part->block_size;
drivers/mtd/rfd_ftl.c
552
offset = (old_addr % part->block_size) / SECTOR_SIZE -
drivers/mtd/rfd_ftl.c
553
part->header_sectors_per_block;
drivers/mtd/rfd_ftl.c
555
addr = part->blocks[block].offset +
drivers/mtd/rfd_ftl.c
557
rc = mtd_write(part->mbd.mtd, addr, sizeof(del), &retlen,
drivers/mtd/rfd_ftl.c
565
"0x%lx\n", part->mbd.mtd->name, addr);
drivers/mtd/rfd_ftl.c
568
if (block == part->current_block)
drivers/mtd/rfd_ftl.c
569
part->header_cache[offset + HEADER_MAP_OFFSET] = del;
drivers/mtd/rfd_ftl.c
571
part->blocks[block].used_sectors--;
drivers/mtd/rfd_ftl.c
573
if (!part->blocks[block].used_sectors &&
drivers/mtd/rfd_ftl.c
574
!part->blocks[block].free_sectors)
drivers/mtd/rfd_ftl.c
575
rc = erase_block(part, block);
drivers/mtd/rfd_ftl.c
581
static int find_free_sector(const struct partition *part, const struct block *block)
drivers/mtd/rfd_ftl.c
585
i = stop = part->data_sectors_per_block - block->free_sectors;
drivers/mtd/rfd_ftl.c
588
if (le16_to_cpu(part->header_cache[HEADER_MAP_OFFSET + i])
drivers/mtd/rfd_ftl.c
592
if (++i == part->data_sectors_per_block)
drivers/mtd/rfd_ftl.c
602
struct partition *part = container_of(dev, struct partition, mbd);
drivers/mtd/rfd_ftl.c
610
if (part->current_block == -1 ||
drivers/mtd/rfd_ftl.c
611
!part->blocks[part->current_block].free_sectors) {
drivers/mtd/rfd_ftl.c
613
rc = find_writable_block(part, old_addr);
drivers/mtd/rfd_ftl.c
618
block = &part->blocks[part->current_block];
drivers/mtd/rfd_ftl.c
620
i = find_free_sector(part, block);
drivers/mtd/rfd_ftl.c
627
addr = (i + part->header_sectors_per_block) * SECTOR_SIZE +
drivers/mtd/rfd_ftl.c
629
rc = mtd_write(part->mbd.mtd, addr, SECTOR_SIZE, &retlen,
drivers/mtd/rfd_ftl.c
637
part->mbd.mtd->name, addr);
drivers/mtd/rfd_ftl.c
641
part->sector_map[sector] = addr;
drivers/mtd/rfd_ftl.c
645
part->header_cache[i + HEADER_MAP_OFFSET] = entry;
drivers/mtd/rfd_ftl.c
648
rc = mtd_write(part->mbd.mtd, addr, sizeof(entry), &retlen,
drivers/mtd/rfd_ftl.c
656
part->mbd.mtd->name, addr);
drivers/mtd/rfd_ftl.c
668
struct partition *part = container_of(dev, struct partition, mbd);
drivers/mtd/rfd_ftl.c
675
if (part->reserved_block == -1) {
drivers/mtd/rfd_ftl.c
680
if (sector >= part->sector_count) {
drivers/mtd/rfd_ftl.c
685
old_addr = part->sector_map[sector];
drivers/mtd/rfd_ftl.c
698
part->sector_map[sector] = -1;
drivers/mtd/rfd_ftl.c
701
rc = mark_sector_deleted(part, old_addr);
drivers/mtd/rfd_ftl.c
710
struct partition *part = container_of(dev, struct partition, mbd);
drivers/mtd/rfd_ftl.c
715
if (sector >= part->sector_count)
drivers/mtd/rfd_ftl.c
718
addr = part->sector_map[sector];
drivers/mtd/rfd_ftl.c
721
rc = mark_sector_deleted(part, addr);
drivers/mtd/rfd_ftl.c
725
part->sector_map[sector] = -1;
drivers/mtd/rfd_ftl.c
737
struct partition *part = container_of(dev, struct partition, mbd);
drivers/mtd/rfd_ftl.c
741
geo->cylinders = part->cylinders;
drivers/mtd/rfd_ftl.c
748
struct partition *part;
drivers/mtd/rfd_ftl.c
754
part = kzalloc_obj(struct partition);
drivers/mtd/rfd_ftl.c
755
if (!part)
drivers/mtd/rfd_ftl.c
758
part->mbd.mtd = mtd;
drivers/mtd/rfd_ftl.c
761
part->block_size = block_size;
drivers/mtd/rfd_ftl.c
767
part->block_size = mtd->erasesize;
drivers/mtd/rfd_ftl.c
770
if (scan_header(part) == 0) {
drivers/mtd/rfd_ftl.c
771
part->mbd.size = part->sector_count;
drivers/mtd/rfd_ftl.c
772
part->mbd.tr = tr;
drivers/mtd/rfd_ftl.c
773
part->mbd.devnum = -1;
drivers/mtd/rfd_ftl.c
775
part->mbd.readonly = 1;
drivers/mtd/rfd_ftl.c
776
else if (part->errors) {
drivers/mtd/rfd_ftl.c
779
part->mbd.readonly = 1;
drivers/mtd/rfd_ftl.c
785
if (!add_mtd_blktrans_dev(&part->mbd))
drivers/mtd/rfd_ftl.c
789
kfree(part);
drivers/mtd/rfd_ftl.c
794
struct partition *part = container_of(dev, struct partition, mbd);
drivers/mtd/rfd_ftl.c
797
for (i=0; i<part->total_blocks; i++) {
drivers/mtd/rfd_ftl.c
799
part->mbd.mtd->name, i, part->blocks[i].erases);
drivers/mtd/rfd_ftl.c
802
vfree(part->sector_map);
drivers/mtd/rfd_ftl.c
803
kfree(part->header_cache);
drivers/mtd/rfd_ftl.c
804
kfree(part->blocks);
drivers/mtd/rfd_ftl.c
805
del_mtd_blktrans_dev(&part->mbd);
drivers/mtd/rfd_ftl.c
93
static int build_block_map(struct partition *part, int block_no)
drivers/mtd/rfd_ftl.c
95
struct block *block = &part->blocks[block_no];
drivers/mtd/rfd_ftl.c
98
block->offset = part->block_size * block_no;
drivers/mtd/spi-nor/core.c
1966
const struct flash_info *part;
drivers/mtd/spi-nor/core.c
1971
part = &manufacturers[i]->parts[j];
drivers/mtd/spi-nor/core.c
1972
if (part->id &&
drivers/mtd/spi-nor/core.c
1973
!memcmp(part->id->bytes, id, part->id->len)) {
drivers/mtd/spi-nor/core.c
1975
return part;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c
401
u32 part;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c
415
part = (speed * 100) / line_speed;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c
417
if (line_speed < speed || !part) {
drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c
425
bp->pending_max = part;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c
427
bnx2x_update_max_mf_config(bp, part);
drivers/net/ethernet/brocade/bna/bfa_defs.h
280
struct bfa_flash_part_attr part[BFA_FLASH_PART_MAX];
drivers/net/ethernet/brocade/bna/bfa_ioc.c
3151
attr->part[i].part_type =
drivers/net/ethernet/brocade/bna/bfa_ioc.c
3152
be32_to_cpu(f->part[i].part_type);
drivers/net/ethernet/brocade/bna/bfa_ioc.c
3153
attr->part[i].part_instance =
drivers/net/ethernet/brocade/bna/bfa_ioc.c
3154
be32_to_cpu(f->part[i].part_instance);
drivers/net/ethernet/brocade/bna/bfa_ioc.c
3155
attr->part[i].part_off =
drivers/net/ethernet/brocade/bna/bfa_ioc.c
3156
be32_to_cpu(f->part[i].part_off);
drivers/net/ethernet/brocade/bna/bfa_ioc.c
3157
attr->part[i].part_size =
drivers/net/ethernet/brocade/bna/bfa_ioc.c
3158
be32_to_cpu(f->part[i].part_size);
drivers/net/ethernet/brocade/bna/bfa_ioc.c
3159
attr->part[i].part_len =
drivers/net/ethernet/brocade/bna/bfa_ioc.c
3160
be32_to_cpu(f->part[i].part_len);
drivers/net/ethernet/brocade/bna/bfa_ioc.c
3161
attr->part[i].part_status =
drivers/net/ethernet/brocade/bna/bfa_ioc.c
3162
be32_to_cpu(f->part[i].part_status);
drivers/net/ethernet/brocade/bna/bnad_ethtool.c
926
if (offset >= flash_attr->part[i].part_off &&
drivers/net/ethernet/brocade/bna/bnad_ethtool.c
927
offset < (flash_attr->part[i].part_off +
drivers/net/ethernet/brocade/bna/bnad_ethtool.c
928
flash_attr->part[i].part_size)) {
drivers/net/ethernet/brocade/bna/bnad_ethtool.c
929
flash_part = flash_attr->part[i].part_type;
drivers/net/ethernet/brocade/bna/bnad_ethtool.c
930
*base_offset = flash_attr->part[i].part_off;
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
8961
unsigned int part, manufacturer;
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
8981
for (part = 0; part < ARRAY_SIZE(supported_flash); part++)
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
8982
if (supported_flash[part].vendor_and_model_id == flashid) {
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
8983
adap->params.sf_size = supported_flash[part].size_mb;
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
118
static int mlxsw_sp1_kvdl_part_alloc(struct mlxsw_sp1_kvdl_part *part,
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
121
const struct mlxsw_sp1_kvdl_part_info *info = &part->info;
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
126
entry_index = find_first_zero_bit(part->usage, nr_entries);
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
129
__set_bit(entry_index, part->usage);
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
136
static void mlxsw_sp1_kvdl_part_free(struct mlxsw_sp1_kvdl_part *part,
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
139
const struct mlxsw_sp1_kvdl_part_info *info = &part->info;
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
143
__clear_bit(entry_index, part->usage);
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
152
struct mlxsw_sp1_kvdl_part *part;
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
157
part = mlxsw_sp1_kvdl_alloc_size_part(kvdl, entry_count);
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
158
if (IS_ERR(part))
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
159
return PTR_ERR(part);
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
161
return mlxsw_sp1_kvdl_part_alloc(part, p_entry_index);
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
169
struct mlxsw_sp1_kvdl_part *part;
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
171
part = mlxsw_sp1_kvdl_index_part(kvdl, entry_index);
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
172
if (IS_ERR(part))
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
174
mlxsw_sp1_kvdl_part_free(part, entry_index);
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
184
struct mlxsw_sp1_kvdl_part *part;
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
186
part = mlxsw_sp1_kvdl_alloc_size_part(kvdl, entry_count);
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
187
if (IS_ERR(part))
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
188
return PTR_ERR(part);
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
190
*p_alloc_size = part->info.alloc_size;
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
195
static void mlxsw_sp1_kvdl_part_update(struct mlxsw_sp1_kvdl_part *part,
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
200
part->info.end_index = size - 1;
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
202
part->info.start_index = part_prev->info.end_index + 1;
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
203
part->info.end_index = part->info.start_index + size - 1;
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
213
struct mlxsw_sp1_kvdl_part *part;
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
227
part = kzalloc_flex(*part, usage, BITS_TO_LONGS(nr_entries));
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
228
if (!part)
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
231
memcpy(&part->info, info, sizeof(part->info));
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
234
mlxsw_sp1_kvdl_part_update(part, part_prev, resource_size);
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
235
return part;
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
238
static void mlxsw_sp1_kvdl_part_fini(struct mlxsw_sp1_kvdl_part *part)
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
240
kfree(part);
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
276
static u64 mlxsw_sp1_kvdl_part_occ(struct mlxsw_sp1_kvdl_part *part)
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
278
const struct mlxsw_sp1_kvdl_part_info *info = &part->info;
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
286
while ((bit = find_next_bit(part->usage, nr_entries, bit + 1))
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
307
struct mlxsw_sp1_kvdl_part *part;
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
309
part = kvdl->parts[MLXSW_SP1_KVDL_PART_ID_SINGLE];
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
310
return mlxsw_sp1_kvdl_part_occ(part);
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
316
struct mlxsw_sp1_kvdl_part *part;
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
318
part = kvdl->parts[MLXSW_SP1_KVDL_PART_ID_CHUNKS];
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
319
return mlxsw_sp1_kvdl_part_occ(part);
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
325
struct mlxsw_sp1_kvdl_part *part;
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
327
part = kvdl->parts[MLXSW_SP1_KVDL_PART_ID_LARGE_CHUNKS];
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
328
return mlxsw_sp1_kvdl_part_occ(part);
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
74
struct mlxsw_sp1_kvdl_part *part, *min_part = NULL;
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
78
part = kvdl->parts[i];
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
79
if (alloc_size <= part->info.alloc_size &&
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
81
part->info.alloc_size <= min_part->info.alloc_size))
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
82
min_part = part;
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
91
struct mlxsw_sp1_kvdl_part *part;
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
95
part = kvdl->parts[i];
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
96
if (kvdl_index >= part->info.start_index &&
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
97
kvdl_index <= part->info.end_index)
drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c
98
return part;
drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c
102
__set_bit(bit + i, part->usage);
drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c
103
*p_kvdl_index = bit * part->indexes_per_usage_bit;
drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c
125
struct mlxsw_sp2_kvdl_part *part,
drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c
134
err = mlxsw_sp2_kvdl_rec_del(mlxsw_sp, part->info->res_type,
drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c
139
bit_count = DIV_ROUND_UP(size, part->indexes_per_usage_bit);
drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c
140
bit = kvdl_index / part->indexes_per_usage_bit;
drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c
142
__clear_bit(bit + i, part->usage);
drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c
152
struct mlxsw_sp2_kvdl_part *part = kvdl->parts[type];
drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c
154
return mlxsw_sp2_kvdl_part_alloc(part, size, p_entry_index);
drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c
164
struct mlxsw_sp2_kvdl_part *part = kvdl->parts[type];
drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c
166
return mlxsw_sp2_kvdl_part_free(mlxsw_sp, part, size, entry_index);
drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c
184
struct mlxsw_sp2_kvdl_part *part;
drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c
206
part = kzalloc(sizeof(*part) + usage_size, GFP_KERNEL);
drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c
207
if (!part)
drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c
209
part->info = info;
drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c
210
part->usage_bit_count = usage_bit_count;
drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c
211
part->indexes_per_usage_bit = indexes_per_usage_bit;
drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c
212
part->last_allocated_bit = usage_bit_count - 1;
drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c
213
return part;
drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c
216
static void mlxsw_sp2_kvdl_part_fini(struct mlxsw_sp2_kvdl_part *part)
drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c
218
kfree(part);
drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c
56
static int mlxsw_sp2_kvdl_part_find_zero_bits(struct mlxsw_sp2_kvdl_part *part,
drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c
65
start_bit = part->last_allocated_bit + 1;
drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c
66
if (start_bit == part->usage_bit_count)
drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c
70
bit = find_next_zero_bit(part->usage, part->usage_bit_count, bit);
drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c
71
if (!wrap && bit + bit_count >= part->usage_bit_count) {
drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c
79
if (test_bit(bit + i, part->usage)) {
drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c
88
static int mlxsw_sp2_kvdl_part_alloc(struct mlxsw_sp2_kvdl_part *part,
drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c
97
bit_count = DIV_ROUND_UP(size, part->indexes_per_usage_bit);
drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c
98
err = mlxsw_sp2_kvdl_part_find_zero_bits(part, bit_count, &bit);
drivers/net/ethernet/netronome/nfp/nfp_devlink.c
242
const char *sn, *vendor, *part;
drivers/net/ethernet/netronome/nfp/nfp_devlink.c
248
part = nfp_hwinfo_lookup(pf->hwinfo, "assembly.partno");
drivers/net/ethernet/netronome/nfp/nfp_devlink.c
250
if (vendor && part && sn) {
drivers/net/ethernet/netronome/nfp/nfp_devlink.c
253
buf = kmalloc(strlen(vendor) + strlen(part) + strlen(sn) + 1,
drivers/net/ethernet/netronome/nfp/nfp_devlink.c
260
strcat(buf, part);
drivers/net/ethernet/sfc/ef10.c
3497
struct efx_mcdi_mtd_partition *part,
drivers/net/ethernet/sfc/ef10.c
3540
part->nvram_type = type;
drivers/net/ethernet/sfc/ef10.c
3551
part->fw_subtype = MCDI_DWORD(outbuf,
drivers/net/ethernet/sfc/ef10.c
3554
part->common.dev_type_name = "EF10 NVRAM manager";
drivers/net/ethernet/sfc/ef10.c
3555
part->common.type_name = info->name;
drivers/net/ethernet/sfc/ef10.c
3557
part->common.mtd.type = MTD_NORFLASH;
drivers/net/ethernet/sfc/ef10.c
3558
part->common.mtd.flags = MTD_CAP_NORFLASH;
drivers/net/ethernet/sfc/ef10.c
3559
part->common.mtd.size = size;
drivers/net/ethernet/sfc/ef10.c
3560
part->common.mtd.erasesize = erase_size;
drivers/net/ethernet/sfc/ef10.c
3563
part->common.mtd.flags |= MTD_NO_ERASE;
drivers/net/ethernet/sfc/ef10.c
3565
part->common.mtd.writesize = write_size;
drivers/net/ethernet/sfc/falcon/falcon.c
740
falcon_spi_slow_wait(struct falcon_mtd_partition *part, bool uninterruptible)
drivers/net/ethernet/sfc/falcon/falcon.c
742
const struct falcon_spi_device *spi = part->spi;
drivers/net/ethernet/sfc/falcon/falcon.c
743
struct ef4_nic *efx = part->common.mtd.priv;
drivers/net/ethernet/sfc/falcon/falcon.c
762
part->common.name, part->common.dev_type_name);
drivers/net/ethernet/sfc/falcon/falcon.c
804
falcon_spi_erase(struct falcon_mtd_partition *part, loff_t start, size_t len)
drivers/net/ethernet/sfc/falcon/falcon.c
806
const struct falcon_spi_device *spi = part->spi;
drivers/net/ethernet/sfc/falcon/falcon.c
807
struct ef4_nic *efx = part->common.mtd.priv;
drivers/net/ethernet/sfc/falcon/falcon.c
829
rc = falcon_spi_slow_wait(part, false);
drivers/net/ethernet/sfc/falcon/falcon.c
851
static void falcon_mtd_rename(struct ef4_mtd_partition *part)
drivers/net/ethernet/sfc/falcon/falcon.c
853
struct ef4_nic *efx = part->mtd.priv;
drivers/net/ethernet/sfc/falcon/falcon.c
855
snprintf(part->name, sizeof(part->name), "%s %s",
drivers/net/ethernet/sfc/falcon/falcon.c
856
efx->name, part->type_name);
drivers/net/ethernet/sfc/falcon/falcon.c
862
struct falcon_mtd_partition *part = to_falcon_mtd_partition(mtd);
drivers/net/ethernet/sfc/falcon/falcon.c
870
rc = falcon_spi_read(efx, part->spi, part->offset + start,
drivers/net/ethernet/sfc/falcon/falcon.c
878
struct falcon_mtd_partition *part = to_falcon_mtd_partition(mtd);
drivers/net/ethernet/sfc/falcon/falcon.c
886
rc = falcon_spi_erase(part, part->offset + start, len);
drivers/net/ethernet/sfc/falcon/falcon.c
894
struct falcon_mtd_partition *part = to_falcon_mtd_partition(mtd);
drivers/net/ethernet/sfc/falcon/falcon.c
902
rc = falcon_spi_write(efx, part->spi, part->offset + start,
drivers/net/ethernet/sfc/falcon/falcon.c
910
struct falcon_mtd_partition *part = to_falcon_mtd_partition(mtd);
drivers/net/ethernet/sfc/falcon/falcon.c
916
rc = falcon_spi_slow_wait(part, true);
drivers/net/ethernet/sfc/falcon/mtd.c
107
list_for_each_entry_safe(part, next, &efx->mtd_list, node)
drivers/net/ethernet/sfc/falcon/mtd.c
108
ef4_mtd_remove_partition(part);
drivers/net/ethernet/sfc/falcon/mtd.c
115
struct ef4_mtd_partition *part;
drivers/net/ethernet/sfc/falcon/mtd.c
119
list_for_each_entry(part, &efx->mtd_list, node)
drivers/net/ethernet/sfc/falcon/mtd.c
120
efx->type->mtd_rename(part);
drivers/net/ethernet/sfc/falcon/mtd.c
30
struct ef4_mtd_partition *part = to_ef4_mtd_partition(mtd);
drivers/net/ethernet/sfc/falcon/mtd.c
37
part->name, part->dev_type_name, rc);
drivers/net/ethernet/sfc/falcon/mtd.c
40
static void ef4_mtd_remove_partition(struct ef4_mtd_partition *part)
drivers/net/ethernet/sfc/falcon/mtd.c
45
rc = mtd_device_unregister(&part->mtd);
drivers/net/ethernet/sfc/falcon/mtd.c
51
list_del(&part->node);
drivers/net/ethernet/sfc/falcon/mtd.c
57
struct ef4_mtd_partition *part;
drivers/net/ethernet/sfc/falcon/mtd.c
61
part = (struct ef4_mtd_partition *)((char *)parts +
drivers/net/ethernet/sfc/falcon/mtd.c
64
part->mtd.writesize = 1;
drivers/net/ethernet/sfc/falcon/mtd.c
66
part->mtd.owner = THIS_MODULE;
drivers/net/ethernet/sfc/falcon/mtd.c
67
part->mtd.priv = efx;
drivers/net/ethernet/sfc/falcon/mtd.c
68
part->mtd.name = part->name;
drivers/net/ethernet/sfc/falcon/mtd.c
69
part->mtd._erase = ef4_mtd_erase;
drivers/net/ethernet/sfc/falcon/mtd.c
70
part->mtd._read = efx->type->mtd_read;
drivers/net/ethernet/sfc/falcon/mtd.c
71
part->mtd._write = efx->type->mtd_write;
drivers/net/ethernet/sfc/falcon/mtd.c
72
part->mtd._sync = ef4_mtd_sync;
drivers/net/ethernet/sfc/falcon/mtd.c
74
efx->type->mtd_rename(part);
drivers/net/ethernet/sfc/falcon/mtd.c
76
if (mtd_device_register(&part->mtd, NULL, 0))
drivers/net/ethernet/sfc/falcon/mtd.c
80
list_add_tail(&part->node, &efx->mtd_list);
drivers/net/ethernet/sfc/falcon/mtd.c
87
part = (struct ef4_mtd_partition *)((char *)parts +
drivers/net/ethernet/sfc/falcon/mtd.c
89
ef4_mtd_remove_partition(part);
drivers/net/ethernet/sfc/falcon/mtd.c
97
struct ef4_mtd_partition *parts, *part, *next;
drivers/net/ethernet/sfc/falcon/net_driver.h
1130
void (*mtd_rename)(struct ef4_mtd_partition *part);
drivers/net/ethernet/sfc/mcdi.c
2376
struct efx_mcdi_mtd_partition *part = to_efx_mcdi_mtd_partition(mtd);
drivers/net/ethernet/sfc/mcdi.c
2385
rc = efx_mcdi_nvram_read(efx, part->nvram_type, offset,
drivers/net/ethernet/sfc/mcdi.c
2399
struct efx_mcdi_mtd_partition *part = to_efx_mcdi_mtd_partition(mtd);
drivers/net/ethernet/sfc/mcdi.c
2403
size_t chunk = part->common.mtd.erasesize;
drivers/net/ethernet/sfc/mcdi.c
2406
if (!part->updating) {
drivers/net/ethernet/sfc/mcdi.c
2407
rc = efx_mcdi_nvram_update_start(efx, part->nvram_type);
drivers/net/ethernet/sfc/mcdi.c
2410
part->updating = true;
drivers/net/ethernet/sfc/mcdi.c
2417
rc = efx_mcdi_nvram_erase(efx, part->nvram_type, offset,
drivers/net/ethernet/sfc/mcdi.c
2430
struct efx_mcdi_mtd_partition *part = to_efx_mcdi_mtd_partition(mtd);
drivers/net/ethernet/sfc/mcdi.c
2437
if (!part->updating) {
drivers/net/ethernet/sfc/mcdi.c
2438
rc = efx_mcdi_nvram_update_start(efx, part->nvram_type);
drivers/net/ethernet/sfc/mcdi.c
2441
part->updating = true;
drivers/net/ethernet/sfc/mcdi.c
2446
rc = efx_mcdi_nvram_write(efx, part->nvram_type, offset,
drivers/net/ethernet/sfc/mcdi.c
2460
struct efx_mcdi_mtd_partition *part = to_efx_mcdi_mtd_partition(mtd);
drivers/net/ethernet/sfc/mcdi.c
2464
if (part->updating) {
drivers/net/ethernet/sfc/mcdi.c
2465
part->updating = false;
drivers/net/ethernet/sfc/mcdi.c
2466
rc = efx_mcdi_nvram_update_finish(efx, part->nvram_type,
drivers/net/ethernet/sfc/mcdi.c
2473
void efx_mcdi_mtd_rename(struct efx_mtd_partition *part)
drivers/net/ethernet/sfc/mcdi.c
2476
container_of(part, struct efx_mcdi_mtd_partition, common);
drivers/net/ethernet/sfc/mcdi.c
2477
struct efx_nic *efx = part->mtd.priv;
drivers/net/ethernet/sfc/mcdi.c
2479
snprintf(part->name, sizeof(part->name), "%s %s:%02x",
drivers/net/ethernet/sfc/mcdi.c
2480
efx->name, part->type_name, mcdi_part->fw_subtype);
drivers/net/ethernet/sfc/mcdi.h
454
void efx_mcdi_mtd_rename(struct efx_mtd_partition *part);
drivers/net/ethernet/sfc/mtd.c
100
struct efx_mtd_partition *parts, *part, *next;
drivers/net/ethernet/sfc/mtd.c
110
list_for_each_entry_safe(part, next, &efx->mtd_list, node)
drivers/net/ethernet/sfc/mtd.c
111
efx_mtd_remove_partition(part);
drivers/net/ethernet/sfc/mtd.c
118
struct efx_mtd_partition *part;
drivers/net/ethernet/sfc/mtd.c
122
list_for_each_entry(part, &efx->mtd_list, node)
drivers/net/ethernet/sfc/mtd.c
123
efx->type->mtd_rename(part);
drivers/net/ethernet/sfc/mtd.c
30
struct efx_mtd_partition *part = to_efx_mtd_partition(mtd);
drivers/net/ethernet/sfc/mtd.c
37
part->name, part->dev_type_name, rc);
drivers/net/ethernet/sfc/mtd.c
40
static void efx_mtd_remove_partition(struct efx_mtd_partition *part)
drivers/net/ethernet/sfc/mtd.c
45
rc = mtd_device_unregister(&part->mtd);
drivers/net/ethernet/sfc/mtd.c
51
list_del(&part->node);
drivers/net/ethernet/sfc/mtd.c
57
struct efx_mtd_partition *part;
drivers/net/ethernet/sfc/mtd.c
61
part = (struct efx_mtd_partition *)((char *)parts +
drivers/net/ethernet/sfc/mtd.c
64
part->mtd.writesize = 1;
drivers/net/ethernet/sfc/mtd.c
66
if (!(part->mtd.flags & MTD_NO_ERASE))
drivers/net/ethernet/sfc/mtd.c
67
part->mtd.flags |= MTD_WRITEABLE;
drivers/net/ethernet/sfc/mtd.c
69
part->mtd.owner = THIS_MODULE;
drivers/net/ethernet/sfc/mtd.c
70
part->mtd.priv = efx;
drivers/net/ethernet/sfc/mtd.c
71
part->mtd.name = part->name;
drivers/net/ethernet/sfc/mtd.c
72
part->mtd._erase = efx_mtd_erase;
drivers/net/ethernet/sfc/mtd.c
73
part->mtd._read = efx->type->mtd_read;
drivers/net/ethernet/sfc/mtd.c
74
part->mtd._write = efx->type->mtd_write;
drivers/net/ethernet/sfc/mtd.c
75
part->mtd._sync = efx_mtd_sync;
drivers/net/ethernet/sfc/mtd.c
77
efx->type->mtd_rename(part);
drivers/net/ethernet/sfc/mtd.c
79
if (mtd_device_register(&part->mtd, NULL, 0))
drivers/net/ethernet/sfc/mtd.c
83
list_add_tail(&part->node, &efx->mtd_list);
drivers/net/ethernet/sfc/mtd.c
90
part = (struct efx_mtd_partition *)((char *)parts +
drivers/net/ethernet/sfc/mtd.c
92
efx_mtd_remove_partition(part);
drivers/net/ethernet/sfc/net_driver.h
1504
void (*mtd_rename)(struct efx_mtd_partition *part);
drivers/net/ethernet/sfc/siena/mcdi.c
2154
struct efx_mcdi_mtd_partition *part = to_efx_mcdi_mtd_partition(mtd);
drivers/net/ethernet/sfc/siena/mcdi.c
2163
rc = efx_mcdi_nvram_read(efx, part->nvram_type, offset,
drivers/net/ethernet/sfc/siena/mcdi.c
2177
struct efx_mcdi_mtd_partition *part = to_efx_mcdi_mtd_partition(mtd);
drivers/net/ethernet/sfc/siena/mcdi.c
2181
size_t chunk = part->common.mtd.erasesize;
drivers/net/ethernet/sfc/siena/mcdi.c
2184
if (!part->updating) {
drivers/net/ethernet/sfc/siena/mcdi.c
2185
rc = efx_mcdi_nvram_update_start(efx, part->nvram_type);
drivers/net/ethernet/sfc/siena/mcdi.c
2188
part->updating = true;
drivers/net/ethernet/sfc/siena/mcdi.c
2195
rc = efx_mcdi_nvram_erase(efx, part->nvram_type, offset,
drivers/net/ethernet/sfc/siena/mcdi.c
2208
struct efx_mcdi_mtd_partition *part = to_efx_mcdi_mtd_partition(mtd);
drivers/net/ethernet/sfc/siena/mcdi.c
2215
if (!part->updating) {
drivers/net/ethernet/sfc/siena/mcdi.c
2216
rc = efx_mcdi_nvram_update_start(efx, part->nvram_type);
drivers/net/ethernet/sfc/siena/mcdi.c
2219
part->updating = true;
drivers/net/ethernet/sfc/siena/mcdi.c
2224
rc = efx_mcdi_nvram_write(efx, part->nvram_type, offset,
drivers/net/ethernet/sfc/siena/mcdi.c
2238
struct efx_mcdi_mtd_partition *part = to_efx_mcdi_mtd_partition(mtd);
drivers/net/ethernet/sfc/siena/mcdi.c
2242
if (part->updating) {
drivers/net/ethernet/sfc/siena/mcdi.c
2243
part->updating = false;
drivers/net/ethernet/sfc/siena/mcdi.c
2244
rc = efx_mcdi_nvram_update_finish(efx, part->nvram_type);
drivers/net/ethernet/sfc/siena/mcdi.c
2250
void efx_siena_mcdi_mtd_rename(struct efx_mtd_partition *part)
drivers/net/ethernet/sfc/siena/mcdi.c
2253
container_of(part, struct efx_mcdi_mtd_partition, common);
drivers/net/ethernet/sfc/siena/mcdi.c
2254
struct efx_nic *efx = part->mtd.priv;
drivers/net/ethernet/sfc/siena/mcdi.c
2256
snprintf(part->name, sizeof(part->name), "%s %s:%02x",
drivers/net/ethernet/sfc/siena/mcdi.c
2257
efx->name, part->type_name, mcdi_part->fw_subtype);
drivers/net/ethernet/sfc/siena/mcdi.h
383
void efx_siena_mcdi_mtd_rename(struct efx_mtd_partition *part);
drivers/net/ethernet/sfc/siena/mtd.c
100
struct efx_mtd_partition *parts, *part, *next;
drivers/net/ethernet/sfc/siena/mtd.c
110
list_for_each_entry_safe(part, next, &efx->mtd_list, node)
drivers/net/ethernet/sfc/siena/mtd.c
111
efx_siena_mtd_remove_partition(part);
drivers/net/ethernet/sfc/siena/mtd.c
118
struct efx_mtd_partition *part;
drivers/net/ethernet/sfc/siena/mtd.c
122
list_for_each_entry(part, &efx->mtd_list, node)
drivers/net/ethernet/sfc/siena/mtd.c
123
efx->type->mtd_rename(part);
drivers/net/ethernet/sfc/siena/mtd.c
30
struct efx_mtd_partition *part = to_efx_mtd_partition(mtd);
drivers/net/ethernet/sfc/siena/mtd.c
37
part->name, part->dev_type_name, rc);
drivers/net/ethernet/sfc/siena/mtd.c
40
static void efx_siena_mtd_remove_partition(struct efx_mtd_partition *part)
drivers/net/ethernet/sfc/siena/mtd.c
45
rc = mtd_device_unregister(&part->mtd);
drivers/net/ethernet/sfc/siena/mtd.c
51
list_del(&part->node);
drivers/net/ethernet/sfc/siena/mtd.c
57
struct efx_mtd_partition *part;
drivers/net/ethernet/sfc/siena/mtd.c
61
part = (struct efx_mtd_partition *)((char *)parts +
drivers/net/ethernet/sfc/siena/mtd.c
64
part->mtd.writesize = 1;
drivers/net/ethernet/sfc/siena/mtd.c
66
if (!(part->mtd.flags & MTD_NO_ERASE))
drivers/net/ethernet/sfc/siena/mtd.c
67
part->mtd.flags |= MTD_WRITEABLE;
drivers/net/ethernet/sfc/siena/mtd.c
69
part->mtd.owner = THIS_MODULE;
drivers/net/ethernet/sfc/siena/mtd.c
70
part->mtd.priv = efx;
drivers/net/ethernet/sfc/siena/mtd.c
71
part->mtd.name = part->name;
drivers/net/ethernet/sfc/siena/mtd.c
72
part->mtd._erase = efx_mtd_erase;
drivers/net/ethernet/sfc/siena/mtd.c
73
part->mtd._read = efx->type->mtd_read;
drivers/net/ethernet/sfc/siena/mtd.c
74
part->mtd._write = efx->type->mtd_write;
drivers/net/ethernet/sfc/siena/mtd.c
75
part->mtd._sync = efx_mtd_sync;
drivers/net/ethernet/sfc/siena/mtd.c
77
efx->type->mtd_rename(part);
drivers/net/ethernet/sfc/siena/mtd.c
79
if (mtd_device_register(&part->mtd, NULL, 0))
drivers/net/ethernet/sfc/siena/mtd.c
83
list_add_tail(&part->node, &efx->mtd_list);
drivers/net/ethernet/sfc/siena/mtd.c
90
part = (struct efx_mtd_partition *)((char *)parts +
drivers/net/ethernet/sfc/siena/mtd.c
92
efx_siena_mtd_remove_partition(part);
drivers/net/ethernet/sfc/siena/net_driver.h
1396
void (*mtd_rename)(struct efx_mtd_partition *part);
drivers/net/ethernet/sfc/siena/siena.c
857
struct efx_mcdi_mtd_partition *part,
drivers/net/ethernet/sfc/siena/siena.c
881
part->nvram_type = type;
drivers/net/ethernet/sfc/siena/siena.c
882
part->common.dev_type_name = "Siena NVRAM manager";
drivers/net/ethernet/sfc/siena/siena.c
883
part->common.type_name = info->name;
drivers/net/ethernet/sfc/siena/siena.c
885
part->common.mtd.type = MTD_NORFLASH;
drivers/net/ethernet/sfc/siena/siena.c
886
part->common.mtd.flags = MTD_CAP_NORFLASH;
drivers/net/ethernet/sfc/siena/siena.c
887
part->common.mtd.size = size;
drivers/net/ethernet/sfc/siena/siena.c
888
part->common.mtd.erasesize = erase_size;
drivers/net/ieee802154/at86rf230.c
1420
unsigned int part, version, val;
drivers/net/ieee802154/at86rf230.c
1435
rc = __at86rf230_read(lp, RG_PART_NUM, &part);
drivers/net/ieee802154/at86rf230.c
1465
switch (part) {
drivers/net/phy/sfp.c
498
{ .vendor = _v, .part = _p, .support = _s, .fixup = _f, }
drivers/net/phy/sfp.c
627
sfp_match(q->part, id->base.vendor_pn, ps))
drivers/net/phy/sfp.h
11
const char *part;
drivers/net/wireguard/selftest/allowedips.c
505
__be64 part;
drivers/net/wireguard/selftest/allowedips.c
647
part = cpu_to_be64(~0LLU << i);
drivers/net/wireguard/selftest/allowedips.c
649
memcpy((u8 *)&ip + 8, &part, 8);
drivers/net/wireguard/selftest/allowedips.c
651
memcpy(&ip, &part, 8);
drivers/net/wireless/mediatek/mt76/eeprom.c
37
const char *part;
drivers/net/wireless/mediatek/mt76/eeprom.c
55
part = of_get_property(np, "label", NULL);
drivers/net/wireless/mediatek/mt76/eeprom.c
56
if (!part)
drivers/net/wireless/mediatek/mt76/eeprom.c
57
part = np->name;
drivers/net/wireless/mediatek/mt76/eeprom.c
59
mtd = get_mtd_device_nm(part);
drivers/net/wireless/mediatek/mt76/eeprom.c
77
part, ret);
drivers/net/wireless/mediatek/mt76/eeprom.c
97
dev->test_mtd.name = devm_kstrdup(dev->dev, part, GFP_KERNEL);
drivers/net/wireless/ti/wlcore/debugfs.c
1029
struct wlcore_partition_set part, old_part;
drivers/net/wireless/ti/wlcore/debugfs.c
1047
memset(&part, 0, sizeof(part));
drivers/net/wireless/ti/wlcore/debugfs.c
1048
part.mem.start = *ppos;
drivers/net/wireless/ti/wlcore/debugfs.c
1049
part.mem.size = bytes;
drivers/net/wireless/ti/wlcore/debugfs.c
1070
ret = wlcore_set_partition(wl, &part);
drivers/net/wireless/ti/wlcore/debugfs.c
1110
struct wlcore_partition_set part, old_part;
drivers/net/wireless/ti/wlcore/debugfs.c
1128
memset(&part, 0, sizeof(part));
drivers/net/wireless/ti/wlcore/debugfs.c
1129
part.mem.start = *ppos;
drivers/net/wireless/ti/wlcore/debugfs.c
1130
part.mem.size = bytes;
drivers/net/wireless/ti/wlcore/debugfs.c
1151
ret = wlcore_set_partition(wl, &part);
drivers/net/wireless/ti/wlcore/io.c
57
struct wlcore_partition_set *part = &wl->curr_part;
drivers/net/wireless/ti/wlcore/io.c
69
if ((addr >= part->mem.start) &&
drivers/net/wireless/ti/wlcore/io.c
70
(addr < part->mem.start + part->mem.size))
drivers/net/wireless/ti/wlcore/io.c
71
return addr - part->mem.start;
drivers/net/wireless/ti/wlcore/io.c
72
else if ((addr >= part->reg.start) &&
drivers/net/wireless/ti/wlcore/io.c
73
(addr < part->reg.start + part->reg.size))
drivers/net/wireless/ti/wlcore/io.c
74
return addr - part->reg.start + part->mem.size;
drivers/net/wireless/ti/wlcore/io.c
75
else if ((addr >= part->mem2.start) &&
drivers/net/wireless/ti/wlcore/io.c
76
(addr < part->mem2.start + part->mem2.size))
drivers/net/wireless/ti/wlcore/io.c
77
return addr - part->mem2.start + part->mem.size +
drivers/net/wireless/ti/wlcore/io.c
78
part->reg.size;
drivers/net/wireless/ti/wlcore/io.c
79
else if ((addr >= part->mem3.start) &&
drivers/net/wireless/ti/wlcore/io.c
80
(addr < part->mem3.start + part->mem3.size))
drivers/net/wireless/ti/wlcore/io.c
81
return addr - part->mem3.start + part->mem.size +
drivers/net/wireless/ti/wlcore/io.c
82
part->reg.size + part->mem2.size;
drivers/ntb/hw/idt/ntb_hw_idt.c
1312
data = SET_FIELD(BARSETUP_TPART, data, ndev->peers[pidx].part);
drivers/ntb/hw/idt/ntb_hw_idt.c
1327
data = SET_FIELD(LUTUDATA_PART, 0, ndev->peers[pidx].part) |
drivers/ntb/hw/idt/ntb_hw_idt.c
1798
SET_FIELD(SWPxMSGCTL_PART, 0, ndev->peers[pidx].part);
drivers/ntb/hw/idt/ntb_hw_idt.c
1803
idt_sw_write(ndev, partdata_tbl[ndev->part].msgctl[midx], swpmsgctl);
drivers/ntb/hw/idt/ntb_hw_idt.c
2357
"Local Port %hhu, Partition %hhu\n", ndev->port, ndev->part);
drivers/ntb/hw/idt/ntb_hw_idt.c
2364
idx, ndev->peers[idx].port, ndev->peers[idx].part);
drivers/ntb/hw/idt/ntb_hw_idt.c
482
unsigned char pidx, port, part;
drivers/ntb/hw/idt/ntb_hw_idt.c
491
ndev->part = GET_FIELD(SWPORTxSTS_SWPART, portsts);
drivers/ntb/hw/idt/ntb_hw_idt.c
510
part = GET_FIELD(SWPORTxSTS_SWPART, portsts);
drivers/ntb/hw/idt/ntb_hw_idt.c
513
partsts = idt_sw_read(ndev, partdata_tbl[part].sts);
drivers/ntb/hw/idt/ntb_hw_idt.c
522
ndev->peers[ndev->peer_cnt].part = part;
drivers/ntb/hw/idt/ntb_hw_idt.c
525
ndev->part_idx_map[part] = ndev->peer_cnt;
drivers/ntb/hw/idt/ntb_hw_idt.c
645
part_mask = ~BIT(ndev->part);
drivers/ntb/hw/idt/ntb_hw_idt.c
648
part_mask &= ~BIT(ndev->peers[pidx].part);
drivers/ntb/hw/idt/ntb_hw_idt.c
746
SET_FIELD(NTMTBLDATA_PART, 0, ndev->part) |
drivers/ntb/hw/idt/ntb_hw_idt.c
749
idt_nt_write(ndev, IDT_NT_NTMTBLADDR, ndev->part);
drivers/ntb/hw/idt/ntb_hw_idt.c
755
idt_sw_write(ndev, IDT_SW_SEGSIGSTS, (u32)1 << ndev->part);
drivers/ntb/hw/idt/ntb_hw_idt.c
776
idt_nt_write(ndev, IDT_NT_NTMTBLADDR, ndev->part);
drivers/ntb/hw/idt/ntb_hw_idt.c
782
idt_sw_write(ndev, IDT_SW_SEGSIGSTS, (u32)1 << ndev->part);
drivers/ntb/hw/idt/ntb_hw_idt.c
815
idt_nt_write(ndev, IDT_NT_NTMTBLADDR, ndev->part);
drivers/ntb/hw/idt/ntb_hw_idt.c
861
idt_nt_write(ndev, IDT_NT_NTMTBLADDR, ndev->peers[pidx].part);
drivers/ntb/hw/idt/ntb_hw_idt.h
1086
unsigned char part;
drivers/ntb/hw/idt/ntb_hw_idt.h
1129
unsigned char part;
drivers/pci/switch/switchtec.c
1120
int i, part;
drivers/pci/switch/switchtec.c
1129
for (part = 0; part < stdev->partition_count; part++) {
drivers/pci/switch/switchtec.c
1130
pcfg = &stdev->mmio_part_cfg_all[part];
drivers/pci/switch/switchtec.c
1131
p.partition = part;
drivers/pci/switch/switchtec.c
908
s->part[i] = reg;
drivers/perf/arm-cmn.c
1340
return sysfs_emit(buf, "%03x%02x\n", cmn->part, cmn->rev);
drivers/perf/arm-cmn.c
1401
bool is_cmn600 = to_cmn(event->pmu)->part == PART_CMN600;
drivers/perf/arm-cmn.c
1801
(cmn->part == PART_CMN700 || cmn->part == PART_CMN_S3)) {
drivers/perf/arm-cmn.c
1894
if (cmn->part == PART_CMN600 && j > 0) {
drivers/perf/arm-cmn.c
2209
if (cmn->part == PART_CMN650 || cmn->part == PART_CI700)
drivers/perf/arm-cmn.c
2255
enum cmn_part part;
drivers/perf/arm-cmn.c
2268
part = FIELD_GET(CMN_CFGM_PID0_PART_0, reg);
drivers/perf/arm-cmn.c
2269
part |= FIELD_GET(CMN_CFGM_PID1_PART_1, reg) << 8;
drivers/perf/arm-cmn.c
2271
if (part == PART_CMN600AE)
drivers/perf/arm-cmn.c
2272
part = PART_CMN600;
drivers/perf/arm-cmn.c
2273
if (cmn->part && cmn->part != part)
drivers/perf/arm-cmn.c
2276
cmn->part, part);
drivers/perf/arm-cmn.c
2277
cmn->part = part;
drivers/perf/arm-cmn.c
2279
dev_warn(cmn->dev, "Unknown part number: 0x%x\n", part);
drivers/perf/arm-cmn.c
2357
if (cmn->part == PART_CMN600)
drivers/perf/arm-cmn.c
2417
if (reg == 0 && cmn->part != PART_CMN600) {
drivers/perf/arm-cmn.c
2519
dev_dbg(cmn->dev, "periph_id part 0x%03x revision %d\n", cmn->part, cmn->rev);
drivers/perf/arm-cmn.c
2574
cmn->part = (unsigned long)device_get_match_data(cmn->dev);
drivers/perf/arm-cmn.c
2578
if (cmn->part == PART_CMN600 && has_acpi_companion(cmn->dev)) {
drivers/perf/arm-cmn.c
2585
if (cmn->part == PART_CMN600)
drivers/perf/arm-cmn.c
339
enum cmn_part part;
drivers/perf/arm-cmn.c
413
switch (cmn->part) {
drivers/perf/arm-cmn.c
431
if (cmn->part == PART_CMN_S3) {
drivers/perf/arm-cmn.c
445
if (cmn->part == PART_CMN600 || cmn->part == PART_CMN650)
drivers/perf/arm-cmn.c
451
if (cmn->part == PART_CI700)
drivers/perf/arm-cmn.c
729
if (chan == 4 && cmn->part == PART_CMN600)
drivers/perf/arm-cmn.c
740
if (cmn->part == PART_CMN600) {
drivers/perf/arm-cmn.c
751
} else if (cmn->part == PART_CMN650) {
drivers/perf/arm-cmn.c
760
} else if (cmn->part == PART_CMN700) {
drivers/perf/arm-ni.c
129
enum ni_part part;
drivers/perf/arm-ni.c
255
return sysfs_emit(buf, "%03x%02x\n", ni->part, version);
drivers/perf/arm-ni.c
328
return ni->part == PART_NI_700 || ni->part == PART_NI_710AE;
drivers/perf/arm-ni.c
679
u32 reg, part;
drivers/perf/arm-ni.c
696
part = FIELD_GET(NI_PIDR0_PART_7_0, reg);
drivers/perf/arm-ni.c
698
part |= FIELD_GET(NI_PIDR1_PART_11_8, reg) << 8;
drivers/perf/arm-ni.c
700
switch (part) {
drivers/perf/arm-ni.c
707
dev_WARN(&pdev->dev, "Unknown part number: 0x%03x, this may go badly\n", part);
drivers/perf/arm-ni.c
729
ni->part = part;
drivers/platform/x86/hp/hp-bioscfg/order-list-attributes.c
142
char *part = NULL;
drivers/platform/x86/hp/hp-bioscfg/order-list-attributes.c
269
part = strsep(&part_tmp, COMMA_SEP);
drivers/platform/x86/hp/hp-bioscfg/order-list-attributes.c
271
for (olist_elem = 0; olist_elem < MAX_ELEMENTS_SIZE && part; olist_elem++) {
drivers/platform/x86/hp/hp-bioscfg/order-list-attributes.c
272
strscpy(ordered_list_data->elements[olist_elem], part);
drivers/platform/x86/hp/hp-bioscfg/order-list-attributes.c
273
part = strsep(&part_tmp, COMMA_SEP);
drivers/platform/x86/intel/speed_select_if/isst_tpmi_core.c
1379
u8 part;
drivers/platform/x86/intel/speed_select_if/isst_tpmi_core.c
1382
pd = map_partition_power_domain_id(sst_inst, i, &part);
drivers/platform/x86/intel/speed_select_if/isst_tpmi_core.c
1386
pd_info = &sst_inst->power_domain_info[part][pd];
drivers/platform/x86/intel/speed_select_if/isst_tpmi_core.c
536
u8 part;
drivers/platform/x86/intel/speed_select_if/isst_tpmi_core.c
545
power_domain_id = map_partition_power_domain_id(sst_inst, power_domain_id, &part);
drivers/platform/x86/intel/speed_select_if/isst_tpmi_core.c
549
power_domain_info = &sst_inst->power_domain_info[part][power_domain_id];
drivers/platform/x86/intel/speed_select_if/isst_tpmi_core.c
727
u8 part;
drivers/platform/x86/intel/speed_select_if/isst_tpmi_core.c
751
punit_id = map_partition_power_domain_id(sst_inst, punit_id, &part);
drivers/platform/x86/intel/speed_select_if/isst_tpmi_core.c
755
power_domain_info = &sst_inst->power_domain_info[part][punit_id];
drivers/rtc/rtc-abx80x.c
776
unsigned int part = id->driver_data;
drivers/rtc/rtc-abx80x.c
818
if (part == RV1805) {
drivers/rtc/rtc-abx80x.c
869
if (part == ABX80X) {
drivers/rtc/rtc-abx80x.c
878
part = i;
drivers/rtc/rtc-abx80x.c
881
if (partnumber != abx80x_caps[part].pn) {
drivers/rtc/rtc-abx80x.c
883
partnumber, abx80x_caps[part].pn);
drivers/rtc/rtc-abx80x.c
887
if (np && abx80x_caps[part].has_tc)
drivers/rtc/rtc-abx80x.c
914
if (abx80x_caps[part].has_wdog) {
drivers/scsi/bfa/bfa_defs.h
1140
struct bfa_flash_part_attr_s part[BFA_FLASH_PART_MAX];
drivers/scsi/bfa/bfa_ioc.c
4355
attr->part[i].part_type =
drivers/scsi/bfa/bfa_ioc.c
4356
be32_to_cpu(f->part[i].part_type);
drivers/scsi/bfa/bfa_ioc.c
4357
attr->part[i].part_instance =
drivers/scsi/bfa/bfa_ioc.c
4358
be32_to_cpu(f->part[i].part_instance);
drivers/scsi/bfa/bfa_ioc.c
4359
attr->part[i].part_off =
drivers/scsi/bfa/bfa_ioc.c
4360
be32_to_cpu(f->part[i].part_off);
drivers/scsi/bfa/bfa_ioc.c
4361
attr->part[i].part_size =
drivers/scsi/bfa/bfa_ioc.c
4362
be32_to_cpu(f->part[i].part_size);
drivers/scsi/bfa/bfa_ioc.c
4363
attr->part[i].part_len =
drivers/scsi/bfa/bfa_ioc.c
4364
be32_to_cpu(f->part[i].part_len);
drivers/scsi/bfa/bfa_ioc.c
4365
attr->part[i].part_status =
drivers/scsi/bfa/bfa_ioc.c
4366
be32_to_cpu(f->part[i].part_status);
drivers/scsi/csiostor/csio_hw.c
774
u32 part, manufacturer;
drivers/scsi/csiostor/csio_hw.c
788
for (part = 0; part < ARRAY_SIZE(supported_flash); part++)
drivers/scsi/csiostor/csio_hw.c
789
if (supported_flash[part].vendor_and_model_id == flashid) {
drivers/scsi/csiostor/csio_hw.c
790
hw->params.sf_size = supported_flash[part].size_mb;
drivers/soc/qcom/smem.c
404
struct smem_partition *part,
drivers/soc/qcom/smem.c
414
phdr = (struct smem_partition_header __force *)part->virt_base;
drivers/soc/qcom/smem.c
415
p_end = (void *)phdr + part->size;
drivers/soc/qcom/smem.c
511
struct smem_partition *part;
drivers/soc/qcom/smem.c
534
part = &__smem->partitions[host];
drivers/soc/qcom/smem.c
535
ret = qcom_smem_alloc_private(__smem, part, item, size);
drivers/soc/qcom/smem.c
537
part = &__smem->global_partition;
drivers/soc/qcom/smem.c
538
ret = qcom_smem_alloc_private(__smem, part, item, size);
drivers/soc/qcom/smem.c
589
struct smem_partition *part,
drivers/soc/qcom/smem.c
599
phdr = (struct smem_partition_header __force *)part->virt_base;
drivers/soc/qcom/smem.c
600
p_end = (void *)phdr + part->size;
drivers/soc/qcom/smem.c
614
if (WARN_ON(e_size > part->size || padding_data > e_size))
drivers/soc/qcom/smem.c
635
e = phdr_to_first_cached_entry(phdr, part->cacheline);
drivers/soc/qcom/smem.c
650
if (WARN_ON(e_size > part->size || padding_data > e_size))
drivers/soc/qcom/smem.c
663
e = cached_entry_next(e, part->cacheline);
drivers/soc/qcom/smem.c
691
struct smem_partition *part;
drivers/soc/qcom/smem.c
701
part = &__smem->partitions[host];
drivers/soc/qcom/smem.c
702
ptr = qcom_smem_get_private(__smem, part, item, size);
drivers/soc/qcom/smem.c
704
part = &__smem->global_partition;
drivers/soc/qcom/smem.c
705
ptr = qcom_smem_get_private(__smem, part, item, size);
drivers/soc/qcom/smem.c
725
struct smem_partition *part;
drivers/soc/qcom/smem.c
734
part = &__smem->partitions[host];
drivers/soc/qcom/smem.c
735
phdr = part->virt_base;
drivers/soc/qcom/smem.c
739
if (ret > le32_to_cpu(part->size))
drivers/soc/qcom/smem.c
742
part = &__smem->global_partition;
drivers/soc/qcom/smem.c
743
phdr = part->virt_base;
drivers/soc/qcom/smem.c
747
if (ret > le32_to_cpu(part->size))
drivers/soc/qcom/smem.c
775
struct smem_partition *part;
drivers/soc/qcom/smem.c
781
part = &__smem->partitions[i];
drivers/soc/qcom/smem.c
783
if (addr_in_range(part->virt_base, part->size, p)) {
drivers/soc/qcom/smem.c
784
offset = p - part->virt_base;
drivers/soc/qcom/smem.c
786
return (phys_addr_t)part->phys_base + offset;
drivers/soc/qcom/smem.c
790
part = &__smem->global_partition;
drivers/soc/qcom/smem.c
792
if (addr_in_range(part->virt_base, part->size, p)) {
drivers/soc/qcom/smem.c
793
offset = p - part->virt_base;
drivers/soc/qcom/smem.c
795
return (phys_addr_t)part->phys_base + offset;
drivers/spi/spi-cadence-quadspi.c
57
#define CQSPI_OP_WIDTH(part) ((part).nbytes ? ilog2((part).buswidth) : 0)
drivers/spi/spi-intel.c
1231
struct mtd_partition *part)
drivers/spi/spi-intel.c
1236
memset(part, 0, sizeof(*part));
drivers/spi/spi-intel.c
1239
part->size = 4096;
drivers/spi/spi-intel.c
1240
part->name = "BIOS";
drivers/spi/spi-intel.c
1266
part->mask_flags |= MTD_WRITEABLE;
drivers/spi/spi-intel.c
1271
if (end > part->size)
drivers/spi/spi-intel.c
1272
part->size = end;
drivers/spi/spi-intel.c
1279
if (ispi->chip0_size && part->size > ispi->chip0_size)
drivers/spi/spi-intel.c
1280
part->size = MTDPART_SIZ_FULL;
drivers/thermal/qcom/tsens.c
197
u8 part = 32 - cell->shift;
drivers/thermal/qcom/tsens.c
200
val |= data[cell->idx + 1] << part;
drivers/usb/gadget/function/uvc_video.c
141
unsigned int sg_left, part = 0;
drivers/usb/gadget/function/uvc_video.c
168
part = min_t(unsigned int, len, sg_left);
drivers/usb/gadget/function/uvc_video.c
170
sg_set_page(iter, sg_page(buf->sg), part, buf->offset);
drivers/usb/gadget/function/uvc_video.c
172
if (part == sg_left) {
drivers/usb/gadget/function/uvc_video.c
176
buf->offset += part;
drivers/usb/gadget/function/uvc_video.c
178
len -= part;
drivers/vhost/vringh.c
268
size_t part, len = sizeof(struct vring_desc);
drivers/vhost/vringh.c
274
part = len;
drivers/vhost/vringh.c
277
if (!rcheck(vrh, addr, &part, range, getrange))
drivers/vhost/vringh.c
280
err = copy(vrh, dst, src, part);
drivers/vhost/vringh.c
284
dst += part;
drivers/vhost/vringh.c
285
src += part;
drivers/vhost/vringh.c
286
len -= part;
drivers/vhost/vringh.c
467
u16 part = vrh->vring.num - off;
drivers/vhost/vringh.c
468
err = putused(vrh, &used_ring->ring[off], used, part);
drivers/vhost/vringh.c
470
err = putused(vrh, &used_ring->ring[0], used + part,
drivers/vhost/vringh.c
471
num_used - part);
fs/btrfs/accessors.c
108
memcpy(kaddr, lebytes, part); \
fs/btrfs/accessors.c
110
memcpy(kaddr, lebytes + part, sizeof(u##bits) - part); \
fs/btrfs/accessors.c
60
const int part = eb->folio_size - oif; \
fs/btrfs/accessors.c
68
likely(sizeof(u##bits) <= part)) \
fs/btrfs/accessors.c
78
part, sizeof(u##bits)); \
fs/btrfs/accessors.c
90
const int part = eb->folio_size - oif; \
fs/btrfs/accessors.c
98
likely(sizeof(u##bits) <= part)) { \
fs/cachefiles/io.c
677
size_t part = len % CACHEFILES_DIO_BLOCK_SIZE;
fs/cachefiles/io.c
678
size_t need = CACHEFILES_DIO_BLOCK_SIZE - part;
fs/cachefiles/io.c
680
if (part && stream->submit_extendable_to >= need) {
fs/char_dev.c
454
static struct kobject *exact_match(dev_t dev, int *part, void *data)
fs/char_dev.c
663
static struct kobject *base_probe(dev_t dev, int *part, void *data)
fs/hfs/hfs_fs.h
125
int session, part;
fs/hfs/part_tbl.c
82
(HFS_SB(sb)->part < 0 || HFS_SB(sb)->part == i)) {
fs/hfs/part_tbl.c
98
(HFS_SB(sb)->part < 0 || HFS_SB(sb)->part == i)) {
fs/hfs/super.c
156
if (sbi->part >= 0)
fs/hfs/super.c
157
seq_printf(seq, ",part=%u", sbi->part);
fs/hfs/super.c
252
hsb->part = result.uint_32;
fs/hfs/super.c
430
hsb->part = -1;
fs/hfsplus/hfsplus_fs.h
155
int part, session;
fs/hfsplus/options.c
108
sbi->part = result.uint_32;
fs/hfsplus/options.c
158
if (sbi->part >= 0)
fs/hfsplus/options.c
159
seq_printf(seq, ",part=%u", sbi->part);
fs/hfsplus/options.c
58
opts->part = -1;
fs/hfsplus/part_tbl.c
102
(sbi->part < 0 || sbi->part == i)) {
fs/hfsplus/part_tbl.c
81
(sbi->part < 0 || sbi->part == i)) {
fs/netfs/buffered_read.c
409
size_t part;
fs/netfs/buffered_read.c
450
part = min_t(size_t, to - off, PAGE_SIZE);
fs/netfs/buffered_read.c
451
bvec_set_folio(&bvec[i++], sink, part, 0);
fs/netfs/buffered_read.c
452
off += part;
fs/netfs/buffered_write.c
158
size_t part; /* Bytes to write to folio */
fs/netfs/buffered_write.c
162
part = min(max_chunk - offset, iov_iter_count(iter));
fs/netfs/buffered_write.c
176
if (unlikely(fault_in_iov_iter_readable(iter, part) == part))
fs/netfs/buffered_write.c
179
folio = netfs_grab_folio_for_write(mapping, pos, part);
fs/netfs/buffered_write.c
188
part = min_t(size_t, flen - offset, part);
fs/netfs/buffered_write.c
222
copied = copy_folio_from_iter_atomic(folio, offset, part, iter);
fs/netfs/buffered_write.c
236
copied = copy_folio_from_iter_atomic(folio, offset, part, iter);
fs/netfs/buffered_write.c
247
if (!maybe_trouble && offset == 0 && part >= flen) {
fs/netfs/buffered_write.c
248
copied = copy_folio_from_iter_atomic(folio, offset, part, iter);
fs/netfs/buffered_write.c
251
if (unlikely(copied < part)) {
fs/netfs/buffered_write.c
276
ret = netfs_prefetch_for_write(file, folio, offset, part);
fs/netfs/buffered_write.c
283
copied = copy_folio_from_iter_atomic(folio, offset, part, iter);
fs/netfs/buffered_write.c
295
copied = copy_folio_from_iter_atomic(folio, offset, part, iter);
fs/netfs/buffered_write.c
325
copied = copy_folio_from_iter_atomic(folio, offset, part, iter);
fs/netfs/buffered_write.c
39
loff_t pos, size_t part)
fs/netfs/buffered_write.c
45
fgp_flags |= fgf_set_order(pos % PAGE_SIZE + part);
fs/netfs/read_pgpriv2.c
71
ssize_t part;
fs/netfs/read_pgpriv2.c
77
part = netfs_advance_write(creq, cache, fpos + cache->submit_off,
fs/netfs/read_pgpriv2.c
79
cache->submit_off += part;
fs/netfs/read_pgpriv2.c
80
if (part > cache->submit_len)
fs/netfs/read_pgpriv2.c
83
cache->submit_len -= part;
fs/netfs/read_retry.c
145
part = umin(len, stream->sreq_max_len);
fs/netfs/read_retry.c
147
part = netfs_limit_iter(&source, 0, part, stream->sreq_max_segs);
fs/netfs/read_retry.c
148
subreq->len = subreq->transferred + part;
fs/netfs/read_retry.c
150
iov_iter_truncate(&subreq->io_iter, part);
fs/netfs/read_retry.c
151
iov_iter_advance(&source, part);
fs/netfs/read_retry.c
152
len -= part;
fs/netfs/read_retry.c
153
start += part;
fs/netfs/read_retry.c
213
part = netfs_limit_iter(&source, 0, part, stream->sreq_max_segs);
fs/netfs/read_retry.c
222
part = umin(len, stream->sreq_max_len);
fs/netfs/read_retry.c
223
subreq->len = subreq->transferred + part;
fs/netfs/read_retry.c
225
iov_iter_truncate(&subreq->io_iter, part);
fs/netfs/read_retry.c
226
iov_iter_advance(&source, part);
fs/netfs/read_retry.c
228
len -= part;
fs/netfs/read_retry.c
229
start += part;
fs/netfs/read_retry.c
79
size_t part;
fs/netfs/write_issue.c
282
size_t part;
fs/netfs/write_issue.c
300
part = umin(stream->sreq_max_len - subreq->len, len);
fs/netfs/write_issue.c
301
_debug("part %zx/%zx %zx/%zx", subreq->len, stream->sreq_max_len, part, len);
fs/netfs/write_issue.c
302
subreq->len += part;
fs/netfs/write_issue.c
304
stream->submit_extendable_to -= part;
fs/netfs/write_issue.c
313
return part;
fs/netfs/write_issue.c
471
ssize_t part;
fs/netfs/write_issue.c
497
part = netfs_advance_write(wreq, stream, fpos + stream->submit_off,
fs/netfs/write_issue.c
499
stream->submit_off += part;
fs/netfs/write_issue.c
500
if (part > stream->submit_len)
fs/netfs/write_issue.c
503
stream->submit_len -= part;
fs/netfs/write_issue.c
504
if (part > 0)
fs/netfs/write_issue.c
765
ssize_t part;
fs/netfs/write_issue.c
791
part = netfs_advance_write(wreq, stream, fpos + stream->submit_off,
fs/netfs/write_issue.c
793
stream->submit_off += part;
fs/netfs/write_issue.c
794
if (part > stream->submit_len)
fs/netfs/write_issue.c
797
stream->submit_len -= part;
fs/netfs/write_issue.c
798
if (part > 0)
fs/netfs/write_issue.c
860
size_t part = umin(folioq_folio_size(fq, slot), size);
fs/netfs/write_issue.c
867
size -= part;
fs/netfs/write_retry.c
107
part = umin(len, stream->sreq_max_len);
fs/netfs/write_retry.c
109
part = netfs_limit_iter(&source, 0, part, stream->sreq_max_segs);
fs/netfs/write_retry.c
110
subreq->len = part;
fs/netfs/write_retry.c
112
len -= part;
fs/netfs/write_retry.c
113
start += part;
fs/netfs/write_retry.c
176
part = umin(len, stream->sreq_max_len);
fs/netfs/write_retry.c
177
subreq->len = subreq->transferred + part;
fs/netfs/write_retry.c
178
len -= part;
fs/netfs/write_retry.c
179
start += part;
fs/netfs/write_retry.c
59
size_t part;
fs/orangefs/dir.c
135
struct orangefs_dir_part *part, *new;
fs/orangefs/dir.c
139
part = od->part;
fs/orangefs/dir.c
140
while (part) {
fs/orangefs/dir.c
142
if (part->next)
fs/orangefs/dir.c
143
part = part->next;
fs/orangefs/dir.c
152
if (!od->part)
fs/orangefs/dir.c
153
od->part = new;
fs/orangefs/dir.c
155
part->next = new;
fs/orangefs/dir.c
18
struct orangefs_dir_part *part;
fs/orangefs/dir.c
189
static int fill_from_part(struct orangefs_dir_part *part,
fs/orangefs/dir.c
200
if (i > part->len)
fs/orangefs/dir.c
210
while (i < part->len) {
fs/orangefs/dir.c
211
if (part->len < i + sizeof *len)
fs/orangefs/dir.c
213
len = (void *)part + offset + i;
fs/orangefs/dir.c
220
if (part->len < i + padlen + sizeof *khandle)
fs/orangefs/dir.c
222
s = (void *)part + offset + i + sizeof *len;
fs/orangefs/dir.c
225
khandle = (void *)part + offset + i + padlen;
fs/orangefs/dir.c
232
BUG_ON(i > part->len);
fs/orangefs/dir.c
243
struct orangefs_dir_part *part;
fs/orangefs/dir.c
248
part = od->part;
fs/orangefs/dir.c
249
while (part->next && count) {
fs/orangefs/dir.c
251
part = part->next;
fs/orangefs/dir.c
259
while (part && part->len) {
fs/orangefs/dir.c
261
r = fill_from_part(part, ctx);
fs/orangefs/dir.c
274
part = part->next;
fs/orangefs/dir.c
289
struct orangefs_dir_part *part = od->part;
fs/orangefs/dir.c
290
while (part) {
fs/orangefs/dir.c
291
struct orangefs_dir_part *next = part->next;
fs/orangefs/dir.c
292
vfree(part);
fs/orangefs/dir.c
293
part = next;
fs/orangefs/dir.c
296
od->part = NULL;
fs/orangefs/dir.c
371
od->part = NULL;
fs/orangefs/dir.c
380
struct orangefs_dir_part *part = od->part;
fs/orangefs/dir.c
381
while (part) {
fs/orangefs/dir.c
382
struct orangefs_dir_part *next = part->next;
fs/orangefs/dir.c
383
vfree(part);
fs/orangefs/dir.c
384
part = next;
fs/pstore/platform.c
284
unsigned int part = 1;
fs/pstore/platform.c
316
record.part = part;
fs/pstore/platform.c
324
oopscount, part);
fs/pstore/platform.c
366
part++;
fs/pstore/ram.c
363
if (record->part != 1)
fs/pstore/zone.c
806
if (record->part != 1)
fs/smb/client/compress.c
166
size_t part = umin(umin(iov_iter_count(&iter), SZ_2K), max);
fs/smb/client/compress.c
169
n = copy_from_iter(sample + s, part, &iter);
fs/smb/client/compress.c
170
if (n != part)
fs/smb/client/smb2ops.c
4526
size_t part = folioq_folio_size(buffer, s);
fs/smb/client/smb2ops.c
4528
part = umin(part, size);
fs/smb/client/smb2ops.c
4530
if (copy_folio_from_iter(folio, 0, part, iter) != part)
fs/smb/client/smb2ops.c
4532
size -= part;
fs/smb/client/smbdirect.c
3270
size_t part = umin(maxsize, fsize - offset);
fs/smb/client/smbdirect.c
3272
if (!smb_set_sge(rdma, folio_page(folio, 0), offset, part))
fs/smb/client/smbdirect.c
3275
offset += part;
fs/smb/client/smbdirect.c
3276
ret += part;
fs/smb/client/smbdirect.c
3277
maxsize -= part;
fs/udf/super.c
2513
unsigned int part = sbi->s_partition;
fs/udf/super.c
2514
int ptype = sbi->s_partmaps[part].s_partition_type;
fs/udf/super.c
2517
part = sbi->s_partmaps[part].s_type_specific.s_metadata.
fs/udf/super.c
2531
if (le32_to_cpu(lvid->numOfPartitions) > part) {
fs/udf/super.c
2533
lvid->freeSpaceTable[part]);
fs/udf/super.c
2542
map = &sbi->s_partmaps[part];
include/linux/bcm963xx_nvram.h
51
#define BCM963XX_NVRAM_NAND_PART_OFFSET(nvram, part) \
include/linux/bcm963xx_nvram.h
52
bcm963xx_nvram_nand_part_offset(nvram, BCM963XX_NVRAM_NAND_PART_ ##part)
include/linux/bcm963xx_nvram.h
56
enum bcm963xx_nvram_nand_part part)
include/linux/bcm963xx_nvram.h
58
return nvram->nand_part_offset[part] * SZ_1K;
include/linux/bcm963xx_nvram.h
61
#define BCM963XX_NVRAM_NAND_PART_SIZE(nvram, part) \
include/linux/bcm963xx_nvram.h
62
bcm963xx_nvram_nand_part_size(nvram, BCM963XX_NVRAM_NAND_PART_ ##part)
include/linux/bcm963xx_nvram.h
66
enum bcm963xx_nvram_nand_part part)
include/linux/bcm963xx_nvram.h
68
return nvram->nand_part_size[part] * SZ_1K;
include/linux/clk/ti.h
255
int (*cm_wait_module_ready)(u8 part, s16 prcm_mod, u16 idlest_reg,
include/linux/iov_iter.h
117
size_t offset = p->bv_offset + skip, part;
include/linux/iov_iter.h
120
part = min3(len,
include/linux/iov_iter.h
123
remain = step(kaddr + offset % PAGE_SIZE, progress, part, priv, priv2);
include/linux/iov_iter.h
125
consumed = part - remain;
include/linux/iov_iter.h
163
size_t part, remain = 0, consumed;
include/linux/iov_iter.h
173
part = umin(len, PAGE_SIZE - skip % PAGE_SIZE);
include/linux/iov_iter.h
174
remain = step(base, progress, part, priv, priv2);
include/linux/iov_iter.h
176
consumed = part - remain;
include/linux/iov_iter.h
215
size_t remain, consumed, offset, part, flen;
include/linux/iov_iter.h
230
part = min_t(size_t, flen,
include/linux/iov_iter.h
232
remain = step(base, progress, part, priv, priv2);
include/linux/iov_iter.h
235
consumed = part - remain;
include/linux/iov_iter.h
49
size_t part = min(len, p->iov_len - skip);
include/linux/iov_iter.h
51
if (likely(part)) {
include/linux/iov_iter.h
52
remain = step(p->iov_base + skip, progress, part, priv, priv2);
include/linux/iov_iter.h
53
consumed = part - remain;
include/linux/iov_iter.h
83
size_t part = min(len, p->iov_len - skip);
include/linux/iov_iter.h
85
if (likely(part)) {
include/linux/iov_iter.h
86
remain = step(p->iov_base + skip, progress, part, priv, priv2);
include/linux/iov_iter.h
87
consumed = part - remain;
include/linux/mmc/card.h
376
struct mmc_part part[MMC_NUM_PHY_PARTITION]; /* physical partitions */
include/linux/mtd/mtd.h
399
struct mtd_part part;
include/linux/mtd/mtd.h
414
ofs += mtd->part.offset;
include/linux/part_stat.h
28
#define part_stat_get_cpu(part, field, cpu) \
include/linux/part_stat.h
29
(per_cpu_ptr((part)->bd_stats, (cpu))->field)
include/linux/part_stat.h
31
#define part_stat_get(part, field) \
include/linux/part_stat.h
32
part_stat_get_cpu(part, field, smp_processor_id())
include/linux/part_stat.h
34
#define part_stat_read(part, field) \
include/linux/part_stat.h
36
TYPEOF_UNQUAL((part)->bd_stats->field) res = 0; \
include/linux/part_stat.h
39
res += per_cpu_ptr((part)->bd_stats, _cpu)->field; \
include/linux/part_stat.h
43
static inline void part_stat_set_all(struct block_device *part, int value)
include/linux/part_stat.h
48
memset(per_cpu_ptr(part->bd_stats, i), value,
include/linux/part_stat.h
52
#define part_stat_read_accum(part, field) \
include/linux/part_stat.h
53
(part_stat_read(part, field[STAT_READ]) + \
include/linux/part_stat.h
54
part_stat_read(part, field[STAT_WRITE]) + \
include/linux/part_stat.h
55
part_stat_read(part, field[STAT_DISCARD]))
include/linux/part_stat.h
57
#define __part_stat_add(part, field, addnd) \
include/linux/part_stat.h
58
__this_cpu_add((part)->bd_stats->field, addnd)
include/linux/part_stat.h
60
#define part_stat_add(part, field, addnd) do { \
include/linux/part_stat.h
61
__part_stat_add((part), field, addnd); \
include/linux/part_stat.h
62
if (bdev_is_partition(part)) \
include/linux/part_stat.h
63
__part_stat_add(bdev_whole(part), field, addnd); \
include/linux/part_stat.h
66
#define part_stat_dec(part, field) \
include/linux/part_stat.h
67
part_stat_add(part, field, -1)
include/linux/part_stat.h
68
#define part_stat_inc(part, field) \
include/linux/part_stat.h
69
part_stat_add(part, field, 1)
include/linux/part_stat.h
70
#define part_stat_sub(part, field, subnd) \
include/linux/part_stat.h
71
part_stat_add(part, field, -subnd)
include/linux/part_stat.h
73
#define part_stat_local_dec(part, field) \
include/linux/part_stat.h
74
local_dec(&(part_stat_get(part, field)))
include/linux/part_stat.h
75
#define part_stat_local_inc(part, field) \
include/linux/part_stat.h
76
local_inc(&(part_stat_get(part, field)))
include/linux/part_stat.h
77
#define part_stat_local_read(part, field) \
include/linux/part_stat.h
78
local_read(&(part_stat_get(part, field)))
include/linux/part_stat.h
79
#define part_stat_local_read_cpu(part, field, cpu) \
include/linux/part_stat.h
80
local_read(&(part_stat_get_cpu(part, field, cpu)))
include/linux/part_stat.h
82
unsigned int bdev_count_inflight(struct block_device *part);
include/linux/pstore.h
84
unsigned int part;
include/uapi/linux/switchtec_ioctl.h
69
__u32 part[48];
include/uapi/linux/switchtec_ioctl.h
78
__u32 part[48];
kernel/trace/trace_events_user.c
1177
char *part, *type, *name;
kernel/trace/trace_events_user.c
1227
while ((part = strsep(&field, " ")) != NULL) {
kernel/trace/trace_events_user.c
1230
type = part;
kernel/trace/trace_events_user.c
1233
name = part;
kernel/trace/trace_events_user.c
1239
if (kstrtou32(part, 10, &size))
lib/iov_iter.c
1528
size_t part = PAGE_SIZE - offset % PAGE_SIZE;
lib/iov_iter.c
1531
part = umin(part, umin(maxsize - extracted, fsize - offset));
lib/iov_iter.c
1532
i->count -= part;
lib/iov_iter.c
1533
i->iov_offset += part;
lib/iov_iter.c
1534
extracted += part;
lib/iov_iter.c
938
size_t part = PAGE_SIZE - offset % PAGE_SIZE;
lib/iov_iter.c
941
part = umin(part, umin(maxsize - extracted, fsize - offset));
lib/iov_iter.c
942
count -= part;
lib/iov_iter.c
943
iov_offset += part;
lib/iov_iter.c
944
extracted += part;
lib/scatterlist.c
1300
size_t part = umin(maxsize - ret, fsize - offset);
lib/scatterlist.c
1302
sg_set_page(sg, folio_page(folio, 0), part, offset);
lib/scatterlist.c
1306
offset += part;
lib/scatterlist.c
1307
ret += part;
lib/tests/kunit_iov_iter.c
743
ssize_t part = min_t(ssize_t, len, PAGE_SIZE - offset0);
lib/tests/kunit_iov_iter.c
746
KUNIT_ASSERT_GE(test, part, 0);
lib/tests/kunit_iov_iter.c
758
from += part;
lib/tests/kunit_iov_iter.c
759
len -= part;
lib/tests/kunit_iov_iter.c
821
ssize_t part = min_t(ssize_t, len, PAGE_SIZE - offset0);
lib/tests/kunit_iov_iter.c
824
KUNIT_ASSERT_GE(test, part, 0);
lib/tests/kunit_iov_iter.c
836
from += part;
lib/tests/kunit_iov_iter.c
837
len -= part;
lib/tests/kunit_iov_iter.c
904
ssize_t part = min_t(ssize_t, len, PAGE_SIZE - offset0);
lib/tests/kunit_iov_iter.c
907
KUNIT_ASSERT_GE(test, part, 0);
lib/tests/kunit_iov_iter.c
913
from += part;
lib/tests/kunit_iov_iter.c
914
len -= part;
lib/tests/kunit_iov_iter.c
982
ssize_t part = min_t(ssize_t, len, PAGE_SIZE - offset0);
lib/tests/kunit_iov_iter.c
985
KUNIT_ASSERT_GE(test, part, 0);
lib/tests/kunit_iov_iter.c
991
from += part;
lib/tests/kunit_iov_iter.c
992
len -= part;
mm/filemap.c
3016
size_t part = min_t(size_t, PAGE_SIZE - offset, size - spliced);
mm/filemap.c
3022
.len = part,
mm/filemap.c
3027
spliced += part;
mm/shmem.c
3560
size_t total_spliced = 0, used, npages, n, part;
mm/shmem.c
3620
part = min_t(loff_t, isize - *ppos, size);
mm/shmem.c
3639
n = splice_folio_into_pipe(pipe, folio, *ppos, part);
mm/shmem.c
3643
n = splice_zeropage_into_pipe(pipe, *ppos, part);
net/core/skbuff.c
7374
size_t part = min_t(size_t, PAGE_SIZE - off, len);
net/core/skbuff.c
7380
ret = skb_append_pagefrags(skb, page, off, part,
net/core/skbuff.c
7388
skb_splice_csum_page(skb, page, off, part);
net/core/skbuff.c
7391
spliced += part;
net/core/skbuff.c
7392
maxsize -= part;
net/core/skbuff.c
7393
len -= part;
net/rds/transport.c
150
unsigned int part;
net/rds/transport.c
161
part = trans->stats_info_copy(iter, avail);
net/rds/transport.c
162
avail -= min(avail, part);
net/rds/transport.c
163
total += part;
net/rxrpc/output.c
333
size_t part = umin(probe_mtu - len, PAGE_SIZE);
net/rxrpc/output.c
336
kv[nr_kv].iov_len = part;
net/rxrpc/output.c
337
len += part;
net/rxrpc/rxgk.c
288
size_t shdr, alloc, limit, part, offset, gap;
net/rxrpc/rxgk.c
313
part = remain;
net/rxrpc/rxgk.c
315
shdr + part, &offset);
net/rxrpc/rxgk.c
318
part = limit - shdr;
net/rxrpc/rxgk.c
332
txb->space = part;
net/rxrpc/rxkad.c
151
size_t shdr, alloc, limit, part;
net/rxrpc/rxkad.c
169
part = remain;
net/rxrpc/rxkad.c
170
alloc = round_up(shdr + part, RXKAD_ALIGN);
net/rxrpc/rxkad.c
172
part = limit;
net/rxrpc/rxkad.c
183
txb->space = part;
net/tls/tls_sw.c
1012
ssize_t part;
net/tls/tls_sw.c
1015
part = iov_iter_extract_pages(&msg->msg_iter, &pages,
net/tls/tls_sw.c
1017
if (part <= 0)
net/tls/tls_sw.c
1018
return part ?: -EIO;
net/tls/tls_sw.c
1021
iov_iter_revert(&msg->msg_iter, part);
net/tls/tls_sw.c
1025
sk_msg_page_add(msg_pl, page, part, off);
net/tls/tls_sw.c
1028
sk_mem_charge(sk, part);
net/tls/tls_sw.c
1029
*copied += part;
net/tls/tls_sw.c
1030
try_to_copy -= part;
sound/soc/codecs/cs35l41.c
1089
dsp->part = "cs35l41";
sound/soc/codecs/cs35l45.c
1351
dsp->part = "cs35l45";
sound/soc/codecs/cs35l56.c
1251
cs35l56->dsp.part = kasprintf(GFP_KERNEL, "cs35l%02x", cs35l56->base.type);
sound/soc/codecs/cs35l56.c
1252
if (!cs35l56->dsp.part)
sound/soc/codecs/cs35l56.c
1321
kfree(cs35l56->dsp.part);
sound/soc/codecs/cs35l56.c
1322
cs35l56->dsp.part = NULL;
sound/soc/codecs/cs47l15.c
1409
cs47l15->core.adsp[0].part = "cs47l15";
sound/soc/codecs/cs47l24.c
1239
cs47l24->core.adsp[i].part = "cs47l24";
sound/soc/codecs/cs47l35.c
1692
cs47l35->core.adsp[i].part = "cs47l35";
sound/soc/codecs/cs47l85.c
2638
cs47l85->core.adsp[i].part = "cs47l85";
sound/soc/codecs/cs47l90.c
2549
cs47l90->core.adsp[i].part = "cs47l90";
sound/soc/codecs/cs47l92.c
2012
cs47l92->core.adsp[0].part = "cs47l92";
sound/soc/codecs/cs48l32.c
3680
dsp->part = "cs48l32";
sound/soc/codecs/wm2200.c
2207
wm2200->dsp[i].part = "wm2200";
sound/soc/codecs/wm5102.c
2064
wm5102->core.adsp[0].part = "wm5102";
sound/soc/codecs/wm5110.c
2418
wm5110->core.adsp[i].part = "wm5110";
sound/soc/codecs/wm_adsp.c
1609
buf->name = kasprintf(GFP_KERNEL, "%s-dsp-%s", dsp->part,
sound/soc/codecs/wm_adsp.c
737
*filename = kasprintf(GFP_KERNEL, "%s%s-%s-%s-%s-%s.%s", dir, dsp->part,
sound/soc/codecs/wm_adsp.c
741
*filename = kasprintf(GFP_KERNEL, "%s%s-%s-%s-%s.%s", dir, dsp->part,
sound/soc/codecs/wm_adsp.c
745
*filename = kasprintf(GFP_KERNEL, "%s%s-%s-%s.%s", dir, dsp->part, fwf,
sound/soc/codecs/wm_adsp.c
854
cirrus_dir, dsp->part,
sound/soc/codecs/wm_adsp.h
29
const char *part;
sound/soc/intel/atom/sst-atom-controls.c
1055
cmd.param.part.cfg.rate = 2; /* 48khz */
sound/soc/intel/atom/sst-atom-controls.c
1057
cmd.param.part.cfg.format = ids->format; /* stereo/Mono */
sound/soc/intel/atom/sst-atom-controls.c
1058
cmd.param.part.cfg.s_length = 1; /* 24bit left justified */
sound/soc/intel/atom/sst-atom-controls.h
425
} part;
sound/soc/intel/atom/sst-mfld-dsp.h
200
} part;
sound/soc/intel/atom/sst-mfld-dsp.h
223
} part;
sound/soc/intel/atom/sst/sst.c
105
sst_is_process_reply(header.p.header_high.part.msg_id);
sound/soc/intel/atom/sst/sst.c
59
if (isr.part.done_interrupt) {
sound/soc/intel/atom/sst/sst.c
64
header.p.header_high.part.done = 0;
sound/soc/intel/atom/sst/sst.c
68
isr.part.done_interrupt = 1;
sound/soc/intel/atom/sst/sst.c
77
if (isr.part.busy_interrupt) {
sound/soc/intel/atom/sst/sst.c
81
imr.part.busy_interrupt = 1;
sound/soc/intel/atom/sst/sst.c
86
if (sst_create_ipc_msg(&msg, header.p.header_high.part.large)) {
sound/soc/intel/atom/sst/sst.c
91
if (header.p.header_high.part.large) {
sound/soc/intel/atom/sst/sst.h
107
} part;
sound/soc/intel/atom/sst/sst.h
127
} part;
sound/soc/intel/atom/sst/sst.h
136
} part;
sound/soc/intel/atom/sst/sst.h
145
} part;
sound/soc/intel/atom/sst/sst.h
80
} part;
sound/soc/intel/atom/sst/sst.h
91
} part;
sound/soc/intel/atom/sst/sst_ipc.c
128
while (header.p.header_high.part.busy) {
sound/soc/intel/atom/sst/sst_ipc.c
148
if (header.p.header_high.part.busy) {
sound/soc/intel/atom/sst/sst_ipc.c
164
if (msg->mrfld_header.p.header_high.part.large)
sound/soc/intel/atom/sst/sst_ipc.c
190
isr.part.busy_interrupt = 1;
sound/soc/intel/atom/sst/sst_ipc.c
196
clear_ipc.p.header_high.part.busy = 0;
sound/soc/intel/atom/sst/sst_ipc.c
197
clear_ipc.p.header_high.part.done = 1;
sound/soc/intel/atom/sst/sst_ipc.c
201
imr.part.busy_interrupt = 0;
sound/soc/intel/atom/sst/sst_ipc.c
334
drv_id = msg_high.part.drv_id;
sound/soc/intel/atom/sst/sst_ipc.c
344
if (msg_high.part.result && !msg_high.part.large) {
sound/soc/intel/atom/sst/sst_ipc.c
347
sst_wake_up_block(sst_drv_ctx, msg_high.part.result,
sound/soc/intel/atom/sst/sst_ipc.c
348
msg_high.part.drv_id,
sound/soc/intel/atom/sst/sst_ipc.c
349
msg_high.part.msg_id, NULL, 0);
sound/soc/intel/atom/sst/sst_ipc.c
358
if (msg_high.part.large) {
sound/soc/intel/atom/sst/sst_ipc.c
365
if (sst_wake_up_block(sst_drv_ctx, msg_high.part.result,
sound/soc/intel/atom/sst/sst_ipc.c
366
msg_high.part.drv_id,
sound/soc/intel/atom/sst/sst_ipc.c
367
msg_high.part.msg_id, data, msg_low))
sound/soc/intel/atom/sst/sst_ipc.c
370
sst_wake_up_block(sst_drv_ctx, msg_high.part.result,
sound/soc/intel/atom/sst/sst_ipc.c
371
msg_high.part.drv_id,
sound/soc/intel/atom/sst/sst_ipc.c
372
msg_high.part.msg_id, NULL, 0);
sound/soc/intel/atom/sst/sst_loader.c
97
csr.part.xt_snoop = 1;
sound/soc/intel/atom/sst/sst_pvt.c
217
msg->mrfld_header.p.header_high.part.res_rqd = !sync;
sound/soc/intel/atom/sst/sst_pvt.c
221
msg->mrfld_header.p.header_high.part.res_rqd);
sound/soc/intel/atom/sst/sst_pvt.c
272
header->p.header_high.part.msg_id = msg;
sound/soc/intel/atom/sst/sst_pvt.c
273
header->p.header_high.part.task_id = task_id;
sound/soc/intel/atom/sst/sst_pvt.c
274
header->p.header_high.part.large = large;
sound/soc/intel/atom/sst/sst_pvt.c
275
header->p.header_high.part.drv_id = drv_id;
sound/soc/intel/atom/sst/sst_pvt.c
276
header->p.header_high.part.done = 0;
sound/soc/intel/atom/sst/sst_pvt.c
277
header->p.header_high.part.busy = 1;
sound/soc/intel/atom/sst/sst_pvt.c
278
header->p.header_high.part.res_rqd = 1;
sound/soc/intel/atom/sst/sst_stream.c
189
msg->mrfld_header.p.header_high.part.res_rqd = bytes_block;
sound/soc/tegra/tegra210_ahub.c
2054
int part;
sound/soc/tegra/tegra210_ahub.c
2059
for (part = 0; part < TEGRA210_XBAR_UPDATE_MAX_REG; part++) {
sound/soc/tegra/tegra210_ahub.c
2060
switch (reg & ~(part * TEGRA210_XBAR_PART1_RX)) {
sound/soc/tegra/tegra210_ahub.c
2082
int part;
sound/soc/tegra/tegra210_ahub.c
2087
for (part = 0; part < TEGRA186_XBAR_UPDATE_MAX_REG; part++) {
sound/soc/tegra/tegra210_ahub.c
2088
switch (reg & ~(part * TEGRA210_XBAR_PART1_RX)) {
sound/soc/tegra/tegra210_ahub.c
2109
int part;
sound/soc/tegra/tegra210_ahub.c
2111
for (part = 0; part < TEGRA264_XBAR_UPDATE_MAX_REG; part++) {
sound/soc/tegra/tegra210_ahub.c
2112
switch (reg & ~(part << 12)) {
tools/perf/util/bpf-filter.c
170
entry[i].part = expr->part;
tools/perf/util/bpf-filter.c
183
entry[i].part = group->part;
tools/perf/util/bpf-filter.c
613
int part,
tools/perf/util/bpf-filter.c
622
expr->part = part;
tools/perf/util/bpf-filter.h
14
int part;
tools/perf/util/bpf-filter.h
27
int part,
tools/perf/util/bpf-filter.y
36
int part;
tools/perf/util/bpf_skel/sample-filter.h
61
__u32 part; /* sub-sample type info when it has multiple values */
tools/perf/util/bpf_skel/sample_filter.bpf.c
113
if (entry->part)
tools/perf/util/bpf_skel/sample_filter.bpf.c
128
if (entry->part == 1)
tools/perf/util/bpf_skel/sample_filter.bpf.c
130
if (entry->part == 2)
tools/perf/util/bpf_skel/sample_filter.bpf.c
132
if (entry->part == 3)
tools/perf/util/bpf_skel/sample_filter.bpf.c
146
if (entry->part == 1)
tools/perf/util/bpf_skel/sample_filter.bpf.c
148
if (entry->part == 2)
tools/perf/util/bpf_skel/sample_filter.bpf.c
150
if (entry->part == 3) {
tools/perf/util/bpf_skel/sample_filter.bpf.c
156
if (entry->part == 4)
tools/perf/util/bpf_skel/sample_filter.bpf.c
158
if (entry->part == 5)
tools/perf/util/bpf_skel/sample_filter.bpf.c
160
if (entry->part == 6)
tools/perf/util/bpf_skel/sample_filter.bpf.c
162
if (entry->part == 7)
tools/perf/util/bpf_skel/sample_filter.bpf.c
164
if (entry->part == 8) {
tools/testing/cxl/test/cxl.c
1098
struct resource *res = &cxlds->part[i].res;
tools/testing/cxl/test/cxl.c
1099
struct cxl_dpa_perf *perf = &cxlds->part[i].perf;
tools/tracing/rtla/src/utils.h
89
#define ns_to_per(total, part) ((part * 100) / (double)total)
tools/usb/usbip/libsrc/vhci_driver.c
190
int part;
tools/usb/usbip/libsrc/vhci_driver.c
221
for (part = 0; part < 3; part++) {
tools/usb/usbip/libsrc/vhci_driver.c
222
end = strchr(start, delim[part]);
tools/usb/usbip/libsrc/vhci_driver.c
223
if (end == NULL || (end - start) > max_len[part]) {