Symbol: num_locks
drivers/hwspinlock/hwspinlock_core.c
411
if (id < 0 || id >= hwlock->bank->num_locks) {
drivers/hwspinlock/hwspinlock_core.c
519
const struct hwspinlock_ops *ops, int base_id, int num_locks)
drivers/hwspinlock/hwspinlock_core.c
524
if (!bank || !ops || !dev || !num_locks || !ops->trylock ||
drivers/hwspinlock/hwspinlock_core.c
533
bank->num_locks = num_locks;
drivers/hwspinlock/hwspinlock_core.c
535
for (i = 0; i < num_locks; i++) {
drivers/hwspinlock/hwspinlock_core.c
571
for (i = 0; i < bank->num_locks; i++) {
drivers/hwspinlock/hwspinlock_core.c
647
int base_id, int num_locks)
drivers/hwspinlock/hwspinlock_core.c
656
ret = hwspin_lock_register(bank, dev, ops, base_id, num_locks);
drivers/hwspinlock/hwspinlock_internal.h
61
int num_locks;
drivers/hwspinlock/omap_hwspinlock.c
112
num_locks = i * 32; /* actual number of locks in this device */
drivers/hwspinlock/omap_hwspinlock.c
114
bank = devm_kzalloc(&pdev->dev, struct_size(bank, lock, num_locks),
drivers/hwspinlock/omap_hwspinlock.c
119
for (i = 0; i < num_locks; i++)
drivers/hwspinlock/omap_hwspinlock.c
123
base_id, num_locks);
drivers/hwspinlock/omap_hwspinlock.c
79
int num_locks, i, ret;
drivers/hwspinlock/u8500_hsem.c
108
bank = devm_kzalloc(&pdev->dev, struct_size(bank, lock, num_locks),
drivers/hwspinlock/u8500_hsem.c
115
for (i = 0, hwlock = &bank->lock[0]; i < num_locks; i++, hwlock++)
drivers/hwspinlock/u8500_hsem.c
120
pdata->base_id, num_locks);
drivers/hwspinlock/u8500_hsem.c
91
int i, num_locks = U8500_MAX_SEMAPHORE;
drivers/iommu/iommufd/io_pagetable.c
770
if (area->num_locks) {
drivers/iommu/iommufd/io_pagetable.h
52
unsigned int num_locks;
drivers/iommu/iommufd/pages.c
2450
area->num_locks++;
drivers/iommu/iommufd/pages.c
2473
area->num_locks++;
drivers/iommu/iommufd/pages.c
2508
WARN_ON(area->num_locks == 0);
drivers/iommu/iommufd/pages.c
2509
area->num_locks--;
drivers/md/dm-bio-prison-v1.c
120
static inline unsigned int lock_nr(struct dm_cell_key *key, unsigned int num_locks)
drivers/md/dm-bio-prison-v1.c
123
num_locks);
drivers/md/dm-bio-prison-v1.c
182
unsigned l = lock_nr(key, prison->num_locks);
drivers/md/dm-bio-prison-v1.c
221
unsigned l = lock_nr(&cell->key, prison->num_locks);
drivers/md/dm-bio-prison-v1.c
244
unsigned l = lock_nr(&cell->key, prison->num_locks);
drivers/md/dm-bio-prison-v1.c
274
unsigned l = lock_nr(&cell->key, prison->num_locks);
drivers/md/dm-bio-prison-v1.c
28
unsigned int num_locks;
drivers/md/dm-bio-prison-v1.c
29
struct prison_region regions[] __counted_by(num_locks);
drivers/md/dm-bio-prison-v1.c
43
unsigned int i, num_locks;
drivers/md/dm-bio-prison-v1.c
46
num_locks = dm_num_hash_locks();
drivers/md/dm-bio-prison-v1.c
47
prison = kzalloc_flex(*prison, regions, num_locks);
drivers/md/dm-bio-prison-v1.c
50
prison->num_locks = num_locks;
drivers/md/dm-bio-prison-v1.c
52
for (i = 0; i < prison->num_locks; i++) {
drivers/md/dm-bufio.c
2502
unsigned int num_locks;
drivers/md/dm-bufio.c
2513
num_locks = dm_num_hash_locks();
drivers/md/dm-bufio.c
2514
c = kzalloc(sizeof(*c) + (num_locks * sizeof(struct buffer_tree)), GFP_KERNEL);
drivers/md/dm-bufio.c
2519
cache_init(&c->cache, num_locks, (flags & DM_BUFIO_CLIENT_NO_SLEEP) != 0);
drivers/md/dm-bufio.c
399
static inline unsigned int cache_index(sector_t block, unsigned int num_locks)
drivers/md/dm-bufio.c
547
bc->num_locks = num_locks;
drivers/md/dm-bufio.c
550
for (i = 0; i < bc->num_locks; i++) {
drivers/md/dm-bufio.c
566
for (i = 0; i < bc->num_locks; i++)
drivers/md/dm-bufio.c
951
for (i = 0; i < bc->num_locks; i++) {
drivers/md/dm.h
238
unsigned int num_locks = roundup_pow_of_two(num_online_cpus()) << 1;
drivers/md/dm.h
240
return min_t(unsigned int, num_locks, DM_HASH_LOCKS_MAX);
drivers/md/dm.h
247
unsigned int num_locks)
drivers/md/dm.h
252
return (h1 ^ h2) & (num_locks - 1);
fs/ocfs2/dlm/dlmcommon.h
552
u8 num_locks; // locks sent in this structure
fs/ocfs2/dlm/dlmrecovery.c
1105
BUG_ON(mres->num_locks > DLM_MAX_MIGRATABLE_LOCKS);
fs/ocfs2/dlm/dlmrecovery.c
1106
if (!mres->num_locks)
fs/ocfs2/dlm/dlmrecovery.c
1122
struct_size(mres, ml, mres->num_locks),
fs/ocfs2/dlm/dlmrecovery.c
1162
mres->num_locks = 0;
fs/ocfs2/dlm/dlmrecovery.c
1209
int lock_num = mres->num_locks;
fs/ocfs2/dlm/dlmrecovery.c
1222
mres->num_locks++;
fs/ocfs2/dlm/dlmrecovery.c
1224
if (mres->num_locks == DLM_MAX_MIGRATABLE_LOCKS)
fs/ocfs2/dlm/dlmrecovery.c
1800
mlog(0, "running %d locks for this lockres\n", mres->num_locks);
fs/ocfs2/dlm/dlmrecovery.c
1801
for (i=0; i<mres->num_locks; i++) {
fs/ocfs2/dlm/dlmrecovery.c
1806
BUG_ON(mres->num_locks != 1);
fs/smb/client/cifsglob.h
1234
atomic_t num_locks;
fs/smb/client/cifssmb.c
2184
cifs_stats_inc(&tcon->stats.cifs_stats.num_locks);
fs/smb/client/cifssmb.c
2257
cifs_stats_inc(&tcon->stats.cifs_stats.num_locks);
fs/smb/client/smb1ops.c
809
atomic_set(&tcon->stats.cifs_stats.num_locks, 0);
fs/smb/client/smb1ops.c
828
atomic_read(&tcon->stats.cifs_stats.num_locks),
fs/smb/client/smb2pdu.c
6267
cifs_stats_inc(&tcon->stats.cifs_stats.num_locks);
include/linux/hwspinlock.h
59
const struct hwspinlock_ops *ops, int base_id, int num_locks);
include/linux/hwspinlock.h
78
int base_id, int num_locks);