num_locks
if (id < 0 || id >= hwlock->bank->num_locks) {
const struct hwspinlock_ops *ops, int base_id, int num_locks)
if (!bank || !ops || !dev || !num_locks || !ops->trylock ||
bank->num_locks = num_locks;
for (i = 0; i < num_locks; i++) {
for (i = 0; i < bank->num_locks; i++) {
int base_id, int num_locks)
ret = hwspin_lock_register(bank, dev, ops, base_id, num_locks);
int num_locks;
num_locks = i * 32; /* actual number of locks in this device */
bank = devm_kzalloc(&pdev->dev, struct_size(bank, lock, num_locks),
for (i = 0; i < num_locks; i++)
base_id, num_locks);
int num_locks, i, ret;
bank = devm_kzalloc(&pdev->dev, struct_size(bank, lock, num_locks),
for (i = 0, hwlock = &bank->lock[0]; i < num_locks; i++, hwlock++)
pdata->base_id, num_locks);
int i, num_locks = U8500_MAX_SEMAPHORE;
if (area->num_locks) {
unsigned int num_locks;
area->num_locks++;
area->num_locks++;
WARN_ON(area->num_locks == 0);
area->num_locks--;
static inline unsigned int lock_nr(struct dm_cell_key *key, unsigned int num_locks)
num_locks);
unsigned l = lock_nr(key, prison->num_locks);
unsigned l = lock_nr(&cell->key, prison->num_locks);
unsigned l = lock_nr(&cell->key, prison->num_locks);
unsigned l = lock_nr(&cell->key, prison->num_locks);
unsigned int num_locks;
struct prison_region regions[] __counted_by(num_locks);
unsigned int i, num_locks;
num_locks = dm_num_hash_locks();
prison = kzalloc_flex(*prison, regions, num_locks);
prison->num_locks = num_locks;
for (i = 0; i < prison->num_locks; i++) {
unsigned int num_locks;
num_locks = dm_num_hash_locks();
c = kzalloc(sizeof(*c) + (num_locks * sizeof(struct buffer_tree)), GFP_KERNEL);
cache_init(&c->cache, num_locks, (flags & DM_BUFIO_CLIENT_NO_SLEEP) != 0);
static inline unsigned int cache_index(sector_t block, unsigned int num_locks)
bc->num_locks = num_locks;
for (i = 0; i < bc->num_locks; i++) {
for (i = 0; i < bc->num_locks; i++)
for (i = 0; i < bc->num_locks; i++) {
unsigned int num_locks = roundup_pow_of_two(num_online_cpus()) << 1;
return min_t(unsigned int, num_locks, DM_HASH_LOCKS_MAX);
unsigned int num_locks)
return (h1 ^ h2) & (num_locks - 1);
u8 num_locks; // locks sent in this structure
BUG_ON(mres->num_locks > DLM_MAX_MIGRATABLE_LOCKS);
if (!mres->num_locks)
struct_size(mres, ml, mres->num_locks),
mres->num_locks = 0;
int lock_num = mres->num_locks;
mres->num_locks++;
if (mres->num_locks == DLM_MAX_MIGRATABLE_LOCKS)
mlog(0, "running %d locks for this lockres\n", mres->num_locks);
for (i=0; i<mres->num_locks; i++) {
BUG_ON(mres->num_locks != 1);
atomic_t num_locks;
cifs_stats_inc(&tcon->stats.cifs_stats.num_locks);
cifs_stats_inc(&tcon->stats.cifs_stats.num_locks);
atomic_set(&tcon->stats.cifs_stats.num_locks, 0);
atomic_read(&tcon->stats.cifs_stats.num_locks),
cifs_stats_inc(&tcon->stats.cifs_stats.num_locks);
const struct hwspinlock_ops *ops, int base_id, int num_locks);
int base_id, int num_locks);