num_possible_nodes
nest_imc_refc = kzalloc_objs(*nest_imc_refc, num_possible_nodes());
for (i = 0; i < num_possible_nodes(); i++) {
numn = num_possible_nodes();
pr_info("UV: Found %d hubs, %d nodes, %d CPUs\n", uv_num_possible_blades(), num_possible_nodes(), num_possible_cpus());
bytes = sizeof(void *) * num_possible_nodes();
sgx_numa_nodes = kmalloc_objs(*sgx_numa_nodes, num_possible_nodes());
node_width = num_digits(num_possible_nodes()) + 1; /* + '#' */
if (num_possible_nodes() > 1 && dev_to_node(&pdev->dev) < 0) {
if (num_possible_nodes() > 1 && dev_to_node(&pdev->dev) < 0) {
if (num_possible_nodes() > 1 && dev_to_node(dev) < 0) {
if (num_possible_nodes() > 1 && dev_to_node(&pdev->dev) < 0) {
if (num_possible_nodes() > 1 && dev_to_node(&pdev->dev) < 0) {
if (num_possible_nodes() > 1 && dev_to_node(&pdev->dev) < 0) {
if (num_possible_nodes() > 1 && dev_to_node(&vdev->dev) < 0) {
num_possible_nodes() > 1 && !params->pages_addr && params->allow_override)
num_possible_nodes() <= 1) ||
num_possible_nodes() > 1)
if (num_possible_nodes() <= 1)
node_affinity.num_possible_nodes = num_possible_nodes();
hfi1_per_node_cntr = kcalloc(node_affinity.num_possible_nodes,
for (node = 0; node < node_affinity.num_possible_nodes; node++)
int num_possible_nodes;
node_affinity.num_possible_nodes);
for (i = 0; i < node_affinity.num_possible_nodes; i++) {
*group_cnt = num_possible_nodes();
size += num_possible_nodes() * sizeof(struct nvme_ns *);
hv_dev->desc.virtual_numa_node < num_possible_nodes())
shost->nr_hw_queues = min_t(int, 2 * num_possible_nodes(),
if (num_possible_nodes() > 1)
num_possible_nodes());
#define BPF_PROG_PACK_SIZE (SZ_2M * num_possible_nodes())
hashsize /= num_possible_nodes();
hashsize, size, num_possible_nodes(), size * num_possible_nodes() / 1024,
unsigned int nnodes = num_possible_nodes();