KVM_MAX_VCPUS
return vgic_present ? kvm_vgic_get_max_vcpus() : KVM_MAX_VCPUS;
r = KVM_MAX_VCPUS;
r = min_t(unsigned int, num_online_cpus(), KVM_MAX_VCPUS);
r = KVM_MAX_VCPUS;
int block = (id / KVM_MAX_VCPUS) * (MAX_SMT_THREADS / stride);
packed_id = (id % KVM_MAX_VCPUS) + block_offsets[block];
if (WARN_ONCE(packed_id >= KVM_MAX_VCPUS, "VCPU ID packing failed"))
#define KVM_MAX_VCPU_IDS KVM_MAX_VCPUS
if (id >= (KVM_MAX_VCPUS * kvm->arch.emul_smt_mode)) {
else if (nr_servers > KVM_MAX_VCPUS)
xive->nr_servers = KVM_MAX_VCPUS;
xive->nr_servers = KVM_MAX_VCPUS;
xive->nr_servers = KVM_MAX_VCPUS;
r = min_t(unsigned int, num_present_cpus(), KVM_MAX_VCPUS);
r = min_t(unsigned int, num_online_cpus(), KVM_MAX_VCPUS);
r = KVM_MAX_VCPUS;
DECLARE_BITMAP(vcpu_mask, KVM_MAX_VCPUS);
bitmap_zero(vcpu_mask, KVM_MAX_VCPUS);
r = min_t(unsigned int, num_online_cpus(), KVM_MAX_VCPUS);
r = KVM_MAX_VCPUS;
DECLARE_BITMAP(sigp_emerg_pending, KVM_MAX_VCPUS);
struct vsie_page *pages[KVM_MAX_VCPUS];
DECLARE_BITMAP(kicked_mask, KVM_MAX_VCPUS);
DECLARE_BITMAP(idle_mask, KVM_MAX_VCPUS);
bitmap_zero(li->sigp_emerg_pending, KVM_MAX_VCPUS);
(ASYNC_PF_PER_VCPU * KVM_MAX_VCPUS)) {
DECLARE_BITMAP(sigp_emerg_pending, KVM_MAX_VCPUS);
for_each_set_bit(cpuaddr, sigp_emerg_pending, KVM_MAX_VCPUS) {
cpu_addr = find_first_bit(li->sigp_emerg_pending, KVM_MAX_VCPUS);
if (bitmap_empty(li->sigp_emerg_pending, KVM_MAX_VCPUS))
return id < KVM_MAX_VCPUS;
(KVM_MAX_VCPUS + LOCAL_IRQS))
r = KVM_MAX_VCPUS;
unsigned long poll_mask[BITS_TO_LONGS(KVM_MAX_VCPUS)];
#define KVM_MAX_VCPU_IDS (KVM_MAX_VCPUS * KVM_VCPU_ID_RATIO)
DECLARE_BITMAP(vcpu_mask, KVM_MAX_VCPUS);
if (!host || new_vp_index >= KVM_MAX_VCPUS)
sizeof(*vcpu_mask) * BITS_TO_LONGS(KVM_MAX_VCPUS));
bitmap_zero(vcpu_mask, KVM_MAX_VCPUS);
if (vpidx >= KVM_MAX_VCPUS)
for_each_set_bit(i, vcpu_mask, KVM_MAX_VCPUS) {
bitmap_zero(vcpu_mask, KVM_MAX_VCPUS);
ent->eax = KVM_MAX_VCPUS;
#define KVM_HV_MAX_SPARSE_VCPU_SET_BITS DIV_ROUND_UP(KVM_MAX_VCPUS, HV_VCPUS_PER_SPARSE_BANK)
DECLARE_BITMAP(vcpu_bitmap, KVM_MAX_VCPUS);
bitmap_zero(vcpu_bitmap, KVM_MAX_VCPUS);
unsigned long i, dest_vcpu_bitmap[BITS_TO_LONGS(KVM_MAX_VCPUS)];
dest_vcpu_bitmap, KVM_MAX_VCPUS);
r = min_t(unsigned int, num_online_cpus(), KVM_MAX_VCPUS);
r = KVM_MAX_VCPUS;
if (data->u.vcpu_id >= KVM_MAX_VCPUS)
if (id < KVM_MAX_VCPUS)
#define KVM_MAX_VCPU_IDS KVM_MAX_VCPUS
static int vcpu_last_completed_iteration[KVM_MAX_VCPUS];
NR_VCPUS_DEF, KVM_MAX_VCPUS);
if (test_args.nr_vcpus > KVM_MAX_VCPUS) {
KVM_MAX_VCPUS);
struct kvm_vcpu *vcpus[KVM_MAX_VCPUS];
struct test_vcpu_shared_data vcpu_shared_data[KVM_MAX_VCPUS];
static pthread_t pt_vcpu_run[KVM_MAX_VCPUS];
static int vcpu_last_completed_iteration[KVM_MAX_VCPUS];
uint32_t vcpu_to_pcpu[KVM_MAX_VCPUS];
struct memstress_vcpu_args vcpu_args[KVM_MAX_VCPUS];
struct kvm_vcpu *vcpus[KVM_MAX_VCPUS];
static struct vcpu_thread vcpu_threads[KVM_MAX_VCPUS];
static struct kvm_vcpu *vcpus[KVM_MAX_VCPUS];
DECLARE_BITMAP(in_use, KVM_MAX_VCPUS);
struct ucall ucalls[KVM_MAX_VCPUS];
for (i = 0; i < KVM_MAX_VCPUS; ++i) {
for (i = 0; i < KVM_MAX_VCPUS; ++i) {
static int vcpu_last_completed_iteration[KVM_MAX_VCPUS];
struct kvm_vcpu *vcpus[KVM_MAX_VCPUS];
pthread_t threads[KVM_MAX_VCPUS];
struct kvm_vcpu *vcpus[KVM_MAX_VCPUS];
kvm_static_assert(KVM_MAX_VCPUS > MAX_XAPIC_ID);
vm = vm_create_with_vcpus(KVM_MAX_VCPUS, NULL, vcpus);
for (i = 0; i < KVM_MAX_VCPUS; i++)
vcpuN = vcpus[KVM_MAX_VCPUS - 1];
kvm->max_vcpus = KVM_MAX_VCPUS;
for_each_set_bit(i, vcpu_bitmap, KVM_MAX_VCPUS) {