cpu_set
{ "cpu", cpu_set, cpu_get, always_valid },
cpu_set_t cpu_set;
CPU_ZERO(&cpu_set);
CPU_SET(cpu, &cpu_set);
if (sched_setaffinity(0, sizeof(cpu_set), &cpu_set)) {
cpu_set_t cpu_set;
CPU_ZERO(&cpu_set);
CPU_SET(perf_cpu_map__cpu(cpus, 0).cpu, &cpu_set);
sched_setaffinity(0, sizeof(cpu_set), &cpu_set);
if (sched_setaffinity(0, sizeof(cpu_set), &cpu_set) < 0) {
cpu_set_t cpu_set;
CPU_ZERO(&cpu_set);
CPU_SET(cpu.cpu, &cpu_set);
if (sched_setaffinity(0, sizeof(cpu_set), &cpu_set) < 0) {
CPU_CLR(cpu.cpu, &cpu_set);
size_t alloc_cpu_set(cpu_set_t **cpu_set)
*cpu_set = _cpu_set;
void free_cpu_set(cpu_set_t *cpu_set)
CPU_FREE(cpu_set);
extern size_t alloc_cpu_set(cpu_set_t **cpu_set);
extern void free_cpu_set(cpu_set_t *cpu_set);
static int parse_cpu_str(char *cpu_str, cpu_set_t *cpu_set, int cpu_set_size)
CPU_SET_S(start, cpu_set_size, cpu_set);
CPU_SET_S(start, cpu_set_size, cpu_set);
int initialize_cpu_set_from_sysfs(cpu_set_t *cpu_set, char *sysfs_path, char *sysfs_file)
if (parse_cpu_str(cpuset_buf, cpu_set, cpu_possible_setsize)) {
void for_all_cpus_in_set(size_t set_size, cpu_set_t *cpu_set, int (func)(int))
if (CPU_ISSET_S(cpu_num, set_size, cpu_set))
int for_all_cpus_in_set_and(size_t set_size, cpu_set_t *cpu_set, int (func)(int))
if (CPU_ISSET_S(cpu_num, set_size, cpu_set))
static int next_cpu(struct cpu_set *cpu_set)
if (cpu_set->cpus) {
for (i = cpu_set->next_cpu; i < cpu_set->cpus_len; i++) {
if (cpu_set->cpus[i]) {
cpu_set->next_cpu = i + 1;
return cpu_set->next_cpu++ % env.nr_cpus;
struct cpu_set prod_cpus;
struct cpu_set cons_cpus;
cpu_set_t cpu_set;
CPU_ZERO(&cpu_set);
CPU_SET(0, &cpu_set);
err = pthread_setaffinity_np(pthread_self(), sizeof(cpu_set), &cpu_set);
CPU_ZERO(&cpu_set);
CPU_SET(0, &cpu_set);
err = pthread_setaffinity_np(pthread_self(), sizeof(cpu_set), &cpu_set);
cpu_set_t cpu_set;
cpu_set_t cpu_set;
CPU_ZERO(&cpu_set);
CPU_SET(0, &cpu_set);
err = pthread_setaffinity_np(pthread_self(), sizeof(cpu_set), &cpu_set);
cpu_set_t cpu_set;
CPU_ZERO(&cpu_set);
CPU_SET(cpu, &cpu_set);
err = pthread_setaffinity_np(pthread_self(), sizeof(cpu_set), &cpu_set);
cpu_set_t cpu_set;
CPU_ZERO(&cpu_set);
CPU_SET(0, &cpu_set);
err = pthread_setaffinity_np(pthread_self(), sizeof(cpu_set), &cpu_set);
cpu_set_t cpu_set;
CPU_ZERO(&cpu_set);
CPU_SET(0, &cpu_set);
err = pthread_setaffinity_np(pthread_self(), sizeof(cpu_set), &cpu_set);
cpu_set_t cpu_set;
CPU_ZERO(&cpu_set);
CPU_SET(0, &cpu_set);
pthread_setaffinity_np(pthread_self(), sizeof(cpu_set), &cpu_set);
check_set_affinity(threads[i], &cpu_set);
check_set_affinity(throw_away, &cpu_set);
static inline void check_set_affinity(pthread_t thread, cpu_set_t *cpu_set)
r = pthread_setaffinity_np(thread, sizeof(cpu_set_t), cpu_set);
cpu_set_t cpu_set;
CPU_ZERO(&cpu_set);
CPU_SET(i, &cpu_set);
CPU_FREE(cpu_set);
cpu_set = CPU_ALLOC(cpus_online);
if (cpu_set == NULL) {
CPU_ZERO_S(cpu_set_size, cpu_set);
cpu_set)) {
cpu_set_size, cpu_set);
CPU_ZERO_S(cpu_set_size, cpu_set);
CPU_SET_S(cpus_to_pin[i], cpu_set_size, cpu_set);
cpu_set);
CPU_FREE(cpu_set);
cpu_set_t *cpu_set;
memset(&cpu_set, 0, sizeof(cpu_set));
CPU_SET(cpu_id, &cpu_set);
if (sched_setaffinity(0, sizeof(cpu_set), &cpu_set) < 0)
cpu_set_t cpu_set;
cpu_set_t cpu_set;
CPU_ZERO(&cpu_set);
CPU_SET(i, &cpu_set);
ASSERT_EQ(CPU_COUNT(&cpu_set), 1);
ASSERT_NE(CPU_ISSET(i, &cpu_set), 0);
ret = sched_setaffinity(0, sizeof(cpu_set), &cpu_set);
cpu_set_t cpu_set;
CPU_ZERO(&cpu_set);
CPU_SET(first_cpu, &cpu_set);
if (sched_setaffinity(getpid(), sizeof(cpu_set), &cpu_set))
CPU_SET(first_cpu, &cpu_set);
if (sched_setaffinity(getpid(), sizeof(cpu_set), &cpu_set))