HK_TYPE_DOMAIN
return __task_cpu_possible_mask(p, housekeeping_cpumask(HK_TYPE_DOMAIN));
if (!housekeeping_cpu(cpu, HK_TYPE_DOMAIN))
housekeeping_cpumask(HK_TYPE_DOMAIN));
# define task_cpu_fallback_mask(p) housekeeping_cpumask(HK_TYPE_DOMAIN)
return !housekeeping_test_cpu(cpu, HK_TYPE_DOMAIN);
housekeeping_cpumask(HK_TYPE_DOMAIN));
housekeeping_cpumask(HK_TYPE_DOMAIN))))
cpumask_and(dp, dp, housekeeping_cpumask(HK_TYPE_DOMAIN));
housekeeping_cpumask(HK_TYPE_DOMAIN));
housekeeping_cpumask(HK_TYPE_DOMAIN)) >= nr_cpu_ids) {
pref = housekeeping_cpumask(HK_TYPE_DOMAIN);
cpumask_and(cpumask, pref, housekeeping_cpumask(HK_TYPE_DOMAIN));
cpumask_copy(cpumask, housekeeping_cpumask(HK_TYPE_DOMAIN));
if (set_cpus_allowed_ptr(current, housekeeping_cpumask(HK_TYPE_DOMAIN)) < 0)
hk_msk = housekeeping_cpumask(HK_TYPE_DOMAIN);
if (housekeeping_enabled(HK_TYPE_DOMAIN)) {
if (!cpumask_equal(housekeeping_cpumask(HK_TYPE_DOMAIN),
old = housekeeping_cpumask_dereference(HK_TYPE_DOMAIN);
rcu_assign_pointer(housekeeping.cpumasks[HK_TYPE_DOMAIN], trial);
err = workqueue_unbound_housekeeping_update(housekeeping_cpumask(HK_TYPE_DOMAIN));
HK_FLAG_DOMAIN = BIT(HK_TYPE_DOMAIN),
if (IS_ENABLED(CONFIG_LOCKDEP) && type == HK_TYPE_DOMAIN) {
for_each_cpu_and(cpu, cpu_possible_mask, housekeeping_cpumask(HK_TYPE_DOMAIN))
cpumask_and(doms_cur[0], cpu_map, housekeeping_cpumask(HK_TYPE_DOMAIN));
housekeeping_cpumask(HK_TYPE_DOMAIN));
housekeeping_cpumask(HK_TYPE_DOMAIN));
if (!housekeeping_enabled(HK_TYPE_DOMAIN))
cpumask_andnot(cpumask, cpu_possible_mask, housekeeping_cpumask(HK_TYPE_DOMAIN));
return (!housekeeping_cpu(cpu, HK_TYPE_DOMAIN) &&
restrict_unbound_cpumask("HK_TYPE_DOMAIN", housekeeping_cpumask(HK_TYPE_DOMAIN));
housekeeping_cpumask(HK_TYPE_DOMAIN));