#include <smp.h>
#include <stdlib.h>
#include <string.h>
#include <arch/atomic.h>
#include <arch/cpu.h>
#include <arch/debug.h>
#include <arch/int.h>
#include <arch/smp.h>
#include <boot/kernel_args.h>
#include <cpu.h>
#include <generic_syscall.h>
#include <interrupts.h>
#include <spinlock_contention.h>
#include <thread.h>
#include <util/atomic.h>
#include "kernel_debug_config.h"
#ifdef TRACE_SMP
# define TRACE(...) dprintf_no_syslog(__VA_ARGS__)
#else
# define TRACE(...) (void)0
#endif
#undef try_acquire_spinlock
#undef acquire_spinlock
#undef release_spinlock
#undef try_acquire_read_spinlock
#undef acquire_read_spinlock
#undef release_read_spinlock
#undef try_acquire_write_spinlock
#undef acquire_write_spinlock
#undef release_write_spinlock
#undef try_acquire_write_seqlock
#undef acquire_write_seqlock
#undef release_write_seqlock
#undef acquire_read_seqlock
#undef release_read_seqlock
#define MSG_ALLOCATE_PER_CPU (4)
#define SPINLOCK_DEADLOCK_COUNT 100000000
#define SPINLOCK_DEADLOCK_COUNT_NO_CHECK 2000000000
struct smp_msg {
struct smp_msg *next;
int32 message;
addr_t data;
addr_t data2;
addr_t data3;
void *data_ptr;
uint32 flags;
int32 ref_count;
int32 done;
CPUSet proc_bitmap;
};
enum mailbox_source {
MAILBOX_LOCAL,
MAILBOX_BCAST,
};
static int32 sBootCPUSpin = 0;
static int32 sEarlyCPUCallCount;
static CPUSet sEarlyCPUCallSet;
static void (*sEarlyCPUCallFunction)(void*, int);
void* sEarlyCPUCallCookie;
static struct smp_msg* sFreeMessages = NULL;
static int32 sFreeMessageCount = 0;
static rw_spinlock sFreeMessageSpinlock = B_RW_SPINLOCK_INITIALIZER;
static struct smp_msg* sCPUMessages[SMP_MAX_CPUS] = { NULL, };
static struct smp_msg* sBroadcastMessages = NULL;
static rw_spinlock sBroadcastMessageSpinlock = B_RW_SPINLOCK_INITIALIZER;
static int32 sBroadcastMessageCounter;
static bool sICIEnabled = false;
static int32 sNumCPUs = 1;
static int32 process_pending_ici(int32 currentCPU);
#if DEBUG_SPINLOCKS
#define NUM_LAST_CALLERS 32
static struct {
void *caller;
spinlock *lock;
} sLastCaller[NUM_LAST_CALLERS];
static int32 sLastIndex = 0;
static void
push_lock_caller(void* caller, spinlock* lock)
{
int32 index = (uint32)atomic_add(&sLastIndex, 1) % NUM_LAST_CALLERS;
sLastCaller[index].caller = caller;
sLastCaller[index].lock = lock;
}
static void*
find_lock_caller(spinlock* lock)
{
int32 lastIndex = (uint32)atomic_get(&sLastIndex) % NUM_LAST_CALLERS;
for (int32 i = 0; i < NUM_LAST_CALLERS; i++) {
int32 index = (NUM_LAST_CALLERS + lastIndex - 1 - i) % NUM_LAST_CALLERS;
if (sLastCaller[index].lock == lock)
return sLastCaller[index].caller;
}
return NULL;
}
int
dump_spinlock(int argc, char** argv)
{
if (argc != 2) {
print_debugger_command_usage(argv[0]);
return 0;
}
uint64 address;
if (!evaluate_debug_expression(argv[1], &address, false))
return 0;
spinlock* lock = (spinlock*)(addr_t)address;
kprintf("spinlock %p:\n", lock);
bool locked = B_SPINLOCK_IS_LOCKED(lock);
if (locked) {
kprintf(" locked from %p\n", find_lock_caller(lock));
} else
kprintf(" not locked\n");
#if B_DEBUG_SPINLOCK_CONTENTION
kprintf(" failed try_acquire(): %d\n", lock->failed_try_acquire);
kprintf(" total wait time: %" B_PRIdBIGTIME "\n", lock->total_wait);
kprintf(" total held time: %" B_PRIdBIGTIME "\n", lock->total_held);
kprintf(" last acquired at: %" B_PRIdBIGTIME "\n", lock->last_acquired);
#endif
return 0;
}
#endif
#if B_DEBUG_SPINLOCK_CONTENTION
static inline void
update_lock_contention(spinlock* lock, bigtime_t start)
{
const bigtime_t now = system_time();
lock->last_acquired = now;
lock->total_wait += (now - start);
}
static inline void
update_lock_held(spinlock* lock)
{
const bigtime_t held = (system_time() - lock->last_acquired);
lock->total_held += held;
#if DEBUG_SPINLOCK_LATENCIES
if (held > DEBUG_SPINLOCK_LATENCIES) {
panic("spinlock %p was held for %" B_PRIdBIGTIME " usecs (%d allowed)\n",
lock, held, DEBUG_SPINLOCK_LATENCIES);
}
#endif
}
#endif
int
dump_ici_messages(int argc, char** argv)
{
int32 count = 0;
int32 doneCount = 0;
int32 unreferencedCount = 0;
smp_msg* message = sBroadcastMessages;
while (message != NULL) {
count++;
if (message->done == 1)
doneCount++;
if (message->ref_count <= 0)
unreferencedCount++;
message = message->next;
}
kprintf("ICI broadcast messages: %" B_PRId32 ", first: %p\n", count,
sBroadcastMessages);
kprintf(" done: %" B_PRId32 "\n", doneCount);
kprintf(" unreferenced: %" B_PRId32 "\n", unreferencedCount);
for (int32 i = 0; i < sNumCPUs; i++) {
count = 0;
message = sCPUMessages[i];
while (message != NULL) {
count++;
message = message->next;
}
kprintf("CPU %" B_PRId32 " messages: %" B_PRId32 ", first: %p\n", i,
count, sCPUMessages[i]);
}
return 0;
}
int
dump_ici_message(int argc, char** argv)
{
if (argc != 2) {
print_debugger_command_usage(argv[0]);
return 0;
}
uint64 address;
if (!evaluate_debug_expression(argv[1], &address, false))
return 0;
smp_msg* message = (smp_msg*)(addr_t)address;
kprintf("ICI message %p:\n", message);
kprintf(" next: %p\n", message->next);
kprintf(" message: %" B_PRId32 "\n", message->message);
kprintf(" data: 0x%lx\n", message->data);
kprintf(" data2: 0x%lx\n", message->data2);
kprintf(" data3: 0x%lx\n", message->data3);
kprintf(" data_ptr: %p\n", message->data_ptr);
kprintf(" flags: %" B_PRIx32 "\n", message->flags);
kprintf(" ref_count: %" B_PRIx32 "\n", message->ref_count);
kprintf(" done: %s\n", message->done == 1 ? "true" : "false");
kprintf(" proc_bitmap: ");
for (int32 i = 0; i < sNumCPUs; i++) {
if (message->proc_bitmap.GetBit(i))
kprintf("%s%" B_PRId32, i != 0 ? ", " : "", i);
}
kprintf("\n");
return 0;
}
static inline void
process_all_pending_ici(int32 currentCPU)
{
while (process_pending_ici(currentCPU) != B_ENTRY_NOT_FOUND)
;
}
bool
try_acquire_spinlock(spinlock* lock)
{
#if DEBUG_SPINLOCKS
if (are_interrupts_enabled()) {
panic("try_acquire_spinlock: attempt to acquire lock %p with "
"interrupts enabled", lock);
}
#endif
if (atomic_get_and_set(&lock->lock, 1) != 0) {
#if B_DEBUG_SPINLOCK_CONTENTION
atomic_add(&lock->failed_try_acquire, 1);
#endif
return false;
}
#if B_DEBUG_SPINLOCK_CONTENTION
update_lock_contention(lock, system_time());
#endif
#if DEBUG_SPINLOCKS
push_lock_caller(arch_debug_get_caller(), lock);
#endif
return true;
}
void
acquire_spinlock(spinlock* lock)
{
#if DEBUG_SPINLOCKS
if (are_interrupts_enabled()) {
panic("acquire_spinlock: attempt to acquire lock %p with interrupts "
"enabled", lock);
}
#endif
if (sNumCPUs > 1) {
#if B_DEBUG_SPINLOCK_CONTENTION
const bigtime_t start = system_time();
#endif
int currentCPU = smp_get_current_cpu();
while (1) {
uint32 count = 0;
while (lock->lock != 0) {
if (++count == SPINLOCK_DEADLOCK_COUNT) {
#if DEBUG_SPINLOCKS
panic("acquire_spinlock(): Failed to acquire spinlock %p "
"for a long time (last caller: %p, value: %" B_PRIx32
")", lock, find_lock_caller(lock), lock->lock);
#else
panic("acquire_spinlock(): Failed to acquire spinlock %p "
"for a long time (value: %" B_PRIx32 ")", lock,
lock->lock);
#endif
count = 0;
}
process_all_pending_ici(currentCPU);
cpu_wait(&lock->lock, 0);
}
if (atomic_get_and_set(&lock->lock, 1) == 0)
break;
}
#if B_DEBUG_SPINLOCK_CONTENTION
update_lock_contention(lock, start);
#endif
#if DEBUG_SPINLOCKS
push_lock_caller(arch_debug_get_caller(), lock);
#endif
} else {
#if B_DEBUG_SPINLOCK_CONTENTION
lock->last_acquired = system_time();
#endif
#if DEBUG_SPINLOCKS
int32 oldValue = atomic_get_and_set(&lock->lock, 1);
if (oldValue != 0) {
panic("acquire_spinlock: attempt to acquire lock %p twice on "
"non-SMP system (last caller: %p, value %" B_PRIx32 ")", lock,
find_lock_caller(lock), oldValue);
}
push_lock_caller(arch_debug_get_caller(), lock);
#endif
}
}
void
release_spinlock(spinlock *lock)
{
#if B_DEBUG_SPINLOCK_CONTENTION
update_lock_held(lock);
#endif
if (sNumCPUs > 1) {
if (are_interrupts_enabled()) {
panic("release_spinlock: attempt to release lock %p with "
"interrupts enabled\n", lock);
}
#if DEBUG_SPINLOCKS
if (atomic_get_and_set(&lock->lock, 0) != 1)
panic("release_spinlock: lock %p was already released\n", lock);
#else
atomic_set(&lock->lock, 0);
#endif
} else {
#if DEBUG_SPINLOCKS
if (are_interrupts_enabled()) {
panic("release_spinlock: attempt to release lock %p with "
"interrupts enabled\n", lock);
}
if (atomic_get_and_set(&lock->lock, 0) != 1)
panic("release_spinlock: lock %p was already released\n", lock);
#endif
}
}
bool
try_acquire_write_spinlock(rw_spinlock* lock)
{
#if DEBUG_SPINLOCKS
if (are_interrupts_enabled()) {
panic("try_acquire_write_spinlock: attempt to acquire lock %p with "
"interrupts enabled", lock);
}
if (sNumCPUs < 2 && lock->lock != 0) {
panic("try_acquire_write_spinlock(): attempt to acquire lock %p twice "
"on non-SMP system", lock);
}
#endif
return atomic_test_and_set(&lock->lock, 1u << 31, 0) == 0;
}
void
acquire_write_spinlock(rw_spinlock* lock)
{
uint32 count = 0;
int32 currentCPU = smp_get_current_cpu();
while (true) {
if (try_acquire_write_spinlock(lock))
break;
while (lock->lock != 0) {
if (++count == SPINLOCK_DEADLOCK_COUNT) {
panic("acquire_write_spinlock(): Failed to acquire spinlock %p "
"for a long time!", lock);
count = 0;
}
process_all_pending_ici(currentCPU);
cpu_wait(&lock->lock, 0);
}
}
}
static void
acquire_write_spinlock_nocheck(rw_spinlock* lock)
{
uint32 count = 0;
while (true) {
if (try_acquire_write_spinlock(lock))
break;
while (lock->lock != 0) {
if (++count == SPINLOCK_DEADLOCK_COUNT_NO_CHECK) {
panic("acquire_write_spinlock(): Failed to acquire spinlock %p "
"for a long time!", lock);
count = 0;
}
cpu_wait(&lock->lock, 0);
}
}
}
static void
acquire_write_spinlock_cpu(int32 currentCPU, rw_spinlock* lock)
{
uint32 count = 0;
while (true) {
if (try_acquire_write_spinlock(lock))
break;
while (lock->lock != 0) {
if (++count == SPINLOCK_DEADLOCK_COUNT) {
panic("acquire_write_spinlock(): Failed to acquire spinlock %p "
"for a long time!", lock);
count = 0;
}
process_all_pending_ici(currentCPU);
cpu_wait(&lock->lock, 0);
}
}
}
void
release_write_spinlock(rw_spinlock* lock)
{
#if DEBUG_SPINLOCKS
uint32 previous = atomic_get_and_set(&lock->lock, 0);
if ((previous & 1u << 31) == 0) {
panic("release_write_spinlock: lock %p was already released (value: "
"%#" B_PRIx32 ")\n", lock, previous);
}
#else
atomic_set(&lock->lock, 0);
#endif
}
bool
try_acquire_read_spinlock(rw_spinlock* lock)
{
#if DEBUG_SPINLOCKS
if (are_interrupts_enabled()) {
panic("try_acquire_read_spinlock: attempt to acquire lock %p with "
"interrupts enabled", lock);
}
if (sNumCPUs < 2 && lock->lock != 0) {
panic("try_acquire_read_spinlock(): attempt to acquire lock %p twice "
"on non-SMP system", lock);
}
#endif
uint32 previous = atomic_add(&lock->lock, 1);
return (previous & (1u << 31)) == 0;
}
void
acquire_read_spinlock(rw_spinlock* lock)
{
uint32 count = 0;
int currentCPU = smp_get_current_cpu();
while (1) {
if (try_acquire_read_spinlock(lock))
break;
while ((lock->lock & (1u << 31)) != 0) {
if (++count == SPINLOCK_DEADLOCK_COUNT) {
panic("acquire_read_spinlock(): Failed to acquire spinlock %p "
"for a long time!", lock);
count = 0;
}
process_all_pending_ici(currentCPU);
cpu_wait(&lock->lock, 0);
}
}
}
static void
acquire_read_spinlock_nocheck(rw_spinlock* lock)
{
uint32 count = 0;
while (1) {
if (try_acquire_read_spinlock(lock))
break;
while ((lock->lock & (1u << 31)) != 0) {
if (++count == SPINLOCK_DEADLOCK_COUNT_NO_CHECK) {
panic("acquire_read_spinlock(): Failed to acquire spinlock %p "
"for a long time!", lock);
count = 0;
}
cpu_wait(&lock->lock, 0);
}
}
}
void
release_read_spinlock(rw_spinlock* lock)
{
#if DEBUG_SPINLOCKS
uint32 previous = atomic_add(&lock->lock, -1);
if ((previous & 1u << 31) != 0) {
panic("release_read_spinlock: lock %p was already released (value:"
" %#" B_PRIx32 ")\n", lock, previous);
}
#else
atomic_add(&lock->lock, -1);
#endif
}
bool
try_acquire_write_seqlock(seqlock* lock)
{
bool succeed = try_acquire_spinlock(&lock->lock);
if (succeed)
atomic_add((int32*)&lock->count, 1);
return succeed;
}
void
acquire_write_seqlock(seqlock* lock)
{
acquire_spinlock(&lock->lock);
atomic_add((int32*)&lock->count, 1);
}
void
release_write_seqlock(seqlock* lock)
{
atomic_add((int32*)&lock->count, 1);
release_spinlock(&lock->lock);
}
uint32
acquire_read_seqlock(seqlock* lock)
{
return atomic_get((int32*)&lock->count);
}
bool
release_read_seqlock(seqlock* lock, uint32 count)
{
memory_read_barrier();
uint32 current = *(volatile int32*)&lock->count;
if (count % 2 == 1 || current != count) {
cpu_pause();
return false;
}
return true;
}
static cpu_status
find_free_message(struct smp_msg** msg)
{
TRACE("find_free_message: entry\n");
cpu_status state;
retry:
while (sFreeMessageCount <= 0) {
ASSERT(are_interrupts_enabled());
cpu_pause();
}
state = disable_interrupts();
acquire_write_spinlock(&sFreeMessageSpinlock);
if (sFreeMessageCount <= 0) {
release_write_spinlock(&sFreeMessageSpinlock);
restore_interrupts(state);
goto retry;
}
*msg = sFreeMessages;
sFreeMessages = (*msg)->next;
sFreeMessageCount--;
release_write_spinlock(&sFreeMessageSpinlock);
TRACE("find_free_message: returning msg %p\n", *msg);
return state;
}
static void
find_free_message_interrupts_disabled(int32 currentCPU,
struct smp_msg** _message)
{
TRACE("find_free_message_interrupts_disabled: entry\n");
acquire_write_spinlock_cpu(currentCPU, &sFreeMessageSpinlock);
while (sFreeMessageCount <= 0) {
release_write_spinlock(&sFreeMessageSpinlock);
process_all_pending_ici(currentCPU);
cpu_pause();
acquire_write_spinlock_cpu(currentCPU, &sFreeMessageSpinlock);
}
*_message = sFreeMessages;
sFreeMessages = (*_message)->next;
sFreeMessageCount--;
release_write_spinlock(&sFreeMessageSpinlock);
TRACE("find_free_message_interrupts_disabled: returning msg %p\n",
*_message);
}
static void
return_free_message(struct smp_msg* msg)
{
TRACE("return_free_message: returning msg %p\n", msg);
acquire_write_spinlock_nocheck(&sFreeMessageSpinlock);
msg->next = sFreeMessages;
sFreeMessages = msg;
sFreeMessageCount++;
release_write_spinlock(&sFreeMessageSpinlock);
}
static void
prepend_message(struct smp_msg*& listHead, struct smp_msg* msg)
{
while (true) {
struct smp_msg* next = atomic_pointer_get(&listHead);
msg->next = next;
if (atomic_pointer_test_and_set(&listHead, msg, next) == next)
break;
cpu_pause();
}
}
static struct smp_msg*
check_for_message(int currentCPU, mailbox_source& sourceMailbox)
{
if (!sICIEnabled)
return NULL;
struct smp_msg* msg = atomic_pointer_get(&sCPUMessages[currentCPU]);
if (msg != NULL) {
while (true) {
if (atomic_pointer_test_and_set(&sCPUMessages[currentCPU], msg->next, msg) == msg)
break;
cpu_pause();
msg = atomic_pointer_get(&sCPUMessages[currentCPU]);
ASSERT(msg != NULL);
}
TRACE(" cpu %d: found msg %p in cpu mailbox\n", currentCPU, msg);
sourceMailbox = MAILBOX_LOCAL;
} else if (atomic_get(&get_cpu_struct()->ici_counter)
!= atomic_get(&sBroadcastMessageCounter)) {
acquire_read_spinlock_nocheck(&sBroadcastMessageSpinlock);
msg = sBroadcastMessages;
while (msg != NULL) {
if (!msg->proc_bitmap.GetBitAtomic(currentCPU)) {
msg = msg->next;
continue;
}
msg->proc_bitmap.ClearBitAtomic(currentCPU);
atomic_add(&gCPU[currentCPU].ici_counter, 1);
sourceMailbox = MAILBOX_BCAST;
break;
}
release_read_spinlock(&sBroadcastMessageSpinlock);
if (msg != NULL) {
TRACE(" cpu %d: found msg %p in broadcast mailbox\n", currentCPU,
msg);
}
}
return msg;
}
static void
finish_message_processing(int currentCPU, struct smp_msg* msg,
mailbox_source sourceMailbox)
{
if (atomic_add(&msg->ref_count, -1) != 1)
return;
if (sourceMailbox == MAILBOX_BCAST)
acquire_write_spinlock_nocheck(&sBroadcastMessageSpinlock);
TRACE("cleaning up message %p\n", msg);
if (sourceMailbox != MAILBOX_BCAST) {
} else if (msg == sBroadcastMessages) {
sBroadcastMessages = msg->next;
} else {
struct smp_msg* last = NULL;
struct smp_msg* msg1;
msg1 = sBroadcastMessages;
while (msg1 != NULL && msg1 != msg) {
last = msg1;
msg1 = msg1->next;
}
if (msg1 == msg && last != NULL)
last->next = msg->next;
else
panic("last == NULL or msg != msg1");
}
if (sourceMailbox == MAILBOX_BCAST)
release_write_spinlock(&sBroadcastMessageSpinlock);
if ((msg->flags & SMP_MSG_FLAG_FREE_ARG) != 0 && msg->data_ptr != NULL)
free(msg->data_ptr);
if ((msg->flags & SMP_MSG_FLAG_SYNC) != 0) {
atomic_set(&msg->done, 1);
} else {
return_free_message(msg);
}
}
static void
invoke_smp_msg(struct smp_msg* msg, int currentCPU, bool* haltCPU)
{
switch (msg->message) {
case SMP_MSG_INVALIDATE_PAGE_RANGE:
arch_cpu_invalidate_tlb_range(msg->data, msg->data2, msg->data3);
break;
case SMP_MSG_INVALIDATE_PAGE_LIST:
arch_cpu_invalidate_tlb_list(msg->data, (addr_t*)msg->data2, (int)msg->data3);
break;
case SMP_MSG_USER_INVALIDATE_PAGES:
arch_cpu_user_tlb_invalidate(msg->data);
break;
case SMP_MSG_GLOBAL_INVALIDATE_PAGES:
arch_cpu_global_tlb_invalidate();
break;
case SMP_MSG_CPU_HALT:
*haltCPU = true;
break;
case SMP_MSG_CALL_FUNCTION:
{
smp_call_func func = (smp_call_func)msg->data_ptr;
func(msg->data, currentCPU, msg->data2, msg->data3);
break;
}
case SMP_MSG_RESCHEDULE:
scheduler_reschedule_ici();
break;
default:
dprintf("smp_intercpu_interrupt_handler: got unknown message %" B_PRId32 "\n",
msg->message);
break;
}
}
static status_t
process_pending_ici(int32 currentCPU)
{
mailbox_source sourceMailbox;
struct smp_msg* msg = check_for_message(currentCPU, sourceMailbox);
if (msg == NULL)
return B_ENTRY_NOT_FOUND;
TRACE(" cpu %" B_PRId32 " message = %" B_PRId32 "\n", currentCPU, msg->message);
bool haltCPU = false;
invoke_smp_msg(msg, currentCPU, &haltCPU);
finish_message_processing(currentCPU, msg, sourceMailbox);
if (haltCPU)
debug_trap_cpu_in_kdl(currentCPU, false);
return B_OK;
}
#if B_DEBUG_SPINLOCK_CONTENTION
static status_t
spinlock_contention_syscall(const char* subsystem, uint32 function,
void* buffer, size_t bufferSize)
{
if (function != GET_SPINLOCK_CONTENTION_INFO)
return B_BAD_VALUE;
if (bufferSize < sizeof(spinlock_contention_info))
return B_BAD_VALUE;
spinlock_contention_info info;
info.thread_creation_spinlock = gThreadCreationLock.total_wait;
if (!IS_USER_ADDRESS(buffer)
|| user_memcpy(buffer, &info, sizeof(info)) != B_OK) {
return B_BAD_ADDRESS;
}
return B_OK;
}
#endif
static void
process_early_cpu_call(int32 cpu)
{
sEarlyCPUCallFunction(sEarlyCPUCallCookie, cpu);
sEarlyCPUCallSet.ClearBitAtomic(cpu);
atomic_add(&sEarlyCPUCallCount, 1);
}
static void
call_all_cpus_early(void (*function)(void*, int), void* cookie)
{
ASSERT(!are_interrupts_enabled());
if (sNumCPUs > 1) {
sEarlyCPUCallFunction = function;
sEarlyCPUCallCookie = cookie;
atomic_set(&sEarlyCPUCallCount, 1);
sEarlyCPUCallSet.SetAll();
sEarlyCPUCallSet.ClearBit(0);
while (sEarlyCPUCallCount < sNumCPUs)
cpu_wait(&sEarlyCPUCallCount, sNumCPUs);
}
function(cookie, 0);
}
int
smp_intercpu_interrupt_handler(int32 cpu)
{
TRACE("smp_intercpu_interrupt_handler: entry on cpu %" B_PRId32 "\n", cpu);
process_all_pending_ici(cpu);
TRACE("smp_intercpu_interrupt_handler: done on cpu %" B_PRId32 "\n", cpu);
return B_HANDLED_INTERRUPT;
}
void
smp_send_ici(int32 targetCPU, int32 message, addr_t data, addr_t data2,
addr_t data3, void* dataPointer, uint32 flags)
{
if (!sICIEnabled)
return;
TRACE("smp_send_ici: target 0x%" B_PRIx32 ", mess 0x%" B_PRIx32 ", data 0x%lx, data2 0x%lx, "
"data3 0x%lx, ptr %p, flags 0x%" B_PRIx32 "\n", targetCPU, message, data, data2,
data3, dataPointer, flags);
struct smp_msg *msg;
cpu_status state = find_free_message(&msg);
int currentCPU = smp_get_current_cpu();
if (targetCPU == currentCPU) {
ASSERT(false);
return_free_message(msg);
restore_interrupts(state);
return;
}
msg->message = message;
msg->data = data;
msg->data2 = data2;
msg->data3 = data3;
msg->data_ptr = dataPointer;
msg->ref_count = 1;
msg->flags = flags;
msg->done = 0;
prepend_message(sCPUMessages[targetCPU], msg);
arch_smp_send_ici(targetCPU);
if ((flags & SMP_MSG_FLAG_SYNC) != 0) {
while (msg->done == 0) {
process_all_pending_ici(currentCPU);
cpu_wait(&msg->done, 1);
}
return_free_message(msg);
}
restore_interrupts(state);
}
void
smp_broadcast_ici(int32 message, addr_t data, addr_t data2, addr_t data3,
void *dataPointer, uint32 flags)
{
if (!sICIEnabled) {
smp_msg dummy;
dummy.message = message;
dummy.data = data;
dummy.data2 = data2;
dummy.data3 = data3;
dummy.data_ptr = dataPointer;
cpu_status state = disable_interrupts();
invoke_smp_msg(&dummy, 0, NULL);
restore_interrupts(state);
return;
}
TRACE("smp_broadcast_ici: cpu %" B_PRId32 " mess 0x%" B_PRIx32 ", data 0x%lx, data2 "
"0x%lx, data3 0x%lx, ptr %p, flags 0x%" B_PRIx32 "\n", smp_get_current_cpu(),
message, data, data2, data3, dataPointer, flags);
struct smp_msg *msg;
cpu_status state = find_free_message(&msg);
int32 currentCPU = smp_get_current_cpu();
msg->message = message;
msg->data = data;
msg->data2 = data2;
msg->data3 = data3;
msg->data_ptr = dataPointer;
msg->ref_count = sNumCPUs;
msg->flags = flags;
msg->proc_bitmap.SetAll();
msg->done = 0;
TRACE("smp_broadcast_ici%d: inserting msg %p into broadcast "
"mbox\n", currentCPU, msg);
acquire_read_spinlock_nocheck(&sBroadcastMessageSpinlock);
prepend_message(sBroadcastMessages, msg);
release_read_spinlock(&sBroadcastMessageSpinlock);
atomic_add(&sBroadcastMessageCounter, 1);
arch_smp_send_broadcast_ici();
TRACE("smp_broadcast_ici: sent interrupt\n");
if ((flags & SMP_MSG_FLAG_SYNC) != 0) {
TRACE("smp_broadcast_ici: waiting for ack\n");
while (msg->done == 0) {
process_all_pending_ici(currentCPU);
cpu_wait(&msg->done, 1);
}
TRACE("smp_broadcast_ici: returning message to free list\n");
return_free_message(msg);
} else {
process_all_pending_ici(currentCPU);
}
restore_interrupts(state);
}
void
smp_multicast_ici(const CPUSet& cpuMask, int32 message, addr_t data,
addr_t data2, addr_t data3, void *dataPointer, uint32 flags)
{
if (!sICIEnabled) {
ASSERT(false);
return;
}
ASSERT(thread_get_current_thread()->pinned_to_cpu);
int32 currentCPU = smp_get_current_cpu();
bool self = cpuMask.GetBit(currentCPU);
int32 targetCPUs = 0, firstNonCurrentCPU = -1;
for (int32 i = 0; i < sNumCPUs; i++) {
if (cpuMask.GetBit(i)) {
targetCPUs++;
if (firstNonCurrentCPU < 0 && i != currentCPU)
firstNonCurrentCPU = i;
}
}
if (targetCPUs == 0) {
panic("smp_multicast_ici(): 0 CPU mask");
return;
}
struct smp_msg *msg;
cpu_status state = find_free_message(&msg);
msg->message = message;
msg->data = data;
msg->data2 = data2;
msg->data3 = data3;
msg->data_ptr = dataPointer;
msg->ref_count = targetCPUs;
msg->flags = flags;
msg->done = 0;
msg->proc_bitmap = cpuMask;
if ((!self && targetCPUs == 1) || (self && targetCPUs == 2)) {
prepend_message(sCPUMessages[firstNonCurrentCPU], msg);
arch_smp_send_ici(firstNonCurrentCPU);
if (self) {
invoke_smp_msg(msg, currentCPU, NULL);
finish_message_processing(currentCPU, msg, MAILBOX_LOCAL);
}
} else {
bool broadcast = (!self && targetCPUs == sNumCPUs - 1)
|| (self && targetCPUs == sNumCPUs);
acquire_read_spinlock_nocheck(&sBroadcastMessageSpinlock);
prepend_message(sBroadcastMessages, msg);
release_read_spinlock(&sBroadcastMessageSpinlock);
atomic_add(&sBroadcastMessageCounter, 1);
for (int32 i = 0; i < sNumCPUs; i++) {
if (!cpuMask.GetBit(i))
atomic_add(&gCPU[i].ici_counter, 1);
}
if (broadcast) {
arch_smp_send_broadcast_ici();
} else {
CPUSet sendMask = cpuMask;
sendMask.ClearBit(currentCPU);
arch_smp_send_multicast_ici(sendMask);
}
}
if ((flags & SMP_MSG_FLAG_SYNC) != 0) {
while (msg->done == 0) {
process_all_pending_ici(currentCPU);
cpu_wait(&msg->done, 1);
}
return_free_message(msg);
} else if (self) {
process_all_pending_ici(currentCPU);
}
restore_interrupts(state);
}
void
smp_multicast_ici_interrupts_disabled(int32 currentCPU, const CPUSet& cpuMask,
int32 message, addr_t data, addr_t data2, addr_t data3, void *dataPointer, uint32 flags)
{
if (!sICIEnabled)
return;
TRACE("smp_multicast_ici_interrupts_disabled: cpu %" B_PRId32 " mess 0x%" B_PRIx32 ", "
"data 0x%lx, data2 0x%lx, data3 0x%lx, ptr %p, flags 0x%" B_PRIx32 "\n",
currentCPU, message, data, data2, data3, dataPointer, flags);
int32 targetCPUs = 0;
for (int32 i = 0; i < sNumCPUs; i++) {
if (cpuMask.GetBit(i))
targetCPUs++;
}
if (targetCPUs == 0) {
panic("smp_multicast_ici_interrupts_disabled(): 0 CPU mask");
return;
}
struct smp_msg *msg;
find_free_message_interrupts_disabled(currentCPU, &msg);
msg->message = message;
msg->data = data;
msg->data2 = data2;
msg->data3 = data3;
msg->data_ptr = dataPointer;
msg->ref_count = targetCPUs;
msg->flags = flags;
msg->done = 0;
msg->proc_bitmap = cpuMask;
bool self = cpuMask.GetBit(currentCPU);
bool broadcast = (!self && targetCPUs == sNumCPUs - 1)
|| (self && targetCPUs == sNumCPUs);
TRACE("smp_multicast_ici_interrupts_disabled %" B_PRId32 ": inserting msg %p "
"into broadcast mbox\n", currentCPU, msg);
acquire_write_spinlock_nocheck(&sBroadcastMessageSpinlock);
msg->next = sBroadcastMessages;
sBroadcastMessages = msg;
release_write_spinlock(&sBroadcastMessageSpinlock);
atomic_add(&sBroadcastMessageCounter, 1);
for (int32 i = 0; i < sNumCPUs; i++) {
if (!cpuMask.GetBit(i))
atomic_add(&gCPU[i].ici_counter, 1);
}
if (broadcast) {
arch_smp_send_broadcast_ici();
} else {
CPUSet sendMask = cpuMask;
sendMask.ClearBit(currentCPU);
arch_smp_send_multicast_ici(sendMask);
}
TRACE("smp_multicast_ici_interrupts_disabled %" B_PRId32 ": sent interrupt\n",
currentCPU);
if ((flags & SMP_MSG_FLAG_SYNC) != 0) {
TRACE("smp_multicast_ici_interrupts_disabled %" B_PRId32 ": waiting for "
"ack\n", currentCPU);
while (msg->done == 0) {
process_all_pending_ici(currentCPU);
cpu_wait(&msg->done, 1);
}
TRACE("smp_multicast_ici_interrupts_disabled %" B_PRId32 ": returning "
"message to free list\n", currentCPU);
return_free_message(msg);
} else if (self) {
process_all_pending_ici(currentCPU);
}
TRACE("smp_multicast_ici_interrupts_disabled: done\n");
}
bool
smp_trap_non_boot_cpus(int32 cpu, uint32* rendezVous)
{
ASSERT(!are_interrupts_enabled());
if (cpu == 0) {
smp_cpu_rendezvous(rendezVous);
return true;
}
smp_cpu_rendezvous(rendezVous);
while (sBootCPUSpin == 0) {
if (sEarlyCPUCallSet.GetBitAtomic(cpu))
process_early_cpu_call(cpu);
cpu_pause();
}
return false;
}
void
smp_wake_up_non_boot_cpus()
{
if (sNumCPUs > 1)
sICIEnabled = true;
atomic_set(&sBootCPUSpin, 1);
}
void
smp_cpu_rendezvous(uint32* var)
{
atomic_add((int32*)var, 1);
while (*var < (uint32)sNumCPUs)
cpu_wait((int32*)var, sNumCPUs);
}
status_t
smp_init(kernel_args* args)
{
TRACE("smp_init: entry\n");
#if DEBUG_SPINLOCKS
add_debugger_command_etc("spinlock", &dump_spinlock,
"Dump info on a spinlock",
"\n"
"Dumps info on a spinlock.\n", 0);
#endif
add_debugger_command_etc("ici", &dump_ici_messages,
"Dump info on pending ICI messages",
"\n"
"Dumps info on pending ICI messages.\n", 0);
add_debugger_command_etc("ici_message", &dump_ici_message,
"Dump info on an ICI message",
"\n"
"Dumps info on an ICI message.\n", 0);
if (args->num_cpus > 1) {
sNumCPUs = args->num_cpus;
struct smp_msg* messages;
size_t size = ROUNDUP(sNumCPUs * MSG_ALLOCATE_PER_CPU * sizeof(smp_msg), B_PAGE_SIZE);
area_id area = create_area("smp ici msgs", (void**)&messages, B_ANY_KERNEL_ADDRESS,
size, B_FULL_LOCK, B_KERNEL_READ_AREA | B_KERNEL_WRITE_AREA);
if (area < 0) {
panic("error creating smp msgs");
return area;
}
memset((void*)messages, 0, size);
for (size_t i = 0; i < (size / sizeof(smp_msg)); i++) {
struct smp_msg* msg = &messages[i];
msg->next = sFreeMessages;
sFreeMessages = msg;
sFreeMessageCount++;
}
}
TRACE("smp_init: calling arch_smp_init\n");
return arch_smp_init(args);
}
status_t
smp_per_cpu_init(kernel_args* args, int32 cpu)
{
return arch_smp_per_cpu_init(args, cpu);
}
status_t
smp_init_post_generic_syscalls(void)
{
#if B_DEBUG_SPINLOCK_CONTENTION
return register_generic_syscall(SPINLOCK_CONTENTION,
&spinlock_contention_syscall, 0, 0);
#else
return B_OK;
#endif
}
void
smp_set_num_cpus(int32 numCPUs)
{
sNumCPUs = numCPUs;
}
int32
smp_get_num_cpus()
{
return sNumCPUs;
}
int32
smp_get_current_cpu(void)
{
return thread_get_current_thread()->cpu->cpu_num;
}
static void
call_single_cpu(uint32 targetCPU, void (*func)(void*, int), void* cookie, bool sync)
{
thread_pin_to_current_cpu(thread_get_current_thread());
if (targetCPU == (uint32)smp_get_current_cpu()) {
cpu_status state = disable_interrupts();
func(cookie, smp_get_current_cpu());
restore_interrupts(state);
thread_unpin_from_current_cpu(thread_get_current_thread());
return;
}
if (!sICIEnabled) {
panic("call_single_cpu is not yet available");
thread_unpin_from_current_cpu(thread_get_current_thread());
return;
}
smp_send_ici(targetCPU, SMP_MSG_CALL_FUNCTION, (addr_t)cookie,
0, 0, (void*)func, sync ? SMP_MSG_FLAG_SYNC : SMP_MSG_FLAG_ASYNC);
thread_unpin_from_current_cpu(thread_get_current_thread());
}
void
call_single_cpu(uint32 targetCPU, void (*func)(void*, int), void* cookie)
{
return call_single_cpu(targetCPU, func, cookie, false);
}
void
call_single_cpu_sync(uint32 targetCPU, void (*func)(void*, int), void* cookie)
{
return call_single_cpu(targetCPU, func, cookie, true);
}
static void
call_all_cpus(void (*function)(void*, int), void* cookie, bool sync)
{
if (sNumCPUs == 1) {
cpu_status state = disable_interrupts();
function(cookie, 0);
restore_interrupts(state);
return;
}
if (!sICIEnabled) {
call_all_cpus_early(function, cookie);
return;
}
smp_broadcast_ici(SMP_MSG_CALL_FUNCTION, (addr_t)cookie,
0, 0, (void*)function, sync ? SMP_MSG_FLAG_SYNC : SMP_MSG_FLAG_ASYNC);
}
void
call_all_cpus(void (*func)(void*, int), void* cookie)
{
call_all_cpus(func, cookie, false);
}
void
call_all_cpus_sync(void (*func)(void*, int), void* cookie)
{
call_all_cpus(func, cookie, true);
}
#undef memory_read_barrier
#undef memory_write_barrier
void
memory_read_barrier()
{
memory_read_barrier_inline();
}
void
memory_write_barrier()
{
memory_write_barrier_inline();
}