__n
void *__memcpy(void *__dest, __const void *__src, size_t __n) __alias(memcpy);
void *memcpy(void *__dest, __const void *__src, size_t __n)
for (i = __n >> 3; i > 0; i--) {
if (__n & 1 << 2) {
if (__n & 1 << 1) {
if (__n & 1)
register unsigned long long __n asm("r0") = *n;
: "+r" (__n), "=r" (__res)
__rem = __n >> 32;
struct fgt_masks __n = { .str = #__m, .res0 = __m ## _RES0, .res1 = __m ## _RES1 }
extern void *memcpy(void *__to, __const__ void *__from, size_t __n);
extern void *memset(void *__to, int c, size_t __n);
extern void *memcpy(void *__to, __const__ void *__from, size_t __n);
extern void *__memcpy(void *__to, __const__ void *__from, size_t __n);
extern void *memmove(void *__dest, __const__ void *__src, size_t __n);
extern void *__memmove(void *__dest, __const__ void *__src, size_t __n);
} __n; \
__n.n64 = (n); \
if ((__upper = __n.n32[0])) { \
: "=d" (__n.n32[0]), "=d" (__upper) \
: "d" (__base), "0" (__n.n32[0])); \
: "=d" (__n.n32[1]), "=d" (__rem) \
: "d" (__base), "1" (__upper), "0" (__n.n32[1])); \
(n) = __n.n64; \
extern void *memcpy(void *__to, __const__ void *__from, size_t __n);
extern void *memmove(void *__dest, __const__ void *__src, size_t __n);
extern size_t __raw_copy_from_user(void *__to, const void *__from, size_t __n);
extern size_t __raw_copy_to_user(void *__to, const void *__from, size_t __n);
void *memcpy(void *__dest, __const void *__src, size_t __n)
for (i = 0; i < __n; i++)
size_t __n)
for (i=0;i<__n;i++) d[i] = s[i];
static inline int strncmp(const char *__cs, const char *__ct, size_t __n)
if (__n == 0)
: "0" (__cs), "1" (__ct), "r" (__cs+__n)
extern void *memcpy(void *__to, __const__ void *__from, size_t __n);
extern void *memmove(void *__dest, __const__ void *__src, size_t __n);
extern void *memchr(const void *__s, int __c, size_t __n);
: "0" (__cs), "1" (__ct), "r" ((uintptr_t)__cs+__n));
extern void *memcpy(void *__to, __const__ void *__from, size_t __n);
extern void *__memcpy(void *__to, __const__ void *__from, size_t __n);
extern void *memmove(void *__dest, __const__ void *__src, size_t __n);
extern void *__memmove(void *__dest, __const__ void *__src, size_t __n);
static inline char *strncpy(char *__dest, const char *__src, size_t __n)
if (__n == 0)
: "0" (__dest), "1" (__src), "r" ((uintptr_t)__src+__n)
static inline int strncmp(const char *__cs, const char *__ct, size_t __n)
void *__memcpy(void *__dest, const void *__src, size_t __n) __alias(memcpy);
#define REG_MTX(__n) (0x4f + (__n) - 1)
#define REG_LCC(__n) (0x62 + (__n) - 1)
#define SOLO_VI_STATUS0_PAGE(__n) ((__n) & 0x07)
#define SOLO_SYS_CFG_OUTDIV(__n) (((__n) & 0x003) << 3)
#define SOLO_SYS_CFG_FEEDBACKDIV(__n) (((__n) & 0x1ff) << 5)
#define SOLO_SYS_CFG_INPUTDIV(__n) (((__n) & 0x01f) << 14)
#define SOLO_SYS_CFG_NCLK_DELAY(__n) (((__n) & 0x003) << 24)
#define SOLO_SYS_CFG_PCLK_DELAY(__n) (((__n) & 0x00f) << 26)
#define GPOENCTLR_GPOEN(__n) (1 << (__n))
#define MCUCTL_REG_ISSR(__n) (MCUCTL_BASE + 0x80 + ((__n) * 4))
#define INTGR0_INTGC(__n) (1 << ((__n) + 16))
#define INTGR0_INTGD(__n) (1 << (__n))
#define INTCR0_INTGC(__n) (1 << ((__n) + 16))
#define INTCR0_INTCD(__n) (1 << ((__n) + 16))
#define INTMR0_INTMC(__n) (1 << ((__n) + 16))
#define INTMR0_INTMD(__n) (1 << (__n))
#define INTSR0_GET_INTSD(x, __n) (((x) >> (__n)) & 0x1)
#define INTSR0_GET_INTSC(x, __n) (((x) >> ((__n) + 16)) & 0x1)
#define INTMSR0_GET_INTMSD(x, __n) (((x) >> (__n)) & 0x1)
#define INTMSR0_GET_INTMSC(x, __n) (((x) >> ((__n) + 16)) & 0x1)
#define INTGR1_INTGC(__n) (1 << (__n))
#define INTCR1_INTCC(__n) (1 << (__n))
#define INTMR1_INTMC(__n) (1 << (__n))
#define INTCR2_INTCC(__n) (1 << ((__n) + 16))
#define INTMR2_INTMCIS(__n) (1 << (__n))
#define GPOCTLR_GPOG(__n) (1 << (__n))
#define NVM_FLREG(__n) (NVM_FRBA + ((__n) * 4))
#define CT2_CSI_MAC_CONTROL_REG(__n) \
(__n) * (CT2_CSI_MAC1_CONTROL_REG - CT2_CSI_MAC0_CONTROL_REG))
#define TSNEP_QUEUE(n) ({ typeof(n) __n = (n); \
(__n) == 0 ? \
TSNEP_MAC_SIZE + TSNEP_QUEUE_SIZE * ((__n) - 1); })
#define CT2_CSI_MAC_CONTROL_REG(__n) \
(__n) * (CT2_CSI_MAC1_CONTROL_REG - CT2_CSI_MAC0_CONTROL_REG))
#define CSIO_VALID_WWN(__n) ((*__n >> 4) == 0x5 ? true : false)
u64 (__n) = ((u64) n_hi) << 32 | (n_lo); \
(r32) = do_div ((__n), (d32)); \
(q32) = (u32) (__n); \
unsigned int __n = (n), __cur_n; \
while (__n) { \
__cur_n = min((it)->iter.bi_size, __n); \
__n -= __cur_n; \
unsigned long __n = (n); \
__kfifo_in_r(__kfifo, __buf, __n, __recsize) : \
__kfifo_in(__kfifo, __buf, __n); \
unsigned long __n = (n); \
__kfifo_out_r(__kfifo, __buf, __n, __recsize) : \
__kfifo_out(__kfifo, __buf, __n); \
unsigned long __n = (n); \
__kfifo_out_peek_r(__kfifo, __buf, __n, __recsize) : \
__kfifo_out_peek(__kfifo, __buf, __n); \
unsigned long __n = (n); \
__kfifo_out_linear_r(__kfifo, __tail, __n, __recsize) : \
__kfifo_out_linear(__kfifo, __tail, __n); \
unsigned int __i, __n; \
for (__i = 0, __n = 0; __i < (size); \
__i += __entry->next_offset, __n++) { \
if (__n < n) \
static int __init __xbc_parse_value(char **__v, char **__n)
*__n = p;
#define for_each_kmem_cache_node(__s, __node, __n) \
if ((__n = get_node(__s, __node)))
struct tipc_crypto *tipc_node_crypto_rx(struct tipc_node *__n)
return (__n) ? __n->crypto_rx : NULL;
struct tipc_crypto *tipc_node_crypto_rx(struct tipc_node *__n);
#define ADD(__n, __v) \
if (__v && bt_ctf_writer_add_environment_field(writer, __n, __v)) \
#define SET(__n, __v) \
if (bt_ctf_clock_set_##__n(clock, __v)) \
#define DEBUG_WRAPPER(__n, __l) \
static int pr_ ## __n ## _wrapper(const char *fmt, ...) \