lstat
linux_to_osf_stat(struct kstat *lstat, struct osf_stat __user *osf_stat)
tmp.st_dev = lstat->dev;
tmp.st_mode = lstat->mode;
tmp.st_nlink = lstat->nlink;
tmp.st_uid = from_kuid_munged(current_user_ns(), lstat->uid);
tmp.st_gid = from_kgid_munged(current_user_ns(), lstat->gid);
tmp.st_rdev = lstat->rdev;
tmp.st_ldev = lstat->rdev;
tmp.st_size = lstat->size;
tmp.st_uatime = lstat->atime.tv_nsec / 1000;
tmp.st_umtime = lstat->mtime.tv_nsec / 1000;
tmp.st_uctime = lstat->ctime.tv_nsec / 1000;
tmp.st_ino = lstat->ino;
tmp.st_atime = lstat->atime.tv_sec;
tmp.st_mtime = lstat->mtime.tv_sec;
tmp.st_ctime = lstat->ctime.tv_sec;
tmp.st_blksize = lstat->blksize;
tmp.st_blocks = lstat->blocks;
lstat[type].samples++;
struct latency_statistic *lstat = iommu->perf_statistic;
u64 val = lstat[i].counter[j];
if (lstat[i].samples)
val = div_u64(val, (lstat[i].samples * 1000));
struct latency_statistic *lstat = iommu->perf_statistic;
return lstat && lstat[type].enabled;
struct latency_statistic *lstat;
iommu->perf_statistic = kzalloc_objs(*lstat, DMAR_LATENCY_NUM,
lstat = iommu->perf_statistic;
if (!lstat[type].enabled) {
lstat[type].enabled = true;
lstat[type].counter[COUNTS_MIN] = UINT_MAX;
struct latency_statistic *lstat = iommu->perf_statistic;
memset(&lstat[type], 0, sizeof(*lstat) * DMAR_LATENCY_NUM);
struct latency_statistic *lstat = iommu->perf_statistic;
lstat[type].counter[COUNTS_10e2]++;
lstat[type].counter[COUNTS_10e3]++;
lstat[type].counter[COUNTS_10e4]++;
lstat[type].counter[COUNTS_10e5]++;
lstat[type].counter[COUNTS_10e6]++;
lstat[type].counter[COUNTS_10e7]++;
lstat[type].counter[COUNTS_10e8_plus]++;
min = lstat[type].counter[COUNTS_MIN];
max = lstat[type].counter[COUNTS_MAX];
lstat[type].counter[COUNTS_MIN] = min_t(u64, min, latency);
lstat[type].counter[COUNTS_MAX] = max_t(u64, max, latency);
lstat[type].counter[COUNTS_SUM] += latency;
static inline void link_status_user_format(u64 lstat,
linfo->link_up = FIELD_GET(RESP_LINKSTAT_UP, lstat);
linfo->full_duplex = FIELD_GET(RESP_LINKSTAT_FDUPLEX, lstat);
linfo->speed = cgx_speed_mbps[FIELD_GET(RESP_LINKSTAT_SPEED, lstat)];
linfo->an = FIELD_GET(RESP_LINKSTAT_AN, lstat);
linfo->fec = FIELD_GET(RESP_LINKSTAT_FEC, lstat);
linfo->lmac_type_id = FIELD_GET(RESP_LINKSTAT_LMAC_TYPE, lstat);
static inline void cgx_link_change_handler(u64 lstat,
link_status_user_format(lstat, &event.link_uinfo, cgx, lmac->lmac_id);
err_type = FIELD_GET(RESP_LINKSTAT_ERRTYPE, lstat);
#define lpfc_qp_spin_lock_irqsave(lock, flag, qp, lstat) \
qp->lock_conflict.lstat++; \
#define lpfc_qp_spin_lock(lock, qp, lstat) \
qp->lock_conflict.lstat++; \
#define lpfc_qp_spin_lock_irqsave(lock, flag, qp, lstat) \
#define lpfc_qp_spin_lock(lock, qp, lstat) spin_lock(lock)
uint32_t lstat, lspeed, lwidth;
pcie_capability_read_dword(ha->pdev, PCI_EXP_LNKCAP, &lstat);
lspeed = FIELD_GET(PCI_EXP_LNKCAP_SLS, lstat);
lwidth = FIELD_GET(PCI_EXP_LNKCAP_MLW, lstat);
struct iscsi_login_stats *lstat = &tiqn->login_stats;
spin_lock(&lstat->lock);
fail_count = (lstat->redirects + lstat->authorize_fails +
lstat->authenticate_fails + lstat->negotiate_fails +
lstat->other_fails);
spin_unlock(&lstat->lock);
struct iscsi_login_stats *lstat = &tiqn->login_stats;
spin_lock(&lstat->lock);
last_fail_time = lstat->last_fail_time ?
(u32)(((u32)lstat->last_fail_time -
spin_unlock(&lstat->lock);
struct iscsi_login_stats *lstat = &tiqn->login_stats;
spin_lock(&lstat->lock);
last_fail_type = lstat->last_fail_type;
spin_unlock(&lstat->lock);
struct iscsi_login_stats *lstat = &tiqn->login_stats;
spin_lock(&lstat->lock);
snprintf(buf, ISCSI_IQN_LEN, "%s", lstat->last_intr_fail_name[0] ?
lstat->last_intr_fail_name : NONE);
spin_unlock(&lstat->lock);
struct iscsi_login_stats *lstat = &tiqn->login_stats;
spin_lock(&lstat->lock);
if (lstat->last_intr_fail_ip_family == AF_INET6)
spin_unlock(&lstat->lock);
struct iscsi_login_stats *lstat = &tiqn->login_stats;
spin_lock(&lstat->lock);
ret = snprintf(page, PAGE_SIZE, "%pISc\n", &lstat->last_intr_fail_sockaddr);
spin_unlock(&lstat->lock);
struct iscsi_login_stats *lstat = &tiqn->login_stats;
spin_lock(&lstat->lock);
ret = snprintf(page, PAGE_SIZE, "%u\n", lstat->accepts);
spin_unlock(&lstat->lock);
struct iscsi_login_stats *lstat = &tiqn->login_stats;
spin_lock(&lstat->lock);
ret = snprintf(page, PAGE_SIZE, "%u\n", lstat->other_fails);
spin_unlock(&lstat->lock);
struct iscsi_login_stats *lstat = &tiqn->login_stats;
spin_lock(&lstat->lock);
ret = snprintf(page, PAGE_SIZE, "%u\n", lstat->redirects);
spin_unlock(&lstat->lock);
struct iscsi_login_stats *lstat = &tiqn->login_stats;
spin_lock(&lstat->lock);
ret = snprintf(page, PAGE_SIZE, "%u\n", lstat->authorize_fails);
spin_unlock(&lstat->lock);
struct iscsi_login_stats *lstat = &tiqn->login_stats;
spin_lock(&lstat->lock);
ret = snprintf(page, PAGE_SIZE, "%u\n", lstat->authenticate_fails);
spin_unlock(&lstat->lock);
struct iscsi_login_stats *lstat = &tiqn->login_stats;
spin_lock(&lstat->lock);
ret = snprintf(page, PAGE_SIZE, "%u\n", lstat->negotiate_fails);
spin_unlock(&lstat->lock);
SYSCALL_DEFINE2(lstat, const char __user *, filename,
if (lstat(linkname, &st) < 0 || st.st_size + 1 > (off_t)sizeof(pathname))
if (!lstat("./tests", &st))
if (!lstat(path_dir, &st))
if (lstat(path2, &st) == 0 && (st.st_mode & S_IFMT) == S_IFLNK) {
ret = lstat(path, &statbuf);
ret = lstat(namebuf, &statbuf);
if (lstat(path, &sb) != 0)
r = lstat(path2, &buf);
r = lstat(path2, &buf);
rv = lstat(pathname, &st);
rv = lstat(pathname, &st);
if (lstat(TMON_LOG_FILE, &logstat) < 0) {