t_start
unsigned long t_start = (unsigned long)_text;
if (overlaps(start, end, t_start, t_end))
unsigned long t_start = (unsigned long)_text;
if (!kernel_set_to_readonly || !overlaps(start, end, t_start, t_end))
ktime_t t_start;
t_start = ktime_get();
elapsed_us = ktime_us_delta(ktime_get(), t_start);
unsigned long t_start = jiffies;
if (time_after(jiffies, t_start + PHY_IAC_TIMEOUT))
smc->hw.t_stop = smc->hw.t_start;
u_long t_start ; /* HWT start */
smc->hw.t_start = 0 ;
if ((tr > smc->hw.t_start) || (is & IS_TIMINT)) {
smc->hw.t_stop = smc->hw.t_start ;
smc->hw.t_stop = smc->hw.t_start - tr ;
smc->hw.t_start = time ;
ktime_t t_start, t_jlck, t_jhd, t_tlck, t_rep;
t_start = ktime_get();
jd->jd_jid, ktime_ms_delta(t_rep, t_start),
ktime_ms_delta(t_jlck, t_start),
commit_transaction->t_start = jiffies;
commit_transaction->t_start);
stats.run.rs_running = jbd2_time_diff(commit_transaction->t_start,
transaction->t_start = jiffies;
if (time_after(transaction->t_start, ts)) {
newts = jbd2_time_diff(ts, transaction->t_start);
current_tail = le32_to_cpu(tl->tl_recs[tail_index].t_start);
index, le32_to_cpu(tl->tl_recs[index].t_start),
tl->tl_recs[index].t_start = cpu_to_le32(start_cluster);
le32_to_cpu(rec.t_start));
i, le32_to_cpu(rec.t_start), num_clusters);
start_cluster = le32_to_cpu(tl->tl_recs[i].t_start);
__le32 t_start; /* 1st cluster in this log */
.start = t_start,
unsigned long t_start;
.start = t_start,
.start = t_start,
.start = t_start,
.start = t_start,
.start = t_start,
.start = t_start,
.start = t_start,
clock_gettime(CLOCK_MONOTONIC, &t_start);
start_ns = t_start.tv_sec * NS_PER_SEC + t_start.tv_nsec;
struct timespec t_start = {0, 0}, t_end = {0, 0};
struct timespec *t_start, struct timespec *t_end)
int64_t ts_delta = timespec_to_ns64(t_end) - timespec_to_ns64(t_start);