Symbol: df
arch/mips/include/uapi/asm/inst.h
784
__BITFIELD_FIELD(unsigned int df : 2,
arch/mips/kernel/unaligned.c
464
enum msa_2b_fmt df;
arch/mips/kernel/unaligned.c
477
df = insn.msa_mi10_format.df;
arch/mips/kernel/unaligned.c
511
write_msa_wr(wd, fpr, df);
arch/mips/kernel/unaligned.c
529
read_msa_wr(wd, fpr, df);
arch/parisc/math-emu/decode_exc.c
60
(current_ir >> ((current_ir>>opclass & twobits) == 1 ? df : fmtbits) & twobits)
arch/parisc/math-emu/fpudispatch.c
279
u_int df; /* for class 1 conversions */
arch/parisc/math-emu/fpudispatch.c
407
df = extru(ir,fpdfpos,2); /* get dest format */
arch/parisc/math-emu/fpudispatch.c
408
if ((df & 2) || (fmt & 2)) {
arch/parisc/math-emu/fpudispatch.c
420
fmt = (fmt << 1) | df;
arch/parisc/math-emu/fpudispatch.c
726
u_int df; /* dest format for class 1 conversions */
arch/parisc/math-emu/fpudispatch.c
840
df = extru(ir,fpdfpos,2); /* get dest format */
arch/parisc/math-emu/fpudispatch.c
845
if (df == DBL) {
arch/parisc/math-emu/fpudispatch.c
848
if ((df & 2) || (fmt & 2))
arch/parisc/math-emu/fpudispatch.c
851
fmt = (fmt << 1) | df;
arch/s390/hypfs/hypfs.h
79
extern void hypfs_dbfs_create_file(struct hypfs_dbfs_file *df);
arch/s390/hypfs/hypfs.h
80
extern void hypfs_dbfs_remove_file(struct hypfs_dbfs_file *df);
arch/s390/hypfs/hypfs_dbfs.c
36
struct hypfs_dbfs_file *df;
arch/s390/hypfs/hypfs_dbfs.c
42
df = file_inode(file)->i_private;
arch/s390/hypfs/hypfs_dbfs.c
43
if (mutex_lock_interruptible(&df->lock))
arch/s390/hypfs/hypfs_dbfs.c
46
data = hypfs_dbfs_data_alloc(df);
arch/s390/hypfs/hypfs_dbfs.c
48
mutex_unlock(&df->lock);
arch/s390/hypfs/hypfs_dbfs.c
51
rc = df->data_create(&data->buf, &data->buf_free_ptr, &data->size);
arch/s390/hypfs/hypfs_dbfs.c
53
mutex_unlock(&df->lock);
arch/s390/hypfs/hypfs_dbfs.c
57
mutex_unlock(&df->lock);
arch/s390/hypfs/hypfs_dbfs.c
66
struct hypfs_dbfs_file *df = file_inode(file)->i_private;
arch/s390/hypfs/hypfs_dbfs.c
69
mutex_lock(&df->lock);
arch/s390/hypfs/hypfs_dbfs.c
70
rc = df->unlocked_ioctl(file, cmd, arg);
arch/s390/hypfs/hypfs_dbfs.c
71
mutex_unlock(&df->lock);
arch/s390/hypfs/hypfs_dbfs.c
84
void hypfs_dbfs_create_file(struct hypfs_dbfs_file *df)
arch/s390/hypfs/hypfs_dbfs.c
88
if (df->unlocked_ioctl && !security_locked_down(LOCKDOWN_DEBUGFS))
arch/s390/hypfs/hypfs_dbfs.c
90
df->dentry = debugfs_create_file(df->name, 0400, dbfs_dir, df, fops);
arch/s390/hypfs/hypfs_dbfs.c
91
mutex_init(&df->lock);
arch/s390/hypfs/hypfs_dbfs.c
94
void hypfs_dbfs_remove_file(struct hypfs_dbfs_file *df)
arch/s390/hypfs/hypfs_dbfs.c
96
debugfs_remove(df->dentry);
arch/x86/coco/sev/vc-shared.c
322
bool df = ((regs->flags & X86_EFLAGS_DF) == X86_EFLAGS_DF);
arch/x86/coco/sev/vc-shared.c
348
exit_info_2, df);
arch/x86/coco/sev/vc-shared.c
370
exit_info_2, df);
arch/x86/coco/sev/vc-shared.c
374
if (df)
arch/x86/coco/sev/vc-shared.c
379
if (df)
arch/x86/kernel/cpu/common.c
693
const struct cpuid_dependent_feature *df;
arch/x86/kernel/cpu/common.c
695
for (df = cpuid_dependent_features; df->feature; df++) {
arch/x86/kernel/cpu/common.c
697
if (!cpu_has(c, df->feature))
arch/x86/kernel/cpu/common.c
706
if (!((s32)df->level < 0 ?
arch/x86/kernel/cpu/common.c
707
(u32)df->level > (u32)c->extended_cpuid_level :
arch/x86/kernel/cpu/common.c
708
(s32)df->level > (s32)c->cpuid_level))
arch/x86/kernel/cpu/common.c
711
clear_cpu_cap(c, df->feature);
arch/x86/kernel/cpu/common.c
716
x86_cap_flags[df->feature], df->level);
arch/x86/kvm/emulate.c
2999
int df = (ctxt->eflags & X86_EFLAGS_DF) ? -op->count : op->count;
arch/x86/kvm/emulate.c
3001
register_address_increment(ctxt, reg, df * op->bytes);
crypto/testmgr.h
1023
be64_to_cpua(df, 28, 8c, 1b, fa, f9, 95, 88),
crypto/testmgr.h
1058
be64_to_cpua(df, d8, 68, a2, 64, 42, 65, f3),
crypto/testmgr.h
1275
be64_to_cpua(e4, df, 9b, 7d, 9e, 47, ca, 33),
crypto/testmgr.h
1284
be64_to_cpua(df, c4, c0, db, 54, db, 8a, 0d),
crypto/testmgr.h
809
be64_to_cpua(28, 6a, df, 97, fd, 82, 76, 24),
crypto/testmgr.h
975
be64_to_cpua(b8, 6d, 87, 81, 43, df, fb, 9f),
drivers/block/drbd/drbd_int.h
1618
enum drbd_force_detach_flags df,
drivers/block/drbd/drbd_int.h
1628
if (df == DRBD_READ_ERROR || df == DRBD_WRITE_ERROR) {
drivers/block/drbd/drbd_int.h
1659
if (df == DRBD_READ_ERROR)
drivers/block/drbd/drbd_int.h
1661
if (df == DRBD_FORCE_DETACH)
drivers/clk/clk-sp7021.c
237
u32 df; /* diff freq */
drivers/clk/clk-sp7021.c
253
df = df0 - df1;
drivers/clk/clk-sp7021.c
255
df = f * (nint) / pp[0];
drivers/clk/clk-sp7021.c
258
df_quotient = df / m;
drivers/clk/clk-sp7021.c
259
df_remainder = ((df % m) * 1000) / m;
drivers/clk/clk-sp7021.c
279
fout = df / m;
drivers/clk/rockchip/clk.h
1049
#define DIV(_id, cname, pname, f, o, s, w, df) \
drivers/clk/rockchip/clk.h
1060
.div_flags = df, \
drivers/clk/rockchip/clk.h
1064
#define DIVTBL(_id, cname, pname, f, o, s, w, df, dt) \
drivers/clk/rockchip/clk.h
1075
.div_flags = df, \
drivers/clk/rockchip/clk.h
1183
df, go, gs, gf) \
drivers/clk/rockchip/clk.h
1197
.div_flags = df, \
drivers/clk/rockchip/clk.h
1204
ds, dw, df) \
drivers/clk/rockchip/clk.h
1218
.div_flags = df, \
drivers/clk/rockchip/clk.h
1222
#define COMPOSITE_NOMUX_HALFDIV(_id, cname, pname, f, mo, ds, dw, df, \
drivers/clk/rockchip/clk.h
1234
.div_flags = df, \
drivers/clk/rockchip/clk.h
1240
#define DIV_HALF(_id, cname, pname, f, o, s, w, df) \
drivers/clk/rockchip/clk.h
1251
.div_flags = df, \
drivers/clk/rockchip/clk.h
797
df, go, gs, gf) \
drivers/clk/rockchip/clk.h
811
.div_flags = df, \
drivers/clk/rockchip/clk.h
818
mf, do, ds, dw, df, go, gs, gf) \
drivers/clk/rockchip/clk.h
833
.div_flags = df, \
drivers/clk/rockchip/clk.h
839
#define COMPOSITE_NOMUX(_id, cname, pname, f, mo, ds, dw, df, \
drivers/clk/rockchip/clk.h
851
.div_flags = df, \
drivers/clk/rockchip/clk.h
858
df, dt, go, gs, gf) \
drivers/clk/rockchip/clk.h
869
.div_flags = df, \
drivers/clk/rockchip/clk.h
895
ds, dw, df) \
drivers/clk/rockchip/clk.h
909
.div_flags = df, \
drivers/clk/rockchip/clk.h
914
mw, mf, ds, dw, df, dt) \
drivers/clk/rockchip/clk.h
928
.div_flags = df, \
drivers/clk/rockchip/clk.h
933
#define COMPOSITE_FRAC(_id, cname, pname, f, mo, df, go, gs, gf)\
drivers/clk/rockchip/clk.h
944
.div_flags = df, \
drivers/clk/rockchip/clk.h
950
#define COMPOSITE_FRACMUX(_id, cname, pname, f, mo, df, go, gs, gf, ch) \
drivers/clk/rockchip/clk.h
961
.div_flags = df, \
drivers/clk/rockchip/clk.h
968
#define COMPOSITE_FRACMUX_NOGATE(_id, cname, pname, f, mo, df, ch) \
drivers/clk/rockchip/clk.h
979
.div_flags = df, \
drivers/clk/rockchip/clk.h
985
ds, dw, df) \
drivers/clk/rockchip/clk.h
998
.div_flags = df, \
drivers/clk/samsung/clk.h
194
#define __DIV(_id, cname, pname, o, s, w, f, df, t) \
drivers/clk/samsung/clk.h
203
.div_flags = df, \
drivers/clk/samsung/clk.h
210
#define DIV_F(_id, cname, pname, o, s, w, f, df) \
drivers/clk/samsung/clk.h
211
__DIV(_id, cname, pname, o, s, w, f, df, NULL)
drivers/devfreq/devfreq.c
1407
struct devfreq *df = to_devfreq(dev);
drivers/devfreq/devfreq.c
1408
return sprintf(buf, "%s\n", dev_name(df->dev.parent));
drivers/devfreq/devfreq.c
1415
struct devfreq *df = to_devfreq(dev);
drivers/devfreq/devfreq.c
1417
if (!df->governor)
drivers/devfreq/devfreq.c
1420
return sprintf(buf, "%s\n", df->governor->name);
drivers/devfreq/devfreq.c
1426
struct devfreq *df = to_devfreq(dev);
drivers/devfreq/devfreq.c
1431
if (!df->governor)
drivers/devfreq/devfreq.c
1444
if (df->governor == governor) {
drivers/devfreq/devfreq.c
1447
} else if (IS_SUPPORTED_FLAG(df->governor->flags, IMMUTABLE)
drivers/devfreq/devfreq.c
1457
ret = df->governor->event_handler(df, DEVFREQ_GOV_STOP, NULL);
drivers/devfreq/devfreq.c
1460
__func__, df->governor->name, ret);
drivers/devfreq/devfreq.c
1463
remove_sysfs_files(df, df->governor);
drivers/devfreq/devfreq.c
1469
prev_governor = df->governor;
drivers/devfreq/devfreq.c
1470
df->governor = governor;
drivers/devfreq/devfreq.c
1471
ret = df->governor->event_handler(df, DEVFREQ_GOV_START, NULL);
drivers/devfreq/devfreq.c
1474
__func__, df->governor->name, ret);
drivers/devfreq/devfreq.c
1477
df->governor = prev_governor;
drivers/devfreq/devfreq.c
1478
ret = df->governor->event_handler(df, DEVFREQ_GOV_START, NULL);
drivers/devfreq/devfreq.c
1483
df->governor = NULL;
drivers/devfreq/devfreq.c
1492
create_sysfs_files(df, df->governor);
drivers/devfreq/devfreq.c
1507
struct devfreq *df = to_devfreq(d);
drivers/devfreq/devfreq.c
1510
if (!df->governor)
drivers/devfreq/devfreq.c
1519
if (IS_SUPPORTED_FLAG(df->governor->flags, IMMUTABLE)) {
drivers/devfreq/devfreq.c
1521
"%s ", df->governor->name);
drivers/devfreq/devfreq.c
1553
struct devfreq *df = to_devfreq(dev);
drivers/devfreq/devfreq.c
1555
if (!df->profile)
drivers/devfreq/devfreq.c
1558
if (df->profile->get_cur_freq &&
drivers/devfreq/devfreq.c
1559
!df->profile->get_cur_freq(df->dev.parent, &freq))
drivers/devfreq/devfreq.c
1562
return sprintf(buf, "%lu\n", df->previous_freq);
drivers/devfreq/devfreq.c
1569
struct devfreq *df = to_devfreq(dev);
drivers/devfreq/devfreq.c
1571
return sprintf(buf, "%lu\n", df->previous_freq);
drivers/devfreq/devfreq.c
1578
struct devfreq *df = to_devfreq(dev);
drivers/devfreq/devfreq.c
1586
if (!dev_pm_qos_request_active(&df->user_min_freq_req))
drivers/devfreq/devfreq.c
1594
ret = dev_pm_qos_update_request(&df->user_min_freq_req,
drivers/devfreq/devfreq.c
1605
struct devfreq *df = to_devfreq(dev);
drivers/devfreq/devfreq.c
1608
mutex_lock(&df->lock);
drivers/devfreq/devfreq.c
1609
devfreq_get_freq_range(df, &min_freq, &max_freq);
drivers/devfreq/devfreq.c
1610
mutex_unlock(&df->lock);
drivers/devfreq/devfreq.c
1619
struct devfreq *df = to_devfreq(dev);
drivers/devfreq/devfreq.c
1627
if (!dev_pm_qos_request_active(&df->user_max_freq_req))
drivers/devfreq/devfreq.c
1649
ret = dev_pm_qos_update_request(&df->user_max_freq_req, value);
drivers/devfreq/devfreq.c
1659
struct devfreq *df = to_devfreq(dev);
drivers/devfreq/devfreq.c
1662
mutex_lock(&df->lock);
drivers/devfreq/devfreq.c
1663
devfreq_get_freq_range(df, &min_freq, &max_freq);
drivers/devfreq/devfreq.c
1664
mutex_unlock(&df->lock);
drivers/devfreq/devfreq.c
1674
struct devfreq *df = to_devfreq(d);
drivers/devfreq/devfreq.c
1678
if (!df->profile)
drivers/devfreq/devfreq.c
1681
mutex_lock(&df->lock);
drivers/devfreq/devfreq.c
1683
for (i = 0; i < df->max_state; i++)
drivers/devfreq/devfreq.c
1685
"%lu ", df->freq_table[i]);
drivers/devfreq/devfreq.c
1687
mutex_unlock(&df->lock);
drivers/devfreq/devfreq.c
1701
struct devfreq *df = to_devfreq(dev);
drivers/devfreq/devfreq.c
1706
if (!df->profile)
drivers/devfreq/devfreq.c
1708
max_state = df->max_state;
drivers/devfreq/devfreq.c
1713
mutex_lock(&df->lock);
drivers/devfreq/devfreq.c
1714
if (!df->stop_polling &&
drivers/devfreq/devfreq.c
1715
devfreq_update_status(df, df->previous_freq)) {
drivers/devfreq/devfreq.c
1716
mutex_unlock(&df->lock);
drivers/devfreq/devfreq.c
1719
mutex_unlock(&df->lock);
drivers/devfreq/devfreq.c
1727
df->freq_table[i]);
drivers/devfreq/devfreq.c
1737
if (df->freq_table[i] == df->previous_freq)
drivers/devfreq/devfreq.c
1743
len += sysfs_emit_at(buf, len, "%10lu:", df->freq_table[i]);
drivers/devfreq/devfreq.c
1748
df->stats.trans_table[(i * max_state) + j]);
drivers/devfreq/devfreq.c
1753
jiffies64_to_msecs(df->stats.time_in_state[i]));
drivers/devfreq/devfreq.c
1758
df->stats.total_trans);
drivers/devfreq/devfreq.c
1771
struct devfreq *df = to_devfreq(dev);
drivers/devfreq/devfreq.c
1774
if (!df->profile)
drivers/devfreq/devfreq.c
1777
if (df->max_state == 0)
drivers/devfreq/devfreq.c
1784
mutex_lock(&df->lock);
drivers/devfreq/devfreq.c
1785
memset(df->stats.time_in_state, 0, (df->max_state *
drivers/devfreq/devfreq.c
1786
sizeof(*df->stats.time_in_state)));
drivers/devfreq/devfreq.c
1787
memset(df->stats.trans_table, 0, array3_size(sizeof(unsigned int),
drivers/devfreq/devfreq.c
1788
df->max_state,
drivers/devfreq/devfreq.c
1789
df->max_state));
drivers/devfreq/devfreq.c
1790
df->stats.total_trans = 0;
drivers/devfreq/devfreq.c
1791
df->stats.last_update = get_jiffies_64();
drivers/devfreq/devfreq.c
1792
mutex_unlock(&df->lock);
drivers/devfreq/devfreq.c
1815
struct devfreq *df = to_devfreq(dev);
drivers/devfreq/devfreq.c
1817
if (!df->profile)
drivers/devfreq/devfreq.c
1820
return sprintf(buf, "%d\n", df->profile->polling_ms);
drivers/devfreq/devfreq.c
1827
struct devfreq *df = to_devfreq(dev);
drivers/devfreq/devfreq.c
1831
if (!df->governor)
drivers/devfreq/devfreq.c
1838
df->governor->event_handler(df, DEVFREQ_GOV_UPDATE_INTERVAL, &value);
drivers/devfreq/devfreq.c
1848
struct devfreq *df = to_devfreq(dev);
drivers/devfreq/devfreq.c
1850
if (!df->profile)
drivers/devfreq/devfreq.c
1853
return sprintf(buf, "%s\n", timer_name[df->profile->timer]);
drivers/devfreq/devfreq.c
1859
struct devfreq *df = to_devfreq(dev);
drivers/devfreq/devfreq.c
1864
if (!df->governor || !df->profile)
drivers/devfreq/devfreq.c
1883
if (df->profile->timer == timer) {
drivers/devfreq/devfreq.c
1888
mutex_lock(&df->lock);
drivers/devfreq/devfreq.c
1889
df->profile->timer = timer;
drivers/devfreq/devfreq.c
1890
mutex_unlock(&df->lock);
drivers/devfreq/devfreq.c
1892
ret = df->governor->event_handler(df, DEVFREQ_GOV_STOP, NULL);
drivers/devfreq/devfreq.c
1895
__func__, df->governor->name, ret);
drivers/devfreq/devfreq.c
1899
ret = df->governor->event_handler(df, DEVFREQ_GOV_START, NULL);
drivers/devfreq/devfreq.c
1902
__func__, df->governor->name, ret);
drivers/devfreq/devfreq.c
1908
#define CREATE_SYSFS_FILE(df, name) \
drivers/devfreq/devfreq.c
1911
ret = sysfs_create_file(&df->dev.kobj, &dev_attr_##name.attr); \
drivers/devfreq/devfreq.c
1913
dev_warn(&df->dev, \
drivers/devfreq/governor_performance.c
13
static int devfreq_performance_func(struct devfreq *df,
drivers/devfreq/governor_powersave.c
13
static int devfreq_powersave_func(struct devfreq *df,
drivers/devfreq/governor_simpleondemand.c
18
static int devfreq_simple_ondemand_func(struct devfreq *df,
drivers/devfreq/governor_simpleondemand.c
26
struct devfreq_simple_ondemand_data *data = df->data;
drivers/devfreq/governor_simpleondemand.c
28
err = devfreq_update_stats(df);
drivers/devfreq/governor_simpleondemand.c
32
stat = &df->last_status;
drivers/devfreq/governor_userspace.c
23
static int devfreq_userspace_func(struct devfreq *df, unsigned long *freq)
drivers/devfreq/governor_userspace.c
25
struct userspace_data *data = df->governor_data;
drivers/devfreq/governor_userspace.c
30
*freq = df->previous_freq; /* No user freq specified yet */
drivers/devfreq/hisi_uncore_freq.c
349
static int hisi_platform_gov_func(struct devfreq *df, unsigned long *freq)
drivers/devfreq/hisi_uncore_freq.c
360
static int hisi_platform_gov_handler(struct devfreq *df, unsigned int event,
drivers/devfreq/hisi_uncore_freq.c
363
struct hisi_uncore_freq *uncore = dev_get_drvdata(df->dev.parent);
drivers/dma/fsl-dpaa2-qdma/dpaa2-qdma.h
24
} df;
drivers/dma/imx-sdma.c
319
u32 df : 1;
drivers/gpu/drm/amd/amdgpu/amdgpu.h
1099
struct amdgpu_df df;
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
3139
adev->df.funcs = &df_v3_6_funcs;
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
3146
adev->df.funcs = &df_v1_7_funcs;
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
3149
adev->df.funcs = &df_v4_3_funcs;
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
3152
adev->df.funcs = &df_v4_6_2_funcs;
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
3156
adev->df.funcs = &df_v4_15_funcs;
drivers/gpu/drm/amd/amdgpu/amdgpu_pmu.c
236
if ((!pe->adev->df.funcs) ||
drivers/gpu/drm/amd/amdgpu/amdgpu_pmu.c
237
(!pe->adev->df.funcs->pmc_start))
drivers/gpu/drm/amd/amdgpu/amdgpu_pmu.c
247
target_cntr = pe->adev->df.funcs->pmc_start(pe->adev,
drivers/gpu/drm/amd/amdgpu/amdgpu_pmu.c
256
pe->adev->df.funcs->pmc_start(pe->adev, hwc->config,
drivers/gpu/drm/amd/amdgpu/amdgpu_pmu.c
275
if ((!pe->adev->df.funcs) ||
drivers/gpu/drm/amd/amdgpu/amdgpu_pmu.c
276
(!pe->adev->df.funcs->pmc_get_count))
drivers/gpu/drm/amd/amdgpu/amdgpu_pmu.c
284
pe->adev->df.funcs->pmc_get_count(pe->adev,
drivers/gpu/drm/amd/amdgpu/amdgpu_pmu.c
307
if ((!pe->adev->df.funcs) ||
drivers/gpu/drm/amd/amdgpu/amdgpu_pmu.c
308
(!pe->adev->df.funcs->pmc_stop))
drivers/gpu/drm/amd/amdgpu/amdgpu_pmu.c
314
pe->adev->df.funcs->pmc_stop(pe->adev, hwc->config, hwc->idx,
drivers/gpu/drm/amd/amdgpu/amdgpu_pmu.c
340
if ((!pe->adev->df.funcs) ||
drivers/gpu/drm/amd/amdgpu/amdgpu_pmu.c
341
(!pe->adev->df.funcs->pmc_start))
drivers/gpu/drm/amd/amdgpu/amdgpu_pmu.c
360
target_cntr = pe->adev->df.funcs->pmc_start(pe->adev,
drivers/gpu/drm/amd/amdgpu/amdgpu_pmu.c
389
if ((!pe->adev->df.funcs) ||
drivers/gpu/drm/amd/amdgpu/amdgpu_pmu.c
390
(!pe->adev->df.funcs->pmc_stop))
drivers/gpu/drm/amd/amdgpu/amdgpu_pmu.c
398
pe->adev->df.funcs->pmc_stop(pe->adev, hwc->config, hwc->idx,
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
4092
} else if (adev->df.funcs &&
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
4093
adev->df.funcs->query_ras_poison_mode &&
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
4097
adev->df.funcs->query_ras_poison_mode(adev);
drivers/gpu/drm/amd/amdgpu/amdgpu_xgmi.c
544
if ((!adev->df.funcs) ||
drivers/gpu/drm/amd/amdgpu/amdgpu_xgmi.c
545
(!adev->df.funcs->get_fica) ||
drivers/gpu/drm/amd/amdgpu/amdgpu_xgmi.c
546
(!adev->df.funcs->set_fica))
drivers/gpu/drm/amd/amdgpu/amdgpu_xgmi.c
549
fica_out = adev->df.funcs->get_fica(adev, ficaa_pie_ctl_in);
drivers/gpu/drm/amd/amdgpu/amdgpu_xgmi.c
553
fica_out = adev->df.funcs->get_fica(adev, ficaa_pie_status_in);
drivers/gpu/drm/amd/amdgpu/amdgpu_xgmi.c
558
adev->df.funcs->set_fica(adev, ficaa_pie_status_in, 0, 0);
drivers/gpu/drm/amd/amdgpu/df_v1_7.c
100
adev->df.funcs->enable_broadcast_mode(adev, false);
drivers/gpu/drm/amd/amdgpu/df_v1_7.c
34
adev->df.hash_status.hash_64k = false;
drivers/gpu/drm/amd/amdgpu/df_v1_7.c
35
adev->df.hash_status.hash_2m = false;
drivers/gpu/drm/amd/amdgpu/df_v1_7.c
36
adev->df.hash_status.hash_1g = false;
drivers/gpu/drm/amd/amdgpu/df_v1_7.c
72
fb_channel_number = adev->df.funcs->get_fb_channel_number(adev);
drivers/gpu/drm/amd/amdgpu/df_v1_7.c
85
adev->df.funcs->enable_broadcast_mode(adev, true);
drivers/gpu/drm/amd/amdgpu/df_v3_6.c
218
adev->df.hash_status.hash_64k = false;
drivers/gpu/drm/amd/amdgpu/df_v3_6.c
219
adev->df.hash_status.hash_2m = false;
drivers/gpu/drm/amd/amdgpu/df_v3_6.c
220
adev->df.hash_status.hash_1g = false;
drivers/gpu/drm/amd/amdgpu/df_v3_6.c
224
adev->df.funcs->get_fb_channel_number(adev) == 0xe) ||
drivers/gpu/drm/amd/amdgpu/df_v3_6.c
226
adev->df.funcs->get_fb_channel_number(adev) == 0x1e)) {
drivers/gpu/drm/amd/amdgpu/df_v3_6.c
228
adev->df.hash_status.hash_64k = REG_GET_FIELD(tmp,
drivers/gpu/drm/amd/amdgpu/df_v3_6.c
231
adev->df.hash_status.hash_2m = REG_GET_FIELD(tmp,
drivers/gpu/drm/amd/amdgpu/df_v3_6.c
234
adev->df.hash_status.hash_1g = REG_GET_FIELD(tmp,
drivers/gpu/drm/amd/amdgpu/df_v3_6.c
297
fb_channel_number = adev->df.funcs->get_fb_channel_number(adev);
drivers/gpu/drm/amd/amdgpu/df_v3_6.c
311
adev->df.funcs->enable_broadcast_mode(adev, true);
drivers/gpu/drm/amd/amdgpu/df_v3_6.c
330
adev->df.funcs->enable_broadcast_mode(adev, false);
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
4007
adev->df.hash_status.hash_64k);
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
4009
adev->df.hash_status.hash_2m);
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
4011
adev->df.hash_status.hash_1g);
drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c
1623
if (adev->df.funcs &&
drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c
1624
adev->df.funcs->enable_ecc_force_par_wr_rmw)
drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c
1625
adev->df.funcs->enable_ecc_force_par_wr_rmw(adev, false);
drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c
1876
if (adev->df.funcs &&
drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c
1877
adev->df.funcs->get_hbm_channel_number) {
drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c
1878
numchan = adev->df.funcs->get_hbm_channel_number(adev);
drivers/gpu/drm/amd/amdgpu/soc15.c
1253
if (adev->df.funcs &&
drivers/gpu/drm/amd/amdgpu/soc15.c
1254
adev->df.funcs->sw_init)
drivers/gpu/drm/amd/amdgpu/soc15.c
1255
adev->df.funcs->sw_init(adev);
drivers/gpu/drm/amd/amdgpu/soc15.c
1264
if (adev->df.funcs &&
drivers/gpu/drm/amd/amdgpu/soc15.c
1265
adev->df.funcs->sw_fini)
drivers/gpu/drm/amd/amdgpu/soc15.c
1266
adev->df.funcs->sw_fini(adev);
drivers/gpu/drm/amd/amdgpu/soc15.c
1436
adev->df.funcs->update_medium_grain_clock_gating(adev,
drivers/gpu/drm/amd/amdgpu/soc15.c
1498
if (adev->df.funcs && adev->df.funcs->get_clockgating_state)
drivers/gpu/drm/amd/amdgpu/soc15.c
1499
adev->df.funcs->get_clockgating_state(adev, flags);
drivers/gpu/drm/amd/amdgpu/soc24.c
487
if (adev->df.funcs->hw_init)
drivers/gpu/drm/amd/amdgpu/soc24.c
488
adev->df.funcs->hw_init(adev);
drivers/gpu/drm/amd/amdgpu/umc_v6_7.h
61
(((pa) >> 20) & 0x1ULL & adev->df.hash_status.hash_64k) ^ \
drivers/gpu/drm/amd/amdgpu/umc_v6_7.h
62
(((pa) >> 25) & 0x1ULL & adev->df.hash_status.hash_2m) ^ \
drivers/gpu/drm/amd/amdgpu/umc_v6_7.h
63
(((pa) >> 34) & 0x1ULL & adev->df.hash_status.hash_1g))
drivers/gpu/drm/msm/msm_gpu_devfreq.c
100
busy_time = busy_cycles - df->busy_cycles;
drivers/gpu/drm/msm/msm_gpu_devfreq.c
101
df->busy_cycles = busy_cycles;
drivers/gpu/drm/msm/msm_gpu_devfreq.c
103
mutex_unlock(&df->lock);
drivers/gpu/drm/msm/msm_gpu_devfreq.c
135
struct msm_gpu_devfreq *df = &gpu->devfreq;
drivers/gpu/drm/msm/msm_gpu_devfreq.c
136
return !!df->devfreq;
drivers/gpu/drm/msm/msm_gpu_devfreq.c
141
struct msm_gpu_devfreq *df = &gpu->devfreq;
drivers/gpu/drm/msm/msm_gpu_devfreq.c
158
mutex_init(&df->lock);
drivers/gpu/drm/msm/msm_gpu_devfreq.c
159
df->suspended = true;
drivers/gpu/drm/msm/msm_gpu_devfreq.c
161
ret = dev_pm_qos_add_request(&gpu->pdev->dev, &df->boost_freq,
drivers/gpu/drm/msm/msm_gpu_devfreq.c
179
df->devfreq = devm_devfreq_add_device(&gpu->pdev->dev,
drivers/gpu/drm/msm/msm_gpu_devfreq.c
183
if (IS_ERR(df->devfreq)) {
drivers/gpu/drm/msm/msm_gpu_devfreq.c
185
dev_pm_qos_remove_request(&df->boost_freq);
drivers/gpu/drm/msm/msm_gpu_devfreq.c
186
df->devfreq = NULL;
drivers/gpu/drm/msm/msm_gpu_devfreq.c
190
devfreq_suspend_device(df->devfreq);
drivers/gpu/drm/msm/msm_gpu_devfreq.c
192
gpu->cooling = of_devfreq_cooling_register(gpu->pdev->dev.of_node, df->devfreq);
drivers/gpu/drm/msm/msm_gpu_devfreq.c
199
msm_hrtimer_work_init(&df->boost_work, gpu->worker, msm_devfreq_boost_work,
drivers/gpu/drm/msm/msm_gpu_devfreq.c
201
msm_hrtimer_work_init(&df->idle_work, gpu->worker, msm_devfreq_idle_work,
drivers/gpu/drm/msm/msm_gpu_devfreq.c
205
static void cancel_idle_work(struct msm_gpu_devfreq *df)
drivers/gpu/drm/msm/msm_gpu_devfreq.c
207
hrtimer_cancel(&df->idle_work.timer);
drivers/gpu/drm/msm/msm_gpu_devfreq.c
208
kthread_cancel_work_sync(&df->idle_work.work);
drivers/gpu/drm/msm/msm_gpu_devfreq.c
211
static void cancel_boost_work(struct msm_gpu_devfreq *df)
drivers/gpu/drm/msm/msm_gpu_devfreq.c
213
hrtimer_cancel(&df->boost_work.timer);
drivers/gpu/drm/msm/msm_gpu_devfreq.c
214
kthread_cancel_work_sync(&df->boost_work.work);
drivers/gpu/drm/msm/msm_gpu_devfreq.c
219
struct msm_gpu_devfreq *df = &gpu->devfreq;
drivers/gpu/drm/msm/msm_gpu_devfreq.c
225
dev_pm_qos_remove_request(&df->boost_freq);
drivers/gpu/drm/msm/msm_gpu_devfreq.c
23
struct msm_gpu_devfreq *df = &gpu->devfreq;
drivers/gpu/drm/msm/msm_gpu_devfreq.c
230
struct msm_gpu_devfreq *df = &gpu->devfreq;
drivers/gpu/drm/msm/msm_gpu_devfreq.c
236
mutex_lock(&df->lock);
drivers/gpu/drm/msm/msm_gpu_devfreq.c
237
df->busy_cycles = gpu->funcs->gpu_busy(gpu, &sample_rate);
drivers/gpu/drm/msm/msm_gpu_devfreq.c
238
df->time = ktime_get();
drivers/gpu/drm/msm/msm_gpu_devfreq.c
239
df->suspended = false;
drivers/gpu/drm/msm/msm_gpu_devfreq.c
240
mutex_unlock(&df->lock);
drivers/gpu/drm/msm/msm_gpu_devfreq.c
242
devfreq_resume_device(df->devfreq);
drivers/gpu/drm/msm/msm_gpu_devfreq.c
247
struct msm_gpu_devfreq *df = &gpu->devfreq;
drivers/gpu/drm/msm/msm_gpu_devfreq.c
252
mutex_lock(&df->lock);
drivers/gpu/drm/msm/msm_gpu_devfreq.c
253
df->suspended = true;
drivers/gpu/drm/msm/msm_gpu_devfreq.c
254
mutex_unlock(&df->lock);
drivers/gpu/drm/msm/msm_gpu_devfreq.c
256
devfreq_suspend_device(df->devfreq);
drivers/gpu/drm/msm/msm_gpu_devfreq.c
258
cancel_idle_work(df);
drivers/gpu/drm/msm/msm_gpu_devfreq.c
259
cancel_boost_work(df);
drivers/gpu/drm/msm/msm_gpu_devfreq.c
264
struct msm_gpu_devfreq *df = container_of(work,
drivers/gpu/drm/msm/msm_gpu_devfreq.c
267
dev_pm_qos_update_request(&df->boost_freq, 0);
drivers/gpu/drm/msm/msm_gpu_devfreq.c
272
struct msm_gpu_devfreq *df = &gpu->devfreq;
drivers/gpu/drm/msm/msm_gpu_devfreq.c
287
dev_pm_qos_update_request(&df->boost_freq, freq);
drivers/gpu/drm/msm/msm_gpu_devfreq.c
289
msm_hrtimer_queue_work(&df->boost_work,
drivers/gpu/drm/msm/msm_gpu_devfreq.c
296
struct msm_gpu_devfreq *df = &gpu->devfreq;
drivers/gpu/drm/msm/msm_gpu_devfreq.c
306
cancel_idle_work(df);
drivers/gpu/drm/msm/msm_gpu_devfreq.c
312
mutex_lock(&df->devfreq->lock);
drivers/gpu/drm/msm/msm_gpu_devfreq.c
314
target_freq = df->idle_freq;
drivers/gpu/drm/msm/msm_gpu_devfreq.c
316
idle_time = ktime_to_ms(ktime_sub(ktime_get(), df->idle_time));
drivers/gpu/drm/msm/msm_gpu_devfreq.c
318
df->idle_freq = 0;
drivers/gpu/drm/msm/msm_gpu_devfreq.c
328
mutex_unlock(&df->devfreq->lock);
drivers/gpu/drm/msm/msm_gpu_devfreq.c
343
struct msm_gpu_devfreq *df = container_of(work,
drivers/gpu/drm/msm/msm_gpu_devfreq.c
345
struct msm_gpu *gpu = container_of(df, struct msm_gpu, devfreq);
drivers/gpu/drm/msm/msm_gpu_devfreq.c
353
mutex_lock(&df->devfreq->lock);
drivers/gpu/drm/msm/msm_gpu_devfreq.c
360
df->idle_time = ktime_get();
drivers/gpu/drm/msm/msm_gpu_devfreq.c
361
df->idle_freq = idle_freq;
drivers/gpu/drm/msm/msm_gpu_devfreq.c
363
mutex_unlock(&df->devfreq->lock);
drivers/gpu/drm/msm/msm_gpu_devfreq.c
368
struct msm_gpu_devfreq *df = &gpu->devfreq;
drivers/gpu/drm/msm/msm_gpu_devfreq.c
373
msm_hrtimer_queue_work(&df->idle_work, ms_to_ktime(1),
drivers/gpu/drm/msm/msm_gpu_devfreq.c
40
if (df->idle_freq) {
drivers/gpu/drm/msm/msm_gpu_devfreq.c
41
df->idle_freq = *freq;
drivers/gpu/drm/msm/msm_gpu_devfreq.c
47
mutex_lock(&df->lock);
drivers/gpu/drm/msm/msm_gpu_devfreq.c
48
gpu->funcs->gpu_set_freq(gpu, opp, df->suspended);
drivers/gpu/drm/msm/msm_gpu_devfreq.c
49
mutex_unlock(&df->lock);
drivers/gpu/drm/msm/msm_gpu_devfreq.c
61
struct msm_gpu_devfreq *df = &gpu->devfreq;
drivers/gpu/drm/msm/msm_gpu_devfreq.c
68
if (df->idle_freq)
drivers/gpu/drm/msm/msm_gpu_devfreq.c
69
return df->idle_freq;
drivers/gpu/drm/msm/msm_gpu_devfreq.c
81
struct msm_gpu_devfreq *df = &gpu->devfreq;
drivers/gpu/drm/msm/msm_gpu_devfreq.c
86
mutex_lock(&df->lock);
drivers/gpu/drm/msm/msm_gpu_devfreq.c
90
status->total_time = ktime_us_delta(time, df->time);
drivers/gpu/drm/msm/msm_gpu_devfreq.c
91
df->time = time;
drivers/gpu/drm/msm/msm_gpu_devfreq.c
93
if (df->suspended) {
drivers/gpu/drm/msm/msm_gpu_devfreq.c
94
mutex_unlock(&df->lock);
drivers/infiniband/sw/rxe/rxe_net.c
309
__u8 tos, __u8 ttl, __be16 df, bool xnet)
drivers/infiniband/sw/rxe/rxe_net.c
327
iph->frag_off = df;
drivers/infiniband/sw/rxe/rxe_net.c
366
__be16 df = htons(IP_DF);
drivers/infiniband/sw/rxe/rxe_net.c
380
av->grh.traffic_class, av->grh.hop_limit, df, xnet);
drivers/ipack/devices/scc2698.h
63
u8 df, cts; /* Stop counter timer of block */
drivers/ipack/devices/scc2698.h
81
u8 df, r3; /* reserved */
drivers/isdn/hardware/mISDN/hfcpci.c
461
struct dfifo *df;
drivers/isdn/hardware/mISDN/hfcpci.c
464
df = &((union fifo_area *)(hc->hw.fifos))->d_chan.d_rx;
drivers/isdn/hardware/mISDN/hfcpci.c
465
while (((df->f1 & D_FREG_MASK) != (df->f2 & D_FREG_MASK)) && count--) {
drivers/isdn/hardware/mISDN/hfcpci.c
466
zp = &df->za[df->f2 & D_FREG_MASK];
drivers/isdn/hardware/mISDN/hfcpci.c
474
df->f1, df->f2,
drivers/isdn/hardware/mISDN/hfcpci.c
480
(df->data[le16_to_cpu(zp->z1)])) {
drivers/isdn/hardware/mISDN/hfcpci.c
486
df->data[le16_to_cpu(zp->z1)]);
drivers/isdn/hardware/mISDN/hfcpci.c
490
df->f2 = ((df->f2 + 1) & MAX_D_FRAMES) |
drivers/isdn/hardware/mISDN/hfcpci.c
492
df->za[df->f2 & D_FREG_MASK].z2 =
drivers/isdn/hardware/mISDN/hfcpci.c
512
ptr1 = df->data + le16_to_cpu(zp->z2);
drivers/isdn/hardware/mISDN/hfcpci.c
519
ptr1 = df->data; /* start of buffer */
drivers/isdn/hardware/mISDN/hfcpci.c
522
df->f2 = ((df->f2 + 1) & MAX_D_FRAMES) |
drivers/isdn/hardware/mISDN/hfcpci.c
524
df->za[df->f2 & D_FREG_MASK].z2 = cpu_to_le16((
drivers/isdn/hardware/mISDN/hfcpci.c
669
struct dfifo *df;
drivers/isdn/hardware/mISDN/hfcpci.c
680
df = &((union fifo_area *) (hc->hw.fifos))->d_chan.d_tx;
drivers/isdn/hardware/mISDN/hfcpci.c
684
df->f1, df->f2,
drivers/isdn/hardware/mISDN/hfcpci.c
685
le16_to_cpu(df->za[df->f1 & D_FREG_MASK].z1));
drivers/isdn/hardware/mISDN/hfcpci.c
686
fcnt = df->f1 - df->f2; /* frame count actually buffered */
drivers/isdn/hardware/mISDN/hfcpci.c
699
maxlen = le16_to_cpu(df->za[df->f2 & D_FREG_MASK].z2) -
drivers/isdn/hardware/mISDN/hfcpci.c
700
le16_to_cpu(df->za[df->f1 & D_FREG_MASK].z1) - 1;
drivers/isdn/hardware/mISDN/hfcpci.c
712
new_z1 = (le16_to_cpu(df->za[df->f1 & D_FREG_MASK].z1) + count) &
drivers/isdn/hardware/mISDN/hfcpci.c
714
new_f1 = ((df->f1 + 1) & D_FREG_MASK) | (D_FREG_MASK + 1);
drivers/isdn/hardware/mISDN/hfcpci.c
716
dst = df->data + le16_to_cpu(df->za[df->f1 & D_FREG_MASK].z1);
drivers/isdn/hardware/mISDN/hfcpci.c
717
maxlen = D_FIFO_SIZE - le16_to_cpu(df->za[df->f1 & D_FREG_MASK].z1);
drivers/isdn/hardware/mISDN/hfcpci.c
725
dst = df->data; /* start of buffer */
drivers/isdn/hardware/mISDN/hfcpci.c
729
df->za[new_f1 & D_FREG_MASK].z1 = cpu_to_le16(new_z1);
drivers/isdn/hardware/mISDN/hfcpci.c
731
df->za[df->f1 & D_FREG_MASK].z1 = cpu_to_le16(new_z1);
drivers/isdn/hardware/mISDN/hfcpci.c
733
df->f1 = new_f1; /* next frame */
drivers/media/platform/samsung/exynos4-is/fimc-m2m.c
106
df = &ctx->d_frame;
drivers/media/platform/samsung/exynos4-is/fimc-m2m.c
111
fimc_prepare_dma_offset(ctx, df);
drivers/media/platform/samsung/exynos4-is/fimc-m2m.c
120
ret = fimc_prepare_addr(ctx, &dst_vb->vb2_buf, df, &df->addr);
drivers/media/platform/samsung/exynos4-is/fimc-m2m.c
153
fimc_hw_set_output_addr(fimc, &df->addr, -1);
drivers/media/platform/samsung/exynos4-is/fimc-m2m.c
93
struct fimc_frame *sf, *df;
drivers/memory/samsung/exynos5422-dmc.c
1391
mutex_lock(&dmc->df->lock);
drivers/memory/samsung/exynos5422-dmc.c
1393
res = update_devfreq(dmc->df);
drivers/memory/samsung/exynos5422-dmc.c
1394
mutex_unlock(&dmc->df->lock);
drivers/memory/samsung/exynos5422-dmc.c
1517
dmc->df = devm_devfreq_add_device(dev, &exynos5_dmc_df_profile,
drivers/memory/samsung/exynos5422-dmc.c
1521
if (IS_ERR(dmc->df)) {
drivers/memory/samsung/exynos5422-dmc.c
1522
ret = PTR_ERR(dmc->df);
drivers/memory/samsung/exynos5422-dmc.c
155
struct devfreq *df;
drivers/misc/apds990x.c
1082
chip->cf.df = 52;
drivers/misc/apds990x.c
278
APDS_PARAM_SCALE * 64) / (chip->cf.ga * chip->cf.df);
drivers/misc/apds990x.c
446
lpc = APDS990X_LUX_OUTPUT_SCALE * (chip->cf.df * chip->cf.ga) /
drivers/mmc/core/mmc_test.c
3166
struct mmc_test_dbgfs_file *df, *dfs;
drivers/mmc/core/mmc_test.c
3170
list_for_each_entry_safe(df, dfs, &mmc_test_file_test, link) {
drivers/mmc/core/mmc_test.c
3171
if (card && df->card != card)
drivers/mmc/core/mmc_test.c
3173
debugfs_remove(df->file);
drivers/mmc/core/mmc_test.c
3174
list_del(&df->link);
drivers/mmc/core/mmc_test.c
3175
kfree(df);
drivers/mmc/core/mmc_test.c
3185
struct mmc_test_dbgfs_file *df;
drivers/mmc/core/mmc_test.c
3191
df = kmalloc_obj(*df);
drivers/mmc/core/mmc_test.c
3192
if (!df) {
drivers/mmc/core/mmc_test.c
3197
df->card = card;
drivers/mmc/core/mmc_test.c
3198
df->file = file;
drivers/mmc/core/mmc_test.c
3200
list_add(&df->link, &mmc_test_file_test);
drivers/net/bareudp.c
314
__be16 sport, df;
drivers/net/bareudp.c
342
df = test_bit(IP_TUNNEL_DONT_FRAGMENT_BIT, key->tun_flags) ?
drivers/net/bareudp.c
363
tos, ttl, df, sport, bareudp->port,
drivers/net/can/rcar/rcar_canfd.c
330
#define RCANFD_C_RFDF(x, df) \
drivers/net/can/rcar/rcar_canfd.c
331
(RCANFD_C_RFOFFSET + 0x08 + (0x10 * (x)) + (0x04 * (df)))
drivers/net/can/rcar/rcar_canfd.c
342
#define RCANFD_C_CFDF(ch, idx, df) \
drivers/net/can/rcar/rcar_canfd.c
343
(RCANFD_C_CFOFFSET + 0x08 + (0x30 * (ch)) + (0x10 * (idx)) + (0x04 * (df)))
drivers/net/can/rcar/rcar_canfd.c
368
#define RCANFD_F_RFDF(gpriv, x, df) \
drivers/net/can/rcar/rcar_canfd.c
369
(RCANFD_F_RFOFFSET(gpriv) + 0x0c + (0x80 * (x)) + (0x04 * (df)))
drivers/net/can/rcar/rcar_canfd.c
383
#define RCANFD_F_CFDF(gpriv, ch, idx, df) \
drivers/net/can/rcar/rcar_canfd.c
385
(0x04 * (df)))
drivers/net/ethernet/amazon/ena/ena_eth_com.c
467
desc->meta_ctrl |= (ena_tx_ctx->df <<
drivers/net/ethernet/amazon/ena/ena_eth_com.h
36
u8 df; /* Don't fragment */
drivers/net/ethernet/amazon/ena/ena_netdev.c
2442
ena_tx_ctx->df = 1;
drivers/net/ethernet/huawei/hinic3/hinic3_cmdq.c
176
enum cmdq_data_format df;
drivers/net/ethernet/huawei/hinic3/hinic3_cmdq.c
179
df = CMDQ_WQE_HDR_GET(header_info, DATA_FMT);
drivers/net/ethernet/huawei/hinic3/hinic3_cmdq.c
180
if (df == CMDQ_DATA_SGE)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_struct.h
182
u64 df : 1;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
1472
sqe_hdr->df = 1;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
809
sqe_hdr->df = 1;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_xsk.c
214
sqe_hdr->df = 1;
drivers/net/geneve.c
1300
__be16 df = 0;
drivers/net/geneve.c
1364
df = test_bit(IP_TUNNEL_DONT_FRAGMENT_BIT, key->tun_flags) ?
drivers/net/geneve.c
1373
if (geneve->cfg.df == GENEVE_DF_SET) {
drivers/net/geneve.c
1374
df = htons(IP_DF);
drivers/net/geneve.c
1375
} else if (geneve->cfg.df == GENEVE_DF_INHERIT) {
drivers/net/geneve.c
1379
df = htons(IP_DF);
drivers/net/geneve.c
1384
df = htons(IP_DF);
drivers/net/geneve.c
1395
tos, ttl, df, sport, geneve->cfg.info.key.tp_dst,
drivers/net/geneve.c
1763
enum ifla_geneve_df df = nla_get_u8(data[IFLA_GENEVE_DF]);
drivers/net/geneve.c
1765
if (df < 0 || df > GENEVE_DF_MAX) {
drivers/net/geneve.c
1990
cfg->df = nla_get_u8(data[IFLA_GENEVE_DF]);
drivers/net/geneve.c
2146
.df = GENEVE_DF_UNSET,
drivers/net/geneve.c
2319
if (nla_put_u8(skb, IFLA_GENEVE_DF, geneve->cfg.df))
drivers/net/geneve.c
2376
.df = GENEVE_DF_UNSET,
drivers/net/geneve.c
80
enum ifla_geneve_df df;
drivers/net/gtp.c
1054
__be16 df;
drivers/net/gtp.c
1074
df = frag_off;
drivers/net/gtp.c
1075
if (df) {
drivers/net/vxlan/vxlan_core.c
2469
__be16 df = 0;
drivers/net/vxlan/vxlan_core.c
2501
if (vxlan->cfg.df == VXLAN_DF_SET) {
drivers/net/vxlan/vxlan_core.c
2502
df = htons(IP_DF);
drivers/net/vxlan/vxlan_core.c
2503
} else if (vxlan->cfg.df == VXLAN_DF_INHERIT) {
drivers/net/vxlan/vxlan_core.c
2509
df = htons(IP_DF);
drivers/net/vxlan/vxlan_core.c
2513
df = htons(IP_DF);
drivers/net/vxlan/vxlan_core.c
2547
pkey->u.ipv4.dst, tos, ttl, df,
drivers/net/vxlan/vxlan_core.c
3502
enum ifla_vxlan_df df = nla_get_u8(data[IFLA_VXLAN_DF]);
drivers/net/vxlan/vxlan_core.c
3504
if (df < 0 || df > VXLAN_DF_MAX) {
drivers/net/vxlan/vxlan_core.c
4376
conf->df = nla_get_u8(data[IFLA_VXLAN_DF]);
drivers/net/vxlan/vxlan_core.c
4606
nla_put_u8(skb, IFLA_VXLAN_DF, vxlan->cfg.df) ||
drivers/pinctrl/pxa/pinctrl-pxa2xx.c
144
struct pxa_desc_function *df;
drivers/pinctrl/pxa/pinctrl-pxa2xx.c
150
df = pxa_desc_by_func_group(pctl, g->name, (pctl->functions + function)->name);
drivers/pinctrl/pxa/pinctrl-pxa2xx.c
151
if (!df)
drivers/pinctrl/pxa/pinctrl-pxa2xx.c
158
pin, df->muxval >> 1, df->muxval & 0x1);
drivers/pinctrl/pxa/pinctrl-pxa2xx.c
163
val = (val & ~(0x3 << shift)) | ((df->muxval >> 1) << shift);
drivers/pinctrl/pxa/pinctrl-pxa2xx.c
167
val = (val & ~BIT(pin % 32)) | ((df->muxval & 1) ? BIT(pin % 32) : 0);
drivers/pinctrl/pxa/pinctrl-pxa2xx.c
267
struct pxa_desc_function *df;
drivers/pinctrl/pxa/pinctrl-pxa2xx.c
281
for (df = pctl->ppins[i].functions; df->name; df++)
drivers/pinctrl/pxa/pinctrl-pxa2xx.c
282
if (!pxa2xx_find_function(pctl, df->name, functions))
drivers/pinctrl/pxa/pinctrl-pxa2xx.c
283
(functions + pctl->nfuncs++)->name = df->name;
drivers/pinctrl/pxa/pinctrl-pxa2xx.c
296
struct pxa_desc_function *df;
drivers/pinctrl/pxa/pinctrl-pxa2xx.c
308
for (df = pctl->ppins[j].functions; df->name;
drivers/pinctrl/pxa/pinctrl-pxa2xx.c
309
df++)
drivers/pinctrl/pxa/pinctrl-pxa2xx.c
311
df->name))
drivers/pinctrl/pxa/pinctrl-pxa2xx.c
69
struct pxa_desc_function *df;
drivers/pinctrl/pxa/pinctrl-pxa2xx.c
75
for (df = pin->functions; df->name; df++)
drivers/pinctrl/pxa/pinctrl-pxa2xx.c
76
if (!strcmp(df->name, func_name))
drivers/pinctrl/pxa/pinctrl-pxa2xx.c
77
return df;
drivers/regulator/ad5398.c
222
const struct ad5398_current_data_format *df =
drivers/regulator/ad5398.c
243
chip->min_uA = df->min_uA;
drivers/regulator/ad5398.c
244
chip->max_uA = df->max_uA;
drivers/regulator/ad5398.c
245
chip->current_level = 1 << df->current_bits;
drivers/regulator/ad5398.c
246
chip->current_offset = df->current_offset;
drivers/staging/rtl8723bs/include/drv_types.h
394
int df = atomic_read(&adapter_to_dvobj(padapter)->disable_func);
drivers/staging/rtl8723bs/include/drv_types.h
395
df &= ~(func_bit);
drivers/staging/rtl8723bs/include/drv_types.h
396
atomic_set(&adapter_to_dvobj(padapter)->disable_func, df);
drivers/thermal/devfreq_cooling.c
148
static unsigned long get_voltage(struct devfreq *df, unsigned long freq)
drivers/thermal/devfreq_cooling.c
150
struct device *dev = df->dev.parent;
drivers/thermal/devfreq_cooling.c
194
struct devfreq *df = dfc->devfreq;
drivers/thermal/devfreq_cooling.c
202
mutex_lock(&df->lock);
drivers/thermal/devfreq_cooling.c
203
status = df->last_status;
drivers/thermal/devfreq_cooling.c
204
mutex_unlock(&df->lock);
drivers/thermal/devfreq_cooling.c
209
voltage = get_voltage(df, freq);
drivers/thermal/devfreq_cooling.c
215
res = dfc->power_ops->get_real_power(df, power, freq, voltage);
drivers/thermal/devfreq_cooling.c
291
struct devfreq *df = dfc->devfreq;
drivers/thermal/devfreq_cooling.c
298
mutex_lock(&df->lock);
drivers/thermal/devfreq_cooling.c
299
status = df->last_status;
drivers/thermal/devfreq_cooling.c
300
mutex_unlock(&df->lock);
drivers/thermal/devfreq_cooling.c
351
struct devfreq *df = dfc->devfreq;
drivers/thermal/devfreq_cooling.c
352
struct device *dev = df->dev.parent;
drivers/thermal/devfreq_cooling.c
393
of_devfreq_cooling_register_power(struct device_node *np, struct devfreq *df,
drivers/thermal/devfreq_cooling.c
397
struct device *dev = df->dev.parent;
drivers/thermal/devfreq_cooling.c
409
dfc->devfreq = df;
drivers/thermal/devfreq_cooling.c
490
of_devfreq_cooling_register(struct device_node *np, struct devfreq *df)
drivers/thermal/devfreq_cooling.c
492
return of_devfreq_cooling_register_power(np, df, NULL);
drivers/thermal/devfreq_cooling.c
500
struct thermal_cooling_device *devfreq_cooling_register(struct devfreq *df)
drivers/thermal/devfreq_cooling.c
502
return of_devfreq_cooling_register(NULL, df);
drivers/thermal/devfreq_cooling.c
523
devfreq_cooling_em_register(struct devfreq *df,
drivers/thermal/devfreq_cooling.c
530
if (IS_ERR_OR_NULL(df))
drivers/thermal/devfreq_cooling.c
533
dev = df->dev.parent;
drivers/thermal/devfreq_cooling.c
540
cdev = of_devfreq_cooling_register_power(dev->of_node, df, dfc_power);
drivers/thermal/devfreq_cooling.c
88
struct devfreq *df = dfc->devfreq;
drivers/thermal/devfreq_cooling.c
89
struct device *dev = df->dev.parent;
drivers/vfio/device_cdev.c
110
if (df->group)
drivers/vfio/device_cdev.c
119
if (df->access_granted) {
drivers/vfio/device_cdev.c
128
df->iommufd = iommufd_ctx_from_fd(bind.iommufd);
drivers/vfio/device_cdev.c
129
if (IS_ERR(df->iommufd)) {
drivers/vfio/device_cdev.c
130
ret = PTR_ERR(df->iommufd);
drivers/vfio/device_cdev.c
131
df->iommufd = NULL;
drivers/vfio/device_cdev.c
141
vfio_df_get_kvm_safe(df);
drivers/vfio/device_cdev.c
143
ret = vfio_df_open(df);
drivers/vfio/device_cdev.c
147
ret = copy_to_user(&arg->out_devid, &df->devid,
drivers/vfio/device_cdev.c
148
sizeof(df->devid)) ? -EFAULT : 0;
drivers/vfio/device_cdev.c
157
smp_store_release(&df->access_granted, true);
drivers/vfio/device_cdev.c
162
vfio_df_close(df);
drivers/vfio/device_cdev.c
165
iommufd_ctx_put(df->iommufd);
drivers/vfio/device_cdev.c
166
df->iommufd = NULL;
drivers/vfio/device_cdev.c
173
void vfio_df_unbind_iommufd(struct vfio_device_file *df)
drivers/vfio/device_cdev.c
175
struct vfio_device *device = df->device;
drivers/vfio/device_cdev.c
182
if (!df->access_granted)
drivers/vfio/device_cdev.c
186
vfio_df_close(df);
drivers/vfio/device_cdev.c
188
iommufd_ctx_put(df->iommufd);
drivers/vfio/device_cdev.c
194
int vfio_df_ioctl_attach_pt(struct vfio_device_file *df,
drivers/vfio/device_cdev.c
198
struct vfio_device *device = df->device;
drivers/vfio/device_cdev.c
253
int vfio_df_ioctl_detach_pt(struct vfio_device_file *df,
drivers/vfio/device_cdev.c
257
struct vfio_device *device = df->device;
drivers/vfio/device_cdev.c
27
struct vfio_device_file *df;
drivers/vfio/device_cdev.c
34
df = vfio_allocate_device_file(device);
drivers/vfio/device_cdev.c
35
if (IS_ERR(df)) {
drivers/vfio/device_cdev.c
36
ret = PTR_ERR(df);
drivers/vfio/device_cdev.c
40
filep->private_data = df;
drivers/vfio/device_cdev.c
56
static void vfio_df_get_kvm_safe(struct vfio_device_file *df)
drivers/vfio/device_cdev.c
58
spin_lock(&df->kvm_ref_lock);
drivers/vfio/device_cdev.c
59
vfio_device_get_kvm_safe(df->device, df->kvm);
drivers/vfio/device_cdev.c
60
spin_unlock(&df->kvm_ref_lock);
drivers/vfio/device_cdev.c
83
long vfio_df_ioctl_bind_iommufd(struct vfio_device_file *df,
drivers/vfio/device_cdev.c
87
struct vfio_device *device = df->device;
drivers/vfio/device_cdev.c
93
static_assert(__same_type(arg->out_devid, df->devid));
drivers/vfio/group.c
165
static int vfio_df_group_open(struct vfio_device_file *df)
drivers/vfio/group.c
167
struct vfio_device *device = df->device;
drivers/vfio/group.c
187
df->iommufd = device->group->iommufd;
drivers/vfio/group.c
188
if (df->iommufd && vfio_device_is_noiommu(device) && device->open_count == 0) {
drivers/vfio/group.c
195
vfio_iommufd_device_has_compat_ioas(device, df->iommufd)) {
drivers/vfio/group.c
201
ret = vfio_df_open(df);
drivers/vfio/group.c
205
if (df->iommufd && device->open_count == 1) {
drivers/vfio/group.c
206
ret = vfio_iommufd_compat_attach_ioas(device, df->iommufd);
drivers/vfio/group.c
215
smp_store_release(&df->access_granted, true);
drivers/vfio/group.c
222
vfio_df_close(df);
drivers/vfio/group.c
224
df->iommufd = NULL;
drivers/vfio/group.c
233
void vfio_df_group_close(struct vfio_device_file *df)
drivers/vfio/group.c
235
struct vfio_device *device = df->device;
drivers/vfio/group.c
240
vfio_df_close(df);
drivers/vfio/group.c
241
df->iommufd = NULL;
drivers/vfio/group.c
252
struct vfio_device_file *df;
drivers/vfio/group.c
256
df = vfio_allocate_device_file(device);
drivers/vfio/group.c
257
if (IS_ERR(df)) {
drivers/vfio/group.c
258
ret = PTR_ERR(df);
drivers/vfio/group.c
262
df->group = device->group;
drivers/vfio/group.c
264
ret = vfio_df_group_open(df);
drivers/vfio/group.c
269
df, O_RDWR, FMODE_PREAD | FMODE_PWRITE);
drivers/vfio/group.c
291
vfio_df_group_close(df);
drivers/vfio/group.c
293
kfree(df);
drivers/vfio/iommufd.c
21
int vfio_df_iommufd_bind(struct vfio_device_file *df)
drivers/vfio/iommufd.c
23
struct vfio_device *vdev = df->device;
drivers/vfio/iommufd.c
24
struct iommufd_ctx *ictx = df->iommufd;
drivers/vfio/iommufd.c
32
return vdev->ops->bind_iommufd(vdev, ictx, &df->devid);
drivers/vfio/iommufd.c
55
void vfio_df_iommufd_unbind(struct vfio_device_file *df)
drivers/vfio/iommufd.c
57
struct vfio_device *vdev = df->device;
drivers/vfio/vfio.h
108
void vfio_df_group_close(struct vfio_device_file *df);
drivers/vfio/vfio.h
160
static inline void vfio_df_group_close(struct vfio_device_file *df)
drivers/vfio/vfio.h
32
int vfio_df_open(struct vfio_device_file *df);
drivers/vfio/vfio.h
323
int vfio_df_iommufd_bind(struct vfio_device_file *df);
drivers/vfio/vfio.h
324
void vfio_df_iommufd_unbind(struct vfio_device_file *df);
drivers/vfio/vfio.h
33
void vfio_df_close(struct vfio_device_file *df);
drivers/vfio/vfio.h
340
static inline void vfio_df_iommufd_unbind(struct vfio_device_file *df)
drivers/vfio/vfio.h
352
int vfio_df_ioctl_attach_pt(struct vfio_device_file *df,
drivers/vfio/vfio.h
354
int vfio_df_ioctl_detach_pt(struct vfio_device_file *df,
drivers/vfio/vfio.h
378
long vfio_df_ioctl_bind_iommufd(struct vfio_device_file *df,
drivers/vfio/vfio.h
380
void vfio_df_unbind_iommufd(struct vfio_device_file *df);
drivers/vfio/vfio.h
404
static inline long vfio_df_ioctl_bind_iommufd(struct vfio_device_file *df,
drivers/vfio/vfio.h
410
static inline void vfio_df_unbind_iommufd(struct vfio_device_file *df)
drivers/vfio/vfio_main.c
1311
struct vfio_device_file *df = filep->private_data;
drivers/vfio/vfio_main.c
1312
struct vfio_device *device = df->device;
drivers/vfio/vfio_main.c
1317
return vfio_df_ioctl_bind_iommufd(df, uptr);
drivers/vfio/vfio_main.c
1320
if (!smp_load_acquire(&df->access_granted))
drivers/vfio/vfio_main.c
1328
if (IS_ENABLED(CONFIG_VFIO_DEVICE_CDEV) && !df->group) {
drivers/vfio/vfio_main.c
1331
ret = vfio_df_ioctl_attach_pt(df, uptr);
drivers/vfio/vfio_main.c
1335
ret = vfio_df_ioctl_detach_pt(df, uptr);
drivers/vfio/vfio_main.c
1364
struct vfio_device_file *df = filep->private_data;
drivers/vfio/vfio_main.c
1365
struct vfio_device *device = df->device;
drivers/vfio/vfio_main.c
1368
if (!smp_load_acquire(&df->access_granted))
drivers/vfio/vfio_main.c
1381
struct vfio_device_file *df = filep->private_data;
drivers/vfio/vfio_main.c
1382
struct vfio_device *device = df->device;
drivers/vfio/vfio_main.c
1385
if (!smp_load_acquire(&df->access_granted))
drivers/vfio/vfio_main.c
1396
struct vfio_device_file *df = filep->private_data;
drivers/vfio/vfio_main.c
1397
struct vfio_device *device = df->device;
drivers/vfio/vfio_main.c
1400
if (!smp_load_acquire(&df->access_granted))
drivers/vfio/vfio_main.c
1413
struct vfio_device_file *df = filep->private_data;
drivers/vfio/vfio_main.c
1414
struct vfio_device *device = df->device;
drivers/vfio/vfio_main.c
1441
struct vfio_device_file *df = file->private_data;
drivers/vfio/vfio_main.c
1445
return df->device;
drivers/vfio/vfio_main.c
1488
struct vfio_device_file *df = file->private_data;
drivers/vfio/vfio_main.c
1495
spin_lock(&df->kvm_ref_lock);
drivers/vfio/vfio_main.c
1496
df->kvm = kvm;
drivers/vfio/vfio_main.c
1497
spin_unlock(&df->kvm_ref_lock);
drivers/vfio/vfio_main.c
496
struct vfio_device_file *df;
drivers/vfio/vfio_main.c
498
df = kzalloc_obj(*df, GFP_KERNEL_ACCOUNT);
drivers/vfio/vfio_main.c
499
if (!df)
drivers/vfio/vfio_main.c
502
df->device = device;
drivers/vfio/vfio_main.c
503
spin_lock_init(&df->kvm_ref_lock);
drivers/vfio/vfio_main.c
505
return df;
drivers/vfio/vfio_main.c
508
static int vfio_df_device_first_open(struct vfio_device_file *df)
drivers/vfio/vfio_main.c
510
struct vfio_device *device = df->device;
drivers/vfio/vfio_main.c
511
struct iommufd_ctx *iommufd = df->iommufd;
drivers/vfio/vfio_main.c
520
ret = vfio_df_iommufd_bind(df);
drivers/vfio/vfio_main.c
535
vfio_df_iommufd_unbind(df);
drivers/vfio/vfio_main.c
543
static void vfio_df_device_last_close(struct vfio_device_file *df)
drivers/vfio/vfio_main.c
545
struct vfio_device *device = df->device;
drivers/vfio/vfio_main.c
546
struct iommufd_ctx *iommufd = df->iommufd;
drivers/vfio/vfio_main.c
553
vfio_df_iommufd_unbind(df);
drivers/vfio/vfio_main.c
559
int vfio_df_open(struct vfio_device_file *df)
drivers/vfio/vfio_main.c
561
struct vfio_device *device = df->device;
drivers/vfio/vfio_main.c
570
if (device->open_count != 0 && !df->group)
drivers/vfio/vfio_main.c
575
ret = vfio_df_device_first_open(df);
drivers/vfio/vfio_main.c
583
void vfio_df_close(struct vfio_device_file *df)
drivers/vfio/vfio_main.c
585
struct vfio_device *device = df->device;
drivers/vfio/vfio_main.c
592
vfio_df_device_last_close(df);
drivers/vfio/vfio_main.c
634
struct vfio_device_file *df = filep->private_data;
drivers/vfio/vfio_main.c
635
struct vfio_device *device = df->device;
drivers/vfio/vfio_main.c
637
if (df->group)
drivers/vfio/vfio_main.c
638
vfio_df_group_close(df);
drivers/vfio/vfio_main.c
640
vfio_df_unbind_iommufd(df);
drivers/vfio/vfio_main.c
644
kfree(df);
drivers/video/fbdev/aty/mach64_gx.c
165
u8 df, vco_div_count, ref_div_count;
drivers/video/fbdev/aty/mach64_gx.c
167
df = pll->ibm514.m >> 6;
drivers/video/fbdev/aty/mach64_gx.c
171
return ((par->ref_clk_per * ref_div_count) << (3 - df))/
drivers/video/fbdev/matrox/matroxfb_maven.c
341
unsigned char df;
drivers/video/fbdev/matrox/matroxfb_maven.c
343
df = (md->version == MGATVO_B?0x40:0x00);
drivers/video/fbdev/matrox/matroxfb_maven.c
349
df |= 0xB1;
drivers/video/fbdev/matrox/matroxfb_maven.c
352
df |= 0xA2;
drivers/video/fbdev/matrox/matroxfb_maven.c
355
return df;
include/asm-generic/vmlinux.lds.h
118
.text.[_0-9A-Za-df-rt-z]* \
include/linux/devfreq-governor.h
100
return df->profile->get_dev_status(df->dev.parent, &df->last_status);
include/linux/devfreq-governor.h
95
static inline int devfreq_update_stats(struct devfreq *df)
include/linux/devfreq-governor.h
97
if (!df->profile->get_dev_status)
include/linux/devfreq.h
453
static inline int devfreq_update_stats(struct devfreq *df)
include/linux/devfreq_cooling.h
38
int (*get_real_power)(struct devfreq *df, u32 *power,
include/linux/devfreq_cooling.h
45
of_devfreq_cooling_register_power(struct device_node *np, struct devfreq *df,
include/linux/devfreq_cooling.h
48
of_devfreq_cooling_register(struct device_node *np, struct devfreq *df);
include/linux/devfreq_cooling.h
49
struct thermal_cooling_device *devfreq_cooling_register(struct devfreq *df);
include/linux/devfreq_cooling.h
52
devfreq_cooling_em_register(struct devfreq *df,
include/linux/devfreq_cooling.h
58
of_devfreq_cooling_register_power(struct device_node *np, struct devfreq *df,
include/linux/devfreq_cooling.h
65
of_devfreq_cooling_register(struct device_node *np, struct devfreq *df)
include/linux/devfreq_cooling.h
71
devfreq_cooling_register(struct devfreq *df)
include/linux/devfreq_cooling.h
77
devfreq_cooling_em_register(struct devfreq *df,
include/linux/platform_data/apds990x.h
41
int df;
include/net/ip_tunnels.h
626
u8 tos, u8 ttl, __be16 df, bool xnet, u16 ipcb_flags);
include/net/udp_tunnel.h
141
__be16 df, __be16 src_port, __be16 dst_port,
include/net/vxlan.h
229
enum ifla_vxlan_df df;
include/uapi/linux/netfilter/nfnetlink_osf.h
57
__u8 ttl, df;
include/xen/interface/hvm/ioreq.h
46
uint8_t df:1;
kernel/auditfilter.c
788
static inline int audit_dupe_lsm_field(struct audit_field *df,
kernel/auditfilter.c
798
df->lsm_str = lsm_str;
kernel/auditfilter.c
801
ret = security_audit_rule_init(df->type, df->op, df->lsm_str,
kernel/auditfilter.c
802
(void **)&df->lsm_rule, GFP_KERNEL);
kernel/auditfilter.c
807
df->lsm_str);
kernel/bpf/helpers.c
1776
struct bpf_dynptr_file_impl *df = ptr->data;
kernel/bpf/helpers.c
1778
return df->size;
kernel/bpf/helpers.c
1787
struct bpf_dynptr_file_impl *df = ptr->data;
kernel/bpf/helpers.c
1789
df->offset += off;
kernel/bpf/helpers.c
1800
struct bpf_dynptr_file_impl *df = ptr->data;
kernel/bpf/helpers.c
1802
df->size = new_size;
kernel/bpf/helpers.c
1813
static int bpf_file_fetch_bytes(struct bpf_dynptr_file_impl *df, u64 offset, void *buf, u64 len)
kernel/bpf/helpers.c
1820
df->freader.buf = buf;
kernel/bpf/helpers.c
1821
df->freader.buf_sz = len;
kernel/bpf/helpers.c
1822
ptr = freader_fetch(&df->freader, offset + df->offset, len);
kernel/bpf/helpers.c
1824
return df->freader.err;
kernel/bpf/helpers.c
4446
struct bpf_dynptr_file_impl *df = ptr->data;
kernel/bpf/helpers.c
4448
if (!df)
kernel/bpf/helpers.c
4451
freader_cleanup(&df->freader);
kernel/bpf/helpers.c
4452
bpf_mem_free(&bpf_global_ma, df);
mm/slub.c
6106
struct defer_free *df = container_of(work, struct defer_free, work);
mm/slub.c
6107
struct llist_head *objs = &df->objects;
mm/slub.c
6139
struct defer_free *df;
mm/slub.c
6145
df = this_cpu_ptr(&defer_free_objects);
mm/slub.c
6146
if (llist_add(head + s->offset, &df->objects))
mm/slub.c
6147
irq_work_queue(&df->work);
mm/slub.c
6901
void **p, struct detached_freelist *df)
mm/slub.c
6916
df->slab = NULL;
mm/slub.c
6920
df->slab = slab;
mm/slub.c
6921
df->s = slab->slab_cache;
mm/slub.c
6923
df->slab = slab;
mm/slub.c
6924
df->s = s;
mm/slub.c
6928
df->tail = object;
mm/slub.c
6929
df->freelist = object;
mm/slub.c
6930
df->cnt = 1;
mm/slub.c
6935
set_freepointer(df->s, object, NULL);
mm/slub.c
6941
if (df->slab == virt_to_slab(object)) {
mm/slub.c
6943
set_freepointer(df->s, object, df->freelist);
mm/slub.c
6944
df->freelist = object;
mm/slub.c
6945
df->cnt++;
mm/slub.c
6970
struct detached_freelist df;
mm/slub.c
6972
size = build_detached_freelist(s, size, p, &df);
mm/slub.c
6973
if (!df.slab)
mm/slub.c
6976
if (kfence_free(df.freelist))
mm/slub.c
6979
__slab_free(df.s, df.slab, df.freelist, df.tail, df.cnt,
mm/slub.c
7000
struct detached_freelist df;
mm/slub.c
7002
size = build_detached_freelist(s, size, p, &df);
mm/slub.c
7003
if (!df.slab)
mm/slub.c
7006
slab_free_bulk(df.s, df.slab, df.freelist, df.tail, &p[size],
mm/slub.c
7007
df.cnt, _RET_IP_);
net/ipv4/ip_output.c
1409
__be16 df = 0;
net/ipv4/ip_output.c
1444
df = htons(IP_DF);
net/ipv4/ip_output.c
1460
iph->frag_off = df;
net/ipv4/ip_tunnel.c
513
struct rtable *rt, __be16 df,
net/ipv4/ip_tunnel.c
525
if (df) {
net/ipv4/ip_tunnel.c
581
__be16 df = 0;
net/ipv4/ip_tunnel.c
630
df = htons(IP_DF);
net/ipv4/ip_tunnel.c
631
if (tnl_update_pmtu(dev, skb, rt, df, inner_iph, tunnel_hlen,
net/ipv4/ip_tunnel.c
657
df, !net_eq(tunnel->net, dev_net(dev)), 0);
net/ipv4/ip_tunnel.c
684
__be16 df;
net/ipv4/ip_tunnel.c
801
df = tnl_params->frag_off;
net/ipv4/ip_tunnel.c
803
df |= (inner_iph->frag_off & htons(IP_DF));
net/ipv4/ip_tunnel.c
805
if (tnl_update_pmtu(dev, skb, rt, df, inner_iph, 0, 0, false)) {
net/ipv4/ip_tunnel.c
846
df, !net_eq(tunnel->net, dev_net(dev)), 0);
net/ipv4/ip_tunnel_core.c
52
__u8 tos, __u8 ttl, __be16 df, bool xnet,
net/ipv4/ip_tunnel_core.c
89
iph->frag_off = ip_mtu_locked(&rt->dst) ? 0 : df;
net/ipv4/udp_tunnel_core.c
176
__be16 df, __be16 src_port, __be16 dst_port,
net/ipv4/udp_tunnel_core.c
193
iptunnel_xmit(sk, rt, skb, src, dst, IPPROTO_UDP, tos, ttl, df, xnet,
net/ipv6/datagram.c
1031
int df;
net/ipv6/datagram.c
1037
df = *(int *)CMSG_DATA(cmsg);
net/ipv6/datagram.c
1038
if (df < 0 || df > 1)
net/ipv6/datagram.c
1042
ipc6->dontfrag = df;
net/ipv6/sit.c
1030
df, !net_eq(tunnel->net, dev_net(dev)), 0);
net/ipv6/sit.c
906
__be16 df = tiph->frag_off;
net/ipv6/sit.c
964
if (df) {
net/ipv6/sit.c
975
df = 0;
net/mptcp/protocol.c
1006
const struct mptcp_data_frag *df)
net/mptcp/protocol.c
1008
return df && pfrag->page == df->page &&
net/mptcp/protocol.c
1010
pfrag->offset == (df->offset + df->data_len) &&
net/mptcp/protocol.c
1011
df->data_seq + df->data_len == msk->write_seq;
net/netfilter/ipvs/ip_vs_xmit.c
1145
__be16 df = 0;
net/netfilter/ipvs/ip_vs_xmit.c
1196
dfp = sysctl_pmtu_disc(ipvs) ? &df : NULL;
net/netfilter/ipvs/ip_vs_xmit.c
1253
iph->frag_off = df;
net/netfilter/ipvs/ip_vs_xmit.c
951
__be16 *df)
net/netfilter/ipvs/ip_vs_xmit.c
982
if (df)
net/netfilter/ipvs/ip_vs_xmit.c
983
*df = 0;
net/netfilter/ipvs/ip_vs_xmit.c
989
if (df)
net/netfilter/ipvs/ip_vs_xmit.c
990
*df = (old_iph->frag_off & htons(IP_DF));
net/netfilter/nfnetlink_osf.c
181
ctx->df = ntohs(ip->frag_off) & IP_DF;
net/netfilter/nfnetlink_osf.c
220
list_for_each_entry_rcu(kf, &nf_osf_fingers[ctx.df], finger_entry) {
net/netfilter/nfnetlink_osf.c
280
list_for_each_entry_rcu(kf, &nf_osf_fingers[ctx.df], finger_entry) {
net/netfilter/nfnetlink_osf.c
345
list_for_each_entry(sf, &nf_osf_fingers[!!f->df], finger_entry) {
net/netfilter/nfnetlink_osf.c
361
list_add_tail_rcu(&kf->finger_entry, &nf_osf_fingers[!!f->df]);
net/netfilter/nfnetlink_osf.c
382
list_for_each_entry(sf, &nf_osf_fingers[!!f->df], finger_entry) {
net/netfilter/nfnetlink_osf.c
57
bool df;
net/sctp/protocol.c
1046
__be16 df = 0;
net/sctp/protocol.c
1067
df = htons(IP_DF);
net/sctp/protocol.c
1074
fl4->daddr, dscp, ip4_dst_hoplimit(dst), df,
scripts/gendwarfksyms/die.c
124
struct die_fragment *df;
scripts/gendwarfksyms/die.c
126
df = xmalloc(sizeof(*df));
scripts/gendwarfksyms/die.c
127
df->type = FRAGMENT_EMPTY;
scripts/gendwarfksyms/die.c
128
list_add_tail(&df->list, &cd->fragments);
scripts/gendwarfksyms/die.c
129
return df;
scripts/gendwarfksyms/die.c
134
struct die_fragment *df;
scripts/gendwarfksyms/die.c
139
df = append_item(cd);
scripts/gendwarfksyms/die.c
140
df->data.str = xstrdup(str);
scripts/gendwarfksyms/die.c
141
df->type = FRAGMENT_STRING;
scripts/gendwarfksyms/die.c
146
struct die_fragment *df;
scripts/gendwarfksyms/die.c
151
df = append_item(cd);
scripts/gendwarfksyms/die.c
152
df->data.linebreak = linebreak;
scripts/gendwarfksyms/die.c
153
df->type = FRAGMENT_LINEBREAK;
scripts/gendwarfksyms/die.c
158
struct die_fragment *df;
scripts/gendwarfksyms/die.c
163
df = append_item(cd);
scripts/gendwarfksyms/die.c
164
df->data.addr = child->addr;
scripts/gendwarfksyms/die.c
165
df->type = FRAGMENT_DIE;
scripts/gendwarfksyms/die.c
74
struct die_fragment *df;
scripts/gendwarfksyms/die.c
76
list_for_each_entry_safe(df, tmp, &cd->fragments, list) {
scripts/gendwarfksyms/die.c
77
if (df->type == FRAGMENT_STRING)
scripts/gendwarfksyms/die.c
78
free(df->data.str);
scripts/gendwarfksyms/die.c
79
free(df);
scripts/gendwarfksyms/dwarf.c
809
struct die_fragment *df;
scripts/gendwarfksyms/dwarf.c
812
list_for_each_entry(df, &cache->fragments, list) {
scripts/gendwarfksyms/dwarf.c
813
switch (df->type) {
scripts/gendwarfksyms/dwarf.c
816
df->data.str);
scripts/gendwarfksyms/dwarf.c
817
process(NULL, df->data.str);
scripts/gendwarfksyms/dwarf.c
820
process_linebreak(NULL, df->data.linebreak);
scripts/gendwarfksyms/dwarf.c
824
(void *)df->data.addr, &child))
scripts/gendwarfksyms/dwarf.c
827
cache, df->data.addr, dwarf_tag(&child));
scripts/gendwarfksyms/types.c
361
struct die_fragment *df;
scripts/gendwarfksyms/types.c
365
list_for_each_entry(df, &cache->fragments, list) {
scripts/gendwarfksyms/types.c
366
switch (df->type) {
scripts/gendwarfksyms/types.c
368
type_expansion_append(type, df->data.str, NULL);
scripts/gendwarfksyms/types.c
372
if (__die_map_get(df->data.addr, DIE_COMPLETE,
scripts/gendwarfksyms/types.c
374
__die_map_get(df->data.addr, DIE_UNEXPANDED,
scripts/gendwarfksyms/types.c
377
df->data.addr);
scripts/gendwarfksyms/types.c
391
if (list_is_last(&df->list, &cache->fragments) ||
scripts/gendwarfksyms/types.c
392
list_next_entry(df, list)->type !=
sound/core/init.c
422
struct snd_monitor_file *df = NULL, *_df;
sound/core/init.c
427
df = _df;
sound/core/init.c
428
list_del_init(&df->shutdown_list);
sound/core/init.c
434
if (likely(df)) {
sound/core/init.c
435
if ((file->f_flags & FASYNC) && df->disconnected_f_op->fasync)
sound/core/init.c
436
df->disconnected_f_op->fasync(-1, file, 0);
sound/core/init.c
437
return df->disconnected_f_op->release(inode, file);
tools/objtool/check.c
3544
if (state->df) {
tools/objtool/check.c
3581
if (state->df) {
tools/objtool/check.c
3979
if (statep->df) {
tools/objtool/check.c
3984
statep->df = true;
tools/objtool/check.c
3989
if (!statep->df && func) {
tools/objtool/check.c
3994
statep->df = false;
tools/objtool/include/objtool/check.h
17
bool df;