e0
be64_to_cpua(c5, da, b4, e3, 93, 07, e0, 99),
be64_to_cpua(e0, 87, 2c, 44, 4b, 5a, ee, af),
be64_to_cpua(46, 1e, 77, 44, 78, e0, d1, 04),
#define INTERLEAVED_RGBX_FMT(fmt, bp, a, r, g, b, e0, e1, e2, e3) \
.element = { (e0), (e1), (e2), (e3) }, \
#define INTERLEAVED_RGBA_DX_FMT(fmt, bp, a, r, g, b, e0, e1, e2, e3) \
.element = { (e0), (e1), (e2), (e3) }, \
#define INTERLEAVED_RGBX_DX_FMT(fmt, bp, a, r, g, b, e0, e1, e2, e3) \
.element = { (e0), (e1), (e2), (e3) }, \
#define INTERLEAVED_RGB_FMT_TILED(fmt, bp, r, g, b, e0, e1, e2) \
.element = { (e0), (e1), (e2), 0 }, \
#define INTERLEAVED_RGBA_FMT_TILED(fmt, bp, a, r, g, b, e0, e1, e2, e3) \
.element = { (e0), (e1), (e2), (e3) }, \
#define INTERLEAVED_RGBX_FMT_TILED(fmt, bp, a, r, g, b, e0, e1, e2, e3) \
.element = { (e0), (e1), (e2), (e3) }, \
#define INTERLEAVED_RGBA_DX_FMT_TILED(fmt, bp, a, r, g, b, e0, e1, e2, e3) \
.element = { (e0), (e1), (e2), (e3) }, \
#define INTERLEAVED_YUV_FMT(fmt, bp, r, g, b, e0, e1, e2, e3, chroma) \
.element = { (e0), (e1), (e2), (e3)}, \
#define PSEUDO_YUV_FMT(fmt, r, g, b, e0, e1, chroma) \
.element = { (e0), (e1), 0, 0 }, \
#define PSEUDO_YUV_FMT_TILED(fmt, r, g, b, e0, e1, chroma, flg, th) \
.element = { (e0), (e1), 0, 0 }, \
#define PSEUDO_YUV_FMT_LOOSE(fmt, r, g, b, e0, e1, chroma) \
.element = { (e0), (e1), 0, 0 }, \
#define PLANAR_YUV_FMT(fmt, bp, r, g, b, e0, e1, e2, chroma) \
.element = { (e0), (e1), (e2), 0 }, \
#define INTERLEAVED_RGB_FMT(fmt, bp, r, g, b, e0, e1, e2) \
.element = { (e0), (e1), (e2), 0 }, \
#define INTERLEAVED_RGBA_FMT(fmt, bp, a, r, g, b, e0, e1, e2, e3) \
.element = { (e0), (e1), (e2), (e3) }, \
e0, e0val, \
((u64)(e0val) << DCC_CFG_SC_VL_TABLE_##range##_ENTRY##e0##_SHIFT) | \
static int handle_ea_bar(u32 e0, int bar, struct pci_bus *bus,
set_val(e0, where, size, val);
int e0 = !!(arg & 2);
err = mtk_hw_set_value(hw, desc, PINCTRL_PIN_REG_DRV_E0, e0);
u32 en, e0, e1;
err = mtk_hw_get_value(hw, desc, PINCTRL_PIN_REG_DRV_E0, &e0);
*val = (en | e0 << 1 | e1 << 2) & 0x7;
struct e_node *e, *e0 = NULL;
e0 = e;
if (!e0)
e0 = e;
if (e0)
kmem_cache_free(ntfs_enode_cachep, e0);
e = e0 ? e0 : kmem_cache_alloc(ntfs_enode_cachep, GFP_ATOMIC);
struct NTFS_DE *e, *e0, *re;
e0 = hdr_first_de(hdr);
for (e = e0;; e = hdr_next_de(hdr, e)) {
re = kmemdup(e0, to_move, GFP_NOFS);
u64 e0 = -(a & 1) & b;
u64 extra_lo = e0 ^ (e1 << 1) ^ (e2 << 2) ^ (e3 << 3);
t2 = e0(a) + Maj(a, b, c); \
t2 = e0(h) + Maj(h, a, b); c += t1; g = t1 + t2;
t2 = e0(g) + Maj(g, h, a); b += t1; f = t1 + t2;
t2 = e0(f) + Maj(f, g, h); a += t1; e = t1 + t2;
t2 = e0(e) + Maj(e, f, g); h += t1; d = t1 + t2;
t2 = e0(d) + Maj(d, e, f); g += t1; c = t1 + t2;
t2 = e0(c) + Maj(c, d, e); f += t1; b = t1 + t2;
t2 = e0(b) + Maj(b, c, d); e += t1; a = t1 + t2;
t2 = e0(a) + Maj(a, b, c); d += t1; h = t1 + t2;
struct s e0;