ixp_clock
ptp_clock_event(ixp_clock->ptp_clock, &event);
struct ixp_clock *ixp_clock = container_of(ptp, struct ixp_clock, caps);
struct ixp46x_ts_regs *regs = ixp_clock->regs;
struct ixp_clock *ixp_clock = container_of(ptp, struct ixp_clock, caps);
struct ixp46x_ts_regs *regs = ixp_clock->regs;
struct ixp_clock *ixp_clock = container_of(ptp, struct ixp_clock, caps);
struct ixp46x_ts_regs *regs = ixp_clock->regs;
struct ixp_clock *ixp_clock = container_of(ptp, struct ixp_clock, caps);
struct ixp46x_ts_regs *regs = ixp_clock->regs;
struct ixp_clock *ixp_clock = container_of(ptp, struct ixp_clock, caps);
ixp_clock->exts0_enabled = on ? 1 : 0;
ixp_clock->exts1_enabled = on ? 1 : 0;
static struct ixp_clock ixp_clock;
*regs = ixp_clock.regs;
*phc_index = ptp_clock_index(ixp_clock.ptp_clock);
if (!ixp_clock.ptp_clock)
ixp_clock.ptp_clock = NULL;
ixp_clock.regs = devm_platform_ioremap_resource(pdev, 0);
ixp_clock.master_irq = platform_get_irq(pdev, 0);
ixp_clock.slave_irq = platform_get_irq(pdev, 1);
if (IS_ERR(ixp_clock.regs) ||
ixp_clock.master_irq < 0 || ixp_clock.slave_irq < 0)
ixp_clock.caps = ptp_ixp_caps;
ixp_clock.ptp_clock = ptp_clock_register(&ixp_clock.caps, NULL);
if (IS_ERR(ixp_clock.ptp_clock))
return PTR_ERR(ixp_clock.ptp_clock);
ixp_clock.ptp_clock);
__raw_writel(DEFAULT_ADDEND, &ixp_clock.regs->addend);
__raw_writel(1, &ixp_clock.regs->trgt_lo);
__raw_writel(0, &ixp_clock.regs->trgt_hi);
__raw_writel(TTIPEND, &ixp_clock.regs->event);
ret = devm_request_irq(dev, ixp_clock.master_irq, isr,
0, DRIVER, &ixp_clock);
ixp_clock.master_irq);
ret = devm_request_irq(dev, ixp_clock.slave_irq, isr,
0, DRIVER, &ixp_clock);
ixp_clock.slave_irq);
struct ixp_clock *ixp_clock = priv;
struct ixp46x_ts_regs *regs = ixp_clock->regs;
if (ixp_clock->exts0_enabled) {
ptp_clock_event(ixp_clock->ptp_clock, &event);
if (ixp_clock->exts1_enabled) {