root/drivers/cxl/core/pmem.c
// SPDX-License-Identifier: GPL-2.0-only
/* Copyright(c) 2020 Intel Corporation. */
#include <linux/device.h>
#include <linux/slab.h>
#include <linux/idr.h>
#include <cxlmem.h>
#include <cxl.h>
#include "core.h"

/**
 * DOC: cxl pmem
 *
 * The core CXL PMEM infrastructure supports persistent memory
 * provisioning and serves as a bridge to the LIBNVDIMM subsystem. A CXL
 * 'bridge' device is added at the root of a CXL device topology if
 * platform firmware advertises at least one persistent memory capable
 * CXL window. That root-level bridge corresponds to a LIBNVDIMM 'bus'
 * device. Then for each cxl_memdev in the CXL device topology a bridge
 * device is added to host a LIBNVDIMM dimm object. When these bridges
 * are registered native LIBNVDIMM uapis are translated to CXL
 * operations, for example, namespace label access commands.
 */

static DEFINE_IDA(cxl_nvdimm_bridge_ida);

static void cxl_nvdimm_bridge_release(struct device *dev)
{
        struct cxl_nvdimm_bridge *cxl_nvb = to_cxl_nvdimm_bridge(dev);

        ida_free(&cxl_nvdimm_bridge_ida, cxl_nvb->id);
        kfree(cxl_nvb);
}

static const struct attribute_group *cxl_nvdimm_bridge_attribute_groups[] = {
        &cxl_base_attribute_group,
        NULL,
};

const struct device_type cxl_nvdimm_bridge_type = {
        .name = "cxl_nvdimm_bridge",
        .release = cxl_nvdimm_bridge_release,
        .groups = cxl_nvdimm_bridge_attribute_groups,
};

struct cxl_nvdimm_bridge *to_cxl_nvdimm_bridge(struct device *dev)
{
        if (dev_WARN_ONCE(dev, dev->type != &cxl_nvdimm_bridge_type,
                          "not a cxl_nvdimm_bridge device\n"))
                return NULL;
        return container_of(dev, struct cxl_nvdimm_bridge, dev);
}
EXPORT_SYMBOL_NS_GPL(to_cxl_nvdimm_bridge, "CXL");

/**
 * cxl_find_nvdimm_bridge() - find a bridge device relative to a port
 * @port: any descendant port of an nvdimm-bridge associated
 *        root-cxl-port
 */
struct cxl_nvdimm_bridge *cxl_find_nvdimm_bridge(struct cxl_port *port)
{
        struct cxl_root *cxl_root __free(put_cxl_root) = find_cxl_root(port);
        struct device *dev;

        if (!cxl_root)
                return NULL;

        dev = device_find_child(&cxl_root->port.dev,
                                &cxl_nvdimm_bridge_type,
                                device_match_type);

        if (!dev)
                return NULL;

        return to_cxl_nvdimm_bridge(dev);
}
EXPORT_SYMBOL_NS_GPL(cxl_find_nvdimm_bridge, "CXL");

static struct lock_class_key cxl_nvdimm_bridge_key;

static struct cxl_nvdimm_bridge *cxl_nvdimm_bridge_alloc(struct cxl_port *port)
{
        struct cxl_nvdimm_bridge *cxl_nvb;
        struct device *dev;
        int rc;

        cxl_nvb = kzalloc_obj(*cxl_nvb);
        if (!cxl_nvb)
                return ERR_PTR(-ENOMEM);

        rc = ida_alloc(&cxl_nvdimm_bridge_ida, GFP_KERNEL);
        if (rc < 0)
                goto err;
        cxl_nvb->id = rc;

        dev = &cxl_nvb->dev;
        cxl_nvb->port = port;
        device_initialize(dev);
        lockdep_set_class(&dev->mutex, &cxl_nvdimm_bridge_key);
        device_set_pm_not_required(dev);
        dev->parent = &port->dev;
        dev->bus = &cxl_bus_type;
        dev->type = &cxl_nvdimm_bridge_type;

        return cxl_nvb;

err:
        kfree(cxl_nvb);
        return ERR_PTR(rc);
}

static void unregister_nvb(void *_cxl_nvb)
{
        struct cxl_nvdimm_bridge *cxl_nvb = _cxl_nvb;

        device_unregister(&cxl_nvb->dev);
}

static bool cxl_nvdimm_bridge_failed_attach(struct cxl_nvdimm_bridge *cxl_nvb)
{
        struct device *dev = &cxl_nvb->dev;

        guard(device)(dev);
        /* If the device has no driver, then it failed to attach. */
        return dev->driver == NULL;
}

struct cxl_nvdimm_bridge *__devm_cxl_add_nvdimm_bridge(struct device *host,
                                                       struct cxl_port *port)
{
        struct cxl_nvdimm_bridge *cxl_nvb;
        struct device *dev;
        int rc;

        if (!IS_ENABLED(CONFIG_CXL_PMEM))
                return ERR_PTR(-ENXIO);

        cxl_nvb = cxl_nvdimm_bridge_alloc(port);
        if (IS_ERR(cxl_nvb))
                return cxl_nvb;

        dev = &cxl_nvb->dev;
        rc = dev_set_name(dev, "nvdimm-bridge%d", cxl_nvb->id);
        if (rc)
                goto err;

        rc = device_add(dev);
        if (rc)
                goto err;

        if (cxl_nvdimm_bridge_failed_attach(cxl_nvb)) {
                unregister_nvb(cxl_nvb);
                return ERR_PTR(-ENODEV);
        }

        rc = devm_add_action_or_reset(host, unregister_nvb, cxl_nvb);
        if (rc)
                return ERR_PTR(rc);

        return cxl_nvb;

err:
        put_device(dev);
        return ERR_PTR(rc);
}
EXPORT_SYMBOL_FOR_MODULES(__devm_cxl_add_nvdimm_bridge, "cxl_pmem");

static void cxl_nvdimm_release(struct device *dev)
{
        struct cxl_nvdimm *cxl_nvd = to_cxl_nvdimm(dev);

        kfree(cxl_nvd);
}

static const struct attribute_group *cxl_nvdimm_attribute_groups[] = {
        &cxl_base_attribute_group,
        NULL,
};

const struct device_type cxl_nvdimm_type = {
        .name = "cxl_nvdimm",
        .release = cxl_nvdimm_release,
        .groups = cxl_nvdimm_attribute_groups,
};

bool is_cxl_nvdimm(struct device *dev)
{
        return dev->type == &cxl_nvdimm_type;
}
EXPORT_SYMBOL_NS_GPL(is_cxl_nvdimm, "CXL");

struct cxl_nvdimm *to_cxl_nvdimm(struct device *dev)
{
        if (dev_WARN_ONCE(dev, !is_cxl_nvdimm(dev),
                          "not a cxl_nvdimm device\n"))
                return NULL;
        return container_of(dev, struct cxl_nvdimm, dev);
}
EXPORT_SYMBOL_NS_GPL(to_cxl_nvdimm, "CXL");

static struct lock_class_key cxl_nvdimm_key;

static struct cxl_nvdimm *cxl_nvdimm_alloc(struct cxl_nvdimm_bridge *cxl_nvb,
                                           struct cxl_memdev *cxlmd)
{
        struct cxl_nvdimm *cxl_nvd;
        struct device *dev;

        cxl_nvd = kzalloc_obj(*cxl_nvd);
        if (!cxl_nvd)
                return ERR_PTR(-ENOMEM);

        dev = &cxl_nvd->dev;
        cxl_nvd->cxlmd = cxlmd;
        cxlmd->cxl_nvd = cxl_nvd;
        device_initialize(dev);
        lockdep_set_class(&dev->mutex, &cxl_nvdimm_key);
        device_set_pm_not_required(dev);
        dev->parent = &cxlmd->dev;
        dev->bus = &cxl_bus_type;
        dev->type = &cxl_nvdimm_type;
        /*
         * A "%llx" string is 17-bytes vs dimm_id that is max
         * NVDIMM_KEY_DESC_LEN
         */
        BUILD_BUG_ON(sizeof(cxl_nvd->dev_id) < 17 ||
                     sizeof(cxl_nvd->dev_id) > NVDIMM_KEY_DESC_LEN);
        sprintf(cxl_nvd->dev_id, "%llx", cxlmd->cxlds->serial);

        return cxl_nvd;
}

static void cxlmd_release_nvdimm(void *_cxlmd)
{
        struct cxl_memdev *cxlmd = _cxlmd;
        struct cxl_nvdimm *cxl_nvd = cxlmd->cxl_nvd;
        struct cxl_nvdimm_bridge *cxl_nvb = cxlmd->cxl_nvb;

        cxl_nvd->cxlmd = NULL;
        cxlmd->cxl_nvd = NULL;
        cxlmd->cxl_nvb = NULL;
        device_unregister(&cxl_nvd->dev);
        put_device(&cxl_nvb->dev);
}

/**
 * devm_cxl_add_nvdimm() - add a bridge between a cxl_memdev and an nvdimm
 * @host: host device for devm operations
 * @port: any port in the CXL topology to find the nvdimm-bridge device
 * @cxlmd: parent of the to be created cxl_nvdimm device
 *
 * Return: 0 on success negative error code on failure.
 */
int devm_cxl_add_nvdimm(struct device *host, struct cxl_port *port,
                        struct cxl_memdev *cxlmd)
{
        struct cxl_nvdimm_bridge *cxl_nvb;
        struct cxl_nvdimm *cxl_nvd;
        struct device *dev;
        int rc;

        cxl_nvb = cxl_find_nvdimm_bridge(port);
        if (!cxl_nvb)
                return -ENODEV;

        /*
         * Take the uport_dev lock to guard against race of nvdimm_bus object.
         * cxl_acpi_probe() registers the nvdimm_bus and is done under the
         * root port uport_dev lock.
         *
         * Take the cxl_nvb device lock to ensure that cxl_nvb driver is in a
         * consistent state. And the driver registers nvdimm_bus.
         */
        guard(device)(cxl_nvb->port->uport_dev);
        guard(device)(&cxl_nvb->dev);
        if (!cxl_nvb->nvdimm_bus) {
                rc = -ENODEV;
                goto err_alloc;
        }

        cxl_nvd = cxl_nvdimm_alloc(cxl_nvb, cxlmd);
        if (IS_ERR(cxl_nvd)) {
                rc = PTR_ERR(cxl_nvd);
                goto err_alloc;
        }
        cxlmd->cxl_nvb = cxl_nvb;

        dev = &cxl_nvd->dev;
        rc = dev_set_name(dev, "pmem%d", cxlmd->id);
        if (rc)
                goto err;

        rc = device_add(dev);
        if (rc)
                goto err;

        dev_dbg(host, "register %s\n", dev_name(dev));

        /* @cxlmd carries a reference on @cxl_nvb until cxlmd_release_nvdimm */
        return devm_add_action_or_reset(host, cxlmd_release_nvdimm, cxlmd);

err:
        put_device(dev);
err_alloc:
        cxlmd->cxl_nvb = NULL;
        cxlmd->cxl_nvd = NULL;
        put_device(&cxl_nvb->dev);

        return rc;
}
EXPORT_SYMBOL_NS_GPL(devm_cxl_add_nvdimm, "CXL");