root/drivers/block/rnbd/rnbd-proto.h
/* SPDX-License-Identifier: GPL-2.0-or-later */
/*
 * RDMA Network Block Driver
 *
 * Copyright (c) 2014 - 2018 ProfitBricks GmbH. All rights reserved.
 * Copyright (c) 2018 - 2019 1&1 IONOS Cloud GmbH. All rights reserved.
 * Copyright (c) 2019 - 2020 1&1 IONOS SE. All rights reserved.
 */
#ifndef RNBD_PROTO_H
#define RNBD_PROTO_H

#include <linux/types.h>
#include <linux/blk-mq.h>
#include <linux/limits.h>
#include <linux/inet.h>
#include <linux/in.h>
#include <linux/in6.h>
#include <rdma/ib.h>

#define RNBD_PROTO_VER_MAJOR 2
#define RNBD_PROTO_VER_MINOR 2

/* The default port number the RTRS server is listening on. */
#define RTRS_PORT 1234

/**
 * enum rnbd_msg_type - RNBD message types
 * @RNBD_MSG_SESS_INFO: initial session info from client to server
 * @RNBD_MSG_SESS_INFO_RSP:     initial session info from server to client
 * @RNBD_MSG_OPEN:              open (map) device request
 * @RNBD_MSG_OPEN_RSP:          response to an @RNBD_MSG_OPEN
 * @RNBD_MSG_IO:                block IO request operation
 * @RNBD_MSG_CLOSE:             close (unmap) device request
 */
enum rnbd_msg_type {
        RNBD_MSG_SESS_INFO,
        RNBD_MSG_SESS_INFO_RSP,
        RNBD_MSG_OPEN,
        RNBD_MSG_OPEN_RSP,
        RNBD_MSG_IO,
        RNBD_MSG_CLOSE,
};

/**
 * struct rnbd_msg_hdr - header of RNBD messages
 * @type:       Message type, valid values see: enum rnbd_msg_types
 */
struct rnbd_msg_hdr {
        __le16          type;
        /* private: */
        __le16          __padding;
};

/*
 * We allow to map RO many times and RW only once. We allow to map yet another
 * time RW, if MIGRATION is provided (second RW export can be required for
 * example for VM migration)
 */
enum rnbd_access_mode {
        RNBD_ACCESS_RO,
        RNBD_ACCESS_RW,
        RNBD_ACCESS_MIGRATION,
};

static const __maybe_unused struct {
        enum rnbd_access_mode mode;
        const char *str;
} rnbd_access_modes[] = {
        [RNBD_ACCESS_RO] = {RNBD_ACCESS_RO, "ro"},
        [RNBD_ACCESS_RW] = {RNBD_ACCESS_RW, "rw"},
        [RNBD_ACCESS_MIGRATION] = {RNBD_ACCESS_MIGRATION, "migration"},
};

/**
 * struct rnbd_msg_sess_info - initial session info from client to server
 * @hdr:                message header
 * @ver:                RNBD protocol version
 */
struct rnbd_msg_sess_info {
        struct rnbd_msg_hdr hdr;
        u8              ver;
        /* private: */
        u8              reserved[31];
};

/**
 * struct rnbd_msg_sess_info_rsp - initial session info from server to client
 * @hdr:                message header
 * @ver:                RNBD protocol version
 */
struct rnbd_msg_sess_info_rsp {
        struct rnbd_msg_hdr hdr;
        u8              ver;
        /* private: */
        u8              reserved[31];
};

/**
 * struct rnbd_msg_open - request to open a remote device.
 * @hdr:                message header
 * @access_mode:        the mode to open remote device, valid values see:
 *                      enum rnbd_access_mode
 * @dev_name:           device path on remote side
 */
struct rnbd_msg_open {
        struct rnbd_msg_hdr hdr;
        u8              access_mode;
        /* private: */
        u8              resv1;
        /* public: */
        s8              dev_name[NAME_MAX];
        /* private: */
        u8              reserved[3];
};

/**
 * struct rnbd_msg_close - request to close a remote device.
 * @hdr:        message header
 * @device_id:  device_id on server side to identify the device
 */
struct rnbd_msg_close {
        struct rnbd_msg_hdr hdr;
        __le32          device_id;
};

enum rnbd_cache_policy {
        RNBD_FUA = 1 << 0,
        RNBD_WRITEBACK = 1 << 1,
};

/**
 * struct rnbd_msg_open_rsp - response message to RNBD_MSG_OPEN
 * @hdr:                message header
 * @device_id:          device_id on server side to identify the device
 * @nsectors:           number of sectors in the usual 512b unit
 * @max_hw_sectors:     max hardware sectors in the usual 512b unit
 * @max_write_zeroes_sectors: max sectors for WRITE ZEROES in the 512b unit
 * @max_discard_sectors: max. sectors that can be discarded at once in 512b
 * unit.
 * @discard_granularity: size of the internal discard allocation unit in bytes
 * @discard_alignment: offset from internal allocation assignment in bytes
 * @physical_block_size: physical block size device supports in bytes
 * @logical_block_size: logical block size device supports in bytes
 * @max_segments:       max segments hardware support in one transfer
 * @secure_discard:     supports secure discard
 * @obsolete_rotational: obsolete, not in used.
 * @cache_policy:       support write-back caching or FUA?
 */
struct rnbd_msg_open_rsp {
        struct rnbd_msg_hdr     hdr;
        __le32                  device_id;
        __le64                  nsectors;
        __le32                  max_hw_sectors;
        __le32                  max_write_zeroes_sectors;
        __le32                  max_discard_sectors;
        __le32                  discard_granularity;
        __le32                  discard_alignment;
        __le16                  physical_block_size;
        __le16                  logical_block_size;
        __le16                  max_segments;
        __le16                  secure_discard;
        u8                      obsolete_rotational;
        u8                      cache_policy;
        /* private: */
        u8                      reserved[10];
};

/**
 * struct rnbd_msg_io - message for I/O read/write
 * @hdr:        message header
 * @device_id:  device_id on server side to find the right device
 * @sector:     bi_sector attribute from struct bio
 * @rw:         valid values are defined in enum rnbd_io_flags
 * @bi_size:    number of bytes for I/O read/write
 * @prio:       priority
 */
struct rnbd_msg_io {
        struct rnbd_msg_hdr hdr;
        __le32          device_id;
        __le64          sector;
        __le32          rw;
        __le32          bi_size;
        __le16          prio;
};

#define RNBD_OP_BITS  8
#define RNBD_OP_MASK  ((1 << RNBD_OP_BITS) - 1)

/**
 * enum rnbd_io_flags - RNBD request types from rq_flag_bits
 * @RNBD_OP_READ:            read sectors from the device
 * @RNBD_OP_WRITE:           write sectors to the device
 * @RNBD_OP_FLUSH:           flush the volatile write cache
 * @RNBD_OP_DISCARD:        discard sectors
 * @RNBD_OP_SECURE_ERASE:   securely erase sectors
 * @RNBD_OP_WRITE_ZEROES:   write zeroes sectors
 *
 * @RNBD_F_SYNC:             request is sync (sync write or read)
 * @RNBD_F_FUA:             forced unit access
 * @RNBD_F_PREFLUSH:        request for cache flush
 * @RNBD_F_NOUNMAP:         do not free blocks when zeroing
 */
enum rnbd_io_flags {

        /* Operations */
        RNBD_OP_READ            = 0,
        RNBD_OP_WRITE           = 1,
        RNBD_OP_FLUSH           = 2,
        RNBD_OP_DISCARD = 3,
        RNBD_OP_SECURE_ERASE    = 4,
        RNBD_OP_WRITE_ZEROES    = 5,

        /* Flags */
        RNBD_F_SYNC  = 1<<(RNBD_OP_BITS + 0),
        RNBD_F_FUA   = 1<<(RNBD_OP_BITS + 1),
        RNBD_F_PREFLUSH = 1<<(RNBD_OP_BITS + 2),
        RNBD_F_NOUNMAP = 1<<(RNBD_OP_BITS + 3)
};

static inline u32 rnbd_op(u32 flags)
{
        return flags & RNBD_OP_MASK;
}

static inline u32 rnbd_flags(u32 flags)
{
        return flags & ~RNBD_OP_MASK;
}

static inline blk_opf_t rnbd_to_bio_flags(u32 rnbd_opf)
{
        blk_opf_t bio_opf;

        switch (rnbd_op(rnbd_opf)) {
        case RNBD_OP_READ:
                bio_opf = REQ_OP_READ;
                break;
        case RNBD_OP_WRITE:
                bio_opf = REQ_OP_WRITE;
                break;
        case RNBD_OP_FLUSH:
                bio_opf = REQ_OP_WRITE | REQ_PREFLUSH;
                break;
        case RNBD_OP_DISCARD:
                bio_opf = REQ_OP_DISCARD;
                break;
        case RNBD_OP_SECURE_ERASE:
                bio_opf = REQ_OP_SECURE_ERASE;
                break;
        case RNBD_OP_WRITE_ZEROES:
                bio_opf = REQ_OP_WRITE_ZEROES;

                if (rnbd_opf & RNBD_F_NOUNMAP)
                        bio_opf |= REQ_NOUNMAP;
                break;
        default:
                WARN(1, "Unknown RNBD type: %d (flags %d)\n",
                     rnbd_op(rnbd_opf), rnbd_opf);
                bio_opf = 0;
        }

        if (rnbd_opf & RNBD_F_SYNC)
                bio_opf |= REQ_SYNC;

        if (rnbd_opf & RNBD_F_FUA)
                bio_opf |= REQ_FUA;

        if (rnbd_opf & RNBD_F_PREFLUSH)
                bio_opf |= REQ_PREFLUSH;

        return bio_opf;
}

static inline u32 rq_to_rnbd_flags(struct request *rq)
{
        u32 rnbd_opf;

        switch (req_op(rq)) {
        case REQ_OP_READ:
                rnbd_opf = RNBD_OP_READ;
                break;
        case REQ_OP_WRITE:
                rnbd_opf = RNBD_OP_WRITE;
                break;
        case REQ_OP_DISCARD:
                rnbd_opf = RNBD_OP_DISCARD;
                break;
        case REQ_OP_SECURE_ERASE:
                rnbd_opf = RNBD_OP_SECURE_ERASE;
                break;
        case REQ_OP_WRITE_ZEROES:
                rnbd_opf = RNBD_OP_WRITE_ZEROES;

                if (rq->cmd_flags & REQ_NOUNMAP)
                        rnbd_opf |= RNBD_F_NOUNMAP;
                break;
        case REQ_OP_FLUSH:
                rnbd_opf = RNBD_OP_FLUSH;
                break;
        default:
                WARN(1, "Unknown request type %d (flags %llu)\n",
                     (__force u32)req_op(rq),
                     (__force unsigned long long)rq->cmd_flags);
                rnbd_opf = 0;
        }

        if (op_is_sync(rq->cmd_flags))
                rnbd_opf |= RNBD_F_SYNC;

        if (op_is_flush(rq->cmd_flags))
                rnbd_opf |= RNBD_F_FUA;

        if (rq->cmd_flags & REQ_PREFLUSH)
                rnbd_opf |= RNBD_F_PREFLUSH;

        return rnbd_opf;
}

const char *rnbd_access_mode_str(enum rnbd_access_mode mode);

#endif /* RNBD_PROTO_H */