dev_type
int dev_type;
if ((vp->dev_type == T_ANY)
|| (vp->dev_type == SID_TYPE(&cam_dev->inq_data)))
if (vp->dev_type != T_SEQUENTIAL)
cs_dev->dev_type = desc->dev_type;
enum cs_dev_type dev_type;
enum cs_dev_type dev_type;
if (cs_dev->dev_type == event->src &&
if (cs_dev->dev_type == CORESIGHT_CPU_DEBUG &&
desc.dev_type = CORESIGHT_CPU_DEBUG;
desc.dev_type = CORESIGHT_ETMV4;
desc.dev_type = CORESIGHT_FUNNEL;
desc.dev_type = CORESIGHT_DYNAMIC_REPLICATOR;
sc->dev_type = CORESIGHT_ETR;
sc->dev_type = CORESIGHT_ETF;
sc->dev_type = CORESIGHT_UNKNOWN;
if (sc->dev_type == CORESIGHT_ETF)
KASSERT(sc->dev_type == CORESIGHT_ETR,
if (sc->dev_type == CORESIGHT_ETF)
KASSERT(sc->dev_type == CORESIGHT_ETR, ("Wrong dev_type"));
if (sc->dev_type == CORESIGHT_ETF)
desc.dev_type = CORESIGHT_TMC;
uint32_t dev_type;
uint8_t dev_type; /* Value for dev type field in EDT */
ctl_be_block_type dev_type;
be_lun->dev_type = CTL_BE_BLOCK_FILE;
be_lun->dev_type = CTL_BE_BLOCK_DEV;
switch (be_lun->dev_type) {
panic("Unexpected backend type %d", be_lun->dev_type);
be_lun->dev_type = CTL_BE_BLOCK_NONE;
uint16_t dev_type;
dev_type = gd_dev.type;
i, dev_type);
if (dev_type == GDMA_DEVICE_HWC)
if (dev_type == GDMA_DEVICE_MANA) {
tgtdev->dev_type = dev_pg0->DeviceForm;
switch (tgtdev->dev_type) {
(targ->dev_type == MPI3_DEVICE_DEVFORM_PCIE) &&
if (targ && (targ->dev_type == MPI3_DEVICE_DEVFORM_PCIE) &&
if (tgtdev && (tgtdev->dev_type == MPI3_DEVICE_DEVFORM_PCIE)) {
if (targ_dev && (targ_dev->dev_type == MPI3_DEVICE_DEVFORM_VD)) {
uint8_t dev_type;
ntb->dev_type = NTB_DEV_DSD;
ntb->dev_type = NTB_DEV_USD;
ntb->dev_type = NTB_DEV_DSD;
ntb->dev_type = NTB_DEV_USD;
ntb->dev_type = NTB_DEV_DSD;
ntb->dev_type = NTB_DEV_USD;
"features 0x%02x\n", ntb->conn_type, ntb->dev_type, ntb->features);
ntb->dev_type = NTB_DEV_USD;
ntb->dev_type = NTB_DEV_DSD;
ntb->dev_type = NTB_DEV_USD;
ntb->dev_type = NTB_DEV_DSD;
"features 0x%02x\n", ntb->conn_type, ntb->dev_type, ntb->features);
if (ntb->dev_type == NTB_DEV_USD)
if (ntb->dev_type == NTB_DEV_USD) {
enum ntb_b2b_direction dev_type;
return (ntb->dev_type == NTB_DEV_USD ? 0 : 1);
return (ntb->dev_type == NTB_DEV_USD ? 1 : 0);
&ntb->dev_type, 0, "0 - USD; 1 - DSD");
unsigned int dev_type;
if (img_ae_mode->dev_type != prod_type) {
img_ae_mode->dev_type);
pfdev_info->dev_type = p_hwfn->p_dev->type;
p_hwfn->p_dev->type = resp->pfdev_info.dev_type;
u8 dev_type;
u8 dev_type /* Device type (disk or tape). use enum fcoe_device_type (use enum fcoe_device_type) */;
NULL, &xbb->dev_type, NULL);
xbb->dev_type = NULL;
if (xbb->dev_type != NULL) {
free(xbb->dev_type, M_XENSTORE);
xbb->dev_type = NULL;
char *dev_type;
int dev_type = if_gettype(dev);
dev_addr->dev_type = ARPHRD_ETHER;
} else if (dev_type == IFT_INFINIBAND)
dev_addr->dev_type = ARPHRD_INFINIBAND;
else if (dev_type == IFT_ETHER || dev_type == IFT_L2VLAN)
dev_addr->dev_type = ARPHRD_ETHER;
dev_addr->dev_type = 0;
rt->addr.dev_addr.dev_type = ARPHRD_INFINIBAND;
id_priv->id.route.addr.dev_addr.dev_type =
dev_addr->dev_type = ARPHRD_INFINIBAND;
const int dev_type = dev_addr->dev_type;
if ((dev_type == ARPHRD_INFINIBAND) && !rdma_protocol_ib(device, port))
if ((dev_type != ARPHRD_INFINIBAND) && rdma_protocol_ib(device, port))
if (dev_type == ARPHRD_ETHER && rdma_protocol_roce(device, port)) {
if (dev_addr->dev_type != ARPHRD_INFINIBAND &&
return dev_addr->dev_type == ARPHRD_INFINIBAND ? 4 : 0;
dev_addr->dev_type != ARPHRD_INFINIBAND)
unsigned short dev_type;
int dev_type;
if (targ->g->dev_type == DEV_TAP) {
} else if (targ->g->dev_type == DEV_PCAP) {
if (targ->g->dev_type == DEV_TAP) {
} else if (targ->g->dev_type == DEV_PCAP) {
if (g->dev_type == DEV_NETMAP) {
if (g->dev_type == DEV_NETMAP) {
int dev_type;
g.dev_type = DEV_NETMAP;
g.dev_type = DEV_TAP;
g.dev_type = DEV_PCAP;
g.dev_type = DEV_NETMAP;
g.dev_type = DEV_TAP;
g.dev_type = DEV_NETMAP;
if (g.dev_type == DEV_TAP) {
} else if (g.dev_type == DEV_PCAP) {
camdd_dev_type dev_type;
camdd_dev_type dev_type;
switch (io_opts[i].dev_type) {
io_opts[i].dev_type, io_opts[i].dev_name);
io_opts->dev_type = CAMDD_DEV_FILE;
io_opts->dev_type = CAMDD_DEV_PASS;
&& (opt_list[0].dev_type != CAMDD_DEV_NONE))
&& (opt_list[1].dev_type != CAMDD_DEV_NONE))) {
if ((opt_list[0].dev_type == CAMDD_DEV_NONE)
|| (opt_list[1].dev_type == CAMDD_DEV_NONE))
struct camdd_dev *camdd_alloc_dev(camdd_dev_type dev_type,
switch (dev->dev_type) {
camdd_alloc_dev(camdd_dev_type dev_type, struct kevent *new_ke, int num_ke,
dev->dev_type = dev_type;