tlsdev_ops
if (!adap->uld[CXGB4_ULD_KTLS].tlsdev_ops) {
ret = adap->uld[CXGB4_ULD_KTLS].tlsdev_ops->tls_dev_add(netdev, sk,
adap->uld[CXGB4_ULD_KTLS].tlsdev_ops->tls_dev_del(netdev, tls_ctx,
static const struct tlsdev_ops cxgb4_ktls_ops = {
netdev->tlsdev_ops = &cxgb4_ktls_ops;
const struct tlsdev_ops *tlsdev_ops;
static const struct tlsdev_ops chcr_ktls_ops = {
.tlsdev_ops = &chcr_ktls_ops,
static const struct tlsdev_ops fun_ktls_ops = {
netdev->tlsdev_ops = &fun_ktls_ops;
netdev->tlsdev_ops = &mlx5e_ktls_ops;
static const struct tlsdev_ops mlx5e_ktls_ops = {
static const struct tlsdev_ops nfp_net_tls_ops = {
netdev->tlsdev_ops = &nfp_net_tls_ops;
struct tlsdev_ops;
const struct tlsdev_ops *tlsdev_ops;
rc = netdev->tlsdev_ops->tls_dev_add(netdev, sk, TLS_OFFLOAD_CTX_DIR_TX,
rc = netdev->tlsdev_ops->tls_dev_add(netdev, sk, TLS_OFFLOAD_CTX_DIR_RX,
netdev->tlsdev_ops->tls_dev_del(netdev, tls_ctx,
netdev->tlsdev_ops->tls_dev_del(netdev, ctx,
netdev->tlsdev_ops->tls_dev_del(netdev, ctx,
if (!dev->tlsdev_ops &&
!dev->tlsdev_ops->tls_dev_resync)
if (dev->tlsdev_ops &&
dev->tlsdev_ops->tls_dev_add &&
dev->tlsdev_ops->tls_dev_del)
err = netdev->tlsdev_ops->tls_dev_resync(netdev, sk, seq,
netdev->tlsdev_ops->tls_dev_resync(netdev, sk, seq, rcd_sn,
netdev->tlsdev_ops->tls_dev_del(netdev, ctx, TLS_OFFLOAD_CTX_DIR_TX);