resync
e = lc_find(device->resync, device->resync_wenr);
if (lc_put(device->resync, &bm_ext->lce) == 0) {
e = lc_try_get(device->resync, enr);
if (device->resync_locked > device->resync->nr_elements-3)
e = lc_get(device->resync, enr);
const unsigned long rs_flags = device->resync->flags;
lc_committed(device->resync);
if (lc_put(device->resync, &bm_ext->lce) == 0) {
e = lc_find(device->resync, enr);
if (lc_put(device->resync, &bm_ext->lce) == 0) {
lc_reset(device->resync);
for (i = 0; i < device->resync->nr_elements; i++) {
e = lc_element_by_index(device->resync, i);
lc_put(device->resync, &bm_ext->lce);
lc_del(device->resync, &bm_ext->lce);
D_ASSERT(device, device->resync->used == 0);
tmp = lc_find(device->resync, enr/AL_EXT_PER_BM_SECT);
tmp = lc_find(device->resync, enr/AL_EXT_PER_BM_SECT);
e = lc_find(device->resync, enr);
e = lc_get(device->resync, enr);
lc_committed(device->resync);
lc_put(device->resync, &ext->lce);
device->resync->nr_elements,
device->resync->flags);
if (device->resync_locked > device->resync->nr_elements/2) {
e = lc_get(device->resync, enr);
lc_committed(device->resync);
rs_flags = device->resync->flags;
if (lc_put(device->resync, &bm_ext->lce) == 0) {
lc_seq_printf_stats(m, device->resync);
lc_seq_dump_details(m, device->resync, "rs_left flags", resync_dump_detail);
struct lru_cache *resync;
lc_destroy(device->resync);
device->resync = resync_lru;
lc_seq_printf_stats(seq, device->resync);
tmp = lc_find(device->resync, BM_SECT_TO_EXT(sector));
lc_reset(device->resync);
lc_destroy(device->resync);
device->resync = NULL;
if (qpriv->resync) {
if (qpriv->resync) {
qpriv->resync = false;
return sizeof(ohdr->u.tid_rdma.resync) / sizeof(u32);
if (qpriv->resync)
qpriv->resync = true;
if (qpriv->resync) {
if (!qpriv->s_nak_state && !qpriv->resync &&
(qpriv->resync &&
le32_to_cpu(eh->tid_rdma.resync.kdeth0),
le32_to_cpu(eh->tid_rdma.resync.kdeth1),
be32_to_cpu(eh->tid_rdma.resync.verbs_qp));
__field(bool, resync)
__entry->resync = priv->resync;
__entry->resync ? "yes" : "no",
u8 resync;
resync(bc, card);
resync(bc, card);
unsigned long resync, dt, db;
resync = mddev->curr_mark_cnt - atomic_read(&mddev->recovery_active);
db = resync - mddev->resync_mark_cnt;
unsigned long long max_sectors, resync;
resync = mddev->curr_resync_completed;
return sprintf(page, "%llu / %llu\n", resync, max_sectors);
sector_t max_sectors, resync, res;
resync = mddev->curr_resync;
if (resync < MD_RESYNC_ACTIVE) {
resync = max_sectors;
} else if (resync > max_sectors) {
resync = max_sectors;
if (resync < res || resync - res < MD_RESYNC_ACTIVE)
resync = MD_RESYNC_ACTIVE;
resync -= res;
if (resync == MD_RESYNC_NONE) {
if (resync < MD_RESYNC_ACTIVE) {
res = (resync>>scale)*1000;
(unsigned long long) resync/2,
rt = max_sectors - resync; /* number of remaining sectors */
priv_rx->resync.seq, false,
struct mlx5e_ktls_rx_resync_ctx *resync;
resync = container_of(work, struct mlx5e_ktls_rx_resync_ctx, work);
priv_rx = container_of(resync, struct mlx5e_ktls_offload_context_rx, resync);
tls_offload_rx_resync_async_request_cancel(&resync->core);
c = resync->priv->channels.c[priv_rx->rxq];
tls_offload_rx_resync_async_request_cancel(&resync->core);
static void resync_init(struct mlx5e_ktls_rx_resync_ctx *resync,
INIT_WORK(&resync->work, resync_handle_work);
resync->priv = priv;
refcount_set(&resync->refcnt, 1);
memcpy(info->rec_seq, &priv_rx->resync.sw_rcd_sn_be,
memcpy(info->rec_seq, &priv_rx->resync.sw_rcd_sn_be,
struct mlx5e_ktls_rx_resync_ctx *resync;
resync = &priv_rx->resync;
if (unlikely(!queue_work(resync->priv->tls->rx_wq, &resync->work))) {
struct mlx5e_ktls_rx_resync_ctx *resync;
resync = &priv_rx->resync;
resync->sw_rcd_sn_be = *(__be64 *)rcd_sn;
resync->seq = seq;
tls_offload_rx_resync_async_request_cancel(&priv_rx->resync.core);
struct mlx5e_ktls_rx_resync_ctx resync;
struct mlx5e_ktls_rx_resync_ctx *resync;
if (!refcount_dec_and_test(&priv_rx->resync.refcnt))
resync = &priv_rx->resync;
resync_init(resync, priv);
tls_offload_ctx_rx(tls_ctx)->resync_async = &resync->core;
struct mlx5e_ktls_rx_resync_ctx *resync;
resync = &priv_rx->resync;
if (cancel_work_sync(&resync->work))
refcount_inc(&priv_rx->resync.refcnt);
raid_attr_ro_fn(resync);
int resync;
DEFINE_RAID_ATTRIBUTE(int, resync)
struct tid_rdma_resync resync;