close_work
struct work_struct close_work;
struct ucma_context *ctx = container_of(work, struct ucma_context, close_work);
INIT_WORK(&ctx->close_work, ucma_close_id);
queue_work(system_dfl_wq, &ctx->close_work);
cancel_work_sync(&ctx->close_work);
ucma_close_id(&ctx->close_work);
INIT_WORK(&clt_path->close_work, rtrs_clt_close_work);
queue_work(rtrs_wq, &clt_path->close_work);
flush_work(&clt_path->close_work);
clt_path = container_of(work, struct rtrs_clt_path, close_work);
struct work_struct close_work;
srv_path = container_of(work, typeof(*srv_path), close_work);
INIT_WORK(&srv_path->close_work, rtrs_srv_close_work);
queue_work(rtrs_wq, &srv_path->close_work);
struct work_struct close_work;
struct cached_fid, close_work);
queue_work(serverclose_wq, &cfid->close_work);
INIT_WORK(&cfid->close_work, cached_dir_offload_close);
WARN_ON(work_pending(&cfid->close_work));
queue_work(serverclose_wq, &cfid->close_work);
struct work_struct close_work;
struct delayed_work close_work;
INIT_WORK(&acct->work, close_work);
static DECLARE_WORK(close_worker, close_work);
struct work_struct close_work; /* peer sent some closing */
if (!queue_work(smc_close_wq, &conn->close_work))
if (cancel_work_sync(&smc->conn.close_work))
close_work);
INIT_WORK(&smc->conn.close_work, smc_close_passive_work);
(!cancel_timeout || cancel_delayed_work(&vsk->close_work))) {
container_of(work, struct vsock_sock, close_work.work);
INIT_DELAYED_WORK(&vsk->close_work, hvs_close_timeout);
schedule_delayed_work(&vsk->close_work, HVS_CLOSE_TIMEOUT);
(!cancel_timeout || cancel_delayed_work(&vsk->close_work))) {
container_of(work, struct vsock_sock, close_work.work);
INIT_DELAYED_WORK(&vsk->close_work,
schedule_delayed_work(&vsk->close_work, VSOCK_CLOSE_TIMEOUT);