btt
enum bug_trap_type btt;
btt = report_bug_entry((struct bug_entry *)regs->gprs[2], regs);
btt = report_bug(regs->psw.addr - (regs->int_code >> 16), regs);
switch (btt) {
static int btt_meta_init(struct btt *btt)
mutex_lock(&btt->init_lock);
list_for_each_entry(arena, &btt->arena_list, list) {
btt->init_state = INIT_READY;
mutex_unlock(&btt->init_lock);
static u32 btt_meta_size(struct btt *btt)
return btt->lbasize - btt->sector_size;
static int lba_to_arena(struct btt *btt, sector_t sector, __u32 *premap,
__u64 lba = div_u64(sector << SECTOR_SHIFT, btt->sector_size);
list_for_each_entry(arena_list, &btt->arena_list, list) {
static int btt_rw_integrity(struct btt *btt, struct bio_integrity_payload *bip,
unsigned int len = btt_meta_size(btt);
meta_nsoff = to_namespace_offset(arena, postmap) + btt->sector_size;
static int btt_rw_integrity(struct btt *btt, struct bio_integrity_payload *bip,
static int btt_read_pg(struct btt *btt, struct bio_integrity_payload *bip,
lane = nd_region_acquire_lane(btt->nd_region);
ret = lba_to_arena(btt, sector, &premap, &arena);
cur_len = min(btt->sector_size, len);
ret = btt_rw_integrity(btt, bip, arena, postmap, READ);
nd_region_release_lane(btt->nd_region, lane);
sector += btt->sector_size >> SECTOR_SHIFT;
nd_region_release_lane(btt->nd_region, lane);
static bool btt_is_badblock(struct btt *btt, struct arena_info *arena,
return is_bad_pmem(btt->phys_bb, phys_sector, arena->internal_lbasize);
static int btt_write_pg(struct btt *btt, struct bio_integrity_payload *bip,
lane = nd_region_acquire_lane(btt->nd_region);
ret = lba_to_arena(btt, sector, &premap, &arena);
cur_len = min(btt->sector_size, len);
if (btt_is_badblock(btt, arena, arena->freelist[lane].block))
nd_region_release_lane(btt->nd_region, lane);
ret = btt_rw_integrity(btt, bip, arena, new_postmap,
nd_region_release_lane(btt->nd_region, lane);
sector += btt->sector_size >> SECTOR_SHIFT;
nd_region_release_lane(btt->nd_region, lane);
static int btt_do_bvec(struct btt *btt, struct bio_integrity_payload *bip,
ret = btt_read_pg(btt, bip, page, off, sector, len);
ret = btt_write_pg(btt, bip, sector, page, off, len);
struct btt *btt = bio->bi_bdev->bd_disk->private_data;
if (len > PAGE_SIZE || len < btt->sector_size ||
len % btt->sector_size) {
dev_err_ratelimited(&btt->nd_btt->dev,
err = btt_do_bvec(btt, bip, bvec.bv_page, len, bvec.bv_offset,
dev_err(&btt->nd_btt->dev,
static int btt_blk_init(struct btt *btt)
struct nd_btt *nd_btt = btt->nd_btt;
.logical_block_size = btt->sector_size,
if (btt_meta_size(btt) && IS_ENABLED(CONFIG_BLK_DEV_INTEGRITY)) {
lim.integrity.metadata_size = btt_meta_size(btt);
lim.integrity.tag_size = btt_meta_size(btt);
btt->btt_disk = blk_alloc_disk(&lim, NUMA_NO_NODE);
if (IS_ERR(btt->btt_disk))
return PTR_ERR(btt->btt_disk);
nvdimm_namespace_disk_name(ndns, btt->btt_disk->disk_name);
btt->btt_disk->first_minor = 0;
btt->btt_disk->fops = &btt_fops;
btt->btt_disk->private_data = btt;
set_capacity(btt->btt_disk, btt->nlba * btt->sector_size >> 9);
rc = device_add_disk(&btt->nd_btt->dev, btt->btt_disk, NULL);
btt->nd_btt->size = btt->nlba * (u64)btt->sector_size;
nvdimm_check_and_set_ro(btt->btt_disk);
put_disk(btt->btt_disk);
static void btt_blk_cleanup(struct btt *btt)
del_gendisk(btt->btt_disk);
put_disk(btt->btt_disk);
static struct btt *btt_init(struct nd_btt *nd_btt, unsigned long long rawsize,
struct btt *btt;
btt = devm_kzalloc(dev, sizeof(struct btt), GFP_KERNEL);
if (!btt)
btt->nd_btt = nd_btt;
btt->rawsize = rawsize;
btt->lbasize = lbasize;
btt->sector_size = ((lbasize >= 4096) ? 4096 : 512);
INIT_LIST_HEAD(&btt->arena_list);
mutex_init(&btt->init_lock);
btt->nd_region = nd_region;
btt->phys_bb = &nsio->bb;
ret = discover_arenas(btt);
if (btt->init_state != INIT_READY && nd_region->ro) {
} else if (btt->init_state != INIT_READY) {
btt->num_arenas = (rawsize / ARENA_MAX_SIZE) +
btt->num_arenas, rawsize);
ret = create_arenas(btt);
ret = btt_meta_init(btt);
ret = btt_blk_init(btt);
btt_debugfs_init(btt);
return btt;
static void btt_fini(struct btt *btt)
if (btt) {
btt_blk_cleanup(btt);
free_arenas(btt);
debugfs_remove_recursive(btt->debugfs_dir);
struct btt *btt;
btt = btt_init(nd_btt, rawsize, nd_btt->lbasize, nd_btt->uuid,
if (!btt)
nd_btt->btt = btt;
struct btt *btt = nd_btt->btt;
btt_fini(btt);
nd_btt->btt = NULL;
static void btt_debugfs_init(struct btt *btt)
btt->debugfs_dir = debugfs_create_dir(dev_name(&btt->nd_btt->dev),
if (IS_ERR_OR_NULL(btt->debugfs_dir))
list_for_each_entry(arena, &btt->arena_list, list) {
arena_debugfs_init(arena, btt->debugfs_dir, i);
static struct arena_info *alloc_arena(struct btt *btt, size_t size,
arena->nd_btt = btt->nd_btt;
arena->sector_size = btt->sector_size;
arena->external_lbasize = btt->lbasize;
arena->version_major = btt->nd_btt->version_major;
arena->version_minor = btt->nd_btt->version_minor;
static void free_arenas(struct btt *btt)
list_for_each_entry_safe(arena, next, &btt->arena_list, list) {
static int discover_arenas(struct btt *btt)
size_t remaining = btt->rawsize;
arena = alloc_arena(btt, 0, 0, 0);
if (!nd_btt_arena_is_valid(btt->nd_btt, super)) {
if (remaining == btt->rawsize) {
btt->init_state = INIT_NOTFOUND;
list_add_tail(&arena->list, &btt->arena_list);
btt->num_arenas = num_arenas;
btt->nlba = cur_nlba;
btt->init_state = INIT_READY;
free_arenas(btt);
static int create_arenas(struct btt *btt)
size_t remaining = btt->rawsize;
arena = alloc_arena(btt, arena_size, btt->nlba, cur_off);
free_arenas(btt);
btt->nlba += arena->external_nlba;
list_add_tail(&arena->list, &btt->arena_list);
struct btt;
struct btt *btt;
struct btt *btt = nd_btt->btt;
nvdimm_check_and_set_ro(btt->btt_disk);
uint8_t *btt;