Message ID | 20210303062829epcms2p678450953f611c340a2b8e88b5542fe73@epcms2p6 |
---|---|
State | Superseded |
Headers | show |
Series | scsi: ufs: Add Host Performance Booster Support | expand |
On 2021-03-03 14:28, Daejun Park wrote: > This is a patch for managing L2P map in HPB module. > > The HPB divides logical addresses into several regions. A region > consists > of several sub-regions. The sub-region is a basic unit where L2P > mapping is > managed. The driver loads L2P mapping data of each sub-region. The > loaded > sub-region is called active-state. The HPB driver unloads L2P mapping > data > as region unit. The unloaded region is called inactive-state. > > Sub-region/region candidates to be loaded and unloaded are delivered > from > the UFS device. The UFS device delivers the recommended active > sub-region > and inactivate region to the driver using sensedata. > The HPB module performs L2P mapping management on the host through the > delivered information. > > A pinned region is a pre-set regions on the UFS device that is always > activate-state. > > The data structure for map data request and L2P map uses mempool API, > minimizing allocation overhead while avoiding static allocation. > > The mininum size of the memory pool used in the HPB is implemented > as a module parameter, so that it can be configurable by the user. > > To gurantee a minimum memory pool size of 4MB: > ufshpb_host_map_kbytes=4096 > > The map_work manages active/inactive by 2 "to-do" lists. > Each hpb lun maintains 2 "to-do" lists: > hpb->lh_inact_rgn - regions to be inactivated, and > hpb->lh_act_srgn - subregions to be activated > Those lists are maintained on IO completion. > > Reviewed-by: Bart Van Assche <bvanassche@acm.org> > Reviewed-by: Can Guo <cang@codeaurora.org> > Acked-by: Avri Altman <Avri.Altman@wdc.com> > Tested-by: Bean Huo <beanhuo@micron.com> > Signed-off-by: Daejun Park <daejun7.park@samsung.com> > --- > drivers/scsi/ufs/ufs.h | 36 ++ > drivers/scsi/ufs/ufshcd.c | 4 + > drivers/scsi/ufs/ufshpb.c | 1091 ++++++++++++++++++++++++++++++++++++- > drivers/scsi/ufs/ufshpb.h | 65 +++ > 4 files changed, 1181 insertions(+), 15 deletions(-) > > diff --git a/drivers/scsi/ufs/ufs.h b/drivers/scsi/ufs/ufs.h > index 65563635e20e..957763db1006 100644 > --- a/drivers/scsi/ufs/ufs.h > +++ b/drivers/scsi/ufs/ufs.h > @@ -472,6 +472,41 @@ struct utp_cmd_rsp { > u8 sense_data[UFS_SENSE_SIZE]; > }; > > +struct ufshpb_active_field { > + __be16 active_rgn; > + __be16 active_srgn; > +}; > +#define HPB_ACT_FIELD_SIZE 4 > + > +/** > + * struct utp_hpb_rsp - Response UPIU structure > + * @residual_transfer_count: Residual transfer count DW-3 > + * @reserved1: Reserved double words DW-4 to DW-7 > + * @sense_data_len: Sense data length DW-8 U16 > + * @desc_type: Descriptor type of sense data > + * @additional_len: Additional length of sense data > + * @hpb_op: HPB operation type > + * @lun: LUN of response UPIU > + * @active_rgn_cnt: Active region count > + * @inactive_rgn_cnt: Inactive region count > + * @hpb_active_field: Recommended to read HPB region and subregion > + * @hpb_inactive_field: To be inactivated HPB region and subregion > + */ > +struct utp_hpb_rsp { > + __be32 residual_transfer_count; > + __be32 reserved1[4]; > + __be16 sense_data_len; > + u8 desc_type; > + u8 additional_len; > + u8 hpb_op; > + u8 lun; > + u8 active_rgn_cnt; > + u8 inactive_rgn_cnt; > + struct ufshpb_active_field hpb_active_field[2]; > + __be16 hpb_inactive_field[2]; > +}; > +#define UTP_HPB_RSP_SIZE 40 > + > /** > * struct utp_upiu_rsp - general upiu response structure > * @header: UPIU header structure DW-0 to DW-2 > @@ -482,6 +517,7 @@ struct utp_upiu_rsp { > struct utp_upiu_header header; > union { > struct utp_cmd_rsp sr; > + struct utp_hpb_rsp hr; > struct utp_upiu_query qr; > }; > }; > diff --git a/drivers/scsi/ufs/ufshcd.c b/drivers/scsi/ufs/ufshcd.c > index 49b3d5d24fa6..5852ff44c3cc 100644 > --- a/drivers/scsi/ufs/ufshcd.c > +++ b/drivers/scsi/ufs/ufshcd.c > @@ -5021,6 +5021,9 @@ ufshcd_transfer_rsp_status(struct ufs_hba *hba, > struct ufshcd_lrb *lrbp) > */ > pm_runtime_get_noresume(hba->dev); > } > + > + if (scsi_status == SAM_STAT_GOOD) > + ufshpb_rsp_upiu(hba, lrbp); > break; > case UPIU_TRANSACTION_REJECT_UPIU: > /* TODO: handle Reject UPIU Response */ > @@ -9221,6 +9224,7 @@ EXPORT_SYMBOL(ufshcd_shutdown); > void ufshcd_remove(struct ufs_hba *hba) > { > ufs_bsg_remove(hba); > + ufshpb_remove(hba); > ufs_sysfs_remove_nodes(hba->dev); > blk_cleanup_queue(hba->tmf_queue); > blk_mq_free_tag_set(&hba->tmf_tag_set); > diff --git a/drivers/scsi/ufs/ufshpb.c b/drivers/scsi/ufs/ufshpb.c > index 1a72f6541510..8abadb0e010a 100644 > --- a/drivers/scsi/ufs/ufshpb.c > +++ b/drivers/scsi/ufs/ufshpb.c > @@ -16,6 +16,16 @@ > #include "ufshpb.h" > #include "../sd.h" > > +/* memory management */ > +static struct kmem_cache *ufshpb_mctx_cache; > +static mempool_t *ufshpb_mctx_pool; > +static mempool_t *ufshpb_page_pool; > +/* A cache size of 2MB can cache ppn in the 1GB range. */ > +static unsigned int ufshpb_host_map_kbytes = 2048; > +static int tot_active_srgn_pages; > + > +static struct workqueue_struct *ufshpb_wq; > + > bool ufshpb_is_allowed(struct ufs_hba *hba) > { > return !(hba->ufshpb_dev.hpb_disabled); > @@ -36,14 +46,892 @@ static void ufshpb_set_state(struct ufshpb_lu > *hpb, int state) > atomic_set(&hpb->hpb_state, state); > } > > +static bool ufshpb_is_general_lun(int lun) > +{ > + return lun < UFS_UPIU_MAX_UNIT_NUM_ID; > +} > + > +static bool > +ufshpb_is_pinned_region(struct ufshpb_lu *hpb, int rgn_idx) > +{ > + if (hpb->lu_pinned_end != PINNED_NOT_SET && > + rgn_idx >= hpb->lu_pinned_start && > + rgn_idx <= hpb->lu_pinned_end) > + return true; > + > + return false; > +} > + > +static void ufshpb_kick_map_work(struct ufshpb_lu *hpb) > +{ > + bool ret = false; > + unsigned long flags; > + > + if (ufshpb_get_state(hpb) != HPB_PRESENT) > + return; > + > + spin_lock_irqsave(&hpb->rsp_list_lock, flags); > + if (!list_empty(&hpb->lh_inact_rgn) || > !list_empty(&hpb->lh_act_srgn)) > + ret = true; > + spin_unlock_irqrestore(&hpb->rsp_list_lock, flags); > + > + if (ret) > + queue_work(ufshpb_wq, &hpb->map_work); > +} > + > +static bool ufshpb_is_hpb_rsp_valid(struct ufs_hba *hba, > + struct ufshcd_lrb *lrbp, > + struct utp_hpb_rsp *rsp_field) > +{ > + /* Check HPB_UPDATE_ALERT */ > + if (!(lrbp->ucd_rsp_ptr->header.dword_2 & > + UPIU_HEADER_DWORD(0, 2, 0, 0))) > + return false; > + > + if (be16_to_cpu(rsp_field->sense_data_len) != DEV_SENSE_SEG_LEN || > + rsp_field->desc_type != DEV_DES_TYPE || > + rsp_field->additional_len != DEV_ADDITIONAL_LEN || > + rsp_field->active_rgn_cnt > MAX_ACTIVE_NUM || > + rsp_field->inactive_rgn_cnt > MAX_INACTIVE_NUM || > + rsp_field->hpb_op == HPB_RSP_NONE || > + (rsp_field->hpb_op == HPB_RSP_REQ_REGION_UPDATE && > + !rsp_field->active_rgn_cnt && !rsp_field->inactive_rgn_cnt)) > + return false; > + > + if (!ufshpb_is_general_lun(rsp_field->lun)) { > + dev_warn(hba->dev, "ufshpb: lun(%d) not supported\n", > + lrbp->lun); > + return false; > + } > + > + return true; > +} > + > +static struct ufshpb_req *ufshpb_get_map_req(struct ufshpb_lu *hpb, > + struct ufshpb_subregion *srgn) > +{ > + struct ufshpb_req *map_req; > + struct request *req; > + struct bio *bio; > + int retries = HPB_MAP_REQ_RETRIES; > + > + map_req = kmem_cache_alloc(hpb->map_req_cache, GFP_KERNEL); > + if (!map_req) > + return NULL; > + > +retry: > + req = blk_get_request(hpb->sdev_ufs_lu->request_queue, > + REQ_OP_SCSI_IN, BLK_MQ_REQ_NOWAIT); > + > + if ((PTR_ERR(req) == -EWOULDBLOCK) && (--retries > 0)) { > + usleep_range(3000, 3100); > + goto retry; > + } > + > + if (IS_ERR(req)) > + goto free_map_req; > + > + bio = bio_alloc(GFP_KERNEL, hpb->pages_per_srgn); > + if (!bio) { > + blk_put_request(req); > + goto free_map_req; > + } > + > + map_req->hpb = hpb; > + map_req->req = req; > + map_req->bio = bio; > + > + map_req->rgn_idx = srgn->rgn_idx; > + map_req->srgn_idx = srgn->srgn_idx; > + map_req->mctx = srgn->mctx; > + > + return map_req; > + > +free_map_req: > + kmem_cache_free(hpb->map_req_cache, map_req); > + return NULL; > +} > + > +static void ufshpb_put_map_req(struct ufshpb_lu *hpb, > + struct ufshpb_req *map_req) > +{ > + bio_put(map_req->bio); > + blk_put_request(map_req->req); > + kmem_cache_free(hpb->map_req_cache, map_req); > +} > + > +static int ufshpb_clear_dirty_bitmap(struct ufshpb_lu *hpb, > + struct ufshpb_subregion *srgn) > +{ > + u32 num_entries = hpb->entries_per_srgn; > + > + if (!srgn->mctx) { > + dev_err(&hpb->sdev_ufs_lu->sdev_dev, > + "no mctx in region %d subregion %d.\n", > + srgn->rgn_idx, srgn->srgn_idx); > + return -1; > + } > + > + if (unlikely(srgn->is_last)) > + num_entries = hpb->last_srgn_entries; > + > + bitmap_zero(srgn->mctx->ppn_dirty, num_entries); > + return 0; > +} > + > +static void ufshpb_update_active_info(struct ufshpb_lu *hpb, int > rgn_idx, > + int srgn_idx) > +{ > + struct ufshpb_region *rgn; > + struct ufshpb_subregion *srgn; > + > + rgn = hpb->rgn_tbl + rgn_idx; > + srgn = rgn->srgn_tbl + srgn_idx; > + > + list_del_init(&rgn->list_inact_rgn); > + > + if (list_empty(&srgn->list_act_srgn)) > + list_add_tail(&srgn->list_act_srgn, &hpb->lh_act_srgn); > +} > + > +static void ufshpb_update_inactive_info(struct ufshpb_lu *hpb, int > rgn_idx) > +{ > + struct ufshpb_region *rgn; > + struct ufshpb_subregion *srgn; > + int srgn_idx; > + > + rgn = hpb->rgn_tbl + rgn_idx; > + > + for_each_sub_region(rgn, srgn_idx, srgn) > + list_del_init(&srgn->list_act_srgn); > + > + if (list_empty(&rgn->list_inact_rgn)) > + list_add_tail(&rgn->list_inact_rgn, &hpb->lh_inact_rgn); > +} > + > +static void ufshpb_activate_subregion(struct ufshpb_lu *hpb, > + struct ufshpb_subregion *srgn) > +{ > + struct ufshpb_region *rgn; > + > + /* > + * If there is no mctx in subregion > + * after I/O progress for HPB_READ_BUFFER, the region to which the > + * subregion belongs was evicted. > + * Make sure the region must not evict in I/O progress > + */ > + if (!srgn->mctx) { > + dev_err(&hpb->sdev_ufs_lu->sdev_dev, > + "no mctx in region %d subregion %d.\n", > + srgn->rgn_idx, srgn->srgn_idx); > + srgn->srgn_state = HPB_SRGN_INVALID; > + return; > + } > + > + rgn = hpb->rgn_tbl + srgn->rgn_idx; > + > + if (unlikely(rgn->rgn_state == HPB_RGN_INACTIVE)) { > + dev_err(&hpb->sdev_ufs_lu->sdev_dev, > + "region %d subregion %d evicted\n", > + srgn->rgn_idx, srgn->srgn_idx); > + srgn->srgn_state = HPB_SRGN_INVALID; > + return; > + } > + srgn->srgn_state = HPB_SRGN_VALID; > +} > + > +static void ufshpb_map_req_compl_fn(struct request *req, blk_status_t > error) > +{ > + struct ufshpb_req *map_req = (struct ufshpb_req *) req->end_io_data; > + struct ufshpb_lu *hpb = map_req->hpb; > + struct ufshpb_subregion *srgn; > + unsigned long flags; > + > + srgn = hpb->rgn_tbl[map_req->rgn_idx].srgn_tbl + > + map_req->srgn_idx; > + > + ufshpb_clear_dirty_bitmap(hpb, srgn); > + spin_lock_irqsave(&hpb->rgn_state_lock, flags); > + ufshpb_activate_subregion(hpb, srgn); > + spin_unlock_irqrestore(&hpb->rgn_state_lock, flags); > + > + ufshpb_put_map_req(map_req->hpb, map_req); > +} > + > +static void ufshpb_set_read_buf_cmd(unsigned char *cdb, int rgn_idx, > + int srgn_idx, int srgn_mem_size) > +{ > + cdb[0] = UFSHPB_READ_BUFFER; > + cdb[1] = UFSHPB_READ_BUFFER_ID; > + > + put_unaligned_be16(rgn_idx, &cdb[2]); > + put_unaligned_be16(srgn_idx, &cdb[4]); > + put_unaligned_be24(srgn_mem_size, &cdb[6]); > + > + cdb[9] = 0x00; > +} > + > +static int ufshpb_execute_map_req(struct ufshpb_lu *hpb, > + struct ufshpb_req *map_req, bool last) > +{ > + struct request_queue *q; > + struct request *req; > + struct scsi_request *rq; > + int mem_size = hpb->srgn_mem_size; > + int ret = 0; > + int i; > + > + q = hpb->sdev_ufs_lu->request_queue; > + for (i = 0; i < hpb->pages_per_srgn; i++) { > + ret = bio_add_pc_page(q, map_req->bio, map_req->mctx->m_page[i], > + PAGE_SIZE, 0); > + if (ret != PAGE_SIZE) { > + dev_err(&hpb->sdev_ufs_lu->sdev_dev, > + "bio_add_pc_page fail %d - %d\n", > + map_req->rgn_idx, map_req->srgn_idx); > + return ret; > + } > + } > + > + req = map_req->req; > + > + blk_rq_append_bio(req, &map_req->bio); > + > + req->end_io_data = map_req; > + > + rq = scsi_req(req); > + > + if (unlikely(last)) > + mem_size = hpb->last_srgn_entries * HPB_ENTRY_SIZE; > + > + ufshpb_set_read_buf_cmd(rq->cmd, map_req->rgn_idx, > + map_req->srgn_idx, mem_size); > + rq->cmd_len = HPB_READ_BUFFER_CMD_LENGTH; > + > + blk_execute_rq_nowait(q, NULL, req, 1, ufshpb_map_req_compl_fn); > + > + hpb->stats.map_req_cnt++; > + return 0; > +} > + > +static struct ufshpb_map_ctx *ufshpb_get_map_ctx(struct ufshpb_lu > *hpb, > + bool last) > +{ > + struct ufshpb_map_ctx *mctx; > + u32 num_entries = hpb->entries_per_srgn; > + int i, j; > + > + mctx = mempool_alloc(ufshpb_mctx_pool, GFP_KERNEL); > + if (!mctx) > + return NULL; > + > + mctx->m_page = kmem_cache_alloc(hpb->m_page_cache, GFP_KERNEL); > + if (!mctx->m_page) > + goto release_mctx; > + > + if (unlikely(last)) > + num_entries = hpb->last_srgn_entries; > + > + mctx->ppn_dirty = bitmap_zalloc(num_entries, GFP_KERNEL); > + if (!mctx->ppn_dirty) > + goto release_m_page; > + > + for (i = 0; i < hpb->pages_per_srgn; i++) { > + mctx->m_page[i] = mempool_alloc(ufshpb_page_pool, GFP_KERNEL); > + if (!mctx->m_page[i]) { > + for (j = 0; j < i; j++) > + mempool_free(mctx->m_page[j], ufshpb_page_pool); > + goto release_ppn_dirty; > + } > + clear_page(page_address(mctx->m_page[i])); > + } > + > + return mctx; > + > +release_ppn_dirty: > + bitmap_free(mctx->ppn_dirty); > +release_m_page: > + kmem_cache_free(hpb->m_page_cache, mctx->m_page); > +release_mctx: > + mempool_free(mctx, ufshpb_mctx_pool); > + return NULL; > +} > + > +static void ufshpb_put_map_ctx(struct ufshpb_lu *hpb, > + struct ufshpb_map_ctx *mctx) > +{ > + int i; > + > + for (i = 0; i < hpb->pages_per_srgn; i++) > + mempool_free(mctx->m_page[i], ufshpb_page_pool); > + > + bitmap_free(mctx->ppn_dirty); > + kmem_cache_free(hpb->m_page_cache, mctx->m_page); > + mempool_free(mctx, ufshpb_mctx_pool); > +} > + > +static int ufshpb_check_srgns_issue_state(struct ufshpb_lu *hpb, > + struct ufshpb_region *rgn) > +{ > + struct ufshpb_subregion *srgn; > + int srgn_idx; > + > + for_each_sub_region(rgn, srgn_idx, srgn) > + if (srgn->srgn_state == HPB_SRGN_ISSUED) > + return -EPERM; > + > + return 0; > +} > + > +static void ufshpb_add_lru_info(struct victim_select_info *lru_info, > + struct ufshpb_region *rgn) > +{ > + rgn->rgn_state = HPB_RGN_ACTIVE; > + list_add_tail(&rgn->list_lru_rgn, &lru_info->lh_lru_rgn); > + atomic_inc(&lru_info->active_cnt); > +} > + > +static void ufshpb_hit_lru_info(struct victim_select_info *lru_info, > + struct ufshpb_region *rgn) > +{ > + list_move_tail(&rgn->list_lru_rgn, &lru_info->lh_lru_rgn); > +} > + > +static struct ufshpb_region *ufshpb_victim_lru_info(struct ufshpb_lu > *hpb) > +{ > + struct victim_select_info *lru_info = &hpb->lru_info; > + struct ufshpb_region *rgn, *victim_rgn = NULL; > + > + list_for_each_entry(rgn, &lru_info->lh_lru_rgn, list_lru_rgn) { > + if (!rgn) { > + dev_err(&hpb->sdev_ufs_lu->sdev_dev, > + "%s: no region allocated\n", > + __func__); > + return NULL; > + } > + if (ufshpb_check_srgns_issue_state(hpb, rgn)) > + continue; > + > + victim_rgn = rgn; > + break; > + } > + > + return victim_rgn; > +} > + > +static void ufshpb_cleanup_lru_info(struct victim_select_info > *lru_info, > + struct ufshpb_region *rgn) > +{ > + list_del_init(&rgn->list_lru_rgn); > + rgn->rgn_state = HPB_RGN_INACTIVE; > + atomic_dec(&lru_info->active_cnt); > +} > + > +static void ufshpb_purge_active_subregion(struct ufshpb_lu *hpb, > + struct ufshpb_subregion *srgn) > +{ > + if (srgn->srgn_state != HPB_SRGN_UNUSED) { > + ufshpb_put_map_ctx(hpb, srgn->mctx); > + srgn->srgn_state = HPB_SRGN_UNUSED; > + srgn->mctx = NULL; > + } > +} > + > +static void __ufshpb_evict_region(struct ufshpb_lu *hpb, > + struct ufshpb_region *rgn) > +{ > + struct victim_select_info *lru_info; > + struct ufshpb_subregion *srgn; > + int srgn_idx; > + > + lru_info = &hpb->lru_info; > + > + dev_dbg(&hpb->sdev_ufs_lu->sdev_dev, "evict region %d\n", > rgn->rgn_idx); > + > + ufshpb_cleanup_lru_info(lru_info, rgn); > + > + for_each_sub_region(rgn, srgn_idx, srgn) > + ufshpb_purge_active_subregion(hpb, srgn); > +} > + > +static int ufshpb_evict_region(struct ufshpb_lu *hpb, struct > ufshpb_region *rgn) > +{ > + unsigned long flags; > + int ret = 0; > + > + spin_lock_irqsave(&hpb->rgn_state_lock, flags); > + if (rgn->rgn_state == HPB_RGN_PINNED) { > + dev_warn(&hpb->sdev_ufs_lu->sdev_dev, > + "pinned region cannot drop-out. region %d\n", > + rgn->rgn_idx); > + goto out; > + } > + if (!list_empty(&rgn->list_lru_rgn)) { > + if (ufshpb_check_srgns_issue_state(hpb, rgn)) { > + ret = -EBUSY; > + goto out; > + } > + > + __ufshpb_evict_region(hpb, rgn); > + } > +out: > + spin_unlock_irqrestore(&hpb->rgn_state_lock, flags); > + return ret; > +} > + > +static int ufshpb_issue_map_req(struct ufshpb_lu *hpb, > + struct ufshpb_region *rgn, > + struct ufshpb_subregion *srgn) > +{ > + struct ufshpb_req *map_req; > + unsigned long flags; > + int ret; > + int err = -EAGAIN; > + bool alloc_required = false; > + enum HPB_SRGN_STATE state = HPB_SRGN_INVALID; > + > + spin_lock_irqsave(&hpb->rgn_state_lock, flags); > + > + if (ufshpb_get_state(hpb) != HPB_PRESENT) { > + dev_notice(&hpb->sdev_ufs_lu->sdev_dev, > + "%s: ufshpb state is not PRESENT\n", __func__); > + goto unlock_out; > + } > + > + if ((rgn->rgn_state == HPB_RGN_INACTIVE) && > + (srgn->srgn_state == HPB_SRGN_INVALID)) { > + err = 0; > + goto unlock_out; > + } > + > + if (srgn->srgn_state == HPB_SRGN_UNUSED) > + alloc_required = true; > + > + /* > + * If the subregion is already ISSUED state, > + * a specific event (e.g., GC or wear-leveling, etc.) occurs in > + * the device and HPB response for map loading is received. > + * In this case, after finishing the HPB_READ_BUFFER, > + * the next HPB_READ_BUFFER is performed again to obtain the latest > + * map data. > + */ > + if (srgn->srgn_state == HPB_SRGN_ISSUED) > + goto unlock_out; > + > + srgn->srgn_state = HPB_SRGN_ISSUED; > + spin_unlock_irqrestore(&hpb->rgn_state_lock, flags); > + > + if (alloc_required) { > + if (srgn->mctx) { Can this really happen? alloc_required is true if srgn->srgn_state == HPB_SRGN_UNUSED. ufshpb_put_map_ctx() is called and srgn->srgn_state is given HPB_SRGN_UNUSED together in ufshpb_purge_active_subregion(). Do I miss anything? Thanks, Can Guo. > + dev_err(&hpb->sdev_ufs_lu->sdev_dev, > + "map_ctx is already allocated. region %d - %d\n", > + rgn->rgn_idx, srgn->srgn_idx); > + ufshpb_put_map_ctx(hpb, srgn->mctx); > + } > + srgn->mctx = ufshpb_get_map_ctx(hpb, srgn->is_last); > + if (!srgn->mctx) { > + dev_err(&hpb->sdev_ufs_lu->sdev_dev, > + "get map_ctx failed. region %d - %d\n", > + rgn->rgn_idx, srgn->srgn_idx); > + state = HPB_SRGN_UNUSED; > + goto change_srgn_state; > + } > + } > + > + map_req = ufshpb_get_map_req(hpb, srgn); > + if (!map_req) > + goto change_srgn_state; > + > + > + ret = ufshpb_execute_map_req(hpb, map_req, srgn->is_last); > + if (ret) { > + dev_err(&hpb->sdev_ufs_lu->sdev_dev, > + "%s: issue map_req failed: %d, region %d - %d\n", > + __func__, ret, srgn->rgn_idx, srgn->srgn_idx); > + goto free_map_req; > + } > + return 0; > + > +free_map_req: > + ufshpb_put_map_req(hpb, map_req); > +change_srgn_state: > + spin_lock_irqsave(&hpb->rgn_state_lock, flags); > + srgn->srgn_state = state; > +unlock_out: > + spin_unlock_irqrestore(&hpb->rgn_state_lock, flags); > + return err; > +} > + > +static int ufshpb_add_region(struct ufshpb_lu *hpb, struct > ufshpb_region *rgn) > +{ > + struct ufshpb_region *victim_rgn; > + struct victim_select_info *lru_info = &hpb->lru_info; > + unsigned long flags; > + int ret = 0; > + > + spin_lock_irqsave(&hpb->rgn_state_lock, flags); > + /* > + * If region belongs to lru_list, just move the region > + * to the front of lru list. because the state of the region > + * is already active-state > + */ > + if (!list_empty(&rgn->list_lru_rgn)) { > + ufshpb_hit_lru_info(lru_info, rgn); > + goto out; > + } > + > + if (rgn->rgn_state == HPB_RGN_INACTIVE) { > + if (atomic_read(&lru_info->active_cnt) == > + lru_info->max_lru_active_cnt) { > + /* > + * If the maximum number of active regions > + * is exceeded, evict the least recently used region. > + * This case may occur when the device responds > + * to the eviction information late. > + * It is okay to evict the least recently used region, > + * because the device could detect this region > + * by not issuing HPB_READ > + */ > + victim_rgn = ufshpb_victim_lru_info(hpb); > + if (!victim_rgn) { > + dev_warn(&hpb->sdev_ufs_lu->sdev_dev, > + "cannot get victim region error\n"); > + ret = -ENOMEM; > + goto out; > + } > + > + dev_dbg(&hpb->sdev_ufs_lu->sdev_dev, > + "LRU full (%d), choose victim %d\n", > + atomic_read(&lru_info->active_cnt), > + victim_rgn->rgn_idx); > + __ufshpb_evict_region(hpb, victim_rgn); > + } > + > + /* > + * When a region is added to lru_info list_head, > + * it is guaranteed that the subregion has been > + * assigned all mctx. If failed, try to receive mctx again > + * without being added to lru_info list_head > + */ > + ufshpb_add_lru_info(lru_info, rgn); > + } > +out: > + spin_unlock_irqrestore(&hpb->rgn_state_lock, flags); > + return ret; > +} > + > +static void ufshpb_rsp_req_region_update(struct ufshpb_lu *hpb, > + struct utp_hpb_rsp *rsp_field) > +{ > + struct ufshpb_region *rgn; > + struct ufshpb_subregion *srgn; > + int i, rgn_i, srgn_i; > + > + BUILD_BUG_ON(sizeof(struct ufshpb_active_field) != > HPB_ACT_FIELD_SIZE); > + /* > + * If the active region and the inactive region are the same, > + * we will inactivate this region. > + * The device could check this (region inactivated) and > + * will response the proper active region information > + */ > + for (i = 0; i < rsp_field->active_rgn_cnt; i++) { > + rgn_i = > + be16_to_cpu(rsp_field->hpb_active_field[i].active_rgn); > + srgn_i = > + be16_to_cpu(rsp_field->hpb_active_field[i].active_srgn); > + > + dev_dbg(&hpb->sdev_ufs_lu->sdev_dev, > + "activate(%d) region %d - %d\n", i, rgn_i, srgn_i); > + > + spin_lock(&hpb->rsp_list_lock); > + ufshpb_update_active_info(hpb, rgn_i, srgn_i); > + spin_unlock(&hpb->rsp_list_lock); > + > + rgn = hpb->rgn_tbl + rgn_i; > + srgn = rgn->srgn_tbl + srgn_i; > + > + /* blocking HPB_READ */ > + spin_lock(&hpb->rgn_state_lock); > + if (srgn->srgn_state == HPB_SRGN_VALID) > + srgn->srgn_state = HPB_SRGN_INVALID; > + spin_unlock(&hpb->rgn_state_lock); > + hpb->stats.rb_active_cnt++; > + } > + > + for (i = 0; i < rsp_field->inactive_rgn_cnt; i++) { > + rgn_i = be16_to_cpu(rsp_field->hpb_inactive_field[i]); > + dev_dbg(&hpb->sdev_ufs_lu->sdev_dev, > + "inactivate(%d) region %d\n", i, rgn_i); > + > + spin_lock(&hpb->rsp_list_lock); > + ufshpb_update_inactive_info(hpb, rgn_i); > + spin_unlock(&hpb->rsp_list_lock); > + > + rgn = hpb->rgn_tbl + rgn_i; > + > + spin_lock(&hpb->rgn_state_lock); > + if (rgn->rgn_state != HPB_RGN_INACTIVE) { > + for (srgn_i = 0; srgn_i < rgn->srgn_cnt; srgn_i++) { > + srgn = rgn->srgn_tbl + srgn_i; > + if (srgn->srgn_state == HPB_SRGN_VALID) > + srgn->srgn_state = HPB_SRGN_INVALID; > + } > + } > + spin_unlock(&hpb->rgn_state_lock); > + > + hpb->stats.rb_inactive_cnt++; > + } > + > + dev_dbg(&hpb->sdev_ufs_lu->sdev_dev, "Noti: #ACT %u #INACT %u\n", > + rsp_field->active_rgn_cnt, rsp_field->inactive_rgn_cnt); > + > + if (ufshpb_get_state(hpb) == HPB_PRESENT) > + queue_work(ufshpb_wq, &hpb->map_work); > +} > + > +/* > + * This function will parse recommended active subregion information > in sense > + * data field of response UPIU with SAM_STAT_GOOD state. > + */ > +void ufshpb_rsp_upiu(struct ufs_hba *hba, struct ufshcd_lrb *lrbp) > +{ > + struct ufshpb_lu *hpb = ufshpb_get_hpb_data(lrbp->cmd->device); > + struct utp_hpb_rsp *rsp_field = &lrbp->ucd_rsp_ptr->hr; > + int data_seg_len; > + > + if (unlikely(lrbp->lun != rsp_field->lun)) { > + struct scsi_device *sdev; > + bool found = false; > + > + __shost_for_each_device(sdev, hba->host) { > + hpb = ufshpb_get_hpb_data(sdev); > + > + if (!hpb) > + continue; > + > + if (rsp_field->lun == hpb->lun) { > + found = true; > + break; > + } > + } > + > + if (!found) > + return; > + } > + > + if (!hpb) > + return; > + > + if ((ufshpb_get_state(hpb) != HPB_PRESENT) && > + (ufshpb_get_state(hpb) != HPB_SUSPEND)) { > + dev_notice(&hpb->sdev_ufs_lu->sdev_dev, > + "%s: ufshpb state is not PRESENT/SUSPEND\n", > + __func__); > + return; > + } > + > + data_seg_len = be32_to_cpu(lrbp->ucd_rsp_ptr->header.dword_2) > + & MASK_RSP_UPIU_DATA_SEG_LEN; > + > + /* To flush remained rsp_list, we queue the map_work task */ > + if (!data_seg_len) { > + if (!ufshpb_is_general_lun(hpb->lun)) > + return; > + > + ufshpb_kick_map_work(hpb); > + return; > + } > + > + BUILD_BUG_ON(sizeof(struct utp_hpb_rsp) != UTP_HPB_RSP_SIZE); > + > + if (!ufshpb_is_hpb_rsp_valid(hba, lrbp, rsp_field)) > + return; > + > + hpb->stats.rb_noti_cnt++; > + > + switch (rsp_field->hpb_op) { > + case HPB_RSP_REQ_REGION_UPDATE: > + if (data_seg_len != DEV_DATA_SEG_LEN) > + dev_warn(&hpb->sdev_ufs_lu->sdev_dev, > + "%s: data seg length is not same.\n", > + __func__); > + ufshpb_rsp_req_region_update(hpb, rsp_field); > + break; > + case HPB_RSP_DEV_RESET: > + dev_warn(&hpb->sdev_ufs_lu->sdev_dev, > + "UFS device lost HPB information during PM.\n"); > + break; > + default: > + dev_notice(&hpb->sdev_ufs_lu->sdev_dev, > + "hpb_op is not available: %d\n", > + rsp_field->hpb_op); > + break; > + } > +} > + > +static void ufshpb_add_active_list(struct ufshpb_lu *hpb, > + struct ufshpb_region *rgn, > + struct ufshpb_subregion *srgn) > +{ > + if (!list_empty(&rgn->list_inact_rgn)) > + return; > + > + if (!list_empty(&srgn->list_act_srgn)) { > + list_move(&srgn->list_act_srgn, &hpb->lh_act_srgn); > + return; > + } > + > + list_add(&srgn->list_act_srgn, &hpb->lh_act_srgn); > +} > + > +static void ufshpb_add_pending_evict_list(struct ufshpb_lu *hpb, > + struct ufshpb_region *rgn, > + struct list_head *pending_list) > +{ > + struct ufshpb_subregion *srgn; > + int srgn_idx; > + > + if (!list_empty(&rgn->list_inact_rgn)) > + return; > + > + for_each_sub_region(rgn, srgn_idx, srgn) > + if (!list_empty(&srgn->list_act_srgn)) > + return; > + > + list_add_tail(&rgn->list_inact_rgn, pending_list); > +} > + > +static void ufshpb_run_active_subregion_list(struct ufshpb_lu *hpb) > +{ > + struct ufshpb_region *rgn; > + struct ufshpb_subregion *srgn; > + unsigned long flags; > + int ret = 0; > + > + spin_lock_irqsave(&hpb->rsp_list_lock, flags); > + while ((srgn = list_first_entry_or_null(&hpb->lh_act_srgn, > + struct ufshpb_subregion, > + list_act_srgn))) { > + if (ufshpb_get_state(hpb) == HPB_SUSPEND) > + break; > + > + list_del_init(&srgn->list_act_srgn); > + spin_unlock_irqrestore(&hpb->rsp_list_lock, flags); > + > + rgn = hpb->rgn_tbl + srgn->rgn_idx; > + ret = ufshpb_add_region(hpb, rgn); > + if (ret) > + goto active_failed; > + > + ret = ufshpb_issue_map_req(hpb, rgn, srgn); > + if (ret) { > + dev_err(&hpb->sdev_ufs_lu->sdev_dev, > + "issue map_req failed. ret %d, region %d - %d\n", > + ret, rgn->rgn_idx, srgn->srgn_idx); > + goto active_failed; > + } > + spin_lock_irqsave(&hpb->rsp_list_lock, flags); > + } > + spin_unlock_irqrestore(&hpb->rsp_list_lock, flags); > + return; > + > +active_failed: > + dev_err(&hpb->sdev_ufs_lu->sdev_dev, "failed to activate region %d - > %d, will retry\n", > + rgn->rgn_idx, srgn->srgn_idx); > + spin_lock_irqsave(&hpb->rsp_list_lock, flags); > + ufshpb_add_active_list(hpb, rgn, srgn); > + spin_unlock_irqrestore(&hpb->rsp_list_lock, flags); > +} > + > +static void ufshpb_run_inactive_region_list(struct ufshpb_lu *hpb) > +{ > + struct ufshpb_region *rgn; > + unsigned long flags; > + int ret; > + LIST_HEAD(pending_list); > + > + spin_lock_irqsave(&hpb->rsp_list_lock, flags); > + while ((rgn = list_first_entry_or_null(&hpb->lh_inact_rgn, > + struct ufshpb_region, > + list_inact_rgn))) { > + if (ufshpb_get_state(hpb) == HPB_SUSPEND) > + break; > + > + list_del_init(&rgn->list_inact_rgn); > + spin_unlock_irqrestore(&hpb->rsp_list_lock, flags); > + > + ret = ufshpb_evict_region(hpb, rgn); > + if (ret) { > + spin_lock_irqsave(&hpb->rsp_list_lock, flags); > + ufshpb_add_pending_evict_list(hpb, rgn, &pending_list); > + spin_unlock_irqrestore(&hpb->rsp_list_lock, flags); > + } > + > + spin_lock_irqsave(&hpb->rsp_list_lock, flags); > + } > + > + list_splice(&pending_list, &hpb->lh_inact_rgn); > + spin_unlock_irqrestore(&hpb->rsp_list_lock, flags); > +} > + > +static void ufshpb_map_work_handler(struct work_struct *work) > +{ > + struct ufshpb_lu *hpb = container_of(work, struct ufshpb_lu, > map_work); > + > + if (ufshpb_get_state(hpb) != HPB_PRESENT) { > + dev_notice(&hpb->sdev_ufs_lu->sdev_dev, > + "%s: ufshpb state is not PRESENT\n", __func__); > + return; > + } > + > + ufshpb_run_inactive_region_list(hpb); > + ufshpb_run_active_subregion_list(hpb); > +} > + > +/* > + * this function doesn't need to hold lock due to be called in init. > + * (rgn_state_lock, rsp_list_lock, etc..) > + */ > +static int ufshpb_init_pinned_active_region(struct ufs_hba *hba, > + struct ufshpb_lu *hpb, > + struct ufshpb_region *rgn) > +{ > + struct ufshpb_subregion *srgn; > + int srgn_idx, i; > + int err = 0; > + > + for_each_sub_region(rgn, srgn_idx, srgn) { > + srgn->mctx = ufshpb_get_map_ctx(hpb, srgn->is_last); > + srgn->srgn_state = HPB_SRGN_INVALID; > + if (!srgn->mctx) { > + err = -ENOMEM; > + dev_err(hba->dev, > + "alloc mctx for pinned region failed\n"); > + goto release; > + } > + > + list_add_tail(&srgn->list_act_srgn, &hpb->lh_act_srgn); > + } > + > + rgn->rgn_state = HPB_RGN_PINNED; > + return 0; > + > +release: > + for (i = 0; i < srgn_idx; i++) { > + srgn = rgn->srgn_tbl + i; > + ufshpb_put_map_ctx(hpb, srgn->mctx); > + } > + return err; > +} > + > static void ufshpb_init_subregion_tbl(struct ufshpb_lu *hpb, > struct ufshpb_region *rgn, bool last) > { > int srgn_idx; > struct ufshpb_subregion *srgn; > > - for (srgn_idx = 0; srgn_idx < rgn->srgn_cnt; srgn_idx++) { > - srgn = rgn->srgn_tbl + srgn_idx; > + for_each_sub_region(rgn, srgn_idx, srgn) { > + INIT_LIST_HEAD(&srgn->list_act_srgn); > > srgn->rgn_idx = rgn->rgn_idx; > srgn->srgn_idx = srgn_idx; > @@ -78,6 +966,8 @@ static void ufshpb_lu_parameter_init(struct ufs_hba > *hba, > hpb->lu_pinned_end = hpb_lu_info->num_pinned ? > (hpb_lu_info->pinned_start + hpb_lu_info->num_pinned - 1) > : PINNED_NOT_SET; > + hpb->lru_info.max_lru_active_cnt = > + hpb_lu_info->max_active_rgns - hpb_lu_info->num_pinned; > > rgn_mem_size = (1ULL << hpb_dev_info->rgn_size) * HPB_RGN_SIZE_UNIT > * HPB_ENTRY_SIZE; > @@ -129,6 +1019,9 @@ static int ufshpb_alloc_region_tbl(struct ufs_hba > *hba, struct ufshpb_lu *hpb) > rgn = rgn_table + rgn_idx; > rgn->rgn_idx = rgn_idx; > > + INIT_LIST_HEAD(&rgn->list_inact_rgn); > + INIT_LIST_HEAD(&rgn->list_lru_rgn); > + > if (rgn_idx == hpb->rgns_per_lu - 1) { > srgn_cnt = ((hpb->srgns_per_lu - 1) % > hpb->srgns_per_rgn) + 1; > @@ -140,7 +1033,13 @@ static int ufshpb_alloc_region_tbl(struct > ufs_hba *hba, struct ufshpb_lu *hpb) > goto release_srgn_table; > ufshpb_init_subregion_tbl(hpb, rgn, last_srgn); > > - rgn->rgn_state = HPB_RGN_INACTIVE; > + if (ufshpb_is_pinned_region(hpb, rgn_idx)) { > + ret = ufshpb_init_pinned_active_region(hba, hpb, rgn); > + if (ret) > + goto release_srgn_table; > + } else { > + rgn->rgn_state = HPB_RGN_INACTIVE; > + } > } > > return 0; > @@ -159,13 +1058,13 @@ static void ufshpb_destroy_subregion_tbl(struct > ufshpb_lu *hpb, > struct ufshpb_region *rgn) > { > int srgn_idx; > + struct ufshpb_subregion *srgn; > > - for (srgn_idx = 0; srgn_idx < rgn->srgn_cnt; srgn_idx++) { > - struct ufshpb_subregion *srgn; > - > - srgn = rgn->srgn_tbl + srgn_idx; > - srgn->srgn_state = HPB_SRGN_UNUSED; > - } > + for_each_sub_region(rgn, srgn_idx, srgn) > + if (srgn->srgn_state != HPB_SRGN_UNUSED) { > + srgn->srgn_state = HPB_SRGN_UNUSED; > + ufshpb_put_map_ctx(hpb, srgn->mctx); > + } > } > > static void ufshpb_destroy_region_tbl(struct ufshpb_lu *hpb) > @@ -239,11 +1138,47 @@ static int ufshpb_lu_hpb_init(struct ufs_hba > *hba, struct ufshpb_lu *hpb) > { > int ret; > > + spin_lock_init(&hpb->rgn_state_lock); > + spin_lock_init(&hpb->rsp_list_lock); > + > + INIT_LIST_HEAD(&hpb->lru_info.lh_lru_rgn); > + INIT_LIST_HEAD(&hpb->lh_act_srgn); > + INIT_LIST_HEAD(&hpb->lh_inact_rgn); > + INIT_LIST_HEAD(&hpb->list_hpb_lu); > + > + INIT_WORK(&hpb->map_work, ufshpb_map_work_handler); > + > + hpb->map_req_cache = kmem_cache_create("ufshpb_req_cache", > + sizeof(struct ufshpb_req), 0, 0, NULL); > + if (!hpb->map_req_cache) { > + dev_err(hba->dev, "ufshpb(%d) ufshpb_req_cache create fail", > + hpb->lun); > + return -ENOMEM; > + } > + > + hpb->m_page_cache = kmem_cache_create("ufshpb_m_page_cache", > + sizeof(struct page *) * hpb->pages_per_srgn, > + 0, 0, NULL); > + if (!hpb->m_page_cache) { > + dev_err(hba->dev, "ufshpb(%d) ufshpb_m_page_cache create fail", > + hpb->lun); > + ret = -ENOMEM; > + goto release_req_cache; > + } > + > ret = ufshpb_alloc_region_tbl(hba, hpb); > + if (ret) > + goto release_m_page_cache; > > ufshpb_stat_init(hpb); > > return 0; > + > +release_m_page_cache: > + kmem_cache_destroy(hpb->m_page_cache); > +release_req_cache: > + kmem_cache_destroy(hpb->map_req_cache); > + return ret; > } > > static struct ufshpb_lu * > @@ -275,6 +1210,33 @@ ufshpb_alloc_hpb_lu(struct ufs_hba *hba, int lun, > return NULL; > } > > +static void ufshpb_discard_rsp_lists(struct ufshpb_lu *hpb) > +{ > + struct ufshpb_region *rgn, *next_rgn; > + struct ufshpb_subregion *srgn, *next_srgn; > + unsigned long flags; > + > + /* > + * If the device reset occurred, the remained HPB region information > + * may be stale. Therefore, by dicarding the lists of HPB response > + * that remained after reset, it prevents unnecessary work. > + */ > + spin_lock_irqsave(&hpb->rsp_list_lock, flags); > + list_for_each_entry_safe(rgn, next_rgn, &hpb->lh_inact_rgn, > + list_inact_rgn) > + list_del_init(&rgn->list_inact_rgn); > + > + list_for_each_entry_safe(srgn, next_srgn, &hpb->lh_act_srgn, > + list_act_srgn) > + list_del_init(&srgn->list_act_srgn); > + spin_unlock_irqrestore(&hpb->rsp_list_lock, flags); > +} > + > +static void ufshpb_cancel_jobs(struct ufshpb_lu *hpb) > +{ > + cancel_work_sync(&hpb->map_work); > +} > + > static bool ufshpb_check_hpb_reset_query(struct ufs_hba *hba) > { > int err = 0; > @@ -318,7 +1280,7 @@ void ufshpb_reset(struct ufs_hba *hba) > struct scsi_device *sdev; > > shost_for_each_device(sdev, hba->host) { > - hpb = sdev->hostdata; > + hpb = ufshpb_get_hpb_data(sdev); > if (!hpb) > continue; > > @@ -335,13 +1297,15 @@ void ufshpb_reset_host(struct ufs_hba *hba) > struct scsi_device *sdev; > > shost_for_each_device(sdev, hba->host) { > - hpb = sdev->hostdata; > + hpb = ufshpb_get_hpb_data(sdev); > if (!hpb) > continue; > > if (ufshpb_get_state(hpb) != HPB_PRESENT) > continue; > ufshpb_set_state(hpb, HPB_RESET); > + ufshpb_cancel_jobs(hpb); > + ufshpb_discard_rsp_lists(hpb); > } > } > > @@ -351,13 +1315,14 @@ void ufshpb_suspend(struct ufs_hba *hba) > struct scsi_device *sdev; > > shost_for_each_device(sdev, hba->host) { > - hpb = sdev->hostdata; > + hpb = ufshpb_get_hpb_data(sdev); > if (!hpb) > continue; > > if (ufshpb_get_state(hpb) != HPB_PRESENT) > continue; > ufshpb_set_state(hpb, HPB_SUSPEND); > + ufshpb_cancel_jobs(hpb); > } > } > > @@ -367,7 +1332,7 @@ void ufshpb_resume(struct ufs_hba *hba) > struct scsi_device *sdev; > > shost_for_each_device(sdev, hba->host) { > - hpb = sdev->hostdata; > + hpb = ufshpb_get_hpb_data(sdev); > if (!hpb) > continue; > > @@ -375,6 +1340,7 @@ void ufshpb_resume(struct ufs_hba *hba) > (ufshpb_get_state(hpb) != HPB_SUSPEND)) > continue; > ufshpb_set_state(hpb, HPB_PRESENT); > + ufshpb_kick_map_work(hpb); > } > } > > @@ -427,7 +1393,7 @@ static int ufshpb_get_lu_info(struct ufs_hba *hba, > int lun, > > void ufshpb_destroy_lu(struct ufs_hba *hba, struct scsi_device *sdev) > { > - struct ufshpb_lu *hpb = sdev->hostdata; > + struct ufshpb_lu *hpb = ufshpb_get_hpb_data(sdev); > > if (!hpb) > return; > @@ -437,8 +1403,13 @@ void ufshpb_destroy_lu(struct ufs_hba *hba, > struct scsi_device *sdev) > sdev = hpb->sdev_ufs_lu; > sdev->hostdata = NULL; > > + ufshpb_cancel_jobs(hpb); > + > ufshpb_destroy_region_tbl(hpb); > > + kmem_cache_destroy(hpb->map_req_cache); > + kmem_cache_destroy(hpb->m_page_cache); > + > list_del_init(&hpb->list_hpb_lu); > > kfree(hpb); > @@ -446,24 +1417,41 @@ void ufshpb_destroy_lu(struct ufs_hba *hba, > struct scsi_device *sdev) > > static void ufshpb_hpb_lu_prepared(struct ufs_hba *hba) > { > + int pool_size; > struct ufshpb_lu *hpb; > struct scsi_device *sdev; > bool init_success; > > + if (tot_active_srgn_pages == 0) { > + ufshpb_remove(hba); > + return; > + } > + > init_success = !ufshpb_check_hpb_reset_query(hba); > > + pool_size = PAGE_ALIGN(ufshpb_host_map_kbytes * 1024) / PAGE_SIZE; > + if (pool_size > tot_active_srgn_pages) { > + mempool_resize(ufshpb_mctx_pool, tot_active_srgn_pages); > + mempool_resize(ufshpb_page_pool, tot_active_srgn_pages); > + } > + > shost_for_each_device(sdev, hba->host) { > - hpb = sdev->hostdata; > + hpb = ufshpb_get_hpb_data(sdev); > if (!hpb) > continue; > > if (init_success) { > ufshpb_set_state(hpb, HPB_PRESENT); > + if ((hpb->lu_pinned_end - hpb->lu_pinned_start) > 0) > + queue_work(ufshpb_wq, &hpb->map_work); > } else { > dev_err(hba->dev, "destroy HPB lu %d\n", hpb->lun); > ufshpb_destroy_lu(hba, sdev); > } > } > + > + if (!init_success) > + ufshpb_remove(hba); > } > > void ufshpb_init_hpb_lu(struct ufs_hba *hba, struct scsi_device *sdev) > @@ -485,6 +1473,9 @@ void ufshpb_init_hpb_lu(struct ufs_hba *hba, > struct scsi_device *sdev) > if (!hpb) > goto out; > > + tot_active_srgn_pages += hpb_lu_info.max_active_rgns * > + hpb->srgns_per_rgn * hpb->pages_per_srgn; > + > hpb->sdev_ufs_lu = sdev; > sdev->hostdata = hpb; > > @@ -494,6 +1485,57 @@ void ufshpb_init_hpb_lu(struct ufs_hba *hba, > struct scsi_device *sdev) > ufshpb_hpb_lu_prepared(hba); > } > > +static int ufshpb_init_mem_wq(struct ufs_hba *hba) > +{ > + int ret; > + unsigned int pool_size; > + > + ufshpb_mctx_cache = kmem_cache_create("ufshpb_mctx_cache", > + sizeof(struct ufshpb_map_ctx), > + 0, 0, NULL); > + if (!ufshpb_mctx_cache) { > + dev_err(hba->dev, "ufshpb: cannot init mctx cache\n"); > + return -ENOMEM; > + } > + > + pool_size = PAGE_ALIGN(ufshpb_host_map_kbytes * 1024) / PAGE_SIZE; > + dev_info(hba->dev, "%s:%d ufshpb_host_map_kbytes %u pool_size %u\n", > + __func__, __LINE__, ufshpb_host_map_kbytes, pool_size); > + > + ufshpb_mctx_pool = mempool_create_slab_pool(pool_size, > + ufshpb_mctx_cache); > + if (!ufshpb_mctx_pool) { > + dev_err(hba->dev, "ufshpb: cannot init mctx pool\n"); > + ret = -ENOMEM; > + goto release_mctx_cache; > + } > + > + ufshpb_page_pool = mempool_create_page_pool(pool_size, 0); > + if (!ufshpb_page_pool) { > + dev_err(hba->dev, "ufshpb: cannot init page pool\n"); > + ret = -ENOMEM; > + goto release_mctx_pool; > + } > + > + ufshpb_wq = alloc_workqueue("ufshpb-wq", > + WQ_UNBOUND | WQ_MEM_RECLAIM, 0); > + if (!ufshpb_wq) { > + dev_err(hba->dev, "ufshpb: alloc workqueue failed\n"); > + ret = -ENOMEM; > + goto release_page_pool; > + } > + > + return 0; > + > +release_page_pool: > + mempool_destroy(ufshpb_page_pool); > +release_mctx_pool: > + mempool_destroy(ufshpb_mctx_pool); > +release_mctx_cache: > + kmem_cache_destroy(ufshpb_mctx_cache); > + return ret; > +} > + > void ufshpb_get_geo_info(struct ufs_hba *hba, u8 *geo_buf) > { > struct ufshpb_dev_info *hpb_info = &hba->ufshpb_dev; > @@ -558,7 +1600,13 @@ void ufshpb_init(struct ufs_hba *hba) > if (!ufshpb_is_allowed(hba) || !hba->dev_info.hpb_enabled) > return; > > + if (ufshpb_init_mem_wq(hba)) { > + hpb_dev_info->hpb_disabled = true; > + return; > + } > + > atomic_set(&hpb_dev_info->slave_conf_cnt, hpb_dev_info->num_lu); > + tot_active_srgn_pages = 0; > /* issue HPB reset query */ > for (try = 0; try < HPB_RESET_REQ_RETRIES; try++) { > ret = ufshcd_query_flag(hba, UPIU_QUERY_OPCODE_SET_FLAG, > @@ -567,3 +1615,16 @@ void ufshpb_init(struct ufs_hba *hba) > break; > } > } > + > +void ufshpb_remove(struct ufs_hba *hba) > +{ > + mempool_destroy(ufshpb_page_pool); > + mempool_destroy(ufshpb_mctx_pool); > + kmem_cache_destroy(ufshpb_mctx_cache); > + > + destroy_workqueue(ufshpb_wq); > +} > + > +module_param(ufshpb_host_map_kbytes, uint, 0644); > +MODULE_PARM_DESC(ufshpb_host_map_kbytes, > + "ufshpb host mapping memory kilo-bytes for ufshpb memory-pool"); > diff --git a/drivers/scsi/ufs/ufshpb.h b/drivers/scsi/ufs/ufshpb.h > index 11f5b018af51..aaffc8968afd 100644 > --- a/drivers/scsi/ufs/ufshpb.h > +++ b/drivers/scsi/ufs/ufshpb.h > @@ -40,6 +40,7 @@ > #define LU_ENABLED_HPB_FUNC 0x02 > > #define HPB_RESET_REQ_RETRIES 10 > +#define HPB_MAP_REQ_RETRIES 5 > > #define HPB_SUPPORT_VERSION 0x100 > > @@ -83,11 +84,19 @@ struct ufshpb_lu_info { > int max_active_rgns; > }; > > +struct ufshpb_map_ctx { > + struct page **m_page; > + unsigned long *ppn_dirty; > +}; > + > struct ufshpb_subregion { > + struct ufshpb_map_ctx *mctx; > enum HPB_SRGN_STATE srgn_state; > int rgn_idx; > int srgn_idx; > bool is_last; > + /* below information is used by rsp_list */ > + struct list_head list_act_srgn; > }; > > struct ufshpb_region { > @@ -95,6 +104,43 @@ struct ufshpb_region { > enum HPB_RGN_STATE rgn_state; > int rgn_idx; > int srgn_cnt; > + > + /* below information is used by rsp_list */ > + struct list_head list_inact_rgn; > + > + /* below information is used by lru */ > + struct list_head list_lru_rgn; > +}; > + > +#define for_each_sub_region(rgn, i, srgn) \ > + for ((i) = 0; \ > + ((i) < (rgn)->srgn_cnt) && ((srgn) = &(rgn)->srgn_tbl[i]); \ > + (i)++) > + > +/** > + * struct ufshpb_req - UFSHPB READ BUFFER (for caching map) request > structure > + * @req: block layer request for READ BUFFER > + * @bio: bio for holding map page > + * @hpb: ufshpb_lu structure that related to the L2P map > + * @mctx: L2P map information > + * @rgn_idx: target region index > + * @srgn_idx: target sub-region index > + * @lun: target logical unit number > + */ > +struct ufshpb_req { > + struct request *req; > + struct bio *bio; > + struct ufshpb_lu *hpb; > + struct ufshpb_map_ctx *mctx; > + > + unsigned int rgn_idx; > + unsigned int srgn_idx; > +}; > + > +struct victim_select_info { > + struct list_head lh_lru_rgn; /* LRU list of regions */ > + int max_lru_active_cnt; /* supported hpb #region - pinned #region */ > + atomic_t active_cnt; > }; > > struct ufshpb_stats { > @@ -109,10 +155,22 @@ struct ufshpb_stats { > struct ufshpb_lu { > int lun; > struct scsi_device *sdev_ufs_lu; > + > + spinlock_t rgn_state_lock; /* for protect rgn/srgn state */ > struct ufshpb_region *rgn_tbl; > > atomic_t hpb_state; > > + spinlock_t rsp_list_lock; > + struct list_head lh_act_srgn; /* hold rsp_list_lock */ > + struct list_head lh_inact_rgn; /* hold rsp_list_lock */ > + > + /* cached L2P map management worker */ > + struct work_struct map_work; > + > + /* for selecting victim */ > + struct victim_select_info lru_info; > + > /* pinned region information */ > u32 lu_pinned_start; > u32 lu_pinned_end; > @@ -132,6 +190,9 @@ struct ufshpb_lu { > > struct ufshpb_stats stats; > > + struct kmem_cache *map_req_cache; > + struct kmem_cache *m_page_cache; > + > struct list_head list_hpb_lu; > }; > > @@ -139,6 +200,7 @@ struct ufs_hba; > struct ufshcd_lrb; > > #ifndef CONFIG_SCSI_UFS_HPB > +static void ufshpb_rsp_upiu(struct ufs_hba *hba, struct ufshcd_lrb > *lrbp) {} > static void ufshpb_resume(struct ufs_hba *hba) {} > static void ufshpb_suspend(struct ufs_hba *hba) {} > static void ufshpb_reset(struct ufs_hba *hba) {} > @@ -146,10 +208,12 @@ static void ufshpb_reset_host(struct ufs_hba > *hba) {} > static void ufshpb_init(struct ufs_hba *hba) {} > static void ufshpb_init_hpb_lu(struct ufs_hba *hba, struct > scsi_device *sdev) {} > static void ufshpb_destroy_lu(struct ufs_hba *hba, struct scsi_device > *sdev) {} > +static void ufshpb_remove(struct ufs_hba *hba) {} > static bool ufshpb_is_allowed(struct ufs_hba *hba) { return false; } > static void ufshpb_get_geo_info(struct ufs_hba *hba, u8 *geo_buf) {} > static void ufshpb_get_dev_info(struct ufs_hba *hba, u8 *desc_buf) {} > #else > +void ufshpb_rsp_upiu(struct ufs_hba *hba, struct ufshcd_lrb *lrbp); > void ufshpb_resume(struct ufs_hba *hba); > void ufshpb_suspend(struct ufs_hba *hba); > void ufshpb_reset(struct ufs_hba *hba); > @@ -157,6 +221,7 @@ void ufshpb_reset_host(struct ufs_hba *hba); > void ufshpb_init(struct ufs_hba *hba); > void ufshpb_init_hpb_lu(struct ufs_hba *hba, struct scsi_device > *sdev); > void ufshpb_destroy_lu(struct ufs_hba *hba, struct scsi_device *sdev); > +void ufshpb_remove(struct ufs_hba *hba); > bool ufshpb_is_allowed(struct ufs_hba *hba); > void ufshpb_get_geo_info(struct ufs_hba *hba, u8 *geo_buf); > void ufshpb_get_dev_info(struct ufs_hba *hba, u8 *desc_buf);
On 2021-03-03 14:28, Daejun Park wrote: > This is a patch for managing L2P map in HPB module. > > The HPB divides logical addresses into several regions. A region > consists > of several sub-regions. The sub-region is a basic unit where L2P > mapping is > managed. The driver loads L2P mapping data of each sub-region. The > loaded > sub-region is called active-state. The HPB driver unloads L2P mapping > data > as region unit. The unloaded region is called inactive-state. > > Sub-region/region candidates to be loaded and unloaded are delivered > from > the UFS device. The UFS device delivers the recommended active > sub-region > and inactivate region to the driver using sensedata. > The HPB module performs L2P mapping management on the host through the > delivered information. > > A pinned region is a pre-set regions on the UFS device that is always > activate-state. > > The data structure for map data request and L2P map uses mempool API, > minimizing allocation overhead while avoiding static allocation. > > The mininum size of the memory pool used in the HPB is implemented > as a module parameter, so that it can be configurable by the user. > > To gurantee a minimum memory pool size of 4MB: > ufshpb_host_map_kbytes=4096 > > The map_work manages active/inactive by 2 "to-do" lists. > Each hpb lun maintains 2 "to-do" lists: > hpb->lh_inact_rgn - regions to be inactivated, and > hpb->lh_act_srgn - subregions to be activated > Those lists are maintained on IO completion. > > Reviewed-by: Bart Van Assche <bvanassche@acm.org> > Reviewed-by: Can Guo <cang@codeaurora.org> > Acked-by: Avri Altman <Avri.Altman@wdc.com> > Tested-by: Bean Huo <beanhuo@micron.com> > Signed-off-by: Daejun Park <daejun7.park@samsung.com> > --- > drivers/scsi/ufs/ufs.h | 36 ++ > drivers/scsi/ufs/ufshcd.c | 4 + > drivers/scsi/ufs/ufshpb.c | 1091 ++++++++++++++++++++++++++++++++++++- > drivers/scsi/ufs/ufshpb.h | 65 +++ > 4 files changed, 1181 insertions(+), 15 deletions(-) > > diff --git a/drivers/scsi/ufs/ufs.h b/drivers/scsi/ufs/ufs.h > index 65563635e20e..957763db1006 100644 > --- a/drivers/scsi/ufs/ufs.h > +++ b/drivers/scsi/ufs/ufs.h > @@ -472,6 +472,41 @@ struct utp_cmd_rsp { > u8 sense_data[UFS_SENSE_SIZE]; > }; > ... > +/* > + * This function will parse recommended active subregion information > in sense > + * data field of response UPIU with SAM_STAT_GOOD state. > + */ > +void ufshpb_rsp_upiu(struct ufs_hba *hba, struct ufshcd_lrb *lrbp) > +{ > + struct ufshpb_lu *hpb = ufshpb_get_hpb_data(lrbp->cmd->device); > + struct utp_hpb_rsp *rsp_field = &lrbp->ucd_rsp_ptr->hr; > + int data_seg_len; > + > + if (unlikely(lrbp->lun != rsp_field->lun)) { > + struct scsi_device *sdev; > + bool found = false; > + > + __shost_for_each_device(sdev, hba->host) { > + hpb = ufshpb_get_hpb_data(sdev); > + > + if (!hpb) > + continue; > + > + if (rsp_field->lun == hpb->lun) { > + found = true; > + break; > + } > + } > + > + if (!found) > + return; > + } > + > + if (!hpb) > + return; > + > + if ((ufshpb_get_state(hpb) != HPB_PRESENT) && > + (ufshpb_get_state(hpb) != HPB_SUSPEND)) { > + dev_notice(&hpb->sdev_ufs_lu->sdev_dev, > + "%s: ufshpb state is not PRESENT/SUSPEND\n", > + __func__); Please mute these prints before hpb is fully initilized, otherwise there can be tons of these prints during bootup. Say set a flag in ufshpb_hpb_lu_prepared() and check for that flag - just a rough idea. Thanks, Can Guo. > + return; > + } > +
> > The HPB divides logical addresses into several regions. A region > > consists > > of several sub-regions. The sub-region is a basic unit where L2P > > mapping is > > managed. The driver loads L2P mapping data of each sub-region. The > > loaded > > sub-region is called active-state. The HPB driver unloads L2P mapping > > data > > as region unit. The unloaded region is called inactive-state. > > > > Sub-region/region candidates to be loaded and unloaded are delivered > > from > > the UFS device. The UFS device delivers the recommended active > > sub-region > > and inactivate region to the driver using sensedata. > > The HPB module performs L2P mapping management on the host through the > > delivered information. > > > > A pinned region is a pre-set regions on the UFS device that is always > > activate-state. > > > > The data structure for map data request and L2P map uses mempool API, > > minimizing allocation overhead while avoiding static allocation. > > > > The mininum size of the memory pool used in the HPB is implemented > > as a module parameter, so that it can be configurable by the user. > > > > To gurantee a minimum memory pool size of 4MB: > > ufshpb_host_map_kbytes=4096 > > > > The map_work manages active/inactive by 2 "to-do" lists. > > Each hpb lun maintains 2 "to-do" lists: > > hpb->lh_inact_rgn - regions to be inactivated, and > > hpb->lh_act_srgn - subregions to be activated > > Those lists are maintained on IO completion. > > > > Reviewed-by: Bart Van Assche <bvanassche@acm.org> > > Reviewed-by: Can Guo <cang@codeaurora.org> > > Acked-by: Avri Altman <Avri.Altman@wdc.com> > > Tested-by: Bean Huo <beanhuo@micron.com> > > Signed-off-by: Daejun Park <daejun7.park@samsung.com> > > --- > > drivers/scsi/ufs/ufs.h | 36 ++ > > drivers/scsi/ufs/ufshcd.c | 4 + > > drivers/scsi/ufs/ufshpb.c | 1091 ++++++++++++++++++++++++++++++++++++- > > drivers/scsi/ufs/ufshpb.h | 65 +++ > > 4 files changed, 1181 insertions(+), 15 deletions(-) > > > > diff --git a/drivers/scsi/ufs/ufs.h b/drivers/scsi/ufs/ufs.h > > index 65563635e20e..957763db1006 100644 > > --- a/drivers/scsi/ufs/ufs.h > > +++ b/drivers/scsi/ufs/ufs.h > > @@ -472,6 +472,41 @@ struct utp_cmd_rsp { > > u8 sense_data[UFS_SENSE_SIZE]; > > }; > > ... > > +/* > > + * This function will parse recommended active subregion information > > in sense > > + * data field of response UPIU with SAM_STAT_GOOD state. > > + */ > > +void ufshpb_rsp_upiu(struct ufs_hba *hba, struct ufshcd_lrb *lrbp) > > +{ > > + struct ufshpb_lu *hpb = ufshpb_get_hpb_data(lrbp->cmd->device); > > + struct utp_hpb_rsp *rsp_field = &lrbp->ucd_rsp_ptr->hr; > > + int data_seg_len; > > + > > + if (unlikely(lrbp->lun != rsp_field->lun)) { > > + struct scsi_device *sdev; > > + bool found = false; > > + > > + __shost_for_each_device(sdev, hba->host) { > > + hpb = ufshpb_get_hpb_data(sdev); > > + > > + if (!hpb) > > + continue; > > + > > + if (rsp_field->lun == hpb->lun) { > > + found = true; > > + break; > > + } > > + } > > + > > + if (!found) > > + return; > > + } > > + > > + if (!hpb) > > + return; > > + > > + if ((ufshpb_get_state(hpb) != HPB_PRESENT) && > > + (ufshpb_get_state(hpb) != HPB_SUSPEND)) { > > + dev_notice(&hpb->sdev_ufs_lu->sdev_dev, > > + "%s: ufshpb state is not PRESENT/SUSPEND\n", > > + __func__); > > Please mute these prints before hpb is fully initilized, otherwise > there can be tons of these prints during bootup. Say set a flag in > ufshpb_hpb_lu_prepared() and check for that flag - just a rough idea. OK, I will change it. Thanks, Daejun
> > This is a patch for managing L2P map in HPB module. > > > > The HPB divides logical addresses into several regions. A region > > consists > > of several sub-regions. The sub-region is a basic unit where L2P > > mapping is > > managed. The driver loads L2P mapping data of each sub-region. The > > loaded > > sub-region is called active-state. The HPB driver unloads L2P mapping > > data > > as region unit. The unloaded region is called inactive-state. > > > > Sub-region/region candidates to be loaded and unloaded are delivered > > from > > the UFS device. The UFS device delivers the recommended active > > sub-region > > and inactivate region to the driver using sensedata. > > The HPB module performs L2P mapping management on the host through the > > delivered information. > > > > A pinned region is a pre-set regions on the UFS device that is always > > activate-state. > > > > The data structure for map data request and L2P map uses mempool API, > > minimizing allocation overhead while avoiding static allocation. > > > > The mininum size of the memory pool used in the HPB is implemented > > as a module parameter, so that it can be configurable by the user. > > > > To gurantee a minimum memory pool size of 4MB: > > ufshpb_host_map_kbytes=4096 > > > > The map_work manages active/inactive by 2 "to-do" lists. > > Each hpb lun maintains 2 "to-do" lists: > > hpb->lh_inact_rgn - regions to be inactivated, and > > hpb->lh_act_srgn - subregions to be activated > > Those lists are maintained on IO completion. > > > > Reviewed-by: Bart Van Assche <bvanassche@acm.org> > > Reviewed-by: Can Guo <cang@codeaurora.org> > > Acked-by: Avri Altman <Avri.Altman@wdc.com> > > Tested-by: Bean Huo <beanhuo@micron.com> > > Signed-off-by: Daejun Park <daejun7.park@samsung.com> > > --- > > drivers/scsi/ufs/ufs.h | 36 ++ > > drivers/scsi/ufs/ufshcd.c | 4 + > > drivers/scsi/ufs/ufshpb.c | 1091 ++++++++++++++++++++++++++++++++++++- > > drivers/scsi/ufs/ufshpb.h | 65 +++ > > 4 files changed, 1181 insertions(+), 15 deletions(-) > > > > diff --git a/drivers/scsi/ufs/ufs.h b/drivers/scsi/ufs/ufs.h > > index 65563635e20e..957763db1006 100644 > > --- a/drivers/scsi/ufs/ufs.h > > +++ b/drivers/scsi/ufs/ufs.h > > @@ -472,6 +472,41 @@ struct utp_cmd_rsp { > > u8 sense_data[UFS_SENSE_SIZE]; > > }; > > > > +struct ufshpb_active_field { > > + __be16 active_rgn; > > + __be16 active_srgn; > > +}; > > +#define HPB_ACT_FIELD_SIZE 4 > > + > > +/** > > + * struct utp_hpb_rsp - Response UPIU structure > > + * @residual_transfer_count: Residual transfer count DW-3 > > + * @reserved1: Reserved double words DW-4 to DW-7 > > + * @sense_data_len: Sense data length DW-8 U16 > > + * @desc_type: Descriptor type of sense data > > + * @additional_len: Additional length of sense data > > + * @hpb_op: HPB operation type > > + * @lun: LUN of response UPIU > > + * @active_rgn_cnt: Active region count > > + * @inactive_rgn_cnt: Inactive region count > > + * @hpb_active_field: Recommended to read HPB region and subregion > > + * @hpb_inactive_field: To be inactivated HPB region and subregion > > + */ > > +struct utp_hpb_rsp { > > + __be32 residual_transfer_count; > > + __be32 reserved1[4]; > > + __be16 sense_data_len; > > + u8 desc_type; > > + u8 additional_len; > > + u8 hpb_op; > > + u8 lun; > > + u8 active_rgn_cnt; > > + u8 inactive_rgn_cnt; > > + struct ufshpb_active_field hpb_active_field[2]; > > + __be16 hpb_inactive_field[2]; > > +}; > > +#define UTP_HPB_RSP_SIZE 40 > > + > > /** > > * struct utp_upiu_rsp - general upiu response structure > > * @header: UPIU header structure DW-0 to DW-2 > > @@ -482,6 +517,7 @@ struct utp_upiu_rsp { > > struct utp_upiu_header header; > > union { > > struct utp_cmd_rsp sr; > > + struct utp_hpb_rsp hr; > > struct utp_upiu_query qr; > > }; > > }; > > diff --git a/drivers/scsi/ufs/ufshcd.c b/drivers/scsi/ufs/ufshcd.c > > index 49b3d5d24fa6..5852ff44c3cc 100644 > > --- a/drivers/scsi/ufs/ufshcd.c > > +++ b/drivers/scsi/ufs/ufshcd.c > > @@ -5021,6 +5021,9 @@ ufshcd_transfer_rsp_status(struct ufs_hba *hba, > > struct ufshcd_lrb *lrbp) > > */ > > pm_runtime_get_noresume(hba->dev); > > } > > + > > + if (scsi_status == SAM_STAT_GOOD) > > + ufshpb_rsp_upiu(hba, lrbp); > > break; > > case UPIU_TRANSACTION_REJECT_UPIU: > > /* TODO: handle Reject UPIU Response */ > > @@ -9221,6 +9224,7 @@ EXPORT_SYMBOL(ufshcd_shutdown); > > void ufshcd_remove(struct ufs_hba *hba) > > { > > ufs_bsg_remove(hba); > > + ufshpb_remove(hba); > > ufs_sysfs_remove_nodes(hba->dev); > > blk_cleanup_queue(hba->tmf_queue); > > blk_mq_free_tag_set(&hba->tmf_tag_set); > > diff --git a/drivers/scsi/ufs/ufshpb.c b/drivers/scsi/ufs/ufshpb.c > > index 1a72f6541510..8abadb0e010a 100644 > > --- a/drivers/scsi/ufs/ufshpb.c > > +++ b/drivers/scsi/ufs/ufshpb.c > > @@ -16,6 +16,16 @@ > > #include "ufshpb.h" > > #include "../sd.h" > > > > +/* memory management */ > > +static struct kmem_cache *ufshpb_mctx_cache; > > +static mempool_t *ufshpb_mctx_pool; > > +static mempool_t *ufshpb_page_pool; > > +/* A cache size of 2MB can cache ppn in the 1GB range. */ > > +static unsigned int ufshpb_host_map_kbytes = 2048; > > +static int tot_active_srgn_pages; > > + > > +static struct workqueue_struct *ufshpb_wq; > > + > > bool ufshpb_is_allowed(struct ufs_hba *hba) > > { > > return !(hba->ufshpb_dev.hpb_disabled); > > @@ -36,14 +46,892 @@ static void ufshpb_set_state(struct ufshpb_lu > > *hpb, int state) > > atomic_set(&hpb->hpb_state, state); > > } > > > > +static bool ufshpb_is_general_lun(int lun) > > +{ > > + return lun < UFS_UPIU_MAX_UNIT_NUM_ID; > > +} > > + > > +static bool > > +ufshpb_is_pinned_region(struct ufshpb_lu *hpb, int rgn_idx) > > +{ > > + if (hpb->lu_pinned_end != PINNED_NOT_SET && > > + rgn_idx >= hpb->lu_pinned_start && > > + rgn_idx <= hpb->lu_pinned_end) > > + return true; > > + > > + return false; > > +} > > + > > +static void ufshpb_kick_map_work(struct ufshpb_lu *hpb) > > +{ > > + bool ret = false; > > + unsigned long flags; > > + > > + if (ufshpb_get_state(hpb) != HPB_PRESENT) > > + return; > > + > > + spin_lock_irqsave(&hpb->rsp_list_lock, flags); > > + if (!list_empty(&hpb->lh_inact_rgn) || > > !list_empty(&hpb->lh_act_srgn)) > > + ret = true; > > + spin_unlock_irqrestore(&hpb->rsp_list_lock, flags); > > + > > + if (ret) > > + queue_work(ufshpb_wq, &hpb->map_work); > > +} > > + > > +static bool ufshpb_is_hpb_rsp_valid(struct ufs_hba *hba, > > + struct ufshcd_lrb *lrbp, > > + struct utp_hpb_rsp *rsp_field) > > +{ > > + /* Check HPB_UPDATE_ALERT */ > > + if (!(lrbp->ucd_rsp_ptr->header.dword_2 & > > + UPIU_HEADER_DWORD(0, 2, 0, 0))) > > + return false; > > + > > + if (be16_to_cpu(rsp_field->sense_data_len) != DEV_SENSE_SEG_LEN || > > + rsp_field->desc_type != DEV_DES_TYPE || > > + rsp_field->additional_len != DEV_ADDITIONAL_LEN || > > + rsp_field->active_rgn_cnt > MAX_ACTIVE_NUM || > > + rsp_field->inactive_rgn_cnt > MAX_INACTIVE_NUM || > > + rsp_field->hpb_op == HPB_RSP_NONE || > > + (rsp_field->hpb_op == HPB_RSP_REQ_REGION_UPDATE && > > + !rsp_field->active_rgn_cnt && !rsp_field->inactive_rgn_cnt)) > > + return false; > > + > > + if (!ufshpb_is_general_lun(rsp_field->lun)) { > > + dev_warn(hba->dev, "ufshpb: lun(%d) not supported\n", > > + lrbp->lun); > > + return false; > > + } > > + > > + return true; > > +} > > + > > +static struct ufshpb_req *ufshpb_get_map_req(struct ufshpb_lu *hpb, > > + struct ufshpb_subregion *srgn) > > +{ > > + struct ufshpb_req *map_req; > > + struct request *req; > > + struct bio *bio; > > + int retries = HPB_MAP_REQ_RETRIES; > > + > > + map_req = kmem_cache_alloc(hpb->map_req_cache, GFP_KERNEL); > > + if (!map_req) > > + return NULL; > > + > > +retry: > > + req = blk_get_request(hpb->sdev_ufs_lu->request_queue, > > + REQ_OP_SCSI_IN, BLK_MQ_REQ_NOWAIT); > > + > > + if ((PTR_ERR(req) == -EWOULDBLOCK) && (--retries > 0)) { > > + usleep_range(3000, 3100); > > + goto retry; > > + } > > + > > + if (IS_ERR(req)) > > + goto free_map_req; > > + > > + bio = bio_alloc(GFP_KERNEL, hpb->pages_per_srgn); > > + if (!bio) { > > + blk_put_request(req); > > + goto free_map_req; > > + } > > + > > + map_req->hpb = hpb; > > + map_req->req = req; > > + map_req->bio = bio; > > + > > + map_req->rgn_idx = srgn->rgn_idx; > > + map_req->srgn_idx = srgn->srgn_idx; > > + map_req->mctx = srgn->mctx; > > + > > + return map_req; > > + > > +free_map_req: > > + kmem_cache_free(hpb->map_req_cache, map_req); > > + return NULL; > > +} > > + > > +static void ufshpb_put_map_req(struct ufshpb_lu *hpb, > > + struct ufshpb_req *map_req) > > +{ > > + bio_put(map_req->bio); > > + blk_put_request(map_req->req); > > + kmem_cache_free(hpb->map_req_cache, map_req); > > +} > > + > > +static int ufshpb_clear_dirty_bitmap(struct ufshpb_lu *hpb, > > + struct ufshpb_subregion *srgn) > > +{ > > + u32 num_entries = hpb->entries_per_srgn; > > + > > + if (!srgn->mctx) { > > + dev_err(&hpb->sdev_ufs_lu->sdev_dev, > > + "no mctx in region %d subregion %d.\n", > > + srgn->rgn_idx, srgn->srgn_idx); > > + return -1; > > + } > > + > > + if (unlikely(srgn->is_last)) > > + num_entries = hpb->last_srgn_entries; > > + > > + bitmap_zero(srgn->mctx->ppn_dirty, num_entries); > > + return 0; > > +} > > + > > +static void ufshpb_update_active_info(struct ufshpb_lu *hpb, int > > rgn_idx, > > + int srgn_idx) > > +{ > > + struct ufshpb_region *rgn; > > + struct ufshpb_subregion *srgn; > > + > > + rgn = hpb->rgn_tbl + rgn_idx; > > + srgn = rgn->srgn_tbl + srgn_idx; > > + > > + list_del_init(&rgn->list_inact_rgn); > > + > > + if (list_empty(&srgn->list_act_srgn)) > > + list_add_tail(&srgn->list_act_srgn, &hpb->lh_act_srgn); > > +} > > + > > +static void ufshpb_update_inactive_info(struct ufshpb_lu *hpb, int > > rgn_idx) > > +{ > > + struct ufshpb_region *rgn; > > + struct ufshpb_subregion *srgn; > > + int srgn_idx; > > + > > + rgn = hpb->rgn_tbl + rgn_idx; > > + > > + for_each_sub_region(rgn, srgn_idx, srgn) > > + list_del_init(&srgn->list_act_srgn); > > + > > + if (list_empty(&rgn->list_inact_rgn)) > > + list_add_tail(&rgn->list_inact_rgn, &hpb->lh_inact_rgn); > > +} > > + > > +static void ufshpb_activate_subregion(struct ufshpb_lu *hpb, > > + struct ufshpb_subregion *srgn) > > +{ > > + struct ufshpb_region *rgn; > > + > > + /* > > + * If there is no mctx in subregion > > + * after I/O progress for HPB_READ_BUFFER, the region to which the > > + * subregion belongs was evicted. > > + * Make sure the region must not evict in I/O progress > > + */ > > + if (!srgn->mctx) { > > + dev_err(&hpb->sdev_ufs_lu->sdev_dev, > > + "no mctx in region %d subregion %d.\n", > > + srgn->rgn_idx, srgn->srgn_idx); > > + srgn->srgn_state = HPB_SRGN_INVALID; > > + return; > > + } > > + > > + rgn = hpb->rgn_tbl + srgn->rgn_idx; > > + > > + if (unlikely(rgn->rgn_state == HPB_RGN_INACTIVE)) { > > + dev_err(&hpb->sdev_ufs_lu->sdev_dev, > > + "region %d subregion %d evicted\n", > > + srgn->rgn_idx, srgn->srgn_idx); > > + srgn->srgn_state = HPB_SRGN_INVALID; > > + return; > > + } > > + srgn->srgn_state = HPB_SRGN_VALID; > > +} > > + > > +static void ufshpb_map_req_compl_fn(struct request *req, blk_status_t > > error) > > +{ > > + struct ufshpb_req *map_req = (struct ufshpb_req *) req->end_io_data; > > + struct ufshpb_lu *hpb = map_req->hpb; > > + struct ufshpb_subregion *srgn; > > + unsigned long flags; > > + > > + srgn = hpb->rgn_tbl[map_req->rgn_idx].srgn_tbl + > > + map_req->srgn_idx; > > + > > + ufshpb_clear_dirty_bitmap(hpb, srgn); > > + spin_lock_irqsave(&hpb->rgn_state_lock, flags); > > + ufshpb_activate_subregion(hpb, srgn); > > + spin_unlock_irqrestore(&hpb->rgn_state_lock, flags); > > + > > + ufshpb_put_map_req(map_req->hpb, map_req); > > +} > > + > > +static void ufshpb_set_read_buf_cmd(unsigned char *cdb, int rgn_idx, > > + int srgn_idx, int srgn_mem_size) > > +{ > > + cdb[0] = UFSHPB_READ_BUFFER; > > + cdb[1] = UFSHPB_READ_BUFFER_ID; > > + > > + put_unaligned_be16(rgn_idx, &cdb[2]); > > + put_unaligned_be16(srgn_idx, &cdb[4]); > > + put_unaligned_be24(srgn_mem_size, &cdb[6]); > > + > > + cdb[9] = 0x00; > > +} > > + > > +static int ufshpb_execute_map_req(struct ufshpb_lu *hpb, > > + struct ufshpb_req *map_req, bool last) > > +{ > > + struct request_queue *q; > > + struct request *req; > > + struct scsi_request *rq; > > + int mem_size = hpb->srgn_mem_size; > > + int ret = 0; > > + int i; > > + > > + q = hpb->sdev_ufs_lu->request_queue; > > + for (i = 0; i < hpb->pages_per_srgn; i++) { > > + ret = bio_add_pc_page(q, map_req->bio, map_req->mctx->m_page[i], > > + PAGE_SIZE, 0); > > + if (ret != PAGE_SIZE) { > > + dev_err(&hpb->sdev_ufs_lu->sdev_dev, > > + "bio_add_pc_page fail %d - %d\n", > > + map_req->rgn_idx, map_req->srgn_idx); > > + return ret; > > + } > > + } > > + > > + req = map_req->req; > > + > > + blk_rq_append_bio(req, &map_req->bio); > > + > > + req->end_io_data = map_req; > > + > > + rq = scsi_req(req); > > + > > + if (unlikely(last)) > > + mem_size = hpb->last_srgn_entries * HPB_ENTRY_SIZE; > > + > > + ufshpb_set_read_buf_cmd(rq->cmd, map_req->rgn_idx, > > + map_req->srgn_idx, mem_size); > > + rq->cmd_len = HPB_READ_BUFFER_CMD_LENGTH; > > + > > + blk_execute_rq_nowait(q, NULL, req, 1, ufshpb_map_req_compl_fn); > > + > > + hpb->stats.map_req_cnt++; > > + return 0; > > +} > > + > > +static struct ufshpb_map_ctx *ufshpb_get_map_ctx(struct ufshpb_lu > > *hpb, > > + bool last) > > +{ > > + struct ufshpb_map_ctx *mctx; > > + u32 num_entries = hpb->entries_per_srgn; > > + int i, j; > > + > > + mctx = mempool_alloc(ufshpb_mctx_pool, GFP_KERNEL); > > + if (!mctx) > > + return NULL; > > + > > + mctx->m_page = kmem_cache_alloc(hpb->m_page_cache, GFP_KERNEL); > > + if (!mctx->m_page) > > + goto release_mctx; > > + > > + if (unlikely(last)) > > + num_entries = hpb->last_srgn_entries; > > + > > + mctx->ppn_dirty = bitmap_zalloc(num_entries, GFP_KERNEL); > > + if (!mctx->ppn_dirty) > > + goto release_m_page; > > + > > + for (i = 0; i < hpb->pages_per_srgn; i++) { > > + mctx->m_page[i] = mempool_alloc(ufshpb_page_pool, GFP_KERNEL); > > + if (!mctx->m_page[i]) { > > + for (j = 0; j < i; j++) > > + mempool_free(mctx->m_page[j], ufshpb_page_pool); > > + goto release_ppn_dirty; > > + } > > + clear_page(page_address(mctx->m_page[i])); > > + } > > + > > + return mctx; > > + > > +release_ppn_dirty: > > + bitmap_free(mctx->ppn_dirty); > > +release_m_page: > > + kmem_cache_free(hpb->m_page_cache, mctx->m_page); > > +release_mctx: > > + mempool_free(mctx, ufshpb_mctx_pool); > > + return NULL; > > +} > > + > > +static void ufshpb_put_map_ctx(struct ufshpb_lu *hpb, > > + struct ufshpb_map_ctx *mctx) > > +{ > > + int i; > > + > > + for (i = 0; i < hpb->pages_per_srgn; i++) > > + mempool_free(mctx->m_page[i], ufshpb_page_pool); > > + > > + bitmap_free(mctx->ppn_dirty); > > + kmem_cache_free(hpb->m_page_cache, mctx->m_page); > > + mempool_free(mctx, ufshpb_mctx_pool); > > +} > > + > > +static int ufshpb_check_srgns_issue_state(struct ufshpb_lu *hpb, > > + struct ufshpb_region *rgn) > > +{ > > + struct ufshpb_subregion *srgn; > > + int srgn_idx; > > + > > + for_each_sub_region(rgn, srgn_idx, srgn) > > + if (srgn->srgn_state == HPB_SRGN_ISSUED) > > + return -EPERM; > > + > > + return 0; > > +} > > + > > +static void ufshpb_add_lru_info(struct victim_select_info *lru_info, > > + struct ufshpb_region *rgn) > > +{ > > + rgn->rgn_state = HPB_RGN_ACTIVE; > > + list_add_tail(&rgn->list_lru_rgn, &lru_info->lh_lru_rgn); > > + atomic_inc(&lru_info->active_cnt); > > +} > > + > > +static void ufshpb_hit_lru_info(struct victim_select_info *lru_info, > > + struct ufshpb_region *rgn) > > +{ > > + list_move_tail(&rgn->list_lru_rgn, &lru_info->lh_lru_rgn); > > +} > > + > > +static struct ufshpb_region *ufshpb_victim_lru_info(struct ufshpb_lu > > *hpb) > > +{ > > + struct victim_select_info *lru_info = &hpb->lru_info; > > + struct ufshpb_region *rgn, *victim_rgn = NULL; > > + > > + list_for_each_entry(rgn, &lru_info->lh_lru_rgn, list_lru_rgn) { > > + if (!rgn) { > > + dev_err(&hpb->sdev_ufs_lu->sdev_dev, > > + "%s: no region allocated\n", > > + __func__); > > + return NULL; > > + } > > + if (ufshpb_check_srgns_issue_state(hpb, rgn)) > > + continue; > > + > > + victim_rgn = rgn; > > + break; > > + } > > + > > + return victim_rgn; > > +} > > + > > +static void ufshpb_cleanup_lru_info(struct victim_select_info > > *lru_info, > > + struct ufshpb_region *rgn) > > +{ > > + list_del_init(&rgn->list_lru_rgn); > > + rgn->rgn_state = HPB_RGN_INACTIVE; > > + atomic_dec(&lru_info->active_cnt); > > +} > > + > > +static void ufshpb_purge_active_subregion(struct ufshpb_lu *hpb, > > + struct ufshpb_subregion *srgn) > > +{ > > + if (srgn->srgn_state != HPB_SRGN_UNUSED) { > > + ufshpb_put_map_ctx(hpb, srgn->mctx); > > + srgn->srgn_state = HPB_SRGN_UNUSED; > > + srgn->mctx = NULL; > > + } > > +} > > + > > +static void __ufshpb_evict_region(struct ufshpb_lu *hpb, > > + struct ufshpb_region *rgn) > > +{ > > + struct victim_select_info *lru_info; > > + struct ufshpb_subregion *srgn; > > + int srgn_idx; > > + > > + lru_info = &hpb->lru_info; > > + > > + dev_dbg(&hpb->sdev_ufs_lu->sdev_dev, "evict region %d\n", > > rgn->rgn_idx); > > + > > + ufshpb_cleanup_lru_info(lru_info, rgn); > > + > > + for_each_sub_region(rgn, srgn_idx, srgn) > > + ufshpb_purge_active_subregion(hpb, srgn); > > +} > > + > > +static int ufshpb_evict_region(struct ufshpb_lu *hpb, struct > > ufshpb_region *rgn) > > +{ > > + unsigned long flags; > > + int ret = 0; > > + > > + spin_lock_irqsave(&hpb->rgn_state_lock, flags); > > + if (rgn->rgn_state == HPB_RGN_PINNED) { > > + dev_warn(&hpb->sdev_ufs_lu->sdev_dev, > > + "pinned region cannot drop-out. region %d\n", > > + rgn->rgn_idx); > > + goto out; > > + } > > + if (!list_empty(&rgn->list_lru_rgn)) { > > + if (ufshpb_check_srgns_issue_state(hpb, rgn)) { > > + ret = -EBUSY; > > + goto out; > > + } > > + > > + __ufshpb_evict_region(hpb, rgn); > > + } > > +out: > > + spin_unlock_irqrestore(&hpb->rgn_state_lock, flags); > > + return ret; > > +} > > + > > +static int ufshpb_issue_map_req(struct ufshpb_lu *hpb, > > + struct ufshpb_region *rgn, > > + struct ufshpb_subregion *srgn) > > +{ > > + struct ufshpb_req *map_req; > > + unsigned long flags; > > + int ret; > > + int err = -EAGAIN; > > + bool alloc_required = false; > > + enum HPB_SRGN_STATE state = HPB_SRGN_INVALID; > > + > > + spin_lock_irqsave(&hpb->rgn_state_lock, flags); > > + > > + if (ufshpb_get_state(hpb) != HPB_PRESENT) { > > + dev_notice(&hpb->sdev_ufs_lu->sdev_dev, > > + "%s: ufshpb state is not PRESENT\n", __func__); > > + goto unlock_out; > > + } > > + > > + if ((rgn->rgn_state == HPB_RGN_INACTIVE) && > > + (srgn->srgn_state == HPB_SRGN_INVALID)) { > > + err = 0; > > + goto unlock_out; > > + } > > + > > + if (srgn->srgn_state == HPB_SRGN_UNUSED) > > + alloc_required = true; > > + > > + /* > > + * If the subregion is already ISSUED state, > > + * a specific event (e.g., GC or wear-leveling, etc.) occurs in > > + * the device and HPB response for map loading is received. > > + * In this case, after finishing the HPB_READ_BUFFER, > > + * the next HPB_READ_BUFFER is performed again to obtain the latest > > + * map data. > > + */ > > + if (srgn->srgn_state == HPB_SRGN_ISSUED) > > + goto unlock_out; > > + > > + srgn->srgn_state = HPB_SRGN_ISSUED; > > + spin_unlock_irqrestore(&hpb->rgn_state_lock, flags); > > + > > + if (alloc_required) { > > + if (srgn->mctx) { > > Can this really happen? > > alloc_required is true if srgn->srgn_state == HPB_SRGN_UNUSED. > ufshpb_put_map_ctx() is called and srgn->srgn_state is given > HPB_SRGN_UNUSED together in ufshpb_purge_active_subregion(). > Do I miss anything? It is not normal case, but it is for preventing memory leak on abnormal case. Thanks, Daejun
On 2021-03-12 09:48, Daejun Park wrote: >> > This is a patch for managing L2P map in HPB module. >> > >> > The HPB divides logical addresses into several regions. A region >> > consists >> > of several sub-regions. The sub-region is a basic unit where L2P >> > mapping is >> > managed. The driver loads L2P mapping data of each sub-region. The >> > loaded >> > sub-region is called active-state. The HPB driver unloads L2P mapping >> > data >> > as region unit. The unloaded region is called inactive-state. >> > >> > Sub-region/region candidates to be loaded and unloaded are delivered >> > from >> > the UFS device. The UFS device delivers the recommended active >> > sub-region >> > and inactivate region to the driver using sensedata. >> > The HPB module performs L2P mapping management on the host through the >> > delivered information. >> > >> > A pinned region is a pre-set regions on the UFS device that is always >> > activate-state. >> > >> > The data structure for map data request and L2P map uses mempool API, >> > minimizing allocation overhead while avoiding static allocation. >> > >> > The mininum size of the memory pool used in the HPB is implemented >> > as a module parameter, so that it can be configurable by the user. >> > >> > To gurantee a minimum memory pool size of 4MB: >> > ufshpb_host_map_kbytes=4096 >> > >> > The map_work manages active/inactive by 2 "to-do" lists. >> > Each hpb lun maintains 2 "to-do" lists: >> > hpb->lh_inact_rgn - regions to be inactivated, and >> > hpb->lh_act_srgn - subregions to be activated >> > Those lists are maintained on IO completion. >> > >> > Reviewed-by: Bart Van Assche <bvanassche@acm.org> >> > Reviewed-by: Can Guo <cang@codeaurora.org> >> > Acked-by: Avri Altman <Avri.Altman@wdc.com> >> > Tested-by: Bean Huo <beanhuo@micron.com> >> > Signed-off-by: Daejun Park <daejun7.park@samsung.com> >> > --- >> > drivers/scsi/ufs/ufs.h | 36 ++ >> > drivers/scsi/ufs/ufshcd.c | 4 + >> > drivers/scsi/ufs/ufshpb.c | 1091 ++++++++++++++++++++++++++++++++++++- >> > drivers/scsi/ufs/ufshpb.h | 65 +++ >> > 4 files changed, 1181 insertions(+), 15 deletions(-) >> > >> > diff --git a/drivers/scsi/ufs/ufs.h b/drivers/scsi/ufs/ufs.h >> > index 65563635e20e..957763db1006 100644 >> > --- a/drivers/scsi/ufs/ufs.h >> > +++ b/drivers/scsi/ufs/ufs.h >> > @@ -472,6 +472,41 @@ struct utp_cmd_rsp { >> > u8 sense_data[UFS_SENSE_SIZE]; >> > }; >> > >> > +struct ufshpb_active_field { >> > + __be16 active_rgn; >> > + __be16 active_srgn; >> > +}; >> > +#define HPB_ACT_FIELD_SIZE 4 >> > + >> > +/** >> > + * struct utp_hpb_rsp - Response UPIU structure >> > + * @residual_transfer_count: Residual transfer count DW-3 >> > + * @reserved1: Reserved double words DW-4 to DW-7 >> > + * @sense_data_len: Sense data length DW-8 U16 >> > + * @desc_type: Descriptor type of sense data >> > + * @additional_len: Additional length of sense data >> > + * @hpb_op: HPB operation type >> > + * @lun: LUN of response UPIU >> > + * @active_rgn_cnt: Active region count >> > + * @inactive_rgn_cnt: Inactive region count >> > + * @hpb_active_field: Recommended to read HPB region and subregion >> > + * @hpb_inactive_field: To be inactivated HPB region and subregion >> > + */ >> > +struct utp_hpb_rsp { >> > + __be32 residual_transfer_count; >> > + __be32 reserved1[4]; >> > + __be16 sense_data_len; >> > + u8 desc_type; >> > + u8 additional_len; >> > + u8 hpb_op; >> > + u8 lun; >> > + u8 active_rgn_cnt; >> > + u8 inactive_rgn_cnt; >> > + struct ufshpb_active_field hpb_active_field[2]; >> > + __be16 hpb_inactive_field[2]; >> > +}; >> > +#define UTP_HPB_RSP_SIZE 40 >> > + >> > /** >> > * struct utp_upiu_rsp - general upiu response structure >> > * @header: UPIU header structure DW-0 to DW-2 >> > @@ -482,6 +517,7 @@ struct utp_upiu_rsp { >> > struct utp_upiu_header header; >> > union { >> > struct utp_cmd_rsp sr; >> > + struct utp_hpb_rsp hr; >> > struct utp_upiu_query qr; >> > }; >> > }; >> > diff --git a/drivers/scsi/ufs/ufshcd.c b/drivers/scsi/ufs/ufshcd.c >> > index 49b3d5d24fa6..5852ff44c3cc 100644 >> > --- a/drivers/scsi/ufs/ufshcd.c >> > +++ b/drivers/scsi/ufs/ufshcd.c >> > @@ -5021,6 +5021,9 @@ ufshcd_transfer_rsp_status(struct ufs_hba *hba, >> > struct ufshcd_lrb *lrbp) >> > */ >> > pm_runtime_get_noresume(hba->dev); >> > } >> > + >> > + if (scsi_status == SAM_STAT_GOOD) >> > + ufshpb_rsp_upiu(hba, lrbp); >> > break; >> > case UPIU_TRANSACTION_REJECT_UPIU: >> > /* TODO: handle Reject UPIU Response */ >> > @@ -9221,6 +9224,7 @@ EXPORT_SYMBOL(ufshcd_shutdown); >> > void ufshcd_remove(struct ufs_hba *hba) >> > { >> > ufs_bsg_remove(hba); >> > + ufshpb_remove(hba); >> > ufs_sysfs_remove_nodes(hba->dev); >> > blk_cleanup_queue(hba->tmf_queue); >> > blk_mq_free_tag_set(&hba->tmf_tag_set); >> > diff --git a/drivers/scsi/ufs/ufshpb.c b/drivers/scsi/ufs/ufshpb.c >> > index 1a72f6541510..8abadb0e010a 100644 >> > --- a/drivers/scsi/ufs/ufshpb.c >> > +++ b/drivers/scsi/ufs/ufshpb.c >> > @@ -16,6 +16,16 @@ >> > #include "ufshpb.h" >> > #include "../sd.h" >> > >> > +/* memory management */ >> > +static struct kmem_cache *ufshpb_mctx_cache; >> > +static mempool_t *ufshpb_mctx_pool; >> > +static mempool_t *ufshpb_page_pool; >> > +/* A cache size of 2MB can cache ppn in the 1GB range. */ >> > +static unsigned int ufshpb_host_map_kbytes = 2048; >> > +static int tot_active_srgn_pages; >> > + >> > +static struct workqueue_struct *ufshpb_wq; >> > + >> > bool ufshpb_is_allowed(struct ufs_hba *hba) >> > { >> > return !(hba->ufshpb_dev.hpb_disabled); >> > @@ -36,14 +46,892 @@ static void ufshpb_set_state(struct ufshpb_lu >> > *hpb, int state) >> > atomic_set(&hpb->hpb_state, state); >> > } >> > >> > +static bool ufshpb_is_general_lun(int lun) >> > +{ >> > + return lun < UFS_UPIU_MAX_UNIT_NUM_ID; >> > +} >> > + >> > +static bool >> > +ufshpb_is_pinned_region(struct ufshpb_lu *hpb, int rgn_idx) >> > +{ >> > + if (hpb->lu_pinned_end != PINNED_NOT_SET && >> > + rgn_idx >= hpb->lu_pinned_start && >> > + rgn_idx <= hpb->lu_pinned_end) >> > + return true; >> > + >> > + return false; >> > +} >> > + >> > +static void ufshpb_kick_map_work(struct ufshpb_lu *hpb) >> > +{ >> > + bool ret = false; >> > + unsigned long flags; >> > + >> > + if (ufshpb_get_state(hpb) != HPB_PRESENT) >> > + return; >> > + >> > + spin_lock_irqsave(&hpb->rsp_list_lock, flags); >> > + if (!list_empty(&hpb->lh_inact_rgn) || >> > !list_empty(&hpb->lh_act_srgn)) >> > + ret = true; >> > + spin_unlock_irqrestore(&hpb->rsp_list_lock, flags); >> > + >> > + if (ret) >> > + queue_work(ufshpb_wq, &hpb->map_work); >> > +} >> > + >> > +static bool ufshpb_is_hpb_rsp_valid(struct ufs_hba *hba, >> > + struct ufshcd_lrb *lrbp, >> > + struct utp_hpb_rsp *rsp_field) >> > +{ >> > + /* Check HPB_UPDATE_ALERT */ >> > + if (!(lrbp->ucd_rsp_ptr->header.dword_2 & >> > + UPIU_HEADER_DWORD(0, 2, 0, 0))) >> > + return false; >> > + >> > + if (be16_to_cpu(rsp_field->sense_data_len) != DEV_SENSE_SEG_LEN || >> > + rsp_field->desc_type != DEV_DES_TYPE || >> > + rsp_field->additional_len != DEV_ADDITIONAL_LEN || >> > + rsp_field->active_rgn_cnt > MAX_ACTIVE_NUM || >> > + rsp_field->inactive_rgn_cnt > MAX_INACTIVE_NUM || >> > + rsp_field->hpb_op == HPB_RSP_NONE || >> > + (rsp_field->hpb_op == HPB_RSP_REQ_REGION_UPDATE && >> > + !rsp_field->active_rgn_cnt && !rsp_field->inactive_rgn_cnt)) >> > + return false; >> > + >> > + if (!ufshpb_is_general_lun(rsp_field->lun)) { >> > + dev_warn(hba->dev, "ufshpb: lun(%d) not supported\n", >> > + lrbp->lun); >> > + return false; >> > + } >> > + >> > + return true; >> > +} >> > + >> > +static struct ufshpb_req *ufshpb_get_map_req(struct ufshpb_lu *hpb, >> > + struct ufshpb_subregion *srgn) >> > +{ >> > + struct ufshpb_req *map_req; >> > + struct request *req; >> > + struct bio *bio; >> > + int retries = HPB_MAP_REQ_RETRIES; >> > + >> > + map_req = kmem_cache_alloc(hpb->map_req_cache, GFP_KERNEL); >> > + if (!map_req) >> > + return NULL; >> > + >> > +retry: >> > + req = blk_get_request(hpb->sdev_ufs_lu->request_queue, >> > + REQ_OP_SCSI_IN, BLK_MQ_REQ_NOWAIT); >> > + >> > + if ((PTR_ERR(req) == -EWOULDBLOCK) && (--retries > 0)) { >> > + usleep_range(3000, 3100); >> > + goto retry; >> > + } >> > + >> > + if (IS_ERR(req)) >> > + goto free_map_req; >> > + >> > + bio = bio_alloc(GFP_KERNEL, hpb->pages_per_srgn); >> > + if (!bio) { >> > + blk_put_request(req); >> > + goto free_map_req; >> > + } >> > + >> > + map_req->hpb = hpb; >> > + map_req->req = req; >> > + map_req->bio = bio; >> > + >> > + map_req->rgn_idx = srgn->rgn_idx; >> > + map_req->srgn_idx = srgn->srgn_idx; >> > + map_req->mctx = srgn->mctx; >> > + >> > + return map_req; >> > + >> > +free_map_req: >> > + kmem_cache_free(hpb->map_req_cache, map_req); >> > + return NULL; >> > +} >> > + >> > +static void ufshpb_put_map_req(struct ufshpb_lu *hpb, >> > + struct ufshpb_req *map_req) >> > +{ >> > + bio_put(map_req->bio); >> > + blk_put_request(map_req->req); >> > + kmem_cache_free(hpb->map_req_cache, map_req); >> > +} >> > + >> > +static int ufshpb_clear_dirty_bitmap(struct ufshpb_lu *hpb, >> > + struct ufshpb_subregion *srgn) >> > +{ >> > + u32 num_entries = hpb->entries_per_srgn; >> > + >> > + if (!srgn->mctx) { >> > + dev_err(&hpb->sdev_ufs_lu->sdev_dev, >> > + "no mctx in region %d subregion %d.\n", >> > + srgn->rgn_idx, srgn->srgn_idx); >> > + return -1; >> > + } >> > + >> > + if (unlikely(srgn->is_last)) >> > + num_entries = hpb->last_srgn_entries; >> > + >> > + bitmap_zero(srgn->mctx->ppn_dirty, num_entries); >> > + return 0; >> > +} >> > + >> > +static void ufshpb_update_active_info(struct ufshpb_lu *hpb, int >> > rgn_idx, >> > + int srgn_idx) >> > +{ >> > + struct ufshpb_region *rgn; >> > + struct ufshpb_subregion *srgn; >> > + >> > + rgn = hpb->rgn_tbl + rgn_idx; >> > + srgn = rgn->srgn_tbl + srgn_idx; >> > + >> > + list_del_init(&rgn->list_inact_rgn); >> > + >> > + if (list_empty(&srgn->list_act_srgn)) >> > + list_add_tail(&srgn->list_act_srgn, &hpb->lh_act_srgn); >> > +} >> > + >> > +static void ufshpb_update_inactive_info(struct ufshpb_lu *hpb, int >> > rgn_idx) >> > +{ >> > + struct ufshpb_region *rgn; >> > + struct ufshpb_subregion *srgn; >> > + int srgn_idx; >> > + >> > + rgn = hpb->rgn_tbl + rgn_idx; >> > + >> > + for_each_sub_region(rgn, srgn_idx, srgn) >> > + list_del_init(&srgn->list_act_srgn); >> > + >> > + if (list_empty(&rgn->list_inact_rgn)) >> > + list_add_tail(&rgn->list_inact_rgn, &hpb->lh_inact_rgn); >> > +} >> > + >> > +static void ufshpb_activate_subregion(struct ufshpb_lu *hpb, >> > + struct ufshpb_subregion *srgn) >> > +{ >> > + struct ufshpb_region *rgn; >> > + >> > + /* >> > + * If there is no mctx in subregion >> > + * after I/O progress for HPB_READ_BUFFER, the region to which the >> > + * subregion belongs was evicted. >> > + * Make sure the region must not evict in I/O progress >> > + */ >> > + if (!srgn->mctx) { >> > + dev_err(&hpb->sdev_ufs_lu->sdev_dev, >> > + "no mctx in region %d subregion %d.\n", >> > + srgn->rgn_idx, srgn->srgn_idx); >> > + srgn->srgn_state = HPB_SRGN_INVALID; >> > + return; >> > + } >> > + >> > + rgn = hpb->rgn_tbl + srgn->rgn_idx; >> > + >> > + if (unlikely(rgn->rgn_state == HPB_RGN_INACTIVE)) { >> > + dev_err(&hpb->sdev_ufs_lu->sdev_dev, >> > + "region %d subregion %d evicted\n", >> > + srgn->rgn_idx, srgn->srgn_idx); >> > + srgn->srgn_state = HPB_SRGN_INVALID; >> > + return; >> > + } >> > + srgn->srgn_state = HPB_SRGN_VALID; >> > +} >> > + >> > +static void ufshpb_map_req_compl_fn(struct request *req, blk_status_t >> > error) >> > +{ >> > + struct ufshpb_req *map_req = (struct ufshpb_req *) req->end_io_data; >> > + struct ufshpb_lu *hpb = map_req->hpb; >> > + struct ufshpb_subregion *srgn; >> > + unsigned long flags; >> > + >> > + srgn = hpb->rgn_tbl[map_req->rgn_idx].srgn_tbl + >> > + map_req->srgn_idx; >> > + >> > + ufshpb_clear_dirty_bitmap(hpb, srgn); >> > + spin_lock_irqsave(&hpb->rgn_state_lock, flags); >> > + ufshpb_activate_subregion(hpb, srgn); >> > + spin_unlock_irqrestore(&hpb->rgn_state_lock, flags); >> > + >> > + ufshpb_put_map_req(map_req->hpb, map_req); >> > +} >> > + >> > +static void ufshpb_set_read_buf_cmd(unsigned char *cdb, int rgn_idx, >> > + int srgn_idx, int srgn_mem_size) >> > +{ >> > + cdb[0] = UFSHPB_READ_BUFFER; >> > + cdb[1] = UFSHPB_READ_BUFFER_ID; >> > + >> > + put_unaligned_be16(rgn_idx, &cdb[2]); >> > + put_unaligned_be16(srgn_idx, &cdb[4]); >> > + put_unaligned_be24(srgn_mem_size, &cdb[6]); >> > + >> > + cdb[9] = 0x00; >> > +} >> > + >> > +static int ufshpb_execute_map_req(struct ufshpb_lu *hpb, >> > + struct ufshpb_req *map_req, bool last) >> > +{ >> > + struct request_queue *q; >> > + struct request *req; >> > + struct scsi_request *rq; >> > + int mem_size = hpb->srgn_mem_size; >> > + int ret = 0; >> > + int i; >> > + >> > + q = hpb->sdev_ufs_lu->request_queue; >> > + for (i = 0; i < hpb->pages_per_srgn; i++) { >> > + ret = bio_add_pc_page(q, map_req->bio, map_req->mctx->m_page[i], >> > + PAGE_SIZE, 0); >> > + if (ret != PAGE_SIZE) { >> > + dev_err(&hpb->sdev_ufs_lu->sdev_dev, >> > + "bio_add_pc_page fail %d - %d\n", >> > + map_req->rgn_idx, map_req->srgn_idx); >> > + return ret; >> > + } >> > + } >> > + >> > + req = map_req->req; >> > + >> > + blk_rq_append_bio(req, &map_req->bio); >> > + >> > + req->end_io_data = map_req; >> > + >> > + rq = scsi_req(req); >> > + >> > + if (unlikely(last)) >> > + mem_size = hpb->last_srgn_entries * HPB_ENTRY_SIZE; >> > + >> > + ufshpb_set_read_buf_cmd(rq->cmd, map_req->rgn_idx, >> > + map_req->srgn_idx, mem_size); >> > + rq->cmd_len = HPB_READ_BUFFER_CMD_LENGTH; >> > + >> > + blk_execute_rq_nowait(q, NULL, req, 1, ufshpb_map_req_compl_fn); >> > + >> > + hpb->stats.map_req_cnt++; >> > + return 0; >> > +} >> > + >> > +static struct ufshpb_map_ctx *ufshpb_get_map_ctx(struct ufshpb_lu >> > *hpb, >> > + bool last) >> > +{ >> > + struct ufshpb_map_ctx *mctx; >> > + u32 num_entries = hpb->entries_per_srgn; >> > + int i, j; >> > + >> > + mctx = mempool_alloc(ufshpb_mctx_pool, GFP_KERNEL); >> > + if (!mctx) >> > + return NULL; >> > + >> > + mctx->m_page = kmem_cache_alloc(hpb->m_page_cache, GFP_KERNEL); >> > + if (!mctx->m_page) >> > + goto release_mctx; >> > + >> > + if (unlikely(last)) >> > + num_entries = hpb->last_srgn_entries; >> > + >> > + mctx->ppn_dirty = bitmap_zalloc(num_entries, GFP_KERNEL); >> > + if (!mctx->ppn_dirty) >> > + goto release_m_page; >> > + >> > + for (i = 0; i < hpb->pages_per_srgn; i++) { >> > + mctx->m_page[i] = mempool_alloc(ufshpb_page_pool, GFP_KERNEL); >> > + if (!mctx->m_page[i]) { >> > + for (j = 0; j < i; j++) >> > + mempool_free(mctx->m_page[j], ufshpb_page_pool); >> > + goto release_ppn_dirty; >> > + } >> > + clear_page(page_address(mctx->m_page[i])); >> > + } >> > + >> > + return mctx; >> > + >> > +release_ppn_dirty: >> > + bitmap_free(mctx->ppn_dirty); >> > +release_m_page: >> > + kmem_cache_free(hpb->m_page_cache, mctx->m_page); >> > +release_mctx: >> > + mempool_free(mctx, ufshpb_mctx_pool); >> > + return NULL; >> > +} >> > + >> > +static void ufshpb_put_map_ctx(struct ufshpb_lu *hpb, >> > + struct ufshpb_map_ctx *mctx) >> > +{ >> > + int i; >> > + >> > + for (i = 0; i < hpb->pages_per_srgn; i++) >> > + mempool_free(mctx->m_page[i], ufshpb_page_pool); >> > + >> > + bitmap_free(mctx->ppn_dirty); >> > + kmem_cache_free(hpb->m_page_cache, mctx->m_page); >> > + mempool_free(mctx, ufshpb_mctx_pool); >> > +} >> > + >> > +static int ufshpb_check_srgns_issue_state(struct ufshpb_lu *hpb, >> > + struct ufshpb_region *rgn) >> > +{ >> > + struct ufshpb_subregion *srgn; >> > + int srgn_idx; >> > + >> > + for_each_sub_region(rgn, srgn_idx, srgn) >> > + if (srgn->srgn_state == HPB_SRGN_ISSUED) >> > + return -EPERM; >> > + >> > + return 0; >> > +} >> > + >> > +static void ufshpb_add_lru_info(struct victim_select_info *lru_info, >> > + struct ufshpb_region *rgn) >> > +{ >> > + rgn->rgn_state = HPB_RGN_ACTIVE; >> > + list_add_tail(&rgn->list_lru_rgn, &lru_info->lh_lru_rgn); >> > + atomic_inc(&lru_info->active_cnt); >> > +} >> > + >> > +static void ufshpb_hit_lru_info(struct victim_select_info *lru_info, >> > + struct ufshpb_region *rgn) >> > +{ >> > + list_move_tail(&rgn->list_lru_rgn, &lru_info->lh_lru_rgn); >> > +} >> > + >> > +static struct ufshpb_region *ufshpb_victim_lru_info(struct ufshpb_lu >> > *hpb) >> > +{ >> > + struct victim_select_info *lru_info = &hpb->lru_info; >> > + struct ufshpb_region *rgn, *victim_rgn = NULL; >> > + >> > + list_for_each_entry(rgn, &lru_info->lh_lru_rgn, list_lru_rgn) { >> > + if (!rgn) { >> > + dev_err(&hpb->sdev_ufs_lu->sdev_dev, >> > + "%s: no region allocated\n", >> > + __func__); >> > + return NULL; >> > + } >> > + if (ufshpb_check_srgns_issue_state(hpb, rgn)) >> > + continue; >> > + >> > + victim_rgn = rgn; >> > + break; >> > + } >> > + >> > + return victim_rgn; >> > +} >> > + >> > +static void ufshpb_cleanup_lru_info(struct victim_select_info >> > *lru_info, >> > + struct ufshpb_region *rgn) >> > +{ >> > + list_del_init(&rgn->list_lru_rgn); >> > + rgn->rgn_state = HPB_RGN_INACTIVE; >> > + atomic_dec(&lru_info->active_cnt); >> > +} >> > + >> > +static void ufshpb_purge_active_subregion(struct ufshpb_lu *hpb, >> > + struct ufshpb_subregion *srgn) >> > +{ >> > + if (srgn->srgn_state != HPB_SRGN_UNUSED) { >> > + ufshpb_put_map_ctx(hpb, srgn->mctx); >> > + srgn->srgn_state = HPB_SRGN_UNUSED; >> > + srgn->mctx = NULL; >> > + } >> > +} >> > + >> > +static void __ufshpb_evict_region(struct ufshpb_lu *hpb, >> > + struct ufshpb_region *rgn) >> > +{ >> > + struct victim_select_info *lru_info; >> > + struct ufshpb_subregion *srgn; >> > + int srgn_idx; >> > + >> > + lru_info = &hpb->lru_info; >> > + >> > + dev_dbg(&hpb->sdev_ufs_lu->sdev_dev, "evict region %d\n", >> > rgn->rgn_idx); >> > + >> > + ufshpb_cleanup_lru_info(lru_info, rgn); >> > + >> > + for_each_sub_region(rgn, srgn_idx, srgn) >> > + ufshpb_purge_active_subregion(hpb, srgn); >> > +} >> > + >> > +static int ufshpb_evict_region(struct ufshpb_lu *hpb, struct >> > ufshpb_region *rgn) >> > +{ >> > + unsigned long flags; >> > + int ret = 0; >> > + >> > + spin_lock_irqsave(&hpb->rgn_state_lock, flags); >> > + if (rgn->rgn_state == HPB_RGN_PINNED) { >> > + dev_warn(&hpb->sdev_ufs_lu->sdev_dev, >> > + "pinned region cannot drop-out. region %d\n", >> > + rgn->rgn_idx); >> > + goto out; >> > + } >> > + if (!list_empty(&rgn->list_lru_rgn)) { >> > + if (ufshpb_check_srgns_issue_state(hpb, rgn)) { >> > + ret = -EBUSY; >> > + goto out; >> > + } >> > + >> > + __ufshpb_evict_region(hpb, rgn); >> > + } >> > +out: >> > + spin_unlock_irqrestore(&hpb->rgn_state_lock, flags); >> > + return ret; >> > +} >> > + >> > +static int ufshpb_issue_map_req(struct ufshpb_lu *hpb, >> > + struct ufshpb_region *rgn, >> > + struct ufshpb_subregion *srgn) >> > +{ >> > + struct ufshpb_req *map_req; >> > + unsigned long flags; >> > + int ret; >> > + int err = -EAGAIN; >> > + bool alloc_required = false; >> > + enum HPB_SRGN_STATE state = HPB_SRGN_INVALID; >> > + >> > + spin_lock_irqsave(&hpb->rgn_state_lock, flags); >> > + >> > + if (ufshpb_get_state(hpb) != HPB_PRESENT) { >> > + dev_notice(&hpb->sdev_ufs_lu->sdev_dev, >> > + "%s: ufshpb state is not PRESENT\n", __func__); >> > + goto unlock_out; >> > + } >> > + >> > + if ((rgn->rgn_state == HPB_RGN_INACTIVE) && >> > + (srgn->srgn_state == HPB_SRGN_INVALID)) { >> > + err = 0; >> > + goto unlock_out; >> > + } >> > + >> > + if (srgn->srgn_state == HPB_SRGN_UNUSED) >> > + alloc_required = true; >> > + >> > + /* >> > + * If the subregion is already ISSUED state, >> > + * a specific event (e.g., GC or wear-leveling, etc.) occurs in >> > + * the device and HPB response for map loading is received. >> > + * In this case, after finishing the HPB_READ_BUFFER, >> > + * the next HPB_READ_BUFFER is performed again to obtain the latest >> > + * map data. >> > + */ >> > + if (srgn->srgn_state == HPB_SRGN_ISSUED) >> > + goto unlock_out; >> > + >> > + srgn->srgn_state = HPB_SRGN_ISSUED; >> > + spin_unlock_irqrestore(&hpb->rgn_state_lock, flags); >> > + >> > + if (alloc_required) { >> > + if (srgn->mctx) { >> >> Can this really happen? >> >> alloc_required is true if srgn->srgn_state == HPB_SRGN_UNUSED. >> ufshpb_put_map_ctx() is called and srgn->srgn_state is given >> HPB_SRGN_UNUSED together in ufshpb_purge_active_subregion(). >> Do I miss anything? > > It is not normal case, but it is for preventing memory leak on abnormal > case. > > Thanks, > Daejun Abnormal how? Could you please give an example? Thanks, Can Guo
>>> > This is a patch for managing L2P map in HPB module. >>> > >>> > The HPB divides logical addresses into several regions. A region >>> > consists >>> > of several sub-regions. The sub-region is a basic unit where L2P >>> > mapping is >>> > managed. The driver loads L2P mapping data of each sub-region. The >>> > loaded >>> > sub-region is called active-state. The HPB driver unloads L2P mapping >>> > data >>> > as region unit. The unloaded region is called inactive-state. >>> > >>> > Sub-region/region candidates to be loaded and unloaded are delivered >>> > from >>> > the UFS device. The UFS device delivers the recommended active >>> > sub-region >>> > and inactivate region to the driver using sensedata. >>> > The HPB module performs L2P mapping management on the host through the >>> > delivered information. >>> > >>> > A pinned region is a pre-set regions on the UFS device that is always >>> > activate-state. >>> > >>> > The data structure for map data request and L2P map uses mempool API, >>> > minimizing allocation overhead while avoiding static allocation. >>> > >>> > The mininum size of the memory pool used in the HPB is implemented >>> > as a module parameter, so that it can be configurable by the user. >>> > >>> > To gurantee a minimum memory pool size of 4MB: >>> > ufshpb_host_map_kbytes=4096 >>> > >>> > The map_work manages active/inactive by 2 "to-do" lists. >>> > Each hpb lun maintains 2 "to-do" lists: >>> > hpb->lh_inact_rgn - regions to be inactivated, and >>> > hpb->lh_act_srgn - subregions to be activated >>> > Those lists are maintained on IO completion. >>> > >>> > Reviewed-by: Bart Van Assche <bvanassche@acm.org> >>> > Reviewed-by: Can Guo <cang@codeaurora.org> >>> > Acked-by: Avri Altman <Avri.Altman@wdc.com> >>> > Tested-by: Bean Huo <beanhuo@micron.com> >>> > Signed-off-by: Daejun Park <daejun7.park@samsung.com> >>> > --- >>> > drivers/scsi/ufs/ufs.h | 36 ++ >>> > drivers/scsi/ufs/ufshcd.c | 4 + >>> > drivers/scsi/ufs/ufshpb.c | 1091 ++++++++++++++++++++++++++++++++++++- >>> > drivers/scsi/ufs/ufshpb.h | 65 +++ >>> > 4 files changed, 1181 insertions(+), 15 deletions(-) >>> > >>> > diff --git a/drivers/scsi/ufs/ufs.h b/drivers/scsi/ufs/ufs.h >>> > index 65563635e20e..957763db1006 100644 >>> > --- a/drivers/scsi/ufs/ufs.h >>> > +++ b/drivers/scsi/ufs/ufs.h >>> > @@ -472,6 +472,41 @@ struct utp_cmd_rsp { >>> > u8 sense_data[UFS_SENSE_SIZE]; >>> > }; >>> > >>> > +struct ufshpb_active_field { >>> > + __be16 active_rgn; >>> > + __be16 active_srgn; >>> > +}; >>> > +#define HPB_ACT_FIELD_SIZE 4 >>> > + >>> > +/** >>> > + * struct utp_hpb_rsp - Response UPIU structure >>> > + * @residual_transfer_count: Residual transfer count DW-3 >>> > + * @reserved1: Reserved double words DW-4 to DW-7 >>> > + * @sense_data_len: Sense data length DW-8 U16 >>> > + * @desc_type: Descriptor type of sense data >>> > + * @additional_len: Additional length of sense data >>> > + * @hpb_op: HPB operation type >>> > + * @lun: LUN of response UPIU >>> > + * @active_rgn_cnt: Active region count >>> > + * @inactive_rgn_cnt: Inactive region count >>> > + * @hpb_active_field: Recommended to read HPB region and subregion >>> > + * @hpb_inactive_field: To be inactivated HPB region and subregion >>> > + */ >>> > +struct utp_hpb_rsp { >>> > + __be32 residual_transfer_count; >>> > + __be32 reserved1[4]; >>> > + __be16 sense_data_len; >>> > + u8 desc_type; >>> > + u8 additional_len; >>> > + u8 hpb_op; >>> > + u8 lun; >>> > + u8 active_rgn_cnt; >>> > + u8 inactive_rgn_cnt; >>> > + struct ufshpb_active_field hpb_active_field[2]; >>> > + __be16 hpb_inactive_field[2]; >>> > +}; >>> > +#define UTP_HPB_RSP_SIZE 40 >>> > + >>> > /** >>> > * struct utp_upiu_rsp - general upiu response structure >>> > * @header: UPIU header structure DW-0 to DW-2 >>> > @@ -482,6 +517,7 @@ struct utp_upiu_rsp { >>> > struct utp_upiu_header header; >>> > union { >>> > struct utp_cmd_rsp sr; >>> > + struct utp_hpb_rsp hr; >>> > struct utp_upiu_query qr; >>> > }; >>> > }; >>> > diff --git a/drivers/scsi/ufs/ufshcd.c b/drivers/scsi/ufs/ufshcd.c >>> > index 49b3d5d24fa6..5852ff44c3cc 100644 >>> > --- a/drivers/scsi/ufs/ufshcd.c >>> > +++ b/drivers/scsi/ufs/ufshcd.c >>> > @@ -5021,6 +5021,9 @@ ufshcd_transfer_rsp_status(struct ufs_hba *hba, >>> > struct ufshcd_lrb *lrbp) >>> > */ >>> > pm_runtime_get_noresume(hba->dev); >>> > } >>> > + >>> > + if (scsi_status == SAM_STAT_GOOD) >>> > + ufshpb_rsp_upiu(hba, lrbp); >>> > break; >>> > case UPIU_TRANSACTION_REJECT_UPIU: >>> > /* TODO: handle Reject UPIU Response */ >>> > @@ -9221,6 +9224,7 @@ EXPORT_SYMBOL(ufshcd_shutdown); >>> > void ufshcd_remove(struct ufs_hba *hba) >>> > { >>> > ufs_bsg_remove(hba); >>> > + ufshpb_remove(hba); >>> > ufs_sysfs_remove_nodes(hba->dev); >>> > blk_cleanup_queue(hba->tmf_queue); >>> > blk_mq_free_tag_set(&hba->tmf_tag_set); >>> > diff --git a/drivers/scsi/ufs/ufshpb.c b/drivers/scsi/ufs/ufshpb.c >>> > index 1a72f6541510..8abadb0e010a 100644 >>> > --- a/drivers/scsi/ufs/ufshpb.c >>> > +++ b/drivers/scsi/ufs/ufshpb.c >>> > @@ -16,6 +16,16 @@ >>> > #include "ufshpb.h" >>> > #include "../sd.h" >>> > >>> > +/* memory management */ >>> > +static struct kmem_cache *ufshpb_mctx_cache; >>> > +static mempool_t *ufshpb_mctx_pool; >>> > +static mempool_t *ufshpb_page_pool; >>> > +/* A cache size of 2MB can cache ppn in the 1GB range. */ >>> > +static unsigned int ufshpb_host_map_kbytes = 2048; >>> > +static int tot_active_srgn_pages; >>> > + >>> > +static struct workqueue_struct *ufshpb_wq; >>> > + >>> > bool ufshpb_is_allowed(struct ufs_hba *hba) >>> > { >>> > return !(hba->ufshpb_dev.hpb_disabled); >>> > @@ -36,14 +46,892 @@ static void ufshpb_set_state(struct ufshpb_lu >>> > *hpb, int state) >>> > atomic_set(&hpb->hpb_state, state); >>> > } >>> > >>> > +static bool ufshpb_is_general_lun(int lun) >>> > +{ >>> > + return lun < UFS_UPIU_MAX_UNIT_NUM_ID; >>> > +} >>> > + >>> > +static bool >>> > +ufshpb_is_pinned_region(struct ufshpb_lu *hpb, int rgn_idx) >>> > +{ >>> > + if (hpb->lu_pinned_end != PINNED_NOT_SET && >>> > + rgn_idx >= hpb->lu_pinned_start && >>> > + rgn_idx <= hpb->lu_pinned_end) >>> > + return true; >>> > + >>> > + return false; >>> > +} >>> > + >>> > +static void ufshpb_kick_map_work(struct ufshpb_lu *hpb) >>> > +{ >>> > + bool ret = false; >>> > + unsigned long flags; >>> > + >>> > + if (ufshpb_get_state(hpb) != HPB_PRESENT) >>> > + return; >>> > + >>> > + spin_lock_irqsave(&hpb->rsp_list_lock, flags); >>> > + if (!list_empty(&hpb->lh_inact_rgn) || >>> > !list_empty(&hpb->lh_act_srgn)) >>> > + ret = true; >>> > + spin_unlock_irqrestore(&hpb->rsp_list_lock, flags); >>> > + >>> > + if (ret) >>> > + queue_work(ufshpb_wq, &hpb->map_work); >>> > +} >>> > + >>> > +static bool ufshpb_is_hpb_rsp_valid(struct ufs_hba *hba, >>> > + struct ufshcd_lrb *lrbp, >>> > + struct utp_hpb_rsp *rsp_field) >>> > +{ >>> > + /* Check HPB_UPDATE_ALERT */ >>> > + if (!(lrbp->ucd_rsp_ptr->header.dword_2 & >>> > + UPIU_HEADER_DWORD(0, 2, 0, 0))) >>> > + return false; >>> > + >>> > + if (be16_to_cpu(rsp_field->sense_data_len) != DEV_SENSE_SEG_LEN || >>> > + rsp_field->desc_type != DEV_DES_TYPE || >>> > + rsp_field->additional_len != DEV_ADDITIONAL_LEN || >>> > + rsp_field->active_rgn_cnt > MAX_ACTIVE_NUM || >>> > + rsp_field->inactive_rgn_cnt > MAX_INACTIVE_NUM || >>> > + rsp_field->hpb_op == HPB_RSP_NONE || >>> > + (rsp_field->hpb_op == HPB_RSP_REQ_REGION_UPDATE && >>> > + !rsp_field->active_rgn_cnt && !rsp_field->inactive_rgn_cnt)) >>> > + return false; >>> > + >>> > + if (!ufshpb_is_general_lun(rsp_field->lun)) { >>> > + dev_warn(hba->dev, "ufshpb: lun(%d) not supported\n", >>> > + lrbp->lun); >>> > + return false; >>> > + } >>> > + >>> > + return true; >>> > +} >>> > + >>> > +static struct ufshpb_req *ufshpb_get_map_req(struct ufshpb_lu *hpb, >>> > + struct ufshpb_subregion *srgn) >>> > +{ >>> > + struct ufshpb_req *map_req; >>> > + struct request *req; >>> > + struct bio *bio; >>> > + int retries = HPB_MAP_REQ_RETRIES; >>> > + >>> > + map_req = kmem_cache_alloc(hpb->map_req_cache, GFP_KERNEL); >>> > + if (!map_req) >>> > + return NULL; >>> > + >>> > +retry: >>> > + req = blk_get_request(hpb->sdev_ufs_lu->request_queue, >>> > + REQ_OP_SCSI_IN, BLK_MQ_REQ_NOWAIT); >>> > + >>> > + if ((PTR_ERR(req) == -EWOULDBLOCK) && (--retries > 0)) { >>> > + usleep_range(3000, 3100); >>> > + goto retry; >>> > + } >>> > + >>> > + if (IS_ERR(req)) >>> > + goto free_map_req; >>> > + >>> > + bio = bio_alloc(GFP_KERNEL, hpb->pages_per_srgn); >>> > + if (!bio) { >>> > + blk_put_request(req); >>> > + goto free_map_req; >>> > + } >>> > + >>> > + map_req->hpb = hpb; >>> > + map_req->req = req; >>> > + map_req->bio = bio; >>> > + >>> > + map_req->rgn_idx = srgn->rgn_idx; >>> > + map_req->srgn_idx = srgn->srgn_idx; >>> > + map_req->mctx = srgn->mctx; >>> > + >>> > + return map_req; >>> > + >>> > +free_map_req: >>> > + kmem_cache_free(hpb->map_req_cache, map_req); >>> > + return NULL; >>> > +} >>> > + >>> > +static void ufshpb_put_map_req(struct ufshpb_lu *hpb, >>> > + struct ufshpb_req *map_req) >>> > +{ >>> > + bio_put(map_req->bio); >>> > + blk_put_request(map_req->req); >>> > + kmem_cache_free(hpb->map_req_cache, map_req); >>> > +} >>> > + >>> > +static int ufshpb_clear_dirty_bitmap(struct ufshpb_lu *hpb, >>> > + struct ufshpb_subregion *srgn) >>> > +{ >>> > + u32 num_entries = hpb->entries_per_srgn; >>> > + >>> > + if (!srgn->mctx) { >>> > + dev_err(&hpb->sdev_ufs_lu->sdev_dev, >>> > + "no mctx in region %d subregion %d.\n", >>> > + srgn->rgn_idx, srgn->srgn_idx); >>> > + return -1; >>> > + } >>> > + >>> > + if (unlikely(srgn->is_last)) >>> > + num_entries = hpb->last_srgn_entries; >>> > + >>> > + bitmap_zero(srgn->mctx->ppn_dirty, num_entries); >>> > + return 0; >>> > +} >>> > + >>> > +static void ufshpb_update_active_info(struct ufshpb_lu *hpb, int >>> > rgn_idx, >>> > + int srgn_idx) >>> > +{ >>> > + struct ufshpb_region *rgn; >>> > + struct ufshpb_subregion *srgn; >>> > + >>> > + rgn = hpb->rgn_tbl + rgn_idx; >>> > + srgn = rgn->srgn_tbl + srgn_idx; >>> > + >>> > + list_del_init(&rgn->list_inact_rgn); >>> > + >>> > + if (list_empty(&srgn->list_act_srgn)) >>> > + list_add_tail(&srgn->list_act_srgn, &hpb->lh_act_srgn); >>> > +} >>> > + >>> > +static void ufshpb_update_inactive_info(struct ufshpb_lu *hpb, int >>> > rgn_idx) >>> > +{ >>> > + struct ufshpb_region *rgn; >>> > + struct ufshpb_subregion *srgn; >>> > + int srgn_idx; >>> > + >>> > + rgn = hpb->rgn_tbl + rgn_idx; >>> > + >>> > + for_each_sub_region(rgn, srgn_idx, srgn) >>> > + list_del_init(&srgn->list_act_srgn); >>> > + >>> > + if (list_empty(&rgn->list_inact_rgn)) >>> > + list_add_tail(&rgn->list_inact_rgn, &hpb->lh_inact_rgn); >>> > +} >>> > + >>> > +static void ufshpb_activate_subregion(struct ufshpb_lu *hpb, >>> > + struct ufshpb_subregion *srgn) >>> > +{ >>> > + struct ufshpb_region *rgn; >>> > + >>> > + /* >>> > + * If there is no mctx in subregion >>> > + * after I/O progress for HPB_READ_BUFFER, the region to which the >>> > + * subregion belongs was evicted. >>> > + * Make sure the region must not evict in I/O progress >>> > + */ >>> > + if (!srgn->mctx) { >>> > + dev_err(&hpb->sdev_ufs_lu->sdev_dev, >>> > + "no mctx in region %d subregion %d.\n", >>> > + srgn->rgn_idx, srgn->srgn_idx); >>> > + srgn->srgn_state = HPB_SRGN_INVALID; >>> > + return; >>> > + } >>> > + >>> > + rgn = hpb->rgn_tbl + srgn->rgn_idx; >>> > + >>> > + if (unlikely(rgn->rgn_state == HPB_RGN_INACTIVE)) { >>> > + dev_err(&hpb->sdev_ufs_lu->sdev_dev, >>> > + "region %d subregion %d evicted\n", >>> > + srgn->rgn_idx, srgn->srgn_idx); >>> > + srgn->srgn_state = HPB_SRGN_INVALID; >>> > + return; >>> > + } >>> > + srgn->srgn_state = HPB_SRGN_VALID; >>> > +} >>> > + >>> > +static void ufshpb_map_req_compl_fn(struct request *req, blk_status_t >>> > error) >>> > +{ >>> > + struct ufshpb_req *map_req = (struct ufshpb_req *) req->end_io_data; >>> > + struct ufshpb_lu *hpb = map_req->hpb; >>> > + struct ufshpb_subregion *srgn; >>> > + unsigned long flags; >>> > + >>> > + srgn = hpb->rgn_tbl[map_req->rgn_idx].srgn_tbl + >>> > + map_req->srgn_idx; >>> > + >>> > + ufshpb_clear_dirty_bitmap(hpb, srgn); >>> > + spin_lock_irqsave(&hpb->rgn_state_lock, flags); >>> > + ufshpb_activate_subregion(hpb, srgn); >>> > + spin_unlock_irqrestore(&hpb->rgn_state_lock, flags); >>> > + >>> > + ufshpb_put_map_req(map_req->hpb, map_req); >>> > +} >>> > + >>> > +static void ufshpb_set_read_buf_cmd(unsigned char *cdb, int rgn_idx, >>> > + int srgn_idx, int srgn_mem_size) >>> > +{ >>> > + cdb[0] = UFSHPB_READ_BUFFER; >>> > + cdb[1] = UFSHPB_READ_BUFFER_ID; >>> > + >>> > + put_unaligned_be16(rgn_idx, &cdb[2]); >>> > + put_unaligned_be16(srgn_idx, &cdb[4]); >>> > + put_unaligned_be24(srgn_mem_size, &cdb[6]); >>> > + >>> > + cdb[9] = 0x00; >>> > +} >>> > + >>> > +static int ufshpb_execute_map_req(struct ufshpb_lu *hpb, >>> > + struct ufshpb_req *map_req, bool last) >>> > +{ >>> > + struct request_queue *q; >>> > + struct request *req; >>> > + struct scsi_request *rq; >>> > + int mem_size = hpb->srgn_mem_size; >>> > + int ret = 0; >>> > + int i; >>> > + >>> > + q = hpb->sdev_ufs_lu->request_queue; >>> > + for (i = 0; i < hpb->pages_per_srgn; i++) { >>> > + ret = bio_add_pc_page(q, map_req->bio, map_req->mctx->m_page[i], >>> > + PAGE_SIZE, 0); >>> > + if (ret != PAGE_SIZE) { >>> > + dev_err(&hpb->sdev_ufs_lu->sdev_dev, >>> > + "bio_add_pc_page fail %d - %d\n", >>> > + map_req->rgn_idx, map_req->srgn_idx); >>> > + return ret; >>> > + } >>> > + } >>> > + >>> > + req = map_req->req; >>> > + >>> > + blk_rq_append_bio(req, &map_req->bio); >>> > + >>> > + req->end_io_data = map_req; >>> > + >>> > + rq = scsi_req(req); >>> > + >>> > + if (unlikely(last)) >>> > + mem_size = hpb->last_srgn_entries * HPB_ENTRY_SIZE; >>> > + >>> > + ufshpb_set_read_buf_cmd(rq->cmd, map_req->rgn_idx, >>> > + map_req->srgn_idx, mem_size); >>> > + rq->cmd_len = HPB_READ_BUFFER_CMD_LENGTH; >>> > + >>> > + blk_execute_rq_nowait(q, NULL, req, 1, ufshpb_map_req_compl_fn); >>> > + >>> > + hpb->stats.map_req_cnt++; >>> > + return 0; >>> > +} >>> > + >>> > +static struct ufshpb_map_ctx *ufshpb_get_map_ctx(struct ufshpb_lu >>> > *hpb, >>> > + bool last) >>> > +{ >>> > + struct ufshpb_map_ctx *mctx; >>> > + u32 num_entries = hpb->entries_per_srgn; >>> > + int i, j; >>> > + >>> > + mctx = mempool_alloc(ufshpb_mctx_pool, GFP_KERNEL); >>> > + if (!mctx) >>> > + return NULL; >>> > + >>> > + mctx->m_page = kmem_cache_alloc(hpb->m_page_cache, GFP_KERNEL); >>> > + if (!mctx->m_page) >>> > + goto release_mctx; >>> > + >>> > + if (unlikely(last)) >>> > + num_entries = hpb->last_srgn_entries; >>> > + >>> > + mctx->ppn_dirty = bitmap_zalloc(num_entries, GFP_KERNEL); >>> > + if (!mctx->ppn_dirty) >>> > + goto release_m_page; >>> > + >>> > + for (i = 0; i < hpb->pages_per_srgn; i++) { >>> > + mctx->m_page[i] = mempool_alloc(ufshpb_page_pool, GFP_KERNEL); >>> > + if (!mctx->m_page[i]) { >>> > + for (j = 0; j < i; j++) >>> > + mempool_free(mctx->m_page[j], ufshpb_page_pool); >>> > + goto release_ppn_dirty; >>> > + } >>> > + clear_page(page_address(mctx->m_page[i])); >>> > + } >>> > + >>> > + return mctx; >>> > + >>> > +release_ppn_dirty: >>> > + bitmap_free(mctx->ppn_dirty); >>> > +release_m_page: >>> > + kmem_cache_free(hpb->m_page_cache, mctx->m_page); >>> > +release_mctx: >>> > + mempool_free(mctx, ufshpb_mctx_pool); >>> > + return NULL; >>> > +} >>> > + >>> > +static void ufshpb_put_map_ctx(struct ufshpb_lu *hpb, >>> > + struct ufshpb_map_ctx *mctx) >>> > +{ >>> > + int i; >>> > + >>> > + for (i = 0; i < hpb->pages_per_srgn; i++) >>> > + mempool_free(mctx->m_page[i], ufshpb_page_pool); >>> > + >>> > + bitmap_free(mctx->ppn_dirty); >>> > + kmem_cache_free(hpb->m_page_cache, mctx->m_page); >>> > + mempool_free(mctx, ufshpb_mctx_pool); >>> > +} >>> > + >>> > +static int ufshpb_check_srgns_issue_state(struct ufshpb_lu *hpb, >>> > + struct ufshpb_region *rgn) >>> > +{ >>> > + struct ufshpb_subregion *srgn; >>> > + int srgn_idx; >>> > + >>> > + for_each_sub_region(rgn, srgn_idx, srgn) >>> > + if (srgn->srgn_state == HPB_SRGN_ISSUED) >>> > + return -EPERM; >>> > + >>> > + return 0; >>> > +} >>> > + >>> > +static void ufshpb_add_lru_info(struct victim_select_info *lru_info, >>> > + struct ufshpb_region *rgn) >>> > +{ >>> > + rgn->rgn_state = HPB_RGN_ACTIVE; >>> > + list_add_tail(&rgn->list_lru_rgn, &lru_info->lh_lru_rgn); >>> > + atomic_inc(&lru_info->active_cnt); >>> > +} >>> > + >>> > +static void ufshpb_hit_lru_info(struct victim_select_info *lru_info, >>> > + struct ufshpb_region *rgn) >>> > +{ >>> > + list_move_tail(&rgn->list_lru_rgn, &lru_info->lh_lru_rgn); >>> > +} >>> > + >>> > +static struct ufshpb_region *ufshpb_victim_lru_info(struct ufshpb_lu >>> > *hpb) >>> > +{ >>> > + struct victim_select_info *lru_info = &hpb->lru_info; >>> > + struct ufshpb_region *rgn, *victim_rgn = NULL; >>> > + >>> > + list_for_each_entry(rgn, &lru_info->lh_lru_rgn, list_lru_rgn) { >>> > + if (!rgn) { >>> > + dev_err(&hpb->sdev_ufs_lu->sdev_dev, >>> > + "%s: no region allocated\n", >>> > + __func__); >>> > + return NULL; >>> > + } >>> > + if (ufshpb_check_srgns_issue_state(hpb, rgn)) >>> > + continue; >>> > + >>> > + victim_rgn = rgn; >>> > + break; >>> > + } >>> > + >>> > + return victim_rgn; >>> > +} >>> > + >>> > +static void ufshpb_cleanup_lru_info(struct victim_select_info >>> > *lru_info, >>> > + struct ufshpb_region *rgn) >>> > +{ >>> > + list_del_init(&rgn->list_lru_rgn); >>> > + rgn->rgn_state = HPB_RGN_INACTIVE; >>> > + atomic_dec(&lru_info->active_cnt); >>> > +} >>> > + >>> > +static void ufshpb_purge_active_subregion(struct ufshpb_lu *hpb, >>> > + struct ufshpb_subregion *srgn) >>> > +{ >>> > + if (srgn->srgn_state != HPB_SRGN_UNUSED) { >>> > + ufshpb_put_map_ctx(hpb, srgn->mctx); >>> > + srgn->srgn_state = HPB_SRGN_UNUSED; >>> > + srgn->mctx = NULL; >>> > + } >>> > +} >>> > + >>> > +static void __ufshpb_evict_region(struct ufshpb_lu *hpb, >>> > + struct ufshpb_region *rgn) >>> > +{ >>> > + struct victim_select_info *lru_info; >>> > + struct ufshpb_subregion *srgn; >>> > + int srgn_idx; >>> > + >>> > + lru_info = &hpb->lru_info; >>> > + >>> > + dev_dbg(&hpb->sdev_ufs_lu->sdev_dev, "evict region %d\n", >>> > rgn->rgn_idx); >>> > + >>> > + ufshpb_cleanup_lru_info(lru_info, rgn); >>> > + >>> > + for_each_sub_region(rgn, srgn_idx, srgn) >>> > + ufshpb_purge_active_subregion(hpb, srgn); >>> > +} >>> > + >>> > +static int ufshpb_evict_region(struct ufshpb_lu *hpb, struct >>> > ufshpb_region *rgn) >>> > +{ >>> > + unsigned long flags; >>> > + int ret = 0; >>> > + >>> > + spin_lock_irqsave(&hpb->rgn_state_lock, flags); >>> > + if (rgn->rgn_state == HPB_RGN_PINNED) { >>> > + dev_warn(&hpb->sdev_ufs_lu->sdev_dev, >>> > + "pinned region cannot drop-out. region %d\n", >>> > + rgn->rgn_idx); >>> > + goto out; >>> > + } >>> > + if (!list_empty(&rgn->list_lru_rgn)) { >>> > + if (ufshpb_check_srgns_issue_state(hpb, rgn)) { >>> > + ret = -EBUSY; >>> > + goto out; >>> > + } >>> > + >>> > + __ufshpb_evict_region(hpb, rgn); >>> > + } >>> > +out: >>> > + spin_unlock_irqrestore(&hpb->rgn_state_lock, flags); >>> > + return ret; >>> > +} >>> > + >>> > +static int ufshpb_issue_map_req(struct ufshpb_lu *hpb, >>> > + struct ufshpb_region *rgn, >>> > + struct ufshpb_subregion *srgn) >>> > +{ >>> > + struct ufshpb_req *map_req; >>> > + unsigned long flags; >>> > + int ret; >>> > + int err = -EAGAIN; >>> > + bool alloc_required = false; >>> > + enum HPB_SRGN_STATE state = HPB_SRGN_INVALID; >>> > + >>> > + spin_lock_irqsave(&hpb->rgn_state_lock, flags); >>> > + >>> > + if (ufshpb_get_state(hpb) != HPB_PRESENT) { >>> > + dev_notice(&hpb->sdev_ufs_lu->sdev_dev, >>> > + "%s: ufshpb state is not PRESENT\n", __func__); >>> > + goto unlock_out; >>> > + } >>> > + >>> > + if ((rgn->rgn_state == HPB_RGN_INACTIVE) && >>> > + (srgn->srgn_state == HPB_SRGN_INVALID)) { >>> > + err = 0; >>> > + goto unlock_out; >>> > + } >>> > + >>> > + if (srgn->srgn_state == HPB_SRGN_UNUSED) >>> > + alloc_required = true; >>> > + >>> > + /* >>> > + * If the subregion is already ISSUED state, >>> > + * a specific event (e.g., GC or wear-leveling, etc.) occurs in >>> > + * the device and HPB response for map loading is received. >>> > + * In this case, after finishing the HPB_READ_BUFFER, >>> > + * the next HPB_READ_BUFFER is performed again to obtain the latest >>> > + * map data. >>> > + */ >>> > + if (srgn->srgn_state == HPB_SRGN_ISSUED) >>> > + goto unlock_out; >>> > + >>> > + srgn->srgn_state = HPB_SRGN_ISSUED; >>> > + spin_unlock_irqrestore(&hpb->rgn_state_lock, flags); >>> > + >>> > + if (alloc_required) { >>> > + if (srgn->mctx) { >>> >>> Can this really happen? >>> >>> alloc_required is true if srgn->srgn_state == HPB_SRGN_UNUSED. >>> ufshpb_put_map_ctx() is called and srgn->srgn_state is given >>> HPB_SRGN_UNUSED together in ufshpb_purge_active_subregion(). >>> Do I miss anything? >> >> It is not normal case, but it is for preventing memory leak on abnormal >> case. >> >> Thanks, >> Daejun > >Abnormal how? Could you please give an example? It is kind of additional memory leak protection. Do you think it is unnecessary? Thanks, Daejun
On 2021-03-12 15:17, Daejun Park wrote: >>>> > This is a patch for managing L2P map in HPB module. >>>> > >>>> > The HPB divides logical addresses into several regions. A region >>>> > consists >>>> > of several sub-regions. The sub-region is a basic unit where L2P >>>> > mapping is >>>> > managed. The driver loads L2P mapping data of each sub-region. The >>>> > loaded >>>> > sub-region is called active-state. The HPB driver unloads L2P mapping >>>> > data >>>> > as region unit. The unloaded region is called inactive-state. >>>> > >>>> > Sub-region/region candidates to be loaded and unloaded are delivered >>>> > from >>>> > the UFS device. The UFS device delivers the recommended active >>>> > sub-region >>>> > and inactivate region to the driver using sensedata. >>>> > The HPB module performs L2P mapping management on the host through the >>>> > delivered information. >>>> > >>>> > A pinned region is a pre-set regions on the UFS device that is always >>>> > activate-state. >>>> > >>>> > The data structure for map data request and L2P map uses mempool API, >>>> > minimizing allocation overhead while avoiding static allocation. >>>> > >>>> > The mininum size of the memory pool used in the HPB is implemented >>>> > as a module parameter, so that it can be configurable by the user. >>>> > >>>> > To gurantee a minimum memory pool size of 4MB: >>>> > ufshpb_host_map_kbytes=4096 >>>> > >>>> > The map_work manages active/inactive by 2 "to-do" lists. >>>> > Each hpb lun maintains 2 "to-do" lists: >>>> > hpb->lh_inact_rgn - regions to be inactivated, and >>>> > hpb->lh_act_srgn - subregions to be activated >>>> > Those lists are maintained on IO completion. >>>> > >>>> > Reviewed-by: Bart Van Assche <bvanassche@acm.org> >>>> > Reviewed-by: Can Guo <cang@codeaurora.org> >>>> > Acked-by: Avri Altman <Avri.Altman@wdc.com> >>>> > Tested-by: Bean Huo <beanhuo@micron.com> >>>> > Signed-off-by: Daejun Park <daejun7.park@samsung.com> >>>> > --- >>>> > drivers/scsi/ufs/ufs.h | 36 ++ >>>> > drivers/scsi/ufs/ufshcd.c | 4 + >>>> > drivers/scsi/ufs/ufshpb.c | 1091 ++++++++++++++++++++++++++++++++++++- >>>> > drivers/scsi/ufs/ufshpb.h | 65 +++ >>>> > 4 files changed, 1181 insertions(+), 15 deletions(-) >>>> > >>>> > diff --git a/drivers/scsi/ufs/ufs.h b/drivers/scsi/ufs/ufs.h >>>> > index 65563635e20e..957763db1006 100644 >>>> > --- a/drivers/scsi/ufs/ufs.h >>>> > +++ b/drivers/scsi/ufs/ufs.h >>>> > @@ -472,6 +472,41 @@ struct utp_cmd_rsp { >>>> > u8 sense_data[UFS_SENSE_SIZE]; >>>> > }; >>>> > >>>> > +struct ufshpb_active_field { >>>> > + __be16 active_rgn; >>>> > + __be16 active_srgn; >>>> > +}; >>>> > +#define HPB_ACT_FIELD_SIZE 4 >>>> > + >>>> > +/** >>>> > + * struct utp_hpb_rsp - Response UPIU structure >>>> > + * @residual_transfer_count: Residual transfer count DW-3 >>>> > + * @reserved1: Reserved double words DW-4 to DW-7 >>>> > + * @sense_data_len: Sense data length DW-8 U16 >>>> > + * @desc_type: Descriptor type of sense data >>>> > + * @additional_len: Additional length of sense data >>>> > + * @hpb_op: HPB operation type >>>> > + * @lun: LUN of response UPIU >>>> > + * @active_rgn_cnt: Active region count >>>> > + * @inactive_rgn_cnt: Inactive region count >>>> > + * @hpb_active_field: Recommended to read HPB region and subregion >>>> > + * @hpb_inactive_field: To be inactivated HPB region and subregion >>>> > + */ >>>> > +struct utp_hpb_rsp { >>>> > + __be32 residual_transfer_count; >>>> > + __be32 reserved1[4]; >>>> > + __be16 sense_data_len; >>>> > + u8 desc_type; >>>> > + u8 additional_len; >>>> > + u8 hpb_op; >>>> > + u8 lun; >>>> > + u8 active_rgn_cnt; >>>> > + u8 inactive_rgn_cnt; >>>> > + struct ufshpb_active_field hpb_active_field[2]; >>>> > + __be16 hpb_inactive_field[2]; >>>> > +}; >>>> > +#define UTP_HPB_RSP_SIZE 40 >>>> > + >>>> > /** >>>> > * struct utp_upiu_rsp - general upiu response structure >>>> > * @header: UPIU header structure DW-0 to DW-2 >>>> > @@ -482,6 +517,7 @@ struct utp_upiu_rsp { >>>> > struct utp_upiu_header header; >>>> > union { >>>> > struct utp_cmd_rsp sr; >>>> > + struct utp_hpb_rsp hr; >>>> > struct utp_upiu_query qr; >>>> > }; >>>> > }; >>>> > diff --git a/drivers/scsi/ufs/ufshcd.c b/drivers/scsi/ufs/ufshcd.c >>>> > index 49b3d5d24fa6..5852ff44c3cc 100644 >>>> > --- a/drivers/scsi/ufs/ufshcd.c >>>> > +++ b/drivers/scsi/ufs/ufshcd.c >>>> > @@ -5021,6 +5021,9 @@ ufshcd_transfer_rsp_status(struct ufs_hba *hba, >>>> > struct ufshcd_lrb *lrbp) >>>> > */ >>>> > pm_runtime_get_noresume(hba->dev); >>>> > } >>>> > + >>>> > + if (scsi_status == SAM_STAT_GOOD) >>>> > + ufshpb_rsp_upiu(hba, lrbp); >>>> > break; >>>> > case UPIU_TRANSACTION_REJECT_UPIU: >>>> > /* TODO: handle Reject UPIU Response */ >>>> > @@ -9221,6 +9224,7 @@ EXPORT_SYMBOL(ufshcd_shutdown); >>>> > void ufshcd_remove(struct ufs_hba *hba) >>>> > { >>>> > ufs_bsg_remove(hba); >>>> > + ufshpb_remove(hba); >>>> > ufs_sysfs_remove_nodes(hba->dev); >>>> > blk_cleanup_queue(hba->tmf_queue); >>>> > blk_mq_free_tag_set(&hba->tmf_tag_set); >>>> > diff --git a/drivers/scsi/ufs/ufshpb.c b/drivers/scsi/ufs/ufshpb.c >>>> > index 1a72f6541510..8abadb0e010a 100644 >>>> > --- a/drivers/scsi/ufs/ufshpb.c >>>> > +++ b/drivers/scsi/ufs/ufshpb.c >>>> > @@ -16,6 +16,16 @@ >>>> > #include "ufshpb.h" >>>> > #include "../sd.h" >>>> > >>>> > +/* memory management */ >>>> > +static struct kmem_cache *ufshpb_mctx_cache; >>>> > +static mempool_t *ufshpb_mctx_pool; >>>> > +static mempool_t *ufshpb_page_pool; >>>> > +/* A cache size of 2MB can cache ppn in the 1GB range. */ >>>> > +static unsigned int ufshpb_host_map_kbytes = 2048; >>>> > +static int tot_active_srgn_pages; >>>> > + >>>> > +static struct workqueue_struct *ufshpb_wq; >>>> > + >>>> > bool ufshpb_is_allowed(struct ufs_hba *hba) >>>> > { >>>> > return !(hba->ufshpb_dev.hpb_disabled); >>>> > @@ -36,14 +46,892 @@ static void ufshpb_set_state(struct ufshpb_lu >>>> > *hpb, int state) >>>> > atomic_set(&hpb->hpb_state, state); >>>> > } >>>> > >>>> > +static bool ufshpb_is_general_lun(int lun) >>>> > +{ >>>> > + return lun < UFS_UPIU_MAX_UNIT_NUM_ID; >>>> > +} >>>> > + >>>> > +static bool >>>> > +ufshpb_is_pinned_region(struct ufshpb_lu *hpb, int rgn_idx) >>>> > +{ >>>> > + if (hpb->lu_pinned_end != PINNED_NOT_SET && >>>> > + rgn_idx >= hpb->lu_pinned_start && >>>> > + rgn_idx <= hpb->lu_pinned_end) >>>> > + return true; >>>> > + >>>> > + return false; >>>> > +} >>>> > + >>>> > +static void ufshpb_kick_map_work(struct ufshpb_lu *hpb) >>>> > +{ >>>> > + bool ret = false; >>>> > + unsigned long flags; >>>> > + >>>> > + if (ufshpb_get_state(hpb) != HPB_PRESENT) >>>> > + return; >>>> > + >>>> > + spin_lock_irqsave(&hpb->rsp_list_lock, flags); >>>> > + if (!list_empty(&hpb->lh_inact_rgn) || >>>> > !list_empty(&hpb->lh_act_srgn)) >>>> > + ret = true; >>>> > + spin_unlock_irqrestore(&hpb->rsp_list_lock, flags); >>>> > + >>>> > + if (ret) >>>> > + queue_work(ufshpb_wq, &hpb->map_work); >>>> > +} >>>> > + >>>> > +static bool ufshpb_is_hpb_rsp_valid(struct ufs_hba *hba, >>>> > + struct ufshcd_lrb *lrbp, >>>> > + struct utp_hpb_rsp *rsp_field) >>>> > +{ >>>> > + /* Check HPB_UPDATE_ALERT */ >>>> > + if (!(lrbp->ucd_rsp_ptr->header.dword_2 & >>>> > + UPIU_HEADER_DWORD(0, 2, 0, 0))) >>>> > + return false; >>>> > + >>>> > + if (be16_to_cpu(rsp_field->sense_data_len) != DEV_SENSE_SEG_LEN || >>>> > + rsp_field->desc_type != DEV_DES_TYPE || >>>> > + rsp_field->additional_len != DEV_ADDITIONAL_LEN || >>>> > + rsp_field->active_rgn_cnt > MAX_ACTIVE_NUM || >>>> > + rsp_field->inactive_rgn_cnt > MAX_INACTIVE_NUM || >>>> > + rsp_field->hpb_op == HPB_RSP_NONE || >>>> > + (rsp_field->hpb_op == HPB_RSP_REQ_REGION_UPDATE && >>>> > + !rsp_field->active_rgn_cnt && !rsp_field->inactive_rgn_cnt)) >>>> > + return false; >>>> > + >>>> > + if (!ufshpb_is_general_lun(rsp_field->lun)) { >>>> > + dev_warn(hba->dev, "ufshpb: lun(%d) not supported\n", >>>> > + lrbp->lun); >>>> > + return false; >>>> > + } >>>> > + >>>> > + return true; >>>> > +} >>>> > + >>>> > +static struct ufshpb_req *ufshpb_get_map_req(struct ufshpb_lu *hpb, >>>> > + struct ufshpb_subregion *srgn) >>>> > +{ >>>> > + struct ufshpb_req *map_req; >>>> > + struct request *req; >>>> > + struct bio *bio; >>>> > + int retries = HPB_MAP_REQ_RETRIES; >>>> > + >>>> > + map_req = kmem_cache_alloc(hpb->map_req_cache, GFP_KERNEL); >>>> > + if (!map_req) >>>> > + return NULL; >>>> > + >>>> > +retry: >>>> > + req = blk_get_request(hpb->sdev_ufs_lu->request_queue, >>>> > + REQ_OP_SCSI_IN, BLK_MQ_REQ_NOWAIT); >>>> > + >>>> > + if ((PTR_ERR(req) == -EWOULDBLOCK) && (--retries > 0)) { >>>> > + usleep_range(3000, 3100); >>>> > + goto retry; >>>> > + } >>>> > + >>>> > + if (IS_ERR(req)) >>>> > + goto free_map_req; >>>> > + >>>> > + bio = bio_alloc(GFP_KERNEL, hpb->pages_per_srgn); >>>> > + if (!bio) { >>>> > + blk_put_request(req); >>>> > + goto free_map_req; >>>> > + } >>>> > + >>>> > + map_req->hpb = hpb; >>>> > + map_req->req = req; >>>> > + map_req->bio = bio; >>>> > + >>>> > + map_req->rgn_idx = srgn->rgn_idx; >>>> > + map_req->srgn_idx = srgn->srgn_idx; >>>> > + map_req->mctx = srgn->mctx; >>>> > + >>>> > + return map_req; >>>> > + >>>> > +free_map_req: >>>> > + kmem_cache_free(hpb->map_req_cache, map_req); >>>> > + return NULL; >>>> > +} >>>> > + >>>> > +static void ufshpb_put_map_req(struct ufshpb_lu *hpb, >>>> > + struct ufshpb_req *map_req) >>>> > +{ >>>> > + bio_put(map_req->bio); >>>> > + blk_put_request(map_req->req); >>>> > + kmem_cache_free(hpb->map_req_cache, map_req); >>>> > +} >>>> > + >>>> > +static int ufshpb_clear_dirty_bitmap(struct ufshpb_lu *hpb, >>>> > + struct ufshpb_subregion *srgn) >>>> > +{ >>>> > + u32 num_entries = hpb->entries_per_srgn; >>>> > + >>>> > + if (!srgn->mctx) { >>>> > + dev_err(&hpb->sdev_ufs_lu->sdev_dev, >>>> > + "no mctx in region %d subregion %d.\n", >>>> > + srgn->rgn_idx, srgn->srgn_idx); >>>> > + return -1; >>>> > + } >>>> > + >>>> > + if (unlikely(srgn->is_last)) >>>> > + num_entries = hpb->last_srgn_entries; >>>> > + >>>> > + bitmap_zero(srgn->mctx->ppn_dirty, num_entries); >>>> > + return 0; >>>> > +} >>>> > + >>>> > +static void ufshpb_update_active_info(struct ufshpb_lu *hpb, int >>>> > rgn_idx, >>>> > + int srgn_idx) >>>> > +{ >>>> > + struct ufshpb_region *rgn; >>>> > + struct ufshpb_subregion *srgn; >>>> > + >>>> > + rgn = hpb->rgn_tbl + rgn_idx; >>>> > + srgn = rgn->srgn_tbl + srgn_idx; >>>> > + >>>> > + list_del_init(&rgn->list_inact_rgn); >>>> > + >>>> > + if (list_empty(&srgn->list_act_srgn)) >>>> > + list_add_tail(&srgn->list_act_srgn, &hpb->lh_act_srgn); >>>> > +} >>>> > + >>>> > +static void ufshpb_update_inactive_info(struct ufshpb_lu *hpb, int >>>> > rgn_idx) >>>> > +{ >>>> > + struct ufshpb_region *rgn; >>>> > + struct ufshpb_subregion *srgn; >>>> > + int srgn_idx; >>>> > + >>>> > + rgn = hpb->rgn_tbl + rgn_idx; >>>> > + >>>> > + for_each_sub_region(rgn, srgn_idx, srgn) >>>> > + list_del_init(&srgn->list_act_srgn); >>>> > + >>>> > + if (list_empty(&rgn->list_inact_rgn)) >>>> > + list_add_tail(&rgn->list_inact_rgn, &hpb->lh_inact_rgn); >>>> > +} >>>> > + >>>> > +static void ufshpb_activate_subregion(struct ufshpb_lu *hpb, >>>> > + struct ufshpb_subregion *srgn) >>>> > +{ >>>> > + struct ufshpb_region *rgn; >>>> > + >>>> > + /* >>>> > + * If there is no mctx in subregion >>>> > + * after I/O progress for HPB_READ_BUFFER, the region to which the >>>> > + * subregion belongs was evicted. >>>> > + * Make sure the region must not evict in I/O progress >>>> > + */ >>>> > + if (!srgn->mctx) { >>>> > + dev_err(&hpb->sdev_ufs_lu->sdev_dev, >>>> > + "no mctx in region %d subregion %d.\n", >>>> > + srgn->rgn_idx, srgn->srgn_idx); >>>> > + srgn->srgn_state = HPB_SRGN_INVALID; >>>> > + return; >>>> > + } >>>> > + >>>> > + rgn = hpb->rgn_tbl + srgn->rgn_idx; >>>> > + >>>> > + if (unlikely(rgn->rgn_state == HPB_RGN_INACTIVE)) { >>>> > + dev_err(&hpb->sdev_ufs_lu->sdev_dev, >>>> > + "region %d subregion %d evicted\n", >>>> > + srgn->rgn_idx, srgn->srgn_idx); >>>> > + srgn->srgn_state = HPB_SRGN_INVALID; >>>> > + return; >>>> > + } >>>> > + srgn->srgn_state = HPB_SRGN_VALID; >>>> > +} >>>> > + >>>> > +static void ufshpb_map_req_compl_fn(struct request *req, blk_status_t >>>> > error) >>>> > +{ >>>> > + struct ufshpb_req *map_req = (struct ufshpb_req *) req->end_io_data; >>>> > + struct ufshpb_lu *hpb = map_req->hpb; >>>> > + struct ufshpb_subregion *srgn; >>>> > + unsigned long flags; >>>> > + >>>> > + srgn = hpb->rgn_tbl[map_req->rgn_idx].srgn_tbl + >>>> > + map_req->srgn_idx; >>>> > + >>>> > + ufshpb_clear_dirty_bitmap(hpb, srgn); >>>> > + spin_lock_irqsave(&hpb->rgn_state_lock, flags); >>>> > + ufshpb_activate_subregion(hpb, srgn); >>>> > + spin_unlock_irqrestore(&hpb->rgn_state_lock, flags); >>>> > + >>>> > + ufshpb_put_map_req(map_req->hpb, map_req); >>>> > +} >>>> > + >>>> > +static void ufshpb_set_read_buf_cmd(unsigned char *cdb, int rgn_idx, >>>> > + int srgn_idx, int srgn_mem_size) >>>> > +{ >>>> > + cdb[0] = UFSHPB_READ_BUFFER; >>>> > + cdb[1] = UFSHPB_READ_BUFFER_ID; >>>> > + >>>> > + put_unaligned_be16(rgn_idx, &cdb[2]); >>>> > + put_unaligned_be16(srgn_idx, &cdb[4]); >>>> > + put_unaligned_be24(srgn_mem_size, &cdb[6]); >>>> > + >>>> > + cdb[9] = 0x00; >>>> > +} >>>> > + >>>> > +static int ufshpb_execute_map_req(struct ufshpb_lu *hpb, >>>> > + struct ufshpb_req *map_req, bool last) >>>> > +{ >>>> > + struct request_queue *q; >>>> > + struct request *req; >>>> > + struct scsi_request *rq; >>>> > + int mem_size = hpb->srgn_mem_size; >>>> > + int ret = 0; >>>> > + int i; >>>> > + >>>> > + q = hpb->sdev_ufs_lu->request_queue; >>>> > + for (i = 0; i < hpb->pages_per_srgn; i++) { >>>> > + ret = bio_add_pc_page(q, map_req->bio, map_req->mctx->m_page[i], >>>> > + PAGE_SIZE, 0); >>>> > + if (ret != PAGE_SIZE) { >>>> > + dev_err(&hpb->sdev_ufs_lu->sdev_dev, >>>> > + "bio_add_pc_page fail %d - %d\n", >>>> > + map_req->rgn_idx, map_req->srgn_idx); >>>> > + return ret; >>>> > + } >>>> > + } >>>> > + >>>> > + req = map_req->req; >>>> > + >>>> > + blk_rq_append_bio(req, &map_req->bio); >>>> > + >>>> > + req->end_io_data = map_req; >>>> > + >>>> > + rq = scsi_req(req); >>>> > + >>>> > + if (unlikely(last)) >>>> > + mem_size = hpb->last_srgn_entries * HPB_ENTRY_SIZE; >>>> > + >>>> > + ufshpb_set_read_buf_cmd(rq->cmd, map_req->rgn_idx, >>>> > + map_req->srgn_idx, mem_size); >>>> > + rq->cmd_len = HPB_READ_BUFFER_CMD_LENGTH; >>>> > + >>>> > + blk_execute_rq_nowait(q, NULL, req, 1, ufshpb_map_req_compl_fn); >>>> > + >>>> > + hpb->stats.map_req_cnt++; >>>> > + return 0; >>>> > +} >>>> > + >>>> > +static struct ufshpb_map_ctx *ufshpb_get_map_ctx(struct ufshpb_lu >>>> > *hpb, >>>> > + bool last) >>>> > +{ >>>> > + struct ufshpb_map_ctx *mctx; >>>> > + u32 num_entries = hpb->entries_per_srgn; >>>> > + int i, j; >>>> > + >>>> > + mctx = mempool_alloc(ufshpb_mctx_pool, GFP_KERNEL); >>>> > + if (!mctx) >>>> > + return NULL; >>>> > + >>>> > + mctx->m_page = kmem_cache_alloc(hpb->m_page_cache, GFP_KERNEL); >>>> > + if (!mctx->m_page) >>>> > + goto release_mctx; >>>> > + >>>> > + if (unlikely(last)) >>>> > + num_entries = hpb->last_srgn_entries; >>>> > + >>>> > + mctx->ppn_dirty = bitmap_zalloc(num_entries, GFP_KERNEL); >>>> > + if (!mctx->ppn_dirty) >>>> > + goto release_m_page; >>>> > + >>>> > + for (i = 0; i < hpb->pages_per_srgn; i++) { >>>> > + mctx->m_page[i] = mempool_alloc(ufshpb_page_pool, GFP_KERNEL); >>>> > + if (!mctx->m_page[i]) { >>>> > + for (j = 0; j < i; j++) >>>> > + mempool_free(mctx->m_page[j], ufshpb_page_pool); >>>> > + goto release_ppn_dirty; >>>> > + } >>>> > + clear_page(page_address(mctx->m_page[i])); >>>> > + } >>>> > + >>>> > + return mctx; >>>> > + >>>> > +release_ppn_dirty: >>>> > + bitmap_free(mctx->ppn_dirty); >>>> > +release_m_page: >>>> > + kmem_cache_free(hpb->m_page_cache, mctx->m_page); >>>> > +release_mctx: >>>> > + mempool_free(mctx, ufshpb_mctx_pool); >>>> > + return NULL; >>>> > +} >>>> > + >>>> > +static void ufshpb_put_map_ctx(struct ufshpb_lu *hpb, >>>> > + struct ufshpb_map_ctx *mctx) >>>> > +{ >>>> > + int i; >>>> > + >>>> > + for (i = 0; i < hpb->pages_per_srgn; i++) >>>> > + mempool_free(mctx->m_page[i], ufshpb_page_pool); >>>> > + >>>> > + bitmap_free(mctx->ppn_dirty); >>>> > + kmem_cache_free(hpb->m_page_cache, mctx->m_page); >>>> > + mempool_free(mctx, ufshpb_mctx_pool); >>>> > +} >>>> > + >>>> > +static int ufshpb_check_srgns_issue_state(struct ufshpb_lu *hpb, >>>> > + struct ufshpb_region *rgn) >>>> > +{ >>>> > + struct ufshpb_subregion *srgn; >>>> > + int srgn_idx; >>>> > + >>>> > + for_each_sub_region(rgn, srgn_idx, srgn) >>>> > + if (srgn->srgn_state == HPB_SRGN_ISSUED) >>>> > + return -EPERM; >>>> > + >>>> > + return 0; >>>> > +} >>>> > + >>>> > +static void ufshpb_add_lru_info(struct victim_select_info *lru_info, >>>> > + struct ufshpb_region *rgn) >>>> > +{ >>>> > + rgn->rgn_state = HPB_RGN_ACTIVE; >>>> > + list_add_tail(&rgn->list_lru_rgn, &lru_info->lh_lru_rgn); >>>> > + atomic_inc(&lru_info->active_cnt); >>>> > +} >>>> > + >>>> > +static void ufshpb_hit_lru_info(struct victim_select_info *lru_info, >>>> > + struct ufshpb_region *rgn) >>>> > +{ >>>> > + list_move_tail(&rgn->list_lru_rgn, &lru_info->lh_lru_rgn); >>>> > +} >>>> > + >>>> > +static struct ufshpb_region *ufshpb_victim_lru_info(struct ufshpb_lu >>>> > *hpb) >>>> > +{ >>>> > + struct victim_select_info *lru_info = &hpb->lru_info; >>>> > + struct ufshpb_region *rgn, *victim_rgn = NULL; >>>> > + >>>> > + list_for_each_entry(rgn, &lru_info->lh_lru_rgn, list_lru_rgn) { >>>> > + if (!rgn) { >>>> > + dev_err(&hpb->sdev_ufs_lu->sdev_dev, >>>> > + "%s: no region allocated\n", >>>> > + __func__); >>>> > + return NULL; >>>> > + } >>>> > + if (ufshpb_check_srgns_issue_state(hpb, rgn)) >>>> > + continue; >>>> > + >>>> > + victim_rgn = rgn; >>>> > + break; >>>> > + } >>>> > + >>>> > + return victim_rgn; >>>> > +} >>>> > + >>>> > +static void ufshpb_cleanup_lru_info(struct victim_select_info >>>> > *lru_info, >>>> > + struct ufshpb_region *rgn) >>>> > +{ >>>> > + list_del_init(&rgn->list_lru_rgn); >>>> > + rgn->rgn_state = HPB_RGN_INACTIVE; >>>> > + atomic_dec(&lru_info->active_cnt); >>>> > +} >>>> > + >>>> > +static void ufshpb_purge_active_subregion(struct ufshpb_lu *hpb, >>>> > + struct ufshpb_subregion *srgn) >>>> > +{ >>>> > + if (srgn->srgn_state != HPB_SRGN_UNUSED) { >>>> > + ufshpb_put_map_ctx(hpb, srgn->mctx); >>>> > + srgn->srgn_state = HPB_SRGN_UNUSED; >>>> > + srgn->mctx = NULL; >>>> > + } >>>> > +} >>>> > + >>>> > +static void __ufshpb_evict_region(struct ufshpb_lu *hpb, >>>> > + struct ufshpb_region *rgn) >>>> > +{ >>>> > + struct victim_select_info *lru_info; >>>> > + struct ufshpb_subregion *srgn; >>>> > + int srgn_idx; >>>> > + >>>> > + lru_info = &hpb->lru_info; >>>> > + >>>> > + dev_dbg(&hpb->sdev_ufs_lu->sdev_dev, "evict region %d\n", >>>> > rgn->rgn_idx); >>>> > + >>>> > + ufshpb_cleanup_lru_info(lru_info, rgn); >>>> > + >>>> > + for_each_sub_region(rgn, srgn_idx, srgn) >>>> > + ufshpb_purge_active_subregion(hpb, srgn); >>>> > +} >>>> > + >>>> > +static int ufshpb_evict_region(struct ufshpb_lu *hpb, struct >>>> > ufshpb_region *rgn) >>>> > +{ >>>> > + unsigned long flags; >>>> > + int ret = 0; >>>> > + >>>> > + spin_lock_irqsave(&hpb->rgn_state_lock, flags); >>>> > + if (rgn->rgn_state == HPB_RGN_PINNED) { >>>> > + dev_warn(&hpb->sdev_ufs_lu->sdev_dev, >>>> > + "pinned region cannot drop-out. region %d\n", >>>> > + rgn->rgn_idx); >>>> > + goto out; >>>> > + } >>>> > + if (!list_empty(&rgn->list_lru_rgn)) { >>>> > + if (ufshpb_check_srgns_issue_state(hpb, rgn)) { >>>> > + ret = -EBUSY; >>>> > + goto out; >>>> > + } >>>> > + >>>> > + __ufshpb_evict_region(hpb, rgn); >>>> > + } >>>> > +out: >>>> > + spin_unlock_irqrestore(&hpb->rgn_state_lock, flags); >>>> > + return ret; >>>> > +} >>>> > + >>>> > +static int ufshpb_issue_map_req(struct ufshpb_lu *hpb, >>>> > + struct ufshpb_region *rgn, >>>> > + struct ufshpb_subregion *srgn) >>>> > +{ >>>> > + struct ufshpb_req *map_req; >>>> > + unsigned long flags; >>>> > + int ret; >>>> > + int err = -EAGAIN; >>>> > + bool alloc_required = false; >>>> > + enum HPB_SRGN_STATE state = HPB_SRGN_INVALID; >>>> > + >>>> > + spin_lock_irqsave(&hpb->rgn_state_lock, flags); >>>> > + >>>> > + if (ufshpb_get_state(hpb) != HPB_PRESENT) { >>>> > + dev_notice(&hpb->sdev_ufs_lu->sdev_dev, >>>> > + "%s: ufshpb state is not PRESENT\n", __func__); >>>> > + goto unlock_out; >>>> > + } >>>> > + >>>> > + if ((rgn->rgn_state == HPB_RGN_INACTIVE) && >>>> > + (srgn->srgn_state == HPB_SRGN_INVALID)) { >>>> > + err = 0; >>>> > + goto unlock_out; >>>> > + } >>>> > + >>>> > + if (srgn->srgn_state == HPB_SRGN_UNUSED) >>>> > + alloc_required = true; >>>> > + >>>> > + /* >>>> > + * If the subregion is already ISSUED state, >>>> > + * a specific event (e.g., GC or wear-leveling, etc.) occurs in >>>> > + * the device and HPB response for map loading is received. >>>> > + * In this case, after finishing the HPB_READ_BUFFER, >>>> > + * the next HPB_READ_BUFFER is performed again to obtain the latest >>>> > + * map data. >>>> > + */ >>>> > + if (srgn->srgn_state == HPB_SRGN_ISSUED) >>>> > + goto unlock_out; >>>> > + >>>> > + srgn->srgn_state = HPB_SRGN_ISSUED; >>>> > + spin_unlock_irqrestore(&hpb->rgn_state_lock, flags); >>>> > + >>>> > + if (alloc_required) { >>>> > + if (srgn->mctx) { >>>> >>>> Can this really happen? >>>> >>>> alloc_required is true if srgn->srgn_state == HPB_SRGN_UNUSED. >>>> ufshpb_put_map_ctx() is called and srgn->srgn_state is given >>>> HPB_SRGN_UNUSED together in ufshpb_purge_active_subregion(). >>>> Do I miss anything? >>> >>> It is not normal case, but it is for preventing memory leak on >>> abnormal >>> case. >>> >>> Thanks, >>> Daejun >> >> Abnormal how? Could you please give an example? > > It is kind of additional memory leak protection. Do you think it is > unnecessary? If this can never happen in this code base, then yes it is unnecessary, since it is dead code. It seems to me like a debug print left there in early development stage of this change. I am just curious about what memory leak can happen here. If you are not confident, then please leave it. Thanks, Can Guo. > > Thanks, > Daejun
>>>>> > This is a patch for managing L2P map in HPB module. >>>>> > >>>>> > The HPB divides logical addresses into several regions. A region >>>>> > consists >>>>> > of several sub-regions. The sub-region is a basic unit where L2P >>>>> > mapping is >>>>> > managed. The driver loads L2P mapping data of each sub-region. The >>>>> > loaded >>>>> > sub-region is called active-state. The HPB driver unloads L2P mapping >>>>> > data >>>>> > as region unit. The unloaded region is called inactive-state. >>>>> > >>>>> > Sub-region/region candidates to be loaded and unloaded are delivered >>>>> > from >>>>> > the UFS device. The UFS device delivers the recommended active >>>>> > sub-region >>>>> > and inactivate region to the driver using sensedata. >>>>> > The HPB module performs L2P mapping management on the host through the >>>>> > delivered information. >>>>> > >>>>> > A pinned region is a pre-set regions on the UFS device that is always >>>>> > activate-state. >>>>> > >>>>> > The data structure for map data request and L2P map uses mempool API, >>>>> > minimizing allocation overhead while avoiding static allocation. >>>>> > >>>>> > The mininum size of the memory pool used in the HPB is implemented >>>>> > as a module parameter, so that it can be configurable by the user. >>>>> > >>>>> > To gurantee a minimum memory pool size of 4MB: >>>>> > ufshpb_host_map_kbytes=4096 >>>>> > >>>>> > The map_work manages active/inactive by 2 "to-do" lists. >>>>> > Each hpb lun maintains 2 "to-do" lists: >>>>> > hpb->lh_inact_rgn - regions to be inactivated, and >>>>> > hpb->lh_act_srgn - subregions to be activated >>>>> > Those lists are maintained on IO completion. >>>>> > >>>>> > Reviewed-by: Bart Van Assche <bvanassche@acm.org> >>>>> > Reviewed-by: Can Guo <cang@codeaurora.org> >>>>> > Acked-by: Avri Altman <Avri.Altman@wdc.com> >>>>> > Tested-by: Bean Huo <beanhuo@micron.com> >>>>> > Signed-off-by: Daejun Park <daejun7.park@samsung.com> >>>>> > --- >>>>> > drivers/scsi/ufs/ufs.h | 36 ++ >>>>> > drivers/scsi/ufs/ufshcd.c | 4 + >>>>> > drivers/scsi/ufs/ufshpb.c | 1091 ++++++++++++++++++++++++++++++++++++- >>>>> > drivers/scsi/ufs/ufshpb.h | 65 +++ >>>>> > 4 files changed, 1181 insertions(+), 15 deletions(-) >>>>> > >>>>> > diff --git a/drivers/scsi/ufs/ufs.h b/drivers/scsi/ufs/ufs.h >>>>> > index 65563635e20e..957763db1006 100644 >>>>> > --- a/drivers/scsi/ufs/ufs.h >>>>> > +++ b/drivers/scsi/ufs/ufs.h >>>>> > @@ -472,6 +472,41 @@ struct utp_cmd_rsp { >>>>> > u8 sense_data[UFS_SENSE_SIZE]; >>>>> > }; >>>>> > >>>>> > +struct ufshpb_active_field { >>>>> > + __be16 active_rgn; >>>>> > + __be16 active_srgn; >>>>> > +}; >>>>> > +#define HPB_ACT_FIELD_SIZE 4 >>>>> > + >>>>> > +/** >>>>> > + * struct utp_hpb_rsp - Response UPIU structure >>>>> > + * @residual_transfer_count: Residual transfer count DW-3 >>>>> > + * @reserved1: Reserved double words DW-4 to DW-7 >>>>> > + * @sense_data_len: Sense data length DW-8 U16 >>>>> > + * @desc_type: Descriptor type of sense data >>>>> > + * @additional_len: Additional length of sense data >>>>> > + * @hpb_op: HPB operation type >>>>> > + * @lun: LUN of response UPIU >>>>> > + * @active_rgn_cnt: Active region count >>>>> > + * @inactive_rgn_cnt: Inactive region count >>>>> > + * @hpb_active_field: Recommended to read HPB region and subregion >>>>> > + * @hpb_inactive_field: To be inactivated HPB region and subregion >>>>> > + */ >>>>> > +struct utp_hpb_rsp { >>>>> > + __be32 residual_transfer_count; >>>>> > + __be32 reserved1[4]; >>>>> > + __be16 sense_data_len; >>>>> > + u8 desc_type; >>>>> > + u8 additional_len; >>>>> > + u8 hpb_op; >>>>> > + u8 lun; >>>>> > + u8 active_rgn_cnt; >>>>> > + u8 inactive_rgn_cnt; >>>>> > + struct ufshpb_active_field hpb_active_field[2]; >>>>> > + __be16 hpb_inactive_field[2]; >>>>> > +}; >>>>> > +#define UTP_HPB_RSP_SIZE 40 >>>>> > + >>>>> > /** >>>>> > * struct utp_upiu_rsp - general upiu response structure >>>>> > * @header: UPIU header structure DW-0 to DW-2 >>>>> > @@ -482,6 +517,7 @@ struct utp_upiu_rsp { >>>>> > struct utp_upiu_header header; >>>>> > union { >>>>> > struct utp_cmd_rsp sr; >>>>> > + struct utp_hpb_rsp hr; >>>>> > struct utp_upiu_query qr; >>>>> > }; >>>>> > }; >>>>> > diff --git a/drivers/scsi/ufs/ufshcd.c b/drivers/scsi/ufs/ufshcd.c >>>>> > index 49b3d5d24fa6..5852ff44c3cc 100644 >>>>> > --- a/drivers/scsi/ufs/ufshcd.c >>>>> > +++ b/drivers/scsi/ufs/ufshcd.c >>>>> > @@ -5021,6 +5021,9 @@ ufshcd_transfer_rsp_status(struct ufs_hba *hba, >>>>> > struct ufshcd_lrb *lrbp) >>>>> > */ >>>>> > pm_runtime_get_noresume(hba->dev); >>>>> > } >>>>> > + >>>>> > + if (scsi_status == SAM_STAT_GOOD) >>>>> > + ufshpb_rsp_upiu(hba, lrbp); >>>>> > break; >>>>> > case UPIU_TRANSACTION_REJECT_UPIU: >>>>> > /* TODO: handle Reject UPIU Response */ >>>>> > @@ -9221,6 +9224,7 @@ EXPORT_SYMBOL(ufshcd_shutdown); >>>>> > void ufshcd_remove(struct ufs_hba *hba) >>>>> > { >>>>> > ufs_bsg_remove(hba); >>>>> > + ufshpb_remove(hba); >>>>> > ufs_sysfs_remove_nodes(hba->dev); >>>>> > blk_cleanup_queue(hba->tmf_queue); >>>>> > blk_mq_free_tag_set(&hba->tmf_tag_set); >>>>> > diff --git a/drivers/scsi/ufs/ufshpb.c b/drivers/scsi/ufs/ufshpb.c >>>>> > index 1a72f6541510..8abadb0e010a 100644 >>>>> > --- a/drivers/scsi/ufs/ufshpb.c >>>>> > +++ b/drivers/scsi/ufs/ufshpb.c >>>>> > @@ -16,6 +16,16 @@ >>>>> > #include "ufshpb.h" >>>>> > #include "../sd.h" >>>>> > >>>>> > +/* memory management */ >>>>> > +static struct kmem_cache *ufshpb_mctx_cache; >>>>> > +static mempool_t *ufshpb_mctx_pool; >>>>> > +static mempool_t *ufshpb_page_pool; >>>>> > +/* A cache size of 2MB can cache ppn in the 1GB range. */ >>>>> > +static unsigned int ufshpb_host_map_kbytes = 2048; >>>>> > +static int tot_active_srgn_pages; >>>>> > + >>>>> > +static struct workqueue_struct *ufshpb_wq; >>>>> > + >>>>> > bool ufshpb_is_allowed(struct ufs_hba *hba) >>>>> > { >>>>> > return !(hba->ufshpb_dev.hpb_disabled); >>>>> > @@ -36,14 +46,892 @@ static void ufshpb_set_state(struct ufshpb_lu >>>>> > *hpb, int state) >>>>> > atomic_set(&hpb->hpb_state, state); >>>>> > } >>>>> > >>>>> > +static bool ufshpb_is_general_lun(int lun) >>>>> > +{ >>>>> > + return lun < UFS_UPIU_MAX_UNIT_NUM_ID; >>>>> > +} >>>>> > + >>>>> > +static bool >>>>> > +ufshpb_is_pinned_region(struct ufshpb_lu *hpb, int rgn_idx) >>>>> > +{ >>>>> > + if (hpb->lu_pinned_end != PINNED_NOT_SET && >>>>> > + rgn_idx >= hpb->lu_pinned_start && >>>>> > + rgn_idx <= hpb->lu_pinned_end) >>>>> > + return true; >>>>> > + >>>>> > + return false; >>>>> > +} >>>>> > + >>>>> > +static void ufshpb_kick_map_work(struct ufshpb_lu *hpb) >>>>> > +{ >>>>> > + bool ret = false; >>>>> > + unsigned long flags; >>>>> > + >>>>> > + if (ufshpb_get_state(hpb) != HPB_PRESENT) >>>>> > + return; >>>>> > + >>>>> > + spin_lock_irqsave(&hpb->rsp_list_lock, flags); >>>>> > + if (!list_empty(&hpb->lh_inact_rgn) || >>>>> > !list_empty(&hpb->lh_act_srgn)) >>>>> > + ret = true; >>>>> > + spin_unlock_irqrestore(&hpb->rsp_list_lock, flags); >>>>> > + >>>>> > + if (ret) >>>>> > + queue_work(ufshpb_wq, &hpb->map_work); >>>>> > +} >>>>> > + >>>>> > +static bool ufshpb_is_hpb_rsp_valid(struct ufs_hba *hba, >>>>> > + struct ufshcd_lrb *lrbp, >>>>> > + struct utp_hpb_rsp *rsp_field) >>>>> > +{ >>>>> > + /* Check HPB_UPDATE_ALERT */ >>>>> > + if (!(lrbp->ucd_rsp_ptr->header.dword_2 & >>>>> > + UPIU_HEADER_DWORD(0, 2, 0, 0))) >>>>> > + return false; >>>>> > + >>>>> > + if (be16_to_cpu(rsp_field->sense_data_len) != DEV_SENSE_SEG_LEN || >>>>> > + rsp_field->desc_type != DEV_DES_TYPE || >>>>> > + rsp_field->additional_len != DEV_ADDITIONAL_LEN || >>>>> > + rsp_field->active_rgn_cnt > MAX_ACTIVE_NUM || >>>>> > + rsp_field->inactive_rgn_cnt > MAX_INACTIVE_NUM || >>>>> > + rsp_field->hpb_op == HPB_RSP_NONE || >>>>> > + (rsp_field->hpb_op == HPB_RSP_REQ_REGION_UPDATE && >>>>> > + !rsp_field->active_rgn_cnt && !rsp_field->inactive_rgn_cnt)) >>>>> > + return false; >>>>> > + >>>>> > + if (!ufshpb_is_general_lun(rsp_field->lun)) { >>>>> > + dev_warn(hba->dev, "ufshpb: lun(%d) not supported\n", >>>>> > + lrbp->lun); >>>>> > + return false; >>>>> > + } >>>>> > + >>>>> > + return true; >>>>> > +} >>>>> > + >>>>> > +static struct ufshpb_req *ufshpb_get_map_req(struct ufshpb_lu *hpb, >>>>> > + struct ufshpb_subregion *srgn) >>>>> > +{ >>>>> > + struct ufshpb_req *map_req; >>>>> > + struct request *req; >>>>> > + struct bio *bio; >>>>> > + int retries = HPB_MAP_REQ_RETRIES; >>>>> > + >>>>> > + map_req = kmem_cache_alloc(hpb->map_req_cache, GFP_KERNEL); >>>>> > + if (!map_req) >>>>> > + return NULL; >>>>> > + >>>>> > +retry: >>>>> > + req = blk_get_request(hpb->sdev_ufs_lu->request_queue, >>>>> > + REQ_OP_SCSI_IN, BLK_MQ_REQ_NOWAIT); >>>>> > + >>>>> > + if ((PTR_ERR(req) == -EWOULDBLOCK) && (--retries > 0)) { >>>>> > + usleep_range(3000, 3100); >>>>> > + goto retry; >>>>> > + } >>>>> > + >>>>> > + if (IS_ERR(req)) >>>>> > + goto free_map_req; >>>>> > + >>>>> > + bio = bio_alloc(GFP_KERNEL, hpb->pages_per_srgn); >>>>> > + if (!bio) { >>>>> > + blk_put_request(req); >>>>> > + goto free_map_req; >>>>> > + } >>>>> > + >>>>> > + map_req->hpb = hpb; >>>>> > + map_req->req = req; >>>>> > + map_req->bio = bio; >>>>> > + >>>>> > + map_req->rgn_idx = srgn->rgn_idx; >>>>> > + map_req->srgn_idx = srgn->srgn_idx; >>>>> > + map_req->mctx = srgn->mctx; >>>>> > + >>>>> > + return map_req; >>>>> > + >>>>> > +free_map_req: >>>>> > + kmem_cache_free(hpb->map_req_cache, map_req); >>>>> > + return NULL; >>>>> > +} >>>>> > + >>>>> > +static void ufshpb_put_map_req(struct ufshpb_lu *hpb, >>>>> > + struct ufshpb_req *map_req) >>>>> > +{ >>>>> > + bio_put(map_req->bio); >>>>> > + blk_put_request(map_req->req); >>>>> > + kmem_cache_free(hpb->map_req_cache, map_req); >>>>> > +} >>>>> > + >>>>> > +static int ufshpb_clear_dirty_bitmap(struct ufshpb_lu *hpb, >>>>> > + struct ufshpb_subregion *srgn) >>>>> > +{ >>>>> > + u32 num_entries = hpb->entries_per_srgn; >>>>> > + >>>>> > + if (!srgn->mctx) { >>>>> > + dev_err(&hpb->sdev_ufs_lu->sdev_dev, >>>>> > + "no mctx in region %d subregion %d.\n", >>>>> > + srgn->rgn_idx, srgn->srgn_idx); >>>>> > + return -1; >>>>> > + } >>>>> > + >>>>> > + if (unlikely(srgn->is_last)) >>>>> > + num_entries = hpb->last_srgn_entries; >>>>> > + >>>>> > + bitmap_zero(srgn->mctx->ppn_dirty, num_entries); >>>>> > + return 0; >>>>> > +} >>>>> > + >>>>> > +static void ufshpb_update_active_info(struct ufshpb_lu *hpb, int >>>>> > rgn_idx, >>>>> > + int srgn_idx) >>>>> > +{ >>>>> > + struct ufshpb_region *rgn; >>>>> > + struct ufshpb_subregion *srgn; >>>>> > + >>>>> > + rgn = hpb->rgn_tbl + rgn_idx; >>>>> > + srgn = rgn->srgn_tbl + srgn_idx; >>>>> > + >>>>> > + list_del_init(&rgn->list_inact_rgn); >>>>> > + >>>>> > + if (list_empty(&srgn->list_act_srgn)) >>>>> > + list_add_tail(&srgn->list_act_srgn, &hpb->lh_act_srgn); >>>>> > +} >>>>> > + >>>>> > +static void ufshpb_update_inactive_info(struct ufshpb_lu *hpb, int >>>>> > rgn_idx) >>>>> > +{ >>>>> > + struct ufshpb_region *rgn; >>>>> > + struct ufshpb_subregion *srgn; >>>>> > + int srgn_idx; >>>>> > + >>>>> > + rgn = hpb->rgn_tbl + rgn_idx; >>>>> > + >>>>> > + for_each_sub_region(rgn, srgn_idx, srgn) >>>>> > + list_del_init(&srgn->list_act_srgn); >>>>> > + >>>>> > + if (list_empty(&rgn->list_inact_rgn)) >>>>> > + list_add_tail(&rgn->list_inact_rgn, &hpb->lh_inact_rgn); >>>>> > +} >>>>> > + >>>>> > +static void ufshpb_activate_subregion(struct ufshpb_lu *hpb, >>>>> > + struct ufshpb_subregion *srgn) >>>>> > +{ >>>>> > + struct ufshpb_region *rgn; >>>>> > + >>>>> > + /* >>>>> > + * If there is no mctx in subregion >>>>> > + * after I/O progress for HPB_READ_BUFFER, the region to which the >>>>> > + * subregion belongs was evicted. >>>>> > + * Make sure the region must not evict in I/O progress >>>>> > + */ >>>>> > + if (!srgn->mctx) { >>>>> > + dev_err(&hpb->sdev_ufs_lu->sdev_dev, >>>>> > + "no mctx in region %d subregion %d.\n", >>>>> > + srgn->rgn_idx, srgn->srgn_idx); >>>>> > + srgn->srgn_state = HPB_SRGN_INVALID; >>>>> > + return; >>>>> > + } >>>>> > + >>>>> > + rgn = hpb->rgn_tbl + srgn->rgn_idx; >>>>> > + >>>>> > + if (unlikely(rgn->rgn_state == HPB_RGN_INACTIVE)) { >>>>> > + dev_err(&hpb->sdev_ufs_lu->sdev_dev, >>>>> > + "region %d subregion %d evicted\n", >>>>> > + srgn->rgn_idx, srgn->srgn_idx); >>>>> > + srgn->srgn_state = HPB_SRGN_INVALID; >>>>> > + return; >>>>> > + } >>>>> > + srgn->srgn_state = HPB_SRGN_VALID; >>>>> > +} >>>>> > + >>>>> > +static void ufshpb_map_req_compl_fn(struct request *req, blk_status_t >>>>> > error) >>>>> > +{ >>>>> > + struct ufshpb_req *map_req = (struct ufshpb_req *) req->end_io_data; >>>>> > + struct ufshpb_lu *hpb = map_req->hpb; >>>>> > + struct ufshpb_subregion *srgn; >>>>> > + unsigned long flags; >>>>> > + >>>>> > + srgn = hpb->rgn_tbl[map_req->rgn_idx].srgn_tbl + >>>>> > + map_req->srgn_idx; >>>>> > + >>>>> > + ufshpb_clear_dirty_bitmap(hpb, srgn); >>>>> > + spin_lock_irqsave(&hpb->rgn_state_lock, flags); >>>>> > + ufshpb_activate_subregion(hpb, srgn); >>>>> > + spin_unlock_irqrestore(&hpb->rgn_state_lock, flags); >>>>> > + >>>>> > + ufshpb_put_map_req(map_req->hpb, map_req); >>>>> > +} >>>>> > + >>>>> > +static void ufshpb_set_read_buf_cmd(unsigned char *cdb, int rgn_idx, >>>>> > + int srgn_idx, int srgn_mem_size) >>>>> > +{ >>>>> > + cdb[0] = UFSHPB_READ_BUFFER; >>>>> > + cdb[1] = UFSHPB_READ_BUFFER_ID; >>>>> > + >>>>> > + put_unaligned_be16(rgn_idx, &cdb[2]); >>>>> > + put_unaligned_be16(srgn_idx, &cdb[4]); >>>>> > + put_unaligned_be24(srgn_mem_size, &cdb[6]); >>>>> > + >>>>> > + cdb[9] = 0x00; >>>>> > +} >>>>> > + >>>>> > +static int ufshpb_execute_map_req(struct ufshpb_lu *hpb, >>>>> > + struct ufshpb_req *map_req, bool last) >>>>> > +{ >>>>> > + struct request_queue *q; >>>>> > + struct request *req; >>>>> > + struct scsi_request *rq; >>>>> > + int mem_size = hpb->srgn_mem_size; >>>>> > + int ret = 0; >>>>> > + int i; >>>>> > + >>>>> > + q = hpb->sdev_ufs_lu->request_queue; >>>>> > + for (i = 0; i < hpb->pages_per_srgn; i++) { >>>>> > + ret = bio_add_pc_page(q, map_req->bio, map_req->mctx->m_page[i], >>>>> > + PAGE_SIZE, 0); >>>>> > + if (ret != PAGE_SIZE) { >>>>> > + dev_err(&hpb->sdev_ufs_lu->sdev_dev, >>>>> > + "bio_add_pc_page fail %d - %d\n", >>>>> > + map_req->rgn_idx, map_req->srgn_idx); >>>>> > + return ret; >>>>> > + } >>>>> > + } >>>>> > + >>>>> > + req = map_req->req; >>>>> > + >>>>> > + blk_rq_append_bio(req, &map_req->bio); >>>>> > + >>>>> > + req->end_io_data = map_req; >>>>> > + >>>>> > + rq = scsi_req(req); >>>>> > + >>>>> > + if (unlikely(last)) >>>>> > + mem_size = hpb->last_srgn_entries * HPB_ENTRY_SIZE; >>>>> > + >>>>> > + ufshpb_set_read_buf_cmd(rq->cmd, map_req->rgn_idx, >>>>> > + map_req->srgn_idx, mem_size); >>>>> > + rq->cmd_len = HPB_READ_BUFFER_CMD_LENGTH; >>>>> > + >>>>> > + blk_execute_rq_nowait(q, NULL, req, 1, ufshpb_map_req_compl_fn); >>>>> > + >>>>> > + hpb->stats.map_req_cnt++; >>>>> > + return 0; >>>>> > +} >>>>> > + >>>>> > +static struct ufshpb_map_ctx *ufshpb_get_map_ctx(struct ufshpb_lu >>>>> > *hpb, >>>>> > + bool last) >>>>> > +{ >>>>> > + struct ufshpb_map_ctx *mctx; >>>>> > + u32 num_entries = hpb->entries_per_srgn; >>>>> > + int i, j; >>>>> > + >>>>> > + mctx = mempool_alloc(ufshpb_mctx_pool, GFP_KERNEL); >>>>> > + if (!mctx) >>>>> > + return NULL; >>>>> > + >>>>> > + mctx->m_page = kmem_cache_alloc(hpb->m_page_cache, GFP_KERNEL); >>>>> > + if (!mctx->m_page) >>>>> > + goto release_mctx; >>>>> > + >>>>> > + if (unlikely(last)) >>>>> > + num_entries = hpb->last_srgn_entries; >>>>> > + >>>>> > + mctx->ppn_dirty = bitmap_zalloc(num_entries, GFP_KERNEL); >>>>> > + if (!mctx->ppn_dirty) >>>>> > + goto release_m_page; >>>>> > + >>>>> > + for (i = 0; i < hpb->pages_per_srgn; i++) { >>>>> > + mctx->m_page[i] = mempool_alloc(ufshpb_page_pool, GFP_KERNEL); >>>>> > + if (!mctx->m_page[i]) { >>>>> > + for (j = 0; j < i; j++) >>>>> > + mempool_free(mctx->m_page[j], ufshpb_page_pool); >>>>> > + goto release_ppn_dirty; >>>>> > + } >>>>> > + clear_page(page_address(mctx->m_page[i])); >>>>> > + } >>>>> > + >>>>> > + return mctx; >>>>> > + >>>>> > +release_ppn_dirty: >>>>> > + bitmap_free(mctx->ppn_dirty); >>>>> > +release_m_page: >>>>> > + kmem_cache_free(hpb->m_page_cache, mctx->m_page); >>>>> > +release_mctx: >>>>> > + mempool_free(mctx, ufshpb_mctx_pool); >>>>> > + return NULL; >>>>> > +} >>>>> > + >>>>> > +static void ufshpb_put_map_ctx(struct ufshpb_lu *hpb, >>>>> > + struct ufshpb_map_ctx *mctx) >>>>> > +{ >>>>> > + int i; >>>>> > + >>>>> > + for (i = 0; i < hpb->pages_per_srgn; i++) >>>>> > + mempool_free(mctx->m_page[i], ufshpb_page_pool); >>>>> > + >>>>> > + bitmap_free(mctx->ppn_dirty); >>>>> > + kmem_cache_free(hpb->m_page_cache, mctx->m_page); >>>>> > + mempool_free(mctx, ufshpb_mctx_pool); >>>>> > +} >>>>> > + >>>>> > +static int ufshpb_check_srgns_issue_state(struct ufshpb_lu *hpb, >>>>> > + struct ufshpb_region *rgn) >>>>> > +{ >>>>> > + struct ufshpb_subregion *srgn; >>>>> > + int srgn_idx; >>>>> > + >>>>> > + for_each_sub_region(rgn, srgn_idx, srgn) >>>>> > + if (srgn->srgn_state == HPB_SRGN_ISSUED) >>>>> > + return -EPERM; >>>>> > + >>>>> > + return 0; >>>>> > +} >>>>> > + >>>>> > +static void ufshpb_add_lru_info(struct victim_select_info *lru_info, >>>>> > + struct ufshpb_region *rgn) >>>>> > +{ >>>>> > + rgn->rgn_state = HPB_RGN_ACTIVE; >>>>> > + list_add_tail(&rgn->list_lru_rgn, &lru_info->lh_lru_rgn); >>>>> > + atomic_inc(&lru_info->active_cnt); >>>>> > +} >>>>> > + >>>>> > +static void ufshpb_hit_lru_info(struct victim_select_info *lru_info, >>>>> > + struct ufshpb_region *rgn) >>>>> > +{ >>>>> > + list_move_tail(&rgn->list_lru_rgn, &lru_info->lh_lru_rgn); >>>>> > +} >>>>> > + >>>>> > +static struct ufshpb_region *ufshpb_victim_lru_info(struct ufshpb_lu >>>>> > *hpb) >>>>> > +{ >>>>> > + struct victim_select_info *lru_info = &hpb->lru_info; >>>>> > + struct ufshpb_region *rgn, *victim_rgn = NULL; >>>>> > + >>>>> > + list_for_each_entry(rgn, &lru_info->lh_lru_rgn, list_lru_rgn) { >>>>> > + if (!rgn) { >>>>> > + dev_err(&hpb->sdev_ufs_lu->sdev_dev, >>>>> > + "%s: no region allocated\n", >>>>> > + __func__); >>>>> > + return NULL; >>>>> > + } >>>>> > + if (ufshpb_check_srgns_issue_state(hpb, rgn)) >>>>> > + continue; >>>>> > + >>>>> > + victim_rgn = rgn; >>>>> > + break; >>>>> > + } >>>>> > + >>>>> > + return victim_rgn; >>>>> > +} >>>>> > + >>>>> > +static void ufshpb_cleanup_lru_info(struct victim_select_info >>>>> > *lru_info, >>>>> > + struct ufshpb_region *rgn) >>>>> > +{ >>>>> > + list_del_init(&rgn->list_lru_rgn); >>>>> > + rgn->rgn_state = HPB_RGN_INACTIVE; >>>>> > + atomic_dec(&lru_info->active_cnt); >>>>> > +} >>>>> > + >>>>> > +static void ufshpb_purge_active_subregion(struct ufshpb_lu *hpb, >>>>> > + struct ufshpb_subregion *srgn) >>>>> > +{ >>>>> > + if (srgn->srgn_state != HPB_SRGN_UNUSED) { >>>>> > + ufshpb_put_map_ctx(hpb, srgn->mctx); >>>>> > + srgn->srgn_state = HPB_SRGN_UNUSED; >>>>> > + srgn->mctx = NULL; >>>>> > + } >>>>> > +} >>>>> > + >>>>> > +static void __ufshpb_evict_region(struct ufshpb_lu *hpb, >>>>> > + struct ufshpb_region *rgn) >>>>> > +{ >>>>> > + struct victim_select_info *lru_info; >>>>> > + struct ufshpb_subregion *srgn; >>>>> > + int srgn_idx; >>>>> > + >>>>> > + lru_info = &hpb->lru_info; >>>>> > + >>>>> > + dev_dbg(&hpb->sdev_ufs_lu->sdev_dev, "evict region %d\n", >>>>> > rgn->rgn_idx); >>>>> > + >>>>> > + ufshpb_cleanup_lru_info(lru_info, rgn); >>>>> > + >>>>> > + for_each_sub_region(rgn, srgn_idx, srgn) >>>>> > + ufshpb_purge_active_subregion(hpb, srgn); >>>>> > +} >>>>> > + >>>>> > +static int ufshpb_evict_region(struct ufshpb_lu *hpb, struct >>>>> > ufshpb_region *rgn) >>>>> > +{ >>>>> > + unsigned long flags; >>>>> > + int ret = 0; >>>>> > + >>>>> > + spin_lock_irqsave(&hpb->rgn_state_lock, flags); >>>>> > + if (rgn->rgn_state == HPB_RGN_PINNED) { >>>>> > + dev_warn(&hpb->sdev_ufs_lu->sdev_dev, >>>>> > + "pinned region cannot drop-out. region %d\n", >>>>> > + rgn->rgn_idx); >>>>> > + goto out; >>>>> > + } >>>>> > + if (!list_empty(&rgn->list_lru_rgn)) { >>>>> > + if (ufshpb_check_srgns_issue_state(hpb, rgn)) { >>>>> > + ret = -EBUSY; >>>>> > + goto out; >>>>> > + } >>>>> > + >>>>> > + __ufshpb_evict_region(hpb, rgn); >>>>> > + } >>>>> > +out: >>>>> > + spin_unlock_irqrestore(&hpb->rgn_state_lock, flags); >>>>> > + return ret; >>>>> > +} >>>>> > + >>>>> > +static int ufshpb_issue_map_req(struct ufshpb_lu *hpb, >>>>> > + struct ufshpb_region *rgn, >>>>> > + struct ufshpb_subregion *srgn) >>>>> > +{ >>>>> > + struct ufshpb_req *map_req; >>>>> > + unsigned long flags; >>>>> > + int ret; >>>>> > + int err = -EAGAIN; >>>>> > + bool alloc_required = false; >>>>> > + enum HPB_SRGN_STATE state = HPB_SRGN_INVALID; >>>>> > + >>>>> > + spin_lock_irqsave(&hpb->rgn_state_lock, flags); >>>>> > + >>>>> > + if (ufshpb_get_state(hpb) != HPB_PRESENT) { >>>>> > + dev_notice(&hpb->sdev_ufs_lu->sdev_dev, >>>>> > + "%s: ufshpb state is not PRESENT\n", __func__); >>>>> > + goto unlock_out; >>>>> > + } >>>>> > + >>>>> > + if ((rgn->rgn_state == HPB_RGN_INACTIVE) && >>>>> > + (srgn->srgn_state == HPB_SRGN_INVALID)) { >>>>> > + err = 0; >>>>> > + goto unlock_out; >>>>> > + } >>>>> > + >>>>> > + if (srgn->srgn_state == HPB_SRGN_UNUSED) >>>>> > + alloc_required = true; >>>>> > + >>>>> > + /* >>>>> > + * If the subregion is already ISSUED state, >>>>> > + * a specific event (e.g., GC or wear-leveling, etc.) occurs in >>>>> > + * the device and HPB response for map loading is received. >>>>> > + * In this case, after finishing the HPB_READ_BUFFER, >>>>> > + * the next HPB_READ_BUFFER is performed again to obtain the latest >>>>> > + * map data. >>>>> > + */ >>>>> > + if (srgn->srgn_state == HPB_SRGN_ISSUED) >>>>> > + goto unlock_out; >>>>> > + >>>>> > + srgn->srgn_state = HPB_SRGN_ISSUED; >>>>> > + spin_unlock_irqrestore(&hpb->rgn_state_lock, flags); >>>>> > + >>>>> > + if (alloc_required) { >>>>> > + if (srgn->mctx) { >>>>> >>>>> Can this really happen? >>>>> >>>>> alloc_required is true if srgn->srgn_state == HPB_SRGN_UNUSED. >>>>> ufshpb_put_map_ctx() is called and srgn->srgn_state is given >>>>> HPB_SRGN_UNUSED together in ufshpb_purge_active_subregion(). >>>>> Do I miss anything? >>>> >>>> It is not normal case, but it is for preventing memory leak on >>>> abnormal >>>> case. >>>> >>>> Thanks, >>>> Daejun >>> >>> Abnormal how? Could you please give an example? >> >> It is kind of additional memory leak protection. Do you think it is >> unnecessary? > >If this can never happen in this code base, then yes it is unnecessary, >since it is dead code. It seems to me like a debug print left there in >early development stage of this change. I am just curious about what >memory leak can happen here. If you are not confident, then please leave >it. OK, I will delete it on next patch. Thanks, Daejun >Thanks, >Can Guo. > >> >> Thanks, >> Daejun > > >
diff --git a/drivers/scsi/ufs/ufs.h b/drivers/scsi/ufs/ufs.h index 65563635e20e..957763db1006 100644 --- a/drivers/scsi/ufs/ufs.h +++ b/drivers/scsi/ufs/ufs.h @@ -472,6 +472,41 @@ struct utp_cmd_rsp { u8 sense_data[UFS_SENSE_SIZE]; }; +struct ufshpb_active_field { + __be16 active_rgn; + __be16 active_srgn; +}; +#define HPB_ACT_FIELD_SIZE 4 + +/** + * struct utp_hpb_rsp - Response UPIU structure + * @residual_transfer_count: Residual transfer count DW-3 + * @reserved1: Reserved double words DW-4 to DW-7 + * @sense_data_len: Sense data length DW-8 U16 + * @desc_type: Descriptor type of sense data + * @additional_len: Additional length of sense data + * @hpb_op: HPB operation type + * @lun: LUN of response UPIU + * @active_rgn_cnt: Active region count + * @inactive_rgn_cnt: Inactive region count + * @hpb_active_field: Recommended to read HPB region and subregion + * @hpb_inactive_field: To be inactivated HPB region and subregion + */ +struct utp_hpb_rsp { + __be32 residual_transfer_count; + __be32 reserved1[4]; + __be16 sense_data_len; + u8 desc_type; + u8 additional_len; + u8 hpb_op; + u8 lun; + u8 active_rgn_cnt; + u8 inactive_rgn_cnt; + struct ufshpb_active_field hpb_active_field[2]; + __be16 hpb_inactive_field[2]; +}; +#define UTP_HPB_RSP_SIZE 40 + /** * struct utp_upiu_rsp - general upiu response structure * @header: UPIU header structure DW-0 to DW-2 @@ -482,6 +517,7 @@ struct utp_upiu_rsp { struct utp_upiu_header header; union { struct utp_cmd_rsp sr; + struct utp_hpb_rsp hr; struct utp_upiu_query qr; }; }; diff --git a/drivers/scsi/ufs/ufshcd.c b/drivers/scsi/ufs/ufshcd.c index 49b3d5d24fa6..5852ff44c3cc 100644 --- a/drivers/scsi/ufs/ufshcd.c +++ b/drivers/scsi/ufs/ufshcd.c @@ -5021,6 +5021,9 @@ ufshcd_transfer_rsp_status(struct ufs_hba *hba, struct ufshcd_lrb *lrbp) */ pm_runtime_get_noresume(hba->dev); } + + if (scsi_status == SAM_STAT_GOOD) + ufshpb_rsp_upiu(hba, lrbp); break; case UPIU_TRANSACTION_REJECT_UPIU: /* TODO: handle Reject UPIU Response */ @@ -9221,6 +9224,7 @@ EXPORT_SYMBOL(ufshcd_shutdown); void ufshcd_remove(struct ufs_hba *hba) { ufs_bsg_remove(hba); + ufshpb_remove(hba); ufs_sysfs_remove_nodes(hba->dev); blk_cleanup_queue(hba->tmf_queue); blk_mq_free_tag_set(&hba->tmf_tag_set); diff --git a/drivers/scsi/ufs/ufshpb.c b/drivers/scsi/ufs/ufshpb.c index 1a72f6541510..8abadb0e010a 100644 --- a/drivers/scsi/ufs/ufshpb.c +++ b/drivers/scsi/ufs/ufshpb.c @@ -16,6 +16,16 @@ #include "ufshpb.h" #include "../sd.h" +/* memory management */ +static struct kmem_cache *ufshpb_mctx_cache; +static mempool_t *ufshpb_mctx_pool; +static mempool_t *ufshpb_page_pool; +/* A cache size of 2MB can cache ppn in the 1GB range. */ +static unsigned int ufshpb_host_map_kbytes = 2048; +static int tot_active_srgn_pages; + +static struct workqueue_struct *ufshpb_wq; + bool ufshpb_is_allowed(struct ufs_hba *hba) { return !(hba->ufshpb_dev.hpb_disabled); @@ -36,14 +46,892 @@ static void ufshpb_set_state(struct ufshpb_lu *hpb, int state) atomic_set(&hpb->hpb_state, state); } +static bool ufshpb_is_general_lun(int lun) +{ + return lun < UFS_UPIU_MAX_UNIT_NUM_ID; +} + +static bool +ufshpb_is_pinned_region(struct ufshpb_lu *hpb, int rgn_idx) +{ + if (hpb->lu_pinned_end != PINNED_NOT_SET && + rgn_idx >= hpb->lu_pinned_start && + rgn_idx <= hpb->lu_pinned_end) + return true; + + return false; +} + +static void ufshpb_kick_map_work(struct ufshpb_lu *hpb) +{ + bool ret = false; + unsigned long flags; + + if (ufshpb_get_state(hpb) != HPB_PRESENT) + return; + + spin_lock_irqsave(&hpb->rsp_list_lock, flags); + if (!list_empty(&hpb->lh_inact_rgn) || !list_empty(&hpb->lh_act_srgn)) + ret = true; + spin_unlock_irqrestore(&hpb->rsp_list_lock, flags); + + if (ret) + queue_work(ufshpb_wq, &hpb->map_work); +} + +static bool ufshpb_is_hpb_rsp_valid(struct ufs_hba *hba, + struct ufshcd_lrb *lrbp, + struct utp_hpb_rsp *rsp_field) +{ + /* Check HPB_UPDATE_ALERT */ + if (!(lrbp->ucd_rsp_ptr->header.dword_2 & + UPIU_HEADER_DWORD(0, 2, 0, 0))) + return false; + + if (be16_to_cpu(rsp_field->sense_data_len) != DEV_SENSE_SEG_LEN || + rsp_field->desc_type != DEV_DES_TYPE || + rsp_field->additional_len != DEV_ADDITIONAL_LEN || + rsp_field->active_rgn_cnt > MAX_ACTIVE_NUM || + rsp_field->inactive_rgn_cnt > MAX_INACTIVE_NUM || + rsp_field->hpb_op == HPB_RSP_NONE || + (rsp_field->hpb_op == HPB_RSP_REQ_REGION_UPDATE && + !rsp_field->active_rgn_cnt && !rsp_field->inactive_rgn_cnt)) + return false; + + if (!ufshpb_is_general_lun(rsp_field->lun)) { + dev_warn(hba->dev, "ufshpb: lun(%d) not supported\n", + lrbp->lun); + return false; + } + + return true; +} + +static struct ufshpb_req *ufshpb_get_map_req(struct ufshpb_lu *hpb, + struct ufshpb_subregion *srgn) +{ + struct ufshpb_req *map_req; + struct request *req; + struct bio *bio; + int retries = HPB_MAP_REQ_RETRIES; + + map_req = kmem_cache_alloc(hpb->map_req_cache, GFP_KERNEL); + if (!map_req) + return NULL; + +retry: + req = blk_get_request(hpb->sdev_ufs_lu->request_queue, + REQ_OP_SCSI_IN, BLK_MQ_REQ_NOWAIT); + + if ((PTR_ERR(req) == -EWOULDBLOCK) && (--retries > 0)) { + usleep_range(3000, 3100); + goto retry; + } + + if (IS_ERR(req)) + goto free_map_req; + + bio = bio_alloc(GFP_KERNEL, hpb->pages_per_srgn); + if (!bio) { + blk_put_request(req); + goto free_map_req; + } + + map_req->hpb = hpb; + map_req->req = req; + map_req->bio = bio; + + map_req->rgn_idx = srgn->rgn_idx; + map_req->srgn_idx = srgn->srgn_idx; + map_req->mctx = srgn->mctx; + + return map_req; + +free_map_req: + kmem_cache_free(hpb->map_req_cache, map_req); + return NULL; +} + +static void ufshpb_put_map_req(struct ufshpb_lu *hpb, + struct ufshpb_req *map_req) +{ + bio_put(map_req->bio); + blk_put_request(map_req->req); + kmem_cache_free(hpb->map_req_cache, map_req); +} + +static int ufshpb_clear_dirty_bitmap(struct ufshpb_lu *hpb, + struct ufshpb_subregion *srgn) +{ + u32 num_entries = hpb->entries_per_srgn; + + if (!srgn->mctx) { + dev_err(&hpb->sdev_ufs_lu->sdev_dev, + "no mctx in region %d subregion %d.\n", + srgn->rgn_idx, srgn->srgn_idx); + return -1; + } + + if (unlikely(srgn->is_last)) + num_entries = hpb->last_srgn_entries; + + bitmap_zero(srgn->mctx->ppn_dirty, num_entries); + return 0; +} + +static void ufshpb_update_active_info(struct ufshpb_lu *hpb, int rgn_idx, + int srgn_idx) +{ + struct ufshpb_region *rgn; + struct ufshpb_subregion *srgn; + + rgn = hpb->rgn_tbl + rgn_idx; + srgn = rgn->srgn_tbl + srgn_idx; + + list_del_init(&rgn->list_inact_rgn); + + if (list_empty(&srgn->list_act_srgn)) + list_add_tail(&srgn->list_act_srgn, &hpb->lh_act_srgn); +} + +static void ufshpb_update_inactive_info(struct ufshpb_lu *hpb, int rgn_idx) +{ + struct ufshpb_region *rgn; + struct ufshpb_subregion *srgn; + int srgn_idx; + + rgn = hpb->rgn_tbl + rgn_idx; + + for_each_sub_region(rgn, srgn_idx, srgn) + list_del_init(&srgn->list_act_srgn); + + if (list_empty(&rgn->list_inact_rgn)) + list_add_tail(&rgn->list_inact_rgn, &hpb->lh_inact_rgn); +} + +static void ufshpb_activate_subregion(struct ufshpb_lu *hpb, + struct ufshpb_subregion *srgn) +{ + struct ufshpb_region *rgn; + + /* + * If there is no mctx in subregion + * after I/O progress for HPB_READ_BUFFER, the region to which the + * subregion belongs was evicted. + * Make sure the region must not evict in I/O progress + */ + if (!srgn->mctx) { + dev_err(&hpb->sdev_ufs_lu->sdev_dev, + "no mctx in region %d subregion %d.\n", + srgn->rgn_idx, srgn->srgn_idx); + srgn->srgn_state = HPB_SRGN_INVALID; + return; + } + + rgn = hpb->rgn_tbl + srgn->rgn_idx; + + if (unlikely(rgn->rgn_state == HPB_RGN_INACTIVE)) { + dev_err(&hpb->sdev_ufs_lu->sdev_dev, + "region %d subregion %d evicted\n", + srgn->rgn_idx, srgn->srgn_idx); + srgn->srgn_state = HPB_SRGN_INVALID; + return; + } + srgn->srgn_state = HPB_SRGN_VALID; +} + +static void ufshpb_map_req_compl_fn(struct request *req, blk_status_t error) +{ + struct ufshpb_req *map_req = (struct ufshpb_req *) req->end_io_data; + struct ufshpb_lu *hpb = map_req->hpb; + struct ufshpb_subregion *srgn; + unsigned long flags; + + srgn = hpb->rgn_tbl[map_req->rgn_idx].srgn_tbl + + map_req->srgn_idx; + + ufshpb_clear_dirty_bitmap(hpb, srgn); + spin_lock_irqsave(&hpb->rgn_state_lock, flags); + ufshpb_activate_subregion(hpb, srgn); + spin_unlock_irqrestore(&hpb->rgn_state_lock, flags); + + ufshpb_put_map_req(map_req->hpb, map_req); +} + +static void ufshpb_set_read_buf_cmd(unsigned char *cdb, int rgn_idx, + int srgn_idx, int srgn_mem_size) +{ + cdb[0] = UFSHPB_READ_BUFFER; + cdb[1] = UFSHPB_READ_BUFFER_ID; + + put_unaligned_be16(rgn_idx, &cdb[2]); + put_unaligned_be16(srgn_idx, &cdb[4]); + put_unaligned_be24(srgn_mem_size, &cdb[6]); + + cdb[9] = 0x00; +} + +static int ufshpb_execute_map_req(struct ufshpb_lu *hpb, + struct ufshpb_req *map_req, bool last) +{ + struct request_queue *q; + struct request *req; + struct scsi_request *rq; + int mem_size = hpb->srgn_mem_size; + int ret = 0; + int i; + + q = hpb->sdev_ufs_lu->request_queue; + for (i = 0; i < hpb->pages_per_srgn; i++) { + ret = bio_add_pc_page(q, map_req->bio, map_req->mctx->m_page[i], + PAGE_SIZE, 0); + if (ret != PAGE_SIZE) { + dev_err(&hpb->sdev_ufs_lu->sdev_dev, + "bio_add_pc_page fail %d - %d\n", + map_req->rgn_idx, map_req->srgn_idx); + return ret; + } + } + + req = map_req->req; + + blk_rq_append_bio(req, &map_req->bio); + + req->end_io_data = map_req; + + rq = scsi_req(req); + + if (unlikely(last)) + mem_size = hpb->last_srgn_entries * HPB_ENTRY_SIZE; + + ufshpb_set_read_buf_cmd(rq->cmd, map_req->rgn_idx, + map_req->srgn_idx, mem_size); + rq->cmd_len = HPB_READ_BUFFER_CMD_LENGTH; + + blk_execute_rq_nowait(q, NULL, req, 1, ufshpb_map_req_compl_fn); + + hpb->stats.map_req_cnt++; + return 0; +} + +static struct ufshpb_map_ctx *ufshpb_get_map_ctx(struct ufshpb_lu *hpb, + bool last) +{ + struct ufshpb_map_ctx *mctx; + u32 num_entries = hpb->entries_per_srgn; + int i, j; + + mctx = mempool_alloc(ufshpb_mctx_pool, GFP_KERNEL); + if (!mctx) + return NULL; + + mctx->m_page = kmem_cache_alloc(hpb->m_page_cache, GFP_KERNEL); + if (!mctx->m_page) + goto release_mctx; + + if (unlikely(last)) + num_entries = hpb->last_srgn_entries; + + mctx->ppn_dirty = bitmap_zalloc(num_entries, GFP_KERNEL); + if (!mctx->ppn_dirty) + goto release_m_page; + + for (i = 0; i < hpb->pages_per_srgn; i++) { + mctx->m_page[i] = mempool_alloc(ufshpb_page_pool, GFP_KERNEL); + if (!mctx->m_page[i]) { + for (j = 0; j < i; j++) + mempool_free(mctx->m_page[j], ufshpb_page_pool); + goto release_ppn_dirty; + } + clear_page(page_address(mctx->m_page[i])); + } + + return mctx; + +release_ppn_dirty: + bitmap_free(mctx->ppn_dirty); +release_m_page: + kmem_cache_free(hpb->m_page_cache, mctx->m_page); +release_mctx: + mempool_free(mctx, ufshpb_mctx_pool); + return NULL; +} + +static void ufshpb_put_map_ctx(struct ufshpb_lu *hpb, + struct ufshpb_map_ctx *mctx) +{ + int i; + + for (i = 0; i < hpb->pages_per_srgn; i++) + mempool_free(mctx->m_page[i], ufshpb_page_pool); + + bitmap_free(mctx->ppn_dirty); + kmem_cache_free(hpb->m_page_cache, mctx->m_page); + mempool_free(mctx, ufshpb_mctx_pool); +} + +static int ufshpb_check_srgns_issue_state(struct ufshpb_lu *hpb, + struct ufshpb_region *rgn) +{ + struct ufshpb_subregion *srgn; + int srgn_idx; + + for_each_sub_region(rgn, srgn_idx, srgn) + if (srgn->srgn_state == HPB_SRGN_ISSUED) + return -EPERM; + + return 0; +} + +static void ufshpb_add_lru_info(struct victim_select_info *lru_info, + struct ufshpb_region *rgn) +{ + rgn->rgn_state = HPB_RGN_ACTIVE; + list_add_tail(&rgn->list_lru_rgn, &lru_info->lh_lru_rgn); + atomic_inc(&lru_info->active_cnt); +} + +static void ufshpb_hit_lru_info(struct victim_select_info *lru_info, + struct ufshpb_region *rgn) +{ + list_move_tail(&rgn->list_lru_rgn, &lru_info->lh_lru_rgn); +} + +static struct ufshpb_region *ufshpb_victim_lru_info(struct ufshpb_lu *hpb) +{ + struct victim_select_info *lru_info = &hpb->lru_info; + struct ufshpb_region *rgn, *victim_rgn = NULL; + + list_for_each_entry(rgn, &lru_info->lh_lru_rgn, list_lru_rgn) { + if (!rgn) { + dev_err(&hpb->sdev_ufs_lu->sdev_dev, + "%s: no region allocated\n", + __func__); + return NULL; + } + if (ufshpb_check_srgns_issue_state(hpb, rgn)) + continue; + + victim_rgn = rgn; + break; + } + + return victim_rgn; +} + +static void ufshpb_cleanup_lru_info(struct victim_select_info *lru_info, + struct ufshpb_region *rgn) +{ + list_del_init(&rgn->list_lru_rgn); + rgn->rgn_state = HPB_RGN_INACTIVE; + atomic_dec(&lru_info->active_cnt); +} + +static void ufshpb_purge_active_subregion(struct ufshpb_lu *hpb, + struct ufshpb_subregion *srgn) +{ + if (srgn->srgn_state != HPB_SRGN_UNUSED) { + ufshpb_put_map_ctx(hpb, srgn->mctx); + srgn->srgn_state = HPB_SRGN_UNUSED; + srgn->mctx = NULL; + } +} + +static void __ufshpb_evict_region(struct ufshpb_lu *hpb, + struct ufshpb_region *rgn) +{ + struct victim_select_info *lru_info; + struct ufshpb_subregion *srgn; + int srgn_idx; + + lru_info = &hpb->lru_info; + + dev_dbg(&hpb->sdev_ufs_lu->sdev_dev, "evict region %d\n", rgn->rgn_idx); + + ufshpb_cleanup_lru_info(lru_info, rgn); + + for_each_sub_region(rgn, srgn_idx, srgn) + ufshpb_purge_active_subregion(hpb, srgn); +} + +static int ufshpb_evict_region(struct ufshpb_lu *hpb, struct ufshpb_region *rgn) +{ + unsigned long flags; + int ret = 0; + + spin_lock_irqsave(&hpb->rgn_state_lock, flags); + if (rgn->rgn_state == HPB_RGN_PINNED) { + dev_warn(&hpb->sdev_ufs_lu->sdev_dev, + "pinned region cannot drop-out. region %d\n", + rgn->rgn_idx); + goto out; + } + if (!list_empty(&rgn->list_lru_rgn)) { + if (ufshpb_check_srgns_issue_state(hpb, rgn)) { + ret = -EBUSY; + goto out; + } + + __ufshpb_evict_region(hpb, rgn); + } +out: + spin_unlock_irqrestore(&hpb->rgn_state_lock, flags); + return ret; +} + +static int ufshpb_issue_map_req(struct ufshpb_lu *hpb, + struct ufshpb_region *rgn, + struct ufshpb_subregion *srgn) +{ + struct ufshpb_req *map_req; + unsigned long flags; + int ret; + int err = -EAGAIN; + bool alloc_required = false; + enum HPB_SRGN_STATE state = HPB_SRGN_INVALID; + + spin_lock_irqsave(&hpb->rgn_state_lock, flags); + + if (ufshpb_get_state(hpb) != HPB_PRESENT) { + dev_notice(&hpb->sdev_ufs_lu->sdev_dev, + "%s: ufshpb state is not PRESENT\n", __func__); + goto unlock_out; + } + + if ((rgn->rgn_state == HPB_RGN_INACTIVE) && + (srgn->srgn_state == HPB_SRGN_INVALID)) { + err = 0; + goto unlock_out; + } + + if (srgn->srgn_state == HPB_SRGN_UNUSED) + alloc_required = true; + + /* + * If the subregion is already ISSUED state, + * a specific event (e.g., GC or wear-leveling, etc.) occurs in + * the device and HPB response for map loading is received. + * In this case, after finishing the HPB_READ_BUFFER, + * the next HPB_READ_BUFFER is performed again to obtain the latest + * map data. + */ + if (srgn->srgn_state == HPB_SRGN_ISSUED) + goto unlock_out; + + srgn->srgn_state = HPB_SRGN_ISSUED; + spin_unlock_irqrestore(&hpb->rgn_state_lock, flags); + + if (alloc_required) { + if (srgn->mctx) { + dev_err(&hpb->sdev_ufs_lu->sdev_dev, + "map_ctx is already allocated. region %d - %d\n", + rgn->rgn_idx, srgn->srgn_idx); + ufshpb_put_map_ctx(hpb, srgn->mctx); + } + srgn->mctx = ufshpb_get_map_ctx(hpb, srgn->is_last); + if (!srgn->mctx) { + dev_err(&hpb->sdev_ufs_lu->sdev_dev, + "get map_ctx failed. region %d - %d\n", + rgn->rgn_idx, srgn->srgn_idx); + state = HPB_SRGN_UNUSED; + goto change_srgn_state; + } + } + + map_req = ufshpb_get_map_req(hpb, srgn); + if (!map_req) + goto change_srgn_state; + + + ret = ufshpb_execute_map_req(hpb, map_req, srgn->is_last); + if (ret) { + dev_err(&hpb->sdev_ufs_lu->sdev_dev, + "%s: issue map_req failed: %d, region %d - %d\n", + __func__, ret, srgn->rgn_idx, srgn->srgn_idx); + goto free_map_req; + } + return 0; + +free_map_req: + ufshpb_put_map_req(hpb, map_req); +change_srgn_state: + spin_lock_irqsave(&hpb->rgn_state_lock, flags); + srgn->srgn_state = state; +unlock_out: + spin_unlock_irqrestore(&hpb->rgn_state_lock, flags); + return err; +} + +static int ufshpb_add_region(struct ufshpb_lu *hpb, struct ufshpb_region *rgn) +{ + struct ufshpb_region *victim_rgn; + struct victim_select_info *lru_info = &hpb->lru_info; + unsigned long flags; + int ret = 0; + + spin_lock_irqsave(&hpb->rgn_state_lock, flags); + /* + * If region belongs to lru_list, just move the region + * to the front of lru list. because the state of the region + * is already active-state + */ + if (!list_empty(&rgn->list_lru_rgn)) { + ufshpb_hit_lru_info(lru_info, rgn); + goto out; + } + + if (rgn->rgn_state == HPB_RGN_INACTIVE) { + if (atomic_read(&lru_info->active_cnt) == + lru_info->max_lru_active_cnt) { + /* + * If the maximum number of active regions + * is exceeded, evict the least recently used region. + * This case may occur when the device responds + * to the eviction information late. + * It is okay to evict the least recently used region, + * because the device could detect this region + * by not issuing HPB_READ + */ + victim_rgn = ufshpb_victim_lru_info(hpb); + if (!victim_rgn) { + dev_warn(&hpb->sdev_ufs_lu->sdev_dev, + "cannot get victim region error\n"); + ret = -ENOMEM; + goto out; + } + + dev_dbg(&hpb->sdev_ufs_lu->sdev_dev, + "LRU full (%d), choose victim %d\n", + atomic_read(&lru_info->active_cnt), + victim_rgn->rgn_idx); + __ufshpb_evict_region(hpb, victim_rgn); + } + + /* + * When a region is added to lru_info list_head, + * it is guaranteed that the subregion has been + * assigned all mctx. If failed, try to receive mctx again + * without being added to lru_info list_head + */ + ufshpb_add_lru_info(lru_info, rgn); + } +out: + spin_unlock_irqrestore(&hpb->rgn_state_lock, flags); + return ret; +} + +static void ufshpb_rsp_req_region_update(struct ufshpb_lu *hpb, + struct utp_hpb_rsp *rsp_field) +{ + struct ufshpb_region *rgn; + struct ufshpb_subregion *srgn; + int i, rgn_i, srgn_i; + + BUILD_BUG_ON(sizeof(struct ufshpb_active_field) != HPB_ACT_FIELD_SIZE); + /* + * If the active region and the inactive region are the same, + * we will inactivate this region. + * The device could check this (region inactivated) and + * will response the proper active region information + */ + for (i = 0; i < rsp_field->active_rgn_cnt; i++) { + rgn_i = + be16_to_cpu(rsp_field->hpb_active_field[i].active_rgn); + srgn_i = + be16_to_cpu(rsp_field->hpb_active_field[i].active_srgn); + + dev_dbg(&hpb->sdev_ufs_lu->sdev_dev, + "activate(%d) region %d - %d\n", i, rgn_i, srgn_i); + + spin_lock(&hpb->rsp_list_lock); + ufshpb_update_active_info(hpb, rgn_i, srgn_i); + spin_unlock(&hpb->rsp_list_lock); + + rgn = hpb->rgn_tbl + rgn_i; + srgn = rgn->srgn_tbl + srgn_i; + + /* blocking HPB_READ */ + spin_lock(&hpb->rgn_state_lock); + if (srgn->srgn_state == HPB_SRGN_VALID) + srgn->srgn_state = HPB_SRGN_INVALID; + spin_unlock(&hpb->rgn_state_lock); + hpb->stats.rb_active_cnt++; + } + + for (i = 0; i < rsp_field->inactive_rgn_cnt; i++) { + rgn_i = be16_to_cpu(rsp_field->hpb_inactive_field[i]); + dev_dbg(&hpb->sdev_ufs_lu->sdev_dev, + "inactivate(%d) region %d\n", i, rgn_i); + + spin_lock(&hpb->rsp_list_lock); + ufshpb_update_inactive_info(hpb, rgn_i); + spin_unlock(&hpb->rsp_list_lock); + + rgn = hpb->rgn_tbl + rgn_i; + + spin_lock(&hpb->rgn_state_lock); + if (rgn->rgn_state != HPB_RGN_INACTIVE) { + for (srgn_i = 0; srgn_i < rgn->srgn_cnt; srgn_i++) { + srgn = rgn->srgn_tbl + srgn_i; + if (srgn->srgn_state == HPB_SRGN_VALID) + srgn->srgn_state = HPB_SRGN_INVALID; + } + } + spin_unlock(&hpb->rgn_state_lock); + + hpb->stats.rb_inactive_cnt++; + } + + dev_dbg(&hpb->sdev_ufs_lu->sdev_dev, "Noti: #ACT %u #INACT %u\n", + rsp_field->active_rgn_cnt, rsp_field->inactive_rgn_cnt); + + if (ufshpb_get_state(hpb) == HPB_PRESENT) + queue_work(ufshpb_wq, &hpb->map_work); +} + +/* + * This function will parse recommended active subregion information in sense + * data field of response UPIU with SAM_STAT_GOOD state. + */ +void ufshpb_rsp_upiu(struct ufs_hba *hba, struct ufshcd_lrb *lrbp) +{ + struct ufshpb_lu *hpb = ufshpb_get_hpb_data(lrbp->cmd->device); + struct utp_hpb_rsp *rsp_field = &lrbp->ucd_rsp_ptr->hr; + int data_seg_len; + + if (unlikely(lrbp->lun != rsp_field->lun)) { + struct scsi_device *sdev; + bool found = false; + + __shost_for_each_device(sdev, hba->host) { + hpb = ufshpb_get_hpb_data(sdev); + + if (!hpb) + continue; + + if (rsp_field->lun == hpb->lun) { + found = true; + break; + } + } + + if (!found) + return; + } + + if (!hpb) + return; + + if ((ufshpb_get_state(hpb) != HPB_PRESENT) && + (ufshpb_get_state(hpb) != HPB_SUSPEND)) { + dev_notice(&hpb->sdev_ufs_lu->sdev_dev, + "%s: ufshpb state is not PRESENT/SUSPEND\n", + __func__); + return; + } + + data_seg_len = be32_to_cpu(lrbp->ucd_rsp_ptr->header.dword_2) + & MASK_RSP_UPIU_DATA_SEG_LEN; + + /* To flush remained rsp_list, we queue the map_work task */ + if (!data_seg_len) { + if (!ufshpb_is_general_lun(hpb->lun)) + return; + + ufshpb_kick_map_work(hpb); + return; + } + + BUILD_BUG_ON(sizeof(struct utp_hpb_rsp) != UTP_HPB_RSP_SIZE); + + if (!ufshpb_is_hpb_rsp_valid(hba, lrbp, rsp_field)) + return; + + hpb->stats.rb_noti_cnt++; + + switch (rsp_field->hpb_op) { + case HPB_RSP_REQ_REGION_UPDATE: + if (data_seg_len != DEV_DATA_SEG_LEN) + dev_warn(&hpb->sdev_ufs_lu->sdev_dev, + "%s: data seg length is not same.\n", + __func__); + ufshpb_rsp_req_region_update(hpb, rsp_field); + break; + case HPB_RSP_DEV_RESET: + dev_warn(&hpb->sdev_ufs_lu->sdev_dev, + "UFS device lost HPB information during PM.\n"); + break; + default: + dev_notice(&hpb->sdev_ufs_lu->sdev_dev, + "hpb_op is not available: %d\n", + rsp_field->hpb_op); + break; + } +} + +static void ufshpb_add_active_list(struct ufshpb_lu *hpb, + struct ufshpb_region *rgn, + struct ufshpb_subregion *srgn) +{ + if (!list_empty(&rgn->list_inact_rgn)) + return; + + if (!list_empty(&srgn->list_act_srgn)) { + list_move(&srgn->list_act_srgn, &hpb->lh_act_srgn); + return; + } + + list_add(&srgn->list_act_srgn, &hpb->lh_act_srgn); +} + +static void ufshpb_add_pending_evict_list(struct ufshpb_lu *hpb, + struct ufshpb_region *rgn, + struct list_head *pending_list) +{ + struct ufshpb_subregion *srgn; + int srgn_idx; + + if (!list_empty(&rgn->list_inact_rgn)) + return; + + for_each_sub_region(rgn, srgn_idx, srgn) + if (!list_empty(&srgn->list_act_srgn)) + return; + + list_add_tail(&rgn->list_inact_rgn, pending_list); +} + +static void ufshpb_run_active_subregion_list(struct ufshpb_lu *hpb) +{ + struct ufshpb_region *rgn; + struct ufshpb_subregion *srgn; + unsigned long flags; + int ret = 0; + + spin_lock_irqsave(&hpb->rsp_list_lock, flags); + while ((srgn = list_first_entry_or_null(&hpb->lh_act_srgn, + struct ufshpb_subregion, + list_act_srgn))) { + if (ufshpb_get_state(hpb) == HPB_SUSPEND) + break; + + list_del_init(&srgn->list_act_srgn); + spin_unlock_irqrestore(&hpb->rsp_list_lock, flags); + + rgn = hpb->rgn_tbl + srgn->rgn_idx; + ret = ufshpb_add_region(hpb, rgn); + if (ret) + goto active_failed; + + ret = ufshpb_issue_map_req(hpb, rgn, srgn); + if (ret) { + dev_err(&hpb->sdev_ufs_lu->sdev_dev, + "issue map_req failed. ret %d, region %d - %d\n", + ret, rgn->rgn_idx, srgn->srgn_idx); + goto active_failed; + } + spin_lock_irqsave(&hpb->rsp_list_lock, flags); + } + spin_unlock_irqrestore(&hpb->rsp_list_lock, flags); + return; + +active_failed: + dev_err(&hpb->sdev_ufs_lu->sdev_dev, "failed to activate region %d - %d, will retry\n", + rgn->rgn_idx, srgn->srgn_idx); + spin_lock_irqsave(&hpb->rsp_list_lock, flags); + ufshpb_add_active_list(hpb, rgn, srgn); + spin_unlock_irqrestore(&hpb->rsp_list_lock, flags); +} + +static void ufshpb_run_inactive_region_list(struct ufshpb_lu *hpb) +{ + struct ufshpb_region *rgn; + unsigned long flags; + int ret; + LIST_HEAD(pending_list); + + spin_lock_irqsave(&hpb->rsp_list_lock, flags); + while ((rgn = list_first_entry_or_null(&hpb->lh_inact_rgn, + struct ufshpb_region, + list_inact_rgn))) { + if (ufshpb_get_state(hpb) == HPB_SUSPEND) + break; + + list_del_init(&rgn->list_inact_rgn); + spin_unlock_irqrestore(&hpb->rsp_list_lock, flags); + + ret = ufshpb_evict_region(hpb, rgn); + if (ret) { + spin_lock_irqsave(&hpb->rsp_list_lock, flags); + ufshpb_add_pending_evict_list(hpb, rgn, &pending_list); + spin_unlock_irqrestore(&hpb->rsp_list_lock, flags); + } + + spin_lock_irqsave(&hpb->rsp_list_lock, flags); + } + + list_splice(&pending_list, &hpb->lh_inact_rgn); + spin_unlock_irqrestore(&hpb->rsp_list_lock, flags); +} + +static void ufshpb_map_work_handler(struct work_struct *work) +{ + struct ufshpb_lu *hpb = container_of(work, struct ufshpb_lu, map_work); + + if (ufshpb_get_state(hpb) != HPB_PRESENT) { + dev_notice(&hpb->sdev_ufs_lu->sdev_dev, + "%s: ufshpb state is not PRESENT\n", __func__); + return; + } + + ufshpb_run_inactive_region_list(hpb); + ufshpb_run_active_subregion_list(hpb); +} + +/* + * this function doesn't need to hold lock due to be called in init. + * (rgn_state_lock, rsp_list_lock, etc..) + */ +static int ufshpb_init_pinned_active_region(struct ufs_hba *hba, + struct ufshpb_lu *hpb, + struct ufshpb_region *rgn) +{ + struct ufshpb_subregion *srgn; + int srgn_idx, i; + int err = 0; + + for_each_sub_region(rgn, srgn_idx, srgn) { + srgn->mctx = ufshpb_get_map_ctx(hpb, srgn->is_last); + srgn->srgn_state = HPB_SRGN_INVALID; + if (!srgn->mctx) { + err = -ENOMEM; + dev_err(hba->dev, + "alloc mctx for pinned region failed\n"); + goto release; + } + + list_add_tail(&srgn->list_act_srgn, &hpb->lh_act_srgn); + } + + rgn->rgn_state = HPB_RGN_PINNED; + return 0; + +release: + for (i = 0; i < srgn_idx; i++) { + srgn = rgn->srgn_tbl + i; + ufshpb_put_map_ctx(hpb, srgn->mctx); + } + return err; +} + static void ufshpb_init_subregion_tbl(struct ufshpb_lu *hpb, struct ufshpb_region *rgn, bool last) { int srgn_idx; struct ufshpb_subregion *srgn; - for (srgn_idx = 0; srgn_idx < rgn->srgn_cnt; srgn_idx++) { - srgn = rgn->srgn_tbl + srgn_idx; + for_each_sub_region(rgn, srgn_idx, srgn) { + INIT_LIST_HEAD(&srgn->list_act_srgn); srgn->rgn_idx = rgn->rgn_idx; srgn->srgn_idx = srgn_idx; @@ -78,6 +966,8 @@ static void ufshpb_lu_parameter_init(struct ufs_hba *hba, hpb->lu_pinned_end = hpb_lu_info->num_pinned ? (hpb_lu_info->pinned_start + hpb_lu_info->num_pinned - 1) : PINNED_NOT_SET; + hpb->lru_info.max_lru_active_cnt = + hpb_lu_info->max_active_rgns - hpb_lu_info->num_pinned; rgn_mem_size = (1ULL << hpb_dev_info->rgn_size) * HPB_RGN_SIZE_UNIT * HPB_ENTRY_SIZE; @@ -129,6 +1019,9 @@ static int ufshpb_alloc_region_tbl(struct ufs_hba *hba, struct ufshpb_lu *hpb) rgn = rgn_table + rgn_idx; rgn->rgn_idx = rgn_idx; + INIT_LIST_HEAD(&rgn->list_inact_rgn); + INIT_LIST_HEAD(&rgn->list_lru_rgn); + if (rgn_idx == hpb->rgns_per_lu - 1) { srgn_cnt = ((hpb->srgns_per_lu - 1) % hpb->srgns_per_rgn) + 1; @@ -140,7 +1033,13 @@ static int ufshpb_alloc_region_tbl(struct ufs_hba *hba, struct ufshpb_lu *hpb) goto release_srgn_table; ufshpb_init_subregion_tbl(hpb, rgn, last_srgn); - rgn->rgn_state = HPB_RGN_INACTIVE; + if (ufshpb_is_pinned_region(hpb, rgn_idx)) { + ret = ufshpb_init_pinned_active_region(hba, hpb, rgn); + if (ret) + goto release_srgn_table; + } else { + rgn->rgn_state = HPB_RGN_INACTIVE; + } } return 0; @@ -159,13 +1058,13 @@ static void ufshpb_destroy_subregion_tbl(struct ufshpb_lu *hpb, struct ufshpb_region *rgn) { int srgn_idx; + struct ufshpb_subregion *srgn; - for (srgn_idx = 0; srgn_idx < rgn->srgn_cnt; srgn_idx++) { - struct ufshpb_subregion *srgn; - - srgn = rgn->srgn_tbl + srgn_idx; - srgn->srgn_state = HPB_SRGN_UNUSED; - } + for_each_sub_region(rgn, srgn_idx, srgn) + if (srgn->srgn_state != HPB_SRGN_UNUSED) { + srgn->srgn_state = HPB_SRGN_UNUSED; + ufshpb_put_map_ctx(hpb, srgn->mctx); + } } static void ufshpb_destroy_region_tbl(struct ufshpb_lu *hpb) @@ -239,11 +1138,47 @@ static int ufshpb_lu_hpb_init(struct ufs_hba *hba, struct ufshpb_lu *hpb) { int ret; + spin_lock_init(&hpb->rgn_state_lock); + spin_lock_init(&hpb->rsp_list_lock); + + INIT_LIST_HEAD(&hpb->lru_info.lh_lru_rgn); + INIT_LIST_HEAD(&hpb->lh_act_srgn); + INIT_LIST_HEAD(&hpb->lh_inact_rgn); + INIT_LIST_HEAD(&hpb->list_hpb_lu); + + INIT_WORK(&hpb->map_work, ufshpb_map_work_handler); + + hpb->map_req_cache = kmem_cache_create("ufshpb_req_cache", + sizeof(struct ufshpb_req), 0, 0, NULL); + if (!hpb->map_req_cache) { + dev_err(hba->dev, "ufshpb(%d) ufshpb_req_cache create fail", + hpb->lun); + return -ENOMEM; + } + + hpb->m_page_cache = kmem_cache_create("ufshpb_m_page_cache", + sizeof(struct page *) * hpb->pages_per_srgn, + 0, 0, NULL); + if (!hpb->m_page_cache) { + dev_err(hba->dev, "ufshpb(%d) ufshpb_m_page_cache create fail", + hpb->lun); + ret = -ENOMEM; + goto release_req_cache; + } + ret = ufshpb_alloc_region_tbl(hba, hpb); + if (ret) + goto release_m_page_cache; ufshpb_stat_init(hpb); return 0; + +release_m_page_cache: + kmem_cache_destroy(hpb->m_page_cache); +release_req_cache: + kmem_cache_destroy(hpb->map_req_cache); + return ret; } static struct ufshpb_lu * @@ -275,6 +1210,33 @@ ufshpb_alloc_hpb_lu(struct ufs_hba *hba, int lun, return NULL; } +static void ufshpb_discard_rsp_lists(struct ufshpb_lu *hpb) +{ + struct ufshpb_region *rgn, *next_rgn; + struct ufshpb_subregion *srgn, *next_srgn; + unsigned long flags; + + /* + * If the device reset occurred, the remained HPB region information + * may be stale. Therefore, by dicarding the lists of HPB response + * that remained after reset, it prevents unnecessary work. + */ + spin_lock_irqsave(&hpb->rsp_list_lock, flags); + list_for_each_entry_safe(rgn, next_rgn, &hpb->lh_inact_rgn, + list_inact_rgn) + list_del_init(&rgn->list_inact_rgn); + + list_for_each_entry_safe(srgn, next_srgn, &hpb->lh_act_srgn, + list_act_srgn) + list_del_init(&srgn->list_act_srgn); + spin_unlock_irqrestore(&hpb->rsp_list_lock, flags); +} + +static void ufshpb_cancel_jobs(struct ufshpb_lu *hpb) +{ + cancel_work_sync(&hpb->map_work); +} + static bool ufshpb_check_hpb_reset_query(struct ufs_hba *hba) { int err = 0; @@ -318,7 +1280,7 @@ void ufshpb_reset(struct ufs_hba *hba) struct scsi_device *sdev; shost_for_each_device(sdev, hba->host) { - hpb = sdev->hostdata; + hpb = ufshpb_get_hpb_data(sdev); if (!hpb) continue; @@ -335,13 +1297,15 @@ void ufshpb_reset_host(struct ufs_hba *hba) struct scsi_device *sdev; shost_for_each_device(sdev, hba->host) { - hpb = sdev->hostdata; + hpb = ufshpb_get_hpb_data(sdev); if (!hpb) continue; if (ufshpb_get_state(hpb) != HPB_PRESENT) continue; ufshpb_set_state(hpb, HPB_RESET); + ufshpb_cancel_jobs(hpb); + ufshpb_discard_rsp_lists(hpb); } } @@ -351,13 +1315,14 @@ void ufshpb_suspend(struct ufs_hba *hba) struct scsi_device *sdev; shost_for_each_device(sdev, hba->host) { - hpb = sdev->hostdata; + hpb = ufshpb_get_hpb_data(sdev); if (!hpb) continue; if (ufshpb_get_state(hpb) != HPB_PRESENT) continue; ufshpb_set_state(hpb, HPB_SUSPEND); + ufshpb_cancel_jobs(hpb); } } @@ -367,7 +1332,7 @@ void ufshpb_resume(struct ufs_hba *hba) struct scsi_device *sdev; shost_for_each_device(sdev, hba->host) { - hpb = sdev->hostdata; + hpb = ufshpb_get_hpb_data(sdev); if (!hpb) continue; @@ -375,6 +1340,7 @@ void ufshpb_resume(struct ufs_hba *hba) (ufshpb_get_state(hpb) != HPB_SUSPEND)) continue; ufshpb_set_state(hpb, HPB_PRESENT); + ufshpb_kick_map_work(hpb); } } @@ -427,7 +1393,7 @@ static int ufshpb_get_lu_info(struct ufs_hba *hba, int lun, void ufshpb_destroy_lu(struct ufs_hba *hba, struct scsi_device *sdev) { - struct ufshpb_lu *hpb = sdev->hostdata; + struct ufshpb_lu *hpb = ufshpb_get_hpb_data(sdev); if (!hpb) return; @@ -437,8 +1403,13 @@ void ufshpb_destroy_lu(struct ufs_hba *hba, struct scsi_device *sdev) sdev = hpb->sdev_ufs_lu; sdev->hostdata = NULL; + ufshpb_cancel_jobs(hpb); + ufshpb_destroy_region_tbl(hpb); + kmem_cache_destroy(hpb->map_req_cache); + kmem_cache_destroy(hpb->m_page_cache); + list_del_init(&hpb->list_hpb_lu); kfree(hpb); @@ -446,24 +1417,41 @@ void ufshpb_destroy_lu(struct ufs_hba *hba, struct scsi_device *sdev) static void ufshpb_hpb_lu_prepared(struct ufs_hba *hba) { + int pool_size; struct ufshpb_lu *hpb; struct scsi_device *sdev; bool init_success; + if (tot_active_srgn_pages == 0) { + ufshpb_remove(hba); + return; + } + init_success = !ufshpb_check_hpb_reset_query(hba); + pool_size = PAGE_ALIGN(ufshpb_host_map_kbytes * 1024) / PAGE_SIZE; + if (pool_size > tot_active_srgn_pages) { + mempool_resize(ufshpb_mctx_pool, tot_active_srgn_pages); + mempool_resize(ufshpb_page_pool, tot_active_srgn_pages); + } + shost_for_each_device(sdev, hba->host) { - hpb = sdev->hostdata; + hpb = ufshpb_get_hpb_data(sdev); if (!hpb) continue; if (init_success) { ufshpb_set_state(hpb, HPB_PRESENT); + if ((hpb->lu_pinned_end - hpb->lu_pinned_start) > 0) + queue_work(ufshpb_wq, &hpb->map_work); } else { dev_err(hba->dev, "destroy HPB lu %d\n", hpb->lun); ufshpb_destroy_lu(hba, sdev); } } + + if (!init_success) + ufshpb_remove(hba); } void ufshpb_init_hpb_lu(struct ufs_hba *hba, struct scsi_device *sdev) @@ -485,6 +1473,9 @@ void ufshpb_init_hpb_lu(struct ufs_hba *hba, struct scsi_device *sdev) if (!hpb) goto out; + tot_active_srgn_pages += hpb_lu_info.max_active_rgns * + hpb->srgns_per_rgn * hpb->pages_per_srgn; + hpb->sdev_ufs_lu = sdev; sdev->hostdata = hpb; @@ -494,6 +1485,57 @@ void ufshpb_init_hpb_lu(struct ufs_hba *hba, struct scsi_device *sdev) ufshpb_hpb_lu_prepared(hba); } +static int ufshpb_init_mem_wq(struct ufs_hba *hba) +{ + int ret; + unsigned int pool_size; + + ufshpb_mctx_cache = kmem_cache_create("ufshpb_mctx_cache", + sizeof(struct ufshpb_map_ctx), + 0, 0, NULL); + if (!ufshpb_mctx_cache) { + dev_err(hba->dev, "ufshpb: cannot init mctx cache\n"); + return -ENOMEM; + } + + pool_size = PAGE_ALIGN(ufshpb_host_map_kbytes * 1024) / PAGE_SIZE; + dev_info(hba->dev, "%s:%d ufshpb_host_map_kbytes %u pool_size %u\n", + __func__, __LINE__, ufshpb_host_map_kbytes, pool_size); + + ufshpb_mctx_pool = mempool_create_slab_pool(pool_size, + ufshpb_mctx_cache); + if (!ufshpb_mctx_pool) { + dev_err(hba->dev, "ufshpb: cannot init mctx pool\n"); + ret = -ENOMEM; + goto release_mctx_cache; + } + + ufshpb_page_pool = mempool_create_page_pool(pool_size, 0); + if (!ufshpb_page_pool) { + dev_err(hba->dev, "ufshpb: cannot init page pool\n"); + ret = -ENOMEM; + goto release_mctx_pool; + } + + ufshpb_wq = alloc_workqueue("ufshpb-wq", + WQ_UNBOUND | WQ_MEM_RECLAIM, 0); + if (!ufshpb_wq) { + dev_err(hba->dev, "ufshpb: alloc workqueue failed\n"); + ret = -ENOMEM; + goto release_page_pool; + } + + return 0; + +release_page_pool: + mempool_destroy(ufshpb_page_pool); +release_mctx_pool: + mempool_destroy(ufshpb_mctx_pool); +release_mctx_cache: + kmem_cache_destroy(ufshpb_mctx_cache); + return ret; +} + void ufshpb_get_geo_info(struct ufs_hba *hba, u8 *geo_buf) { struct ufshpb_dev_info *hpb_info = &hba->ufshpb_dev; @@ -558,7 +1600,13 @@ void ufshpb_init(struct ufs_hba *hba) if (!ufshpb_is_allowed(hba) || !hba->dev_info.hpb_enabled) return; + if (ufshpb_init_mem_wq(hba)) { + hpb_dev_info->hpb_disabled = true; + return; + } + atomic_set(&hpb_dev_info->slave_conf_cnt, hpb_dev_info->num_lu); + tot_active_srgn_pages = 0; /* issue HPB reset query */ for (try = 0; try < HPB_RESET_REQ_RETRIES; try++) { ret = ufshcd_query_flag(hba, UPIU_QUERY_OPCODE_SET_FLAG, @@ -567,3 +1615,16 @@ void ufshpb_init(struct ufs_hba *hba) break; } } + +void ufshpb_remove(struct ufs_hba *hba) +{ + mempool_destroy(ufshpb_page_pool); + mempool_destroy(ufshpb_mctx_pool); + kmem_cache_destroy(ufshpb_mctx_cache); + + destroy_workqueue(ufshpb_wq); +} + +module_param(ufshpb_host_map_kbytes, uint, 0644); +MODULE_PARM_DESC(ufshpb_host_map_kbytes, + "ufshpb host mapping memory kilo-bytes for ufshpb memory-pool"); diff --git a/drivers/scsi/ufs/ufshpb.h b/drivers/scsi/ufs/ufshpb.h index 11f5b018af51..aaffc8968afd 100644 --- a/drivers/scsi/ufs/ufshpb.h +++ b/drivers/scsi/ufs/ufshpb.h @@ -40,6 +40,7 @@ #define LU_ENABLED_HPB_FUNC 0x02 #define HPB_RESET_REQ_RETRIES 10 +#define HPB_MAP_REQ_RETRIES 5 #define HPB_SUPPORT_VERSION 0x100 @@ -83,11 +84,19 @@ struct ufshpb_lu_info { int max_active_rgns; }; +struct ufshpb_map_ctx { + struct page **m_page; + unsigned long *ppn_dirty; +}; + struct ufshpb_subregion { + struct ufshpb_map_ctx *mctx; enum HPB_SRGN_STATE srgn_state; int rgn_idx; int srgn_idx; bool is_last; + /* below information is used by rsp_list */ + struct list_head list_act_srgn; }; struct ufshpb_region { @@ -95,6 +104,43 @@ struct ufshpb_region { enum HPB_RGN_STATE rgn_state; int rgn_idx; int srgn_cnt; + + /* below information is used by rsp_list */ + struct list_head list_inact_rgn; + + /* below information is used by lru */ + struct list_head list_lru_rgn; +}; + +#define for_each_sub_region(rgn, i, srgn) \ + for ((i) = 0; \ + ((i) < (rgn)->srgn_cnt) && ((srgn) = &(rgn)->srgn_tbl[i]); \ + (i)++) + +/** + * struct ufshpb_req - UFSHPB READ BUFFER (for caching map) request structure + * @req: block layer request for READ BUFFER + * @bio: bio for holding map page + * @hpb: ufshpb_lu structure that related to the L2P map + * @mctx: L2P map information + * @rgn_idx: target region index + * @srgn_idx: target sub-region index + * @lun: target logical unit number + */ +struct ufshpb_req { + struct request *req; + struct bio *bio; + struct ufshpb_lu *hpb; + struct ufshpb_map_ctx *mctx; + + unsigned int rgn_idx; + unsigned int srgn_idx; +}; + +struct victim_select_info { + struct list_head lh_lru_rgn; /* LRU list of regions */ + int max_lru_active_cnt; /* supported hpb #region - pinned #region */ + atomic_t active_cnt; }; struct ufshpb_stats { @@ -109,10 +155,22 @@ struct ufshpb_stats { struct ufshpb_lu { int lun; struct scsi_device *sdev_ufs_lu; + + spinlock_t rgn_state_lock; /* for protect rgn/srgn state */ struct ufshpb_region *rgn_tbl; atomic_t hpb_state; + spinlock_t rsp_list_lock; + struct list_head lh_act_srgn; /* hold rsp_list_lock */ + struct list_head lh_inact_rgn; /* hold rsp_list_lock */ + + /* cached L2P map management worker */ + struct work_struct map_work; + + /* for selecting victim */ + struct victim_select_info lru_info; + /* pinned region information */ u32 lu_pinned_start; u32 lu_pinned_end; @@ -132,6 +190,9 @@ struct ufshpb_lu { struct ufshpb_stats stats; + struct kmem_cache *map_req_cache; + struct kmem_cache *m_page_cache; + struct list_head list_hpb_lu; }; @@ -139,6 +200,7 @@ struct ufs_hba; struct ufshcd_lrb; #ifndef CONFIG_SCSI_UFS_HPB +static void ufshpb_rsp_upiu(struct ufs_hba *hba, struct ufshcd_lrb *lrbp) {} static void ufshpb_resume(struct ufs_hba *hba) {} static void ufshpb_suspend(struct ufs_hba *hba) {} static void ufshpb_reset(struct ufs_hba *hba) {} @@ -146,10 +208,12 @@ static void ufshpb_reset_host(struct ufs_hba *hba) {} static void ufshpb_init(struct ufs_hba *hba) {} static void ufshpb_init_hpb_lu(struct ufs_hba *hba, struct scsi_device *sdev) {} static void ufshpb_destroy_lu(struct ufs_hba *hba, struct scsi_device *sdev) {} +static void ufshpb_remove(struct ufs_hba *hba) {} static bool ufshpb_is_allowed(struct ufs_hba *hba) { return false; } static void ufshpb_get_geo_info(struct ufs_hba *hba, u8 *geo_buf) {} static void ufshpb_get_dev_info(struct ufs_hba *hba, u8 *desc_buf) {} #else +void ufshpb_rsp_upiu(struct ufs_hba *hba, struct ufshcd_lrb *lrbp); void ufshpb_resume(struct ufs_hba *hba); void ufshpb_suspend(struct ufs_hba *hba); void ufshpb_reset(struct ufs_hba *hba); @@ -157,6 +221,7 @@ void ufshpb_reset_host(struct ufs_hba *hba); void ufshpb_init(struct ufs_hba *hba); void ufshpb_init_hpb_lu(struct ufs_hba *hba, struct scsi_device *sdev); void ufshpb_destroy_lu(struct ufs_hba *hba, struct scsi_device *sdev); +void ufshpb_remove(struct ufs_hba *hba); bool ufshpb_is_allowed(struct ufs_hba *hba); void ufshpb_get_geo_info(struct ufs_hba *hba, u8 *geo_buf); void ufshpb_get_dev_info(struct ufs_hba *hba, u8 *desc_buf);