1 // SPDX-License-Identifier: GPL-2.0
3 * Universal Flash Storage Host Performance Booster
5 * Copyright (C) 2017-2021 Samsung Electronics Co., Ltd.
8 * Yongmyung Lee <ymhungry.lee@samsung.com>
9 * Jinyoung Choi <j-young.choi@samsung.com>
12 #include <asm/unaligned.h>
13 #include <linux/delay.h>
14 #include <linux/device.h>
15 #include <linux/module.h>
16 #include <scsi/scsi_cmnd.h>
18 #include "ufshcd-priv.h"
20 #include "../../scsi/sd.h"
22 #define ACTIVATION_THRESHOLD 8 /* 8 IOs */
23 #define READ_TO_MS 1000
24 #define READ_TO_EXPIRIES 100
25 #define POLLING_INTERVAL_MS 200
26 #define THROTTLE_MAP_REQ_DEFAULT 1
28 /* memory management */
29 static struct kmem_cache *ufshpb_mctx_cache;
30 static mempool_t *ufshpb_mctx_pool;
31 static mempool_t *ufshpb_page_pool;
32 /* A cache size of 2MB can cache ppn in the 1GB range. */
33 static unsigned int ufshpb_host_map_kbytes = 2048;
34 static int tot_active_srgn_pages;
36 static struct workqueue_struct *ufshpb_wq;
38 static void ufshpb_update_active_info(struct ufshpb_lu *hpb, int rgn_idx,
41 bool ufshpb_is_allowed(struct ufs_hba *hba)
43 return !(hba->ufshpb_dev.hpb_disabled);
46 /* HPB version 1.0 is called as legacy version. */
47 bool ufshpb_is_legacy(struct ufs_hba *hba)
49 return hba->ufshpb_dev.is_legacy;
52 static struct ufshpb_lu *ufshpb_get_hpb_data(struct scsi_device *sdev)
54 return sdev->hostdata;
57 static int ufshpb_get_state(struct ufshpb_lu *hpb)
59 return atomic_read(&hpb->hpb_state);
62 static void ufshpb_set_state(struct ufshpb_lu *hpb, int state)
64 atomic_set(&hpb->hpb_state, state);
67 static int ufshpb_is_valid_srgn(struct ufshpb_region *rgn,
68 struct ufshpb_subregion *srgn)
70 return rgn->rgn_state != HPB_RGN_INACTIVE &&
71 srgn->srgn_state == HPB_SRGN_VALID;
74 static bool ufshpb_is_read_cmd(struct scsi_cmnd *cmd)
76 return req_op(scsi_cmd_to_rq(cmd)) == REQ_OP_READ;
79 static bool ufshpb_is_write_or_discard(struct scsi_cmnd *cmd)
81 return op_is_write(req_op(scsi_cmd_to_rq(cmd))) ||
82 op_is_discard(req_op(scsi_cmd_to_rq(cmd)));
85 static bool ufshpb_is_supported_chunk(struct ufshpb_lu *hpb, int transfer_len)
87 return transfer_len <= hpb->pre_req_max_tr_len;
90 static bool ufshpb_is_general_lun(int lun)
92 return lun < UFS_UPIU_MAX_UNIT_NUM_ID;
95 static bool ufshpb_is_pinned_region(struct ufshpb_lu *hpb, int rgn_idx)
97 return hpb->lu_pinned_end != PINNED_NOT_SET &&
98 rgn_idx >= hpb->lu_pinned_start && rgn_idx <= hpb->lu_pinned_end;
101 static void ufshpb_kick_map_work(struct ufshpb_lu *hpb)
106 if (ufshpb_get_state(hpb) != HPB_PRESENT)
109 spin_lock_irqsave(&hpb->rsp_list_lock, flags);
110 if (!list_empty(&hpb->lh_inact_rgn) || !list_empty(&hpb->lh_act_srgn))
112 spin_unlock_irqrestore(&hpb->rsp_list_lock, flags);
115 queue_work(ufshpb_wq, &hpb->map_work);
118 static bool ufshpb_is_hpb_rsp_valid(struct ufs_hba *hba,
119 struct ufshcd_lrb *lrbp,
120 struct utp_hpb_rsp *rsp_field)
122 /* Check HPB_UPDATE_ALERT */
123 if (!(lrbp->ucd_rsp_ptr->header.dword_2 &
124 UPIU_HEADER_DWORD(0, 2, 0, 0)))
127 if (be16_to_cpu(rsp_field->sense_data_len) != DEV_SENSE_SEG_LEN ||
128 rsp_field->desc_type != DEV_DES_TYPE ||
129 rsp_field->additional_len != DEV_ADDITIONAL_LEN ||
130 rsp_field->active_rgn_cnt > MAX_ACTIVE_NUM ||
131 rsp_field->inactive_rgn_cnt > MAX_INACTIVE_NUM ||
132 rsp_field->hpb_op == HPB_RSP_NONE ||
133 (rsp_field->hpb_op == HPB_RSP_REQ_REGION_UPDATE &&
134 !rsp_field->active_rgn_cnt && !rsp_field->inactive_rgn_cnt))
137 if (!ufshpb_is_general_lun(rsp_field->lun)) {
138 dev_warn(hba->dev, "ufshpb: lun(%d) not supported\n",
146 static void ufshpb_iterate_rgn(struct ufshpb_lu *hpb, int rgn_idx, int srgn_idx,
147 int srgn_offset, int cnt, bool set_dirty)
149 struct ufshpb_region *rgn;
150 struct ufshpb_subregion *srgn, *prev_srgn = NULL;
156 rgn = hpb->rgn_tbl + rgn_idx;
157 srgn = rgn->srgn_tbl + srgn_idx;
159 if (likely(!srgn->is_last))
160 bitmap_len = hpb->entries_per_srgn;
162 bitmap_len = hpb->last_srgn_entries;
164 if ((srgn_offset + cnt) > bitmap_len)
165 set_bit_len = bitmap_len - srgn_offset;
169 spin_lock_irqsave(&hpb->rgn_state_lock, flags);
170 if (rgn->rgn_state != HPB_RGN_INACTIVE) {
172 if (srgn->srgn_state == HPB_SRGN_VALID)
173 bitmap_set(srgn->mctx->ppn_dirty, srgn_offset,
175 } else if (hpb->is_hcm) {
176 /* rewind the read timer for lru regions */
177 rgn->read_timeout = ktime_add_ms(ktime_get(),
178 rgn->hpb->params.read_timeout_ms);
179 rgn->read_timeout_expiries =
180 rgn->hpb->params.read_timeout_expiries;
183 spin_unlock_irqrestore(&hpb->rgn_state_lock, flags);
185 if (hpb->is_hcm && prev_srgn != srgn) {
186 bool activate = false;
188 spin_lock(&rgn->rgn_lock);
190 rgn->reads -= srgn->reads;
192 set_bit(RGN_FLAG_DIRTY, &rgn->rgn_flags);
196 if (srgn->reads == hpb->params.activation_thld)
199 spin_unlock(&rgn->rgn_lock);
202 test_and_clear_bit(RGN_FLAG_UPDATE, &rgn->rgn_flags)) {
203 spin_lock_irqsave(&hpb->rsp_list_lock, flags);
204 ufshpb_update_active_info(hpb, rgn_idx, srgn_idx);
205 spin_unlock_irqrestore(&hpb->rsp_list_lock, flags);
206 dev_dbg(&hpb->sdev_ufs_lu->sdev_dev,
207 "activate region %d-%d\n", rgn_idx, srgn_idx);
214 if (++srgn_idx == hpb->srgns_per_rgn) {
224 static bool ufshpb_test_ppn_dirty(struct ufshpb_lu *hpb, int rgn_idx,
225 int srgn_idx, int srgn_offset, int cnt)
227 struct ufshpb_region *rgn;
228 struct ufshpb_subregion *srgn;
233 rgn = hpb->rgn_tbl + rgn_idx;
234 srgn = rgn->srgn_tbl + srgn_idx;
236 if (likely(!srgn->is_last))
237 bitmap_len = hpb->entries_per_srgn;
239 bitmap_len = hpb->last_srgn_entries;
241 if (!ufshpb_is_valid_srgn(rgn, srgn))
245 * If the region state is active, mctx must be allocated.
246 * In this case, check whether the region is evicted or
247 * mctx allocation fail.
249 if (unlikely(!srgn->mctx)) {
250 dev_err(&hpb->sdev_ufs_lu->sdev_dev,
251 "no mctx in region %d subregion %d.\n",
252 srgn->rgn_idx, srgn->srgn_idx);
256 if ((srgn_offset + cnt) > bitmap_len)
257 bit_len = bitmap_len - srgn_offset;
261 if (find_next_bit(srgn->mctx->ppn_dirty, bit_len + srgn_offset,
262 srgn_offset) < bit_len + srgn_offset)
266 if (++srgn_idx == hpb->srgns_per_rgn) {
278 static inline bool is_rgn_dirty(struct ufshpb_region *rgn)
280 return test_bit(RGN_FLAG_DIRTY, &rgn->rgn_flags);
283 static int ufshpb_fill_ppn_from_page(struct ufshpb_lu *hpb,
284 struct ufshpb_map_ctx *mctx, int pos,
285 int len, __be64 *ppn_buf)
291 index = pos / (PAGE_SIZE / HPB_ENTRY_SIZE);
292 offset = pos % (PAGE_SIZE / HPB_ENTRY_SIZE);
294 if ((offset + len) <= (PAGE_SIZE / HPB_ENTRY_SIZE))
297 copied = (PAGE_SIZE / HPB_ENTRY_SIZE) - offset;
299 page = mctx->m_page[index];
300 if (unlikely(!page)) {
301 dev_err(&hpb->sdev_ufs_lu->sdev_dev,
302 "error. cannot find page in mctx\n");
306 memcpy(ppn_buf, page_address(page) + (offset * HPB_ENTRY_SIZE),
307 copied * HPB_ENTRY_SIZE);
313 ufshpb_get_pos_from_lpn(struct ufshpb_lu *hpb, unsigned long lpn, int *rgn_idx,
314 int *srgn_idx, int *offset)
318 *rgn_idx = lpn >> hpb->entries_per_rgn_shift;
319 rgn_offset = lpn & hpb->entries_per_rgn_mask;
320 *srgn_idx = rgn_offset >> hpb->entries_per_srgn_shift;
321 *offset = rgn_offset & hpb->entries_per_srgn_mask;
325 ufshpb_set_hpb_read_to_upiu(struct ufs_hba *hba, struct ufshcd_lrb *lrbp,
326 __be64 ppn, u8 transfer_len)
328 unsigned char *cdb = lrbp->cmd->cmnd;
329 __be64 ppn_tmp = ppn;
330 cdb[0] = UFSHPB_READ;
332 if (hba->dev_quirks & UFS_DEVICE_QUIRK_SWAP_L2P_ENTRY_FOR_HPB_READ)
333 ppn_tmp = (__force __be64)swab64((__force u64)ppn);
335 /* ppn value is stored as big-endian in the host memory */
336 memcpy(&cdb[6], &ppn_tmp, sizeof(__be64));
337 cdb[14] = transfer_len;
340 lrbp->cmd->cmd_len = UFS_CDB_SIZE;
344 * This function will set up HPB read command using host-side L2P map data.
346 int ufshpb_prep(struct ufs_hba *hba, struct ufshcd_lrb *lrbp)
348 struct ufshpb_lu *hpb;
349 struct ufshpb_region *rgn;
350 struct ufshpb_subregion *srgn;
351 struct scsi_cmnd *cmd = lrbp->cmd;
355 int transfer_len, rgn_idx, srgn_idx, srgn_offset;
358 hpb = ufshpb_get_hpb_data(cmd->device);
362 if (ufshpb_get_state(hpb) == HPB_INIT)
365 if (ufshpb_get_state(hpb) != HPB_PRESENT) {
366 dev_notice(&hpb->sdev_ufs_lu->sdev_dev,
367 "%s: ufshpb state is not PRESENT", __func__);
371 if (blk_rq_is_passthrough(scsi_cmd_to_rq(cmd)) ||
372 (!ufshpb_is_write_or_discard(cmd) &&
373 !ufshpb_is_read_cmd(cmd)))
376 transfer_len = sectors_to_logical(cmd->device,
377 blk_rq_sectors(scsi_cmd_to_rq(cmd)));
378 if (unlikely(!transfer_len))
381 lpn = sectors_to_logical(cmd->device, blk_rq_pos(scsi_cmd_to_rq(cmd)));
382 ufshpb_get_pos_from_lpn(hpb, lpn, &rgn_idx, &srgn_idx, &srgn_offset);
383 rgn = hpb->rgn_tbl + rgn_idx;
384 srgn = rgn->srgn_tbl + srgn_idx;
386 /* If command type is WRITE or DISCARD, set bitmap as drity */
387 if (ufshpb_is_write_or_discard(cmd)) {
388 ufshpb_iterate_rgn(hpb, rgn_idx, srgn_idx, srgn_offset,
393 if (!ufshpb_is_supported_chunk(hpb, transfer_len))
398 * in host control mode, reads are the main source for
401 ufshpb_iterate_rgn(hpb, rgn_idx, srgn_idx, srgn_offset,
402 transfer_len, false);
404 /* keep those counters normalized */
405 if (rgn->reads > hpb->entries_per_srgn)
406 schedule_work(&hpb->ufshpb_normalization_work);
409 spin_lock_irqsave(&hpb->rgn_state_lock, flags);
410 if (ufshpb_test_ppn_dirty(hpb, rgn_idx, srgn_idx, srgn_offset,
412 hpb->stats.miss_cnt++;
413 spin_unlock_irqrestore(&hpb->rgn_state_lock, flags);
417 err = ufshpb_fill_ppn_from_page(hpb, srgn->mctx, srgn_offset, 1, &ppn);
418 spin_unlock_irqrestore(&hpb->rgn_state_lock, flags);
419 if (unlikely(err < 0)) {
421 * In this case, the region state is active,
422 * but the ppn table is not allocated.
423 * Make sure that ppn table must be allocated on
426 dev_err(hba->dev, "get ppn failed. err %d\n", err);
430 ufshpb_set_hpb_read_to_upiu(hba, lrbp, ppn, transfer_len);
432 hpb->stats.hit_cnt++;
436 static struct ufshpb_req *ufshpb_get_req(struct ufshpb_lu *hpb, int rgn_idx,
437 enum req_op op, bool atomic)
439 struct ufshpb_req *rq;
441 int retries = HPB_MAP_REQ_RETRIES;
443 rq = kmem_cache_alloc(hpb->map_req_cache, GFP_KERNEL);
448 req = blk_mq_alloc_request(hpb->sdev_ufs_lu->request_queue, op,
451 if (!atomic && (PTR_ERR(req) == -EWOULDBLOCK) && (--retries > 0)) {
452 usleep_range(3000, 3100);
461 rq->rb.rgn_idx = rgn_idx;
466 kmem_cache_free(hpb->map_req_cache, rq);
470 static void ufshpb_put_req(struct ufshpb_lu *hpb, struct ufshpb_req *rq)
472 blk_mq_free_request(rq->req);
473 kmem_cache_free(hpb->map_req_cache, rq);
476 static struct ufshpb_req *ufshpb_get_map_req(struct ufshpb_lu *hpb,
477 struct ufshpb_subregion *srgn)
479 struct ufshpb_req *map_req;
484 hpb->num_inflight_map_req >= hpb->params.inflight_map_req) {
485 dev_info(&hpb->sdev_ufs_lu->sdev_dev,
486 "map_req throttle. inflight %d throttle %d",
487 hpb->num_inflight_map_req,
488 hpb->params.inflight_map_req);
492 map_req = ufshpb_get_req(hpb, srgn->rgn_idx, REQ_OP_DRV_IN, false);
496 bio = bio_alloc(NULL, hpb->pages_per_srgn, 0, GFP_KERNEL);
498 ufshpb_put_req(hpb, map_req);
504 map_req->rb.srgn_idx = srgn->srgn_idx;
505 map_req->rb.mctx = srgn->mctx;
507 spin_lock_irqsave(&hpb->param_lock, flags);
508 hpb->num_inflight_map_req++;
509 spin_unlock_irqrestore(&hpb->param_lock, flags);
514 static void ufshpb_put_map_req(struct ufshpb_lu *hpb,
515 struct ufshpb_req *map_req)
519 bio_put(map_req->bio);
520 ufshpb_put_req(hpb, map_req);
522 spin_lock_irqsave(&hpb->param_lock, flags);
523 hpb->num_inflight_map_req--;
524 spin_unlock_irqrestore(&hpb->param_lock, flags);
527 static int ufshpb_clear_dirty_bitmap(struct ufshpb_lu *hpb,
528 struct ufshpb_subregion *srgn)
530 struct ufshpb_region *rgn;
531 u32 num_entries = hpb->entries_per_srgn;
534 dev_err(&hpb->sdev_ufs_lu->sdev_dev,
535 "no mctx in region %d subregion %d.\n",
536 srgn->rgn_idx, srgn->srgn_idx);
540 if (unlikely(srgn->is_last))
541 num_entries = hpb->last_srgn_entries;
543 bitmap_zero(srgn->mctx->ppn_dirty, num_entries);
545 rgn = hpb->rgn_tbl + srgn->rgn_idx;
546 clear_bit(RGN_FLAG_DIRTY, &rgn->rgn_flags);
551 static void ufshpb_update_active_info(struct ufshpb_lu *hpb, int rgn_idx,
554 struct ufshpb_region *rgn;
555 struct ufshpb_subregion *srgn;
557 rgn = hpb->rgn_tbl + rgn_idx;
558 srgn = rgn->srgn_tbl + srgn_idx;
560 list_del_init(&rgn->list_inact_rgn);
562 if (list_empty(&srgn->list_act_srgn))
563 list_add_tail(&srgn->list_act_srgn, &hpb->lh_act_srgn);
565 hpb->stats.rcmd_active_cnt++;
568 static void ufshpb_update_inactive_info(struct ufshpb_lu *hpb, int rgn_idx)
570 struct ufshpb_region *rgn;
571 struct ufshpb_subregion *srgn;
574 rgn = hpb->rgn_tbl + rgn_idx;
576 for_each_sub_region(rgn, srgn_idx, srgn)
577 list_del_init(&srgn->list_act_srgn);
579 if (list_empty(&rgn->list_inact_rgn))
580 list_add_tail(&rgn->list_inact_rgn, &hpb->lh_inact_rgn);
582 hpb->stats.rcmd_inactive_cnt++;
585 static void ufshpb_activate_subregion(struct ufshpb_lu *hpb,
586 struct ufshpb_subregion *srgn)
588 struct ufshpb_region *rgn;
591 * If there is no mctx in subregion
592 * after I/O progress for HPB_READ_BUFFER, the region to which the
593 * subregion belongs was evicted.
594 * Make sure the region must not evict in I/O progress
597 dev_err(&hpb->sdev_ufs_lu->sdev_dev,
598 "no mctx in region %d subregion %d.\n",
599 srgn->rgn_idx, srgn->srgn_idx);
600 srgn->srgn_state = HPB_SRGN_INVALID;
604 rgn = hpb->rgn_tbl + srgn->rgn_idx;
606 if (unlikely(rgn->rgn_state == HPB_RGN_INACTIVE)) {
607 dev_err(&hpb->sdev_ufs_lu->sdev_dev,
608 "region %d subregion %d evicted\n",
609 srgn->rgn_idx, srgn->srgn_idx);
610 srgn->srgn_state = HPB_SRGN_INVALID;
613 srgn->srgn_state = HPB_SRGN_VALID;
616 static void ufshpb_umap_req_compl_fn(struct request *req, blk_status_t error)
618 struct ufshpb_req *umap_req = (struct ufshpb_req *)req->end_io_data;
620 ufshpb_put_req(umap_req->hpb, umap_req);
623 static void ufshpb_map_req_compl_fn(struct request *req, blk_status_t error)
625 struct ufshpb_req *map_req = (struct ufshpb_req *) req->end_io_data;
626 struct ufshpb_lu *hpb = map_req->hpb;
627 struct ufshpb_subregion *srgn;
630 srgn = hpb->rgn_tbl[map_req->rb.rgn_idx].srgn_tbl +
631 map_req->rb.srgn_idx;
633 ufshpb_clear_dirty_bitmap(hpb, srgn);
634 spin_lock_irqsave(&hpb->rgn_state_lock, flags);
635 ufshpb_activate_subregion(hpb, srgn);
636 spin_unlock_irqrestore(&hpb->rgn_state_lock, flags);
638 ufshpb_put_map_req(map_req->hpb, map_req);
641 static void ufshpb_set_unmap_cmd(unsigned char *cdb, struct ufshpb_region *rgn)
643 cdb[0] = UFSHPB_WRITE_BUFFER;
644 cdb[1] = rgn ? UFSHPB_WRITE_BUFFER_INACT_SINGLE_ID :
645 UFSHPB_WRITE_BUFFER_INACT_ALL_ID;
647 put_unaligned_be16(rgn->rgn_idx, &cdb[2]);
651 static void ufshpb_set_read_buf_cmd(unsigned char *cdb, int rgn_idx,
652 int srgn_idx, int srgn_mem_size)
654 cdb[0] = UFSHPB_READ_BUFFER;
655 cdb[1] = UFSHPB_READ_BUFFER_ID;
657 put_unaligned_be16(rgn_idx, &cdb[2]);
658 put_unaligned_be16(srgn_idx, &cdb[4]);
659 put_unaligned_be24(srgn_mem_size, &cdb[6]);
664 static void ufshpb_execute_umap_req(struct ufshpb_lu *hpb,
665 struct ufshpb_req *umap_req,
666 struct ufshpb_region *rgn)
668 struct request *req = umap_req->req;
669 struct scsi_cmnd *scmd = blk_mq_rq_to_pdu(req);
672 req->end_io_data = umap_req;
673 req->end_io = ufshpb_umap_req_compl_fn;
675 ufshpb_set_unmap_cmd(scmd->cmnd, rgn);
676 scmd->cmd_len = HPB_WRITE_BUFFER_CMD_LENGTH;
678 blk_execute_rq_nowait(req, true);
680 hpb->stats.umap_req_cnt++;
683 static int ufshpb_execute_map_req(struct ufshpb_lu *hpb,
684 struct ufshpb_req *map_req, bool last)
686 struct request_queue *q;
688 struct scsi_cmnd *scmd;
689 int mem_size = hpb->srgn_mem_size;
693 q = hpb->sdev_ufs_lu->request_queue;
694 for (i = 0; i < hpb->pages_per_srgn; i++) {
695 ret = bio_add_pc_page(q, map_req->bio, map_req->rb.mctx->m_page[i],
697 if (ret != PAGE_SIZE) {
698 dev_err(&hpb->sdev_ufs_lu->sdev_dev,
699 "bio_add_pc_page fail %d - %d\n",
700 map_req->rb.rgn_idx, map_req->rb.srgn_idx);
707 blk_rq_append_bio(req, map_req->bio);
709 req->end_io_data = map_req;
710 req->end_io = ufshpb_map_req_compl_fn;
713 mem_size = hpb->last_srgn_entries * HPB_ENTRY_SIZE;
715 scmd = blk_mq_rq_to_pdu(req);
716 ufshpb_set_read_buf_cmd(scmd->cmnd, map_req->rb.rgn_idx,
717 map_req->rb.srgn_idx, mem_size);
718 scmd->cmd_len = HPB_READ_BUFFER_CMD_LENGTH;
720 blk_execute_rq_nowait(req, true);
722 hpb->stats.map_req_cnt++;
726 static struct ufshpb_map_ctx *ufshpb_get_map_ctx(struct ufshpb_lu *hpb,
729 struct ufshpb_map_ctx *mctx;
730 u32 num_entries = hpb->entries_per_srgn;
733 mctx = mempool_alloc(ufshpb_mctx_pool, GFP_KERNEL);
737 mctx->m_page = kmem_cache_alloc(hpb->m_page_cache, GFP_KERNEL);
742 num_entries = hpb->last_srgn_entries;
744 mctx->ppn_dirty = bitmap_zalloc(num_entries, GFP_KERNEL);
745 if (!mctx->ppn_dirty)
748 for (i = 0; i < hpb->pages_per_srgn; i++) {
749 mctx->m_page[i] = mempool_alloc(ufshpb_page_pool, GFP_KERNEL);
750 if (!mctx->m_page[i]) {
751 for (j = 0; j < i; j++)
752 mempool_free(mctx->m_page[j], ufshpb_page_pool);
753 goto release_ppn_dirty;
755 clear_page(page_address(mctx->m_page[i]));
761 bitmap_free(mctx->ppn_dirty);
763 kmem_cache_free(hpb->m_page_cache, mctx->m_page);
765 mempool_free(mctx, ufshpb_mctx_pool);
769 static void ufshpb_put_map_ctx(struct ufshpb_lu *hpb,
770 struct ufshpb_map_ctx *mctx)
774 for (i = 0; i < hpb->pages_per_srgn; i++)
775 mempool_free(mctx->m_page[i], ufshpb_page_pool);
777 bitmap_free(mctx->ppn_dirty);
778 kmem_cache_free(hpb->m_page_cache, mctx->m_page);
779 mempool_free(mctx, ufshpb_mctx_pool);
782 static int ufshpb_check_srgns_issue_state(struct ufshpb_lu *hpb,
783 struct ufshpb_region *rgn)
785 struct ufshpb_subregion *srgn;
788 for_each_sub_region(rgn, srgn_idx, srgn)
789 if (srgn->srgn_state == HPB_SRGN_ISSUED)
795 static void ufshpb_read_to_handler(struct work_struct *work)
797 struct ufshpb_lu *hpb = container_of(work, struct ufshpb_lu,
798 ufshpb_read_to_work.work);
799 struct victim_select_info *lru_info = &hpb->lru_info;
800 struct ufshpb_region *rgn, *next_rgn;
803 LIST_HEAD(expired_list);
805 if (test_and_set_bit(TIMEOUT_WORK_RUNNING, &hpb->work_data_bits))
808 spin_lock_irqsave(&hpb->rgn_state_lock, flags);
810 list_for_each_entry_safe(rgn, next_rgn, &lru_info->lh_lru_rgn,
812 bool timedout = ktime_after(ktime_get(), rgn->read_timeout);
815 rgn->read_timeout_expiries--;
816 if (is_rgn_dirty(rgn) ||
817 rgn->read_timeout_expiries == 0)
818 list_add(&rgn->list_expired_rgn, &expired_list);
820 rgn->read_timeout = ktime_add_ms(ktime_get(),
821 hpb->params.read_timeout_ms);
825 spin_unlock_irqrestore(&hpb->rgn_state_lock, flags);
827 list_for_each_entry_safe(rgn, next_rgn, &expired_list,
829 list_del_init(&rgn->list_expired_rgn);
830 spin_lock_irqsave(&hpb->rsp_list_lock, flags);
831 ufshpb_update_inactive_info(hpb, rgn->rgn_idx);
832 spin_unlock_irqrestore(&hpb->rsp_list_lock, flags);
835 ufshpb_kick_map_work(hpb);
837 clear_bit(TIMEOUT_WORK_RUNNING, &hpb->work_data_bits);
839 poll = hpb->params.timeout_polling_interval_ms;
840 schedule_delayed_work(&hpb->ufshpb_read_to_work,
841 msecs_to_jiffies(poll));
844 static void ufshpb_add_lru_info(struct victim_select_info *lru_info,
845 struct ufshpb_region *rgn)
847 rgn->rgn_state = HPB_RGN_ACTIVE;
848 list_add_tail(&rgn->list_lru_rgn, &lru_info->lh_lru_rgn);
849 atomic_inc(&lru_info->active_cnt);
850 if (rgn->hpb->is_hcm) {
852 ktime_add_ms(ktime_get(),
853 rgn->hpb->params.read_timeout_ms);
854 rgn->read_timeout_expiries =
855 rgn->hpb->params.read_timeout_expiries;
859 static void ufshpb_hit_lru_info(struct victim_select_info *lru_info,
860 struct ufshpb_region *rgn)
862 list_move_tail(&rgn->list_lru_rgn, &lru_info->lh_lru_rgn);
865 static struct ufshpb_region *ufshpb_victim_lru_info(struct ufshpb_lu *hpb)
867 struct victim_select_info *lru_info = &hpb->lru_info;
868 struct ufshpb_region *rgn, *victim_rgn = NULL;
870 list_for_each_entry(rgn, &lru_info->lh_lru_rgn, list_lru_rgn) {
871 if (ufshpb_check_srgns_issue_state(hpb, rgn))
875 * in host control mode, verify that the exiting region
879 rgn->reads > hpb->params.eviction_thld_exit)
887 dev_err(&hpb->sdev_ufs_lu->sdev_dev,
888 "%s: no region allocated\n",
894 static void ufshpb_cleanup_lru_info(struct victim_select_info *lru_info,
895 struct ufshpb_region *rgn)
897 list_del_init(&rgn->list_lru_rgn);
898 rgn->rgn_state = HPB_RGN_INACTIVE;
899 atomic_dec(&lru_info->active_cnt);
902 static void ufshpb_purge_active_subregion(struct ufshpb_lu *hpb,
903 struct ufshpb_subregion *srgn)
905 if (srgn->srgn_state != HPB_SRGN_UNUSED) {
906 ufshpb_put_map_ctx(hpb, srgn->mctx);
907 srgn->srgn_state = HPB_SRGN_UNUSED;
912 static int ufshpb_issue_umap_req(struct ufshpb_lu *hpb,
913 struct ufshpb_region *rgn,
916 struct ufshpb_req *umap_req;
917 int rgn_idx = rgn ? rgn->rgn_idx : 0;
919 umap_req = ufshpb_get_req(hpb, rgn_idx, REQ_OP_DRV_OUT, atomic);
923 ufshpb_execute_umap_req(hpb, umap_req, rgn);
928 static int ufshpb_issue_umap_single_req(struct ufshpb_lu *hpb,
929 struct ufshpb_region *rgn)
931 return ufshpb_issue_umap_req(hpb, rgn, true);
934 static void __ufshpb_evict_region(struct ufshpb_lu *hpb,
935 struct ufshpb_region *rgn)
937 struct victim_select_info *lru_info;
938 struct ufshpb_subregion *srgn;
941 lru_info = &hpb->lru_info;
943 dev_dbg(&hpb->sdev_ufs_lu->sdev_dev, "evict region %d\n", rgn->rgn_idx);
945 ufshpb_cleanup_lru_info(lru_info, rgn);
947 for_each_sub_region(rgn, srgn_idx, srgn)
948 ufshpb_purge_active_subregion(hpb, srgn);
951 static int ufshpb_evict_region(struct ufshpb_lu *hpb, struct ufshpb_region *rgn)
956 spin_lock_irqsave(&hpb->rgn_state_lock, flags);
957 if (rgn->rgn_state == HPB_RGN_PINNED) {
958 dev_warn(&hpb->sdev_ufs_lu->sdev_dev,
959 "pinned region cannot drop-out. region %d\n",
964 if (!list_empty(&rgn->list_lru_rgn)) {
965 if (ufshpb_check_srgns_issue_state(hpb, rgn)) {
971 spin_unlock_irqrestore(&hpb->rgn_state_lock, flags);
972 ret = ufshpb_issue_umap_single_req(hpb, rgn);
973 spin_lock_irqsave(&hpb->rgn_state_lock, flags);
978 __ufshpb_evict_region(hpb, rgn);
981 spin_unlock_irqrestore(&hpb->rgn_state_lock, flags);
985 static int ufshpb_issue_map_req(struct ufshpb_lu *hpb,
986 struct ufshpb_region *rgn,
987 struct ufshpb_subregion *srgn)
989 struct ufshpb_req *map_req;
993 bool alloc_required = false;
994 enum HPB_SRGN_STATE state = HPB_SRGN_INVALID;
996 spin_lock_irqsave(&hpb->rgn_state_lock, flags);
998 if (ufshpb_get_state(hpb) != HPB_PRESENT) {
999 dev_notice(&hpb->sdev_ufs_lu->sdev_dev,
1000 "%s: ufshpb state is not PRESENT\n", __func__);
1004 if ((rgn->rgn_state == HPB_RGN_INACTIVE) &&
1005 (srgn->srgn_state == HPB_SRGN_INVALID)) {
1010 if (srgn->srgn_state == HPB_SRGN_UNUSED)
1011 alloc_required = true;
1014 * If the subregion is already ISSUED state,
1015 * a specific event (e.g., GC or wear-leveling, etc.) occurs in
1016 * the device and HPB response for map loading is received.
1017 * In this case, after finishing the HPB_READ_BUFFER,
1018 * the next HPB_READ_BUFFER is performed again to obtain the latest
1021 if (srgn->srgn_state == HPB_SRGN_ISSUED)
1024 srgn->srgn_state = HPB_SRGN_ISSUED;
1025 spin_unlock_irqrestore(&hpb->rgn_state_lock, flags);
1027 if (alloc_required) {
1028 srgn->mctx = ufshpb_get_map_ctx(hpb, srgn->is_last);
1030 dev_err(&hpb->sdev_ufs_lu->sdev_dev,
1031 "get map_ctx failed. region %d - %d\n",
1032 rgn->rgn_idx, srgn->srgn_idx);
1033 state = HPB_SRGN_UNUSED;
1034 goto change_srgn_state;
1038 map_req = ufshpb_get_map_req(hpb, srgn);
1040 goto change_srgn_state;
1043 ret = ufshpb_execute_map_req(hpb, map_req, srgn->is_last);
1045 dev_err(&hpb->sdev_ufs_lu->sdev_dev,
1046 "%s: issue map_req failed: %d, region %d - %d\n",
1047 __func__, ret, srgn->rgn_idx, srgn->srgn_idx);
1053 ufshpb_put_map_req(hpb, map_req);
1055 spin_lock_irqsave(&hpb->rgn_state_lock, flags);
1056 srgn->srgn_state = state;
1058 spin_unlock_irqrestore(&hpb->rgn_state_lock, flags);
1062 static int ufshpb_add_region(struct ufshpb_lu *hpb, struct ufshpb_region *rgn)
1064 struct ufshpb_region *victim_rgn = NULL;
1065 struct victim_select_info *lru_info = &hpb->lru_info;
1066 unsigned long flags;
1069 spin_lock_irqsave(&hpb->rgn_state_lock, flags);
1071 * If region belongs to lru_list, just move the region
1072 * to the front of lru list because the state of the region
1073 * is already active-state.
1075 if (!list_empty(&rgn->list_lru_rgn)) {
1076 ufshpb_hit_lru_info(lru_info, rgn);
1080 if (rgn->rgn_state == HPB_RGN_INACTIVE) {
1081 if (atomic_read(&lru_info->active_cnt) ==
1082 lru_info->max_lru_active_cnt) {
1084 * If the maximum number of active regions
1085 * is exceeded, evict the least recently used region.
1086 * This case may occur when the device responds
1087 * to the eviction information late.
1088 * It is okay to evict the least recently used region,
1089 * because the device could detect this region
1090 * by not issuing HPB_READ
1092 * in host control mode, verify that the entering
1093 * region has enough reads
1096 rgn->reads < hpb->params.eviction_thld_enter) {
1101 victim_rgn = ufshpb_victim_lru_info(hpb);
1103 dev_warn(&hpb->sdev_ufs_lu->sdev_dev,
1104 "cannot get victim region %s\n",
1105 hpb->is_hcm ? "" : "error");
1110 dev_dbg(&hpb->sdev_ufs_lu->sdev_dev,
1111 "LRU full (%d), choose victim %d\n",
1112 atomic_read(&lru_info->active_cnt),
1113 victim_rgn->rgn_idx);
1116 spin_unlock_irqrestore(&hpb->rgn_state_lock,
1118 ret = ufshpb_issue_umap_single_req(hpb,
1120 spin_lock_irqsave(&hpb->rgn_state_lock,
1126 __ufshpb_evict_region(hpb, victim_rgn);
1130 * When a region is added to lru_info list_head,
1131 * it is guaranteed that the subregion has been
1132 * assigned all mctx. If failed, try to receive mctx again
1133 * without being added to lru_info list_head
1135 ufshpb_add_lru_info(lru_info, rgn);
1138 spin_unlock_irqrestore(&hpb->rgn_state_lock, flags);
1142 *ufshpb_submit_region_inactive() - submit a region to be inactivated later
1143 *@hpb: per-LU HPB instance
1144 *@region_index: the index associated with the region that will be inactivated later
1146 static void ufshpb_submit_region_inactive(struct ufshpb_lu *hpb, int region_index)
1148 int subregion_index;
1149 struct ufshpb_region *rgn;
1150 struct ufshpb_subregion *srgn;
1153 * Remove this region from active region list and add it to inactive list
1155 spin_lock(&hpb->rsp_list_lock);
1156 ufshpb_update_inactive_info(hpb, region_index);
1157 spin_unlock(&hpb->rsp_list_lock);
1159 rgn = hpb->rgn_tbl + region_index;
1162 * Set subregion state to be HPB_SRGN_INVALID, there will no HPB read on this subregion
1164 spin_lock(&hpb->rgn_state_lock);
1165 if (rgn->rgn_state != HPB_RGN_INACTIVE) {
1166 for (subregion_index = 0; subregion_index < rgn->srgn_cnt; subregion_index++) {
1167 srgn = rgn->srgn_tbl + subregion_index;
1168 if (srgn->srgn_state == HPB_SRGN_VALID)
1169 srgn->srgn_state = HPB_SRGN_INVALID;
1172 spin_unlock(&hpb->rgn_state_lock);
1175 static void ufshpb_rsp_req_region_update(struct ufshpb_lu *hpb,
1176 struct utp_hpb_rsp *rsp_field)
1178 struct ufshpb_region *rgn;
1179 struct ufshpb_subregion *srgn;
1180 int i, rgn_i, srgn_i;
1182 BUILD_BUG_ON(sizeof(struct ufshpb_active_field) != HPB_ACT_FIELD_SIZE);
1184 * If the active region and the inactive region are the same,
1185 * we will inactivate this region.
1186 * The device could check this (region inactivated) and
1187 * will response the proper active region information
1189 for (i = 0; i < rsp_field->active_rgn_cnt; i++) {
1191 be16_to_cpu(rsp_field->hpb_active_field[i].active_rgn);
1193 be16_to_cpu(rsp_field->hpb_active_field[i].active_srgn);
1195 rgn = hpb->rgn_tbl + rgn_i;
1197 (rgn->rgn_state != HPB_RGN_ACTIVE || is_rgn_dirty(rgn))) {
1199 * in host control mode, subregion activation
1200 * recommendations are only allowed to active regions.
1201 * Also, ignore recommendations for dirty regions - the
1202 * host will make decisions concerning those by himself
1207 dev_dbg(&hpb->sdev_ufs_lu->sdev_dev,
1208 "activate(%d) region %d - %d\n", i, rgn_i, srgn_i);
1210 spin_lock(&hpb->rsp_list_lock);
1211 ufshpb_update_active_info(hpb, rgn_i, srgn_i);
1212 spin_unlock(&hpb->rsp_list_lock);
1214 srgn = rgn->srgn_tbl + srgn_i;
1216 /* blocking HPB_READ */
1217 spin_lock(&hpb->rgn_state_lock);
1218 if (srgn->srgn_state == HPB_SRGN_VALID)
1219 srgn->srgn_state = HPB_SRGN_INVALID;
1220 spin_unlock(&hpb->rgn_state_lock);
1225 * in host control mode the device is not allowed to inactivate
1231 for (i = 0; i < rsp_field->inactive_rgn_cnt; i++) {
1232 rgn_i = be16_to_cpu(rsp_field->hpb_inactive_field[i]);
1233 dev_dbg(&hpb->sdev_ufs_lu->sdev_dev, "inactivate(%d) region %d\n", i, rgn_i);
1234 ufshpb_submit_region_inactive(hpb, rgn_i);
1238 dev_dbg(&hpb->sdev_ufs_lu->sdev_dev, "Noti: #ACT %u #INACT %u\n",
1239 rsp_field->active_rgn_cnt, rsp_field->inactive_rgn_cnt);
1241 if (ufshpb_get_state(hpb) == HPB_PRESENT)
1242 queue_work(ufshpb_wq, &hpb->map_work);
1246 * Set the flags of all active regions to RGN_FLAG_UPDATE to let host side reload L2P entries later
1248 static void ufshpb_set_regions_update(struct ufshpb_lu *hpb)
1250 struct victim_select_info *lru_info = &hpb->lru_info;
1251 struct ufshpb_region *rgn;
1252 unsigned long flags;
1254 spin_lock_irqsave(&hpb->rgn_state_lock, flags);
1256 list_for_each_entry(rgn, &lru_info->lh_lru_rgn, list_lru_rgn)
1257 set_bit(RGN_FLAG_UPDATE, &rgn->rgn_flags);
1259 spin_unlock_irqrestore(&hpb->rgn_state_lock, flags);
1262 static void ufshpb_dev_reset_handler(struct ufs_hba *hba)
1264 struct scsi_device *sdev;
1265 struct ufshpb_lu *hpb;
1267 __shost_for_each_device(sdev, hba->host) {
1268 hpb = ufshpb_get_hpb_data(sdev);
1274 * For the HPB host control mode, in case device powered up and lost HPB
1275 * information, we will set the region flag to be RGN_FLAG_UPDATE, it will
1276 * let host reload its L2P entries(reactivate region in the UFS device).
1278 ufshpb_set_regions_update(hpb);
1281 * For the HPB device control mode, if host side receives 02h:HPB Operation
1282 * in UPIU response, which means device recommends the host side should
1283 * inactivate all active regions. Here we add all active regions to inactive
1284 * list, they will be inactivated later in ufshpb_map_work_handler().
1286 struct victim_select_info *lru_info = &hpb->lru_info;
1287 struct ufshpb_region *rgn;
1289 list_for_each_entry(rgn, &lru_info->lh_lru_rgn, list_lru_rgn)
1290 ufshpb_submit_region_inactive(hpb, rgn->rgn_idx);
1292 if (ufshpb_get_state(hpb) == HPB_PRESENT)
1293 queue_work(ufshpb_wq, &hpb->map_work);
1299 * This function will parse recommended active subregion information in sense
1300 * data field of response UPIU with SAM_STAT_GOOD state.
1302 void ufshpb_rsp_upiu(struct ufs_hba *hba, struct ufshcd_lrb *lrbp)
1304 struct ufshpb_lu *hpb = ufshpb_get_hpb_data(lrbp->cmd->device);
1305 struct utp_hpb_rsp *rsp_field = &lrbp->ucd_rsp_ptr->hr;
1308 data_seg_len = be32_to_cpu(lrbp->ucd_rsp_ptr->header.dword_2)
1309 & MASK_RSP_UPIU_DATA_SEG_LEN;
1311 /* If data segment length is zero, rsp_field is not valid */
1315 if (unlikely(lrbp->lun != rsp_field->lun)) {
1316 struct scsi_device *sdev;
1319 __shost_for_each_device(sdev, hba->host) {
1320 hpb = ufshpb_get_hpb_data(sdev);
1325 if (rsp_field->lun == hpb->lun) {
1338 if (ufshpb_get_state(hpb) == HPB_INIT)
1341 if ((ufshpb_get_state(hpb) != HPB_PRESENT) &&
1342 (ufshpb_get_state(hpb) != HPB_SUSPEND)) {
1343 dev_notice(&hpb->sdev_ufs_lu->sdev_dev,
1344 "%s: ufshpb state is not PRESENT/SUSPEND\n",
1349 BUILD_BUG_ON(sizeof(struct utp_hpb_rsp) != UTP_HPB_RSP_SIZE);
1351 if (!ufshpb_is_hpb_rsp_valid(hba, lrbp, rsp_field))
1354 hpb->stats.rcmd_noti_cnt++;
1356 switch (rsp_field->hpb_op) {
1357 case HPB_RSP_REQ_REGION_UPDATE:
1358 if (data_seg_len != DEV_DATA_SEG_LEN)
1359 dev_warn(&hpb->sdev_ufs_lu->sdev_dev,
1360 "%s: data seg length is not same.\n",
1362 ufshpb_rsp_req_region_update(hpb, rsp_field);
1364 case HPB_RSP_DEV_RESET:
1365 dev_warn(&hpb->sdev_ufs_lu->sdev_dev,
1366 "UFS device lost HPB information during PM.\n");
1367 ufshpb_dev_reset_handler(hba);
1371 dev_notice(&hpb->sdev_ufs_lu->sdev_dev,
1372 "hpb_op is not available: %d\n",
1378 static void ufshpb_add_active_list(struct ufshpb_lu *hpb,
1379 struct ufshpb_region *rgn,
1380 struct ufshpb_subregion *srgn)
1382 if (!list_empty(&rgn->list_inact_rgn))
1385 if (!list_empty(&srgn->list_act_srgn)) {
1386 list_move(&srgn->list_act_srgn, &hpb->lh_act_srgn);
1390 list_add(&srgn->list_act_srgn, &hpb->lh_act_srgn);
1393 static void ufshpb_add_pending_evict_list(struct ufshpb_lu *hpb,
1394 struct ufshpb_region *rgn,
1395 struct list_head *pending_list)
1397 struct ufshpb_subregion *srgn;
1400 if (!list_empty(&rgn->list_inact_rgn))
1403 for_each_sub_region(rgn, srgn_idx, srgn)
1404 if (!list_empty(&srgn->list_act_srgn))
1407 list_add_tail(&rgn->list_inact_rgn, pending_list);
1410 static void ufshpb_run_active_subregion_list(struct ufshpb_lu *hpb)
1412 struct ufshpb_region *rgn;
1413 struct ufshpb_subregion *srgn;
1414 unsigned long flags;
1417 spin_lock_irqsave(&hpb->rsp_list_lock, flags);
1418 while ((srgn = list_first_entry_or_null(&hpb->lh_act_srgn,
1419 struct ufshpb_subregion,
1421 if (ufshpb_get_state(hpb) == HPB_SUSPEND)
1424 list_del_init(&srgn->list_act_srgn);
1425 spin_unlock_irqrestore(&hpb->rsp_list_lock, flags);
1427 rgn = hpb->rgn_tbl + srgn->rgn_idx;
1428 ret = ufshpb_add_region(hpb, rgn);
1432 ret = ufshpb_issue_map_req(hpb, rgn, srgn);
1434 dev_err(&hpb->sdev_ufs_lu->sdev_dev,
1435 "issue map_req failed. ret %d, region %d - %d\n",
1436 ret, rgn->rgn_idx, srgn->srgn_idx);
1439 spin_lock_irqsave(&hpb->rsp_list_lock, flags);
1441 spin_unlock_irqrestore(&hpb->rsp_list_lock, flags);
1445 dev_err(&hpb->sdev_ufs_lu->sdev_dev, "failed to activate region %d - %d, will retry\n",
1446 rgn->rgn_idx, srgn->srgn_idx);
1447 spin_lock_irqsave(&hpb->rsp_list_lock, flags);
1448 ufshpb_add_active_list(hpb, rgn, srgn);
1449 spin_unlock_irqrestore(&hpb->rsp_list_lock, flags);
1452 static void ufshpb_run_inactive_region_list(struct ufshpb_lu *hpb)
1454 struct ufshpb_region *rgn;
1455 unsigned long flags;
1457 LIST_HEAD(pending_list);
1459 spin_lock_irqsave(&hpb->rsp_list_lock, flags);
1460 while ((rgn = list_first_entry_or_null(&hpb->lh_inact_rgn,
1461 struct ufshpb_region,
1463 if (ufshpb_get_state(hpb) == HPB_SUSPEND)
1466 list_del_init(&rgn->list_inact_rgn);
1467 spin_unlock_irqrestore(&hpb->rsp_list_lock, flags);
1469 ret = ufshpb_evict_region(hpb, rgn);
1471 spin_lock_irqsave(&hpb->rsp_list_lock, flags);
1472 ufshpb_add_pending_evict_list(hpb, rgn, &pending_list);
1473 spin_unlock_irqrestore(&hpb->rsp_list_lock, flags);
1476 spin_lock_irqsave(&hpb->rsp_list_lock, flags);
1479 list_splice(&pending_list, &hpb->lh_inact_rgn);
1480 spin_unlock_irqrestore(&hpb->rsp_list_lock, flags);
1483 static void ufshpb_normalization_work_handler(struct work_struct *work)
1485 struct ufshpb_lu *hpb = container_of(work, struct ufshpb_lu,
1486 ufshpb_normalization_work);
1488 u8 factor = hpb->params.normalization_factor;
1490 for (rgn_idx = 0; rgn_idx < hpb->rgns_per_lu; rgn_idx++) {
1491 struct ufshpb_region *rgn = hpb->rgn_tbl + rgn_idx;
1494 spin_lock(&rgn->rgn_lock);
1496 for (srgn_idx = 0; srgn_idx < hpb->srgns_per_rgn; srgn_idx++) {
1497 struct ufshpb_subregion *srgn = rgn->srgn_tbl + srgn_idx;
1499 srgn->reads >>= factor;
1500 rgn->reads += srgn->reads;
1502 spin_unlock(&rgn->rgn_lock);
1504 if (rgn->rgn_state != HPB_RGN_ACTIVE || rgn->reads)
1507 /* if region is active but has no reads - inactivate it */
1508 spin_lock(&hpb->rsp_list_lock);
1509 ufshpb_update_inactive_info(hpb, rgn->rgn_idx);
1510 spin_unlock(&hpb->rsp_list_lock);
1514 static void ufshpb_map_work_handler(struct work_struct *work)
1516 struct ufshpb_lu *hpb = container_of(work, struct ufshpb_lu, map_work);
1518 if (ufshpb_get_state(hpb) != HPB_PRESENT) {
1519 dev_notice(&hpb->sdev_ufs_lu->sdev_dev,
1520 "%s: ufshpb state is not PRESENT\n", __func__);
1524 ufshpb_run_inactive_region_list(hpb);
1525 ufshpb_run_active_subregion_list(hpb);
1529 * this function doesn't need to hold lock due to be called in init.
1530 * (rgn_state_lock, rsp_list_lock, etc..)
1532 static int ufshpb_init_pinned_active_region(struct ufs_hba *hba,
1533 struct ufshpb_lu *hpb,
1534 struct ufshpb_region *rgn)
1536 struct ufshpb_subregion *srgn;
1540 for_each_sub_region(rgn, srgn_idx, srgn) {
1541 srgn->mctx = ufshpb_get_map_ctx(hpb, srgn->is_last);
1542 srgn->srgn_state = HPB_SRGN_INVALID;
1546 "alloc mctx for pinned region failed\n");
1550 list_add_tail(&srgn->list_act_srgn, &hpb->lh_act_srgn);
1553 rgn->rgn_state = HPB_RGN_PINNED;
1557 for (i = 0; i < srgn_idx; i++) {
1558 srgn = rgn->srgn_tbl + i;
1559 ufshpb_put_map_ctx(hpb, srgn->mctx);
1564 static void ufshpb_init_subregion_tbl(struct ufshpb_lu *hpb,
1565 struct ufshpb_region *rgn, bool last)
1568 struct ufshpb_subregion *srgn;
1570 for_each_sub_region(rgn, srgn_idx, srgn) {
1571 INIT_LIST_HEAD(&srgn->list_act_srgn);
1573 srgn->rgn_idx = rgn->rgn_idx;
1574 srgn->srgn_idx = srgn_idx;
1575 srgn->srgn_state = HPB_SRGN_UNUSED;
1578 if (unlikely(last && hpb->last_srgn_entries))
1579 srgn->is_last = true;
1582 static int ufshpb_alloc_subregion_tbl(struct ufshpb_lu *hpb,
1583 struct ufshpb_region *rgn, int srgn_cnt)
1585 rgn->srgn_tbl = kvcalloc(srgn_cnt, sizeof(struct ufshpb_subregion),
1590 rgn->srgn_cnt = srgn_cnt;
1594 static void ufshpb_lu_parameter_init(struct ufs_hba *hba,
1595 struct ufshpb_lu *hpb,
1596 struct ufshpb_dev_info *hpb_dev_info,
1597 struct ufshpb_lu_info *hpb_lu_info)
1599 u32 entries_per_rgn;
1600 u64 rgn_mem_size, tmp;
1602 if (ufshpb_is_legacy(hba))
1603 hpb->pre_req_max_tr_len = HPB_LEGACY_CHUNK_HIGH;
1605 hpb->pre_req_max_tr_len = hpb_dev_info->max_hpb_single_cmd;
1607 hpb->lu_pinned_start = hpb_lu_info->pinned_start;
1608 hpb->lu_pinned_end = hpb_lu_info->num_pinned ?
1609 (hpb_lu_info->pinned_start + hpb_lu_info->num_pinned - 1)
1611 hpb->lru_info.max_lru_active_cnt =
1612 hpb_lu_info->max_active_rgns - hpb_lu_info->num_pinned;
1614 rgn_mem_size = (1ULL << hpb_dev_info->rgn_size) * HPB_RGN_SIZE_UNIT
1616 do_div(rgn_mem_size, HPB_ENTRY_BLOCK_SIZE);
1617 hpb->srgn_mem_size = (1ULL << hpb_dev_info->srgn_size)
1618 * HPB_RGN_SIZE_UNIT / HPB_ENTRY_BLOCK_SIZE * HPB_ENTRY_SIZE;
1621 do_div(tmp, HPB_ENTRY_SIZE);
1622 entries_per_rgn = (u32)tmp;
1623 hpb->entries_per_rgn_shift = ilog2(entries_per_rgn);
1624 hpb->entries_per_rgn_mask = entries_per_rgn - 1;
1626 hpb->entries_per_srgn = hpb->srgn_mem_size / HPB_ENTRY_SIZE;
1627 hpb->entries_per_srgn_shift = ilog2(hpb->entries_per_srgn);
1628 hpb->entries_per_srgn_mask = hpb->entries_per_srgn - 1;
1631 do_div(tmp, hpb->srgn_mem_size);
1632 hpb->srgns_per_rgn = (int)tmp;
1634 hpb->rgns_per_lu = DIV_ROUND_UP(hpb_lu_info->num_blocks,
1636 hpb->srgns_per_lu = DIV_ROUND_UP(hpb_lu_info->num_blocks,
1637 (hpb->srgn_mem_size / HPB_ENTRY_SIZE));
1638 hpb->last_srgn_entries = hpb_lu_info->num_blocks
1639 % (hpb->srgn_mem_size / HPB_ENTRY_SIZE);
1641 hpb->pages_per_srgn = DIV_ROUND_UP(hpb->srgn_mem_size, PAGE_SIZE);
1643 if (hpb_dev_info->control_mode == HPB_HOST_CONTROL)
1647 static int ufshpb_alloc_region_tbl(struct ufs_hba *hba, struct ufshpb_lu *hpb)
1649 struct ufshpb_region *rgn_table, *rgn;
1653 rgn_table = kvcalloc(hpb->rgns_per_lu, sizeof(struct ufshpb_region),
1658 for (rgn_idx = 0; rgn_idx < hpb->rgns_per_lu; rgn_idx++) {
1659 int srgn_cnt = hpb->srgns_per_rgn;
1660 bool last_srgn = false;
1662 rgn = rgn_table + rgn_idx;
1663 rgn->rgn_idx = rgn_idx;
1665 spin_lock_init(&rgn->rgn_lock);
1667 INIT_LIST_HEAD(&rgn->list_inact_rgn);
1668 INIT_LIST_HEAD(&rgn->list_lru_rgn);
1669 INIT_LIST_HEAD(&rgn->list_expired_rgn);
1671 if (rgn_idx == hpb->rgns_per_lu - 1) {
1672 srgn_cnt = ((hpb->srgns_per_lu - 1) %
1673 hpb->srgns_per_rgn) + 1;
1677 ret = ufshpb_alloc_subregion_tbl(hpb, rgn, srgn_cnt);
1679 goto release_srgn_table;
1680 ufshpb_init_subregion_tbl(hpb, rgn, last_srgn);
1682 if (ufshpb_is_pinned_region(hpb, rgn_idx)) {
1683 ret = ufshpb_init_pinned_active_region(hba, hpb, rgn);
1685 goto release_srgn_table;
1687 rgn->rgn_state = HPB_RGN_INACTIVE;
1694 hpb->rgn_tbl = rgn_table;
1699 for (i = 0; i <= rgn_idx; i++)
1700 kvfree(rgn_table[i].srgn_tbl);
1706 static void ufshpb_destroy_subregion_tbl(struct ufshpb_lu *hpb,
1707 struct ufshpb_region *rgn)
1710 struct ufshpb_subregion *srgn;
1712 for_each_sub_region(rgn, srgn_idx, srgn)
1713 if (srgn->srgn_state != HPB_SRGN_UNUSED) {
1714 srgn->srgn_state = HPB_SRGN_UNUSED;
1715 ufshpb_put_map_ctx(hpb, srgn->mctx);
1719 static void ufshpb_destroy_region_tbl(struct ufshpb_lu *hpb)
1723 for (rgn_idx = 0; rgn_idx < hpb->rgns_per_lu; rgn_idx++) {
1724 struct ufshpb_region *rgn;
1726 rgn = hpb->rgn_tbl + rgn_idx;
1727 if (rgn->rgn_state != HPB_RGN_INACTIVE) {
1728 rgn->rgn_state = HPB_RGN_INACTIVE;
1730 ufshpb_destroy_subregion_tbl(hpb, rgn);
1733 kvfree(rgn->srgn_tbl);
1736 kvfree(hpb->rgn_tbl);
1739 /* SYSFS functions */
1740 #define ufshpb_sysfs_attr_show_func(__name) \
1741 static ssize_t __name##_show(struct device *dev, \
1742 struct device_attribute *attr, char *buf) \
1744 struct scsi_device *sdev = to_scsi_device(dev); \
1745 struct ufshpb_lu *hpb = ufshpb_get_hpb_data(sdev); \
1750 return sysfs_emit(buf, "%llu\n", hpb->stats.__name); \
1753 static DEVICE_ATTR_RO(__name)
1755 ufshpb_sysfs_attr_show_func(hit_cnt);
1756 ufshpb_sysfs_attr_show_func(miss_cnt);
1757 ufshpb_sysfs_attr_show_func(rcmd_noti_cnt);
1758 ufshpb_sysfs_attr_show_func(rcmd_active_cnt);
1759 ufshpb_sysfs_attr_show_func(rcmd_inactive_cnt);
1760 ufshpb_sysfs_attr_show_func(map_req_cnt);
1761 ufshpb_sysfs_attr_show_func(umap_req_cnt);
1763 static struct attribute *hpb_dev_stat_attrs[] = {
1764 &dev_attr_hit_cnt.attr,
1765 &dev_attr_miss_cnt.attr,
1766 &dev_attr_rcmd_noti_cnt.attr,
1767 &dev_attr_rcmd_active_cnt.attr,
1768 &dev_attr_rcmd_inactive_cnt.attr,
1769 &dev_attr_map_req_cnt.attr,
1770 &dev_attr_umap_req_cnt.attr,
1774 struct attribute_group ufs_sysfs_hpb_stat_group = {
1775 .name = "hpb_stats",
1776 .attrs = hpb_dev_stat_attrs,
1779 /* SYSFS functions */
1780 #define ufshpb_sysfs_param_show_func(__name) \
1781 static ssize_t __name##_show(struct device *dev, \
1782 struct device_attribute *attr, char *buf) \
1784 struct scsi_device *sdev = to_scsi_device(dev); \
1785 struct ufshpb_lu *hpb = ufshpb_get_hpb_data(sdev); \
1790 return sysfs_emit(buf, "%d\n", hpb->params.__name); \
1793 ufshpb_sysfs_param_show_func(requeue_timeout_ms);
1795 requeue_timeout_ms_store(struct device *dev, struct device_attribute *attr,
1796 const char *buf, size_t count)
1798 struct scsi_device *sdev = to_scsi_device(dev);
1799 struct ufshpb_lu *hpb = ufshpb_get_hpb_data(sdev);
1805 if (kstrtouint(buf, 0, &val))
1811 hpb->params.requeue_timeout_ms = val;
1815 static DEVICE_ATTR_RW(requeue_timeout_ms);
1817 ufshpb_sysfs_param_show_func(activation_thld);
1819 activation_thld_store(struct device *dev, struct device_attribute *attr,
1820 const char *buf, size_t count)
1822 struct scsi_device *sdev = to_scsi_device(dev);
1823 struct ufshpb_lu *hpb = ufshpb_get_hpb_data(sdev);
1832 if (kstrtouint(buf, 0, &val))
1838 hpb->params.activation_thld = val;
1842 static DEVICE_ATTR_RW(activation_thld);
1844 ufshpb_sysfs_param_show_func(normalization_factor);
1846 normalization_factor_store(struct device *dev, struct device_attribute *attr,
1847 const char *buf, size_t count)
1849 struct scsi_device *sdev = to_scsi_device(dev);
1850 struct ufshpb_lu *hpb = ufshpb_get_hpb_data(sdev);
1859 if (kstrtouint(buf, 0, &val))
1862 if (val <= 0 || val > ilog2(hpb->entries_per_srgn))
1865 hpb->params.normalization_factor = val;
1869 static DEVICE_ATTR_RW(normalization_factor);
1871 ufshpb_sysfs_param_show_func(eviction_thld_enter);
1873 eviction_thld_enter_store(struct device *dev, struct device_attribute *attr,
1874 const char *buf, size_t count)
1876 struct scsi_device *sdev = to_scsi_device(dev);
1877 struct ufshpb_lu *hpb = ufshpb_get_hpb_data(sdev);
1886 if (kstrtouint(buf, 0, &val))
1889 if (val <= hpb->params.eviction_thld_exit)
1892 hpb->params.eviction_thld_enter = val;
1896 static DEVICE_ATTR_RW(eviction_thld_enter);
1898 ufshpb_sysfs_param_show_func(eviction_thld_exit);
1900 eviction_thld_exit_store(struct device *dev, struct device_attribute *attr,
1901 const char *buf, size_t count)
1903 struct scsi_device *sdev = to_scsi_device(dev);
1904 struct ufshpb_lu *hpb = ufshpb_get_hpb_data(sdev);
1913 if (kstrtouint(buf, 0, &val))
1916 if (val <= hpb->params.activation_thld)
1919 hpb->params.eviction_thld_exit = val;
1923 static DEVICE_ATTR_RW(eviction_thld_exit);
1925 ufshpb_sysfs_param_show_func(read_timeout_ms);
1927 read_timeout_ms_store(struct device *dev, struct device_attribute *attr,
1928 const char *buf, size_t count)
1930 struct scsi_device *sdev = to_scsi_device(dev);
1931 struct ufshpb_lu *hpb = ufshpb_get_hpb_data(sdev);
1940 if (kstrtouint(buf, 0, &val))
1943 /* read_timeout >> timeout_polling_interval */
1944 if (val < hpb->params.timeout_polling_interval_ms * 2)
1947 hpb->params.read_timeout_ms = val;
1951 static DEVICE_ATTR_RW(read_timeout_ms);
1953 ufshpb_sysfs_param_show_func(read_timeout_expiries);
1955 read_timeout_expiries_store(struct device *dev, struct device_attribute *attr,
1956 const char *buf, size_t count)
1958 struct scsi_device *sdev = to_scsi_device(dev);
1959 struct ufshpb_lu *hpb = ufshpb_get_hpb_data(sdev);
1968 if (kstrtouint(buf, 0, &val))
1974 hpb->params.read_timeout_expiries = val;
1978 static DEVICE_ATTR_RW(read_timeout_expiries);
1980 ufshpb_sysfs_param_show_func(timeout_polling_interval_ms);
1982 timeout_polling_interval_ms_store(struct device *dev,
1983 struct device_attribute *attr,
1984 const char *buf, size_t count)
1986 struct scsi_device *sdev = to_scsi_device(dev);
1987 struct ufshpb_lu *hpb = ufshpb_get_hpb_data(sdev);
1996 if (kstrtouint(buf, 0, &val))
1999 /* timeout_polling_interval << read_timeout */
2000 if (val <= 0 || val > hpb->params.read_timeout_ms / 2)
2003 hpb->params.timeout_polling_interval_ms = val;
2007 static DEVICE_ATTR_RW(timeout_polling_interval_ms);
2009 ufshpb_sysfs_param_show_func(inflight_map_req);
2010 static ssize_t inflight_map_req_store(struct device *dev,
2011 struct device_attribute *attr,
2012 const char *buf, size_t count)
2014 struct scsi_device *sdev = to_scsi_device(dev);
2015 struct ufshpb_lu *hpb = ufshpb_get_hpb_data(sdev);
2024 if (kstrtouint(buf, 0, &val))
2027 if (val <= 0 || val > hpb->sdev_ufs_lu->queue_depth - 1)
2030 hpb->params.inflight_map_req = val;
2034 static DEVICE_ATTR_RW(inflight_map_req);
2036 static void ufshpb_hcm_param_init(struct ufshpb_lu *hpb)
2038 hpb->params.activation_thld = ACTIVATION_THRESHOLD;
2039 hpb->params.normalization_factor = 1;
2040 hpb->params.eviction_thld_enter = (ACTIVATION_THRESHOLD << 5);
2041 hpb->params.eviction_thld_exit = (ACTIVATION_THRESHOLD << 4);
2042 hpb->params.read_timeout_ms = READ_TO_MS;
2043 hpb->params.read_timeout_expiries = READ_TO_EXPIRIES;
2044 hpb->params.timeout_polling_interval_ms = POLLING_INTERVAL_MS;
2045 hpb->params.inflight_map_req = THROTTLE_MAP_REQ_DEFAULT;
2048 static struct attribute *hpb_dev_param_attrs[] = {
2049 &dev_attr_requeue_timeout_ms.attr,
2050 &dev_attr_activation_thld.attr,
2051 &dev_attr_normalization_factor.attr,
2052 &dev_attr_eviction_thld_enter.attr,
2053 &dev_attr_eviction_thld_exit.attr,
2054 &dev_attr_read_timeout_ms.attr,
2055 &dev_attr_read_timeout_expiries.attr,
2056 &dev_attr_timeout_polling_interval_ms.attr,
2057 &dev_attr_inflight_map_req.attr,
2061 struct attribute_group ufs_sysfs_hpb_param_group = {
2062 .name = "hpb_params",
2063 .attrs = hpb_dev_param_attrs,
2066 static int ufshpb_pre_req_mempool_init(struct ufshpb_lu *hpb)
2068 struct ufshpb_req *pre_req = NULL, *t;
2069 int qd = hpb->sdev_ufs_lu->queue_depth / 2;
2072 INIT_LIST_HEAD(&hpb->lh_pre_req_free);
2074 hpb->pre_req = kcalloc(qd, sizeof(struct ufshpb_req), GFP_KERNEL);
2075 hpb->throttle_pre_req = qd;
2076 hpb->num_inflight_pre_req = 0;
2081 for (i = 0; i < qd; i++) {
2082 pre_req = hpb->pre_req + i;
2083 INIT_LIST_HEAD(&pre_req->list_req);
2084 pre_req->req = NULL;
2086 pre_req->bio = bio_alloc(NULL, 1, 0, GFP_KERNEL);
2090 pre_req->wb.m_page = alloc_page(GFP_KERNEL | __GFP_ZERO);
2091 if (!pre_req->wb.m_page) {
2092 bio_put(pre_req->bio);
2096 list_add_tail(&pre_req->list_req, &hpb->lh_pre_req_free);
2101 list_for_each_entry_safe(pre_req, t, &hpb->lh_pre_req_free, list_req) {
2102 list_del_init(&pre_req->list_req);
2103 bio_put(pre_req->bio);
2104 __free_page(pre_req->wb.m_page);
2107 kfree(hpb->pre_req);
2111 static void ufshpb_pre_req_mempool_destroy(struct ufshpb_lu *hpb)
2113 struct ufshpb_req *pre_req = NULL;
2116 for (i = 0; i < hpb->throttle_pre_req; i++) {
2117 pre_req = hpb->pre_req + i;
2118 bio_put(hpb->pre_req[i].bio);
2119 if (!pre_req->wb.m_page)
2120 __free_page(hpb->pre_req[i].wb.m_page);
2121 list_del_init(&pre_req->list_req);
2124 kfree(hpb->pre_req);
2127 static void ufshpb_stat_init(struct ufshpb_lu *hpb)
2129 hpb->stats.hit_cnt = 0;
2130 hpb->stats.miss_cnt = 0;
2131 hpb->stats.rcmd_noti_cnt = 0;
2132 hpb->stats.rcmd_active_cnt = 0;
2133 hpb->stats.rcmd_inactive_cnt = 0;
2134 hpb->stats.map_req_cnt = 0;
2135 hpb->stats.umap_req_cnt = 0;
2138 static void ufshpb_param_init(struct ufshpb_lu *hpb)
2140 hpb->params.requeue_timeout_ms = HPB_REQUEUE_TIME_MS;
2142 ufshpb_hcm_param_init(hpb);
2145 static int ufshpb_lu_hpb_init(struct ufs_hba *hba, struct ufshpb_lu *hpb)
2149 spin_lock_init(&hpb->rgn_state_lock);
2150 spin_lock_init(&hpb->rsp_list_lock);
2151 spin_lock_init(&hpb->param_lock);
2153 INIT_LIST_HEAD(&hpb->lru_info.lh_lru_rgn);
2154 INIT_LIST_HEAD(&hpb->lh_act_srgn);
2155 INIT_LIST_HEAD(&hpb->lh_inact_rgn);
2156 INIT_LIST_HEAD(&hpb->list_hpb_lu);
2158 INIT_WORK(&hpb->map_work, ufshpb_map_work_handler);
2160 INIT_WORK(&hpb->ufshpb_normalization_work,
2161 ufshpb_normalization_work_handler);
2162 INIT_DELAYED_WORK(&hpb->ufshpb_read_to_work,
2163 ufshpb_read_to_handler);
2166 hpb->map_req_cache = kmem_cache_create("ufshpb_req_cache",
2167 sizeof(struct ufshpb_req), 0, 0, NULL);
2168 if (!hpb->map_req_cache) {
2169 dev_err(hba->dev, "ufshpb(%d) ufshpb_req_cache create fail",
2174 hpb->m_page_cache = kmem_cache_create("ufshpb_m_page_cache",
2175 sizeof(struct page *) * hpb->pages_per_srgn,
2177 if (!hpb->m_page_cache) {
2178 dev_err(hba->dev, "ufshpb(%d) ufshpb_m_page_cache create fail",
2181 goto release_req_cache;
2184 ret = ufshpb_pre_req_mempool_init(hpb);
2186 dev_err(hba->dev, "ufshpb(%d) pre_req_mempool init fail",
2188 goto release_m_page_cache;
2191 ret = ufshpb_alloc_region_tbl(hba, hpb);
2193 goto release_pre_req_mempool;
2195 ufshpb_stat_init(hpb);
2196 ufshpb_param_init(hpb);
2201 poll = hpb->params.timeout_polling_interval_ms;
2202 schedule_delayed_work(&hpb->ufshpb_read_to_work,
2203 msecs_to_jiffies(poll));
2208 release_pre_req_mempool:
2209 ufshpb_pre_req_mempool_destroy(hpb);
2210 release_m_page_cache:
2211 kmem_cache_destroy(hpb->m_page_cache);
2213 kmem_cache_destroy(hpb->map_req_cache);
2217 static struct ufshpb_lu *
2218 ufshpb_alloc_hpb_lu(struct ufs_hba *hba, struct scsi_device *sdev,
2219 struct ufshpb_dev_info *hpb_dev_info,
2220 struct ufshpb_lu_info *hpb_lu_info)
2222 struct ufshpb_lu *hpb;
2225 hpb = kzalloc(sizeof(struct ufshpb_lu), GFP_KERNEL);
2229 hpb->lun = sdev->lun;
2230 hpb->sdev_ufs_lu = sdev;
2232 ufshpb_lu_parameter_init(hba, hpb, hpb_dev_info, hpb_lu_info);
2234 ret = ufshpb_lu_hpb_init(hba, hpb);
2236 dev_err(hba->dev, "hpb lu init failed. ret %d", ret);
2240 sdev->hostdata = hpb;
2248 static void ufshpb_discard_rsp_lists(struct ufshpb_lu *hpb)
2250 struct ufshpb_region *rgn, *next_rgn;
2251 struct ufshpb_subregion *srgn, *next_srgn;
2252 unsigned long flags;
2255 * If the device reset occurred, the remaining HPB region information
2256 * may be stale. Therefore, by discarding the lists of HPB response
2257 * that remained after reset, we prevent unnecessary work.
2259 spin_lock_irqsave(&hpb->rsp_list_lock, flags);
2260 list_for_each_entry_safe(rgn, next_rgn, &hpb->lh_inact_rgn,
2262 list_del_init(&rgn->list_inact_rgn);
2264 list_for_each_entry_safe(srgn, next_srgn, &hpb->lh_act_srgn,
2266 list_del_init(&srgn->list_act_srgn);
2267 spin_unlock_irqrestore(&hpb->rsp_list_lock, flags);
2270 static void ufshpb_cancel_jobs(struct ufshpb_lu *hpb)
2273 cancel_delayed_work_sync(&hpb->ufshpb_read_to_work);
2274 cancel_work_sync(&hpb->ufshpb_normalization_work);
2276 cancel_work_sync(&hpb->map_work);
2279 static bool ufshpb_check_hpb_reset_query(struct ufs_hba *hba)
2282 bool flag_res = true;
2285 /* wait for the device to complete HPB reset query */
2286 for (try = 0; try < HPB_RESET_REQ_RETRIES; try++) {
2288 "%s start flag reset polling %d times\n",
2291 /* Poll fHpbReset flag to be cleared */
2292 err = ufshcd_query_flag(hba, UPIU_QUERY_OPCODE_READ_FLAG,
2293 QUERY_FLAG_IDN_HPB_RESET, 0, &flag_res);
2297 "%s reading fHpbReset flag failed with error %d\n",
2305 usleep_range(1000, 1100);
2309 "%s fHpbReset was not cleared by the device\n",
2317 * ufshpb_toggle_state - switch HPB state of all LUs
2318 * @hba: per-adapter instance
2319 * @src: expected current HPB state
2320 * @dest: target HPB state to switch to
2322 void ufshpb_toggle_state(struct ufs_hba *hba, enum UFSHPB_STATE src, enum UFSHPB_STATE dest)
2324 struct ufshpb_lu *hpb;
2325 struct scsi_device *sdev;
2327 shost_for_each_device(sdev, hba->host) {
2328 hpb = ufshpb_get_hpb_data(sdev);
2330 if (!hpb || ufshpb_get_state(hpb) != src)
2332 ufshpb_set_state(hpb, dest);
2334 if (dest == HPB_RESET) {
2335 ufshpb_cancel_jobs(hpb);
2336 ufshpb_discard_rsp_lists(hpb);
2341 void ufshpb_suspend(struct ufs_hba *hba)
2343 struct ufshpb_lu *hpb;
2344 struct scsi_device *sdev;
2346 shost_for_each_device(sdev, hba->host) {
2347 hpb = ufshpb_get_hpb_data(sdev);
2348 if (!hpb || ufshpb_get_state(hpb) != HPB_PRESENT)
2351 ufshpb_set_state(hpb, HPB_SUSPEND);
2352 ufshpb_cancel_jobs(hpb);
2356 void ufshpb_resume(struct ufs_hba *hba)
2358 struct ufshpb_lu *hpb;
2359 struct scsi_device *sdev;
2361 shost_for_each_device(sdev, hba->host) {
2362 hpb = ufshpb_get_hpb_data(sdev);
2363 if (!hpb || ufshpb_get_state(hpb) != HPB_SUSPEND)
2366 ufshpb_set_state(hpb, HPB_PRESENT);
2367 ufshpb_kick_map_work(hpb);
2369 unsigned int poll = hpb->params.timeout_polling_interval_ms;
2371 schedule_delayed_work(&hpb->ufshpb_read_to_work, msecs_to_jiffies(poll));
2376 static int ufshpb_get_lu_info(struct ufs_hba *hba, int lun,
2377 struct ufshpb_lu_info *hpb_lu_info)
2379 u16 max_active_rgns;
2383 char desc_buf[QUERY_DESC_MAX_SIZE];
2385 ufshcd_map_desc_id_to_length(hba, QUERY_DESC_IDN_UNIT, &size);
2387 ufshcd_rpm_get_sync(hba);
2388 ret = ufshcd_query_descriptor_retry(hba, UPIU_QUERY_OPCODE_READ_DESC,
2389 QUERY_DESC_IDN_UNIT, lun, 0,
2391 ufshcd_rpm_put_sync(hba);
2395 "%s: idn: %d lun: %d query request failed",
2396 __func__, QUERY_DESC_IDN_UNIT, lun);
2400 lu_enable = desc_buf[UNIT_DESC_PARAM_LU_ENABLE];
2401 if (lu_enable != LU_ENABLED_HPB_FUNC)
2404 max_active_rgns = get_unaligned_be16(
2405 desc_buf + UNIT_DESC_PARAM_HPB_LU_MAX_ACTIVE_RGNS);
2406 if (!max_active_rgns) {
2408 "lun %d wrong number of max active regions\n", lun);
2412 hpb_lu_info->num_blocks = get_unaligned_be64(
2413 desc_buf + UNIT_DESC_PARAM_LOGICAL_BLK_COUNT);
2414 hpb_lu_info->pinned_start = get_unaligned_be16(
2415 desc_buf + UNIT_DESC_PARAM_HPB_PIN_RGN_START_OFF);
2416 hpb_lu_info->num_pinned = get_unaligned_be16(
2417 desc_buf + UNIT_DESC_PARAM_HPB_NUM_PIN_RGNS);
2418 hpb_lu_info->max_active_rgns = max_active_rgns;
2423 void ufshpb_destroy_lu(struct ufs_hba *hba, struct scsi_device *sdev)
2425 struct ufshpb_lu *hpb = ufshpb_get_hpb_data(sdev);
2430 ufshpb_set_state(hpb, HPB_FAILED);
2432 sdev = hpb->sdev_ufs_lu;
2433 sdev->hostdata = NULL;
2435 ufshpb_cancel_jobs(hpb);
2437 ufshpb_pre_req_mempool_destroy(hpb);
2438 ufshpb_destroy_region_tbl(hpb);
2440 kmem_cache_destroy(hpb->map_req_cache);
2441 kmem_cache_destroy(hpb->m_page_cache);
2443 list_del_init(&hpb->list_hpb_lu);
2448 static void ufshpb_hpb_lu_prepared(struct ufs_hba *hba)
2451 struct ufshpb_lu *hpb;
2452 struct scsi_device *sdev;
2455 if (tot_active_srgn_pages == 0) {
2460 init_success = !ufshpb_check_hpb_reset_query(hba);
2462 pool_size = PAGE_ALIGN(ufshpb_host_map_kbytes * 1024) / PAGE_SIZE;
2463 if (pool_size > tot_active_srgn_pages) {
2464 mempool_resize(ufshpb_mctx_pool, tot_active_srgn_pages);
2465 mempool_resize(ufshpb_page_pool, tot_active_srgn_pages);
2468 shost_for_each_device(sdev, hba->host) {
2469 hpb = ufshpb_get_hpb_data(sdev);
2474 ufshpb_set_state(hpb, HPB_PRESENT);
2475 if ((hpb->lu_pinned_end - hpb->lu_pinned_start) > 0)
2476 queue_work(ufshpb_wq, &hpb->map_work);
2478 dev_err(hba->dev, "destroy HPB lu %d\n", hpb->lun);
2479 ufshpb_destroy_lu(hba, sdev);
2487 void ufshpb_init_hpb_lu(struct ufs_hba *hba, struct scsi_device *sdev)
2489 struct ufshpb_lu *hpb;
2491 struct ufshpb_lu_info hpb_lu_info = { 0 };
2492 int lun = sdev->lun;
2494 if (lun >= hba->dev_info.max_lu_supported)
2497 ret = ufshpb_get_lu_info(hba, lun, &hpb_lu_info);
2501 hpb = ufshpb_alloc_hpb_lu(hba, sdev, &hba->ufshpb_dev,
2506 tot_active_srgn_pages += hpb_lu_info.max_active_rgns *
2507 hpb->srgns_per_rgn * hpb->pages_per_srgn;
2510 /* All LUs are initialized */
2511 if (atomic_dec_and_test(&hba->ufshpb_dev.slave_conf_cnt))
2512 ufshpb_hpb_lu_prepared(hba);
2515 static int ufshpb_init_mem_wq(struct ufs_hba *hba)
2518 unsigned int pool_size;
2520 ufshpb_mctx_cache = kmem_cache_create("ufshpb_mctx_cache",
2521 sizeof(struct ufshpb_map_ctx),
2523 if (!ufshpb_mctx_cache) {
2524 dev_err(hba->dev, "ufshpb: cannot init mctx cache\n");
2528 pool_size = PAGE_ALIGN(ufshpb_host_map_kbytes * 1024) / PAGE_SIZE;
2529 dev_info(hba->dev, "%s:%d ufshpb_host_map_kbytes %u pool_size %u\n",
2530 __func__, __LINE__, ufshpb_host_map_kbytes, pool_size);
2532 ufshpb_mctx_pool = mempool_create_slab_pool(pool_size,
2534 if (!ufshpb_mctx_pool) {
2535 dev_err(hba->dev, "ufshpb: cannot init mctx pool\n");
2537 goto release_mctx_cache;
2540 ufshpb_page_pool = mempool_create_page_pool(pool_size, 0);
2541 if (!ufshpb_page_pool) {
2542 dev_err(hba->dev, "ufshpb: cannot init page pool\n");
2544 goto release_mctx_pool;
2547 ufshpb_wq = alloc_workqueue("ufshpb-wq",
2548 WQ_UNBOUND | WQ_MEM_RECLAIM, 0);
2550 dev_err(hba->dev, "ufshpb: alloc workqueue failed\n");
2552 goto release_page_pool;
2558 mempool_destroy(ufshpb_page_pool);
2560 mempool_destroy(ufshpb_mctx_pool);
2562 kmem_cache_destroy(ufshpb_mctx_cache);
2566 void ufshpb_get_geo_info(struct ufs_hba *hba, u8 *geo_buf)
2568 struct ufshpb_dev_info *hpb_info = &hba->ufshpb_dev;
2569 int max_active_rgns = 0;
2572 hpb_num_lu = geo_buf[GEOMETRY_DESC_PARAM_HPB_NUMBER_LU];
2573 if (hpb_num_lu == 0) {
2574 dev_err(hba->dev, "No HPB LU supported\n");
2575 hpb_info->hpb_disabled = true;
2579 hpb_info->rgn_size = geo_buf[GEOMETRY_DESC_PARAM_HPB_REGION_SIZE];
2580 hpb_info->srgn_size = geo_buf[GEOMETRY_DESC_PARAM_HPB_SUBREGION_SIZE];
2581 max_active_rgns = get_unaligned_be16(geo_buf +
2582 GEOMETRY_DESC_PARAM_HPB_MAX_ACTIVE_REGS);
2584 if (hpb_info->rgn_size == 0 || hpb_info->srgn_size == 0 ||
2585 max_active_rgns == 0) {
2586 dev_err(hba->dev, "No HPB supported device\n");
2587 hpb_info->hpb_disabled = true;
2592 void ufshpb_get_dev_info(struct ufs_hba *hba, u8 *desc_buf)
2594 struct ufshpb_dev_info *hpb_dev_info = &hba->ufshpb_dev;
2598 hpb_dev_info->control_mode = desc_buf[DEVICE_DESC_PARAM_HPB_CONTROL];
2600 version = get_unaligned_be16(desc_buf + DEVICE_DESC_PARAM_HPB_VER);
2601 if ((version != HPB_SUPPORT_VERSION) &&
2602 (version != HPB_SUPPORT_LEGACY_VERSION)) {
2603 dev_err(hba->dev, "%s: HPB %x version is not supported.\n",
2605 hpb_dev_info->hpb_disabled = true;
2609 if (version == HPB_SUPPORT_LEGACY_VERSION)
2610 hpb_dev_info->is_legacy = true;
2613 * Get the number of user logical unit to check whether all
2614 * scsi_device finish initialization
2616 hpb_dev_info->num_lu = desc_buf[DEVICE_DESC_PARAM_NUM_LU];
2618 if (hpb_dev_info->is_legacy)
2621 ret = ufshcd_query_attr_retry(hba, UPIU_QUERY_OPCODE_READ_ATTR,
2622 QUERY_ATTR_IDN_MAX_HPB_SINGLE_CMD, 0, 0, &max_single_cmd);
2625 hpb_dev_info->max_hpb_single_cmd = HPB_LEGACY_CHUNK_HIGH;
2627 hpb_dev_info->max_hpb_single_cmd = min(max_single_cmd + 1, HPB_MULTI_CHUNK_HIGH);
2630 void ufshpb_init(struct ufs_hba *hba)
2632 struct ufshpb_dev_info *hpb_dev_info = &hba->ufshpb_dev;
2636 if (!ufshpb_is_allowed(hba) || !hba->dev_info.hpb_enabled)
2639 if (ufshpb_init_mem_wq(hba)) {
2640 hpb_dev_info->hpb_disabled = true;
2644 atomic_set(&hpb_dev_info->slave_conf_cnt, hpb_dev_info->num_lu);
2645 tot_active_srgn_pages = 0;
2646 /* issue HPB reset query */
2647 for (try = 0; try < HPB_RESET_REQ_RETRIES; try++) {
2648 ret = ufshcd_query_flag(hba, UPIU_QUERY_OPCODE_SET_FLAG,
2649 QUERY_FLAG_IDN_HPB_RESET, 0, NULL);
2655 void ufshpb_remove(struct ufs_hba *hba)
2657 mempool_destroy(ufshpb_page_pool);
2658 mempool_destroy(ufshpb_mctx_pool);
2659 kmem_cache_destroy(ufshpb_mctx_cache);
2661 destroy_workqueue(ufshpb_wq);
2664 module_param(ufshpb_host_map_kbytes, uint, 0644);
2665 MODULE_PARM_DESC(ufshpb_host_map_kbytes,
2666 "ufshpb host mapping memory kilo-bytes for ufshpb memory-pool");