From patchwork Wed Jun 15 13:28:46 2022 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Chao Yu X-Patchwork-Id: 582170 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by smtp.lore.kernel.org (Postfix) with ESMTP id CDC06C433EF for ; Wed, 15 Jun 2022 13:29:26 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S238254AbiFON3Y (ORCPT ); Wed, 15 Jun 2022 09:29:24 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:35882 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1352214AbiFON25 (ORCPT ); Wed, 15 Jun 2022 09:28:57 -0400 Received: from dfw.source.kernel.org (dfw.source.kernel.org [IPv6:2604:1380:4641:c500::1]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 2BE1860F3; Wed, 15 Jun 2022 06:28:56 -0700 (PDT) Received: from smtp.kernel.org (relay.kernel.org [52.25.139.140]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by dfw.source.kernel.org (Postfix) with ESMTPS id BC6ED619E4; Wed, 15 Jun 2022 13:28:55 +0000 (UTC) Received: by smtp.kernel.org (Postfix) with ESMTPSA id 8FFADC3411C; Wed, 15 Jun 2022 13:28:53 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=kernel.org; s=k20201202; t=1655299735; bh=S+lkWBMeibl/wRB+W+22IKyiUIA6vhzXK2JHutJeaMQ=; h=From:To:Cc:Subject:Date:From; b=Mdkc/dapTOik9/cI6wxdBJtwsiBsWK+VlAwwI454s8XhAYKYRyliDiKHid9FgusGs 5jzOkk1h/qoAsdMqGN/LZl9x1by2P6Z/+9m910A4gynyBqcSOceU7mUmeOTnVxOep/ fQzx+8waRtjdSC2sIdy+DfKhAxyYsd2sVjrBrnSPyYBjmN3iDrBwu2XRwkd62gMiXI v5pDZj1XGElf6mlypLRI5pHPVVAhsxBbOK9rRv7gh0W8qgQ2aKdC+Jph6lZZsSBGZF Asd4rpqkEh2GJRV+DNyk/RARMIzd7I6M+OO2PSH5N/Fm0qtvNaBHLGYbvsXvDNxZ7e fQNvrEzajJyuw== From: Chao Yu To: jejb@linux.ibm.com, martin.petersen@oracle.com Cc: linux-scsi@vger.kernel.org, linux-kernel@vger.kernel.org, chao@kernel.org Subject: [PATCH v2] scsi: support packing multi-segment in UNMAP command Date: Wed, 15 Jun 2022 21:28:46 +0800 Message-Id: <20220615132846.2232920-1-chao@kernel.org> X-Mailer: git-send-email 2.25.1 MIME-Version: 1.0 Precedence: bulk List-ID: X-Mailing-List: linux-scsi@vger.kernel.org As SPEC describes that it can support unmapping one or more LBA range in single UNMAP command, however, previously we only pack one LBA range in UNMAP command by default no matter device gives the block limits that says it can support in-batch UNMAP. This patch tries to set max_discard_segments config according to block limits of device, and supports in-batch UNMAP. Signed-off-by: Chao Yu --- v2: - rebase the code. drivers/scsi/sd.c | 30 +++++++++++++++++++----------- drivers/scsi/sd.h | 1 + 2 files changed, 20 insertions(+), 11 deletions(-) diff --git a/drivers/scsi/sd.c b/drivers/scsi/sd.c index 895b56c8f25e..143b4eecf657 100644 --- a/drivers/scsi/sd.c +++ b/drivers/scsi/sd.c @@ -790,6 +790,7 @@ static void sd_config_discard(struct scsi_disk *sdkp, unsigned int mode) q->limits.discard_granularity = max(sdkp->physical_block_size, sdkp->unmap_granularity * logical_block_size); + blk_queue_max_discard_segments(q, sdkp->max_block_desc_count); sdkp->provisioning_mode = mode; switch (mode) { @@ -835,10 +836,10 @@ static blk_status_t sd_setup_unmap_cmnd(struct scsi_cmnd *cmd) { struct scsi_device *sdp = cmd->device; struct request *rq = scsi_cmd_to_rq(cmd); + struct bio *bio; struct scsi_disk *sdkp = scsi_disk(rq->q->disk); - u64 lba = sectors_to_logical(sdp, blk_rq_pos(rq)); - u32 nr_blocks = sectors_to_logical(sdp, blk_rq_sectors(rq)); - unsigned int data_len = 24; + unsigned short segments = blk_rq_nr_discard_segments(rq); + unsigned int data_len = 8 + 16 * segments, i = 0; char *buf; rq->special_vec.bv_page = mempool_alloc(sd_page_pool, GFP_ATOMIC); @@ -851,13 +852,20 @@ static blk_status_t sd_setup_unmap_cmnd(struct scsi_cmnd *cmd) cmd->cmd_len = 10; cmd->cmnd[0] = UNMAP; - cmd->cmnd[8] = 24; + cmd->cmnd[8] = data_len; buf = bvec_virt(&rq->special_vec); - put_unaligned_be16(6 + 16, &buf[0]); - put_unaligned_be16(16, &buf[2]); - put_unaligned_be64(lba, &buf[8]); - put_unaligned_be32(nr_blocks, &buf[16]); + put_unaligned_be16(6 + 16 * segments, &buf[0]); + put_unaligned_be16(16 * segments, &buf[2]); + + __rq_for_each_bio(bio, rq) { + u64 lba = sectors_to_logical(sdp, bio->bi_iter.bi_sector); + u32 nr_blocks = sectors_to_logical(sdp, bio_sectors(bio)); + + put_unaligned_be64(lba, &buf[8 + 16 * i]); + put_unaligned_be32(nr_blocks, &buf[8 + 16 * i + 8]); + i++; + } cmd->allowed = sdkp->max_retries; cmd->transfersize = data_len; @@ -2862,7 +2870,7 @@ static void sd_read_block_limits(struct scsi_disk *sdkp) sdkp->opt_xfer_blocks = get_unaligned_be32(&vpd->data[12]); if (vpd->len >= 64) { - unsigned int lba_count, desc_count; + unsigned int lba_count; sdkp->max_ws_blocks = (u32)get_unaligned_be64(&vpd->data[36]); @@ -2870,9 +2878,9 @@ static void sd_read_block_limits(struct scsi_disk *sdkp) goto out; lba_count = get_unaligned_be32(&vpd->data[20]); - desc_count = get_unaligned_be32(&vpd->data[24]); + sdkp->max_block_desc_count = get_unaligned_be32(&vpd->data[24]); - if (lba_count && desc_count) + if (lba_count && sdkp->max_block_desc_count) sdkp->max_unmap_blocks = lba_count; sdkp->unmap_granularity = get_unaligned_be32(&vpd->data[28]); diff --git a/drivers/scsi/sd.h b/drivers/scsi/sd.h index 5eea762f84d1..bda9db5e2322 100644 --- a/drivers/scsi/sd.h +++ b/drivers/scsi/sd.h @@ -119,6 +119,7 @@ struct scsi_disk { u32 opt_xfer_blocks; u32 max_ws_blocks; u32 max_unmap_blocks; + u32 max_block_desc_count; u32 unmap_granularity; u32 unmap_alignment; u32 index;