M7350/kernel/drivers/scsi/sd_dif.c

206 lines
5.3 KiB
C
Raw Normal View History

2024-09-09 08:52:07 +00:00
/*
* sd_dif.c - SCSI Data Integrity Field
*
* Copyright (C) 2007, 2008 Oracle Corporation
* Written by: Martin K. Petersen <martin.petersen@oracle.com>
*
* This program is free software; you can redistribute it and/or
* modify it under the terms of the GNU General Public License version
* 2 as published by the Free Software Foundation.
*
* This program is distributed in the hope that it will be useful, but
* WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* General Public License for more details.
*
* You should have received a copy of the GNU General Public License
* along with this program; see the file COPYING. If not, write to
* the Free Software Foundation, 675 Mass Ave, Cambridge, MA 02139,
* USA.
*
*/
#include <linux/blkdev.h>
2024-09-09 08:57:42 +00:00
#include <linux/t10-pi.h>
2024-09-09 08:52:07 +00:00
#include <scsi/scsi.h>
#include <scsi/scsi_cmnd.h>
#include <scsi/scsi_dbg.h>
#include <scsi/scsi_device.h>
#include <scsi/scsi_driver.h>
#include <scsi/scsi_eh.h>
#include <scsi/scsi_host.h>
#include <scsi/scsi_ioctl.h>
#include <scsi/scsicam.h>
#include "sd.h"
/*
* Configure exchange of protection information between OS and HBA.
*/
void sd_dif_config_host(struct scsi_disk *sdkp)
{
struct scsi_device *sdp = sdkp->device;
struct gendisk *disk = sdkp->disk;
u8 type = sdkp->protection_type;
int dif, dix;
dif = scsi_host_dif_capable(sdp->host, type);
dix = scsi_host_dix_capable(sdp->host, type);
if (!dix && scsi_host_dix_capable(sdp->host, 0)) {
dif = 0; dix = 1;
}
if (!dix)
return;
/* Enable DMA of protection information */
2024-09-09 08:57:42 +00:00
if (scsi_host_get_guard(sdkp->device->host) & SHOST_DIX_GUARD_IP) {
2024-09-09 08:52:07 +00:00
if (type == SD_DIF_TYPE3_PROTECTION)
2024-09-09 08:57:42 +00:00
blk_integrity_register(disk, &t10_pi_type3_ip);
2024-09-09 08:52:07 +00:00
else
2024-09-09 08:57:42 +00:00
blk_integrity_register(disk, &t10_pi_type1_ip);
disk->integrity->flags |= BLK_INTEGRITY_IP_CHECKSUM;
} else
2024-09-09 08:52:07 +00:00
if (type == SD_DIF_TYPE3_PROTECTION)
2024-09-09 08:57:42 +00:00
blk_integrity_register(disk, &t10_pi_type3_crc);
2024-09-09 08:52:07 +00:00
else
2024-09-09 08:57:42 +00:00
blk_integrity_register(disk, &t10_pi_type1_crc);
2024-09-09 08:52:07 +00:00
sd_printk(KERN_NOTICE, sdkp,
"Enabling DIX %s protection\n", disk->integrity->name);
/* Signal to block layer that we support sector tagging */
2024-09-09 08:57:42 +00:00
if (dif && type) {
disk->integrity->flags |= BLK_INTEGRITY_DEVICE_CAPABLE;
if (!sdkp->ATO)
return;
2024-09-09 08:52:07 +00:00
if (type == SD_DIF_TYPE3_PROTECTION)
disk->integrity->tag_size = sizeof(u16) + sizeof(u32);
else
disk->integrity->tag_size = sizeof(u16);
sd_printk(KERN_NOTICE, sdkp, "DIF application tag size %u\n",
disk->integrity->tag_size);
}
}
/*
* The virtual start sector is the one that was originally submitted
* by the block layer. Due to partitioning, MD/DM cloning, etc. the
* actual physical start sector is likely to be different. Remap
* protection information to match the physical LBA.
*
* From a protocol perspective there's a slight difference between
* Type 1 and 2. The latter uses 32-byte CDBs exclusively, and the
* reference tag is seeded in the CDB. This gives us the potential to
* avoid virt->phys remapping during write. However, at read time we
* don't know whether the virt sector is the same as when we wrote it
* (we could be reading from real disk as opposed to MD/DM device. So
* we always remap Type 2 making it identical to Type 1.
*
* Type 3 does not have a reference tag so no remapping is required.
*/
2024-09-09 08:57:42 +00:00
void sd_dif_prepare(struct scsi_cmnd *scmd)
2024-09-09 08:52:07 +00:00
{
2024-09-09 08:57:42 +00:00
const int tuple_sz = sizeof(struct t10_pi_tuple);
2024-09-09 08:52:07 +00:00
struct bio *bio;
struct scsi_disk *sdkp;
2024-09-09 08:57:42 +00:00
struct t10_pi_tuple *pi;
2024-09-09 08:52:07 +00:00
u32 phys, virt;
2024-09-09 08:57:42 +00:00
sdkp = scsi_disk(scmd->request->rq_disk);
2024-09-09 08:52:07 +00:00
if (sdkp->protection_type == SD_DIF_TYPE3_PROTECTION)
2024-09-09 08:57:42 +00:00
return;
2024-09-09 08:52:07 +00:00
2024-09-09 08:57:42 +00:00
phys = scsi_prot_ref_tag(scmd);
2024-09-09 08:52:07 +00:00
2024-09-09 08:57:42 +00:00
__rq_for_each_bio(bio, scmd->request) {
struct bio_integrity_payload *bip = bio_integrity(bio);
struct bio_vec iv;
struct bvec_iter iter;
unsigned int j;
2024-09-09 08:52:07 +00:00
/* Already remapped? */
2024-09-09 08:57:42 +00:00
if (bip->bip_flags & BIP_MAPPED_INTEGRITY)
2024-09-09 08:52:07 +00:00
break;
2024-09-09 08:57:42 +00:00
virt = bip_get_seed(bip) & 0xffffffff;
2024-09-09 08:52:07 +00:00
2024-09-09 08:57:42 +00:00
bip_for_each_vec(iv, bip, iter) {
pi = kmap_atomic(iv.bv_page) + iv.bv_offset;
2024-09-09 08:52:07 +00:00
2024-09-09 08:57:42 +00:00
for (j = 0; j < iv.bv_len; j += tuple_sz, pi++) {
2024-09-09 08:52:07 +00:00
2024-09-09 08:57:42 +00:00
if (be32_to_cpu(pi->ref_tag) == virt)
pi->ref_tag = cpu_to_be32(phys);
2024-09-09 08:52:07 +00:00
virt++;
phys++;
}
2024-09-09 08:57:42 +00:00
kunmap_atomic(pi);
2024-09-09 08:52:07 +00:00
}
2024-09-09 08:57:42 +00:00
bip->bip_flags |= BIP_MAPPED_INTEGRITY;
2024-09-09 08:52:07 +00:00
}
}
/*
* Remap physical sector values in the reference tag to the virtual
* values expected by the block layer.
*/
void sd_dif_complete(struct scsi_cmnd *scmd, unsigned int good_bytes)
{
2024-09-09 08:57:42 +00:00
const int tuple_sz = sizeof(struct t10_pi_tuple);
2024-09-09 08:52:07 +00:00
struct scsi_disk *sdkp;
struct bio *bio;
2024-09-09 08:57:42 +00:00
struct t10_pi_tuple *pi;
unsigned int j, intervals;
2024-09-09 08:52:07 +00:00
u32 phys, virt;
sdkp = scsi_disk(scmd->request->rq_disk);
if (sdkp->protection_type == SD_DIF_TYPE3_PROTECTION || good_bytes == 0)
return;
2024-09-09 08:57:42 +00:00
intervals = good_bytes / scsi_prot_interval(scmd);
phys = scsi_prot_ref_tag(scmd);
2024-09-09 08:52:07 +00:00
__rq_for_each_bio(bio, scmd->request) {
2024-09-09 08:57:42 +00:00
struct bio_integrity_payload *bip = bio_integrity(bio);
struct bio_vec iv;
struct bvec_iter iter;
2024-09-09 08:52:07 +00:00
2024-09-09 08:57:42 +00:00
virt = bip_get_seed(bip) & 0xffffffff;
2024-09-09 08:52:07 +00:00
2024-09-09 08:57:42 +00:00
bip_for_each_vec(iv, bip, iter) {
pi = kmap_atomic(iv.bv_page) + iv.bv_offset;
2024-09-09 08:52:07 +00:00
2024-09-09 08:57:42 +00:00
for (j = 0; j < iv.bv_len; j += tuple_sz, pi++) {
2024-09-09 08:52:07 +00:00
2024-09-09 08:57:42 +00:00
if (intervals == 0) {
kunmap_atomic(pi);
2024-09-09 08:52:07 +00:00
return;
}
2024-09-09 08:57:42 +00:00
if (be32_to_cpu(pi->ref_tag) == phys)
pi->ref_tag = cpu_to_be32(virt);
2024-09-09 08:52:07 +00:00
virt++;
phys++;
2024-09-09 08:57:42 +00:00
intervals--;
2024-09-09 08:52:07 +00:00
}
2024-09-09 08:57:42 +00:00
kunmap_atomic(pi);
2024-09-09 08:52:07 +00:00
}
}
}