target/rd: always chain S/G list

The rd sg lists are never passed to hardware, so use S/G chaining
unonditionally.

Signed-off-by: Christoph Hellwig <hch@lst.de>
Signed-off-by: Jens Axboe <axboe@fb.com>
This commit is contained in:
Christoph Hellwig 2015-08-07 18:15:12 +02:00 committed by Jens Axboe
parent 10c95ed9aa
commit 02c4de53ad

View File

@ -138,16 +138,12 @@ static int rd_allocate_sgl_table(struct rd_dev *rd_dev, struct rd_dev_sg_table *
sg_per_table = (total_sg_needed > max_sg_per_table) ? sg_per_table = (total_sg_needed > max_sg_per_table) ?
max_sg_per_table : total_sg_needed; max_sg_per_table : total_sg_needed;
#ifdef CONFIG_ARCH_HAS_SG_CHAIN
/* /*
* Reserve extra element for chain entry * Reserve extra element for chain entry
*/ */
if (sg_per_table < total_sg_needed) if (sg_per_table < total_sg_needed)
chain_entry = 1; chain_entry = 1;
#endif /* CONFIG_ARCH_HAS_SG_CHAIN */
sg = kcalloc(sg_per_table + chain_entry, sizeof(*sg), sg = kcalloc(sg_per_table + chain_entry, sizeof(*sg),
GFP_KERNEL); GFP_KERNEL);
if (!sg) { if (!sg) {
@ -158,15 +154,11 @@ static int rd_allocate_sgl_table(struct rd_dev *rd_dev, struct rd_dev_sg_table *
sg_init_table(sg, sg_per_table + chain_entry); sg_init_table(sg, sg_per_table + chain_entry);
#ifdef CONFIG_ARCH_HAS_SG_CHAIN
if (i > 0) { if (i > 0) {
sg_chain(sg_table[i - 1].sg_table, sg_chain(sg_table[i - 1].sg_table,
max_sg_per_table + 1, sg); max_sg_per_table + 1, sg);
} }
#endif /* CONFIG_ARCH_HAS_SG_CHAIN */
sg_table[i].sg_table = sg; sg_table[i].sg_table = sg;
sg_table[i].rd_sg_count = sg_per_table; sg_table[i].rd_sg_count = sg_per_table;
sg_table[i].page_start_offset = page_offset; sg_table[i].page_start_offset = page_offset;
@ -429,42 +421,6 @@ static sense_reason_t rd_do_prot_rw(struct se_cmd *cmd, bool is_read)
prot_sg = &prot_table->sg_table[prot_page - prot_sg = &prot_table->sg_table[prot_page -
prot_table->page_start_offset]; prot_table->page_start_offset];
#ifndef CONFIG_ARCH_HAS_SG_CHAIN
prot_npages = DIV_ROUND_UP(prot_offset + sectors * se_dev->prot_length,
PAGE_SIZE);
/*
* Allocate temporaly contiguous scatterlist entries if prot pages
* straddles multiple scatterlist tables.
*/
if (prot_table->page_end_offset < prot_page + prot_npages - 1) {
int i;
prot_sg = kcalloc(prot_npages, sizeof(*prot_sg), GFP_KERNEL);
if (!prot_sg)
return TCM_LOGICAL_UNIT_COMMUNICATION_FAILURE;
need_to_release = true;
sg_init_table(prot_sg, prot_npages);
for (i = 0; i < prot_npages; i++) {
if (prot_page + i > prot_table->page_end_offset) {
prot_table = rd_get_prot_table(dev,
prot_page + i);
if (!prot_table) {
kfree(prot_sg);
return rc;
}
sg_unmark_end(&prot_sg[i - 1]);
}
prot_sg[i] = prot_table->sg_table[prot_page + i -
prot_table->page_start_offset];
}
}
#endif /* !CONFIG_ARCH_HAS_SG_CHAIN */
if (is_read) if (is_read)
rc = sbc_dif_verify(cmd, cmd->t_task_lba, sectors, 0, rc = sbc_dif_verify(cmd, cmd->t_task_lba, sectors, 0,
prot_sg, prot_offset); prot_sg, prot_offset);