mirror of
https://github.com/torvalds/linux.git
synced 2024-11-10 06:01:57 +00:00
ata: scsi: rename flag ATA_QCFLAG_FAILED to ATA_QCFLAG_EH
The name ATA_QCFLAG_FAILED is misleading since it does not mean that a QC completed in error, or that it didn't complete at all. It means that libata decided to schedule EH for the QC, so the QC is now owned by the libata error handler (EH). The normal execution path is responsible for not accessing a QC owned by EH. libata core enforces the rule by returning NULL from ata_qc_from_tag() for QCs owned by EH. It is quite easy to mistake that a QC marked with ATA_QCFLAG_FAILED was an error. However, a QC that was actually an error is instead indicated by having qc->err_mask set. E.g. when we have a NCQ error, we abort all QCs, which currently will mark all QCs as ATA_QCFLAG_FAILED. However, it will only be a single QC that is an error (i.e. has qc->err_mask set). Rename ATA_QCFLAG_FAILED to ATA_QCFLAG_EH to more clearly highlight that this flag simply means that a QC is now owned by EH. This new name will not mislead to think that the QC was an error (which is instead indicated by having qc->err_mask set). This also makes it more obvious that the EH code skips all QCs that do not have ATA_QCFLAG_EH set (rather than ATA_QCFLAG_FAILED), since the EH code should simply only care about QCs that are owned by EH itself. Signed-off-by: Niklas Cassel <niklas.cassel@wdc.com> Reviewed-by: John Garry <john.g.garry@oracle.com> Signed-off-by: Damien Le Moal <damien.lemoal@opensource.wdc.com>
This commit is contained in:
parent
b83ad9eec3
commit
876293121f
@ -263,7 +263,7 @@ static bool acard_ahci_qc_fill_rtf(struct ata_queued_cmd *qc)
|
||||
* Setup FIS.
|
||||
*/
|
||||
if (qc->tf.protocol == ATA_PROT_PIO && qc->dma_dir == DMA_FROM_DEVICE &&
|
||||
!(qc->flags & ATA_QCFLAG_FAILED)) {
|
||||
!(qc->flags & ATA_QCFLAG_EH)) {
|
||||
ata_tf_from_fis(rx_fis + RX_FIS_PIO_SETUP, &qc->result_tf);
|
||||
qc->result_tf.status = (rx_fis + RX_FIS_PIO_SETUP)[15];
|
||||
} else
|
||||
|
@ -2068,7 +2068,7 @@ static bool ahci_qc_fill_rtf(struct ata_queued_cmd *qc)
|
||||
* Setup FIS.
|
||||
*/
|
||||
if (qc->tf.protocol == ATA_PROT_PIO && qc->dma_dir == DMA_FROM_DEVICE &&
|
||||
!(qc->flags & ATA_QCFLAG_FAILED)) {
|
||||
!(qc->flags & ATA_QCFLAG_EH)) {
|
||||
ata_tf_from_fis(rx_fis + RX_FIS_PIO_SETUP, &qc->result_tf);
|
||||
qc->result_tf.status = (rx_fis + RX_FIS_PIO_SETUP)[15];
|
||||
|
||||
@ -2138,7 +2138,7 @@ static void ahci_post_internal_cmd(struct ata_queued_cmd *qc)
|
||||
struct ata_port *ap = qc->ap;
|
||||
|
||||
/* make DMA engine forget about the failed command */
|
||||
if (qc->flags & ATA_QCFLAG_FAILED)
|
||||
if (qc->flags & ATA_QCFLAG_EH)
|
||||
ahci_kick_engine(ap);
|
||||
}
|
||||
|
||||
|
@ -1590,7 +1590,7 @@ static unsigned ata_exec_internal_sg(struct ata_device *dev,
|
||||
ap->ops->post_internal_cmd(qc);
|
||||
|
||||
/* perform minimal error analysis */
|
||||
if (qc->flags & ATA_QCFLAG_FAILED) {
|
||||
if (qc->flags & ATA_QCFLAG_EH) {
|
||||
if (qc->result_tf.status & (ATA_ERR | ATA_DF))
|
||||
qc->err_mask |= AC_ERR_DEV;
|
||||
|
||||
@ -4683,10 +4683,10 @@ void ata_qc_complete(struct ata_queued_cmd *qc)
|
||||
/* XXX: New EH and old EH use different mechanisms to
|
||||
* synchronize EH with regular execution path.
|
||||
*
|
||||
* In new EH, a failed qc is marked with ATA_QCFLAG_FAILED.
|
||||
* In new EH, a qc owned by EH is marked with ATA_QCFLAG_EH.
|
||||
* Normal execution path is responsible for not accessing a
|
||||
* failed qc. libata core enforces the rule by returning NULL
|
||||
* from ata_qc_from_tag() for failed qcs.
|
||||
* qc owned by EH. libata core enforces the rule by returning NULL
|
||||
* from ata_qc_from_tag() for qcs owned by EH.
|
||||
*
|
||||
* Old EH depends on ata_qc_complete() nullifying completion
|
||||
* requests if ATA_QCFLAG_EH_SCHEDULED is set. Old EH does
|
||||
@ -4698,7 +4698,7 @@ void ata_qc_complete(struct ata_queued_cmd *qc)
|
||||
struct ata_eh_info *ehi = &dev->link->eh_info;
|
||||
|
||||
if (unlikely(qc->err_mask))
|
||||
qc->flags |= ATA_QCFLAG_FAILED;
|
||||
qc->flags |= ATA_QCFLAG_EH;
|
||||
|
||||
/*
|
||||
* Finish internal commands without any further processing
|
||||
@ -4715,7 +4715,7 @@ void ata_qc_complete(struct ata_queued_cmd *qc)
|
||||
* Non-internal qc has failed. Fill the result TF and
|
||||
* summon EH.
|
||||
*/
|
||||
if (unlikely(qc->flags & ATA_QCFLAG_FAILED)) {
|
||||
if (unlikely(qc->flags & ATA_QCFLAG_EH)) {
|
||||
fill_result_tf(qc);
|
||||
trace_ata_qc_complete_failed(qc);
|
||||
ata_qc_schedule_eh(qc);
|
||||
|
@ -581,7 +581,7 @@ void ata_scsi_cmd_error_handler(struct Scsi_Host *host, struct ata_port *ap,
|
||||
* normal completion, error completion, and SCSI timeout.
|
||||
* Both completions can race against SCSI timeout. When normal
|
||||
* completion wins, the qc never reaches EH. When error
|
||||
* completion wins, the qc has ATA_QCFLAG_FAILED set.
|
||||
* completion wins, the qc has ATA_QCFLAG_EH set.
|
||||
*
|
||||
* When SCSI timeout wins, things are a bit more complex.
|
||||
* Normal or error completion can occur after the timeout but
|
||||
@ -615,10 +615,10 @@ void ata_scsi_cmd_error_handler(struct Scsi_Host *host, struct ata_port *ap,
|
||||
|
||||
if (i < ATA_MAX_QUEUE) {
|
||||
/* the scmd has an associated qc */
|
||||
if (!(qc->flags & ATA_QCFLAG_FAILED)) {
|
||||
if (!(qc->flags & ATA_QCFLAG_EH)) {
|
||||
/* which hasn't failed yet, timeout */
|
||||
qc->err_mask |= AC_ERR_TIMEOUT;
|
||||
qc->flags |= ATA_QCFLAG_FAILED;
|
||||
qc->flags |= ATA_QCFLAG_EH;
|
||||
nr_timedout++;
|
||||
}
|
||||
} else {
|
||||
@ -636,7 +636,7 @@ void ata_scsi_cmd_error_handler(struct Scsi_Host *host, struct ata_port *ap,
|
||||
* this point but the state of the controller is
|
||||
* unknown. Freeze the port to make sure the IRQ
|
||||
* handler doesn't diddle with those qcs. This must
|
||||
* be done atomically w.r.t. setting QCFLAG_FAILED.
|
||||
* be done atomically w.r.t. setting ATA_QCFLAG_EH.
|
||||
*/
|
||||
if (nr_timedout)
|
||||
__ata_port_freeze(ap);
|
||||
@ -914,12 +914,12 @@ void ata_qc_schedule_eh(struct ata_queued_cmd *qc)
|
||||
|
||||
WARN_ON(!ap->ops->error_handler);
|
||||
|
||||
qc->flags |= ATA_QCFLAG_FAILED;
|
||||
qc->flags |= ATA_QCFLAG_EH;
|
||||
ata_eh_set_pending(ap, 1);
|
||||
|
||||
/* The following will fail if timeout has already expired.
|
||||
* ata_scsi_error() takes care of such scmds on EH entry.
|
||||
* Note that ATA_QCFLAG_FAILED is unconditionally set after
|
||||
* Note that ATA_QCFLAG_EH is unconditionally set after
|
||||
* this function completes.
|
||||
*/
|
||||
blk_abort_request(scsi_cmd_to_rq(qc->scsicmd));
|
||||
@ -997,7 +997,7 @@ static int ata_do_link_abort(struct ata_port *ap, struct ata_link *link)
|
||||
/* include internal tag in iteration */
|
||||
ata_qc_for_each_with_internal(ap, qc, tag) {
|
||||
if (qc && (!link || qc->dev->link == link)) {
|
||||
qc->flags |= ATA_QCFLAG_FAILED;
|
||||
qc->flags |= ATA_QCFLAG_EH;
|
||||
ata_qc_complete(qc);
|
||||
nr_aborted++;
|
||||
}
|
||||
@ -1957,7 +1957,7 @@ static void ata_eh_link_autopsy(struct ata_link *link)
|
||||
all_err_mask |= ehc->i.err_mask;
|
||||
|
||||
ata_qc_for_each_raw(ap, qc, tag) {
|
||||
if (!(qc->flags & ATA_QCFLAG_FAILED) ||
|
||||
if (!(qc->flags & ATA_QCFLAG_EH) ||
|
||||
qc->flags & ATA_QCFLAG_RETRY ||
|
||||
ata_dev_phys_link(qc->dev) != link)
|
||||
continue;
|
||||
@ -2235,7 +2235,7 @@ static void ata_eh_link_report(struct ata_link *link)
|
||||
desc = ehc->i.desc;
|
||||
|
||||
ata_qc_for_each_raw(ap, qc, tag) {
|
||||
if (!(qc->flags & ATA_QCFLAG_FAILED) ||
|
||||
if (!(qc->flags & ATA_QCFLAG_EH) ||
|
||||
ata_dev_phys_link(qc->dev) != link ||
|
||||
((qc->flags & ATA_QCFLAG_QUIET) &&
|
||||
qc->err_mask == AC_ERR_DEV))
|
||||
@ -2301,7 +2301,7 @@ static void ata_eh_link_report(struct ata_link *link)
|
||||
char data_buf[20] = "";
|
||||
char cdb_buf[70] = "";
|
||||
|
||||
if (!(qc->flags & ATA_QCFLAG_FAILED) ||
|
||||
if (!(qc->flags & ATA_QCFLAG_EH) ||
|
||||
ata_dev_phys_link(qc->dev) != link || !qc->err_mask)
|
||||
continue;
|
||||
|
||||
@ -3805,7 +3805,7 @@ void ata_eh_finish(struct ata_port *ap)
|
||||
|
||||
/* retry or finish qcs */
|
||||
ata_qc_for_each_raw(ap, qc, tag) {
|
||||
if (!(qc->flags & ATA_QCFLAG_FAILED))
|
||||
if (!(qc->flags & ATA_QCFLAG_EH))
|
||||
continue;
|
||||
|
||||
if (qc->err_mask) {
|
||||
|
@ -1429,7 +1429,7 @@ void ata_eh_analyze_ncq_error(struct ata_link *link)
|
||||
|
||||
/* has LLDD analyzed already? */
|
||||
ata_qc_for_each_raw(ap, qc, tag) {
|
||||
if (!(qc->flags & ATA_QCFLAG_FAILED))
|
||||
if (!(qc->flags & ATA_QCFLAG_EH))
|
||||
continue;
|
||||
|
||||
if (qc->err_mask)
|
||||
@ -1477,7 +1477,7 @@ void ata_eh_analyze_ncq_error(struct ata_link *link)
|
||||
}
|
||||
|
||||
ata_qc_for_each_raw(ap, qc, tag) {
|
||||
if (!(qc->flags & ATA_QCFLAG_FAILED) ||
|
||||
if (!(qc->flags & ATA_QCFLAG_EH) ||
|
||||
ata_dev_phys_link(qc->dev) != link)
|
||||
continue;
|
||||
|
||||
|
@ -2073,7 +2073,7 @@ void ata_sff_error_handler(struct ata_port *ap)
|
||||
unsigned long flags;
|
||||
|
||||
qc = __ata_qc_from_tag(ap, ap->link.active_tag);
|
||||
if (qc && !(qc->flags & ATA_QCFLAG_FAILED))
|
||||
if (qc && !(qc->flags & ATA_QCFLAG_EH))
|
||||
qc = NULL;
|
||||
|
||||
spin_lock_irqsave(ap->lock, flags);
|
||||
@ -2796,7 +2796,7 @@ void ata_bmdma_error_handler(struct ata_port *ap)
|
||||
bool thaw = false;
|
||||
|
||||
qc = __ata_qc_from_tag(ap, ap->link.active_tag);
|
||||
if (qc && !(qc->flags & ATA_QCFLAG_FAILED))
|
||||
if (qc && !(qc->flags & ATA_QCFLAG_EH))
|
||||
qc = NULL;
|
||||
|
||||
/* reset PIO HSM and stop DMA engine */
|
||||
|
@ -142,7 +142,7 @@ libata_trace_parse_qc_flags(struct trace_seq *p, unsigned int qc_flags)
|
||||
trace_seq_printf(p, "QUIET ");
|
||||
if (qc_flags & ATA_QCFLAG_RETRY)
|
||||
trace_seq_printf(p, "RETRY ");
|
||||
if (qc_flags & ATA_QCFLAG_FAILED)
|
||||
if (qc_flags & ATA_QCFLAG_EH)
|
||||
trace_seq_printf(p, "FAILED ");
|
||||
if (qc_flags & ATA_QCFLAG_SENSE_VALID)
|
||||
trace_seq_printf(p, "SENSE_VALID ");
|
||||
|
@ -1042,7 +1042,7 @@ static void sata_fsl_error_handler(struct ata_port *ap)
|
||||
|
||||
static void sata_fsl_post_internal_cmd(struct ata_queued_cmd *qc)
|
||||
{
|
||||
if (qc->flags & ATA_QCFLAG_FAILED)
|
||||
if (qc->flags & ATA_QCFLAG_EH)
|
||||
qc->err_mask |= AC_ERR_OTHER;
|
||||
|
||||
if (qc->err_mask) {
|
||||
|
@ -672,7 +672,7 @@ static void inic_error_handler(struct ata_port *ap)
|
||||
static void inic_post_internal_cmd(struct ata_queued_cmd *qc)
|
||||
{
|
||||
/* make DMA engine forget about the failed command */
|
||||
if (qc->flags & ATA_QCFLAG_FAILED)
|
||||
if (qc->flags & ATA_QCFLAG_EH)
|
||||
inic_reset_port(inic_port_base(qc->ap));
|
||||
}
|
||||
|
||||
|
@ -828,7 +828,7 @@ static void pdc_post_internal_cmd(struct ata_queued_cmd *qc)
|
||||
struct ata_port *ap = qc->ap;
|
||||
|
||||
/* make DMA engine forget about the failed command */
|
||||
if (qc->flags & ATA_QCFLAG_FAILED)
|
||||
if (qc->flags & ATA_QCFLAG_EH)
|
||||
pdc_reset_port(ap);
|
||||
}
|
||||
|
||||
|
@ -1185,7 +1185,7 @@ static void sil24_post_internal_cmd(struct ata_queued_cmd *qc)
|
||||
struct ata_port *ap = qc->ap;
|
||||
|
||||
/* make DMA engine forget about the failed command */
|
||||
if ((qc->flags & ATA_QCFLAG_FAILED) && sil24_init_port(ap))
|
||||
if ((qc->flags & ATA_QCFLAG_EH) && sil24_init_port(ap))
|
||||
ata_eh_freeze_port(ap);
|
||||
}
|
||||
|
||||
|
@ -866,7 +866,7 @@ static void pdc_post_internal_cmd(struct ata_queued_cmd *qc)
|
||||
struct ata_port *ap = qc->ap;
|
||||
|
||||
/* make DMA engine forget about the failed command */
|
||||
if (qc->flags & ATA_QCFLAG_FAILED)
|
||||
if (qc->flags & ATA_QCFLAG_EH)
|
||||
pdc_reset_port(ap);
|
||||
}
|
||||
|
||||
|
@ -5370,9 +5370,9 @@ static int __ipr_eh_dev_reset(struct scsi_cmnd *scsi_cmd)
|
||||
continue;
|
||||
|
||||
ipr_cmd->done = ipr_sata_eh_done;
|
||||
if (!(ipr_cmd->qc->flags & ATA_QCFLAG_FAILED)) {
|
||||
if (!(ipr_cmd->qc->flags & ATA_QCFLAG_EH)) {
|
||||
ipr_cmd->qc->err_mask |= AC_ERR_TIMEOUT;
|
||||
ipr_cmd->qc->flags |= ATA_QCFLAG_FAILED;
|
||||
ipr_cmd->qc->flags |= ATA_QCFLAG_EH;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
@ -125,7 +125,7 @@ static void sas_ata_task_done(struct sas_task *task)
|
||||
} else {
|
||||
link->eh_info.err_mask |= ac_err_mask(dev->sata_dev.fis[2]);
|
||||
if (unlikely(link->eh_info.err_mask))
|
||||
qc->flags |= ATA_QCFLAG_FAILED;
|
||||
qc->flags |= ATA_QCFLAG_EH;
|
||||
}
|
||||
} else {
|
||||
ac = sas_to_ata_err(stat);
|
||||
@ -136,7 +136,7 @@ static void sas_ata_task_done(struct sas_task *task)
|
||||
qc->err_mask = ac;
|
||||
} else {
|
||||
link->eh_info.err_mask |= AC_ERR_DEV;
|
||||
qc->flags |= ATA_QCFLAG_FAILED;
|
||||
qc->flags |= ATA_QCFLAG_EH;
|
||||
}
|
||||
|
||||
dev->sata_dev.fis[2] = ATA_ERR | ATA_DRDY; /* tf status */
|
||||
@ -476,7 +476,7 @@ static void sas_ata_internal_abort(struct sas_task *task)
|
||||
|
||||
static void sas_ata_post_internal(struct ata_queued_cmd *qc)
|
||||
{
|
||||
if (qc->flags & ATA_QCFLAG_FAILED)
|
||||
if (qc->flags & ATA_QCFLAG_EH)
|
||||
qc->err_mask |= AC_ERR_OTHER;
|
||||
|
||||
if (qc->err_mask) {
|
||||
@ -631,7 +631,7 @@ void sas_ata_task_abort(struct sas_task *task)
|
||||
|
||||
/* Internal command, fake a timeout and complete. */
|
||||
qc->flags &= ~ATA_QCFLAG_ACTIVE;
|
||||
qc->flags |= ATA_QCFLAG_FAILED;
|
||||
qc->flags |= ATA_QCFLAG_EH;
|
||||
qc->err_mask |= AC_ERR_TIMEOUT;
|
||||
waiting = qc->private_data;
|
||||
complete(waiting);
|
||||
|
@ -206,7 +206,7 @@ enum {
|
||||
ATA_QCFLAG_QUIET = (1 << 6), /* don't report device error */
|
||||
ATA_QCFLAG_RETRY = (1 << 7), /* retry after failure */
|
||||
|
||||
ATA_QCFLAG_FAILED = (1 << 16), /* cmd failed and is owned by EH */
|
||||
ATA_QCFLAG_EH = (1 << 16), /* cmd aborted and owned by EH */
|
||||
ATA_QCFLAG_SENSE_VALID = (1 << 17), /* sense data valid */
|
||||
ATA_QCFLAG_EH_SCHEDULED = (1 << 18), /* EH scheduled (obsolete) */
|
||||
|
||||
@ -1756,7 +1756,7 @@ static inline struct ata_queued_cmd *ata_qc_from_tag(struct ata_port *ap,
|
||||
return qc;
|
||||
|
||||
if ((qc->flags & (ATA_QCFLAG_ACTIVE |
|
||||
ATA_QCFLAG_FAILED)) == ATA_QCFLAG_ACTIVE)
|
||||
ATA_QCFLAG_EH)) == ATA_QCFLAG_ACTIVE)
|
||||
return qc;
|
||||
|
||||
return NULL;
|
||||
|
Loading…
Reference in New Issue
Block a user