Message ID | 20240122054308.23901-16-baolu.lu@linux.intel.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | iommu: Prepare to deliver page faults to user space | expand |
On Mon, Jan 22, 2024 at 01:43:07PM +0800, Lu Baolu wrote: > The iopf_group_response() should return void, as nothing can do anything > with the failure. This implies that ops->page_response() must also return > void; this is consistent with what the drivers do. The failure paths, > which are all integrity validations of the fault, should be WARN_ON'd, > not return codes. > > If the iommu core fails to enqueue the fault, it should respond the fault > directly by calling ops->page_response() instead of returning an error > number and relying on the iommu drivers to do so. Consolidate the error > fault handling code in the core. > > Co-developed-by: Jason Gunthorpe <jgg@nvidia.com> > Signed-off-by: Jason Gunthorpe <jgg@nvidia.com> > Signed-off-by: Lu Baolu <baolu.lu@linux.intel.com> > --- > include/linux/iommu.h | 14 +-- > drivers/iommu/intel/iommu.h | 4 +- > drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c | 50 +++----- > drivers/iommu/intel/svm.c | 18 +-- > drivers/iommu/io-pgfault.c | 132 +++++++++++--------- > 5 files changed, 99 insertions(+), 119 deletions(-) > > diff --git a/include/linux/iommu.h b/include/linux/iommu.h > index 48196efc9327..d7b6f4017254 100644 > --- a/include/linux/iommu.h > +++ b/include/linux/iommu.h > @@ -578,9 +578,8 @@ struct iommu_ops { > int (*dev_enable_feat)(struct device *dev, enum iommu_dev_features f); > int (*dev_disable_feat)(struct device *dev, enum iommu_dev_features f); > > - int (*page_response)(struct device *dev, > - struct iopf_fault *evt, > - struct iommu_page_response *msg); > + void (*page_response)(struct device *dev, struct iopf_fault *evt, > + struct iommu_page_response *msg); > > int (*def_domain_type)(struct device *dev); > void (*remove_dev_pasid)(struct device *dev, ioasid_t pasid); > @@ -1551,8 +1550,8 @@ void iopf_queue_free(struct iopf_queue *queue); > int iopf_queue_discard_partial(struct iopf_queue *queue); > void iopf_free_group(struct iopf_group *group); > int iommu_report_device_fault(struct device *dev, struct iopf_fault *evt); > -int iopf_group_response(struct iopf_group *group, > - enum iommu_page_response_code status); > +void iopf_group_response(struct iopf_group *group, > + enum iommu_page_response_code status); > #else > static inline int > iopf_queue_add_device(struct iopf_queue *queue, struct device *dev) > @@ -1594,10 +1593,9 @@ iommu_report_device_fault(struct device *dev, struct iopf_fault *evt) > return -ENODEV; > } > > -static inline int iopf_group_response(struct iopf_group *group, > - enum iommu_page_response_code status) > +static inline void iopf_group_response(struct iopf_group *group, > + enum iommu_page_response_code status) > { > - return -ENODEV; > } > #endif /* CONFIG_IOMMU_IOPF */ > #endif /* __LINUX_IOMMU_H */ > diff --git a/drivers/iommu/intel/iommu.h b/drivers/iommu/intel/iommu.h > index 696d95293a69..cf9a28c7fab8 100644 > --- a/drivers/iommu/intel/iommu.h > +++ b/drivers/iommu/intel/iommu.h > @@ -1079,8 +1079,8 @@ struct iommu_domain *intel_nested_domain_alloc(struct iommu_domain *parent, > void intel_svm_check(struct intel_iommu *iommu); > int intel_svm_enable_prq(struct intel_iommu *iommu); > int intel_svm_finish_prq(struct intel_iommu *iommu); > -int intel_svm_page_response(struct device *dev, struct iopf_fault *evt, > - struct iommu_page_response *msg); > +void intel_svm_page_response(struct device *dev, struct iopf_fault *evt, > + struct iommu_page_response *msg); > struct iommu_domain *intel_svm_domain_alloc(void); > void intel_svm_remove_dev_pasid(struct device *dev, ioasid_t pasid); > void intel_drain_pasid_prq(struct device *dev, u32 pasid); > diff --git a/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c b/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c > index 4e93e845458c..42eb59cb99f4 100644 > --- a/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c > +++ b/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c > @@ -920,31 +920,29 @@ static int arm_smmu_cmdq_batch_submit(struct arm_smmu_device *smmu, > return arm_smmu_cmdq_issue_cmdlist(smmu, cmds->cmds, cmds->num, true); > } > > -static int arm_smmu_page_response(struct device *dev, > - struct iopf_fault *unused, > - struct iommu_page_response *resp) > +static void arm_smmu_page_response(struct device *dev, struct iopf_fault *unused, > + struct iommu_page_response *resp) > { > struct arm_smmu_cmdq_ent cmd = {0}; > struct arm_smmu_master *master = dev_iommu_priv_get(dev); > int sid = master->streams[0].id; > > - if (master->stall_enabled) { > - cmd.opcode = CMDQ_OP_RESUME; > - cmd.resume.sid = sid; > - cmd.resume.stag = resp->grpid; > - switch (resp->code) { > - case IOMMU_PAGE_RESP_INVALID: > - case IOMMU_PAGE_RESP_FAILURE: > - cmd.resume.resp = CMDQ_RESUME_0_RESP_ABORT; > - break; > - case IOMMU_PAGE_RESP_SUCCESS: > - cmd.resume.resp = CMDQ_RESUME_0_RESP_RETRY; > - break; > - default: > - return -EINVAL; > - } > - } else { > - return -ENODEV; > + if (WARN_ON(!master->stall_enabled)) > + return; > + > + cmd.opcode = CMDQ_OP_RESUME; > + cmd.resume.sid = sid; > + cmd.resume.stag = resp->grpid; > + switch (resp->code) { > + case IOMMU_PAGE_RESP_INVALID: > + case IOMMU_PAGE_RESP_FAILURE: > + cmd.resume.resp = CMDQ_RESUME_0_RESP_ABORT; > + break; > + case IOMMU_PAGE_RESP_SUCCESS: > + cmd.resume.resp = CMDQ_RESUME_0_RESP_RETRY; > + break; > + default: > + break; > } > > arm_smmu_cmdq_issue_cmd(master->smmu, &cmd); > @@ -954,8 +952,6 @@ static int arm_smmu_page_response(struct device *dev, > * terminated... at some point in the future. PRI_RESP is fire and > * forget. > */ > - > - return 0; > } > > /* Context descriptor manipulation functions */ > @@ -1516,16 +1512,6 @@ static int arm_smmu_handle_evt(struct arm_smmu_device *smmu, u64 *evt) > } > > ret = iommu_report_device_fault(master->dev, &fault_evt); > - if (ret && flt->type == IOMMU_FAULT_PAGE_REQ) { > - /* Nobody cared, abort the access */ > - struct iommu_page_response resp = { > - .pasid = flt->prm.pasid, > - .grpid = flt->prm.grpid, > - .code = IOMMU_PAGE_RESP_FAILURE, > - }; > - arm_smmu_page_response(master->dev, &fault_evt, &resp); > - } > - > out_unlock: > mutex_unlock(&smmu->streams_mutex); > return ret; > diff --git a/drivers/iommu/intel/svm.c b/drivers/iommu/intel/svm.c > index e1cbcb9515f0..2f8716636dbb 100644 > --- a/drivers/iommu/intel/svm.c > +++ b/drivers/iommu/intel/svm.c > @@ -740,9 +740,8 @@ static irqreturn_t prq_event_thread(int irq, void *d) > return IRQ_RETVAL(handled); > } > > -int intel_svm_page_response(struct device *dev, > - struct iopf_fault *evt, > - struct iommu_page_response *msg) > +void intel_svm_page_response(struct device *dev, struct iopf_fault *evt, > + struct iommu_page_response *msg) > { > struct device_domain_info *info = dev_iommu_priv_get(dev); > struct intel_iommu *iommu = info->iommu; > @@ -751,7 +750,6 @@ int intel_svm_page_response(struct device *dev, > bool private_present; > bool pasid_present; > bool last_page; > - int ret = 0; > u16 sid; > > prm = &evt->fault.prm; > @@ -760,16 +758,6 @@ int intel_svm_page_response(struct device *dev, > private_present = prm->flags & IOMMU_FAULT_PAGE_REQUEST_PRIV_DATA; > last_page = prm->flags & IOMMU_FAULT_PAGE_REQUEST_LAST_PAGE; > > - if (!pasid_present) { > - ret = -EINVAL; > - goto out; > - } > - > - if (prm->pasid == 0 || prm->pasid >= PASID_MAX) { > - ret = -EINVAL; > - goto out; > - } > - > /* > * Per VT-d spec. v3.0 ch7.7, system software must respond > * with page group response if private data is present (PDP) > @@ -798,8 +786,6 @@ int intel_svm_page_response(struct device *dev, > > qi_submit_sync(iommu, &desc, 1, 0); > } > -out: > - return ret; > } > > static int intel_svm_set_dev_pasid(struct iommu_domain *domain, > diff --git a/drivers/iommu/io-pgfault.c b/drivers/iommu/io-pgfault.c > index c22e13df84c2..6e63e5a02884 100644 > --- a/drivers/iommu/io-pgfault.c > +++ b/drivers/iommu/io-pgfault.c > @@ -39,7 +39,7 @@ static void iopf_put_dev_fault_param(struct iommu_fault_param *fault_param) > kfree_rcu(fault_param, rcu); > } > > -void iopf_free_group(struct iopf_group *group) > +static void __iopf_free_group(struct iopf_group *group) > { > struct iopf_fault *iopf, *next; > > @@ -50,6 +50,11 @@ void iopf_free_group(struct iopf_group *group) > > /* Pair with iommu_report_device_fault(). */ > iopf_put_dev_fault_param(group->fault_param); > +} > + > +void iopf_free_group(struct iopf_group *group) > +{ > + __iopf_free_group(group); > kfree(group); > } > EXPORT_SYMBOL_GPL(iopf_free_group); > @@ -97,14 +102,49 @@ static int report_partial_fault(struct iommu_fault_param *fault_param, > return 0; > } > > +static struct iopf_group *iopf_group_alloc(struct iommu_fault_param *iopf_param, > + struct iopf_fault *evt, > + struct iopf_group *abort_group) > +{ > + struct iopf_fault *iopf, *next; > + struct iopf_group *group; > + > + group = kzalloc(sizeof(*group), GFP_KERNEL); > + if (!group) { > + /* > + * We always need to construct the group as we need it to abort > + * the request at the driver if it cfan't be handled. > + */ > + group = abort_group; > + } > + > + group->fault_param = iopf_param; > + group->last_fault.fault = evt->fault; > + INIT_LIST_HEAD(&group->faults); > + INIT_LIST_HEAD(&group->pending_node); > + list_add(&group->last_fault.list, &group->faults); > + > + /* See if we have partial faults for this group */ > + mutex_lock(&iopf_param->lock); > + list_for_each_entry_safe(iopf, next, &iopf_param->partial, list) { > + if (iopf->fault.prm.grpid == evt->fault.prm.grpid) > + /* Insert *before* the last fault */ > + list_move(&iopf->list, &group->faults); > + } > + list_add(&group->pending_node, &iopf_param->faults); > + mutex_unlock(&iopf_param->lock); > + > + return group; > +} > + > /** > * iommu_report_device_fault() - Report fault event to device driver > * @dev: the device > * @evt: fault event data > * > * Called by IOMMU drivers when a fault is detected, typically in a threaded IRQ > - * handler. When this function fails and the fault is recoverable, it is the > - * caller's responsibility to complete the fault. > + * handler. If this function fails then ops->page_response() was called to > + * complete evt if required. > * > * This module doesn't handle PCI PASID Stop Marker; IOMMU drivers must discard > * them before reporting faults. A PASID Stop Marker (LRW = 0b100) doesn't > @@ -143,22 +183,18 @@ int iommu_report_device_fault(struct device *dev, struct iopf_fault *evt) > { > struct iommu_fault *fault = &evt->fault; > struct iommu_fault_param *iopf_param; > - struct iopf_fault *iopf, *next; > - struct iommu_domain *domain; > + struct iopf_group abort_group = {}; > struct iopf_group *group; > int ret; > > - if (fault->type != IOMMU_FAULT_PAGE_REQ) > - return -EOPNOTSUPP; > - > iopf_param = iopf_get_dev_fault_param(dev); > - if (!iopf_param) > + if (WARN_ON(!iopf_param)) > return -ENODEV; > > if (!(fault->prm.flags & IOMMU_FAULT_PAGE_REQUEST_LAST_PAGE)) { > ret = report_partial_fault(iopf_param, fault); > iopf_put_dev_fault_param(iopf_param); > - > + /* A request that is not the last does not need to be ack'd */ > return ret; > } > > @@ -170,56 +206,33 @@ int iommu_report_device_fault(struct device *dev, struct iopf_fault *evt) > * will send a response to the hardware. We need to clean up before > * leaving, otherwise partial faults will be stuck. > */ > - domain = get_domain_for_iopf(dev, fault); > - if (!domain) { > + group = iopf_group_alloc(iopf_param, evt, &abort_group); > + if (group == &abort_group) { > + ret = -ENOMEM; > + goto err_abort; > + } > + > + group->domain = get_domain_for_iopf(dev, fault); > + if (!group->domain) { > ret = -EINVAL; > - goto cleanup_partial; > + goto err_abort; > } > > - group = kzalloc(sizeof(*group), GFP_KERNEL); > - if (!group) { > - ret = -ENOMEM; > - goto cleanup_partial; > - } > - > - group->fault_param = iopf_param; > - group->last_fault.fault = *fault; > - INIT_LIST_HEAD(&group->faults); > - INIT_LIST_HEAD(&group->pending_node); > - group->domain = domain; > - list_add(&group->last_fault.list, &group->faults); > - > - /* See if we have partial faults for this group */ > - mutex_lock(&iopf_param->lock); > - list_for_each_entry_safe(iopf, next, &iopf_param->partial, list) { > - if (iopf->fault.prm.grpid == fault->prm.grpid) > - /* Insert *before* the last fault */ > - list_move(&iopf->list, &group->faults); > - } > - list_add(&group->pending_node, &iopf_param->faults); > - mutex_unlock(&iopf_param->lock); > + /* > + * On success iopf_handler must call iopf_group_response() and > + * iopf_free_group() > + */ > + ret = group->domain->iopf_handler(group); > + if (ret) > + goto err_abort; > + return 0; > > - ret = domain->iopf_handler(group); > - if (ret) { > - mutex_lock(&iopf_param->lock); > - list_del_init(&group->pending_node); > - mutex_unlock(&iopf_param->lock); > +err_abort: > + iopf_group_response(group, IOMMU_PAGE_RESP_FAILURE); > + if (group == &abort_group) > + __iopf_free_group(group); > + else > iopf_free_group(group); > - } > - > - return ret; > - > -cleanup_partial: > - mutex_lock(&iopf_param->lock); > - list_for_each_entry_safe(iopf, next, &iopf_param->partial, list) { > - if (iopf->fault.prm.grpid == fault->prm.grpid) { > - list_del(&iopf->list); > - kfree(iopf); > - } > - } > - mutex_unlock(&iopf_param->lock); > - iopf_put_dev_fault_param(iopf_param); > - > return ret; > } > EXPORT_SYMBOL_GPL(iommu_report_device_fault); > @@ -262,8 +275,8 @@ EXPORT_SYMBOL_GPL(iopf_queue_flush_dev); > * > * Return 0 on success and <0 on error. > */ Should you adjust the docs as well? > -int iopf_group_response(struct iopf_group *group, > - enum iommu_page_response_code status) > +void iopf_group_response(struct iopf_group *group, > + enum iommu_page_response_code status) > { > struct iommu_fault_param *fault_param = group->fault_param; > struct iopf_fault *iopf = &group->last_fault; > @@ -274,7 +287,6 @@ int iopf_group_response(struct iopf_group *group, > .grpid = iopf->fault.prm.grpid, > .code = status, > }; > - int ret = -EINVAL; > > if ((iopf->fault.prm.flags & IOMMU_FAULT_PAGE_REQUEST_PASID_VALID) && > (iopf->fault.prm.flags & IOMMU_FAULT_PAGE_RESPONSE_NEEDS_PASID)) > @@ -283,12 +295,10 @@ int iopf_group_response(struct iopf_group *group, > /* Only send response if there is a fault report pending */ > mutex_lock(&fault_param->lock); > if (!list_empty(&group->pending_node)) { > - ret = ops->page_response(dev, &group->last_fault, &resp); > + ops->page_response(dev, &group->last_fault, &resp); > list_del_init(&group->pending_node); > } > mutex_unlock(&fault_param->lock); > - > - return ret; > } > EXPORT_SYMBOL_GPL(iopf_group_response); > > -- > 2.34.1 >
On 2024/1/26 0:27, Joel Granados wrote: > On Mon, Jan 22, 2024 at 01:43:07PM +0800, Lu Baolu wrote: >> The iopf_group_response() should return void, as nothing can do anything >> with the failure. This implies that ops->page_response() must also return >> void; this is consistent with what the drivers do. The failure paths, >> which are all integrity validations of the fault, should be WARN_ON'd, >> not return codes. >> >> If the iommu core fails to enqueue the fault, it should respond the fault >> directly by calling ops->page_response() instead of returning an error >> number and relying on the iommu drivers to do so. Consolidate the error >> fault handling code in the core. >> >> Co-developed-by: Jason Gunthorpe<jgg@nvidia.com> >> Signed-off-by: Jason Gunthorpe<jgg@nvidia.com> >> Signed-off-by: Lu Baolu<baolu.lu@linux.intel.com> >> --- >> include/linux/iommu.h | 14 +-- >> drivers/iommu/intel/iommu.h | 4 +- >> drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c | 50 +++----- >> drivers/iommu/intel/svm.c | 18 +-- >> drivers/iommu/io-pgfault.c | 132 +++++++++++--------- >> 5 files changed, 99 insertions(+), 119 deletions(-) >> >> diff --git a/include/linux/iommu.h b/include/linux/iommu.h >> index 48196efc9327..d7b6f4017254 100644 >> --- a/include/linux/iommu.h >> +++ b/include/linux/iommu.h >> @@ -578,9 +578,8 @@ struct iommu_ops { >> int (*dev_enable_feat)(struct device *dev, enum iommu_dev_features f); >> int (*dev_disable_feat)(struct device *dev, enum iommu_dev_features f); >> >> - int (*page_response)(struct device *dev, >> - struct iopf_fault *evt, >> - struct iommu_page_response *msg); >> + void (*page_response)(struct device *dev, struct iopf_fault *evt, >> + struct iommu_page_response *msg); >> >> int (*def_domain_type)(struct device *dev); >> void (*remove_dev_pasid)(struct device *dev, ioasid_t pasid); >> @@ -1551,8 +1550,8 @@ void iopf_queue_free(struct iopf_queue *queue); >> int iopf_queue_discard_partial(struct iopf_queue *queue); >> void iopf_free_group(struct iopf_group *group); >> int iommu_report_device_fault(struct device *dev, struct iopf_fault *evt); >> -int iopf_group_response(struct iopf_group *group, >> - enum iommu_page_response_code status); >> +void iopf_group_response(struct iopf_group *group, >> + enum iommu_page_response_code status); >> #else >> static inline int >> iopf_queue_add_device(struct iopf_queue *queue, struct device *dev) >> @@ -1594,10 +1593,9 @@ iommu_report_device_fault(struct device *dev, struct iopf_fault *evt) >> return -ENODEV; >> } >> >> -static inline int iopf_group_response(struct iopf_group *group, >> - enum iommu_page_response_code status) >> +static inline void iopf_group_response(struct iopf_group *group, >> + enum iommu_page_response_code status) >> { >> - return -ENODEV; >> } >> #endif /* CONFIG_IOMMU_IOPF */ >> #endif /* __LINUX_IOMMU_H */ >> diff --git a/drivers/iommu/intel/iommu.h b/drivers/iommu/intel/iommu.h >> index 696d95293a69..cf9a28c7fab8 100644 >> --- a/drivers/iommu/intel/iommu.h >> +++ b/drivers/iommu/intel/iommu.h >> @@ -1079,8 +1079,8 @@ struct iommu_domain *intel_nested_domain_alloc(struct iommu_domain *parent, >> void intel_svm_check(struct intel_iommu *iommu); >> int intel_svm_enable_prq(struct intel_iommu *iommu); >> int intel_svm_finish_prq(struct intel_iommu *iommu); >> -int intel_svm_page_response(struct device *dev, struct iopf_fault *evt, >> - struct iommu_page_response *msg); >> +void intel_svm_page_response(struct device *dev, struct iopf_fault *evt, >> + struct iommu_page_response *msg); >> struct iommu_domain *intel_svm_domain_alloc(void); >> void intel_svm_remove_dev_pasid(struct device *dev, ioasid_t pasid); >> void intel_drain_pasid_prq(struct device *dev, u32 pasid); >> diff --git a/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c b/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c >> index 4e93e845458c..42eb59cb99f4 100644 >> --- a/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c >> +++ b/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c >> @@ -920,31 +920,29 @@ static int arm_smmu_cmdq_batch_submit(struct arm_smmu_device *smmu, >> return arm_smmu_cmdq_issue_cmdlist(smmu, cmds->cmds, cmds->num, true); >> } >> >> -static int arm_smmu_page_response(struct device *dev, >> - struct iopf_fault *unused, >> - struct iommu_page_response *resp) >> +static void arm_smmu_page_response(struct device *dev, struct iopf_fault *unused, >> + struct iommu_page_response *resp) >> { >> struct arm_smmu_cmdq_ent cmd = {0}; >> struct arm_smmu_master *master = dev_iommu_priv_get(dev); >> int sid = master->streams[0].id; >> >> - if (master->stall_enabled) { >> - cmd.opcode = CMDQ_OP_RESUME; >> - cmd.resume.sid = sid; >> - cmd.resume.stag = resp->grpid; >> - switch (resp->code) { >> - case IOMMU_PAGE_RESP_INVALID: >> - case IOMMU_PAGE_RESP_FAILURE: >> - cmd.resume.resp = CMDQ_RESUME_0_RESP_ABORT; >> - break; >> - case IOMMU_PAGE_RESP_SUCCESS: >> - cmd.resume.resp = CMDQ_RESUME_0_RESP_RETRY; >> - break; >> - default: >> - return -EINVAL; >> - } >> - } else { >> - return -ENODEV; >> + if (WARN_ON(!master->stall_enabled)) >> + return; >> + >> + cmd.opcode = CMDQ_OP_RESUME; >> + cmd.resume.sid = sid; >> + cmd.resume.stag = resp->grpid; >> + switch (resp->code) { >> + case IOMMU_PAGE_RESP_INVALID: >> + case IOMMU_PAGE_RESP_FAILURE: >> + cmd.resume.resp = CMDQ_RESUME_0_RESP_ABORT; >> + break; >> + case IOMMU_PAGE_RESP_SUCCESS: >> + cmd.resume.resp = CMDQ_RESUME_0_RESP_RETRY; >> + break; >> + default: >> + break; >> } >> >> arm_smmu_cmdq_issue_cmd(master->smmu, &cmd); >> @@ -954,8 +952,6 @@ static int arm_smmu_page_response(struct device *dev, >> * terminated... at some point in the future. PRI_RESP is fire and >> * forget. >> */ >> - >> - return 0; >> } >> >> /* Context descriptor manipulation functions */ >> @@ -1516,16 +1512,6 @@ static int arm_smmu_handle_evt(struct arm_smmu_device *smmu, u64 *evt) >> } >> >> ret = iommu_report_device_fault(master->dev, &fault_evt); >> - if (ret && flt->type == IOMMU_FAULT_PAGE_REQ) { >> - /* Nobody cared, abort the access */ >> - struct iommu_page_response resp = { >> - .pasid = flt->prm.pasid, >> - .grpid = flt->prm.grpid, >> - .code = IOMMU_PAGE_RESP_FAILURE, >> - }; >> - arm_smmu_page_response(master->dev, &fault_evt, &resp); >> - } >> - >> out_unlock: >> mutex_unlock(&smmu->streams_mutex); >> return ret; >> diff --git a/drivers/iommu/intel/svm.c b/drivers/iommu/intel/svm.c >> index e1cbcb9515f0..2f8716636dbb 100644 >> --- a/drivers/iommu/intel/svm.c >> +++ b/drivers/iommu/intel/svm.c >> @@ -740,9 +740,8 @@ static irqreturn_t prq_event_thread(int irq, void *d) >> return IRQ_RETVAL(handled); >> } >> >> -int intel_svm_page_response(struct device *dev, >> - struct iopf_fault *evt, >> - struct iommu_page_response *msg) >> +void intel_svm_page_response(struct device *dev, struct iopf_fault *evt, >> + struct iommu_page_response *msg) >> { >> struct device_domain_info *info = dev_iommu_priv_get(dev); >> struct intel_iommu *iommu = info->iommu; >> @@ -751,7 +750,6 @@ int intel_svm_page_response(struct device *dev, >> bool private_present; >> bool pasid_present; >> bool last_page; >> - int ret = 0; >> u16 sid; >> >> prm = &evt->fault.prm; >> @@ -760,16 +758,6 @@ int intel_svm_page_response(struct device *dev, >> private_present = prm->flags & IOMMU_FAULT_PAGE_REQUEST_PRIV_DATA; >> last_page = prm->flags & IOMMU_FAULT_PAGE_REQUEST_LAST_PAGE; >> >> - if (!pasid_present) { >> - ret = -EINVAL; >> - goto out; >> - } >> - >> - if (prm->pasid == 0 || prm->pasid >= PASID_MAX) { >> - ret = -EINVAL; >> - goto out; >> - } >> - >> /* >> * Per VT-d spec. v3.0 ch7.7, system software must respond >> * with page group response if private data is present (PDP) >> @@ -798,8 +786,6 @@ int intel_svm_page_response(struct device *dev, >> >> qi_submit_sync(iommu, &desc, 1, 0); >> } >> -out: >> - return ret; >> } >> >> static int intel_svm_set_dev_pasid(struct iommu_domain *domain, >> diff --git a/drivers/iommu/io-pgfault.c b/drivers/iommu/io-pgfault.c >> index c22e13df84c2..6e63e5a02884 100644 >> --- a/drivers/iommu/io-pgfault.c >> +++ b/drivers/iommu/io-pgfault.c >> @@ -39,7 +39,7 @@ static void iopf_put_dev_fault_param(struct iommu_fault_param *fault_param) >> kfree_rcu(fault_param, rcu); >> } >> >> -void iopf_free_group(struct iopf_group *group) >> +static void __iopf_free_group(struct iopf_group *group) >> { >> struct iopf_fault *iopf, *next; >> >> @@ -50,6 +50,11 @@ void iopf_free_group(struct iopf_group *group) >> >> /* Pair with iommu_report_device_fault(). */ >> iopf_put_dev_fault_param(group->fault_param); >> +} >> + >> +void iopf_free_group(struct iopf_group *group) >> +{ >> + __iopf_free_group(group); >> kfree(group); >> } >> EXPORT_SYMBOL_GPL(iopf_free_group); >> @@ -97,14 +102,49 @@ static int report_partial_fault(struct iommu_fault_param *fault_param, >> return 0; >> } >> >> +static struct iopf_group *iopf_group_alloc(struct iommu_fault_param *iopf_param, >> + struct iopf_fault *evt, >> + struct iopf_group *abort_group) >> +{ >> + struct iopf_fault *iopf, *next; >> + struct iopf_group *group; >> + >> + group = kzalloc(sizeof(*group), GFP_KERNEL); >> + if (!group) { >> + /* >> + * We always need to construct the group as we need it to abort >> + * the request at the driver if it cfan't be handled. >> + */ >> + group = abort_group; >> + } >> + >> + group->fault_param = iopf_param; >> + group->last_fault.fault = evt->fault; >> + INIT_LIST_HEAD(&group->faults); >> + INIT_LIST_HEAD(&group->pending_node); >> + list_add(&group->last_fault.list, &group->faults); >> + >> + /* See if we have partial faults for this group */ >> + mutex_lock(&iopf_param->lock); >> + list_for_each_entry_safe(iopf, next, &iopf_param->partial, list) { >> + if (iopf->fault.prm.grpid == evt->fault.prm.grpid) >> + /* Insert*before* the last fault */ >> + list_move(&iopf->list, &group->faults); >> + } >> + list_add(&group->pending_node, &iopf_param->faults); >> + mutex_unlock(&iopf_param->lock); >> + >> + return group; >> +} >> + >> /** >> * iommu_report_device_fault() - Report fault event to device driver >> * @dev: the device >> * @evt: fault event data >> * >> * Called by IOMMU drivers when a fault is detected, typically in a threaded IRQ >> - * handler. When this function fails and the fault is recoverable, it is the >> - * caller's responsibility to complete the fault. >> + * handler. If this function fails then ops->page_response() was called to >> + * complete evt if required. >> * >> * This module doesn't handle PCI PASID Stop Marker; IOMMU drivers must discard >> * them before reporting faults. A PASID Stop Marker (LRW = 0b100) doesn't >> @@ -143,22 +183,18 @@ int iommu_report_device_fault(struct device *dev, struct iopf_fault *evt) >> { >> struct iommu_fault *fault = &evt->fault; >> struct iommu_fault_param *iopf_param; >> - struct iopf_fault *iopf, *next; >> - struct iommu_domain *domain; >> + struct iopf_group abort_group = {}; >> struct iopf_group *group; >> int ret; >> >> - if (fault->type != IOMMU_FAULT_PAGE_REQ) >> - return -EOPNOTSUPP; >> - >> iopf_param = iopf_get_dev_fault_param(dev); >> - if (!iopf_param) >> + if (WARN_ON(!iopf_param)) >> return -ENODEV; >> >> if (!(fault->prm.flags & IOMMU_FAULT_PAGE_REQUEST_LAST_PAGE)) { >> ret = report_partial_fault(iopf_param, fault); >> iopf_put_dev_fault_param(iopf_param); >> - >> + /* A request that is not the last does not need to be ack'd */ >> return ret; >> } >> >> @@ -170,56 +206,33 @@ int iommu_report_device_fault(struct device *dev, struct iopf_fault *evt) >> * will send a response to the hardware. We need to clean up before >> * leaving, otherwise partial faults will be stuck. >> */ >> - domain = get_domain_for_iopf(dev, fault); >> - if (!domain) { >> + group = iopf_group_alloc(iopf_param, evt, &abort_group); >> + if (group == &abort_group) { >> + ret = -ENOMEM; >> + goto err_abort; >> + } >> + >> + group->domain = get_domain_for_iopf(dev, fault); >> + if (!group->domain) { >> ret = -EINVAL; >> - goto cleanup_partial; >> + goto err_abort; >> } >> >> - group = kzalloc(sizeof(*group), GFP_KERNEL); >> - if (!group) { >> - ret = -ENOMEM; >> - goto cleanup_partial; >> - } >> - >> - group->fault_param = iopf_param; >> - group->last_fault.fault = *fault; >> - INIT_LIST_HEAD(&group->faults); >> - INIT_LIST_HEAD(&group->pending_node); >> - group->domain = domain; >> - list_add(&group->last_fault.list, &group->faults); >> - >> - /* See if we have partial faults for this group */ >> - mutex_lock(&iopf_param->lock); >> - list_for_each_entry_safe(iopf, next, &iopf_param->partial, list) { >> - if (iopf->fault.prm.grpid == fault->prm.grpid) >> - /* Insert*before* the last fault */ >> - list_move(&iopf->list, &group->faults); >> - } >> - list_add(&group->pending_node, &iopf_param->faults); >> - mutex_unlock(&iopf_param->lock); >> + /* >> + * On success iopf_handler must call iopf_group_response() and >> + * iopf_free_group() >> + */ >> + ret = group->domain->iopf_handler(group); >> + if (ret) >> + goto err_abort; >> + return 0; >> >> - ret = domain->iopf_handler(group); >> - if (ret) { >> - mutex_lock(&iopf_param->lock); >> - list_del_init(&group->pending_node); >> - mutex_unlock(&iopf_param->lock); >> +err_abort: >> + iopf_group_response(group, IOMMU_PAGE_RESP_FAILURE); >> + if (group == &abort_group) >> + __iopf_free_group(group); >> + else >> iopf_free_group(group); >> - } >> - >> - return ret; >> - >> -cleanup_partial: >> - mutex_lock(&iopf_param->lock); >> - list_for_each_entry_safe(iopf, next, &iopf_param->partial, list) { >> - if (iopf->fault.prm.grpid == fault->prm.grpid) { >> - list_del(&iopf->list); >> - kfree(iopf); >> - } >> - } >> - mutex_unlock(&iopf_param->lock); >> - iopf_put_dev_fault_param(iopf_param); >> - >> return ret; >> } >> EXPORT_SYMBOL_GPL(iommu_report_device_fault); >> @@ -262,8 +275,8 @@ EXPORT_SYMBOL_GPL(iopf_queue_flush_dev); >> * >> * Return 0 on success and <0 on error. >> */ > Should you adjust the docs as well? Yes. Will do. Best regards, baolu
On Mon, Jan 22, 2024 at 01:43:07PM +0800, Lu Baolu wrote: > The iopf_group_response() should return void, as nothing can do anything > with the failure. This implies that ops->page_response() must also return > void; this is consistent with what the drivers do. The failure paths, > which are all integrity validations of the fault, should be WARN_ON'd, > not return codes. > > If the iommu core fails to enqueue the fault, it should respond the fault > directly by calling ops->page_response() instead of returning an error > number and relying on the iommu drivers to do so. Consolidate the error > fault handling code in the core. > > Co-developed-by: Jason Gunthorpe <jgg@nvidia.com> > Signed-off-by: Jason Gunthorpe <jgg@nvidia.com> > Signed-off-by: Lu Baolu <baolu.lu@linux.intel.com> > --- > include/linux/iommu.h | 14 +-- > drivers/iommu/intel/iommu.h | 4 +- > drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c | 50 +++----- > drivers/iommu/intel/svm.c | 18 +-- > drivers/iommu/io-pgfault.c | 132 +++++++++++--------- > 5 files changed, 99 insertions(+), 119 deletions(-) Reviewed-by: Jason Gunthorpe <jgg@nvidia.com> > > +static struct iopf_group *iopf_group_alloc(struct iommu_fault_param *iopf_param, > + struct iopf_fault *evt, > + struct iopf_group *abort_group) > +{ > + struct iopf_fault *iopf, *next; > + struct iopf_group *group; > + > + group = kzalloc(sizeof(*group), GFP_KERNEL); > + if (!group) { > + /* > + * We always need to construct the group as we need it to abort > + * the request at the driver if it cfan't be handled. ^^^^^^ can't Jason
diff --git a/include/linux/iommu.h b/include/linux/iommu.h index 48196efc9327..d7b6f4017254 100644 --- a/include/linux/iommu.h +++ b/include/linux/iommu.h @@ -578,9 +578,8 @@ struct iommu_ops { int (*dev_enable_feat)(struct device *dev, enum iommu_dev_features f); int (*dev_disable_feat)(struct device *dev, enum iommu_dev_features f); - int (*page_response)(struct device *dev, - struct iopf_fault *evt, - struct iommu_page_response *msg); + void (*page_response)(struct device *dev, struct iopf_fault *evt, + struct iommu_page_response *msg); int (*def_domain_type)(struct device *dev); void (*remove_dev_pasid)(struct device *dev, ioasid_t pasid); @@ -1551,8 +1550,8 @@ void iopf_queue_free(struct iopf_queue *queue); int iopf_queue_discard_partial(struct iopf_queue *queue); void iopf_free_group(struct iopf_group *group); int iommu_report_device_fault(struct device *dev, struct iopf_fault *evt); -int iopf_group_response(struct iopf_group *group, - enum iommu_page_response_code status); +void iopf_group_response(struct iopf_group *group, + enum iommu_page_response_code status); #else static inline int iopf_queue_add_device(struct iopf_queue *queue, struct device *dev) @@ -1594,10 +1593,9 @@ iommu_report_device_fault(struct device *dev, struct iopf_fault *evt) return -ENODEV; } -static inline int iopf_group_response(struct iopf_group *group, - enum iommu_page_response_code status) +static inline void iopf_group_response(struct iopf_group *group, + enum iommu_page_response_code status) { - return -ENODEV; } #endif /* CONFIG_IOMMU_IOPF */ #endif /* __LINUX_IOMMU_H */ diff --git a/drivers/iommu/intel/iommu.h b/drivers/iommu/intel/iommu.h index 696d95293a69..cf9a28c7fab8 100644 --- a/drivers/iommu/intel/iommu.h +++ b/drivers/iommu/intel/iommu.h @@ -1079,8 +1079,8 @@ struct iommu_domain *intel_nested_domain_alloc(struct iommu_domain *parent, void intel_svm_check(struct intel_iommu *iommu); int intel_svm_enable_prq(struct intel_iommu *iommu); int intel_svm_finish_prq(struct intel_iommu *iommu); -int intel_svm_page_response(struct device *dev, struct iopf_fault *evt, - struct iommu_page_response *msg); +void intel_svm_page_response(struct device *dev, struct iopf_fault *evt, + struct iommu_page_response *msg); struct iommu_domain *intel_svm_domain_alloc(void); void intel_svm_remove_dev_pasid(struct device *dev, ioasid_t pasid); void intel_drain_pasid_prq(struct device *dev, u32 pasid); diff --git a/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c b/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c index 4e93e845458c..42eb59cb99f4 100644 --- a/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c +++ b/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c @@ -920,31 +920,29 @@ static int arm_smmu_cmdq_batch_submit(struct arm_smmu_device *smmu, return arm_smmu_cmdq_issue_cmdlist(smmu, cmds->cmds, cmds->num, true); } -static int arm_smmu_page_response(struct device *dev, - struct iopf_fault *unused, - struct iommu_page_response *resp) +static void arm_smmu_page_response(struct device *dev, struct iopf_fault *unused, + struct iommu_page_response *resp) { struct arm_smmu_cmdq_ent cmd = {0}; struct arm_smmu_master *master = dev_iommu_priv_get(dev); int sid = master->streams[0].id; - if (master->stall_enabled) { - cmd.opcode = CMDQ_OP_RESUME; - cmd.resume.sid = sid; - cmd.resume.stag = resp->grpid; - switch (resp->code) { - case IOMMU_PAGE_RESP_INVALID: - case IOMMU_PAGE_RESP_FAILURE: - cmd.resume.resp = CMDQ_RESUME_0_RESP_ABORT; - break; - case IOMMU_PAGE_RESP_SUCCESS: - cmd.resume.resp = CMDQ_RESUME_0_RESP_RETRY; - break; - default: - return -EINVAL; - } - } else { - return -ENODEV; + if (WARN_ON(!master->stall_enabled)) + return; + + cmd.opcode = CMDQ_OP_RESUME; + cmd.resume.sid = sid; + cmd.resume.stag = resp->grpid; + switch (resp->code) { + case IOMMU_PAGE_RESP_INVALID: + case IOMMU_PAGE_RESP_FAILURE: + cmd.resume.resp = CMDQ_RESUME_0_RESP_ABORT; + break; + case IOMMU_PAGE_RESP_SUCCESS: + cmd.resume.resp = CMDQ_RESUME_0_RESP_RETRY; + break; + default: + break; } arm_smmu_cmdq_issue_cmd(master->smmu, &cmd); @@ -954,8 +952,6 @@ static int arm_smmu_page_response(struct device *dev, * terminated... at some point in the future. PRI_RESP is fire and * forget. */ - - return 0; } /* Context descriptor manipulation functions */ @@ -1516,16 +1512,6 @@ static int arm_smmu_handle_evt(struct arm_smmu_device *smmu, u64 *evt) } ret = iommu_report_device_fault(master->dev, &fault_evt); - if (ret && flt->type == IOMMU_FAULT_PAGE_REQ) { - /* Nobody cared, abort the access */ - struct iommu_page_response resp = { - .pasid = flt->prm.pasid, - .grpid = flt->prm.grpid, - .code = IOMMU_PAGE_RESP_FAILURE, - }; - arm_smmu_page_response(master->dev, &fault_evt, &resp); - } - out_unlock: mutex_unlock(&smmu->streams_mutex); return ret; diff --git a/drivers/iommu/intel/svm.c b/drivers/iommu/intel/svm.c index e1cbcb9515f0..2f8716636dbb 100644 --- a/drivers/iommu/intel/svm.c +++ b/drivers/iommu/intel/svm.c @@ -740,9 +740,8 @@ static irqreturn_t prq_event_thread(int irq, void *d) return IRQ_RETVAL(handled); } -int intel_svm_page_response(struct device *dev, - struct iopf_fault *evt, - struct iommu_page_response *msg) +void intel_svm_page_response(struct device *dev, struct iopf_fault *evt, + struct iommu_page_response *msg) { struct device_domain_info *info = dev_iommu_priv_get(dev); struct intel_iommu *iommu = info->iommu; @@ -751,7 +750,6 @@ int intel_svm_page_response(struct device *dev, bool private_present; bool pasid_present; bool last_page; - int ret = 0; u16 sid; prm = &evt->fault.prm; @@ -760,16 +758,6 @@ int intel_svm_page_response(struct device *dev, private_present = prm->flags & IOMMU_FAULT_PAGE_REQUEST_PRIV_DATA; last_page = prm->flags & IOMMU_FAULT_PAGE_REQUEST_LAST_PAGE; - if (!pasid_present) { - ret = -EINVAL; - goto out; - } - - if (prm->pasid == 0 || prm->pasid >= PASID_MAX) { - ret = -EINVAL; - goto out; - } - /* * Per VT-d spec. v3.0 ch7.7, system software must respond * with page group response if private data is present (PDP) @@ -798,8 +786,6 @@ int intel_svm_page_response(struct device *dev, qi_submit_sync(iommu, &desc, 1, 0); } -out: - return ret; } static int intel_svm_set_dev_pasid(struct iommu_domain *domain, diff --git a/drivers/iommu/io-pgfault.c b/drivers/iommu/io-pgfault.c index c22e13df84c2..6e63e5a02884 100644 --- a/drivers/iommu/io-pgfault.c +++ b/drivers/iommu/io-pgfault.c @@ -39,7 +39,7 @@ static void iopf_put_dev_fault_param(struct iommu_fault_param *fault_param) kfree_rcu(fault_param, rcu); } -void iopf_free_group(struct iopf_group *group) +static void __iopf_free_group(struct iopf_group *group) { struct iopf_fault *iopf, *next; @@ -50,6 +50,11 @@ void iopf_free_group(struct iopf_group *group) /* Pair with iommu_report_device_fault(). */ iopf_put_dev_fault_param(group->fault_param); +} + +void iopf_free_group(struct iopf_group *group) +{ + __iopf_free_group(group); kfree(group); } EXPORT_SYMBOL_GPL(iopf_free_group); @@ -97,14 +102,49 @@ static int report_partial_fault(struct iommu_fault_param *fault_param, return 0; } +static struct iopf_group *iopf_group_alloc(struct iommu_fault_param *iopf_param, + struct iopf_fault *evt, + struct iopf_group *abort_group) +{ + struct iopf_fault *iopf, *next; + struct iopf_group *group; + + group = kzalloc(sizeof(*group), GFP_KERNEL); + if (!group) { + /* + * We always need to construct the group as we need it to abort + * the request at the driver if it cfan't be handled. + */ + group = abort_group; + } + + group->fault_param = iopf_param; + group->last_fault.fault = evt->fault; + INIT_LIST_HEAD(&group->faults); + INIT_LIST_HEAD(&group->pending_node); + list_add(&group->last_fault.list, &group->faults); + + /* See if we have partial faults for this group */ + mutex_lock(&iopf_param->lock); + list_for_each_entry_safe(iopf, next, &iopf_param->partial, list) { + if (iopf->fault.prm.grpid == evt->fault.prm.grpid) + /* Insert *before* the last fault */ + list_move(&iopf->list, &group->faults); + } + list_add(&group->pending_node, &iopf_param->faults); + mutex_unlock(&iopf_param->lock); + + return group; +} + /** * iommu_report_device_fault() - Report fault event to device driver * @dev: the device * @evt: fault event data * * Called by IOMMU drivers when a fault is detected, typically in a threaded IRQ - * handler. When this function fails and the fault is recoverable, it is the - * caller's responsibility to complete the fault. + * handler. If this function fails then ops->page_response() was called to + * complete evt if required. * * This module doesn't handle PCI PASID Stop Marker; IOMMU drivers must discard * them before reporting faults. A PASID Stop Marker (LRW = 0b100) doesn't @@ -143,22 +183,18 @@ int iommu_report_device_fault(struct device *dev, struct iopf_fault *evt) { struct iommu_fault *fault = &evt->fault; struct iommu_fault_param *iopf_param; - struct iopf_fault *iopf, *next; - struct iommu_domain *domain; + struct iopf_group abort_group = {}; struct iopf_group *group; int ret; - if (fault->type != IOMMU_FAULT_PAGE_REQ) - return -EOPNOTSUPP; - iopf_param = iopf_get_dev_fault_param(dev); - if (!iopf_param) + if (WARN_ON(!iopf_param)) return -ENODEV; if (!(fault->prm.flags & IOMMU_FAULT_PAGE_REQUEST_LAST_PAGE)) { ret = report_partial_fault(iopf_param, fault); iopf_put_dev_fault_param(iopf_param); - + /* A request that is not the last does not need to be ack'd */ return ret; } @@ -170,56 +206,33 @@ int iommu_report_device_fault(struct device *dev, struct iopf_fault *evt) * will send a response to the hardware. We need to clean up before * leaving, otherwise partial faults will be stuck. */ - domain = get_domain_for_iopf(dev, fault); - if (!domain) { + group = iopf_group_alloc(iopf_param, evt, &abort_group); + if (group == &abort_group) { + ret = -ENOMEM; + goto err_abort; + } + + group->domain = get_domain_for_iopf(dev, fault); + if (!group->domain) { ret = -EINVAL; - goto cleanup_partial; + goto err_abort; } - group = kzalloc(sizeof(*group), GFP_KERNEL); - if (!group) { - ret = -ENOMEM; - goto cleanup_partial; - } - - group->fault_param = iopf_param; - group->last_fault.fault = *fault; - INIT_LIST_HEAD(&group->faults); - INIT_LIST_HEAD(&group->pending_node); - group->domain = domain; - list_add(&group->last_fault.list, &group->faults); - - /* See if we have partial faults for this group */ - mutex_lock(&iopf_param->lock); - list_for_each_entry_safe(iopf, next, &iopf_param->partial, list) { - if (iopf->fault.prm.grpid == fault->prm.grpid) - /* Insert *before* the last fault */ - list_move(&iopf->list, &group->faults); - } - list_add(&group->pending_node, &iopf_param->faults); - mutex_unlock(&iopf_param->lock); + /* + * On success iopf_handler must call iopf_group_response() and + * iopf_free_group() + */ + ret = group->domain->iopf_handler(group); + if (ret) + goto err_abort; + return 0; - ret = domain->iopf_handler(group); - if (ret) { - mutex_lock(&iopf_param->lock); - list_del_init(&group->pending_node); - mutex_unlock(&iopf_param->lock); +err_abort: + iopf_group_response(group, IOMMU_PAGE_RESP_FAILURE); + if (group == &abort_group) + __iopf_free_group(group); + else iopf_free_group(group); - } - - return ret; - -cleanup_partial: - mutex_lock(&iopf_param->lock); - list_for_each_entry_safe(iopf, next, &iopf_param->partial, list) { - if (iopf->fault.prm.grpid == fault->prm.grpid) { - list_del(&iopf->list); - kfree(iopf); - } - } - mutex_unlock(&iopf_param->lock); - iopf_put_dev_fault_param(iopf_param); - return ret; } EXPORT_SYMBOL_GPL(iommu_report_device_fault); @@ -262,8 +275,8 @@ EXPORT_SYMBOL_GPL(iopf_queue_flush_dev); * * Return 0 on success and <0 on error. */ -int iopf_group_response(struct iopf_group *group, - enum iommu_page_response_code status) +void iopf_group_response(struct iopf_group *group, + enum iommu_page_response_code status) { struct iommu_fault_param *fault_param = group->fault_param; struct iopf_fault *iopf = &group->last_fault; @@ -274,7 +287,6 @@ int iopf_group_response(struct iopf_group *group, .grpid = iopf->fault.prm.grpid, .code = status, }; - int ret = -EINVAL; if ((iopf->fault.prm.flags & IOMMU_FAULT_PAGE_REQUEST_PASID_VALID) && (iopf->fault.prm.flags & IOMMU_FAULT_PAGE_RESPONSE_NEEDS_PASID)) @@ -283,12 +295,10 @@ int iopf_group_response(struct iopf_group *group, /* Only send response if there is a fault report pending */ mutex_lock(&fault_param->lock); if (!list_empty(&group->pending_node)) { - ret = ops->page_response(dev, &group->last_fault, &resp); + ops->page_response(dev, &group->last_fault, &resp); list_del_init(&group->pending_node); } mutex_unlock(&fault_param->lock); - - return ret; } EXPORT_SYMBOL_GPL(iopf_group_response);