Message ID | 20230213211825.3507034-3-tanmay.shah@amd.com (mailing list archive) |
---|---|
State | Superseded |
Headers | show |
Series | drivers: remoteproc: xilinx: add mailbox support | expand |
On Mon, Feb 13, 2023 at 01:18:25PM -0800, Tanmay Shah wrote: > If the unit address is appended to node name of memory-region, > then adding rproc carveouts fails as node name and unit-address > both are passed as carveout name (i.e. vdev0vring0@xxxxxxxx). However, > only node name is expected by remoteproc framework. This patch moves > memory-region node parsing from driver probe to prepare and > only passes node-name and not unit-address > > Fixes: 6b291e8020a8 ("drivers: remoteproc: Add Xilinx r5 remoteproc driver") > Signed-off-by: Tanmay Shah <tanmay.shah@amd.com> > --- > > Changelog: > - This is first version of this change, however posting as part of the series > that has version v3. The v2 of the series could be found at following link. > > v2: https://lore.kernel.org/all/20230126213154.1707300-1-tanmay.shah@amd.com/ > > drivers/remoteproc/xlnx_r5_remoteproc.c | 87 ++++++------------------- > 1 file changed, 20 insertions(+), 67 deletions(-) > > diff --git a/drivers/remoteproc/xlnx_r5_remoteproc.c b/drivers/remoteproc/xlnx_r5_remoteproc.c > index 2db57d394155..81af2dea56c2 100644 > --- a/drivers/remoteproc/xlnx_r5_remoteproc.c > +++ b/drivers/remoteproc/xlnx_r5_remoteproc.c > @@ -61,8 +61,6 @@ static const struct mem_bank_data zynqmp_tcm_banks[] = { > * @np: device node of RPU instance > * @tcm_bank_count: number TCM banks accessible to this RPU > * @tcm_banks: array of each TCM bank data > - * @rmem_count: Number of reserved mem regions > - * @rmem: reserved memory region nodes from device tree > * @rproc: rproc handle > * @pm_domain_id: RPU CPU power domain id > */ > @@ -71,8 +69,6 @@ struct zynqmp_r5_core { > struct device_node *np; > int tcm_bank_count; > struct mem_bank_data **tcm_banks; > - int rmem_count; > - struct reserved_mem **rmem; > struct rproc *rproc; > u32 pm_domain_id; > }; > @@ -239,21 +235,31 @@ static int add_mem_regions_carveout(struct rproc *rproc) > { > struct rproc_mem_entry *rproc_mem; > struct zynqmp_r5_core *r5_core; > + struct device_node *rmem_np; > struct reserved_mem *rmem; > int i, num_mem_regions; > > r5_core = (struct zynqmp_r5_core *)rproc->priv; > - num_mem_regions = r5_core->rmem_count; > + > + num_mem_regions = of_property_count_elems_of_size(r5_core->np, "memory-region", > + sizeof(phandle)); > > for (i = 0; i < num_mem_regions; i++) { > - rmem = r5_core->rmem[i]; > Extra line Everyone else in the remoteproc subsystem is using of_phandle_iterator_next(), please do the same. It is easier to maintain and you don't have to call of_node_put() after each iteration. > - if (!strncmp(rmem->name, "vdev0buffer", strlen("vdev0buffer"))) { > + rmem_np = of_parse_phandle(r5_core->np, "memory-region", i); > + > + rmem = of_reserved_mem_lookup(rmem_np); > + if (!rmem) { > + of_node_put(rmem_np); > + return -EINVAL; > + } > + > + if (!strcmp(rmem_np->name, "vdev0buffer")) { > /* Init reserved memory for vdev buffer */ > rproc_mem = rproc_of_resm_mem_entry_init(&rproc->dev, i, > rmem->size, > rmem->base, > - rmem->name); > + rmem_np->name); > } else { > /* Register associated reserved memory regions */ > rproc_mem = rproc_mem_entry_init(&rproc->dev, NULL, > @@ -261,16 +267,20 @@ static int add_mem_regions_carveout(struct rproc *rproc) > rmem->size, rmem->base, > zynqmp_r5_mem_region_map, > zynqmp_r5_mem_region_unmap, > - rmem->name); > + rmem_np->name); > } > > - if (!rproc_mem) > + if (!rproc_mem) { > + of_node_put(rmem_np); When moving to of_phandle_iterator_next(), of_node_put(it.node) has to be called on error conditions. Other drivers don't do it, something I will fix in the next cycle. > return -ENOMEM; > + } > > rproc_add_carveout(rproc, rproc_mem); > > dev_dbg(&rproc->dev, "reserved mem carveout %s addr=%llx, size=0x%llx", > rmem->name, rmem->base, rmem->size); > + > + of_node_put(rmem_np); > } > > return 0; > @@ -726,59 +736,6 @@ static int zynqmp_r5_get_tcm_node(struct zynqmp_r5_cluster *cluster) > return 0; > } > > -/** > - * zynqmp_r5_get_mem_region_node() > - * parse memory-region property and get reserved mem regions > - * > - * @r5_core: pointer to zynqmp_r5_core type object > - * > - * Return: 0 for success and error code for failure. > - */ > -static int zynqmp_r5_get_mem_region_node(struct zynqmp_r5_core *r5_core) > -{ > - struct device_node *np, *rmem_np; > - struct reserved_mem **rmem; > - int res_mem_count, i; > - struct device *dev; > - > - dev = r5_core->dev; > - np = r5_core->np; > - > - res_mem_count = of_property_count_elems_of_size(np, "memory-region", > - sizeof(phandle)); > - if (res_mem_count <= 0) { > - dev_warn(dev, "failed to get memory-region property %d\n", > - res_mem_count); > - return 0; > - } > - > - rmem = devm_kcalloc(dev, res_mem_count, > - sizeof(struct reserved_mem *), GFP_KERNEL); > - if (!rmem) > - return -ENOMEM; > - > - for (i = 0; i < res_mem_count; i++) { > - rmem_np = of_parse_phandle(np, "memory-region", i); > - if (!rmem_np) > - goto release_rmem; > - > - rmem[i] = of_reserved_mem_lookup(rmem_np); > - if (!rmem[i]) { > - of_node_put(rmem_np); > - goto release_rmem; > - } > - > - of_node_put(rmem_np); > - } > - > - r5_core->rmem_count = res_mem_count; > - r5_core->rmem = rmem; > - return 0; > - > -release_rmem: > - return -EINVAL; > -} > - > /* > * zynqmp_r5_core_init() > * Create and initialize zynqmp_r5_core type object > @@ -806,10 +763,6 @@ static int zynqmp_r5_core_init(struct zynqmp_r5_cluster *cluster, > for (i = 0; i < cluster->core_count; i++) { > r5_core = cluster->r5_cores[i]; > > - ret = zynqmp_r5_get_mem_region_node(r5_core); > - if (ret) > - dev_warn(dev, "memory-region prop failed %d\n", ret); > - > /* Initialize r5 cores with power-domains parsed from dts */ > ret = of_property_read_u32_index(r5_core->np, "power-domains", > 1, &r5_core->pm_domain_id); > -- > 2.25.1 >
Hi Mathieu, Thanks for the reviews. I agree to all the comments and will address in the next revision accordingly. Thanks, Tanmay On 2/22/23 10:41 AM, Mathieu Poirier wrote: > On Mon, Feb 13, 2023 at 01:18:25PM -0800, Tanmay Shah wrote: >> If the unit address is appended to node name of memory-region, >> then adding rproc carveouts fails as node name and unit-address >> both are passed as carveout name (i.e. vdev0vring0@xxxxxxxx). However, >> only node name is expected by remoteproc framework. This patch moves >> memory-region node parsing from driver probe to prepare and >> only passes node-name and not unit-address >> >> Fixes: 6b291e8020a8 ("drivers: remoteproc: Add Xilinx r5 remoteproc driver") >> Signed-off-by: Tanmay Shah <tanmay.shah@amd.com> >> --- >> >> Changelog: >> - This is first version of this change, however posting as part of the series >> that has version v3. The v2 of the series could be found at following link. >> >> v2: https://lore.kernel.org/all/20230126213154.1707300-1-tanmay.shah@amd.com/ >> >> drivers/remoteproc/xlnx_r5_remoteproc.c | 87 ++++++------------------- >> 1 file changed, 20 insertions(+), 67 deletions(-) >> >> diff --git a/drivers/remoteproc/xlnx_r5_remoteproc.c b/drivers/remoteproc/xlnx_r5_remoteproc.c >> index 2db57d394155..81af2dea56c2 100644 >> --- a/drivers/remoteproc/xlnx_r5_remoteproc.c >> +++ b/drivers/remoteproc/xlnx_r5_remoteproc.c >> @@ -61,8 +61,6 @@ static const struct mem_bank_data zynqmp_tcm_banks[] = { >> * @np: device node of RPU instance >> * @tcm_bank_count: number TCM banks accessible to this RPU >> * @tcm_banks: array of each TCM bank data >> - * @rmem_count: Number of reserved mem regions >> - * @rmem: reserved memory region nodes from device tree >> * @rproc: rproc handle >> * @pm_domain_id: RPU CPU power domain id >> */ >> @@ -71,8 +69,6 @@ struct zynqmp_r5_core { >> struct device_node *np; >> int tcm_bank_count; >> struct mem_bank_data **tcm_banks; >> - int rmem_count; >> - struct reserved_mem **rmem; >> struct rproc *rproc; >> u32 pm_domain_id; >> }; >> @@ -239,21 +235,31 @@ static int add_mem_regions_carveout(struct rproc *rproc) >> { >> struct rproc_mem_entry *rproc_mem; >> struct zynqmp_r5_core *r5_core; >> + struct device_node *rmem_np; >> struct reserved_mem *rmem; >> int i, num_mem_regions; >> >> r5_core = (struct zynqmp_r5_core *)rproc->priv; >> - num_mem_regions = r5_core->rmem_count; >> + >> + num_mem_regions = of_property_count_elems_of_size(r5_core->np, "memory-region", >> + sizeof(phandle)); >> >> for (i = 0; i < num_mem_regions; i++) { >> - rmem = r5_core->rmem[i]; >> > Extra line > > Everyone else in the remoteproc subsystem is using of_phandle_iterator_next(), > please do the same. It is easier to maintain and you don't have to call > of_node_put() after each iteration. > > >> - if (!strncmp(rmem->name, "vdev0buffer", strlen("vdev0buffer"))) { >> + rmem_np = of_parse_phandle(r5_core->np, "memory-region", i); >> + >> + rmem = of_reserved_mem_lookup(rmem_np); >> + if (!rmem) { >> + of_node_put(rmem_np); >> + return -EINVAL; >> + } >> + >> + if (!strcmp(rmem_np->name, "vdev0buffer")) { >> /* Init reserved memory for vdev buffer */ >> rproc_mem = rproc_of_resm_mem_entry_init(&rproc->dev, i, >> rmem->size, >> rmem->base, >> - rmem->name); >> + rmem_np->name); >> } else { >> /* Register associated reserved memory regions */ >> rproc_mem = rproc_mem_entry_init(&rproc->dev, NULL, >> @@ -261,16 +267,20 @@ static int add_mem_regions_carveout(struct rproc *rproc) >> rmem->size, rmem->base, >> zynqmp_r5_mem_region_map, >> zynqmp_r5_mem_region_unmap, >> - rmem->name); >> + rmem_np->name); >> } >> >> - if (!rproc_mem) >> + if (!rproc_mem) { >> + of_node_put(rmem_np); > When moving to of_phandle_iterator_next(), of_node_put(it.node) has to be > called on error conditions. Other drivers don't do it, something I will fix in > the next cycle. > >> return -ENOMEM; >> + } >> >> rproc_add_carveout(rproc, rproc_mem); >> >> dev_dbg(&rproc->dev, "reserved mem carveout %s addr=%llx, size=0x%llx", >> rmem->name, rmem->base, rmem->size); >> + >> + of_node_put(rmem_np); >> } >> >> return 0; >> @@ -726,59 +736,6 @@ static int zynqmp_r5_get_tcm_node(struct zynqmp_r5_cluster *cluster) >> return 0; >> } >> >> -/** >> - * zynqmp_r5_get_mem_region_node() >> - * parse memory-region property and get reserved mem regions >> - * >> - * @r5_core: pointer to zynqmp_r5_core type object >> - * >> - * Return: 0 for success and error code for failure. >> - */ >> -static int zynqmp_r5_get_mem_region_node(struct zynqmp_r5_core *r5_core) >> -{ >> - struct device_node *np, *rmem_np; >> - struct reserved_mem **rmem; >> - int res_mem_count, i; >> - struct device *dev; >> - >> - dev = r5_core->dev; >> - np = r5_core->np; >> - >> - res_mem_count = of_property_count_elems_of_size(np, "memory-region", >> - sizeof(phandle)); >> - if (res_mem_count <= 0) { >> - dev_warn(dev, "failed to get memory-region property %d\n", >> - res_mem_count); >> - return 0; >> - } >> - >> - rmem = devm_kcalloc(dev, res_mem_count, >> - sizeof(struct reserved_mem *), GFP_KERNEL); >> - if (!rmem) >> - return -ENOMEM; >> - >> - for (i = 0; i < res_mem_count; i++) { >> - rmem_np = of_parse_phandle(np, "memory-region", i); >> - if (!rmem_np) >> - goto release_rmem; >> - >> - rmem[i] = of_reserved_mem_lookup(rmem_np); >> - if (!rmem[i]) { >> - of_node_put(rmem_np); >> - goto release_rmem; >> - } >> - >> - of_node_put(rmem_np); >> - } >> - >> - r5_core->rmem_count = res_mem_count; >> - r5_core->rmem = rmem; >> - return 0; >> - >> -release_rmem: >> - return -EINVAL; >> -} >> - >> /* >> * zynqmp_r5_core_init() >> * Create and initialize zynqmp_r5_core type object >> @@ -806,10 +763,6 @@ static int zynqmp_r5_core_init(struct zynqmp_r5_cluster *cluster, >> for (i = 0; i < cluster->core_count; i++) { >> r5_core = cluster->r5_cores[i]; >> >> - ret = zynqmp_r5_get_mem_region_node(r5_core); >> - if (ret) >> - dev_warn(dev, "memory-region prop failed %d\n", ret); >> - >> /* Initialize r5 cores with power-domains parsed from dts */ >> ret = of_property_read_u32_index(r5_core->np, "power-domains", >> 1, &r5_core->pm_domain_id); >> -- >> 2.25.1 >>
diff --git a/drivers/remoteproc/xlnx_r5_remoteproc.c b/drivers/remoteproc/xlnx_r5_remoteproc.c index 2db57d394155..81af2dea56c2 100644 --- a/drivers/remoteproc/xlnx_r5_remoteproc.c +++ b/drivers/remoteproc/xlnx_r5_remoteproc.c @@ -61,8 +61,6 @@ static const struct mem_bank_data zynqmp_tcm_banks[] = { * @np: device node of RPU instance * @tcm_bank_count: number TCM banks accessible to this RPU * @tcm_banks: array of each TCM bank data - * @rmem_count: Number of reserved mem regions - * @rmem: reserved memory region nodes from device tree * @rproc: rproc handle * @pm_domain_id: RPU CPU power domain id */ @@ -71,8 +69,6 @@ struct zynqmp_r5_core { struct device_node *np; int tcm_bank_count; struct mem_bank_data **tcm_banks; - int rmem_count; - struct reserved_mem **rmem; struct rproc *rproc; u32 pm_domain_id; }; @@ -239,21 +235,31 @@ static int add_mem_regions_carveout(struct rproc *rproc) { struct rproc_mem_entry *rproc_mem; struct zynqmp_r5_core *r5_core; + struct device_node *rmem_np; struct reserved_mem *rmem; int i, num_mem_regions; r5_core = (struct zynqmp_r5_core *)rproc->priv; - num_mem_regions = r5_core->rmem_count; + + num_mem_regions = of_property_count_elems_of_size(r5_core->np, "memory-region", + sizeof(phandle)); for (i = 0; i < num_mem_regions; i++) { - rmem = r5_core->rmem[i]; - if (!strncmp(rmem->name, "vdev0buffer", strlen("vdev0buffer"))) { + rmem_np = of_parse_phandle(r5_core->np, "memory-region", i); + + rmem = of_reserved_mem_lookup(rmem_np); + if (!rmem) { + of_node_put(rmem_np); + return -EINVAL; + } + + if (!strcmp(rmem_np->name, "vdev0buffer")) { /* Init reserved memory for vdev buffer */ rproc_mem = rproc_of_resm_mem_entry_init(&rproc->dev, i, rmem->size, rmem->base, - rmem->name); + rmem_np->name); } else { /* Register associated reserved memory regions */ rproc_mem = rproc_mem_entry_init(&rproc->dev, NULL, @@ -261,16 +267,20 @@ static int add_mem_regions_carveout(struct rproc *rproc) rmem->size, rmem->base, zynqmp_r5_mem_region_map, zynqmp_r5_mem_region_unmap, - rmem->name); + rmem_np->name); } - if (!rproc_mem) + if (!rproc_mem) { + of_node_put(rmem_np); return -ENOMEM; + } rproc_add_carveout(rproc, rproc_mem); dev_dbg(&rproc->dev, "reserved mem carveout %s addr=%llx, size=0x%llx", rmem->name, rmem->base, rmem->size); + + of_node_put(rmem_np); } return 0; @@ -726,59 +736,6 @@ static int zynqmp_r5_get_tcm_node(struct zynqmp_r5_cluster *cluster) return 0; } -/** - * zynqmp_r5_get_mem_region_node() - * parse memory-region property and get reserved mem regions - * - * @r5_core: pointer to zynqmp_r5_core type object - * - * Return: 0 for success and error code for failure. - */ -static int zynqmp_r5_get_mem_region_node(struct zynqmp_r5_core *r5_core) -{ - struct device_node *np, *rmem_np; - struct reserved_mem **rmem; - int res_mem_count, i; - struct device *dev; - - dev = r5_core->dev; - np = r5_core->np; - - res_mem_count = of_property_count_elems_of_size(np, "memory-region", - sizeof(phandle)); - if (res_mem_count <= 0) { - dev_warn(dev, "failed to get memory-region property %d\n", - res_mem_count); - return 0; - } - - rmem = devm_kcalloc(dev, res_mem_count, - sizeof(struct reserved_mem *), GFP_KERNEL); - if (!rmem) - return -ENOMEM; - - for (i = 0; i < res_mem_count; i++) { - rmem_np = of_parse_phandle(np, "memory-region", i); - if (!rmem_np) - goto release_rmem; - - rmem[i] = of_reserved_mem_lookup(rmem_np); - if (!rmem[i]) { - of_node_put(rmem_np); - goto release_rmem; - } - - of_node_put(rmem_np); - } - - r5_core->rmem_count = res_mem_count; - r5_core->rmem = rmem; - return 0; - -release_rmem: - return -EINVAL; -} - /* * zynqmp_r5_core_init() * Create and initialize zynqmp_r5_core type object @@ -806,10 +763,6 @@ static int zynqmp_r5_core_init(struct zynqmp_r5_cluster *cluster, for (i = 0; i < cluster->core_count; i++) { r5_core = cluster->r5_cores[i]; - ret = zynqmp_r5_get_mem_region_node(r5_core); - if (ret) - dev_warn(dev, "memory-region prop failed %d\n", ret); - /* Initialize r5 cores with power-domains parsed from dts */ ret = of_property_read_u32_index(r5_core->np, "power-domains", 1, &r5_core->pm_domain_id);
If the unit address is appended to node name of memory-region, then adding rproc carveouts fails as node name and unit-address both are passed as carveout name (i.e. vdev0vring0@xxxxxxxx). However, only node name is expected by remoteproc framework. This patch moves memory-region node parsing from driver probe to prepare and only passes node-name and not unit-address Fixes: 6b291e8020a8 ("drivers: remoteproc: Add Xilinx r5 remoteproc driver") Signed-off-by: Tanmay Shah <tanmay.shah@amd.com> --- Changelog: - This is first version of this change, however posting as part of the series that has version v3. The v2 of the series could be found at following link. v2: https://lore.kernel.org/all/20230126213154.1707300-1-tanmay.shah@amd.com/ drivers/remoteproc/xlnx_r5_remoteproc.c | 87 ++++++------------------- 1 file changed, 20 insertions(+), 67 deletions(-)