@@ -700,6 +700,11 @@ static inline int dma_set_seg_boundary(struct device *dev, unsigned long mask)
return -EIO;
}
+extern int dmam_set_dma_parms(struct device *dev, unsigned int max_seg_size,
+ unsigned long seg_bound_mask);
+
+extern void dmam_free_dma_parms(struct device *dev);
+
#ifndef dma_max_pfn
static inline unsigned long dma_max_pfn(struct device *dev)
{
@@ -198,6 +198,56 @@ EXPORT_SYMBOL(dmam_release_declared_memory);
#endif
+static void dmam_release_dma_parms(struct device *dev, void *res)
+{
+ dev->dma_parms = NULL;
+}
+
+static int dmam_match_dma_parms(struct device *dev, void *res, void *data)
+{
+ return res == data;
+}
+
+/**
+ * dmam_set_dma_parms - Managed setting of dma_parms
+ * @dev: device to set dma_parms for
+ * @max_seg_size: the maximum segment size for this device
+ * @seg_bound_mask: the segment boundary mask for this device
+ *
+ * RETURNS:
+ * 0 on success, errno on failure.
+ */
+int dmam_set_dma_parms(struct device *dev, unsigned int max_seg_size,
+ unsigned long seg_bound_mask)
+{
+ struct device_dma_parameters *parms;
+
+ parms = devres_alloc(dmam_release_dma_parms,
+ sizeof(struct device_dma_parameters), GFP_KERNEL);
+ if (!parms)
+ return -ENOMEM;
+
+ dev->dma_parms = parms;
+ dma_set_max_seg_size(dev, max_seg_size);
+ dma_set_seg_boundary(dev, seg_bound_mask);
+
+ devres_add(dev, parms);
+ return 0;
+}
+EXPORT_SYMBOL_GPL(dmam_set_dma_parms);
+
+/**
+ * dmam_free_dma_parms - free dma_parms allocated with dmam_set_dma_parms
+ * @dev: device with dma_parms allocated by dmam_set_dma_parms()
+ */
+void dmam_free_dma_parms(struct device *dev)
+{
+ int rc = devres_destroy(dev, dmam_release_dma_parms, dmam_match_dma_parms,
+ dev->dma_parms);
+ WARN_ON(rc);
+}
+EXPORT_SYMBOL_GPL(dmam_free_dma_parms);
+
/*
* Create scatter-list for the already allocated DMA buffer.
*/
Setting up dma_parms seems not so trivial because it is easy to miss that its pointer has the life cycle of its containing 'struct device' while the allocation of dma_parms usually happens during the bind/unbind life cycle. Which can lead to dangling pointers. If coding this correctly in drivers, this results in boilerplate code. So, this patch adds a devm_* style helper which is easy to use and make sure the allocation and the pointer are always handled at the same time. Signed-off-by: Wolfram Sang <wsa+renesas@sang-engineering.com> --- include/linux/dma-mapping.h | 5 ++++ kernel/dma/mapping.c | 50 +++++++++++++++++++++++++++++++++++++ 2 files changed, 55 insertions(+)