diff mbox series

[v2,08/29] drm/xe: Add dma_addr res cursor

Message ID 20241016032518.539495-9-matthew.brost@intel.com (mailing list archive)
State New, archived
Headers show
Series Introduce GPU SVM and Xe SVM implementation | expand

Commit Message

Matthew Brost Oct. 16, 2024, 3:24 a.m. UTC
From: Thomas Hellström <thomas.hellstrom@linux.intel.com>

Useful for SVM ranges in SRAM and programing page tables.

Signed-off-by: Matthew Brost <matthew.brost@intel.com>
Signed-off-by: Thomas Hellström <thomas.hellstrom@linux.intel.com>
---
 drivers/gpu/drm/xe/xe_res_cursor.h | 116 ++++++++++++++++++++++++++++-
 drivers/gpu/drm/xe/xe_svm.h        |   4 +
 2 files changed, 118 insertions(+), 2 deletions(-)

Comments

Thomas Hellstrom Nov. 19, 2024, 12:15 p.m. UTC | #1
On Tue, 2024-10-15 at 20:24 -0700, Matthew Brost wrote:
> From: Thomas Hellström <thomas.hellstrom@linux.intel.com>
> 
> Useful for SVM ranges in SRAM and programing page tables.

We should look at providing a better commit message.

> 
> Signed-off-by: Matthew Brost <matthew.brost@intel.com>
> Signed-off-by: Thomas Hellström <thomas.hellstrom@linux.intel.com>
> ---
>  drivers/gpu/drm/xe/xe_res_cursor.h | 116
> ++++++++++++++++++++++++++++-
>  drivers/gpu/drm/xe/xe_svm.h        |   4 +
>  2 files changed, 118 insertions(+), 2 deletions(-)
> 
> diff --git a/drivers/gpu/drm/xe/xe_res_cursor.h
> b/drivers/gpu/drm/xe/xe_res_cursor.h
> index dca374b6521c..3faa3d9adb82 100644
> --- a/drivers/gpu/drm/xe/xe_res_cursor.h
> +++ b/drivers/gpu/drm/xe/xe_res_cursor.h
> @@ -30,13 +30,18 @@
>  #include <drm/ttm/ttm_range_manager.h>
>  #include <drm/ttm/ttm_resource.h>
>  #include <drm/ttm/ttm_tt.h>
> +#include "drm_pagemap.h"
>  
>  #include "xe_bo.h"
>  #include "xe_device.h"
>  #include "xe_macros.h"
> +#include "xe_svm.h"
>  #include "xe_ttm_vram_mgr.h"
>  
> -/* state back for walking over vram_mgr, stolen_mgr, and gtt_mgr
> allocations */
> +/**
> + * struct xe_res_cursor - state for walking over vram_mgr,
> stolen_mgr,
> + * and gtt_mgr allocations
> + */
>  struct xe_res_cursor {
>  	u64 start;
>  	u64 size;
> @@ -44,7 +49,17 @@ struct xe_res_cursor {
>  	void *node;
>  	u32 mem_type;
>  	struct scatterlist *sgl;
> +	/** @dma_addr: Current element in a struct
> drm_pagemap_dma_addr array */
> +	const struct drm_pagemap_dma_addr *dma_addr;
>  	struct drm_buddy *mm;
> +	/**
> +	 * @dma_start: DMA start address for the current segment.
> +	 * This may be different to @dma_addr.addr since elements in
> +	 * the array may be coalesced to a single segment.
> +	 */
> +	u64 dma_start;
> +	/** @dma_seg_size: Size of the current segment. */
> +	u64 dma_seg_size;
>  };
>  
>  static struct drm_buddy *xe_res_get_buddy(struct ttm_resource *res)
> @@ -70,6 +85,7 @@ static inline void xe_res_first(struct ttm_resource
> *res,
>  				struct xe_res_cursor *cur)
>  {
>  	cur->sgl = NULL;
> +	cur->dma_addr = NULL;
>  	if (!res)
>  		goto fallback;
>  
> @@ -141,6 +157,36 @@ static inline void __xe_res_sg_next(struct
> xe_res_cursor *cur)
>  	cur->sgl = sgl;
>  }
>  
> +/**
> + * __xe_res_dma_next() - Advance the cursor when end-of-segment is
> reached
> + * @cur: The cursor
> + */
> +static inline void __xe_res_dma_next(struct xe_res_cursor *cur)
> +{
> +	const struct drm_pagemap_dma_addr *addr = cur->dma_addr;
> +	u64 start = cur->start;
> +
> +	while (start >= cur->dma_seg_size) {
> +		start -= cur->dma_seg_size;
> +		addr++;
> +		cur->dma_seg_size = PAGE_SIZE << addr->order;
> +	}
> +	cur->dma_start = addr->addr;
> +
> +	/* Coalesce array_elements */
> +	while (cur->dma_seg_size - start < cur->remaining) {
> +		if (cur->dma_start + cur->dma_seg_size !=
> addr[1].addr ||
> +		    addr->proto != addr[1].proto)
> +			break;
> +		addr++;
> +		cur->dma_seg_size += PAGE_SIZE << addr->order;
> +	}
> +
> +	cur->dma_addr = addr;
> +	cur->start = start;
> +	cur->size = cur->dma_seg_size - start;
> +}
> +
>  /**
>   * xe_res_first_sg - initialize a xe_res_cursor with a scatter
> gather table
>   *
> @@ -160,11 +206,42 @@ static inline void xe_res_first_sg(const struct
> sg_table *sg,
>  	cur->start = start;
>  	cur->remaining = size;
>  	cur->size = 0;
> +	cur->dma_addr = NULL;
>  	cur->sgl = sg->sgl;
>  	cur->mem_type = XE_PL_TT;
>  	__xe_res_sg_next(cur);
>  }
>  
> +/**
> + * xe_res_first_dma - initialize a xe_res_cursor with dma_addr array
> + *
> + * @dma_addr: struct drm_pagemap_dma_addr array to walk
> + * @start: Start of the range
> + * @size: Size of the range
> + * @cur: cursor object to initialize
> + *
> + * Start walking over the range of allocations between @start and
> @size.
> + */
> +static inline void xe_res_first_dma(const struct
> drm_pagemap_dma_addr *dma_addr,
> +				    u64 start, u64 size,
> +				    struct xe_res_cursor *cur)
> +{
> +	XE_WARN_ON(!dma_addr);
> +	XE_WARN_ON(!IS_ALIGNED(start, PAGE_SIZE) ||
> +		   !IS_ALIGNED(size, PAGE_SIZE));
> +
> +	cur->node = NULL;
> +	cur->start = start;
> +	cur->remaining = size;
> +	cur->dma_seg_size = PAGE_SIZE << dma_addr->order;
> +	cur->dma_start = 0;
> +	cur->size = 0;
> +	cur->dma_addr = dma_addr;
> +	__xe_res_dma_next(cur);
> +	cur->sgl = NULL;
> +	cur->mem_type = XE_PL_TT;
> +}
> +
>  /**
>   * xe_res_next - advance the cursor
>   *
> @@ -191,6 +268,12 @@ static inline void xe_res_next(struct
> xe_res_cursor *cur, u64 size)
>  		return;
>  	}
>  
> +	if (cur->dma_addr) {
> +		cur->start += size;
> +		__xe_res_dma_next(cur);
> +		return;
> +	}
> +
>  	if (cur->sgl) {
>  		cur->start += size;
>  		__xe_res_sg_next(cur);
> @@ -232,6 +315,35 @@ static inline void xe_res_next(struct
> xe_res_cursor *cur, u64 size)
>   */
>  static inline u64 xe_res_dma(const struct xe_res_cursor *cur)
>  {
> -	return cur->sgl ? sg_dma_address(cur->sgl) + cur->start :
> cur->start;
> +	if (cur->dma_addr)
> +		return cur->dma_start + cur->start;
> +	else if (cur->sgl)
> +		return sg_dma_address(cur->sgl) + cur->start;
> +	else
> +		return cur->start;
> +}
> +
> +/**
> + * xe_res_is_vram() - Whether the cursor current dma address points
> to
> + * same-device VRAM
> + * @cur: The cursor.
> + *
> + * Return: true iff the address returned by xe_res_dma() points to
> internal vram.
> + */
> +static inline bool xe_res_is_vram(const struct xe_res_cursor *cur)
> +{
> +	if (cur->dma_addr)
> +		return cur->dma_addr->proto == XE_INTERCONNECT_VRAM;
> +
> +	switch (cur->mem_type) {
> +	case XE_PL_STOLEN:
> +	case XE_PL_VRAM0:
> +	case XE_PL_VRAM1:
> +		return true;
> +	default:
> +		break;
> +	}
> +
> +	return false;
>  }
>  #endif
> diff --git a/drivers/gpu/drm/xe/xe_svm.h
> b/drivers/gpu/drm/xe/xe_svm.h
> index 979f2322eeba..376e86876a11 100644
> --- a/drivers/gpu/drm/xe/xe_svm.h
> +++ b/drivers/gpu/drm/xe/xe_svm.h
> @@ -6,6 +6,10 @@
>  #ifndef _XE_SVM_H_
>  #define _XE_SVM_H_
>  
> +#include "drm_pagemap.h"
> +
> +#define XE_INTERCONNECT_VRAM DRM_INTERCONNECT_DRIVER
> +
>  struct xe_vm;
>  
>  int xe_svm_init(struct xe_vm *vm);
Matthew Brost Nov. 19, 2024, 4:24 p.m. UTC | #2
On Tue, Nov 19, 2024 at 01:15:12PM +0100, Thomas Hellström wrote:
> On Tue, 2024-10-15 at 20:24 -0700, Matthew Brost wrote:
> > From: Thomas Hellström <thomas.hellstrom@linux.intel.com>
> > 
> > Useful for SVM ranges in SRAM and programing page tables.
> 
> We should look at providing a better commit message.
> 

Yes, this is pretty poor. Will scrub all commit messages in next and
ensure higher quality.

Matt

> > 
> > Signed-off-by: Matthew Brost <matthew.brost@intel.com>
> > Signed-off-by: Thomas Hellström <thomas.hellstrom@linux.intel.com>
> > ---
> >  drivers/gpu/drm/xe/xe_res_cursor.h | 116
> > ++++++++++++++++++++++++++++-
> >  drivers/gpu/drm/xe/xe_svm.h        |   4 +
> >  2 files changed, 118 insertions(+), 2 deletions(-)
> > 
> > diff --git a/drivers/gpu/drm/xe/xe_res_cursor.h
> > b/drivers/gpu/drm/xe/xe_res_cursor.h
> > index dca374b6521c..3faa3d9adb82 100644
> > --- a/drivers/gpu/drm/xe/xe_res_cursor.h
> > +++ b/drivers/gpu/drm/xe/xe_res_cursor.h
> > @@ -30,13 +30,18 @@
> >  #include <drm/ttm/ttm_range_manager.h>
> >  #include <drm/ttm/ttm_resource.h>
> >  #include <drm/ttm/ttm_tt.h>
> > +#include "drm_pagemap.h"
> >  
> >  #include "xe_bo.h"
> >  #include "xe_device.h"
> >  #include "xe_macros.h"
> > +#include "xe_svm.h"
> >  #include "xe_ttm_vram_mgr.h"
> >  
> > -/* state back for walking over vram_mgr, stolen_mgr, and gtt_mgr
> > allocations */
> > +/**
> > + * struct xe_res_cursor - state for walking over vram_mgr,
> > stolen_mgr,
> > + * and gtt_mgr allocations
> > + */
> >  struct xe_res_cursor {
> >  	u64 start;
> >  	u64 size;
> > @@ -44,7 +49,17 @@ struct xe_res_cursor {
> >  	void *node;
> >  	u32 mem_type;
> >  	struct scatterlist *sgl;
> > +	/** @dma_addr: Current element in a struct
> > drm_pagemap_dma_addr array */
> > +	const struct drm_pagemap_dma_addr *dma_addr;
> >  	struct drm_buddy *mm;
> > +	/**
> > +	 * @dma_start: DMA start address for the current segment.
> > +	 * This may be different to @dma_addr.addr since elements in
> > +	 * the array may be coalesced to a single segment.
> > +	 */
> > +	u64 dma_start;
> > +	/** @dma_seg_size: Size of the current segment. */
> > +	u64 dma_seg_size;
> >  };
> >  
> >  static struct drm_buddy *xe_res_get_buddy(struct ttm_resource *res)
> > @@ -70,6 +85,7 @@ static inline void xe_res_first(struct ttm_resource
> > *res,
> >  				struct xe_res_cursor *cur)
> >  {
> >  	cur->sgl = NULL;
> > +	cur->dma_addr = NULL;
> >  	if (!res)
> >  		goto fallback;
> >  
> > @@ -141,6 +157,36 @@ static inline void __xe_res_sg_next(struct
> > xe_res_cursor *cur)
> >  	cur->sgl = sgl;
> >  }
> >  
> > +/**
> > + * __xe_res_dma_next() - Advance the cursor when end-of-segment is
> > reached
> > + * @cur: The cursor
> > + */
> > +static inline void __xe_res_dma_next(struct xe_res_cursor *cur)
> > +{
> > +	const struct drm_pagemap_dma_addr *addr = cur->dma_addr;
> > +	u64 start = cur->start;
> > +
> > +	while (start >= cur->dma_seg_size) {
> > +		start -= cur->dma_seg_size;
> > +		addr++;
> > +		cur->dma_seg_size = PAGE_SIZE << addr->order;
> > +	}
> > +	cur->dma_start = addr->addr;
> > +
> > +	/* Coalesce array_elements */
> > +	while (cur->dma_seg_size - start < cur->remaining) {
> > +		if (cur->dma_start + cur->dma_seg_size !=
> > addr[1].addr ||
> > +		    addr->proto != addr[1].proto)
> > +			break;
> > +		addr++;
> > +		cur->dma_seg_size += PAGE_SIZE << addr->order;
> > +	}
> > +
> > +	cur->dma_addr = addr;
> > +	cur->start = start;
> > +	cur->size = cur->dma_seg_size - start;
> > +}
> > +
> >  /**
> >   * xe_res_first_sg - initialize a xe_res_cursor with a scatter
> > gather table
> >   *
> > @@ -160,11 +206,42 @@ static inline void xe_res_first_sg(const struct
> > sg_table *sg,
> >  	cur->start = start;
> >  	cur->remaining = size;
> >  	cur->size = 0;
> > +	cur->dma_addr = NULL;
> >  	cur->sgl = sg->sgl;
> >  	cur->mem_type = XE_PL_TT;
> >  	__xe_res_sg_next(cur);
> >  }
> >  
> > +/**
> > + * xe_res_first_dma - initialize a xe_res_cursor with dma_addr array
> > + *
> > + * @dma_addr: struct drm_pagemap_dma_addr array to walk
> > + * @start: Start of the range
> > + * @size: Size of the range
> > + * @cur: cursor object to initialize
> > + *
> > + * Start walking over the range of allocations between @start and
> > @size.
> > + */
> > +static inline void xe_res_first_dma(const struct
> > drm_pagemap_dma_addr *dma_addr,
> > +				    u64 start, u64 size,
> > +				    struct xe_res_cursor *cur)
> > +{
> > +	XE_WARN_ON(!dma_addr);
> > +	XE_WARN_ON(!IS_ALIGNED(start, PAGE_SIZE) ||
> > +		   !IS_ALIGNED(size, PAGE_SIZE));
> > +
> > +	cur->node = NULL;
> > +	cur->start = start;
> > +	cur->remaining = size;
> > +	cur->dma_seg_size = PAGE_SIZE << dma_addr->order;
> > +	cur->dma_start = 0;
> > +	cur->size = 0;
> > +	cur->dma_addr = dma_addr;
> > +	__xe_res_dma_next(cur);
> > +	cur->sgl = NULL;
> > +	cur->mem_type = XE_PL_TT;
> > +}
> > +
> >  /**
> >   * xe_res_next - advance the cursor
> >   *
> > @@ -191,6 +268,12 @@ static inline void xe_res_next(struct
> > xe_res_cursor *cur, u64 size)
> >  		return;
> >  	}
> >  
> > +	if (cur->dma_addr) {
> > +		cur->start += size;
> > +		__xe_res_dma_next(cur);
> > +		return;
> > +	}
> > +
> >  	if (cur->sgl) {
> >  		cur->start += size;
> >  		__xe_res_sg_next(cur);
> > @@ -232,6 +315,35 @@ static inline void xe_res_next(struct
> > xe_res_cursor *cur, u64 size)
> >   */
> >  static inline u64 xe_res_dma(const struct xe_res_cursor *cur)
> >  {
> > -	return cur->sgl ? sg_dma_address(cur->sgl) + cur->start :
> > cur->start;
> > +	if (cur->dma_addr)
> > +		return cur->dma_start + cur->start;
> > +	else if (cur->sgl)
> > +		return sg_dma_address(cur->sgl) + cur->start;
> > +	else
> > +		return cur->start;
> > +}
> > +
> > +/**
> > + * xe_res_is_vram() - Whether the cursor current dma address points
> > to
> > + * same-device VRAM
> > + * @cur: The cursor.
> > + *
> > + * Return: true iff the address returned by xe_res_dma() points to
> > internal vram.
> > + */
> > +static inline bool xe_res_is_vram(const struct xe_res_cursor *cur)
> > +{
> > +	if (cur->dma_addr)
> > +		return cur->dma_addr->proto == XE_INTERCONNECT_VRAM;
> > +
> > +	switch (cur->mem_type) {
> > +	case XE_PL_STOLEN:
> > +	case XE_PL_VRAM0:
> > +	case XE_PL_VRAM1:
> > +		return true;
> > +	default:
> > +		break;
> > +	}
> > +
> > +	return false;
> >  }
> >  #endif
> > diff --git a/drivers/gpu/drm/xe/xe_svm.h
> > b/drivers/gpu/drm/xe/xe_svm.h
> > index 979f2322eeba..376e86876a11 100644
> > --- a/drivers/gpu/drm/xe/xe_svm.h
> > +++ b/drivers/gpu/drm/xe/xe_svm.h
> > @@ -6,6 +6,10 @@
> >  #ifndef _XE_SVM_H_
> >  #define _XE_SVM_H_
> >  
> > +#include "drm_pagemap.h"
> > +
> > +#define XE_INTERCONNECT_VRAM DRM_INTERCONNECT_DRIVER
> > +
> >  struct xe_vm;
> >  
> >  int xe_svm_init(struct xe_vm *vm);
>
diff mbox series

Patch

diff --git a/drivers/gpu/drm/xe/xe_res_cursor.h b/drivers/gpu/drm/xe/xe_res_cursor.h
index dca374b6521c..3faa3d9adb82 100644
--- a/drivers/gpu/drm/xe/xe_res_cursor.h
+++ b/drivers/gpu/drm/xe/xe_res_cursor.h
@@ -30,13 +30,18 @@ 
 #include <drm/ttm/ttm_range_manager.h>
 #include <drm/ttm/ttm_resource.h>
 #include <drm/ttm/ttm_tt.h>
+#include "drm_pagemap.h"
 
 #include "xe_bo.h"
 #include "xe_device.h"
 #include "xe_macros.h"
+#include "xe_svm.h"
 #include "xe_ttm_vram_mgr.h"
 
-/* state back for walking over vram_mgr, stolen_mgr, and gtt_mgr allocations */
+/**
+ * struct xe_res_cursor - state for walking over vram_mgr, stolen_mgr,
+ * and gtt_mgr allocations
+ */
 struct xe_res_cursor {
 	u64 start;
 	u64 size;
@@ -44,7 +49,17 @@  struct xe_res_cursor {
 	void *node;
 	u32 mem_type;
 	struct scatterlist *sgl;
+	/** @dma_addr: Current element in a struct drm_pagemap_dma_addr array */
+	const struct drm_pagemap_dma_addr *dma_addr;
 	struct drm_buddy *mm;
+	/**
+	 * @dma_start: DMA start address for the current segment.
+	 * This may be different to @dma_addr.addr since elements in
+	 * the array may be coalesced to a single segment.
+	 */
+	u64 dma_start;
+	/** @dma_seg_size: Size of the current segment. */
+	u64 dma_seg_size;
 };
 
 static struct drm_buddy *xe_res_get_buddy(struct ttm_resource *res)
@@ -70,6 +85,7 @@  static inline void xe_res_first(struct ttm_resource *res,
 				struct xe_res_cursor *cur)
 {
 	cur->sgl = NULL;
+	cur->dma_addr = NULL;
 	if (!res)
 		goto fallback;
 
@@ -141,6 +157,36 @@  static inline void __xe_res_sg_next(struct xe_res_cursor *cur)
 	cur->sgl = sgl;
 }
 
+/**
+ * __xe_res_dma_next() - Advance the cursor when end-of-segment is reached
+ * @cur: The cursor
+ */
+static inline void __xe_res_dma_next(struct xe_res_cursor *cur)
+{
+	const struct drm_pagemap_dma_addr *addr = cur->dma_addr;
+	u64 start = cur->start;
+
+	while (start >= cur->dma_seg_size) {
+		start -= cur->dma_seg_size;
+		addr++;
+		cur->dma_seg_size = PAGE_SIZE << addr->order;
+	}
+	cur->dma_start = addr->addr;
+
+	/* Coalesce array_elements */
+	while (cur->dma_seg_size - start < cur->remaining) {
+		if (cur->dma_start + cur->dma_seg_size != addr[1].addr ||
+		    addr->proto != addr[1].proto)
+			break;
+		addr++;
+		cur->dma_seg_size += PAGE_SIZE << addr->order;
+	}
+
+	cur->dma_addr = addr;
+	cur->start = start;
+	cur->size = cur->dma_seg_size - start;
+}
+
 /**
  * xe_res_first_sg - initialize a xe_res_cursor with a scatter gather table
  *
@@ -160,11 +206,42 @@  static inline void xe_res_first_sg(const struct sg_table *sg,
 	cur->start = start;
 	cur->remaining = size;
 	cur->size = 0;
+	cur->dma_addr = NULL;
 	cur->sgl = sg->sgl;
 	cur->mem_type = XE_PL_TT;
 	__xe_res_sg_next(cur);
 }
 
+/**
+ * xe_res_first_dma - initialize a xe_res_cursor with dma_addr array
+ *
+ * @dma_addr: struct drm_pagemap_dma_addr array to walk
+ * @start: Start of the range
+ * @size: Size of the range
+ * @cur: cursor object to initialize
+ *
+ * Start walking over the range of allocations between @start and @size.
+ */
+static inline void xe_res_first_dma(const struct drm_pagemap_dma_addr *dma_addr,
+				    u64 start, u64 size,
+				    struct xe_res_cursor *cur)
+{
+	XE_WARN_ON(!dma_addr);
+	XE_WARN_ON(!IS_ALIGNED(start, PAGE_SIZE) ||
+		   !IS_ALIGNED(size, PAGE_SIZE));
+
+	cur->node = NULL;
+	cur->start = start;
+	cur->remaining = size;
+	cur->dma_seg_size = PAGE_SIZE << dma_addr->order;
+	cur->dma_start = 0;
+	cur->size = 0;
+	cur->dma_addr = dma_addr;
+	__xe_res_dma_next(cur);
+	cur->sgl = NULL;
+	cur->mem_type = XE_PL_TT;
+}
+
 /**
  * xe_res_next - advance the cursor
  *
@@ -191,6 +268,12 @@  static inline void xe_res_next(struct xe_res_cursor *cur, u64 size)
 		return;
 	}
 
+	if (cur->dma_addr) {
+		cur->start += size;
+		__xe_res_dma_next(cur);
+		return;
+	}
+
 	if (cur->sgl) {
 		cur->start += size;
 		__xe_res_sg_next(cur);
@@ -232,6 +315,35 @@  static inline void xe_res_next(struct xe_res_cursor *cur, u64 size)
  */
 static inline u64 xe_res_dma(const struct xe_res_cursor *cur)
 {
-	return cur->sgl ? sg_dma_address(cur->sgl) + cur->start : cur->start;
+	if (cur->dma_addr)
+		return cur->dma_start + cur->start;
+	else if (cur->sgl)
+		return sg_dma_address(cur->sgl) + cur->start;
+	else
+		return cur->start;
+}
+
+/**
+ * xe_res_is_vram() - Whether the cursor current dma address points to
+ * same-device VRAM
+ * @cur: The cursor.
+ *
+ * Return: true iff the address returned by xe_res_dma() points to internal vram.
+ */
+static inline bool xe_res_is_vram(const struct xe_res_cursor *cur)
+{
+	if (cur->dma_addr)
+		return cur->dma_addr->proto == XE_INTERCONNECT_VRAM;
+
+	switch (cur->mem_type) {
+	case XE_PL_STOLEN:
+	case XE_PL_VRAM0:
+	case XE_PL_VRAM1:
+		return true;
+	default:
+		break;
+	}
+
+	return false;
 }
 #endif
diff --git a/drivers/gpu/drm/xe/xe_svm.h b/drivers/gpu/drm/xe/xe_svm.h
index 979f2322eeba..376e86876a11 100644
--- a/drivers/gpu/drm/xe/xe_svm.h
+++ b/drivers/gpu/drm/xe/xe_svm.h
@@ -6,6 +6,10 @@ 
 #ifndef _XE_SVM_H_
 #define _XE_SVM_H_
 
+#include "drm_pagemap.h"
+
+#define XE_INTERCONNECT_VRAM DRM_INTERCONNECT_DRIVER
+
 struct xe_vm;
 
 int xe_svm_init(struct xe_vm *vm);