@@ -87,6 +87,7 @@ config PPC
select ARCH_HAS_DMA_SET_COHERENT_MASK
select ARCH_HAS_ELF_RANDOMIZE
select ARCH_HAS_GCOV_PROFILE_ALL
+ select ARCH_HAS_PMEM_API
select ARCH_HAS_SCALED_CPUTIME if VIRT_CPU_ACCOUNTING_NATIVE
select ARCH_HAS_SG_CHAIN
select ARCH_HAS_TICK_BROADCAST if GENERIC_CLOCKEVENTS_BROADCAST
new file mode 100644
@@ -0,0 +1,109 @@
+/*
+ * Copyright(c) 2017 IBM Corporation. All rights reserved.
+ *
+ * Based on the x86 version.
+ *
+ * This program is free software; you can redistribute it and/or modify
+ * it under the terms of version 2 of the GNU General Public License as
+ * published by the Free Software Foundation.
+ *
+ * This program is distributed in the hope that it will be useful, but
+ * WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ * General Public License for more details.
+ */
+#ifndef __ASM_POWERPC_PMEM_H__
+#define __ASM_POWERPC_PMEM_H__
+
+#include <linux/uio.h>
+#include <linux/uaccess.h>
+#include <asm/cacheflush.h>
+
+/*
+ * See include/linux/pmem.h for API documentation
+ *
+ * PPC specific notes:
+ *
+ * 1. PPC has no non-temporal (cache bypassing) stores so we're stuck with
+ * doing cache writebacks.
+ *
+ * 2. DCBST is a suggestion. DCBF *will* force a writeback.
+ *
+ */
+
+static inline void arch_wb_cache_pmem(void *addr, size_t size)
+{
+ unsigned long iaddr = (unsigned long) addr;
+
+ /* NB: contains a barrier */
+ flush_inval_dcache_range(iaddr, iaddr + size);
+}
+
+/* invalidate and writeback are functionally identical */
+#define arch_invalidate_pmem arch_wb_cache_pmem
+
+static inline void arch_memcpy_to_pmem(void *dst, const void *src, size_t n)
+{
+ int unwritten;
+
+ /*
+ * We are copying between two kernel buffers, if
+ * __copy_from_user_inatomic_nocache() returns an error (page
+ * fault) we would have already reported a general protection fault
+ * before the WARN+BUG.
+ *
+ * XXX: replace this with a hand-rolled memcpy+dcbf
+ */
+ unwritten = __copy_from_user_inatomic(dst, (void __user *) src, n);
+ if (WARN(unwritten, "%s: fault copying %p <- %p unwritten: %d\n",
+ __func__, dst, src, unwritten))
+ BUG();
+
+ arch_wb_cache_pmem(dst, n);
+}
+
+static inline int arch_memcpy_from_pmem(void *dst, const void *src, size_t n)
+{
+ /*
+ * TODO: We should have most of the infrastructure for MCE handling
+ * but it needs to be made slightly smarter.
+ */
+ memcpy(dst, src, n);
+ return 0;
+}
+
+static inline size_t arch_copy_from_iter_pmem(void *addr, size_t bytes,
+ struct iov_iter *i)
+{
+ size_t len;
+
+ /* XXX: under what conditions would this return len < size? */
+ len = copy_from_iter(addr, bytes, i);
+ arch_wb_cache_pmem(addr, bytes - len);
+
+ return len;
+}
+
+static inline void arch_clear_pmem(void *addr, size_t size)
+{
+ void *start = addr;
+
+ /*
+ * XXX: A hand rolled dcbz+dcbf loop would probably be better.
+ */
+
+ if (((uintptr_t) addr & ~PAGE_MASK) == 0) {
+ while (size >= PAGE_SIZE) {
+ clear_page(addr);
+ addr += PAGE_SIZE;
+ size -= PAGE_SIZE;
+ }
+ }
+
+ if (size)
+ memset(addr, 0, size);
+
+ arch_wb_cache_pmem(start, size);
+}
+
+#endif /* __ASM_POWERPC_PMEM_H__ */
@@ -182,7 +182,7 @@ _GLOBAL(flush_dcache_phys_range)
isync
blr
-_GLOBAL(flush_inval_dcache_range)
+_GLOBAL_TOC(flush_inval_dcache_range)
ld r10,PPC64_CACHES@toc(r2)
lwz r7,DCACHEL1BLOCKSIZE(r10) /* Get dcache block size */
addi r5,r7,-1
Initial powerpc support for the arch-specific bit of the persistent memory API. Nothing fancy here. Signed-off-by: Oliver O'Halloran <oohall@gmail.com> --- arch/powerpc/Kconfig | 1 + arch/powerpc/include/asm/pmem.h | 109 ++++++++++++++++++++++++++++++++++++++++ arch/powerpc/kernel/misc_64.S | 2 +- 3 files changed, 111 insertions(+), 1 deletion(-) create mode 100644 arch/powerpc/include/asm/pmem.h