Message ID | 20200209212609.7928-11-jarkko.sakkinen@linux.intel.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | None | expand |
On 2020-02-09 22:25, Jarkko Sakkinen wrote: > Intel Software Guard eXtensions (SGX) is a set of CPU instructions that > can be used by applications to set aside private regions of code and > data. The code outside the SGX hosted software entity is disallowed to > access the memory inside the enclave enforced by the CPU. We call these > entities as enclaves. > > This commit implements a driver that provides an ioctl API to construct > and run enclaves. Enclaves are constructed from pages residing in > reserved physical memory areas. The contents of these pages can only be > accessed when they are mapped as part of an enclave, by a hardware > thread running inside the enclave. > > The starting state of an enclave consists of a fixed measured set of > pages that are copied to the EPC during the construction process by > using ENCLS leaf functions and Software Enclave Control Structure (SECS) > that defines the enclave properties. > > Enclave are constructed by using ENCLS leaf functions ECREATE, EADD and > EINIT. ECREATE initializes SECS, EADD copies pages from system memory to > the EPC and EINIT check a given signed measurement and moves the enclave > into a state ready for execution. > > An initialized enclave can only be accessed through special Thread Control > Structure (TCS) pages by using ENCLU (ring-3 only) leaf EENTER. This leaf > function converts a thread into enclave mode and continues the execution in > the offset defined by the TCS provided to EENTER. An enclave is exited > through syscall, exception, interrupts or by explicitly calling another > ENCLU leaf EEXIT. > > The permissions, which enclave page is added will set the limit for maximum > permissions that can be set for mmap() and mprotect(). This will > effectively allow to build different security schemes between producers and > consumers of enclaves. Later on we can increase granularity with LSM hooks > for page addition (i.e. for producers) and mapping of the enclave (i.e. for > consumers) > > Cc: linux-security-module@vger.kernel.org > Cc: Nathaniel McCallum <npmccallum@redhat.com> > Co-developed-by: Sean Christopherson <sean.j.christopherson@intel.com> > Signed-off-by: Sean Christopherson <sean.j.christopherson@intel.com> > Co-developed-by: Suresh Siddha <suresh.b.siddha@intel.com> > Signed-off-by: Suresh Siddha <suresh.b.siddha@intel.com> > Tested-by: Haitao Huang <haitao.huang@linux.intel.com> > Signed-off-by: Jarkko Sakkinen <jarkko.sakkinen@linux.intel.com> > --- > .../userspace-api/ioctl/ioctl-number.rst | 1 + > arch/x86/include/uapi/asm/sgx.h | 66 ++ > arch/x86/kernel/cpu/sgx/Makefile | 3 + > arch/x86/kernel/cpu/sgx/driver.c | 194 +++++ > arch/x86/kernel/cpu/sgx/driver.h | 30 + > arch/x86/kernel/cpu/sgx/encl.c | 329 +++++++++ > arch/x86/kernel/cpu/sgx/encl.h | 87 +++ > arch/x86/kernel/cpu/sgx/ioctl.c | 697 ++++++++++++++++++ > arch/x86/kernel/cpu/sgx/main.c | 12 +- > arch/x86/kernel/cpu/sgx/reclaim.c | 1 + > 10 files changed, 1419 insertions(+), 1 deletion(-) > create mode 100644 arch/x86/include/uapi/asm/sgx.h > create mode 100644 arch/x86/kernel/cpu/sgx/driver.c > create mode 100644 arch/x86/kernel/cpu/sgx/driver.h > create mode 100644 arch/x86/kernel/cpu/sgx/encl.c > create mode 100644 arch/x86/kernel/cpu/sgx/encl.h > create mode 100644 arch/x86/kernel/cpu/sgx/ioctl.c > > diff --git a/Documentation/userspace-api/ioctl/ioctl-number.rst b/Documentation/userspace-api/ioctl/ioctl-number.rst > index 2e91370dc159..1c54dd2704db 100644 > --- a/Documentation/userspace-api/ioctl/ioctl-number.rst > +++ b/Documentation/userspace-api/ioctl/ioctl-number.rst > @@ -321,6 +321,7 @@ Code Seq# Include File Comments > <mailto:tlewis@mindspring.com> > 0xA3 90-9F linux/dtlk.h > 0xA4 00-1F uapi/linux/tee.h Generic TEE subsystem > +0xA4 00-1F uapi/asm/sgx.h Intel SGX subsystem (a legit conflict as TEE and SGX do not co-exist) > 0xAA 00-3F linux/uapi/linux/userfaultfd.h > 0xAB 00-1F linux/nbd.h > 0xAC 00-1F linux/raw.h > diff --git a/arch/x86/include/uapi/asm/sgx.h b/arch/x86/include/uapi/asm/sgx.h > new file mode 100644 > index 000000000000..5edb08ab8fd0 > --- /dev/null > +++ b/arch/x86/include/uapi/asm/sgx.h > @@ -0,0 +1,66 @@ > +/* SPDX-License-Identifier: (GPL-2.0 OR BSD-3-Clause) WITH Linux-syscall-note */ > +/* > + * Copyright(c) 2016-19 Intel Corporation. > + */ > +#ifndef _UAPI_ASM_X86_SGX_H > +#define _UAPI_ASM_X86_SGX_H > + > +#include <linux/types.h> > +#include <linux/ioctl.h> > + > +/** > + * enum sgx_epage_flags - page control flags > + * %SGX_PAGE_MEASURE: Measure the page contents with a sequence of > + * ENCLS[EEXTEND] operations. > + */ > +enum sgx_page_flags { > + SGX_PAGE_MEASURE = 0x01, > +}; > + > +#define SGX_MAGIC 0xA4 > + > +#define SGX_IOC_ENCLAVE_CREATE \ > + _IOW(SGX_MAGIC, 0x00, struct sgx_enclave_create) > +#define SGX_IOC_ENCLAVE_ADD_PAGES \ > + _IOWR(SGX_MAGIC, 0x01, struct sgx_enclave_add_pages) > +#define SGX_IOC_ENCLAVE_INIT \ > + _IOW(SGX_MAGIC, 0x02, struct sgx_enclave_init) > + > +/** > + * struct sgx_enclave_create - parameter structure for the > + * %SGX_IOC_ENCLAVE_CREATE ioctl > + * @src: address for the SECS page data > + */ > +struct sgx_enclave_create { > + __u64 src; > +}; > + > +/** > + * struct sgx_enclave_add_pages - parameter structure for the > + * %SGX_IOC_ENCLAVE_ADD_PAGE ioctl > + * @src: start address for the page data > + * @offset: starting page offset > + * @length: length of the data (multiple of the page size) > + * @secinfo: address for the SECINFO data > + * @flags: page control flags > + * @count: number of bytes added (multiple of the page size) > + */ > +struct sgx_enclave_add_pages { > + __u64 src; > + __u64 offset; > + __u64 length; > + __u64 secinfo; > + __u64 flags; > + __u64 count; > +}; Compared to the last time I looked at the patch set, this API removes the ability to measure individual pages chunks. That is not acceptable. On 2019-10-11 16:37, Sean Christopherson wrote: > Hiding the 256-byte granualarity from userspace is a good idea as it's not > intrinsically tied to the SGX architecture and exists only because of > latency requirements. What do you mean by "it's not intrinsically tied to the SGX architecture"? This is a fundamental part of the SGX instruction set. This is the instruction definition from the SDM: "EEXTEND—Extend Uninitialized Enclave Measurement by 256 Bytes". The exact sequence of EADD/EEXTEND calls is part of the enclave hash. The OS mustn't arbitrarily restrict how an enclave may be loaded. If the enclave loader were to follows OS-specific restrictions, that would result in effectively different enclaves. Because of these interoperability concerns, 256-byte granularity *must* be exposed through the UAPI. Besides only partially measuring a page, there are some other fringe cases that are technically possible, although I haven't seen any toolchains that do that. These include not interleaving EADD and EEXTEND, not using logical ordering for the EEXTENDs, and call EEXTEND multiple times on the same chunk. Maximum interoperability would require supporting any EADD/EEXTEND sequence. Maybe we should just add an EEXTEND@offset ioctl? This would give fine-grained control when needed (one could set flags=0 in the add pages ioctl and interleave with EEXTEND as needed). If you're ok adding an EEXTEND ioctl I don't think this issue needs to block landing the driver in its current form, in which case: Tested-by: Jethro Beekman <jethro@fortanix.com> Sorry for being super late with this, I know you asked me for feedback about this specific point in October. However, I did previously mention several times that being able to measure individual 256-byte chunks is necessary. -- Jethro Beekman | Fortanix
On Thu, Feb 13, 2020 at 02:59:52PM +0100, Jethro Beekman wrote: > On 2020-02-09 22:25, Jarkko Sakkinen wrote: > > +/** > > + * struct sgx_enclave_add_pages - parameter structure for the > > + * %SGX_IOC_ENCLAVE_ADD_PAGE ioctl > > + * @src: start address for the page data > > + * @offset: starting page offset > > + * @length: length of the data (multiple of the page size) > > + * @secinfo: address for the SECINFO data > > + * @flags: page control flags > > + * @count: number of bytes added (multiple of the page size) > > + */ > > +struct sgx_enclave_add_pages { > > + __u64 src; > > + __u64 offset; > > + __u64 length; > > + __u64 secinfo; > > + __u64 flags; > > + __u64 count; > > +}; > > Compared to the last time I looked at the patch set, this API removes the > ability to measure individual pages chunks. That is not acceptable. Why is it not acceptable? E.g. what specific use case do you have that _requires_ on measuring partial 4k pages of an enclave? > On 2019-10-11 16:37, Sean Christopherson wrote: > > Hiding the 256-byte granualarity from userspace is a good idea as it's not > > intrinsically tied to the SGX architecture and exists only because of > > latency requirements. > > What do you mean by "it's not intrinsically tied to the SGX architecture"? > This is a fundamental part of the SGX instruction set. This is the > instruction definition from the SDM: "EEXTEND—Extend Uninitialized Enclave > Measurement by 256 Bytes". SGX fundamentally works at a 4k granularity. EEXTEND is special cased because extending the measurement is a slow operation, i.e. EEXTEND on more than 256 byte chunks, *with the current implementation*, would exceeded latency requirements, e.g. block interrupts for too long and hose the kernel. A future implementation of SGX could change the latency of extending the measurement, e.g. a different algorithm that is slower/faster, and so could introduce EEXTEND2 which would work at a different granularity than EEXTEND. EEXTEND could have avoided the latency problems via other methods, e.g. by being interruptible a la EINIT and/or by being restartable. But that ship has sailed, so to avoid future complication in the kernel's ABI we're proposing/advocating supporting only measuring at a 4k granularity. > The exact sequence of EADD/EEXTEND calls is part of the enclave hash. The OS > mustn't arbitrarily restrict how an enclave may be loaded. If the enclave It's not arbitrary, there are good reasons for wanting to work with 4k granularity. Regardless, there are many examples of the kernel arbitrarily restricting what can be done relative to what is physically possible in hardware. > loader were to follows OS-specific restrictions, that would result in > effectively different enclaves. Because of these interoperability concerns, > 256-byte granularity *must* be exposed through the UAPI. Interoperability with what? Other OSes? > Besides only partially measuring a page, there are some other fringe cases > that are technically possible, although I haven't seen any toolchains that do > that. These include not interleaving EADD and EEXTEND, not using logical > ordering for the EEXTENDs, and call EEXTEND multiple times on the same chunk. > Maximum interoperability would require supporting any EADD/EEXTEND sequence. Same interoperability question as above. > Maybe we should just add an EEXTEND@offset ioctl? This would give > fine-grained control when needed (one could set flags=0 in the add pages > ioctl and interleave with EEXTEND as needed). If you're ok adding an EEXTEND > ioctl I don't think this issue needs to block landing the driver in its > current form, in which case: We've also discussed an EEXTEND ioctl(), but ultimately couldn't come up with a use case that _required_ partial page measurement. > Tested-by: Jethro Beekman <jethro@fortanix.com> > > Sorry for being super late with this, I know you asked me for feedback about > this specific point in October. However, I did previously mention several > times that being able to measure individual 256-byte chunks is necessary. > > -- Jethro Beekman | Fortanix
On 2020-02-13 19:07, Sean Christopherson wrote: > On Thu, Feb 13, 2020 at 02:59:52PM +0100, Jethro Beekman wrote: >> On 2020-02-09 22:25, Jarkko Sakkinen wrote: >>> +/** >>> + * struct sgx_enclave_add_pages - parameter structure for the >>> + * %SGX_IOC_ENCLAVE_ADD_PAGE ioctl >>> + * @src: start address for the page data >>> + * @offset: starting page offset >>> + * @length: length of the data (multiple of the page size) >>> + * @secinfo: address for the SECINFO data >>> + * @flags: page control flags >>> + * @count: number of bytes added (multiple of the page size) >>> + */ >>> +struct sgx_enclave_add_pages { >>> + __u64 src; >>> + __u64 offset; >>> + __u64 length; >>> + __u64 secinfo; >>> + __u64 flags; >>> + __u64 count; >>> +}; >> >> Compared to the last time I looked at the patch set, this API removes the >> ability to measure individual pages chunks. That is not acceptable. > > Why is it not acceptable? E.g. what specific use case do you have that > _requires_ on measuring partial 4k pages of an enclave? The use case is someone gives me an enclave and I want to load it. If I don't load it exactly as the enclave author specified, the enclave hash will be different, and it won't work. >> On 2019-10-11 16:37, Sean Christopherson wrote: >>> Hiding the 256-byte granualarity from userspace is a good idea as it's not >>> intrinsically tied to the SGX architecture and exists only because of >>> latency requirements. >> >> What do you mean by "it's not intrinsically tied to the SGX architecture"? >> This is a fundamental part of the SGX instruction set. This is the >> instruction definition from the SDM: "EEXTEND—Extend Uninitialized Enclave >> Measurement by 256 Bytes". > > SGX fundamentally works at a 4k granularity. EEXTEND is special cased > because extending the measurement is a slow operation, i.e. EEXTEND on more > than 256 byte chunks, *with the current implementation*, would exceeded > latency requirements, e.g. block interrupts for too long and hose the > kernel. > > A future implementation of SGX could change the latency of extending the > measurement, e.g. a different algorithm that is slower/faster, and so could > introduce EEXTEND2 which would work at a different granularity than EEXTEND. > > EEXTEND could have avoided the latency problems via other methods, e.g. by > being interruptible a la EINIT and/or by being restartable. But that ship > has sailed, so to avoid future complication in the kernel's ABI we're > proposing/advocating supporting only measuring at a 4k granularity. It doesn't really matter what the reason for the current EEXTEND implementation is. It's there now in the ISA, it needs to be supported. If EEXTEND2 (or whatever) is added to the ISA, it will likely influence the enclave hash, so userspace would need to specify what instruction is used for measuring anyway. -- Jethro Beekman | Fortanix
On Fri, Feb 14, 2020 at 10:24:10AM +0100, Jethro Beekman wrote: > On 2020-02-13 19:07, Sean Christopherson wrote: > > On Thu, Feb 13, 2020 at 02:59:52PM +0100, Jethro Beekman wrote: > >> On 2020-02-09 22:25, Jarkko Sakkinen wrote: > >>> +/** > >>> + * struct sgx_enclave_add_pages - parameter structure for the > >>> + * %SGX_IOC_ENCLAVE_ADD_PAGE ioctl > >>> + * @src: start address for the page data > >>> + * @offset: starting page offset > >>> + * @length: length of the data (multiple of the page size) > >>> + * @secinfo: address for the SECINFO data > >>> + * @flags: page control flags > >>> + * @count: number of bytes added (multiple of the page size) > >>> + */ > >>> +struct sgx_enclave_add_pages { > >>> + __u64 src; > >>> + __u64 offset; > >>> + __u64 length; > >>> + __u64 secinfo; > >>> + __u64 flags; > >>> + __u64 count; > >>> +}; > >> > >> Compared to the last time I looked at the patch set, this API removes the > >> ability to measure individual pages chunks. That is not acceptable. > > > > Why is it not acceptable? E.g. what specific use case do you have that > > _requires_ on measuring partial 4k pages of an enclave? > > The use case is someone gives me an enclave and I want to load it. If I don't > load it exactly as the enclave author specified, the enclave hash will be > different, and it won't work. And if our ABI says "thou shall measure in 4k chunks", then it's an invalid enclave if its author generated MRENCLAVE using a different granularity.
> On Feb 14, 2020, at 9:11 AM, Sean Christopherson <sean.j.christopherson@intel.com> wrote: > > On Fri, Feb 14, 2020 at 10:24:10AM +0100, Jethro Beekman wrote: >>> On 2020-02-13 19:07, Sean Christopherson wrote: >>> On Thu, Feb 13, 2020 at 02:59:52PM +0100, Jethro Beekman wrote: >>>> On 2020-02-09 22:25, Jarkko Sakkinen wrote: >>>>> +/** >>>>> + * struct sgx_enclave_add_pages - parameter structure for the >>>>> + * %SGX_IOC_ENCLAVE_ADD_PAGE ioctl >>>>> + * @src: start address for the page data >>>>> + * @offset: starting page offset >>>>> + * @length: length of the data (multiple of the page size) >>>>> + * @secinfo: address for the SECINFO data >>>>> + * @flags: page control flags >>>>> + * @count: number of bytes added (multiple of the page size) >>>>> + */ >>>>> +struct sgx_enclave_add_pages { >>>>> + __u64 src; >>>>> + __u64 offset; >>>>> + __u64 length; >>>>> + __u64 secinfo; >>>>> + __u64 flags; >>>>> + __u64 count; >>>>> +}; >>>> >>>> Compared to the last time I looked at the patch set, this API removes the >>>> ability to measure individual pages chunks. That is not acceptable. >>> >>> Why is it not acceptable? E.g. what specific use case do you have that >>> _requires_ on measuring partial 4k pages of an enclave? >> >> The use case is someone gives me an enclave and I want to load it. If I don't >> load it exactly as the enclave author specified, the enclave hash will be >> different, and it won't work. > > And if our ABI says "thou shall measure in 4k chunks", then it's an invalid > enclave if its author generated MRENCLAVE using a different granularity. ISTM, unless there’s a particularly compelling reason, if an enclave is valid, we should be able to load it.
On Fri, Feb 14, 2020 at 09:40:00AM -0800, Andy Lutomirski wrote: > > > > On Feb 14, 2020, at 9:11 AM, Sean Christopherson <sean.j.christopherson@intel.com> wrote: > > > > On Fri, Feb 14, 2020 at 10:24:10AM +0100, Jethro Beekman wrote: > >>> On 2020-02-13 19:07, Sean Christopherson wrote: > >>> On Thu, Feb 13, 2020 at 02:59:52PM +0100, Jethro Beekman wrote: > >>>> On 2020-02-09 22:25, Jarkko Sakkinen wrote: > >>>>> +/** > >>>>> + * struct sgx_enclave_add_pages - parameter structure for the > >>>>> + * %SGX_IOC_ENCLAVE_ADD_PAGE ioctl > >>>>> + * @src: start address for the page data > >>>>> + * @offset: starting page offset > >>>>> + * @length: length of the data (multiple of the page size) > >>>>> + * @secinfo: address for the SECINFO data > >>>>> + * @flags: page control flags > >>>>> + * @count: number of bytes added (multiple of the page size) > >>>>> + */ > >>>>> +struct sgx_enclave_add_pages { > >>>>> + __u64 src; > >>>>> + __u64 offset; > >>>>> + __u64 length; > >>>>> + __u64 secinfo; > >>>>> + __u64 flags; > >>>>> + __u64 count; > >>>>> +}; > >>>> > >>>> Compared to the last time I looked at the patch set, this API removes the > >>>> ability to measure individual pages chunks. That is not acceptable. > >>> > >>> Why is it not acceptable? E.g. what specific use case do you have that > >>> _requires_ on measuring partial 4k pages of an enclave? > >> > >> The use case is someone gives me an enclave and I want to load it. If I don't > >> load it exactly as the enclave author specified, the enclave hash will be > >> different, and it won't work. > > > > And if our ABI says "thou shall measure in 4k chunks", then it's an invalid > > enclave if its author generated MRENCLAVE using a different granularity. > > ISTM, unless there’s a particularly compelling reason, if an enclave is > valid, we should be able to load it. That means we have to have a separate ioctl() for EEXTEND, otherwise we can't handle EADD[0]->EADD[1]->EADD[2]->EEXTEND[0]->EEXTEND[1]->EEXTEND[2]. I think we'd still want to keep the MEASURE flag for SGX_IOC_ENCLAVE_ADD_PAGE so that we can optimize EADD[0]->EEXTEND[0]->EADD[1]->EEXTEND[1].
On Thu, Feb 13, 2020 at 02:59:52PM +0100, Jethro Beekman wrote: > Besides only partially measuring a page, there are some other fringe > cases that are technically possible, although I haven't seen any > toolchains that do that. These include not interleaving EADD and > EEXTEND, not using logical ordering for the EEXTENDs, and call EEXTEND > multiple times on the same chunk. Maximum interoperability would > require supporting any EADD/EEXTEND sequence. The reason why EEXTEND deals with chunks is nothing to do with the granularity but just to amortize the algorithm. I did ask about this and this is the answer that I got. I think it is perfectly sane for Linux to define ABI here in at this level. > Tested-by: Jethro Beekman <jethro@fortanix.com> Thanks you. > Sorry for being super late with this, I know you asked me for feedback > about this specific point in October. However, I did previously > mention several times that being able to measure individual 256-byte > chunks is necessary. NP. /Jarkko
On Sat, Feb 15, 2020 at 09:32:34AM +0200, Jarkko Sakkinen wrote: > On Thu, Feb 13, 2020 at 02:59:52PM +0100, Jethro Beekman wrote: > > Besides only partially measuring a page, there are some other fringe > > cases that are technically possible, although I haven't seen any > > toolchains that do that. These include not interleaving EADD and > > EEXTEND, not using logical ordering for the EEXTENDs, and call EEXTEND > > multiple times on the same chunk. Maximum interoperability would > > require supporting any EADD/EEXTEND sequence. > > The reason why EEXTEND deals with chunks is nothing to do with the > granularity but just to amortize the algorithm. I did ask about this > and this is the answer that I got. I mean asked from people who have been involed in the development of this CPU feature. /Jarkko
> On Feb 14, 2020, at 9:52 AM, Sean Christopherson <sean.j.christopherson@intel.com> wrote: > > On Fri, Feb 14, 2020 at 09:40:00AM -0800, Andy Lutomirski wrote: >> >> >>>> On Feb 14, 2020, at 9:11 AM, Sean Christopherson <sean.j.christopherson@intel.com> wrote: >>> >>> On Fri, Feb 14, 2020 at 10:24:10AM +0100, Jethro Beekman wrote: >>>>> On 2020-02-13 19:07, Sean Christopherson wrote: >>>>> On Thu, Feb 13, 2020 at 02:59:52PM +0100, Jethro Beekman wrote: >>>>>> On 2020-02-09 22:25, Jarkko Sakkinen wrote: >>>>>>> +/** >>>>>>> + * struct sgx_enclave_add_pages - parameter structure for the >>>>>>> + * %SGX_IOC_ENCLAVE_ADD_PAGE ioctl >>>>>>> + * @src: start address for the page data >>>>>>> + * @offset: starting page offset >>>>>>> + * @length: length of the data (multiple of the page size) >>>>>>> + * @secinfo: address for the SECINFO data >>>>>>> + * @flags: page control flags >>>>>>> + * @count: number of bytes added (multiple of the page size) >>>>>>> + */ >>>>>>> +struct sgx_enclave_add_pages { >>>>>>> + __u64 src; >>>>>>> + __u64 offset; >>>>>>> + __u64 length; >>>>>>> + __u64 secinfo; >>>>>>> + __u64 flags; >>>>>>> + __u64 count; >>>>>>> +}; >>>>>> >>>>>> Compared to the last time I looked at the patch set, this API removes the >>>>>> ability to measure individual pages chunks. That is not acceptable. >>>>> >>>>> Why is it not acceptable? E.g. what specific use case do you have that >>>>> _requires_ on measuring partial 4k pages of an enclave? >>>> >>>> The use case is someone gives me an enclave and I want to load it. If I don't >>>> load it exactly as the enclave author specified, the enclave hash will be >>>> different, and it won't work. >>> >>> And if our ABI says "thou shall measure in 4k chunks", then it's an invalid >>> enclave if its author generated MRENCLAVE using a different granularity. >> >> ISTM, unless there’s a particularly compelling reason, if an enclave is >> valid, we should be able to load it. > > That means we have to have a separate ioctl() for EEXTEND, otherwise we > can't handle EADD[0]->EADD[1]->EADD[2]->EEXTEND[0]->EEXTEND[1]->EEXTEND[2]. > > I think we'd still want to keep the MEASURE flag for SGX_IOC_ENCLAVE_ADD_PAGE > so that we can optimize EADD[0]->EEXTEND[0]->EADD[1]->EEXTEND[1]. Seems reasonable to me. I suppose such as ioctl could also be added later if there’s a need.
On Fri, Feb 14, 2020 at 09:11:46AM -0800, Sean Christopherson wrote: Good morning to everyone, I hope the weekend is going well. > On Fri, Feb 14, 2020 at 10:24:10AM +0100, Jethro Beekman wrote: > > On 2020-02-13 19:07, Sean Christopherson wrote: > > > On Thu, Feb 13, 2020 at 02:59:52PM +0100, Jethro Beekman wrote: > > >> On 2020-02-09 22:25, Jarkko Sakkinen wrote: > > >>> +/** > > >>> + * struct sgx_enclave_add_pages - parameter structure for the > > >>> + * %SGX_IOC_ENCLAVE_ADD_PAGE ioctl > > >>> + * @src: start address for the page data > > >>> + * @offset: starting page offset > > >>> + * @length: length of the data (multiple of the page size) > > >>> + * @secinfo: address for the SECINFO data > > >>> + * @flags: page control flags > > >>> + * @count: number of bytes added (multiple of the page size) > > >>> + */ > > >>> +struct sgx_enclave_add_pages { > > >>> + __u64 src; > > >>> + __u64 offset; > > >>> + __u64 length; > > >>> + __u64 secinfo; > > >>> + __u64 flags; > > >>> + __u64 count; > > >>> +}; > > >> > > >> Compared to the last time I looked at the patch set, this API > > >> removes the ability to measure individual pages chunks. That is > > >> not acceptable. > > > > > > Why is it not acceptable? E.g. what specific use case do you > > > have that _requires_ on measuring partial 4k pages of an > > > enclave? > > > > The use case is someone gives me an enclave and I want to load > > it. If I don't load it exactly as the enclave author specified, > > the enclave hash will be different, and it won't work. > And if our ABI says "thou shall measure in 4k chunks", then it's an > invalid enclave if its author generated MRENCLAVE using a different > granularity. The enclave isn't invalid with respect to the hardware ISA or a potential application or business case need. It is only invalid with respect to how a small group of kernel developers have decided that runtime/application developers, and ultimately their users, should use hardware that they have purchased and own. Interestingly, the very antithesis of what started the open source movement. If Jethro/Fortanix have a business case for measuring partial pages, which incidentally he may not be able to divulge at this time, it seems the driver should support it if the hardware does. An interesting phenomenon is evolving with respect to Linux. With secure boot, kernel module signing, and now the lockdown patches; the major Linux vendors are in a position to use cryptographic constraints to limit what the general Linux user community has available to it, and particularly in the case of SGX, how the Linux application eco-system can evolve. I find this situation particularly fascinating. Intel has choreographed 30+ million dollars of capital investment in Fortanix in order to advance the development of an SGX software eco-system. Given who is authoring the driver, one would think that the Fortanix engineering desires/needs would be given careful consideration before the hardware capabilities are limited by the driver ABI, an ABI that will be subsequently cryptographically constrained from innovation. My apologies in advance for any intended or perceived indelicacies on these issues. Have a good weekend. Dr. Greg As always, Dr. Greg Wettstein, Ph.D, Worker IDfusion, LLC SGX secured infrastructure and 4206 N. 19th Ave. autonomously self-defensive platforms. Fargo, ND 58102 PH: 701-281-1686 EMAIL: greg@idfusion.net ------------------------------------------------------------------------------ "Snow removal teaches all the important elements of succesful corporate politics: 1.) Be the first one to work. 2.) Always signal your intentions before moving. 3.) Be damn sure you're driving something big enough to deal with anything that decides not to get out of your way." -- Dr. G.W. Wettstein Guerrilla Tactics for Corporate Survival
On Sat, Feb 15, 2020 at 08:56:54AM -0800, Andy Lutomirski wrote: > > > On Feb 14, 2020, at 9:52 AM, Sean Christopherson <sean.j.christopherson@intel.com> wrote: > > > > On Fri, Feb 14, 2020 at 09:40:00AM -0800, Andy Lutomirski wrote: > >> > >> > >>>> On Feb 14, 2020, at 9:11 AM, Sean Christopherson <sean.j.christopherson@intel.com> wrote: > >>> > >>> On Fri, Feb 14, 2020 at 10:24:10AM +0100, Jethro Beekman wrote: > >>>>> On 2020-02-13 19:07, Sean Christopherson wrote: > >>>>> On Thu, Feb 13, 2020 at 02:59:52PM +0100, Jethro Beekman wrote: > >>>>>> On 2020-02-09 22:25, Jarkko Sakkinen wrote: > >>>>>>> +/** > >>>>>>> + * struct sgx_enclave_add_pages - parameter structure for the > >>>>>>> + * %SGX_IOC_ENCLAVE_ADD_PAGE ioctl > >>>>>>> + * @src: start address for the page data > >>>>>>> + * @offset: starting page offset > >>>>>>> + * @length: length of the data (multiple of the page size) > >>>>>>> + * @secinfo: address for the SECINFO data > >>>>>>> + * @flags: page control flags > >>>>>>> + * @count: number of bytes added (multiple of the page size) > >>>>>>> + */ > >>>>>>> +struct sgx_enclave_add_pages { > >>>>>>> + __u64 src; > >>>>>>> + __u64 offset; > >>>>>>> + __u64 length; > >>>>>>> + __u64 secinfo; > >>>>>>> + __u64 flags; > >>>>>>> + __u64 count; > >>>>>>> +}; > >>>>>> > >>>>>> Compared to the last time I looked at the patch set, this API removes the > >>>>>> ability to measure individual pages chunks. That is not acceptable. > >>>>> > >>>>> Why is it not acceptable? E.g. what specific use case do you have that > >>>>> _requires_ on measuring partial 4k pages of an enclave? > >>>> > >>>> The use case is someone gives me an enclave and I want to load it. If I don't > >>>> load it exactly as the enclave author specified, the enclave hash will be > >>>> different, and it won't work. > >>> > >>> And if our ABI says "thou shall measure in 4k chunks", then it's an invalid > >>> enclave if its author generated MRENCLAVE using a different granularity. > >> > >> ISTM, unless there’s a particularly compelling reason, if an enclave is > >> valid, we should be able to load it. > > > > That means we have to have a separate ioctl() for EEXTEND, otherwise we > > can't handle EADD[0]->EADD[1]->EADD[2]->EEXTEND[0]->EEXTEND[1]->EEXTEND[2]. > > > > I think we'd still want to keep the MEASURE flag for SGX_IOC_ENCLAVE_ADD_PAGE > > so that we can optimize EADD[0]->EEXTEND[0]->EADD[1]->EEXTEND[1]. > > Seems reasonable to me. I suppose such as ioctl could also be added later if > there’s a need. Assuming you're referring to the separate EEXTEND ioctl()... Everyone: Is a dedicated EEXTEND ioctl() needed now, i.e. is there an existing or in-flight use case that would break by having only the MEASURE flag in ADD_PAGE? If the answer is 'no' from all parties, my preference would be to hold off on adding it until there is an actual end user. To be clear, I'm not against adding such an ioctl(), I just don't want to add code where the only user is a kernel selftest.
I ran our validation tests for the Open Enclave SDK against this patch set and came across a potential issue. On 2/9/20 1:25 PM, Jarkko Sakkinen wrote: > +/** > + * sgx_encl_may_map() - Check if a requested VMA mapping is allowed > + * @encl: an enclave > + * @start: lower bound of the address range, inclusive > + * @end: upper bound of the address range, exclusive > + * @vm_prot_bits: requested protections of the address range > + * > + * Iterate through the enclave pages contained within [@start, @end) to verify > + * the permissions requested by @vm_prot_bits do not exceed that of any enclave > + * page to be mapped. Page addresses that do not have an associated enclave > + * page are interpreted to zero permissions. > + * > + * Return: > + * 0 on success, > + * -EACCES if VMA permissions exceed enclave page permissions > + */ > +int sgx_encl_may_map(struct sgx_encl *encl, unsigned long start, > + unsigned long end, unsigned long vm_prot_bits) > +{ > + unsigned long idx, idx_start, idx_end; > + struct sgx_encl_page *page; > + > + /* PROT_NONE always succeeds. */ > + if (!vm_prot_bits) > + return 0; > + > + idx_start = PFN_DOWN(start); > + idx_end = PFN_DOWN(end - 1); > + > + for (idx = idx_start; idx <= idx_end; ++idx) { > + mutex_lock(&encl->lock); > + page = radix_tree_lookup(&encl->page_tree, idx); > + mutex_unlock(&encl->lock); > + > + if (!page || (~page->vm_max_prot_bits & vm_prot_bits)) > + return -EACCES; > + } > + > + return 0; > +} > +static struct sgx_encl_page *sgx_encl_page_alloc(struct sgx_encl *encl, > + unsigned long offset, > + u64 secinfo_flags) > +{ > + struct sgx_encl_page *encl_page; > + unsigned long prot; > + > + encl_page = kzalloc(sizeof(*encl_page), GFP_KERNEL); > + if (!encl_page) > + return ERR_PTR(-ENOMEM); > + > + encl_page->desc = encl->base + offset; > + encl_page->encl = encl; > + > + prot = _calc_vm_trans(secinfo_flags, SGX_SECINFO_R, PROT_READ) | > + _calc_vm_trans(secinfo_flags, SGX_SECINFO_W, PROT_WRITE) | > + _calc_vm_trans(secinfo_flags, SGX_SECINFO_X, PROT_EXEC); > + > + /* > + * TCS pages must always RW set for CPU access while the SECINFO > + * permissions are *always* zero - the CPU ignores the user provided > + * values and silently overwrites them with zero permissions. > + */ > + if ((secinfo_flags & SGX_SECINFO_PAGE_TYPE_MASK) == SGX_SECINFO_TCS) > + prot |= PROT_READ | PROT_WRITE; > + > + /* Calculate maximum of the VM flags for the page. */ > + encl_page->vm_max_prot_bits = calc_vm_prot_bits(prot, 0); During mprotect (in mm/mprotect.c line 525) the following checks if READ_IMPLIES_EXECUTE and a PROT_READ is being requested. If so and VM_MAYEXEC is set, it also adds PROT_EXEC to the request. if (rier && (vma->vm_flags & VM_MAYEXEC)) prot |= PROT_EXEC; But if we look at sgx_encl_page_alloc(), we see vm_max_prot_bits is set without taking VM_MAYEXEC into account: encl_page->vm_max_prot_bits = calc_vm_prot_bits(prot, 0); sgx_encl_may_map() checks that the requested protection can be added with: if (!page || (~page->vm_max_prot_bits & vm_prot_bits)) return -EACCESS This means that for any process where READ_IMPLIES_EXECUTE is set and page where (vma->vm_flags & VM_MAYEXEC) == true, mmap/mprotect calls to that request PROT_READ on a page that was not added with PROT_EXEC will fail. - Jordan
On Tue, Feb 18, 2020 at 07:26:31PM -0800, Jordan Hand wrote: > I ran our validation tests for the Open Enclave SDK against this patch > set and came across a potential issue. > > On 2/9/20 1:25 PM, Jarkko Sakkinen wrote: > > +/** > > + * sgx_encl_may_map() - Check if a requested VMA mapping is allowed > > + * @encl: an enclave > > + * @start: lower bound of the address range, inclusive > > + * @end: upper bound of the address range, exclusive > > + * @vm_prot_bits: requested protections of the address range > > + * > > + * Iterate through the enclave pages contained within [@start, @end) to verify > > + * the permissions requested by @vm_prot_bits do not exceed that of any enclave > > + * page to be mapped. Page addresses that do not have an associated enclave > > + * page are interpreted to zero permissions. > > + * > > + * Return: > > + * 0 on success, > > + * -EACCES if VMA permissions exceed enclave page permissions > > + */ > > +int sgx_encl_may_map(struct sgx_encl *encl, unsigned long start, > > + unsigned long end, unsigned long vm_prot_bits) > > +{ > > + unsigned long idx, idx_start, idx_end; > > + struct sgx_encl_page *page; > > + > > + /* PROT_NONE always succeeds. */ > > + if (!vm_prot_bits) > > + return 0; > > + > > + idx_start = PFN_DOWN(start); > > + idx_end = PFN_DOWN(end - 1); > > + > > + for (idx = idx_start; idx <= idx_end; ++idx) { > > + mutex_lock(&encl->lock); > > + page = radix_tree_lookup(&encl->page_tree, idx); > > + mutex_unlock(&encl->lock); > > + > > + if (!page || (~page->vm_max_prot_bits & vm_prot_bits)) > > + return -EACCES; > > + } > > + > > + return 0; > > +} > > +static struct sgx_encl_page *sgx_encl_page_alloc(struct sgx_encl *encl, > > + unsigned long offset, > > + u64 secinfo_flags) > > +{ > > + struct sgx_encl_page *encl_page; > > + unsigned long prot; > > + > > + encl_page = kzalloc(sizeof(*encl_page), GFP_KERNEL); > > + if (!encl_page) > > + return ERR_PTR(-ENOMEM); > > + > > + encl_page->desc = encl->base + offset; > > + encl_page->encl = encl; > > + > > + prot = _calc_vm_trans(secinfo_flags, SGX_SECINFO_R, PROT_READ) | > > + _calc_vm_trans(secinfo_flags, SGX_SECINFO_W, PROT_WRITE) | > > + _calc_vm_trans(secinfo_flags, SGX_SECINFO_X, PROT_EXEC); > > + > > + /* > > + * TCS pages must always RW set for CPU access while the SECINFO > > + * permissions are *always* zero - the CPU ignores the user provided > > + * values and silently overwrites them with zero permissions. > > + */ > > + if ((secinfo_flags & SGX_SECINFO_PAGE_TYPE_MASK) == SGX_SECINFO_TCS) > > + prot |= PROT_READ | PROT_WRITE; > > + > > + /* Calculate maximum of the VM flags for the page. */ > > + encl_page->vm_max_prot_bits = calc_vm_prot_bits(prot, 0); > > During mprotect (in mm/mprotect.c line 525) the following checks if > READ_IMPLIES_EXECUTE and a PROT_READ is being requested. If so and > VM_MAYEXEC is set, it also adds PROT_EXEC to the request. > > if (rier && (vma->vm_flags & VM_MAYEXEC)) > prot |= PROT_EXEC; > > But if we look at sgx_encl_page_alloc(), we see vm_max_prot_bits is set > without taking VM_MAYEXEC into account: > > encl_page->vm_max_prot_bits = calc_vm_prot_bits(prot, 0); > > sgx_encl_may_map() checks that the requested protection can be added with: > > if (!page || (~page->vm_max_prot_bits & vm_prot_bits)) > return -EACCESS > > This means that for any process where READ_IMPLIES_EXECUTE is set and > page where (vma->vm_flags & VM_MAYEXEC) == true, mmap/mprotect calls to > that request PROT_READ on a page that was not added with PROT_EXEC will > fail. I could've sworn this was discussed on the SGX list at one point, but apparently we only discussed it internally. Anyways... More than likely, the READ_IMPLIES_EXECUTE (RIE) crud rears its head because part of the enclave loader is written in assembly. Unless explicitly told otherwise, the linker assumes that any program with assembly code may need an executable stack, which leads to the RIE personality being set for the process. Here's a fantastic write up for more details: https://www.airs.com/blog/archives/518 There are essentially two paths we can take: 1) Exempt EPC pages from RIE during mmap()/mprotect(), i.e. don't add PROT_EXEC for enclaves. 2) Punt the issue to userspace. Option (1) is desirable in some ways: - Enclaves will get an executable stack if and only if the loader/creator intentionally configures it to have an executable stack. - Separates enclaves from the personality of the loader. - Userspace doesn't have to do anything for the common case of not wanting an executable stack for its enclaves. The big down side to (1) is that it'd require an ugly hook in architecture agnostic code. And arguably, it reduces the overall security of the platform (more below). For (2), userspace has a few options: a) Tell the linker the enclave loader doesn't need RIE, either via a .note in assembly files or via the global "-z noexecstack" flag. b) Spawn a separate process to run/map the enclave if the enclave loader needs RIE. c) Require enclaves to allow PROT_EXEC on all pages. Note, this is an absolutely terrible idea and only included for completeness. As shown by the lack of a mmap()/mprotect() hook in this series to squash RIE, we chose option (2). Given that enclave loaders are not legacy code and hopefully following decent coding practices, option (2a) should suffice for all loaders. The security benefit mentioned above is that forcing enclave loaders to squash RIE eliminates an exectuable stack as an attack vector on the loader. If for some reason a loader "needs" an executable stack, requiring such a beast to jump through a few hoops to run sane enclaves doesn't seem too onerous. This obviously needs to be called out in the kernel docs.
On 2/20/20 10:13 AM, Sean Christopherson wrote: > On Tue, Feb 18, 2020 at 07:26:31PM -0800, Jordan Hand wrote: >> During mprotect (in mm/mprotect.c line 525) the following checks if >> READ_IMPLIES_EXECUTE and a PROT_READ is being requested. If so and >> VM_MAYEXEC is set, it also adds PROT_EXEC to the request. >> >> if (rier && (vma->vm_flags & VM_MAYEXEC)) >> prot |= PROT_EXEC; >> >> But if we look at sgx_encl_page_alloc(), we see vm_max_prot_bits is set >> without taking VM_MAYEXEC into account: >> >> encl_page->vm_max_prot_bits = calc_vm_prot_bits(prot, 0); >> >> sgx_encl_may_map() checks that the requested protection can be added with: >> >> if (!page || (~page->vm_max_prot_bits & vm_prot_bits)) >> return -EACCESS >> >> This means that for any process where READ_IMPLIES_EXECUTE is set and >> page where (vma->vm_flags & VM_MAYEXEC) == true, mmap/mprotect calls to >> that request PROT_READ on a page that was not added with PROT_EXEC will >> fail. > > I could've sworn this was discussed on the SGX list at one point, but > apparently we only discussed it internally. Anyways... > > More than likely, the READ_IMPLIES_EXECUTE (RIE) crud rears its head > because part of the enclave loader is written in assembly. Unless > explicitly told otherwise, the linker assumes that any program with > assembly code may need an executable stack, which leads to the RIE > personality being set for the process. Here's a fantastic write up for > more details: https://www.airs.com/blog/archives/518 > > There are essentially two paths we can take: > > 1) Exempt EPC pages from RIE during mmap()/mprotect(), i.e. don't add > PROT_EXEC for enclaves. > > 2) Punt the issue to userspace. > > Option (1) is desirable in some ways: > > - Enclaves will get an executable stack if and only if the loader/creator > intentionally configures it to have an executable stack. > > - Separates enclaves from the personality of the loader. > > - Userspace doesn't have to do anything for the common case of not > wanting an executable stack for its enclaves. > > The big down side to (1) is that it'd require an ugly hook in architecture > agnostic code. And arguably, it reduces the overall security of the > platform (more below). > > For (2), userspace has a few options: > > a) Tell the linker the enclave loader doesn't need RIE, either via a .note > in assembly files or via the global "-z noexecstack" flag. > > b) Spawn a separate process to run/map the enclave if the enclave loader > needs RIE. > > c) Require enclaves to allow PROT_EXEC on all pages. Note, this is an > absolutely terrible idea and only included for completeness. > > As shown by the lack of a mmap()/mprotect() hook in this series to squash > RIE, we chose option (2). Given that enclave loaders are not legacy code > and hopefully following decent coding practices, option (2a) should suffice > for all loaders. The security benefit mentioned above is that forcing > enclave loaders to squash RIE eliminates an exectuable stack as an attack > vector on the loader. I see your point and I do agree that there are security benefits to (2a) and I think we could do that for our loader. That said, it does concern me that this breaks perfectly valid userspace behavior. If a userspace process decides to use RIE, I don't know that the SGX driver should disobey that decision. So option (3) would be to just honor RIE for enclave pages and when page permissions are set to PROT_READ in sgx_encl_page_alloc and RIE is set, also add PROT_EXEC. I understand your concerns that this using RIE is bad security practice and I'm not convinced that (3) is the way to go, but from a philosophy perspective I don't know that the kernel should be in the business of stopping userspace from doing valid things. If option (3) can't/shouldn't be done for some reason, option (1) at least keeps from breaking expected userspace behavior. But I do agree that (1) is ugly to implement. -Jordan
On Thu, Feb 20, 2020 at 10:33:36AM -0800, Jordan Hand wrote: > On 2/20/20 10:13 AM, Sean Christopherson wrote: > > There are essentially two paths we can take: > > > > 1) Exempt EPC pages from RIE during mmap()/mprotect(), i.e. don't add > > PROT_EXEC for enclaves. > > > > 2) Punt the issue to userspace. > > > > Option (1) is desirable in some ways: > > > > - Enclaves will get an executable stack if and only if the loader/creator > > intentionally configures it to have an executable stack. > > > > - Separates enclaves from the personality of the loader. > > > > - Userspace doesn't have to do anything for the common case of not > > wanting an executable stack for its enclaves. > > > > The big down side to (1) is that it'd require an ugly hook in architecture > > agnostic code. And arguably, it reduces the overall security of the > > platform (more below). > > > > For (2), userspace has a few options: > > > > a) Tell the linker the enclave loader doesn't need RIE, either via a .note > > in assembly files or via the global "-z noexecstack" flag. > > > > b) Spawn a separate process to run/map the enclave if the enclave loader > > needs RIE. > > > > c) Require enclaves to allow PROT_EXEC on all pages. Note, this is an > > absolutely terrible idea and only included for completeness. > > > > As shown by the lack of a mmap()/mprotect() hook in this series to squash > > RIE, we chose option (2). Given that enclave loaders are not legacy code > > and hopefully following decent coding practices, option (2a) should suffice > > for all loaders. The security benefit mentioned above is that forcing > > enclave loaders to squash RIE eliminates an exectuable stack as an attack > > vector on the loader. > > I see your point and I do agree that there are security benefits to (2a) > and I think we could do that for our loader. That said, it does concern > me that this breaks perfectly valid userspace behavior. If a userspace > process decides to use RIE, I don't know that the SGX driver should > disobey that decision. > > So option (3) would be to just honor RIE for enclave pages and when page > permissions are set to PROT_READ in sgx_encl_page_alloc and RIE is set, > also add PROT_EXEC. Ah, right, IIRC that idea also came up in our internal discussions. Note, SGX would need to implement this option by checking for RIE in sgx_encl_may_map(), as the process that built the enclave may not be the same process that is running the enclave. > I understand your concerns that this using RIE is bad security practice > and I'm not convinced that (3) is the way to go, but from a philosophy > perspective I don't know that the kernel should be in the business of > stopping userspace from doing valid things. > > If option (3) can't/shouldn't be done for some reason, option (1) at > least keeps from breaking expected userspace behavior. But I do agree > that (1) is ugly to implement. My biggest concern for allowing PROT_EXEC if RIE is that it would result in #PF(SGX) (#GP on Skylake) due to an EPCM violation if the enclave actually tried to execute from such a page. This isn't a problem for the kernel as the fault will be reported cleanly through the vDSO (or get delivered as a SIGSEGV if the enclave isn't entered through the vDSO), but it's a bit weird for userspace as userspace will see the #PF(SGX) and likely assume the EPC was lost, e.g. silently restart the enclave instead of logging an error that the enclave is broken.
On Thu, Feb 20, 2020 at 10:13 AM Sean Christopherson <sean.j.christopherson@intel.com> wrote: > > On Tue, Feb 18, 2020 at 07:26:31PM -0800, Jordan Hand wrote: > > I ran our validation tests for the Open Enclave SDK against this patch > > set and came across a potential issue. > > > > On 2/9/20 1:25 PM, Jarkko Sakkinen wrote: > > > +/** > > > + * sgx_encl_may_map() - Check if a requested VMA mapping is allowed > > > + * @encl: an enclave > > > + * @start: lower bound of the address range, inclusive > > > + * @end: upper bound of the address range, exclusive > > > + * @vm_prot_bits: requested protections of the address range > > > + * > > > + * Iterate through the enclave pages contained within [@start, @end) to verify > > > + * the permissions requested by @vm_prot_bits do not exceed that of any enclave > > > + * page to be mapped. Page addresses that do not have an associated enclave > > > + * page are interpreted to zero permissions. > > > + * > > > + * Return: > > > + * 0 on success, > > > + * -EACCES if VMA permissions exceed enclave page permissions > > > + */ > > > +int sgx_encl_may_map(struct sgx_encl *encl, unsigned long start, > > > + unsigned long end, unsigned long vm_prot_bits) > > > +{ > > > + unsigned long idx, idx_start, idx_end; > > > + struct sgx_encl_page *page; > > > + > > > + /* PROT_NONE always succeeds. */ > > > + if (!vm_prot_bits) > > > + return 0; > > > + > > > + idx_start = PFN_DOWN(start); > > > + idx_end = PFN_DOWN(end - 1); > > > + > > > + for (idx = idx_start; idx <= idx_end; ++idx) { > > > + mutex_lock(&encl->lock); > > > + page = radix_tree_lookup(&encl->page_tree, idx); > > > + mutex_unlock(&encl->lock); > > > + > > > + if (!page || (~page->vm_max_prot_bits & vm_prot_bits)) > > > + return -EACCES; > > > + } > > > + > > > + return 0; > > > +} > > > +static struct sgx_encl_page *sgx_encl_page_alloc(struct sgx_encl *encl, > > > + unsigned long offset, > > > + u64 secinfo_flags) > > > +{ > > > + struct sgx_encl_page *encl_page; > > > + unsigned long prot; > > > + > > > + encl_page = kzalloc(sizeof(*encl_page), GFP_KERNEL); > > > + if (!encl_page) > > > + return ERR_PTR(-ENOMEM); > > > + > > > + encl_page->desc = encl->base + offset; > > > + encl_page->encl = encl; > > > + > > > + prot = _calc_vm_trans(secinfo_flags, SGX_SECINFO_R, PROT_READ) | > > > + _calc_vm_trans(secinfo_flags, SGX_SECINFO_W, PROT_WRITE) | > > > + _calc_vm_trans(secinfo_flags, SGX_SECINFO_X, PROT_EXEC); > > > + > > > + /* > > > + * TCS pages must always RW set for CPU access while the SECINFO > > > + * permissions are *always* zero - the CPU ignores the user provided > > > + * values and silently overwrites them with zero permissions. > > > + */ > > > + if ((secinfo_flags & SGX_SECINFO_PAGE_TYPE_MASK) == SGX_SECINFO_TCS) > > > + prot |= PROT_READ | PROT_WRITE; > > > + > > > + /* Calculate maximum of the VM flags for the page. */ > > > + encl_page->vm_max_prot_bits = calc_vm_prot_bits(prot, 0); > > > > During mprotect (in mm/mprotect.c line 525) the following checks if > > READ_IMPLIES_EXECUTE and a PROT_READ is being requested. If so and > > VM_MAYEXEC is set, it also adds PROT_EXEC to the request. > > > > if (rier && (vma->vm_flags & VM_MAYEXEC)) > > prot |= PROT_EXEC; > > > > But if we look at sgx_encl_page_alloc(), we see vm_max_prot_bits is set > > without taking VM_MAYEXEC into account: > > > > encl_page->vm_max_prot_bits = calc_vm_prot_bits(prot, 0); > > > > sgx_encl_may_map() checks that the requested protection can be added with: > > > > if (!page || (~page->vm_max_prot_bits & vm_prot_bits)) > > return -EACCESS > > > > This means that for any process where READ_IMPLIES_EXECUTE is set and > > page where (vma->vm_flags & VM_MAYEXEC) == true, mmap/mprotect calls to > > that request PROT_READ on a page that was not added with PROT_EXEC will > > fail. > > I could've sworn this was discussed on the SGX list at one point, but > apparently we only discussed it internally. Anyways... > > More than likely, the READ_IMPLIES_EXECUTE (RIE) crud rears its head > because part of the enclave loader is written in assembly. Unless > explicitly told otherwise, the linker assumes that any program with > assembly code may need an executable stack, which leads to the RIE > personality being set for the process. Here's a fantastic write up for > more details: https://www.airs.com/blog/archives/518 > > There are essentially two paths we can take: > > 1) Exempt EPC pages from RIE during mmap()/mprotect(), i.e. don't add > PROT_EXEC for enclaves. Seems reasonable. Honestly, it probably makes sense to try to exempt almost everything from RIE. I'd be a bit surprised if RIE is actually useful for anything other than plain anonymous pages and private file mappings.
On Thu, Feb 20, 2020 at 10:51:37AM -0800, Andy Lutomirski wrote: > On Thu, Feb 20, 2020 at 10:13 AM Sean Christopherson > <sean.j.christopherson@intel.com> wrote: > > More than likely, the READ_IMPLIES_EXECUTE (RIE) crud rears its head > > because part of the enclave loader is written in assembly. Unless > > explicitly told otherwise, the linker assumes that any program with > > assembly code may need an executable stack, which leads to the RIE > > personality being set for the process. Here's a fantastic write up for > > more details: https://www.airs.com/blog/archives/518 > > > > There are essentially two paths we can take: > > > > 1) Exempt EPC pages from RIE during mmap()/mprotect(), i.e. don't add > > PROT_EXEC for enclaves. > > Seems reasonable. > > Honestly, it probably makes sense to try to exempt almost everything > from RIE. I'd be a bit surprised if RIE is actually useful for > anything other than plain anonymous pages and private file mappings. Hmm, last I looked at this I was focused on adding a generic protections manipulator, e.g. vm_ops->mprotect_adjust() and f_op->???, and I thought those options were too ugly to pursue. But if we want to start killing RIE specifically, adding a boolean flag to and f_op wouldn't be _that_ heinous, e.g. static int do_mprotect_pkey(...) { ... /* Does the application expect PROT_READ to imply PROT_EXEC */ if (rier && (vma->vm_flags & VM_MAYEXEC) && (!vma->vm_file || !vma->vm_file->f_op->no_read_implies_exec)) prot |= PROT_EXEC; } unsigned long do_mmap(...) { if ((prot & PROT_READ) && (current->personality & READ_IMPLIES_EXEC)) if (!file || (!path_noexec(&file->f_path) && !file->f_op->no_read_implies_exec)) prot |= PROT_EXEC; }
On Tue, Feb 18, 2020 at 07:26:31PM -0800, Jordan Hand wrote: > if (!page || (~page->vm_max_prot_bits & vm_prot_bits)) > return -EACCESS > > This means that for any process where READ_IMPLIES_EXECUTE is set and > page where (vma->vm_flags & VM_MAYEXEC) == true, mmap/mprotect calls to > that request PROT_READ on a page that was not added with PROT_EXEC will > fail. Right. You would end up requesting RX from a R region. And you are suggesting that we tweak it along the lines of to make RIE processes work: unsigned long max_prot_bits = page->vm_max_prot_bits; if (!!(current->personality & READ_IMPLIES_EXEC) && vma->vm_flags & VM_MAY_EXEC) max_prot_bits |= VM_EXEC; /* ... */ if (!page || (~max_prot_bits & vm_prot_bits)) return -EACCESS ? /Jarkko
On Thu, Feb 20, 2020 at 10:48:42AM -0800, Sean Christopherson wrote: > My biggest concern for allowing PROT_EXEC if RIE is that it would result > in #PF(SGX) (#GP on Skylake) due to an EPCM violation if the enclave > actually tried to execute from such a page. This isn't a problem for the > kernel as the fault will be reported cleanly through the vDSO (or get > delivered as a SIGSEGV if the enclave isn't entered through the vDSO), but > it's a bit weird for userspace as userspace will see the #PF(SGX) and > likely assume the EPC was lost, e.g. silently restart the enclave instead > of logging an error that the enclave is broken. I think right way to fix the current implementation is to -EACCES mmap() (and mprotect) when !!(current->personality & READ_IMPLIES_EXEC). This way supporting RIE can be reconsidered later on without any potential ABI bottlenecks. /Jarkko
On 2/20/20 2:16 PM, Jarkko Sakkinen wrote: > On Thu, Feb 20, 2020 at 10:48:42AM -0800, Sean Christopherson wrote: >> My biggest concern for allowing PROT_EXEC if RIE is that it would result >> in #PF(SGX) (#GP on Skylake) due to an EPCM violation if the enclave >> actually tried to execute from such a page. This isn't a problem for the >> kernel as the fault will be reported cleanly through the vDSO (or get >> delivered as a SIGSEGV if the enclave isn't entered through the vDSO), but >> it's a bit weird for userspace as userspace will see the #PF(SGX) and >> likely assume the EPC was lost, e.g. silently restart the enclave instead >> of logging an error that the enclave is broken. > > I think right way to fix the current implementation is to -EACCES mmap() > (and mprotect) when !!(current->personality & READ_IMPLIES_EXEC). > I agree. It still means userspace code with an executable stack can't mmap/mprotect enclave pages and request PROT_READ but the check you've proposed would more consistently enforce this which is easier to understand from userspace perspective. -Jordan
> On Feb 20, 2020, at 2:16 PM, Jarkko Sakkinen <jarkko.sakkinen@linux.intel.com> wrote: > > On Thu, Feb 20, 2020 at 10:48:42AM -0800, Sean Christopherson wrote: >> My biggest concern for allowing PROT_EXEC if RIE is that it would result >> in #PF(SGX) (#GP on Skylake) due to an EPCM violation if the enclave >> actually tried to execute from such a page. This isn't a problem for the >> kernel as the fault will be reported cleanly through the vDSO (or get >> delivered as a SIGSEGV if the enclave isn't entered through the vDSO), but >> it's a bit weird for userspace as userspace will see the #PF(SGX) and >> likely assume the EPC was lost, e.g. silently restart the enclave instead >> of logging an error that the enclave is broken. > > I think right way to fix the current implementation is to -EACCES mmap() > (and mprotect) when !!(current->personality & READ_IMPLIES_EXEC). > > This way supporting RIE can be reconsidered later on without any > potential ABI bottlenecks. > Sounds good to me. I see no credible reason why anyone would use RIE and SGX. > /Jarkko
On Thu, Feb 20, 2020 at 04:11:04PM -0800, Jordan Hand wrote: > On 2/20/20 2:16 PM, Jarkko Sakkinen wrote: > > On Thu, Feb 20, 2020 at 10:48:42AM -0800, Sean Christopherson wrote: > >> My biggest concern for allowing PROT_EXEC if RIE is that it would result > >> in #PF(SGX) (#GP on Skylake) due to an EPCM violation if the enclave > >> actually tried to execute from such a page. This isn't a problem for the > >> kernel as the fault will be reported cleanly through the vDSO (or get > >> delivered as a SIGSEGV if the enclave isn't entered through the vDSO), but > >> it's a bit weird for userspace as userspace will see the #PF(SGX) and > >> likely assume the EPC was lost, e.g. silently restart the enclave instead > >> of logging an error that the enclave is broken. > > > > I think right way to fix the current implementation is to -EACCES mmap() > > (and mprotect) when !!(current->personality & READ_IMPLIES_EXEC). > > > > I agree. It still means userspace code with an executable stack can't > mmap/mprotect enclave pages and request PROT_READ but the check you've > proposed would more consistently enforce this which is easier to > understand from userspace perspective. Thank you. Your observation was really important because having half working RIE support hanging around would only have potential to cause unnecessary maintenance burden. It would even make adding a legit RIE support later on somewhat more difficult. I updated the commit under discussion in my tree [*] with a fix that adds the following to the beginning of sgx_encl_may_map(): /* * Disallow RIE tasks as their VMA permissions might conflict with the * enclave page permissions. */ if (!!(current->personality & READ_IMPLIES_EXEC)) return -EACCES; [*] https://github.com/jsakkine-intel/linux-sgx.git /Jarkko
On Thu, Feb 20, 2020 at 04:32:22PM -0800, Andy Lutomirski wrote: > > On Feb 20, 2020, at 2:16 PM, Jarkko Sakkinen <jarkko.sakkinen@linux.intel.com> wrote: > > > > On Thu, Feb 20, 2020 at 10:48:42AM -0800, Sean Christopherson wrote: > >> My biggest concern for allowing PROT_EXEC if RIE is that it would result > >> in #PF(SGX) (#GP on Skylake) due to an EPCM violation if the enclave > >> actually tried to execute from such a page. This isn't a problem for the > >> kernel as the fault will be reported cleanly through the vDSO (or get > >> delivered as a SIGSEGV if the enclave isn't entered through the vDSO), but > >> it's a bit weird for userspace as userspace will see the #PF(SGX) and > >> likely assume the EPC was lost, e.g. silently restart the enclave instead > >> of logging an error that the enclave is broken. > > > > I think right way to fix the current implementation is to -EACCES mmap() > > (and mprotect) when !!(current->personality & READ_IMPLIES_EXEC). > > > > This way supporting RIE can be reconsidered later on without any > > potential ABI bottlenecks. > > > > Sounds good to me. I see no credible reason why anyone would use RIE and SGX. Great, thanks Andy. /Jarkko
diff --git a/Documentation/userspace-api/ioctl/ioctl-number.rst b/Documentation/userspace-api/ioctl/ioctl-number.rst index 2e91370dc159..1c54dd2704db 100644 --- a/Documentation/userspace-api/ioctl/ioctl-number.rst +++ b/Documentation/userspace-api/ioctl/ioctl-number.rst @@ -321,6 +321,7 @@ Code Seq# Include File Comments <mailto:tlewis@mindspring.com> 0xA3 90-9F linux/dtlk.h 0xA4 00-1F uapi/linux/tee.h Generic TEE subsystem +0xA4 00-1F uapi/asm/sgx.h Intel SGX subsystem (a legit conflict as TEE and SGX do not co-exist) 0xAA 00-3F linux/uapi/linux/userfaultfd.h 0xAB 00-1F linux/nbd.h 0xAC 00-1F linux/raw.h diff --git a/arch/x86/include/uapi/asm/sgx.h b/arch/x86/include/uapi/asm/sgx.h new file mode 100644 index 000000000000..5edb08ab8fd0 --- /dev/null +++ b/arch/x86/include/uapi/asm/sgx.h @@ -0,0 +1,66 @@ +/* SPDX-License-Identifier: (GPL-2.0 OR BSD-3-Clause) WITH Linux-syscall-note */ +/* + * Copyright(c) 2016-19 Intel Corporation. + */ +#ifndef _UAPI_ASM_X86_SGX_H +#define _UAPI_ASM_X86_SGX_H + +#include <linux/types.h> +#include <linux/ioctl.h> + +/** + * enum sgx_epage_flags - page control flags + * %SGX_PAGE_MEASURE: Measure the page contents with a sequence of + * ENCLS[EEXTEND] operations. + */ +enum sgx_page_flags { + SGX_PAGE_MEASURE = 0x01, +}; + +#define SGX_MAGIC 0xA4 + +#define SGX_IOC_ENCLAVE_CREATE \ + _IOW(SGX_MAGIC, 0x00, struct sgx_enclave_create) +#define SGX_IOC_ENCLAVE_ADD_PAGES \ + _IOWR(SGX_MAGIC, 0x01, struct sgx_enclave_add_pages) +#define SGX_IOC_ENCLAVE_INIT \ + _IOW(SGX_MAGIC, 0x02, struct sgx_enclave_init) + +/** + * struct sgx_enclave_create - parameter structure for the + * %SGX_IOC_ENCLAVE_CREATE ioctl + * @src: address for the SECS page data + */ +struct sgx_enclave_create { + __u64 src; +}; + +/** + * struct sgx_enclave_add_pages - parameter structure for the + * %SGX_IOC_ENCLAVE_ADD_PAGE ioctl + * @src: start address for the page data + * @offset: starting page offset + * @length: length of the data (multiple of the page size) + * @secinfo: address for the SECINFO data + * @flags: page control flags + * @count: number of bytes added (multiple of the page size) + */ +struct sgx_enclave_add_pages { + __u64 src; + __u64 offset; + __u64 length; + __u64 secinfo; + __u64 flags; + __u64 count; +}; + +/** + * struct sgx_enclave_init - parameter structure for the + * %SGX_IOC_ENCLAVE_INIT ioctl + * @sigstruct: address for the SIGSTRUCT data + */ +struct sgx_enclave_init { + __u64 sigstruct; +}; + +#endif /* _UAPI_ASM_X86_SGX_H */ diff --git a/arch/x86/kernel/cpu/sgx/Makefile b/arch/x86/kernel/cpu/sgx/Makefile index 2dec75916a5e..f8d32da3a67a 100644 --- a/arch/x86/kernel/cpu/sgx/Makefile +++ b/arch/x86/kernel/cpu/sgx/Makefile @@ -1,3 +1,6 @@ obj-y += \ + driver.o \ + encl.o \ + ioctl.o \ main.o \ reclaim.o diff --git a/arch/x86/kernel/cpu/sgx/driver.c b/arch/x86/kernel/cpu/sgx/driver.c new file mode 100644 index 000000000000..b4aa7b9f8376 --- /dev/null +++ b/arch/x86/kernel/cpu/sgx/driver.c @@ -0,0 +1,194 @@ +// SPDX-License-Identifier: (GPL-2.0 OR BSD-3-Clause) +// Copyright(c) 2016-18 Intel Corporation. + +#include <linux/acpi.h> +#include <linux/miscdevice.h> +#include <linux/mman.h> +#include <linux/security.h> +#include <linux/suspend.h> +#include <asm/traps.h> +#include "driver.h" +#include "encl.h" + +MODULE_DESCRIPTION("Intel SGX Enclave Driver"); +MODULE_AUTHOR("Jarkko Sakkinen <jarkko.sakkinen@linux.intel.com>"); +MODULE_LICENSE("Dual BSD/GPL"); + +u64 sgx_encl_size_max_32; +u64 sgx_encl_size_max_64; +u32 sgx_misc_reserved_mask; +u64 sgx_attributes_reserved_mask; +u64 sgx_xfrm_reserved_mask = ~0x3; +u32 sgx_xsave_size_tbl[64]; + +static int sgx_open(struct inode *inode, struct file *file) +{ + struct sgx_encl *encl; + int ret; + + encl = kzalloc(sizeof(*encl), GFP_KERNEL); + if (!encl) + return -ENOMEM; + + atomic_set(&encl->flags, 0); + kref_init(&encl->refcount); + INIT_RADIX_TREE(&encl->page_tree, GFP_KERNEL); + mutex_init(&encl->lock); + INIT_LIST_HEAD(&encl->mm_list); + spin_lock_init(&encl->mm_lock); + + ret = init_srcu_struct(&encl->srcu); + if (ret) { + kfree(encl); + return ret; + } + + file->private_data = encl; + + return 0; +} + +static int sgx_release(struct inode *inode, struct file *file) +{ + struct sgx_encl *encl = file->private_data; + struct sgx_encl_mm *encl_mm; + + for ( ; ; ) { + spin_lock(&encl->mm_lock); + + if (list_empty(&encl->mm_list)) { + encl_mm = NULL; + } else { + encl_mm = list_first_entry(&encl->mm_list, + struct sgx_encl_mm, list); + list_del_rcu(&encl_mm->list); + } + + spin_unlock(&encl->mm_lock); + + /* The list is empty, ready to go. */ + if (!encl_mm) + break; + + synchronize_srcu(&encl->srcu); + mmu_notifier_unregister(&encl_mm->mmu_notifier, encl_mm->mm); + kfree(encl_mm); + }; + + mutex_lock(&encl->lock); + atomic_or(SGX_ENCL_DEAD, &encl->flags); + mutex_unlock(&encl->lock); + + kref_put(&encl->refcount, sgx_encl_release); + return 0; +} + +#ifdef CONFIG_COMPAT +static long sgx_compat_ioctl(struct file *filep, unsigned int cmd, + unsigned long arg) +{ + return sgx_ioctl(filep, cmd, arg); +} +#endif + +static int sgx_mmap(struct file *file, struct vm_area_struct *vma) +{ + struct sgx_encl *encl = file->private_data; + int ret; + + ret = sgx_encl_may_map(encl, vma->vm_start, vma->vm_end, + vma->vm_flags & (VM_READ | VM_WRITE | VM_EXEC)); + if (ret) + return ret; + + ret = sgx_encl_mm_add(encl, vma->vm_mm); + if (ret) + return ret; + + vma->vm_ops = &sgx_vm_ops; + vma->vm_flags |= VM_PFNMAP | VM_DONTEXPAND | VM_DONTDUMP | VM_IO; + vma->vm_private_data = encl; + + return 0; +} + +static unsigned long sgx_get_unmapped_area(struct file *file, + unsigned long addr, + unsigned long len, + unsigned long pgoff, + unsigned long flags) +{ + if (flags & MAP_PRIVATE) + return -EINVAL; + + if (flags & MAP_FIXED) + return addr; + + return current->mm->get_unmapped_area(file, addr, len, pgoff, flags); +} + +static const struct file_operations sgx_encl_fops = { + .owner = THIS_MODULE, + .open = sgx_open, + .release = sgx_release, + .unlocked_ioctl = sgx_ioctl, +#ifdef CONFIG_COMPAT + .compat_ioctl = sgx_compat_ioctl, +#endif + .mmap = sgx_mmap, + .get_unmapped_area = sgx_get_unmapped_area, +}; + +const struct file_operations sgx_provision_fops = { + .owner = THIS_MODULE, +}; + +static struct miscdevice sgx_dev_enclave = { + .minor = MISC_DYNAMIC_MINOR, + .name = "enclave", + .nodename = "sgx/enclave", + .fops = &sgx_encl_fops, +}; + +int __init sgx_drv_init(void) +{ + unsigned int eax, ebx, ecx, edx; + u64 attr_mask, xfrm_mask; + int ret; + int i; + + if (!boot_cpu_has(X86_FEATURE_SGX_LC)) { + pr_info("The public key MSRs are not writable.\n"); + return -ENODEV; + } + + cpuid_count(SGX_CPUID, 0, &eax, &ebx, &ecx, &edx); + sgx_misc_reserved_mask = ~ebx | SGX_MISC_RESERVED_MASK; + sgx_encl_size_max_64 = 1ULL << ((edx >> 8) & 0xFF); + sgx_encl_size_max_32 = 1ULL << (edx & 0xFF); + + cpuid_count(SGX_CPUID, 1, &eax, &ebx, &ecx, &edx); + + attr_mask = (((u64)ebx) << 32) + (u64)eax; + sgx_attributes_reserved_mask = ~attr_mask | SGX_ATTR_RESERVED_MASK; + + if (boot_cpu_has(X86_FEATURE_OSXSAVE)) { + xfrm_mask = (((u64)edx) << 32) + (u64)ecx; + + for (i = 2; i < 64; i++) { + cpuid_count(0x0D, i, &eax, &ebx, &ecx, &edx); + if ((1 << i) & xfrm_mask) + sgx_xsave_size_tbl[i] = eax + ebx; + } + + sgx_xfrm_reserved_mask = ~xfrm_mask; + } + + ret = misc_register(&sgx_dev_enclave); + if (ret) { + pr_err("Creating /dev/sgx/enclave failed with %d.\n", ret); + return ret; + } + + return 0; +} diff --git a/arch/x86/kernel/cpu/sgx/driver.h b/arch/x86/kernel/cpu/sgx/driver.h new file mode 100644 index 000000000000..e4063923115b --- /dev/null +++ b/arch/x86/kernel/cpu/sgx/driver.h @@ -0,0 +1,30 @@ +/* SPDX-License-Identifier: (GPL-2.0 OR BSD-3-Clause) */ +#ifndef __ARCH_SGX_DRIVER_H__ +#define __ARCH_SGX_DRIVER_H__ + +#include <crypto/hash.h> +#include <linux/kref.h> +#include <linux/mmu_notifier.h> +#include <linux/radix-tree.h> +#include <linux/rwsem.h> +#include <linux/sched.h> +#include <linux/workqueue.h> +#include <uapi/asm/sgx.h> +#include "sgx.h" + +#define SGX_EINIT_SPIN_COUNT 20 +#define SGX_EINIT_SLEEP_COUNT 50 +#define SGX_EINIT_SLEEP_TIME 20 + +extern u64 sgx_encl_size_max_32; +extern u64 sgx_encl_size_max_64; +extern u32 sgx_misc_reserved_mask; +extern u64 sgx_attributes_reserved_mask; +extern u64 sgx_xfrm_reserved_mask; +extern u32 sgx_xsave_size_tbl[64]; + +long sgx_ioctl(struct file *filep, unsigned int cmd, unsigned long arg); + +int sgx_drv_init(void); + +#endif /* __ARCH_X86_SGX_DRIVER_H__ */ diff --git a/arch/x86/kernel/cpu/sgx/encl.c b/arch/x86/kernel/cpu/sgx/encl.c new file mode 100644 index 000000000000..cd2b8dbb0eca --- /dev/null +++ b/arch/x86/kernel/cpu/sgx/encl.c @@ -0,0 +1,329 @@ +// SPDX-License-Identifier: (GPL-2.0 OR BSD-3-Clause) +// Copyright(c) 2016-18 Intel Corporation. + +#include <linux/lockdep.h> +#include <linux/mm.h> +#include <linux/mman.h> +#include <linux/shmem_fs.h> +#include <linux/suspend.h> +#include <linux/sched/mm.h> +#include "arch.h" +#include "encl.h" +#include "sgx.h" + +static struct sgx_encl_page *sgx_encl_load_page(struct sgx_encl *encl, + unsigned long addr) +{ + struct sgx_encl_page *entry; + unsigned int flags; + + /* If process was forked, VMA is still there but vm_private_data is set + * to NULL. + */ + if (!encl) + return ERR_PTR(-EFAULT); + + flags = atomic_read(&encl->flags); + + if ((flags & SGX_ENCL_DEAD) || !(flags & SGX_ENCL_INITIALIZED)) + return ERR_PTR(-EFAULT); + + entry = radix_tree_lookup(&encl->page_tree, addr >> PAGE_SHIFT); + if (!entry) + return ERR_PTR(-EFAULT); + + /* Page is already resident in the EPC. */ + if (entry->epc_page) + return entry; + + return ERR_PTR(-EFAULT); +} + +static void sgx_mmu_notifier_release(struct mmu_notifier *mn, + struct mm_struct *mm) +{ + struct sgx_encl_mm *encl_mm = + container_of(mn, struct sgx_encl_mm, mmu_notifier); + struct sgx_encl_mm *tmp = NULL; + + /* + * The enclave itself can remove encl_mm. Note, objects can't be moved + * off an RCU protected list, but deletion is ok. + */ + spin_lock(&encl_mm->encl->mm_lock); + list_for_each_entry(tmp, &encl_mm->encl->mm_list, list) { + if (tmp == encl_mm) { + list_del_rcu(&encl_mm->list); + break; + } + } + spin_unlock(&encl_mm->encl->mm_lock); + + if (tmp == encl_mm) { + synchronize_srcu(&encl_mm->encl->srcu); + mmu_notifier_put(mn); + } +} + +static void sgx_mmu_notifier_free(struct mmu_notifier *mn) +{ + struct sgx_encl_mm *encl_mm = + container_of(mn, struct sgx_encl_mm, mmu_notifier); + + kfree(encl_mm); +} + +static const struct mmu_notifier_ops sgx_mmu_notifier_ops = { + .release = sgx_mmu_notifier_release, + .free_notifier = sgx_mmu_notifier_free, +}; + +static struct sgx_encl_mm *sgx_encl_find_mm(struct sgx_encl *encl, + struct mm_struct *mm) +{ + struct sgx_encl_mm *encl_mm = NULL; + struct sgx_encl_mm *tmp; + int idx; + + idx = srcu_read_lock(&encl->srcu); + + list_for_each_entry_rcu(tmp, &encl->mm_list, list) { + if (tmp->mm == mm) { + encl_mm = tmp; + break; + } + } + + srcu_read_unlock(&encl->srcu, idx); + + return encl_mm; +} + +int sgx_encl_mm_add(struct sgx_encl *encl, struct mm_struct *mm) +{ + struct sgx_encl_mm *encl_mm; + int ret; + + if (atomic_read(&encl->flags) & SGX_ENCL_DEAD) + return -EINVAL; + + /* + * mm_structs are kept on mm_list until the mm or the enclave dies, + * i.e. once an mm is off the list, it's gone for good, therefore it's + * impossible to get a false positive on @mm due to a stale mm_list. + */ + if (sgx_encl_find_mm(encl, mm)) + return 0; + + encl_mm = kzalloc(sizeof(*encl_mm), GFP_KERNEL); + if (!encl_mm) + return -ENOMEM; + + encl_mm->encl = encl; + encl_mm->mm = mm; + encl_mm->mmu_notifier.ops = &sgx_mmu_notifier_ops; + + ret = __mmu_notifier_register(&encl_mm->mmu_notifier, mm); + if (ret) { + kfree(encl_mm); + return ret; + } + + spin_lock(&encl->mm_lock); + list_add_rcu(&encl_mm->list, &encl->mm_list); + spin_unlock(&encl->mm_lock); + + synchronize_srcu(&encl->srcu); + + return 0; +} + +static void sgx_vma_open(struct vm_area_struct *vma) +{ + struct sgx_encl *encl = vma->vm_private_data; + + if (!encl) + return; + + if (sgx_encl_mm_add(encl, vma->vm_mm)) + vma->vm_private_data = NULL; +} + +static unsigned int sgx_vma_fault(struct vm_fault *vmf) +{ + unsigned long addr = (unsigned long)vmf->address; + struct vm_area_struct *vma = vmf->vma; + struct sgx_encl *encl = vma->vm_private_data; + struct sgx_encl_page *entry; + int ret = VM_FAULT_NOPAGE; + unsigned long pfn; + + if (!encl) + return VM_FAULT_SIGBUS; + + mutex_lock(&encl->lock); + + entry = sgx_encl_load_page(encl, addr); + if (IS_ERR(entry)) { + if (unlikely(PTR_ERR(entry) != -EBUSY)) + ret = VM_FAULT_SIGBUS; + + goto out; + } + + if (!follow_pfn(vma, addr, &pfn)) + goto out; + + ret = vmf_insert_pfn(vma, addr, PFN_DOWN(entry->epc_page->desc)); + if (ret != VM_FAULT_NOPAGE) { + ret = VM_FAULT_SIGBUS; + goto out; + } + +out: + mutex_unlock(&encl->lock); + return ret; +} + +/** + * sgx_encl_may_map() - Check if a requested VMA mapping is allowed + * @encl: an enclave + * @start: lower bound of the address range, inclusive + * @end: upper bound of the address range, exclusive + * @vm_prot_bits: requested protections of the address range + * + * Iterate through the enclave pages contained within [@start, @end) to verify + * the permissions requested by @vm_prot_bits do not exceed that of any enclave + * page to be mapped. Page addresses that do not have an associated enclave + * page are interpreted to zero permissions. + * + * Return: + * 0 on success, + * -EACCES if VMA permissions exceed enclave page permissions + */ +int sgx_encl_may_map(struct sgx_encl *encl, unsigned long start, + unsigned long end, unsigned long vm_prot_bits) +{ + unsigned long idx, idx_start, idx_end; + struct sgx_encl_page *page; + + /* PROT_NONE always succeeds. */ + if (!vm_prot_bits) + return 0; + + idx_start = PFN_DOWN(start); + idx_end = PFN_DOWN(end - 1); + + for (idx = idx_start; idx <= idx_end; ++idx) { + mutex_lock(&encl->lock); + page = radix_tree_lookup(&encl->page_tree, idx); + mutex_unlock(&encl->lock); + + if (!page || (~page->vm_max_prot_bits & vm_prot_bits)) + return -EACCES; + } + + return 0; +} + +static int sgx_vma_mprotect(struct vm_area_struct *vma, unsigned long start, + unsigned long end, unsigned long prot) +{ + return sgx_encl_may_map(vma->vm_private_data, start, end, + calc_vm_prot_bits(prot, 0)); +} + +const struct vm_operations_struct sgx_vm_ops = { + .open = sgx_vma_open, + .fault = sgx_vma_fault, + .may_mprotect = sgx_vma_mprotect, +}; + +/** + * sgx_encl_find - find an enclave + * @mm: mm struct of the current process + * @addr: address in the ELRANGE + * @vma: the resulting VMA + * + * Find an enclave identified by the given address. Give back a VMA that is + * part of the enclave and located in that address. The VMA is given back if it + * is a proper enclave VMA even if an &sgx_encl instance does not exist yet + * (enclave creation has not been performed). + * + * Return: + * 0 on success, + * -EINVAL if an enclave was not found, + * -ENOENT if the enclave has not been created yet + */ +int sgx_encl_find(struct mm_struct *mm, unsigned long addr, + struct vm_area_struct **vma) +{ + struct vm_area_struct *result; + struct sgx_encl *encl; + + result = find_vma(mm, addr); + if (!result || result->vm_ops != &sgx_vm_ops || addr < result->vm_start) + return -EINVAL; + + encl = result->vm_private_data; + *vma = result; + + return encl ? 0 : -ENOENT; +} + +/** + * sgx_encl_destroy() - destroy enclave resources + * @encl: an &sgx_encl instance + */ +void sgx_encl_destroy(struct sgx_encl *encl) +{ + struct sgx_encl_page *entry; + struct radix_tree_iter iter; + void **slot; + + atomic_or(SGX_ENCL_DEAD, &encl->flags); + + radix_tree_for_each_slot(slot, &encl->page_tree, &iter, 0) { + entry = *slot; + + if (entry->epc_page) { + sgx_free_page(entry->epc_page); + encl->secs_child_cnt--; + entry->epc_page = NULL; + } + + radix_tree_delete(&entry->encl->page_tree, + PFN_DOWN(entry->desc)); + kfree(entry); + } + + if (!encl->secs_child_cnt && encl->secs.epc_page) { + sgx_free_page(encl->secs.epc_page); + encl->secs.epc_page = NULL; + } +} + +/** + * sgx_encl_release - Destroy an enclave instance + * @kref: address of a kref inside &sgx_encl + * + * Used together with kref_put(). Frees all the resources associated with the + * enclave and the instance itself. + */ +void sgx_encl_release(struct kref *ref) +{ + struct sgx_encl *encl = container_of(ref, struct sgx_encl, refcount); + + sgx_encl_destroy(encl); + + if (encl->backing) + fput(encl->backing); + + WARN_ON_ONCE(!list_empty(&encl->mm_list)); + + /* Detect EPC page leak's. */ + WARN_ON_ONCE(encl->secs_child_cnt); + WARN_ON_ONCE(encl->secs.epc_page); + + kfree(encl); +} diff --git a/arch/x86/kernel/cpu/sgx/encl.h b/arch/x86/kernel/cpu/sgx/encl.h new file mode 100644 index 000000000000..1d1bc5d590ee --- /dev/null +++ b/arch/x86/kernel/cpu/sgx/encl.h @@ -0,0 +1,87 @@ +/* SPDX-License-Identifier: (GPL-2.0 OR BSD-3-Clause) */ +/** + * Copyright(c) 2016-19 Intel Corporation. + */ +#ifndef _X86_ENCL_H +#define _X86_ENCL_H + +#include <linux/cpumask.h> +#include <linux/kref.h> +#include <linux/list.h> +#include <linux/mm_types.h> +#include <linux/mmu_notifier.h> +#include <linux/mutex.h> +#include <linux/notifier.h> +#include <linux/radix-tree.h> +#include <linux/srcu.h> +#include <linux/workqueue.h> +#include "sgx.h" + +/** + * enum sgx_encl_page_desc - defines bits for an enclave page's descriptor + * %SGX_ENCL_PAGE_ADDR_MASK: Holds the virtual address of the page. + * + * The page address for SECS is zero and is used by the subsystem to recognize + * the SECS page. + */ +enum sgx_encl_page_desc { + /* Bits 11:3 are available when the page is not swapped. */ + SGX_ENCL_PAGE_ADDR_MASK = PAGE_MASK, +}; + +#define SGX_ENCL_PAGE_ADDR(page) \ + ((page)->desc & SGX_ENCL_PAGE_ADDR_MASK) + +struct sgx_encl_page { + unsigned long desc; + unsigned long vm_max_prot_bits; + struct sgx_epc_page *epc_page; + struct sgx_encl *encl; +}; + +enum sgx_encl_flags { + SGX_ENCL_CREATED = BIT(0), + SGX_ENCL_INITIALIZED = BIT(1), + SGX_ENCL_DEBUG = BIT(2), + SGX_ENCL_DEAD = BIT(3), + SGX_ENCL_IOCTL = BIT(4), +}; + +struct sgx_encl_mm { + struct sgx_encl *encl; + struct mm_struct *mm; + struct list_head list; + struct mmu_notifier mmu_notifier; +}; + +struct sgx_encl { + atomic_t flags; + u64 secs_attributes; + u64 allowed_attributes; + unsigned int page_cnt; + unsigned int secs_child_cnt; + struct mutex lock; + struct list_head mm_list; + spinlock_t mm_lock; + struct file *backing; + struct kref refcount; + struct srcu_struct srcu; + unsigned long base; + unsigned long size; + unsigned long ssaframesize; + struct radix_tree_root page_tree; + struct sgx_encl_page secs; + cpumask_t cpumask; +}; + +extern const struct vm_operations_struct sgx_vm_ops; + +int sgx_encl_find(struct mm_struct *mm, unsigned long addr, + struct vm_area_struct **vma); +void sgx_encl_destroy(struct sgx_encl *encl); +void sgx_encl_release(struct kref *ref); +int sgx_encl_mm_add(struct sgx_encl *encl, struct mm_struct *mm); +int sgx_encl_may_map(struct sgx_encl *encl, unsigned long start, + unsigned long end, unsigned long vm_prot_bits); + +#endif /* _X86_ENCL_H */ diff --git a/arch/x86/kernel/cpu/sgx/ioctl.c b/arch/x86/kernel/cpu/sgx/ioctl.c new file mode 100644 index 000000000000..83513cdfd1c0 --- /dev/null +++ b/arch/x86/kernel/cpu/sgx/ioctl.c @@ -0,0 +1,697 @@ +// SPDX-License-Identifier: (GPL-2.0 OR BSD-3-Clause) +// Copyright(c) 2016-19 Intel Corporation. + +#include <asm/mman.h> +#include <linux/mman.h> +#include <linux/delay.h> +#include <linux/file.h> +#include <linux/hashtable.h> +#include <linux/highmem.h> +#include <linux/ratelimit.h> +#include <linux/sched/signal.h> +#include <linux/shmem_fs.h> +#include <linux/slab.h> +#include <linux/suspend.h> +#include "driver.h" +#include "encl.h" +#include "encls.h" + +/* A per-cpu cache for the last known values of IA32_SGXLEPUBKEYHASHx MSRs. */ +static DEFINE_PER_CPU(u64 [4], sgx_lepubkeyhash_cache); + +static u32 sgx_calc_ssaframesize(u32 miscselect, u64 xfrm) +{ + u32 size_max = PAGE_SIZE; + u32 size; + int i; + + for (i = 2; i < 64; i++) { + if (!((1 << i) & xfrm)) + continue; + + size = SGX_SSA_GPRS_SIZE + sgx_xsave_size_tbl[i]; + if (miscselect & SGX_MISC_EXINFO) + size += SGX_SSA_MISC_EXINFO_SIZE; + + if (size > size_max) + size_max = size; + } + + return PFN_UP(size_max); +} + +static int sgx_validate_secs(const struct sgx_secs *secs, + unsigned long ssaframesize) +{ + if (secs->size < (2 * PAGE_SIZE) || !is_power_of_2(secs->size)) + return -EINVAL; + + if (secs->base & (secs->size - 1)) + return -EINVAL; + + if (secs->miscselect & sgx_misc_reserved_mask || + secs->attributes & sgx_attributes_reserved_mask || + secs->xfrm & sgx_xfrm_reserved_mask) + return -EINVAL; + + if (secs->attributes & SGX_ATTR_MODE64BIT) { + if (secs->size > sgx_encl_size_max_64) + return -EINVAL; + } else if (secs->size > sgx_encl_size_max_32) + return -EINVAL; + + if (!(secs->xfrm & XFEATURE_MASK_FP) || + !(secs->xfrm & XFEATURE_MASK_SSE) || + (((secs->xfrm >> XFEATURE_BNDREGS) & 1) != + ((secs->xfrm >> XFEATURE_BNDCSR) & 1))) + return -EINVAL; + + if (!secs->ssa_frame_size || ssaframesize > secs->ssa_frame_size) + return -EINVAL; + + if (memchr_inv(secs->reserved1, 0, sizeof(secs->reserved1)) || + memchr_inv(secs->reserved2, 0, sizeof(secs->reserved2)) || + memchr_inv(secs->reserved3, 0, sizeof(secs->reserved3)) || + memchr_inv(secs->reserved4, 0, sizeof(secs->reserved4))) + return -EINVAL; + + return 0; +} + +static struct sgx_encl_page *sgx_encl_page_alloc(struct sgx_encl *encl, + unsigned long offset, + u64 secinfo_flags) +{ + struct sgx_encl_page *encl_page; + unsigned long prot; + + encl_page = kzalloc(sizeof(*encl_page), GFP_KERNEL); + if (!encl_page) + return ERR_PTR(-ENOMEM); + + encl_page->desc = encl->base + offset; + encl_page->encl = encl; + + prot = _calc_vm_trans(secinfo_flags, SGX_SECINFO_R, PROT_READ) | + _calc_vm_trans(secinfo_flags, SGX_SECINFO_W, PROT_WRITE) | + _calc_vm_trans(secinfo_flags, SGX_SECINFO_X, PROT_EXEC); + + /* + * TCS pages must always RW set for CPU access while the SECINFO + * permissions are *always* zero - the CPU ignores the user provided + * values and silently overwrites them with zero permissions. + */ + if ((secinfo_flags & SGX_SECINFO_PAGE_TYPE_MASK) == SGX_SECINFO_TCS) + prot |= PROT_READ | PROT_WRITE; + + /* Calculate maximum of the VM flags for the page. */ + encl_page->vm_max_prot_bits = calc_vm_prot_bits(prot, 0); + + return encl_page; +} + +static int sgx_encl_create(struct sgx_encl *encl, struct sgx_secs *secs) +{ + unsigned long encl_size = secs->size + PAGE_SIZE; + struct sgx_epc_page *secs_epc; + unsigned long ssaframesize; + struct sgx_pageinfo pginfo; + struct sgx_secinfo secinfo; + struct file *backing; + long ret; + + if (atomic_read(&encl->flags) & SGX_ENCL_CREATED) + return -EINVAL; + + ssaframesize = sgx_calc_ssaframesize(secs->miscselect, secs->xfrm); + if (sgx_validate_secs(secs, ssaframesize)) { + pr_debug("invalid SECS\n"); + return -EINVAL; + } + + backing = shmem_file_setup("SGX backing", encl_size + (encl_size >> 5), + VM_NORESERVE); + if (IS_ERR(backing)) + return PTR_ERR(backing); + + encl->backing = backing; + + secs_epc = sgx_try_alloc_page(); + if (IS_ERR(secs_epc)) { + ret = PTR_ERR(secs_epc); + goto err_out_backing; + } + + encl->secs.epc_page = secs_epc; + + pginfo.addr = 0; + pginfo.contents = (unsigned long)secs; + pginfo.metadata = (unsigned long)&secinfo; + pginfo.secs = 0; + memset(&secinfo, 0, sizeof(secinfo)); + + ret = __ecreate((void *)&pginfo, sgx_epc_addr(secs_epc)); + if (ret) { + pr_debug("ECREATE returned %ld\n", ret); + goto err_out; + } + + if (secs->attributes & SGX_ATTR_DEBUG) + atomic_or(SGX_ENCL_DEBUG, &encl->flags); + + encl->secs.encl = encl; + encl->secs_attributes = secs->attributes; + encl->allowed_attributes |= SGX_ATTR_ALLOWED_MASK; + encl->base = secs->base; + encl->size = secs->size; + encl->ssaframesize = secs->ssa_frame_size; + + /* + * Set SGX_ENCL_CREATED only after the enclave is fully prepped. This + * allows setting and checking enclave creation without having to take + * encl->lock. + */ + atomic_or(SGX_ENCL_CREATED, &encl->flags); + + return 0; + +err_out: + sgx_free_page(encl->secs.epc_page); + encl->secs.epc_page = NULL; + +err_out_backing: + fput(encl->backing); + encl->backing = NULL; + + return ret; +} + +/** + * sgx_ioc_enclave_create - handler for %SGX_IOC_ENCLAVE_CREATE + * @filep: open file to /dev/sgx + * @arg: userspace pointer to a struct sgx_enclave_create instance + * + * Allocate kernel data structures for a new enclave and execute ECREATE after + * verifying the correctness of the provided SECS. + * + * Note, enforcement of restricted and disallowed attributes is deferred until + * sgx_ioc_enclave_init(), only the architectural correctness of the SECS is + * checked by sgx_ioc_enclave_create(). + * + * Return: + * 0 on success, + * -errno otherwise + */ +static long sgx_ioc_enclave_create(struct sgx_encl *encl, void __user *arg) +{ + struct sgx_enclave_create ecreate; + struct page *secs_page; + struct sgx_secs *secs; + int ret; + + if (copy_from_user(&ecreate, arg, sizeof(ecreate))) + return -EFAULT; + + secs_page = alloc_page(GFP_KERNEL); + if (!secs_page) + return -ENOMEM; + + secs = kmap(secs_page); + if (copy_from_user(secs, (void __user *)ecreate.src, sizeof(*secs))) { + ret = -EFAULT; + goto out; + } + + ret = sgx_encl_create(encl, secs); + +out: + kunmap(secs_page); + __free_page(secs_page); + return ret; +} + +static int sgx_validate_secinfo(struct sgx_secinfo *secinfo) +{ + u64 perm = secinfo->flags & SGX_SECINFO_PERMISSION_MASK; + u64 pt = secinfo->flags & SGX_SECINFO_PAGE_TYPE_MASK; + + if (pt != SGX_SECINFO_REG && pt != SGX_SECINFO_TCS) + return -EINVAL; + + if ((perm & SGX_SECINFO_W) && !(perm & SGX_SECINFO_R)) + return -EINVAL; + + /* + * CPU will silently overwrite the permissions as zero, which means + * that we need to validate it ourselves. + */ + if (pt == SGX_SECINFO_TCS && perm) + return -EINVAL; + + if (secinfo->flags & SGX_SECINFO_RESERVED_MASK) + return -EINVAL; + + if (memchr_inv(secinfo->reserved, 0, sizeof(secinfo->reserved))) + return -EINVAL; + + return 0; +} + +static int __sgx_encl_add_page(struct sgx_encl *encl, + struct sgx_encl_page *encl_page, + struct sgx_epc_page *epc_page, + struct sgx_secinfo *secinfo, unsigned long src) +{ + struct sgx_pageinfo pginfo; + struct vm_area_struct *vma; + struct page *src_page; + int ret; + + /* Query vma's VM_MAYEXEC as an indirect path_noexec() check. */ + if (encl_page->vm_max_prot_bits & VM_EXEC) { + vma = find_vma(current->mm, src); + if (!vma) + return -EFAULT; + + if (!(vma->vm_flags & VM_MAYEXEC)) + return -EACCES; + } + + ret = get_user_pages(src, 1, 0, &src_page, NULL); + if (ret < 1) + return ret; + + pginfo.secs = (unsigned long)sgx_epc_addr(encl->secs.epc_page); + pginfo.addr = SGX_ENCL_PAGE_ADDR(encl_page); + pginfo.metadata = (unsigned long)secinfo; + pginfo.contents = (unsigned long)kmap_atomic(src_page); + + ret = __eadd(&pginfo, sgx_epc_addr(epc_page)); + + kunmap_atomic((void *)pginfo.contents); + put_page(src_page); + + return ret ? -EIO : 0; +} + +static int __sgx_encl_extend(struct sgx_encl *encl, + struct sgx_epc_page *epc_page) +{ + int ret; + int i; + + for (i = 0; i < 16; i++) { + ret = __eextend(sgx_epc_addr(encl->secs.epc_page), + sgx_epc_addr(epc_page) + (i * 0x100)); + if (ret) { + if (encls_failed(ret)) + ENCLS_WARN(ret, "EEXTEND"); + return -EIO; + } + } + + return 0; +} + +static int sgx_encl_add_page(struct sgx_encl *encl, unsigned long src, + unsigned long offset, unsigned long length, + struct sgx_secinfo *secinfo, unsigned long flags) +{ + struct sgx_encl_page *encl_page; + struct sgx_epc_page *epc_page; + int ret; + + encl_page = sgx_encl_page_alloc(encl, offset, secinfo->flags); + if (IS_ERR(encl_page)) + return PTR_ERR(encl_page); + + epc_page = sgx_try_alloc_page(); + if (IS_ERR(epc_page)) { + kfree(encl_page); + return PTR_ERR(epc_page); + } + + if (atomic_read(&encl->flags) & + (SGX_ENCL_INITIALIZED | SGX_ENCL_DEAD)) { + ret = -EFAULT; + goto err_out_free; + } + + down_read(¤t->mm->mmap_sem); + mutex_lock(&encl->lock); + + /* + * Insert prior to EADD in case of OOM. EADD modifies MRENCLAVE, i.e. + * can't be gracefully unwound, while failure on EADD/EXTEND is limited + * to userspace errors (or kernel/hardware bugs). + */ + ret = radix_tree_insert(&encl->page_tree, PFN_DOWN(encl_page->desc), + encl_page); + if (ret) + goto err_out_unlock; + + ret = __sgx_encl_add_page(encl, encl_page, epc_page, secinfo, + src); + if (ret) + goto err_out; + + /* + * Complete the "add" before doing the "extend" so that the "add" + * isn't in a half-baked state in the extremely unlikely scenario the + * the enclave will be destroyed in response to EEXTEND failure. + */ + encl_page->encl = encl; + encl_page->epc_page = epc_page; + encl->secs_child_cnt++; + + if (flags & SGX_PAGE_MEASURE) { + ret = __sgx_encl_extend(encl, epc_page); + if (ret) + goto err_out; + } + + mutex_unlock(&encl->lock); + up_read(¤t->mm->mmap_sem); + return ret; + +err_out: + radix_tree_delete(&encl_page->encl->page_tree, + PFN_DOWN(encl_page->desc)); + +err_out_unlock: + mutex_unlock(&encl->lock); + up_read(¤t->mm->mmap_sem); + +err_out_free: + sgx_free_page(epc_page); + kfree(encl_page); + + /* + * Destroy enclave on ENCLS failure as this means that EPC has been + * invalidated. + */ + if (ret == -EIO) + sgx_encl_destroy(encl); + + return ret; +} + +/** + * sgx_ioc_enclave_add_pages() - The handler for %SGX_IOC_ENCLAVE_ADD_PAGES + * @encl: pointer to an enclave instance (via ioctl() file pointer) + * @arg: a user pointer to a struct sgx_enclave_add_pages instance + * + * Add one or more pages to an uninitialized enclave, and optionally extend the + * measurement with the contents of the page. The address range of pages must + * be contiguous. The SECINFO and measurement mask are applied to all pages. + * + * A SECINFO for a TCS is required to always contain zero permissions because + * CPU silently zeros them. Allowing anything else would cause a mismatch in + * the measurement. + * + * mmap()'s protection bits are capped by the page permissions. For each page + * address, the maximum protection bits are computed with the following + * heuristics: + * + * 1. A regular page: PROT_R, PROT_W and PROT_X match the SECINFO permissions. + * 2. A TCS page: PROT_R | PROT_W. + * 3. No page: PROT_NONE. + * + * mmap() is not allowed to surpass the minimum of the maximum protection bits + * within the given address range. + * + * As stated above, a non-existent page is interpreted as a page with no + * permissions. In effect, this allows mmap() with PROT_NONE to be used to seek + * an address range for the enclave that can be then populated into SECS. + * + * If ENCLS opcode fails, that effectively means that EPC has been invalidated. + * When this happens the enclave is destroyed and -EIO is returned to the + * caller. + * + * Return: + * 0 on success, + * -EACCES if an executable source page is located in a noexec partition, + * -EIO if either ENCLS[EADD] or ENCLS[EEXTEND] fails + * -errno otherwise + */ +static long sgx_ioc_enclave_add_pages(struct sgx_encl *encl, void __user *arg) +{ + struct sgx_enclave_add_pages addp; + struct sgx_secinfo secinfo; + unsigned long c; + int ret; + + if (!(atomic_read(&encl->flags) & SGX_ENCL_CREATED)) + return -EINVAL; + + if (copy_from_user(&addp, arg, sizeof(addp))) + return -EFAULT; + + if (!IS_ALIGNED(addp.offset, PAGE_SIZE) || + !IS_ALIGNED(addp.src, PAGE_SIZE)) + return -EINVAL; + + if (!(access_ok(addp.src, PAGE_SIZE))) + return -EFAULT; + + if (addp.length & (PAGE_SIZE - 1)) + return -EINVAL; + + if (addp.offset + addp.length - PAGE_SIZE >= encl->size) + return -EINVAL; + + if (copy_from_user(&secinfo, (void __user *)addp.secinfo, + sizeof(secinfo))) + return -EFAULT; + + if (sgx_validate_secinfo(&secinfo)) + return -EINVAL; + + for (c = 0 ; c < addp.length; c += PAGE_SIZE) { + if (signal_pending(current)) { + ret = -EINTR; + break; + } + + if (need_resched()) + cond_resched(); + + ret = sgx_encl_add_page(encl, addp.src + c, addp.offset + c, + addp.length - c, &secinfo, addp.flags); + if (ret) + break; + } + + addp.count = c; + + if (copy_to_user(arg, &addp, sizeof(addp))) + return -EFAULT; + + return ret; +} + +static int __sgx_get_key_hash(struct crypto_shash *tfm, const void *modulus, + void *hash) +{ + SHASH_DESC_ON_STACK(shash, tfm); + + shash->tfm = tfm; + + return crypto_shash_digest(shash, modulus, SGX_MODULUS_SIZE, hash); +} + +static int sgx_get_key_hash(const void *modulus, void *hash) +{ + struct crypto_shash *tfm; + int ret; + + tfm = crypto_alloc_shash("sha256", 0, CRYPTO_ALG_ASYNC); + if (IS_ERR(tfm)) + return PTR_ERR(tfm); + + ret = __sgx_get_key_hash(tfm, modulus, hash); + + crypto_free_shash(tfm); + return ret; +} + +static void sgx_update_lepubkeyhash_msrs(u64 *lepubkeyhash, bool enforce) +{ + u64 *cache; + int i; + + cache = per_cpu(sgx_lepubkeyhash_cache, smp_processor_id()); + for (i = 0; i < 4; i++) { + if (enforce || (lepubkeyhash[i] != cache[i])) { + wrmsrl(MSR_IA32_SGXLEPUBKEYHASH0 + i, lepubkeyhash[i]); + cache[i] = lepubkeyhash[i]; + } + } +} + +static int sgx_einit(struct sgx_sigstruct *sigstruct, + struct sgx_einittoken *token, + struct sgx_epc_page *secs, u64 *lepubkeyhash) +{ + int ret; + + if (!boot_cpu_has(X86_FEATURE_SGX_LC)) + return __einit(sigstruct, token, sgx_epc_addr(secs)); + + preempt_disable(); + sgx_update_lepubkeyhash_msrs(lepubkeyhash, false); + ret = __einit(sigstruct, token, sgx_epc_addr(secs)); + if (ret == SGX_INVALID_EINITTOKEN) { + sgx_update_lepubkeyhash_msrs(lepubkeyhash, true); + ret = __einit(sigstruct, token, sgx_epc_addr(secs)); + } + preempt_enable(); + return ret; +} + +static int sgx_encl_init(struct sgx_encl *encl, struct sgx_sigstruct *sigstruct, + struct sgx_einittoken *token) +{ + u64 mrsigner[4]; + int ret; + int i; + int j; + + /* Check that the required attributes have been authorized. */ + if (encl->secs_attributes & ~encl->allowed_attributes) + return -EACCES; + + ret = sgx_get_key_hash(sigstruct->modulus, mrsigner); + if (ret) + return ret; + + mutex_lock(&encl->lock); + + if (atomic_read(&encl->flags) & SGX_ENCL_INITIALIZED) { + ret = -EFAULT; + goto err_out; + } + + for (i = 0; i < SGX_EINIT_SLEEP_COUNT; i++) { + for (j = 0; j < SGX_EINIT_SPIN_COUNT; j++) { + ret = sgx_einit(sigstruct, token, encl->secs.epc_page, + mrsigner); + if (ret == SGX_UNMASKED_EVENT) + continue; + else + break; + } + + if (ret != SGX_UNMASKED_EVENT) + break; + + msleep_interruptible(SGX_EINIT_SLEEP_TIME); + + if (signal_pending(current)) { + ret = -ERESTARTSYS; + goto err_out; + } + } + + if (ret & ENCLS_FAULT_FLAG) { + if (encls_failed(ret)) + ENCLS_WARN(ret, "EINIT"); + + sgx_encl_destroy(encl); + ret = -EFAULT; + } else if (ret) { + pr_debug("EINIT returned %d\n", ret); + ret = -EPERM; + } else { + atomic_or(SGX_ENCL_INITIALIZED, &encl->flags); + } + +err_out: + mutex_unlock(&encl->lock); + return ret; +} + +/** + * sgx_ioc_enclave_init - handler for %SGX_IOC_ENCLAVE_INIT + * + * @filep: open file to /dev/sgx + * @arg: userspace pointer to a struct sgx_enclave_init instance + * + * Flush any outstanding enqueued EADD operations and perform EINIT. The + * Launch Enclave Public Key Hash MSRs are rewritten as necessary to match + * the enclave's MRSIGNER, which is caculated from the provided sigstruct. + * + * Return: + * 0 on success, + * SGX error code on EINIT failure, + * -errno otherwise + */ +static long sgx_ioc_enclave_init(struct sgx_encl *encl, void __user *arg) +{ + struct sgx_einittoken *einittoken; + struct sgx_sigstruct *sigstruct; + struct sgx_enclave_init einit; + struct page *initp_page; + int ret; + + if (!(atomic_read(&encl->flags) & SGX_ENCL_CREATED)) + return -EINVAL; + + if (copy_from_user(&einit, arg, sizeof(einit))) + return -EFAULT; + + initp_page = alloc_page(GFP_KERNEL); + if (!initp_page) + return -ENOMEM; + + sigstruct = kmap(initp_page); + einittoken = (struct sgx_einittoken *) + ((unsigned long)sigstruct + PAGE_SIZE / 2); + memset(einittoken, 0, sizeof(*einittoken)); + + if (copy_from_user(sigstruct, (void __user *)einit.sigstruct, + sizeof(*sigstruct))) { + ret = -EFAULT; + goto out; + } + + ret = sgx_encl_init(encl, sigstruct, einittoken); + +out: + kunmap(initp_page); + __free_page(initp_page); + return ret; +} + + +long sgx_ioctl(struct file *filep, unsigned int cmd, unsigned long arg) +{ + struct sgx_encl *encl = filep->private_data; + int ret, encl_flags; + + encl_flags = atomic_fetch_or(SGX_ENCL_IOCTL, &encl->flags); + if (encl_flags & SGX_ENCL_IOCTL) + return -EBUSY; + + if (encl_flags & SGX_ENCL_DEAD) + return -EFAULT; + + switch (cmd) { + case SGX_IOC_ENCLAVE_CREATE: + ret = sgx_ioc_enclave_create(encl, (void __user *)arg); + break; + case SGX_IOC_ENCLAVE_ADD_PAGES: + ret = sgx_ioc_enclave_add_pages(encl, (void __user *)arg); + break; + case SGX_IOC_ENCLAVE_INIT: + ret = sgx_ioc_enclave_init(encl, (void __user *)arg); + break; + default: + ret = -ENOIOCTLCMD; + break; + } + + atomic_andnot(SGX_ENCL_IOCTL, &encl->flags); + + return ret; +} diff --git a/arch/x86/kernel/cpu/sgx/main.c b/arch/x86/kernel/cpu/sgx/main.c index 60d82e7537c8..842f9abba1c0 100644 --- a/arch/x86/kernel/cpu/sgx/main.c +++ b/arch/x86/kernel/cpu/sgx/main.c @@ -8,6 +8,7 @@ #include <linux/ratelimit.h> #include <linux/sched/signal.h> #include <linux/slab.h> +#include "driver.h" #include "encls.h" struct sgx_epc_section sgx_epc_sections[SGX_MAX_EPC_SECTIONS]; @@ -193,6 +194,8 @@ static bool __init sgx_page_cache_init(void) static void __init sgx_init(void) { + int ret; + if (!boot_cpu_has(X86_FEATURE_SGX)) return; @@ -202,10 +205,17 @@ static void __init sgx_init(void) if (!sgx_page_reclaimer_init()) goto err_page_cache; + ret = sgx_drv_init(); + if (ret) + goto err_kthread; + return; +err_kthread: + kthread_stop(ksgxswapd_tsk); + err_page_cache: sgx_page_cache_teardown(); } -arch_initcall(sgx_init); +device_initcall(sgx_init); diff --git a/arch/x86/kernel/cpu/sgx/reclaim.c b/arch/x86/kernel/cpu/sgx/reclaim.c index 215371588a25..9e6d3e147aa2 100644 --- a/arch/x86/kernel/cpu/sgx/reclaim.c +++ b/arch/x86/kernel/cpu/sgx/reclaim.c @@ -10,6 +10,7 @@ #include <linux/sched/mm.h> #include <linux/sched/signal.h> #include "encls.h" +#include "driver.h" struct task_struct *ksgxswapd_tsk;