763b78040d
config handling stack overflow CVE-2015-3259-xsa137.patch - Upstream patches from Jan 558bfaa0-x86-traps-avoid-using-current-too-early.patch 5592a116-nested-EPT-fix-the-handling-of-nested-EPT.patch 559b9dd6-x86-p2m-ept-don-t-unmap-in-use-EPT-pagetable.patch 559bdde5-pull-in-latest-linux-earlycpio.patch - Upstream patches from Jan pending review 552d0fd2-x86-hvm-don-t-include-asm-spinlock-h.patch 552d0fe8-x86-mtrr-include-asm-atomic.h.patch 552d293b-x86-vMSI-X-honor-all-mask-requests.patch 552d2966-x86-vMSI-X-add-valid-bits-for-read-acceleration.patch 554c7aee-x86-provide-arch_fetch_and_add.patch 554c7b00-arm-provide-arch_fetch_and_add.patch 55534b0a-x86-provide-add_sized.patch 55534b25-arm-provide-add_sized.patch 5555a4f8-use-ticket-locks-for-spin-locks.patch 5555a5b9-x86-arm-remove-asm-spinlock-h.patch 5555a8ec-introduce-non-contiguous-allocation.patch 55795a52-x86-vMSI-X-support-qword-MMIO-access.patch 557eb55f-gnttab-per-active-entry-locking.patch 557eb5b6-gnttab-introduce-maptrack-lock.patch 557eb620-gnttab-make-the-grant-table-lock-a-read-write-lock.patch 557ffab8-evtchn-factor-out-freeing-an-event-channel.patch 5582bf43-evtchn-simplify-port_is_valid.patch 5582bf81-evtchn-remove-the-locking-when-unmasking-an-event-channel.patch 5583d9c5-x86-MSI-X-cleanup.patch 5583da09-x86-MSI-track-host-and-guest-masking-separately.patch 5583da64-gnttab-use-per-VCPU-maptrack-free-lists.patch OBS-URL: https://build.opensuse.org/package/show/Virtualization/xen?expand=0&rev=369
142 lines
4.3 KiB
Diff
142 lines
4.3 KiB
Diff
# Commit f278fcf19ce15f7b7ee69181560b5884a5e12b66
|
|
# Date 2015-05-15 10:06:04 +0200
|
|
# Author Roger Pau Monné <roger.pau@citrix.com>
|
|
# Committer Jan Beulich <jbeulich@suse.com>
|
|
introduce a helper to allocate non-contiguous memory
|
|
|
|
The allocator uses independent calls to alloc_domheap_pages in order to get
|
|
the desired amount of memory and then maps all the independent physical
|
|
addresses into a contiguous virtual address space.
|
|
|
|
Signed-off-by: Roger Pau Monné <roger.pau@citrix.com>
|
|
Tested-by: Julien Grall <julien.grall@citrix.com> (ARM)
|
|
Reviewed-by: Tim Deegan <tim@xen.org>
|
|
|
|
# Commit 640f891eb258563bb155e577389e8c5e6541a59a
|
|
# Date 2015-05-21 08:57:19 +0200
|
|
# Author Andrew Cooper <andrew.cooper3@citrix.com>
|
|
# Committer Jan Beulich <jbeulich@suse.com>
|
|
vmap: avoid hitting an ASSERT with vfree(NULL)
|
|
|
|
and unconditionally defer the vm_size() call, as it doesn't have a NULL
|
|
short circuit.
|
|
|
|
Reported-by: Wei Liu <wei.liu2@citrix.com>
|
|
Signed-off-by: Andrew Cooper <andrew.cooper3@citrix.com>
|
|
Tested-by: Wei Liu <wei.liu2@citrix.com>
|
|
Reviewed-by: Roger Pau Monné <roger.pau@citrix.com>
|
|
Acked-by: Tim Deegan <tim@xen.org>
|
|
|
|
--- sle12sp1.orig/xen/common/vmap.c 2013-10-31 22:33:32.000000000 +0100
|
|
+++ sle12sp1/xen/common/vmap.c 2015-07-08 14:18:50.000000000 +0200
|
|
@@ -215,4 +215,75 @@ void vunmap(const void *va)
|
|
#endif
|
|
vm_free(va);
|
|
}
|
|
+
|
|
+void *vmalloc(size_t size)
|
|
+{
|
|
+ unsigned long *mfn;
|
|
+ size_t pages, i;
|
|
+ struct page_info *pg;
|
|
+ void *va;
|
|
+
|
|
+ ASSERT(size);
|
|
+
|
|
+ pages = PFN_UP(size);
|
|
+ mfn = xmalloc_array(unsigned long, pages);
|
|
+ if ( mfn == NULL )
|
|
+ return NULL;
|
|
+
|
|
+ for ( i = 0; i < pages; i++ )
|
|
+ {
|
|
+ pg = alloc_domheap_page(NULL, 0);
|
|
+ if ( pg == NULL )
|
|
+ goto error;
|
|
+ mfn[i] = page_to_mfn(pg);
|
|
+ }
|
|
+
|
|
+ va = vmap(mfn, pages);
|
|
+ if ( va == NULL )
|
|
+ goto error;
|
|
+
|
|
+ xfree(mfn);
|
|
+ return va;
|
|
+
|
|
+ error:
|
|
+ while ( i-- )
|
|
+ free_domheap_page(mfn_to_page(mfn[i]));
|
|
+ xfree(mfn);
|
|
+ return NULL;
|
|
+}
|
|
+
|
|
+void *vzalloc(size_t size)
|
|
+{
|
|
+ void *p = vmalloc(size);
|
|
+ int i;
|
|
+
|
|
+ if ( p == NULL )
|
|
+ return NULL;
|
|
+
|
|
+ for ( i = 0; i < size; i += PAGE_SIZE )
|
|
+ clear_page(p + i);
|
|
+
|
|
+ return p;
|
|
+}
|
|
+
|
|
+void vfree(void *va)
|
|
+{
|
|
+ unsigned int i, pages;
|
|
+ struct page_info *pg;
|
|
+ PAGE_LIST_HEAD(pg_list);
|
|
+
|
|
+ if ( !va )
|
|
+ return;
|
|
+
|
|
+ pages = vm_size(va);
|
|
+ ASSERT(pages);
|
|
+
|
|
+ for ( i = 0; i < pages; i++ )
|
|
+ page_list_add(vmap_to_page(va + i * PAGE_SIZE), &pg_list);
|
|
+
|
|
+ vunmap(va);
|
|
+
|
|
+ while ( (pg = page_list_remove_head(&pg_list)) != NULL )
|
|
+ free_domheap_page(pg);
|
|
+}
|
|
#endif
|
|
--- sle12sp1.orig/xen/include/asm-arm/mm.h 2015-01-14 18:44:18.000000000 +0100
|
|
+++ sle12sp1/xen/include/asm-arm/mm.h 2015-07-08 14:18:50.000000000 +0200
|
|
@@ -208,6 +208,8 @@ static inline void __iomem *ioremap_wc(p
|
|
#define pfn_to_paddr(pfn) ((paddr_t)(pfn) << PAGE_SHIFT)
|
|
#define paddr_to_pfn(pa) ((unsigned long)((pa) >> PAGE_SHIFT))
|
|
#define paddr_to_pdx(pa) pfn_to_pdx(paddr_to_pfn(pa))
|
|
+#define vmap_to_mfn(va) paddr_to_pfn(virt_to_maddr((vaddr_t)va))
|
|
+#define vmap_to_page(va) mfn_to_page(vmap_to_mfn(va))
|
|
|
|
/* Page-align address and convert to frame number format */
|
|
#define paddr_to_pfn_aligned(paddr) paddr_to_pfn(PAGE_ALIGN(paddr))
|
|
--- sle12sp1.orig/xen/include/asm-x86/page.h 2015-06-03 16:55:05.000000000 +0200
|
|
+++ sle12sp1/xen/include/asm-x86/page.h 2015-07-08 14:18:50.000000000 +0200
|
|
@@ -262,6 +262,8 @@ void copy_page_sse2(void *, const void *
|
|
#define pfn_to_paddr(pfn) __pfn_to_paddr(pfn)
|
|
#define paddr_to_pfn(pa) __paddr_to_pfn(pa)
|
|
#define paddr_to_pdx(pa) pfn_to_pdx(paddr_to_pfn(pa))
|
|
+#define vmap_to_mfn(va) l1e_get_pfn(*virt_to_xen_l1e((unsigned long)(va)))
|
|
+#define vmap_to_page(va) mfn_to_page(vmap_to_mfn(va))
|
|
|
|
#endif /* !defined(__ASSEMBLY__) */
|
|
|
|
--- sle12sp1.orig/xen/include/xen/vmap.h 2013-07-09 20:57:12.000000000 +0200
|
|
+++ sle12sp1/xen/include/xen/vmap.h 2015-07-08 14:18:50.000000000 +0200
|
|
@@ -11,6 +11,9 @@ void *__vmap(const unsigned long *mfn, u
|
|
unsigned int nr, unsigned int align, unsigned int flags);
|
|
void *vmap(const unsigned long *mfn, unsigned int nr);
|
|
void vunmap(const void *);
|
|
+void *vmalloc(size_t size);
|
|
+void *vzalloc(size_t size);
|
|
+void vfree(void *va);
|
|
|
|
void __iomem *ioremap(paddr_t, size_t);
|
|
|