97a0425e04
xen-4.5.1-testing-src.tar.bz2 - Dropped patches now contained in tarball 556c2cf2-x86-don-t-crash-mapping-a-page-using-EFI-rt-page-tables.patch 556d9718-efi-fix-allocation-problems-if-ExitBootServices-fails.patch 556eabf7-x86-apic-Disable-the-LAPIC-later-in-smp_send_stop.patch 556eac15-x86-crash-don-t-use-set_fixmap-in-the-crash-path.patch 55780aaa-efi-avoid-calling-boot-services-after-ExitBootServices.patch 55780aff-x86-EFI-fix-EFI_MEMORY_WP-handling.patch 55780b43-EFI-early-add-mapbs-to-map-EfiBootServices-Code-Data.patch 55780b97-EFI-support-default-attributes-to-map-Runtime-service-areas.patch - Replace 5124efbe-add-qxl-support.patch with the variant that finally made it upstream, 554cc211-libxl-add-qxl.patch - bsc#931627 - VUL-0: CVE-2015-4105: XSA-130: xen: Guest triggerable qemu MSI-X pass-through error messages qemu-MSI-X-latch-writes.patch - bsc#907514 - Bus fatal error & sles12 sudden reboot has been observed - bsc#910258 - SLES12 Xen host crashes with FATAL NMI after shutdown of guest with VT-d NIC - bsc#918984 - Bus fatal error & sles11-SP4 sudden reboot has been observed - bsc#923967 - Partner-L3: Bus fatal error & sles11-SP3 sudden reboot has been observed x86-MSI-X-teardown.patch x86-MSI-X-enable.patch x86-MSI-X-guest-mask.patch x86-MSI-X-maskall.patch qemu-MSI-X-enable-maskall.patch OBS-URL: https://build.opensuse.org/package/show/Virtualization/xen?expand=0&rev=364
251 lines
10 KiB
Diff
251 lines
10 KiB
Diff
# Commit d72a4605e18d3a61c4469ff092dbbbfa4ac919f7
|
|
# Date 2015-05-05 18:01:33 +0200
|
|
# Author Jan Beulich <jbeulich@suse.com>
|
|
# Committer Jan Beulich <jbeulich@suse.com>
|
|
x86: allow 64-bit PV guest kernels to suppress user mode exposure of M2P
|
|
|
|
Xen L4 entries being uniformly installed into any L4 table and 64-bit
|
|
PV kernels running in ring 3 means that user mode was able to see the
|
|
read-only M2P presented by Xen to the guests. While apparently not
|
|
really representing an exploitable information leak, this still very
|
|
certainly was never meant to be that way.
|
|
|
|
Building on the fact that these guests already have separate kernel and
|
|
user mode page tables we can allow guest kernels to tell Xen that they
|
|
don't want user mode to see this table. We can't, however, do this by
|
|
default: There is no ABI requirement that kernel and user mode page
|
|
tables be separate. Therefore introduce a new VM-assist flag allowing
|
|
the guest to control respective hypervisor behavior:
|
|
- when not set, L4 tables get created with the respective slot blank,
|
|
and whenever the L4 table gets used as a kernel one the missing
|
|
mapping gets inserted,
|
|
- when set, L4 tables get created with the respective slot initialized
|
|
as before, and whenever the L4 table gets used as a user one the
|
|
mapping gets zapped.
|
|
|
|
Signed-off-by: Jan Beulich <jbeulich@suse.com>
|
|
Reviewed-by: Andrew Cooper <andrew.cooper3@citrix.com>
|
|
Reviewed-by: Tim Deegan <tim@xen.org>
|
|
|
|
--- a/xen/arch/x86/domain.c
|
|
+++ b/xen/arch/x86/domain.c
|
|
@@ -338,7 +338,7 @@ static int setup_compat_l4(struct vcpu *
|
|
|
|
l4tab = __map_domain_page(pg);
|
|
clear_page(l4tab);
|
|
- init_guest_l4_table(l4tab, v->domain);
|
|
+ init_guest_l4_table(l4tab, v->domain, 1);
|
|
unmap_domain_page(l4tab);
|
|
|
|
v->arch.guest_table = pagetable_from_page(pg);
|
|
@@ -977,7 +977,11 @@ int arch_set_info_guest(
|
|
case -EINTR:
|
|
rc = -ERESTART;
|
|
case -ERESTART:
|
|
+ break;
|
|
case 0:
|
|
+ if ( !compat && !VM_ASSIST(d, VMASST_TYPE_m2p_strict) &&
|
|
+ !paging_mode_refcounts(d) )
|
|
+ fill_ro_mpt(cr3_gfn);
|
|
break;
|
|
default:
|
|
if ( cr3_page == current->arch.old_guest_table )
|
|
@@ -1012,7 +1016,10 @@ int arch_set_info_guest(
|
|
default:
|
|
if ( cr3_page == current->arch.old_guest_table )
|
|
cr3_page = NULL;
|
|
+ break;
|
|
case 0:
|
|
+ if ( VM_ASSIST(d, VMASST_TYPE_m2p_strict) )
|
|
+ zap_ro_mpt(cr3_gfn);
|
|
break;
|
|
}
|
|
}
|
|
--- a/xen/arch/x86/domain_build.c
|
|
+++ b/xen/arch/x86/domain_build.c
|
|
@@ -1092,7 +1092,7 @@ int __init construct_dom0(
|
|
l3start = __va(mpt_alloc); mpt_alloc += PAGE_SIZE;
|
|
}
|
|
clear_page(l4tab);
|
|
- init_guest_l4_table(l4tab, d);
|
|
+ init_guest_l4_table(l4tab, d, 0);
|
|
v->arch.guest_table = pagetable_from_paddr(__pa(l4start));
|
|
if ( is_pv_32on64_domain(d) )
|
|
v->arch.guest_table_user = v->arch.guest_table;
|
|
--- a/xen/arch/x86/mm.c
|
|
+++ b/xen/arch/x86/mm.c
|
|
@@ -1380,7 +1380,8 @@ static int alloc_l3_table(struct page_in
|
|
return rc > 0 ? 0 : rc;
|
|
}
|
|
|
|
-void init_guest_l4_table(l4_pgentry_t l4tab[], const struct domain *d)
|
|
+void init_guest_l4_table(l4_pgentry_t l4tab[], const struct domain *d,
|
|
+ bool_t zap_ro_mpt)
|
|
{
|
|
/* Xen private mappings. */
|
|
memcpy(&l4tab[ROOT_PAGETABLE_FIRST_XEN_SLOT],
|
|
@@ -1395,6 +1396,25 @@ void init_guest_l4_table(l4_pgentry_t l4
|
|
l4e_from_pfn(domain_page_map_to_mfn(l4tab), __PAGE_HYPERVISOR);
|
|
l4tab[l4_table_offset(PERDOMAIN_VIRT_START)] =
|
|
l4e_from_page(d->arch.perdomain_l3_pg, __PAGE_HYPERVISOR);
|
|
+ if ( zap_ro_mpt || is_pv_32on64_domain(d) || paging_mode_refcounts(d) )
|
|
+ l4tab[l4_table_offset(RO_MPT_VIRT_START)] = l4e_empty();
|
|
+}
|
|
+
|
|
+void fill_ro_mpt(unsigned long mfn)
|
|
+{
|
|
+ l4_pgentry_t *l4tab = map_domain_page(mfn);
|
|
+
|
|
+ l4tab[l4_table_offset(RO_MPT_VIRT_START)] =
|
|
+ idle_pg_table[l4_table_offset(RO_MPT_VIRT_START)];
|
|
+ unmap_domain_page(l4tab);
|
|
+}
|
|
+
|
|
+void zap_ro_mpt(unsigned long mfn)
|
|
+{
|
|
+ l4_pgentry_t *l4tab = map_domain_page(mfn);
|
|
+
|
|
+ l4tab[l4_table_offset(RO_MPT_VIRT_START)] = l4e_empty();
|
|
+ unmap_domain_page(l4tab);
|
|
}
|
|
|
|
static int alloc_l4_table(struct page_info *page)
|
|
@@ -1444,7 +1464,7 @@ static int alloc_l4_table(struct page_in
|
|
adjust_guest_l4e(pl4e[i], d);
|
|
}
|
|
|
|
- init_guest_l4_table(pl4e, d);
|
|
+ init_guest_l4_table(pl4e, d, !VM_ASSIST(d, VMASST_TYPE_m2p_strict));
|
|
unmap_domain_page(pl4e);
|
|
|
|
return rc > 0 ? 0 : rc;
|
|
@@ -2755,6 +2775,8 @@ int new_guest_cr3(unsigned long mfn)
|
|
|
|
invalidate_shadow_ldt(curr, 0);
|
|
|
|
+ if ( !VM_ASSIST(d, VMASST_TYPE_m2p_strict) && !paging_mode_refcounts(d) )
|
|
+ fill_ro_mpt(mfn);
|
|
curr->arch.guest_table = pagetable_from_pfn(mfn);
|
|
update_cr3(curr);
|
|
|
|
@@ -3111,6 +3133,9 @@ long do_mmuext_op(
|
|
op.arg1.mfn);
|
|
break;
|
|
}
|
|
+ if ( VM_ASSIST(d, VMASST_TYPE_m2p_strict) &&
|
|
+ !paging_mode_refcounts(d) )
|
|
+ zap_ro_mpt(op.arg1.mfn);
|
|
}
|
|
|
|
curr->arch.guest_table_user = pagetable_from_pfn(op.arg1.mfn);
|
|
--- a/xen/arch/x86/mm/shadow/multi.c
|
|
+++ b/xen/arch/x86/mm/shadow/multi.c
|
|
@@ -1438,6 +1438,13 @@ void sh_install_xen_entries_in_l4(struct
|
|
shadow_l4e_from_mfn(page_to_mfn(d->arch.perdomain_l3_pg),
|
|
__PAGE_HYPERVISOR);
|
|
|
|
+ if ( !shadow_mode_external(d) && !is_pv_32on64_domain(d) &&
|
|
+ !VM_ASSIST(d, VMASST_TYPE_m2p_strict) )
|
|
+ {
|
|
+ /* open coded zap_ro_mpt(mfn_x(sl4mfn)): */
|
|
+ sl4e[shadow_l4_table_offset(RO_MPT_VIRT_START)] = shadow_l4e_empty();
|
|
+ }
|
|
+
|
|
/* Shadow linear mapping for 4-level shadows. N.B. for 3-level
|
|
* shadows on 64-bit xen, this linear mapping is later replaced by the
|
|
* monitor pagetable structure, which is built in make_monitor_table
|
|
@@ -4062,6 +4069,17 @@ sh_update_cr3(struct vcpu *v, int do_loc
|
|
if ( sh_remove_write_access(v, gmfn, 4, 0) != 0 )
|
|
flush_tlb_mask(d->domain_dirty_cpumask);
|
|
sh_set_toplevel_shadow(v, 0, gmfn, SH_type_l4_shadow);
|
|
+ if ( !shadow_mode_external(d) && !is_pv_32on64_domain(d) )
|
|
+ {
|
|
+ mfn_t smfn = pagetable_get_mfn(v->arch.shadow_table[0]);
|
|
+
|
|
+ if ( !(v->arch.flags & TF_kernel_mode) &&
|
|
+ VM_ASSIST(d, VMASST_TYPE_m2p_strict) )
|
|
+ zap_ro_mpt(mfn_x(smfn));
|
|
+ else if ( (v->arch.flags & TF_kernel_mode) &&
|
|
+ !VM_ASSIST(d, VMASST_TYPE_m2p_strict) )
|
|
+ fill_ro_mpt(mfn_x(smfn));
|
|
+ }
|
|
#else
|
|
#error This should never happen
|
|
#endif
|
|
--- a/xen/arch/x86/x86_64/mm.c
|
|
+++ b/xen/arch/x86/x86_64/mm.c
|
|
@@ -480,7 +480,7 @@ static int setup_m2p_table(struct mem_ho
|
|
l2_ro_mpt += l2_table_offset(va);
|
|
}
|
|
|
|
- /* NB. Cannot be GLOBAL as shadow_mode_translate reuses this area. */
|
|
+ /* NB. Cannot be GLOBAL: guest user mode should not see it. */
|
|
l2e_write(l2_ro_mpt, l2e_from_pfn(mfn,
|
|
/*_PAGE_GLOBAL|*/_PAGE_PSE|_PAGE_USER|_PAGE_PRESENT));
|
|
}
|
|
@@ -583,7 +583,7 @@ void __init paging_init(void)
|
|
0x77, 1UL << L3_PAGETABLE_SHIFT);
|
|
|
|
ASSERT(!l2_table_offset(va));
|
|
- /* NB. Cannot be GLOBAL as shadow_mode_translate reuses this area. */
|
|
+ /* NB. Cannot be GLOBAL: guest user mode should not see it. */
|
|
l3e_write(&l3_ro_mpt[l3_table_offset(va)],
|
|
l3e_from_page(l1_pg,
|
|
/*_PAGE_GLOBAL|*/_PAGE_PSE|_PAGE_USER|_PAGE_PRESENT));
|
|
@@ -621,7 +621,7 @@ void __init paging_init(void)
|
|
l3e_from_page(l2_pg, __PAGE_HYPERVISOR | _PAGE_USER));
|
|
ASSERT(!l2_table_offset(va));
|
|
}
|
|
- /* NB. Cannot be GLOBAL as shadow_mode_translate reuses this area. */
|
|
+ /* NB. Cannot be GLOBAL: guest user mode should not see it. */
|
|
if ( l1_pg )
|
|
l2e_write(l2_ro_mpt, l2e_from_page(
|
|
l1_pg, /*_PAGE_GLOBAL|*/_PAGE_PSE|_PAGE_USER|_PAGE_PRESENT));
|
|
--- a/xen/include/asm-x86/config.h
|
|
+++ b/xen/include/asm-x86/config.h
|
|
@@ -330,7 +330,8 @@ extern unsigned long xen_phys_start;
|
|
#define NATIVE_VM_ASSIST_VALID ((1UL << VMASST_TYPE_4gb_segments) | \
|
|
(1UL << VMASST_TYPE_4gb_segments_notify) | \
|
|
(1UL << VMASST_TYPE_writable_pagetables) | \
|
|
- (1UL << VMASST_TYPE_pae_extended_cr3))
|
|
+ (1UL << VMASST_TYPE_pae_extended_cr3) | \
|
|
+ (1UL << VMASST_TYPE_m2p_strict))
|
|
#define VM_ASSIST_VALID NATIVE_VM_ASSIST_VALID
|
|
#define COMPAT_VM_ASSIST_VALID (NATIVE_VM_ASSIST_VALID & \
|
|
((1UL << COMPAT_BITS_PER_LONG) - 1))
|
|
--- a/xen/include/asm-x86/mm.h
|
|
+++ b/xen/include/asm-x86/mm.h
|
|
@@ -314,7 +314,10 @@ static inline void *__page_to_virt(const
|
|
int free_page_type(struct page_info *page, unsigned long type,
|
|
int preemptible);
|
|
|
|
-void init_guest_l4_table(l4_pgentry_t[], const struct domain *);
|
|
+void init_guest_l4_table(l4_pgentry_t[], const struct domain *,
|
|
+ bool_t zap_ro_mpt);
|
|
+void fill_ro_mpt(unsigned long mfn);
|
|
+void zap_ro_mpt(unsigned long mfn);
|
|
|
|
int is_iomem_page(unsigned long mfn);
|
|
|
|
--- a/xen/include/public/xen.h
|
|
+++ b/xen/include/public/xen.h
|
|
@@ -486,6 +486,18 @@ DEFINE_XEN_GUEST_HANDLE(mmuext_op_t);
|
|
/* x86/PAE guests: support PDPTs above 4GB. */
|
|
#define VMASST_TYPE_pae_extended_cr3 3
|
|
|
|
+/*
|
|
+ * x86/64 guests: strictly hide M2P from user mode.
|
|
+ * This allows the guest to control respective hypervisor behavior:
|
|
+ * - when not set, L4 tables get created with the respective slot blank,
|
|
+ * and whenever the L4 table gets used as a kernel one the missing
|
|
+ * mapping gets inserted,
|
|
+ * - when set, L4 tables get created with the respective slot initialized
|
|
+ * as before, and whenever the L4 table gets used as a user one the
|
|
+ * mapping gets zapped.
|
|
+ */
|
|
+#define VMASST_TYPE_m2p_strict 32
|
|
+
|
|
#if __XEN_INTERFACE_VERSION__ < 0x00040600
|
|
#define MAX_VMASST_TYPE 3
|
|
#endif
|