2015-07-10 17:21:29 +02:00
|
|
|
# Commit a753f0e53ff973a8a066e86c1cb3d6dd5c68d59f
|
|
|
|
# Date 2015-06-22 11:38:01 +0200
|
|
|
|
# Author David Vrabel <david.vrabel@citrix.com>
|
|
|
|
# Committer Jan Beulich <jbeulich@suse.com>
|
|
|
|
evtchn: defer freeing struct evtchn's until evtchn_destroy_final()
|
|
|
|
|
|
|
|
notify_via_xen_event_channel() and free_xen_event_channel() had to
|
|
|
|
check if the domain was dying because they may be called while the
|
|
|
|
domain is being destroyed and the struct evtchn's are being freed.
|
|
|
|
|
|
|
|
By deferring the freeing of the struct evtchn's until all references
|
|
|
|
to the domain are dropped, these functions can rely on the channel
|
|
|
|
state being present and valid.
|
|
|
|
|
|
|
|
Signed-off-by: David Vrabel <david.vrabel@citrix.com>
|
|
|
|
|
2015-08-27 00:28:15 +02:00
|
|
|
--- a/xen/common/event_channel.c
|
|
|
|
+++ b/xen/common/event_channel.c
|
|
|
|
@@ -1174,22 +1174,8 @@ int alloc_unbound_xen_event_channel(
|
2015-07-10 17:21:29 +02:00
|
|
|
void free_xen_event_channel(
|
|
|
|
struct vcpu *local_vcpu, int port)
|
|
|
|
{
|
|
|
|
- struct evtchn *chn;
|
|
|
|
struct domain *d = local_vcpu->domain;
|
|
|
|
-
|
|
|
|
- spin_lock(&d->event_lock);
|
|
|
|
-
|
|
|
|
- if ( unlikely(d->is_dying) )
|
|
|
|
- {
|
|
|
|
- spin_unlock(&d->event_lock);
|
|
|
|
- return;
|
|
|
|
- }
|
|
|
|
-
|
|
|
|
BUG_ON(!port_is_valid(d, port));
|
|
|
|
- chn = evtchn_from_port(d, port);
|
|
|
|
- BUG_ON(!consumer_is_xen(chn));
|
|
|
|
-
|
|
|
|
- spin_unlock(&d->event_lock);
|
|
|
|
|
|
|
|
evtchn_close(d, port, 0);
|
|
|
|
}
|
2015-08-27 00:28:15 +02:00
|
|
|
@@ -1203,18 +1189,12 @@ void notify_via_xen_event_channel(struct
|
2015-07-10 17:21:29 +02:00
|
|
|
|
|
|
|
spin_lock(&ld->event_lock);
|
|
|
|
|
|
|
|
- if ( unlikely(ld->is_dying) )
|
|
|
|
- {
|
|
|
|
- spin_unlock(&ld->event_lock);
|
|
|
|
- return;
|
|
|
|
- }
|
|
|
|
-
|
|
|
|
ASSERT(port_is_valid(ld, lport));
|
|
|
|
lchn = evtchn_from_port(ld, lport);
|
|
|
|
- ASSERT(consumer_is_xen(lchn));
|
|
|
|
|
|
|
|
if ( likely(lchn->state == ECS_INTERDOMAIN) )
|
|
|
|
{
|
|
|
|
+ ASSERT(consumer_is_xen(lchn));
|
|
|
|
rd = lchn->u.interdomain.remote_dom;
|
|
|
|
rport = lchn->u.interdomain.remote_port;
|
|
|
|
rchn = evtchn_from_port(rd, rport);
|
2015-08-27 00:28:15 +02:00
|
|
|
@@ -1282,7 +1262,7 @@ int evtchn_init(struct domain *d)
|
2015-07-10 17:21:29 +02:00
|
|
|
|
|
|
|
void evtchn_destroy(struct domain *d)
|
|
|
|
{
|
|
|
|
- unsigned int i, j;
|
|
|
|
+ unsigned int i;
|
|
|
|
|
|
|
|
/* After this barrier no new event-channel allocations can occur. */
|
|
|
|
BUG_ON(!d->is_dying);
|
2015-08-27 00:28:15 +02:00
|
|
|
@@ -1292,8 +1272,17 @@ void evtchn_destroy(struct domain *d)
|
2015-07-10 17:21:29 +02:00
|
|
|
for ( i = 0; port_is_valid(d, i); i++ )
|
|
|
|
evtchn_close(d, i, 0);
|
|
|
|
|
|
|
|
+ clear_global_virq_handlers(d);
|
|
|
|
+
|
|
|
|
+ evtchn_fifo_destroy(d);
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+
|
|
|
|
+void evtchn_destroy_final(struct domain *d)
|
|
|
|
+{
|
|
|
|
+ unsigned int i, j;
|
|
|
|
+
|
|
|
|
/* Free all event-channel buckets. */
|
|
|
|
- spin_lock(&d->event_lock);
|
|
|
|
for ( i = 0; i < NR_EVTCHN_GROUPS; i++ )
|
|
|
|
{
|
|
|
|
if ( !d->evtchn_group[i] )
|
2015-08-27 00:28:15 +02:00
|
|
|
@@ -1301,20 +1290,9 @@ void evtchn_destroy(struct domain *d)
|
2015-07-10 17:21:29 +02:00
|
|
|
for ( j = 0; j < BUCKETS_PER_GROUP; j++ )
|
|
|
|
free_evtchn_bucket(d, d->evtchn_group[i][j]);
|
|
|
|
xfree(d->evtchn_group[i]);
|
|
|
|
- d->evtchn_group[i] = NULL;
|
|
|
|
}
|
|
|
|
free_evtchn_bucket(d, d->evtchn);
|
|
|
|
- d->evtchn = NULL;
|
|
|
|
- spin_unlock(&d->event_lock);
|
|
|
|
|
|
|
|
- clear_global_virq_handlers(d);
|
|
|
|
-
|
|
|
|
- evtchn_fifo_destroy(d);
|
|
|
|
-}
|
|
|
|
-
|
|
|
|
-
|
|
|
|
-void evtchn_destroy_final(struct domain *d)
|
|
|
|
-{
|
|
|
|
#if MAX_VIRT_CPUS > BITS_PER_LONG
|
|
|
|
xfree(d->poll_mask);
|
|
|
|
d->poll_mask = NULL;
|