1
0
mirror of https://gitlab.gnome.org/GNOME/glib.git synced 2025-03-22 07:30:06 +01:00
glib/gio/inotify/inotify-kernel.c
Alexander Larsson d8473705b0 Removed. Removed. Removed the included copies of the inotify headers. We
2008-01-21  Alexander Larsson  <alexl@redhat.com>

        * inotify/Makefile.am:
        * inotify/inotify-helper.c:
        * inotify/inotify-kernel.c:
        * inotify/inotify-path.c:
        * inotify/local_inotify.h: Removed.
        * inotify/local_inotify_syscalls.h: Removed.
	Removed the included copies of the inotify
	headers. We now only use the <sys/inotify.h>
	header which exists on modern systems.
	This fixes problems on ARM and SH5 ()
	but is also generally much cleaner and future
	safe. For instance, if other OSes add support
	for inotify it should "just work".



svn path=/trunk/; revision=6338
2008-01-21 10:35:44 +00:00

674 lines
16 KiB
C

/*
Copyright (C) 2005 John McCutchan
The Gnome Library is free software; you can redistribute it and/or
modify it under the terms of the GNU Library General Public License as
published by the Free Software Foundation; either version 2 of the
License, or (at your option) any later version.
The Gnome Library is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
Library General Public License for more details.
You should have received a copy of the GNU Library General Public
License along with the Gnome Library; see the file COPYING.LIB. If not,
write to the Free Software Foundation, Inc., 59 Temple Place - Suite 330,
Boston, MA 02111-1307, USA.
Authors:.
John McCutchan <john@johnmccutchan.com>
*/
#include "config.h"
#include <stdio.h>
#include <sys/ioctl.h>
#include <unistd.h>
#include <errno.h>
#include <string.h>
#include <glib.h>
#include "inotify-kernel.h"
#include <sys/inotify.h>
/* Timings for pairing MOVED_TO / MOVED_FROM events */
#define PROCESS_EVENTS_TIME 1000 /* milliseconds (1 hz) */
#define DEFAULT_HOLD_UNTIL_TIME 0 /* 0 millisecond */
#define MOVE_HOLD_UNTIL_TIME 0 /* 0 milliseconds */
static int inotify_instance_fd = -1;
static GQueue *events_to_process = NULL;
static GQueue *event_queue = NULL;
static GHashTable * cookie_hash = NULL;
static GIOChannel *inotify_read_ioc;
static GPollFD ik_poll_fd;
static gboolean ik_poll_fd_enabled = TRUE;
static void (*user_cb)(ik_event_t *event);
static gboolean ik_read_callback (gpointer user_data);
static gboolean ik_process_eq_callback (gpointer user_data);
static guint32 ik_move_matches = 0;
static guint32 ik_move_misses = 0;
static gboolean process_eq_running = FALSE;
/* We use the lock from inotify-helper.c
*
* There are two places that we take this lock
*
* 1) In ik_read_callback
*
* 2) ik_process_eq_callback.
*
*
* The rest of locking is taken care of in inotify-helper.c
*/
G_LOCK_EXTERN (inotify_lock);
typedef struct ik_event_internal {
ik_event_t *event;
gboolean seen;
gboolean sent;
GTimeVal hold_until;
struct ik_event_internal *pair;
} ik_event_internal_t;
/* In order to perform non-sleeping inotify event chunking we need
* a custom GSource
*/
static gboolean
ik_source_prepare (GSource *source,
gint *timeout)
{
return FALSE;
}
static gboolean
ik_source_timeout (gpointer data)
{
GSource *source = (GSource *)data;
/* Re-active the PollFD */
g_source_add_poll (source, &ik_poll_fd);
g_source_unref (source);
ik_poll_fd_enabled = TRUE;
return FALSE;
}
#define MAX_PENDING_COUNT 2
#define PENDING_THRESHOLD(qsize) ((qsize) >> 1)
#define PENDING_MARGINAL_COST(p) ((unsigned int)(1 << (p)))
#define MAX_QUEUED_EVENTS 2048
#define AVERAGE_EVENT_SIZE sizeof (struct inotify_event) + 16
#define TIMEOUT_MILLISECONDS 10
static gboolean
ik_source_check (GSource *source)
{
static int prev_pending = 0, pending_count = 0;
/* We already disabled the PollFD or
* nothing to be read from inotify */
if (!ik_poll_fd_enabled || !(ik_poll_fd.revents & G_IO_IN))
return FALSE;
if (pending_count < MAX_PENDING_COUNT)
{
unsigned int pending;
if (ioctl (inotify_instance_fd, FIONREAD, &pending) == -1)
goto do_read;
pending /= AVERAGE_EVENT_SIZE;
/* Don't wait if the number of pending events is too close
* to the maximum queue size.
*/
if (pending > PENDING_THRESHOLD (MAX_QUEUED_EVENTS))
goto do_read;
/* With each successive iteration, the minimum rate for
* further sleep doubles.
*/
if (pending-prev_pending < PENDING_MARGINAL_COST (pending_count))
goto do_read;
prev_pending = pending;
pending_count++;
/* We are going to wait to read the events: */
/* Remove the PollFD from the source */
g_source_remove_poll (source, &ik_poll_fd);
/* To avoid threading issues we need to flag that we've done that */
ik_poll_fd_enabled = FALSE;
/* Set a timeout to re-add the PollFD to the source */
g_source_ref (source);
g_timeout_add (TIMEOUT_MILLISECONDS, ik_source_timeout, source);
return FALSE;
}
do_read:
/* We are ready to read events from inotify */
prev_pending = 0;
pending_count = 0;
return TRUE;
}
static gboolean
ik_source_dispatch (GSource *source,
GSourceFunc callback,
gpointer user_data)
{
if (callback)
return callback (user_data);
return TRUE;
}
static GSourceFuncs ik_source_funcs =
{
ik_source_prepare,
ik_source_check,
ik_source_dispatch,
NULL
};
gboolean _ik_startup (void (*cb)(ik_event_t *event))
{
static gboolean initialized = FALSE;
GSource *source;
user_cb = cb;
/* Ignore multi-calls */
if (initialized)
return inotify_instance_fd >= 0;
initialized = TRUE;
inotify_instance_fd = inotify_init ();
if (inotify_instance_fd < 0)
return FALSE;
inotify_read_ioc = g_io_channel_unix_new (inotify_instance_fd);
ik_poll_fd.fd = inotify_instance_fd;
ik_poll_fd.events = G_IO_IN | G_IO_HUP | G_IO_ERR;
g_io_channel_set_encoding (inotify_read_ioc, NULL, NULL);
g_io_channel_set_flags (inotify_read_ioc, G_IO_FLAG_NONBLOCK, NULL);
source = g_source_new (&ik_source_funcs, sizeof (GSource));
g_source_add_poll (source, &ik_poll_fd);
g_source_set_callback (source, ik_read_callback, NULL, NULL);
g_source_attach (source, NULL);
g_source_unref (source);
cookie_hash = g_hash_table_new (g_direct_hash, g_direct_equal);
event_queue = g_queue_new ();
events_to_process = g_queue_new ();
return TRUE;
}
static ik_event_internal_t *
ik_event_internal_new (ik_event_t *event)
{
ik_event_internal_t *internal_event = g_new0 (ik_event_internal_t, 1);
GTimeVal tv;
g_assert (event);
g_get_current_time (&tv);
g_time_val_add (&tv, DEFAULT_HOLD_UNTIL_TIME);
internal_event->event = event;
internal_event->hold_until = tv;
return internal_event;
}
static ik_event_t *
ik_event_new (char *buffer)
{
struct inotify_event *kevent = (struct inotify_event *)buffer;
ik_event_t *event = g_new0 (ik_event_t, 1);
g_assert (buffer);
event->wd = kevent->wd;
event->mask = kevent->mask;
event->cookie = kevent->cookie;
event->len = kevent->len;
if (event->len)
event->name = g_strdup (kevent->name);
else
event->name = g_strdup ("");
return event;
}
ik_event_t *
_ik_event_new_dummy (const char *name,
gint32 wd,
guint32 mask)
{
ik_event_t *event = g_new0 (ik_event_t, 1);
event->wd = wd;
event->mask = mask;
event->cookie = 0;
if (name)
event->name = g_strdup (name);
else
event->name = g_strdup("");
event->len = strlen (event->name);
return event;
}
void
_ik_event_free (ik_event_t *event)
{
if (event->pair)
_ik_event_free (event->pair);
g_free (event->name);
g_free (event);
}
gint32
_ik_watch (const char *path,
guint32 mask,
int *err)
{
gint32 wd = -1;
g_assert (path != NULL);
g_assert (inotify_instance_fd >= 0);
wd = inotify_add_watch (inotify_instance_fd, path, mask);
if (wd < 0)
{
int e = errno;
/* FIXME: debug msg failed to add watch */
if (err)
*err = e;
return wd;
}
g_assert (wd >= 0);
return wd;
}
int
_ik_ignore (const char *path,
gint32 wd)
{
g_assert (wd >= 0);
g_assert (inotify_instance_fd >= 0);
if (inotify_rm_watch (inotify_instance_fd, wd) < 0)
{
/* int e = errno; */
/* failed to rm watch */
return -1;
}
return 0;
}
void
_ik_move_stats (guint32 *matches,
guint32 *misses)
{
if (matches)
*matches = ik_move_matches;
if (misses)
*misses = ik_move_misses;
}
const char *
_ik_mask_to_string (guint32 mask)
{
gboolean is_dir = mask & IN_ISDIR;
mask &= ~IN_ISDIR;
if (is_dir)
{
switch (mask)
{
case IN_ACCESS:
return "ACCESS (dir)";
case IN_MODIFY:
return "MODIFY (dir)";
case IN_ATTRIB:
return "ATTRIB (dir)";
case IN_CLOSE_WRITE:
return "CLOSE_WRITE (dir)";
case IN_CLOSE_NOWRITE:
return "CLOSE_NOWRITE (dir)";
case IN_OPEN:
return "OPEN (dir)";
case IN_MOVED_FROM:
return "MOVED_FROM (dir)";
case IN_MOVED_TO:
return "MOVED_TO (dir)";
case IN_DELETE:
return "DELETE (dir)";
case IN_CREATE:
return "CREATE (dir)";
case IN_DELETE_SELF:
return "DELETE_SELF (dir)";
case IN_UNMOUNT:
return "UNMOUNT (dir)";
case IN_Q_OVERFLOW:
return "Q_OVERFLOW (dir)";
case IN_IGNORED:
return "IGNORED (dir)";
default:
return "UNKNOWN_EVENT (dir)";
}
}
else
{
switch (mask)
{
case IN_ACCESS:
return "ACCESS";
case IN_MODIFY:
return "MODIFY";
case IN_ATTRIB:
return "ATTRIB";
case IN_CLOSE_WRITE:
return "CLOSE_WRITE";
case IN_CLOSE_NOWRITE:
return "CLOSE_NOWRITE";
case IN_OPEN:
return "OPEN";
case IN_MOVED_FROM:
return "MOVED_FROM";
case IN_MOVED_TO:
return "MOVED_TO";
case IN_DELETE:
return "DELETE";
case IN_CREATE:
return "CREATE";
case IN_DELETE_SELF:
return "DELETE_SELF";
case IN_UNMOUNT:
return "UNMOUNT";
case IN_Q_OVERFLOW:
return "Q_OVERFLOW";
case IN_IGNORED:
return "IGNORED";
default:
return "UNKNOWN_EVENT";
}
}
}
static void
ik_read_events (gsize *buffer_size_out,
gchar **buffer_out)
{
static gchar *buffer = NULL;
static gsize buffer_size;
/* Initialize the buffer on our first call */
if (buffer == NULL)
{
buffer_size = AVERAGE_EVENT_SIZE;
buffer_size *= MAX_QUEUED_EVENTS;
buffer = g_malloc (buffer_size);
}
*buffer_size_out = 0;
*buffer_out = NULL;
memset (buffer, 0, buffer_size);
if (g_io_channel_read_chars (inotify_read_ioc, (char *)buffer, buffer_size, buffer_size_out, NULL) != G_IO_STATUS_NORMAL) {
/* error reading */
}
*buffer_out = buffer;
}
static gboolean
ik_read_callback (gpointer user_data)
{
gchar *buffer;
gsize buffer_size, buffer_i, events;
G_LOCK (inotify_lock);
ik_read_events (&buffer_size, &buffer);
buffer_i = 0;
events = 0;
while (buffer_i < buffer_size)
{
struct inotify_event *event;
gsize event_size;
event = (struct inotify_event *)&buffer[buffer_i];
event_size = sizeof(struct inotify_event) + event->len;
g_queue_push_tail (events_to_process, ik_event_internal_new (ik_event_new (&buffer[buffer_i])));
buffer_i += event_size;
events++;
}
/* If the event process callback is off, turn it back on */
if (!process_eq_running && events)
{
process_eq_running = TRUE;
g_timeout_add (PROCESS_EVENTS_TIME, ik_process_eq_callback, NULL);
}
G_UNLOCK (inotify_lock);
return TRUE;
}
static gboolean
g_timeval_lt (GTimeVal *val1,
GTimeVal *val2)
{
if (val1->tv_sec < val2->tv_sec)
return TRUE;
if (val1->tv_sec > val2->tv_sec)
return FALSE;
/* val1->tv_sec == val2->tv_sec */
if (val1->tv_usec < val2->tv_usec)
return TRUE;
return FALSE;
}
static gboolean
g_timeval_eq (GTimeVal *val1,
GTimeVal *val2)
{
return (val1->tv_sec == val2->tv_sec) && (val1->tv_usec == val2->tv_usec);
}
static void
ik_pair_events (ik_event_internal_t *event1,
ik_event_internal_t *event2)
{
g_assert (event1 && event2);
/* We should only be pairing events that have the same cookie */
g_assert (event1->event->cookie == event2->event->cookie);
/* We shouldn't pair an event that already is paired */
g_assert (event1->pair == NULL && event2->pair == NULL);
/* Pair the internal structures and the ik_event_t structures */
event1->pair = event2;
event1->event->pair = event2->event;
if (g_timeval_lt (&event1->hold_until, &event2->hold_until))
event1->hold_until = event2->hold_until;
event2->hold_until = event1->hold_until;
}
static void
ik_event_add_microseconds (ik_event_internal_t *event,
glong ms)
{
g_assert (event);
g_time_val_add (&event->hold_until, ms);
}
static gboolean
ik_event_ready (ik_event_internal_t *event)
{
GTimeVal tv;
g_assert (event);
g_get_current_time (&tv);
/* An event is ready if,
*
* it has no cookie -- there is nothing to be gained by holding it
* or, it is already paired -- we don't need to hold it anymore
* or, we have held it long enough
*/
return
event->event->cookie == 0 ||
event->pair != NULL ||
g_timeval_lt (&event->hold_until, &tv) ||
g_timeval_eq (&event->hold_until, &tv);
}
static void
ik_pair_moves (gpointer data,
gpointer user_data)
{
ik_event_internal_t *event = (ik_event_internal_t *)data;
if (event->seen == TRUE || event->sent == TRUE)
return;
if (event->event->cookie != 0)
{
/* When we get a MOVED_FROM event we delay sending the event by
* MOVE_HOLD_UNTIL_TIME microseconds. We need to do this because a
* MOVED_TO pair _might_ be coming in the near future */
if (event->event->mask & IN_MOVED_FROM)
{
g_hash_table_insert (cookie_hash, GINT_TO_POINTER (event->event->cookie), event);
/* because we don't deliver move events there is no point in waiting for the match right now. */
ik_event_add_microseconds (event, MOVE_HOLD_UNTIL_TIME);
}
else if (event->event->mask & IN_MOVED_TO)
{
/* We need to check if we are waiting for this MOVED_TO events cookie to pair it with
* a MOVED_FROM */
ik_event_internal_t *match = NULL;
match = g_hash_table_lookup (cookie_hash, GINT_TO_POINTER (event->event->cookie));
if (match)
{
g_hash_table_remove (cookie_hash, GINT_TO_POINTER (event->event->cookie));
ik_pair_events (match, event);
}
}
}
event->seen = TRUE;
}
static void
ik_process_events (void)
{
g_queue_foreach (events_to_process, ik_pair_moves, NULL);
while (!g_queue_is_empty (events_to_process))
{
ik_event_internal_t *event = g_queue_peek_head (events_to_process);
/* This must have been sent as part of a MOVED_TO/MOVED_FROM */
if (event->sent)
{
/* Pop event */
g_queue_pop_head (events_to_process);
/* Free the internal event structure */
g_free (event);
continue;
}
/* The event isn't ready yet */
if (!ik_event_ready (event))
break;
/* Pop it */
event = g_queue_pop_head (events_to_process);
/* Check if this is a MOVED_FROM that is also sitting in the cookie_hash */
if (event->event->cookie && event->pair == NULL &&
g_hash_table_lookup (cookie_hash, GINT_TO_POINTER (event->event->cookie)))
g_hash_table_remove (cookie_hash, GINT_TO_POINTER (event->event->cookie));
if (event->pair)
{
/* We send out paired MOVED_FROM/MOVED_TO events in the same event buffer */
/* g_assert (event->event->mask == IN_MOVED_FROM && event->pair->event->mask == IN_MOVED_TO); */
/* Copy the paired data */
event->pair->sent = TRUE;
event->sent = TRUE;
ik_move_matches++;
}
else if (event->event->cookie)
{
/* If we couldn't pair a MOVED_FROM and MOVED_TO together, we change
* the event masks */
/* Changeing MOVED_FROM to DELETE and MOVED_TO to create lets us make
* the gaurantee that you will never see a non-matched MOVE event */
if (event->event->mask & IN_MOVED_FROM)
{
event->event->mask = IN_DELETE|(event->event->mask & IN_ISDIR);
ik_move_misses++; /* not super accurate, if we aren't watching the destination it still counts as a miss */
}
if (event->event->mask & IN_MOVED_TO)
event->event->mask = IN_CREATE|(event->event->mask & IN_ISDIR);
}
/* Push the ik_event_t onto the event queue */
g_queue_push_tail (event_queue, event->event);
/* Free the internal event structure */
g_free (event);
}
}
static gboolean
ik_process_eq_callback (gpointer user_data)
{
gboolean res;
/* Try and move as many events to the event queue */
G_LOCK (inotify_lock);
ik_process_events ();
while (!g_queue_is_empty (event_queue))
{
ik_event_t *event = g_queue_pop_head (event_queue);
user_cb (event);
}
res = TRUE;
if (g_queue_get_length (events_to_process) == 0)
{
process_eq_running = FALSE;
res = FALSE;
}
G_UNLOCK (inotify_lock);
return res;
}