binutils-gdb/gdb/testsuite/gdb.threads/slow-waitpid.c
Andrew Burgess 7010835a6c gdb: Don't drop SIGSTOP during stop_all_threads
This patch fixes an issue where GDB would sometimes hang when
attaching to a multi-threaded process.  This issue was especially
likely to trigger if the machine (running the inferior) was under
load.

In summary, the problem is an imbalance between two functions in
linux-nat.c, stop_callback and stop_wait_callback.  In stop_callback
we send SIGSTOP to a thread, but _only_ if the thread is not already
stopped, and if it is not signalled, which means it should stop soon.
In stop_wait_callback we wait for the SIGSTOP to arrive, however, we
are aware that the thread might have been signalled for some other
reason, and so if a signal other than SIGSTOP causes the thread to
stop then we stash that signal away so it can be reported back later.
If we get a SIGSTOP then this is discarded, after all, this signal was
sent from stop_callback.  Except that this might not be the case, it
could be that SIGSTOP was sent to a thread from elsewhere in GDB, in
which case we would not have sent another SIGSTOP from stop_callback
and the SIGSTOP received in stop_wait_callback should not be ignored.

Below I've laid out the exact sequence of events that I saw that lead
me to track down the above diagnosis.

After attaching to the inferior GDB sends a SIGSTOP to all of the
threads and then returns to the event loop waiting for interesting
things to happen.

Eventually the first target event is detected (this will be the first
SIGSTOP arriving) and GDB calls inferior_event_handler which calls
fetch_inferior_event.  Inside fetch_inferior_event GDB calls
do_target_wait which calls target_wait to find a thread with an event.

The target_wait call ends up in linux_nat_wait_1, which first checks
to see if any threads already have stashed stop events to report, and
if there are none then we enter a loop fetching as many events as
possible out of the kernel.  This event fetching is non-blocking, and
we give up once the kernel has no more events ready to give us.

All of the events from the kernel are passed through
linux_nat_filter_event which stashes the wait status for all of the
threads that reported a SIGSTOP, these will be returned by future
calls to linux_nat_wait_1.

Lets assume for a moment that we've attached to a multi-threaded
inferior, and that all but one thread has reported its stop during the
initial wait call in linux_nat_wait_1.  The other thread will be
reporting a SIGSTOP, but the kernel has not yet managed to deliver
that signal to GDB before GDB gave up waiting and continued handling
the events it already had.  GDB selects one of the threads that has
reported a SIGSTOP and passes this thread ID back to
fetch_inferior_event.

To handle the thread's SIGSTOP, GDB calls handle_signal_stop, which
calls stop_all_threads, this calls wait_one, which in turn calls
target_wait.

The first call to target_wait at this point will result in a stashed
wait status being returned, at which point we call setup_inferior.
The call to setup_inferior leads to a call into try_thread_db_load_1
which results in a call to linux_stop_and_wait_all_lwps.  This in turn
calls stop_callback on each thread followed by stop_wait_callback on
each thread.

We're now ready to make the mistake.  In stop_callback we see that our
problem thread is not stopped, but is signalled, so it should stop
soon.  As a result we don't send another SIGSTOP.

We then enter stop_wait_callback, eventually the problem thread stops
with SIGSTOP which we _incorrectly_ assume came from stop_callback,
and we discard.

Once stop_wait_callback has done its damage we return from
linux_stop_and_wait_all_lwps, finish in try_thread_db_load_1, and
eventually unwind back to the call to setup_inferior in
stop_all_threads.  GDB now loops around, and performs another
target_wait to get the next event from the inferior.

The target_wait calls causes us to once again reach linux_nat_wait_1,
and we pass through some code that calls resume_stopped_resumed_lwps.
This allows GDB to resume threads that are physically stopped, but
which GDB doesn't see any good reason for the thread to remain
stopped.  In our case, the problem thread which had its SIGSTOP
discarded is stopped, but doesn't have a stashed wait status to
report, and so GDB sets the thread going again.

We are now stuck waiting for an event on the problem thread that might
never arrive.

When considering how to write a test for this bug I struggled.  The
issue was only spotted _randomly_ when a machine was heavily loaded
with many multi-threaded applications, and GDB was being attached (by
script) to all of these applications in parallel.  In one reproducer I
required around 5 applications each of 5 threads per machine core in
order to reproduce the bug 2 out of 3 times.

What we really want to do though is simulate the kernel being slow to
report events through waitpid during the initial attach.  The solution
I came up with was to write an LD_PRELOAD library that intercepts
(some) waitpid calls and rate limits them to one per-second.  Any more
than that simply return 0 indicating there's no event available.
Obviously this can only be applied to waitpid calls that have the
WNOHANG flag set.

Unfortunately, once you ignore a waitpid call GDB can get a bit stuck.
Usually, once the kernel has made a child status available to waitpid
GDB will be sent a SIGCHLD signal.  However, if the kernel makes 5
child statuses available but, due to the preload library we only
collect one of them, then the kernel will not send any further SIGCHLD
signals, and so, when GDB, thinking that the remaining statuses have
not yet arrived sits waiting for a SIGCHLD it will be disappointed.

The solution, implemented within the preload library, is that, when we
hold back a waitpid result from GDB we spawn a new thread.  This
thread delays for a short period, and then sends GDB a SIGCHLD.  This
causes GDB to retry the waitpid, at which point sufficient time has
passed and our library allows the waitpid call to complete.

gdb/ChangeLog:

	* linux-nat.c (stop_wait_callback): Don't discard SIGSTOP if it
	was requested by GDB.

gdb/testsuite/ChangeLog:

	* gdb.threads/attach-slow-waitpid.c: New file.
	* gdb.threads/attach-slow-waitpid.exp: New file.
	* gdb.threads/slow-waitpid.c: New file.
2018-06-16 01:03:57 +01:00

343 lines
10 KiB
C

/* This testcase is part of GDB, the GNU debugger.
Copyright 2018 Free Software Foundation, Inc.
This program is free software; you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation; either version 3 of the License, or
(at your option) any later version.
This program is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with this program. If not, see <http://www.gnu.org/licenses/>. */
/* This file contains a library that can be preloaded into GDB on Linux
using the LD_PRELOAD technique.
The library intercepts calls to WAITPID and SIGSUSPEND in order to
simulate the behaviour of a heavily loaded kernel.
When GDB wants to stop all threads in an inferior each thread is sent a
SIGSTOP, GDB will then wait for the signal to be received by the thread
with a waitpid call.
If the kernel is slow in either delivering the signal, or making the
result available to the waitpid call then GDB will enter a sigsuspend
call in order to wait for the inferior threads to change state, this is
signalled to GDB with a SIGCHLD.
A bug in GDB meant that in some cases we would deadlock during this
process. This was rarely seen as the kernel is usually quick at
delivering signals and making the results available to waitpid, so quick
that GDB would gather the statuses from all inferior threads in the
original pass.
The idea in this library is to rate limit calls to waitpid (where pid is
-1 and the WNOHANG option is set) so that only 1 per second can return
an answer. Any additional calls will report that no threads are
currently ready. This should match the behaviour we see on a slow
kernel.
However, given that usually when using this library, the kernel does
have the waitpid result ready this means that the kernel will never send
GDB a SIGCHLD. This means that when GDB enters sigsuspend it will block
forever. Alternatively, if GDB enters its polling loop the lack of
SIGCHLD means that we will never see an event on the child threads. To
resolve these problems the library intercepts calls to sigsuspend and
forces the call to exit if there is a pending waitpid result. Also,
when we know that there's a waitpid result that we've ignored, we create
a new thread which, after a short delay, will send GDB a SIGCHLD. */
#define _GNU_SOURCE
#include <sys/types.h>
#include <sys/wait.h>
#include <sys/time.h>
#include <stdlib.h>
#include <stdio.h>
#include <dlfcn.h>
#include <string.h>
#include <stdarg.h>
#include <signal.h>
#include <errno.h>
#include <pthread.h>
#include <unistd.h>
/* Logging. */
static void
log_msg (const char *fmt, ...)
{
#ifdef LOGGING
va_list ap;
va_start (ap, fmt);
vfprintf (stderr, fmt, ap);
va_end (ap);
#endif /* LOGGING */
}
/* Error handling, message and exit. */
static void
error (const char *fmt, ...)
{
va_list ap;
va_start (ap, fmt);
vfprintf (stderr, fmt, ap);
va_end (ap);
exit (EXIT_FAILURE);
}
/* Cache the result of a waitpid call that has not been reported back to
GDB yet. We only ever cache a single result. Once we have a result
cached then later calls to waitpid with the WNOHANG option will return a
result of 0. */
static struct
{
/* Flag to indicate when we have a result cached. */
int cached_p;
/* The cached result fields from a waitpid call. */
pid_t pid;
int wstatus;
} cached_wait_status;
/* Lock to hold when modifying SIGNAL_THREAD_ACTIVE_P. */
static pthread_mutex_t thread_creation_lock_obj = PTHREAD_MUTEX_INITIALIZER;
#define thread_creation_lock (&thread_creation_lock_obj)
/* This flag is only modified while holding the THREAD_CREATION_LOCK mutex.
When this flag is true then there is a signal thread alive that will be
sending a SIGCHLD at some point in the future. */
static int signal_thread_active_p;
/* When we last allowed a waitpid to complete. */
static struct timeval last_waitpid_time = { 0, 0 };
/* The number of seconds that must elapse between calls to waitpid where
the pid is -1 and the WNOHANG option is set. If calls occur faster than
this then we force a result of 0 to be returned from waitpid. */
#define WAITPID_MIN_TIME (1)
/* Return true (non-zero) if we should skip this call to waitpid, or false
(zero) if this waitpid call should be handled with a call to the "real"
waitpid function. Allows 1 waitpid call per second. */
static int
should_skip_waitpid (void)
{
struct timeval *tv = &last_waitpid_time;
if (tv->tv_sec == 0)
{
if (gettimeofday (tv, NULL) < 0)
error ("error: gettimeofday failed\n");
return 0; /* Don't skip. */
}
else
{
struct timeval new_tv;
if (gettimeofday (&new_tv, NULL) < 0)
error ("error: gettimeofday failed\n");
if ((new_tv.tv_sec - tv->tv_sec) < WAITPID_MIN_TIME)
return 1; /* Skip. */
*tv = new_tv;
}
/* Don't skip. */
return 0;
}
/* Perform a real waitpid call. */
static pid_t
real_waitpid (pid_t pid, int *wstatus, int options)
{
typedef pid_t (*fptr_t) (pid_t, int *, int);
static fptr_t real_func = NULL;
if (real_func == NULL)
{
real_func = dlsym (RTLD_NEXT, "waitpid");
if (real_func == NULL)
error ("error: failed to find real waitpid\n");
}
return (*real_func) (pid, wstatus, options);
}
/* Thread worker created when we cache a waitpid result. Delays for a
short period of time and then sends SIGCHLD to the GDB process. This
should trigger GDB to call waitpid again, at which point we will make
the cached waitpid result available. */
static void*
send_sigchld_thread (void *arg)
{
/* Delay one second longer than WAITPID_MIN_TIME so that there can be no
chance that a call to SHOULD_SKIP_WAITPID will return true once the
SIGCHLD is delivered and handled. */
sleep (WAITPID_MIN_TIME + 1);
pthread_mutex_lock (thread_creation_lock);
signal_thread_active_p = 0;
if (cached_wait_status.cached_p)
{
log_msg ("signal-thread: sending SIGCHLD\n");
kill (getpid (), SIGCHLD);
}
pthread_mutex_unlock (thread_creation_lock);
return NULL;
}
/* The waitpid entry point function. */
pid_t
waitpid (pid_t pid, int *wstatus, int options)
{
log_msg ("waitpid: waitpid (%d, %p, 0x%x)\n", pid, wstatus, options);
if ((options & WNOHANG) != 0
&& pid == -1
&& should_skip_waitpid ())
{
if (!cached_wait_status.cached_p)
{
/* Do the waitpid call, but hold the result back. */
pid_t tmp_pid;
int tmp_wstatus;
tmp_pid = real_waitpid (-1, &tmp_wstatus, options);
if (tmp_pid > 0)
{
log_msg ("waitpid: delaying waitpid result (pid = %d)\n",
tmp_pid);
/* Cache the result. */
cached_wait_status.pid = tmp_pid;
cached_wait_status.wstatus = tmp_wstatus;
cached_wait_status.cached_p = 1;
/* Is there a thread around that will be sending a signal in
the near future? The prevents us from creating one
thread per call to waitpid when the calls occur in a
sequence. */
pthread_mutex_lock (thread_creation_lock);
if (!signal_thread_active_p)
{
sigset_t old_ss, new_ss;
pthread_t thread_id;
pthread_attr_t attr;
/* Create the new signal sending thread in detached
state. This means that the thread doesn't need to be
pthread_join'ed. Which is fine as there's no result
we care about. */
pthread_attr_init (&attr);
pthread_attr_setdetachstate (&attr, PTHREAD_CREATE_DETACHED);
/* Ensure the signal sending thread has all signals
blocked. We don't want any signals to GDB to be
handled in that thread. */
sigfillset (&new_ss);
sigprocmask (SIG_BLOCK, &new_ss, &old_ss);
log_msg ("waitpid: spawn thread to signal us\n");
if (pthread_create (&thread_id, &attr,
send_sigchld_thread, NULL) != 0)
error ("error: pthread_create failed\n");
signal_thread_active_p = 1;
sigprocmask (SIG_SETMASK, &old_ss, NULL);
pthread_attr_destroy (&attr);
}
pthread_mutex_unlock (thread_creation_lock);
}
}
log_msg ("waitpid: skipping\n");
return 0;
}
/* If we have a cached result that is a suitable reply for this call to
waitpid then send that cached result back now. */
if (cached_wait_status.cached_p
&& (pid == -1 || pid == cached_wait_status.pid))
{
pid_t pid;
pid = cached_wait_status.pid;
log_msg ("waitpid: return cached result (%d)\n", pid);
*wstatus = cached_wait_status.wstatus;
cached_wait_status.cached_p = 0;
return pid;
}
log_msg ("waitpid: real waitpid call\n");
return real_waitpid (pid, wstatus, options);
}
/* Perform a real sigsuspend call. */
static int
real_sigsuspend (const sigset_t *mask)
{
typedef int (*fptr_t) (const sigset_t *);
static fptr_t real_func = NULL;
if (real_func == NULL)
{
real_func = dlsym (RTLD_NEXT, "sigsuspend");
if (real_func == NULL)
error ("error: failed to find real sigsuspend\n");
}
return (*real_func) (mask);
}
/* The sigsuspend entry point function. */
int
sigsuspend (const sigset_t *mask)
{
log_msg ("sigsuspend: sigsuspend (0x%p)\n", ((void *) mask));
/* If SIGCHLD is _not_ in MASK, and is therefore deliverable, then if we
have a pending wait status pretend that a signal arrived. We will
have a thread alive that is going to deliver a signal but doing this
will boost the speed as we don't have to wait for a signal. If the
signal ends up being delivered then it should be harmless, we'll just
perform an additional waitpid call. */
if (!sigismember (mask, SIGCHLD))
{
if (cached_wait_status.cached_p)
{
log_msg ("sigsuspend: interrupt for cached waitstatus\n");
last_waitpid_time.tv_sec = 0;
last_waitpid_time.tv_usec = 0;
errno = EINTR;
return -1;
}
}
log_msg ("sigsuspend: real sigsuspend call\n");
return real_sigsuspend (mask);
}