tech-toolchain archive

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index][Old Index]

Re: ld.elf_so locking



On Mon, Mar 14, 2011 at 04:37:21PM +0100, Joerg Sonnenberger wrote:

> Comments?

Only a few right now..

> There are currently two situations left where a recursion would currently
> result in a dead lock: a signal handler or constructor using __tls_get_addr
> requiring memory allocation when already in ld.elf_so and a signal
> handler occuring during exclusively locked calls like dlopen trying to
> access a shared lock e.g. to lazily resolve a relocation.
> 
> The second can be mostly be adressed by using sigprocmask in the

For stuff like that one idea is to have the signal mask directly in the
_lwp_ctl area, so that sigprocmask() isn't any kind of performance concern. 
The signal mask is thread private state in the kernel with a couple of
exceptions, Oe can be worked around and I think the other relates to SA (and
so can die).  As a perhipheral benefit this could be a nice boost to some
workloads since sigprocmask() is called a lot.

On a related note is there a thread-private area that we can reserve to
point to the _lwp_ctl block?

> +static volatile bool _rtld_mutex_may_recurse;
> +

Prefer "int" or sigatomic_t as bool may be sub-word sized and thus cause
atomicity problems.  Maybe this is not a concern (although volatile
suggests it is).

> +#ifdef __HAVE_FUNCTION_DESCRIPTORS
> +#define      lookup_mutex_enter()    _rtld_exclusive_enter()
> +#define      lookup_mutex_exit()     _rtld_exclusive_exit()
> +#else
> +#define      lookup_mutex_enter()    _rtld_shared_enter()
> +#define      lookup_mutex_exit()     _rtld_shared_exit()
> +#endif

Support for named lock objects would be nice (e.g. rtldlock_t *lock as
parameter, or "int lock" if the former would cause additional relocs).

> +void
> +_rtld_shared_enter(void)
> +{
> +     unsigned int cur;
> +     lwpid_t waiter, self = 0;
> +
> +     membar_enter();

membar_enter() should be after lock acquire point.  Suggest wrapping in
#ifndef __HAVE_ATOMIC_AS_MEMBAR.

> +     for (;;) {
> +             cur = _rtld_mutex;
> +             /*
> +              * First check if we are currently not exclusively locked.
> +              */
> +             if ((cur & RTLD_EXCLUSIVE_MASK) == 0) {
> +                     /* Yes, so increment use counter */
> +                     if (atomic_cas_uint(&_rtld_mutex, cur, cur + 1) != cur)
> +                             continue;
> +                     return;
> +             }
> +             /*
> +              * Someone has an exclusive lock.  Puts us on the waiter list.
> +              */
> +             if (!self)
> +                     self = _lwp_self();
> +             if (cur == (self | RTLD_EXCLUSIVE_MASK)) {
> +                     if (_rtld_mutex_may_recurse)
> +                             return;
> +                     _rtld_error("dead lock detected");
> +                     _rtld_die();
> +             }
> +             waiter = atomic_swap_uint(&_rtld_waiter_shared, self);
> +             /*
> +              * Check for race against _rtld_exclusive_exit before sleeping.
> +              */
> +             if ((_rtld_mutex & RTLD_EXCLUSIVE_MASK) ||
> +                 _rtld_waiter_exclusive)
> +                     _lwp_park(NULL, -1, __UNVOLATILE(&_rtld_mutex), NULL);
> +             /* Try to remove us from the waiter list. */
> +             atomic_cas_uint(&_rtld_waiter_shared, self, 0);
> +             if (waiter)
> +                     _lwp_unpark(waiter, __UNVOLATILE(&_rtld_mutex));
> +     }
> +}

This makes me a bit nervous since it's sort of hard to follow whats going
on.  Some further comments would be nice.  A comment about the realtime
restriction that we can't just go and do a dumb spin on the lock would be
good too (as described in pthread_mutex.c).


Home | Main Index | Thread Index | Old Index