This is the mail archive of the libc-alpha@sourceware.org mailing list for the glibc project.


Index Nav: [Date Index] [Subject Index] [Author Index] [Thread Index]
Message Nav: [Date Prev] [Date Next] [Thread Prev] [Thread Next]
Other format: [Raw text]

Re: [PATCH RFC] introduce dl_iterate_phdr_parallel


On Mon, 2016-08-01 at 23:41 +0300, Gleb Natapov wrote:
> On Mon, Aug 01, 2016 at 10:19:55PM +0200, Torvald Riegel wrote:
> > On Mon, 2016-08-01 at 23:07 +0300, Gleb Natapov wrote:
> > > On Mon, Aug 01, 2016 at 09:46:35PM +0200, Torvald Riegel wrote:
> > > > If we built something custom for this and are willing to make the
> > > > wrlock / exclusive-access case much more costly, we can decrease this
> > > > overhead.  This could be roughly similar to one lock per thread or a set
> > > > of rwlocks as you mentioned, but with less space overhead.
> > > > 
> > > IMO space overhead is negligible. More efficient rwlock is, of course,
> > > better and can be useful in many more places. If you have something to
> > > test I am willing to do so, but if custom rwlock will take time to
> > > materialize we may start with lock array and change it later. The lock
> > > array is not part of the interface, but implementation detail. What I
> > > would like to avoid is stalling the afford while waiting for something
> > > better. Exception scalability is very pressing issue for us.
> > 
> > I haven't looked at the workload in detail, and so can't say off-hand
> > what would be best and how long it would take.  Is there a summary
> > describing the synchronization that is necessary?  Once I know that, I
> > can give a closer estimate.  It could be simple to do, so it might be
> > worth having a look before trying the (rw?)lock-array approach.
> > 
> rwlock semantics is pretty much what is needed. There is a list of
> loaded objects that changes rarely (only during dlopen/dlclose), but it
> is accessed for read on each exception.

Does "for read" mean that the only thing you are interested in is some
consistent snapshot of the list, or something like a yes/no result for a
query about whether some element is contained?
Is this list in memory that is never unmapped?  Can we construct it in
such a way that we'll follow only a few pointers for which we don't
quite know whether they point to proper elements of the list?

If we can do that, than we should be able to run an atomic snapshot
speculatively.  IOW, do what many STMs do, or seqlocks, etc.

> Exceptions on different threads
> are independent, so they should not interfere with each other. Having
> only one lock (even rw one) means that all threads will access same
> atomic variable for synchronisation which has a lot of overhead on
> modern cpus. lock-array approach is a way to allow multiple threads to
> process completely independently without any inter-core communication at
> all. It is hard to beat.

You can try to never modify memory in the common case, which is better.
Additionally, the problem with the lock array is that you have to put
the locks somewhere, preferably so that they are in a page that is local
to where the thread runs.  This makes a simple lock array suboptimal, at
least on bigger systems.  TLS space would be most likely to be backed by
local memory, but as Adhemerval has mentioned, we might have lots of
threads, and if we do TLS, we need to sync between thread destruction
and the heavyweight operation that wants to acquire all locks.

> New rwlock is substantially slower on only 4
> cores (it is much better than current rwlock since it avoids entering
> kernel if there are only readers).

Is this a microbenchmark just for acquisition/release of the rwlock, or
does this significant difference also show up in the real scenario you
are concerned about (the exception storm you mentioned)?

> > The POSIX rwlock's complexity is partly due to having to support the
> > different modes (eg, reader preference), so a custom rwlock for a
> > specific use case could be much simpler.  I would guess that it looks
> > more like reference counting, or if the snapshots of the readers are
> > suitable, it could also be seqlock based (and thus require no memory
> > modification by readers at all).
> > 
> I evaluated seqlock approach, but snapshoting of the object list requires
> pretty deep copy (for exception handling case)

Why do you need to copy?  Can you point me to the list datastructure you
need to snapshot?

> and the data accessed
> varies for each user of dl_iterate_phdr (elements of the list point into
> mmaped parts of .so files).

That doesn't matter.  The only overhead is how often we need to validate
so that we know that a subsequent pointer dereference is safe.  The
number of these valuations compared to a lock acquisition+release is the
deciding factor here.


Index Nav: [Date Index] [Subject Index] [Author Index] [Thread Index]
Message Nav: [Date Prev] [Date Next] [Thread Prev] [Thread Next]