From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.1 (2015-04-28) on dcvr.yhbt.net X-Spam-Level: X-Spam-ASN: AS31976 209.132.180.0/23 X-Spam-Status: No, score=-3.9 required=3.0 tests=AWL,BAYES_00,DKIM_SIGNED, DKIM_VALID,HEADER_FROM_DIFFERENT_DOMAINS,MAILING_LIST_MULTI,RCVD_IN_DNSWL_MED, SPF_HELO_PASS,SPF_PASS shortcircuit=no autolearn=ham autolearn_force=no version=3.4.1 Received: from sourceware.org (server1.sourceware.org [209.132.180.131]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by dcvr.yhbt.net (Postfix) with ESMTPS id 433351F597 for ; Wed, 1 Aug 2018 06:24:00 +0000 (UTC) DomainKey-Signature: a=rsa-sha1; c=nofws; d=sourceware.org; h=list-id :list-unsubscribe:list-subscribe:list-archive:list-post :list-help:sender:date:from:to:cc:subject:message-id:references :mime-version:content-type:in-reply-to; q=dns; s=default; b=emjj ZoEXKarPxQ+0phZUCnWn8ZXXS7FoYjwLoPlaRkILrHWwG+ydHmrdIAt0pQPlZWOu Zr7mFokO1/EB/x+RYGDb9Pd+vnXzaFvp0VtKnNhOcalxoPa5r9Y3raaXRLVGE41U v/avMDFOAhuDOkNYDkAgcShEAS9E9kBPjuxmg64= DKIM-Signature: v=1; a=rsa-sha1; c=relaxed; d=sourceware.org; h=list-id :list-unsubscribe:list-subscribe:list-archive:list-post :list-help:sender:date:from:to:cc:subject:message-id:references :mime-version:content-type:in-reply-to; s=default; bh=P4S4y+HaRh 0VJ9IUEjOv002Q+DI=; b=bO2tuITeyl4yJQV/4sAAF5KggqJHlEbvgVTAPqDJqR K3yWkzjh/86F1455xnquMixwjGHRdHHzq1vEuahn07Gf8erynarFPOiqBb40f2Eb Ua0filIFXXzhN8VD/qX3NLzMNJTX1iMcTzga780q7NhvpFRkjkAMcEWPMhrkIkId o= Received: (qmail 72234 invoked by alias); 1 Aug 2018 06:23:57 -0000 Mailing-List: contact libc-alpha-help@sourceware.org; run by ezmlm Precedence: bulk List-Id: List-Unsubscribe: List-Subscribe: List-Archive: List-Post: List-Help: , Sender: libc-alpha-owner@sourceware.org Received: (qmail 72220 invoked by uid 89); 1 Aug 2018 06:23:56 -0000 Authentication-Results: sourceware.org; auth=none X-HELO: dcvr.yhbt.net Date: Wed, 1 Aug 2018 06:23:52 +0000 From: Eric Wong To: Carlos O'Donell Cc: libc-alpha@sourceware.org Subject: Re: [RFC/PoC] malloc: use wfcqueue to speed up remote frees Message-ID: <20180801062352.rlrjqmsszntkzlfe@untitled> References: <20180731084936.g4yw6wnvt677miti@dcvr> <0cfdccea-d173-486c-85f4-27e285a30a1a@redhat.com> <20180731231819.57xsqvdfdyfxrzy5@whir> MIME-Version: 1.0 Content-Type: text/plain; charset=utf-8 Content-Disposition: inline In-Reply-To: Carlos O'Donell wrote: > On 07/31/2018 07:18 PM, Eric Wong wrote: > >> - Can you explain the RSS reduction given this patch? You > >> might think that just adding the frees to a queue wouldn't > >> result in any RSS gains. > > > > At least two reasons I can see: > > > > 1) With lock contention, the freeing thread can lose to the > > allocating thread. This makes the allocating thread hit > > sysmalloc since it prevented the freeing thread from doing > > its job. sysmalloc is the slow path, so the lock gets held > > even longer and the problem compounds from there. > > How does this impact RSS? It would only block the remote thread > from freeing in a timely fashion, but it would eventually make > progress. Blocking the freeing thread causes the allocating thread to sysmalloc more. If the freeing thread could always beat the allocating thread, then the freed memory would be available in the arena by the time the allocating thread takes the lock. > > 2) thread caching - memory ends up in the wrong thread and > > could never get used in some cases. Fortunately this is > > bounded, but still a waste. > > We can't have memory end up in the wrong thread. The remote thread > computes the arena from the chunk it has, and then frees back to > the appropriate arena, even if it's not the arena that the thread > is attached to. Really? I see: __libc_free -> MAYBE_INIT_TCACHE && _int_free -> tcache_put I am not seeing anything in _int_free which makes the tcache_put arena-aware. If we drop MAYBE_INIT_TCACHE from __libc_free, then the tcache_put could be avoided. > > I'm still new to the code, but it looks like threads are pinned > > to the arena and the memory used for arenas never gets released. > > Is that correct? > > Threads are pinned to their arenas, but they can move in the event > of allocation failures, particularly to the main arena to attempt > sbrk to get more memory. OK. > > I was wondering if there was another possibility: the allocating > > thread gives up the arena and creates a new one because the > > freeing thread locked it, but I don't think that's the case. > > No. > > > Also, if I spawn a bunch of threads and get a bunch of > > arenas early in the program lifetime; and then only have few > > threads later, there can be a lot of idle arenas. > > Yes. That is true. We don't coalesce arenas to match the thread > demand. Eep :< If contention can be avoided (which tcache seems to work well for), limiting arenas to CPU count seems desirable and worth trying. > >> - Adding urcu as a build-time dependency is not acceptable for > >> bootstrap, instead we would bundle a copy of urcu and keep it > >> in sync with upstream. Would that make your work easier? > > > > Yes, bundling that sounds great. I assume it's something for > > you or one of the regular contributors to work on (build systems > > scare me :x) > > Yes, that is something we'd have to do. OK, I noticed my patch fails conformance tests because (despite my use of __cds_wfcq_splice_nonblocking) it references poll(), despite poll() being in an impossible code path: __cds_wfcq_splice_nonblocking -> ___cds_wfcq_splice -> ___cds_wfcq_busy_wait -> poll The poll call is impossible because the `blocking' parameter is 0; but I guess the linker doesn't know that? > >> - What problems are you having with `make -j4 check?' Try > >> master and report back. We are about to release 2.28 so it > >> should build and pass. > > > > My fault. It seems like tests aren't automatically rerun when I > > change the code; so some of my broken work-in-progress changes > > ended up being false positives :x. When working on this, I made > > the mistake of doing remote_free_step inside malloc_consolidate, > > which could recurse into _int_free or _int_malloc > > This depends a bit on what you touch. Alright, I'll keep that in mind. Thanks!