From: Jeff King <firstname.lastname@example.org>
To: ardi <email@example.com>
Subject: Re: Settings for minimizing repacking (and keeping 'rsync' happy)
Date: Mon, 29 Jul 2019 05:42:29 -0400 [thread overview]
Message-ID: <20190729094229.GA2415@sigill.intra.peff.net> (raw)
On Sun, Jul 28, 2019 at 01:41:34AM +0200, ardi wrote:
> Some of my Git repositories have mirrors, maintained with 'rsync'. I
> want to have some level of repacking, so that the repositories are
> efficient, but I also want it to minimize it, so that 'rsync' never
> has to perform a big transfer for the repositories.
Yes, this is a common problem. The solutions I've seen/used are:
- use a git-aware transport like git-fetch that can negotiate which
objects to send
- use a tool that can find duplicated chunks across files. Many
de-duping backup systems (e.g., borg) use a rolling hash similar to
rsync to find moveable chunks, but then look up those chunks in a
master index (whereas rsync is always looking to match chunks in a
file of the same name). This works well in practice because Git is
not usually rewriting most of the data, but just shuffling it around
In theory it shouldn't be that hard to tell the receiving rsync to
look for source chunks not just in the file of the same name, but
from a set of existing packfiles (say, everything already in
.git/objects/pack/ on the receiver). But I don't know offhand of an
option to rsync to do so.
> For example, I think it would be fine if files are repacked just once
> in their lifetimes, and then that resulting pack file is never
> repacked again. I did read the gc.bigPackThreshold and
> gc.autoPackLimit settings, but I don't think they would accomplish
> Basically, what I'm describing is the behaviour of not packing files
> until the resulting pack would be a given size (say 10MB for example),
> and then never repack such ~10MB packs again, ever.
> Can this be done with some Git settings? And do you foresee any kind
> of serious drawback or potential problem with this kind of behaviour?
You can mark a pack to be kept forever by creating a matching
"pack-1234abcd.keep" file. That doesn't do your automatic "I want 10MB
packs" thing, but if you did it occasionally at the right frequency,
you'd end up with a bunch of 10MB-ish packs.
But there are downsides to having a bunch of packs:
- object lookups are O(log n) within a single pack, but O(n) over the
number of packs. So if you get a very large number of packs, normal
operations will start to suffer. This is mitigated by the new "midx"
feature, which generates an index for multiple packs.
- git doesn't allow delta compression across packs. So imagine you
have ten versions of a file that's 5kb, and each version changes
about 100 bytes. In a single pack, we'd store one base object, plus
9 deltas, for a total of about 6kb (5000 + 9*100). Across two packs,
we'd store ~11kb (2*5000 + 8*100). And the worst case is ten packs
As a more real-world example, try this:
git -c pack.packsizelimit=10M repack -ad
In a fresh clone of git.git, the size of the pack directory jumps
from 88MB to 168MB. And in a time-based split (i.e., creating a new
10MB pack every week), it may be even worse. The command above
ordered the objects optimally to keep deltas together and _then_
split things. Whereas a time-based scheme would likely sprinkle
versions of a file across more packs.
It should be possible to loosen this restriction and allow
cross-pack deltas, but it would be very risky. The assumption that
packs are independent of each other is implicit in much of Git's
repacking code, so it would be easy to introduce a bug where we
generate a circular dependency (object A in pack X is a delta
against object B in pack Y, which is a delta against object A --
oops, we don't have a full copy anymore).
next prev parent reply other threads:[~2019-07-29 9:42 UTC|newest]
Thread overview: 5+ messages / expand[flat|nested] mbox.gz Atom feed top
2019-07-27 23:41 Settings for minimizing repacking (and keeping 'rsync' happy) ardi
2019-07-29 9:42 ` Jeff King [this message]
2019-07-29 12:56 ` Ævar Arnfjörð Bjarmason
2019-07-29 20:16 ` Jeff King
2019-07-29 14:35 ` Konstantin Ryabitsev
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
List information: http://vger.kernel.org/majordomo-info.html
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
Code repositories for project(s) associated with this public inbox
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for read-only IMAP folder(s) and NNTP newsgroup(s).