git-bigfiles - Git for big files

git-bigfiles is our fork of Git. It has two goals:

  • make life bearable for people using Git on projects hosting very large files (hundreds of megabytes)
  • merge back as many changes as possible into upstream Git once they’re of acceptable quality


git-bigfiles already features the following fixes:

  • git-config: add core.bigFileThreshold option to treat files larger than this value as big files
  • git-p4: fix poor string building performance when importing big files (git-p4 is now only marginally faster on Linux but 4 to 10 times faster on Win32)
  • git-p4: do not perform keyword replacements on big files
  • git-fast-import: do not perform delta search on big files and deflate them on-the-fly (fast-import is now three times as fast and uses 5 times less memory with big files)
  • git-pack-objects: do not perform delta search on big files (pack-objects is now 10% faster, but there is no memory gain yet)


Git fast-import memory usage and running time when importing a repository with binary files up to 150 MiB, then git-gc on the same repository:


git-bigfiles development is centralised in a Git repository.

Clone the repository:

git clone git://

If you already have a working copy of upstream git, you may save a lot of bandwidth by doing:

git clone git:// --reference /path/to/git/

The main Git repository is constantly merged into git-bigfiles. See the git-bigfiles repository on

To do

The following commands are real memory hogs with big files:

  • git diff-tree --stat --summary -M <sha1> HEAD
  • git gc (as called by git merge) fails with warning: suboptimal pack - out of memory


Last modified 15 years ago Last modified on 03/24/2009 09:45:21 PM

Attachments (3)

Download all attachments as: .zip