Version 9 (modified by 16 years ago) (diff) | ,
---|
git-bigfiles - Git for big files
git-bigfiles is our fork of Git. It has two goals:
- make life bearable for people using Git on projects hosting very large files (hundreds of megabytes)
- merge back as many changes as possible into upstream Git
Features
git-bigfiles already features the following fixes:
- config: add
core.bigFileThreshold
option to treat files larger than this value as big files - git-p4: fix abysmal performance when importing big files (git-p4 is now generally 4 times faster with big files, but can be up to 8000 times faster on Win32)
- fast-import: avoid computing deltas on big files and deflate them on-the-fly (fast-import is now twice as fast and uses 3.7 times less memory with big files)
Results
Git fast-import
memory usage and running time when importing a repository with binary files up to 150 MiB:
Download
Clone the git-bigfiles repository:
git clone git://repo.or.cz/git/git-bigfiles.git
If you already have a working copy of upstream git, you may save a lot of bandwidth by doing:
git clone git://repo.or.cz/git/git-bigfiles.git --reference /path/to/git/
The main Git repository is constantly merged into git-bigfiles. See the git-bigfiles repository on repo.or.cz.
Attachments (3)
- git-bigfiles.png (57.8 KB) - added by 16 years ago.
- fast-import-memory.png (26.9 KB) - added by 16 years ago.
- pack-objects-memory.png (17.4 KB) - added by 16 years ago.
Download all attachments as: .zip