Version 4 (modified by 16 years ago) (diff) | ,
---|
git-bigfiles - Git for big files
git-bigfiles is our fork of Git. It has two goals:
- make life bearable for people using Git on projects hosting very large files (hundreds of megabytes)
- merge back as many changes as possible into upstream Git
Features
git-bigfiles already features the following fixes:
- config: add
core.bigFileThreshold
option to treat files larger than this value as big files - git-p4: fix abysmal performance when importing big files (operation is now up to 8000 times faster, especially on Win32)
- fast-import: do not compute deltas on big files and deflate them on-the-fly (operation is now twice as fast and use 3.7 times less memory)
Download
Clone the git-bigfiles repository:
git clone git://repo.or.cz/git/git-bigfiles.git
If you already have a working copy of upstream git, you may save a lot of bandwidth by doing:
git clone git://repo.or.cz/git/git-bigfiles.git --reference /path/to/git/
The main Git repository is constantly merged into git-bigfiles. See the git-bigfiles repository on repo.or.cz.
Attachments (3)
- git-bigfiles.png (57.8 KB) - added by 16 years ago.
- fast-import-memory.png (26.9 KB) - added by 16 years ago.
- pack-objects-memory.png (17.4 KB) - added by 16 years ago.
Download all attachments as: .zip