We are using git for a multi site project and both the repo and the team have grown substantially. Therefore remote sites are suffering. I would like to add new git servers to the remote sites to decrease the load on our git server and to make the downloads faster. However I am not sure how to keep these synchronized. Is there any way to replicate git servers atomically? And if not any other suggestions?
相关问题
- How to account for clock offsets in a distributed
- Why does recursive submodule update from github fa
- Extended message for commit via Visual Studio Code
- Emacs shell: save commit message
- Can I organize Git submodules in a flat hierarchy?
相关文章
- 请教Git如何克隆本地库?
- GitHub:Enterprise post-receive hook
- Git Clone Fails: Server Certificate Verification F
- SSIS solution on GIT?
- Is there a version control system abstraction for
- ssh: Could not resolve hostname git: Name or servi
- Cannot commit changes with gitextensions
- git: retry if http request failed
You can't replicate atomically. I'm kind of hoping that that's not what you mean.
I wrote gitmirror for automatically triggering a remote pull-based replication on repository change. It requires a tiny bit of setup, but I've got tons of stuff coming through github into it (and a few that don't use github).
You could have a main "write" repo, and multiple remote, read only, repos pull from that. In turn, remote users would pull from those. That would give you increased download speeds from remote sites. Uploads would still be the same, but it seems like download speed is your main concern.