1
0
Fork 0
mirror of https://git.sr.ht/~seirdy/seirdy.one synced 2024-12-26 02:22:09 +00:00
seirdy.one/content/posts/git-workflow-1.md

85 lines
3.9 KiB
Markdown
Raw Normal View History

2020-11-19 02:25:33 +00:00
---
2020-11-19 02:31:17 +00:00
date: "2020-11-18T18:31:15-08:00"
description: Efficient redundancy via repository mirroring with nothing but git.
2020-11-19 02:25:33 +00:00
outputs:
- html
- gemtext
2020-11-19 02:25:33 +00:00
tags:
- git
- foss
2020-11-19 02:25:33 +00:00
title: "Resilient Git, Part 1: Hydra Hosting"
---
<div role="note">
This is Part 1 of a series called [Resilient Git](/2020/11/17/git-workflow-0.html "{itemprop='relatedLink'}").
</div>
<section role="doc-introduction">
Intro&shy;duction {#introduction}
-----------------
2020-11-19 02:25:33 +00:00
2021-01-24 22:42:40 +00:00
The most important part of a project is its code. Resilient projects should have their code in multiple places of equal weight so that work continues normally if a single remote goes down.
2020-11-19 02:25:33 +00:00
2021-01-24 22:42:40 +00:00
Many projects already do something similar: they have one "primary" remote and several mirrors. I'm suggesting something different. Treating a remote as a "mirror" implies that the remote is a second-class citizen. Mirrors are often out of date and aren't usually the preferred place to fetch code. Instead of setting up a primary remote and mirrors, I propose **hydra hosting:** setting up multiple primary remotes of equal status and pushing to/fetching from them in parallel.
2020-11-19 02:25:33 +00:00
2021-01-24 22:42:40 +00:00
Having multiple primary remotes of equal status might sound like a bad idea. If there are multiple remotes, how do people know which one to use? Where do they file bug reports, get code, or send patches? Do maintainers need to check multiple places?
2020-11-19 02:25:33 +00:00
2021-01-24 22:42:40 +00:00
No. Of course not. A good distributed system should automatically keep its nodes in sync to avoid the hassle of checking multiple places for updates.
2020-11-19 02:25:33 +00:00
</section>
Adding remotes
--------------
2020-11-19 02:25:33 +00:00
2021-01-24 22:42:40 +00:00
This process should pretty straightforward. You can run `git remote add` (see `git-remote(1)`) or edit your repo's `.git/config` directly:
2020-11-19 02:25:33 +00:00
2022-04-26 00:29:57 +00:00
```
2020-11-19 02:25:33 +00:00
[remote "origin"]
url = git@git.sr.ht:~seirdy/seirdy.one
fetch = +refs/heads/*:refs/remotes/origin/*
[remote "gl_upstream"]
url = git@gitlab.com:seirdy/seirdy.one.git
fetch = +refs/heads/*:refs/remotes/gl_upstream/*
[remote "gh_upstream"]
url = git@github.com:seirdy/seirdy.one.git
fetch = +refs/heads/*:refs/remotes/gh_upstream/*
```
2021-01-24 22:42:40 +00:00
If that's too much work--a perfectly understandable complaint--automating the process is trivial. Here's [an example from my dotfiles](https://git.sr.ht/~seirdy/dotfiles/tree/master/Executables/shell-scripts/bin/git-remote-setup).
2020-11-19 02:25:33 +00:00
Seamless pushing and pulling
----------------------------
2020-11-19 02:25:33 +00:00
2021-01-24 22:42:40 +00:00
Having multiple remotes is fine, but pushing to and fetching from all of them can be slow. Two simple git aliases fix that:
2020-11-19 02:25:33 +00:00
2022-04-26 00:29:57 +00:00
```
2020-11-19 02:25:33 +00:00
[alias]
pushall = !git remote | grep -E 'origin|upstream' | xargs -L1 -P 0 git push --all --follow-tags
fetchall = !git remote | grep -E 'origin|upstream' | xargs -L1 -P 0 git fetch
```
2021-01-24 22:42:40 +00:00
Now, `git pushall` and `git fetchall` will push to and fetch from all remotes in parallel, respectively. Only one remote needs to be online for project members to keep working.
2020-11-19 02:25:33 +00:00
Advertising remotes
-------------------
2020-11-19 02:25:33 +00:00
2021-01-24 22:42:40 +00:00
I'd recommend advertising at least three remotes in your README: your personal favorite and two determined by popularity. Tell users to run `git remote set-url` to switch remote locations if one goes down.
2020-11-19 02:25:33 +00:00
Before you ask...
-----------------
2020-11-19 02:25:33 +00:00
Q: Why not use a cloud service to automate mirroring?
2021-01-24 22:42:40 +00:00
A: Such a setup depends upon the cloud service and a primary repo for that service to watch, defeating the purpose (resiliency). Hydra hosting automates this without introducing new tools, dependencies, or closed platforms to the mix.
2020-11-19 02:25:33 +00:00
Q: What about issues, patches, etc.?
2020-11-19 20:23:18 +00:00
A: Stay tuned for Parts 2 and 3, coming soon to a weblog/gemlog near you™.
2020-11-19 02:25:33 +00:00
Q: Why did you call this "hydra hosting"?
2021-01-24 22:42:40 +00:00
A: It's a reference to the Hydra of Lerna from Greek Mythology, famous for keeping its brain in a nested RAID array to protect against disk failures and beheading. It could also be a reference to a fictional organization of the same name from Marvel Comics named after the Greek monster for [similar reasons](https://www.youtube.com/watch?v=assccoyvntI&t=37) ([direct webm](https://seirdy.one/misc/hail_hydra.webm)).