+19
-24
.forgejo/workflows/ci.yaml
+19
-24
.forgejo/workflows/ci.yaml
···
7
7
8
8
jobs:
9
9
check:
10
-
runs-on: codeberg-small-lazy
10
+
runs-on: debian-trixie
11
11
container:
12
-
image: docker.io/library/node:24-trixie-slim@sha256:fcdfd7bcd8f641c8c76a8950343c73912d68ba341e8dd1074e663b784d3e76f4
12
+
image: docker.io/library/node:24-trixie-slim@sha256:b05474903f463ce4064c09986525e6588c3e66c51b69be9c93a39fb359f883ce
13
13
steps:
14
14
- name: Check out source code
15
-
uses: https://code.forgejo.org/actions/checkout@1af3b93b6815bc44a9784bd300feb67ff0d1eeb3 # v6.0.0
15
+
uses: https://code.forgejo.org/actions/checkout@de0fac2e4500dabe0009e67214ff5f5447ce83dd # v6.0.2
16
16
- name: Set up toolchain
17
17
uses: https://code.forgejo.org/actions/setup-go@4dc6199c7b1a012772edbd06daecab0f50c9053c # v6.1.0
18
18
with:
···
34
34
# IMPORTANT: This workflow step will not work without the Releases unit enabled!
35
35
if: ${{ forge.ref == 'refs/heads/main' || startsWith(forge.event.ref, 'refs/tags/v') }}
36
36
needs: [check]
37
-
runs-on: codeberg-small-lazy
37
+
runs-on: debian-trixie
38
38
container:
39
-
image: docker.io/library/node:24-trixie-slim@sha256:fcdfd7bcd8f641c8c76a8950343c73912d68ba341e8dd1074e663b784d3e76f4
39
+
image: docker.io/library/node:24-trixie-slim@sha256:b05474903f463ce4064c09986525e6588c3e66c51b69be9c93a39fb359f883ce
40
40
steps:
41
41
- name: Check out source code
42
-
uses: https://code.forgejo.org/actions/checkout@1af3b93b6815bc44a9784bd300feb67ff0d1eeb3 # v6.0.0
42
+
uses: https://code.forgejo.org/actions/checkout@de0fac2e4500dabe0009e67214ff5f5447ce83dd # v6.0.2
43
43
- name: Set up toolchain
44
44
uses: https://code.forgejo.org/actions/setup-go@4dc6199c7b1a012772edbd06daecab0f50c9053c # v6.1.0
45
45
with:
···
70
70
package:
71
71
if: ${{ forge.ref == 'refs/heads/main' || startsWith(forge.event.ref, 'refs/tags/v') }}
72
72
needs: [check]
73
-
runs-on: codeberg-small-lazy
73
+
runs-on: debian-trixie
74
74
container:
75
-
image: docker.io/library/node:24-trixie-slim@sha256:fcdfd7bcd8f641c8c76a8950343c73912d68ba341e8dd1074e663b784d3e76f4
75
+
image: docker.io/library/node:24-trixie-slim@sha256:b05474903f463ce4064c09986525e6588c3e66c51b69be9c93a39fb359f883ce
76
76
steps:
77
77
- name: Install dependencies
78
78
run: |
79
79
apt-get -y update
80
-
apt-get -y install buildah ca-certificates
80
+
apt-get -y install ca-certificates buildah qemu-user-binfmt
81
81
- name: Check out source code
82
-
uses: https://code.forgejo.org/actions/checkout@1af3b93b6815bc44a9784bd300feb67ff0d1eeb3 # v6.0.0
82
+
uses: https://code.forgejo.org/actions/checkout@de0fac2e4500dabe0009e67214ff5f5447ce83dd # v6.0.2
83
83
- name: Build container
84
84
run: |
85
85
printf '[storage]\ndriver="vfs"\nrunroot="/run/containers/storage"\ngraphroot="/var/lib/containers/storage"\n' | tee /etc/containers/storage.conf
86
-
buildah build --arch=amd64 --tag=container:${VER}-amd64 .
87
-
buildah build --arch=arm64 --tag=container:${VER}-arm64 .
88
-
buildah manifest create container:${VER} \
89
-
container:${VER}-amd64 \
90
-
container:${VER}-arm64
86
+
buildah build --arch=amd64 --tag=container:amd64
87
+
buildah build --arch=arm64 --tag=container:arm64
88
+
buildah manifest create container container:amd64 container:arm64
91
89
env:
92
90
BUILDAH_ISOLATION: chroot
93
-
VER: ${{ startsWith(forge.event.ref, 'refs/tags/v') && forge.ref_name || 'latest' }}
94
91
- if: ${{ forge.repository == 'git-pages/git-pages-cli' }}
95
92
name: Push container to Codeberg
96
93
run: |
97
-
buildah login --authfile=/tmp/authfile-${FORGE}.json \
94
+
buildah login --authfile=/tmp/authfile.json \
98
95
-u ${{ vars.PACKAGES_USER }} -p ${{ secrets.PACKAGES_TOKEN }} ${FORGE}
99
-
buildah manifest push --authfile=/tmp/authfile-${FORGE}.json \
100
-
--all container:${VER} "docker://${FORGE}/${{ forge.repository }}:${VER/v/}"
96
+
buildah manifest push --authfile=/tmp/authfile.json \
97
+
--all container "docker://${FORGE}/${{ forge.repository }}:${VER/v/}"
101
98
env:
102
-
BUILDAH_ISOLATION: chroot
103
99
FORGE: codeberg.org
104
100
VER: ${{ startsWith(forge.event.ref, 'refs/tags/v') && forge.ref_name || 'latest' }}
105
101
- if: ${{ forge.repository == 'git-pages/git-pages-cli' }}
106
102
name: Push container to code.forgejo.org
107
103
run: |
108
-
buildah login --authfile=/tmp/authfile-${FORGE}.json \
104
+
buildah login --authfile=/tmp/authfile.json \
109
105
-u ${{ vars.PACKAGES_USER }} -p ${{ secrets.CFO_PACKAGES_TOKEN }} ${FORGE}
110
-
buildah manifest push --authfile=/tmp/authfile-${FORGE}.json \
111
-
--all container:${VER} "docker://${FORGE}/${{ forge.repository }}:${VER/v/}"
106
+
buildah manifest push --authfile=/tmp/authfile.json \
107
+
--all container "docker://${FORGE}/${{ forge.repository }}:${VER/v/}"
112
108
env:
113
-
BUILDAH_ISOLATION: chroot
114
109
FORGE: code.forgejo.org
115
110
VER: ${{ startsWith(forge.event.ref, 'refs/tags/v') && forge.ref_name || 'latest' }}
+1
-1
Dockerfile
+1
-1
Dockerfile
···
1
-
FROM --platform=$BUILDPLATFORM docker.io/library/golang:1.25-alpine@sha256:d3f0cf7723f3429e3f9ed846243970b20a2de7bae6a5b66fc5914e228d831bbb AS builder
1
+
FROM --platform=$BUILDPLATFORM docker.io/library/golang:1.25-alpine@sha256:ac09a5f469f307e5da71e766b0bd59c9c49ea460a528cc3e6686513d64a6f1fb AS builder
2
2
ARG TARGETOS TARGETARCH
3
3
RUN apk --no-cache add ca-certificates git
4
4
WORKDIR /build
+3
-2
README.md
+3
-2
README.md
···
3
3
4
4
_git-pages-cli_ is a command-line application for publishing sites to [git-pages].
5
5
6
-
If you want to publish a site from a Forgejo Actions workflow, use [git-pages/action] instead.
6
+
> [!TIP]
7
+
> If you want to publish a site from a CI workflow, use the [Forgejo Action][git-pages-action] instead.
7
8
8
9
[git-pages]: https://codeberg.org/git-pages/git-pages
9
-
[git-pages/action]: https://codeberg.org/git-pages/action
10
+
[git-pages-action]: https://codeberg.org/git-pages/action
10
11
11
12
12
13
Installation
+24
flake.lock
+24
flake.lock
···
18
18
"type": "github"
19
19
}
20
20
},
21
+
"gomod2nix": {
22
+
"inputs": {
23
+
"flake-utils": [
24
+
"flake-utils"
25
+
],
26
+
"nixpkgs": [
27
+
"nixpkgs"
28
+
]
29
+
},
30
+
"locked": {
31
+
"lastModified": 1763982521,
32
+
"narHash": "sha256-ur4QIAHwgFc0vXiaxn5No/FuZicxBr2p0gmT54xZkUQ=",
33
+
"owner": "nix-community",
34
+
"repo": "gomod2nix",
35
+
"rev": "02e63a239d6eabd595db56852535992c898eba72",
36
+
"type": "github"
37
+
},
38
+
"original": {
39
+
"owner": "nix-community",
40
+
"repo": "gomod2nix",
41
+
"type": "github"
42
+
}
43
+
},
21
44
"nix-filter": {
22
45
"locked": {
23
46
"lastModified": 1757882181,
···
52
75
"root": {
53
76
"inputs": {
54
77
"flake-utils": "flake-utils",
78
+
"gomod2nix": "gomod2nix",
55
79
"nix-filter": "nix-filter",
56
80
"nixpkgs": "nixpkgs"
57
81
}
+22
-4
flake.nix
+22
-4
flake.nix
···
3
3
nixpkgs.url = "github:NixOS/nixpkgs/nixpkgs-unstable";
4
4
flake-utils.url = "github:numtide/flake-utils";
5
5
nix-filter.url = "github:numtide/nix-filter";
6
+
7
+
gomod2nix = {
8
+
url = "github:nix-community/gomod2nix";
9
+
inputs.nixpkgs.follows = "nixpkgs";
10
+
inputs.flake-utils.follows = "flake-utils";
11
+
};
6
12
};
7
13
8
14
outputs =
···
11
17
nixpkgs,
12
18
flake-utils,
13
19
nix-filter,
14
-
}:
20
+
...
21
+
}@inputs:
15
22
flake-utils.lib.eachDefaultSystem (
16
23
system:
17
24
let
18
-
pkgs = nixpkgs.legacyPackages.${system};
25
+
pkgs = import nixpkgs {
26
+
inherit system;
19
27
20
-
git-pages-cli = pkgs.buildGo125Module {
28
+
overlays = [
29
+
inputs.gomod2nix.overlays.default
30
+
];
31
+
};
32
+
33
+
git-pages-cli = pkgs.buildGoApplication {
21
34
pname = "git-pages-cli";
22
35
version = "0";
23
36
···
41
54
"-s -w"
42
55
];
43
56
44
-
vendorHash = "sha256-5vjUhN3lCr41q91lOD7v0F9c6a8GJj7wBGnnzgFBhJU=";
57
+
go = pkgs.go_1_25;
58
+
modules = ./gomod2nix.toml;
45
59
};
46
60
in
47
61
{
···
50
64
devShells.default = pkgs.mkShell {
51
65
inputsFrom = [
52
66
git-pages-cli
67
+
];
68
+
69
+
packages = with pkgs; [
70
+
gomod2nix
53
71
];
54
72
};
55
73
+2
-2
go.mod
+2
-2
go.mod
+2
-2
go.sum
+2
-2
go.sum
···
1
1
github.com/google/uuid v1.6.0 h1:NIvaJDMOsjHA8n1jAhLSgzrAzy1Hgr+hNrb57e+94F0=
2
2
github.com/google/uuid v1.6.0/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo=
3
-
github.com/klauspost/compress v1.18.1 h1:bcSGx7UbpBqMChDtsF28Lw6v/G94LPrrbMbdC3JH2co=
4
-
github.com/klauspost/compress v1.18.1/go.mod h1:ZQFFVG+MdnR0P+l6wpXgIL4NTtwiKIdBnrBd8Nrxr+0=
3
+
github.com/klauspost/compress v1.18.2 h1:iiPHWW0YrcFgpBYhsA6D1+fqHssJscY/Tm/y2Uqnapk=
4
+
github.com/klauspost/compress v1.18.2/go.mod h1:R0h/fSBs8DE4ENlcrlib3PsXS61voFxhIs2DeRhCvJ4=
5
5
github.com/spf13/pflag v1.0.10 h1:4EBh2KAYBwaONj6b2Ye1GiHfwjqyROoF4RwYO+vPwFk=
6
6
github.com/spf13/pflag v1.0.10/go.mod h1:McXfInJRrz4CZXVZOBLb0bTZqETkiAhM9Iw0y3An2Bg=
+12
gomod2nix.toml
+12
gomod2nix.toml
···
1
+
schema = 3
2
+
3
+
[mod]
4
+
[mod."github.com/google/uuid"]
5
+
version = "v1.6.0"
6
+
hash = "sha256-VWl9sqUzdOuhW0KzQlv0gwwUQClYkmZwSydHG2sALYw="
7
+
[mod."github.com/klauspost/compress"]
8
+
version = "v1.18.2"
9
+
hash = "sha256-mRa+6qEi5joqQao13ZFogmq67rOQzHCVbCCjKA+HKEc="
10
+
[mod."github.com/spf13/pflag"]
11
+
version = "v1.0.10"
12
+
hash = "sha256-uDPnWjHpSrzXr17KEYEA1yAbizfcsfo5AyztY2tS6ZU="
+204
-58
main.go
+204
-58
main.go
···
2
2
3
3
import (
4
4
"archive/tar"
5
+
"bufio"
5
6
"bytes"
7
+
"crypto"
6
8
"crypto/sha256"
9
+
"encoding/hex"
10
+
"errors"
7
11
"fmt"
8
12
"io"
9
13
"io/fs"
···
11
15
"net/url"
12
16
"os"
13
17
"runtime/debug"
18
+
"strconv"
19
+
"strings"
14
20
15
21
"github.com/google/uuid"
16
22
"github.com/klauspost/compress/zstd"
···
37
43
var challengeFlag = pflag.Bool("challenge", false, "compute DNS challenge entry from password (output zone file record)")
38
44
var challengeBareFlag = pflag.Bool("challenge-bare", false, "compute DNS challenge entry from password (output bare TXT value)")
39
45
var uploadGitFlag = pflag.String("upload-git", "", "replace site with contents of specified git repository")
40
-
var uploadDirFlag = pflag.String("upload-dir", "", "replace site with contents of specified directory")
41
-
var deleteFlag = pflag.Bool("delete", false, "delete site")
46
+
var uploadDirFlag = pflag.String("upload-dir", "", "replace whole site or a path with contents of specified directory")
47
+
var deleteFlag = pflag.Bool("delete", false, "delete whole site or a path")
42
48
var debugManifestFlag = pflag.Bool("debug-manifest", false, "retrieve site manifest as ProtoJSON, for debugging")
43
49
var serverFlag = pflag.String("server", "", "hostname of server to connect to")
44
-
var verboseFlag = pflag.Bool("verbose", false, "display more information for debugging")
45
-
var versionFlag = pflag.Bool("version", false, "display version information")
50
+
var pathFlag = pflag.String("path", "", "partially update site at specified path")
51
+
var parentsFlag = pflag.Bool("parents", false, "create parent directories of --path")
52
+
var atomicFlag = pflag.Bool("atomic", false, "require partial updates to be atomic")
53
+
var incrementalFlag = pflag.Bool("incremental", false, "make --upload-dir only upload changed files")
54
+
var verboseFlag = pflag.BoolP("verbose", "v", false, "display more information for debugging")
55
+
var versionFlag = pflag.BoolP("version", "V", false, "display version information")
46
56
47
57
func singleOperation() bool {
48
58
operations := 0
···
70
80
return operations == 1
71
81
}
72
82
73
-
func displayFS(root fs.FS) error {
83
+
func gitBlobSHA256(data []byte) string {
84
+
h := crypto.SHA256.New()
85
+
h.Write([]byte("blob "))
86
+
h.Write([]byte(strconv.FormatInt(int64(len(data)), 10)))
87
+
h.Write([]byte{0})
88
+
h.Write(data)
89
+
return hex.EncodeToString(h.Sum(nil))
90
+
}
91
+
92
+
func displayFS(root fs.FS, prefix string) error {
74
93
return fs.WalkDir(root, ".", func(name string, entry fs.DirEntry, err error) error {
75
94
if err != nil {
76
95
return err
77
96
}
78
97
switch {
79
-
case entry.Type() == 0:
80
-
fmt.Fprintln(os.Stderr, "file", name)
81
-
case entry.Type() == fs.ModeDir:
82
-
fmt.Fprintln(os.Stderr, "dir", name)
98
+
case entry.Type().IsDir():
99
+
fmt.Fprintf(os.Stderr, "dir %s%s\n", prefix, name)
100
+
case entry.Type().IsRegular():
101
+
fmt.Fprintf(os.Stderr, "file %s%s\n", prefix, name)
83
102
case entry.Type() == fs.ModeSymlink:
84
-
fmt.Fprintln(os.Stderr, "symlink", name)
103
+
fmt.Fprintf(os.Stderr, "symlink %s%s\n", prefix, name)
85
104
default:
86
-
fmt.Fprintln(os.Stderr, "other", name)
105
+
fmt.Fprintf(os.Stderr, "other %s%s\n", prefix, name)
87
106
}
88
107
return nil
89
108
})
90
109
}
91
110
92
-
func archiveFS(writer io.Writer, root fs.FS) (err error) {
111
+
// It doesn't make sense to use incremental updates for very small files since the cost of
112
+
// repeating a request to fill in a missing blob is likely to be higher than any savings gained.
113
+
const incrementalSizeThreshold = 256
114
+
115
+
func archiveFS(writer io.Writer, root fs.FS, prefix string, needBlobs []string) (err error) {
116
+
requestedSet := make(map[string]struct{})
117
+
for _, hash := range needBlobs {
118
+
requestedSet[hash] = struct{}{}
119
+
}
93
120
zstdWriter, _ := zstd.NewWriter(writer)
94
121
tarWriter := tar.NewWriter(zstdWriter)
95
-
err = tarWriter.AddFS(root)
96
-
if err != nil {
122
+
if err = fs.WalkDir(root, ".", func(name string, entry fs.DirEntry, err error) error {
123
+
if err != nil {
124
+
return err
125
+
}
126
+
header := &tar.Header{}
127
+
data := []byte{}
128
+
if prefix == "" && name == "." {
129
+
return nil
130
+
} else if name == "." {
131
+
header.Name = prefix
132
+
} else {
133
+
header.Name = prefix + name
134
+
}
135
+
switch {
136
+
case entry.Type().IsDir():
137
+
header.Typeflag = tar.TypeDir
138
+
header.Name += "/"
139
+
case entry.Type().IsRegular():
140
+
header.Typeflag = tar.TypeReg
141
+
if data, err = fs.ReadFile(root, name); err != nil {
142
+
return err
143
+
}
144
+
if *incrementalFlag && len(data) > incrementalSizeThreshold {
145
+
hash := gitBlobSHA256(data)
146
+
if _, requested := requestedSet[hash]; !requested {
147
+
header.Typeflag = tar.TypeSymlink
148
+
header.Linkname = "/git/blobs/" + hash
149
+
data = nil
150
+
}
151
+
}
152
+
case entry.Type() == fs.ModeSymlink:
153
+
header.Typeflag = tar.TypeSymlink
154
+
if header.Linkname, err = fs.ReadLink(root, name); err != nil {
155
+
return err
156
+
}
157
+
default:
158
+
return errors.New("tar: cannot add non-regular file")
159
+
}
160
+
header.Size = int64(len(data))
161
+
if err = tarWriter.WriteHeader(header); err != nil {
162
+
return err
163
+
}
164
+
if _, err = tarWriter.Write(data); err != nil {
165
+
return err
166
+
}
167
+
return err
168
+
}); err != nil {
97
169
return
98
170
}
99
-
err = tarWriter.Close()
100
-
if err != nil {
171
+
if err = tarWriter.Close(); err != nil {
101
172
return
102
173
}
103
-
err = zstdWriter.Close()
104
-
if err != nil {
174
+
if err = zstdWriter.Close(); err != nil {
105
175
return
106
176
}
107
177
return
178
+
}
179
+
180
+
// Stream archive data without ever loading the entire working set into RAM.
181
+
func streamArchiveFS(root fs.FS, prefix string, needBlobs []string) io.ReadCloser {
182
+
reader, writer := io.Pipe()
183
+
go func() {
184
+
err := archiveFS(writer, root, prefix, needBlobs)
185
+
if err != nil {
186
+
writer.CloseWithError(err)
187
+
} else {
188
+
writer.Close()
189
+
}
190
+
}()
191
+
return reader
192
+
}
193
+
194
+
func makeWhiteout(path string) (reader io.Reader) {
195
+
buffer := &bytes.Buffer{}
196
+
tarWriter := tar.NewWriter(buffer)
197
+
tarWriter.WriteHeader(&tar.Header{
198
+
Typeflag: tar.TypeChar,
199
+
Name: path,
200
+
})
201
+
tarWriter.Flush()
202
+
return buffer
108
203
}
109
204
110
205
const usageExitCode = 125
···
135
230
os.Exit(usageExitCode)
136
231
}
137
232
233
+
var pathPrefix string
234
+
if *pathFlag != "" {
235
+
if *uploadDirFlag == "" && !*deleteFlag {
236
+
fmt.Fprintf(os.Stderr, "--path requires --upload-dir or --delete")
237
+
os.Exit(usageExitCode)
238
+
} else {
239
+
pathPrefix = strings.Trim(*pathFlag, "/") + "/"
240
+
}
241
+
}
242
+
138
243
var err error
139
244
siteURL, err := url.Parse(pflag.Args()[0])
140
245
if err != nil {
···
143
248
}
144
249
145
250
var request *http.Request
251
+
var uploadDir *os.Root
146
252
switch {
147
253
case *challengeFlag || *challengeBareFlag:
148
254
if *passwordFlag == "" {
···
174
280
request.Header.Add("Content-Type", "application/x-www-form-urlencoded")
175
281
176
282
case *uploadDirFlag != "":
177
-
uploadDirFS, err := os.OpenRoot(*uploadDirFlag)
283
+
uploadDir, err = os.OpenRoot(*uploadDirFlag)
178
284
if err != nil {
179
285
fmt.Fprintf(os.Stderr, "error: invalid directory: %s\n", err)
180
286
os.Exit(1)
181
287
}
182
288
183
289
if *verboseFlag {
184
-
err := displayFS(uploadDirFS.FS())
290
+
err := displayFS(uploadDir.FS(), pathPrefix)
185
291
if err != nil {
186
292
fmt.Fprintf(os.Stderr, "error: %s\n", err)
187
293
os.Exit(1)
188
294
}
189
295
}
190
296
191
-
// Stream archive data without ever loading the entire working set into RAM.
192
-
reader, writer := io.Pipe()
193
-
go func() {
194
-
err = archiveFS(writer, uploadDirFS.FS())
195
-
if err != nil {
196
-
fmt.Fprintf(os.Stderr, "error: %s\n", err)
197
-
os.Exit(1)
198
-
}
199
-
writer.Close()
200
-
}()
201
-
202
-
request, err = http.NewRequest("PUT", siteURL.String(), reader)
297
+
if *pathFlag == "" {
298
+
request, err = http.NewRequest("PUT", siteURL.String(), nil)
299
+
} else {
300
+
request, err = http.NewRequest("PATCH", siteURL.String(), nil)
301
+
}
203
302
if err != nil {
204
303
fmt.Fprintf(os.Stderr, "error: %s\n", err)
205
304
os.Exit(1)
206
305
}
306
+
request.Body = streamArchiveFS(uploadDir.FS(), pathPrefix, []string{})
207
307
request.ContentLength = -1
208
308
request.Header.Add("Content-Type", "application/x-tar+zstd")
309
+
request.Header.Add("Accept", "application/vnd.git-pages.unresolved;q=1.0, text/plain;q=0.9")
310
+
if *parentsFlag {
311
+
request.Header.Add("Create-Parents", "yes")
312
+
} else {
313
+
request.Header.Add("Create-Parents", "no")
314
+
}
209
315
210
316
case *deleteFlag:
211
-
request, err = http.NewRequest("DELETE", siteURL.String(), nil)
212
-
if err != nil {
213
-
fmt.Fprintf(os.Stderr, "error: %s\n", err)
214
-
os.Exit(1)
317
+
if *pathFlag == "" {
318
+
request, err = http.NewRequest("DELETE", siteURL.String(), nil)
319
+
if err != nil {
320
+
fmt.Fprintf(os.Stderr, "error: %s\n", err)
321
+
os.Exit(1)
322
+
}
323
+
} else {
324
+
request, err = http.NewRequest("PATCH", siteURL.String(), makeWhiteout(pathPrefix))
325
+
if err != nil {
326
+
fmt.Fprintf(os.Stderr, "error: %s\n", err)
327
+
os.Exit(1)
328
+
}
329
+
request.Header.Add("Content-Type", "application/x-tar")
215
330
}
216
331
217
332
case *debugManifestFlag:
···
226
341
panic("no operation chosen")
227
342
}
228
343
request.Header.Add("User-Agent", versionInfo())
344
+
if request.Method == "PATCH" {
345
+
if *atomicFlag {
346
+
request.Header.Add("Atomic", "yes")
347
+
request.Header.Add("Race-Free", "yes") // deprecated name, to be removed soon
348
+
} else {
349
+
request.Header.Add("Atomic", "no")
350
+
request.Header.Add("Race-Free", "no") // deprecated name, to be removed soon
351
+
}
352
+
}
229
353
switch {
230
354
case *passwordFlag != "":
231
355
request.Header.Add("Authorization", fmt.Sprintf("Pages %s", *passwordFlag))
···
243
367
request.Header.Set("Host", siteURL.Host)
244
368
}
245
369
246
-
response, err := http.DefaultClient.Do(request)
247
-
if err != nil {
248
-
fmt.Fprintf(os.Stderr, "error: %s\n", err)
249
-
os.Exit(1)
250
-
}
251
-
if *verboseFlag {
252
-
fmt.Fprintf(os.Stderr, "server: %s\n", response.Header.Get("Server"))
253
-
}
254
-
if *debugManifestFlag {
255
-
if response.StatusCode == 200 {
256
-
io.Copy(os.Stdout, response.Body)
257
-
fmt.Fprintf(os.Stdout, "\n")
258
-
} else {
259
-
io.Copy(os.Stderr, response.Body)
370
+
displayServer := *verboseFlag
371
+
for {
372
+
response, err := http.DefaultClient.Do(request)
373
+
if err != nil {
374
+
fmt.Fprintf(os.Stderr, "error: %s\n", err)
260
375
os.Exit(1)
261
376
}
262
-
} else { // an update operation
263
-
if response.StatusCode == 200 {
264
-
fmt.Fprintf(os.Stdout, "result: %s\n", response.Header.Get("Update-Result"))
265
-
io.Copy(os.Stdout, response.Body)
266
-
} else {
267
-
fmt.Fprintf(os.Stderr, "result: error\n")
268
-
io.Copy(os.Stderr, response.Body)
269
-
os.Exit(1)
377
+
if displayServer {
378
+
fmt.Fprintf(os.Stderr, "server: %s\n", response.Header.Get("Server"))
379
+
displayServer = false
380
+
}
381
+
if *debugManifestFlag {
382
+
if response.StatusCode == http.StatusOK {
383
+
io.Copy(os.Stdout, response.Body)
384
+
fmt.Fprintf(os.Stdout, "\n")
385
+
} else {
386
+
io.Copy(os.Stderr, response.Body)
387
+
os.Exit(1)
388
+
}
389
+
} else { // an update operation
390
+
if *verboseFlag {
391
+
fmt.Fprintf(os.Stderr, "response: %d %s\n",
392
+
response.StatusCode, response.Header.Get("Content-Type"))
393
+
}
394
+
if response.StatusCode == http.StatusUnprocessableEntity &&
395
+
response.Header.Get("Content-Type") == "application/vnd.git-pages.unresolved" {
396
+
needBlobs := []string{}
397
+
scanner := bufio.NewScanner(response.Body)
398
+
for scanner.Scan() {
399
+
needBlobs = append(needBlobs, scanner.Text())
400
+
}
401
+
response.Body.Close()
402
+
if *verboseFlag {
403
+
fmt.Fprintf(os.Stderr, "incremental: need %d blobs\n", len(needBlobs))
404
+
}
405
+
request.Body = streamArchiveFS(uploadDir.FS(), pathPrefix, needBlobs)
406
+
continue // resubmit
407
+
} else if response.StatusCode == http.StatusOK {
408
+
fmt.Fprintf(os.Stdout, "result: %s\n", response.Header.Get("Update-Result"))
409
+
io.Copy(os.Stdout, response.Body)
410
+
} else {
411
+
fmt.Fprintf(os.Stderr, "result: error\n")
412
+
io.Copy(os.Stderr, response.Body)
413
+
os.Exit(1)
414
+
}
270
415
}
416
+
break
271
417
}
272
418
}