[mirror] Command-line application for uploading a site to a git-pages server

Compare changes

Choose any two refs to compare.

+19 -24
.forgejo/workflows/ci.yaml
··· 7 7 8 8 jobs: 9 9 check: 10 - runs-on: codeberg-small-lazy 10 + runs-on: debian-trixie 11 11 container: 12 - image: docker.io/library/node:24-trixie-slim@sha256:fcdfd7bcd8f641c8c76a8950343c73912d68ba341e8dd1074e663b784d3e76f4 12 + image: docker.io/library/node:24-trixie-slim@sha256:b05474903f463ce4064c09986525e6588c3e66c51b69be9c93a39fb359f883ce 13 13 steps: 14 14 - name: Check out source code 15 - uses: https://code.forgejo.org/actions/checkout@1af3b93b6815bc44a9784bd300feb67ff0d1eeb3 # v6.0.0 15 + uses: https://code.forgejo.org/actions/checkout@de0fac2e4500dabe0009e67214ff5f5447ce83dd # v6.0.2 16 16 - name: Set up toolchain 17 17 uses: https://code.forgejo.org/actions/setup-go@4dc6199c7b1a012772edbd06daecab0f50c9053c # v6.1.0 18 18 with: ··· 34 34 # IMPORTANT: This workflow step will not work without the Releases unit enabled! 35 35 if: ${{ forge.ref == 'refs/heads/main' || startsWith(forge.event.ref, 'refs/tags/v') }} 36 36 needs: [check] 37 - runs-on: codeberg-small-lazy 37 + runs-on: debian-trixie 38 38 container: 39 - image: docker.io/library/node:24-trixie-slim@sha256:fcdfd7bcd8f641c8c76a8950343c73912d68ba341e8dd1074e663b784d3e76f4 39 + image: docker.io/library/node:24-trixie-slim@sha256:b05474903f463ce4064c09986525e6588c3e66c51b69be9c93a39fb359f883ce 40 40 steps: 41 41 - name: Check out source code 42 - uses: https://code.forgejo.org/actions/checkout@1af3b93b6815bc44a9784bd300feb67ff0d1eeb3 # v6.0.0 42 + uses: https://code.forgejo.org/actions/checkout@de0fac2e4500dabe0009e67214ff5f5447ce83dd # v6.0.2 43 43 - name: Set up toolchain 44 44 uses: https://code.forgejo.org/actions/setup-go@4dc6199c7b1a012772edbd06daecab0f50c9053c # v6.1.0 45 45 with: ··· 70 70 package: 71 71 if: ${{ forge.ref == 'refs/heads/main' || startsWith(forge.event.ref, 'refs/tags/v') }} 72 72 needs: [check] 73 - runs-on: codeberg-small-lazy 73 + runs-on: debian-trixie 74 74 container: 75 - image: docker.io/library/node:24-trixie-slim@sha256:fcdfd7bcd8f641c8c76a8950343c73912d68ba341e8dd1074e663b784d3e76f4 75 + image: docker.io/library/node:24-trixie-slim@sha256:b05474903f463ce4064c09986525e6588c3e66c51b69be9c93a39fb359f883ce 76 76 steps: 77 77 - name: Install dependencies 78 78 run: | 79 79 apt-get -y update 80 - apt-get -y install buildah ca-certificates 80 + apt-get -y install ca-certificates buildah qemu-user-binfmt 81 81 - name: Check out source code 82 - uses: https://code.forgejo.org/actions/checkout@1af3b93b6815bc44a9784bd300feb67ff0d1eeb3 # v6.0.0 82 + uses: https://code.forgejo.org/actions/checkout@de0fac2e4500dabe0009e67214ff5f5447ce83dd # v6.0.2 83 83 - name: Build container 84 84 run: | 85 85 printf '[storage]\ndriver="vfs"\nrunroot="/run/containers/storage"\ngraphroot="/var/lib/containers/storage"\n' | tee /etc/containers/storage.conf 86 - buildah build --arch=amd64 --tag=container:${VER}-amd64 . 87 - buildah build --arch=arm64 --tag=container:${VER}-arm64 . 88 - buildah manifest create container:${VER} \ 89 - container:${VER}-amd64 \ 90 - container:${VER}-arm64 86 + buildah build --arch=amd64 --tag=container:amd64 87 + buildah build --arch=arm64 --tag=container:arm64 88 + buildah manifest create container container:amd64 container:arm64 91 89 env: 92 90 BUILDAH_ISOLATION: chroot 93 - VER: ${{ startsWith(forge.event.ref, 'refs/tags/v') && forge.ref_name || 'latest' }} 94 91 - if: ${{ forge.repository == 'git-pages/git-pages-cli' }} 95 92 name: Push container to Codeberg 96 93 run: | 97 - buildah login --authfile=/tmp/authfile-${FORGE}.json \ 94 + buildah login --authfile=/tmp/authfile.json \ 98 95 -u ${{ vars.PACKAGES_USER }} -p ${{ secrets.PACKAGES_TOKEN }} ${FORGE} 99 - buildah manifest push --authfile=/tmp/authfile-${FORGE}.json \ 100 - --all container:${VER} "docker://${FORGE}/${{ forge.repository }}:${VER/v/}" 96 + buildah manifest push --authfile=/tmp/authfile.json \ 97 + --all container "docker://${FORGE}/${{ forge.repository }}:${VER/v/}" 101 98 env: 102 - BUILDAH_ISOLATION: chroot 103 99 FORGE: codeberg.org 104 100 VER: ${{ startsWith(forge.event.ref, 'refs/tags/v') && forge.ref_name || 'latest' }} 105 101 - if: ${{ forge.repository == 'git-pages/git-pages-cli' }} 106 102 name: Push container to code.forgejo.org 107 103 run: | 108 - buildah login --authfile=/tmp/authfile-${FORGE}.json \ 104 + buildah login --authfile=/tmp/authfile.json \ 109 105 -u ${{ vars.PACKAGES_USER }} -p ${{ secrets.CFO_PACKAGES_TOKEN }} ${FORGE} 110 - buildah manifest push --authfile=/tmp/authfile-${FORGE}.json \ 111 - --all container:${VER} "docker://${FORGE}/${{ forge.repository }}:${VER/v/}" 106 + buildah manifest push --authfile=/tmp/authfile.json \ 107 + --all container "docker://${FORGE}/${{ forge.repository }}:${VER/v/}" 112 108 env: 113 - BUILDAH_ISOLATION: chroot 114 109 FORGE: code.forgejo.org 115 110 VER: ${{ startsWith(forge.event.ref, 'refs/tags/v') && forge.ref_name || 'latest' }}
+1 -1
Dockerfile
··· 1 - FROM --platform=$BUILDPLATFORM docker.io/library/golang:1.25-alpine@sha256:d3f0cf7723f3429e3f9ed846243970b20a2de7bae6a5b66fc5914e228d831bbb AS builder 1 + FROM --platform=$BUILDPLATFORM docker.io/library/golang:1.25-alpine@sha256:ac09a5f469f307e5da71e766b0bd59c9c49ea460a528cc3e6686513d64a6f1fb AS builder 2 2 ARG TARGETOS TARGETARCH 3 3 RUN apk --no-cache add ca-certificates git 4 4 WORKDIR /build
+3 -2
README.md
··· 3 3 4 4 _git-pages-cli_ is a command-line application for publishing sites to [git-pages]. 5 5 6 - If you want to publish a site from a Forgejo Actions workflow, use [git-pages/action] instead. 6 + > [!TIP] 7 + > If you want to publish a site from a CI workflow, use the [Forgejo Action][git-pages-action] instead. 7 8 8 9 [git-pages]: https://codeberg.org/git-pages/git-pages 9 - [git-pages/action]: https://codeberg.org/git-pages/action 10 + [git-pages-action]: https://codeberg.org/git-pages/action 10 11 11 12 12 13 Installation
+24
flake.lock
··· 18 18 "type": "github" 19 19 } 20 20 }, 21 + "gomod2nix": { 22 + "inputs": { 23 + "flake-utils": [ 24 + "flake-utils" 25 + ], 26 + "nixpkgs": [ 27 + "nixpkgs" 28 + ] 29 + }, 30 + "locked": { 31 + "lastModified": 1763982521, 32 + "narHash": "sha256-ur4QIAHwgFc0vXiaxn5No/FuZicxBr2p0gmT54xZkUQ=", 33 + "owner": "nix-community", 34 + "repo": "gomod2nix", 35 + "rev": "02e63a239d6eabd595db56852535992c898eba72", 36 + "type": "github" 37 + }, 38 + "original": { 39 + "owner": "nix-community", 40 + "repo": "gomod2nix", 41 + "type": "github" 42 + } 43 + }, 21 44 "nix-filter": { 22 45 "locked": { 23 46 "lastModified": 1757882181, ··· 52 75 "root": { 53 76 "inputs": { 54 77 "flake-utils": "flake-utils", 78 + "gomod2nix": "gomod2nix", 55 79 "nix-filter": "nix-filter", 56 80 "nixpkgs": "nixpkgs" 57 81 }
+22 -4
flake.nix
··· 3 3 nixpkgs.url = "github:NixOS/nixpkgs/nixpkgs-unstable"; 4 4 flake-utils.url = "github:numtide/flake-utils"; 5 5 nix-filter.url = "github:numtide/nix-filter"; 6 + 7 + gomod2nix = { 8 + url = "github:nix-community/gomod2nix"; 9 + inputs.nixpkgs.follows = "nixpkgs"; 10 + inputs.flake-utils.follows = "flake-utils"; 11 + }; 6 12 }; 7 13 8 14 outputs = ··· 11 17 nixpkgs, 12 18 flake-utils, 13 19 nix-filter, 14 - }: 20 + ... 21 + }@inputs: 15 22 flake-utils.lib.eachDefaultSystem ( 16 23 system: 17 24 let 18 - pkgs = nixpkgs.legacyPackages.${system}; 25 + pkgs = import nixpkgs { 26 + inherit system; 19 27 20 - git-pages-cli = pkgs.buildGo125Module { 28 + overlays = [ 29 + inputs.gomod2nix.overlays.default 30 + ]; 31 + }; 32 + 33 + git-pages-cli = pkgs.buildGoApplication { 21 34 pname = "git-pages-cli"; 22 35 version = "0"; 23 36 ··· 41 54 "-s -w" 42 55 ]; 43 56 44 - vendorHash = "sha256-5vjUhN3lCr41q91lOD7v0F9c6a8GJj7wBGnnzgFBhJU="; 57 + go = pkgs.go_1_25; 58 + modules = ./gomod2nix.toml; 45 59 }; 46 60 in 47 61 { ··· 50 64 devShells.default = pkgs.mkShell { 51 65 inputsFrom = [ 52 66 git-pages-cli 67 + ]; 68 + 69 + packages = with pkgs; [ 70 + gomod2nix 53 71 ]; 54 72 }; 55 73
+2 -2
go.mod
··· 3 3 go 1.25.0 4 4 5 5 require ( 6 - github.com/klauspost/compress v1.18.1 7 - github.com/spf13/pflag v1.0.10 8 6 github.com/google/uuid v1.6.0 7 + github.com/klauspost/compress v1.18.2 8 + github.com/spf13/pflag v1.0.10 9 9 )
+2 -2
go.sum
··· 1 1 github.com/google/uuid v1.6.0 h1:NIvaJDMOsjHA8n1jAhLSgzrAzy1Hgr+hNrb57e+94F0= 2 2 github.com/google/uuid v1.6.0/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo= 3 - github.com/klauspost/compress v1.18.1 h1:bcSGx7UbpBqMChDtsF28Lw6v/G94LPrrbMbdC3JH2co= 4 - github.com/klauspost/compress v1.18.1/go.mod h1:ZQFFVG+MdnR0P+l6wpXgIL4NTtwiKIdBnrBd8Nrxr+0= 3 + github.com/klauspost/compress v1.18.2 h1:iiPHWW0YrcFgpBYhsA6D1+fqHssJscY/Tm/y2Uqnapk= 4 + github.com/klauspost/compress v1.18.2/go.mod h1:R0h/fSBs8DE4ENlcrlib3PsXS61voFxhIs2DeRhCvJ4= 5 5 github.com/spf13/pflag v1.0.10 h1:4EBh2KAYBwaONj6b2Ye1GiHfwjqyROoF4RwYO+vPwFk= 6 6 github.com/spf13/pflag v1.0.10/go.mod h1:McXfInJRrz4CZXVZOBLb0bTZqETkiAhM9Iw0y3An2Bg=
+12
gomod2nix.toml
··· 1 + schema = 3 2 + 3 + [mod] 4 + [mod."github.com/google/uuid"] 5 + version = "v1.6.0" 6 + hash = "sha256-VWl9sqUzdOuhW0KzQlv0gwwUQClYkmZwSydHG2sALYw=" 7 + [mod."github.com/klauspost/compress"] 8 + version = "v1.18.2" 9 + hash = "sha256-mRa+6qEi5joqQao13ZFogmq67rOQzHCVbCCjKA+HKEc=" 10 + [mod."github.com/spf13/pflag"] 11 + version = "v1.0.10" 12 + hash = "sha256-uDPnWjHpSrzXr17KEYEA1yAbizfcsfo5AyztY2tS6ZU="
+214 -62
main.go
··· 2 2 3 3 import ( 4 4 "archive/tar" 5 + "bufio" 5 6 "bytes" 7 + "crypto" 6 8 "crypto/sha256" 9 + "encoding/hex" 10 + "errors" 7 11 "fmt" 8 12 "io" 9 13 "io/fs" ··· 11 15 "net/url" 12 16 "os" 13 17 "runtime/debug" 18 + "strconv" 19 + "strings" 14 20 15 21 "github.com/google/uuid" 16 22 "github.com/klauspost/compress/zstd" ··· 37 43 var challengeFlag = pflag.Bool("challenge", false, "compute DNS challenge entry from password (output zone file record)") 38 44 var challengeBareFlag = pflag.Bool("challenge-bare", false, "compute DNS challenge entry from password (output bare TXT value)") 39 45 var uploadGitFlag = pflag.String("upload-git", "", "replace site with contents of specified git repository") 40 - var uploadDirFlag = pflag.String("upload-dir", "", "replace site with contents of specified directory") 41 - var deleteFlag = pflag.Bool("delete", false, "delete site") 46 + var uploadDirFlag = pflag.String("upload-dir", "", "replace whole site or a path with contents of specified directory") 47 + var deleteFlag = pflag.Bool("delete", false, "delete whole site or a path") 42 48 var debugManifestFlag = pflag.Bool("debug-manifest", false, "retrieve site manifest as ProtoJSON, for debugging") 43 49 var serverFlag = pflag.String("server", "", "hostname of server to connect to") 44 - var verboseFlag = pflag.Bool("verbose", false, "display more information for debugging") 45 - var versionFlag = pflag.Bool("version", false, "display version information") 50 + var pathFlag = pflag.String("path", "", "partially update site at specified path") 51 + var parentsFlag = pflag.Bool("parents", false, "create parent directories of --path") 52 + var atomicFlag = pflag.Bool("atomic", false, "require partial updates to be atomic") 53 + var incrementalFlag = pflag.Bool("incremental", false, "make --upload-dir only upload changed files") 54 + var verboseFlag = pflag.BoolP("verbose", "v", false, "display more information for debugging") 55 + var versionFlag = pflag.BoolP("version", "V", false, "display version information") 46 56 47 57 func singleOperation() bool { 48 58 operations := 0 ··· 70 80 return operations == 1 71 81 } 72 82 73 - func displayFS(root fs.FS) error { 83 + func gitBlobSHA256(data []byte) string { 84 + h := crypto.SHA256.New() 85 + h.Write([]byte("blob ")) 86 + h.Write([]byte(strconv.FormatInt(int64(len(data)), 10))) 87 + h.Write([]byte{0}) 88 + h.Write(data) 89 + return hex.EncodeToString(h.Sum(nil)) 90 + } 91 + 92 + func displayFS(root fs.FS, prefix string) error { 74 93 return fs.WalkDir(root, ".", func(name string, entry fs.DirEntry, err error) error { 75 94 if err != nil { 76 95 return err 77 96 } 78 97 switch { 79 - case entry.Type() == 0: 80 - fmt.Fprintln(os.Stderr, "file", name) 81 - case entry.Type() == fs.ModeDir: 82 - fmt.Fprintln(os.Stderr, "dir", name) 98 + case entry.Type().IsDir(): 99 + fmt.Fprintf(os.Stderr, "dir %s%s\n", prefix, name) 100 + case entry.Type().IsRegular(): 101 + fmt.Fprintf(os.Stderr, "file %s%s\n", prefix, name) 83 102 case entry.Type() == fs.ModeSymlink: 84 - fmt.Fprintln(os.Stderr, "symlink", name) 103 + fmt.Fprintf(os.Stderr, "symlink %s%s\n", prefix, name) 85 104 default: 86 - fmt.Fprintln(os.Stderr, "other", name) 105 + fmt.Fprintf(os.Stderr, "other %s%s\n", prefix, name) 87 106 } 88 107 return nil 89 108 }) 90 109 } 91 110 92 - func archiveFS(writer io.Writer, root fs.FS) (err error) { 111 + // It doesn't make sense to use incremental updates for very small files since the cost of 112 + // repeating a request to fill in a missing blob is likely to be higher than any savings gained. 113 + const incrementalSizeThreshold = 256 114 + 115 + func archiveFS(writer io.Writer, root fs.FS, prefix string, needBlobs []string) (err error) { 116 + requestedSet := make(map[string]struct{}) 117 + for _, hash := range needBlobs { 118 + requestedSet[hash] = struct{}{} 119 + } 93 120 zstdWriter, _ := zstd.NewWriter(writer) 94 121 tarWriter := tar.NewWriter(zstdWriter) 95 - err = tarWriter.AddFS(root) 96 - if err != nil { 122 + if err = fs.WalkDir(root, ".", func(name string, entry fs.DirEntry, err error) error { 123 + if err != nil { 124 + return err 125 + } 126 + header := &tar.Header{} 127 + data := []byte{} 128 + if prefix == "" && name == "." { 129 + return nil 130 + } else if name == "." { 131 + header.Name = prefix 132 + } else { 133 + header.Name = prefix + name 134 + } 135 + switch { 136 + case entry.Type().IsDir(): 137 + header.Typeflag = tar.TypeDir 138 + header.Name += "/" 139 + case entry.Type().IsRegular(): 140 + header.Typeflag = tar.TypeReg 141 + if data, err = fs.ReadFile(root, name); err != nil { 142 + return err 143 + } 144 + if *incrementalFlag && len(data) > incrementalSizeThreshold { 145 + hash := gitBlobSHA256(data) 146 + if _, requested := requestedSet[hash]; !requested { 147 + header.Typeflag = tar.TypeSymlink 148 + header.Linkname = "/git/blobs/" + hash 149 + data = nil 150 + } 151 + } 152 + case entry.Type() == fs.ModeSymlink: 153 + header.Typeflag = tar.TypeSymlink 154 + if header.Linkname, err = fs.ReadLink(root, name); err != nil { 155 + return err 156 + } 157 + default: 158 + return errors.New("tar: cannot add non-regular file") 159 + } 160 + header.Size = int64(len(data)) 161 + if err = tarWriter.WriteHeader(header); err != nil { 162 + return err 163 + } 164 + if _, err = tarWriter.Write(data); err != nil { 165 + return err 166 + } 167 + return err 168 + }); err != nil { 97 169 return 98 170 } 99 - err = tarWriter.Close() 100 - if err != nil { 171 + if err = tarWriter.Close(); err != nil { 101 172 return 102 173 } 103 - err = zstdWriter.Close() 104 - if err != nil { 174 + if err = zstdWriter.Close(); err != nil { 105 175 return 106 176 } 107 177 return 108 178 } 109 179 180 + // Stream archive data without ever loading the entire working set into RAM. 181 + func streamArchiveFS(root fs.FS, prefix string, needBlobs []string) io.ReadCloser { 182 + reader, writer := io.Pipe() 183 + go func() { 184 + err := archiveFS(writer, root, prefix, needBlobs) 185 + if err != nil { 186 + writer.CloseWithError(err) 187 + } else { 188 + writer.Close() 189 + } 190 + }() 191 + return reader 192 + } 193 + 194 + func makeWhiteout(path string) (reader io.Reader) { 195 + buffer := &bytes.Buffer{} 196 + tarWriter := tar.NewWriter(buffer) 197 + tarWriter.WriteHeader(&tar.Header{ 198 + Typeflag: tar.TypeChar, 199 + Name: path, 200 + }) 201 + tarWriter.Flush() 202 + return buffer 203 + } 204 + 110 205 const usageExitCode = 125 111 206 207 + func usage() { 208 + fmt.Fprintf(os.Stderr, 209 + "Usage: %s <site-url> {--challenge|--upload-git url|--upload-dir path|--delete} [options...]\n", 210 + os.Args[0], 211 + ) 212 + pflag.PrintDefaults() 213 + } 214 + 112 215 func main() { 216 + pflag.Usage = usage 113 217 pflag.Parse() 114 218 if !singleOperation() || (!*versionFlag && len(pflag.Args()) != 1) { 115 - fmt.Fprintf(os.Stderr, 116 - "Usage: %s <site-url> [--challenge|--upload-git url|--upload-dir path|--delete]\n", 117 - os.Args[0], 118 - ) 219 + pflag.Usage() 119 220 os.Exit(usageExitCode) 120 221 } 121 222 ··· 129 230 os.Exit(usageExitCode) 130 231 } 131 232 233 + var pathPrefix string 234 + if *pathFlag != "" { 235 + if *uploadDirFlag == "" && !*deleteFlag { 236 + fmt.Fprintf(os.Stderr, "--path requires --upload-dir or --delete") 237 + os.Exit(usageExitCode) 238 + } else { 239 + pathPrefix = strings.Trim(*pathFlag, "/") + "/" 240 + } 241 + } 242 + 132 243 var err error 133 244 siteURL, err := url.Parse(pflag.Args()[0]) 134 245 if err != nil { ··· 137 248 } 138 249 139 250 var request *http.Request 251 + var uploadDir *os.Root 140 252 switch { 141 253 case *challengeFlag || *challengeBareFlag: 142 254 if *passwordFlag == "" { ··· 168 280 request.Header.Add("Content-Type", "application/x-www-form-urlencoded") 169 281 170 282 case *uploadDirFlag != "": 171 - uploadDirFS, err := os.OpenRoot(*uploadDirFlag) 283 + uploadDir, err = os.OpenRoot(*uploadDirFlag) 172 284 if err != nil { 173 285 fmt.Fprintf(os.Stderr, "error: invalid directory: %s\n", err) 174 286 os.Exit(1) 175 287 } 176 288 177 289 if *verboseFlag { 178 - err := displayFS(uploadDirFS.FS()) 290 + err := displayFS(uploadDir.FS(), pathPrefix) 179 291 if err != nil { 180 292 fmt.Fprintf(os.Stderr, "error: %s\n", err) 181 293 os.Exit(1) 182 294 } 183 295 } 184 296 185 - // Stream archive data without ever loading the entire working set into RAM. 186 - reader, writer := io.Pipe() 187 - go func() { 188 - err = archiveFS(writer, uploadDirFS.FS()) 189 - if err != nil { 190 - fmt.Fprintf(os.Stderr, "error: %s\n", err) 191 - os.Exit(1) 192 - } 193 - writer.Close() 194 - }() 195 - 196 - request, err = http.NewRequest("PUT", siteURL.String(), reader) 297 + if *pathFlag == "" { 298 + request, err = http.NewRequest("PUT", siteURL.String(), nil) 299 + } else { 300 + request, err = http.NewRequest("PATCH", siteURL.String(), nil) 301 + } 197 302 if err != nil { 198 303 fmt.Fprintf(os.Stderr, "error: %s\n", err) 199 304 os.Exit(1) 200 305 } 306 + request.Body = streamArchiveFS(uploadDir.FS(), pathPrefix, []string{}) 201 307 request.ContentLength = -1 202 308 request.Header.Add("Content-Type", "application/x-tar+zstd") 309 + request.Header.Add("Accept", "application/vnd.git-pages.unresolved;q=1.0, text/plain;q=0.9") 310 + if *parentsFlag { 311 + request.Header.Add("Create-Parents", "yes") 312 + } else { 313 + request.Header.Add("Create-Parents", "no") 314 + } 203 315 204 316 case *deleteFlag: 205 - request, err = http.NewRequest("DELETE", siteURL.String(), nil) 206 - if err != nil { 207 - fmt.Fprintf(os.Stderr, "error: %s\n", err) 208 - os.Exit(1) 317 + if *pathFlag == "" { 318 + request, err = http.NewRequest("DELETE", siteURL.String(), nil) 319 + if err != nil { 320 + fmt.Fprintf(os.Stderr, "error: %s\n", err) 321 + os.Exit(1) 322 + } 323 + } else { 324 + request, err = http.NewRequest("PATCH", siteURL.String(), makeWhiteout(pathPrefix)) 325 + if err != nil { 326 + fmt.Fprintf(os.Stderr, "error: %s\n", err) 327 + os.Exit(1) 328 + } 329 + request.Header.Add("Content-Type", "application/x-tar") 209 330 } 210 331 211 332 case *debugManifestFlag: ··· 220 341 panic("no operation chosen") 221 342 } 222 343 request.Header.Add("User-Agent", versionInfo()) 344 + if request.Method == "PATCH" { 345 + if *atomicFlag { 346 + request.Header.Add("Atomic", "yes") 347 + request.Header.Add("Race-Free", "yes") // deprecated name, to be removed soon 348 + } else { 349 + request.Header.Add("Atomic", "no") 350 + request.Header.Add("Race-Free", "no") // deprecated name, to be removed soon 351 + } 352 + } 223 353 switch { 224 354 case *passwordFlag != "": 225 355 request.Header.Add("Authorization", fmt.Sprintf("Pages %s", *passwordFlag)) ··· 237 367 request.Header.Set("Host", siteURL.Host) 238 368 } 239 369 240 - response, err := http.DefaultClient.Do(request) 241 - if err != nil { 242 - fmt.Fprintf(os.Stderr, "error: %s\n", err) 243 - os.Exit(1) 244 - } 245 - if *verboseFlag { 246 - fmt.Fprintf(os.Stderr, "server: %s\n", response.Header.Get("Server")) 247 - } 248 - if *debugManifestFlag { 249 - if response.StatusCode == 200 { 250 - io.Copy(os.Stdout, response.Body) 251 - fmt.Fprintf(os.Stdout, "\n") 252 - } else { 253 - io.Copy(os.Stderr, response.Body) 370 + displayServer := *verboseFlag 371 + for { 372 + response, err := http.DefaultClient.Do(request) 373 + if err != nil { 374 + fmt.Fprintf(os.Stderr, "error: %s\n", err) 254 375 os.Exit(1) 255 376 } 256 - } else { // an update operation 257 - if response.StatusCode == 200 { 258 - fmt.Fprintf(os.Stdout, "result: %s\n", response.Header.Get("Update-Result")) 259 - io.Copy(os.Stdout, response.Body) 260 - } else { 261 - fmt.Fprintf(os.Stderr, "result: error\n") 262 - io.Copy(os.Stderr, response.Body) 263 - os.Exit(1) 377 + if displayServer { 378 + fmt.Fprintf(os.Stderr, "server: %s\n", response.Header.Get("Server")) 379 + displayServer = false 380 + } 381 + if *debugManifestFlag { 382 + if response.StatusCode == http.StatusOK { 383 + io.Copy(os.Stdout, response.Body) 384 + fmt.Fprintf(os.Stdout, "\n") 385 + } else { 386 + io.Copy(os.Stderr, response.Body) 387 + os.Exit(1) 388 + } 389 + } else { // an update operation 390 + if *verboseFlag { 391 + fmt.Fprintf(os.Stderr, "response: %d %s\n", 392 + response.StatusCode, response.Header.Get("Content-Type")) 393 + } 394 + if response.StatusCode == http.StatusUnprocessableEntity && 395 + response.Header.Get("Content-Type") == "application/vnd.git-pages.unresolved" { 396 + needBlobs := []string{} 397 + scanner := bufio.NewScanner(response.Body) 398 + for scanner.Scan() { 399 + needBlobs = append(needBlobs, scanner.Text()) 400 + } 401 + response.Body.Close() 402 + if *verboseFlag { 403 + fmt.Fprintf(os.Stderr, "incremental: need %d blobs\n", len(needBlobs)) 404 + } 405 + request.Body = streamArchiveFS(uploadDir.FS(), pathPrefix, needBlobs) 406 + continue // resubmit 407 + } else if response.StatusCode == http.StatusOK { 408 + fmt.Fprintf(os.Stdout, "result: %s\n", response.Header.Get("Update-Result")) 409 + io.Copy(os.Stdout, response.Body) 410 + } else { 411 + fmt.Fprintf(os.Stderr, "result: error\n") 412 + io.Copy(os.Stderr, response.Body) 413 + os.Exit(1) 414 + } 264 415 } 416 + break 265 417 } 266 418 }
+3 -1
renovate.json
··· 9 9 "lockFileMaintenance": { 10 10 "enabled": true, 11 11 "automerge": false 12 - } 12 + }, 13 + "semanticCommits": "disabled", 14 + "commitMessagePrefix": "[Renovate]" 13 15 }