···3232 # https://github.com/NixOS/nixpkgs/blob/master/pkgs/tools/misc/ollama/default.nix
33333434 pname = "tabby";
3535- version = "0.11.0";
3535+ version = "0.11.1";
363637373838 availableAccelerations = flatten [
···7878 # to use a specific device type as it is relying on llama-cpp only being
7979 # built to use one type of device.
8080 #
8181- # See: https://github.com/TabbyML/tabby/blob/v0.11.0/crates/llama-cpp-bindings/include/engine.h#L20
8181+ # See: https://github.com/TabbyML/tabby/blob/v0.11.1/crates/llama-cpp-bindings/include/engine.h#L20
8282 #
8383 llamaccpPackage = llama-cpp.override {
8484 rocmSupport = enableRocm;
···108108 owner = "TabbyML";
109109 repo = "tabby";
110110 rev = "v${version}";
111111- hash = "sha256-7PHCvI2/QSFVs6SAXrx8rohU5Wu8QRaqsI1CimnUGQY=";
111111+ hash = "sha256-OgAE526aW3mVqf6fVmBmL5/B4gH9B54QLEITQk9Kgsg=";
112112 fetchSubmodules = true;
113113 };
114114