lol
0
fork

Configure Feed

Select the types of activity you want to include in your feed.

openai-whisper: 20230918 -> 20231117

https://github.com/openai/whisper/blob/v20231117/CHANGELOG.md

Always propagate openai-triton, since whether or not it is free is
decided by `nixpkgs.config.cudaSupport`.

This allows us to drop the local `cudaSupport` flags in favor of the
nixpkgs-wide one.

+20 -32
+19 -28
pkgs/development/python-modules/openai-whisper/default.nix
··· 2 2 , fetchFromGitHub 3 3 , buildPythonPackage 4 4 , substituteAll 5 - , cudaSupport ? false 5 + 6 + # build-system 7 + , setuptools 6 8 7 9 # runtime 8 10 , ffmpeg-headless 9 11 10 12 # propagates 11 - , numpy 12 - , torch 13 - , torchWithCuda 14 - , tqdm 15 13 , more-itertools 16 - , transformers 17 14 , numba 15 + , numpy 18 16 , openai-triton 19 17 , scipy 20 18 , tiktoken 19 + , torch 20 + , tqdm 21 + , transformers 21 22 22 23 # tests 23 24 , pytestCheckHook ··· 25 26 26 27 buildPythonPackage rec { 27 28 pname = "whisper"; 28 - version = "20230918"; 29 - format = "setuptools"; 29 + version = "20231117"; 30 + pyproject = true; 30 31 31 32 src = fetchFromGitHub { 32 33 owner = "openai"; 33 34 repo = pname; 34 35 rev = "refs/tags/v${version}"; 35 - hash = "sha256-wBAanFVEIIzTcoX40P9eI26UdEu0SC/xuife/zi2Xho="; 36 + hash = "sha256-MJ1XjB/GuYUiECCuuHS0NWHvvs+ko0oTvLuDI7zLNiY="; 36 37 }; 37 38 38 39 patches = [ ··· 42 43 }) 43 44 ]; 44 45 46 + nativeBuildInputs = [ 47 + setuptools 48 + ]; 49 + 45 50 propagatedBuildInputs = [ 46 - numpy 47 - tqdm 48 51 more-itertools 49 - transformers 50 52 numba 53 + numpy 54 + openai-triton 51 55 scipy 52 56 tiktoken 53 - ] ++ lib.optionals (!cudaSupport) [ 54 57 torch 55 - ] ++ lib.optionals (cudaSupport) [ 56 - openai-triton 57 - torchWithCuda 58 + tqdm 59 + transformers 58 60 ]; 59 61 60 - postPatch = '' 61 - substituteInPlace requirements.txt \ 62 - --replace "tiktoken==0.3.3" "tiktoken>=0.3.3" 63 - '' 64 - # openai-triton is only needed for CUDA support. 65 - # triton needs CUDA to be build. 66 - # -> by making it optional, we can build whisper without unfree packages enabled 67 - + lib.optionalString (!cudaSupport) '' 68 - sed -i '/if sys.platform.startswith("linux") and platform.machine() == "x86_64":/{N;d}' setup.py 69 - ''; 70 - 71 62 preCheck = '' 72 63 export HOME=$TMPDIR 73 64 ''; ··· 85 76 ]; 86 77 87 78 meta = with lib; { 88 - changelog = "https://github.com/openai/whisper/blob/v$[version}/CHANGELOG.md"; 79 + changelog = "https://github.com/openai/whisper/blob/v${version}/CHANGELOG.md"; 89 80 description = "General-purpose speech recognition model"; 90 81 homepage = "https://github.com/openai/whisper"; 91 82 license = licenses.mit;
+1 -4
pkgs/top-level/python-packages.nix
··· 8456 8456 8457 8457 openai-triton-bin = callPackage ../development/python-modules/openai-triton/bin.nix { }; 8458 8458 8459 - openai-whisper = callPackage ../development/python-modules/openai-whisper { 8460 - inherit (pkgs.config) cudaSupport; 8461 - openai-triton = self.openai-triton-cuda; 8462 - }; 8459 + openai-whisper = callPackage ../development/python-modules/openai-whisper { }; 8463 8460 8464 8461 openant = callPackage ../development/python-modules/openant { }; 8465 8462