lol

python310Packages.pyspark: 3.3.0 -> 3.3.1

+24 -3
+24 -3
pkgs/development/python-modules/pyspark/default.nix
··· 1 { lib 2 , buildPythonPackage 3 , fetchPypi 4 , py4j 5 }: 6 7 buildPythonPackage rec { 8 pname = "pyspark"; 9 - version = "3.3.0"; 10 11 src = fetchPypi { 12 inherit pname version; 13 - sha256 = "sha256-fr6OlQVke00STVqC/KYN/TiRAhz4rWxeyId37uzpLPc="; 14 }; 15 16 # pypandoc is broken with pandoc2, so we just lose docs. ··· 25 py4j 26 ]; 27 28 # Tests assume running spark instance 29 doCheck = false; 30 ··· 40 binaryBytecode 41 ]; 42 license = licenses.asl20; 43 - maintainers = [ maintainers.shlevy ]; 44 }; 45 }
··· 1 { lib 2 , buildPythonPackage 3 , fetchPypi 4 + , numpy 5 + , pandas 6 , py4j 7 + , pyarrow 8 + , pythonOlder 9 }: 10 11 buildPythonPackage rec { 12 pname = "pyspark"; 13 + version = "3.3.1"; 14 + format = "setuptools"; 15 + 16 + disabled = pythonOlder "3.7"; 17 18 src = fetchPypi { 19 inherit pname version; 20 + hash = "sha256-6Z+n3pK+QGiEv9gxwyuTBqOpneRM/Dmi7vtu0HRF1fo="; 21 }; 22 23 # pypandoc is broken with pandoc2, so we just lose docs. ··· 32 py4j 33 ]; 34 35 + passthru.optional-dependencies = { 36 + ml = [ 37 + numpy 38 + ]; 39 + mllib = [ 40 + numpy 41 + ]; 42 + sql = [ 43 + numpy 44 + pandas 45 + pyarrow 46 + ]; 47 + }; 48 + 49 # Tests assume running spark instance 50 doCheck = false; 51 ··· 61 binaryBytecode 62 ]; 63 license = licenses.asl20; 64 + maintainers = with maintainers; [ shlevy ]; 65 }; 66 }