2021-02-05 17:12:51 +00:00
|
|
|
{ buildPythonPackage, fetchPypi, lib, py4j }:
|
2020-04-24 23:36:52 +00:00
|
|
|
|
|
|
|
buildPythonPackage rec {
|
|
|
|
pname = "pyspark";
|
2021-04-26 19:14:03 +00:00
|
|
|
version = "3.1.1";
|
2020-04-24 23:36:52 +00:00
|
|
|
|
|
|
|
src = fetchPypi {
|
|
|
|
inherit pname version;
|
2021-04-26 19:14:03 +00:00
|
|
|
sha256 = "104abc146d4ffb72d4c683d25d7af5a6bf955d94590a76f542ee23185670aa7e";
|
2020-04-24 23:36:52 +00:00
|
|
|
};
|
|
|
|
|
|
|
|
# pypandoc is broken with pandoc2, so we just lose docs.
|
|
|
|
postPatch = ''
|
|
|
|
sed -i "s/'pypandoc'//" setup.py
|
|
|
|
|
2020-11-24 20:58:05 +00:00
|
|
|
substituteInPlace setup.py --replace py4j==0.10.9 'py4j>=0.10.9,<0.11'
|
2020-04-24 23:36:52 +00:00
|
|
|
'';
|
|
|
|
|
|
|
|
propagatedBuildInputs = [ py4j ];
|
|
|
|
|
|
|
|
# Tests assume running spark...
|
|
|
|
doCheck = false;
|
|
|
|
|
2021-01-15 22:18:51 +00:00
|
|
|
meta = with lib; {
|
2020-04-24 23:36:52 +00:00
|
|
|
description = "Apache Spark";
|
|
|
|
homepage = "https://github.com/apache/spark/tree/master/python";
|
|
|
|
license = licenses.asl20;
|
|
|
|
maintainers = [ maintainers.shlevy ];
|
|
|
|
};
|
|
|
|
}
|