2020-04-24 23:36:52 +00:00
|
|
|
{ buildPythonPackage, fetchPypi, stdenv, py4j }:
|
|
|
|
|
|
|
|
buildPythonPackage rec {
|
|
|
|
pname = "pyspark";
|
2020-10-19 00:13:06 +00:00
|
|
|
version = "3.0.1";
|
2020-04-24 23:36:52 +00:00
|
|
|
|
|
|
|
src = fetchPypi {
|
|
|
|
inherit pname version;
|
2020-10-19 00:13:06 +00:00
|
|
|
sha256 = "38b485d3634a86c9a2923c39c8f08f003fdd0e0a3d7f07114b2fb4392ce60479";
|
2020-04-24 23:36:52 +00:00
|
|
|
};
|
|
|
|
|
|
|
|
# pypandoc is broken with pandoc2, so we just lose docs.
|
|
|
|
postPatch = ''
|
|
|
|
sed -i "s/'pypandoc'//" setup.py
|
|
|
|
|
2020-11-24 20:58:05 +00:00
|
|
|
substituteInPlace setup.py --replace py4j==0.10.9 'py4j>=0.10.9,<0.11'
|
2020-04-24 23:36:52 +00:00
|
|
|
'';
|
|
|
|
|
|
|
|
propagatedBuildInputs = [ py4j ];
|
|
|
|
|
|
|
|
# Tests assume running spark...
|
|
|
|
doCheck = false;
|
|
|
|
|
|
|
|
meta = with stdenv.lib; {
|
|
|
|
description = "Apache Spark";
|
|
|
|
homepage = "https://github.com/apache/spark/tree/master/python";
|
|
|
|
license = licenses.asl20;
|
|
|
|
maintainers = [ maintainers.shlevy ];
|
|
|
|
};
|
|
|
|
}
|