* Automatically install latest pyspark version
* Better text
* Do not use shutil to keep behaviour
* Make setup_script cwd independent
* Use _get_program_version to calculate spark version
* Update setup_spark.py reqs
* Update setup_spark.py
* Add info about HADOOP_VERSION
* Add customization back
* Better text
* Specify build args when they are actually needed
* Better text
* Better code
* Better code
* Better text
* Get rid of warning
* Improve code
* Remove information about checksum
* Better text
* add grpcio grpcio_status to support spark connect
* Sort install list
* Fix package name
* Update pyspark docs with new deps grpcio and grpcio-status
* set grpcio and grpcio-status version as 1.56
* exclude grpcio and grpcio-status in test_packages.py
* Update selecting.md
* Update test_packages.py
* Update Dockerfile
---------
Co-authored-by: Ayaz Salikhov <mathbunnyru@users.noreply.github.com>
* 1.
* add note for pandas version
* [pre-commit.ci] auto fixes from pre-commit.com hooks
for more information, see https://pre-commit.ci
* Update images/pyspark-notebook/Dockerfile
Co-authored-by: Ayaz Salikhov <mathbunnyru@users.noreply.github.com>
---------
Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
Co-authored-by: Ayaz Salikhov <mathbunnyru@users.noreply.github.com>