pyspark.sql.SparkSession.addArtifacts¶
-
SparkSession.
addArtifacts
(*path: str, pyfile: bool = False, archive: bool = False, file: bool = False) → None[source]¶ Add artifact(s) to the client session. Currently only local files are supported.
New in version 3.5.0.
- Parameters
- *pathtuple of str
Artifact’s URIs to add.
- pyfilebool
Whether to add them as Python dependencies such as .py, .egg, .zip or .jar files. The pyfiles are directly inserted into the path when executing Python functions in executors.
- archivebool
Whether to add them as archives such as .zip, .jar, .tar.gz, .tgz, or .tar files. The archives are unpacked on the executor side automatically.
- filebool
Add a file to be downloaded with this Spark job on every node. The
path
passed can only be a local file for now.
Notes
This is an API dedicated to Spark Connect client only. With regular Spark Session, it throws an exception.