<<>strong>TLDR: 除非我的Jupyter说明书能够查阅数据地点,否则我为什么没有填写德尔塔表格,这与我预期“Se”公司应当独立于Jupyter数据查阅处理书面材料相反?
I ve set up a PySpark Jupyter Notebook connected to a Spark cluster, where the Spark instance is intended to perform writes to a Delta table. However, I m observing that the Spark instance fails to complete the writes if the Jupyter Notebook doesn t have access to the data location. Repo for reproducibility. Specific PR that reproduces the bug.
<>Setup:
version: 3
services:
spark:
image: com/data_lake_spark:latest
# Spark service configuration details...
spark-worker-1:
# Configuration details...
spark-worker-2:
# Configuration details...
jupyter:
image: com/data_lake_notebook:latest
# Jupyter Notebook service configuration details...
www.un.org/Depts/DGACM/index_spanish.htm 届会的配置:
# Spark session setup...
http://www.hchr.org。
# Write initial test data to Delta table
owner_df.write.format("delta").mode("overwrite").save(delta_output_path)
移除Jupyter进入Docker Compose配置的名录/data
,在试图写到Delta表格时,导致了DeltaIOException。 但是,如果能够查阅<代码>/数据目录,则可以成功撰写。
www.un.org/Depts/DGACM/index_spanish.htm Error 信:
Py4JJavaError: An error occurred while calling o56.save.
: org.apache.spark.sql.delta.DeltaIOException: [DELTA_CANNOT_CREATE_LOG_PATH] Cannot create file:/data/delta_table_of_dog_owners/_delta_log
at org.apache.spark.sql.delta.DeltaErrorsBase.cannotCreateLogPathException(DeltaErrors.scala:1534)
at org.apache.spark.sql.delta.DeltaErrorsBase.cannotCreateLogPathException$(DeltaErrors.scala:1533)
at org.apache.spark.sql.delta.DeltaErrors$.cannotCreateLogPathException(DeltaErrors.scala:3203)
at org.apache.spark.sql.delta.DeltaLog.createDirIfNotExists$1(DeltaLog.scala:443)
我期望“花旗”能够独立处理Jupyter数据存取。 为解决这一问题寻求见解或建议。 希望得到任何指导。