Skip to content

Conversation

@sarutak
Copy link
Member

@sarutak sarutak commented Mar 10, 2017

What changes were proposed in this pull request?

When we try to run Structured Streaming in local mode but use HDFS for the storage, batches will be fail because of error like as follows.

val handle = stream.writeStream.format("console").start()
17/03/09 16:54:45 ERROR StreamMetadata: Error writing stream metadata StreamMetadata(fc07a0b1-5423-483e-a59d-b2206a49491e) to /private/var/folders/4y/tmspvv353y59p3w4lknrf7cc0000gn/T/temporary-79d4fe05-4301-4b6d-a902-dff642d0ddca/metadata
org.apache.hadoop.security.AccessControlException: Permission denied: user=kou, access=WRITE, inode="/private/var/folders/4y/tmspvv353y59p3w4lknrf7cc0000gn/T/temporary-79d4fe05-4301-4b6d-a902-dff642d0ddca/metadata":hdfs:supergroup:drwxr-xr-x

It's because that a temporary checkpoint directory is created on local file system but metadata whose path is based on the checkpoint directory will be created on HDFS.

This PR will fixe this issue.

How was this patch tested?

I tested manually in local mode with HDFS.

@SparkQA
Copy link

SparkQA commented Mar 10, 2017

Test build #74334 has finished for PR 17248 at commit cc7a3f8.

  • This patch passes all tests.
  • This patch merges cleanly.
  • This patch adds no public classes.

@mgaido91
Copy link
Contributor

I don't think this PR resolve properly the issue. Indeed, it somewhat forces the metadata to be written in a local dir instead of the configured default filesystem.
Of course, this fixes the exception, but we loose all the benefits of a distributed file system, as fault tolerance.
Thus in my opinion, it would be better to let the metadata be written on the default file system, but changing the default location.

@gatorsmile
Copy link
Member

We are closing it due to inactivity. please do reopen if you want to push it forward. Thanks!

@sarutak
Copy link
Member Author

sarutak commented Jun 27, 2017

O.K. I'll close this PR. Thanks!

@sarutak sarutak closed this Jun 27, 2017
ghost pushed a commit to dbtsai/spark that referenced this pull request Jun 27, 2017
## What changes were proposed in this pull request?

This PR proposes to close stale PRs, mostly the same instances with apache#18017

I believe the author in apache#14807 removed his account.

Closes apache#7075
Closes apache#8927
Closes apache#9202
Closes apache#9366
Closes apache#10861
Closes apache#11420
Closes apache#12356
Closes apache#13028
Closes apache#13506
Closes apache#14191
Closes apache#14198
Closes apache#14330
Closes apache#14807
Closes apache#15839
Closes apache#16225
Closes apache#16685
Closes apache#16692
Closes apache#16995
Closes apache#17181
Closes apache#17211
Closes apache#17235
Closes apache#17237
Closes apache#17248
Closes apache#17341
Closes apache#17708
Closes apache#17716
Closes apache#17721
Closes apache#17937

Added:
Closes apache#14739
Closes apache#17139
Closes apache#17445
Closes apache#18042
Closes apache#18359

Added:
Closes apache#16450
Closes apache#16525
Closes apache#17738

Added:
Closes apache#16458
Closes apache#16508
Closes apache#17714

Added:
Closes apache#17830
Closes apache#14742

## How was this patch tested?

N/A

Author: hyukjinkwon <[email protected]>

Closes apache#18417 from HyukjinKwon/close-stale-pr.
@sarutak sarutak deleted the SPARK-19909 branch June 4, 2021 20:45
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

4 participants