Skip to content

Conversation

yeweizhang
Copy link

When we do the size estimation in joining, spark calls hive metastore to get the size in bytes. Add callback so that we will call hadoop file system directly.

The initial attempt to use the shuffle partition size in local properties. If we can get the adaptive execution work, then this work is redundant.

@markhamstra @mbautin

markhamstra added a commit that referenced this pull request Nov 4, 2015
SKIPME: Add logic to get the size estimation from Hadoop filesystem.
@markhamstra markhamstra merged commit 9f90518 into alteryx:csd-1.5 Nov 4, 2015
markhamstra pushed a commit to markhamstra/spark that referenced this pull request Nov 7, 2017
* Add -DskipTests to dev docs

* Remove extraneous skipTests
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants