Skip to content

Conversation

@liancheng
Copy link
Contributor

For Hadoop 1.x, TaskAttemptContext constructor clones the Configuration argument, thus configurations done in HadoopFsRelation.prepareForWriteJob() are not populated to driver side TaskAttemptContext (executor side configurations are properly populated). Currently this should only affect Parquet output committer class configuration.

@liancheng
Copy link
Contributor Author

cc @yhuai

@liancheng liancheng force-pushed the spark-8121 branch 3 times, most recently from 046a01e to 910b8aa Compare June 5, 2015 09:50
@SparkQA
Copy link

SparkQA commented Jun 5, 2015

Test build #34264 has finished for PR 6669 at commit c46b044.

  • This patch fails Spark unit tests.
  • This patch does not merge cleanly.
  • This patch adds no public classes.

@SparkQA
Copy link

SparkQA commented Jun 5, 2015

Test build #34267 has finished for PR 6669 at commit 910b8aa.

  • This patch passes all tests.
  • This patch merges cleanly.
  • This patch adds no public classes.

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Seems we have protected def configuration = sparkContext.hadoopConfiguration in SQLTestUtils and we are not cloning it at here?

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Oops, thanks!

@SparkQA
Copy link

SparkQA commented Jun 6, 2015

Test build #34344 has finished for PR 6669 at commit e1ede07.

  • This patch fails Spark unit tests.
  • This patch merges cleanly.
  • This patch adds no public classes.

@SparkQA
Copy link

SparkQA commented Jun 6, 2015

Test build #34355 has finished for PR 6669 at commit 1340f25.

  • This patch passes all tests.
  • This patch merges cleanly.
  • This patch adds no public classes.

@SparkQA
Copy link

SparkQA commented Jun 7, 2015

Test build #34390 has finished for PR 6669 at commit b6f78a6.

  • This patch passes all tests.
  • This patch merges cleanly.
  • This patch adds no public classes.

@yhuai
Copy link
Contributor

yhuai commented Jun 7, 2015

Should we log the output committer used by parquet?

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

This one seems not right?

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Maybe we can get the instance of outputCommitter first and then log its class?

val outputCommitter = ...
logInfo(...)
outputCommitter

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Oops, thanks!

@SparkQA
Copy link

SparkQA commented Jun 8, 2015

Test build #34407 has finished for PR 6669 at commit fce089c.

  • This patch passes all tests.
  • This patch merges cleanly.
  • This patch adds the following public classes (experimental):
    • logInfo(s"Using user defined output committer class $
    • logInfo(s"Using output committer class $

liancheng added a commit to liancheng/spark that referenced this pull request Jun 8, 2015
@SparkQA
Copy link

SparkQA commented Jun 8, 2015

Test build #34431 has finished for PR 6669 at commit 73819e8.

  • This patch passes all tests.
  • This patch merges cleanly.
  • This patch adds the following public classes (experimental):
    • logInfo(s"Using user defined output committer class $
    • logInfo(s"Using output committer class $

@yhuai
Copy link
Contributor

yhuai commented Jun 8, 2015

LGTM. I am merging it to master.

@asfgit asfgit closed this in bbdfc0a Jun 8, 2015
asfgit pushed a commit that referenced this pull request Jun 8, 2015
nemccarthy pushed a commit to nemccarthy/spark that referenced this pull request Jun 19, 2015
…n for Hadoop 1.x

For Hadoop 1.x, `TaskAttemptContext` constructor clones the `Configuration` argument, thus configurations done in `HadoopFsRelation.prepareForWriteJob()` are not populated to *driver* side `TaskAttemptContext` (executor side configurations are properly populated). Currently this should only affect Parquet output committer class configuration.

Author: Cheng Lian <[email protected]>

Closes apache#6669 from liancheng/spark-8121 and squashes the following commits:

73819e8 [Cheng Lian] Minor logging fix
fce089c [Cheng Lian] Adds more logging
b6f78a6 [Cheng Lian] Fixes compilation error introduced while rebasing
963a1aa [Cheng Lian] Addresses @yhuai's comment
c3a0b1a [Cheng Lian] Fixes InsertIntoHadoopFsRelation job initialization
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants