Skip to content

Conversation

@wangyum
Copy link
Member

@wangyum wangyum commented Feb 22, 2018

What changes were proposed in this pull request?

Check if there is any test failed.

* Update Hive to 2.3.2
@SparkQA
Copy link

SparkQA commented Feb 22, 2018

Test build #87611 has finished for PR 20659 at commit 915e68f.

  • This patch fails build dependency tests.
  • This patch merges cleanly.
  • This patch adds no public classes.

@SparkQA
Copy link

SparkQA commented Feb 22, 2018

Test build #87613 has finished for PR 20659 at commit a5bb731.

  • This patch fails to generate documentation.
  • This patch merges cleanly.
  • This patch adds no public classes.

@SparkQA
Copy link

SparkQA commented Feb 22, 2018

Test build #87616 has finished for PR 20659 at commit 80fd8a8.

  • This patch fails Spark unit tests.
  • This patch merges cleanly.
  • This patch adds no public classes.

@wangyum wangyum closed this Feb 22, 2018
# Conflicts:
#	sql/hive/src/main/scala/org/apache/spark/sql/hive/client/HiveClientImpl.scala
#	sql/hive/src/main/scala/org/apache/spark/sql/hive/client/HiveShim.scala
#	sql/hive/src/main/scala/org/apache/spark/sql/hive/client/IsolatedClientLoader.scala
#	sql/hive/src/main/scala/org/apache/spark/sql/hive/client/package.scala
#	sql/hive/src/main/scala/org/apache/spark/sql/hive/execution/SaveAsHiveFile.scala
#	sql/hive/src/test/scala/org/apache/spark/sql/hive/client/VersionsSuite.scala
@wangyum wangyum reopened this Mar 5, 2018
@SparkQA
Copy link

SparkQA commented Mar 5, 2018

Test build #87960 has finished for PR 20659 at commit 1110ede.

  • This patch fails Spark unit tests.
  • This patch merges cleanly.
  • This patch adds no public classes.

@dongjoon-hyun
Copy link
Member

Nice try! Could you fix the remaining failure?

@wangyum
Copy link
Member Author

wangyum commented Mar 6, 2018

Yes, I'm doing it

<dependency>
<groupId>org.apache.hive</groupId>
<artifactId>hive-storage-api</artifactId>
</dependency>
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

BTW, you should not introduce hive into sql/core.
Apache ORC should be used as Apache ORC.

<dependency>
<groupId>org.apache.orc</groupId>
<artifactId>orc-core</artifactId>
<classifier>${orc.classifier}</classifier>
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Please don't remove this.

import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.hive.common.type.HiveDecimal;
import org.apache.hadoop.hive.ql.exec.vector.*;
import org.apache.hadoop.hive.serde2.io.HiveDecimalWritable;
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Please revert this.

withTable("jt2") {
withView("testView") {
sql("CREATE VIEW testView AS SELECT id FROM jt")
sql("CREATE VIEW testView AS SELECT 1 as c1 FROM jt")
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

It is unrelated change and looks wrong.

sql("ALTER TABLE t1 ADD COLUMNS (C2 string)")
assert(spark.table("t1").schema ==
new StructType().add("c1", IntegerType).add("C1", StringType))
new StructType().add("c1", IntegerType).add("C2", StringType))
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

You are intentionally removing case sensitive test.

@dongjoon-hyun
Copy link
Member

In general, this should keep all existing test coverage. Otherwise, it will not considered as a feasible solution.

@HyukjinKwon
Copy link
Member

just out of curiosity, what does DNM stands for?

@dongjoon-hyun
Copy link
Member

I guess Do Not Merge. Actually, this is a try to see the result. :)

@gatorsmile
Copy link
Member

This is a very interesting work. Thanks for your effort!

When we implement such a feature, please remember we have to ensure it is configurable.

@wangyum wangyum changed the title [DNM] Try to update Hive to 2.3.2 [DO-NOT-MERGE] Try to update Hive to 2.3.2 Mar 10, 2018
@jerryshao
Copy link
Contributor

@gatorsmile @cloud-fan what is our plan for such thing? Currently we're using our own forked version, which blocks the support of Hadoop 3 (SPARK-18673). I'm wondering if we're going to upgrade Hive like what is done in this PR, or we still uses forked branch, and patch that branch.

What is your thought and plan on it?

@jerryshao
Copy link
Contributor

@wangyum can you please create a relative JIRA, so that we discuss on the JIRA. I think it is better to discuss and get a conclusion first.

@SparkQA
Copy link

SparkQA commented Mar 13, 2018

Test build #88206 has finished for PR 20659 at commit b35daa0.

  • This patch fails Spark unit tests.
  • This patch merges cleanly.
  • This patch adds no public classes.

@SparkQA
Copy link

SparkQA commented Mar 14, 2018

Test build #88217 has finished for PR 20659 at commit b418909.

  • This patch fails to build.
  • This patch merges cleanly.
  • This patch adds no public classes.

@wangyum
Copy link
Member Author

wangyum commented Mar 14, 2018

retest this please

@SparkQA
Copy link

SparkQA commented Mar 15, 2018

Test build #88260 has finished for PR 20659 at commit 03138d8.

  • This patch fails Spark unit tests.
  • This patch merges cleanly.
  • This patch adds no public classes.

@SparkQA
Copy link

SparkQA commented Mar 15, 2018

Test build #88269 has finished for PR 20659 at commit e041704.

  • This patch fails Spark unit tests.
  • This patch merges cleanly.
  • This patch adds no public classes.

@wangyum
Copy link
Member Author

wangyum commented Mar 15, 2018

retest this please

@SparkQA
Copy link

SparkQA commented Mar 16, 2018

Test build #88275 has finished for PR 20659 at commit e041704.

  • This patch passes all tests.
  • This patch merges cleanly.
  • This patch adds no public classes.

@SparkQA
Copy link

SparkQA commented Mar 16, 2018

Test build #88299 has finished for PR 20659 at commit b92918b.

  • This patch passes all tests.
  • This patch merges cleanly.
  • This patch adds no public classes.

@SparkQA
Copy link

SparkQA commented Mar 16, 2018

Test build #88310 has finished for PR 20659 at commit 71bcb25.

  • This patch fails build dependency tests.
  • This patch merges cleanly.
  • This patch adds no public classes.

@SparkQA
Copy link

SparkQA commented Mar 16, 2018

Test build #88311 has finished for PR 20659 at commit 75e4cc9.

  • This patch fails Spark unit tests.
  • This patch merges cleanly.
  • This patch adds no public classes.

@wangyum
Copy link
Member Author

wangyum commented Mar 16, 2018

retest this please

@SparkQA
Copy link

SparkQA commented Mar 16, 2018

Test build #88317 has finished for PR 20659 at commit 75e4cc9.

  • This patch passes all tests.
  • This patch merges cleanly.
  • This patch adds no public classes.

@wangyum
Copy link
Member Author

wangyum commented Mar 16, 2018

Thanks everyone, we can move to SPARK-23710 to discuss.

@dongjoon-hyun
Copy link
Member

dongjoon-hyun commented Mar 17, 2018

@wangyum .
Please read again my previous comment.

Although I support your efforts, Spark must provide ORC data source without -Phive.


package org.apache.spark.sql.execution.datasources.orc

import org.apache.hadoop.hive.serde2.io.{DateWritable, HiveDecimalWritable}
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Please revert this.

import org.apache.orc.storage.serde2.io.HiveDecimalWritable
import org.apache.hadoop.hive.ql.io.sarg.{PredicateLeaf, SearchArgument, SearchArgumentFactory}
import org.apache.hadoop.hive.ql.io.sarg.SearchArgument.Builder
import org.apache.hadoop.hive.serde2.io.HiveDecimalWritable
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Please revert this.

import org.apache.orc.TypeDescription
import org.apache.orc.mapred.{OrcList, OrcMap, OrcStruct, OrcTimestamp}
import org.apache.orc.storage.common.`type`.HiveDecimal
import org.apache.orc.storage.serde2.io.{DateWritable, HiveDecimalWritable}
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Please revert this.

import scala.collection.JavaConverters._

import org.apache.orc.storage.ql.io.sarg.{PredicateLeaf, SearchArgument}
import org.apache.hadoop.hive.ql.io.sarg.{PredicateLeaf, SearchArgument}
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Please revert this.

@dongjoon-hyun
Copy link
Member

I added more specific comments about ORC parts, @wangyum .

@SparkQA
Copy link

SparkQA commented Mar 17, 2018

Test build #88337 has finished for PR 20659 at commit e4c0624.

  • This patch fails due to an unknown error code, -9.
  • This patch merges cleanly.
  • This patch adds no public classes.

@wangyum
Copy link
Member Author

wangyum commented Mar 17, 2018

retest this please

@SparkQA
Copy link

SparkQA commented Mar 17, 2018

Test build #88338 has finished for PR 20659 at commit e4c0624.

  • This patch passes all tests.
  • This patch merges cleanly.
  • This patch adds no public classes.

@Bidek56
Copy link
Contributor

Bidek56 commented Mar 28, 2018

What will take to merge this PR in? It has passed all the tests.

@srowen
Copy link
Member

srowen commented Apr 23, 2018

@wangyum you can close this experiment?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

8 participants