-
Notifications
You must be signed in to change notification settings - Fork 2.5k
Closed
Labels
area:awsAWS ecosystem supportAWS ecosystem supportdependenciesDependency updatesDependency updatespriority:mediumModerate impact; usability gapsModerate impact; usability gaps
Description
See the Following exception in the logs after running deltastreamer. I believe this is shown when we have metrics enabled.
java.lang.NoSuchMethodError: 'int org.apache.hudi.avro.model.HoodieRollbackMetadata.getTotalFilesDeleted()'
A clear and concise description of what you expected to happen.
Environment Description
-
Hudi version : 0.12.1
-
Spark version : 3.3.1
-
Hive version :
-
Hadoop version : 3
-
Storage (HDFS/S3/GCS..) : S3
-
Running on Docker? (yes/no) : yes, kubernetes
Additional context
Add any other context about the problem here.
The following hudi jars are in the classpath
-hudi-aws-bundle.jar
- hudi-cli.jar
-hudi-hive-sync-bundle.jar
-hudi-spark-bundle.jar
-hudi-utilities-slim-bundle.jar
hudi-utilities-slim-bundle does show that class.
10851 Mon Oct 17 22:29:30 UTC 2022 org/apache/hudi/avro/model/HoodieRollbackMetadata.class
Stacktrace
22/12/07 16:27:35 DEBUG org.apache.spark.internal.Logging: Executor 1 on 10.22.164.85 lost, but reason not yet known.
22/12/07 16:27:35 DEBUG org.apache.spark.internal.Logging: Cleaning up TaskScheduler state for tasks [] on failed executor 1
22/12/07 16:27:35 ERROR org.apache.spark.internal.Logging: Uncaught exception in thread dispatcher-CoarseGrainedScheduler
org.apache.spark.SparkException: Could not find CoarseGrainedScheduler.
at org.apache.spark.rpc.netty.Dispatcher.postMessage(Dispatcher.scala:178)
at org.apache.spark.rpc.netty.Dispatcher.postOneWayMessage(Dispatcher.scala:150)
at org.apache.spark.rpc.netty.NettyRpcEnv.send(NettyRpcEnv.scala:193)
at org.apache.spark.rpc.netty.NettyRpcEndpointRef.send(NettyRpcEnv.scala:563)
at org.apache.spark.scheduler.cluster.CoarseGrainedSchedulerBackend.$anonfun$reviveOffers$1(CoarseGrainedSchedulerBackend.scala:630)
at org.apache.spark.util.Utils$.tryLogNonFatalError(Utils.scala:1484)
at org.apache.spark.scheduler.cluster.CoarseGrainedSchedulerBackend.reviveOffers(CoarseGrainedSchedulerBackend.scala:630)
at org.apache.spark.scheduler.TaskSchedulerImpl.executorLost(TaskSchedulerImpl.scala:1004)
at org.apache.spark.scheduler.cluster.CoarseGrainedSchedulerBackend$DriverEndpoint.disableExecutor(CoarseGrainedSchedulerBackend.scala:482)
at org.apache.spark.scheduler.cluster.k8s.KubernetesClusterSchedulerBackend$KubernetesDriverEndpoint.$anonfun$onDisconnected$1(KubernetesClusterSchedulerBackend.scala:328)
at org.apache.spark.scheduler.cluster.k8s.KubernetesClusterSchedulerBackend$KubernetesDriverEndpoint.$anonfun$onDisconnected$1$adapted(KubernetesClusterSchedulerBackend.scala:328)
at scala.Option.foreach(Option.scala:407)
at org.apache.spark.scheduler.cluster.k8s.KubernetesClusterSchedulerBackend$KubernetesDriverEndpoint.onDisconnected(KubernetesClusterSchedulerBackend.scala:328)
at org.apache.spark.rpc.netty.Inbox.$anonfun$process$1(Inbox.scala:141)
at org.apache.spark.rpc.netty.Inbox.safelyCall(Inbox.scala:213)
at org.apache.spark.rpc.netty.Inbox.process(Inbox.scala:100)
at org.apache.spark.rpc.netty.MessageLoop.org$apache$spark$rpc$netty$MessageLoop$$receiveLoop(MessageLoop.scala:75)
at org.apache.spark.rpc.netty.MessageLoop$$anon$1.run(MessageLoop.scala:41)
at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)
at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
at java.base/java.lang.Thread.run(Thread.java:829)
22/12/07 16:27:35 WARN org.apache.spark.internal.Logging: Kubernetes client has been closed.
22/12/07 16:27:36 INFO org.apache.spark.internal.Logging: MapOutputTrackerMasterEndpoint stopped!
22/12/07 16:27:36 INFO org.apache.spark.internal.Logging: MemoryStore cleared
22/12/07 16:27:36 INFO org.apache.spark.internal.Logging: BlockManager stopped
22/12/07 16:27:36 INFO org.apache.spark.internal.Logging: BlockManagerMaster stopped
22/12/07 16:27:36 INFO org.apache.spark.internal.Logging: OutputCommitCoordinator stopped!
22/12/07 16:27:36 INFO org.apache.spark.internal.Logging: Successfully stopped SparkContext
Exception in thread "main" java.lang.NoSuchMethodError: 'int org.apache.hudi.avro.model.HoodieRollbackMetadata.getTotalFilesDeleted()'
at org.apache.hudi.client.BaseHoodieWriteClient.rollback(BaseHoodieWriteClient.java:774)
at org.apache.hudi.client.BaseHoodieWriteClient.rollbackFailedWrites(BaseHoodieWriteClient.java:1194)
at org.apache.hudi.client.BaseHoodieWriteClient.rollbackFailedWrites(BaseHoodieWriteClient.java:1177)
at org.apache.hudi.client.BaseHoodieWriteClient.rollbackFailedWrites(BaseHoodieWriteClient.java:1165)
at org.apache.hudi.client.BaseHoodieWriteClient.lambda$startCommitWithTime$97cdbdca$1(BaseHoodieWriteClient.java:966)
at org.apache.hudi.common.util.CleanerUtils.rollbackFailedWrites(CleanerUtils.java:153)
at org.apache.hudi.client.BaseHoodieWriteClient.startCommitWithTime(BaseHoodieWriteClient.java:965)
at org.apache.hudi.client.BaseHoodieWriteClient.startCommitWithTime(BaseHoodieWriteClient.java:958)
at org.apache.hudi.utilities.deltastreamer.DeltaSync.startCommit(DeltaSync.java:677)
at org.apache.hudi.utilities.deltastreamer.DeltaSync.writeToSink(DeltaSync.java:581)
at org.apache.hudi.utilities.deltastreamer.DeltaSync.syncOnce(DeltaSync.java:336)
at org.apache.hudi.utilities.deltastreamer.HoodieDeltaStreamer.lambda$sync$2(HoodieDeltaStreamer.java:204)
at org.apache.hudi.common.util.Option.ifPresent(Option.java:97)
at org.apache.hudi.utilities.deltastreamer.HoodieDeltaStreamer.sync(HoodieDeltaStreamer.java:202)
at org.apache.hudi.utilities.deltastreamer.HoodieDeltaStreamer.main(HoodieDeltaStreamer.java:571)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:566)
at org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52)
at org.apache.spark.deploy.SparkSubmit.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:958)
at org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:180)
at org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:203)
at org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:90)
at org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:1046)
at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:1055)
at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
22/12/07 16:27:36 INFO org.apache.spark.internal.Logging: Shutdown hook called
22/12/07 16:27:36 INFO org.apache.spark.internal.Logging: Deleting directory /tmp/spark-7a432e99-1f82-46a4-a555-10bd6090a2ba
22/12/07 16:27:36 INFO org.apache.spark.internal.Logging: Deleting directory /tmp/spark-42983bb3-9835-4bf4-9ea5-c68004077e7c
22/12/07 16:27:36 INFO org.apache.spark.internal.Logging: Deleting directory /tmp/spark-6098afd8-d431-4b91-a81e-3ebea9419630
22/12/07 16:27:36 INFO org.apache.spark.internal.Logging: Deleting directory /tmp/spark-f394dce8-567a-461d-b559-c953966adb93
22/12/07 16:27:36 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Stopping s3a-file-system metrics system...
22/12/07 16:27:36 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: s3a-file-system metrics system stopped.
22/12/07 16:27:36 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: s3a-file-system metrics system shutdown complete.```
Metadata
Metadata
Assignees
Labels
area:awsAWS ecosystem supportAWS ecosystem supportdependenciesDependency updatesDependency updatespriority:mediumModerate impact; usability gapsModerate impact; usability gaps
Type
Projects
Status
✅ Done